var/home/core/zuul-output/0000755000175000017500000000000015071524113014524 5ustar corecorevar/home/core/zuul-output/logs/0000755000175000017500000000000015071551626015501 5ustar corecorevar/home/core/zuul-output/logs/kubelet.log0000644000000000000000006455507115071551616017721 0ustar rootrootOct 08 18:10:53 crc systemd[1]: Starting Kubernetes Kubelet... Oct 08 18:10:53 crc restorecon[4721]: Relabeled /var/lib/kubelet/config.json from system_u:object_r:unlabeled_t:s0 to system_u:object_r:container_var_lib_t:s0 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/device-plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/device-plugins/kubelet.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/volumes/kubernetes.io~configmap/nginx-conf/..2025_02_23_05_40_35.4114275528/nginx.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/22e96971 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/21c98286 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/0f1869e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/46889d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/5b6a5969 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/6c7921f5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4804f443 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/2a46b283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/a6b5573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4f88ee5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/5a4eee4b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/cd87c521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/38602af4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/1483b002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/0346718b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/d3ed4ada not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/3bb473a5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/8cd075a9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/00ab4760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/54a21c09 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/70478888 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/43802770 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/955a0edc not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/bca2d009 not reset as customized by admin to system_u:object_r:container_file_t:s0:c140,c1009 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/b295f9bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/bc46ea27 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5731fc1b not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5e1b2a3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/943f0936 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/3f764ee4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/8695e3f9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/aed7aa86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/c64d7448 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/0ba16bd2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/207a939f not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/54aa8cdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/1f5fa595 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/bf9c8153 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/47fba4ea not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/7ae55ce9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7906a268 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/ce43fa69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7fc7ea3a not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/d8c38b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/9ef015fb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/b9db6a41 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/b1733d79 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/afccd338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/9df0a185 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/18938cf8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/7ab4eb23 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/56930be6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_35.630010865 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/0d8e3722 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/d22b2e76 not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/e036759f not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/2734c483 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/57878fe7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/3f3c2e58 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/375bec3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/7bc41e08 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/48c7a72d not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/4b66701f not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/a5a1c202 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_40.1388695756 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/26f3df5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/6d8fb21d not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/50e94777 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208473b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/ec9e08ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3b787c39 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208eaed5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/93aa3a2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3c697968 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/ba950ec9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/cb5cdb37 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/f2df9827 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/fedaa673 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/9ca2df95 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/b2d7460e not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2207853c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/241c1c29 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2d910eaf not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/c6c0f2e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/399edc97 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8049f7cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/0cec5484 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/312446d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c406,c828 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8e56a35d not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/2d30ddb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/eca8053d not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/c3a25c9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c168,c522 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/b9609c22 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/e8b0eca9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/b36a9c3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/38af7b07 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/ae821620 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/baa23338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/2c534809 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/59b29eae not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/c91a8e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/4d87494a not reset as customized by admin to system_u:object_r:container_file_t:s0:c442,c857 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/1e33ca63 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/8dea7be2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d0b04a99 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d84f01e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/4109059b not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/a7258a3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/05bdf2b6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/f3261b51 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/315d045e not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/5fdcf278 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/d053f757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/c2850dc7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fcfb0b2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c7ac9b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fa0c0d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c609b6ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/2be6c296 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/89a32653 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/4eb9afeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/13af6efa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/b03f9724 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/e3d105cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/3aed4d83 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/0765fa6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/2cefc627 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/3dcc6345 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/365af391 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b1130c0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/236a5913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b9432e26 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/5ddb0e3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/986dc4fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/8a23ff9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/9728ae68 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/665f31d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/136c9b42 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/98a1575b not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/cac69136 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/5deb77a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/2ae53400 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/e46f2326 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/dc688d3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/3497c3cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/177eb008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/af5a2afa not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/d780cb1f not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/49b0f374 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/26fbb125 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/cf14125a not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/b7f86972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/e51d739c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/88ba6a69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/669a9acf not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/5cd51231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/75349ec7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/15c26839 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/45023dcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/2bb66a50 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/64d03bdd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/ab8e7ca0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/bb9be25f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/9a0b61d3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/d471b9d2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/8cb76b8e not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/11a00840 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/ec355a92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/992f735e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d59cdbbc not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/72133ff0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/c56c834c not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d13724c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/0a498258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa471982 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fc900d92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa7d68da not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/4bacf9b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/424021b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/fc2e31a3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/f51eefac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/c8997f2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/7481f599 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/fdafea19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/d0e1c571 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/ee398915 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/682bb6b8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a3e67855 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a989f289 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/915431bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/7796fdab not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/dcdb5f19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/a3aaa88c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/5508e3e6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/160585de not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/e99f8da3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/8bc85570 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/a5861c91 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/84db1135 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/9e1a6043 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/c1aba1c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/d55ccd6d not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/971cc9f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/8f2e3dcf not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/ceb35e9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/1c192745 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/5209e501 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/f83de4df not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/e7b978ac not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/c64304a1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/5384386b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/cce3e3ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/8fb75465 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/740f573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/32fd1134 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/0a861bd3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/80363026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/bfa952a8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..2025_02_23_05_33_31.333075221 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/793bf43d not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/7db1bb6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/4f6a0368 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/c12c7d86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/36c4a773 not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/4c1e98ae not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/a4c8115c not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/setup/7db1802e not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver/a008a7ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-syncer/2c836bac not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-regeneration-controller/0ce62299 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-insecure-readyz/945d2457 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-check-endpoints/7d5c1dd8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/index.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/bundle-v1.15.0.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/channel.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/package.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/bc8d0691 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/6b76097a not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/34d1af30 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/312ba61c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/645d5dd1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/16e825f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/4cf51fc9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/2a23d348 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/075dbd49 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/dd585ddd not reset as customized by admin to system_u:object_r:container_file_t:s0:c377,c642 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/17ebd0ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c343 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/005579f4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_23_11.1287037894 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/bf5f3b9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/af276eb7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/ea28e322 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/692e6683 not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/871746a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/4eb2e958 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/ca9b62da not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/0edd6fce not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/containers/controller-manager/89b4555f not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/655fcd71 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/0d43c002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/e68efd17 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/9acf9b65 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/5ae3ff11 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/1e59206a not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/27af16d1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c304,c1017 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/7918e729 not reset as customized by admin to system_u:object_r:container_file_t:s0:c853,c893 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/5d976d0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c585,c981 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/d7f55cbb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/f0812073 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/1a56cbeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/7fdd437e not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/cdfb5652 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/fix-audit-permissions/fb93119e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver/f1e8fc0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver-check-endpoints/218511f3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server/serving-certs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/ca8af7b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/72cc8a75 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/6e8a3760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4c3455c0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/2278acb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4b453e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/3ec09bda not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2/cacerts.bin not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java/cacerts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl/ca-bundle.trust.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/email-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/objsign-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:53 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2ae6433e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fde84897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75680d2e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/openshift-service-serving-signer_1740288168.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/facfc4fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f5a969c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CFCA_EV_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9ef4a08a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ingress-operator_1740288202.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2f332aed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/248c8271.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d10a21f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ACCVRAIZ1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a94d09e5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c9a4d3b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40193066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd8c0d63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b936d1c6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CA_Disig_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4fd49c6c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM_SERVIDORES_SEGUROS.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b81b93f0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f9a69fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b30d5fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ANF_Secure_Server_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b433981b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93851c9e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9282e51c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7dd1bc4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Actalis_Authentication_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/930ac5d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f47b495.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e113c810.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5931b5bc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Commercial.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2b349938.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e48193cf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/302904dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a716d4ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Networking.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93bc0acc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/86212b19.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b727005e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbc54cab.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f51bb24c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c28a8a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9c8dfbd4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ccc52f49.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cb1c3204.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ce5e74ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd08c599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6d41d539.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb5fa911.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e35234b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8cb5ee0f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a7c655d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f8fc53da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/de6d66f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d41b5e2a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/41a3f684.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1df5a75f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_2011.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e36a6752.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b872f2b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9576d26b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/228f89db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_ECC_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb717492.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d21b73c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b1b94ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/595e996b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_RSA_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b46e03d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/128f4b91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_3_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81f2d2b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Autoridad_de_Certificacion_Firmaprofesional_CIF_A62634068.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3bde41ac.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d16a5865.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_EC-384_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0179095f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ffa7f1eb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9482e63a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4dae3dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e359ba6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7e067d03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/95aff9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7746a63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Baltimore_CyberTrust_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/653b494a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3ad48a91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_2_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/54657681.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/82223c44.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8de2f56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d9dafe4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d96b65e2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee64a828.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40547a79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5a3f0ff8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a780d93.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/34d996fb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/eed8c118.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/89c02a45.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b1159c4c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d6325660.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4c339cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8312c4c1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_E1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8508e720.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5fdd185d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48bec511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/69105f4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b9bc432.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/32888f65.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b03dec0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/219d9499.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5acf816d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbf06781.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc99f41e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AAA_Certificate_Services.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/985c1f52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8794b4e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_BR_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7c037b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ef954a4e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_EV_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2add47b6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/90c5a3c8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0f3e76e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/53a1b57a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_EV_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5ad8a5d6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/68dd7389.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d04f354.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d6437c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/062cdee6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bd43e1dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7f3d5d1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c491639e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3513523f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/399e7759.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/feffd413.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d18e9066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/607986c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c90bc37d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1b0f7e5c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e08bfd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dd8e9d41.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed39abd0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a3418fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bc3f2570.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_High_Assurance_EV_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/244b5494.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81b9768f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4be590e0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_ECC_P384_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9846683b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/252252d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e8e7201.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_RSA4096_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d52c538d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c44cc0c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Trusted_Root_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75d1b2ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a2c66da8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ecccd8db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust.net_Certification_Authority__2048_.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/aee5f10d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e7271e8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0e59380.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4c3982f2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b99d060.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf64f35b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0a775a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/002c0b4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cc450945.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_EC1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/106f3e4d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b3fb433b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4042bcee.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/02265526.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/455f1b52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0d69c7e1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9f727ac7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5e98733a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0cd152c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc4d6a89.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6187b673.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/FIRMAPROFESIONAL_CA_ROOT-A_WEB.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ba8887ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/068570d1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f081611a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48a195d8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GDCA_TrustAUTH_R5_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f6fa695.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab59055e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b92fd57f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GLOBALTRUST_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fa5da96b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ec40989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7719f463.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1001acf7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f013ecaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/626dceaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c559d742.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1d3472b9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9479c8c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a81e292b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4bfab552.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e071171e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/57bcb2da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_ECC_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab5346f4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5046c355.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_RSA_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/865fbdf9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da0cfd1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/85cde254.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_ECC_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbb3f32b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureSign_RootCA11.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5860aaa6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/31188b5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HiPKI_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c7f1359b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f15c80c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hongkong_Post_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/09789157.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/18856ac4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e09d511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Commercial_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cf701eeb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d06393bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Public_Sector_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/10531352.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Izenpe.com.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureTrust_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0ed035a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsec_e-Szigno_Root_CA_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8160b96c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8651083.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2c63f966.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_ECC_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d89cda1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/01419da9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_RSA_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7a5b843.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_RSA_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf53fb88.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9591a472.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3afde786.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Gold_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NAVER_Global_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3fb36b73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d39b0a2c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a89d74c2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd58d51e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7db1890.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NetLock_Arany__Class_Gold__F__tan__s__tv__ny.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/988a38cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/60afe812.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f39fc864.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5443e9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GB_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e73d606e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dfc0fe80.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b66938e9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e1eab7c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GC_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/773e07ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c899c73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d59297b8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ddcda989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_1_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/749e9e03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/52b525c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7e8dc79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a819ef2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/08063a00.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b483515.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/064e0aa9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1f58a078.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6f7454b3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7fa05551.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76faf6c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9339512a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f387163d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee37c333.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e18bfb83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e442e424.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fe8a2cd8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/23f4c490.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5cd81ad7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0c70a8d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7892ad52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SZAFIR_ROOT_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4f316efb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_RSA_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/06dc52d5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/583d0756.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0bf05006.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/88950faa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9046744a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c860d51.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_RSA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6fa5da56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/33ee480d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Secure_Global_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/63a2c897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_ECC_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bdacca6f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ff34af3f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbff3a01.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_ECC_RootCA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_C1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/406c9bb1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_C3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Services_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Silver_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/99e1b953.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/14bc7599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TUBITAK_Kamu_SM_SSL_Kok_Sertifikasi_-_Surum_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a3adc42.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f459871d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_ECC_Root_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_RSA_Root_2023.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TeliaSonera_Root_CA_v1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telia_Root_CA_v2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f103249.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f058632f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-certificates.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9bf03295.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/98aaf404.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1cef98f5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/073bfcc5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2923b3f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f249de83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/edcbddb5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P256_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b5697b0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ae85e5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b74d2bd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P384_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d887a5bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9aef356c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TunTrust_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd64f3fc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e13665f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Extended_Validation_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f5dc4f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da7377f6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Global_G2_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c01eb047.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/304d27c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed858448.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f30dd6ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/04f60c28.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_ECC_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fc5a8f99.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/35105088.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee532fd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/XRamp_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/706f604c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76579174.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d86cdd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/882de061.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f618aec.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a9d40e02.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e-Szigno_Root_CA_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e868b802.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/83e9984f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ePKI_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca6e4ad9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d6523ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4b718d9b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/869fbf79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/containers/registry/f8d22bdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/6e8bbfac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/54dd7996 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/a4f1bb05 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/207129da not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/c1df39e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/15b8f1cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/77bd6913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/2382c1b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/704ce128 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/70d16fe0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/bfb95535 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/57a8e8e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/1b9d3e5e not reset as customized by admin to system_u:object_r:container_file_t:s0:c107,c917 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/fddb173c not reset as customized by admin to system_u:object_r:container_file_t:s0:c202,c983 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/95d3c6c4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/bfb5fff5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/2aef40aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/c0391cad not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/1119e69d not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/660608b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/8220bd53 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/85f99d5c not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/4b0225f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/9c2a3394 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/e820b243 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/1ca52ea0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/e6988e45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/6655f00b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/98bc3986 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/08e3458a not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/2a191cb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/6c4eeefb not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/f61a549c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/24891863 not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/fbdfd89c not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/9b63b3bc not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/8acde6d6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/node-driver-registrar/59ecbba3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/csi-provisioner/685d4be3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/containers/route-controller-manager/feaea55e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/63709497 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/d966b7fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/f5773757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/81c9edb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/57bf57ee not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/86f5e6aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/0aabe31d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/d2af85c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/09d157d9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c0fe7256 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c30319e4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/e6b1dd45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/2bb643f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/920de426 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/70fa1e87 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/a1c12a2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/9442e6c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/5b45ec72 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/3c9f3a59 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/1091c11b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/9a6821c6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/ec0c35e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/517f37e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/6214fe78 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/ba189c8b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/351e4f31 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/c0f219ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/8069f607 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/559c3d82 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/605ad488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/148df488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/3bf6dcb4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/022a2feb not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/938c3924 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/729fe23e not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/1fd5cbd4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/a96697e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/e155ddca not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/10dd0e0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/6f2c8392 not reset as customized by admin to system_u:object_r:container_file_t:s0:c267,c588 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/bd241ad9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/plugins/csi-hostpath not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/plugins/csi-hostpath/csi.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/plugins/kubernetes.io not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/plugins/kubernetes.io/csi not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983 not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/vol_data.json not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 08 18:10:54 crc restorecon[4721]: /var/lib/kubelet/plugins_registry not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 08 18:10:54 crc restorecon[4721]: Relabeled /var/usrlocal/bin/kubenswrapper from system_u:object_r:bin_t:s0 to system_u:object_r:kubelet_exec_t:s0 Oct 08 18:10:54 crc kubenswrapper[4988]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Oct 08 18:10:54 crc kubenswrapper[4988]: Flag --minimum-container-ttl-duration has been deprecated, Use --eviction-hard or --eviction-soft instead. Will be removed in a future version. Oct 08 18:10:54 crc kubenswrapper[4988]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Oct 08 18:10:54 crc kubenswrapper[4988]: Flag --register-with-taints has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Oct 08 18:10:54 crc kubenswrapper[4988]: Flag --pod-infra-container-image has been deprecated, will be removed in a future release. Image garbage collector will get sandbox image information from CRI. Oct 08 18:10:54 crc kubenswrapper[4988]: Flag --system-reserved has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Oct 08 18:10:54 crc kubenswrapper[4988]: I1008 18:10:54.984861 4988 server.go:211] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Oct 08 18:10:54 crc kubenswrapper[4988]: W1008 18:10:54.991003 4988 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Oct 08 18:10:54 crc kubenswrapper[4988]: W1008 18:10:54.991029 4988 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Oct 08 18:10:54 crc kubenswrapper[4988]: W1008 18:10:54.991037 4988 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Oct 08 18:10:54 crc kubenswrapper[4988]: W1008 18:10:54.991044 4988 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Oct 08 18:10:54 crc kubenswrapper[4988]: W1008 18:10:54.991052 4988 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Oct 08 18:10:54 crc kubenswrapper[4988]: W1008 18:10:54.991059 4988 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Oct 08 18:10:54 crc kubenswrapper[4988]: W1008 18:10:54.991065 4988 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Oct 08 18:10:54 crc kubenswrapper[4988]: W1008 18:10:54.991071 4988 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Oct 08 18:10:54 crc kubenswrapper[4988]: W1008 18:10:54.991078 4988 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Oct 08 18:10:54 crc kubenswrapper[4988]: W1008 18:10:54.991084 4988 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Oct 08 18:10:54 crc kubenswrapper[4988]: W1008 18:10:54.991091 4988 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Oct 08 18:10:54 crc kubenswrapper[4988]: W1008 18:10:54.991097 4988 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Oct 08 18:10:54 crc kubenswrapper[4988]: W1008 18:10:54.991105 4988 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Oct 08 18:10:54 crc kubenswrapper[4988]: W1008 18:10:54.991115 4988 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Oct 08 18:10:54 crc kubenswrapper[4988]: W1008 18:10:54.991136 4988 feature_gate.go:330] unrecognized feature gate: GatewayAPI Oct 08 18:10:54 crc kubenswrapper[4988]: W1008 18:10:54.991144 4988 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Oct 08 18:10:54 crc kubenswrapper[4988]: W1008 18:10:54.991152 4988 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Oct 08 18:10:54 crc kubenswrapper[4988]: W1008 18:10:54.991159 4988 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Oct 08 18:10:54 crc kubenswrapper[4988]: W1008 18:10:54.991166 4988 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Oct 08 18:10:54 crc kubenswrapper[4988]: W1008 18:10:54.991173 4988 feature_gate.go:330] unrecognized feature gate: Example Oct 08 18:10:54 crc kubenswrapper[4988]: W1008 18:10:54.991180 4988 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Oct 08 18:10:54 crc kubenswrapper[4988]: W1008 18:10:54.991187 4988 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Oct 08 18:10:54 crc kubenswrapper[4988]: W1008 18:10:54.991193 4988 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Oct 08 18:10:54 crc kubenswrapper[4988]: W1008 18:10:54.991199 4988 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Oct 08 18:10:54 crc kubenswrapper[4988]: W1008 18:10:54.991205 4988 feature_gate.go:330] unrecognized feature gate: InsightsConfig Oct 08 18:10:54 crc kubenswrapper[4988]: W1008 18:10:54.991211 4988 feature_gate.go:330] unrecognized feature gate: PlatformOperators Oct 08 18:10:54 crc kubenswrapper[4988]: W1008 18:10:54.991217 4988 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Oct 08 18:10:54 crc kubenswrapper[4988]: W1008 18:10:54.991223 4988 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Oct 08 18:10:54 crc kubenswrapper[4988]: W1008 18:10:54.991229 4988 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Oct 08 18:10:54 crc kubenswrapper[4988]: W1008 18:10:54.991236 4988 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Oct 08 18:10:54 crc kubenswrapper[4988]: W1008 18:10:54.991242 4988 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Oct 08 18:10:54 crc kubenswrapper[4988]: W1008 18:10:54.991248 4988 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Oct 08 18:10:54 crc kubenswrapper[4988]: W1008 18:10:54.991255 4988 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Oct 08 18:10:54 crc kubenswrapper[4988]: W1008 18:10:54.991261 4988 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Oct 08 18:10:54 crc kubenswrapper[4988]: W1008 18:10:54.991270 4988 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Oct 08 18:10:54 crc kubenswrapper[4988]: W1008 18:10:54.991277 4988 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Oct 08 18:10:54 crc kubenswrapper[4988]: W1008 18:10:54.991284 4988 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Oct 08 18:10:54 crc kubenswrapper[4988]: W1008 18:10:54.991291 4988 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Oct 08 18:10:54 crc kubenswrapper[4988]: W1008 18:10:54.991298 4988 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Oct 08 18:10:54 crc kubenswrapper[4988]: W1008 18:10:54.991307 4988 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Oct 08 18:10:54 crc kubenswrapper[4988]: W1008 18:10:54.991314 4988 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Oct 08 18:10:54 crc kubenswrapper[4988]: W1008 18:10:54.991321 4988 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Oct 08 18:10:54 crc kubenswrapper[4988]: W1008 18:10:54.991327 4988 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Oct 08 18:10:54 crc kubenswrapper[4988]: W1008 18:10:54.991334 4988 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Oct 08 18:10:54 crc kubenswrapper[4988]: W1008 18:10:54.991340 4988 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Oct 08 18:10:54 crc kubenswrapper[4988]: W1008 18:10:54.991347 4988 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Oct 08 18:10:54 crc kubenswrapper[4988]: W1008 18:10:54.991353 4988 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Oct 08 18:10:54 crc kubenswrapper[4988]: W1008 18:10:54.991359 4988 feature_gate.go:330] unrecognized feature gate: SignatureStores Oct 08 18:10:54 crc kubenswrapper[4988]: W1008 18:10:54.991365 4988 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Oct 08 18:10:54 crc kubenswrapper[4988]: W1008 18:10:54.991372 4988 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Oct 08 18:10:54 crc kubenswrapper[4988]: W1008 18:10:54.991378 4988 feature_gate.go:330] unrecognized feature gate: PinnedImages Oct 08 18:10:54 crc kubenswrapper[4988]: W1008 18:10:54.991408 4988 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Oct 08 18:10:54 crc kubenswrapper[4988]: W1008 18:10:54.991416 4988 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Oct 08 18:10:54 crc kubenswrapper[4988]: W1008 18:10:54.991424 4988 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Oct 08 18:10:54 crc kubenswrapper[4988]: W1008 18:10:54.991433 4988 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Oct 08 18:10:54 crc kubenswrapper[4988]: W1008 18:10:54.991441 4988 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Oct 08 18:10:54 crc kubenswrapper[4988]: W1008 18:10:54.991448 4988 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Oct 08 18:10:54 crc kubenswrapper[4988]: W1008 18:10:54.991455 4988 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Oct 08 18:10:54 crc kubenswrapper[4988]: W1008 18:10:54.991461 4988 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Oct 08 18:10:54 crc kubenswrapper[4988]: W1008 18:10:54.991467 4988 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Oct 08 18:10:54 crc kubenswrapper[4988]: W1008 18:10:54.991474 4988 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Oct 08 18:10:54 crc kubenswrapper[4988]: W1008 18:10:54.991480 4988 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Oct 08 18:10:54 crc kubenswrapper[4988]: W1008 18:10:54.991486 4988 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Oct 08 18:10:54 crc kubenswrapper[4988]: W1008 18:10:54.991492 4988 feature_gate.go:330] unrecognized feature gate: OVNObservability Oct 08 18:10:54 crc kubenswrapper[4988]: W1008 18:10:54.991498 4988 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Oct 08 18:10:54 crc kubenswrapper[4988]: W1008 18:10:54.991504 4988 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Oct 08 18:10:54 crc kubenswrapper[4988]: W1008 18:10:54.991510 4988 feature_gate.go:330] unrecognized feature gate: NewOLM Oct 08 18:10:54 crc kubenswrapper[4988]: W1008 18:10:54.991517 4988 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Oct 08 18:10:54 crc kubenswrapper[4988]: W1008 18:10:54.991525 4988 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Oct 08 18:10:54 crc kubenswrapper[4988]: W1008 18:10:54.991531 4988 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Oct 08 18:10:54 crc kubenswrapper[4988]: W1008 18:10:54.991537 4988 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Oct 08 18:10:54 crc kubenswrapper[4988]: I1008 18:10:54.991677 4988 flags.go:64] FLAG: --address="0.0.0.0" Oct 08 18:10:54 crc kubenswrapper[4988]: I1008 18:10:54.991693 4988 flags.go:64] FLAG: --allowed-unsafe-sysctls="[]" Oct 08 18:10:54 crc kubenswrapper[4988]: I1008 18:10:54.991706 4988 flags.go:64] FLAG: --anonymous-auth="true" Oct 08 18:10:54 crc kubenswrapper[4988]: I1008 18:10:54.991715 4988 flags.go:64] FLAG: --application-metrics-count-limit="100" Oct 08 18:10:54 crc kubenswrapper[4988]: I1008 18:10:54.991725 4988 flags.go:64] FLAG: --authentication-token-webhook="false" Oct 08 18:10:54 crc kubenswrapper[4988]: I1008 18:10:54.991734 4988 flags.go:64] FLAG: --authentication-token-webhook-cache-ttl="2m0s" Oct 08 18:10:54 crc kubenswrapper[4988]: I1008 18:10:54.991745 4988 flags.go:64] FLAG: --authorization-mode="AlwaysAllow" Oct 08 18:10:54 crc kubenswrapper[4988]: I1008 18:10:54.991754 4988 flags.go:64] FLAG: --authorization-webhook-cache-authorized-ttl="5m0s" Oct 08 18:10:54 crc kubenswrapper[4988]: I1008 18:10:54.991761 4988 flags.go:64] FLAG: --authorization-webhook-cache-unauthorized-ttl="30s" Oct 08 18:10:54 crc kubenswrapper[4988]: I1008 18:10:54.991769 4988 flags.go:64] FLAG: --boot-id-file="/proc/sys/kernel/random/boot_id" Oct 08 18:10:54 crc kubenswrapper[4988]: I1008 18:10:54.991777 4988 flags.go:64] FLAG: --bootstrap-kubeconfig="/etc/kubernetes/kubeconfig" Oct 08 18:10:54 crc kubenswrapper[4988]: I1008 18:10:54.991786 4988 flags.go:64] FLAG: --cert-dir="/var/lib/kubelet/pki" Oct 08 18:10:54 crc kubenswrapper[4988]: I1008 18:10:54.991794 4988 flags.go:64] FLAG: --cgroup-driver="cgroupfs" Oct 08 18:10:54 crc kubenswrapper[4988]: I1008 18:10:54.991802 4988 flags.go:64] FLAG: --cgroup-root="" Oct 08 18:10:54 crc kubenswrapper[4988]: I1008 18:10:54.991812 4988 flags.go:64] FLAG: --cgroups-per-qos="true" Oct 08 18:10:54 crc kubenswrapper[4988]: I1008 18:10:54.991820 4988 flags.go:64] FLAG: --client-ca-file="" Oct 08 18:10:54 crc kubenswrapper[4988]: I1008 18:10:54.991827 4988 flags.go:64] FLAG: --cloud-config="" Oct 08 18:10:54 crc kubenswrapper[4988]: I1008 18:10:54.991835 4988 flags.go:64] FLAG: --cloud-provider="" Oct 08 18:10:54 crc kubenswrapper[4988]: I1008 18:10:54.991842 4988 flags.go:64] FLAG: --cluster-dns="[]" Oct 08 18:10:54 crc kubenswrapper[4988]: I1008 18:10:54.991855 4988 flags.go:64] FLAG: --cluster-domain="" Oct 08 18:10:54 crc kubenswrapper[4988]: I1008 18:10:54.991862 4988 flags.go:64] FLAG: --config="/etc/kubernetes/kubelet.conf" Oct 08 18:10:54 crc kubenswrapper[4988]: I1008 18:10:54.991870 4988 flags.go:64] FLAG: --config-dir="" Oct 08 18:10:54 crc kubenswrapper[4988]: I1008 18:10:54.991877 4988 flags.go:64] FLAG: --container-hints="/etc/cadvisor/container_hints.json" Oct 08 18:10:54 crc kubenswrapper[4988]: I1008 18:10:54.991886 4988 flags.go:64] FLAG: --container-log-max-files="5" Oct 08 18:10:54 crc kubenswrapper[4988]: I1008 18:10:54.991895 4988 flags.go:64] FLAG: --container-log-max-size="10Mi" Oct 08 18:10:54 crc kubenswrapper[4988]: I1008 18:10:54.991902 4988 flags.go:64] FLAG: --container-runtime-endpoint="/var/run/crio/crio.sock" Oct 08 18:10:54 crc kubenswrapper[4988]: I1008 18:10:54.991909 4988 flags.go:64] FLAG: --containerd="/run/containerd/containerd.sock" Oct 08 18:10:54 crc kubenswrapper[4988]: I1008 18:10:54.991917 4988 flags.go:64] FLAG: --containerd-namespace="k8s.io" Oct 08 18:10:54 crc kubenswrapper[4988]: I1008 18:10:54.991924 4988 flags.go:64] FLAG: --contention-profiling="false" Oct 08 18:10:54 crc kubenswrapper[4988]: I1008 18:10:54.991932 4988 flags.go:64] FLAG: --cpu-cfs-quota="true" Oct 08 18:10:54 crc kubenswrapper[4988]: I1008 18:10:54.991940 4988 flags.go:64] FLAG: --cpu-cfs-quota-period="100ms" Oct 08 18:10:54 crc kubenswrapper[4988]: I1008 18:10:54.991948 4988 flags.go:64] FLAG: --cpu-manager-policy="none" Oct 08 18:10:54 crc kubenswrapper[4988]: I1008 18:10:54.991958 4988 flags.go:64] FLAG: --cpu-manager-policy-options="" Oct 08 18:10:54 crc kubenswrapper[4988]: I1008 18:10:54.991968 4988 flags.go:64] FLAG: --cpu-manager-reconcile-period="10s" Oct 08 18:10:54 crc kubenswrapper[4988]: I1008 18:10:54.991976 4988 flags.go:64] FLAG: --enable-controller-attach-detach="true" Oct 08 18:10:54 crc kubenswrapper[4988]: I1008 18:10:54.991983 4988 flags.go:64] FLAG: --enable-debugging-handlers="true" Oct 08 18:10:54 crc kubenswrapper[4988]: I1008 18:10:54.991991 4988 flags.go:64] FLAG: --enable-load-reader="false" Oct 08 18:10:54 crc kubenswrapper[4988]: I1008 18:10:54.991998 4988 flags.go:64] FLAG: --enable-server="true" Oct 08 18:10:54 crc kubenswrapper[4988]: I1008 18:10:54.992006 4988 flags.go:64] FLAG: --enforce-node-allocatable="[pods]" Oct 08 18:10:54 crc kubenswrapper[4988]: I1008 18:10:54.992016 4988 flags.go:64] FLAG: --event-burst="100" Oct 08 18:10:54 crc kubenswrapper[4988]: I1008 18:10:54.992024 4988 flags.go:64] FLAG: --event-qps="50" Oct 08 18:10:54 crc kubenswrapper[4988]: I1008 18:10:54.992031 4988 flags.go:64] FLAG: --event-storage-age-limit="default=0" Oct 08 18:10:54 crc kubenswrapper[4988]: I1008 18:10:54.992038 4988 flags.go:64] FLAG: --event-storage-event-limit="default=0" Oct 08 18:10:54 crc kubenswrapper[4988]: I1008 18:10:54.992045 4988 flags.go:64] FLAG: --eviction-hard="" Oct 08 18:10:54 crc kubenswrapper[4988]: I1008 18:10:54.992054 4988 flags.go:64] FLAG: --eviction-max-pod-grace-period="0" Oct 08 18:10:54 crc kubenswrapper[4988]: I1008 18:10:54.992062 4988 flags.go:64] FLAG: --eviction-minimum-reclaim="" Oct 08 18:10:54 crc kubenswrapper[4988]: I1008 18:10:54.992070 4988 flags.go:64] FLAG: --eviction-pressure-transition-period="5m0s" Oct 08 18:10:54 crc kubenswrapper[4988]: I1008 18:10:54.992078 4988 flags.go:64] FLAG: --eviction-soft="" Oct 08 18:10:54 crc kubenswrapper[4988]: I1008 18:10:54.992085 4988 flags.go:64] FLAG: --eviction-soft-grace-period="" Oct 08 18:10:54 crc kubenswrapper[4988]: I1008 18:10:54.992093 4988 flags.go:64] FLAG: --exit-on-lock-contention="false" Oct 08 18:10:54 crc kubenswrapper[4988]: I1008 18:10:54.992100 4988 flags.go:64] FLAG: --experimental-allocatable-ignore-eviction="false" Oct 08 18:10:54 crc kubenswrapper[4988]: I1008 18:10:54.992108 4988 flags.go:64] FLAG: --experimental-mounter-path="" Oct 08 18:10:54 crc kubenswrapper[4988]: I1008 18:10:54.992116 4988 flags.go:64] FLAG: --fail-cgroupv1="false" Oct 08 18:10:54 crc kubenswrapper[4988]: I1008 18:10:54.992123 4988 flags.go:64] FLAG: --fail-swap-on="true" Oct 08 18:10:54 crc kubenswrapper[4988]: I1008 18:10:54.992130 4988 flags.go:64] FLAG: --feature-gates="" Oct 08 18:10:54 crc kubenswrapper[4988]: I1008 18:10:54.992140 4988 flags.go:64] FLAG: --file-check-frequency="20s" Oct 08 18:10:54 crc kubenswrapper[4988]: I1008 18:10:54.992148 4988 flags.go:64] FLAG: --global-housekeeping-interval="1m0s" Oct 08 18:10:54 crc kubenswrapper[4988]: I1008 18:10:54.992156 4988 flags.go:64] FLAG: --hairpin-mode="promiscuous-bridge" Oct 08 18:10:54 crc kubenswrapper[4988]: I1008 18:10:54.992164 4988 flags.go:64] FLAG: --healthz-bind-address="127.0.0.1" Oct 08 18:10:54 crc kubenswrapper[4988]: I1008 18:10:54.992171 4988 flags.go:64] FLAG: --healthz-port="10248" Oct 08 18:10:54 crc kubenswrapper[4988]: I1008 18:10:54.992179 4988 flags.go:64] FLAG: --help="false" Oct 08 18:10:54 crc kubenswrapper[4988]: I1008 18:10:54.992187 4988 flags.go:64] FLAG: --hostname-override="" Oct 08 18:10:54 crc kubenswrapper[4988]: I1008 18:10:54.992195 4988 flags.go:64] FLAG: --housekeeping-interval="10s" Oct 08 18:10:54 crc kubenswrapper[4988]: I1008 18:10:54.992202 4988 flags.go:64] FLAG: --http-check-frequency="20s" Oct 08 18:10:54 crc kubenswrapper[4988]: I1008 18:10:54.992210 4988 flags.go:64] FLAG: --image-credential-provider-bin-dir="" Oct 08 18:10:54 crc kubenswrapper[4988]: I1008 18:10:54.992218 4988 flags.go:64] FLAG: --image-credential-provider-config="" Oct 08 18:10:54 crc kubenswrapper[4988]: I1008 18:10:54.992225 4988 flags.go:64] FLAG: --image-gc-high-threshold="85" Oct 08 18:10:54 crc kubenswrapper[4988]: I1008 18:10:54.992233 4988 flags.go:64] FLAG: --image-gc-low-threshold="80" Oct 08 18:10:54 crc kubenswrapper[4988]: I1008 18:10:54.992242 4988 flags.go:64] FLAG: --image-service-endpoint="" Oct 08 18:10:54 crc kubenswrapper[4988]: I1008 18:10:54.992250 4988 flags.go:64] FLAG: --kernel-memcg-notification="false" Oct 08 18:10:54 crc kubenswrapper[4988]: I1008 18:10:54.992257 4988 flags.go:64] FLAG: --kube-api-burst="100" Oct 08 18:10:54 crc kubenswrapper[4988]: I1008 18:10:54.992266 4988 flags.go:64] FLAG: --kube-api-content-type="application/vnd.kubernetes.protobuf" Oct 08 18:10:54 crc kubenswrapper[4988]: I1008 18:10:54.992276 4988 flags.go:64] FLAG: --kube-api-qps="50" Oct 08 18:10:54 crc kubenswrapper[4988]: I1008 18:10:54.992283 4988 flags.go:64] FLAG: --kube-reserved="" Oct 08 18:10:54 crc kubenswrapper[4988]: I1008 18:10:54.992291 4988 flags.go:64] FLAG: --kube-reserved-cgroup="" Oct 08 18:10:54 crc kubenswrapper[4988]: I1008 18:10:54.992298 4988 flags.go:64] FLAG: --kubeconfig="/var/lib/kubelet/kubeconfig" Oct 08 18:10:54 crc kubenswrapper[4988]: I1008 18:10:54.992306 4988 flags.go:64] FLAG: --kubelet-cgroups="" Oct 08 18:10:54 crc kubenswrapper[4988]: I1008 18:10:54.992314 4988 flags.go:64] FLAG: --local-storage-capacity-isolation="true" Oct 08 18:10:54 crc kubenswrapper[4988]: I1008 18:10:54.992322 4988 flags.go:64] FLAG: --lock-file="" Oct 08 18:10:54 crc kubenswrapper[4988]: I1008 18:10:54.992329 4988 flags.go:64] FLAG: --log-cadvisor-usage="false" Oct 08 18:10:54 crc kubenswrapper[4988]: I1008 18:10:54.992337 4988 flags.go:64] FLAG: --log-flush-frequency="5s" Oct 08 18:10:54 crc kubenswrapper[4988]: I1008 18:10:54.992345 4988 flags.go:64] FLAG: --log-json-info-buffer-size="0" Oct 08 18:10:54 crc kubenswrapper[4988]: I1008 18:10:54.992357 4988 flags.go:64] FLAG: --log-json-split-stream="false" Oct 08 18:10:54 crc kubenswrapper[4988]: I1008 18:10:54.992365 4988 flags.go:64] FLAG: --log-text-info-buffer-size="0" Oct 08 18:10:54 crc kubenswrapper[4988]: I1008 18:10:54.992373 4988 flags.go:64] FLAG: --log-text-split-stream="false" Oct 08 18:10:54 crc kubenswrapper[4988]: I1008 18:10:54.992380 4988 flags.go:64] FLAG: --logging-format="text" Oct 08 18:10:54 crc kubenswrapper[4988]: I1008 18:10:54.992412 4988 flags.go:64] FLAG: --machine-id-file="/etc/machine-id,/var/lib/dbus/machine-id" Oct 08 18:10:54 crc kubenswrapper[4988]: I1008 18:10:54.992421 4988 flags.go:64] FLAG: --make-iptables-util-chains="true" Oct 08 18:10:54 crc kubenswrapper[4988]: I1008 18:10:54.992429 4988 flags.go:64] FLAG: --manifest-url="" Oct 08 18:10:54 crc kubenswrapper[4988]: I1008 18:10:54.992436 4988 flags.go:64] FLAG: --manifest-url-header="" Oct 08 18:10:54 crc kubenswrapper[4988]: I1008 18:10:54.992447 4988 flags.go:64] FLAG: --max-housekeeping-interval="15s" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:54.992454 4988 flags.go:64] FLAG: --max-open-files="1000000" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:54.992464 4988 flags.go:64] FLAG: --max-pods="110" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:54.992472 4988 flags.go:64] FLAG: --maximum-dead-containers="-1" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:54.992480 4988 flags.go:64] FLAG: --maximum-dead-containers-per-container="1" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:54.992487 4988 flags.go:64] FLAG: --memory-manager-policy="None" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:54.992494 4988 flags.go:64] FLAG: --minimum-container-ttl-duration="6m0s" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:54.992502 4988 flags.go:64] FLAG: --minimum-image-ttl-duration="2m0s" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:54.992509 4988 flags.go:64] FLAG: --node-ip="192.168.126.11" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:54.992517 4988 flags.go:64] FLAG: --node-labels="node-role.kubernetes.io/control-plane=,node-role.kubernetes.io/master=,node.openshift.io/os_id=rhcos" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:54.992535 4988 flags.go:64] FLAG: --node-status-max-images="50" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:54.992543 4988 flags.go:64] FLAG: --node-status-update-frequency="10s" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:54.992550 4988 flags.go:64] FLAG: --oom-score-adj="-999" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:54.992557 4988 flags.go:64] FLAG: --pod-cidr="" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:54.992567 4988 flags.go:64] FLAG: --pod-infra-container-image="quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:33549946e22a9ffa738fd94b1345f90921bc8f92fa6137784cb33c77ad806f9d" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:54.992579 4988 flags.go:64] FLAG: --pod-manifest-path="" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:54.992586 4988 flags.go:64] FLAG: --pod-max-pids="-1" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:54.992594 4988 flags.go:64] FLAG: --pods-per-core="0" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:54.992601 4988 flags.go:64] FLAG: --port="10250" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:54.992609 4988 flags.go:64] FLAG: --protect-kernel-defaults="false" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:54.992617 4988 flags.go:64] FLAG: --provider-id="" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:54.992625 4988 flags.go:64] FLAG: --qos-reserved="" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:54.992632 4988 flags.go:64] FLAG: --read-only-port="10255" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:54.992640 4988 flags.go:64] FLAG: --register-node="true" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:54.992647 4988 flags.go:64] FLAG: --register-schedulable="true" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:54.992654 4988 flags.go:64] FLAG: --register-with-taints="node-role.kubernetes.io/master=:NoSchedule" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:54.992666 4988 flags.go:64] FLAG: --registry-burst="10" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:54.992673 4988 flags.go:64] FLAG: --registry-qps="5" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:54.992680 4988 flags.go:64] FLAG: --reserved-cpus="" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:54.992688 4988 flags.go:64] FLAG: --reserved-memory="" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:54.992697 4988 flags.go:64] FLAG: --resolv-conf="/etc/resolv.conf" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:54.992705 4988 flags.go:64] FLAG: --root-dir="/var/lib/kubelet" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:54.992713 4988 flags.go:64] FLAG: --rotate-certificates="false" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:54.992720 4988 flags.go:64] FLAG: --rotate-server-certificates="false" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:54.992728 4988 flags.go:64] FLAG: --runonce="false" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:54.992735 4988 flags.go:64] FLAG: --runtime-cgroups="/system.slice/crio.service" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:54.992743 4988 flags.go:64] FLAG: --runtime-request-timeout="2m0s" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:54.992751 4988 flags.go:64] FLAG: --seccomp-default="false" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:54.992759 4988 flags.go:64] FLAG: --serialize-image-pulls="true" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:54.992768 4988 flags.go:64] FLAG: --storage-driver-buffer-duration="1m0s" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:54.992776 4988 flags.go:64] FLAG: --storage-driver-db="cadvisor" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:54.992784 4988 flags.go:64] FLAG: --storage-driver-host="localhost:8086" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:54.992792 4988 flags.go:64] FLAG: --storage-driver-password="root" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:54.992799 4988 flags.go:64] FLAG: --storage-driver-secure="false" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:54.992807 4988 flags.go:64] FLAG: --storage-driver-table="stats" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:54.992814 4988 flags.go:64] FLAG: --storage-driver-user="root" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:54.992821 4988 flags.go:64] FLAG: --streaming-connection-idle-timeout="4h0m0s" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:54.992829 4988 flags.go:64] FLAG: --sync-frequency="1m0s" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:54.992837 4988 flags.go:64] FLAG: --system-cgroups="" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:54.992845 4988 flags.go:64] FLAG: --system-reserved="cpu=200m,ephemeral-storage=350Mi,memory=350Mi" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:54.992869 4988 flags.go:64] FLAG: --system-reserved-cgroup="" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:54.992876 4988 flags.go:64] FLAG: --tls-cert-file="" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:54.992884 4988 flags.go:64] FLAG: --tls-cipher-suites="[]" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:54.992896 4988 flags.go:64] FLAG: --tls-min-version="" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:54.992903 4988 flags.go:64] FLAG: --tls-private-key-file="" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:54.992910 4988 flags.go:64] FLAG: --topology-manager-policy="none" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:54.992917 4988 flags.go:64] FLAG: --topology-manager-policy-options="" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:54.992925 4988 flags.go:64] FLAG: --topology-manager-scope="container" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:54.992933 4988 flags.go:64] FLAG: --v="2" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:54.992943 4988 flags.go:64] FLAG: --version="false" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:54.992953 4988 flags.go:64] FLAG: --vmodule="" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:54.992962 4988 flags.go:64] FLAG: --volume-plugin-dir="/etc/kubernetes/kubelet-plugins/volume/exec" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:54.992970 4988 flags.go:64] FLAG: --volume-stats-agg-period="1m0s" Oct 08 18:10:55 crc kubenswrapper[4988]: W1008 18:10:54.993178 4988 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Oct 08 18:10:55 crc kubenswrapper[4988]: W1008 18:10:54.993190 4988 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Oct 08 18:10:55 crc kubenswrapper[4988]: W1008 18:10:54.993198 4988 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Oct 08 18:10:55 crc kubenswrapper[4988]: W1008 18:10:54.993206 4988 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Oct 08 18:10:55 crc kubenswrapper[4988]: W1008 18:10:54.993213 4988 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Oct 08 18:10:55 crc kubenswrapper[4988]: W1008 18:10:54.993221 4988 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Oct 08 18:10:55 crc kubenswrapper[4988]: W1008 18:10:54.993229 4988 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Oct 08 18:10:55 crc kubenswrapper[4988]: W1008 18:10:54.993236 4988 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Oct 08 18:10:55 crc kubenswrapper[4988]: W1008 18:10:54.993243 4988 feature_gate.go:330] unrecognized feature gate: PinnedImages Oct 08 18:10:55 crc kubenswrapper[4988]: W1008 18:10:54.993250 4988 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Oct 08 18:10:55 crc kubenswrapper[4988]: W1008 18:10:54.993257 4988 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Oct 08 18:10:55 crc kubenswrapper[4988]: W1008 18:10:54.993263 4988 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Oct 08 18:10:55 crc kubenswrapper[4988]: W1008 18:10:54.993269 4988 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Oct 08 18:10:55 crc kubenswrapper[4988]: W1008 18:10:54.993276 4988 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Oct 08 18:10:55 crc kubenswrapper[4988]: W1008 18:10:54.993285 4988 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Oct 08 18:10:55 crc kubenswrapper[4988]: W1008 18:10:54.993293 4988 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Oct 08 18:10:55 crc kubenswrapper[4988]: W1008 18:10:54.993299 4988 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Oct 08 18:10:55 crc kubenswrapper[4988]: W1008 18:10:54.993306 4988 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Oct 08 18:10:55 crc kubenswrapper[4988]: W1008 18:10:54.993312 4988 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Oct 08 18:10:55 crc kubenswrapper[4988]: W1008 18:10:54.993319 4988 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Oct 08 18:10:55 crc kubenswrapper[4988]: W1008 18:10:54.993328 4988 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Oct 08 18:10:55 crc kubenswrapper[4988]: W1008 18:10:54.993339 4988 feature_gate.go:330] unrecognized feature gate: PlatformOperators Oct 08 18:10:55 crc kubenswrapper[4988]: W1008 18:10:54.993346 4988 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Oct 08 18:10:55 crc kubenswrapper[4988]: W1008 18:10:54.993354 4988 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Oct 08 18:10:55 crc kubenswrapper[4988]: W1008 18:10:54.993368 4988 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Oct 08 18:10:55 crc kubenswrapper[4988]: W1008 18:10:54.993375 4988 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Oct 08 18:10:55 crc kubenswrapper[4988]: W1008 18:10:54.993382 4988 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Oct 08 18:10:55 crc kubenswrapper[4988]: W1008 18:10:54.993416 4988 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Oct 08 18:10:55 crc kubenswrapper[4988]: W1008 18:10:54.993423 4988 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Oct 08 18:10:55 crc kubenswrapper[4988]: W1008 18:10:54.993429 4988 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Oct 08 18:10:55 crc kubenswrapper[4988]: W1008 18:10:54.993436 4988 feature_gate.go:330] unrecognized feature gate: InsightsConfig Oct 08 18:10:55 crc kubenswrapper[4988]: W1008 18:10:54.993442 4988 feature_gate.go:330] unrecognized feature gate: OVNObservability Oct 08 18:10:55 crc kubenswrapper[4988]: W1008 18:10:54.993448 4988 feature_gate.go:330] unrecognized feature gate: Example Oct 08 18:10:55 crc kubenswrapper[4988]: W1008 18:10:54.993454 4988 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Oct 08 18:10:55 crc kubenswrapper[4988]: W1008 18:10:54.993461 4988 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Oct 08 18:10:55 crc kubenswrapper[4988]: W1008 18:10:54.993467 4988 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Oct 08 18:10:55 crc kubenswrapper[4988]: W1008 18:10:54.993476 4988 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Oct 08 18:10:55 crc kubenswrapper[4988]: W1008 18:10:54.993483 4988 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Oct 08 18:10:55 crc kubenswrapper[4988]: W1008 18:10:54.993489 4988 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Oct 08 18:10:55 crc kubenswrapper[4988]: W1008 18:10:54.993495 4988 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Oct 08 18:10:55 crc kubenswrapper[4988]: W1008 18:10:54.993502 4988 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Oct 08 18:10:55 crc kubenswrapper[4988]: W1008 18:10:54.993509 4988 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Oct 08 18:10:55 crc kubenswrapper[4988]: W1008 18:10:54.993516 4988 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Oct 08 18:10:55 crc kubenswrapper[4988]: W1008 18:10:54.993522 4988 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Oct 08 18:10:55 crc kubenswrapper[4988]: W1008 18:10:54.993529 4988 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Oct 08 18:10:55 crc kubenswrapper[4988]: W1008 18:10:54.993535 4988 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Oct 08 18:10:55 crc kubenswrapper[4988]: W1008 18:10:54.993544 4988 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Oct 08 18:10:55 crc kubenswrapper[4988]: W1008 18:10:54.993551 4988 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Oct 08 18:10:55 crc kubenswrapper[4988]: W1008 18:10:54.993557 4988 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Oct 08 18:10:55 crc kubenswrapper[4988]: W1008 18:10:54.993564 4988 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Oct 08 18:10:55 crc kubenswrapper[4988]: W1008 18:10:54.993570 4988 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Oct 08 18:10:55 crc kubenswrapper[4988]: W1008 18:10:54.993577 4988 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Oct 08 18:10:55 crc kubenswrapper[4988]: W1008 18:10:54.993584 4988 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Oct 08 18:10:55 crc kubenswrapper[4988]: W1008 18:10:54.993594 4988 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Oct 08 18:10:55 crc kubenswrapper[4988]: W1008 18:10:54.993600 4988 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Oct 08 18:10:55 crc kubenswrapper[4988]: W1008 18:10:54.993607 4988 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Oct 08 18:10:55 crc kubenswrapper[4988]: W1008 18:10:54.993616 4988 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Oct 08 18:10:55 crc kubenswrapper[4988]: W1008 18:10:54.993623 4988 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Oct 08 18:10:55 crc kubenswrapper[4988]: W1008 18:10:54.993630 4988 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Oct 08 18:10:55 crc kubenswrapper[4988]: W1008 18:10:54.993637 4988 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Oct 08 18:10:55 crc kubenswrapper[4988]: W1008 18:10:54.993644 4988 feature_gate.go:330] unrecognized feature gate: GatewayAPI Oct 08 18:10:55 crc kubenswrapper[4988]: W1008 18:10:54.993651 4988 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Oct 08 18:10:55 crc kubenswrapper[4988]: W1008 18:10:54.993659 4988 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Oct 08 18:10:55 crc kubenswrapper[4988]: W1008 18:10:54.993666 4988 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Oct 08 18:10:55 crc kubenswrapper[4988]: W1008 18:10:54.993673 4988 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Oct 08 18:10:55 crc kubenswrapper[4988]: W1008 18:10:54.993680 4988 feature_gate.go:330] unrecognized feature gate: NewOLM Oct 08 18:10:55 crc kubenswrapper[4988]: W1008 18:10:54.993686 4988 feature_gate.go:330] unrecognized feature gate: SignatureStores Oct 08 18:10:55 crc kubenswrapper[4988]: W1008 18:10:54.993693 4988 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Oct 08 18:10:55 crc kubenswrapper[4988]: W1008 18:10:54.993699 4988 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Oct 08 18:10:55 crc kubenswrapper[4988]: W1008 18:10:54.993722 4988 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Oct 08 18:10:55 crc kubenswrapper[4988]: W1008 18:10:54.993729 4988 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:54.993750 4988 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:55.006133 4988 server.go:491] "Kubelet version" kubeletVersion="v1.31.5" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:55.006173 4988 server.go:493] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Oct 08 18:10:55 crc kubenswrapper[4988]: W1008 18:10:55.006269 4988 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Oct 08 18:10:55 crc kubenswrapper[4988]: W1008 18:10:55.006280 4988 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Oct 08 18:10:55 crc kubenswrapper[4988]: W1008 18:10:55.006291 4988 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Oct 08 18:10:55 crc kubenswrapper[4988]: W1008 18:10:55.006297 4988 feature_gate.go:330] unrecognized feature gate: SignatureStores Oct 08 18:10:55 crc kubenswrapper[4988]: W1008 18:10:55.006303 4988 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Oct 08 18:10:55 crc kubenswrapper[4988]: W1008 18:10:55.006308 4988 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Oct 08 18:10:55 crc kubenswrapper[4988]: W1008 18:10:55.006313 4988 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Oct 08 18:10:55 crc kubenswrapper[4988]: W1008 18:10:55.006318 4988 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Oct 08 18:10:55 crc kubenswrapper[4988]: W1008 18:10:55.006323 4988 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Oct 08 18:10:55 crc kubenswrapper[4988]: W1008 18:10:55.006328 4988 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Oct 08 18:10:55 crc kubenswrapper[4988]: W1008 18:10:55.006333 4988 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Oct 08 18:10:55 crc kubenswrapper[4988]: W1008 18:10:55.006339 4988 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Oct 08 18:10:55 crc kubenswrapper[4988]: W1008 18:10:55.006346 4988 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Oct 08 18:10:55 crc kubenswrapper[4988]: W1008 18:10:55.006352 4988 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Oct 08 18:10:55 crc kubenswrapper[4988]: W1008 18:10:55.006357 4988 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Oct 08 18:10:55 crc kubenswrapper[4988]: W1008 18:10:55.006363 4988 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Oct 08 18:10:55 crc kubenswrapper[4988]: W1008 18:10:55.006368 4988 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Oct 08 18:10:55 crc kubenswrapper[4988]: W1008 18:10:55.006373 4988 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Oct 08 18:10:55 crc kubenswrapper[4988]: W1008 18:10:55.006378 4988 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Oct 08 18:10:55 crc kubenswrapper[4988]: W1008 18:10:55.006403 4988 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Oct 08 18:10:55 crc kubenswrapper[4988]: W1008 18:10:55.006409 4988 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Oct 08 18:10:55 crc kubenswrapper[4988]: W1008 18:10:55.006414 4988 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Oct 08 18:10:55 crc kubenswrapper[4988]: W1008 18:10:55.006419 4988 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Oct 08 18:10:55 crc kubenswrapper[4988]: W1008 18:10:55.006423 4988 feature_gate.go:330] unrecognized feature gate: GatewayAPI Oct 08 18:10:55 crc kubenswrapper[4988]: W1008 18:10:55.006428 4988 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Oct 08 18:10:55 crc kubenswrapper[4988]: W1008 18:10:55.006433 4988 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Oct 08 18:10:55 crc kubenswrapper[4988]: W1008 18:10:55.006438 4988 feature_gate.go:330] unrecognized feature gate: OVNObservability Oct 08 18:10:55 crc kubenswrapper[4988]: W1008 18:10:55.006443 4988 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Oct 08 18:10:55 crc kubenswrapper[4988]: W1008 18:10:55.006448 4988 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Oct 08 18:10:55 crc kubenswrapper[4988]: W1008 18:10:55.006453 4988 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Oct 08 18:10:55 crc kubenswrapper[4988]: W1008 18:10:55.006458 4988 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Oct 08 18:10:55 crc kubenswrapper[4988]: W1008 18:10:55.006466 4988 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Oct 08 18:10:55 crc kubenswrapper[4988]: W1008 18:10:55.006473 4988 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Oct 08 18:10:55 crc kubenswrapper[4988]: W1008 18:10:55.006479 4988 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Oct 08 18:10:55 crc kubenswrapper[4988]: W1008 18:10:55.006487 4988 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Oct 08 18:10:55 crc kubenswrapper[4988]: W1008 18:10:55.006492 4988 feature_gate.go:330] unrecognized feature gate: InsightsConfig Oct 08 18:10:55 crc kubenswrapper[4988]: W1008 18:10:55.006497 4988 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Oct 08 18:10:55 crc kubenswrapper[4988]: W1008 18:10:55.006503 4988 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Oct 08 18:10:55 crc kubenswrapper[4988]: W1008 18:10:55.006507 4988 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Oct 08 18:10:55 crc kubenswrapper[4988]: W1008 18:10:55.006512 4988 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Oct 08 18:10:55 crc kubenswrapper[4988]: W1008 18:10:55.006517 4988 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Oct 08 18:10:55 crc kubenswrapper[4988]: W1008 18:10:55.006522 4988 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Oct 08 18:10:55 crc kubenswrapper[4988]: W1008 18:10:55.006527 4988 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Oct 08 18:10:55 crc kubenswrapper[4988]: W1008 18:10:55.006532 4988 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Oct 08 18:10:55 crc kubenswrapper[4988]: W1008 18:10:55.006537 4988 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Oct 08 18:10:55 crc kubenswrapper[4988]: W1008 18:10:55.006542 4988 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Oct 08 18:10:55 crc kubenswrapper[4988]: W1008 18:10:55.006547 4988 feature_gate.go:330] unrecognized feature gate: PlatformOperators Oct 08 18:10:55 crc kubenswrapper[4988]: W1008 18:10:55.006552 4988 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Oct 08 18:10:55 crc kubenswrapper[4988]: W1008 18:10:55.006557 4988 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Oct 08 18:10:55 crc kubenswrapper[4988]: W1008 18:10:55.006562 4988 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Oct 08 18:10:55 crc kubenswrapper[4988]: W1008 18:10:55.006567 4988 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Oct 08 18:10:55 crc kubenswrapper[4988]: W1008 18:10:55.006572 4988 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Oct 08 18:10:55 crc kubenswrapper[4988]: W1008 18:10:55.006577 4988 feature_gate.go:330] unrecognized feature gate: NewOLM Oct 08 18:10:55 crc kubenswrapper[4988]: W1008 18:10:55.006582 4988 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Oct 08 18:10:55 crc kubenswrapper[4988]: W1008 18:10:55.006587 4988 feature_gate.go:330] unrecognized feature gate: Example Oct 08 18:10:55 crc kubenswrapper[4988]: W1008 18:10:55.006591 4988 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Oct 08 18:10:55 crc kubenswrapper[4988]: W1008 18:10:55.006597 4988 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Oct 08 18:10:55 crc kubenswrapper[4988]: W1008 18:10:55.006601 4988 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Oct 08 18:10:55 crc kubenswrapper[4988]: W1008 18:10:55.006606 4988 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Oct 08 18:10:55 crc kubenswrapper[4988]: W1008 18:10:55.006611 4988 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Oct 08 18:10:55 crc kubenswrapper[4988]: W1008 18:10:55.006616 4988 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Oct 08 18:10:55 crc kubenswrapper[4988]: W1008 18:10:55.006621 4988 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Oct 08 18:10:55 crc kubenswrapper[4988]: W1008 18:10:55.006626 4988 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Oct 08 18:10:55 crc kubenswrapper[4988]: W1008 18:10:55.006634 4988 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Oct 08 18:10:55 crc kubenswrapper[4988]: W1008 18:10:55.006639 4988 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Oct 08 18:10:55 crc kubenswrapper[4988]: W1008 18:10:55.006646 4988 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Oct 08 18:10:55 crc kubenswrapper[4988]: W1008 18:10:55.006652 4988 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Oct 08 18:10:55 crc kubenswrapper[4988]: W1008 18:10:55.006657 4988 feature_gate.go:330] unrecognized feature gate: PinnedImages Oct 08 18:10:55 crc kubenswrapper[4988]: W1008 18:10:55.006663 4988 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Oct 08 18:10:55 crc kubenswrapper[4988]: W1008 18:10:55.006669 4988 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Oct 08 18:10:55 crc kubenswrapper[4988]: W1008 18:10:55.006675 4988 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:55.006685 4988 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Oct 08 18:10:55 crc kubenswrapper[4988]: W1008 18:10:55.006839 4988 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Oct 08 18:10:55 crc kubenswrapper[4988]: W1008 18:10:55.006849 4988 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Oct 08 18:10:55 crc kubenswrapper[4988]: W1008 18:10:55.006855 4988 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Oct 08 18:10:55 crc kubenswrapper[4988]: W1008 18:10:55.006862 4988 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Oct 08 18:10:55 crc kubenswrapper[4988]: W1008 18:10:55.006870 4988 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Oct 08 18:10:55 crc kubenswrapper[4988]: W1008 18:10:55.006876 4988 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Oct 08 18:10:55 crc kubenswrapper[4988]: W1008 18:10:55.006883 4988 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Oct 08 18:10:55 crc kubenswrapper[4988]: W1008 18:10:55.006909 4988 feature_gate.go:330] unrecognized feature gate: OVNObservability Oct 08 18:10:55 crc kubenswrapper[4988]: W1008 18:10:55.006917 4988 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Oct 08 18:10:55 crc kubenswrapper[4988]: W1008 18:10:55.006927 4988 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Oct 08 18:10:55 crc kubenswrapper[4988]: W1008 18:10:55.006934 4988 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Oct 08 18:10:55 crc kubenswrapper[4988]: W1008 18:10:55.006940 4988 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Oct 08 18:10:55 crc kubenswrapper[4988]: W1008 18:10:55.006947 4988 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Oct 08 18:10:55 crc kubenswrapper[4988]: W1008 18:10:55.006953 4988 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Oct 08 18:10:55 crc kubenswrapper[4988]: W1008 18:10:55.006959 4988 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Oct 08 18:10:55 crc kubenswrapper[4988]: W1008 18:10:55.006966 4988 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Oct 08 18:10:55 crc kubenswrapper[4988]: W1008 18:10:55.006971 4988 feature_gate.go:330] unrecognized feature gate: GatewayAPI Oct 08 18:10:55 crc kubenswrapper[4988]: W1008 18:10:55.006978 4988 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Oct 08 18:10:55 crc kubenswrapper[4988]: W1008 18:10:55.006984 4988 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Oct 08 18:10:55 crc kubenswrapper[4988]: W1008 18:10:55.006990 4988 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Oct 08 18:10:55 crc kubenswrapper[4988]: W1008 18:10:55.006997 4988 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Oct 08 18:10:55 crc kubenswrapper[4988]: W1008 18:10:55.007005 4988 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Oct 08 18:10:55 crc kubenswrapper[4988]: W1008 18:10:55.007014 4988 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Oct 08 18:10:55 crc kubenswrapper[4988]: W1008 18:10:55.007022 4988 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Oct 08 18:10:55 crc kubenswrapper[4988]: W1008 18:10:55.007030 4988 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Oct 08 18:10:55 crc kubenswrapper[4988]: W1008 18:10:55.007038 4988 feature_gate.go:330] unrecognized feature gate: InsightsConfig Oct 08 18:10:55 crc kubenswrapper[4988]: W1008 18:10:55.007044 4988 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Oct 08 18:10:55 crc kubenswrapper[4988]: W1008 18:10:55.007050 4988 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Oct 08 18:10:55 crc kubenswrapper[4988]: W1008 18:10:55.007056 4988 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Oct 08 18:10:55 crc kubenswrapper[4988]: W1008 18:10:55.007063 4988 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Oct 08 18:10:55 crc kubenswrapper[4988]: W1008 18:10:55.007069 4988 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Oct 08 18:10:55 crc kubenswrapper[4988]: W1008 18:10:55.007075 4988 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Oct 08 18:10:55 crc kubenswrapper[4988]: W1008 18:10:55.007080 4988 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Oct 08 18:10:55 crc kubenswrapper[4988]: W1008 18:10:55.007086 4988 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Oct 08 18:10:55 crc kubenswrapper[4988]: W1008 18:10:55.007101 4988 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Oct 08 18:10:55 crc kubenswrapper[4988]: W1008 18:10:55.007108 4988 feature_gate.go:330] unrecognized feature gate: PinnedImages Oct 08 18:10:55 crc kubenswrapper[4988]: W1008 18:10:55.007115 4988 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Oct 08 18:10:55 crc kubenswrapper[4988]: W1008 18:10:55.007122 4988 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Oct 08 18:10:55 crc kubenswrapper[4988]: W1008 18:10:55.007130 4988 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Oct 08 18:10:55 crc kubenswrapper[4988]: W1008 18:10:55.007138 4988 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Oct 08 18:10:55 crc kubenswrapper[4988]: W1008 18:10:55.007144 4988 feature_gate.go:330] unrecognized feature gate: Example Oct 08 18:10:55 crc kubenswrapper[4988]: W1008 18:10:55.007151 4988 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Oct 08 18:10:55 crc kubenswrapper[4988]: W1008 18:10:55.007157 4988 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Oct 08 18:10:55 crc kubenswrapper[4988]: W1008 18:10:55.007163 4988 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Oct 08 18:10:55 crc kubenswrapper[4988]: W1008 18:10:55.007169 4988 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Oct 08 18:10:55 crc kubenswrapper[4988]: W1008 18:10:55.007175 4988 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Oct 08 18:10:55 crc kubenswrapper[4988]: W1008 18:10:55.007181 4988 feature_gate.go:330] unrecognized feature gate: SignatureStores Oct 08 18:10:55 crc kubenswrapper[4988]: W1008 18:10:55.007186 4988 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Oct 08 18:10:55 crc kubenswrapper[4988]: W1008 18:10:55.007191 4988 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Oct 08 18:10:55 crc kubenswrapper[4988]: W1008 18:10:55.007197 4988 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Oct 08 18:10:55 crc kubenswrapper[4988]: W1008 18:10:55.007202 4988 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Oct 08 18:10:55 crc kubenswrapper[4988]: W1008 18:10:55.007207 4988 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Oct 08 18:10:55 crc kubenswrapper[4988]: W1008 18:10:55.007214 4988 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Oct 08 18:10:55 crc kubenswrapper[4988]: W1008 18:10:55.007220 4988 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Oct 08 18:10:55 crc kubenswrapper[4988]: W1008 18:10:55.007225 4988 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Oct 08 18:10:55 crc kubenswrapper[4988]: W1008 18:10:55.007230 4988 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Oct 08 18:10:55 crc kubenswrapper[4988]: W1008 18:10:55.007236 4988 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Oct 08 18:10:55 crc kubenswrapper[4988]: W1008 18:10:55.007241 4988 feature_gate.go:330] unrecognized feature gate: PlatformOperators Oct 08 18:10:55 crc kubenswrapper[4988]: W1008 18:10:55.007246 4988 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Oct 08 18:10:55 crc kubenswrapper[4988]: W1008 18:10:55.007251 4988 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Oct 08 18:10:55 crc kubenswrapper[4988]: W1008 18:10:55.007256 4988 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Oct 08 18:10:55 crc kubenswrapper[4988]: W1008 18:10:55.007261 4988 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Oct 08 18:10:55 crc kubenswrapper[4988]: W1008 18:10:55.007266 4988 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Oct 08 18:10:55 crc kubenswrapper[4988]: W1008 18:10:55.007271 4988 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Oct 08 18:10:55 crc kubenswrapper[4988]: W1008 18:10:55.007276 4988 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Oct 08 18:10:55 crc kubenswrapper[4988]: W1008 18:10:55.007280 4988 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Oct 08 18:10:55 crc kubenswrapper[4988]: W1008 18:10:55.007287 4988 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Oct 08 18:10:55 crc kubenswrapper[4988]: W1008 18:10:55.007294 4988 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Oct 08 18:10:55 crc kubenswrapper[4988]: W1008 18:10:55.007299 4988 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Oct 08 18:10:55 crc kubenswrapper[4988]: W1008 18:10:55.007305 4988 feature_gate.go:330] unrecognized feature gate: NewOLM Oct 08 18:10:55 crc kubenswrapper[4988]: W1008 18:10:55.007313 4988 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:55.007322 4988 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:55.009091 4988 server.go:940] "Client rotation is on, will bootstrap in background" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:55.014060 4988 bootstrap.go:85] "Current kubeconfig file contents are still valid, no bootstrap necessary" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:55.014152 4988 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-client-current.pem". Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:55.016412 4988 server.go:997] "Starting client certificate rotation" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:55.016441 4988 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate rotation is enabled Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:55.017571 4988 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate expiration is 2026-02-24 05:52:08 +0000 UTC, rotation deadline is 2025-12-07 17:51:45.113622763 +0000 UTC Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:55.017680 4988 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Waiting 1439h40m50.095948142s for next certificate rotation Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:55.046653 4988 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:55.054358 4988 dynamic_cafile_content.go:161] "Starting controller" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:55.074137 4988 log.go:25] "Validated CRI v1 runtime API" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:55.116767 4988 log.go:25] "Validated CRI v1 image API" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:55.119334 4988 server.go:1437] "Using cgroup driver setting received from the CRI runtime" cgroupDriver="systemd" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:55.127158 4988 fs.go:133] Filesystem UUIDs: map[0b076daa-c26a-46d2-b3a6-72a8dbc6e257:/dev/vda4 2025-10-08-18-06-12-00:/dev/sr0 7B77-95E7:/dev/vda2 de0497b0-db1b-465a-b278-03db02455c71:/dev/vda3] Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:55.127210 4988 fs.go:134] Filesystem partitions: map[/dev/shm:{mountpoint:/dev/shm major:0 minor:22 fsType:tmpfs blockSize:0} /dev/vda3:{mountpoint:/boot major:252 minor:3 fsType:ext4 blockSize:0} /dev/vda4:{mountpoint:/var major:252 minor:4 fsType:xfs blockSize:0} /run:{mountpoint:/run major:0 minor:24 fsType:tmpfs blockSize:0} /run/user/1000:{mountpoint:/run/user/1000 major:0 minor:42 fsType:tmpfs blockSize:0} /tmp:{mountpoint:/tmp major:0 minor:30 fsType:tmpfs blockSize:0} /var/lib/etcd:{mountpoint:/var/lib/etcd major:0 minor:43 fsType:tmpfs blockSize:0}] Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:55.146432 4988 manager.go:217] Machine: {Timestamp:2025-10-08 18:10:55.143109353 +0000 UTC m=+0.592952163 CPUVendorID:AuthenticAMD NumCores:12 NumPhysicalCores:1 NumSockets:12 CpuFrequency:2800000 MemoryCapacity:33654120448 SwapCapacity:0 MemoryByType:map[] NVMInfo:{MemoryModeCapacity:0 AppDirectModeCapacity:0 AvgPowerBudget:0} HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] MachineID:21801e6708c44f15b81395eb736a7cec SystemUUID:bd9489c4-fb7d-486d-b7ce-a13b6d7de36a BootID:95ee9fd0-8bc9-4ca5-9d79-5ee94d58bfeb Filesystems:[{Device:/dev/vda3 DeviceMajor:252 DeviceMinor:3 Capacity:366869504 Type:vfs Inodes:98304 HasInodes:true} {Device:/run/user/1000 DeviceMajor:0 DeviceMinor:42 Capacity:3365408768 Type:vfs Inodes:821633 HasInodes:true} {Device:/var/lib/etcd DeviceMajor:0 DeviceMinor:43 Capacity:1073741824 Type:vfs Inodes:4108169 HasInodes:true} {Device:/dev/shm DeviceMajor:0 DeviceMinor:22 Capacity:16827060224 Type:vfs Inodes:4108169 HasInodes:true} {Device:/run DeviceMajor:0 DeviceMinor:24 Capacity:6730825728 Type:vfs Inodes:819200 HasInodes:true} {Device:/dev/vda4 DeviceMajor:252 DeviceMinor:4 Capacity:85292941312 Type:vfs Inodes:41679680 HasInodes:true} {Device:/tmp DeviceMajor:0 DeviceMinor:30 Capacity:16827060224 Type:vfs Inodes:1048576 HasInodes:true}] DiskMap:map[252:0:{Name:vda Major:252 Minor:0 Size:214748364800 Scheduler:none}] NetworkDevices:[{Name:br-ex MacAddress:fa:16:3e:9e:1f:7d Speed:0 Mtu:1500} {Name:br-int MacAddress:d6:39:55:2e:22:71 Speed:0 Mtu:1400} {Name:ens3 MacAddress:fa:16:3e:9e:1f:7d Speed:-1 Mtu:1500} {Name:ens7 MacAddress:fa:16:3e:bb:80:cb Speed:-1 Mtu:1500} {Name:ens7.20 MacAddress:52:54:00:91:b4:f6 Speed:-1 Mtu:1496} {Name:ens7.21 MacAddress:52:54:00:14:c7:fa Speed:-1 Mtu:1496} {Name:ens7.22 MacAddress:52:54:00:e1:b2:53 Speed:-1 Mtu:1496} {Name:ens7.23 MacAddress:52:54:00:b9:a8:10 Speed:-1 Mtu:1496} {Name:eth10 MacAddress:82:69:a2:dd:67:f0 Speed:0 Mtu:1500} {Name:ovn-k8s-mp0 MacAddress:0a:58:0a:d9:00:02 Speed:0 Mtu:1400} {Name:ovs-system MacAddress:06:30:ca:48:6c:c5 Speed:0 Mtu:1500}] Topology:[{Id:0 Memory:33654120448 HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] Cores:[{Id:0 Threads:[0] Caches:[{Id:0 Size:32768 Type:Data Level:1} {Id:0 Size:32768 Type:Instruction Level:1} {Id:0 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:0 Size:16777216 Type:Unified Level:3}] SocketID:0 BookID: DrawerID:} {Id:0 Threads:[1] Caches:[{Id:1 Size:32768 Type:Data Level:1} {Id:1 Size:32768 Type:Instruction Level:1} {Id:1 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:1 Size:16777216 Type:Unified Level:3}] SocketID:1 BookID: DrawerID:} {Id:0 Threads:[10] Caches:[{Id:10 Size:32768 Type:Data Level:1} {Id:10 Size:32768 Type:Instruction Level:1} {Id:10 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:10 Size:16777216 Type:Unified Level:3}] SocketID:10 BookID: DrawerID:} {Id:0 Threads:[11] Caches:[{Id:11 Size:32768 Type:Data Level:1} {Id:11 Size:32768 Type:Instruction Level:1} {Id:11 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:11 Size:16777216 Type:Unified Level:3}] SocketID:11 BookID: DrawerID:} {Id:0 Threads:[2] Caches:[{Id:2 Size:32768 Type:Data Level:1} {Id:2 Size:32768 Type:Instruction Level:1} {Id:2 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:2 Size:16777216 Type:Unified Level:3}] SocketID:2 BookID: DrawerID:} {Id:0 Threads:[3] Caches:[{Id:3 Size:32768 Type:Data Level:1} {Id:3 Size:32768 Type:Instruction Level:1} {Id:3 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:3 Size:16777216 Type:Unified Level:3}] SocketID:3 BookID: DrawerID:} {Id:0 Threads:[4] Caches:[{Id:4 Size:32768 Type:Data Level:1} {Id:4 Size:32768 Type:Instruction Level:1} {Id:4 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:4 Size:16777216 Type:Unified Level:3}] SocketID:4 BookID: DrawerID:} {Id:0 Threads:[5] Caches:[{Id:5 Size:32768 Type:Data Level:1} {Id:5 Size:32768 Type:Instruction Level:1} {Id:5 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:5 Size:16777216 Type:Unified Level:3}] SocketID:5 BookID: DrawerID:} {Id:0 Threads:[6] Caches:[{Id:6 Size:32768 Type:Data Level:1} {Id:6 Size:32768 Type:Instruction Level:1} {Id:6 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:6 Size:16777216 Type:Unified Level:3}] SocketID:6 BookID: DrawerID:} {Id:0 Threads:[7] Caches:[{Id:7 Size:32768 Type:Data Level:1} {Id:7 Size:32768 Type:Instruction Level:1} {Id:7 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:7 Size:16777216 Type:Unified Level:3}] SocketID:7 BookID: DrawerID:} {Id:0 Threads:[8] Caches:[{Id:8 Size:32768 Type:Data Level:1} {Id:8 Size:32768 Type:Instruction Level:1} {Id:8 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:8 Size:16777216 Type:Unified Level:3}] SocketID:8 BookID: DrawerID:} {Id:0 Threads:[9] Caches:[{Id:9 Size:32768 Type:Data Level:1} {Id:9 Size:32768 Type:Instruction Level:1} {Id:9 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:9 Size:16777216 Type:Unified Level:3}] SocketID:9 BookID: DrawerID:}] Caches:[] Distances:[10]}] CloudProvider:Unknown InstanceType:Unknown InstanceID:None} Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:55.147044 4988 manager_no_libpfm.go:29] cAdvisor is build without cgo and/or libpfm support. Perf event counters are not available. Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:55.147279 4988 manager.go:233] Version: {KernelVersion:5.14.0-427.50.2.el9_4.x86_64 ContainerOsVersion:Red Hat Enterprise Linux CoreOS 418.94.202502100215-0 DockerVersion: DockerAPIVersion: CadvisorVersion: CadvisorRevision:} Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:55.150953 4988 swap_util.go:113] "Swap is on" /proc/swaps contents="Filename\t\t\t\tType\t\tSize\t\tUsed\t\tPriority" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:55.151278 4988 container_manager_linux.go:267] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:55.151338 4988 container_manager_linux.go:272] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"crc","RuntimeCgroupsName":"/system.slice/crio.service","SystemCgroupsName":"/system.slice","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":true,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":{"cpu":"200m","ephemeral-storage":"350Mi","memory":"350Mi"},"HardEvictionThresholds":[{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":4096,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:55.151879 4988 topology_manager.go:138] "Creating topology manager with none policy" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:55.151898 4988 container_manager_linux.go:303] "Creating device plugin manager" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:55.152850 4988 manager.go:142] "Creating Device Plugin manager" path="/var/lib/kubelet/device-plugins/kubelet.sock" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:55.152902 4988 server.go:66] "Creating device plugin registration server" version="v1beta1" socket="/var/lib/kubelet/device-plugins/kubelet.sock" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:55.153157 4988 state_mem.go:36] "Initialized new in-memory state store" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:55.153294 4988 server.go:1245] "Using root directory" path="/var/lib/kubelet" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:55.157312 4988 kubelet.go:418] "Attempting to sync node with API server" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:55.157358 4988 kubelet.go:313] "Adding static pod path" path="/etc/kubernetes/manifests" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:55.157456 4988 file.go:69] "Watching path" path="/etc/kubernetes/manifests" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:55.157488 4988 kubelet.go:324] "Adding apiserver pod source" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:55.157516 4988 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:55.167190 4988 kuberuntime_manager.go:262] "Container runtime initialized" containerRuntime="cri-o" version="1.31.5-4.rhaos4.18.gitdad78d5.el9" apiVersion="v1" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:55.168317 4988 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-server-current.pem". Oct 08 18:10:55 crc kubenswrapper[4988]: W1008 18:10:55.168906 4988 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.102:6443: connect: connection refused Oct 08 18:10:55 crc kubenswrapper[4988]: W1008 18:10:55.168967 4988 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.102:6443: connect: connection refused Oct 08 18:10:55 crc kubenswrapper[4988]: E1008 18:10:55.169133 4988 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.102:6443: connect: connection refused" logger="UnhandledError" Oct 08 18:10:55 crc kubenswrapper[4988]: E1008 18:10:55.169134 4988 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.102:6443: connect: connection refused" logger="UnhandledError" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:55.172451 4988 kubelet.go:854] "Not starting ClusterTrustBundle informer because we are in static kubelet mode" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:55.174150 4988 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/portworx-volume" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:55.174176 4988 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/empty-dir" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:55.174183 4988 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/git-repo" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:55.174190 4988 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/host-path" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:55.174203 4988 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/nfs" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:55.174211 4988 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/secret" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:55.174219 4988 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/iscsi" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:55.174231 4988 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/downward-api" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:55.174240 4988 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/fc" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:55.174249 4988 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/configmap" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:55.174263 4988 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/projected" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:55.174273 4988 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/local-volume" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:55.178335 4988 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/csi" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:55.178865 4988 server.go:1280] "Started kubelet" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:55.179705 4988 server.go:163] "Starting to listen" address="0.0.0.0" port=10250 Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:55.179734 4988 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Oct 08 18:10:55 crc systemd[1]: Started Kubernetes Kubelet. Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:55.180595 4988 server.go:236] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:55.181527 4988 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.102:6443: connect: connection refused Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:55.182897 4988 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate rotation is enabled Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:55.182940 4988 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:55.182993 4988 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-22 15:56:12.404434464 +0000 UTC Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:55.183180 4988 certificate_manager.go:356] kubernetes.io/kubelet-serving: Waiting 1077h45m17.221264026s for next certificate rotation Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:55.183186 4988 volume_manager.go:287] "The desired_state_of_world populator starts" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:55.183222 4988 volume_manager.go:289] "Starting Kubelet Volume Manager" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:55.183317 4988 desired_state_of_world_populator.go:146] "Desired state populator starts to run" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:55.183677 4988 server.go:460] "Adding debug handlers to kubelet server" Oct 08 18:10:55 crc kubenswrapper[4988]: E1008 18:10:55.183295 4988 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Oct 08 18:10:55 crc kubenswrapper[4988]: W1008 18:10:55.183880 4988 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.102:6443: connect: connection refused Oct 08 18:10:55 crc kubenswrapper[4988]: E1008 18:10:55.183955 4988 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.102:6443: connect: connection refused" logger="UnhandledError" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:55.184721 4988 factory.go:55] Registering systemd factory Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:55.184745 4988 factory.go:221] Registration of the systemd container factory successfully Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:55.186581 4988 factory.go:153] Registering CRI-O factory Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:55.186631 4988 factory.go:221] Registration of the crio container factory successfully Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:55.186774 4988 factory.go:219] Registration of the containerd container factory failed: unable to create containerd client: containerd: cannot unix dial containerd api service: dial unix /run/containerd/containerd.sock: connect: no such file or directory Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:55.186845 4988 factory.go:103] Registering Raw factory Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:55.186871 4988 manager.go:1196] Started watching for new ooms in manager Oct 08 18:10:55 crc kubenswrapper[4988]: E1008 18:10:55.187204 4988 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.102:6443: connect: connection refused" interval="200ms" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:55.187602 4988 manager.go:319] Starting recovery of all containers Oct 08 18:10:55 crc kubenswrapper[4988]: E1008 18:10:55.193494 4988 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": dial tcp 38.102.83.102:6443: connect: connection refused" event="&Event{ObjectMeta:{crc.186c9685a76dd185 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-10-08 18:10:55.178830213 +0000 UTC m=+0.628672993,LastTimestamp:2025-10-08 18:10:55.178830213 +0000 UTC m=+0.628672993,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:55.201722 4988 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" seLinuxMountContext="" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:55.201815 4988 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" seLinuxMountContext="" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:55.201838 4988 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" seLinuxMountContext="" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:55.201855 4988 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" seLinuxMountContext="" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:55.201871 4988 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script" seLinuxMountContext="" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:55.201885 4988 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" seLinuxMountContext="" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:55.201898 4988 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" seLinuxMountContext="" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:55.201910 4988 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" seLinuxMountContext="" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:55.201926 4988 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" seLinuxMountContext="" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:55.201939 4988 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" seLinuxMountContext="" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:55.201957 4988 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" seLinuxMountContext="" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:55.201968 4988 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" seLinuxMountContext="" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:55.201981 4988 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" seLinuxMountContext="" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:55.201997 4988 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" seLinuxMountContext="" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:55.202011 4988 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" seLinuxMountContext="" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:55.202024 4988 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" seLinuxMountContext="" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:55.202040 4988 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" seLinuxMountContext="" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:55.202051 4988 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" seLinuxMountContext="" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:55.202061 4988 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" seLinuxMountContext="" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:55.202072 4988 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" seLinuxMountContext="" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:55.202083 4988 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" seLinuxMountContext="" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:55.202097 4988 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" seLinuxMountContext="" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:55.202110 4988 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" seLinuxMountContext="" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:55.202124 4988 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" seLinuxMountContext="" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:55.202136 4988 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" seLinuxMountContext="" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:55.202147 4988 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" seLinuxMountContext="" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:55.202159 4988 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" seLinuxMountContext="" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:55.202172 4988 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" seLinuxMountContext="" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:55.202184 4988 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" seLinuxMountContext="" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:55.202200 4988 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" seLinuxMountContext="" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:55.202219 4988 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" seLinuxMountContext="" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:55.202235 4988 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" seLinuxMountContext="" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:55.206025 4988 reconstruct.go:144] "Volume is marked device as uncertain and added into the actual state" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" deviceMountPath="/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:55.206081 4988 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" seLinuxMountContext="" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:55.206096 4988 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" seLinuxMountContext="" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:55.206110 4988 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" seLinuxMountContext="" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:55.206125 4988 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" seLinuxMountContext="" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:55.206137 4988 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" seLinuxMountContext="" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:55.206155 4988 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" seLinuxMountContext="" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:55.206168 4988 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" seLinuxMountContext="" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:55.206210 4988 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" seLinuxMountContext="" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:55.206223 4988 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" seLinuxMountContext="" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:55.206238 4988 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" seLinuxMountContext="" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:55.206252 4988 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" seLinuxMountContext="" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:55.206266 4988 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" seLinuxMountContext="" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:55.206276 4988 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" seLinuxMountContext="" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:55.206290 4988 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" seLinuxMountContext="" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:55.206305 4988 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" seLinuxMountContext="" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:55.206320 4988 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" seLinuxMountContext="" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:55.206335 4988 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" seLinuxMountContext="" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:55.206350 4988 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" seLinuxMountContext="" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:55.206362 4988 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" seLinuxMountContext="" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:55.206377 4988 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" volumeName="kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" seLinuxMountContext="" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:55.206422 4988 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" seLinuxMountContext="" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:55.206445 4988 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" seLinuxMountContext="" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:55.206466 4988 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" seLinuxMountContext="" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:55.206484 4988 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" seLinuxMountContext="" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:55.206498 4988 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" seLinuxMountContext="" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:55.206513 4988 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" seLinuxMountContext="" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:55.206525 4988 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides" seLinuxMountContext="" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:55.206565 4988 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" seLinuxMountContext="" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:55.206579 4988 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" seLinuxMountContext="" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:55.206591 4988 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" seLinuxMountContext="" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:55.206604 4988 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" seLinuxMountContext="" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:55.206618 4988 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" seLinuxMountContext="" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:55.206630 4988 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" seLinuxMountContext="" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:55.206643 4988 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" seLinuxMountContext="" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:55.206656 4988 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" seLinuxMountContext="" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:55.206669 4988 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" seLinuxMountContext="" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:55.206682 4988 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" seLinuxMountContext="" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:55.206696 4988 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" seLinuxMountContext="" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:55.206709 4988 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" seLinuxMountContext="" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:55.206724 4988 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" seLinuxMountContext="" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:55.206735 4988 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" seLinuxMountContext="" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:55.206747 4988 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" seLinuxMountContext="" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:55.206759 4988 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" seLinuxMountContext="" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:55.206770 4988 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" seLinuxMountContext="" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:55.206783 4988 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" seLinuxMountContext="" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:55.206795 4988 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" seLinuxMountContext="" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:55.206811 4988 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" seLinuxMountContext="" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:55.206829 4988 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" seLinuxMountContext="" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:55.206845 4988 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" seLinuxMountContext="" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:55.206890 4988 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" seLinuxMountContext="" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:55.206942 4988 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3b6479f0-333b-4a96-9adf-2099afdc2447" volumeName="kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr" seLinuxMountContext="" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:55.206960 4988 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" seLinuxMountContext="" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:55.206972 4988 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" seLinuxMountContext="" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:55.206988 4988 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" seLinuxMountContext="" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:55.207002 4988 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" seLinuxMountContext="" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:55.207014 4988 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" seLinuxMountContext="" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:55.207028 4988 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert" seLinuxMountContext="" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:55.207041 4988 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" seLinuxMountContext="" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:55.207054 4988 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" seLinuxMountContext="" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:55.207067 4988 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" seLinuxMountContext="" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:55.207081 4988 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" seLinuxMountContext="" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:55.207094 4988 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" seLinuxMountContext="" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:55.207104 4988 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" seLinuxMountContext="" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:55.207118 4988 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" seLinuxMountContext="" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:55.207132 4988 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" seLinuxMountContext="" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:55.207145 4988 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" seLinuxMountContext="" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:55.207157 4988 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d751cbb-f2e2-430d-9754-c882a5e924a5" volumeName="kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl" seLinuxMountContext="" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:55.207168 4988 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" seLinuxMountContext="" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:55.207179 4988 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" seLinuxMountContext="" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:55.207194 4988 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" seLinuxMountContext="" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:55.207208 4988 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" seLinuxMountContext="" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:55.207220 4988 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" seLinuxMountContext="" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:55.207272 4988 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" seLinuxMountContext="" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:55.207286 4988 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" seLinuxMountContext="" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:55.207300 4988 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" seLinuxMountContext="" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:55.207313 4988 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" seLinuxMountContext="" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:55.207327 4988 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" seLinuxMountContext="" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:55.207341 4988 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" seLinuxMountContext="" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:55.207353 4988 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" seLinuxMountContext="" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:55.207366 4988 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb" seLinuxMountContext="" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:55.207379 4988 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" seLinuxMountContext="" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:55.207408 4988 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="44663579-783b-4372-86d6-acf235a62d72" volumeName="kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" seLinuxMountContext="" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:55.207443 4988 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" seLinuxMountContext="" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:55.207455 4988 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" seLinuxMountContext="" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:55.207465 4988 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" seLinuxMountContext="" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:55.207477 4988 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" seLinuxMountContext="" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:55.207488 4988 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" seLinuxMountContext="" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:55.207501 4988 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" seLinuxMountContext="" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:55.207512 4988 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" volumeName="kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" seLinuxMountContext="" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:55.207524 4988 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" seLinuxMountContext="" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:55.207536 4988 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" seLinuxMountContext="" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:55.207554 4988 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" seLinuxMountContext="" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:55.207571 4988 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" seLinuxMountContext="" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:55.207618 4988 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" seLinuxMountContext="" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:55.207631 4988 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" seLinuxMountContext="" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:55.207643 4988 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" seLinuxMountContext="" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:55.207657 4988 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" seLinuxMountContext="" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:55.207669 4988 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" seLinuxMountContext="" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:55.207683 4988 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5" seLinuxMountContext="" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:55.207696 4988 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" seLinuxMountContext="" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:55.207709 4988 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" seLinuxMountContext="" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:55.207723 4988 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" seLinuxMountContext="" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:55.207737 4988 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" seLinuxMountContext="" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:55.207750 4988 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" seLinuxMountContext="" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:55.207766 4988 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" seLinuxMountContext="" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:55.207781 4988 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf" seLinuxMountContext="" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:55.207796 4988 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" seLinuxMountContext="" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:55.207812 4988 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" seLinuxMountContext="" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:55.207828 4988 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" seLinuxMountContext="" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:55.207841 4988 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" seLinuxMountContext="" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:55.207856 4988 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" seLinuxMountContext="" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:55.207870 4988 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" seLinuxMountContext="" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:55.207882 4988 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" seLinuxMountContext="" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:55.207894 4988 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" seLinuxMountContext="" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:55.207909 4988 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" seLinuxMountContext="" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:55.207921 4988 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" seLinuxMountContext="" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:55.207935 4988 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" seLinuxMountContext="" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:55.207954 4988 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" seLinuxMountContext="" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:55.207971 4988 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" seLinuxMountContext="" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:55.207987 4988 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" seLinuxMountContext="" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:55.208003 4988 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" seLinuxMountContext="" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:55.208020 4988 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" seLinuxMountContext="" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:55.208036 4988 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" seLinuxMountContext="" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:55.208053 4988 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" seLinuxMountContext="" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:55.208067 4988 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" seLinuxMountContext="" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:55.208080 4988 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" seLinuxMountContext="" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:55.208093 4988 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" seLinuxMountContext="" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:55.208108 4988 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert" seLinuxMountContext="" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:55.208126 4988 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" seLinuxMountContext="" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:55.208141 4988 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" seLinuxMountContext="" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:55.208159 4988 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" seLinuxMountContext="" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:55.208175 4988 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" seLinuxMountContext="" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:55.208192 4988 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" seLinuxMountContext="" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:55.208210 4988 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" seLinuxMountContext="" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:55.208227 4988 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49ef4625-1d3a-4a9f-b595-c2433d32326d" volumeName="kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" seLinuxMountContext="" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:55.208243 4988 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" seLinuxMountContext="" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:55.208258 4988 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" seLinuxMountContext="" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:55.208276 4988 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" seLinuxMountContext="" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:55.208290 4988 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" seLinuxMountContext="" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:55.208305 4988 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" seLinuxMountContext="" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:55.208321 4988 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls" seLinuxMountContext="" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:55.208337 4988 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" seLinuxMountContext="" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:55.208354 4988 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" seLinuxMountContext="" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:55.208369 4988 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" seLinuxMountContext="" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:55.208407 4988 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" seLinuxMountContext="" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:55.208426 4988 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" seLinuxMountContext="" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:55.208442 4988 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm" seLinuxMountContext="" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:55.208457 4988 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" seLinuxMountContext="" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:55.208473 4988 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf" seLinuxMountContext="" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:55.208488 4988 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" seLinuxMountContext="" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:55.208502 4988 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" seLinuxMountContext="" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:55.208516 4988 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" seLinuxMountContext="" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:55.208528 4988 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" seLinuxMountContext="" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:55.208540 4988 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" seLinuxMountContext="" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:55.208558 4988 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" seLinuxMountContext="" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:55.208573 4988 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" seLinuxMountContext="" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:55.208587 4988 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" seLinuxMountContext="" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:55.208602 4988 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" seLinuxMountContext="" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:55.208617 4988 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" seLinuxMountContext="" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:55.208635 4988 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" seLinuxMountContext="" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:55.208649 4988 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" seLinuxMountContext="" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:55.208663 4988 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" seLinuxMountContext="" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:55.208680 4988 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" seLinuxMountContext="" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:55.208695 4988 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" seLinuxMountContext="" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:55.208712 4988 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" seLinuxMountContext="" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:55.208729 4988 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" seLinuxMountContext="" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:55.208744 4988 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" seLinuxMountContext="" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:55.208757 4988 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" seLinuxMountContext="" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:55.208770 4988 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" seLinuxMountContext="" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:55.208786 4988 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" seLinuxMountContext="" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:55.208801 4988 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" seLinuxMountContext="" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:55.208816 4988 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" seLinuxMountContext="" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:55.208831 4988 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" seLinuxMountContext="" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:55.208845 4988 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" seLinuxMountContext="" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:55.208859 4988 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" seLinuxMountContext="" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:55.208872 4988 reconstruct.go:97] "Volume reconstruction finished" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:55.208882 4988 reconciler.go:26] "Reconciler: start to sync state" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:55.215881 4988 manager.go:324] Recovery completed Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:55.226371 4988 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:55.228737 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:55.228811 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:55.228828 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:55.230357 4988 cpu_manager.go:225] "Starting CPU manager" policy="none" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:55.230477 4988 cpu_manager.go:226] "Reconciling" reconcilePeriod="10s" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:55.230506 4988 state_mem.go:36] "Initialized new in-memory state store" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:55.233847 4988 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:55.236437 4988 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:55.236483 4988 status_manager.go:217] "Starting to sync pod status with apiserver" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:55.236513 4988 kubelet.go:2335] "Starting kubelet main sync loop" Oct 08 18:10:55 crc kubenswrapper[4988]: E1008 18:10:55.236651 4988 kubelet.go:2359] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Oct 08 18:10:55 crc kubenswrapper[4988]: W1008 18:10:55.239149 4988 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.102:6443: connect: connection refused Oct 08 18:10:55 crc kubenswrapper[4988]: E1008 18:10:55.239242 4988 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.102:6443: connect: connection refused" logger="UnhandledError" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:55.252340 4988 policy_none.go:49] "None policy: Start" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:55.253509 4988 memory_manager.go:170] "Starting memorymanager" policy="None" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:55.253560 4988 state_mem.go:35] "Initializing new in-memory state store" Oct 08 18:10:55 crc kubenswrapper[4988]: E1008 18:10:55.284316 4988 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:55.308863 4988 manager.go:334] "Starting Device Plugin manager" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:55.308927 4988 manager.go:513] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:55.308944 4988 server.go:79] "Starting device plugin registration server" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:55.309577 4988 eviction_manager.go:189] "Eviction manager: starting control loop" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:55.309602 4988 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:55.309850 4988 plugin_watcher.go:51] "Plugin Watcher Start" path="/var/lib/kubelet/plugins_registry" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:55.309945 4988 plugin_manager.go:116] "The desired_state_of_world populator (plugin watcher) starts" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:55.309951 4988 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Oct 08 18:10:55 crc kubenswrapper[4988]: E1008 18:10:55.317906 4988 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:55.337279 4988 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-machine-config-operator/kube-rbac-proxy-crio-crc","openshift-etcd/etcd-crc","openshift-kube-apiserver/kube-apiserver-crc","openshift-kube-controller-manager/kube-controller-manager-crc","openshift-kube-scheduler/openshift-kube-scheduler-crc"] Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:55.337407 4988 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:55.339017 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:55.339063 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:55.339074 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:55.339266 4988 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:55.339732 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:55.339833 4988 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:55.340343 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:55.340372 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:55.340403 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:55.340584 4988 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:55.340748 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:55.340786 4988 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:55.341182 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:55.341199 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:55.341209 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:55.341684 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:55.341710 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:55.341720 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:55.341836 4988 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:55.341994 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:55.342037 4988 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:55.342759 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:55.342781 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:55.342795 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:55.342804 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:55.342825 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:55.342835 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:55.342861 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:55.342901 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:55.342920 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:55.343012 4988 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:55.343059 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:55.343113 4988 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:55.344198 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:55.344232 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:55.344248 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:55.344653 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:55.344891 4988 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:55.347114 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:55.347192 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:55.347214 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:55.349093 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:55.349119 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:55.349131 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:10:55 crc kubenswrapper[4988]: E1008 18:10:55.387801 4988 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.102:6443: connect: connection refused" interval="400ms" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:55.409895 4988 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:55.410880 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:55.410954 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:55.410993 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:55.411181 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:55.411228 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:55.411283 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:55.411343 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:55.411409 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:55.411446 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:55.411480 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:55.411513 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:55.411566 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:55.411593 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:55.411615 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:55.411635 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:55.412082 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:55.412132 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:55.412141 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:55.412163 4988 kubelet_node_status.go:76] "Attempting to register node" node="crc" Oct 08 18:10:55 crc kubenswrapper[4988]: E1008 18:10:55.412528 4988 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.102:6443: connect: connection refused" node="crc" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:55.512598 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:55.512958 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:55.513033 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:55.513142 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:55.513162 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:55.513179 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:55.513199 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:55.513217 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:55.513232 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:55.513248 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:55.513265 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:55.513283 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:55.513300 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:55.513316 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:55.513332 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:55.513349 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:55.513351 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:55.513465 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:55.513482 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:55.513550 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:55.513605 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:55.513609 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:55.513631 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:55.513633 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:55.513653 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:55.513649 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:55.513700 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:55.513699 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:55.513644 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:55.513655 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:55.613049 4988 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:55.616032 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:55.616099 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:55.616108 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:55.616159 4988 kubelet_node_status.go:76] "Attempting to register node" node="crc" Oct 08 18:10:55 crc kubenswrapper[4988]: E1008 18:10:55.616915 4988 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.102:6443: connect: connection refused" node="crc" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:55.679810 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:55.687670 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:55.713319 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:55.733818 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 08 18:10:55 crc kubenswrapper[4988]: W1008 18:10:55.738890 4988 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf4b27818a5e8e43d0dc095d08835c792.slice/crio-b8a46f64980785ecb83551cd7fff13e611b4cf33214d291667d14956a8ce59ed WatchSource:0}: Error finding container b8a46f64980785ecb83551cd7fff13e611b4cf33214d291667d14956a8ce59ed: Status 404 returned error can't find the container with id b8a46f64980785ecb83551cd7fff13e611b4cf33214d291667d14956a8ce59ed Oct 08 18:10:55 crc kubenswrapper[4988]: I1008 18:10:55.739290 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 08 18:10:55 crc kubenswrapper[4988]: W1008 18:10:55.739645 4988 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd1b160f5dda77d281dd8e69ec8d817f9.slice/crio-cca9af02840a6322fdd48a6bb7eafd69542491cdeb242291277572f6bbd5cb2a WatchSource:0}: Error finding container cca9af02840a6322fdd48a6bb7eafd69542491cdeb242291277572f6bbd5cb2a: Status 404 returned error can't find the container with id cca9af02840a6322fdd48a6bb7eafd69542491cdeb242291277572f6bbd5cb2a Oct 08 18:10:55 crc kubenswrapper[4988]: W1008 18:10:55.749897 4988 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2139d3e2895fc6797b9c76a1b4c9886d.slice/crio-6a70db63009ade06760d74abd8f4ebc6166e11e6daf2d989b01386d4f386f2ef WatchSource:0}: Error finding container 6a70db63009ade06760d74abd8f4ebc6166e11e6daf2d989b01386d4f386f2ef: Status 404 returned error can't find the container with id 6a70db63009ade06760d74abd8f4ebc6166e11e6daf2d989b01386d4f386f2ef Oct 08 18:10:55 crc kubenswrapper[4988]: W1008 18:10:55.751896 4988 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf614b9022728cf315e60c057852e563e.slice/crio-1456a66beb8303ada02cb2cad31924fb177b6abbdf4c528a3947235450948f99 WatchSource:0}: Error finding container 1456a66beb8303ada02cb2cad31924fb177b6abbdf4c528a3947235450948f99: Status 404 returned error can't find the container with id 1456a66beb8303ada02cb2cad31924fb177b6abbdf4c528a3947235450948f99 Oct 08 18:10:55 crc kubenswrapper[4988]: W1008 18:10:55.755966 4988 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3dcd261975c3d6b9a6ad6367fd4facd3.slice/crio-678d9647488c878f5552b02c35d21866cceb53acdd9df89a5436f6bf94b93d70 WatchSource:0}: Error finding container 678d9647488c878f5552b02c35d21866cceb53acdd9df89a5436f6bf94b93d70: Status 404 returned error can't find the container with id 678d9647488c878f5552b02c35d21866cceb53acdd9df89a5436f6bf94b93d70 Oct 08 18:10:55 crc kubenswrapper[4988]: E1008 18:10:55.788727 4988 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.102:6443: connect: connection refused" interval="800ms" Oct 08 18:10:56 crc kubenswrapper[4988]: I1008 18:10:56.017829 4988 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 08 18:10:56 crc kubenswrapper[4988]: I1008 18:10:56.019685 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:10:56 crc kubenswrapper[4988]: I1008 18:10:56.019764 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:10:56 crc kubenswrapper[4988]: I1008 18:10:56.019791 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:10:56 crc kubenswrapper[4988]: I1008 18:10:56.019848 4988 kubelet_node_status.go:76] "Attempting to register node" node="crc" Oct 08 18:10:56 crc kubenswrapper[4988]: E1008 18:10:56.020455 4988 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.102:6443: connect: connection refused" node="crc" Oct 08 18:10:56 crc kubenswrapper[4988]: W1008 18:10:56.122016 4988 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.102:6443: connect: connection refused Oct 08 18:10:56 crc kubenswrapper[4988]: E1008 18:10:56.122178 4988 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.102:6443: connect: connection refused" logger="UnhandledError" Oct 08 18:10:56 crc kubenswrapper[4988]: W1008 18:10:56.156691 4988 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.102:6443: connect: connection refused Oct 08 18:10:56 crc kubenswrapper[4988]: E1008 18:10:56.156842 4988 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.102:6443: connect: connection refused" logger="UnhandledError" Oct 08 18:10:56 crc kubenswrapper[4988]: I1008 18:10:56.183373 4988 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.102:6443: connect: connection refused Oct 08 18:10:56 crc kubenswrapper[4988]: I1008 18:10:56.241445 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"678d9647488c878f5552b02c35d21866cceb53acdd9df89a5436f6bf94b93d70"} Oct 08 18:10:56 crc kubenswrapper[4988]: I1008 18:10:56.242644 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"1456a66beb8303ada02cb2cad31924fb177b6abbdf4c528a3947235450948f99"} Oct 08 18:10:56 crc kubenswrapper[4988]: I1008 18:10:56.243809 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"6a70db63009ade06760d74abd8f4ebc6166e11e6daf2d989b01386d4f386f2ef"} Oct 08 18:10:56 crc kubenswrapper[4988]: I1008 18:10:56.245126 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"b8a46f64980785ecb83551cd7fff13e611b4cf33214d291667d14956a8ce59ed"} Oct 08 18:10:56 crc kubenswrapper[4988]: I1008 18:10:56.246149 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"cca9af02840a6322fdd48a6bb7eafd69542491cdeb242291277572f6bbd5cb2a"} Oct 08 18:10:56 crc kubenswrapper[4988]: W1008 18:10:56.496902 4988 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.102:6443: connect: connection refused Oct 08 18:10:56 crc kubenswrapper[4988]: E1008 18:10:56.497037 4988 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.102:6443: connect: connection refused" logger="UnhandledError" Oct 08 18:10:56 crc kubenswrapper[4988]: E1008 18:10:56.590621 4988 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.102:6443: connect: connection refused" interval="1.6s" Oct 08 18:10:56 crc kubenswrapper[4988]: W1008 18:10:56.642932 4988 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.102:6443: connect: connection refused Oct 08 18:10:56 crc kubenswrapper[4988]: E1008 18:10:56.643362 4988 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.102:6443: connect: connection refused" logger="UnhandledError" Oct 08 18:10:56 crc kubenswrapper[4988]: I1008 18:10:56.821361 4988 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 08 18:10:56 crc kubenswrapper[4988]: I1008 18:10:56.823436 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:10:56 crc kubenswrapper[4988]: I1008 18:10:56.823505 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:10:56 crc kubenswrapper[4988]: I1008 18:10:56.823526 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:10:56 crc kubenswrapper[4988]: I1008 18:10:56.823580 4988 kubelet_node_status.go:76] "Attempting to register node" node="crc" Oct 08 18:10:56 crc kubenswrapper[4988]: E1008 18:10:56.824274 4988 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.102:6443: connect: connection refused" node="crc" Oct 08 18:10:57 crc kubenswrapper[4988]: I1008 18:10:57.184024 4988 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.102:6443: connect: connection refused Oct 08 18:10:57 crc kubenswrapper[4988]: I1008 18:10:57.255839 4988 generic.go:334] "Generic (PLEG): container finished" podID="3dcd261975c3d6b9a6ad6367fd4facd3" containerID="6eb46d76cf7cfb1d4ecc4fcee398dad8904732c1664c317865864b65f5703589" exitCode=0 Oct 08 18:10:57 crc kubenswrapper[4988]: I1008 18:10:57.255944 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerDied","Data":"6eb46d76cf7cfb1d4ecc4fcee398dad8904732c1664c317865864b65f5703589"} Oct 08 18:10:57 crc kubenswrapper[4988]: I1008 18:10:57.255993 4988 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 08 18:10:57 crc kubenswrapper[4988]: I1008 18:10:57.257304 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:10:57 crc kubenswrapper[4988]: I1008 18:10:57.257339 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:10:57 crc kubenswrapper[4988]: I1008 18:10:57.257350 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:10:57 crc kubenswrapper[4988]: I1008 18:10:57.259940 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"16f23c4c7163d32dc0df8d6a7ee17e2864fb261f8a40ecf05556774b173fc868"} Oct 08 18:10:57 crc kubenswrapper[4988]: I1008 18:10:57.259985 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"90b1a9e14c82bf0f75f08fdb342eed35a503020f9db641e5e9ceff8b122cb3bd"} Oct 08 18:10:57 crc kubenswrapper[4988]: I1008 18:10:57.260009 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"30bec350fefee3d19758b3ed735ede4e9aa085edb57aa3f3ec24249a67dc7ca2"} Oct 08 18:10:57 crc kubenswrapper[4988]: I1008 18:10:57.260018 4988 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 08 18:10:57 crc kubenswrapper[4988]: I1008 18:10:57.260024 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"a2d850fd27d4ed65891ff69b9add65fb78535357f54819c2a8d196f5a340c58e"} Oct 08 18:10:57 crc kubenswrapper[4988]: I1008 18:10:57.261186 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:10:57 crc kubenswrapper[4988]: I1008 18:10:57.261242 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:10:57 crc kubenswrapper[4988]: I1008 18:10:57.261259 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:10:57 crc kubenswrapper[4988]: I1008 18:10:57.262127 4988 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="967e0b3a3b31296cf65abfe5efad2072d679917927c1f5dd79544aded1818947" exitCode=0 Oct 08 18:10:57 crc kubenswrapper[4988]: I1008 18:10:57.262204 4988 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 08 18:10:57 crc kubenswrapper[4988]: I1008 18:10:57.262197 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"967e0b3a3b31296cf65abfe5efad2072d679917927c1f5dd79544aded1818947"} Oct 08 18:10:57 crc kubenswrapper[4988]: I1008 18:10:57.263609 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:10:57 crc kubenswrapper[4988]: I1008 18:10:57.263708 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:10:57 crc kubenswrapper[4988]: I1008 18:10:57.263727 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:10:57 crc kubenswrapper[4988]: I1008 18:10:57.264964 4988 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="28d21731af0f16d6a8ae6c6e9d0bebf0560d4a533db4c78c10bf1837890394ef" exitCode=0 Oct 08 18:10:57 crc kubenswrapper[4988]: I1008 18:10:57.265061 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"28d21731af0f16d6a8ae6c6e9d0bebf0560d4a533db4c78c10bf1837890394ef"} Oct 08 18:10:57 crc kubenswrapper[4988]: I1008 18:10:57.265253 4988 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 08 18:10:57 crc kubenswrapper[4988]: I1008 18:10:57.266354 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:10:57 crc kubenswrapper[4988]: I1008 18:10:57.266426 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:10:57 crc kubenswrapper[4988]: I1008 18:10:57.266446 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:10:57 crc kubenswrapper[4988]: I1008 18:10:57.268327 4988 generic.go:334] "Generic (PLEG): container finished" podID="d1b160f5dda77d281dd8e69ec8d817f9" containerID="7b7185463f2a962f3214a2524d783a54b8fbc76f385ea27001087bfb7d023561" exitCode=0 Oct 08 18:10:57 crc kubenswrapper[4988]: I1008 18:10:57.268438 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerDied","Data":"7b7185463f2a962f3214a2524d783a54b8fbc76f385ea27001087bfb7d023561"} Oct 08 18:10:57 crc kubenswrapper[4988]: I1008 18:10:57.270045 4988 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 08 18:10:57 crc kubenswrapper[4988]: I1008 18:10:57.270159 4988 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 08 18:10:57 crc kubenswrapper[4988]: I1008 18:10:57.271433 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:10:57 crc kubenswrapper[4988]: I1008 18:10:57.271463 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:10:57 crc kubenswrapper[4988]: I1008 18:10:57.271473 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:10:57 crc kubenswrapper[4988]: I1008 18:10:57.272778 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:10:57 crc kubenswrapper[4988]: I1008 18:10:57.272824 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:10:57 crc kubenswrapper[4988]: I1008 18:10:57.272842 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:10:57 crc kubenswrapper[4988]: W1008 18:10:57.884606 4988 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.102:6443: connect: connection refused Oct 08 18:10:57 crc kubenswrapper[4988]: E1008 18:10:57.884710 4988 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.102:6443: connect: connection refused" logger="UnhandledError" Oct 08 18:10:58 crc kubenswrapper[4988]: W1008 18:10:58.155900 4988 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.102:6443: connect: connection refused Oct 08 18:10:58 crc kubenswrapper[4988]: E1008 18:10:58.156042 4988 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.102:6443: connect: connection refused" logger="UnhandledError" Oct 08 18:10:58 crc kubenswrapper[4988]: I1008 18:10:58.183352 4988 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.102:6443: connect: connection refused Oct 08 18:10:58 crc kubenswrapper[4988]: E1008 18:10:58.192378 4988 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.102:6443: connect: connection refused" interval="3.2s" Oct 08 18:10:58 crc kubenswrapper[4988]: I1008 18:10:58.275246 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"5cb03de583c14ad6751227c21166ac5ac4c6e11b67d5b709361ced8a0877e6d7"} Oct 08 18:10:58 crc kubenswrapper[4988]: I1008 18:10:58.275299 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"75a2bd2757133e79c0c7f2939eb433dc45291f14808e3c3f1daaf90a239c6a71"} Oct 08 18:10:58 crc kubenswrapper[4988]: I1008 18:10:58.275311 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"3fd3b2ca80233400d2798b035a65b833f735632efd76d1160f6bb906e07b80a8"} Oct 08 18:10:58 crc kubenswrapper[4988]: I1008 18:10:58.275410 4988 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 08 18:10:58 crc kubenswrapper[4988]: I1008 18:10:58.278733 4988 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="690ba09c4e63352ae782035c7242352baabb8f477e7d14664470649f65a9acb4" exitCode=0 Oct 08 18:10:58 crc kubenswrapper[4988]: I1008 18:10:58.278813 4988 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 08 18:10:58 crc kubenswrapper[4988]: I1008 18:10:58.278812 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"690ba09c4e63352ae782035c7242352baabb8f477e7d14664470649f65a9acb4"} Oct 08 18:10:58 crc kubenswrapper[4988]: I1008 18:10:58.278927 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:10:58 crc kubenswrapper[4988]: I1008 18:10:58.278964 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:10:58 crc kubenswrapper[4988]: I1008 18:10:58.278975 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:10:58 crc kubenswrapper[4988]: I1008 18:10:58.279663 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:10:58 crc kubenswrapper[4988]: I1008 18:10:58.279736 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:10:58 crc kubenswrapper[4988]: I1008 18:10:58.279751 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:10:58 crc kubenswrapper[4988]: I1008 18:10:58.291186 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"c79ef704fda1bbcc486f4c7afa3039a3ab38669833f0f66bafe24dd4cc679e60"} Oct 08 18:10:58 crc kubenswrapper[4988]: I1008 18:10:58.291240 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"a560ee84c2405227de71fda127566213cd4f017e73d5581442ba267781b1c539"} Oct 08 18:10:58 crc kubenswrapper[4988]: I1008 18:10:58.292032 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"6100accc3bd08928bc1ff88310f29fa5da37fd2d5c315c94279801c9778c4745"} Oct 08 18:10:58 crc kubenswrapper[4988]: I1008 18:10:58.292108 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"77c68f3ecb8a0724499e495cb26ef3db5beecb140f2558d8c0b25f41928b84ba"} Oct 08 18:10:58 crc kubenswrapper[4988]: I1008 18:10:58.303683 4988 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 08 18:10:58 crc kubenswrapper[4988]: I1008 18:10:58.304303 4988 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 08 18:10:58 crc kubenswrapper[4988]: I1008 18:10:58.304621 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"adbdfab7dbc007ce91c26fa7a6535ef27949ad2567f1bc6b4c52b6a3f62789f8"} Oct 08 18:10:58 crc kubenswrapper[4988]: I1008 18:10:58.305152 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:10:58 crc kubenswrapper[4988]: I1008 18:10:58.305219 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:10:58 crc kubenswrapper[4988]: I1008 18:10:58.305237 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:10:58 crc kubenswrapper[4988]: I1008 18:10:58.306235 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:10:58 crc kubenswrapper[4988]: I1008 18:10:58.306266 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:10:58 crc kubenswrapper[4988]: I1008 18:10:58.306276 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:10:58 crc kubenswrapper[4988]: I1008 18:10:58.425170 4988 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 08 18:10:58 crc kubenswrapper[4988]: I1008 18:10:58.427470 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:10:58 crc kubenswrapper[4988]: I1008 18:10:58.427534 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:10:58 crc kubenswrapper[4988]: I1008 18:10:58.427556 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:10:58 crc kubenswrapper[4988]: I1008 18:10:58.427598 4988 kubelet_node_status.go:76] "Attempting to register node" node="crc" Oct 08 18:10:58 crc kubenswrapper[4988]: E1008 18:10:58.428240 4988 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.102:6443: connect: connection refused" node="crc" Oct 08 18:10:59 crc kubenswrapper[4988]: I1008 18:10:59.020225 4988 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 08 18:10:59 crc kubenswrapper[4988]: I1008 18:10:59.029139 4988 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 08 18:10:59 crc kubenswrapper[4988]: I1008 18:10:59.310816 4988 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="e94b9f46109c961b4dcb9d1ecf9801f59d8b32d04674ebba2e15cd89156d8ca1" exitCode=0 Oct 08 18:10:59 crc kubenswrapper[4988]: I1008 18:10:59.310940 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"e94b9f46109c961b4dcb9d1ecf9801f59d8b32d04674ebba2e15cd89156d8ca1"} Oct 08 18:10:59 crc kubenswrapper[4988]: I1008 18:10:59.311056 4988 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 08 18:10:59 crc kubenswrapper[4988]: I1008 18:10:59.312814 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:10:59 crc kubenswrapper[4988]: I1008 18:10:59.312876 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:10:59 crc kubenswrapper[4988]: I1008 18:10:59.312901 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:10:59 crc kubenswrapper[4988]: I1008 18:10:59.317060 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"5bd35aef24be034a0b305a40bb53ca171a4f76db3b93b4f78ae95099b0a4cde3"} Oct 08 18:10:59 crc kubenswrapper[4988]: I1008 18:10:59.317158 4988 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 08 18:10:59 crc kubenswrapper[4988]: I1008 18:10:59.317208 4988 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 08 18:10:59 crc kubenswrapper[4988]: I1008 18:10:59.317213 4988 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 08 18:10:59 crc kubenswrapper[4988]: I1008 18:10:59.317377 4988 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 08 18:10:59 crc kubenswrapper[4988]: I1008 18:10:59.317373 4988 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 08 18:10:59 crc kubenswrapper[4988]: I1008 18:10:59.317359 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 08 18:10:59 crc kubenswrapper[4988]: I1008 18:10:59.318754 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:10:59 crc kubenswrapper[4988]: I1008 18:10:59.318793 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:10:59 crc kubenswrapper[4988]: I1008 18:10:59.318804 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:10:59 crc kubenswrapper[4988]: I1008 18:10:59.318844 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:10:59 crc kubenswrapper[4988]: I1008 18:10:59.318892 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:10:59 crc kubenswrapper[4988]: I1008 18:10:59.318913 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:10:59 crc kubenswrapper[4988]: I1008 18:10:59.319928 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:10:59 crc kubenswrapper[4988]: I1008 18:10:59.319946 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:10:59 crc kubenswrapper[4988]: I1008 18:10:59.319982 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:10:59 crc kubenswrapper[4988]: I1008 18:10:59.319993 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:10:59 crc kubenswrapper[4988]: I1008 18:10:59.320026 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:10:59 crc kubenswrapper[4988]: I1008 18:10:59.319996 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:11:00 crc kubenswrapper[4988]: I1008 18:11:00.204120 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 08 18:11:00 crc kubenswrapper[4988]: I1008 18:11:00.324299 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"40565a22db5378b968d0c8ee38663a7d3f869d211dc83f584c7d9fbf90395e8f"} Oct 08 18:11:00 crc kubenswrapper[4988]: I1008 18:11:00.324345 4988 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 08 18:11:00 crc kubenswrapper[4988]: I1008 18:11:00.324416 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"c28467fbb44e38bc14ae11dcba70d755966e677fa6f00763487076977c510fbf"} Oct 08 18:11:00 crc kubenswrapper[4988]: I1008 18:11:00.324464 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"66b4d85b77880c3b6e7270bb8a5f6cee54878f864d66020112b43e5678243f35"} Oct 08 18:11:00 crc kubenswrapper[4988]: I1008 18:11:00.324501 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"a8af097fb13c9d04954647ef80a1724e40c742a1af21ec6fdbc6a943fb496919"} Oct 08 18:11:00 crc kubenswrapper[4988]: I1008 18:11:00.324409 4988 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 08 18:11:00 crc kubenswrapper[4988]: I1008 18:11:00.324622 4988 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 08 18:11:00 crc kubenswrapper[4988]: I1008 18:11:00.325266 4988 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 08 18:11:00 crc kubenswrapper[4988]: I1008 18:11:00.325638 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:11:00 crc kubenswrapper[4988]: I1008 18:11:00.325699 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:11:00 crc kubenswrapper[4988]: I1008 18:11:00.325717 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:11:00 crc kubenswrapper[4988]: I1008 18:11:00.326238 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:11:00 crc kubenswrapper[4988]: I1008 18:11:00.326280 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:11:00 crc kubenswrapper[4988]: I1008 18:11:00.326294 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:11:00 crc kubenswrapper[4988]: I1008 18:11:00.327154 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:11:00 crc kubenswrapper[4988]: I1008 18:11:00.327198 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:11:00 crc kubenswrapper[4988]: I1008 18:11:00.327213 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:11:00 crc kubenswrapper[4988]: I1008 18:11:00.982147 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 08 18:11:01 crc kubenswrapper[4988]: I1008 18:11:01.101200 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 08 18:11:01 crc kubenswrapper[4988]: I1008 18:11:01.331014 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"ebd72de8c7c4cf26cbe8dee49c35eae93e20f20d4a5fcb30adbbb001a6037355"} Oct 08 18:11:01 crc kubenswrapper[4988]: I1008 18:11:01.331108 4988 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 08 18:11:01 crc kubenswrapper[4988]: I1008 18:11:01.331203 4988 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 08 18:11:01 crc kubenswrapper[4988]: I1008 18:11:01.332152 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:11:01 crc kubenswrapper[4988]: I1008 18:11:01.332177 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:11:01 crc kubenswrapper[4988]: I1008 18:11:01.332191 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:11:01 crc kubenswrapper[4988]: I1008 18:11:01.332427 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:11:01 crc kubenswrapper[4988]: I1008 18:11:01.332480 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:11:01 crc kubenswrapper[4988]: I1008 18:11:01.332491 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:11:01 crc kubenswrapper[4988]: I1008 18:11:01.628378 4988 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 08 18:11:01 crc kubenswrapper[4988]: I1008 18:11:01.630519 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:11:01 crc kubenswrapper[4988]: I1008 18:11:01.630592 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:11:01 crc kubenswrapper[4988]: I1008 18:11:01.630613 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:11:01 crc kubenswrapper[4988]: I1008 18:11:01.630662 4988 kubelet_node_status.go:76] "Attempting to register node" node="crc" Oct 08 18:11:01 crc kubenswrapper[4988]: I1008 18:11:01.638701 4988 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 08 18:11:01 crc kubenswrapper[4988]: I1008 18:11:01.638992 4988 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 08 18:11:01 crc kubenswrapper[4988]: I1008 18:11:01.640643 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:11:01 crc kubenswrapper[4988]: I1008 18:11:01.640706 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:11:01 crc kubenswrapper[4988]: I1008 18:11:01.640733 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:11:02 crc kubenswrapper[4988]: I1008 18:11:02.334331 4988 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 08 18:11:02 crc kubenswrapper[4988]: I1008 18:11:02.334355 4988 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 08 18:11:02 crc kubenswrapper[4988]: I1008 18:11:02.337017 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:11:02 crc kubenswrapper[4988]: I1008 18:11:02.337084 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:11:02 crc kubenswrapper[4988]: I1008 18:11:02.337105 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:11:02 crc kubenswrapper[4988]: I1008 18:11:02.337212 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:11:02 crc kubenswrapper[4988]: I1008 18:11:02.337262 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:11:02 crc kubenswrapper[4988]: I1008 18:11:02.337281 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:11:02 crc kubenswrapper[4988]: I1008 18:11:02.690884 4988 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-etcd/etcd-crc" Oct 08 18:11:02 crc kubenswrapper[4988]: I1008 18:11:02.965003 4988 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 08 18:11:03 crc kubenswrapper[4988]: I1008 18:11:03.338569 4988 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 08 18:11:03 crc kubenswrapper[4988]: I1008 18:11:03.338637 4988 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 08 18:11:03 crc kubenswrapper[4988]: I1008 18:11:03.340175 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:11:03 crc kubenswrapper[4988]: I1008 18:11:03.340222 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:11:03 crc kubenswrapper[4988]: I1008 18:11:03.340239 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:11:03 crc kubenswrapper[4988]: I1008 18:11:03.340491 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:11:03 crc kubenswrapper[4988]: I1008 18:11:03.340538 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:11:03 crc kubenswrapper[4988]: I1008 18:11:03.340556 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:11:04 crc kubenswrapper[4988]: I1008 18:11:04.557469 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 08 18:11:04 crc kubenswrapper[4988]: I1008 18:11:04.557722 4988 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 08 18:11:04 crc kubenswrapper[4988]: I1008 18:11:04.559026 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:11:04 crc kubenswrapper[4988]: I1008 18:11:04.559062 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:11:04 crc kubenswrapper[4988]: I1008 18:11:04.559072 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:11:04 crc kubenswrapper[4988]: I1008 18:11:04.639074 4988 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Oct 08 18:11:04 crc kubenswrapper[4988]: I1008 18:11:04.639156 4988 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Oct 08 18:11:04 crc kubenswrapper[4988]: I1008 18:11:04.828437 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-etcd/etcd-crc" Oct 08 18:11:04 crc kubenswrapper[4988]: I1008 18:11:04.828714 4988 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 08 18:11:04 crc kubenswrapper[4988]: I1008 18:11:04.830584 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:11:04 crc kubenswrapper[4988]: I1008 18:11:04.830618 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:11:04 crc kubenswrapper[4988]: I1008 18:11:04.830629 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:11:05 crc kubenswrapper[4988]: E1008 18:11:05.318156 4988 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Oct 08 18:11:07 crc kubenswrapper[4988]: I1008 18:11:07.122571 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 08 18:11:07 crc kubenswrapper[4988]: I1008 18:11:07.122721 4988 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 08 18:11:07 crc kubenswrapper[4988]: I1008 18:11:07.124124 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:11:07 crc kubenswrapper[4988]: I1008 18:11:07.124178 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:11:07 crc kubenswrapper[4988]: I1008 18:11:07.124193 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:11:09 crc kubenswrapper[4988]: I1008 18:11:09.184337 4988 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": net/http: TLS handshake timeout Oct 08 18:11:09 crc kubenswrapper[4988]: E1008 18:11:09.186594 4988 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": net/http: TLS handshake timeout" event="&Event{ObjectMeta:{crc.186c9685a76dd185 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-10-08 18:10:55.178830213 +0000 UTC m=+0.628672993,LastTimestamp:2025-10-08 18:10:55.178830213 +0000 UTC m=+0.628672993,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Oct 08 18:11:09 crc kubenswrapper[4988]: W1008 18:11:09.211017 4988 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": net/http: TLS handshake timeout Oct 08 18:11:09 crc kubenswrapper[4988]: I1008 18:11:09.211246 4988 trace.go:236] Trace[2060990686]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (08-Oct-2025 18:10:59.209) (total time: 10001ms): Oct 08 18:11:09 crc kubenswrapper[4988]: Trace[2060990686]: ---"Objects listed" error:Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": net/http: TLS handshake timeout 10001ms (18:11:09.211) Oct 08 18:11:09 crc kubenswrapper[4988]: Trace[2060990686]: [10.001736212s] [10.001736212s] END Oct 08 18:11:09 crc kubenswrapper[4988]: E1008 18:11:09.211556 4988 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": net/http: TLS handshake timeout" logger="UnhandledError" Oct 08 18:11:09 crc kubenswrapper[4988]: W1008 18:11:09.410628 4988 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": net/http: TLS handshake timeout Oct 08 18:11:09 crc kubenswrapper[4988]: I1008 18:11:09.410830 4988 trace.go:236] Trace[2100699147]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (08-Oct-2025 18:10:59.409) (total time: 10001ms): Oct 08 18:11:09 crc kubenswrapper[4988]: Trace[2100699147]: ---"Objects listed" error:Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": net/http: TLS handshake timeout 10001ms (18:11:09.410) Oct 08 18:11:09 crc kubenswrapper[4988]: Trace[2100699147]: [10.001627804s] [10.001627804s] END Oct 08 18:11:09 crc kubenswrapper[4988]: E1008 18:11:09.410880 4988 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": net/http: TLS handshake timeout" logger="UnhandledError" Oct 08 18:11:09 crc kubenswrapper[4988]: I1008 18:11:09.602304 4988 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Oct 08 18:11:09 crc kubenswrapper[4988]: I1008 18:11:09.602402 4988 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Oct 08 18:11:09 crc kubenswrapper[4988]: I1008 18:11:09.621230 4988 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Oct 08 18:11:09 crc kubenswrapper[4988]: I1008 18:11:09.621315 4988 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Oct 08 18:11:12 crc kubenswrapper[4988]: I1008 18:11:12.727054 4988 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-etcd/etcd-crc" Oct 08 18:11:12 crc kubenswrapper[4988]: I1008 18:11:12.727433 4988 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 08 18:11:12 crc kubenswrapper[4988]: I1008 18:11:12.729244 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:11:12 crc kubenswrapper[4988]: I1008 18:11:12.729292 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:11:12 crc kubenswrapper[4988]: I1008 18:11:12.729304 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:11:12 crc kubenswrapper[4988]: I1008 18:11:12.746991 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-etcd/etcd-crc" Oct 08 18:11:12 crc kubenswrapper[4988]: I1008 18:11:12.971573 4988 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 08 18:11:12 crc kubenswrapper[4988]: I1008 18:11:12.971748 4988 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 08 18:11:12 crc kubenswrapper[4988]: I1008 18:11:12.973628 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:11:12 crc kubenswrapper[4988]: I1008 18:11:12.973709 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:11:12 crc kubenswrapper[4988]: I1008 18:11:12.973740 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:11:12 crc kubenswrapper[4988]: I1008 18:11:12.978184 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 08 18:11:13 crc kubenswrapper[4988]: I1008 18:11:13.367800 4988 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 08 18:11:13 crc kubenswrapper[4988]: I1008 18:11:13.367910 4988 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 08 18:11:13 crc kubenswrapper[4988]: I1008 18:11:13.367829 4988 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 08 18:11:13 crc kubenswrapper[4988]: I1008 18:11:13.369720 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:11:13 crc kubenswrapper[4988]: I1008 18:11:13.369777 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:11:13 crc kubenswrapper[4988]: I1008 18:11:13.369799 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:11:13 crc kubenswrapper[4988]: I1008 18:11:13.370054 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:11:13 crc kubenswrapper[4988]: I1008 18:11:13.370144 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:11:13 crc kubenswrapper[4988]: I1008 18:11:13.370170 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:11:13 crc kubenswrapper[4988]: I1008 18:11:13.623955 4988 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Oct 08 18:11:14 crc kubenswrapper[4988]: E1008 18:11:14.601371 4988 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": context deadline exceeded" interval="6.4s" Oct 08 18:11:14 crc kubenswrapper[4988]: I1008 18:11:14.604045 4988 trace.go:236] Trace[1707469102]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (08-Oct-2025 18:11:03.878) (total time: 10725ms): Oct 08 18:11:14 crc kubenswrapper[4988]: Trace[1707469102]: ---"Objects listed" error: 10725ms (18:11:14.603) Oct 08 18:11:14 crc kubenswrapper[4988]: Trace[1707469102]: [10.725080461s] [10.725080461s] END Oct 08 18:11:14 crc kubenswrapper[4988]: I1008 18:11:14.604288 4988 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Oct 08 18:11:14 crc kubenswrapper[4988]: I1008 18:11:14.604628 4988 trace.go:236] Trace[1970404420]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (08-Oct-2025 18:11:03.452) (total time: 11151ms): Oct 08 18:11:14 crc kubenswrapper[4988]: Trace[1970404420]: ---"Objects listed" error: 11151ms (18:11:14.604) Oct 08 18:11:14 crc kubenswrapper[4988]: Trace[1970404420]: [11.151896407s] [11.151896407s] END Oct 08 18:11:14 crc kubenswrapper[4988]: I1008 18:11:14.604679 4988 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Oct 08 18:11:14 crc kubenswrapper[4988]: I1008 18:11:14.606466 4988 reconstruct.go:205] "DevicePaths of reconstructed volumes updated" Oct 08 18:11:14 crc kubenswrapper[4988]: E1008 18:11:14.609297 4988 kubelet_node_status.go:99] "Unable to register node with API server" err="nodes \"crc\" is forbidden: autoscaling.openshift.io/ManagedNode infra config cache not synchronized" node="crc" Oct 08 18:11:14 crc kubenswrapper[4988]: I1008 18:11:14.639677 4988 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Oct 08 18:11:14 crc kubenswrapper[4988]: I1008 18:11:14.639861 4988 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Oct 08 18:11:14 crc kubenswrapper[4988]: I1008 18:11:14.655157 4988 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Readiness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": read tcp 192.168.126.11:48014->192.168.126.11:17697: read: connection reset by peer" start-of-body= Oct 08 18:11:14 crc kubenswrapper[4988]: I1008 18:11:14.655258 4988 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": read tcp 192.168.126.11:48014->192.168.126.11:17697: read: connection reset by peer" Oct 08 18:11:14 crc kubenswrapper[4988]: I1008 18:11:14.655368 4988 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Liveness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": read tcp 192.168.126.11:37318->192.168.126.11:17697: read: connection reset by peer" start-of-body= Oct 08 18:11:14 crc kubenswrapper[4988]: I1008 18:11:14.655569 4988 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": read tcp 192.168.126.11:37318->192.168.126.11:17697: read: connection reset by peer" Oct 08 18:11:14 crc kubenswrapper[4988]: I1008 18:11:14.655667 4988 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Readiness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" start-of-body= Oct 08 18:11:14 crc kubenswrapper[4988]: I1008 18:11:14.655693 4988 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.177257 4988 apiserver.go:52] "Watching apiserver" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.192555 4988 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66 Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.193130 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-network-node-identity/network-node-identity-vrzqb","openshift-network-operator/iptables-alerter-4ln5h","openshift-network-operator/network-operator-58b4c7f79c-55gtf","openshift-network-console/networking-console-plugin-85b44fc459-gdk6g","openshift-network-diagnostics/network-check-source-55646444c4-trplf","openshift-network-diagnostics/network-check-target-xd92c"] Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.193803 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.194019 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 18:11:15 crc kubenswrapper[4988]: E1008 18:11:15.194211 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.194677 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.194855 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 08 18:11:15 crc kubenswrapper[4988]: E1008 18:11:15.195071 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.194985 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.195331 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 18:11:15 crc kubenswrapper[4988]: E1008 18:11:15.195565 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.196948 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.197044 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.197754 4988 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.197779 4988 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.197908 4988 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.198256 4988 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.199573 4988 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.199700 4988 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.201178 4988 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.236272 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.261636 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.277812 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.284478 4988 desired_state_of_world_populator.go:154] "Finished populating initial desired state of world" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.293060 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.308108 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.311665 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.311726 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.311751 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.311776 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.311797 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.311820 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.311840 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.311866 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.311887 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.311907 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.311934 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.311958 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.311979 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.312003 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.312025 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.312045 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.312131 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.312161 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.312183 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.312487 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.312670 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.312711 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.312528 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" (OuterVolumeSpecName: "kube-api-access-kfwg7") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "kube-api-access-kfwg7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.312840 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" (OuterVolumeSpecName: "control-plane-machine-set-operator-tls") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "control-plane-machine-set-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.313005 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" (OuterVolumeSpecName: "kube-api-access-sb6h7") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "kube-api-access-sb6h7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.313166 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.313321 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" (OuterVolumeSpecName: "kube-api-access-nzwt7") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "kube-api-access-nzwt7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.313500 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.314127 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") pod \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\" (UID: \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\") " Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.313497 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.314169 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.314211 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.313511 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.313528 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" (OuterVolumeSpecName: "config-volume") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.313131 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" (OuterVolumeSpecName: "kube-api-access-9xfj7") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "kube-api-access-9xfj7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.313688 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" (OuterVolumeSpecName: "config") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.313660 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" (OuterVolumeSpecName: "config") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.313785 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.313816 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" (OuterVolumeSpecName: "kube-api-access-d6qdx") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "kube-api-access-d6qdx". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.313881 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.313983 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" (OuterVolumeSpecName: "etcd-service-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.314243 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.314563 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.314013 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" (OuterVolumeSpecName: "config") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.314617 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.314142 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.314210 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.314484 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" (OuterVolumeSpecName: "kube-api-access-6ccd8") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "kube-api-access-6ccd8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.314506 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" (OuterVolumeSpecName: "kube-api-access-x2m85") pod "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" (UID: "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d"). InnerVolumeSpecName "kube-api-access-x2m85". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.314588 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" (OuterVolumeSpecName: "kube-api-access-jkwtn") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "kube-api-access-jkwtn". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.314665 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.314782 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.314903 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.314935 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" (OuterVolumeSpecName: "kube-api-access-2w9zh") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "kube-api-access-2w9zh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.314946 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.314989 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.315041 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.315029 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 18:11:15 crc kubenswrapper[4988]: E1008 18:11:15.315102 4988 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 18:11:15.815069311 +0000 UTC m=+21.264912081 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.315067 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.315083 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.315150 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.315184 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.316262 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.315216 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.316630 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.316673 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.316711 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.316741 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.316784 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.316833 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.316756 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" (OuterVolumeSpecName: "kube-api-access-4d4hj") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "kube-api-access-4d4hj". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.316804 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" (OuterVolumeSpecName: "utilities") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.317185 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" (OuterVolumeSpecName: "kube-api-access-7c4vf") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "kube-api-access-7c4vf". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.316814 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.316824 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.316892 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" (OuterVolumeSpecName: "utilities") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.316867 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.318139 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.318205 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.318267 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.318325 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.318373 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.318506 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.318617 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.318608 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" (OuterVolumeSpecName: "client-ca") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.318704 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.318801 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.318623 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.319013 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" (OuterVolumeSpecName: "webhook-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "webhook-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.319003 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" (OuterVolumeSpecName: "cert") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.319062 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" (OuterVolumeSpecName: "kube-api-access-mnrrd") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "kube-api-access-mnrrd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.319028 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.319281 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.319363 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" (OuterVolumeSpecName: "kube-api-access-d4lsv") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "kube-api-access-d4lsv". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.319432 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.319438 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.319611 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.319691 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.319736 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.319796 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.319862 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.319920 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.319979 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.320032 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.320095 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.320149 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.320199 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.320259 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.320300 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.320348 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.320425 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.320494 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.320550 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.320610 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.320674 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.320730 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.320793 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.320842 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.320889 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.320944 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.320997 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.321049 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.321098 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.321143 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.321199 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.321244 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.321299 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.321340 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.321412 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.321461 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.321510 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.321564 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.322172 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.322309 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.319862 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" (OuterVolumeSpecName: "kube-api-access-lz9wn") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "kube-api-access-lz9wn". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.320011 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" (OuterVolumeSpecName: "node-bootstrap-token") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "node-bootstrap-token". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.319864 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" (OuterVolumeSpecName: "kube-api-access-rnphk") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "kube-api-access-rnphk". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.320078 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.320110 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" (OuterVolumeSpecName: "config") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.320365 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" (OuterVolumeSpecName: "certs") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.320367 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.323080 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" (OuterVolumeSpecName: "machine-api-operator-tls") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "machine-api-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.323243 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.323433 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" (OuterVolumeSpecName: "multus-daemon-config") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "multus-daemon-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.323446 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" (OuterVolumeSpecName: "service-ca") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.320582 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" (OuterVolumeSpecName: "client-ca") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.320857 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.321263 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" (OuterVolumeSpecName: "config") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.321280 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.321446 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" (OuterVolumeSpecName: "kube-api-access-xcphl") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "kube-api-access-xcphl". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.321941 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" (OuterVolumeSpecName: "kube-api-access-qs4fp") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "kube-api-access-qs4fp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.324014 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.322144 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" (OuterVolumeSpecName: "kube-api-access-fqsjt") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "kube-api-access-fqsjt". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.324029 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.322836 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.322911 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" (OuterVolumeSpecName: "kube-api-access-w9rds") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "kube-api-access-w9rds". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.323694 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" (OuterVolumeSpecName: "utilities") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.323898 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" (OuterVolumeSpecName: "kube-api-access-6g6sz") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "kube-api-access-6g6sz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.323996 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.324176 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.324193 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" (OuterVolumeSpecName: "kube-api-access-bf2bz") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "kube-api-access-bf2bz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.324246 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.324316 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.324542 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" (OuterVolumeSpecName: "kube-api-access-htfz6") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "kube-api-access-htfz6". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.324684 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.324815 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" (OuterVolumeSpecName: "kube-api-access-wxkg8") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "kube-api-access-wxkg8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.324792 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" (OuterVolumeSpecName: "mcd-auth-proxy-config") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "mcd-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.324995 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" (OuterVolumeSpecName: "kube-api-access-fcqwp") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "kube-api-access-fcqwp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.325185 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.322730 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.325419 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" (OuterVolumeSpecName: "tmpfs") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "tmpfs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.325444 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.325531 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.325629 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.325855 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" (OuterVolumeSpecName: "kube-api-access-zkvpv") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "kube-api-access-zkvpv". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.326041 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.325760 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.326202 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.326248 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.326290 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.326292 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.326332 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.326544 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.326072 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.326741 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" (OuterVolumeSpecName: "kube-api-access-mg5zb") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "kube-api-access-mg5zb". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.326741 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.326706 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.326733 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.326779 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" (OuterVolumeSpecName: "config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.326818 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" (OuterVolumeSpecName: "config") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.327290 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.328187 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" (OuterVolumeSpecName: "kube-api-access-2d4wz") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "kube-api-access-2d4wz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.328503 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" (OuterVolumeSpecName: "image-registry-operator-tls") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "image-registry-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.328795 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.328980 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" (OuterVolumeSpecName: "kube-api-access-xcgwh") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "kube-api-access-xcgwh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.328996 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.329064 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.329106 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.329140 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.329172 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.329202 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.329231 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.329469 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.329081 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" (OuterVolumeSpecName: "etcd-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.329618 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.329709 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.329829 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.329197 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" (OuterVolumeSpecName: "webhook-certs") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "webhook-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.329907 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.330105 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.330157 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.330236 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.329457 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" (OuterVolumeSpecName: "config") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.330311 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.330535 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.330612 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.330674 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.330744 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" (OuterVolumeSpecName: "service-ca") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.329399 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" (OuterVolumeSpecName: "images") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.331031 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" (OuterVolumeSpecName: "signing-key") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.331051 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.331061 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.331138 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.331175 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.331243 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.331505 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" (OuterVolumeSpecName: "kube-api-access-v47cf") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "kube-api-access-v47cf". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.332127 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.332331 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.332368 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.332415 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.332311 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.332449 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.332536 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.332566 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.332597 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.332626 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.332627 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.332650 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.332996 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.333151 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" (OuterVolumeSpecName: "images") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.333315 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" (OuterVolumeSpecName: "kube-api-access-pj782") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "kube-api-access-pj782". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.333377 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.333370 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" (OuterVolumeSpecName: "signing-cabundle") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-cabundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.333583 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.333599 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" (OuterVolumeSpecName: "kube-api-access-cfbct") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "kube-api-access-cfbct". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.332894 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.333813 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.334186 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.334262 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.334310 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.334353 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.334447 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.334489 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.334528 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") pod \"44663579-783b-4372-86d6-acf235a62d72\" (UID: \"44663579-783b-4372-86d6-acf235a62d72\") " Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.334563 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.334599 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.334635 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.334675 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.334704 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.334717 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.334756 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.334797 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.334836 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.334875 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.334876 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.334913 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.334951 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.334987 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.335096 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.335134 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.335173 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.335210 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.335246 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.335283 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.335329 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.335373 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.335435 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.335472 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.335508 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.335544 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.335581 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.335620 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.335654 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.335693 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.335729 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") pod \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\" (UID: \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\") " Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.335776 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.335822 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.335856 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.335892 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.335926 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.335967 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.336002 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.336037 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.336043 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.336078 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") pod \"49ef4625-1d3a-4a9f-b595-c2433d32326d\" (UID: \"49ef4625-1d3a-4a9f-b595-c2433d32326d\") " Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.336120 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.336157 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.336194 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.336238 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.336698 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.336740 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.336791 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.336827 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.336870 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.336905 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.336941 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.336979 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.337015 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.337061 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.337098 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.337136 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.337170 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.337204 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.337291 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.337340 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.337409 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.337465 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.337515 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.337561 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.337605 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.337653 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.337697 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.337741 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.337780 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.337818 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.337854 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.337899 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.338070 4988 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") on node \"crc\" DevicePath \"\"" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.338095 4988 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") on node \"crc\" DevicePath \"\"" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.338115 4988 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.338137 4988 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") on node \"crc\" DevicePath \"\"" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.338159 4988 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") on node \"crc\" DevicePath \"\"" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.338181 4988 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") on node \"crc\" DevicePath \"\"" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.338202 4988 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") on node \"crc\" DevicePath \"\"" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.338223 4988 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") on node \"crc\" DevicePath \"\"" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.338243 4988 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.338262 4988 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") on node \"crc\" DevicePath \"\"" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.338282 4988 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") on node \"crc\" DevicePath \"\"" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.338302 4988 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") on node \"crc\" DevicePath \"\"" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.338320 4988 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") on node \"crc\" DevicePath \"\"" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.338339 4988 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") on node \"crc\" DevicePath \"\"" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.338358 4988 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.338378 4988 reconciler_common.go:293] "Volume detached for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") on node \"crc\" DevicePath \"\"" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.338433 4988 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") on node \"crc\" DevicePath \"\"" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.338455 4988 reconciler_common.go:293] "Volume detached for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") on node \"crc\" DevicePath \"\"" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.338475 4988 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") on node \"crc\" DevicePath \"\"" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.338495 4988 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") on node \"crc\" DevicePath \"\"" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.336056 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.336285 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.336421 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.336530 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.336881 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" (OuterVolumeSpecName: "kube-api-access-w7l8j") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "kube-api-access-w7l8j". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.337006 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" (OuterVolumeSpecName: "config") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.337559 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.337934 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" (OuterVolumeSpecName: "kube-api-access-w4xd4") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "kube-api-access-w4xd4". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.338039 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.338171 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" (OuterVolumeSpecName: "apiservice-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "apiservice-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.338261 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" (OuterVolumeSpecName: "config") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.338685 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" (OuterVolumeSpecName: "kube-api-access-zgdk5") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "kube-api-access-zgdk5". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.338845 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.338968 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.339261 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" (OuterVolumeSpecName: "package-server-manager-serving-cert") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "package-server-manager-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.338514 4988 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") on node \"crc\" DevicePath \"\"" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.339305 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" (OuterVolumeSpecName: "kube-api-access-pjr6v") pod "49ef4625-1d3a-4a9f-b595-c2433d32326d" (UID: "49ef4625-1d3a-4a9f-b595-c2433d32326d"). InnerVolumeSpecName "kube-api-access-pjr6v". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.339354 4988 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.339376 4988 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") on node \"crc\" DevicePath \"\"" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.339416 4988 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.339438 4988 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") on node \"crc\" DevicePath \"\"" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.339461 4988 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") on node \"crc\" DevicePath \"\"" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.339457 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" (OuterVolumeSpecName: "kube-api-access-x7zkh") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "kube-api-access-x7zkh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.339478 4988 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.339528 4988 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") on node \"crc\" DevicePath \"\"" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.339555 4988 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") on node \"crc\" DevicePath \"\"" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.339577 4988 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.339584 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.339598 4988 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") on node \"crc\" DevicePath \"\"" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.339620 4988 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") on node \"crc\" DevicePath \"\"" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.339641 4988 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.339722 4988 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") on node \"crc\" DevicePath \"\"" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.339750 4988 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") on node \"crc\" DevicePath \"\"" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.339773 4988 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") on node \"crc\" DevicePath \"\"" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.339797 4988 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.339804 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" (OuterVolumeSpecName: "utilities") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.339822 4988 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") on node \"crc\" DevicePath \"\"" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.339847 4988 reconciler_common.go:293] "Volume detached for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") on node \"crc\" DevicePath \"\"" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.339869 4988 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") on node \"crc\" DevicePath \"\"" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.339891 4988 reconciler_common.go:293] "Volume detached for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") on node \"crc\" DevicePath \"\"" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.339912 4988 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") on node \"crc\" DevicePath \"\"" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.339936 4988 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") on node \"crc\" DevicePath \"\"" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.339959 4988 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") on node \"crc\" DevicePath \"\"" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.339980 4988 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.340003 4988 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") on node \"crc\" DevicePath \"\"" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.340026 4988 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") on node \"crc\" DevicePath \"\"" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.340050 4988 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") on node \"crc\" DevicePath \"\"" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.340073 4988 reconciler_common.go:293] "Volume detached for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") on node \"crc\" DevicePath \"\"" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.340097 4988 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.340123 4988 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.340150 4988 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") on node \"crc\" DevicePath \"\"" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.340173 4988 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.340194 4988 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") on node \"crc\" DevicePath \"\"" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.340216 4988 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") on node \"crc\" DevicePath \"\"" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.340239 4988 reconciler_common.go:293] "Volume detached for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") on node \"crc\" DevicePath \"\"" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.340261 4988 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.340283 4988 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") on node \"crc\" DevicePath \"\"" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.340304 4988 reconciler_common.go:293] "Volume detached for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") on node \"crc\" DevicePath \"\"" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.340323 4988 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") on node \"crc\" DevicePath \"\"" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.340342 4988 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.340363 4988 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") on node \"crc\" DevicePath \"\"" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.340381 4988 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.340430 4988 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") on node \"crc\" DevicePath \"\"" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.340452 4988 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") on node \"crc\" DevicePath \"\"" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.340471 4988 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.340555 4988 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.340628 4988 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") on node \"crc\" DevicePath \"\"" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.340648 4988 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") on node \"crc\" DevicePath \"\"" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.340669 4988 reconciler_common.go:293] "Volume detached for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") on node \"crc\" DevicePath \"\"" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.340688 4988 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") on node \"crc\" DevicePath \"\"" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.340710 4988 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") on node \"crc\" DevicePath \"\"" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.340729 4988 reconciler_common.go:293] "Volume detached for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.340749 4988 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") on node \"crc\" DevicePath \"\"" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.340768 4988 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.340788 4988 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") on node \"crc\" DevicePath \"\"" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.340806 4988 reconciler_common.go:293] "Volume detached for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") on node \"crc\" DevicePath \"\"" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.340825 4988 reconciler_common.go:293] "Volume detached for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") on node \"crc\" DevicePath \"\"" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.340843 4988 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") on node \"crc\" DevicePath \"\"" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.340862 4988 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.340881 4988 reconciler_common.go:293] "Volume detached for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") on node \"crc\" DevicePath \"\"" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.340901 4988 reconciler_common.go:293] "Volume detached for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") on node \"crc\" DevicePath \"\"" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.340920 4988 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.340939 4988 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.340956 4988 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") on node \"crc\" DevicePath \"\"" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.340974 4988 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") on node \"crc\" DevicePath \"\"" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.340993 4988 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") on node \"crc\" DevicePath \"\"" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.341014 4988 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") on node \"crc\" DevicePath \"\"" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.341038 4988 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") on node \"crc\" DevicePath \"\"" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.341059 4988 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") on node \"crc\" DevicePath \"\"" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.341078 4988 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.341096 4988 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.341116 4988 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") on node \"crc\" DevicePath \"\"" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.341135 4988 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") on node \"crc\" DevicePath \"\"" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.341155 4988 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") on node \"crc\" DevicePath \"\"" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.341174 4988 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") on node \"crc\" DevicePath \"\"" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.341192 4988 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") on node \"crc\" DevicePath \"\"" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.341212 4988 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.341236 4988 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.341254 4988 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") on node \"crc\" DevicePath \"\"" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.341273 4988 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") on node \"crc\" DevicePath \"\"" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.341292 4988 reconciler_common.go:293] "Volume detached for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") on node \"crc\" DevicePath \"\"" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.341313 4988 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") on node \"crc\" DevicePath \"\"" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.341332 4988 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.341351 4988 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") on node \"crc\" DevicePath \"\"" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.341371 4988 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.341411 4988 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") on node \"crc\" DevicePath \"\"" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.341429 4988 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.341449 4988 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") on node \"crc\" DevicePath \"\"" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.341470 4988 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") on node \"crc\" DevicePath \"\"" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.341488 4988 reconciler_common.go:293] "Volume detached for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") on node \"crc\" DevicePath \"\"" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.341507 4988 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.341525 4988 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") on node \"crc\" DevicePath \"\"" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.341545 4988 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") on node \"crc\" DevicePath \"\"" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.341563 4988 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") on node \"crc\" DevicePath \"\"" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.341580 4988 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") on node \"crc\" DevicePath \"\"" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.341599 4988 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.339906 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.340084 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" (OuterVolumeSpecName: "kube-api-access-qg5z5") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "kube-api-access-qg5z5". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.340158 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" (OuterVolumeSpecName: "machine-approver-tls") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "machine-approver-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:11:15 crc kubenswrapper[4988]: E1008 18:11:15.340226 4988 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.341695 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.341710 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" (OuterVolumeSpecName: "console-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.340885 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" (OuterVolumeSpecName: "kube-api-access-tk88c") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "kube-api-access-tk88c". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:11:15 crc kubenswrapper[4988]: E1008 18:11:15.341761 4988 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-08 18:11:15.841730212 +0000 UTC m=+21.291573022 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.342078 4988 swap_util.go:74] "error creating dir to test if tmpfs noswap is enabled. Assuming not supported" mount path="" error="stat /var/lib/kubelet/plugins/kubernetes.io/empty-dir: no such file or directory" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.342610 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 08 18:11:15 crc kubenswrapper[4988]: E1008 18:11:15.343379 4988 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 08 18:11:15 crc kubenswrapper[4988]: E1008 18:11:15.343478 4988 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-08 18:11:15.843452804 +0000 UTC m=+21.293295814 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.340880 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" (OuterVolumeSpecName: "config") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.341142 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.344672 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" (OuterVolumeSpecName: "kube-api-access-jhbk2") pod "bd23aa5c-e532-4e53-bccf-e79f130c5ae8" (UID: "bd23aa5c-e532-4e53-bccf-e79f130c5ae8"). InnerVolumeSpecName "kube-api-access-jhbk2". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.344165 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" (OuterVolumeSpecName: "samples-operator-tls") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "samples-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.344406 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" (OuterVolumeSpecName: "available-featuregates") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "available-featuregates". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.344550 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.344532 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" (OuterVolumeSpecName: "kube-api-access-249nr") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "kube-api-access-249nr". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.344664 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.344955 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" (OuterVolumeSpecName: "kube-api-access-279lb") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "kube-api-access-279lb". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.345097 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" (OuterVolumeSpecName: "kube-api-access-x4zgh") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "kube-api-access-x4zgh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.345126 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" (OuterVolumeSpecName: "kube-api-access-8tdtz") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "kube-api-access-8tdtz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.345411 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" (OuterVolumeSpecName: "kube-api-access-vt5rc") pod "44663579-783b-4372-86d6-acf235a62d72" (UID: "44663579-783b-4372-86d6-acf235a62d72"). InnerVolumeSpecName "kube-api-access-vt5rc". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.346295 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.346566 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" (OuterVolumeSpecName: "audit") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "audit". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.346930 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" (OuterVolumeSpecName: "kube-api-access-gf66m") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "kube-api-access-gf66m". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.347092 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" (OuterVolumeSpecName: "config") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.347570 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.348018 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" (OuterVolumeSpecName: "config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.348172 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.348237 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.348520 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" (OuterVolumeSpecName: "image-import-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "image-import-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.348809 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.351822 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.358806 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.360856 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.361888 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.362089 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" (OuterVolumeSpecName: "ovn-control-plane-metrics-cert") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovn-control-plane-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.362691 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.362828 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.362967 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" (OuterVolumeSpecName: "kube-api-access-s4n52") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "kube-api-access-s4n52". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:11:15 crc kubenswrapper[4988]: E1008 18:11:15.363423 4988 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 08 18:11:15 crc kubenswrapper[4988]: E1008 18:11:15.363449 4988 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 08 18:11:15 crc kubenswrapper[4988]: E1008 18:11:15.363468 4988 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 08 18:11:15 crc kubenswrapper[4988]: E1008 18:11:15.363554 4988 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-08 18:11:15.863529415 +0000 UTC m=+21.313372395 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.364258 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.364607 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.365868 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" (OuterVolumeSpecName: "default-certificate") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "default-certificate". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.366005 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" (OuterVolumeSpecName: "stats-auth") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "stats-auth". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.366882 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.366892 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" (OuterVolumeSpecName: "config") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.366991 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" (OuterVolumeSpecName: "kube-api-access-dbsvg") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "kube-api-access-dbsvg". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:11:15 crc kubenswrapper[4988]: E1008 18:11:15.367378 4988 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 08 18:11:15 crc kubenswrapper[4988]: E1008 18:11:15.367425 4988 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 08 18:11:15 crc kubenswrapper[4988]: E1008 18:11:15.367445 4988 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 08 18:11:15 crc kubenswrapper[4988]: E1008 18:11:15.367522 4988 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-08 18:11:15.867495326 +0000 UTC m=+21.317338326 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.367635 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" (OuterVolumeSpecName: "kube-api-access-lzf88") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "kube-api-access-lzf88". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.368669 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.368968 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.369110 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" (OuterVolumeSpecName: "kube-api-access-ngvvp") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "kube-api-access-ngvvp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.369620 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" (OuterVolumeSpecName: "mcc-auth-proxy-config") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "mcc-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.369650 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.369672 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" (OuterVolumeSpecName: "config") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.369757 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.369982 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" (OuterVolumeSpecName: "kube-api-access-pcxfs") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "kube-api-access-pcxfs". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.370206 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" (OuterVolumeSpecName: "cni-sysctl-allowlist") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-sysctl-allowlist". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.370518 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" (OuterVolumeSpecName: "serviceca") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "serviceca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.370546 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.370749 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.370853 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.370880 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.373129 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.373715 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.373911 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.374425 4988 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.375676 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.377162 4988 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="5bd35aef24be034a0b305a40bb53ca171a4f76db3b93b4f78ae95099b0a4cde3" exitCode=255 Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.377211 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"5bd35aef24be034a0b305a40bb53ca171a4f76db3b93b4f78ae95099b0a4cde3"} Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.378593 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.384030 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.389217 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.391316 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.391875 4988 scope.go:117] "RemoveContainer" containerID="5bd35aef24be034a0b305a40bb53ca171a4f76db3b93b4f78ae95099b0a4cde3" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.393171 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.400710 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.403243 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.403266 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.413099 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.424866 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.434969 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.442638 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.442708 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.442771 4988 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.442789 4988 reconciler_common.go:293] "Volume detached for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") on node \"crc\" DevicePath \"\"" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.442801 4988 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") on node \"crc\" DevicePath \"\"" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.442812 4988 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.442823 4988 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") on node \"crc\" DevicePath \"\"" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.442832 4988 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.442843 4988 reconciler_common.go:293] "Volume detached for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.442853 4988 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") on node \"crc\" DevicePath \"\"" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.442864 4988 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.442875 4988 reconciler_common.go:293] "Volume detached for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") on node \"crc\" DevicePath \"\"" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.442885 4988 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.442896 4988 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.442908 4988 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") on node \"crc\" DevicePath \"\"" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.442922 4988 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") on node \"crc\" DevicePath \"\"" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.442810 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.442935 4988 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") on node \"crc\" DevicePath \"\"" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.442985 4988 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") on node \"crc\" DevicePath \"\"" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.442997 4988 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.443008 4988 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") on node \"crc\" DevicePath \"\"" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.443018 4988 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") on node \"crc\" DevicePath \"\"" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.443030 4988 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.443040 4988 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.443052 4988 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") on node \"crc\" DevicePath \"\"" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.443063 4988 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") on node \"crc\" DevicePath \"\"" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.442762 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.443073 4988 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") on node \"crc\" DevicePath \"\"" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.443173 4988 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.443191 4988 reconciler_common.go:293] "Volume detached for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.443205 4988 reconciler_common.go:293] "Volume detached for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") on node \"crc\" DevicePath \"\"" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.443219 4988 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.443228 4988 reconciler_common.go:293] "Volume detached for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") on node \"crc\" DevicePath \"\"" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.443238 4988 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.443249 4988 reconciler_common.go:293] "Volume detached for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") on node \"crc\" DevicePath \"\"" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.443258 4988 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") on node \"crc\" DevicePath \"\"" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.443269 4988 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") on node \"crc\" DevicePath \"\"" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.443280 4988 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.443290 4988 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.443300 4988 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") on node \"crc\" DevicePath \"\"" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.443310 4988 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") on node \"crc\" DevicePath \"\"" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.443320 4988 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") on node \"crc\" DevicePath \"\"" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.443329 4988 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") on node \"crc\" DevicePath \"\"" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.443339 4988 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") on node \"crc\" DevicePath \"\"" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.443350 4988 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") on node \"crc\" DevicePath \"\"" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.443360 4988 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") on node \"crc\" DevicePath \"\"" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.443369 4988 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.443378 4988 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.443400 4988 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.443410 4988 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.443430 4988 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") on node \"crc\" DevicePath \"\"" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.443443 4988 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") on node \"crc\" DevicePath \"\"" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.443453 4988 reconciler_common.go:293] "Volume detached for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") on node \"crc\" DevicePath \"\"" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.443462 4988 reconciler_common.go:293] "Volume detached for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") on node \"crc\" DevicePath \"\"" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.443472 4988 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") on node \"crc\" DevicePath \"\"" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.443482 4988 reconciler_common.go:293] "Volume detached for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") on node \"crc\" DevicePath \"\"" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.443491 4988 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") on node \"crc\" DevicePath \"\"" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.443501 4988 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.443511 4988 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") on node \"crc\" DevicePath \"\"" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.443521 4988 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") on node \"crc\" DevicePath \"\"" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.443530 4988 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") on node \"crc\" DevicePath \"\"" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.443540 4988 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") on node \"crc\" DevicePath \"\"" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.443549 4988 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") on node \"crc\" DevicePath \"\"" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.443558 4988 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.443567 4988 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") on node \"crc\" DevicePath \"\"" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.443577 4988 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") on node \"crc\" DevicePath \"\"" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.443586 4988 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") on node \"crc\" DevicePath \"\"" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.443595 4988 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") on node \"crc\" DevicePath \"\"" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.443604 4988 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") on node \"crc\" DevicePath \"\"" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.443613 4988 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") on node \"crc\" DevicePath \"\"" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.443623 4988 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.443632 4988 reconciler_common.go:293] "Volume detached for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") on node \"crc\" DevicePath \"\"" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.443643 4988 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") on node \"crc\" DevicePath \"\"" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.443652 4988 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") on node \"crc\" DevicePath \"\"" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.443661 4988 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") on node \"crc\" DevicePath \"\"" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.443670 4988 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") on node \"crc\" DevicePath \"\"" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.443679 4988 reconciler_common.go:293] "Volume detached for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") on node \"crc\" DevicePath \"\"" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.443688 4988 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.443698 4988 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.443710 4988 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") on node \"crc\" DevicePath \"\"" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.443720 4988 reconciler_common.go:293] "Volume detached for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") on node \"crc\" DevicePath \"\"" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.446029 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.469557 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.500486 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aef71be1-1d12-46e5-95b2-bd399040fd4d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:10:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:10:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:10:55Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:10:55Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:10:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77c68f3ecb8a0724499e495cb26ef3db5beecb140f2558d8c0b25f41928b84ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:10:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a560ee84c2405227de71fda127566213cd4f017e73d5581442ba267781b1c539\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:10:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6100accc3bd08928bc1ff88310f29fa5da37fd2d5c315c94279801c9778c4745\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:10:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5bd35aef24be034a0b305a40bb53ca171a4f76db3b93b4f78ae95099b0a4cde3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5bd35aef24be034a0b305a40bb53ca171a4f76db3b93b4f78ae95099b0a4cde3\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-08T18:11:14Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1008 18:11:09.091034 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1008 18:11:09.091984 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3472374062/tls.crt::/tmp/serving-cert-3472374062/tls.key\\\\\\\"\\\\nI1008 18:11:14.619696 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1008 18:11:14.628837 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1008 18:11:14.628898 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1008 18:11:14.628994 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1008 18:11:14.629036 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1008 18:11:14.641318 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1008 18:11:14.641363 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1008 18:11:14.641372 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1008 18:11:14.641382 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1008 18:11:14.641415 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1008 18:11:14.641423 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1008 18:11:14.641431 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1008 18:11:14.641484 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1008 18:11:14.646451 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T18:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c79ef704fda1bbcc486f4c7afa3039a3ab38669833f0f66bafe24dd4cc679e60\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:10:58Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://28d21731af0f16d6a8ae6c6e9d0bebf0560d4a533db4c78c10bf1837890394ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://28d21731af0f16d6a8ae6c6e9d0bebf0560d4a533db4c78c10bf1837890394ef\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:10:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:10:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:10:55Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.513981 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.518145 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.524735 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 08 18:11:15 crc kubenswrapper[4988]: W1008 18:11:15.534835 4988 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod37a5e44f_9a88_4405_be8a_b645485e7312.slice/crio-ba985648bde602e9f2b2c11d837b2b158288c7c366898e0fe9e3de9d9d6775e0 WatchSource:0}: Error finding container ba985648bde602e9f2b2c11d837b2b158288c7c366898e0fe9e3de9d9d6775e0: Status 404 returned error can't find the container with id ba985648bde602e9f2b2c11d837b2b158288c7c366898e0fe9e3de9d9d6775e0 Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.534936 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.541884 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.550557 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.556624 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 08 18:11:15 crc kubenswrapper[4988]: W1008 18:11:15.559645 4988 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podef543e1b_8068_4ea3_b32a_61027b32e95d.slice/crio-e1c7e2c2ce97311d5ef22eea3c0e0f9a50a2a66b9b6ffce9040f3538c2e99dbe WatchSource:0}: Error finding container e1c7e2c2ce97311d5ef22eea3c0e0f9a50a2a66b9b6ffce9040f3538c2e99dbe: Status 404 returned error can't find the container with id e1c7e2c2ce97311d5ef22eea3c0e0f9a50a2a66b9b6ffce9040f3538c2e99dbe Oct 08 18:11:15 crc kubenswrapper[4988]: W1008 18:11:15.568220 4988 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd75a4c96_2883_4a0b_bab2_0fab2b6c0b49.slice/crio-6f366aa46454c618b420e92e9b12118bf147c3384afd2176eb031b9770663d15 WatchSource:0}: Error finding container 6f366aa46454c618b420e92e9b12118bf147c3384afd2176eb031b9770663d15: Status 404 returned error can't find the container with id 6f366aa46454c618b420e92e9b12118bf147c3384afd2176eb031b9770663d15 Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.569666 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.777622 4988 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.847789 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.847873 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.847913 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 18:11:15 crc kubenswrapper[4988]: E1008 18:11:15.848018 4988 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 08 18:11:15 crc kubenswrapper[4988]: E1008 18:11:15.848054 4988 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 18:11:16.848024483 +0000 UTC m=+22.297867253 (durationBeforeRetry 1s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 18:11:15 crc kubenswrapper[4988]: E1008 18:11:15.848104 4988 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-08 18:11:16.848084275 +0000 UTC m=+22.297927045 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 08 18:11:15 crc kubenswrapper[4988]: E1008 18:11:15.848145 4988 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 08 18:11:15 crc kubenswrapper[4988]: E1008 18:11:15.848266 4988 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-08 18:11:16.84824131 +0000 UTC m=+22.298084080 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.930202 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/node-resolver-s4f2k"] Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.930665 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-s4f2k" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.933661 4988 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.933886 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.934562 4988 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.948642 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.948724 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 18:11:15 crc kubenswrapper[4988]: E1008 18:11:15.948917 4988 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 08 18:11:15 crc kubenswrapper[4988]: E1008 18:11:15.948978 4988 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 08 18:11:15 crc kubenswrapper[4988]: E1008 18:11:15.948996 4988 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 08 18:11:15 crc kubenswrapper[4988]: E1008 18:11:15.948933 4988 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 08 18:11:15 crc kubenswrapper[4988]: E1008 18:11:15.949082 4988 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 08 18:11:15 crc kubenswrapper[4988]: E1008 18:11:15.949089 4988 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-08 18:11:16.949061456 +0000 UTC m=+22.398904406 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 08 18:11:15 crc kubenswrapper[4988]: E1008 18:11:15.949097 4988 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 08 18:11:15 crc kubenswrapper[4988]: E1008 18:11:15.949175 4988 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-08 18:11:16.949152019 +0000 UTC m=+22.398994789 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.952438 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.962277 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.972948 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.982041 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-s4f2k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ba5e11d7-7bde-4c3d-983b-884232219652\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hn5jf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:11:15Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-s4f2k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 08 18:11:15 crc kubenswrapper[4988]: I1008 18:11:15.993191 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 08 18:11:16 crc kubenswrapper[4988]: I1008 18:11:16.004213 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 08 18:11:16 crc kubenswrapper[4988]: I1008 18:11:16.030741 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/node-ca-dzxs2"] Oct 08 18:11:16 crc kubenswrapper[4988]: I1008 18:11:16.031162 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-dzxs2" Oct 08 18:11:16 crc kubenswrapper[4988]: I1008 18:11:16.036541 4988 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Oct 08 18:11:16 crc kubenswrapper[4988]: I1008 18:11:16.036768 4988 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Oct 08 18:11:16 crc kubenswrapper[4988]: I1008 18:11:16.036854 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Oct 08 18:11:16 crc kubenswrapper[4988]: I1008 18:11:16.036975 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 08 18:11:16 crc kubenswrapper[4988]: I1008 18:11:16.037927 4988 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Oct 08 18:11:16 crc kubenswrapper[4988]: I1008 18:11:16.049473 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/ba5e11d7-7bde-4c3d-983b-884232219652-hosts-file\") pod \"node-resolver-s4f2k\" (UID: \"ba5e11d7-7bde-4c3d-983b-884232219652\") " pod="openshift-dns/node-resolver-s4f2k" Oct 08 18:11:16 crc kubenswrapper[4988]: I1008 18:11:16.049521 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hn5jf\" (UniqueName: \"kubernetes.io/projected/ba5e11d7-7bde-4c3d-983b-884232219652-kube-api-access-hn5jf\") pod \"node-resolver-s4f2k\" (UID: \"ba5e11d7-7bde-4c3d-983b-884232219652\") " pod="openshift-dns/node-resolver-s4f2k" Oct 08 18:11:16 crc kubenswrapper[4988]: I1008 18:11:16.060193 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aef71be1-1d12-46e5-95b2-bd399040fd4d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:10:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:10:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:10:55Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:10:55Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:10:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77c68f3ecb8a0724499e495cb26ef3db5beecb140f2558d8c0b25f41928b84ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:10:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a560ee84c2405227de71fda127566213cd4f017e73d5581442ba267781b1c539\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:10:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6100accc3bd08928bc1ff88310f29fa5da37fd2d5c315c94279801c9778c4745\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:10:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5bd35aef24be034a0b305a40bb53ca171a4f76db3b93b4f78ae95099b0a4cde3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5bd35aef24be034a0b305a40bb53ca171a4f76db3b93b4f78ae95099b0a4cde3\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-08T18:11:14Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1008 18:11:09.091034 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1008 18:11:09.091984 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3472374062/tls.crt::/tmp/serving-cert-3472374062/tls.key\\\\\\\"\\\\nI1008 18:11:14.619696 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1008 18:11:14.628837 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1008 18:11:14.628898 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1008 18:11:14.628994 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1008 18:11:14.629036 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1008 18:11:14.641318 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1008 18:11:14.641363 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1008 18:11:14.641372 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1008 18:11:14.641382 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1008 18:11:14.641415 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1008 18:11:14.641423 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1008 18:11:14.641431 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1008 18:11:14.641484 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1008 18:11:14.646451 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T18:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c79ef704fda1bbcc486f4c7afa3039a3ab38669833f0f66bafe24dd4cc679e60\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:10:58Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://28d21731af0f16d6a8ae6c6e9d0bebf0560d4a533db4c78c10bf1837890394ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://28d21731af0f16d6a8ae6c6e9d0bebf0560d4a533db4c78c10bf1837890394ef\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:10:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:10:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:10:55Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 08 18:11:16 crc kubenswrapper[4988]: I1008 18:11:16.072466 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-dzxs2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8f307b20-5e33-4725-99cb-210e25113532\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q65f7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:11:16Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-dzxs2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 08 18:11:16 crc kubenswrapper[4988]: I1008 18:11:16.105561 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aef71be1-1d12-46e5-95b2-bd399040fd4d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:10:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:10:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:10:55Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:10:55Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:10:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77c68f3ecb8a0724499e495cb26ef3db5beecb140f2558d8c0b25f41928b84ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:10:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a560ee84c2405227de71fda127566213cd4f017e73d5581442ba267781b1c539\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:10:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6100accc3bd08928bc1ff88310f29fa5da37fd2d5c315c94279801c9778c4745\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:10:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5bd35aef24be034a0b305a40bb53ca171a4f76db3b93b4f78ae95099b0a4cde3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5bd35aef24be034a0b305a40bb53ca171a4f76db3b93b4f78ae95099b0a4cde3\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-08T18:11:14Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1008 18:11:09.091034 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1008 18:11:09.091984 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3472374062/tls.crt::/tmp/serving-cert-3472374062/tls.key\\\\\\\"\\\\nI1008 18:11:14.619696 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1008 18:11:14.628837 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1008 18:11:14.628898 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1008 18:11:14.628994 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1008 18:11:14.629036 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1008 18:11:14.641318 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1008 18:11:14.641363 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1008 18:11:14.641372 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1008 18:11:14.641382 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1008 18:11:14.641415 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1008 18:11:14.641423 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1008 18:11:14.641431 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1008 18:11:14.641484 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1008 18:11:14.646451 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T18:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c79ef704fda1bbcc486f4c7afa3039a3ab38669833f0f66bafe24dd4cc679e60\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:10:58Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://28d21731af0f16d6a8ae6c6e9d0bebf0560d4a533db4c78c10bf1837890394ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://28d21731af0f16d6a8ae6c6e9d0bebf0560d4a533db4c78c10bf1837890394ef\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:10:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:10:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:10:55Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 08 18:11:16 crc kubenswrapper[4988]: I1008 18:11:16.121269 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 08 18:11:16 crc kubenswrapper[4988]: I1008 18:11:16.142530 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 08 18:11:16 crc kubenswrapper[4988]: I1008 18:11:16.150363 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/ba5e11d7-7bde-4c3d-983b-884232219652-hosts-file\") pod \"node-resolver-s4f2k\" (UID: \"ba5e11d7-7bde-4c3d-983b-884232219652\") " pod="openshift-dns/node-resolver-s4f2k" Oct 08 18:11:16 crc kubenswrapper[4988]: I1008 18:11:16.150419 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hn5jf\" (UniqueName: \"kubernetes.io/projected/ba5e11d7-7bde-4c3d-983b-884232219652-kube-api-access-hn5jf\") pod \"node-resolver-s4f2k\" (UID: \"ba5e11d7-7bde-4c3d-983b-884232219652\") " pod="openshift-dns/node-resolver-s4f2k" Oct 08 18:11:16 crc kubenswrapper[4988]: I1008 18:11:16.150463 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q65f7\" (UniqueName: \"kubernetes.io/projected/8f307b20-5e33-4725-99cb-210e25113532-kube-api-access-q65f7\") pod \"node-ca-dzxs2\" (UID: \"8f307b20-5e33-4725-99cb-210e25113532\") " pod="openshift-image-registry/node-ca-dzxs2" Oct 08 18:11:16 crc kubenswrapper[4988]: I1008 18:11:16.150486 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/8f307b20-5e33-4725-99cb-210e25113532-host\") pod \"node-ca-dzxs2\" (UID: \"8f307b20-5e33-4725-99cb-210e25113532\") " pod="openshift-image-registry/node-ca-dzxs2" Oct 08 18:11:16 crc kubenswrapper[4988]: I1008 18:11:16.150500 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/8f307b20-5e33-4725-99cb-210e25113532-serviceca\") pod \"node-ca-dzxs2\" (UID: \"8f307b20-5e33-4725-99cb-210e25113532\") " pod="openshift-image-registry/node-ca-dzxs2" Oct 08 18:11:16 crc kubenswrapper[4988]: I1008 18:11:16.150542 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/ba5e11d7-7bde-4c3d-983b-884232219652-hosts-file\") pod \"node-resolver-s4f2k\" (UID: \"ba5e11d7-7bde-4c3d-983b-884232219652\") " pod="openshift-dns/node-resolver-s4f2k" Oct 08 18:11:16 crc kubenswrapper[4988]: I1008 18:11:16.156806 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 08 18:11:16 crc kubenswrapper[4988]: I1008 18:11:16.170835 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 08 18:11:16 crc kubenswrapper[4988]: I1008 18:11:16.176295 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hn5jf\" (UniqueName: \"kubernetes.io/projected/ba5e11d7-7bde-4c3d-983b-884232219652-kube-api-access-hn5jf\") pod \"node-resolver-s4f2k\" (UID: \"ba5e11d7-7bde-4c3d-983b-884232219652\") " pod="openshift-dns/node-resolver-s4f2k" Oct 08 18:11:16 crc kubenswrapper[4988]: I1008 18:11:16.183565 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 08 18:11:16 crc kubenswrapper[4988]: I1008 18:11:16.194439 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 08 18:11:16 crc kubenswrapper[4988]: I1008 18:11:16.203283 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-s4f2k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ba5e11d7-7bde-4c3d-983b-884232219652\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hn5jf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:11:15Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-s4f2k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 08 18:11:16 crc kubenswrapper[4988]: I1008 18:11:16.242679 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-s4f2k" Oct 08 18:11:16 crc kubenswrapper[4988]: I1008 18:11:16.251044 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q65f7\" (UniqueName: \"kubernetes.io/projected/8f307b20-5e33-4725-99cb-210e25113532-kube-api-access-q65f7\") pod \"node-ca-dzxs2\" (UID: \"8f307b20-5e33-4725-99cb-210e25113532\") " pod="openshift-image-registry/node-ca-dzxs2" Oct 08 18:11:16 crc kubenswrapper[4988]: I1008 18:11:16.251119 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/8f307b20-5e33-4725-99cb-210e25113532-host\") pod \"node-ca-dzxs2\" (UID: \"8f307b20-5e33-4725-99cb-210e25113532\") " pod="openshift-image-registry/node-ca-dzxs2" Oct 08 18:11:16 crc kubenswrapper[4988]: I1008 18:11:16.251147 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/8f307b20-5e33-4725-99cb-210e25113532-serviceca\") pod \"node-ca-dzxs2\" (UID: \"8f307b20-5e33-4725-99cb-210e25113532\") " pod="openshift-image-registry/node-ca-dzxs2" Oct 08 18:11:16 crc kubenswrapper[4988]: I1008 18:11:16.251346 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/8f307b20-5e33-4725-99cb-210e25113532-host\") pod \"node-ca-dzxs2\" (UID: \"8f307b20-5e33-4725-99cb-210e25113532\") " pod="openshift-image-registry/node-ca-dzxs2" Oct 08 18:11:16 crc kubenswrapper[4988]: I1008 18:11:16.252245 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/8f307b20-5e33-4725-99cb-210e25113532-serviceca\") pod \"node-ca-dzxs2\" (UID: \"8f307b20-5e33-4725-99cb-210e25113532\") " pod="openshift-image-registry/node-ca-dzxs2" Oct 08 18:11:16 crc kubenswrapper[4988]: W1008 18:11:16.256216 4988 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podba5e11d7_7bde_4c3d_983b_884232219652.slice/crio-8cb732e5b5f1b04c0cc364e91ad69a24451b884c8fd57ff34e52a35ffb14842c WatchSource:0}: Error finding container 8cb732e5b5f1b04c0cc364e91ad69a24451b884c8fd57ff34e52a35ffb14842c: Status 404 returned error can't find the container with id 8cb732e5b5f1b04c0cc364e91ad69a24451b884c8fd57ff34e52a35ffb14842c Oct 08 18:11:16 crc kubenswrapper[4988]: I1008 18:11:16.278003 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q65f7\" (UniqueName: \"kubernetes.io/projected/8f307b20-5e33-4725-99cb-210e25113532-kube-api-access-q65f7\") pod \"node-ca-dzxs2\" (UID: \"8f307b20-5e33-4725-99cb-210e25113532\") " pod="openshift-image-registry/node-ca-dzxs2" Oct 08 18:11:16 crc kubenswrapper[4988]: I1008 18:11:16.342589 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-dzxs2" Oct 08 18:11:16 crc kubenswrapper[4988]: I1008 18:11:16.416750 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-s4f2k" event={"ID":"ba5e11d7-7bde-4c3d-983b-884232219652","Type":"ContainerStarted","Data":"8cb732e5b5f1b04c0cc364e91ad69a24451b884c8fd57ff34e52a35ffb14842c"} Oct 08 18:11:16 crc kubenswrapper[4988]: I1008 18:11:16.422592 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"6f366aa46454c618b420e92e9b12118bf147c3384afd2176eb031b9770663d15"} Oct 08 18:11:16 crc kubenswrapper[4988]: I1008 18:11:16.424953 4988 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Oct 08 18:11:16 crc kubenswrapper[4988]: I1008 18:11:16.426581 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"b0e596d7ce4e9341c44499af79d3b3caf0e5e2c156902ea40cf2a3dca672be9e"} Oct 08 18:11:16 crc kubenswrapper[4988]: I1008 18:11:16.427331 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 08 18:11:16 crc kubenswrapper[4988]: I1008 18:11:16.428733 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-dzxs2" event={"ID":"8f307b20-5e33-4725-99cb-210e25113532","Type":"ContainerStarted","Data":"93d4251f6d68fc6cdd6e14c86512eeddca81b7bd875593e9edd9e66ae538adb7"} Oct 08 18:11:16 crc kubenswrapper[4988]: I1008 18:11:16.431640 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"a35a5e989ba69c8ef36ce4cea272aadf67ce20473fc257adfcfa86a2faba6ae0"} Oct 08 18:11:16 crc kubenswrapper[4988]: I1008 18:11:16.431691 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"a522e937cfe03fb9f2e60cb1bea07214cf63c69bdecd5ba65a2dfdba882d28dd"} Oct 08 18:11:16 crc kubenswrapper[4988]: I1008 18:11:16.431704 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"e1c7e2c2ce97311d5ef22eea3c0e0f9a50a2a66b9b6ffce9040f3538c2e99dbe"} Oct 08 18:11:16 crc kubenswrapper[4988]: I1008 18:11:16.434459 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"e8906182d55c285b3cf095e2d9e1dac397af4618a514f11e26b44c172dd573f3"} Oct 08 18:11:16 crc kubenswrapper[4988]: I1008 18:11:16.434500 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"ba985648bde602e9f2b2c11d837b2b158288c7c366898e0fe9e3de9d9d6775e0"} Oct 08 18:11:16 crc kubenswrapper[4988]: I1008 18:11:16.447671 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:16Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:16 crc kubenswrapper[4988]: I1008 18:11:16.472510 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:16Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:16 crc kubenswrapper[4988]: I1008 18:11:16.486406 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:16Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:16 crc kubenswrapper[4988]: I1008 18:11:16.498907 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-dzxs2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8f307b20-5e33-4725-99cb-210e25113532\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q65f7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:11:16Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-dzxs2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:16Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:16 crc kubenswrapper[4988]: I1008 18:11:16.519354 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aef71be1-1d12-46e5-95b2-bd399040fd4d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:10:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:10:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:10:55Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:10:55Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:10:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77c68f3ecb8a0724499e495cb26ef3db5beecb140f2558d8c0b25f41928b84ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:10:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a560ee84c2405227de71fda127566213cd4f017e73d5581442ba267781b1c539\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:10:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6100accc3bd08928bc1ff88310f29fa5da37fd2d5c315c94279801c9778c4745\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:10:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b0e596d7ce4e9341c44499af79d3b3caf0e5e2c156902ea40cf2a3dca672be9e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5bd35aef24be034a0b305a40bb53ca171a4f76db3b93b4f78ae95099b0a4cde3\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-08T18:11:14Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1008 18:11:09.091034 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1008 18:11:09.091984 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3472374062/tls.crt::/tmp/serving-cert-3472374062/tls.key\\\\\\\"\\\\nI1008 18:11:14.619696 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1008 18:11:14.628837 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1008 18:11:14.628898 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1008 18:11:14.628994 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1008 18:11:14.629036 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1008 18:11:14.641318 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1008 18:11:14.641363 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1008 18:11:14.641372 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1008 18:11:14.641382 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1008 18:11:14.641415 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1008 18:11:14.641423 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1008 18:11:14.641431 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1008 18:11:14.641484 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1008 18:11:14.646451 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T18:10:58Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c79ef704fda1bbcc486f4c7afa3039a3ab38669833f0f66bafe24dd4cc679e60\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:10:58Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://28d21731af0f16d6a8ae6c6e9d0bebf0560d4a533db4c78c10bf1837890394ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://28d21731af0f16d6a8ae6c6e9d0bebf0560d4a533db4c78c10bf1837890394ef\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:10:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:10:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:10:55Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:16Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:16 crc kubenswrapper[4988]: I1008 18:11:16.537706 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:16Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:16 crc kubenswrapper[4988]: I1008 18:11:16.574037 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:16Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:16 crc kubenswrapper[4988]: I1008 18:11:16.588405 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:16Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:16 crc kubenswrapper[4988]: I1008 18:11:16.599436 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-s4f2k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ba5e11d7-7bde-4c3d-983b-884232219652\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hn5jf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:11:15Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-s4f2k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:16Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:16 crc kubenswrapper[4988]: I1008 18:11:16.619116 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a35a5e989ba69c8ef36ce4cea272aadf67ce20473fc257adfcfa86a2faba6ae0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a522e937cfe03fb9f2e60cb1bea07214cf63c69bdecd5ba65a2dfdba882d28dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:16Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:16 crc kubenswrapper[4988]: I1008 18:11:16.635705 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:16Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:16 crc kubenswrapper[4988]: I1008 18:11:16.654134 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:16Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:16 crc kubenswrapper[4988]: I1008 18:11:16.667478 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-s4f2k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ba5e11d7-7bde-4c3d-983b-884232219652\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hn5jf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:11:15Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-s4f2k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:16Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:16 crc kubenswrapper[4988]: I1008 18:11:16.684885 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aef71be1-1d12-46e5-95b2-bd399040fd4d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:10:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:10:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:10:55Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:10:55Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:10:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77c68f3ecb8a0724499e495cb26ef3db5beecb140f2558d8c0b25f41928b84ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:10:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a560ee84c2405227de71fda127566213cd4f017e73d5581442ba267781b1c539\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:10:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6100accc3bd08928bc1ff88310f29fa5da37fd2d5c315c94279801c9778c4745\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:10:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b0e596d7ce4e9341c44499af79d3b3caf0e5e2c156902ea40cf2a3dca672be9e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5bd35aef24be034a0b305a40bb53ca171a4f76db3b93b4f78ae95099b0a4cde3\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-08T18:11:14Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1008 18:11:09.091034 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1008 18:11:09.091984 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3472374062/tls.crt::/tmp/serving-cert-3472374062/tls.key\\\\\\\"\\\\nI1008 18:11:14.619696 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1008 18:11:14.628837 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1008 18:11:14.628898 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1008 18:11:14.628994 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1008 18:11:14.629036 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1008 18:11:14.641318 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1008 18:11:14.641363 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1008 18:11:14.641372 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1008 18:11:14.641382 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1008 18:11:14.641415 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1008 18:11:14.641423 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1008 18:11:14.641431 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1008 18:11:14.641484 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1008 18:11:14.646451 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T18:10:58Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c79ef704fda1bbcc486f4c7afa3039a3ab38669833f0f66bafe24dd4cc679e60\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:10:58Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://28d21731af0f16d6a8ae6c6e9d0bebf0560d4a533db4c78c10bf1837890394ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://28d21731af0f16d6a8ae6c6e9d0bebf0560d4a533db4c78c10bf1837890394ef\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:10:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:10:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:10:55Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:16Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:16 crc kubenswrapper[4988]: I1008 18:11:16.707602 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e8906182d55c285b3cf095e2d9e1dac397af4618a514f11e26b44c172dd573f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:16Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:16 crc kubenswrapper[4988]: I1008 18:11:16.729427 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:16Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:16 crc kubenswrapper[4988]: I1008 18:11:16.758179 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:16Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:16 crc kubenswrapper[4988]: I1008 18:11:16.800726 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-dzxs2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8f307b20-5e33-4725-99cb-210e25113532\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q65f7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:11:16Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-dzxs2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:16Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:16 crc kubenswrapper[4988]: I1008 18:11:16.857005 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 18:11:16 crc kubenswrapper[4988]: I1008 18:11:16.857117 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 18:11:16 crc kubenswrapper[4988]: I1008 18:11:16.857179 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 18:11:16 crc kubenswrapper[4988]: E1008 18:11:16.857231 4988 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 18:11:18.857193881 +0000 UTC m=+24.307036651 (durationBeforeRetry 2s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 18:11:16 crc kubenswrapper[4988]: E1008 18:11:16.857283 4988 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 08 18:11:16 crc kubenswrapper[4988]: E1008 18:11:16.857339 4988 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 08 18:11:16 crc kubenswrapper[4988]: E1008 18:11:16.857357 4988 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-08 18:11:18.857336355 +0000 UTC m=+24.307179125 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 08 18:11:16 crc kubenswrapper[4988]: E1008 18:11:16.857443 4988 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-08 18:11:18.857428698 +0000 UTC m=+24.307271618 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 08 18:11:16 crc kubenswrapper[4988]: I1008 18:11:16.859970 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-h9jwp"] Oct 08 18:11:16 crc kubenswrapper[4988]: I1008 18:11:16.860361 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-h9jwp" Oct 08 18:11:16 crc kubenswrapper[4988]: I1008 18:11:16.861809 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-daemon-nm54f"] Oct 08 18:11:16 crc kubenswrapper[4988]: I1008 18:11:16.862114 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-additional-cni-plugins-jnh7m"] Oct 08 18:11:16 crc kubenswrapper[4988]: I1008 18:11:16.862823 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-jnh7m" Oct 08 18:11:16 crc kubenswrapper[4988]: I1008 18:11:16.863168 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" Oct 08 18:11:16 crc kubenswrapper[4988]: I1008 18:11:16.864491 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-nsc65"] Oct 08 18:11:16 crc kubenswrapper[4988]: I1008 18:11:16.866353 4988 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Oct 08 18:11:16 crc kubenswrapper[4988]: I1008 18:11:16.866628 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Oct 08 18:11:16 crc kubenswrapper[4988]: I1008 18:11:16.866970 4988 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Oct 08 18:11:16 crc kubenswrapper[4988]: I1008 18:11:16.867550 4988 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Oct 08 18:11:16 crc kubenswrapper[4988]: I1008 18:11:16.867960 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Oct 08 18:11:16 crc kubenswrapper[4988]: I1008 18:11:16.867963 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Oct 08 18:11:16 crc kubenswrapper[4988]: I1008 18:11:16.868302 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Oct 08 18:11:16 crc kubenswrapper[4988]: I1008 18:11:16.868508 4988 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Oct 08 18:11:16 crc kubenswrapper[4988]: I1008 18:11:16.868574 4988 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Oct 08 18:11:16 crc kubenswrapper[4988]: I1008 18:11:16.868506 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-nsc65" Oct 08 18:11:16 crc kubenswrapper[4988]: I1008 18:11:16.877988 4988 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Oct 08 18:11:16 crc kubenswrapper[4988]: I1008 18:11:16.878661 4988 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Oct 08 18:11:16 crc kubenswrapper[4988]: I1008 18:11:16.879018 4988 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Oct 08 18:11:16 crc kubenswrapper[4988]: I1008 18:11:16.881868 4988 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Oct 08 18:11:16 crc kubenswrapper[4988]: I1008 18:11:16.886317 4988 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Oct 08 18:11:16 crc kubenswrapper[4988]: I1008 18:11:16.886606 4988 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Oct 08 18:11:16 crc kubenswrapper[4988]: I1008 18:11:16.886787 4988 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Oct 08 18:11:16 crc kubenswrapper[4988]: I1008 18:11:16.886819 4988 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Oct 08 18:11:16 crc kubenswrapper[4988]: I1008 18:11:16.886791 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Oct 08 18:11:16 crc kubenswrapper[4988]: I1008 18:11:16.890185 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Oct 08 18:11:16 crc kubenswrapper[4988]: I1008 18:11:16.905310 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:16Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:16 crc kubenswrapper[4988]: I1008 18:11:16.921611 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:16Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:16 crc kubenswrapper[4988]: I1008 18:11:16.938445 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-h9jwp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4490d846-ca32-4cd5-9202-87cc6d499fb4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7k6rk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:11:16Z\\\"}}\" for pod \"openshift-multus\"/\"multus-h9jwp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:16Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:16 crc kubenswrapper[4988]: I1008 18:11:16.950310 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-s4f2k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ba5e11d7-7bde-4c3d-983b-884232219652\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hn5jf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:11:15Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-s4f2k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:16Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:16 crc kubenswrapper[4988]: I1008 18:11:16.958080 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 18:11:16 crc kubenswrapper[4988]: I1008 18:11:16.958147 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/556fb048-8afb-44a5-9ea0-2b7c89e6cf4b-cnibin\") pod \"multus-additional-cni-plugins-jnh7m\" (UID: \"556fb048-8afb-44a5-9ea0-2b7c89e6cf4b\") " pod="openshift-multus/multus-additional-cni-plugins-jnh7m" Oct 08 18:11:16 crc kubenswrapper[4988]: I1008 18:11:16.958180 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/4490d846-ca32-4cd5-9202-87cc6d499fb4-cnibin\") pod \"multus-h9jwp\" (UID: \"4490d846-ca32-4cd5-9202-87cc6d499fb4\") " pod="openshift-multus/multus-h9jwp" Oct 08 18:11:16 crc kubenswrapper[4988]: I1008 18:11:16.958207 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4490d846-ca32-4cd5-9202-87cc6d499fb4-cni-binary-copy\") pod \"multus-h9jwp\" (UID: \"4490d846-ca32-4cd5-9202-87cc6d499fb4\") " pod="openshift-multus/multus-h9jwp" Oct 08 18:11:16 crc kubenswrapper[4988]: I1008 18:11:16.958233 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/4490d846-ca32-4cd5-9202-87cc6d499fb4-multus-socket-dir-parent\") pod \"multus-h9jwp\" (UID: \"4490d846-ca32-4cd5-9202-87cc6d499fb4\") " pod="openshift-multus/multus-h9jwp" Oct 08 18:11:16 crc kubenswrapper[4988]: I1008 18:11:16.958259 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/4490d846-ca32-4cd5-9202-87cc6d499fb4-host-var-lib-cni-bin\") pod \"multus-h9jwp\" (UID: \"4490d846-ca32-4cd5-9202-87cc6d499fb4\") " pod="openshift-multus/multus-h9jwp" Oct 08 18:11:16 crc kubenswrapper[4988]: I1008 18:11:16.958285 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/556fb048-8afb-44a5-9ea0-2b7c89e6cf4b-os-release\") pod \"multus-additional-cni-plugins-jnh7m\" (UID: \"556fb048-8afb-44a5-9ea0-2b7c89e6cf4b\") " pod="openshift-multus/multus-additional-cni-plugins-jnh7m" Oct 08 18:11:16 crc kubenswrapper[4988]: I1008 18:11:16.958308 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/f010a463-69f2-4082-acc3-161899a939c3-run-ovn\") pod \"ovnkube-node-nsc65\" (UID: \"f010a463-69f2-4082-acc3-161899a939c3\") " pod="openshift-ovn-kubernetes/ovnkube-node-nsc65" Oct 08 18:11:16 crc kubenswrapper[4988]: I1008 18:11:16.958334 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/f010a463-69f2-4082-acc3-161899a939c3-ovn-node-metrics-cert\") pod \"ovnkube-node-nsc65\" (UID: \"f010a463-69f2-4082-acc3-161899a939c3\") " pod="openshift-ovn-kubernetes/ovnkube-node-nsc65" Oct 08 18:11:16 crc kubenswrapper[4988]: I1008 18:11:16.958358 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/4490d846-ca32-4cd5-9202-87cc6d499fb4-host-run-netns\") pod \"multus-h9jwp\" (UID: \"4490d846-ca32-4cd5-9202-87cc6d499fb4\") " pod="openshift-multus/multus-h9jwp" Oct 08 18:11:16 crc kubenswrapper[4988]: I1008 18:11:16.958399 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/f328a20b-5283-407e-8a93-2ab2da599f18-mcd-auth-proxy-config\") pod \"machine-config-daemon-nm54f\" (UID: \"f328a20b-5283-407e-8a93-2ab2da599f18\") " pod="openshift-machine-config-operator/machine-config-daemon-nm54f" Oct 08 18:11:16 crc kubenswrapper[4988]: I1008 18:11:16.958424 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/4490d846-ca32-4cd5-9202-87cc6d499fb4-multus-conf-dir\") pod \"multus-h9jwp\" (UID: \"4490d846-ca32-4cd5-9202-87cc6d499fb4\") " pod="openshift-multus/multus-h9jwp" Oct 08 18:11:16 crc kubenswrapper[4988]: I1008 18:11:16.958461 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 18:11:16 crc kubenswrapper[4988]: I1008 18:11:16.958496 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z2gh4\" (UniqueName: \"kubernetes.io/projected/f328a20b-5283-407e-8a93-2ab2da599f18-kube-api-access-z2gh4\") pod \"machine-config-daemon-nm54f\" (UID: \"f328a20b-5283-407e-8a93-2ab2da599f18\") " pod="openshift-machine-config-operator/machine-config-daemon-nm54f" Oct 08 18:11:16 crc kubenswrapper[4988]: I1008 18:11:16.958521 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7k6rk\" (UniqueName: \"kubernetes.io/projected/4490d846-ca32-4cd5-9202-87cc6d499fb4-kube-api-access-7k6rk\") pod \"multus-h9jwp\" (UID: \"4490d846-ca32-4cd5-9202-87cc6d499fb4\") " pod="openshift-multus/multus-h9jwp" Oct 08 18:11:16 crc kubenswrapper[4988]: I1008 18:11:16.958547 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/556fb048-8afb-44a5-9ea0-2b7c89e6cf4b-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-jnh7m\" (UID: \"556fb048-8afb-44a5-9ea0-2b7c89e6cf4b\") " pod="openshift-multus/multus-additional-cni-plugins-jnh7m" Oct 08 18:11:16 crc kubenswrapper[4988]: I1008 18:11:16.958568 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/f010a463-69f2-4082-acc3-161899a939c3-systemd-units\") pod \"ovnkube-node-nsc65\" (UID: \"f010a463-69f2-4082-acc3-161899a939c3\") " pod="openshift-ovn-kubernetes/ovnkube-node-nsc65" Oct 08 18:11:16 crc kubenswrapper[4988]: I1008 18:11:16.958588 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/f010a463-69f2-4082-acc3-161899a939c3-run-systemd\") pod \"ovnkube-node-nsc65\" (UID: \"f010a463-69f2-4082-acc3-161899a939c3\") " pod="openshift-ovn-kubernetes/ovnkube-node-nsc65" Oct 08 18:11:16 crc kubenswrapper[4988]: I1008 18:11:16.958609 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/4490d846-ca32-4cd5-9202-87cc6d499fb4-os-release\") pod \"multus-h9jwp\" (UID: \"4490d846-ca32-4cd5-9202-87cc6d499fb4\") " pod="openshift-multus/multus-h9jwp" Oct 08 18:11:16 crc kubenswrapper[4988]: I1008 18:11:16.958634 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/f010a463-69f2-4082-acc3-161899a939c3-host-kubelet\") pod \"ovnkube-node-nsc65\" (UID: \"f010a463-69f2-4082-acc3-161899a939c3\") " pod="openshift-ovn-kubernetes/ovnkube-node-nsc65" Oct 08 18:11:16 crc kubenswrapper[4988]: I1008 18:11:16.958659 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/4490d846-ca32-4cd5-9202-87cc6d499fb4-multus-cni-dir\") pod \"multus-h9jwp\" (UID: \"4490d846-ca32-4cd5-9202-87cc6d499fb4\") " pod="openshift-multus/multus-h9jwp" Oct 08 18:11:16 crc kubenswrapper[4988]: I1008 18:11:16.958684 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gmlt2\" (UniqueName: \"kubernetes.io/projected/556fb048-8afb-44a5-9ea0-2b7c89e6cf4b-kube-api-access-gmlt2\") pod \"multus-additional-cni-plugins-jnh7m\" (UID: \"556fb048-8afb-44a5-9ea0-2b7c89e6cf4b\") " pod="openshift-multus/multus-additional-cni-plugins-jnh7m" Oct 08 18:11:16 crc kubenswrapper[4988]: I1008 18:11:16.958715 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/f328a20b-5283-407e-8a93-2ab2da599f18-proxy-tls\") pod \"machine-config-daemon-nm54f\" (UID: \"f328a20b-5283-407e-8a93-2ab2da599f18\") " pod="openshift-machine-config-operator/machine-config-daemon-nm54f" Oct 08 18:11:16 crc kubenswrapper[4988]: I1008 18:11:16.958737 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/4490d846-ca32-4cd5-9202-87cc6d499fb4-host-run-k8s-cni-cncf-io\") pod \"multus-h9jwp\" (UID: \"4490d846-ca32-4cd5-9202-87cc6d499fb4\") " pod="openshift-multus/multus-h9jwp" Oct 08 18:11:16 crc kubenswrapper[4988]: I1008 18:11:16.958759 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dnwgr\" (UniqueName: \"kubernetes.io/projected/f010a463-69f2-4082-acc3-161899a939c3-kube-api-access-dnwgr\") pod \"ovnkube-node-nsc65\" (UID: \"f010a463-69f2-4082-acc3-161899a939c3\") " pod="openshift-ovn-kubernetes/ovnkube-node-nsc65" Oct 08 18:11:16 crc kubenswrapper[4988]: I1008 18:11:16.958783 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/4490d846-ca32-4cd5-9202-87cc6d499fb4-host-var-lib-kubelet\") pod \"multus-h9jwp\" (UID: \"4490d846-ca32-4cd5-9202-87cc6d499fb4\") " pod="openshift-multus/multus-h9jwp" Oct 08 18:11:16 crc kubenswrapper[4988]: I1008 18:11:16.958805 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/4490d846-ca32-4cd5-9202-87cc6d499fb4-etc-kubernetes\") pod \"multus-h9jwp\" (UID: \"4490d846-ca32-4cd5-9202-87cc6d499fb4\") " pod="openshift-multus/multus-h9jwp" Oct 08 18:11:16 crc kubenswrapper[4988]: I1008 18:11:16.958825 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/f010a463-69f2-4082-acc3-161899a939c3-host-slash\") pod \"ovnkube-node-nsc65\" (UID: \"f010a463-69f2-4082-acc3-161899a939c3\") " pod="openshift-ovn-kubernetes/ovnkube-node-nsc65" Oct 08 18:11:16 crc kubenswrapper[4988]: I1008 18:11:16.958846 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/f010a463-69f2-4082-acc3-161899a939c3-log-socket\") pod \"ovnkube-node-nsc65\" (UID: \"f010a463-69f2-4082-acc3-161899a939c3\") " pod="openshift-ovn-kubernetes/ovnkube-node-nsc65" Oct 08 18:11:16 crc kubenswrapper[4988]: I1008 18:11:16.958871 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/f010a463-69f2-4082-acc3-161899a939c3-env-overrides\") pod \"ovnkube-node-nsc65\" (UID: \"f010a463-69f2-4082-acc3-161899a939c3\") " pod="openshift-ovn-kubernetes/ovnkube-node-nsc65" Oct 08 18:11:16 crc kubenswrapper[4988]: I1008 18:11:16.958912 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/556fb048-8afb-44a5-9ea0-2b7c89e6cf4b-cni-binary-copy\") pod \"multus-additional-cni-plugins-jnh7m\" (UID: \"556fb048-8afb-44a5-9ea0-2b7c89e6cf4b\") " pod="openshift-multus/multus-additional-cni-plugins-jnh7m" Oct 08 18:11:16 crc kubenswrapper[4988]: I1008 18:11:16.958932 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/f010a463-69f2-4082-acc3-161899a939c3-host-run-netns\") pod \"ovnkube-node-nsc65\" (UID: \"f010a463-69f2-4082-acc3-161899a939c3\") " pod="openshift-ovn-kubernetes/ovnkube-node-nsc65" Oct 08 18:11:16 crc kubenswrapper[4988]: I1008 18:11:16.958955 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/f010a463-69f2-4082-acc3-161899a939c3-etc-openvswitch\") pod \"ovnkube-node-nsc65\" (UID: \"f010a463-69f2-4082-acc3-161899a939c3\") " pod="openshift-ovn-kubernetes/ovnkube-node-nsc65" Oct 08 18:11:16 crc kubenswrapper[4988]: I1008 18:11:16.958975 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/f010a463-69f2-4082-acc3-161899a939c3-run-openvswitch\") pod \"ovnkube-node-nsc65\" (UID: \"f010a463-69f2-4082-acc3-161899a939c3\") " pod="openshift-ovn-kubernetes/ovnkube-node-nsc65" Oct 08 18:11:16 crc kubenswrapper[4988]: I1008 18:11:16.958995 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/4490d846-ca32-4cd5-9202-87cc6d499fb4-hostroot\") pod \"multus-h9jwp\" (UID: \"4490d846-ca32-4cd5-9202-87cc6d499fb4\") " pod="openshift-multus/multus-h9jwp" Oct 08 18:11:16 crc kubenswrapper[4988]: I1008 18:11:16.959013 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/556fb048-8afb-44a5-9ea0-2b7c89e6cf4b-system-cni-dir\") pod \"multus-additional-cni-plugins-jnh7m\" (UID: \"556fb048-8afb-44a5-9ea0-2b7c89e6cf4b\") " pod="openshift-multus/multus-additional-cni-plugins-jnh7m" Oct 08 18:11:16 crc kubenswrapper[4988]: I1008 18:11:16.959033 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/f010a463-69f2-4082-acc3-161899a939c3-var-lib-openvswitch\") pod \"ovnkube-node-nsc65\" (UID: \"f010a463-69f2-4082-acc3-161899a939c3\") " pod="openshift-ovn-kubernetes/ovnkube-node-nsc65" Oct 08 18:11:16 crc kubenswrapper[4988]: I1008 18:11:16.959053 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/f010a463-69f2-4082-acc3-161899a939c3-node-log\") pod \"ovnkube-node-nsc65\" (UID: \"f010a463-69f2-4082-acc3-161899a939c3\") " pod="openshift-ovn-kubernetes/ovnkube-node-nsc65" Oct 08 18:11:16 crc kubenswrapper[4988]: I1008 18:11:16.959076 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/f010a463-69f2-4082-acc3-161899a939c3-host-cni-netd\") pod \"ovnkube-node-nsc65\" (UID: \"f010a463-69f2-4082-acc3-161899a939c3\") " pod="openshift-ovn-kubernetes/ovnkube-node-nsc65" Oct 08 18:11:16 crc kubenswrapper[4988]: I1008 18:11:16.959105 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/f010a463-69f2-4082-acc3-161899a939c3-ovnkube-config\") pod \"ovnkube-node-nsc65\" (UID: \"f010a463-69f2-4082-acc3-161899a939c3\") " pod="openshift-ovn-kubernetes/ovnkube-node-nsc65" Oct 08 18:11:16 crc kubenswrapper[4988]: I1008 18:11:16.959127 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/f010a463-69f2-4082-acc3-161899a939c3-ovnkube-script-lib\") pod \"ovnkube-node-nsc65\" (UID: \"f010a463-69f2-4082-acc3-161899a939c3\") " pod="openshift-ovn-kubernetes/ovnkube-node-nsc65" Oct 08 18:11:16 crc kubenswrapper[4988]: I1008 18:11:16.959149 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/f328a20b-5283-407e-8a93-2ab2da599f18-rootfs\") pod \"machine-config-daemon-nm54f\" (UID: \"f328a20b-5283-407e-8a93-2ab2da599f18\") " pod="openshift-machine-config-operator/machine-config-daemon-nm54f" Oct 08 18:11:16 crc kubenswrapper[4988]: I1008 18:11:16.959171 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/4490d846-ca32-4cd5-9202-87cc6d499fb4-system-cni-dir\") pod \"multus-h9jwp\" (UID: \"4490d846-ca32-4cd5-9202-87cc6d499fb4\") " pod="openshift-multus/multus-h9jwp" Oct 08 18:11:16 crc kubenswrapper[4988]: I1008 18:11:16.959192 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4490d846-ca32-4cd5-9202-87cc6d499fb4-multus-daemon-config\") pod \"multus-h9jwp\" (UID: \"4490d846-ca32-4cd5-9202-87cc6d499fb4\") " pod="openshift-multus/multus-h9jwp" Oct 08 18:11:16 crc kubenswrapper[4988]: I1008 18:11:16.959215 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/4490d846-ca32-4cd5-9202-87cc6d499fb4-host-run-multus-certs\") pod \"multus-h9jwp\" (UID: \"4490d846-ca32-4cd5-9202-87cc6d499fb4\") " pod="openshift-multus/multus-h9jwp" Oct 08 18:11:16 crc kubenswrapper[4988]: I1008 18:11:16.959237 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/f010a463-69f2-4082-acc3-161899a939c3-host-run-ovn-kubernetes\") pod \"ovnkube-node-nsc65\" (UID: \"f010a463-69f2-4082-acc3-161899a939c3\") " pod="openshift-ovn-kubernetes/ovnkube-node-nsc65" Oct 08 18:11:16 crc kubenswrapper[4988]: I1008 18:11:16.959261 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/f010a463-69f2-4082-acc3-161899a939c3-host-cni-bin\") pod \"ovnkube-node-nsc65\" (UID: \"f010a463-69f2-4082-acc3-161899a939c3\") " pod="openshift-ovn-kubernetes/ovnkube-node-nsc65" Oct 08 18:11:16 crc kubenswrapper[4988]: I1008 18:11:16.959284 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/4490d846-ca32-4cd5-9202-87cc6d499fb4-host-var-lib-cni-multus\") pod \"multus-h9jwp\" (UID: \"4490d846-ca32-4cd5-9202-87cc6d499fb4\") " pod="openshift-multus/multus-h9jwp" Oct 08 18:11:16 crc kubenswrapper[4988]: I1008 18:11:16.959308 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/556fb048-8afb-44a5-9ea0-2b7c89e6cf4b-tuning-conf-dir\") pod \"multus-additional-cni-plugins-jnh7m\" (UID: \"556fb048-8afb-44a5-9ea0-2b7c89e6cf4b\") " pod="openshift-multus/multus-additional-cni-plugins-jnh7m" Oct 08 18:11:16 crc kubenswrapper[4988]: I1008 18:11:16.959331 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/f010a463-69f2-4082-acc3-161899a939c3-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-nsc65\" (UID: \"f010a463-69f2-4082-acc3-161899a939c3\") " pod="openshift-ovn-kubernetes/ovnkube-node-nsc65" Oct 08 18:11:16 crc kubenswrapper[4988]: E1008 18:11:16.959552 4988 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 08 18:11:16 crc kubenswrapper[4988]: E1008 18:11:16.959575 4988 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 08 18:11:16 crc kubenswrapper[4988]: E1008 18:11:16.959589 4988 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 08 18:11:16 crc kubenswrapper[4988]: E1008 18:11:16.959649 4988 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-08 18:11:18.959624686 +0000 UTC m=+24.409467456 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 08 18:11:16 crc kubenswrapper[4988]: E1008 18:11:16.959842 4988 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 08 18:11:16 crc kubenswrapper[4988]: E1008 18:11:16.959858 4988 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 08 18:11:16 crc kubenswrapper[4988]: E1008 18:11:16.959869 4988 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 08 18:11:16 crc kubenswrapper[4988]: E1008 18:11:16.959909 4988 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-08 18:11:18.959897435 +0000 UTC m=+24.409740205 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 08 18:11:16 crc kubenswrapper[4988]: I1008 18:11:16.972964 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e8906182d55c285b3cf095e2d9e1dac397af4618a514f11e26b44c172dd573f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:16Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:16 crc kubenswrapper[4988]: I1008 18:11:16.998104 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-dzxs2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8f307b20-5e33-4725-99cb-210e25113532\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q65f7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:11:16Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-dzxs2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:16Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:17 crc kubenswrapper[4988]: I1008 18:11:17.016255 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aef71be1-1d12-46e5-95b2-bd399040fd4d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:10:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:10:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:10:55Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:10:55Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:10:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77c68f3ecb8a0724499e495cb26ef3db5beecb140f2558d8c0b25f41928b84ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:10:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a560ee84c2405227de71fda127566213cd4f017e73d5581442ba267781b1c539\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:10:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6100accc3bd08928bc1ff88310f29fa5da37fd2d5c315c94279801c9778c4745\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:10:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b0e596d7ce4e9341c44499af79d3b3caf0e5e2c156902ea40cf2a3dca672be9e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5bd35aef24be034a0b305a40bb53ca171a4f76db3b93b4f78ae95099b0a4cde3\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-08T18:11:14Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1008 18:11:09.091034 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1008 18:11:09.091984 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3472374062/tls.crt::/tmp/serving-cert-3472374062/tls.key\\\\\\\"\\\\nI1008 18:11:14.619696 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1008 18:11:14.628837 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1008 18:11:14.628898 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1008 18:11:14.628994 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1008 18:11:14.629036 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1008 18:11:14.641318 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1008 18:11:14.641363 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1008 18:11:14.641372 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1008 18:11:14.641382 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1008 18:11:14.641415 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1008 18:11:14.641423 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1008 18:11:14.641431 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1008 18:11:14.641484 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1008 18:11:14.646451 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T18:10:58Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c79ef704fda1bbcc486f4c7afa3039a3ab38669833f0f66bafe24dd4cc679e60\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:10:58Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://28d21731af0f16d6a8ae6c6e9d0bebf0560d4a533db4c78c10bf1837890394ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://28d21731af0f16d6a8ae6c6e9d0bebf0560d4a533db4c78c10bf1837890394ef\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:10:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:10:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:10:55Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:17Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:17 crc kubenswrapper[4988]: I1008 18:11:17.042326 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a35a5e989ba69c8ef36ce4cea272aadf67ce20473fc257adfcfa86a2faba6ae0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a522e937cfe03fb9f2e60cb1bea07214cf63c69bdecd5ba65a2dfdba882d28dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:17Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:17 crc kubenswrapper[4988]: I1008 18:11:17.060283 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:17Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:17 crc kubenswrapper[4988]: I1008 18:11:17.060554 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/4490d846-ca32-4cd5-9202-87cc6d499fb4-hostroot\") pod \"multus-h9jwp\" (UID: \"4490d846-ca32-4cd5-9202-87cc6d499fb4\") " pod="openshift-multus/multus-h9jwp" Oct 08 18:11:17 crc kubenswrapper[4988]: I1008 18:11:17.060599 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/556fb048-8afb-44a5-9ea0-2b7c89e6cf4b-system-cni-dir\") pod \"multus-additional-cni-plugins-jnh7m\" (UID: \"556fb048-8afb-44a5-9ea0-2b7c89e6cf4b\") " pod="openshift-multus/multus-additional-cni-plugins-jnh7m" Oct 08 18:11:17 crc kubenswrapper[4988]: I1008 18:11:17.060623 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/f010a463-69f2-4082-acc3-161899a939c3-var-lib-openvswitch\") pod \"ovnkube-node-nsc65\" (UID: \"f010a463-69f2-4082-acc3-161899a939c3\") " pod="openshift-ovn-kubernetes/ovnkube-node-nsc65" Oct 08 18:11:17 crc kubenswrapper[4988]: I1008 18:11:17.060647 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/f010a463-69f2-4082-acc3-161899a939c3-ovnkube-config\") pod \"ovnkube-node-nsc65\" (UID: \"f010a463-69f2-4082-acc3-161899a939c3\") " pod="openshift-ovn-kubernetes/ovnkube-node-nsc65" Oct 08 18:11:17 crc kubenswrapper[4988]: I1008 18:11:17.060668 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/f010a463-69f2-4082-acc3-161899a939c3-ovnkube-script-lib\") pod \"ovnkube-node-nsc65\" (UID: \"f010a463-69f2-4082-acc3-161899a939c3\") " pod="openshift-ovn-kubernetes/ovnkube-node-nsc65" Oct 08 18:11:17 crc kubenswrapper[4988]: I1008 18:11:17.060683 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/4490d846-ca32-4cd5-9202-87cc6d499fb4-hostroot\") pod \"multus-h9jwp\" (UID: \"4490d846-ca32-4cd5-9202-87cc6d499fb4\") " pod="openshift-multus/multus-h9jwp" Oct 08 18:11:17 crc kubenswrapper[4988]: I1008 18:11:17.060691 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/f010a463-69f2-4082-acc3-161899a939c3-node-log\") pod \"ovnkube-node-nsc65\" (UID: \"f010a463-69f2-4082-acc3-161899a939c3\") " pod="openshift-ovn-kubernetes/ovnkube-node-nsc65" Oct 08 18:11:17 crc kubenswrapper[4988]: I1008 18:11:17.060751 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/f010a463-69f2-4082-acc3-161899a939c3-host-cni-netd\") pod \"ovnkube-node-nsc65\" (UID: \"f010a463-69f2-4082-acc3-161899a939c3\") " pod="openshift-ovn-kubernetes/ovnkube-node-nsc65" Oct 08 18:11:17 crc kubenswrapper[4988]: I1008 18:11:17.060766 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/f010a463-69f2-4082-acc3-161899a939c3-var-lib-openvswitch\") pod \"ovnkube-node-nsc65\" (UID: \"f010a463-69f2-4082-acc3-161899a939c3\") " pod="openshift-ovn-kubernetes/ovnkube-node-nsc65" Oct 08 18:11:17 crc kubenswrapper[4988]: I1008 18:11:17.060777 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/4490d846-ca32-4cd5-9202-87cc6d499fb4-host-run-multus-certs\") pod \"multus-h9jwp\" (UID: \"4490d846-ca32-4cd5-9202-87cc6d499fb4\") " pod="openshift-multus/multus-h9jwp" Oct 08 18:11:17 crc kubenswrapper[4988]: I1008 18:11:17.060804 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/f010a463-69f2-4082-acc3-161899a939c3-host-run-ovn-kubernetes\") pod \"ovnkube-node-nsc65\" (UID: \"f010a463-69f2-4082-acc3-161899a939c3\") " pod="openshift-ovn-kubernetes/ovnkube-node-nsc65" Oct 08 18:11:17 crc kubenswrapper[4988]: I1008 18:11:17.060809 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/556fb048-8afb-44a5-9ea0-2b7c89e6cf4b-system-cni-dir\") pod \"multus-additional-cni-plugins-jnh7m\" (UID: \"556fb048-8afb-44a5-9ea0-2b7c89e6cf4b\") " pod="openshift-multus/multus-additional-cni-plugins-jnh7m" Oct 08 18:11:17 crc kubenswrapper[4988]: I1008 18:11:17.060825 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/f010a463-69f2-4082-acc3-161899a939c3-host-cni-bin\") pod \"ovnkube-node-nsc65\" (UID: \"f010a463-69f2-4082-acc3-161899a939c3\") " pod="openshift-ovn-kubernetes/ovnkube-node-nsc65" Oct 08 18:11:17 crc kubenswrapper[4988]: I1008 18:11:17.060845 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/f010a463-69f2-4082-acc3-161899a939c3-node-log\") pod \"ovnkube-node-nsc65\" (UID: \"f010a463-69f2-4082-acc3-161899a939c3\") " pod="openshift-ovn-kubernetes/ovnkube-node-nsc65" Oct 08 18:11:17 crc kubenswrapper[4988]: I1008 18:11:17.060858 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/f328a20b-5283-407e-8a93-2ab2da599f18-rootfs\") pod \"machine-config-daemon-nm54f\" (UID: \"f328a20b-5283-407e-8a93-2ab2da599f18\") " pod="openshift-machine-config-operator/machine-config-daemon-nm54f" Oct 08 18:11:17 crc kubenswrapper[4988]: I1008 18:11:17.060883 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/4490d846-ca32-4cd5-9202-87cc6d499fb4-system-cni-dir\") pod \"multus-h9jwp\" (UID: \"4490d846-ca32-4cd5-9202-87cc6d499fb4\") " pod="openshift-multus/multus-h9jwp" Oct 08 18:11:17 crc kubenswrapper[4988]: I1008 18:11:17.060907 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4490d846-ca32-4cd5-9202-87cc6d499fb4-multus-daemon-config\") pod \"multus-h9jwp\" (UID: \"4490d846-ca32-4cd5-9202-87cc6d499fb4\") " pod="openshift-multus/multus-h9jwp" Oct 08 18:11:17 crc kubenswrapper[4988]: I1008 18:11:17.060936 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/4490d846-ca32-4cd5-9202-87cc6d499fb4-host-var-lib-cni-multus\") pod \"multus-h9jwp\" (UID: \"4490d846-ca32-4cd5-9202-87cc6d499fb4\") " pod="openshift-multus/multus-h9jwp" Oct 08 18:11:17 crc kubenswrapper[4988]: I1008 18:11:17.060960 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/556fb048-8afb-44a5-9ea0-2b7c89e6cf4b-tuning-conf-dir\") pod \"multus-additional-cni-plugins-jnh7m\" (UID: \"556fb048-8afb-44a5-9ea0-2b7c89e6cf4b\") " pod="openshift-multus/multus-additional-cni-plugins-jnh7m" Oct 08 18:11:17 crc kubenswrapper[4988]: I1008 18:11:17.060984 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/f010a463-69f2-4082-acc3-161899a939c3-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-nsc65\" (UID: \"f010a463-69f2-4082-acc3-161899a939c3\") " pod="openshift-ovn-kubernetes/ovnkube-node-nsc65" Oct 08 18:11:17 crc kubenswrapper[4988]: I1008 18:11:17.061021 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4490d846-ca32-4cd5-9202-87cc6d499fb4-cni-binary-copy\") pod \"multus-h9jwp\" (UID: \"4490d846-ca32-4cd5-9202-87cc6d499fb4\") " pod="openshift-multus/multus-h9jwp" Oct 08 18:11:17 crc kubenswrapper[4988]: I1008 18:11:17.061041 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/4490d846-ca32-4cd5-9202-87cc6d499fb4-multus-socket-dir-parent\") pod \"multus-h9jwp\" (UID: \"4490d846-ca32-4cd5-9202-87cc6d499fb4\") " pod="openshift-multus/multus-h9jwp" Oct 08 18:11:17 crc kubenswrapper[4988]: I1008 18:11:17.061065 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/4490d846-ca32-4cd5-9202-87cc6d499fb4-host-var-lib-cni-bin\") pod \"multus-h9jwp\" (UID: \"4490d846-ca32-4cd5-9202-87cc6d499fb4\") " pod="openshift-multus/multus-h9jwp" Oct 08 18:11:17 crc kubenswrapper[4988]: I1008 18:11:17.061086 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/556fb048-8afb-44a5-9ea0-2b7c89e6cf4b-cnibin\") pod \"multus-additional-cni-plugins-jnh7m\" (UID: \"556fb048-8afb-44a5-9ea0-2b7c89e6cf4b\") " pod="openshift-multus/multus-additional-cni-plugins-jnh7m" Oct 08 18:11:17 crc kubenswrapper[4988]: I1008 18:11:17.061109 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/4490d846-ca32-4cd5-9202-87cc6d499fb4-cnibin\") pod \"multus-h9jwp\" (UID: \"4490d846-ca32-4cd5-9202-87cc6d499fb4\") " pod="openshift-multus/multus-h9jwp" Oct 08 18:11:17 crc kubenswrapper[4988]: I1008 18:11:17.061130 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/556fb048-8afb-44a5-9ea0-2b7c89e6cf4b-os-release\") pod \"multus-additional-cni-plugins-jnh7m\" (UID: \"556fb048-8afb-44a5-9ea0-2b7c89e6cf4b\") " pod="openshift-multus/multus-additional-cni-plugins-jnh7m" Oct 08 18:11:17 crc kubenswrapper[4988]: I1008 18:11:17.061150 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/f010a463-69f2-4082-acc3-161899a939c3-run-ovn\") pod \"ovnkube-node-nsc65\" (UID: \"f010a463-69f2-4082-acc3-161899a939c3\") " pod="openshift-ovn-kubernetes/ovnkube-node-nsc65" Oct 08 18:11:17 crc kubenswrapper[4988]: I1008 18:11:17.061170 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/f010a463-69f2-4082-acc3-161899a939c3-ovn-node-metrics-cert\") pod \"ovnkube-node-nsc65\" (UID: \"f010a463-69f2-4082-acc3-161899a939c3\") " pod="openshift-ovn-kubernetes/ovnkube-node-nsc65" Oct 08 18:11:17 crc kubenswrapper[4988]: I1008 18:11:17.061190 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/f328a20b-5283-407e-8a93-2ab2da599f18-mcd-auth-proxy-config\") pod \"machine-config-daemon-nm54f\" (UID: \"f328a20b-5283-407e-8a93-2ab2da599f18\") " pod="openshift-machine-config-operator/machine-config-daemon-nm54f" Oct 08 18:11:17 crc kubenswrapper[4988]: I1008 18:11:17.061210 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/4490d846-ca32-4cd5-9202-87cc6d499fb4-host-run-netns\") pod \"multus-h9jwp\" (UID: \"4490d846-ca32-4cd5-9202-87cc6d499fb4\") " pod="openshift-multus/multus-h9jwp" Oct 08 18:11:17 crc kubenswrapper[4988]: I1008 18:11:17.061240 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/4490d846-ca32-4cd5-9202-87cc6d499fb4-multus-conf-dir\") pod \"multus-h9jwp\" (UID: \"4490d846-ca32-4cd5-9202-87cc6d499fb4\") " pod="openshift-multus/multus-h9jwp" Oct 08 18:11:17 crc kubenswrapper[4988]: I1008 18:11:17.061271 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z2gh4\" (UniqueName: \"kubernetes.io/projected/f328a20b-5283-407e-8a93-2ab2da599f18-kube-api-access-z2gh4\") pod \"machine-config-daemon-nm54f\" (UID: \"f328a20b-5283-407e-8a93-2ab2da599f18\") " pod="openshift-machine-config-operator/machine-config-daemon-nm54f" Oct 08 18:11:17 crc kubenswrapper[4988]: I1008 18:11:17.061293 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/f010a463-69f2-4082-acc3-161899a939c3-systemd-units\") pod \"ovnkube-node-nsc65\" (UID: \"f010a463-69f2-4082-acc3-161899a939c3\") " pod="openshift-ovn-kubernetes/ovnkube-node-nsc65" Oct 08 18:11:17 crc kubenswrapper[4988]: I1008 18:11:17.061313 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/f010a463-69f2-4082-acc3-161899a939c3-run-systemd\") pod \"ovnkube-node-nsc65\" (UID: \"f010a463-69f2-4082-acc3-161899a939c3\") " pod="openshift-ovn-kubernetes/ovnkube-node-nsc65" Oct 08 18:11:17 crc kubenswrapper[4988]: I1008 18:11:17.061333 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7k6rk\" (UniqueName: \"kubernetes.io/projected/4490d846-ca32-4cd5-9202-87cc6d499fb4-kube-api-access-7k6rk\") pod \"multus-h9jwp\" (UID: \"4490d846-ca32-4cd5-9202-87cc6d499fb4\") " pod="openshift-multus/multus-h9jwp" Oct 08 18:11:17 crc kubenswrapper[4988]: I1008 18:11:17.061360 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/556fb048-8afb-44a5-9ea0-2b7c89e6cf4b-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-jnh7m\" (UID: \"556fb048-8afb-44a5-9ea0-2b7c89e6cf4b\") " pod="openshift-multus/multus-additional-cni-plugins-jnh7m" Oct 08 18:11:17 crc kubenswrapper[4988]: I1008 18:11:17.061401 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/4490d846-ca32-4cd5-9202-87cc6d499fb4-os-release\") pod \"multus-h9jwp\" (UID: \"4490d846-ca32-4cd5-9202-87cc6d499fb4\") " pod="openshift-multus/multus-h9jwp" Oct 08 18:11:17 crc kubenswrapper[4988]: I1008 18:11:17.061425 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/f010a463-69f2-4082-acc3-161899a939c3-host-kubelet\") pod \"ovnkube-node-nsc65\" (UID: \"f010a463-69f2-4082-acc3-161899a939c3\") " pod="openshift-ovn-kubernetes/ovnkube-node-nsc65" Oct 08 18:11:17 crc kubenswrapper[4988]: I1008 18:11:17.061448 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gmlt2\" (UniqueName: \"kubernetes.io/projected/556fb048-8afb-44a5-9ea0-2b7c89e6cf4b-kube-api-access-gmlt2\") pod \"multus-additional-cni-plugins-jnh7m\" (UID: \"556fb048-8afb-44a5-9ea0-2b7c89e6cf4b\") " pod="openshift-multus/multus-additional-cni-plugins-jnh7m" Oct 08 18:11:17 crc kubenswrapper[4988]: I1008 18:11:17.061469 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/f328a20b-5283-407e-8a93-2ab2da599f18-proxy-tls\") pod \"machine-config-daemon-nm54f\" (UID: \"f328a20b-5283-407e-8a93-2ab2da599f18\") " pod="openshift-machine-config-operator/machine-config-daemon-nm54f" Oct 08 18:11:17 crc kubenswrapper[4988]: I1008 18:11:17.061492 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/4490d846-ca32-4cd5-9202-87cc6d499fb4-multus-cni-dir\") pod \"multus-h9jwp\" (UID: \"4490d846-ca32-4cd5-9202-87cc6d499fb4\") " pod="openshift-multus/multus-h9jwp" Oct 08 18:11:17 crc kubenswrapper[4988]: I1008 18:11:17.061512 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/4490d846-ca32-4cd5-9202-87cc6d499fb4-host-run-k8s-cni-cncf-io\") pod \"multus-h9jwp\" (UID: \"4490d846-ca32-4cd5-9202-87cc6d499fb4\") " pod="openshift-multus/multus-h9jwp" Oct 08 18:11:17 crc kubenswrapper[4988]: I1008 18:11:17.061566 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/f010a463-69f2-4082-acc3-161899a939c3-ovnkube-config\") pod \"ovnkube-node-nsc65\" (UID: \"f010a463-69f2-4082-acc3-161899a939c3\") " pod="openshift-ovn-kubernetes/ovnkube-node-nsc65" Oct 08 18:11:17 crc kubenswrapper[4988]: I1008 18:11:17.061536 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dnwgr\" (UniqueName: \"kubernetes.io/projected/f010a463-69f2-4082-acc3-161899a939c3-kube-api-access-dnwgr\") pod \"ovnkube-node-nsc65\" (UID: \"f010a463-69f2-4082-acc3-161899a939c3\") " pod="openshift-ovn-kubernetes/ovnkube-node-nsc65" Oct 08 18:11:17 crc kubenswrapper[4988]: I1008 18:11:17.061643 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/f010a463-69f2-4082-acc3-161899a939c3-log-socket\") pod \"ovnkube-node-nsc65\" (UID: \"f010a463-69f2-4082-acc3-161899a939c3\") " pod="openshift-ovn-kubernetes/ovnkube-node-nsc65" Oct 08 18:11:17 crc kubenswrapper[4988]: I1008 18:11:17.061659 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/f010a463-69f2-4082-acc3-161899a939c3-env-overrides\") pod \"ovnkube-node-nsc65\" (UID: \"f010a463-69f2-4082-acc3-161899a939c3\") " pod="openshift-ovn-kubernetes/ovnkube-node-nsc65" Oct 08 18:11:17 crc kubenswrapper[4988]: I1008 18:11:17.061682 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/4490d846-ca32-4cd5-9202-87cc6d499fb4-host-var-lib-kubelet\") pod \"multus-h9jwp\" (UID: \"4490d846-ca32-4cd5-9202-87cc6d499fb4\") " pod="openshift-multus/multus-h9jwp" Oct 08 18:11:17 crc kubenswrapper[4988]: I1008 18:11:17.061705 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/4490d846-ca32-4cd5-9202-87cc6d499fb4-etc-kubernetes\") pod \"multus-h9jwp\" (UID: \"4490d846-ca32-4cd5-9202-87cc6d499fb4\") " pod="openshift-multus/multus-h9jwp" Oct 08 18:11:17 crc kubenswrapper[4988]: I1008 18:11:17.061728 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/f010a463-69f2-4082-acc3-161899a939c3-host-slash\") pod \"ovnkube-node-nsc65\" (UID: \"f010a463-69f2-4082-acc3-161899a939c3\") " pod="openshift-ovn-kubernetes/ovnkube-node-nsc65" Oct 08 18:11:17 crc kubenswrapper[4988]: I1008 18:11:17.061749 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/f010a463-69f2-4082-acc3-161899a939c3-host-run-netns\") pod \"ovnkube-node-nsc65\" (UID: \"f010a463-69f2-4082-acc3-161899a939c3\") " pod="openshift-ovn-kubernetes/ovnkube-node-nsc65" Oct 08 18:11:17 crc kubenswrapper[4988]: I1008 18:11:17.061773 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/f010a463-69f2-4082-acc3-161899a939c3-etc-openvswitch\") pod \"ovnkube-node-nsc65\" (UID: \"f010a463-69f2-4082-acc3-161899a939c3\") " pod="openshift-ovn-kubernetes/ovnkube-node-nsc65" Oct 08 18:11:17 crc kubenswrapper[4988]: I1008 18:11:17.061798 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/f010a463-69f2-4082-acc3-161899a939c3-run-openvswitch\") pod \"ovnkube-node-nsc65\" (UID: \"f010a463-69f2-4082-acc3-161899a939c3\") " pod="openshift-ovn-kubernetes/ovnkube-node-nsc65" Oct 08 18:11:17 crc kubenswrapper[4988]: I1008 18:11:17.061827 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/4490d846-ca32-4cd5-9202-87cc6d499fb4-host-var-lib-kubelet\") pod \"multus-h9jwp\" (UID: \"4490d846-ca32-4cd5-9202-87cc6d499fb4\") " pod="openshift-multus/multus-h9jwp" Oct 08 18:11:17 crc kubenswrapper[4988]: I1008 18:11:17.061847 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/556fb048-8afb-44a5-9ea0-2b7c89e6cf4b-cni-binary-copy\") pod \"multus-additional-cni-plugins-jnh7m\" (UID: \"556fb048-8afb-44a5-9ea0-2b7c89e6cf4b\") " pod="openshift-multus/multus-additional-cni-plugins-jnh7m" Oct 08 18:11:17 crc kubenswrapper[4988]: I1008 18:11:17.061962 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/556fb048-8afb-44a5-9ea0-2b7c89e6cf4b-os-release\") pod \"multus-additional-cni-plugins-jnh7m\" (UID: \"556fb048-8afb-44a5-9ea0-2b7c89e6cf4b\") " pod="openshift-multus/multus-additional-cni-plugins-jnh7m" Oct 08 18:11:17 crc kubenswrapper[4988]: I1008 18:11:17.061973 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/4490d846-ca32-4cd5-9202-87cc6d499fb4-etc-kubernetes\") pod \"multus-h9jwp\" (UID: \"4490d846-ca32-4cd5-9202-87cc6d499fb4\") " pod="openshift-multus/multus-h9jwp" Oct 08 18:11:17 crc kubenswrapper[4988]: I1008 18:11:17.061988 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/f010a463-69f2-4082-acc3-161899a939c3-host-slash\") pod \"ovnkube-node-nsc65\" (UID: \"f010a463-69f2-4082-acc3-161899a939c3\") " pod="openshift-ovn-kubernetes/ovnkube-node-nsc65" Oct 08 18:11:17 crc kubenswrapper[4988]: I1008 18:11:17.062024 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/f010a463-69f2-4082-acc3-161899a939c3-run-openvswitch\") pod \"ovnkube-node-nsc65\" (UID: \"f010a463-69f2-4082-acc3-161899a939c3\") " pod="openshift-ovn-kubernetes/ovnkube-node-nsc65" Oct 08 18:11:17 crc kubenswrapper[4988]: I1008 18:11:17.062000 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/f010a463-69f2-4082-acc3-161899a939c3-etc-openvswitch\") pod \"ovnkube-node-nsc65\" (UID: \"f010a463-69f2-4082-acc3-161899a939c3\") " pod="openshift-ovn-kubernetes/ovnkube-node-nsc65" Oct 08 18:11:17 crc kubenswrapper[4988]: I1008 18:11:17.061988 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/4490d846-ca32-4cd5-9202-87cc6d499fb4-host-run-multus-certs\") pod \"multus-h9jwp\" (UID: \"4490d846-ca32-4cd5-9202-87cc6d499fb4\") " pod="openshift-multus/multus-h9jwp" Oct 08 18:11:17 crc kubenswrapper[4988]: I1008 18:11:17.062027 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/f010a463-69f2-4082-acc3-161899a939c3-host-run-ovn-kubernetes\") pod \"ovnkube-node-nsc65\" (UID: \"f010a463-69f2-4082-acc3-161899a939c3\") " pod="openshift-ovn-kubernetes/ovnkube-node-nsc65" Oct 08 18:11:17 crc kubenswrapper[4988]: I1008 18:11:17.062062 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/f010a463-69f2-4082-acc3-161899a939c3-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-nsc65\" (UID: \"f010a463-69f2-4082-acc3-161899a939c3\") " pod="openshift-ovn-kubernetes/ovnkube-node-nsc65" Oct 08 18:11:17 crc kubenswrapper[4988]: I1008 18:11:17.062078 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/4490d846-ca32-4cd5-9202-87cc6d499fb4-host-var-lib-cni-bin\") pod \"multus-h9jwp\" (UID: \"4490d846-ca32-4cd5-9202-87cc6d499fb4\") " pod="openshift-multus/multus-h9jwp" Oct 08 18:11:17 crc kubenswrapper[4988]: I1008 18:11:17.062052 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/556fb048-8afb-44a5-9ea0-2b7c89e6cf4b-cnibin\") pod \"multus-additional-cni-plugins-jnh7m\" (UID: \"556fb048-8afb-44a5-9ea0-2b7c89e6cf4b\") " pod="openshift-multus/multus-additional-cni-plugins-jnh7m" Oct 08 18:11:17 crc kubenswrapper[4988]: I1008 18:11:17.062070 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/4490d846-ca32-4cd5-9202-87cc6d499fb4-multus-socket-dir-parent\") pod \"multus-h9jwp\" (UID: \"4490d846-ca32-4cd5-9202-87cc6d499fb4\") " pod="openshift-multus/multus-h9jwp" Oct 08 18:11:17 crc kubenswrapper[4988]: I1008 18:11:17.062122 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/4490d846-ca32-4cd5-9202-87cc6d499fb4-cnibin\") pod \"multus-h9jwp\" (UID: \"4490d846-ca32-4cd5-9202-87cc6d499fb4\") " pod="openshift-multus/multus-h9jwp" Oct 08 18:11:17 crc kubenswrapper[4988]: I1008 18:11:17.062173 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/4490d846-ca32-4cd5-9202-87cc6d499fb4-os-release\") pod \"multus-h9jwp\" (UID: \"4490d846-ca32-4cd5-9202-87cc6d499fb4\") " pod="openshift-multus/multus-h9jwp" Oct 08 18:11:17 crc kubenswrapper[4988]: I1008 18:11:17.062168 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/4490d846-ca32-4cd5-9202-87cc6d499fb4-multus-cni-dir\") pod \"multus-h9jwp\" (UID: \"4490d846-ca32-4cd5-9202-87cc6d499fb4\") " pod="openshift-multus/multus-h9jwp" Oct 08 18:11:17 crc kubenswrapper[4988]: I1008 18:11:17.062457 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/f010a463-69f2-4082-acc3-161899a939c3-host-kubelet\") pod \"ovnkube-node-nsc65\" (UID: \"f010a463-69f2-4082-acc3-161899a939c3\") " pod="openshift-ovn-kubernetes/ovnkube-node-nsc65" Oct 08 18:11:17 crc kubenswrapper[4988]: I1008 18:11:17.062491 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/f010a463-69f2-4082-acc3-161899a939c3-host-cni-netd\") pod \"ovnkube-node-nsc65\" (UID: \"f010a463-69f2-4082-acc3-161899a939c3\") " pod="openshift-ovn-kubernetes/ovnkube-node-nsc65" Oct 08 18:11:17 crc kubenswrapper[4988]: I1008 18:11:17.062522 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/f010a463-69f2-4082-acc3-161899a939c3-systemd-units\") pod \"ovnkube-node-nsc65\" (UID: \"f010a463-69f2-4082-acc3-161899a939c3\") " pod="openshift-ovn-kubernetes/ovnkube-node-nsc65" Oct 08 18:11:17 crc kubenswrapper[4988]: I1008 18:11:17.062544 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/4490d846-ca32-4cd5-9202-87cc6d499fb4-host-var-lib-cni-multus\") pod \"multus-h9jwp\" (UID: \"4490d846-ca32-4cd5-9202-87cc6d499fb4\") " pod="openshift-multus/multus-h9jwp" Oct 08 18:11:17 crc kubenswrapper[4988]: I1008 18:11:17.062758 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/556fb048-8afb-44a5-9ea0-2b7c89e6cf4b-tuning-conf-dir\") pod \"multus-additional-cni-plugins-jnh7m\" (UID: \"556fb048-8afb-44a5-9ea0-2b7c89e6cf4b\") " pod="openshift-multus/multus-additional-cni-plugins-jnh7m" Oct 08 18:11:17 crc kubenswrapper[4988]: I1008 18:11:17.062823 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/f010a463-69f2-4082-acc3-161899a939c3-log-socket\") pod \"ovnkube-node-nsc65\" (UID: \"f010a463-69f2-4082-acc3-161899a939c3\") " pod="openshift-ovn-kubernetes/ovnkube-node-nsc65" Oct 08 18:11:17 crc kubenswrapper[4988]: I1008 18:11:17.062830 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/4490d846-ca32-4cd5-9202-87cc6d499fb4-host-run-netns\") pod \"multus-h9jwp\" (UID: \"4490d846-ca32-4cd5-9202-87cc6d499fb4\") " pod="openshift-multus/multus-h9jwp" Oct 08 18:11:17 crc kubenswrapper[4988]: I1008 18:11:17.062862 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/f010a463-69f2-4082-acc3-161899a939c3-run-ovn\") pod \"ovnkube-node-nsc65\" (UID: \"f010a463-69f2-4082-acc3-161899a939c3\") " pod="openshift-ovn-kubernetes/ovnkube-node-nsc65" Oct 08 18:11:17 crc kubenswrapper[4988]: I1008 18:11:17.062873 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/f010a463-69f2-4082-acc3-161899a939c3-run-systemd\") pod \"ovnkube-node-nsc65\" (UID: \"f010a463-69f2-4082-acc3-161899a939c3\") " pod="openshift-ovn-kubernetes/ovnkube-node-nsc65" Oct 08 18:11:17 crc kubenswrapper[4988]: I1008 18:11:17.062905 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/556fb048-8afb-44a5-9ea0-2b7c89e6cf4b-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-jnh7m\" (UID: \"556fb048-8afb-44a5-9ea0-2b7c89e6cf4b\") " pod="openshift-multus/multus-additional-cni-plugins-jnh7m" Oct 08 18:11:17 crc kubenswrapper[4988]: I1008 18:11:17.062911 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/f010a463-69f2-4082-acc3-161899a939c3-host-cni-bin\") pod \"ovnkube-node-nsc65\" (UID: \"f010a463-69f2-4082-acc3-161899a939c3\") " pod="openshift-ovn-kubernetes/ovnkube-node-nsc65" Oct 08 18:11:17 crc kubenswrapper[4988]: I1008 18:11:17.062951 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/4490d846-ca32-4cd5-9202-87cc6d499fb4-host-run-k8s-cni-cncf-io\") pod \"multus-h9jwp\" (UID: \"4490d846-ca32-4cd5-9202-87cc6d499fb4\") " pod="openshift-multus/multus-h9jwp" Oct 08 18:11:17 crc kubenswrapper[4988]: I1008 18:11:17.062968 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/f010a463-69f2-4082-acc3-161899a939c3-host-run-netns\") pod \"ovnkube-node-nsc65\" (UID: \"f010a463-69f2-4082-acc3-161899a939c3\") " pod="openshift-ovn-kubernetes/ovnkube-node-nsc65" Oct 08 18:11:17 crc kubenswrapper[4988]: I1008 18:11:17.062974 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4490d846-ca32-4cd5-9202-87cc6d499fb4-cni-binary-copy\") pod \"multus-h9jwp\" (UID: \"4490d846-ca32-4cd5-9202-87cc6d499fb4\") " pod="openshift-multus/multus-h9jwp" Oct 08 18:11:17 crc kubenswrapper[4988]: I1008 18:11:17.062991 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/4490d846-ca32-4cd5-9202-87cc6d499fb4-multus-conf-dir\") pod \"multus-h9jwp\" (UID: \"4490d846-ca32-4cd5-9202-87cc6d499fb4\") " pod="openshift-multus/multus-h9jwp" Oct 08 18:11:17 crc kubenswrapper[4988]: I1008 18:11:17.063177 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/f328a20b-5283-407e-8a93-2ab2da599f18-rootfs\") pod \"machine-config-daemon-nm54f\" (UID: \"f328a20b-5283-407e-8a93-2ab2da599f18\") " pod="openshift-machine-config-operator/machine-config-daemon-nm54f" Oct 08 18:11:17 crc kubenswrapper[4988]: I1008 18:11:17.063209 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/4490d846-ca32-4cd5-9202-87cc6d499fb4-system-cni-dir\") pod \"multus-h9jwp\" (UID: \"4490d846-ca32-4cd5-9202-87cc6d499fb4\") " pod="openshift-multus/multus-h9jwp" Oct 08 18:11:17 crc kubenswrapper[4988]: I1008 18:11:17.063291 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/f010a463-69f2-4082-acc3-161899a939c3-env-overrides\") pod \"ovnkube-node-nsc65\" (UID: \"f010a463-69f2-4082-acc3-161899a939c3\") " pod="openshift-ovn-kubernetes/ovnkube-node-nsc65" Oct 08 18:11:17 crc kubenswrapper[4988]: I1008 18:11:17.063296 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/f328a20b-5283-407e-8a93-2ab2da599f18-mcd-auth-proxy-config\") pod \"machine-config-daemon-nm54f\" (UID: \"f328a20b-5283-407e-8a93-2ab2da599f18\") " pod="openshift-machine-config-operator/machine-config-daemon-nm54f" Oct 08 18:11:17 crc kubenswrapper[4988]: I1008 18:11:17.063558 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/f010a463-69f2-4082-acc3-161899a939c3-ovnkube-script-lib\") pod \"ovnkube-node-nsc65\" (UID: \"f010a463-69f2-4082-acc3-161899a939c3\") " pod="openshift-ovn-kubernetes/ovnkube-node-nsc65" Oct 08 18:11:17 crc kubenswrapper[4988]: I1008 18:11:17.063956 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/556fb048-8afb-44a5-9ea0-2b7c89e6cf4b-cni-binary-copy\") pod \"multus-additional-cni-plugins-jnh7m\" (UID: \"556fb048-8afb-44a5-9ea0-2b7c89e6cf4b\") " pod="openshift-multus/multus-additional-cni-plugins-jnh7m" Oct 08 18:11:17 crc kubenswrapper[4988]: I1008 18:11:17.066950 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4490d846-ca32-4cd5-9202-87cc6d499fb4-multus-daemon-config\") pod \"multus-h9jwp\" (UID: \"4490d846-ca32-4cd5-9202-87cc6d499fb4\") " pod="openshift-multus/multus-h9jwp" Oct 08 18:11:17 crc kubenswrapper[4988]: I1008 18:11:17.073211 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/f010a463-69f2-4082-acc3-161899a939c3-ovn-node-metrics-cert\") pod \"ovnkube-node-nsc65\" (UID: \"f010a463-69f2-4082-acc3-161899a939c3\") " pod="openshift-ovn-kubernetes/ovnkube-node-nsc65" Oct 08 18:11:17 crc kubenswrapper[4988]: I1008 18:11:17.073337 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/f328a20b-5283-407e-8a93-2ab2da599f18-proxy-tls\") pod \"machine-config-daemon-nm54f\" (UID: \"f328a20b-5283-407e-8a93-2ab2da599f18\") " pod="openshift-machine-config-operator/machine-config-daemon-nm54f" Oct 08 18:11:17 crc kubenswrapper[4988]: I1008 18:11:17.087347 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dnwgr\" (UniqueName: \"kubernetes.io/projected/f010a463-69f2-4082-acc3-161899a939c3-kube-api-access-dnwgr\") pod \"ovnkube-node-nsc65\" (UID: \"f010a463-69f2-4082-acc3-161899a939c3\") " pod="openshift-ovn-kubernetes/ovnkube-node-nsc65" Oct 08 18:11:17 crc kubenswrapper[4988]: I1008 18:11:17.088603 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:17Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:17 crc kubenswrapper[4988]: I1008 18:11:17.090002 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7k6rk\" (UniqueName: \"kubernetes.io/projected/4490d846-ca32-4cd5-9202-87cc6d499fb4-kube-api-access-7k6rk\") pod \"multus-h9jwp\" (UID: \"4490d846-ca32-4cd5-9202-87cc6d499fb4\") " pod="openshift-multus/multus-h9jwp" Oct 08 18:11:17 crc kubenswrapper[4988]: I1008 18:11:17.090068 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gmlt2\" (UniqueName: \"kubernetes.io/projected/556fb048-8afb-44a5-9ea0-2b7c89e6cf4b-kube-api-access-gmlt2\") pod \"multus-additional-cni-plugins-jnh7m\" (UID: \"556fb048-8afb-44a5-9ea0-2b7c89e6cf4b\") " pod="openshift-multus/multus-additional-cni-plugins-jnh7m" Oct 08 18:11:17 crc kubenswrapper[4988]: I1008 18:11:17.094608 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z2gh4\" (UniqueName: \"kubernetes.io/projected/f328a20b-5283-407e-8a93-2ab2da599f18-kube-api-access-z2gh4\") pod \"machine-config-daemon-nm54f\" (UID: \"f328a20b-5283-407e-8a93-2ab2da599f18\") " pod="openshift-machine-config-operator/machine-config-daemon-nm54f" Oct 08 18:11:17 crc kubenswrapper[4988]: I1008 18:11:17.107581 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:17Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:17 crc kubenswrapper[4988]: I1008 18:11:17.121152 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:17Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:17 crc kubenswrapper[4988]: I1008 18:11:17.139660 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-h9jwp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4490d846-ca32-4cd5-9202-87cc6d499fb4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7k6rk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:11:16Z\\\"}}\" for pod \"openshift-multus\"/\"multus-h9jwp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:17Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:17 crc kubenswrapper[4988]: I1008 18:11:17.167788 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-nsc65" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f010a463-69f2-4082-acc3-161899a939c3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dnwgr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dnwgr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dnwgr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dnwgr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dnwgr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dnwgr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dnwgr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dnwgr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dnwgr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:11:16Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-nsc65\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:17Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:17 crc kubenswrapper[4988]: I1008 18:11:17.179396 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-h9jwp" Oct 08 18:11:17 crc kubenswrapper[4988]: I1008 18:11:17.180860 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-s4f2k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ba5e11d7-7bde-4c3d-983b-884232219652\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hn5jf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:11:15Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-s4f2k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:17Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:17 crc kubenswrapper[4988]: W1008 18:11:17.189411 4988 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4490d846_ca32_4cd5_9202_87cc6d499fb4.slice/crio-e6ee134ade672116cb12c2517358fb8ad7341c83dbbef8ed480c69e915c1ffaf WatchSource:0}: Error finding container e6ee134ade672116cb12c2517358fb8ad7341c83dbbef8ed480c69e915c1ffaf: Status 404 returned error can't find the container with id e6ee134ade672116cb12c2517358fb8ad7341c83dbbef8ed480c69e915c1ffaf Oct 08 18:11:17 crc kubenswrapper[4988]: I1008 18:11:17.190846 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-jnh7m" Oct 08 18:11:17 crc kubenswrapper[4988]: I1008 18:11:17.195691 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e8906182d55c285b3cf095e2d9e1dac397af4618a514f11e26b44c172dd573f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:17Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:17 crc kubenswrapper[4988]: I1008 18:11:17.198935 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" Oct 08 18:11:17 crc kubenswrapper[4988]: I1008 18:11:17.203600 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-nsc65" Oct 08 18:11:17 crc kubenswrapper[4988]: I1008 18:11:17.209185 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-dzxs2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8f307b20-5e33-4725-99cb-210e25113532\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q65f7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:11:16Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-dzxs2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:17Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:17 crc kubenswrapper[4988]: I1008 18:11:17.233035 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aef71be1-1d12-46e5-95b2-bd399040fd4d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:10:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:10:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:10:55Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:10:55Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:10:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77c68f3ecb8a0724499e495cb26ef3db5beecb140f2558d8c0b25f41928b84ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:10:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a560ee84c2405227de71fda127566213cd4f017e73d5581442ba267781b1c539\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:10:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6100accc3bd08928bc1ff88310f29fa5da37fd2d5c315c94279801c9778c4745\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:10:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b0e596d7ce4e9341c44499af79d3b3caf0e5e2c156902ea40cf2a3dca672be9e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5bd35aef24be034a0b305a40bb53ca171a4f76db3b93b4f78ae95099b0a4cde3\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-08T18:11:14Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1008 18:11:09.091034 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1008 18:11:09.091984 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3472374062/tls.crt::/tmp/serving-cert-3472374062/tls.key\\\\\\\"\\\\nI1008 18:11:14.619696 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1008 18:11:14.628837 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1008 18:11:14.628898 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1008 18:11:14.628994 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1008 18:11:14.629036 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1008 18:11:14.641318 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1008 18:11:14.641363 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1008 18:11:14.641372 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1008 18:11:14.641382 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1008 18:11:14.641415 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1008 18:11:14.641423 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1008 18:11:14.641431 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1008 18:11:14.641484 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1008 18:11:14.646451 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T18:10:58Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c79ef704fda1bbcc486f4c7afa3039a3ab38669833f0f66bafe24dd4cc679e60\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:10:58Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://28d21731af0f16d6a8ae6c6e9d0bebf0560d4a533db4c78c10bf1837890394ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://28d21731af0f16d6a8ae6c6e9d0bebf0560d4a533db4c78c10bf1837890394ef\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:10:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:10:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:10:55Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:17Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:17 crc kubenswrapper[4988]: I1008 18:11:17.237376 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 18:11:17 crc kubenswrapper[4988]: I1008 18:11:17.237424 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 18:11:17 crc kubenswrapper[4988]: I1008 18:11:17.237433 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 18:11:17 crc kubenswrapper[4988]: E1008 18:11:17.237548 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 08 18:11:17 crc kubenswrapper[4988]: E1008 18:11:17.237647 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 08 18:11:17 crc kubenswrapper[4988]: E1008 18:11:17.237719 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 08 18:11:17 crc kubenswrapper[4988]: I1008 18:11:17.253602 4988 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="01ab3dd5-8196-46d0-ad33-122e2ca51def" path="/var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes" Oct 08 18:11:17 crc kubenswrapper[4988]: I1008 18:11:17.255037 4988 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" path="/var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes" Oct 08 18:11:17 crc kubenswrapper[4988]: I1008 18:11:17.255447 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a35a5e989ba69c8ef36ce4cea272aadf67ce20473fc257adfcfa86a2faba6ae0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a522e937cfe03fb9f2e60cb1bea07214cf63c69bdecd5ba65a2dfdba882d28dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:17Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:17 crc kubenswrapper[4988]: I1008 18:11:17.256139 4988 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09efc573-dbb6-4249-bd59-9b87aba8dd28" path="/var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes" Oct 08 18:11:17 crc kubenswrapper[4988]: I1008 18:11:17.257729 4988 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b574797-001e-440a-8f4e-c0be86edad0f" path="/var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes" Oct 08 18:11:17 crc kubenswrapper[4988]: I1008 18:11:17.258690 4988 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b78653f-4ff9-4508-8672-245ed9b561e3" path="/var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes" Oct 08 18:11:17 crc kubenswrapper[4988]: I1008 18:11:17.260501 4988 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1386a44e-36a2-460c-96d0-0359d2b6f0f5" path="/var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes" Oct 08 18:11:17 crc kubenswrapper[4988]: I1008 18:11:17.261711 4988 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1bf7eb37-55a3-4c65-b768-a94c82151e69" path="/var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes" Oct 08 18:11:17 crc kubenswrapper[4988]: I1008 18:11:17.263178 4988 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1d611f23-29be-4491-8495-bee1670e935f" path="/var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes" Oct 08 18:11:17 crc kubenswrapper[4988]: I1008 18:11:17.264243 4988 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="20b0d48f-5fd6-431c-a545-e3c800c7b866" path="/var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/volumes" Oct 08 18:11:17 crc kubenswrapper[4988]: I1008 18:11:17.268504 4988 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" path="/var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes" Oct 08 18:11:17 crc kubenswrapper[4988]: I1008 18:11:17.269906 4988 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="22c825df-677d-4ca6-82db-3454ed06e783" path="/var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes" Oct 08 18:11:17 crc kubenswrapper[4988]: I1008 18:11:17.271983 4988 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="25e176fe-21b4-4974-b1ed-c8b94f112a7f" path="/var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes" Oct 08 18:11:17 crc kubenswrapper[4988]: I1008 18:11:17.273098 4988 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" path="/var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes" Oct 08 18:11:17 crc kubenswrapper[4988]: I1008 18:11:17.274432 4988 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="31d8b7a1-420e-4252-a5b7-eebe8a111292" path="/var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes" Oct 08 18:11:17 crc kubenswrapper[4988]: I1008 18:11:17.275924 4988 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3ab1a177-2de0-46d9-b765-d0d0649bb42e" path="/var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/volumes" Oct 08 18:11:17 crc kubenswrapper[4988]: I1008 18:11:17.277143 4988 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" path="/var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes" Oct 08 18:11:17 crc kubenswrapper[4988]: I1008 18:11:17.278410 4988 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="43509403-f426-496e-be36-56cef71462f5" path="/var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes" Oct 08 18:11:17 crc kubenswrapper[4988]: I1008 18:11:17.279115 4988 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="44663579-783b-4372-86d6-acf235a62d72" path="/var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/volumes" Oct 08 18:11:17 crc kubenswrapper[4988]: I1008 18:11:17.280205 4988 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="496e6271-fb68-4057-954e-a0d97a4afa3f" path="/var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes" Oct 08 18:11:17 crc kubenswrapper[4988]: I1008 18:11:17.280828 4988 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" path="/var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes" Oct 08 18:11:17 crc kubenswrapper[4988]: I1008 18:11:17.281308 4988 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49ef4625-1d3a-4a9f-b595-c2433d32326d" path="/var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/volumes" Oct 08 18:11:17 crc kubenswrapper[4988]: I1008 18:11:17.282348 4988 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4bb40260-dbaa-4fb0-84df-5e680505d512" path="/var/lib/kubelet/pods/4bb40260-dbaa-4fb0-84df-5e680505d512/volumes" Oct 08 18:11:17 crc kubenswrapper[4988]: I1008 18:11:17.284844 4988 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5225d0e4-402f-4861-b410-819f433b1803" path="/var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes" Oct 08 18:11:17 crc kubenswrapper[4988]: I1008 18:11:17.285505 4988 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5441d097-087c-4d9a-baa8-b210afa90fc9" path="/var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes" Oct 08 18:11:17 crc kubenswrapper[4988]: I1008 18:11:17.286546 4988 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="57a731c4-ef35-47a8-b875-bfb08a7f8011" path="/var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes" Oct 08 18:11:17 crc kubenswrapper[4988]: I1008 18:11:17.287413 4988 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5b88f790-22fa-440e-b583-365168c0b23d" path="/var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/volumes" Oct 08 18:11:17 crc kubenswrapper[4988]: I1008 18:11:17.288694 4988 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5fe579f8-e8a6-4643-bce5-a661393c4dde" path="/var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/volumes" Oct 08 18:11:17 crc kubenswrapper[4988]: I1008 18:11:17.289519 4988 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6402fda4-df10-493c-b4e5-d0569419652d" path="/var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes" Oct 08 18:11:17 crc kubenswrapper[4988]: I1008 18:11:17.290851 4988 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6509e943-70c6-444c-bc41-48a544e36fbd" path="/var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes" Oct 08 18:11:17 crc kubenswrapper[4988]: I1008 18:11:17.291613 4988 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6731426b-95fe-49ff-bb5f-40441049fde2" path="/var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/volumes" Oct 08 18:11:17 crc kubenswrapper[4988]: I1008 18:11:17.292227 4988 kubelet_volumes.go:152] "Cleaned up orphaned volume subpath from pod" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volume-subpaths/run-systemd/ovnkube-controller/6" Oct 08 18:11:17 crc kubenswrapper[4988]: I1008 18:11:17.292768 4988 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volumes" Oct 08 18:11:17 crc kubenswrapper[4988]: I1008 18:11:17.295010 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:17Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:17 crc kubenswrapper[4988]: I1008 18:11:17.296440 4988 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7539238d-5fe0-46ed-884e-1c3b566537ec" path="/var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes" Oct 08 18:11:17 crc kubenswrapper[4988]: I1008 18:11:17.296977 4988 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7583ce53-e0fe-4a16-9e4d-50516596a136" path="/var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes" Oct 08 18:11:17 crc kubenswrapper[4988]: I1008 18:11:17.297666 4988 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7bb08738-c794-4ee8-9972-3a62ca171029" path="/var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes" Oct 08 18:11:17 crc kubenswrapper[4988]: I1008 18:11:17.299313 4988 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="87cf06ed-a83f-41a7-828d-70653580a8cb" path="/var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes" Oct 08 18:11:17 crc kubenswrapper[4988]: I1008 18:11:17.300310 4988 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" path="/var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes" Oct 08 18:11:17 crc kubenswrapper[4988]: I1008 18:11:17.300878 4988 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="925f1c65-6136-48ba-85aa-3a3b50560753" path="/var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes" Oct 08 18:11:17 crc kubenswrapper[4988]: I1008 18:11:17.301857 4988 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" path="/var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/volumes" Oct 08 18:11:17 crc kubenswrapper[4988]: I1008 18:11:17.302565 4988 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9d4552c7-cd75-42dd-8880-30dd377c49a4" path="/var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes" Oct 08 18:11:17 crc kubenswrapper[4988]: I1008 18:11:17.303589 4988 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" path="/var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/volumes" Oct 08 18:11:17 crc kubenswrapper[4988]: I1008 18:11:17.304179 4988 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a31745f5-9847-4afe-82a5-3161cc66ca93" path="/var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes" Oct 08 18:11:17 crc kubenswrapper[4988]: I1008 18:11:17.305235 4988 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" path="/var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes" Oct 08 18:11:17 crc kubenswrapper[4988]: I1008 18:11:17.305891 4988 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6312bbd-5731-4ea0-a20f-81d5a57df44a" path="/var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/volumes" Oct 08 18:11:17 crc kubenswrapper[4988]: I1008 18:11:17.306836 4988 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" path="/var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes" Oct 08 18:11:17 crc kubenswrapper[4988]: I1008 18:11:17.307528 4988 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" path="/var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes" Oct 08 18:11:17 crc kubenswrapper[4988]: I1008 18:11:17.308614 4988 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" path="/var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/volumes" Oct 08 18:11:17 crc kubenswrapper[4988]: I1008 18:11:17.309379 4988 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bf126b07-da06-4140-9a57-dfd54fc6b486" path="/var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes" Oct 08 18:11:17 crc kubenswrapper[4988]: I1008 18:11:17.310347 4988 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c03ee662-fb2f-4fc4-a2c1-af487c19d254" path="/var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes" Oct 08 18:11:17 crc kubenswrapper[4988]: I1008 18:11:17.310852 4988 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" path="/var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/volumes" Oct 08 18:11:17 crc kubenswrapper[4988]: I1008 18:11:17.311749 4988 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e7e6199b-1264-4501-8953-767f51328d08" path="/var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes" Oct 08 18:11:17 crc kubenswrapper[4988]: I1008 18:11:17.312279 4988 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="efdd0498-1daa-4136-9a4a-3b948c2293fc" path="/var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/volumes" Oct 08 18:11:17 crc kubenswrapper[4988]: I1008 18:11:17.312902 4988 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" path="/var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/volumes" Oct 08 18:11:17 crc kubenswrapper[4988]: I1008 18:11:17.313782 4988 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fda69060-fa79-4696-b1a6-7980f124bf7c" path="/var/lib/kubelet/pods/fda69060-fa79-4696-b1a6-7980f124bf7c/volumes" Oct 08 18:11:17 crc kubenswrapper[4988]: I1008 18:11:17.335193 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:17Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:17 crc kubenswrapper[4988]: I1008 18:11:17.376967 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-jnh7m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"556fb048-8afb-44a5-9ea0-2b7c89e6cf4b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmlt2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmlt2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmlt2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmlt2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmlt2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmlt2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmlt2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:11:16Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-jnh7m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:17Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:17 crc kubenswrapper[4988]: I1008 18:11:17.394906 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f328a20b-5283-407e-8a93-2ab2da599f18\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z2gh4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z2gh4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:11:16Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-nm54f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:17Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:17 crc kubenswrapper[4988]: I1008 18:11:17.439568 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-s4f2k" event={"ID":"ba5e11d7-7bde-4c3d-983b-884232219652","Type":"ContainerStarted","Data":"824bc34a2807068d3fceb10651803ae9ae5651e940f7aed4877c544d76dae4d6"} Oct 08 18:11:17 crc kubenswrapper[4988]: I1008 18:11:17.441532 4988 generic.go:334] "Generic (PLEG): container finished" podID="f010a463-69f2-4082-acc3-161899a939c3" containerID="befe684cea9bd6dfda3ec8e85bca431c26bd8f82823b0a69dc0d04319a3ca385" exitCode=0 Oct 08 18:11:17 crc kubenswrapper[4988]: I1008 18:11:17.441597 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-nsc65" event={"ID":"f010a463-69f2-4082-acc3-161899a939c3","Type":"ContainerDied","Data":"befe684cea9bd6dfda3ec8e85bca431c26bd8f82823b0a69dc0d04319a3ca385"} Oct 08 18:11:17 crc kubenswrapper[4988]: I1008 18:11:17.441627 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-nsc65" event={"ID":"f010a463-69f2-4082-acc3-161899a939c3","Type":"ContainerStarted","Data":"937461074d800725cfae8d0f2f3c1a96fcb0473c00b305eb9c50b400d5f92549"} Oct 08 18:11:17 crc kubenswrapper[4988]: I1008 18:11:17.446066 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" event={"ID":"f328a20b-5283-407e-8a93-2ab2da599f18","Type":"ContainerStarted","Data":"3cdb2a757332bdcdc2a2f3362913fe8f06003b0597de20b72627f9badd5e30af"} Oct 08 18:11:17 crc kubenswrapper[4988]: I1008 18:11:17.446116 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" event={"ID":"f328a20b-5283-407e-8a93-2ab2da599f18","Type":"ContainerStarted","Data":"95311fda6d15ffa79bbebafc7883ac9f7dd751def0e68b5f1baccd6cb6b6ebe3"} Oct 08 18:11:17 crc kubenswrapper[4988]: I1008 18:11:17.447208 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-jnh7m" event={"ID":"556fb048-8afb-44a5-9ea0-2b7c89e6cf4b","Type":"ContainerStarted","Data":"fbc0fffded37a25567810830009b8f890c49301b25d6eaee9b675d48ab019c71"} Oct 08 18:11:17 crc kubenswrapper[4988]: I1008 18:11:17.448836 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-dzxs2" event={"ID":"8f307b20-5e33-4725-99cb-210e25113532","Type":"ContainerStarted","Data":"72ee803c6ef88524e0ce0391476f8e5a7dfbc858b095a3c2ae31bad10991ce3f"} Oct 08 18:11:17 crc kubenswrapper[4988]: I1008 18:11:17.453509 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-h9jwp" event={"ID":"4490d846-ca32-4cd5-9202-87cc6d499fb4","Type":"ContainerStarted","Data":"00b7f6db043c28efda3a09fdc77a6ccded3eae8b6a3d2852aa500789d5f01194"} Oct 08 18:11:17 crc kubenswrapper[4988]: I1008 18:11:17.453616 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-h9jwp" event={"ID":"4490d846-ca32-4cd5-9202-87cc6d499fb4","Type":"ContainerStarted","Data":"e6ee134ade672116cb12c2517358fb8ad7341c83dbbef8ed480c69e915c1ffaf"} Oct 08 18:11:17 crc kubenswrapper[4988]: I1008 18:11:17.457423 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:17Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:17 crc kubenswrapper[4988]: I1008 18:11:17.475422 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:17Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:17 crc kubenswrapper[4988]: I1008 18:11:17.492149 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-h9jwp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4490d846-ca32-4cd5-9202-87cc6d499fb4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7k6rk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:11:16Z\\\"}}\" for pod \"openshift-multus\"/\"multus-h9jwp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:17Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:17 crc kubenswrapper[4988]: I1008 18:11:17.544097 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-nsc65" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f010a463-69f2-4082-acc3-161899a939c3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dnwgr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dnwgr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dnwgr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dnwgr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dnwgr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dnwgr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dnwgr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dnwgr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dnwgr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:11:16Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-nsc65\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:17Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:17 crc kubenswrapper[4988]: I1008 18:11:17.573949 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-s4f2k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ba5e11d7-7bde-4c3d-983b-884232219652\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://824bc34a2807068d3fceb10651803ae9ae5651e940f7aed4877c544d76dae4d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hn5jf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:11:15Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-s4f2k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:17Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:17 crc kubenswrapper[4988]: I1008 18:11:17.611061 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aef71be1-1d12-46e5-95b2-bd399040fd4d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:10:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:10:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:10:55Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:10:55Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:10:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77c68f3ecb8a0724499e495cb26ef3db5beecb140f2558d8c0b25f41928b84ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:10:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a560ee84c2405227de71fda127566213cd4f017e73d5581442ba267781b1c539\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:10:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6100accc3bd08928bc1ff88310f29fa5da37fd2d5c315c94279801c9778c4745\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:10:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b0e596d7ce4e9341c44499af79d3b3caf0e5e2c156902ea40cf2a3dca672be9e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5bd35aef24be034a0b305a40bb53ca171a4f76db3b93b4f78ae95099b0a4cde3\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-08T18:11:14Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1008 18:11:09.091034 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1008 18:11:09.091984 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3472374062/tls.crt::/tmp/serving-cert-3472374062/tls.key\\\\\\\"\\\\nI1008 18:11:14.619696 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1008 18:11:14.628837 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1008 18:11:14.628898 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1008 18:11:14.628994 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1008 18:11:14.629036 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1008 18:11:14.641318 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1008 18:11:14.641363 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1008 18:11:14.641372 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1008 18:11:14.641382 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1008 18:11:14.641415 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1008 18:11:14.641423 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1008 18:11:14.641431 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1008 18:11:14.641484 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1008 18:11:14.646451 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T18:10:58Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c79ef704fda1bbcc486f4c7afa3039a3ab38669833f0f66bafe24dd4cc679e60\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:10:58Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://28d21731af0f16d6a8ae6c6e9d0bebf0560d4a533db4c78c10bf1837890394ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://28d21731af0f16d6a8ae6c6e9d0bebf0560d4a533db4c78c10bf1837890394ef\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:10:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:10:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:10:55Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:17Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:17 crc kubenswrapper[4988]: I1008 18:11:17.652008 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e8906182d55c285b3cf095e2d9e1dac397af4618a514f11e26b44c172dd573f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:17Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:17 crc kubenswrapper[4988]: I1008 18:11:17.692965 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-dzxs2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8f307b20-5e33-4725-99cb-210e25113532\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q65f7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:11:16Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-dzxs2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:17Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:17 crc kubenswrapper[4988]: I1008 18:11:17.731514 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f328a20b-5283-407e-8a93-2ab2da599f18\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z2gh4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z2gh4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:11:16Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-nm54f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:17Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:17 crc kubenswrapper[4988]: I1008 18:11:17.772187 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a35a5e989ba69c8ef36ce4cea272aadf67ce20473fc257adfcfa86a2faba6ae0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a522e937cfe03fb9f2e60cb1bea07214cf63c69bdecd5ba65a2dfdba882d28dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:17Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:17 crc kubenswrapper[4988]: I1008 18:11:17.819902 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:17Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:17 crc kubenswrapper[4988]: I1008 18:11:17.852288 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:17Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:17 crc kubenswrapper[4988]: I1008 18:11:17.895900 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-jnh7m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"556fb048-8afb-44a5-9ea0-2b7c89e6cf4b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmlt2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmlt2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmlt2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmlt2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmlt2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmlt2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmlt2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:11:16Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-jnh7m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:17Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:17 crc kubenswrapper[4988]: I1008 18:11:17.932342 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-h9jwp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4490d846-ca32-4cd5-9202-87cc6d499fb4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://00b7f6db043c28efda3a09fdc77a6ccded3eae8b6a3d2852aa500789d5f01194\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7k6rk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:11:16Z\\\"}}\" for pod \"openshift-multus\"/\"multus-h9jwp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:17Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:18 crc kubenswrapper[4988]: I1008 18:11:18.004141 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-nsc65" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f010a463-69f2-4082-acc3-161899a939c3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dnwgr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dnwgr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dnwgr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dnwgr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dnwgr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dnwgr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dnwgr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dnwgr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://befe684cea9bd6dfda3ec8e85bca431c26bd8f82823b0a69dc0d04319a3ca385\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://befe684cea9bd6dfda3ec8e85bca431c26bd8f82823b0a69dc0d04319a3ca385\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dnwgr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:11:16Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-nsc65\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:18Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:18 crc kubenswrapper[4988]: I1008 18:11:18.038934 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:18Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:18 crc kubenswrapper[4988]: I1008 18:11:18.058640 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:18Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:18 crc kubenswrapper[4988]: I1008 18:11:18.091136 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-s4f2k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ba5e11d7-7bde-4c3d-983b-884232219652\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://824bc34a2807068d3fceb10651803ae9ae5651e940f7aed4877c544d76dae4d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hn5jf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:11:15Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-s4f2k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:18Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:18 crc kubenswrapper[4988]: I1008 18:11:18.128901 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-dzxs2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8f307b20-5e33-4725-99cb-210e25113532\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72ee803c6ef88524e0ce0391476f8e5a7dfbc858b095a3c2ae31bad10991ce3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q65f7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:11:16Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-dzxs2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:18Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:18 crc kubenswrapper[4988]: I1008 18:11:18.169965 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aef71be1-1d12-46e5-95b2-bd399040fd4d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:10:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:10:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:10:55Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:10:55Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:10:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77c68f3ecb8a0724499e495cb26ef3db5beecb140f2558d8c0b25f41928b84ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:10:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a560ee84c2405227de71fda127566213cd4f017e73d5581442ba267781b1c539\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:10:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6100accc3bd08928bc1ff88310f29fa5da37fd2d5c315c94279801c9778c4745\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:10:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b0e596d7ce4e9341c44499af79d3b3caf0e5e2c156902ea40cf2a3dca672be9e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5bd35aef24be034a0b305a40bb53ca171a4f76db3b93b4f78ae95099b0a4cde3\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-08T18:11:14Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1008 18:11:09.091034 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1008 18:11:09.091984 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3472374062/tls.crt::/tmp/serving-cert-3472374062/tls.key\\\\\\\"\\\\nI1008 18:11:14.619696 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1008 18:11:14.628837 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1008 18:11:14.628898 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1008 18:11:14.628994 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1008 18:11:14.629036 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1008 18:11:14.641318 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1008 18:11:14.641363 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1008 18:11:14.641372 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1008 18:11:14.641382 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1008 18:11:14.641415 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1008 18:11:14.641423 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1008 18:11:14.641431 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1008 18:11:14.641484 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1008 18:11:14.646451 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T18:10:58Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c79ef704fda1bbcc486f4c7afa3039a3ab38669833f0f66bafe24dd4cc679e60\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:10:58Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://28d21731af0f16d6a8ae6c6e9d0bebf0560d4a533db4c78c10bf1837890394ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://28d21731af0f16d6a8ae6c6e9d0bebf0560d4a533db4c78c10bf1837890394ef\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:10:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:10:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:10:55Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:18Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:18 crc kubenswrapper[4988]: I1008 18:11:18.214319 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e8906182d55c285b3cf095e2d9e1dac397af4618a514f11e26b44c172dd573f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:18Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:18 crc kubenswrapper[4988]: I1008 18:11:18.250145 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:18Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:18 crc kubenswrapper[4988]: I1008 18:11:18.290844 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-jnh7m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"556fb048-8afb-44a5-9ea0-2b7c89e6cf4b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmlt2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmlt2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmlt2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmlt2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmlt2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmlt2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmlt2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:11:16Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-jnh7m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:18Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:18 crc kubenswrapper[4988]: I1008 18:11:18.327855 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f328a20b-5283-407e-8a93-2ab2da599f18\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z2gh4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z2gh4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:11:16Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-nm54f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:18Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:18 crc kubenswrapper[4988]: I1008 18:11:18.368297 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a35a5e989ba69c8ef36ce4cea272aadf67ce20473fc257adfcfa86a2faba6ae0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a522e937cfe03fb9f2e60cb1bea07214cf63c69bdecd5ba65a2dfdba882d28dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:18Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:18 crc kubenswrapper[4988]: I1008 18:11:18.416570 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:18Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:18 crc kubenswrapper[4988]: I1008 18:11:18.460278 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"9590ca0059f0ebbe86343cccb3475338c317382c3231e082655b61e67eece375"} Oct 08 18:11:18 crc kubenswrapper[4988]: I1008 18:11:18.463199 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" event={"ID":"f328a20b-5283-407e-8a93-2ab2da599f18","Type":"ContainerStarted","Data":"8622d56fb198dadee31fb0ca8fc80d56828b6710648055f264533e2d5bdb3037"} Oct 08 18:11:18 crc kubenswrapper[4988]: I1008 18:11:18.465070 4988 generic.go:334] "Generic (PLEG): container finished" podID="556fb048-8afb-44a5-9ea0-2b7c89e6cf4b" containerID="a1a825886f105ccf9a7cc6f28a0cd012fb859f4990dade0e6aae932482146028" exitCode=0 Oct 08 18:11:18 crc kubenswrapper[4988]: I1008 18:11:18.465168 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-jnh7m" event={"ID":"556fb048-8afb-44a5-9ea0-2b7c89e6cf4b","Type":"ContainerDied","Data":"a1a825886f105ccf9a7cc6f28a0cd012fb859f4990dade0e6aae932482146028"} Oct 08 18:11:18 crc kubenswrapper[4988]: I1008 18:11:18.470576 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-nsc65" event={"ID":"f010a463-69f2-4082-acc3-161899a939c3","Type":"ContainerStarted","Data":"9fbbeae0409d2e47e8de9a4306f63b0d13acbfd73cd86f83303fec14352ed718"} Oct 08 18:11:18 crc kubenswrapper[4988]: I1008 18:11:18.470638 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-nsc65" event={"ID":"f010a463-69f2-4082-acc3-161899a939c3","Type":"ContainerStarted","Data":"55791327a370c0009c523b86b2234e82d3794985f2eaccf1922eb90f46b2b970"} Oct 08 18:11:18 crc kubenswrapper[4988]: I1008 18:11:18.470662 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-nsc65" event={"ID":"f010a463-69f2-4082-acc3-161899a939c3","Type":"ContainerStarted","Data":"03406e726f817bcedb841f81959ab12808ba35a530240f9abc5fa5c608b9a6ae"} Oct 08 18:11:18 crc kubenswrapper[4988]: I1008 18:11:18.470681 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-nsc65" event={"ID":"f010a463-69f2-4082-acc3-161899a939c3","Type":"ContainerStarted","Data":"48a516f6c153743bd1a21ec6945fb368a37f4d2b9d6a326b68edc096a192ed05"} Oct 08 18:11:18 crc kubenswrapper[4988]: I1008 18:11:18.470697 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-nsc65" event={"ID":"f010a463-69f2-4082-acc3-161899a939c3","Type":"ContainerStarted","Data":"cee5de93bc68345f9b737857f02832f56582d58757c50a8798eec25d19b2c20d"} Oct 08 18:11:18 crc kubenswrapper[4988]: I1008 18:11:18.470713 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-nsc65" event={"ID":"f010a463-69f2-4082-acc3-161899a939c3","Type":"ContainerStarted","Data":"d314eb248e90371fed068b1173e26d4c9f9765417a70b341e1a5b136ec17353a"} Oct 08 18:11:18 crc kubenswrapper[4988]: I1008 18:11:18.477449 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-dzxs2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8f307b20-5e33-4725-99cb-210e25113532\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72ee803c6ef88524e0ce0391476f8e5a7dfbc858b095a3c2ae31bad10991ce3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q65f7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:11:16Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-dzxs2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:18Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:18 crc kubenswrapper[4988]: I1008 18:11:18.504503 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aef71be1-1d12-46e5-95b2-bd399040fd4d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:10:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:10:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:10:55Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:10:55Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:10:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77c68f3ecb8a0724499e495cb26ef3db5beecb140f2558d8c0b25f41928b84ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:10:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a560ee84c2405227de71fda127566213cd4f017e73d5581442ba267781b1c539\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:10:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6100accc3bd08928bc1ff88310f29fa5da37fd2d5c315c94279801c9778c4745\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:10:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b0e596d7ce4e9341c44499af79d3b3caf0e5e2c156902ea40cf2a3dca672be9e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5bd35aef24be034a0b305a40bb53ca171a4f76db3b93b4f78ae95099b0a4cde3\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-08T18:11:14Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1008 18:11:09.091034 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1008 18:11:09.091984 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3472374062/tls.crt::/tmp/serving-cert-3472374062/tls.key\\\\\\\"\\\\nI1008 18:11:14.619696 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1008 18:11:14.628837 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1008 18:11:14.628898 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1008 18:11:14.628994 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1008 18:11:14.629036 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1008 18:11:14.641318 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1008 18:11:14.641363 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1008 18:11:14.641372 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1008 18:11:14.641382 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1008 18:11:14.641415 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1008 18:11:14.641423 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1008 18:11:14.641431 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1008 18:11:14.641484 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1008 18:11:14.646451 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T18:10:58Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c79ef704fda1bbcc486f4c7afa3039a3ab38669833f0f66bafe24dd4cc679e60\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:10:58Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://28d21731af0f16d6a8ae6c6e9d0bebf0560d4a533db4c78c10bf1837890394ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://28d21731af0f16d6a8ae6c6e9d0bebf0560d4a533db4c78c10bf1837890394ef\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:10:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:10:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:10:55Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:18Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:18 crc kubenswrapper[4988]: I1008 18:11:18.529597 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e8906182d55c285b3cf095e2d9e1dac397af4618a514f11e26b44c172dd573f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:18Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:18 crc kubenswrapper[4988]: I1008 18:11:18.568367 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:18Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:18 crc kubenswrapper[4988]: I1008 18:11:18.610447 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-jnh7m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"556fb048-8afb-44a5-9ea0-2b7c89e6cf4b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmlt2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmlt2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmlt2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmlt2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmlt2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmlt2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmlt2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:11:16Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-jnh7m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:18Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:18 crc kubenswrapper[4988]: I1008 18:11:18.650274 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f328a20b-5283-407e-8a93-2ab2da599f18\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z2gh4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z2gh4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:11:16Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-nm54f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:18Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:18 crc kubenswrapper[4988]: I1008 18:11:18.691029 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a35a5e989ba69c8ef36ce4cea272aadf67ce20473fc257adfcfa86a2faba6ae0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a522e937cfe03fb9f2e60cb1bea07214cf63c69bdecd5ba65a2dfdba882d28dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:18Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:18 crc kubenswrapper[4988]: I1008 18:11:18.739014 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9590ca0059f0ebbe86343cccb3475338c317382c3231e082655b61e67eece375\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:18Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:18 crc kubenswrapper[4988]: I1008 18:11:18.778050 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-h9jwp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4490d846-ca32-4cd5-9202-87cc6d499fb4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://00b7f6db043c28efda3a09fdc77a6ccded3eae8b6a3d2852aa500789d5f01194\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7k6rk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:11:16Z\\\"}}\" for pod \"openshift-multus\"/\"multus-h9jwp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:18Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:18 crc kubenswrapper[4988]: I1008 18:11:18.823288 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-nsc65" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f010a463-69f2-4082-acc3-161899a939c3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dnwgr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dnwgr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dnwgr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dnwgr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dnwgr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dnwgr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dnwgr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dnwgr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://befe684cea9bd6dfda3ec8e85bca431c26bd8f82823b0a69dc0d04319a3ca385\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://befe684cea9bd6dfda3ec8e85bca431c26bd8f82823b0a69dc0d04319a3ca385\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dnwgr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:11:16Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-nsc65\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:18Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:18 crc kubenswrapper[4988]: I1008 18:11:18.851742 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:18Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:18 crc kubenswrapper[4988]: I1008 18:11:18.882819 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 18:11:18 crc kubenswrapper[4988]: E1008 18:11:18.883014 4988 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 18:11:22.882972722 +0000 UTC m=+28.332815492 (durationBeforeRetry 4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 18:11:18 crc kubenswrapper[4988]: I1008 18:11:18.883117 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 18:11:18 crc kubenswrapper[4988]: I1008 18:11:18.883224 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 18:11:18 crc kubenswrapper[4988]: E1008 18:11:18.883272 4988 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 08 18:11:18 crc kubenswrapper[4988]: E1008 18:11:18.883357 4988 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-08 18:11:22.883332683 +0000 UTC m=+28.333175653 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 08 18:11:18 crc kubenswrapper[4988]: E1008 18:11:18.883547 4988 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 08 18:11:18 crc kubenswrapper[4988]: E1008 18:11:18.883681 4988 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-08 18:11:22.883654553 +0000 UTC m=+28.333497323 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 08 18:11:18 crc kubenswrapper[4988]: I1008 18:11:18.891194 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:18Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:18 crc kubenswrapper[4988]: I1008 18:11:18.927936 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-s4f2k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ba5e11d7-7bde-4c3d-983b-884232219652\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://824bc34a2807068d3fceb10651803ae9ae5651e940f7aed4877c544d76dae4d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hn5jf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:11:15Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-s4f2k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:18Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:18 crc kubenswrapper[4988]: I1008 18:11:18.967726 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-s4f2k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ba5e11d7-7bde-4c3d-983b-884232219652\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://824bc34a2807068d3fceb10651803ae9ae5651e940f7aed4877c544d76dae4d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hn5jf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:11:15Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-s4f2k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:18Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:18 crc kubenswrapper[4988]: I1008 18:11:18.984400 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 18:11:18 crc kubenswrapper[4988]: I1008 18:11:18.984483 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 18:11:18 crc kubenswrapper[4988]: E1008 18:11:18.984616 4988 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 08 18:11:18 crc kubenswrapper[4988]: E1008 18:11:18.984654 4988 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 08 18:11:18 crc kubenswrapper[4988]: E1008 18:11:18.984669 4988 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 08 18:11:18 crc kubenswrapper[4988]: E1008 18:11:18.984708 4988 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 08 18:11:18 crc kubenswrapper[4988]: E1008 18:11:18.984736 4988 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 08 18:11:18 crc kubenswrapper[4988]: E1008 18:11:18.984752 4988 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 08 18:11:18 crc kubenswrapper[4988]: E1008 18:11:18.984751 4988 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-08 18:11:22.984726727 +0000 UTC m=+28.434569497 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 08 18:11:18 crc kubenswrapper[4988]: E1008 18:11:18.984855 4988 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-08 18:11:22.98483153 +0000 UTC m=+28.434674510 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 08 18:11:19 crc kubenswrapper[4988]: I1008 18:11:19.007865 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-dzxs2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8f307b20-5e33-4725-99cb-210e25113532\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72ee803c6ef88524e0ce0391476f8e5a7dfbc858b095a3c2ae31bad10991ce3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q65f7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:11:16Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-dzxs2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:19Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:19 crc kubenswrapper[4988]: I1008 18:11:19.052032 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aef71be1-1d12-46e5-95b2-bd399040fd4d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:10:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:10:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:10:55Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:10:55Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:10:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77c68f3ecb8a0724499e495cb26ef3db5beecb140f2558d8c0b25f41928b84ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:10:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a560ee84c2405227de71fda127566213cd4f017e73d5581442ba267781b1c539\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:10:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6100accc3bd08928bc1ff88310f29fa5da37fd2d5c315c94279801c9778c4745\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:10:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b0e596d7ce4e9341c44499af79d3b3caf0e5e2c156902ea40cf2a3dca672be9e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5bd35aef24be034a0b305a40bb53ca171a4f76db3b93b4f78ae95099b0a4cde3\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-08T18:11:14Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1008 18:11:09.091034 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1008 18:11:09.091984 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3472374062/tls.crt::/tmp/serving-cert-3472374062/tls.key\\\\\\\"\\\\nI1008 18:11:14.619696 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1008 18:11:14.628837 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1008 18:11:14.628898 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1008 18:11:14.628994 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1008 18:11:14.629036 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1008 18:11:14.641318 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1008 18:11:14.641363 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1008 18:11:14.641372 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1008 18:11:14.641382 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1008 18:11:14.641415 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1008 18:11:14.641423 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1008 18:11:14.641431 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1008 18:11:14.641484 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1008 18:11:14.646451 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T18:10:58Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c79ef704fda1bbcc486f4c7afa3039a3ab38669833f0f66bafe24dd4cc679e60\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:10:58Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://28d21731af0f16d6a8ae6c6e9d0bebf0560d4a533db4c78c10bf1837890394ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://28d21731af0f16d6a8ae6c6e9d0bebf0560d4a533db4c78c10bf1837890394ef\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:10:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:10:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:10:55Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:19Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:19 crc kubenswrapper[4988]: I1008 18:11:19.088456 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e8906182d55c285b3cf095e2d9e1dac397af4618a514f11e26b44c172dd573f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:19Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:19 crc kubenswrapper[4988]: I1008 18:11:19.129302 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:19Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:19 crc kubenswrapper[4988]: I1008 18:11:19.174966 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-jnh7m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"556fb048-8afb-44a5-9ea0-2b7c89e6cf4b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmlt2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1a825886f105ccf9a7cc6f28a0cd012fb859f4990dade0e6aae932482146028\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1a825886f105ccf9a7cc6f28a0cd012fb859f4990dade0e6aae932482146028\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmlt2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmlt2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmlt2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmlt2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmlt2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmlt2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:11:16Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-jnh7m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:19Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:19 crc kubenswrapper[4988]: I1008 18:11:19.208596 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f328a20b-5283-407e-8a93-2ab2da599f18\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8622d56fb198dadee31fb0ca8fc80d56828b6710648055f264533e2d5bdb3037\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z2gh4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3cdb2a757332bdcdc2a2f3362913fe8f06003b0597de20b72627f9badd5e30af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z2gh4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:11:16Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-nm54f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:19Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:19 crc kubenswrapper[4988]: I1008 18:11:19.236942 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 18:11:19 crc kubenswrapper[4988]: I1008 18:11:19.236967 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 18:11:19 crc kubenswrapper[4988]: E1008 18:11:19.237090 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 08 18:11:19 crc kubenswrapper[4988]: I1008 18:11:19.237149 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 18:11:19 crc kubenswrapper[4988]: E1008 18:11:19.237282 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 08 18:11:19 crc kubenswrapper[4988]: E1008 18:11:19.237396 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 08 18:11:19 crc kubenswrapper[4988]: I1008 18:11:19.249785 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a35a5e989ba69c8ef36ce4cea272aadf67ce20473fc257adfcfa86a2faba6ae0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a522e937cfe03fb9f2e60cb1bea07214cf63c69bdecd5ba65a2dfdba882d28dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:19Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:19 crc kubenswrapper[4988]: I1008 18:11:19.291355 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9590ca0059f0ebbe86343cccb3475338c317382c3231e082655b61e67eece375\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:19Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:19 crc kubenswrapper[4988]: I1008 18:11:19.333935 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-h9jwp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4490d846-ca32-4cd5-9202-87cc6d499fb4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://00b7f6db043c28efda3a09fdc77a6ccded3eae8b6a3d2852aa500789d5f01194\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7k6rk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:11:16Z\\\"}}\" for pod \"openshift-multus\"/\"multus-h9jwp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:19Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:19 crc kubenswrapper[4988]: I1008 18:11:19.373574 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-nsc65" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f010a463-69f2-4082-acc3-161899a939c3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dnwgr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dnwgr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dnwgr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dnwgr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dnwgr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dnwgr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dnwgr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dnwgr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://befe684cea9bd6dfda3ec8e85bca431c26bd8f82823b0a69dc0d04319a3ca385\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://befe684cea9bd6dfda3ec8e85bca431c26bd8f82823b0a69dc0d04319a3ca385\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dnwgr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:11:16Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-nsc65\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:19Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:19 crc kubenswrapper[4988]: I1008 18:11:19.408657 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:19Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:19 crc kubenswrapper[4988]: I1008 18:11:19.451719 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:19Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:19 crc kubenswrapper[4988]: I1008 18:11:19.475650 4988 generic.go:334] "Generic (PLEG): container finished" podID="556fb048-8afb-44a5-9ea0-2b7c89e6cf4b" containerID="0ab2fe6e65418efa75bae72259157a2323bbb4b7d11d1d012053524e14135888" exitCode=0 Oct 08 18:11:19 crc kubenswrapper[4988]: I1008 18:11:19.475780 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-jnh7m" event={"ID":"556fb048-8afb-44a5-9ea0-2b7c89e6cf4b","Type":"ContainerDied","Data":"0ab2fe6e65418efa75bae72259157a2323bbb4b7d11d1d012053524e14135888"} Oct 08 18:11:19 crc kubenswrapper[4988]: I1008 18:11:19.507139 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-jnh7m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"556fb048-8afb-44a5-9ea0-2b7c89e6cf4b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"message\\\":\\\"containers with incomplete status: [bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmlt2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1a825886f105ccf9a7cc6f28a0cd012fb859f4990dade0e6aae932482146028\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1a825886f105ccf9a7cc6f28a0cd012fb859f4990dade0e6aae932482146028\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmlt2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ab2fe6e65418efa75bae72259157a2323bbb4b7d11d1d012053524e14135888\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0ab2fe6e65418efa75bae72259157a2323bbb4b7d11d1d012053524e14135888\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:11:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:11:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmlt2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmlt2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmlt2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmlt2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmlt2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:11:16Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-jnh7m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:19Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:19 crc kubenswrapper[4988]: I1008 18:11:19.529315 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f328a20b-5283-407e-8a93-2ab2da599f18\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8622d56fb198dadee31fb0ca8fc80d56828b6710648055f264533e2d5bdb3037\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z2gh4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3cdb2a757332bdcdc2a2f3362913fe8f06003b0597de20b72627f9badd5e30af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z2gh4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:11:16Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-nm54f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:19Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:19 crc kubenswrapper[4988]: I1008 18:11:19.575373 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a35a5e989ba69c8ef36ce4cea272aadf67ce20473fc257adfcfa86a2faba6ae0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a522e937cfe03fb9f2e60cb1bea07214cf63c69bdecd5ba65a2dfdba882d28dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:19Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:19 crc kubenswrapper[4988]: I1008 18:11:19.609662 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9590ca0059f0ebbe86343cccb3475338c317382c3231e082655b61e67eece375\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:19Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:19 crc kubenswrapper[4988]: I1008 18:11:19.649807 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:19Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:19 crc kubenswrapper[4988]: I1008 18:11:19.698092 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-nsc65" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f010a463-69f2-4082-acc3-161899a939c3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dnwgr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dnwgr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dnwgr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dnwgr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dnwgr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dnwgr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dnwgr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dnwgr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://befe684cea9bd6dfda3ec8e85bca431c26bd8f82823b0a69dc0d04319a3ca385\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://befe684cea9bd6dfda3ec8e85bca431c26bd8f82823b0a69dc0d04319a3ca385\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dnwgr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:11:16Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-nsc65\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:19Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:19 crc kubenswrapper[4988]: I1008 18:11:19.732483 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:19Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:19 crc kubenswrapper[4988]: I1008 18:11:19.779625 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:19Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:19 crc kubenswrapper[4988]: I1008 18:11:19.815685 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-h9jwp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4490d846-ca32-4cd5-9202-87cc6d499fb4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://00b7f6db043c28efda3a09fdc77a6ccded3eae8b6a3d2852aa500789d5f01194\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7k6rk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:11:16Z\\\"}}\" for pod \"openshift-multus\"/\"multus-h9jwp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:19Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:19 crc kubenswrapper[4988]: I1008 18:11:19.850036 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-s4f2k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ba5e11d7-7bde-4c3d-983b-884232219652\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://824bc34a2807068d3fceb10651803ae9ae5651e940f7aed4877c544d76dae4d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hn5jf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:11:15Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-s4f2k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:19Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:19 crc kubenswrapper[4988]: I1008 18:11:19.889235 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aef71be1-1d12-46e5-95b2-bd399040fd4d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:10:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:10:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:10:55Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:10:55Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:10:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77c68f3ecb8a0724499e495cb26ef3db5beecb140f2558d8c0b25f41928b84ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:10:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a560ee84c2405227de71fda127566213cd4f017e73d5581442ba267781b1c539\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:10:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6100accc3bd08928bc1ff88310f29fa5da37fd2d5c315c94279801c9778c4745\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:10:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b0e596d7ce4e9341c44499af79d3b3caf0e5e2c156902ea40cf2a3dca672be9e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5bd35aef24be034a0b305a40bb53ca171a4f76db3b93b4f78ae95099b0a4cde3\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-08T18:11:14Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1008 18:11:09.091034 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1008 18:11:09.091984 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3472374062/tls.crt::/tmp/serving-cert-3472374062/tls.key\\\\\\\"\\\\nI1008 18:11:14.619696 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1008 18:11:14.628837 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1008 18:11:14.628898 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1008 18:11:14.628994 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1008 18:11:14.629036 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1008 18:11:14.641318 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1008 18:11:14.641363 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1008 18:11:14.641372 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1008 18:11:14.641382 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1008 18:11:14.641415 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1008 18:11:14.641423 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1008 18:11:14.641431 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1008 18:11:14.641484 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1008 18:11:14.646451 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T18:10:58Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c79ef704fda1bbcc486f4c7afa3039a3ab38669833f0f66bafe24dd4cc679e60\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:10:58Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://28d21731af0f16d6a8ae6c6e9d0bebf0560d4a533db4c78c10bf1837890394ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://28d21731af0f16d6a8ae6c6e9d0bebf0560d4a533db4c78c10bf1837890394ef\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:10:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:10:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:10:55Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:19Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:19 crc kubenswrapper[4988]: I1008 18:11:19.929426 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e8906182d55c285b3cf095e2d9e1dac397af4618a514f11e26b44c172dd573f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:19Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:19 crc kubenswrapper[4988]: I1008 18:11:19.964927 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-dzxs2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8f307b20-5e33-4725-99cb-210e25113532\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72ee803c6ef88524e0ce0391476f8e5a7dfbc858b095a3c2ae31bad10991ce3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q65f7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:11:16Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-dzxs2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:19Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:20 crc kubenswrapper[4988]: I1008 18:11:20.484397 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-nsc65" event={"ID":"f010a463-69f2-4082-acc3-161899a939c3","Type":"ContainerStarted","Data":"74fcb812b84291763cc6e5ccb20c12aaad580ad968dda4734a0bbfe33d2770ca"} Oct 08 18:11:20 crc kubenswrapper[4988]: I1008 18:11:20.487533 4988 generic.go:334] "Generic (PLEG): container finished" podID="556fb048-8afb-44a5-9ea0-2b7c89e6cf4b" containerID="368a303309be112d9c0b7ccd6538fbaca7085d97a8ec8259f36092f07147e7cb" exitCode=0 Oct 08 18:11:20 crc kubenswrapper[4988]: I1008 18:11:20.487593 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-jnh7m" event={"ID":"556fb048-8afb-44a5-9ea0-2b7c89e6cf4b","Type":"ContainerDied","Data":"368a303309be112d9c0b7ccd6538fbaca7085d97a8ec8259f36092f07147e7cb"} Oct 08 18:11:20 crc kubenswrapper[4988]: I1008 18:11:20.502633 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:20Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:20 crc kubenswrapper[4988]: I1008 18:11:20.518504 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-h9jwp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4490d846-ca32-4cd5-9202-87cc6d499fb4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://00b7f6db043c28efda3a09fdc77a6ccded3eae8b6a3d2852aa500789d5f01194\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7k6rk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:11:16Z\\\"}}\" for pod \"openshift-multus\"/\"multus-h9jwp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:20Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:20 crc kubenswrapper[4988]: I1008 18:11:20.538822 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-nsc65" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f010a463-69f2-4082-acc3-161899a939c3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dnwgr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dnwgr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dnwgr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dnwgr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dnwgr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dnwgr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dnwgr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dnwgr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://befe684cea9bd6dfda3ec8e85bca431c26bd8f82823b0a69dc0d04319a3ca385\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://befe684cea9bd6dfda3ec8e85bca431c26bd8f82823b0a69dc0d04319a3ca385\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dnwgr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:11:16Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-nsc65\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:20Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:20 crc kubenswrapper[4988]: I1008 18:11:20.557230 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:20Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:20 crc kubenswrapper[4988]: I1008 18:11:20.569116 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-s4f2k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ba5e11d7-7bde-4c3d-983b-884232219652\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://824bc34a2807068d3fceb10651803ae9ae5651e940f7aed4877c544d76dae4d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hn5jf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:11:15Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-s4f2k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:20Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:20 crc kubenswrapper[4988]: I1008 18:11:20.586661 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e8906182d55c285b3cf095e2d9e1dac397af4618a514f11e26b44c172dd573f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:20Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:20 crc kubenswrapper[4988]: I1008 18:11:20.602182 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-dzxs2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8f307b20-5e33-4725-99cb-210e25113532\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72ee803c6ef88524e0ce0391476f8e5a7dfbc858b095a3c2ae31bad10991ce3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q65f7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:11:16Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-dzxs2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:20Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:20 crc kubenswrapper[4988]: I1008 18:11:20.626026 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aef71be1-1d12-46e5-95b2-bd399040fd4d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:10:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:10:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:10:55Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:10:55Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:10:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77c68f3ecb8a0724499e495cb26ef3db5beecb140f2558d8c0b25f41928b84ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:10:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a560ee84c2405227de71fda127566213cd4f017e73d5581442ba267781b1c539\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:10:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6100accc3bd08928bc1ff88310f29fa5da37fd2d5c315c94279801c9778c4745\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:10:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b0e596d7ce4e9341c44499af79d3b3caf0e5e2c156902ea40cf2a3dca672be9e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5bd35aef24be034a0b305a40bb53ca171a4f76db3b93b4f78ae95099b0a4cde3\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-08T18:11:14Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1008 18:11:09.091034 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1008 18:11:09.091984 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3472374062/tls.crt::/tmp/serving-cert-3472374062/tls.key\\\\\\\"\\\\nI1008 18:11:14.619696 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1008 18:11:14.628837 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1008 18:11:14.628898 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1008 18:11:14.628994 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1008 18:11:14.629036 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1008 18:11:14.641318 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1008 18:11:14.641363 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1008 18:11:14.641372 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1008 18:11:14.641382 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1008 18:11:14.641415 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1008 18:11:14.641423 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1008 18:11:14.641431 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1008 18:11:14.641484 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1008 18:11:14.646451 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T18:10:58Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c79ef704fda1bbcc486f4c7afa3039a3ab38669833f0f66bafe24dd4cc679e60\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:10:58Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://28d21731af0f16d6a8ae6c6e9d0bebf0560d4a533db4c78c10bf1837890394ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://28d21731af0f16d6a8ae6c6e9d0bebf0560d4a533db4c78c10bf1837890394ef\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:10:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:10:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:10:55Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:20Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:20 crc kubenswrapper[4988]: I1008 18:11:20.643275 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9590ca0059f0ebbe86343cccb3475338c317382c3231e082655b61e67eece375\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:20Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:20 crc kubenswrapper[4988]: I1008 18:11:20.660418 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:20Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:20 crc kubenswrapper[4988]: I1008 18:11:20.694954 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-jnh7m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"556fb048-8afb-44a5-9ea0-2b7c89e6cf4b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"message\\\":\\\"containers with incomplete status: [routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmlt2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1a825886f105ccf9a7cc6f28a0cd012fb859f4990dade0e6aae932482146028\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1a825886f105ccf9a7cc6f28a0cd012fb859f4990dade0e6aae932482146028\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmlt2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ab2fe6e65418efa75bae72259157a2323bbb4b7d11d1d012053524e14135888\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0ab2fe6e65418efa75bae72259157a2323bbb4b7d11d1d012053524e14135888\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:11:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:11:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmlt2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://368a303309be112d9c0b7ccd6538fbaca7085d97a8ec8259f36092f07147e7cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://368a303309be112d9c0b7ccd6538fbaca7085d97a8ec8259f36092f07147e7cb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:11:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:11:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmlt2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmlt2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmlt2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmlt2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:11:16Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-jnh7m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:20Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:20 crc kubenswrapper[4988]: I1008 18:11:20.709778 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f328a20b-5283-407e-8a93-2ab2da599f18\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8622d56fb198dadee31fb0ca8fc80d56828b6710648055f264533e2d5bdb3037\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z2gh4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3cdb2a757332bdcdc2a2f3362913fe8f06003b0597de20b72627f9badd5e30af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z2gh4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:11:16Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-nm54f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:20Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:20 crc kubenswrapper[4988]: I1008 18:11:20.729946 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a35a5e989ba69c8ef36ce4cea272aadf67ce20473fc257adfcfa86a2faba6ae0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a522e937cfe03fb9f2e60cb1bea07214cf63c69bdecd5ba65a2dfdba882d28dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:20Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:21 crc kubenswrapper[4988]: I1008 18:11:21.010137 4988 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 08 18:11:21 crc kubenswrapper[4988]: I1008 18:11:21.012312 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:11:21 crc kubenswrapper[4988]: I1008 18:11:21.012357 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:11:21 crc kubenswrapper[4988]: I1008 18:11:21.012371 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:11:21 crc kubenswrapper[4988]: I1008 18:11:21.012553 4988 kubelet_node_status.go:76] "Attempting to register node" node="crc" Oct 08 18:11:21 crc kubenswrapper[4988]: I1008 18:11:21.021593 4988 kubelet_node_status.go:115] "Node was previously registered" node="crc" Oct 08 18:11:21 crc kubenswrapper[4988]: I1008 18:11:21.021925 4988 kubelet_node_status.go:79] "Successfully registered node" node="crc" Oct 08 18:11:21 crc kubenswrapper[4988]: I1008 18:11:21.023041 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:11:21 crc kubenswrapper[4988]: I1008 18:11:21.023072 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:11:21 crc kubenswrapper[4988]: I1008 18:11:21.023083 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:11:21 crc kubenswrapper[4988]: I1008 18:11:21.023102 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:11:21 crc kubenswrapper[4988]: I1008 18:11:21.023115 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:11:21Z","lastTransitionTime":"2025-10-08T18:11:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:11:21 crc kubenswrapper[4988]: E1008 18:11:21.035026 4988 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T18:11:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:21Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T18:11:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:21Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T18:11:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:21Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T18:11:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:21Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"95ee9fd0-8bc9-4ca5-9d79-5ee94d58bfeb\\\",\\\"systemUUID\\\":\\\"bd9489c4-fb7d-486d-b7ce-a13b6d7de36a\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:21Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:21 crc kubenswrapper[4988]: I1008 18:11:21.039423 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:11:21 crc kubenswrapper[4988]: I1008 18:11:21.039464 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:11:21 crc kubenswrapper[4988]: I1008 18:11:21.039478 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:11:21 crc kubenswrapper[4988]: I1008 18:11:21.039498 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:11:21 crc kubenswrapper[4988]: I1008 18:11:21.039510 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:11:21Z","lastTransitionTime":"2025-10-08T18:11:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:11:21 crc kubenswrapper[4988]: E1008 18:11:21.053898 4988 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T18:11:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:21Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T18:11:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:21Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T18:11:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:21Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T18:11:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:21Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"95ee9fd0-8bc9-4ca5-9d79-5ee94d58bfeb\\\",\\\"systemUUID\\\":\\\"bd9489c4-fb7d-486d-b7ce-a13b6d7de36a\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:21Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:21 crc kubenswrapper[4988]: I1008 18:11:21.058223 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:11:21 crc kubenswrapper[4988]: I1008 18:11:21.058265 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:11:21 crc kubenswrapper[4988]: I1008 18:11:21.058275 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:11:21 crc kubenswrapper[4988]: I1008 18:11:21.058294 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:11:21 crc kubenswrapper[4988]: I1008 18:11:21.058305 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:11:21Z","lastTransitionTime":"2025-10-08T18:11:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:11:21 crc kubenswrapper[4988]: E1008 18:11:21.075120 4988 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T18:11:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:21Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T18:11:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:21Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T18:11:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:21Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T18:11:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:21Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"95ee9fd0-8bc9-4ca5-9d79-5ee94d58bfeb\\\",\\\"systemUUID\\\":\\\"bd9489c4-fb7d-486d-b7ce-a13b6d7de36a\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:21Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:21 crc kubenswrapper[4988]: I1008 18:11:21.079045 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:11:21 crc kubenswrapper[4988]: I1008 18:11:21.079095 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:11:21 crc kubenswrapper[4988]: I1008 18:11:21.079106 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:11:21 crc kubenswrapper[4988]: I1008 18:11:21.079127 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:11:21 crc kubenswrapper[4988]: I1008 18:11:21.079140 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:11:21Z","lastTransitionTime":"2025-10-08T18:11:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:11:21 crc kubenswrapper[4988]: E1008 18:11:21.094545 4988 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T18:11:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:21Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T18:11:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:21Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T18:11:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:21Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T18:11:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:21Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"95ee9fd0-8bc9-4ca5-9d79-5ee94d58bfeb\\\",\\\"systemUUID\\\":\\\"bd9489c4-fb7d-486d-b7ce-a13b6d7de36a\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:21Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:21 crc kubenswrapper[4988]: I1008 18:11:21.098796 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:11:21 crc kubenswrapper[4988]: I1008 18:11:21.098840 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:11:21 crc kubenswrapper[4988]: I1008 18:11:21.098851 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:11:21 crc kubenswrapper[4988]: I1008 18:11:21.098870 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:11:21 crc kubenswrapper[4988]: I1008 18:11:21.098881 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:11:21Z","lastTransitionTime":"2025-10-08T18:11:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:11:21 crc kubenswrapper[4988]: E1008 18:11:21.111456 4988 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T18:11:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:21Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T18:11:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:21Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T18:11:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:21Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T18:11:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:21Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"95ee9fd0-8bc9-4ca5-9d79-5ee94d58bfeb\\\",\\\"systemUUID\\\":\\\"bd9489c4-fb7d-486d-b7ce-a13b6d7de36a\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:21Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:21 crc kubenswrapper[4988]: E1008 18:11:21.111585 4988 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Oct 08 18:11:21 crc kubenswrapper[4988]: I1008 18:11:21.113728 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:11:21 crc kubenswrapper[4988]: I1008 18:11:21.113779 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:11:21 crc kubenswrapper[4988]: I1008 18:11:21.113798 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:11:21 crc kubenswrapper[4988]: I1008 18:11:21.113820 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:11:21 crc kubenswrapper[4988]: I1008 18:11:21.113837 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:11:21Z","lastTransitionTime":"2025-10-08T18:11:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:11:21 crc kubenswrapper[4988]: I1008 18:11:21.217442 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:11:21 crc kubenswrapper[4988]: I1008 18:11:21.217490 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:11:21 crc kubenswrapper[4988]: I1008 18:11:21.217500 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:11:21 crc kubenswrapper[4988]: I1008 18:11:21.217519 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:11:21 crc kubenswrapper[4988]: I1008 18:11:21.217529 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:11:21Z","lastTransitionTime":"2025-10-08T18:11:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:11:21 crc kubenswrapper[4988]: I1008 18:11:21.237726 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 18:11:21 crc kubenswrapper[4988]: I1008 18:11:21.237799 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 18:11:21 crc kubenswrapper[4988]: I1008 18:11:21.237730 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 18:11:21 crc kubenswrapper[4988]: E1008 18:11:21.237870 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 08 18:11:21 crc kubenswrapper[4988]: E1008 18:11:21.238006 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 08 18:11:21 crc kubenswrapper[4988]: E1008 18:11:21.238251 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 08 18:11:21 crc kubenswrapper[4988]: I1008 18:11:21.321308 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:11:21 crc kubenswrapper[4988]: I1008 18:11:21.321373 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:11:21 crc kubenswrapper[4988]: I1008 18:11:21.321415 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:11:21 crc kubenswrapper[4988]: I1008 18:11:21.321445 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:11:21 crc kubenswrapper[4988]: I1008 18:11:21.321463 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:11:21Z","lastTransitionTime":"2025-10-08T18:11:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:11:21 crc kubenswrapper[4988]: I1008 18:11:21.424161 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:11:21 crc kubenswrapper[4988]: I1008 18:11:21.424219 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:11:21 crc kubenswrapper[4988]: I1008 18:11:21.424232 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:11:21 crc kubenswrapper[4988]: I1008 18:11:21.424249 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:11:21 crc kubenswrapper[4988]: I1008 18:11:21.424262 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:11:21Z","lastTransitionTime":"2025-10-08T18:11:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:11:21 crc kubenswrapper[4988]: I1008 18:11:21.500790 4988 generic.go:334] "Generic (PLEG): container finished" podID="556fb048-8afb-44a5-9ea0-2b7c89e6cf4b" containerID="9a8ca69001dddef643f982a97c74c291e30cb9fac5b0e894e80066b885ea9151" exitCode=0 Oct 08 18:11:21 crc kubenswrapper[4988]: I1008 18:11:21.500875 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-jnh7m" event={"ID":"556fb048-8afb-44a5-9ea0-2b7c89e6cf4b","Type":"ContainerDied","Data":"9a8ca69001dddef643f982a97c74c291e30cb9fac5b0e894e80066b885ea9151"} Oct 08 18:11:21 crc kubenswrapper[4988]: I1008 18:11:21.516560 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-s4f2k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ba5e11d7-7bde-4c3d-983b-884232219652\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://824bc34a2807068d3fceb10651803ae9ae5651e940f7aed4877c544d76dae4d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hn5jf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:11:15Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-s4f2k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:21Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:21 crc kubenswrapper[4988]: I1008 18:11:21.538696 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:11:21 crc kubenswrapper[4988]: I1008 18:11:21.538760 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:11:21 crc kubenswrapper[4988]: I1008 18:11:21.538776 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:11:21 crc kubenswrapper[4988]: I1008 18:11:21.538803 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:11:21 crc kubenswrapper[4988]: I1008 18:11:21.538823 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:11:21Z","lastTransitionTime":"2025-10-08T18:11:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:11:21 crc kubenswrapper[4988]: I1008 18:11:21.586922 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aef71be1-1d12-46e5-95b2-bd399040fd4d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:10:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:10:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:10:55Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:10:55Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:10:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77c68f3ecb8a0724499e495cb26ef3db5beecb140f2558d8c0b25f41928b84ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:10:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a560ee84c2405227de71fda127566213cd4f017e73d5581442ba267781b1c539\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:10:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6100accc3bd08928bc1ff88310f29fa5da37fd2d5c315c94279801c9778c4745\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:10:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b0e596d7ce4e9341c44499af79d3b3caf0e5e2c156902ea40cf2a3dca672be9e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5bd35aef24be034a0b305a40bb53ca171a4f76db3b93b4f78ae95099b0a4cde3\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-08T18:11:14Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1008 18:11:09.091034 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1008 18:11:09.091984 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3472374062/tls.crt::/tmp/serving-cert-3472374062/tls.key\\\\\\\"\\\\nI1008 18:11:14.619696 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1008 18:11:14.628837 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1008 18:11:14.628898 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1008 18:11:14.628994 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1008 18:11:14.629036 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1008 18:11:14.641318 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1008 18:11:14.641363 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1008 18:11:14.641372 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1008 18:11:14.641382 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1008 18:11:14.641415 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1008 18:11:14.641423 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1008 18:11:14.641431 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1008 18:11:14.641484 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1008 18:11:14.646451 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T18:10:58Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c79ef704fda1bbcc486f4c7afa3039a3ab38669833f0f66bafe24dd4cc679e60\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:10:58Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://28d21731af0f16d6a8ae6c6e9d0bebf0560d4a533db4c78c10bf1837890394ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://28d21731af0f16d6a8ae6c6e9d0bebf0560d4a533db4c78c10bf1837890394ef\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:10:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:10:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:10:55Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:21Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:21 crc kubenswrapper[4988]: I1008 18:11:21.608198 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e8906182d55c285b3cf095e2d9e1dac397af4618a514f11e26b44c172dd573f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:21Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:21 crc kubenswrapper[4988]: I1008 18:11:21.625887 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-dzxs2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8f307b20-5e33-4725-99cb-210e25113532\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72ee803c6ef88524e0ce0391476f8e5a7dfbc858b095a3c2ae31bad10991ce3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q65f7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:11:16Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-dzxs2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:21Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:21 crc kubenswrapper[4988]: I1008 18:11:21.640440 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-jnh7m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"556fb048-8afb-44a5-9ea0-2b7c89e6cf4b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmlt2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1a825886f105ccf9a7cc6f28a0cd012fb859f4990dade0e6aae932482146028\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1a825886f105ccf9a7cc6f28a0cd012fb859f4990dade0e6aae932482146028\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmlt2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ab2fe6e65418efa75bae72259157a2323bbb4b7d11d1d012053524e14135888\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0ab2fe6e65418efa75bae72259157a2323bbb4b7d11d1d012053524e14135888\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:11:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:11:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmlt2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://368a303309be112d9c0b7ccd6538fbaca7085d97a8ec8259f36092f07147e7cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://368a303309be112d9c0b7ccd6538fbaca7085d97a8ec8259f36092f07147e7cb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:11:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:11:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmlt2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a8ca69001dddef643f982a97c74c291e30cb9fac5b0e894e80066b885ea9151\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9a8ca69001dddef643f982a97c74c291e30cb9fac5b0e894e80066b885ea9151\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:11:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:11:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmlt2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmlt2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmlt2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:11:16Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-jnh7m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:21Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:21 crc kubenswrapper[4988]: I1008 18:11:21.641253 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:11:21 crc kubenswrapper[4988]: I1008 18:11:21.641293 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:11:21 crc kubenswrapper[4988]: I1008 18:11:21.641307 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:11:21 crc kubenswrapper[4988]: I1008 18:11:21.641329 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:11:21 crc kubenswrapper[4988]: I1008 18:11:21.641342 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:11:21Z","lastTransitionTime":"2025-10-08T18:11:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:11:21 crc kubenswrapper[4988]: I1008 18:11:21.643641 4988 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 08 18:11:21 crc kubenswrapper[4988]: I1008 18:11:21.647603 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 08 18:11:21 crc kubenswrapper[4988]: I1008 18:11:21.654329 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/kube-controller-manager-crc"] Oct 08 18:11:21 crc kubenswrapper[4988]: I1008 18:11:21.656300 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f328a20b-5283-407e-8a93-2ab2da599f18\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8622d56fb198dadee31fb0ca8fc80d56828b6710648055f264533e2d5bdb3037\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z2gh4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3cdb2a757332bdcdc2a2f3362913fe8f06003b0597de20b72627f9badd5e30af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z2gh4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:11:16Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-nm54f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:21Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:21 crc kubenswrapper[4988]: I1008 18:11:21.672793 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a35a5e989ba69c8ef36ce4cea272aadf67ce20473fc257adfcfa86a2faba6ae0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a522e937cfe03fb9f2e60cb1bea07214cf63c69bdecd5ba65a2dfdba882d28dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:21Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:21 crc kubenswrapper[4988]: I1008 18:11:21.685876 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9590ca0059f0ebbe86343cccb3475338c317382c3231e082655b61e67eece375\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:21Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:21 crc kubenswrapper[4988]: I1008 18:11:21.699470 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:21Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:21 crc kubenswrapper[4988]: I1008 18:11:21.719560 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-nsc65" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f010a463-69f2-4082-acc3-161899a939c3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dnwgr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dnwgr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dnwgr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dnwgr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dnwgr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dnwgr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dnwgr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dnwgr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://befe684cea9bd6dfda3ec8e85bca431c26bd8f82823b0a69dc0d04319a3ca385\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://befe684cea9bd6dfda3ec8e85bca431c26bd8f82823b0a69dc0d04319a3ca385\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dnwgr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:11:16Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-nsc65\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:21Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:21 crc kubenswrapper[4988]: I1008 18:11:21.731631 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:21Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:21 crc kubenswrapper[4988]: I1008 18:11:21.744658 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:11:21 crc kubenswrapper[4988]: I1008 18:11:21.744740 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:11:21 crc kubenswrapper[4988]: I1008 18:11:21.744766 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:11:21 crc kubenswrapper[4988]: I1008 18:11:21.744792 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:11:21 crc kubenswrapper[4988]: I1008 18:11:21.744808 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:11:21Z","lastTransitionTime":"2025-10-08T18:11:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:11:21 crc kubenswrapper[4988]: I1008 18:11:21.748702 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:21Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:21 crc kubenswrapper[4988]: I1008 18:11:21.765270 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-h9jwp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4490d846-ca32-4cd5-9202-87cc6d499fb4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://00b7f6db043c28efda3a09fdc77a6ccded3eae8b6a3d2852aa500789d5f01194\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7k6rk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:11:16Z\\\"}}\" for pod \"openshift-multus\"/\"multus-h9jwp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:21Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:21 crc kubenswrapper[4988]: I1008 18:11:21.782240 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a35a5e989ba69c8ef36ce4cea272aadf67ce20473fc257adfcfa86a2faba6ae0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a522e937cfe03fb9f2e60cb1bea07214cf63c69bdecd5ba65a2dfdba882d28dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:21Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:21 crc kubenswrapper[4988]: I1008 18:11:21.795717 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9590ca0059f0ebbe86343cccb3475338c317382c3231e082655b61e67eece375\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:21Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:21 crc kubenswrapper[4988]: I1008 18:11:21.809599 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:21Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:21 crc kubenswrapper[4988]: I1008 18:11:21.825569 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-jnh7m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"556fb048-8afb-44a5-9ea0-2b7c89e6cf4b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmlt2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1a825886f105ccf9a7cc6f28a0cd012fb859f4990dade0e6aae932482146028\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1a825886f105ccf9a7cc6f28a0cd012fb859f4990dade0e6aae932482146028\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmlt2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ab2fe6e65418efa75bae72259157a2323bbb4b7d11d1d012053524e14135888\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0ab2fe6e65418efa75bae72259157a2323bbb4b7d11d1d012053524e14135888\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:11:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:11:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmlt2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://368a303309be112d9c0b7ccd6538fbaca7085d97a8ec8259f36092f07147e7cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://368a303309be112d9c0b7ccd6538fbaca7085d97a8ec8259f36092f07147e7cb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:11:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:11:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmlt2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a8ca69001dddef643f982a97c74c291e30cb9fac5b0e894e80066b885ea9151\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9a8ca69001dddef643f982a97c74c291e30cb9fac5b0e894e80066b885ea9151\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:11:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:11:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmlt2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmlt2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmlt2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:11:16Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-jnh7m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:21Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:21 crc kubenswrapper[4988]: I1008 18:11:21.838020 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f328a20b-5283-407e-8a93-2ab2da599f18\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8622d56fb198dadee31fb0ca8fc80d56828b6710648055f264533e2d5bdb3037\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z2gh4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3cdb2a757332bdcdc2a2f3362913fe8f06003b0597de20b72627f9badd5e30af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z2gh4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:11:16Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-nm54f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:21Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:21 crc kubenswrapper[4988]: I1008 18:11:21.847451 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:11:21 crc kubenswrapper[4988]: I1008 18:11:21.847514 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:11:21 crc kubenswrapper[4988]: I1008 18:11:21.847528 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:11:21 crc kubenswrapper[4988]: I1008 18:11:21.847550 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:11:21 crc kubenswrapper[4988]: I1008 18:11:21.847564 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:11:21Z","lastTransitionTime":"2025-10-08T18:11:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:11:21 crc kubenswrapper[4988]: I1008 18:11:21.851593 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:21Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:21 crc kubenswrapper[4988]: I1008 18:11:21.863561 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:21Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:21 crc kubenswrapper[4988]: I1008 18:11:21.879120 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-h9jwp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4490d846-ca32-4cd5-9202-87cc6d499fb4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://00b7f6db043c28efda3a09fdc77a6ccded3eae8b6a3d2852aa500789d5f01194\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7k6rk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:11:16Z\\\"}}\" for pod \"openshift-multus\"/\"multus-h9jwp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:21Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:21 crc kubenswrapper[4988]: I1008 18:11:21.900018 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-nsc65" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f010a463-69f2-4082-acc3-161899a939c3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dnwgr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dnwgr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dnwgr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dnwgr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dnwgr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dnwgr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dnwgr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dnwgr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://befe684cea9bd6dfda3ec8e85bca431c26bd8f82823b0a69dc0d04319a3ca385\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://befe684cea9bd6dfda3ec8e85bca431c26bd8f82823b0a69dc0d04319a3ca385\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dnwgr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:11:16Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-nsc65\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:21Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:21 crc kubenswrapper[4988]: I1008 18:11:21.913259 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-s4f2k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ba5e11d7-7bde-4c3d-983b-884232219652\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://824bc34a2807068d3fceb10651803ae9ae5651e940f7aed4877c544d76dae4d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hn5jf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:11:15Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-s4f2k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:21Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:21 crc kubenswrapper[4988]: I1008 18:11:21.925685 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aef71be1-1d12-46e5-95b2-bd399040fd4d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:10:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:10:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:10:55Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:10:55Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:10:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77c68f3ecb8a0724499e495cb26ef3db5beecb140f2558d8c0b25f41928b84ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:10:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a560ee84c2405227de71fda127566213cd4f017e73d5581442ba267781b1c539\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:10:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6100accc3bd08928bc1ff88310f29fa5da37fd2d5c315c94279801c9778c4745\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:10:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b0e596d7ce4e9341c44499af79d3b3caf0e5e2c156902ea40cf2a3dca672be9e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5bd35aef24be034a0b305a40bb53ca171a4f76db3b93b4f78ae95099b0a4cde3\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-08T18:11:14Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1008 18:11:09.091034 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1008 18:11:09.091984 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3472374062/tls.crt::/tmp/serving-cert-3472374062/tls.key\\\\\\\"\\\\nI1008 18:11:14.619696 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1008 18:11:14.628837 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1008 18:11:14.628898 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1008 18:11:14.628994 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1008 18:11:14.629036 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1008 18:11:14.641318 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1008 18:11:14.641363 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1008 18:11:14.641372 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1008 18:11:14.641382 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1008 18:11:14.641415 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1008 18:11:14.641423 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1008 18:11:14.641431 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1008 18:11:14.641484 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1008 18:11:14.646451 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T18:10:58Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c79ef704fda1bbcc486f4c7afa3039a3ab38669833f0f66bafe24dd4cc679e60\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:10:58Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://28d21731af0f16d6a8ae6c6e9d0bebf0560d4a533db4c78c10bf1837890394ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://28d21731af0f16d6a8ae6c6e9d0bebf0560d4a533db4c78c10bf1837890394ef\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:10:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:10:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:10:55Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:21Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:21 crc kubenswrapper[4988]: I1008 18:11:21.945993 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7e0ecefc-359f-4785-bb79-0343a62a25da\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:10:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:10:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:10:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://30bec350fefee3d19758b3ed735ede4e9aa085edb57aa3f3ec24249a67dc7ca2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:10:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2d850fd27d4ed65891ff69b9add65fb78535357f54819c2a8d196f5a340c58e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:10:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://90b1a9e14c82bf0f75f08fdb342eed35a503020f9db641e5e9ceff8b122cb3bd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:10:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://16f23c4c7163d32dc0df8d6a7ee17e2864fb261f8a40ecf05556774b173fc868\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:10:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:10:55Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:21Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:21 crc kubenswrapper[4988]: I1008 18:11:21.949945 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:11:21 crc kubenswrapper[4988]: I1008 18:11:21.949988 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:11:21 crc kubenswrapper[4988]: I1008 18:11:21.950003 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:11:21 crc kubenswrapper[4988]: I1008 18:11:21.950027 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:11:21 crc kubenswrapper[4988]: I1008 18:11:21.950041 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:11:21Z","lastTransitionTime":"2025-10-08T18:11:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:11:21 crc kubenswrapper[4988]: I1008 18:11:21.965464 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e8906182d55c285b3cf095e2d9e1dac397af4618a514f11e26b44c172dd573f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:21Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:21 crc kubenswrapper[4988]: I1008 18:11:21.978813 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-dzxs2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8f307b20-5e33-4725-99cb-210e25113532\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72ee803c6ef88524e0ce0391476f8e5a7dfbc858b095a3c2ae31bad10991ce3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q65f7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:11:16Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-dzxs2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:21Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:22 crc kubenswrapper[4988]: I1008 18:11:22.053282 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:11:22 crc kubenswrapper[4988]: I1008 18:11:22.053332 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:11:22 crc kubenswrapper[4988]: I1008 18:11:22.053344 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:11:22 crc kubenswrapper[4988]: I1008 18:11:22.053363 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:11:22 crc kubenswrapper[4988]: I1008 18:11:22.053377 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:11:22Z","lastTransitionTime":"2025-10-08T18:11:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:11:22 crc kubenswrapper[4988]: I1008 18:11:22.155792 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:11:22 crc kubenswrapper[4988]: I1008 18:11:22.155863 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:11:22 crc kubenswrapper[4988]: I1008 18:11:22.155886 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:11:22 crc kubenswrapper[4988]: I1008 18:11:22.155915 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:11:22 crc kubenswrapper[4988]: I1008 18:11:22.155939 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:11:22Z","lastTransitionTime":"2025-10-08T18:11:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:11:22 crc kubenswrapper[4988]: I1008 18:11:22.262846 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:11:22 crc kubenswrapper[4988]: I1008 18:11:22.262909 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:11:22 crc kubenswrapper[4988]: I1008 18:11:22.262921 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:11:22 crc kubenswrapper[4988]: I1008 18:11:22.262956 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:11:22 crc kubenswrapper[4988]: I1008 18:11:22.262975 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:11:22Z","lastTransitionTime":"2025-10-08T18:11:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:11:22 crc kubenswrapper[4988]: I1008 18:11:22.405807 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:11:22 crc kubenswrapper[4988]: I1008 18:11:22.405907 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:11:22 crc kubenswrapper[4988]: I1008 18:11:22.405974 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:11:22 crc kubenswrapper[4988]: I1008 18:11:22.406053 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:11:22 crc kubenswrapper[4988]: I1008 18:11:22.406083 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:11:22Z","lastTransitionTime":"2025-10-08T18:11:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:11:22 crc kubenswrapper[4988]: I1008 18:11:22.508972 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:11:22 crc kubenswrapper[4988]: I1008 18:11:22.509055 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:11:22 crc kubenswrapper[4988]: I1008 18:11:22.509081 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:11:22 crc kubenswrapper[4988]: I1008 18:11:22.509120 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:11:22 crc kubenswrapper[4988]: I1008 18:11:22.509149 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:11:22Z","lastTransitionTime":"2025-10-08T18:11:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:11:22 crc kubenswrapper[4988]: I1008 18:11:22.511139 4988 generic.go:334] "Generic (PLEG): container finished" podID="556fb048-8afb-44a5-9ea0-2b7c89e6cf4b" containerID="145303c46b48603ed1aae4b9bfe06332a3ffad28f206f8394574011a87b58cd3" exitCode=0 Oct 08 18:11:22 crc kubenswrapper[4988]: I1008 18:11:22.511204 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-jnh7m" event={"ID":"556fb048-8afb-44a5-9ea0-2b7c89e6cf4b","Type":"ContainerDied","Data":"145303c46b48603ed1aae4b9bfe06332a3ffad28f206f8394574011a87b58cd3"} Oct 08 18:11:22 crc kubenswrapper[4988]: E1008 18:11:22.522758 4988 kubelet.go:1929] "Failed creating a mirror pod for" err="pods \"kube-controller-manager-crc\" already exists" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 08 18:11:22 crc kubenswrapper[4988]: I1008 18:11:22.533493 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:22Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:22 crc kubenswrapper[4988]: I1008 18:11:22.556249 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-h9jwp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4490d846-ca32-4cd5-9202-87cc6d499fb4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://00b7f6db043c28efda3a09fdc77a6ccded3eae8b6a3d2852aa500789d5f01194\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7k6rk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:11:16Z\\\"}}\" for pod \"openshift-multus\"/\"multus-h9jwp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:22Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:22 crc kubenswrapper[4988]: I1008 18:11:22.575183 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-nsc65" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f010a463-69f2-4082-acc3-161899a939c3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dnwgr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dnwgr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dnwgr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dnwgr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dnwgr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dnwgr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dnwgr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dnwgr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://befe684cea9bd6dfda3ec8e85bca431c26bd8f82823b0a69dc0d04319a3ca385\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://befe684cea9bd6dfda3ec8e85bca431c26bd8f82823b0a69dc0d04319a3ca385\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dnwgr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:11:16Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-nsc65\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:22Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:22 crc kubenswrapper[4988]: I1008 18:11:22.594646 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:22Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:22 crc kubenswrapper[4988]: I1008 18:11:22.608439 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-s4f2k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ba5e11d7-7bde-4c3d-983b-884232219652\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://824bc34a2807068d3fceb10651803ae9ae5651e940f7aed4877c544d76dae4d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hn5jf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:11:15Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-s4f2k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:22Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:22 crc kubenswrapper[4988]: I1008 18:11:22.612964 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:11:22 crc kubenswrapper[4988]: I1008 18:11:22.613037 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:11:22 crc kubenswrapper[4988]: I1008 18:11:22.613056 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:11:22 crc kubenswrapper[4988]: I1008 18:11:22.613086 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:11:22 crc kubenswrapper[4988]: I1008 18:11:22.613108 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:11:22Z","lastTransitionTime":"2025-10-08T18:11:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:11:22 crc kubenswrapper[4988]: I1008 18:11:22.626755 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e8906182d55c285b3cf095e2d9e1dac397af4618a514f11e26b44c172dd573f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:22Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:22 crc kubenswrapper[4988]: I1008 18:11:22.643615 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-dzxs2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8f307b20-5e33-4725-99cb-210e25113532\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72ee803c6ef88524e0ce0391476f8e5a7dfbc858b095a3c2ae31bad10991ce3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q65f7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:11:16Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-dzxs2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:22Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:22 crc kubenswrapper[4988]: I1008 18:11:22.661264 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aef71be1-1d12-46e5-95b2-bd399040fd4d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:10:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:10:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:10:55Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:10:55Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:10:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77c68f3ecb8a0724499e495cb26ef3db5beecb140f2558d8c0b25f41928b84ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:10:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a560ee84c2405227de71fda127566213cd4f017e73d5581442ba267781b1c539\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:10:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6100accc3bd08928bc1ff88310f29fa5da37fd2d5c315c94279801c9778c4745\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:10:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b0e596d7ce4e9341c44499af79d3b3caf0e5e2c156902ea40cf2a3dca672be9e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5bd35aef24be034a0b305a40bb53ca171a4f76db3b93b4f78ae95099b0a4cde3\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-08T18:11:14Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1008 18:11:09.091034 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1008 18:11:09.091984 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3472374062/tls.crt::/tmp/serving-cert-3472374062/tls.key\\\\\\\"\\\\nI1008 18:11:14.619696 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1008 18:11:14.628837 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1008 18:11:14.628898 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1008 18:11:14.628994 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1008 18:11:14.629036 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1008 18:11:14.641318 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1008 18:11:14.641363 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1008 18:11:14.641372 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1008 18:11:14.641382 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1008 18:11:14.641415 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1008 18:11:14.641423 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1008 18:11:14.641431 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1008 18:11:14.641484 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1008 18:11:14.646451 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T18:10:58Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c79ef704fda1bbcc486f4c7afa3039a3ab38669833f0f66bafe24dd4cc679e60\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:10:58Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://28d21731af0f16d6a8ae6c6e9d0bebf0560d4a533db4c78c10bf1837890394ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://28d21731af0f16d6a8ae6c6e9d0bebf0560d4a533db4c78c10bf1837890394ef\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:10:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:10:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:10:55Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:22Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:22 crc kubenswrapper[4988]: I1008 18:11:22.672914 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7e0ecefc-359f-4785-bb79-0343a62a25da\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:10:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:10:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:10:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://30bec350fefee3d19758b3ed735ede4e9aa085edb57aa3f3ec24249a67dc7ca2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:10:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2d850fd27d4ed65891ff69b9add65fb78535357f54819c2a8d196f5a340c58e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:10:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://90b1a9e14c82bf0f75f08fdb342eed35a503020f9db641e5e9ceff8b122cb3bd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:10:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://16f23c4c7163d32dc0df8d6a7ee17e2864fb261f8a40ecf05556774b173fc868\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:10:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:10:55Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:22Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:22 crc kubenswrapper[4988]: I1008 18:11:22.686145 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9590ca0059f0ebbe86343cccb3475338c317382c3231e082655b61e67eece375\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:22Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:22 crc kubenswrapper[4988]: I1008 18:11:22.704078 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:22Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:22 crc kubenswrapper[4988]: I1008 18:11:22.716418 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:11:22 crc kubenswrapper[4988]: I1008 18:11:22.716471 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:11:22 crc kubenswrapper[4988]: I1008 18:11:22.716488 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:11:22 crc kubenswrapper[4988]: I1008 18:11:22.716513 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:11:22 crc kubenswrapper[4988]: I1008 18:11:22.716530 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:11:22Z","lastTransitionTime":"2025-10-08T18:11:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:11:22 crc kubenswrapper[4988]: I1008 18:11:22.727244 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-jnh7m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"556fb048-8afb-44a5-9ea0-2b7c89e6cf4b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmlt2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1a825886f105ccf9a7cc6f28a0cd012fb859f4990dade0e6aae932482146028\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1a825886f105ccf9a7cc6f28a0cd012fb859f4990dade0e6aae932482146028\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmlt2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ab2fe6e65418efa75bae72259157a2323bbb4b7d11d1d012053524e14135888\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0ab2fe6e65418efa75bae72259157a2323bbb4b7d11d1d012053524e14135888\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:11:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:11:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmlt2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://368a303309be112d9c0b7ccd6538fbaca7085d97a8ec8259f36092f07147e7cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://368a303309be112d9c0b7ccd6538fbaca7085d97a8ec8259f36092f07147e7cb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:11:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:11:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmlt2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a8ca69001dddef643f982a97c74c291e30cb9fac5b0e894e80066b885ea9151\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9a8ca69001dddef643f982a97c74c291e30cb9fac5b0e894e80066b885ea9151\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:11:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:11:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmlt2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://145303c46b48603ed1aae4b9bfe06332a3ffad28f206f8394574011a87b58cd3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://145303c46b48603ed1aae4b9bfe06332a3ffad28f206f8394574011a87b58cd3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:11:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:11:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmlt2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmlt2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:11:16Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-jnh7m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:22Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:22 crc kubenswrapper[4988]: I1008 18:11:22.747550 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f328a20b-5283-407e-8a93-2ab2da599f18\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8622d56fb198dadee31fb0ca8fc80d56828b6710648055f264533e2d5bdb3037\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z2gh4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3cdb2a757332bdcdc2a2f3362913fe8f06003b0597de20b72627f9badd5e30af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z2gh4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:11:16Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-nm54f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:22Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:22 crc kubenswrapper[4988]: I1008 18:11:22.766118 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a35a5e989ba69c8ef36ce4cea272aadf67ce20473fc257adfcfa86a2faba6ae0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a522e937cfe03fb9f2e60cb1bea07214cf63c69bdecd5ba65a2dfdba882d28dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:22Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:22 crc kubenswrapper[4988]: I1008 18:11:22.820211 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:11:22 crc kubenswrapper[4988]: I1008 18:11:22.820280 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:11:22 crc kubenswrapper[4988]: I1008 18:11:22.820299 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:11:22 crc kubenswrapper[4988]: I1008 18:11:22.820329 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:11:22 crc kubenswrapper[4988]: I1008 18:11:22.820350 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:11:22Z","lastTransitionTime":"2025-10-08T18:11:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:11:22 crc kubenswrapper[4988]: I1008 18:11:22.923968 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:11:22 crc kubenswrapper[4988]: I1008 18:11:22.924036 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:11:22 crc kubenswrapper[4988]: I1008 18:11:22.924052 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:11:22 crc kubenswrapper[4988]: I1008 18:11:22.924077 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:11:22 crc kubenswrapper[4988]: I1008 18:11:22.924094 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:11:22Z","lastTransitionTime":"2025-10-08T18:11:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:11:22 crc kubenswrapper[4988]: I1008 18:11:22.924245 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 18:11:22 crc kubenswrapper[4988]: E1008 18:11:22.924402 4988 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 18:11:30.924359314 +0000 UTC m=+36.374202084 (durationBeforeRetry 8s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 18:11:22 crc kubenswrapper[4988]: I1008 18:11:22.924490 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 18:11:22 crc kubenswrapper[4988]: I1008 18:11:22.924536 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 18:11:22 crc kubenswrapper[4988]: E1008 18:11:22.924688 4988 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 08 18:11:22 crc kubenswrapper[4988]: E1008 18:11:22.924692 4988 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 08 18:11:22 crc kubenswrapper[4988]: E1008 18:11:22.924742 4988 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-08 18:11:30.924732625 +0000 UTC m=+36.374575395 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 08 18:11:22 crc kubenswrapper[4988]: E1008 18:11:22.924795 4988 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-08 18:11:30.924767587 +0000 UTC m=+36.374610367 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 08 18:11:23 crc kubenswrapper[4988]: I1008 18:11:23.025963 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 18:11:23 crc kubenswrapper[4988]: E1008 18:11:23.026245 4988 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 08 18:11:23 crc kubenswrapper[4988]: E1008 18:11:23.026626 4988 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 08 18:11:23 crc kubenswrapper[4988]: E1008 18:11:23.026649 4988 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 08 18:11:23 crc kubenswrapper[4988]: I1008 18:11:23.026554 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 18:11:23 crc kubenswrapper[4988]: E1008 18:11:23.026703 4988 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 08 18:11:23 crc kubenswrapper[4988]: E1008 18:11:23.026720 4988 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 08 18:11:23 crc kubenswrapper[4988]: E1008 18:11:23.026726 4988 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-08 18:11:31.026703298 +0000 UTC m=+36.476546068 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 08 18:11:23 crc kubenswrapper[4988]: E1008 18:11:23.026731 4988 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 08 18:11:23 crc kubenswrapper[4988]: E1008 18:11:23.026768 4988 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-08 18:11:31.02675706 +0000 UTC m=+36.476599830 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 08 18:11:23 crc kubenswrapper[4988]: I1008 18:11:23.028092 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:11:23 crc kubenswrapper[4988]: I1008 18:11:23.028128 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:11:23 crc kubenswrapper[4988]: I1008 18:11:23.028143 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:11:23 crc kubenswrapper[4988]: I1008 18:11:23.028171 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:11:23 crc kubenswrapper[4988]: I1008 18:11:23.028187 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:11:23Z","lastTransitionTime":"2025-10-08T18:11:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:11:23 crc kubenswrapper[4988]: I1008 18:11:23.130677 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:11:23 crc kubenswrapper[4988]: I1008 18:11:23.130725 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:11:23 crc kubenswrapper[4988]: I1008 18:11:23.130735 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:11:23 crc kubenswrapper[4988]: I1008 18:11:23.130753 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:11:23 crc kubenswrapper[4988]: I1008 18:11:23.130766 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:11:23Z","lastTransitionTime":"2025-10-08T18:11:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:11:23 crc kubenswrapper[4988]: I1008 18:11:23.233775 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:11:23 crc kubenswrapper[4988]: I1008 18:11:23.233823 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:11:23 crc kubenswrapper[4988]: I1008 18:11:23.233837 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:11:23 crc kubenswrapper[4988]: I1008 18:11:23.233855 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:11:23 crc kubenswrapper[4988]: I1008 18:11:23.233866 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:11:23Z","lastTransitionTime":"2025-10-08T18:11:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:11:23 crc kubenswrapper[4988]: I1008 18:11:23.237156 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 18:11:23 crc kubenswrapper[4988]: E1008 18:11:23.237273 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 08 18:11:23 crc kubenswrapper[4988]: I1008 18:11:23.237767 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 18:11:23 crc kubenswrapper[4988]: E1008 18:11:23.237832 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 08 18:11:23 crc kubenswrapper[4988]: I1008 18:11:23.237959 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 18:11:23 crc kubenswrapper[4988]: E1008 18:11:23.238007 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 08 18:11:23 crc kubenswrapper[4988]: I1008 18:11:23.337194 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:11:23 crc kubenswrapper[4988]: I1008 18:11:23.337270 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:11:23 crc kubenswrapper[4988]: I1008 18:11:23.337290 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:11:23 crc kubenswrapper[4988]: I1008 18:11:23.337321 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:11:23 crc kubenswrapper[4988]: I1008 18:11:23.337342 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:11:23Z","lastTransitionTime":"2025-10-08T18:11:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:11:23 crc kubenswrapper[4988]: I1008 18:11:23.440672 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:11:23 crc kubenswrapper[4988]: I1008 18:11:23.440744 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:11:23 crc kubenswrapper[4988]: I1008 18:11:23.440764 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:11:23 crc kubenswrapper[4988]: I1008 18:11:23.440796 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:11:23 crc kubenswrapper[4988]: I1008 18:11:23.440815 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:11:23Z","lastTransitionTime":"2025-10-08T18:11:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:11:23 crc kubenswrapper[4988]: I1008 18:11:23.529302 4988 generic.go:334] "Generic (PLEG): container finished" podID="556fb048-8afb-44a5-9ea0-2b7c89e6cf4b" containerID="d08e5cfc7cbe5ec1a7a0265ed7c466363c3abeec492263170860689b46c7ff95" exitCode=0 Oct 08 18:11:23 crc kubenswrapper[4988]: I1008 18:11:23.529472 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-jnh7m" event={"ID":"556fb048-8afb-44a5-9ea0-2b7c89e6cf4b","Type":"ContainerDied","Data":"d08e5cfc7cbe5ec1a7a0265ed7c466363c3abeec492263170860689b46c7ff95"} Oct 08 18:11:23 crc kubenswrapper[4988]: I1008 18:11:23.535912 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-nsc65" event={"ID":"f010a463-69f2-4082-acc3-161899a939c3","Type":"ContainerStarted","Data":"9353d08254f16c1a3cd3b66904805f54467292dd2c052268336aaca5678aeef7"} Oct 08 18:11:23 crc kubenswrapper[4988]: I1008 18:11:23.537790 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-nsc65" Oct 08 18:11:23 crc kubenswrapper[4988]: I1008 18:11:23.538002 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-nsc65" Oct 08 18:11:23 crc kubenswrapper[4988]: I1008 18:11:23.544085 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:11:23 crc kubenswrapper[4988]: I1008 18:11:23.544135 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:11:23 crc kubenswrapper[4988]: I1008 18:11:23.544154 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:11:23 crc kubenswrapper[4988]: I1008 18:11:23.544182 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:11:23 crc kubenswrapper[4988]: I1008 18:11:23.544202 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:11:23Z","lastTransitionTime":"2025-10-08T18:11:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:11:23 crc kubenswrapper[4988]: I1008 18:11:23.551734 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:23Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:23 crc kubenswrapper[4988]: I1008 18:11:23.578751 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:23Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:23 crc kubenswrapper[4988]: I1008 18:11:23.581038 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-nsc65" Oct 08 18:11:23 crc kubenswrapper[4988]: I1008 18:11:23.584404 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-nsc65" Oct 08 18:11:23 crc kubenswrapper[4988]: I1008 18:11:23.600088 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-h9jwp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4490d846-ca32-4cd5-9202-87cc6d499fb4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://00b7f6db043c28efda3a09fdc77a6ccded3eae8b6a3d2852aa500789d5f01194\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7k6rk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:11:16Z\\\"}}\" for pod \"openshift-multus\"/\"multus-h9jwp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:23Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:23 crc kubenswrapper[4988]: I1008 18:11:23.624850 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-nsc65" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f010a463-69f2-4082-acc3-161899a939c3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dnwgr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dnwgr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dnwgr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dnwgr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dnwgr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dnwgr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dnwgr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dnwgr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://befe684cea9bd6dfda3ec8e85bca431c26bd8f82823b0a69dc0d04319a3ca385\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://befe684cea9bd6dfda3ec8e85bca431c26bd8f82823b0a69dc0d04319a3ca385\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dnwgr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:11:16Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-nsc65\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:23Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:23 crc kubenswrapper[4988]: I1008 18:11:23.639574 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-s4f2k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ba5e11d7-7bde-4c3d-983b-884232219652\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://824bc34a2807068d3fceb10651803ae9ae5651e940f7aed4877c544d76dae4d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hn5jf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:11:15Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-s4f2k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:23Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:23 crc kubenswrapper[4988]: I1008 18:11:23.646924 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:11:23 crc kubenswrapper[4988]: I1008 18:11:23.646963 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:11:23 crc kubenswrapper[4988]: I1008 18:11:23.646974 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:11:23 crc kubenswrapper[4988]: I1008 18:11:23.646992 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:11:23 crc kubenswrapper[4988]: I1008 18:11:23.647002 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:11:23Z","lastTransitionTime":"2025-10-08T18:11:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:11:23 crc kubenswrapper[4988]: I1008 18:11:23.657871 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7e0ecefc-359f-4785-bb79-0343a62a25da\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:10:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:10:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:10:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://30bec350fefee3d19758b3ed735ede4e9aa085edb57aa3f3ec24249a67dc7ca2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:10:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2d850fd27d4ed65891ff69b9add65fb78535357f54819c2a8d196f5a340c58e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:10:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://90b1a9e14c82bf0f75f08fdb342eed35a503020f9db641e5e9ceff8b122cb3bd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:10:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://16f23c4c7163d32dc0df8d6a7ee17e2864fb261f8a40ecf05556774b173fc868\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:10:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:10:55Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:23Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:23 crc kubenswrapper[4988]: I1008 18:11:23.674287 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e8906182d55c285b3cf095e2d9e1dac397af4618a514f11e26b44c172dd573f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:23Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:23 crc kubenswrapper[4988]: I1008 18:11:23.687193 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-dzxs2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8f307b20-5e33-4725-99cb-210e25113532\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72ee803c6ef88524e0ce0391476f8e5a7dfbc858b095a3c2ae31bad10991ce3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q65f7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:11:16Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-dzxs2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:23Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:23 crc kubenswrapper[4988]: I1008 18:11:23.701033 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aef71be1-1d12-46e5-95b2-bd399040fd4d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:10:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:10:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:10:55Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:10:55Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:10:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77c68f3ecb8a0724499e495cb26ef3db5beecb140f2558d8c0b25f41928b84ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:10:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a560ee84c2405227de71fda127566213cd4f017e73d5581442ba267781b1c539\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:10:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6100accc3bd08928bc1ff88310f29fa5da37fd2d5c315c94279801c9778c4745\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:10:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b0e596d7ce4e9341c44499af79d3b3caf0e5e2c156902ea40cf2a3dca672be9e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5bd35aef24be034a0b305a40bb53ca171a4f76db3b93b4f78ae95099b0a4cde3\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-08T18:11:14Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1008 18:11:09.091034 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1008 18:11:09.091984 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3472374062/tls.crt::/tmp/serving-cert-3472374062/tls.key\\\\\\\"\\\\nI1008 18:11:14.619696 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1008 18:11:14.628837 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1008 18:11:14.628898 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1008 18:11:14.628994 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1008 18:11:14.629036 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1008 18:11:14.641318 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1008 18:11:14.641363 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1008 18:11:14.641372 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1008 18:11:14.641382 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1008 18:11:14.641415 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1008 18:11:14.641423 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1008 18:11:14.641431 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1008 18:11:14.641484 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1008 18:11:14.646451 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T18:10:58Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c79ef704fda1bbcc486f4c7afa3039a3ab38669833f0f66bafe24dd4cc679e60\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:10:58Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://28d21731af0f16d6a8ae6c6e9d0bebf0560d4a533db4c78c10bf1837890394ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://28d21731af0f16d6a8ae6c6e9d0bebf0560d4a533db4c78c10bf1837890394ef\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:10:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:10:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:10:55Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:23Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:23 crc kubenswrapper[4988]: I1008 18:11:23.723312 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a35a5e989ba69c8ef36ce4cea272aadf67ce20473fc257adfcfa86a2faba6ae0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a522e937cfe03fb9f2e60cb1bea07214cf63c69bdecd5ba65a2dfdba882d28dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:23Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:23 crc kubenswrapper[4988]: I1008 18:11:23.737429 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9590ca0059f0ebbe86343cccb3475338c317382c3231e082655b61e67eece375\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:23Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:23 crc kubenswrapper[4988]: I1008 18:11:23.749666 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:11:23 crc kubenswrapper[4988]: I1008 18:11:23.749724 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:11:23 crc kubenswrapper[4988]: I1008 18:11:23.749736 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:11:23 crc kubenswrapper[4988]: I1008 18:11:23.749757 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:11:23 crc kubenswrapper[4988]: I1008 18:11:23.749772 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:11:23Z","lastTransitionTime":"2025-10-08T18:11:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:11:23 crc kubenswrapper[4988]: I1008 18:11:23.760796 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:23Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:23 crc kubenswrapper[4988]: I1008 18:11:23.778129 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-jnh7m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"556fb048-8afb-44a5-9ea0-2b7c89e6cf4b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmlt2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1a825886f105ccf9a7cc6f28a0cd012fb859f4990dade0e6aae932482146028\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1a825886f105ccf9a7cc6f28a0cd012fb859f4990dade0e6aae932482146028\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmlt2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ab2fe6e65418efa75bae72259157a2323bbb4b7d11d1d012053524e14135888\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0ab2fe6e65418efa75bae72259157a2323bbb4b7d11d1d012053524e14135888\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:11:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:11:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmlt2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://368a303309be112d9c0b7ccd6538fbaca7085d97a8ec8259f36092f07147e7cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://368a303309be112d9c0b7ccd6538fbaca7085d97a8ec8259f36092f07147e7cb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:11:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:11:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmlt2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a8ca69001dddef643f982a97c74c291e30cb9fac5b0e894e80066b885ea9151\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9a8ca69001dddef643f982a97c74c291e30cb9fac5b0e894e80066b885ea9151\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:11:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:11:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmlt2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://145303c46b48603ed1aae4b9bfe06332a3ffad28f206f8394574011a87b58cd3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://145303c46b48603ed1aae4b9bfe06332a3ffad28f206f8394574011a87b58cd3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:11:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:11:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmlt2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d08e5cfc7cbe5ec1a7a0265ed7c466363c3abeec492263170860689b46c7ff95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d08e5cfc7cbe5ec1a7a0265ed7c466363c3abeec492263170860689b46c7ff95\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:11:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:11:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmlt2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:11:16Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-jnh7m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:23Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:23 crc kubenswrapper[4988]: I1008 18:11:23.790166 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f328a20b-5283-407e-8a93-2ab2da599f18\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8622d56fb198dadee31fb0ca8fc80d56828b6710648055f264533e2d5bdb3037\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z2gh4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3cdb2a757332bdcdc2a2f3362913fe8f06003b0597de20b72627f9badd5e30af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z2gh4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:11:16Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-nm54f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:23Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:23 crc kubenswrapper[4988]: I1008 18:11:23.805099 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a35a5e989ba69c8ef36ce4cea272aadf67ce20473fc257adfcfa86a2faba6ae0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a522e937cfe03fb9f2e60cb1bea07214cf63c69bdecd5ba65a2dfdba882d28dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:23Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:23 crc kubenswrapper[4988]: I1008 18:11:23.819751 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9590ca0059f0ebbe86343cccb3475338c317382c3231e082655b61e67eece375\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:23Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:23 crc kubenswrapper[4988]: I1008 18:11:23.834550 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:23Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:23 crc kubenswrapper[4988]: I1008 18:11:23.850058 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-jnh7m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"556fb048-8afb-44a5-9ea0-2b7c89e6cf4b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmlt2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1a825886f105ccf9a7cc6f28a0cd012fb859f4990dade0e6aae932482146028\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1a825886f105ccf9a7cc6f28a0cd012fb859f4990dade0e6aae932482146028\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmlt2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ab2fe6e65418efa75bae72259157a2323bbb4b7d11d1d012053524e14135888\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0ab2fe6e65418efa75bae72259157a2323bbb4b7d11d1d012053524e14135888\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:11:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:11:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmlt2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://368a303309be112d9c0b7ccd6538fbaca7085d97a8ec8259f36092f07147e7cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://368a303309be112d9c0b7ccd6538fbaca7085d97a8ec8259f36092f07147e7cb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:11:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:11:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmlt2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a8ca69001dddef643f982a97c74c291e30cb9fac5b0e894e80066b885ea9151\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9a8ca69001dddef643f982a97c74c291e30cb9fac5b0e894e80066b885ea9151\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:11:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:11:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmlt2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://145303c46b48603ed1aae4b9bfe06332a3ffad28f206f8394574011a87b58cd3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://145303c46b48603ed1aae4b9bfe06332a3ffad28f206f8394574011a87b58cd3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:11:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:11:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmlt2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d08e5cfc7cbe5ec1a7a0265ed7c466363c3abeec492263170860689b46c7ff95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d08e5cfc7cbe5ec1a7a0265ed7c466363c3abeec492263170860689b46c7ff95\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:11:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:11:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmlt2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:11:16Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-jnh7m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:23Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:23 crc kubenswrapper[4988]: I1008 18:11:23.852080 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:11:23 crc kubenswrapper[4988]: I1008 18:11:23.852128 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:11:23 crc kubenswrapper[4988]: I1008 18:11:23.852140 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:11:23 crc kubenswrapper[4988]: I1008 18:11:23.852161 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:11:23 crc kubenswrapper[4988]: I1008 18:11:23.852175 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:11:23Z","lastTransitionTime":"2025-10-08T18:11:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:11:23 crc kubenswrapper[4988]: I1008 18:11:23.863340 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f328a20b-5283-407e-8a93-2ab2da599f18\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8622d56fb198dadee31fb0ca8fc80d56828b6710648055f264533e2d5bdb3037\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z2gh4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3cdb2a757332bdcdc2a2f3362913fe8f06003b0597de20b72627f9badd5e30af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z2gh4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:11:16Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-nm54f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:23Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:23 crc kubenswrapper[4988]: I1008 18:11:23.877353 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:23Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:23 crc kubenswrapper[4988]: I1008 18:11:23.892648 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:23Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:23 crc kubenswrapper[4988]: I1008 18:11:23.909031 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-h9jwp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4490d846-ca32-4cd5-9202-87cc6d499fb4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://00b7f6db043c28efda3a09fdc77a6ccded3eae8b6a3d2852aa500789d5f01194\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7k6rk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:11:16Z\\\"}}\" for pod \"openshift-multus\"/\"multus-h9jwp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:23Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:23 crc kubenswrapper[4988]: I1008 18:11:23.931264 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-nsc65" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f010a463-69f2-4082-acc3-161899a939c3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://48a516f6c153743bd1a21ec6945fb368a37f4d2b9d6a326b68edc096a192ed05\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dnwgr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://03406e726f817bcedb841f81959ab12808ba35a530240f9abc5fa5c608b9a6ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dnwgr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9fbbeae0409d2e47e8de9a4306f63b0d13acbfd73cd86f83303fec14352ed718\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dnwgr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://55791327a370c0009c523b86b2234e82d3794985f2eaccf1922eb90f46b2b970\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dnwgr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cee5de93bc68345f9b737857f02832f56582d58757c50a8798eec25d19b2c20d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dnwgr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d314eb248e90371fed068b1173e26d4c9f9765417a70b341e1a5b136ec17353a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dnwgr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9353d08254f16c1a3cd3b66904805f54467292dd2c052268336aaca5678aeef7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dnwgr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://74fcb812b84291763cc6e5ccb20c12aaad580ad968dda4734a0bbfe33d2770ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dnwgr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://befe684cea9bd6dfda3ec8e85bca431c26bd8f82823b0a69dc0d04319a3ca385\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://befe684cea9bd6dfda3ec8e85bca431c26bd8f82823b0a69dc0d04319a3ca385\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dnwgr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:11:16Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-nsc65\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:23Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:23 crc kubenswrapper[4988]: I1008 18:11:23.945032 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-s4f2k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ba5e11d7-7bde-4c3d-983b-884232219652\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://824bc34a2807068d3fceb10651803ae9ae5651e940f7aed4877c544d76dae4d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hn5jf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:11:15Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-s4f2k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:23Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:23 crc kubenswrapper[4988]: I1008 18:11:23.956221 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:11:23 crc kubenswrapper[4988]: I1008 18:11:23.956283 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:11:23 crc kubenswrapper[4988]: I1008 18:11:23.956302 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:11:23 crc kubenswrapper[4988]: I1008 18:11:23.956328 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:11:23 crc kubenswrapper[4988]: I1008 18:11:23.956346 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:11:23Z","lastTransitionTime":"2025-10-08T18:11:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:11:23 crc kubenswrapper[4988]: I1008 18:11:23.968038 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aef71be1-1d12-46e5-95b2-bd399040fd4d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:10:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:10:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:10:55Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:10:55Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:10:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77c68f3ecb8a0724499e495cb26ef3db5beecb140f2558d8c0b25f41928b84ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:10:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a560ee84c2405227de71fda127566213cd4f017e73d5581442ba267781b1c539\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:10:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6100accc3bd08928bc1ff88310f29fa5da37fd2d5c315c94279801c9778c4745\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:10:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b0e596d7ce4e9341c44499af79d3b3caf0e5e2c156902ea40cf2a3dca672be9e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5bd35aef24be034a0b305a40bb53ca171a4f76db3b93b4f78ae95099b0a4cde3\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-08T18:11:14Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1008 18:11:09.091034 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1008 18:11:09.091984 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3472374062/tls.crt::/tmp/serving-cert-3472374062/tls.key\\\\\\\"\\\\nI1008 18:11:14.619696 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1008 18:11:14.628837 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1008 18:11:14.628898 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1008 18:11:14.628994 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1008 18:11:14.629036 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1008 18:11:14.641318 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1008 18:11:14.641363 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1008 18:11:14.641372 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1008 18:11:14.641382 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1008 18:11:14.641415 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1008 18:11:14.641423 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1008 18:11:14.641431 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1008 18:11:14.641484 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1008 18:11:14.646451 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T18:10:58Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c79ef704fda1bbcc486f4c7afa3039a3ab38669833f0f66bafe24dd4cc679e60\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:10:58Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://28d21731af0f16d6a8ae6c6e9d0bebf0560d4a533db4c78c10bf1837890394ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://28d21731af0f16d6a8ae6c6e9d0bebf0560d4a533db4c78c10bf1837890394ef\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:10:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:10:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:10:55Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:23Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:23 crc kubenswrapper[4988]: I1008 18:11:23.980358 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7e0ecefc-359f-4785-bb79-0343a62a25da\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:10:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:10:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:10:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://30bec350fefee3d19758b3ed735ede4e9aa085edb57aa3f3ec24249a67dc7ca2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:10:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2d850fd27d4ed65891ff69b9add65fb78535357f54819c2a8d196f5a340c58e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:10:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://90b1a9e14c82bf0f75f08fdb342eed35a503020f9db641e5e9ceff8b122cb3bd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:10:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://16f23c4c7163d32dc0df8d6a7ee17e2864fb261f8a40ecf05556774b173fc868\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:10:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:10:55Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:23Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:23 crc kubenswrapper[4988]: I1008 18:11:23.995227 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e8906182d55c285b3cf095e2d9e1dac397af4618a514f11e26b44c172dd573f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:23Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:24 crc kubenswrapper[4988]: I1008 18:11:24.011019 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-dzxs2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8f307b20-5e33-4725-99cb-210e25113532\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72ee803c6ef88524e0ce0391476f8e5a7dfbc858b095a3c2ae31bad10991ce3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q65f7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:11:16Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-dzxs2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:24Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:24 crc kubenswrapper[4988]: I1008 18:11:24.060118 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:11:24 crc kubenswrapper[4988]: I1008 18:11:24.060171 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:11:24 crc kubenswrapper[4988]: I1008 18:11:24.060182 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:11:24 crc kubenswrapper[4988]: I1008 18:11:24.060204 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:11:24 crc kubenswrapper[4988]: I1008 18:11:24.060219 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:11:24Z","lastTransitionTime":"2025-10-08T18:11:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:11:24 crc kubenswrapper[4988]: I1008 18:11:24.162800 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:11:24 crc kubenswrapper[4988]: I1008 18:11:24.162857 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:11:24 crc kubenswrapper[4988]: I1008 18:11:24.162874 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:11:24 crc kubenswrapper[4988]: I1008 18:11:24.162897 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:11:24 crc kubenswrapper[4988]: I1008 18:11:24.162915 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:11:24Z","lastTransitionTime":"2025-10-08T18:11:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:11:24 crc kubenswrapper[4988]: I1008 18:11:24.265536 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:11:24 crc kubenswrapper[4988]: I1008 18:11:24.265586 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:11:24 crc kubenswrapper[4988]: I1008 18:11:24.265598 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:11:24 crc kubenswrapper[4988]: I1008 18:11:24.265618 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:11:24 crc kubenswrapper[4988]: I1008 18:11:24.265669 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:11:24Z","lastTransitionTime":"2025-10-08T18:11:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:11:24 crc kubenswrapper[4988]: I1008 18:11:24.368594 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:11:24 crc kubenswrapper[4988]: I1008 18:11:24.368650 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:11:24 crc kubenswrapper[4988]: I1008 18:11:24.368665 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:11:24 crc kubenswrapper[4988]: I1008 18:11:24.368688 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:11:24 crc kubenswrapper[4988]: I1008 18:11:24.368704 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:11:24Z","lastTransitionTime":"2025-10-08T18:11:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:11:24 crc kubenswrapper[4988]: I1008 18:11:24.471767 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:11:24 crc kubenswrapper[4988]: I1008 18:11:24.471827 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:11:24 crc kubenswrapper[4988]: I1008 18:11:24.471840 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:11:24 crc kubenswrapper[4988]: I1008 18:11:24.471863 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:11:24 crc kubenswrapper[4988]: I1008 18:11:24.471881 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:11:24Z","lastTransitionTime":"2025-10-08T18:11:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:11:24 crc kubenswrapper[4988]: I1008 18:11:24.547448 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-jnh7m" event={"ID":"556fb048-8afb-44a5-9ea0-2b7c89e6cf4b","Type":"ContainerStarted","Data":"bfa2f3382b23c3f788a1b31e80e8a90b3b6868870dccc46d86c3fc11ad027f52"} Oct 08 18:11:24 crc kubenswrapper[4988]: I1008 18:11:24.547535 4988 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 08 18:11:24 crc kubenswrapper[4988]: I1008 18:11:24.569472 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-s4f2k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ba5e11d7-7bde-4c3d-983b-884232219652\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://824bc34a2807068d3fceb10651803ae9ae5651e940f7aed4877c544d76dae4d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hn5jf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:11:15Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-s4f2k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:24Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:24 crc kubenswrapper[4988]: I1008 18:11:24.574763 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:11:24 crc kubenswrapper[4988]: I1008 18:11:24.574818 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:11:24 crc kubenswrapper[4988]: I1008 18:11:24.574833 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:11:24 crc kubenswrapper[4988]: I1008 18:11:24.574856 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:11:24 crc kubenswrapper[4988]: I1008 18:11:24.574869 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:11:24Z","lastTransitionTime":"2025-10-08T18:11:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:11:24 crc kubenswrapper[4988]: I1008 18:11:24.591862 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aef71be1-1d12-46e5-95b2-bd399040fd4d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:10:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:10:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:10:55Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:10:55Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:10:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77c68f3ecb8a0724499e495cb26ef3db5beecb140f2558d8c0b25f41928b84ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:10:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a560ee84c2405227de71fda127566213cd4f017e73d5581442ba267781b1c539\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:10:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6100accc3bd08928bc1ff88310f29fa5da37fd2d5c315c94279801c9778c4745\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:10:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b0e596d7ce4e9341c44499af79d3b3caf0e5e2c156902ea40cf2a3dca672be9e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5bd35aef24be034a0b305a40bb53ca171a4f76db3b93b4f78ae95099b0a4cde3\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-08T18:11:14Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1008 18:11:09.091034 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1008 18:11:09.091984 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3472374062/tls.crt::/tmp/serving-cert-3472374062/tls.key\\\\\\\"\\\\nI1008 18:11:14.619696 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1008 18:11:14.628837 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1008 18:11:14.628898 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1008 18:11:14.628994 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1008 18:11:14.629036 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1008 18:11:14.641318 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1008 18:11:14.641363 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1008 18:11:14.641372 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1008 18:11:14.641382 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1008 18:11:14.641415 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1008 18:11:14.641423 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1008 18:11:14.641431 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1008 18:11:14.641484 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1008 18:11:14.646451 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T18:10:58Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c79ef704fda1bbcc486f4c7afa3039a3ab38669833f0f66bafe24dd4cc679e60\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:10:58Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://28d21731af0f16d6a8ae6c6e9d0bebf0560d4a533db4c78c10bf1837890394ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://28d21731af0f16d6a8ae6c6e9d0bebf0560d4a533db4c78c10bf1837890394ef\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:10:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:10:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:10:55Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:24Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:24 crc kubenswrapper[4988]: I1008 18:11:24.607241 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7e0ecefc-359f-4785-bb79-0343a62a25da\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:10:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:10:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:10:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://30bec350fefee3d19758b3ed735ede4e9aa085edb57aa3f3ec24249a67dc7ca2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:10:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2d850fd27d4ed65891ff69b9add65fb78535357f54819c2a8d196f5a340c58e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:10:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://90b1a9e14c82bf0f75f08fdb342eed35a503020f9db641e5e9ceff8b122cb3bd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:10:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://16f23c4c7163d32dc0df8d6a7ee17e2864fb261f8a40ecf05556774b173fc868\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:10:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:10:55Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:24Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:24 crc kubenswrapper[4988]: I1008 18:11:24.647433 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e8906182d55c285b3cf095e2d9e1dac397af4618a514f11e26b44c172dd573f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:24Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:24 crc kubenswrapper[4988]: I1008 18:11:24.663154 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-dzxs2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8f307b20-5e33-4725-99cb-210e25113532\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72ee803c6ef88524e0ce0391476f8e5a7dfbc858b095a3c2ae31bad10991ce3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q65f7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:11:16Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-dzxs2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:24Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:24 crc kubenswrapper[4988]: I1008 18:11:24.677339 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a35a5e989ba69c8ef36ce4cea272aadf67ce20473fc257adfcfa86a2faba6ae0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a522e937cfe03fb9f2e60cb1bea07214cf63c69bdecd5ba65a2dfdba882d28dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:24Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:24 crc kubenswrapper[4988]: I1008 18:11:24.678066 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:11:24 crc kubenswrapper[4988]: I1008 18:11:24.678117 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:11:24 crc kubenswrapper[4988]: I1008 18:11:24.678133 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:11:24 crc kubenswrapper[4988]: I1008 18:11:24.678159 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:11:24 crc kubenswrapper[4988]: I1008 18:11:24.678178 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:11:24Z","lastTransitionTime":"2025-10-08T18:11:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:11:24 crc kubenswrapper[4988]: I1008 18:11:24.696945 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9590ca0059f0ebbe86343cccb3475338c317382c3231e082655b61e67eece375\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:24Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:24 crc kubenswrapper[4988]: I1008 18:11:24.713900 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:24Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:24 crc kubenswrapper[4988]: I1008 18:11:24.731443 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-jnh7m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"556fb048-8afb-44a5-9ea0-2b7c89e6cf4b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bfa2f3382b23c3f788a1b31e80e8a90b3b6868870dccc46d86c3fc11ad027f52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmlt2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1a825886f105ccf9a7cc6f28a0cd012fb859f4990dade0e6aae932482146028\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1a825886f105ccf9a7cc6f28a0cd012fb859f4990dade0e6aae932482146028\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmlt2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ab2fe6e65418efa75bae72259157a2323bbb4b7d11d1d012053524e14135888\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0ab2fe6e65418efa75bae72259157a2323bbb4b7d11d1d012053524e14135888\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:11:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:11:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmlt2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://368a303309be112d9c0b7ccd6538fbaca7085d97a8ec8259f36092f07147e7cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://368a303309be112d9c0b7ccd6538fbaca7085d97a8ec8259f36092f07147e7cb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:11:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:11:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmlt2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a8ca69001dddef643f982a97c74c291e30cb9fac5b0e894e80066b885ea9151\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9a8ca69001dddef643f982a97c74c291e30cb9fac5b0e894e80066b885ea9151\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:11:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:11:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmlt2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://145303c46b48603ed1aae4b9bfe06332a3ffad28f206f8394574011a87b58cd3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://145303c46b48603ed1aae4b9bfe06332a3ffad28f206f8394574011a87b58cd3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:11:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:11:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmlt2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d08e5cfc7cbe5ec1a7a0265ed7c466363c3abeec492263170860689b46c7ff95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d08e5cfc7cbe5ec1a7a0265ed7c466363c3abeec492263170860689b46c7ff95\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:11:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:11:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmlt2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:11:16Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-jnh7m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:24Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:24 crc kubenswrapper[4988]: I1008 18:11:24.744634 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f328a20b-5283-407e-8a93-2ab2da599f18\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8622d56fb198dadee31fb0ca8fc80d56828b6710648055f264533e2d5bdb3037\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z2gh4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3cdb2a757332bdcdc2a2f3362913fe8f06003b0597de20b72627f9badd5e30af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z2gh4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:11:16Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-nm54f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:24Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:24 crc kubenswrapper[4988]: I1008 18:11:24.762861 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:24Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:24 crc kubenswrapper[4988]: I1008 18:11:24.776663 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:24Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:24 crc kubenswrapper[4988]: I1008 18:11:24.781502 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:11:24 crc kubenswrapper[4988]: I1008 18:11:24.781546 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:11:24 crc kubenswrapper[4988]: I1008 18:11:24.781560 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:11:24 crc kubenswrapper[4988]: I1008 18:11:24.781582 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:11:24 crc kubenswrapper[4988]: I1008 18:11:24.781595 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:11:24Z","lastTransitionTime":"2025-10-08T18:11:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:11:24 crc kubenswrapper[4988]: I1008 18:11:24.791819 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-nsc65" Oct 08 18:11:24 crc kubenswrapper[4988]: I1008 18:11:24.796671 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-h9jwp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4490d846-ca32-4cd5-9202-87cc6d499fb4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://00b7f6db043c28efda3a09fdc77a6ccded3eae8b6a3d2852aa500789d5f01194\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7k6rk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:11:16Z\\\"}}\" for pod \"openshift-multus\"/\"multus-h9jwp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:24Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:24 crc kubenswrapper[4988]: I1008 18:11:24.817054 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-nsc65" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f010a463-69f2-4082-acc3-161899a939c3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://48a516f6c153743bd1a21ec6945fb368a37f4d2b9d6a326b68edc096a192ed05\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dnwgr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://03406e726f817bcedb841f81959ab12808ba35a530240f9abc5fa5c608b9a6ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dnwgr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9fbbeae0409d2e47e8de9a4306f63b0d13acbfd73cd86f83303fec14352ed718\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dnwgr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://55791327a370c0009c523b86b2234e82d3794985f2eaccf1922eb90f46b2b970\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dnwgr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cee5de93bc68345f9b737857f02832f56582d58757c50a8798eec25d19b2c20d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dnwgr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d314eb248e90371fed068b1173e26d4c9f9765417a70b341e1a5b136ec17353a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dnwgr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9353d08254f16c1a3cd3b66904805f54467292dd2c052268336aaca5678aeef7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dnwgr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://74fcb812b84291763cc6e5ccb20c12aaad580ad968dda4734a0bbfe33d2770ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dnwgr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://befe684cea9bd6dfda3ec8e85bca431c26bd8f82823b0a69dc0d04319a3ca385\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://befe684cea9bd6dfda3ec8e85bca431c26bd8f82823b0a69dc0d04319a3ca385\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dnwgr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:11:16Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-nsc65\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:24Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:24 crc kubenswrapper[4988]: I1008 18:11:24.884698 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:11:24 crc kubenswrapper[4988]: I1008 18:11:24.884746 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:11:24 crc kubenswrapper[4988]: I1008 18:11:24.884759 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:11:24 crc kubenswrapper[4988]: I1008 18:11:24.884776 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:11:24 crc kubenswrapper[4988]: I1008 18:11:24.884790 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:11:24Z","lastTransitionTime":"2025-10-08T18:11:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:11:24 crc kubenswrapper[4988]: I1008 18:11:24.987799 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:11:24 crc kubenswrapper[4988]: I1008 18:11:24.987847 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:11:24 crc kubenswrapper[4988]: I1008 18:11:24.987857 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:11:24 crc kubenswrapper[4988]: I1008 18:11:24.987873 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:11:24 crc kubenswrapper[4988]: I1008 18:11:24.987886 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:11:24Z","lastTransitionTime":"2025-10-08T18:11:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:11:25 crc kubenswrapper[4988]: I1008 18:11:25.091479 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:11:25 crc kubenswrapper[4988]: I1008 18:11:25.091551 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:11:25 crc kubenswrapper[4988]: I1008 18:11:25.091573 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:11:25 crc kubenswrapper[4988]: I1008 18:11:25.091606 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:11:25 crc kubenswrapper[4988]: I1008 18:11:25.091624 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:11:25Z","lastTransitionTime":"2025-10-08T18:11:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:11:25 crc kubenswrapper[4988]: I1008 18:11:25.194152 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:11:25 crc kubenswrapper[4988]: I1008 18:11:25.194194 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:11:25 crc kubenswrapper[4988]: I1008 18:11:25.194204 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:11:25 crc kubenswrapper[4988]: I1008 18:11:25.194221 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:11:25 crc kubenswrapper[4988]: I1008 18:11:25.194233 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:11:25Z","lastTransitionTime":"2025-10-08T18:11:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:11:25 crc kubenswrapper[4988]: I1008 18:11:25.237519 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 18:11:25 crc kubenswrapper[4988]: I1008 18:11:25.237590 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 18:11:25 crc kubenswrapper[4988]: E1008 18:11:25.237743 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 08 18:11:25 crc kubenswrapper[4988]: I1008 18:11:25.237789 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 18:11:25 crc kubenswrapper[4988]: E1008 18:11:25.237992 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 08 18:11:25 crc kubenswrapper[4988]: E1008 18:11:25.238182 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 08 18:11:25 crc kubenswrapper[4988]: I1008 18:11:25.256118 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:25Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:25 crc kubenswrapper[4988]: I1008 18:11:25.276700 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:25Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:25 crc kubenswrapper[4988]: I1008 18:11:25.297188 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:11:25 crc kubenswrapper[4988]: I1008 18:11:25.297262 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:11:25 crc kubenswrapper[4988]: I1008 18:11:25.297287 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:11:25 crc kubenswrapper[4988]: I1008 18:11:25.297321 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:11:25 crc kubenswrapper[4988]: I1008 18:11:25.297349 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:11:25Z","lastTransitionTime":"2025-10-08T18:11:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:11:25 crc kubenswrapper[4988]: I1008 18:11:25.299809 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-h9jwp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4490d846-ca32-4cd5-9202-87cc6d499fb4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://00b7f6db043c28efda3a09fdc77a6ccded3eae8b6a3d2852aa500789d5f01194\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7k6rk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:11:16Z\\\"}}\" for pod \"openshift-multus\"/\"multus-h9jwp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:25Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:25 crc kubenswrapper[4988]: I1008 18:11:25.323467 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-nsc65" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f010a463-69f2-4082-acc3-161899a939c3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://48a516f6c153743bd1a21ec6945fb368a37f4d2b9d6a326b68edc096a192ed05\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dnwgr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://03406e726f817bcedb841f81959ab12808ba35a530240f9abc5fa5c608b9a6ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dnwgr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9fbbeae0409d2e47e8de9a4306f63b0d13acbfd73cd86f83303fec14352ed718\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dnwgr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://55791327a370c0009c523b86b2234e82d3794985f2eaccf1922eb90f46b2b970\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dnwgr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cee5de93bc68345f9b737857f02832f56582d58757c50a8798eec25d19b2c20d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dnwgr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d314eb248e90371fed068b1173e26d4c9f9765417a70b341e1a5b136ec17353a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dnwgr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9353d08254f16c1a3cd3b66904805f54467292dd2c052268336aaca5678aeef7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dnwgr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://74fcb812b84291763cc6e5ccb20c12aaad580ad968dda4734a0bbfe33d2770ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dnwgr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://befe684cea9bd6dfda3ec8e85bca431c26bd8f82823b0a69dc0d04319a3ca385\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://befe684cea9bd6dfda3ec8e85bca431c26bd8f82823b0a69dc0d04319a3ca385\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dnwgr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:11:16Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-nsc65\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:25Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:25 crc kubenswrapper[4988]: I1008 18:11:25.336046 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-s4f2k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ba5e11d7-7bde-4c3d-983b-884232219652\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://824bc34a2807068d3fceb10651803ae9ae5651e940f7aed4877c544d76dae4d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hn5jf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:11:15Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-s4f2k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:25Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:25 crc kubenswrapper[4988]: I1008 18:11:25.349921 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7e0ecefc-359f-4785-bb79-0343a62a25da\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:10:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:10:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:10:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://30bec350fefee3d19758b3ed735ede4e9aa085edb57aa3f3ec24249a67dc7ca2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:10:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2d850fd27d4ed65891ff69b9add65fb78535357f54819c2a8d196f5a340c58e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:10:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://90b1a9e14c82bf0f75f08fdb342eed35a503020f9db641e5e9ceff8b122cb3bd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:10:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://16f23c4c7163d32dc0df8d6a7ee17e2864fb261f8a40ecf05556774b173fc868\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:10:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:10:55Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:25Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:25 crc kubenswrapper[4988]: I1008 18:11:25.402607 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:11:25 crc kubenswrapper[4988]: I1008 18:11:25.402653 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:11:25 crc kubenswrapper[4988]: I1008 18:11:25.402663 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:11:25 crc kubenswrapper[4988]: I1008 18:11:25.402679 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:11:25 crc kubenswrapper[4988]: I1008 18:11:25.402693 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:11:25Z","lastTransitionTime":"2025-10-08T18:11:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:11:25 crc kubenswrapper[4988]: I1008 18:11:25.417742 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e8906182d55c285b3cf095e2d9e1dac397af4618a514f11e26b44c172dd573f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:25Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:25 crc kubenswrapper[4988]: I1008 18:11:25.445616 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-dzxs2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8f307b20-5e33-4725-99cb-210e25113532\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72ee803c6ef88524e0ce0391476f8e5a7dfbc858b095a3c2ae31bad10991ce3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q65f7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:11:16Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-dzxs2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:25Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:25 crc kubenswrapper[4988]: I1008 18:11:25.471023 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aef71be1-1d12-46e5-95b2-bd399040fd4d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:10:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:10:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:10:55Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:10:55Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:10:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77c68f3ecb8a0724499e495cb26ef3db5beecb140f2558d8c0b25f41928b84ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:10:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a560ee84c2405227de71fda127566213cd4f017e73d5581442ba267781b1c539\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:10:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6100accc3bd08928bc1ff88310f29fa5da37fd2d5c315c94279801c9778c4745\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:10:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b0e596d7ce4e9341c44499af79d3b3caf0e5e2c156902ea40cf2a3dca672be9e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5bd35aef24be034a0b305a40bb53ca171a4f76db3b93b4f78ae95099b0a4cde3\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-08T18:11:14Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1008 18:11:09.091034 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1008 18:11:09.091984 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3472374062/tls.crt::/tmp/serving-cert-3472374062/tls.key\\\\\\\"\\\\nI1008 18:11:14.619696 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1008 18:11:14.628837 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1008 18:11:14.628898 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1008 18:11:14.628994 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1008 18:11:14.629036 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1008 18:11:14.641318 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1008 18:11:14.641363 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1008 18:11:14.641372 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1008 18:11:14.641382 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1008 18:11:14.641415 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1008 18:11:14.641423 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1008 18:11:14.641431 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1008 18:11:14.641484 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1008 18:11:14.646451 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T18:10:58Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c79ef704fda1bbcc486f4c7afa3039a3ab38669833f0f66bafe24dd4cc679e60\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:10:58Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://28d21731af0f16d6a8ae6c6e9d0bebf0560d4a533db4c78c10bf1837890394ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://28d21731af0f16d6a8ae6c6e9d0bebf0560d4a533db4c78c10bf1837890394ef\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:10:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:10:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:10:55Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:25Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:25 crc kubenswrapper[4988]: I1008 18:11:25.491589 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a35a5e989ba69c8ef36ce4cea272aadf67ce20473fc257adfcfa86a2faba6ae0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a522e937cfe03fb9f2e60cb1bea07214cf63c69bdecd5ba65a2dfdba882d28dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:25Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:25 crc kubenswrapper[4988]: I1008 18:11:25.505927 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9590ca0059f0ebbe86343cccb3475338c317382c3231e082655b61e67eece375\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:25Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:25 crc kubenswrapper[4988]: I1008 18:11:25.506023 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:11:25 crc kubenswrapper[4988]: I1008 18:11:25.506061 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:11:25 crc kubenswrapper[4988]: I1008 18:11:25.506072 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:11:25 crc kubenswrapper[4988]: I1008 18:11:25.506087 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:11:25 crc kubenswrapper[4988]: I1008 18:11:25.506096 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:11:25Z","lastTransitionTime":"2025-10-08T18:11:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:11:25 crc kubenswrapper[4988]: I1008 18:11:25.520548 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:25Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:25 crc kubenswrapper[4988]: I1008 18:11:25.540136 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-jnh7m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"556fb048-8afb-44a5-9ea0-2b7c89e6cf4b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bfa2f3382b23c3f788a1b31e80e8a90b3b6868870dccc46d86c3fc11ad027f52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmlt2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1a825886f105ccf9a7cc6f28a0cd012fb859f4990dade0e6aae932482146028\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1a825886f105ccf9a7cc6f28a0cd012fb859f4990dade0e6aae932482146028\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmlt2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ab2fe6e65418efa75bae72259157a2323bbb4b7d11d1d012053524e14135888\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0ab2fe6e65418efa75bae72259157a2323bbb4b7d11d1d012053524e14135888\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:11:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:11:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmlt2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://368a303309be112d9c0b7ccd6538fbaca7085d97a8ec8259f36092f07147e7cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://368a303309be112d9c0b7ccd6538fbaca7085d97a8ec8259f36092f07147e7cb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:11:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:11:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmlt2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a8ca69001dddef643f982a97c74c291e30cb9fac5b0e894e80066b885ea9151\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9a8ca69001dddef643f982a97c74c291e30cb9fac5b0e894e80066b885ea9151\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:11:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:11:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmlt2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://145303c46b48603ed1aae4b9bfe06332a3ffad28f206f8394574011a87b58cd3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://145303c46b48603ed1aae4b9bfe06332a3ffad28f206f8394574011a87b58cd3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:11:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:11:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmlt2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d08e5cfc7cbe5ec1a7a0265ed7c466363c3abeec492263170860689b46c7ff95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d08e5cfc7cbe5ec1a7a0265ed7c466363c3abeec492263170860689b46c7ff95\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:11:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:11:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmlt2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:11:16Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-jnh7m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:25Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:25 crc kubenswrapper[4988]: I1008 18:11:25.559795 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f328a20b-5283-407e-8a93-2ab2da599f18\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8622d56fb198dadee31fb0ca8fc80d56828b6710648055f264533e2d5bdb3037\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z2gh4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3cdb2a757332bdcdc2a2f3362913fe8f06003b0597de20b72627f9badd5e30af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z2gh4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:11:16Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-nm54f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:25Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:25 crc kubenswrapper[4988]: I1008 18:11:25.608706 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:11:25 crc kubenswrapper[4988]: I1008 18:11:25.608750 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:11:25 crc kubenswrapper[4988]: I1008 18:11:25.608765 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:11:25 crc kubenswrapper[4988]: I1008 18:11:25.608782 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:11:25 crc kubenswrapper[4988]: I1008 18:11:25.608792 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:11:25Z","lastTransitionTime":"2025-10-08T18:11:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:11:25 crc kubenswrapper[4988]: I1008 18:11:25.724618 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:11:25 crc kubenswrapper[4988]: I1008 18:11:25.724674 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:11:25 crc kubenswrapper[4988]: I1008 18:11:25.724693 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:11:25 crc kubenswrapper[4988]: I1008 18:11:25.724720 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:11:25 crc kubenswrapper[4988]: I1008 18:11:25.724740 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:11:25Z","lastTransitionTime":"2025-10-08T18:11:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:11:25 crc kubenswrapper[4988]: I1008 18:11:25.828559 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:11:25 crc kubenswrapper[4988]: I1008 18:11:25.828918 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:11:25 crc kubenswrapper[4988]: I1008 18:11:25.828927 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:11:25 crc kubenswrapper[4988]: I1008 18:11:25.828944 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:11:25 crc kubenswrapper[4988]: I1008 18:11:25.828954 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:11:25Z","lastTransitionTime":"2025-10-08T18:11:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:11:25 crc kubenswrapper[4988]: I1008 18:11:25.932460 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:11:25 crc kubenswrapper[4988]: I1008 18:11:25.932544 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:11:25 crc kubenswrapper[4988]: I1008 18:11:25.932571 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:11:25 crc kubenswrapper[4988]: I1008 18:11:25.932612 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:11:25 crc kubenswrapper[4988]: I1008 18:11:25.932638 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:11:25Z","lastTransitionTime":"2025-10-08T18:11:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:11:26 crc kubenswrapper[4988]: I1008 18:11:26.036646 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:11:26 crc kubenswrapper[4988]: I1008 18:11:26.036747 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:11:26 crc kubenswrapper[4988]: I1008 18:11:26.036791 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:11:26 crc kubenswrapper[4988]: I1008 18:11:26.036826 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:11:26 crc kubenswrapper[4988]: I1008 18:11:26.036850 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:11:26Z","lastTransitionTime":"2025-10-08T18:11:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:11:26 crc kubenswrapper[4988]: I1008 18:11:26.139645 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:11:26 crc kubenswrapper[4988]: I1008 18:11:26.139693 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:11:26 crc kubenswrapper[4988]: I1008 18:11:26.139705 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:11:26 crc kubenswrapper[4988]: I1008 18:11:26.139724 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:11:26 crc kubenswrapper[4988]: I1008 18:11:26.139737 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:11:26Z","lastTransitionTime":"2025-10-08T18:11:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:11:26 crc kubenswrapper[4988]: I1008 18:11:26.243093 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:11:26 crc kubenswrapper[4988]: I1008 18:11:26.243177 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:11:26 crc kubenswrapper[4988]: I1008 18:11:26.243201 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:11:26 crc kubenswrapper[4988]: I1008 18:11:26.243234 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:11:26 crc kubenswrapper[4988]: I1008 18:11:26.243258 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:11:26Z","lastTransitionTime":"2025-10-08T18:11:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:11:26 crc kubenswrapper[4988]: I1008 18:11:26.346099 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:11:26 crc kubenswrapper[4988]: I1008 18:11:26.346152 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:11:26 crc kubenswrapper[4988]: I1008 18:11:26.346166 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:11:26 crc kubenswrapper[4988]: I1008 18:11:26.346186 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:11:26 crc kubenswrapper[4988]: I1008 18:11:26.346200 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:11:26Z","lastTransitionTime":"2025-10-08T18:11:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:11:26 crc kubenswrapper[4988]: I1008 18:11:26.449334 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:11:26 crc kubenswrapper[4988]: I1008 18:11:26.449457 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:11:26 crc kubenswrapper[4988]: I1008 18:11:26.449495 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:11:26 crc kubenswrapper[4988]: I1008 18:11:26.449530 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:11:26 crc kubenswrapper[4988]: I1008 18:11:26.449554 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:11:26Z","lastTransitionTime":"2025-10-08T18:11:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:11:26 crc kubenswrapper[4988]: I1008 18:11:26.553510 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:11:26 crc kubenswrapper[4988]: I1008 18:11:26.553582 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:11:26 crc kubenswrapper[4988]: I1008 18:11:26.553599 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:11:26 crc kubenswrapper[4988]: I1008 18:11:26.553625 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:11:26 crc kubenswrapper[4988]: I1008 18:11:26.553643 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:11:26Z","lastTransitionTime":"2025-10-08T18:11:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:11:26 crc kubenswrapper[4988]: I1008 18:11:26.557337 4988 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-nsc65_f010a463-69f2-4082-acc3-161899a939c3/ovnkube-controller/0.log" Oct 08 18:11:26 crc kubenswrapper[4988]: I1008 18:11:26.562039 4988 generic.go:334] "Generic (PLEG): container finished" podID="f010a463-69f2-4082-acc3-161899a939c3" containerID="9353d08254f16c1a3cd3b66904805f54467292dd2c052268336aaca5678aeef7" exitCode=1 Oct 08 18:11:26 crc kubenswrapper[4988]: I1008 18:11:26.562117 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-nsc65" event={"ID":"f010a463-69f2-4082-acc3-161899a939c3","Type":"ContainerDied","Data":"9353d08254f16c1a3cd3b66904805f54467292dd2c052268336aaca5678aeef7"} Oct 08 18:11:26 crc kubenswrapper[4988]: I1008 18:11:26.563270 4988 scope.go:117] "RemoveContainer" containerID="9353d08254f16c1a3cd3b66904805f54467292dd2c052268336aaca5678aeef7" Oct 08 18:11:26 crc kubenswrapper[4988]: I1008 18:11:26.591225 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:26Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:26 crc kubenswrapper[4988]: I1008 18:11:26.615439 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:26Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:26 crc kubenswrapper[4988]: I1008 18:11:26.637701 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-h9jwp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4490d846-ca32-4cd5-9202-87cc6d499fb4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://00b7f6db043c28efda3a09fdc77a6ccded3eae8b6a3d2852aa500789d5f01194\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7k6rk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:11:16Z\\\"}}\" for pod \"openshift-multus\"/\"multus-h9jwp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:26Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:26 crc kubenswrapper[4988]: I1008 18:11:26.658481 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:11:26 crc kubenswrapper[4988]: I1008 18:11:26.658535 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:11:26 crc kubenswrapper[4988]: I1008 18:11:26.658549 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:11:26 crc kubenswrapper[4988]: I1008 18:11:26.658573 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:11:26 crc kubenswrapper[4988]: I1008 18:11:26.658592 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:11:26Z","lastTransitionTime":"2025-10-08T18:11:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:11:26 crc kubenswrapper[4988]: I1008 18:11:26.674294 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-nsc65" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f010a463-69f2-4082-acc3-161899a939c3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://48a516f6c153743bd1a21ec6945fb368a37f4d2b9d6a326b68edc096a192ed05\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dnwgr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://03406e726f817bcedb841f81959ab12808ba35a530240f9abc5fa5c608b9a6ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dnwgr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9fbbeae0409d2e47e8de9a4306f63b0d13acbfd73cd86f83303fec14352ed718\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dnwgr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://55791327a370c0009c523b86b2234e82d3794985f2eaccf1922eb90f46b2b970\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dnwgr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cee5de93bc68345f9b737857f02832f56582d58757c50a8798eec25d19b2c20d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dnwgr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d314eb248e90371fed068b1173e26d4c9f9765417a70b341e1a5b136ec17353a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dnwgr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9353d08254f16c1a3cd3b66904805f54467292dd2c052268336aaca5678aeef7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9353d08254f16c1a3cd3b66904805f54467292dd2c052268336aaca5678aeef7\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-08T18:11:26Z\\\",\\\"message\\\":\\\"6248 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1008 18:11:26.234639 6248 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1008 18:11:26.234656 6248 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1008 18:11:26.234670 6248 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1008 18:11:26.234760 6248 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1008 18:11:26.235089 6248 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1008 18:11:26.235124 6248 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1008 18:11:26.235165 6248 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1008 18:11:26.235169 6248 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1008 18:11:26.235180 6248 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1008 18:11:26.235210 6248 handler.go:208] Removed *v1.Node event handler 7\\\\nI1008 18:11:26.235213 6248 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1008 18:11:26.235245 6248 handler.go:208] Removed *v1.Node event handler 2\\\\nI1008 18:11:26.235284 6248 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1008 18:11:26.235326 6248 factory.go:656] Stopping \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T18:11:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dnwgr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://74fcb812b84291763cc6e5ccb20c12aaad580ad968dda4734a0bbfe33d2770ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dnwgr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://befe684cea9bd6dfda3ec8e85bca431c26bd8f82823b0a69dc0d04319a3ca385\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://befe684cea9bd6dfda3ec8e85bca431c26bd8f82823b0a69dc0d04319a3ca385\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dnwgr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:11:16Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-nsc65\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:26Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:26 crc kubenswrapper[4988]: I1008 18:11:26.691922 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-s4f2k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ba5e11d7-7bde-4c3d-983b-884232219652\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://824bc34a2807068d3fceb10651803ae9ae5651e940f7aed4877c544d76dae4d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hn5jf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:11:15Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-s4f2k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:26Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:26 crc kubenswrapper[4988]: I1008 18:11:26.717526 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7e0ecefc-359f-4785-bb79-0343a62a25da\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:10:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:10:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:10:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://30bec350fefee3d19758b3ed735ede4e9aa085edb57aa3f3ec24249a67dc7ca2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:10:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2d850fd27d4ed65891ff69b9add65fb78535357f54819c2a8d196f5a340c58e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:10:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://90b1a9e14c82bf0f75f08fdb342eed35a503020f9db641e5e9ceff8b122cb3bd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:10:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://16f23c4c7163d32dc0df8d6a7ee17e2864fb261f8a40ecf05556774b173fc868\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:10:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:10:55Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:26Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:26 crc kubenswrapper[4988]: I1008 18:11:26.740095 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e8906182d55c285b3cf095e2d9e1dac397af4618a514f11e26b44c172dd573f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:26Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:26 crc kubenswrapper[4988]: I1008 18:11:26.753444 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-dzxs2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8f307b20-5e33-4725-99cb-210e25113532\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72ee803c6ef88524e0ce0391476f8e5a7dfbc858b095a3c2ae31bad10991ce3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q65f7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:11:16Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-dzxs2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:26Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:26 crc kubenswrapper[4988]: I1008 18:11:26.761237 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:11:26 crc kubenswrapper[4988]: I1008 18:11:26.761280 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:11:26 crc kubenswrapper[4988]: I1008 18:11:26.761296 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:11:26 crc kubenswrapper[4988]: I1008 18:11:26.761323 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:11:26 crc kubenswrapper[4988]: I1008 18:11:26.761346 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:11:26Z","lastTransitionTime":"2025-10-08T18:11:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:11:26 crc kubenswrapper[4988]: I1008 18:11:26.766144 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aef71be1-1d12-46e5-95b2-bd399040fd4d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:10:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:10:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:10:55Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:10:55Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:10:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77c68f3ecb8a0724499e495cb26ef3db5beecb140f2558d8c0b25f41928b84ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:10:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a560ee84c2405227de71fda127566213cd4f017e73d5581442ba267781b1c539\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:10:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6100accc3bd08928bc1ff88310f29fa5da37fd2d5c315c94279801c9778c4745\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:10:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b0e596d7ce4e9341c44499af79d3b3caf0e5e2c156902ea40cf2a3dca672be9e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5bd35aef24be034a0b305a40bb53ca171a4f76db3b93b4f78ae95099b0a4cde3\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-08T18:11:14Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1008 18:11:09.091034 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1008 18:11:09.091984 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3472374062/tls.crt::/tmp/serving-cert-3472374062/tls.key\\\\\\\"\\\\nI1008 18:11:14.619696 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1008 18:11:14.628837 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1008 18:11:14.628898 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1008 18:11:14.628994 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1008 18:11:14.629036 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1008 18:11:14.641318 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1008 18:11:14.641363 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1008 18:11:14.641372 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1008 18:11:14.641382 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1008 18:11:14.641415 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1008 18:11:14.641423 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1008 18:11:14.641431 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1008 18:11:14.641484 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1008 18:11:14.646451 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T18:10:58Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c79ef704fda1bbcc486f4c7afa3039a3ab38669833f0f66bafe24dd4cc679e60\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:10:58Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://28d21731af0f16d6a8ae6c6e9d0bebf0560d4a533db4c78c10bf1837890394ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://28d21731af0f16d6a8ae6c6e9d0bebf0560d4a533db4c78c10bf1837890394ef\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:10:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:10:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:10:55Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:26Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:26 crc kubenswrapper[4988]: I1008 18:11:26.781758 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a35a5e989ba69c8ef36ce4cea272aadf67ce20473fc257adfcfa86a2faba6ae0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a522e937cfe03fb9f2e60cb1bea07214cf63c69bdecd5ba65a2dfdba882d28dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:26Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:26 crc kubenswrapper[4988]: I1008 18:11:26.792974 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9590ca0059f0ebbe86343cccb3475338c317382c3231e082655b61e67eece375\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:26Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:26 crc kubenswrapper[4988]: I1008 18:11:26.807103 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:26Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:26 crc kubenswrapper[4988]: I1008 18:11:26.820862 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-jnh7m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"556fb048-8afb-44a5-9ea0-2b7c89e6cf4b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bfa2f3382b23c3f788a1b31e80e8a90b3b6868870dccc46d86c3fc11ad027f52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmlt2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1a825886f105ccf9a7cc6f28a0cd012fb859f4990dade0e6aae932482146028\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1a825886f105ccf9a7cc6f28a0cd012fb859f4990dade0e6aae932482146028\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmlt2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ab2fe6e65418efa75bae72259157a2323bbb4b7d11d1d012053524e14135888\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0ab2fe6e65418efa75bae72259157a2323bbb4b7d11d1d012053524e14135888\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:11:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:11:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmlt2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://368a303309be112d9c0b7ccd6538fbaca7085d97a8ec8259f36092f07147e7cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://368a303309be112d9c0b7ccd6538fbaca7085d97a8ec8259f36092f07147e7cb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:11:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:11:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmlt2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a8ca69001dddef643f982a97c74c291e30cb9fac5b0e894e80066b885ea9151\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9a8ca69001dddef643f982a97c74c291e30cb9fac5b0e894e80066b885ea9151\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:11:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:11:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmlt2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://145303c46b48603ed1aae4b9bfe06332a3ffad28f206f8394574011a87b58cd3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://145303c46b48603ed1aae4b9bfe06332a3ffad28f206f8394574011a87b58cd3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:11:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:11:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmlt2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d08e5cfc7cbe5ec1a7a0265ed7c466363c3abeec492263170860689b46c7ff95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d08e5cfc7cbe5ec1a7a0265ed7c466363c3abeec492263170860689b46c7ff95\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:11:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:11:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmlt2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:11:16Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-jnh7m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:26Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:26 crc kubenswrapper[4988]: I1008 18:11:26.831313 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f328a20b-5283-407e-8a93-2ab2da599f18\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8622d56fb198dadee31fb0ca8fc80d56828b6710648055f264533e2d5bdb3037\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z2gh4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3cdb2a757332bdcdc2a2f3362913fe8f06003b0597de20b72627f9badd5e30af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z2gh4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:11:16Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-nm54f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:26Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:26 crc kubenswrapper[4988]: I1008 18:11:26.864441 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:11:26 crc kubenswrapper[4988]: I1008 18:11:26.864493 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:11:26 crc kubenswrapper[4988]: I1008 18:11:26.864505 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:11:26 crc kubenswrapper[4988]: I1008 18:11:26.864523 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:11:26 crc kubenswrapper[4988]: I1008 18:11:26.864537 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:11:26Z","lastTransitionTime":"2025-10-08T18:11:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:11:26 crc kubenswrapper[4988]: I1008 18:11:26.968777 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:11:26 crc kubenswrapper[4988]: I1008 18:11:26.968828 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:11:26 crc kubenswrapper[4988]: I1008 18:11:26.968838 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:11:26 crc kubenswrapper[4988]: I1008 18:11:26.968855 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:11:26 crc kubenswrapper[4988]: I1008 18:11:26.968868 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:11:26Z","lastTransitionTime":"2025-10-08T18:11:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:11:27 crc kubenswrapper[4988]: I1008 18:11:27.071371 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:11:27 crc kubenswrapper[4988]: I1008 18:11:27.071453 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:11:27 crc kubenswrapper[4988]: I1008 18:11:27.071462 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:11:27 crc kubenswrapper[4988]: I1008 18:11:27.071476 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:11:27 crc kubenswrapper[4988]: I1008 18:11:27.071486 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:11:27Z","lastTransitionTime":"2025-10-08T18:11:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:11:27 crc kubenswrapper[4988]: I1008 18:11:27.174347 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:11:27 crc kubenswrapper[4988]: I1008 18:11:27.174419 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:11:27 crc kubenswrapper[4988]: I1008 18:11:27.174436 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:11:27 crc kubenswrapper[4988]: I1008 18:11:27.174470 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:11:27 crc kubenswrapper[4988]: I1008 18:11:27.174489 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:11:27Z","lastTransitionTime":"2025-10-08T18:11:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:11:27 crc kubenswrapper[4988]: I1008 18:11:27.237704 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 18:11:27 crc kubenswrapper[4988]: I1008 18:11:27.237757 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 18:11:27 crc kubenswrapper[4988]: I1008 18:11:27.237808 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 18:11:27 crc kubenswrapper[4988]: E1008 18:11:27.237941 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 08 18:11:27 crc kubenswrapper[4988]: E1008 18:11:27.238088 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 08 18:11:27 crc kubenswrapper[4988]: E1008 18:11:27.238217 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 08 18:11:27 crc kubenswrapper[4988]: I1008 18:11:27.277504 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:11:27 crc kubenswrapper[4988]: I1008 18:11:27.277559 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:11:27 crc kubenswrapper[4988]: I1008 18:11:27.277570 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:11:27 crc kubenswrapper[4988]: I1008 18:11:27.277588 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:11:27 crc kubenswrapper[4988]: I1008 18:11:27.277598 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:11:27Z","lastTransitionTime":"2025-10-08T18:11:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:11:27 crc kubenswrapper[4988]: I1008 18:11:27.380601 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:11:27 crc kubenswrapper[4988]: I1008 18:11:27.380655 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:11:27 crc kubenswrapper[4988]: I1008 18:11:27.380665 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:11:27 crc kubenswrapper[4988]: I1008 18:11:27.380684 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:11:27 crc kubenswrapper[4988]: I1008 18:11:27.380699 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:11:27Z","lastTransitionTime":"2025-10-08T18:11:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:11:27 crc kubenswrapper[4988]: I1008 18:11:27.484279 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:11:27 crc kubenswrapper[4988]: I1008 18:11:27.484380 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:11:27 crc kubenswrapper[4988]: I1008 18:11:27.484452 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:11:27 crc kubenswrapper[4988]: I1008 18:11:27.484487 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:11:27 crc kubenswrapper[4988]: I1008 18:11:27.484512 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:11:27Z","lastTransitionTime":"2025-10-08T18:11:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:11:27 crc kubenswrapper[4988]: I1008 18:11:27.588026 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:11:27 crc kubenswrapper[4988]: I1008 18:11:27.588084 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:11:27 crc kubenswrapper[4988]: I1008 18:11:27.588098 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:11:27 crc kubenswrapper[4988]: I1008 18:11:27.588121 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:11:27 crc kubenswrapper[4988]: I1008 18:11:27.588135 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:11:27Z","lastTransitionTime":"2025-10-08T18:11:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:11:27 crc kubenswrapper[4988]: I1008 18:11:27.590778 4988 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-nsc65_f010a463-69f2-4082-acc3-161899a939c3/ovnkube-controller/0.log" Oct 08 18:11:27 crc kubenswrapper[4988]: I1008 18:11:27.595312 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-nsc65" event={"ID":"f010a463-69f2-4082-acc3-161899a939c3","Type":"ContainerStarted","Data":"08c1ae9c6aaca1379bd4b347fa26858dc50cbee1ae9cb9751a21351448257e11"} Oct 08 18:11:27 crc kubenswrapper[4988]: I1008 18:11:27.596428 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-nsc65" Oct 08 18:11:27 crc kubenswrapper[4988]: I1008 18:11:27.613907 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e8906182d55c285b3cf095e2d9e1dac397af4618a514f11e26b44c172dd573f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:27Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:27 crc kubenswrapper[4988]: I1008 18:11:27.628417 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-dzxs2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8f307b20-5e33-4725-99cb-210e25113532\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72ee803c6ef88524e0ce0391476f8e5a7dfbc858b095a3c2ae31bad10991ce3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q65f7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:11:16Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-dzxs2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:27Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:27 crc kubenswrapper[4988]: I1008 18:11:27.647674 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aef71be1-1d12-46e5-95b2-bd399040fd4d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:10:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:10:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:10:55Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:10:55Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:10:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77c68f3ecb8a0724499e495cb26ef3db5beecb140f2558d8c0b25f41928b84ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:10:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a560ee84c2405227de71fda127566213cd4f017e73d5581442ba267781b1c539\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:10:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6100accc3bd08928bc1ff88310f29fa5da37fd2d5c315c94279801c9778c4745\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:10:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b0e596d7ce4e9341c44499af79d3b3caf0e5e2c156902ea40cf2a3dca672be9e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5bd35aef24be034a0b305a40bb53ca171a4f76db3b93b4f78ae95099b0a4cde3\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-08T18:11:14Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1008 18:11:09.091034 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1008 18:11:09.091984 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3472374062/tls.crt::/tmp/serving-cert-3472374062/tls.key\\\\\\\"\\\\nI1008 18:11:14.619696 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1008 18:11:14.628837 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1008 18:11:14.628898 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1008 18:11:14.628994 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1008 18:11:14.629036 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1008 18:11:14.641318 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1008 18:11:14.641363 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1008 18:11:14.641372 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1008 18:11:14.641382 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1008 18:11:14.641415 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1008 18:11:14.641423 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1008 18:11:14.641431 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1008 18:11:14.641484 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1008 18:11:14.646451 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T18:10:58Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c79ef704fda1bbcc486f4c7afa3039a3ab38669833f0f66bafe24dd4cc679e60\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:10:58Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://28d21731af0f16d6a8ae6c6e9d0bebf0560d4a533db4c78c10bf1837890394ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://28d21731af0f16d6a8ae6c6e9d0bebf0560d4a533db4c78c10bf1837890394ef\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:10:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:10:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:10:55Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:27Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:27 crc kubenswrapper[4988]: I1008 18:11:27.665151 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7e0ecefc-359f-4785-bb79-0343a62a25da\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:10:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:10:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:10:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://30bec350fefee3d19758b3ed735ede4e9aa085edb57aa3f3ec24249a67dc7ca2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:10:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2d850fd27d4ed65891ff69b9add65fb78535357f54819c2a8d196f5a340c58e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:10:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://90b1a9e14c82bf0f75f08fdb342eed35a503020f9db641e5e9ceff8b122cb3bd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:10:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://16f23c4c7163d32dc0df8d6a7ee17e2864fb261f8a40ecf05556774b173fc868\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:10:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:10:55Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:27Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:27 crc kubenswrapper[4988]: I1008 18:11:27.682613 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9590ca0059f0ebbe86343cccb3475338c317382c3231e082655b61e67eece375\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:27Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:27 crc kubenswrapper[4988]: I1008 18:11:27.690520 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:11:27 crc kubenswrapper[4988]: I1008 18:11:27.690560 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:11:27 crc kubenswrapper[4988]: I1008 18:11:27.690569 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:11:27 crc kubenswrapper[4988]: I1008 18:11:27.690587 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:11:27 crc kubenswrapper[4988]: I1008 18:11:27.690599 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:11:27Z","lastTransitionTime":"2025-10-08T18:11:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:11:27 crc kubenswrapper[4988]: I1008 18:11:27.698659 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:27Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:27 crc kubenswrapper[4988]: I1008 18:11:27.715652 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-jnh7m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"556fb048-8afb-44a5-9ea0-2b7c89e6cf4b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bfa2f3382b23c3f788a1b31e80e8a90b3b6868870dccc46d86c3fc11ad027f52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmlt2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1a825886f105ccf9a7cc6f28a0cd012fb859f4990dade0e6aae932482146028\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1a825886f105ccf9a7cc6f28a0cd012fb859f4990dade0e6aae932482146028\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmlt2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ab2fe6e65418efa75bae72259157a2323bbb4b7d11d1d012053524e14135888\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0ab2fe6e65418efa75bae72259157a2323bbb4b7d11d1d012053524e14135888\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:11:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:11:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmlt2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://368a303309be112d9c0b7ccd6538fbaca7085d97a8ec8259f36092f07147e7cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://368a303309be112d9c0b7ccd6538fbaca7085d97a8ec8259f36092f07147e7cb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:11:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:11:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmlt2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a8ca69001dddef643f982a97c74c291e30cb9fac5b0e894e80066b885ea9151\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9a8ca69001dddef643f982a97c74c291e30cb9fac5b0e894e80066b885ea9151\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:11:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:11:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmlt2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://145303c46b48603ed1aae4b9bfe06332a3ffad28f206f8394574011a87b58cd3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://145303c46b48603ed1aae4b9bfe06332a3ffad28f206f8394574011a87b58cd3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:11:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:11:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmlt2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d08e5cfc7cbe5ec1a7a0265ed7c466363c3abeec492263170860689b46c7ff95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d08e5cfc7cbe5ec1a7a0265ed7c466363c3abeec492263170860689b46c7ff95\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:11:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:11:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmlt2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:11:16Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-jnh7m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:27Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:27 crc kubenswrapper[4988]: I1008 18:11:27.731808 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f328a20b-5283-407e-8a93-2ab2da599f18\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8622d56fb198dadee31fb0ca8fc80d56828b6710648055f264533e2d5bdb3037\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z2gh4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3cdb2a757332bdcdc2a2f3362913fe8f06003b0597de20b72627f9badd5e30af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z2gh4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:11:16Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-nm54f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:27Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:27 crc kubenswrapper[4988]: I1008 18:11:27.749905 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a35a5e989ba69c8ef36ce4cea272aadf67ce20473fc257adfcfa86a2faba6ae0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a522e937cfe03fb9f2e60cb1bea07214cf63c69bdecd5ba65a2dfdba882d28dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:27Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:27 crc kubenswrapper[4988]: I1008 18:11:27.765705 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:27Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:27 crc kubenswrapper[4988]: I1008 18:11:27.781302 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-h9jwp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4490d846-ca32-4cd5-9202-87cc6d499fb4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://00b7f6db043c28efda3a09fdc77a6ccded3eae8b6a3d2852aa500789d5f01194\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7k6rk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:11:16Z\\\"}}\" for pod \"openshift-multus\"/\"multus-h9jwp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:27Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:27 crc kubenswrapper[4988]: I1008 18:11:27.793144 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:11:27 crc kubenswrapper[4988]: I1008 18:11:27.793217 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:11:27 crc kubenswrapper[4988]: I1008 18:11:27.793237 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:11:27 crc kubenswrapper[4988]: I1008 18:11:27.793267 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:11:27 crc kubenswrapper[4988]: I1008 18:11:27.793290 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:11:27Z","lastTransitionTime":"2025-10-08T18:11:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:11:27 crc kubenswrapper[4988]: I1008 18:11:27.804587 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-nsc65" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f010a463-69f2-4082-acc3-161899a939c3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://48a516f6c153743bd1a21ec6945fb368a37f4d2b9d6a326b68edc096a192ed05\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dnwgr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://03406e726f817bcedb841f81959ab12808ba35a530240f9abc5fa5c608b9a6ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dnwgr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9fbbeae0409d2e47e8de9a4306f63b0d13acbfd73cd86f83303fec14352ed718\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dnwgr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://55791327a370c0009c523b86b2234e82d3794985f2eaccf1922eb90f46b2b970\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dnwgr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cee5de93bc68345f9b737857f02832f56582d58757c50a8798eec25d19b2c20d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dnwgr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d314eb248e90371fed068b1173e26d4c9f9765417a70b341e1a5b136ec17353a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dnwgr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://08c1ae9c6aaca1379bd4b347fa26858dc50cbee1ae9cb9751a21351448257e11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9353d08254f16c1a3cd3b66904805f54467292dd2c052268336aaca5678aeef7\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-08T18:11:26Z\\\",\\\"message\\\":\\\"6248 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1008 18:11:26.234639 6248 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1008 18:11:26.234656 6248 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1008 18:11:26.234670 6248 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1008 18:11:26.234760 6248 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1008 18:11:26.235089 6248 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1008 18:11:26.235124 6248 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1008 18:11:26.235165 6248 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1008 18:11:26.235169 6248 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1008 18:11:26.235180 6248 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1008 18:11:26.235210 6248 handler.go:208] Removed *v1.Node event handler 7\\\\nI1008 18:11:26.235213 6248 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1008 18:11:26.235245 6248 handler.go:208] Removed *v1.Node event handler 2\\\\nI1008 18:11:26.235284 6248 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1008 18:11:26.235326 6248 factory.go:656] Stopping \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T18:11:22Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dnwgr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://74fcb812b84291763cc6e5ccb20c12aaad580ad968dda4734a0bbfe33d2770ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dnwgr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://befe684cea9bd6dfda3ec8e85bca431c26bd8f82823b0a69dc0d04319a3ca385\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://befe684cea9bd6dfda3ec8e85bca431c26bd8f82823b0a69dc0d04319a3ca385\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dnwgr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:11:16Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-nsc65\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:27Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:27 crc kubenswrapper[4988]: I1008 18:11:27.821521 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:27Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:27 crc kubenswrapper[4988]: I1008 18:11:27.833779 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-s4f2k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ba5e11d7-7bde-4c3d-983b-884232219652\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://824bc34a2807068d3fceb10651803ae9ae5651e940f7aed4877c544d76dae4d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hn5jf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:11:15Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-s4f2k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:27Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:27 crc kubenswrapper[4988]: I1008 18:11:27.897541 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:11:27 crc kubenswrapper[4988]: I1008 18:11:27.897614 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:11:27 crc kubenswrapper[4988]: I1008 18:11:27.897634 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:11:27 crc kubenswrapper[4988]: I1008 18:11:27.897666 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:11:27 crc kubenswrapper[4988]: I1008 18:11:27.897687 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:11:27Z","lastTransitionTime":"2025-10-08T18:11:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:11:28 crc kubenswrapper[4988]: I1008 18:11:28.001501 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:11:28 crc kubenswrapper[4988]: I1008 18:11:28.001580 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:11:28 crc kubenswrapper[4988]: I1008 18:11:28.001599 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:11:28 crc kubenswrapper[4988]: I1008 18:11:28.001628 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:11:28 crc kubenswrapper[4988]: I1008 18:11:28.001649 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:11:28Z","lastTransitionTime":"2025-10-08T18:11:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:11:28 crc kubenswrapper[4988]: I1008 18:11:28.105580 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:11:28 crc kubenswrapper[4988]: I1008 18:11:28.105640 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:11:28 crc kubenswrapper[4988]: I1008 18:11:28.105659 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:11:28 crc kubenswrapper[4988]: I1008 18:11:28.105711 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:11:28 crc kubenswrapper[4988]: I1008 18:11:28.105730 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:11:28Z","lastTransitionTime":"2025-10-08T18:11:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:11:28 crc kubenswrapper[4988]: I1008 18:11:28.208832 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:11:28 crc kubenswrapper[4988]: I1008 18:11:28.208887 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:11:28 crc kubenswrapper[4988]: I1008 18:11:28.208910 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:11:28 crc kubenswrapper[4988]: I1008 18:11:28.208942 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:11:28 crc kubenswrapper[4988]: I1008 18:11:28.208963 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:11:28Z","lastTransitionTime":"2025-10-08T18:11:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:11:28 crc kubenswrapper[4988]: I1008 18:11:28.317570 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:11:28 crc kubenswrapper[4988]: I1008 18:11:28.318457 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:11:28 crc kubenswrapper[4988]: I1008 18:11:28.319041 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:11:28 crc kubenswrapper[4988]: I1008 18:11:28.319089 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:11:28 crc kubenswrapper[4988]: I1008 18:11:28.319117 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:11:28Z","lastTransitionTime":"2025-10-08T18:11:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:11:28 crc kubenswrapper[4988]: I1008 18:11:28.422355 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:11:28 crc kubenswrapper[4988]: I1008 18:11:28.422434 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:11:28 crc kubenswrapper[4988]: I1008 18:11:28.422455 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:11:28 crc kubenswrapper[4988]: I1008 18:11:28.422520 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:11:28 crc kubenswrapper[4988]: I1008 18:11:28.422538 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:11:28Z","lastTransitionTime":"2025-10-08T18:11:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:11:28 crc kubenswrapper[4988]: I1008 18:11:28.526437 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:11:28 crc kubenswrapper[4988]: I1008 18:11:28.526515 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:11:28 crc kubenswrapper[4988]: I1008 18:11:28.526539 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:11:28 crc kubenswrapper[4988]: I1008 18:11:28.526592 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:11:28 crc kubenswrapper[4988]: I1008 18:11:28.526629 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:11:28Z","lastTransitionTime":"2025-10-08T18:11:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:11:28 crc kubenswrapper[4988]: I1008 18:11:28.602684 4988 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-nsc65_f010a463-69f2-4082-acc3-161899a939c3/ovnkube-controller/1.log" Oct 08 18:11:28 crc kubenswrapper[4988]: I1008 18:11:28.603751 4988 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-nsc65_f010a463-69f2-4082-acc3-161899a939c3/ovnkube-controller/0.log" Oct 08 18:11:28 crc kubenswrapper[4988]: I1008 18:11:28.609755 4988 generic.go:334] "Generic (PLEG): container finished" podID="f010a463-69f2-4082-acc3-161899a939c3" containerID="08c1ae9c6aaca1379bd4b347fa26858dc50cbee1ae9cb9751a21351448257e11" exitCode=1 Oct 08 18:11:28 crc kubenswrapper[4988]: I1008 18:11:28.609836 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-nsc65" event={"ID":"f010a463-69f2-4082-acc3-161899a939c3","Type":"ContainerDied","Data":"08c1ae9c6aaca1379bd4b347fa26858dc50cbee1ae9cb9751a21351448257e11"} Oct 08 18:11:28 crc kubenswrapper[4988]: I1008 18:11:28.609986 4988 scope.go:117] "RemoveContainer" containerID="9353d08254f16c1a3cd3b66904805f54467292dd2c052268336aaca5678aeef7" Oct 08 18:11:28 crc kubenswrapper[4988]: I1008 18:11:28.612968 4988 scope.go:117] "RemoveContainer" containerID="08c1ae9c6aaca1379bd4b347fa26858dc50cbee1ae9cb9751a21351448257e11" Oct 08 18:11:28 crc kubenswrapper[4988]: E1008 18:11:28.613522 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-nsc65_openshift-ovn-kubernetes(f010a463-69f2-4082-acc3-161899a939c3)\"" pod="openshift-ovn-kubernetes/ovnkube-node-nsc65" podUID="f010a463-69f2-4082-acc3-161899a939c3" Oct 08 18:11:28 crc kubenswrapper[4988]: I1008 18:11:28.630728 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:11:28 crc kubenswrapper[4988]: I1008 18:11:28.630779 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:11:28 crc kubenswrapper[4988]: I1008 18:11:28.630796 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:11:28 crc kubenswrapper[4988]: I1008 18:11:28.630823 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:11:28 crc kubenswrapper[4988]: I1008 18:11:28.630842 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:11:28Z","lastTransitionTime":"2025-10-08T18:11:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:11:28 crc kubenswrapper[4988]: I1008 18:11:28.631981 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f328a20b-5283-407e-8a93-2ab2da599f18\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8622d56fb198dadee31fb0ca8fc80d56828b6710648055f264533e2d5bdb3037\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z2gh4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3cdb2a757332bdcdc2a2f3362913fe8f06003b0597de20b72627f9badd5e30af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z2gh4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:11:16Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-nm54f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:28Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:28 crc kubenswrapper[4988]: I1008 18:11:28.648999 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a35a5e989ba69c8ef36ce4cea272aadf67ce20473fc257adfcfa86a2faba6ae0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a522e937cfe03fb9f2e60cb1bea07214cf63c69bdecd5ba65a2dfdba882d28dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:28Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:28 crc kubenswrapper[4988]: I1008 18:11:28.664655 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9590ca0059f0ebbe86343cccb3475338c317382c3231e082655b61e67eece375\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:28Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:28 crc kubenswrapper[4988]: I1008 18:11:28.681916 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:28Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:28 crc kubenswrapper[4988]: I1008 18:11:28.708205 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-jnh7m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"556fb048-8afb-44a5-9ea0-2b7c89e6cf4b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bfa2f3382b23c3f788a1b31e80e8a90b3b6868870dccc46d86c3fc11ad027f52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmlt2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1a825886f105ccf9a7cc6f28a0cd012fb859f4990dade0e6aae932482146028\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1a825886f105ccf9a7cc6f28a0cd012fb859f4990dade0e6aae932482146028\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmlt2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ab2fe6e65418efa75bae72259157a2323bbb4b7d11d1d012053524e14135888\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0ab2fe6e65418efa75bae72259157a2323bbb4b7d11d1d012053524e14135888\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:11:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:11:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmlt2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://368a303309be112d9c0b7ccd6538fbaca7085d97a8ec8259f36092f07147e7cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://368a303309be112d9c0b7ccd6538fbaca7085d97a8ec8259f36092f07147e7cb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:11:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:11:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmlt2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a8ca69001dddef643f982a97c74c291e30cb9fac5b0e894e80066b885ea9151\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9a8ca69001dddef643f982a97c74c291e30cb9fac5b0e894e80066b885ea9151\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:11:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:11:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmlt2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://145303c46b48603ed1aae4b9bfe06332a3ffad28f206f8394574011a87b58cd3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://145303c46b48603ed1aae4b9bfe06332a3ffad28f206f8394574011a87b58cd3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:11:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:11:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmlt2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d08e5cfc7cbe5ec1a7a0265ed7c466363c3abeec492263170860689b46c7ff95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d08e5cfc7cbe5ec1a7a0265ed7c466363c3abeec492263170860689b46c7ff95\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:11:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:11:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmlt2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:11:16Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-jnh7m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:28Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:28 crc kubenswrapper[4988]: I1008 18:11:28.730587 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:28Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:28 crc kubenswrapper[4988]: I1008 18:11:28.733980 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:11:28 crc kubenswrapper[4988]: I1008 18:11:28.734019 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:11:28 crc kubenswrapper[4988]: I1008 18:11:28.734030 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:11:28 crc kubenswrapper[4988]: I1008 18:11:28.734049 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:11:28 crc kubenswrapper[4988]: I1008 18:11:28.734064 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:11:28Z","lastTransitionTime":"2025-10-08T18:11:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:11:28 crc kubenswrapper[4988]: I1008 18:11:28.753099 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:28Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:28 crc kubenswrapper[4988]: I1008 18:11:28.772638 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-h9jwp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4490d846-ca32-4cd5-9202-87cc6d499fb4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://00b7f6db043c28efda3a09fdc77a6ccded3eae8b6a3d2852aa500789d5f01194\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7k6rk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:11:16Z\\\"}}\" for pod \"openshift-multus\"/\"multus-h9jwp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:28Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:28 crc kubenswrapper[4988]: I1008 18:11:28.807169 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-nsc65" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f010a463-69f2-4082-acc3-161899a939c3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://48a516f6c153743bd1a21ec6945fb368a37f4d2b9d6a326b68edc096a192ed05\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dnwgr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://03406e726f817bcedb841f81959ab12808ba35a530240f9abc5fa5c608b9a6ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dnwgr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9fbbeae0409d2e47e8de9a4306f63b0d13acbfd73cd86f83303fec14352ed718\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dnwgr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://55791327a370c0009c523b86b2234e82d3794985f2eaccf1922eb90f46b2b970\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dnwgr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cee5de93bc68345f9b737857f02832f56582d58757c50a8798eec25d19b2c20d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dnwgr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d314eb248e90371fed068b1173e26d4c9f9765417a70b341e1a5b136ec17353a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dnwgr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://08c1ae9c6aaca1379bd4b347fa26858dc50cbee1ae9cb9751a21351448257e11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9353d08254f16c1a3cd3b66904805f54467292dd2c052268336aaca5678aeef7\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-08T18:11:26Z\\\",\\\"message\\\":\\\"6248 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1008 18:11:26.234639 6248 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1008 18:11:26.234656 6248 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1008 18:11:26.234670 6248 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1008 18:11:26.234760 6248 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1008 18:11:26.235089 6248 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1008 18:11:26.235124 6248 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1008 18:11:26.235165 6248 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1008 18:11:26.235169 6248 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1008 18:11:26.235180 6248 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1008 18:11:26.235210 6248 handler.go:208] Removed *v1.Node event handler 7\\\\nI1008 18:11:26.235213 6248 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1008 18:11:26.235245 6248 handler.go:208] Removed *v1.Node event handler 2\\\\nI1008 18:11:26.235284 6248 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1008 18:11:26.235326 6248 factory.go:656] Stopping \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T18:11:22Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://08c1ae9c6aaca1379bd4b347fa26858dc50cbee1ae9cb9751a21351448257e11\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-08T18:11:27Z\\\",\\\"message\\\":\\\"ntil: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1008 18:11:27.764809 6411 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1008 18:11:27.764817 6411 handler.go:208] Removed *v1.Node event handler 2\\\\nI1008 18:11:27.764773 6411 loadbalancer.go:304] Deleted 0 stale LBs for map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-ingress-operator/metrics\\\\\\\"}\\\\nI1008 18:11:27.765603 6411 transact.go:42] Configuring OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-machine-api/control-plane-machine-set-operator]} name:Service_openshift-machine-api/control-plane-machine-set-operator_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.41:9443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {589f95f7-f3e2-4140-80ed-9a0717201481}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1008 18:11:27.765621 6411 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1008 18:11:27.765753 6411 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T18:11:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dnwgr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://74fcb812b84291763cc6e5ccb20c12aaad580ad968dda4734a0bbfe33d2770ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dnwgr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://befe684cea9bd6dfda3ec8e85bca431c26bd8f82823b0a69dc0d04319a3ca385\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://befe684cea9bd6dfda3ec8e85bca431c26bd8f82823b0a69dc0d04319a3ca385\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dnwgr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:11:16Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-nsc65\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:28Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:28 crc kubenswrapper[4988]: I1008 18:11:28.830639 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-s4f2k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ba5e11d7-7bde-4c3d-983b-884232219652\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://824bc34a2807068d3fceb10651803ae9ae5651e940f7aed4877c544d76dae4d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hn5jf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:11:15Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-s4f2k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:28Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:28 crc kubenswrapper[4988]: I1008 18:11:28.838713 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:11:28 crc kubenswrapper[4988]: I1008 18:11:28.838774 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:11:28 crc kubenswrapper[4988]: I1008 18:11:28.838793 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:11:28 crc kubenswrapper[4988]: I1008 18:11:28.838822 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:11:28 crc kubenswrapper[4988]: I1008 18:11:28.838841 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:11:28Z","lastTransitionTime":"2025-10-08T18:11:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:11:28 crc kubenswrapper[4988]: I1008 18:11:28.851902 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aef71be1-1d12-46e5-95b2-bd399040fd4d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:10:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:10:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:10:55Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:10:55Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:10:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77c68f3ecb8a0724499e495cb26ef3db5beecb140f2558d8c0b25f41928b84ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:10:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a560ee84c2405227de71fda127566213cd4f017e73d5581442ba267781b1c539\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:10:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6100accc3bd08928bc1ff88310f29fa5da37fd2d5c315c94279801c9778c4745\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:10:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b0e596d7ce4e9341c44499af79d3b3caf0e5e2c156902ea40cf2a3dca672be9e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5bd35aef24be034a0b305a40bb53ca171a4f76db3b93b4f78ae95099b0a4cde3\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-08T18:11:14Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1008 18:11:09.091034 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1008 18:11:09.091984 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3472374062/tls.crt::/tmp/serving-cert-3472374062/tls.key\\\\\\\"\\\\nI1008 18:11:14.619696 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1008 18:11:14.628837 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1008 18:11:14.628898 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1008 18:11:14.628994 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1008 18:11:14.629036 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1008 18:11:14.641318 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1008 18:11:14.641363 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1008 18:11:14.641372 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1008 18:11:14.641382 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1008 18:11:14.641415 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1008 18:11:14.641423 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1008 18:11:14.641431 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1008 18:11:14.641484 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1008 18:11:14.646451 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T18:10:58Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c79ef704fda1bbcc486f4c7afa3039a3ab38669833f0f66bafe24dd4cc679e60\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:10:58Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://28d21731af0f16d6a8ae6c6e9d0bebf0560d4a533db4c78c10bf1837890394ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://28d21731af0f16d6a8ae6c6e9d0bebf0560d4a533db4c78c10bf1837890394ef\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:10:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:10:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:10:55Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:28Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:28 crc kubenswrapper[4988]: I1008 18:11:28.876533 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7e0ecefc-359f-4785-bb79-0343a62a25da\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:10:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:10:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:10:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://30bec350fefee3d19758b3ed735ede4e9aa085edb57aa3f3ec24249a67dc7ca2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:10:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2d850fd27d4ed65891ff69b9add65fb78535357f54819c2a8d196f5a340c58e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:10:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://90b1a9e14c82bf0f75f08fdb342eed35a503020f9db641e5e9ceff8b122cb3bd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:10:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://16f23c4c7163d32dc0df8d6a7ee17e2864fb261f8a40ecf05556774b173fc868\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:10:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:10:55Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:28Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:28 crc kubenswrapper[4988]: I1008 18:11:28.901673 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e8906182d55c285b3cf095e2d9e1dac397af4618a514f11e26b44c172dd573f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:28Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:28 crc kubenswrapper[4988]: I1008 18:11:28.919539 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-dzxs2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8f307b20-5e33-4725-99cb-210e25113532\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72ee803c6ef88524e0ce0391476f8e5a7dfbc858b095a3c2ae31bad10991ce3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q65f7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:11:16Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-dzxs2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:28Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:28 crc kubenswrapper[4988]: I1008 18:11:28.941848 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:11:28 crc kubenswrapper[4988]: I1008 18:11:28.941898 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:11:28 crc kubenswrapper[4988]: I1008 18:11:28.941908 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:11:28 crc kubenswrapper[4988]: I1008 18:11:28.941926 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:11:28 crc kubenswrapper[4988]: I1008 18:11:28.941939 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:11:28Z","lastTransitionTime":"2025-10-08T18:11:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:11:29 crc kubenswrapper[4988]: I1008 18:11:29.045032 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:11:29 crc kubenswrapper[4988]: I1008 18:11:29.045417 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:11:29 crc kubenswrapper[4988]: I1008 18:11:29.045491 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:11:29 crc kubenswrapper[4988]: I1008 18:11:29.045616 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:11:29 crc kubenswrapper[4988]: I1008 18:11:29.045688 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:11:29Z","lastTransitionTime":"2025-10-08T18:11:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:11:29 crc kubenswrapper[4988]: I1008 18:11:29.148616 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:11:29 crc kubenswrapper[4988]: I1008 18:11:29.148711 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:11:29 crc kubenswrapper[4988]: I1008 18:11:29.148736 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:11:29 crc kubenswrapper[4988]: I1008 18:11:29.148770 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:11:29 crc kubenswrapper[4988]: I1008 18:11:29.148791 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:11:29Z","lastTransitionTime":"2025-10-08T18:11:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:11:29 crc kubenswrapper[4988]: I1008 18:11:29.236839 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 18:11:29 crc kubenswrapper[4988]: I1008 18:11:29.236924 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 18:11:29 crc kubenswrapper[4988]: I1008 18:11:29.237042 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 18:11:29 crc kubenswrapper[4988]: E1008 18:11:29.237271 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 08 18:11:29 crc kubenswrapper[4988]: E1008 18:11:29.237566 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 08 18:11:29 crc kubenswrapper[4988]: E1008 18:11:29.237787 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 08 18:11:29 crc kubenswrapper[4988]: I1008 18:11:29.252295 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:11:29 crc kubenswrapper[4988]: I1008 18:11:29.252372 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:11:29 crc kubenswrapper[4988]: I1008 18:11:29.252420 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:11:29 crc kubenswrapper[4988]: I1008 18:11:29.252455 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:11:29 crc kubenswrapper[4988]: I1008 18:11:29.252477 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:11:29Z","lastTransitionTime":"2025-10-08T18:11:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:11:29 crc kubenswrapper[4988]: I1008 18:11:29.356097 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:11:29 crc kubenswrapper[4988]: I1008 18:11:29.356168 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:11:29 crc kubenswrapper[4988]: I1008 18:11:29.356187 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:11:29 crc kubenswrapper[4988]: I1008 18:11:29.356219 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:11:29 crc kubenswrapper[4988]: I1008 18:11:29.356269 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:11:29Z","lastTransitionTime":"2025-10-08T18:11:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:11:29 crc kubenswrapper[4988]: I1008 18:11:29.459716 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:11:29 crc kubenswrapper[4988]: I1008 18:11:29.459787 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:11:29 crc kubenswrapper[4988]: I1008 18:11:29.459806 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:11:29 crc kubenswrapper[4988]: I1008 18:11:29.459834 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:11:29 crc kubenswrapper[4988]: I1008 18:11:29.459855 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:11:29Z","lastTransitionTime":"2025-10-08T18:11:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:11:29 crc kubenswrapper[4988]: I1008 18:11:29.563696 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:11:29 crc kubenswrapper[4988]: I1008 18:11:29.563774 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:11:29 crc kubenswrapper[4988]: I1008 18:11:29.563803 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:11:29 crc kubenswrapper[4988]: I1008 18:11:29.563839 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:11:29 crc kubenswrapper[4988]: I1008 18:11:29.563859 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:11:29Z","lastTransitionTime":"2025-10-08T18:11:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:11:29 crc kubenswrapper[4988]: I1008 18:11:29.616799 4988 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-nsc65_f010a463-69f2-4082-acc3-161899a939c3/ovnkube-controller/1.log" Oct 08 18:11:29 crc kubenswrapper[4988]: I1008 18:11:29.622245 4988 scope.go:117] "RemoveContainer" containerID="08c1ae9c6aaca1379bd4b347fa26858dc50cbee1ae9cb9751a21351448257e11" Oct 08 18:11:29 crc kubenswrapper[4988]: E1008 18:11:29.622457 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-nsc65_openshift-ovn-kubernetes(f010a463-69f2-4082-acc3-161899a939c3)\"" pod="openshift-ovn-kubernetes/ovnkube-node-nsc65" podUID="f010a463-69f2-4082-acc3-161899a939c3" Oct 08 18:11:29 crc kubenswrapper[4988]: I1008 18:11:29.642461 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:29Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:29 crc kubenswrapper[4988]: I1008 18:11:29.661215 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:29Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:29 crc kubenswrapper[4988]: I1008 18:11:29.666303 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:11:29 crc kubenswrapper[4988]: I1008 18:11:29.666361 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:11:29 crc kubenswrapper[4988]: I1008 18:11:29.666423 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:11:29 crc kubenswrapper[4988]: I1008 18:11:29.666461 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:11:29 crc kubenswrapper[4988]: I1008 18:11:29.666483 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:11:29Z","lastTransitionTime":"2025-10-08T18:11:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:11:29 crc kubenswrapper[4988]: I1008 18:11:29.681820 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-h9jwp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4490d846-ca32-4cd5-9202-87cc6d499fb4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://00b7f6db043c28efda3a09fdc77a6ccded3eae8b6a3d2852aa500789d5f01194\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7k6rk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:11:16Z\\\"}}\" for pod \"openshift-multus\"/\"multus-h9jwp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:29Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:29 crc kubenswrapper[4988]: I1008 18:11:29.713238 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-nsc65" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f010a463-69f2-4082-acc3-161899a939c3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://48a516f6c153743bd1a21ec6945fb368a37f4d2b9d6a326b68edc096a192ed05\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dnwgr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://03406e726f817bcedb841f81959ab12808ba35a530240f9abc5fa5c608b9a6ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dnwgr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9fbbeae0409d2e47e8de9a4306f63b0d13acbfd73cd86f83303fec14352ed718\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dnwgr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://55791327a370c0009c523b86b2234e82d3794985f2eaccf1922eb90f46b2b970\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dnwgr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cee5de93bc68345f9b737857f02832f56582d58757c50a8798eec25d19b2c20d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dnwgr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d314eb248e90371fed068b1173e26d4c9f9765417a70b341e1a5b136ec17353a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dnwgr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://08c1ae9c6aaca1379bd4b347fa26858dc50cbee1ae9cb9751a21351448257e11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://08c1ae9c6aaca1379bd4b347fa26858dc50cbee1ae9cb9751a21351448257e11\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-08T18:11:27Z\\\",\\\"message\\\":\\\"ntil: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1008 18:11:27.764809 6411 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1008 18:11:27.764817 6411 handler.go:208] Removed *v1.Node event handler 2\\\\nI1008 18:11:27.764773 6411 loadbalancer.go:304] Deleted 0 stale LBs for map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-ingress-operator/metrics\\\\\\\"}\\\\nI1008 18:11:27.765603 6411 transact.go:42] Configuring OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-machine-api/control-plane-machine-set-operator]} name:Service_openshift-machine-api/control-plane-machine-set-operator_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.41:9443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {589f95f7-f3e2-4140-80ed-9a0717201481}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1008 18:11:27.765621 6411 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1008 18:11:27.765753 6411 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T18:11:26Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-nsc65_openshift-ovn-kubernetes(f010a463-69f2-4082-acc3-161899a939c3)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dnwgr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://74fcb812b84291763cc6e5ccb20c12aaad580ad968dda4734a0bbfe33d2770ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dnwgr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://befe684cea9bd6dfda3ec8e85bca431c26bd8f82823b0a69dc0d04319a3ca385\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://befe684cea9bd6dfda3ec8e85bca431c26bd8f82823b0a69dc0d04319a3ca385\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dnwgr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:11:16Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-nsc65\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:29Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:29 crc kubenswrapper[4988]: I1008 18:11:29.729076 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-s4f2k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ba5e11d7-7bde-4c3d-983b-884232219652\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://824bc34a2807068d3fceb10651803ae9ae5651e940f7aed4877c544d76dae4d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hn5jf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:11:15Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-s4f2k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:29Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:29 crc kubenswrapper[4988]: I1008 18:11:29.753155 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aef71be1-1d12-46e5-95b2-bd399040fd4d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:10:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:10:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:10:55Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:10:55Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:10:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77c68f3ecb8a0724499e495cb26ef3db5beecb140f2558d8c0b25f41928b84ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:10:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a560ee84c2405227de71fda127566213cd4f017e73d5581442ba267781b1c539\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:10:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6100accc3bd08928bc1ff88310f29fa5da37fd2d5c315c94279801c9778c4745\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:10:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b0e596d7ce4e9341c44499af79d3b3caf0e5e2c156902ea40cf2a3dca672be9e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5bd35aef24be034a0b305a40bb53ca171a4f76db3b93b4f78ae95099b0a4cde3\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-08T18:11:14Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1008 18:11:09.091034 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1008 18:11:09.091984 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3472374062/tls.crt::/tmp/serving-cert-3472374062/tls.key\\\\\\\"\\\\nI1008 18:11:14.619696 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1008 18:11:14.628837 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1008 18:11:14.628898 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1008 18:11:14.628994 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1008 18:11:14.629036 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1008 18:11:14.641318 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1008 18:11:14.641363 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1008 18:11:14.641372 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1008 18:11:14.641382 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1008 18:11:14.641415 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1008 18:11:14.641423 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1008 18:11:14.641431 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1008 18:11:14.641484 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1008 18:11:14.646451 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T18:10:58Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c79ef704fda1bbcc486f4c7afa3039a3ab38669833f0f66bafe24dd4cc679e60\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:10:58Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://28d21731af0f16d6a8ae6c6e9d0bebf0560d4a533db4c78c10bf1837890394ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://28d21731af0f16d6a8ae6c6e9d0bebf0560d4a533db4c78c10bf1837890394ef\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:10:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:10:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:10:55Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:29Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:29 crc kubenswrapper[4988]: I1008 18:11:29.769523 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:11:29 crc kubenswrapper[4988]: I1008 18:11:29.769724 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:11:29 crc kubenswrapper[4988]: I1008 18:11:29.769819 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:11:29 crc kubenswrapper[4988]: I1008 18:11:29.769862 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:11:29 crc kubenswrapper[4988]: I1008 18:11:29.769891 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:11:29Z","lastTransitionTime":"2025-10-08T18:11:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:11:29 crc kubenswrapper[4988]: I1008 18:11:29.772294 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7e0ecefc-359f-4785-bb79-0343a62a25da\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:10:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:10:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:10:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://30bec350fefee3d19758b3ed735ede4e9aa085edb57aa3f3ec24249a67dc7ca2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:10:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2d850fd27d4ed65891ff69b9add65fb78535357f54819c2a8d196f5a340c58e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:10:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://90b1a9e14c82bf0f75f08fdb342eed35a503020f9db641e5e9ceff8b122cb3bd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:10:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://16f23c4c7163d32dc0df8d6a7ee17e2864fb261f8a40ecf05556774b173fc868\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:10:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:10:55Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:29Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:29 crc kubenswrapper[4988]: I1008 18:11:29.793830 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e8906182d55c285b3cf095e2d9e1dac397af4618a514f11e26b44c172dd573f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:29Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:29 crc kubenswrapper[4988]: I1008 18:11:29.809611 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-dzxs2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8f307b20-5e33-4725-99cb-210e25113532\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72ee803c6ef88524e0ce0391476f8e5a7dfbc858b095a3c2ae31bad10991ce3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q65f7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:11:16Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-dzxs2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:29Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:29 crc kubenswrapper[4988]: I1008 18:11:29.827814 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a35a5e989ba69c8ef36ce4cea272aadf67ce20473fc257adfcfa86a2faba6ae0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a522e937cfe03fb9f2e60cb1bea07214cf63c69bdecd5ba65a2dfdba882d28dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:29Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:29 crc kubenswrapper[4988]: I1008 18:11:29.849588 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9590ca0059f0ebbe86343cccb3475338c317382c3231e082655b61e67eece375\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:29Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:29 crc kubenswrapper[4988]: I1008 18:11:29.869099 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:29Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:29 crc kubenswrapper[4988]: I1008 18:11:29.874287 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:11:29 crc kubenswrapper[4988]: I1008 18:11:29.874348 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:11:29 crc kubenswrapper[4988]: I1008 18:11:29.874365 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:11:29 crc kubenswrapper[4988]: I1008 18:11:29.874422 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:11:29 crc kubenswrapper[4988]: I1008 18:11:29.874437 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:11:29Z","lastTransitionTime":"2025-10-08T18:11:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:11:29 crc kubenswrapper[4988]: I1008 18:11:29.893003 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-jnh7m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"556fb048-8afb-44a5-9ea0-2b7c89e6cf4b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bfa2f3382b23c3f788a1b31e80e8a90b3b6868870dccc46d86c3fc11ad027f52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmlt2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1a825886f105ccf9a7cc6f28a0cd012fb859f4990dade0e6aae932482146028\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1a825886f105ccf9a7cc6f28a0cd012fb859f4990dade0e6aae932482146028\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmlt2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ab2fe6e65418efa75bae72259157a2323bbb4b7d11d1d012053524e14135888\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0ab2fe6e65418efa75bae72259157a2323bbb4b7d11d1d012053524e14135888\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:11:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:11:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmlt2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://368a303309be112d9c0b7ccd6538fbaca7085d97a8ec8259f36092f07147e7cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://368a303309be112d9c0b7ccd6538fbaca7085d97a8ec8259f36092f07147e7cb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:11:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:11:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmlt2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a8ca69001dddef643f982a97c74c291e30cb9fac5b0e894e80066b885ea9151\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9a8ca69001dddef643f982a97c74c291e30cb9fac5b0e894e80066b885ea9151\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:11:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:11:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmlt2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://145303c46b48603ed1aae4b9bfe06332a3ffad28f206f8394574011a87b58cd3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://145303c46b48603ed1aae4b9bfe06332a3ffad28f206f8394574011a87b58cd3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:11:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:11:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmlt2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d08e5cfc7cbe5ec1a7a0265ed7c466363c3abeec492263170860689b46c7ff95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d08e5cfc7cbe5ec1a7a0265ed7c466363c3abeec492263170860689b46c7ff95\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:11:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:11:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmlt2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:11:16Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-jnh7m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:29Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:29 crc kubenswrapper[4988]: I1008 18:11:29.911587 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f328a20b-5283-407e-8a93-2ab2da599f18\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8622d56fb198dadee31fb0ca8fc80d56828b6710648055f264533e2d5bdb3037\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z2gh4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3cdb2a757332bdcdc2a2f3362913fe8f06003b0597de20b72627f9badd5e30af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z2gh4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:11:16Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-nm54f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:29Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:29 crc kubenswrapper[4988]: I1008 18:11:29.978222 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:11:29 crc kubenswrapper[4988]: I1008 18:11:29.978275 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:11:29 crc kubenswrapper[4988]: I1008 18:11:29.978291 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:11:29 crc kubenswrapper[4988]: I1008 18:11:29.978316 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:11:29 crc kubenswrapper[4988]: I1008 18:11:29.978334 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:11:29Z","lastTransitionTime":"2025-10-08T18:11:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:11:29 crc kubenswrapper[4988]: I1008 18:11:29.989622 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-7wl8v"] Oct 08 18:11:29 crc kubenswrapper[4988]: I1008 18:11:29.990563 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-7wl8v" Oct 08 18:11:29 crc kubenswrapper[4988]: I1008 18:11:29.994032 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert" Oct 08 18:11:29 crc kubenswrapper[4988]: I1008 18:11:29.997580 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd" Oct 08 18:11:30 crc kubenswrapper[4988]: I1008 18:11:30.019982 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a35a5e989ba69c8ef36ce4cea272aadf67ce20473fc257adfcfa86a2faba6ae0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a522e937cfe03fb9f2e60cb1bea07214cf63c69bdecd5ba65a2dfdba882d28dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:30Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:30 crc kubenswrapper[4988]: I1008 18:11:30.036663 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9590ca0059f0ebbe86343cccb3475338c317382c3231e082655b61e67eece375\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:30Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:30 crc kubenswrapper[4988]: I1008 18:11:30.056608 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:30Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:30 crc kubenswrapper[4988]: I1008 18:11:30.071968 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-jnh7m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"556fb048-8afb-44a5-9ea0-2b7c89e6cf4b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bfa2f3382b23c3f788a1b31e80e8a90b3b6868870dccc46d86c3fc11ad027f52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmlt2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1a825886f105ccf9a7cc6f28a0cd012fb859f4990dade0e6aae932482146028\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1a825886f105ccf9a7cc6f28a0cd012fb859f4990dade0e6aae932482146028\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmlt2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ab2fe6e65418efa75bae72259157a2323bbb4b7d11d1d012053524e14135888\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0ab2fe6e65418efa75bae72259157a2323bbb4b7d11d1d012053524e14135888\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:11:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:11:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmlt2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://368a303309be112d9c0b7ccd6538fbaca7085d97a8ec8259f36092f07147e7cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://368a303309be112d9c0b7ccd6538fbaca7085d97a8ec8259f36092f07147e7cb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:11:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:11:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmlt2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a8ca69001dddef643f982a97c74c291e30cb9fac5b0e894e80066b885ea9151\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9a8ca69001dddef643f982a97c74c291e30cb9fac5b0e894e80066b885ea9151\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:11:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:11:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmlt2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://145303c46b48603ed1aae4b9bfe06332a3ffad28f206f8394574011a87b58cd3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://145303c46b48603ed1aae4b9bfe06332a3ffad28f206f8394574011a87b58cd3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:11:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:11:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmlt2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d08e5cfc7cbe5ec1a7a0265ed7c466363c3abeec492263170860689b46c7ff95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d08e5cfc7cbe5ec1a7a0265ed7c466363c3abeec492263170860689b46c7ff95\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:11:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:11:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmlt2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:11:16Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-jnh7m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:30Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:30 crc kubenswrapper[4988]: I1008 18:11:30.082351 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:11:30 crc kubenswrapper[4988]: I1008 18:11:30.082493 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:11:30 crc kubenswrapper[4988]: I1008 18:11:30.082512 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:11:30 crc kubenswrapper[4988]: I1008 18:11:30.082911 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:11:30 crc kubenswrapper[4988]: I1008 18:11:30.083086 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:11:30Z","lastTransitionTime":"2025-10-08T18:11:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:11:30 crc kubenswrapper[4988]: I1008 18:11:30.088914 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f328a20b-5283-407e-8a93-2ab2da599f18\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8622d56fb198dadee31fb0ca8fc80d56828b6710648055f264533e2d5bdb3037\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z2gh4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3cdb2a757332bdcdc2a2f3362913fe8f06003b0597de20b72627f9badd5e30af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z2gh4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:11:16Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-nm54f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:30Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:30 crc kubenswrapper[4988]: I1008 18:11:30.111198 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:30Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:30 crc kubenswrapper[4988]: I1008 18:11:30.115985 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/e433488d-db52-435f-b89c-f63cd477c5d3-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-7wl8v\" (UID: \"e433488d-db52-435f-b89c-f63cd477c5d3\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-7wl8v" Oct 08 18:11:30 crc kubenswrapper[4988]: I1008 18:11:30.116136 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/e433488d-db52-435f-b89c-f63cd477c5d3-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-7wl8v\" (UID: \"e433488d-db52-435f-b89c-f63cd477c5d3\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-7wl8v" Oct 08 18:11:30 crc kubenswrapper[4988]: I1008 18:11:30.116260 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/e433488d-db52-435f-b89c-f63cd477c5d3-env-overrides\") pod \"ovnkube-control-plane-749d76644c-7wl8v\" (UID: \"e433488d-db52-435f-b89c-f63cd477c5d3\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-7wl8v" Oct 08 18:11:30 crc kubenswrapper[4988]: I1008 18:11:30.116421 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dbz6l\" (UniqueName: \"kubernetes.io/projected/e433488d-db52-435f-b89c-f63cd477c5d3-kube-api-access-dbz6l\") pod \"ovnkube-control-plane-749d76644c-7wl8v\" (UID: \"e433488d-db52-435f-b89c-f63cd477c5d3\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-7wl8v" Oct 08 18:11:30 crc kubenswrapper[4988]: I1008 18:11:30.126720 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:30Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:30 crc kubenswrapper[4988]: I1008 18:11:30.147859 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-h9jwp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4490d846-ca32-4cd5-9202-87cc6d499fb4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://00b7f6db043c28efda3a09fdc77a6ccded3eae8b6a3d2852aa500789d5f01194\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7k6rk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:11:16Z\\\"}}\" for pod \"openshift-multus\"/\"multus-h9jwp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:30Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:30 crc kubenswrapper[4988]: I1008 18:11:30.168205 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-nsc65" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f010a463-69f2-4082-acc3-161899a939c3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://48a516f6c153743bd1a21ec6945fb368a37f4d2b9d6a326b68edc096a192ed05\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dnwgr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://03406e726f817bcedb841f81959ab12808ba35a530240f9abc5fa5c608b9a6ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dnwgr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9fbbeae0409d2e47e8de9a4306f63b0d13acbfd73cd86f83303fec14352ed718\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dnwgr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://55791327a370c0009c523b86b2234e82d3794985f2eaccf1922eb90f46b2b970\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dnwgr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cee5de93bc68345f9b737857f02832f56582d58757c50a8798eec25d19b2c20d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dnwgr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d314eb248e90371fed068b1173e26d4c9f9765417a70b341e1a5b136ec17353a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dnwgr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://08c1ae9c6aaca1379bd4b347fa26858dc50cbee1ae9cb9751a21351448257e11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://08c1ae9c6aaca1379bd4b347fa26858dc50cbee1ae9cb9751a21351448257e11\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-08T18:11:27Z\\\",\\\"message\\\":\\\"ntil: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1008 18:11:27.764809 6411 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1008 18:11:27.764817 6411 handler.go:208] Removed *v1.Node event handler 2\\\\nI1008 18:11:27.764773 6411 loadbalancer.go:304] Deleted 0 stale LBs for map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-ingress-operator/metrics\\\\\\\"}\\\\nI1008 18:11:27.765603 6411 transact.go:42] Configuring OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-machine-api/control-plane-machine-set-operator]} name:Service_openshift-machine-api/control-plane-machine-set-operator_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.41:9443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {589f95f7-f3e2-4140-80ed-9a0717201481}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1008 18:11:27.765621 6411 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1008 18:11:27.765753 6411 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T18:11:26Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-nsc65_openshift-ovn-kubernetes(f010a463-69f2-4082-acc3-161899a939c3)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dnwgr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://74fcb812b84291763cc6e5ccb20c12aaad580ad968dda4734a0bbfe33d2770ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dnwgr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://befe684cea9bd6dfda3ec8e85bca431c26bd8f82823b0a69dc0d04319a3ca385\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://befe684cea9bd6dfda3ec8e85bca431c26bd8f82823b0a69dc0d04319a3ca385\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dnwgr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:11:16Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-nsc65\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:30Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:30 crc kubenswrapper[4988]: I1008 18:11:30.180818 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-7wl8v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e433488d-db52-435f-b89c-f63cd477c5d3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:29Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:29Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dbz6l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dbz6l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:11:29Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-7wl8v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:30Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:30 crc kubenswrapper[4988]: I1008 18:11:30.186670 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:11:30 crc kubenswrapper[4988]: I1008 18:11:30.186728 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:11:30 crc kubenswrapper[4988]: I1008 18:11:30.186746 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:11:30 crc kubenswrapper[4988]: I1008 18:11:30.186773 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:11:30 crc kubenswrapper[4988]: I1008 18:11:30.186791 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:11:30Z","lastTransitionTime":"2025-10-08T18:11:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:11:30 crc kubenswrapper[4988]: I1008 18:11:30.193347 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-s4f2k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ba5e11d7-7bde-4c3d-983b-884232219652\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://824bc34a2807068d3fceb10651803ae9ae5651e940f7aed4877c544d76dae4d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hn5jf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:11:15Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-s4f2k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:30Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:30 crc kubenswrapper[4988]: I1008 18:11:30.208488 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7e0ecefc-359f-4785-bb79-0343a62a25da\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:10:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:10:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:10:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://30bec350fefee3d19758b3ed735ede4e9aa085edb57aa3f3ec24249a67dc7ca2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:10:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2d850fd27d4ed65891ff69b9add65fb78535357f54819c2a8d196f5a340c58e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:10:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://90b1a9e14c82bf0f75f08fdb342eed35a503020f9db641e5e9ceff8b122cb3bd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:10:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://16f23c4c7163d32dc0df8d6a7ee17e2864fb261f8a40ecf05556774b173fc868\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:10:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:10:55Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:30Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:30 crc kubenswrapper[4988]: I1008 18:11:30.217903 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dbz6l\" (UniqueName: \"kubernetes.io/projected/e433488d-db52-435f-b89c-f63cd477c5d3-kube-api-access-dbz6l\") pod \"ovnkube-control-plane-749d76644c-7wl8v\" (UID: \"e433488d-db52-435f-b89c-f63cd477c5d3\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-7wl8v" Oct 08 18:11:30 crc kubenswrapper[4988]: I1008 18:11:30.217975 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/e433488d-db52-435f-b89c-f63cd477c5d3-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-7wl8v\" (UID: \"e433488d-db52-435f-b89c-f63cd477c5d3\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-7wl8v" Oct 08 18:11:30 crc kubenswrapper[4988]: I1008 18:11:30.218074 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/e433488d-db52-435f-b89c-f63cd477c5d3-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-7wl8v\" (UID: \"e433488d-db52-435f-b89c-f63cd477c5d3\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-7wl8v" Oct 08 18:11:30 crc kubenswrapper[4988]: I1008 18:11:30.218137 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/e433488d-db52-435f-b89c-f63cd477c5d3-env-overrides\") pod \"ovnkube-control-plane-749d76644c-7wl8v\" (UID: \"e433488d-db52-435f-b89c-f63cd477c5d3\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-7wl8v" Oct 08 18:11:30 crc kubenswrapper[4988]: I1008 18:11:30.219153 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/e433488d-db52-435f-b89c-f63cd477c5d3-env-overrides\") pod \"ovnkube-control-plane-749d76644c-7wl8v\" (UID: \"e433488d-db52-435f-b89c-f63cd477c5d3\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-7wl8v" Oct 08 18:11:30 crc kubenswrapper[4988]: I1008 18:11:30.219966 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/e433488d-db52-435f-b89c-f63cd477c5d3-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-7wl8v\" (UID: \"e433488d-db52-435f-b89c-f63cd477c5d3\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-7wl8v" Oct 08 18:11:30 crc kubenswrapper[4988]: I1008 18:11:30.226049 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e8906182d55c285b3cf095e2d9e1dac397af4618a514f11e26b44c172dd573f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:30Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:30 crc kubenswrapper[4988]: I1008 18:11:30.227408 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/e433488d-db52-435f-b89c-f63cd477c5d3-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-7wl8v\" (UID: \"e433488d-db52-435f-b89c-f63cd477c5d3\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-7wl8v" Oct 08 18:11:30 crc kubenswrapper[4988]: I1008 18:11:30.237772 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-dzxs2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8f307b20-5e33-4725-99cb-210e25113532\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72ee803c6ef88524e0ce0391476f8e5a7dfbc858b095a3c2ae31bad10991ce3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q65f7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:11:16Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-dzxs2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:30Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:30 crc kubenswrapper[4988]: I1008 18:11:30.241323 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dbz6l\" (UniqueName: \"kubernetes.io/projected/e433488d-db52-435f-b89c-f63cd477c5d3-kube-api-access-dbz6l\") pod \"ovnkube-control-plane-749d76644c-7wl8v\" (UID: \"e433488d-db52-435f-b89c-f63cd477c5d3\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-7wl8v" Oct 08 18:11:30 crc kubenswrapper[4988]: I1008 18:11:30.254569 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aef71be1-1d12-46e5-95b2-bd399040fd4d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:10:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:10:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:10:55Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:10:55Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:10:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77c68f3ecb8a0724499e495cb26ef3db5beecb140f2558d8c0b25f41928b84ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:10:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a560ee84c2405227de71fda127566213cd4f017e73d5581442ba267781b1c539\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:10:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6100accc3bd08928bc1ff88310f29fa5da37fd2d5c315c94279801c9778c4745\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:10:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b0e596d7ce4e9341c44499af79d3b3caf0e5e2c156902ea40cf2a3dca672be9e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5bd35aef24be034a0b305a40bb53ca171a4f76db3b93b4f78ae95099b0a4cde3\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-08T18:11:14Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1008 18:11:09.091034 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1008 18:11:09.091984 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3472374062/tls.crt::/tmp/serving-cert-3472374062/tls.key\\\\\\\"\\\\nI1008 18:11:14.619696 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1008 18:11:14.628837 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1008 18:11:14.628898 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1008 18:11:14.628994 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1008 18:11:14.629036 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1008 18:11:14.641318 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1008 18:11:14.641363 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1008 18:11:14.641372 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1008 18:11:14.641382 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1008 18:11:14.641415 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1008 18:11:14.641423 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1008 18:11:14.641431 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1008 18:11:14.641484 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1008 18:11:14.646451 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T18:10:58Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c79ef704fda1bbcc486f4c7afa3039a3ab38669833f0f66bafe24dd4cc679e60\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:10:58Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://28d21731af0f16d6a8ae6c6e9d0bebf0560d4a533db4c78c10bf1837890394ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://28d21731af0f16d6a8ae6c6e9d0bebf0560d4a533db4c78c10bf1837890394ef\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:10:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:10:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:10:55Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:30Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:30 crc kubenswrapper[4988]: I1008 18:11:30.290761 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:11:30 crc kubenswrapper[4988]: I1008 18:11:30.290845 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:11:30 crc kubenswrapper[4988]: I1008 18:11:30.290862 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:11:30 crc kubenswrapper[4988]: I1008 18:11:30.290883 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:11:30 crc kubenswrapper[4988]: I1008 18:11:30.290898 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:11:30Z","lastTransitionTime":"2025-10-08T18:11:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:11:30 crc kubenswrapper[4988]: I1008 18:11:30.309920 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-7wl8v" Oct 08 18:11:30 crc kubenswrapper[4988]: W1008 18:11:30.330349 4988 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode433488d_db52_435f_b89c_f63cd477c5d3.slice/crio-749c887f5fe9932f94a43795df129bb83941cd92cc975dcf11107b8dd57a69b6 WatchSource:0}: Error finding container 749c887f5fe9932f94a43795df129bb83941cd92cc975dcf11107b8dd57a69b6: Status 404 returned error can't find the container with id 749c887f5fe9932f94a43795df129bb83941cd92cc975dcf11107b8dd57a69b6 Oct 08 18:11:30 crc kubenswrapper[4988]: I1008 18:11:30.394406 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:11:30 crc kubenswrapper[4988]: I1008 18:11:30.394581 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:11:30 crc kubenswrapper[4988]: I1008 18:11:30.394593 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:11:30 crc kubenswrapper[4988]: I1008 18:11:30.394635 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:11:30 crc kubenswrapper[4988]: I1008 18:11:30.394647 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:11:30Z","lastTransitionTime":"2025-10-08T18:11:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:11:30 crc kubenswrapper[4988]: I1008 18:11:30.497012 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:11:30 crc kubenswrapper[4988]: I1008 18:11:30.497055 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:11:30 crc kubenswrapper[4988]: I1008 18:11:30.497064 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:11:30 crc kubenswrapper[4988]: I1008 18:11:30.497081 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:11:30 crc kubenswrapper[4988]: I1008 18:11:30.497094 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:11:30Z","lastTransitionTime":"2025-10-08T18:11:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:11:30 crc kubenswrapper[4988]: I1008 18:11:30.601720 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:11:30 crc kubenswrapper[4988]: I1008 18:11:30.601769 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:11:30 crc kubenswrapper[4988]: I1008 18:11:30.601786 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:11:30 crc kubenswrapper[4988]: I1008 18:11:30.601809 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:11:30 crc kubenswrapper[4988]: I1008 18:11:30.601826 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:11:30Z","lastTransitionTime":"2025-10-08T18:11:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:11:30 crc kubenswrapper[4988]: I1008 18:11:30.626336 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-7wl8v" event={"ID":"e433488d-db52-435f-b89c-f63cd477c5d3","Type":"ContainerStarted","Data":"76b30d183684740b647c1af11dc49712232cc8b0c9a392b491580e5ddb75ea7f"} Oct 08 18:11:30 crc kubenswrapper[4988]: I1008 18:11:30.626414 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-7wl8v" event={"ID":"e433488d-db52-435f-b89c-f63cd477c5d3","Type":"ContainerStarted","Data":"749c887f5fe9932f94a43795df129bb83941cd92cc975dcf11107b8dd57a69b6"} Oct 08 18:11:30 crc kubenswrapper[4988]: I1008 18:11:30.704349 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:11:30 crc kubenswrapper[4988]: I1008 18:11:30.704427 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:11:30 crc kubenswrapper[4988]: I1008 18:11:30.704445 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:11:30 crc kubenswrapper[4988]: I1008 18:11:30.704470 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:11:30 crc kubenswrapper[4988]: I1008 18:11:30.704517 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:11:30Z","lastTransitionTime":"2025-10-08T18:11:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:11:30 crc kubenswrapper[4988]: I1008 18:11:30.808265 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:11:30 crc kubenswrapper[4988]: I1008 18:11:30.808331 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:11:30 crc kubenswrapper[4988]: I1008 18:11:30.808349 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:11:30 crc kubenswrapper[4988]: I1008 18:11:30.808372 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:11:30 crc kubenswrapper[4988]: I1008 18:11:30.808418 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:11:30Z","lastTransitionTime":"2025-10-08T18:11:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:11:30 crc kubenswrapper[4988]: I1008 18:11:30.911129 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:11:30 crc kubenswrapper[4988]: I1008 18:11:30.911181 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:11:30 crc kubenswrapper[4988]: I1008 18:11:30.911190 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:11:30 crc kubenswrapper[4988]: I1008 18:11:30.911209 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:11:30 crc kubenswrapper[4988]: I1008 18:11:30.911231 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:11:30Z","lastTransitionTime":"2025-10-08T18:11:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:11:30 crc kubenswrapper[4988]: I1008 18:11:30.926914 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 18:11:30 crc kubenswrapper[4988]: E1008 18:11:30.927161 4988 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 18:11:46.927118477 +0000 UTC m=+52.376961257 (durationBeforeRetry 16s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 18:11:30 crc kubenswrapper[4988]: I1008 18:11:30.927266 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 18:11:30 crc kubenswrapper[4988]: I1008 18:11:30.927331 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 18:11:30 crc kubenswrapper[4988]: E1008 18:11:30.927521 4988 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 08 18:11:30 crc kubenswrapper[4988]: E1008 18:11:30.927565 4988 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 08 18:11:30 crc kubenswrapper[4988]: E1008 18:11:30.927650 4988 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-08 18:11:46.927609822 +0000 UTC m=+52.377452592 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 08 18:11:30 crc kubenswrapper[4988]: E1008 18:11:30.927678 4988 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-08 18:11:46.927668974 +0000 UTC m=+52.377511984 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 08 18:11:31 crc kubenswrapper[4988]: I1008 18:11:31.014733 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:11:31 crc kubenswrapper[4988]: I1008 18:11:31.014783 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:11:31 crc kubenswrapper[4988]: I1008 18:11:31.014794 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:11:31 crc kubenswrapper[4988]: I1008 18:11:31.014811 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:11:31 crc kubenswrapper[4988]: I1008 18:11:31.014826 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:11:31Z","lastTransitionTime":"2025-10-08T18:11:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:11:31 crc kubenswrapper[4988]: I1008 18:11:31.028453 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 18:11:31 crc kubenswrapper[4988]: I1008 18:11:31.028526 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 18:11:31 crc kubenswrapper[4988]: E1008 18:11:31.028687 4988 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 08 18:11:31 crc kubenswrapper[4988]: E1008 18:11:31.028716 4988 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 08 18:11:31 crc kubenswrapper[4988]: E1008 18:11:31.028743 4988 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 08 18:11:31 crc kubenswrapper[4988]: E1008 18:11:31.028758 4988 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 08 18:11:31 crc kubenswrapper[4988]: E1008 18:11:31.028811 4988 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-08 18:11:47.02879122 +0000 UTC m=+52.478634000 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 08 18:11:31 crc kubenswrapper[4988]: E1008 18:11:31.028722 4988 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 08 18:11:31 crc kubenswrapper[4988]: E1008 18:11:31.028838 4988 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 08 18:11:31 crc kubenswrapper[4988]: E1008 18:11:31.028945 4988 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-08 18:11:47.028922914 +0000 UTC m=+52.478765684 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 08 18:11:31 crc kubenswrapper[4988]: I1008 18:11:31.106102 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 08 18:11:31 crc kubenswrapper[4988]: I1008 18:11:31.117706 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:11:31 crc kubenswrapper[4988]: I1008 18:11:31.117764 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:11:31 crc kubenswrapper[4988]: I1008 18:11:31.117778 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:11:31 crc kubenswrapper[4988]: I1008 18:11:31.117800 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:11:31 crc kubenswrapper[4988]: I1008 18:11:31.117814 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:11:31Z","lastTransitionTime":"2025-10-08T18:11:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:11:31 crc kubenswrapper[4988]: I1008 18:11:31.132590 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e8906182d55c285b3cf095e2d9e1dac397af4618a514f11e26b44c172dd573f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:31Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:31 crc kubenswrapper[4988]: I1008 18:11:31.136480 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/network-metrics-daemon-jn2sx"] Oct 08 18:11:31 crc kubenswrapper[4988]: I1008 18:11:31.137166 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jn2sx" Oct 08 18:11:31 crc kubenswrapper[4988]: E1008 18:11:31.137260 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jn2sx" podUID="171e0902-95ee-4a39-8b1b-a58704c30574" Oct 08 18:11:31 crc kubenswrapper[4988]: I1008 18:11:31.149979 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-dzxs2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8f307b20-5e33-4725-99cb-210e25113532\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72ee803c6ef88524e0ce0391476f8e5a7dfbc858b095a3c2ae31bad10991ce3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q65f7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:11:16Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-dzxs2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:31Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:31 crc kubenswrapper[4988]: I1008 18:11:31.171015 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aef71be1-1d12-46e5-95b2-bd399040fd4d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:10:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:10:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:10:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77c68f3ecb8a0724499e495cb26ef3db5beecb140f2558d8c0b25f41928b84ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:10:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a560ee84c2405227de71fda127566213cd4f017e73d5581442ba267781b1c539\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:10:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6100accc3bd08928bc1ff88310f29fa5da37fd2d5c315c94279801c9778c4745\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:10:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b0e596d7ce4e9341c44499af79d3b3caf0e5e2c156902ea40cf2a3dca672be9e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5bd35aef24be034a0b305a40bb53ca171a4f76db3b93b4f78ae95099b0a4cde3\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-08T18:11:14Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1008 18:11:09.091034 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1008 18:11:09.091984 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3472374062/tls.crt::/tmp/serving-cert-3472374062/tls.key\\\\\\\"\\\\nI1008 18:11:14.619696 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1008 18:11:14.628837 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1008 18:11:14.628898 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1008 18:11:14.628994 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1008 18:11:14.629036 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1008 18:11:14.641318 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1008 18:11:14.641363 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1008 18:11:14.641372 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1008 18:11:14.641382 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1008 18:11:14.641415 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1008 18:11:14.641423 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1008 18:11:14.641431 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1008 18:11:14.641484 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1008 18:11:14.646451 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T18:10:58Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c79ef704fda1bbcc486f4c7afa3039a3ab38669833f0f66bafe24dd4cc679e60\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:10:58Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://28d21731af0f16d6a8ae6c6e9d0bebf0560d4a533db4c78c10bf1837890394ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://28d21731af0f16d6a8ae6c6e9d0bebf0560d4a533db4c78c10bf1837890394ef\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:10:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:10:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:10:55Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:31Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:31 crc kubenswrapper[4988]: I1008 18:11:31.189377 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7e0ecefc-359f-4785-bb79-0343a62a25da\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:10:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:10:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:10:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://30bec350fefee3d19758b3ed735ede4e9aa085edb57aa3f3ec24249a67dc7ca2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:10:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2d850fd27d4ed65891ff69b9add65fb78535357f54819c2a8d196f5a340c58e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:10:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://90b1a9e14c82bf0f75f08fdb342eed35a503020f9db641e5e9ceff8b122cb3bd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:10:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://16f23c4c7163d32dc0df8d6a7ee17e2864fb261f8a40ecf05556774b173fc868\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:10:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:10:55Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:31Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:31 crc kubenswrapper[4988]: I1008 18:11:31.207881 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9590ca0059f0ebbe86343cccb3475338c317382c3231e082655b61e67eece375\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:31Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:31 crc kubenswrapper[4988]: I1008 18:11:31.220484 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:11:31 crc kubenswrapper[4988]: I1008 18:11:31.220566 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:11:31 crc kubenswrapper[4988]: I1008 18:11:31.220576 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:11:31 crc kubenswrapper[4988]: I1008 18:11:31.220599 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:11:31 crc kubenswrapper[4988]: I1008 18:11:31.220619 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:11:31Z","lastTransitionTime":"2025-10-08T18:11:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:11:31 crc kubenswrapper[4988]: I1008 18:11:31.228548 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:31Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:31 crc kubenswrapper[4988]: I1008 18:11:31.231122 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9f4q2\" (UniqueName: \"kubernetes.io/projected/171e0902-95ee-4a39-8b1b-a58704c30574-kube-api-access-9f4q2\") pod \"network-metrics-daemon-jn2sx\" (UID: \"171e0902-95ee-4a39-8b1b-a58704c30574\") " pod="openshift-multus/network-metrics-daemon-jn2sx" Oct 08 18:11:31 crc kubenswrapper[4988]: I1008 18:11:31.231223 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/171e0902-95ee-4a39-8b1b-a58704c30574-metrics-certs\") pod \"network-metrics-daemon-jn2sx\" (UID: \"171e0902-95ee-4a39-8b1b-a58704c30574\") " pod="openshift-multus/network-metrics-daemon-jn2sx" Oct 08 18:11:31 crc kubenswrapper[4988]: I1008 18:11:31.237520 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 18:11:31 crc kubenswrapper[4988]: I1008 18:11:31.237561 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 18:11:31 crc kubenswrapper[4988]: I1008 18:11:31.237568 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 18:11:31 crc kubenswrapper[4988]: E1008 18:11:31.237697 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 08 18:11:31 crc kubenswrapper[4988]: E1008 18:11:31.237816 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 08 18:11:31 crc kubenswrapper[4988]: E1008 18:11:31.237964 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 08 18:11:31 crc kubenswrapper[4988]: I1008 18:11:31.256644 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-jnh7m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"556fb048-8afb-44a5-9ea0-2b7c89e6cf4b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bfa2f3382b23c3f788a1b31e80e8a90b3b6868870dccc46d86c3fc11ad027f52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmlt2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1a825886f105ccf9a7cc6f28a0cd012fb859f4990dade0e6aae932482146028\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1a825886f105ccf9a7cc6f28a0cd012fb859f4990dade0e6aae932482146028\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmlt2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ab2fe6e65418efa75bae72259157a2323bbb4b7d11d1d012053524e14135888\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0ab2fe6e65418efa75bae72259157a2323bbb4b7d11d1d012053524e14135888\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:11:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:11:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmlt2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://368a303309be112d9c0b7ccd6538fbaca7085d97a8ec8259f36092f07147e7cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://368a303309be112d9c0b7ccd6538fbaca7085d97a8ec8259f36092f07147e7cb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:11:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:11:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmlt2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a8ca69001dddef643f982a97c74c291e30cb9fac5b0e894e80066b885ea9151\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9a8ca69001dddef643f982a97c74c291e30cb9fac5b0e894e80066b885ea9151\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:11:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:11:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmlt2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://145303c46b48603ed1aae4b9bfe06332a3ffad28f206f8394574011a87b58cd3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://145303c46b48603ed1aae4b9bfe06332a3ffad28f206f8394574011a87b58cd3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:11:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:11:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmlt2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d08e5cfc7cbe5ec1a7a0265ed7c466363c3abeec492263170860689b46c7ff95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d08e5cfc7cbe5ec1a7a0265ed7c466363c3abeec492263170860689b46c7ff95\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:11:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:11:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmlt2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:11:16Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-jnh7m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:31Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:31 crc kubenswrapper[4988]: I1008 18:11:31.279077 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f328a20b-5283-407e-8a93-2ab2da599f18\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8622d56fb198dadee31fb0ca8fc80d56828b6710648055f264533e2d5bdb3037\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z2gh4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3cdb2a757332bdcdc2a2f3362913fe8f06003b0597de20b72627f9badd5e30af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z2gh4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:11:16Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-nm54f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:31Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:31 crc kubenswrapper[4988]: I1008 18:11:31.296791 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a35a5e989ba69c8ef36ce4cea272aadf67ce20473fc257adfcfa86a2faba6ae0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a522e937cfe03fb9f2e60cb1bea07214cf63c69bdecd5ba65a2dfdba882d28dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:31Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:31 crc kubenswrapper[4988]: I1008 18:11:31.312991 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:31Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:31 crc kubenswrapper[4988]: I1008 18:11:31.323647 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:11:31 crc kubenswrapper[4988]: I1008 18:11:31.323728 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:11:31 crc kubenswrapper[4988]: I1008 18:11:31.323751 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:11:31 crc kubenswrapper[4988]: I1008 18:11:31.323783 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:11:31 crc kubenswrapper[4988]: I1008 18:11:31.323806 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:11:31Z","lastTransitionTime":"2025-10-08T18:11:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:11:31 crc kubenswrapper[4988]: I1008 18:11:31.332225 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-h9jwp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4490d846-ca32-4cd5-9202-87cc6d499fb4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://00b7f6db043c28efda3a09fdc77a6ccded3eae8b6a3d2852aa500789d5f01194\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7k6rk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:11:16Z\\\"}}\" for pod \"openshift-multus\"/\"multus-h9jwp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:31Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:31 crc kubenswrapper[4988]: I1008 18:11:31.332633 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9f4q2\" (UniqueName: \"kubernetes.io/projected/171e0902-95ee-4a39-8b1b-a58704c30574-kube-api-access-9f4q2\") pod \"network-metrics-daemon-jn2sx\" (UID: \"171e0902-95ee-4a39-8b1b-a58704c30574\") " pod="openshift-multus/network-metrics-daemon-jn2sx" Oct 08 18:11:31 crc kubenswrapper[4988]: I1008 18:11:31.332722 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/171e0902-95ee-4a39-8b1b-a58704c30574-metrics-certs\") pod \"network-metrics-daemon-jn2sx\" (UID: \"171e0902-95ee-4a39-8b1b-a58704c30574\") " pod="openshift-multus/network-metrics-daemon-jn2sx" Oct 08 18:11:31 crc kubenswrapper[4988]: E1008 18:11:31.332898 4988 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 08 18:11:31 crc kubenswrapper[4988]: E1008 18:11:31.332984 4988 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/171e0902-95ee-4a39-8b1b-a58704c30574-metrics-certs podName:171e0902-95ee-4a39-8b1b-a58704c30574 nodeName:}" failed. No retries permitted until 2025-10-08 18:11:31.832958772 +0000 UTC m=+37.282801552 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/171e0902-95ee-4a39-8b1b-a58704c30574-metrics-certs") pod "network-metrics-daemon-jn2sx" (UID: "171e0902-95ee-4a39-8b1b-a58704c30574") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 08 18:11:31 crc kubenswrapper[4988]: I1008 18:11:31.351740 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9f4q2\" (UniqueName: \"kubernetes.io/projected/171e0902-95ee-4a39-8b1b-a58704c30574-kube-api-access-9f4q2\") pod \"network-metrics-daemon-jn2sx\" (UID: \"171e0902-95ee-4a39-8b1b-a58704c30574\") " pod="openshift-multus/network-metrics-daemon-jn2sx" Oct 08 18:11:31 crc kubenswrapper[4988]: I1008 18:11:31.366420 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-nsc65" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f010a463-69f2-4082-acc3-161899a939c3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://48a516f6c153743bd1a21ec6945fb368a37f4d2b9d6a326b68edc096a192ed05\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dnwgr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://03406e726f817bcedb841f81959ab12808ba35a530240f9abc5fa5c608b9a6ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dnwgr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9fbbeae0409d2e47e8de9a4306f63b0d13acbfd73cd86f83303fec14352ed718\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dnwgr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://55791327a370c0009c523b86b2234e82d3794985f2eaccf1922eb90f46b2b970\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dnwgr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cee5de93bc68345f9b737857f02832f56582d58757c50a8798eec25d19b2c20d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dnwgr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d314eb248e90371fed068b1173e26d4c9f9765417a70b341e1a5b136ec17353a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dnwgr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://08c1ae9c6aaca1379bd4b347fa26858dc50cbee1ae9cb9751a21351448257e11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://08c1ae9c6aaca1379bd4b347fa26858dc50cbee1ae9cb9751a21351448257e11\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-08T18:11:27Z\\\",\\\"message\\\":\\\"ntil: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1008 18:11:27.764809 6411 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1008 18:11:27.764817 6411 handler.go:208] Removed *v1.Node event handler 2\\\\nI1008 18:11:27.764773 6411 loadbalancer.go:304] Deleted 0 stale LBs for map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-ingress-operator/metrics\\\\\\\"}\\\\nI1008 18:11:27.765603 6411 transact.go:42] Configuring OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-machine-api/control-plane-machine-set-operator]} name:Service_openshift-machine-api/control-plane-machine-set-operator_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.41:9443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {589f95f7-f3e2-4140-80ed-9a0717201481}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1008 18:11:27.765621 6411 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1008 18:11:27.765753 6411 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T18:11:26Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-nsc65_openshift-ovn-kubernetes(f010a463-69f2-4082-acc3-161899a939c3)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dnwgr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://74fcb812b84291763cc6e5ccb20c12aaad580ad968dda4734a0bbfe33d2770ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dnwgr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://befe684cea9bd6dfda3ec8e85bca431c26bd8f82823b0a69dc0d04319a3ca385\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://befe684cea9bd6dfda3ec8e85bca431c26bd8f82823b0a69dc0d04319a3ca385\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dnwgr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:11:16Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-nsc65\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:31Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:31 crc kubenswrapper[4988]: I1008 18:11:31.384310 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-7wl8v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e433488d-db52-435f-b89c-f63cd477c5d3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:29Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:29Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dbz6l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dbz6l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:11:29Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-7wl8v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:31Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:31 crc kubenswrapper[4988]: I1008 18:11:31.400828 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:31Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:31 crc kubenswrapper[4988]: I1008 18:11:31.414009 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-s4f2k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ba5e11d7-7bde-4c3d-983b-884232219652\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://824bc34a2807068d3fceb10651803ae9ae5651e940f7aed4877c544d76dae4d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hn5jf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:11:15Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-s4f2k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:31Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:31 crc kubenswrapper[4988]: I1008 18:11:31.424073 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:11:31 crc kubenswrapper[4988]: I1008 18:11:31.424131 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:11:31 crc kubenswrapper[4988]: I1008 18:11:31.424145 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:11:31 crc kubenswrapper[4988]: I1008 18:11:31.424168 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:11:31 crc kubenswrapper[4988]: I1008 18:11:31.424183 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:11:31Z","lastTransitionTime":"2025-10-08T18:11:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:11:31 crc kubenswrapper[4988]: I1008 18:11:31.426314 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-s4f2k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ba5e11d7-7bde-4c3d-983b-884232219652\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://824bc34a2807068d3fceb10651803ae9ae5651e940f7aed4877c544d76dae4d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hn5jf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:11:15Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-s4f2k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:31Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:31 crc kubenswrapper[4988]: E1008 18:11:31.440839 4988 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T18:11:31Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:31Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T18:11:31Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:31Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T18:11:31Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:31Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T18:11:31Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:31Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"95ee9fd0-8bc9-4ca5-9d79-5ee94d58bfeb\\\",\\\"systemUUID\\\":\\\"bd9489c4-fb7d-486d-b7ce-a13b6d7de36a\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:31Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:31 crc kubenswrapper[4988]: I1008 18:11:31.442191 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aef71be1-1d12-46e5-95b2-bd399040fd4d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:10:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:10:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:10:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77c68f3ecb8a0724499e495cb26ef3db5beecb140f2558d8c0b25f41928b84ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:10:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a560ee84c2405227de71fda127566213cd4f017e73d5581442ba267781b1c539\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:10:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6100accc3bd08928bc1ff88310f29fa5da37fd2d5c315c94279801c9778c4745\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:10:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b0e596d7ce4e9341c44499af79d3b3caf0e5e2c156902ea40cf2a3dca672be9e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5bd35aef24be034a0b305a40bb53ca171a4f76db3b93b4f78ae95099b0a4cde3\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-08T18:11:14Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1008 18:11:09.091034 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1008 18:11:09.091984 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3472374062/tls.crt::/tmp/serving-cert-3472374062/tls.key\\\\\\\"\\\\nI1008 18:11:14.619696 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1008 18:11:14.628837 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1008 18:11:14.628898 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1008 18:11:14.628994 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1008 18:11:14.629036 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1008 18:11:14.641318 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1008 18:11:14.641363 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1008 18:11:14.641372 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1008 18:11:14.641382 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1008 18:11:14.641415 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1008 18:11:14.641423 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1008 18:11:14.641431 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1008 18:11:14.641484 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1008 18:11:14.646451 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T18:10:58Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c79ef704fda1bbcc486f4c7afa3039a3ab38669833f0f66bafe24dd4cc679e60\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:10:58Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://28d21731af0f16d6a8ae6c6e9d0bebf0560d4a533db4c78c10bf1837890394ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://28d21731af0f16d6a8ae6c6e9d0bebf0560d4a533db4c78c10bf1837890394ef\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:10:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:10:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:10:55Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:31Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:31 crc kubenswrapper[4988]: I1008 18:11:31.445311 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:11:31 crc kubenswrapper[4988]: I1008 18:11:31.445337 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:11:31 crc kubenswrapper[4988]: I1008 18:11:31.445347 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:11:31 crc kubenswrapper[4988]: I1008 18:11:31.445361 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:11:31 crc kubenswrapper[4988]: I1008 18:11:31.445375 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:11:31Z","lastTransitionTime":"2025-10-08T18:11:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:11:31 crc kubenswrapper[4988]: I1008 18:11:31.456685 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7e0ecefc-359f-4785-bb79-0343a62a25da\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:10:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:10:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:10:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://30bec350fefee3d19758b3ed735ede4e9aa085edb57aa3f3ec24249a67dc7ca2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:10:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2d850fd27d4ed65891ff69b9add65fb78535357f54819c2a8d196f5a340c58e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:10:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://90b1a9e14c82bf0f75f08fdb342eed35a503020f9db641e5e9ceff8b122cb3bd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:10:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://16f23c4c7163d32dc0df8d6a7ee17e2864fb261f8a40ecf05556774b173fc868\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:10:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:10:55Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:31Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:31 crc kubenswrapper[4988]: E1008 18:11:31.458223 4988 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T18:11:31Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:31Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T18:11:31Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:31Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T18:11:31Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:31Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T18:11:31Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:31Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"95ee9fd0-8bc9-4ca5-9d79-5ee94d58bfeb\\\",\\\"systemUUID\\\":\\\"bd9489c4-fb7d-486d-b7ce-a13b6d7de36a\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:31Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:31 crc kubenswrapper[4988]: I1008 18:11:31.462514 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:11:31 crc kubenswrapper[4988]: I1008 18:11:31.462577 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:11:31 crc kubenswrapper[4988]: I1008 18:11:31.462593 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:11:31 crc kubenswrapper[4988]: I1008 18:11:31.462618 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:11:31 crc kubenswrapper[4988]: I1008 18:11:31.462656 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:11:31Z","lastTransitionTime":"2025-10-08T18:11:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:11:31 crc kubenswrapper[4988]: I1008 18:11:31.472754 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e8906182d55c285b3cf095e2d9e1dac397af4618a514f11e26b44c172dd573f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:31Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:31 crc kubenswrapper[4988]: E1008 18:11:31.476631 4988 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T18:11:31Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:31Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T18:11:31Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:31Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T18:11:31Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:31Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T18:11:31Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:31Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"95ee9fd0-8bc9-4ca5-9d79-5ee94d58bfeb\\\",\\\"systemUUID\\\":\\\"bd9489c4-fb7d-486d-b7ce-a13b6d7de36a\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:31Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:31 crc kubenswrapper[4988]: I1008 18:11:31.481345 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:11:31 crc kubenswrapper[4988]: I1008 18:11:31.481417 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:11:31 crc kubenswrapper[4988]: I1008 18:11:31.481430 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:11:31 crc kubenswrapper[4988]: I1008 18:11:31.481455 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:11:31 crc kubenswrapper[4988]: I1008 18:11:31.481473 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:11:31Z","lastTransitionTime":"2025-10-08T18:11:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:11:31 crc kubenswrapper[4988]: I1008 18:11:31.485071 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-dzxs2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8f307b20-5e33-4725-99cb-210e25113532\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72ee803c6ef88524e0ce0391476f8e5a7dfbc858b095a3c2ae31bad10991ce3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q65f7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:11:16Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-dzxs2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:31Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:31 crc kubenswrapper[4988]: E1008 18:11:31.495267 4988 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T18:11:31Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:31Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T18:11:31Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:31Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T18:11:31Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:31Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T18:11:31Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:31Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"95ee9fd0-8bc9-4ca5-9d79-5ee94d58bfeb\\\",\\\"systemUUID\\\":\\\"bd9489c4-fb7d-486d-b7ce-a13b6d7de36a\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:31Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:31 crc kubenswrapper[4988]: I1008 18:11:31.499725 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:11:31 crc kubenswrapper[4988]: I1008 18:11:31.499786 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:11:31 crc kubenswrapper[4988]: I1008 18:11:31.499803 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:11:31 crc kubenswrapper[4988]: I1008 18:11:31.499824 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:11:31 crc kubenswrapper[4988]: I1008 18:11:31.499841 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:11:31Z","lastTransitionTime":"2025-10-08T18:11:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:11:31 crc kubenswrapper[4988]: I1008 18:11:31.500410 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a35a5e989ba69c8ef36ce4cea272aadf67ce20473fc257adfcfa86a2faba6ae0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a522e937cfe03fb9f2e60cb1bea07214cf63c69bdecd5ba65a2dfdba882d28dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:31Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:31 crc kubenswrapper[4988]: E1008 18:11:31.513060 4988 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T18:11:31Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:31Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T18:11:31Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:31Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T18:11:31Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:31Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T18:11:31Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:31Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"95ee9fd0-8bc9-4ca5-9d79-5ee94d58bfeb\\\",\\\"systemUUID\\\":\\\"bd9489c4-fb7d-486d-b7ce-a13b6d7de36a\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:31Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:31 crc kubenswrapper[4988]: E1008 18:11:31.513318 4988 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Oct 08 18:11:31 crc kubenswrapper[4988]: I1008 18:11:31.515164 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9590ca0059f0ebbe86343cccb3475338c317382c3231e082655b61e67eece375\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:31Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:31 crc kubenswrapper[4988]: I1008 18:11:31.515531 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:11:31 crc kubenswrapper[4988]: I1008 18:11:31.515589 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:11:31 crc kubenswrapper[4988]: I1008 18:11:31.515599 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:11:31 crc kubenswrapper[4988]: I1008 18:11:31.515619 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:11:31 crc kubenswrapper[4988]: I1008 18:11:31.515632 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:11:31Z","lastTransitionTime":"2025-10-08T18:11:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:11:31 crc kubenswrapper[4988]: I1008 18:11:31.530290 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:31Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:31 crc kubenswrapper[4988]: I1008 18:11:31.547577 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-jnh7m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"556fb048-8afb-44a5-9ea0-2b7c89e6cf4b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bfa2f3382b23c3f788a1b31e80e8a90b3b6868870dccc46d86c3fc11ad027f52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmlt2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1a825886f105ccf9a7cc6f28a0cd012fb859f4990dade0e6aae932482146028\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1a825886f105ccf9a7cc6f28a0cd012fb859f4990dade0e6aae932482146028\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmlt2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ab2fe6e65418efa75bae72259157a2323bbb4b7d11d1d012053524e14135888\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0ab2fe6e65418efa75bae72259157a2323bbb4b7d11d1d012053524e14135888\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:11:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:11:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmlt2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://368a303309be112d9c0b7ccd6538fbaca7085d97a8ec8259f36092f07147e7cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://368a303309be112d9c0b7ccd6538fbaca7085d97a8ec8259f36092f07147e7cb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:11:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:11:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmlt2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a8ca69001dddef643f982a97c74c291e30cb9fac5b0e894e80066b885ea9151\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9a8ca69001dddef643f982a97c74c291e30cb9fac5b0e894e80066b885ea9151\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:11:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:11:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmlt2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://145303c46b48603ed1aae4b9bfe06332a3ffad28f206f8394574011a87b58cd3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://145303c46b48603ed1aae4b9bfe06332a3ffad28f206f8394574011a87b58cd3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:11:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:11:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmlt2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d08e5cfc7cbe5ec1a7a0265ed7c466363c3abeec492263170860689b46c7ff95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d08e5cfc7cbe5ec1a7a0265ed7c466363c3abeec492263170860689b46c7ff95\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:11:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:11:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmlt2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:11:16Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-jnh7m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:31Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:31 crc kubenswrapper[4988]: I1008 18:11:31.561677 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f328a20b-5283-407e-8a93-2ab2da599f18\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8622d56fb198dadee31fb0ca8fc80d56828b6710648055f264533e2d5bdb3037\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z2gh4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3cdb2a757332bdcdc2a2f3362913fe8f06003b0597de20b72627f9badd5e30af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z2gh4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:11:16Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-nm54f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:31Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:31 crc kubenswrapper[4988]: I1008 18:11:31.574284 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-jn2sx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"171e0902-95ee-4a39-8b1b-a58704c30574\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9f4q2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9f4q2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:11:31Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-jn2sx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:31Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:31 crc kubenswrapper[4988]: I1008 18:11:31.588789 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:31Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:31 crc kubenswrapper[4988]: I1008 18:11:31.603627 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:31Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:31 crc kubenswrapper[4988]: I1008 18:11:31.619497 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:11:31 crc kubenswrapper[4988]: I1008 18:11:31.619546 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:11:31 crc kubenswrapper[4988]: I1008 18:11:31.619557 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:11:31 crc kubenswrapper[4988]: I1008 18:11:31.619578 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:11:31 crc kubenswrapper[4988]: I1008 18:11:31.619593 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:11:31Z","lastTransitionTime":"2025-10-08T18:11:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:11:31 crc kubenswrapper[4988]: I1008 18:11:31.622358 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-h9jwp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4490d846-ca32-4cd5-9202-87cc6d499fb4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://00b7f6db043c28efda3a09fdc77a6ccded3eae8b6a3d2852aa500789d5f01194\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7k6rk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:11:16Z\\\"}}\" for pod \"openshift-multus\"/\"multus-h9jwp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:31Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:31 crc kubenswrapper[4988]: I1008 18:11:31.633422 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-7wl8v" event={"ID":"e433488d-db52-435f-b89c-f63cd477c5d3","Type":"ContainerStarted","Data":"e55016c3e1b97d1b3a857130a5c8e2ec6da2333f786032118a61e3e2ad4c5651"} Oct 08 18:11:31 crc kubenswrapper[4988]: I1008 18:11:31.652207 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-nsc65" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f010a463-69f2-4082-acc3-161899a939c3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://48a516f6c153743bd1a21ec6945fb368a37f4d2b9d6a326b68edc096a192ed05\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dnwgr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://03406e726f817bcedb841f81959ab12808ba35a530240f9abc5fa5c608b9a6ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dnwgr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9fbbeae0409d2e47e8de9a4306f63b0d13acbfd73cd86f83303fec14352ed718\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dnwgr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://55791327a370c0009c523b86b2234e82d3794985f2eaccf1922eb90f46b2b970\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dnwgr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cee5de93bc68345f9b737857f02832f56582d58757c50a8798eec25d19b2c20d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dnwgr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d314eb248e90371fed068b1173e26d4c9f9765417a70b341e1a5b136ec17353a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dnwgr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://08c1ae9c6aaca1379bd4b347fa26858dc50cbee1ae9cb9751a21351448257e11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://08c1ae9c6aaca1379bd4b347fa26858dc50cbee1ae9cb9751a21351448257e11\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-08T18:11:27Z\\\",\\\"message\\\":\\\"ntil: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1008 18:11:27.764809 6411 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1008 18:11:27.764817 6411 handler.go:208] Removed *v1.Node event handler 2\\\\nI1008 18:11:27.764773 6411 loadbalancer.go:304] Deleted 0 stale LBs for map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-ingress-operator/metrics\\\\\\\"}\\\\nI1008 18:11:27.765603 6411 transact.go:42] Configuring OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-machine-api/control-plane-machine-set-operator]} name:Service_openshift-machine-api/control-plane-machine-set-operator_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.41:9443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {589f95f7-f3e2-4140-80ed-9a0717201481}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1008 18:11:27.765621 6411 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1008 18:11:27.765753 6411 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T18:11:26Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-nsc65_openshift-ovn-kubernetes(f010a463-69f2-4082-acc3-161899a939c3)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dnwgr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://74fcb812b84291763cc6e5ccb20c12aaad580ad968dda4734a0bbfe33d2770ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dnwgr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://befe684cea9bd6dfda3ec8e85bca431c26bd8f82823b0a69dc0d04319a3ca385\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://befe684cea9bd6dfda3ec8e85bca431c26bd8f82823b0a69dc0d04319a3ca385\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dnwgr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:11:16Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-nsc65\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:31Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:31 crc kubenswrapper[4988]: I1008 18:11:31.670399 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-7wl8v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e433488d-db52-435f-b89c-f63cd477c5d3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:29Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:29Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dbz6l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dbz6l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:11:29Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-7wl8v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:31Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:31 crc kubenswrapper[4988]: I1008 18:11:31.682676 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-s4f2k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ba5e11d7-7bde-4c3d-983b-884232219652\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://824bc34a2807068d3fceb10651803ae9ae5651e940f7aed4877c544d76dae4d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hn5jf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:11:15Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-s4f2k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:31Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:31 crc kubenswrapper[4988]: I1008 18:11:31.697326 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e8906182d55c285b3cf095e2d9e1dac397af4618a514f11e26b44c172dd573f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:31Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:31 crc kubenswrapper[4988]: I1008 18:11:31.710105 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-dzxs2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8f307b20-5e33-4725-99cb-210e25113532\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72ee803c6ef88524e0ce0391476f8e5a7dfbc858b095a3c2ae31bad10991ce3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q65f7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:11:16Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-dzxs2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:31Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:31 crc kubenswrapper[4988]: I1008 18:11:31.723214 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:11:31 crc kubenswrapper[4988]: I1008 18:11:31.723328 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:11:31 crc kubenswrapper[4988]: I1008 18:11:31.723341 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:11:31 crc kubenswrapper[4988]: I1008 18:11:31.723394 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:11:31 crc kubenswrapper[4988]: I1008 18:11:31.723408 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:11:31Z","lastTransitionTime":"2025-10-08T18:11:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:11:31 crc kubenswrapper[4988]: I1008 18:11:31.724818 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aef71be1-1d12-46e5-95b2-bd399040fd4d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:10:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:10:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:10:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77c68f3ecb8a0724499e495cb26ef3db5beecb140f2558d8c0b25f41928b84ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:10:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a560ee84c2405227de71fda127566213cd4f017e73d5581442ba267781b1c539\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:10:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6100accc3bd08928bc1ff88310f29fa5da37fd2d5c315c94279801c9778c4745\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:10:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b0e596d7ce4e9341c44499af79d3b3caf0e5e2c156902ea40cf2a3dca672be9e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5bd35aef24be034a0b305a40bb53ca171a4f76db3b93b4f78ae95099b0a4cde3\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-08T18:11:14Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1008 18:11:09.091034 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1008 18:11:09.091984 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3472374062/tls.crt::/tmp/serving-cert-3472374062/tls.key\\\\\\\"\\\\nI1008 18:11:14.619696 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1008 18:11:14.628837 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1008 18:11:14.628898 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1008 18:11:14.628994 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1008 18:11:14.629036 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1008 18:11:14.641318 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1008 18:11:14.641363 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1008 18:11:14.641372 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1008 18:11:14.641382 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1008 18:11:14.641415 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1008 18:11:14.641423 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1008 18:11:14.641431 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1008 18:11:14.641484 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1008 18:11:14.646451 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T18:10:58Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c79ef704fda1bbcc486f4c7afa3039a3ab38669833f0f66bafe24dd4cc679e60\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:10:58Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://28d21731af0f16d6a8ae6c6e9d0bebf0560d4a533db4c78c10bf1837890394ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://28d21731af0f16d6a8ae6c6e9d0bebf0560d4a533db4c78c10bf1837890394ef\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:10:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:10:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:10:55Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:31Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:31 crc kubenswrapper[4988]: I1008 18:11:31.744282 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7e0ecefc-359f-4785-bb79-0343a62a25da\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:10:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:10:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:10:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://30bec350fefee3d19758b3ed735ede4e9aa085edb57aa3f3ec24249a67dc7ca2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:10:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2d850fd27d4ed65891ff69b9add65fb78535357f54819c2a8d196f5a340c58e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:10:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://90b1a9e14c82bf0f75f08fdb342eed35a503020f9db641e5e9ceff8b122cb3bd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:10:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://16f23c4c7163d32dc0df8d6a7ee17e2864fb261f8a40ecf05556774b173fc868\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:10:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:10:55Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:31Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:31 crc kubenswrapper[4988]: I1008 18:11:31.763925 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9590ca0059f0ebbe86343cccb3475338c317382c3231e082655b61e67eece375\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:31Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:31 crc kubenswrapper[4988]: I1008 18:11:31.794356 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:31Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:31 crc kubenswrapper[4988]: I1008 18:11:31.826449 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:11:31 crc kubenswrapper[4988]: I1008 18:11:31.826514 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:11:31 crc kubenswrapper[4988]: I1008 18:11:31.826526 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:11:31 crc kubenswrapper[4988]: I1008 18:11:31.826550 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:11:31 crc kubenswrapper[4988]: I1008 18:11:31.826568 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:11:31Z","lastTransitionTime":"2025-10-08T18:11:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:11:31 crc kubenswrapper[4988]: I1008 18:11:31.827524 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-jnh7m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"556fb048-8afb-44a5-9ea0-2b7c89e6cf4b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bfa2f3382b23c3f788a1b31e80e8a90b3b6868870dccc46d86c3fc11ad027f52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmlt2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1a825886f105ccf9a7cc6f28a0cd012fb859f4990dade0e6aae932482146028\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1a825886f105ccf9a7cc6f28a0cd012fb859f4990dade0e6aae932482146028\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmlt2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ab2fe6e65418efa75bae72259157a2323bbb4b7d11d1d012053524e14135888\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0ab2fe6e65418efa75bae72259157a2323bbb4b7d11d1d012053524e14135888\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:11:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:11:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmlt2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://368a303309be112d9c0b7ccd6538fbaca7085d97a8ec8259f36092f07147e7cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://368a303309be112d9c0b7ccd6538fbaca7085d97a8ec8259f36092f07147e7cb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:11:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:11:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmlt2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a8ca69001dddef643f982a97c74c291e30cb9fac5b0e894e80066b885ea9151\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9a8ca69001dddef643f982a97c74c291e30cb9fac5b0e894e80066b885ea9151\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:11:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:11:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmlt2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://145303c46b48603ed1aae4b9bfe06332a3ffad28f206f8394574011a87b58cd3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://145303c46b48603ed1aae4b9bfe06332a3ffad28f206f8394574011a87b58cd3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:11:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:11:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmlt2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d08e5cfc7cbe5ec1a7a0265ed7c466363c3abeec492263170860689b46c7ff95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d08e5cfc7cbe5ec1a7a0265ed7c466363c3abeec492263170860689b46c7ff95\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:11:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:11:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmlt2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:11:16Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-jnh7m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:31Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:31 crc kubenswrapper[4988]: I1008 18:11:31.838925 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/171e0902-95ee-4a39-8b1b-a58704c30574-metrics-certs\") pod \"network-metrics-daemon-jn2sx\" (UID: \"171e0902-95ee-4a39-8b1b-a58704c30574\") " pod="openshift-multus/network-metrics-daemon-jn2sx" Oct 08 18:11:31 crc kubenswrapper[4988]: E1008 18:11:31.839213 4988 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 08 18:11:31 crc kubenswrapper[4988]: E1008 18:11:31.839351 4988 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/171e0902-95ee-4a39-8b1b-a58704c30574-metrics-certs podName:171e0902-95ee-4a39-8b1b-a58704c30574 nodeName:}" failed. No retries permitted until 2025-10-08 18:11:32.839321655 +0000 UTC m=+38.289164605 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/171e0902-95ee-4a39-8b1b-a58704c30574-metrics-certs") pod "network-metrics-daemon-jn2sx" (UID: "171e0902-95ee-4a39-8b1b-a58704c30574") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 08 18:11:31 crc kubenswrapper[4988]: I1008 18:11:31.843104 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f328a20b-5283-407e-8a93-2ab2da599f18\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8622d56fb198dadee31fb0ca8fc80d56828b6710648055f264533e2d5bdb3037\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z2gh4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3cdb2a757332bdcdc2a2f3362913fe8f06003b0597de20b72627f9badd5e30af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z2gh4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:11:16Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-nm54f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:31Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:31 crc kubenswrapper[4988]: I1008 18:11:31.856526 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a35a5e989ba69c8ef36ce4cea272aadf67ce20473fc257adfcfa86a2faba6ae0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a522e937cfe03fb9f2e60cb1bea07214cf63c69bdecd5ba65a2dfdba882d28dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:31Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:31 crc kubenswrapper[4988]: I1008 18:11:31.869867 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:31Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:31 crc kubenswrapper[4988]: I1008 18:11:31.885078 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-h9jwp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4490d846-ca32-4cd5-9202-87cc6d499fb4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://00b7f6db043c28efda3a09fdc77a6ccded3eae8b6a3d2852aa500789d5f01194\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7k6rk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:11:16Z\\\"}}\" for pod \"openshift-multus\"/\"multus-h9jwp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:31Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:31 crc kubenswrapper[4988]: I1008 18:11:31.904623 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-nsc65" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f010a463-69f2-4082-acc3-161899a939c3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://48a516f6c153743bd1a21ec6945fb368a37f4d2b9d6a326b68edc096a192ed05\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dnwgr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://03406e726f817bcedb841f81959ab12808ba35a530240f9abc5fa5c608b9a6ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dnwgr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9fbbeae0409d2e47e8de9a4306f63b0d13acbfd73cd86f83303fec14352ed718\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dnwgr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://55791327a370c0009c523b86b2234e82d3794985f2eaccf1922eb90f46b2b970\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dnwgr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cee5de93bc68345f9b737857f02832f56582d58757c50a8798eec25d19b2c20d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dnwgr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d314eb248e90371fed068b1173e26d4c9f9765417a70b341e1a5b136ec17353a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dnwgr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://08c1ae9c6aaca1379bd4b347fa26858dc50cbee1ae9cb9751a21351448257e11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://08c1ae9c6aaca1379bd4b347fa26858dc50cbee1ae9cb9751a21351448257e11\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-08T18:11:27Z\\\",\\\"message\\\":\\\"ntil: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1008 18:11:27.764809 6411 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1008 18:11:27.764817 6411 handler.go:208] Removed *v1.Node event handler 2\\\\nI1008 18:11:27.764773 6411 loadbalancer.go:304] Deleted 0 stale LBs for map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-ingress-operator/metrics\\\\\\\"}\\\\nI1008 18:11:27.765603 6411 transact.go:42] Configuring OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-machine-api/control-plane-machine-set-operator]} name:Service_openshift-machine-api/control-plane-machine-set-operator_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.41:9443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {589f95f7-f3e2-4140-80ed-9a0717201481}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1008 18:11:27.765621 6411 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1008 18:11:27.765753 6411 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T18:11:26Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-nsc65_openshift-ovn-kubernetes(f010a463-69f2-4082-acc3-161899a939c3)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dnwgr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://74fcb812b84291763cc6e5ccb20c12aaad580ad968dda4734a0bbfe33d2770ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dnwgr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://befe684cea9bd6dfda3ec8e85bca431c26bd8f82823b0a69dc0d04319a3ca385\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://befe684cea9bd6dfda3ec8e85bca431c26bd8f82823b0a69dc0d04319a3ca385\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dnwgr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:11:16Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-nsc65\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:31Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:31 crc kubenswrapper[4988]: I1008 18:11:31.920646 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-7wl8v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e433488d-db52-435f-b89c-f63cd477c5d3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76b30d183684740b647c1af11dc49712232cc8b0c9a392b491580e5ddb75ea7f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dbz6l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e55016c3e1b97d1b3a857130a5c8e2ec6da2333f786032118a61e3e2ad4c5651\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dbz6l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:11:29Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-7wl8v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:31Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:31 crc kubenswrapper[4988]: I1008 18:11:31.929374 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:11:31 crc kubenswrapper[4988]: I1008 18:11:31.929449 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:11:31 crc kubenswrapper[4988]: I1008 18:11:31.929464 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:11:31 crc kubenswrapper[4988]: I1008 18:11:31.929483 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:11:31 crc kubenswrapper[4988]: I1008 18:11:31.929498 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:11:31Z","lastTransitionTime":"2025-10-08T18:11:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:11:31 crc kubenswrapper[4988]: I1008 18:11:31.933406 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-jn2sx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"171e0902-95ee-4a39-8b1b-a58704c30574\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9f4q2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9f4q2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:11:31Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-jn2sx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:31Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:31 crc kubenswrapper[4988]: I1008 18:11:31.950071 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:31Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:32 crc kubenswrapper[4988]: I1008 18:11:32.033352 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:11:32 crc kubenswrapper[4988]: I1008 18:11:32.033470 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:11:32 crc kubenswrapper[4988]: I1008 18:11:32.033490 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:11:32 crc kubenswrapper[4988]: I1008 18:11:32.033515 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:11:32 crc kubenswrapper[4988]: I1008 18:11:32.033535 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:11:32Z","lastTransitionTime":"2025-10-08T18:11:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:11:32 crc kubenswrapper[4988]: I1008 18:11:32.136989 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:11:32 crc kubenswrapper[4988]: I1008 18:11:32.137046 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:11:32 crc kubenswrapper[4988]: I1008 18:11:32.137062 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:11:32 crc kubenswrapper[4988]: I1008 18:11:32.137093 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:11:32 crc kubenswrapper[4988]: I1008 18:11:32.137105 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:11:32Z","lastTransitionTime":"2025-10-08T18:11:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:11:32 crc kubenswrapper[4988]: I1008 18:11:32.241558 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:11:32 crc kubenswrapper[4988]: I1008 18:11:32.241639 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:11:32 crc kubenswrapper[4988]: I1008 18:11:32.241658 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:11:32 crc kubenswrapper[4988]: I1008 18:11:32.241690 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:11:32 crc kubenswrapper[4988]: I1008 18:11:32.241715 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:11:32Z","lastTransitionTime":"2025-10-08T18:11:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:11:32 crc kubenswrapper[4988]: I1008 18:11:32.345950 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:11:32 crc kubenswrapper[4988]: I1008 18:11:32.346023 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:11:32 crc kubenswrapper[4988]: I1008 18:11:32.346042 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:11:32 crc kubenswrapper[4988]: I1008 18:11:32.346075 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:11:32 crc kubenswrapper[4988]: I1008 18:11:32.346098 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:11:32Z","lastTransitionTime":"2025-10-08T18:11:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:11:32 crc kubenswrapper[4988]: I1008 18:11:32.449489 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:11:32 crc kubenswrapper[4988]: I1008 18:11:32.449553 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:11:32 crc kubenswrapper[4988]: I1008 18:11:32.449567 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:11:32 crc kubenswrapper[4988]: I1008 18:11:32.449586 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:11:32 crc kubenswrapper[4988]: I1008 18:11:32.449598 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:11:32Z","lastTransitionTime":"2025-10-08T18:11:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:11:32 crc kubenswrapper[4988]: I1008 18:11:32.553471 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:11:32 crc kubenswrapper[4988]: I1008 18:11:32.553553 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:11:32 crc kubenswrapper[4988]: I1008 18:11:32.553576 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:11:32 crc kubenswrapper[4988]: I1008 18:11:32.553613 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:11:32 crc kubenswrapper[4988]: I1008 18:11:32.553637 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:11:32Z","lastTransitionTime":"2025-10-08T18:11:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:11:32 crc kubenswrapper[4988]: I1008 18:11:32.656563 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:11:32 crc kubenswrapper[4988]: I1008 18:11:32.656608 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:11:32 crc kubenswrapper[4988]: I1008 18:11:32.656617 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:11:32 crc kubenswrapper[4988]: I1008 18:11:32.656631 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:11:32 crc kubenswrapper[4988]: I1008 18:11:32.656642 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:11:32Z","lastTransitionTime":"2025-10-08T18:11:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:11:32 crc kubenswrapper[4988]: I1008 18:11:32.759886 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:11:32 crc kubenswrapper[4988]: I1008 18:11:32.759960 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:11:32 crc kubenswrapper[4988]: I1008 18:11:32.759989 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:11:32 crc kubenswrapper[4988]: I1008 18:11:32.760024 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:11:32 crc kubenswrapper[4988]: I1008 18:11:32.760081 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:11:32Z","lastTransitionTime":"2025-10-08T18:11:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:11:32 crc kubenswrapper[4988]: I1008 18:11:32.853958 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/171e0902-95ee-4a39-8b1b-a58704c30574-metrics-certs\") pod \"network-metrics-daemon-jn2sx\" (UID: \"171e0902-95ee-4a39-8b1b-a58704c30574\") " pod="openshift-multus/network-metrics-daemon-jn2sx" Oct 08 18:11:32 crc kubenswrapper[4988]: E1008 18:11:32.854232 4988 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 08 18:11:32 crc kubenswrapper[4988]: E1008 18:11:32.854433 4988 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/171e0902-95ee-4a39-8b1b-a58704c30574-metrics-certs podName:171e0902-95ee-4a39-8b1b-a58704c30574 nodeName:}" failed. No retries permitted until 2025-10-08 18:11:34.854356962 +0000 UTC m=+40.304199912 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/171e0902-95ee-4a39-8b1b-a58704c30574-metrics-certs") pod "network-metrics-daemon-jn2sx" (UID: "171e0902-95ee-4a39-8b1b-a58704c30574") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 08 18:11:32 crc kubenswrapper[4988]: I1008 18:11:32.863248 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:11:32 crc kubenswrapper[4988]: I1008 18:11:32.863316 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:11:32 crc kubenswrapper[4988]: I1008 18:11:32.863334 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:11:32 crc kubenswrapper[4988]: I1008 18:11:32.863361 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:11:32 crc kubenswrapper[4988]: I1008 18:11:32.863380 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:11:32Z","lastTransitionTime":"2025-10-08T18:11:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:11:32 crc kubenswrapper[4988]: I1008 18:11:32.966987 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:11:32 crc kubenswrapper[4988]: I1008 18:11:32.967046 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:11:32 crc kubenswrapper[4988]: I1008 18:11:32.967063 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:11:32 crc kubenswrapper[4988]: I1008 18:11:32.967085 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:11:32 crc kubenswrapper[4988]: I1008 18:11:32.967099 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:11:32Z","lastTransitionTime":"2025-10-08T18:11:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:11:33 crc kubenswrapper[4988]: I1008 18:11:33.071833 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:11:33 crc kubenswrapper[4988]: I1008 18:11:33.071897 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:11:33 crc kubenswrapper[4988]: I1008 18:11:33.071918 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:11:33 crc kubenswrapper[4988]: I1008 18:11:33.071947 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:11:33 crc kubenswrapper[4988]: I1008 18:11:33.071966 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:11:33Z","lastTransitionTime":"2025-10-08T18:11:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:11:33 crc kubenswrapper[4988]: I1008 18:11:33.175876 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:11:33 crc kubenswrapper[4988]: I1008 18:11:33.175991 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:11:33 crc kubenswrapper[4988]: I1008 18:11:33.176018 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:11:33 crc kubenswrapper[4988]: I1008 18:11:33.176053 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:11:33 crc kubenswrapper[4988]: I1008 18:11:33.176083 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:11:33Z","lastTransitionTime":"2025-10-08T18:11:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:11:33 crc kubenswrapper[4988]: I1008 18:11:33.237979 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 18:11:33 crc kubenswrapper[4988]: E1008 18:11:33.238218 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 08 18:11:33 crc kubenswrapper[4988]: I1008 18:11:33.238679 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 18:11:33 crc kubenswrapper[4988]: I1008 18:11:33.238832 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 18:11:33 crc kubenswrapper[4988]: E1008 18:11:33.238992 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 08 18:11:33 crc kubenswrapper[4988]: I1008 18:11:33.239054 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jn2sx" Oct 08 18:11:33 crc kubenswrapper[4988]: E1008 18:11:33.239168 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 08 18:11:33 crc kubenswrapper[4988]: E1008 18:11:33.239369 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jn2sx" podUID="171e0902-95ee-4a39-8b1b-a58704c30574" Oct 08 18:11:33 crc kubenswrapper[4988]: I1008 18:11:33.279975 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:11:33 crc kubenswrapper[4988]: I1008 18:11:33.280024 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:11:33 crc kubenswrapper[4988]: I1008 18:11:33.280036 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:11:33 crc kubenswrapper[4988]: I1008 18:11:33.280055 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:11:33 crc kubenswrapper[4988]: I1008 18:11:33.280067 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:11:33Z","lastTransitionTime":"2025-10-08T18:11:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:11:33 crc kubenswrapper[4988]: I1008 18:11:33.383577 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:11:33 crc kubenswrapper[4988]: I1008 18:11:33.383642 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:11:33 crc kubenswrapper[4988]: I1008 18:11:33.383660 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:11:33 crc kubenswrapper[4988]: I1008 18:11:33.383685 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:11:33 crc kubenswrapper[4988]: I1008 18:11:33.383703 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:11:33Z","lastTransitionTime":"2025-10-08T18:11:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:11:33 crc kubenswrapper[4988]: I1008 18:11:33.486314 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:11:33 crc kubenswrapper[4988]: I1008 18:11:33.486582 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:11:33 crc kubenswrapper[4988]: I1008 18:11:33.486599 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:11:33 crc kubenswrapper[4988]: I1008 18:11:33.486627 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:11:33 crc kubenswrapper[4988]: I1008 18:11:33.486642 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:11:33Z","lastTransitionTime":"2025-10-08T18:11:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:11:33 crc kubenswrapper[4988]: I1008 18:11:33.590253 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:11:33 crc kubenswrapper[4988]: I1008 18:11:33.590894 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:11:33 crc kubenswrapper[4988]: I1008 18:11:33.590914 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:11:33 crc kubenswrapper[4988]: I1008 18:11:33.590945 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:11:33 crc kubenswrapper[4988]: I1008 18:11:33.590966 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:11:33Z","lastTransitionTime":"2025-10-08T18:11:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:11:33 crc kubenswrapper[4988]: I1008 18:11:33.694116 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:11:33 crc kubenswrapper[4988]: I1008 18:11:33.694180 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:11:33 crc kubenswrapper[4988]: I1008 18:11:33.694198 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:11:33 crc kubenswrapper[4988]: I1008 18:11:33.694226 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:11:33 crc kubenswrapper[4988]: I1008 18:11:33.694244 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:11:33Z","lastTransitionTime":"2025-10-08T18:11:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:11:33 crc kubenswrapper[4988]: I1008 18:11:33.797305 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:11:33 crc kubenswrapper[4988]: I1008 18:11:33.797378 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:11:33 crc kubenswrapper[4988]: I1008 18:11:33.797447 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:11:33 crc kubenswrapper[4988]: I1008 18:11:33.797480 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:11:33 crc kubenswrapper[4988]: I1008 18:11:33.797504 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:11:33Z","lastTransitionTime":"2025-10-08T18:11:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:11:33 crc kubenswrapper[4988]: I1008 18:11:33.901352 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:11:33 crc kubenswrapper[4988]: I1008 18:11:33.901444 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:11:33 crc kubenswrapper[4988]: I1008 18:11:33.901469 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:11:33 crc kubenswrapper[4988]: I1008 18:11:33.901502 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:11:33 crc kubenswrapper[4988]: I1008 18:11:33.901524 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:11:33Z","lastTransitionTime":"2025-10-08T18:11:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:11:34 crc kubenswrapper[4988]: I1008 18:11:34.005121 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:11:34 crc kubenswrapper[4988]: I1008 18:11:34.005203 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:11:34 crc kubenswrapper[4988]: I1008 18:11:34.005223 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:11:34 crc kubenswrapper[4988]: I1008 18:11:34.005259 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:11:34 crc kubenswrapper[4988]: I1008 18:11:34.005282 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:11:34Z","lastTransitionTime":"2025-10-08T18:11:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:11:34 crc kubenswrapper[4988]: I1008 18:11:34.110126 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:11:34 crc kubenswrapper[4988]: I1008 18:11:34.110493 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:11:34 crc kubenswrapper[4988]: I1008 18:11:34.110544 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:11:34 crc kubenswrapper[4988]: I1008 18:11:34.110586 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:11:34 crc kubenswrapper[4988]: I1008 18:11:34.110611 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:11:34Z","lastTransitionTime":"2025-10-08T18:11:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:11:34 crc kubenswrapper[4988]: I1008 18:11:34.214806 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:11:34 crc kubenswrapper[4988]: I1008 18:11:34.214878 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:11:34 crc kubenswrapper[4988]: I1008 18:11:34.214896 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:11:34 crc kubenswrapper[4988]: I1008 18:11:34.214924 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:11:34 crc kubenswrapper[4988]: I1008 18:11:34.214983 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:11:34Z","lastTransitionTime":"2025-10-08T18:11:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:11:34 crc kubenswrapper[4988]: I1008 18:11:34.319139 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:11:34 crc kubenswrapper[4988]: I1008 18:11:34.319228 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:11:34 crc kubenswrapper[4988]: I1008 18:11:34.319257 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:11:34 crc kubenswrapper[4988]: I1008 18:11:34.319291 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:11:34 crc kubenswrapper[4988]: I1008 18:11:34.319318 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:11:34Z","lastTransitionTime":"2025-10-08T18:11:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:11:34 crc kubenswrapper[4988]: I1008 18:11:34.422202 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:11:34 crc kubenswrapper[4988]: I1008 18:11:34.422251 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:11:34 crc kubenswrapper[4988]: I1008 18:11:34.422260 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:11:34 crc kubenswrapper[4988]: I1008 18:11:34.422279 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:11:34 crc kubenswrapper[4988]: I1008 18:11:34.422290 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:11:34Z","lastTransitionTime":"2025-10-08T18:11:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:11:34 crc kubenswrapper[4988]: I1008 18:11:34.526475 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:11:34 crc kubenswrapper[4988]: I1008 18:11:34.526534 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:11:34 crc kubenswrapper[4988]: I1008 18:11:34.526552 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:11:34 crc kubenswrapper[4988]: I1008 18:11:34.526578 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:11:34 crc kubenswrapper[4988]: I1008 18:11:34.526597 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:11:34Z","lastTransitionTime":"2025-10-08T18:11:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:11:34 crc kubenswrapper[4988]: I1008 18:11:34.630075 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:11:34 crc kubenswrapper[4988]: I1008 18:11:34.630157 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:11:34 crc kubenswrapper[4988]: I1008 18:11:34.630182 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:11:34 crc kubenswrapper[4988]: I1008 18:11:34.630212 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:11:34 crc kubenswrapper[4988]: I1008 18:11:34.630233 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:11:34Z","lastTransitionTime":"2025-10-08T18:11:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:11:34 crc kubenswrapper[4988]: I1008 18:11:34.733912 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:11:34 crc kubenswrapper[4988]: I1008 18:11:34.733960 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:11:34 crc kubenswrapper[4988]: I1008 18:11:34.733977 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:11:34 crc kubenswrapper[4988]: I1008 18:11:34.733995 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:11:34 crc kubenswrapper[4988]: I1008 18:11:34.734010 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:11:34Z","lastTransitionTime":"2025-10-08T18:11:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:11:34 crc kubenswrapper[4988]: I1008 18:11:34.837360 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:11:34 crc kubenswrapper[4988]: I1008 18:11:34.837461 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:11:34 crc kubenswrapper[4988]: I1008 18:11:34.837487 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:11:34 crc kubenswrapper[4988]: I1008 18:11:34.837522 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:11:34 crc kubenswrapper[4988]: I1008 18:11:34.837549 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:11:34Z","lastTransitionTime":"2025-10-08T18:11:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:11:34 crc kubenswrapper[4988]: I1008 18:11:34.880565 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/171e0902-95ee-4a39-8b1b-a58704c30574-metrics-certs\") pod \"network-metrics-daemon-jn2sx\" (UID: \"171e0902-95ee-4a39-8b1b-a58704c30574\") " pod="openshift-multus/network-metrics-daemon-jn2sx" Oct 08 18:11:34 crc kubenswrapper[4988]: E1008 18:11:34.880804 4988 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 08 18:11:34 crc kubenswrapper[4988]: E1008 18:11:34.880882 4988 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/171e0902-95ee-4a39-8b1b-a58704c30574-metrics-certs podName:171e0902-95ee-4a39-8b1b-a58704c30574 nodeName:}" failed. No retries permitted until 2025-10-08 18:11:38.880857674 +0000 UTC m=+44.330700444 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/171e0902-95ee-4a39-8b1b-a58704c30574-metrics-certs") pod "network-metrics-daemon-jn2sx" (UID: "171e0902-95ee-4a39-8b1b-a58704c30574") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 08 18:11:34 crc kubenswrapper[4988]: I1008 18:11:34.940355 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:11:34 crc kubenswrapper[4988]: I1008 18:11:34.940437 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:11:34 crc kubenswrapper[4988]: I1008 18:11:34.940456 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:11:34 crc kubenswrapper[4988]: I1008 18:11:34.940477 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:11:34 crc kubenswrapper[4988]: I1008 18:11:34.940491 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:11:34Z","lastTransitionTime":"2025-10-08T18:11:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:11:35 crc kubenswrapper[4988]: I1008 18:11:35.043941 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:11:35 crc kubenswrapper[4988]: I1008 18:11:35.043992 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:11:35 crc kubenswrapper[4988]: I1008 18:11:35.044002 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:11:35 crc kubenswrapper[4988]: I1008 18:11:35.044020 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:11:35 crc kubenswrapper[4988]: I1008 18:11:35.044032 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:11:35Z","lastTransitionTime":"2025-10-08T18:11:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:11:35 crc kubenswrapper[4988]: I1008 18:11:35.146888 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:11:35 crc kubenswrapper[4988]: I1008 18:11:35.146966 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:11:35 crc kubenswrapper[4988]: I1008 18:11:35.146984 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:11:35 crc kubenswrapper[4988]: I1008 18:11:35.147013 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:11:35 crc kubenswrapper[4988]: I1008 18:11:35.147030 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:11:35Z","lastTransitionTime":"2025-10-08T18:11:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:11:35 crc kubenswrapper[4988]: I1008 18:11:35.237683 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jn2sx" Oct 08 18:11:35 crc kubenswrapper[4988]: E1008 18:11:35.238051 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jn2sx" podUID="171e0902-95ee-4a39-8b1b-a58704c30574" Oct 08 18:11:35 crc kubenswrapper[4988]: I1008 18:11:35.237767 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 18:11:35 crc kubenswrapper[4988]: E1008 18:11:35.238238 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 08 18:11:35 crc kubenswrapper[4988]: I1008 18:11:35.237713 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 18:11:35 crc kubenswrapper[4988]: E1008 18:11:35.238537 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 08 18:11:35 crc kubenswrapper[4988]: I1008 18:11:35.237777 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 18:11:35 crc kubenswrapper[4988]: E1008 18:11:35.238895 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 08 18:11:35 crc kubenswrapper[4988]: I1008 18:11:35.251583 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:11:35 crc kubenswrapper[4988]: I1008 18:11:35.251654 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:11:35 crc kubenswrapper[4988]: I1008 18:11:35.251679 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:11:35 crc kubenswrapper[4988]: I1008 18:11:35.251703 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:11:35 crc kubenswrapper[4988]: I1008 18:11:35.251721 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:11:35Z","lastTransitionTime":"2025-10-08T18:11:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:11:35 crc kubenswrapper[4988]: I1008 18:11:35.261589 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aef71be1-1d12-46e5-95b2-bd399040fd4d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:10:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:10:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:10:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77c68f3ecb8a0724499e495cb26ef3db5beecb140f2558d8c0b25f41928b84ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:10:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a560ee84c2405227de71fda127566213cd4f017e73d5581442ba267781b1c539\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:10:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6100accc3bd08928bc1ff88310f29fa5da37fd2d5c315c94279801c9778c4745\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:10:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b0e596d7ce4e9341c44499af79d3b3caf0e5e2c156902ea40cf2a3dca672be9e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5bd35aef24be034a0b305a40bb53ca171a4f76db3b93b4f78ae95099b0a4cde3\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-08T18:11:14Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1008 18:11:09.091034 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1008 18:11:09.091984 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3472374062/tls.crt::/tmp/serving-cert-3472374062/tls.key\\\\\\\"\\\\nI1008 18:11:14.619696 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1008 18:11:14.628837 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1008 18:11:14.628898 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1008 18:11:14.628994 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1008 18:11:14.629036 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1008 18:11:14.641318 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1008 18:11:14.641363 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1008 18:11:14.641372 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1008 18:11:14.641382 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1008 18:11:14.641415 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1008 18:11:14.641423 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1008 18:11:14.641431 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1008 18:11:14.641484 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1008 18:11:14.646451 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T18:10:58Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c79ef704fda1bbcc486f4c7afa3039a3ab38669833f0f66bafe24dd4cc679e60\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:10:58Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://28d21731af0f16d6a8ae6c6e9d0bebf0560d4a533db4c78c10bf1837890394ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://28d21731af0f16d6a8ae6c6e9d0bebf0560d4a533db4c78c10bf1837890394ef\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:10:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:10:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:10:55Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:35Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:35 crc kubenswrapper[4988]: I1008 18:11:35.277934 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7e0ecefc-359f-4785-bb79-0343a62a25da\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:10:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:10:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:10:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://30bec350fefee3d19758b3ed735ede4e9aa085edb57aa3f3ec24249a67dc7ca2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:10:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2d850fd27d4ed65891ff69b9add65fb78535357f54819c2a8d196f5a340c58e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:10:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://90b1a9e14c82bf0f75f08fdb342eed35a503020f9db641e5e9ceff8b122cb3bd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:10:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://16f23c4c7163d32dc0df8d6a7ee17e2864fb261f8a40ecf05556774b173fc868\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:10:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:10:55Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:35Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:35 crc kubenswrapper[4988]: I1008 18:11:35.294943 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e8906182d55c285b3cf095e2d9e1dac397af4618a514f11e26b44c172dd573f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:35Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:35 crc kubenswrapper[4988]: I1008 18:11:35.309983 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-dzxs2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8f307b20-5e33-4725-99cb-210e25113532\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72ee803c6ef88524e0ce0391476f8e5a7dfbc858b095a3c2ae31bad10991ce3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q65f7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:11:16Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-dzxs2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:35Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:35 crc kubenswrapper[4988]: I1008 18:11:35.333340 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-jnh7m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"556fb048-8afb-44a5-9ea0-2b7c89e6cf4b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bfa2f3382b23c3f788a1b31e80e8a90b3b6868870dccc46d86c3fc11ad027f52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmlt2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1a825886f105ccf9a7cc6f28a0cd012fb859f4990dade0e6aae932482146028\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1a825886f105ccf9a7cc6f28a0cd012fb859f4990dade0e6aae932482146028\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmlt2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ab2fe6e65418efa75bae72259157a2323bbb4b7d11d1d012053524e14135888\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0ab2fe6e65418efa75bae72259157a2323bbb4b7d11d1d012053524e14135888\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:11:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:11:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmlt2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://368a303309be112d9c0b7ccd6538fbaca7085d97a8ec8259f36092f07147e7cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://368a303309be112d9c0b7ccd6538fbaca7085d97a8ec8259f36092f07147e7cb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:11:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:11:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmlt2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a8ca69001dddef643f982a97c74c291e30cb9fac5b0e894e80066b885ea9151\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9a8ca69001dddef643f982a97c74c291e30cb9fac5b0e894e80066b885ea9151\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:11:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:11:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmlt2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://145303c46b48603ed1aae4b9bfe06332a3ffad28f206f8394574011a87b58cd3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://145303c46b48603ed1aae4b9bfe06332a3ffad28f206f8394574011a87b58cd3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:11:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:11:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmlt2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d08e5cfc7cbe5ec1a7a0265ed7c466363c3abeec492263170860689b46c7ff95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d08e5cfc7cbe5ec1a7a0265ed7c466363c3abeec492263170860689b46c7ff95\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:11:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:11:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmlt2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:11:16Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-jnh7m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:35Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:35 crc kubenswrapper[4988]: I1008 18:11:35.351514 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f328a20b-5283-407e-8a93-2ab2da599f18\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8622d56fb198dadee31fb0ca8fc80d56828b6710648055f264533e2d5bdb3037\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z2gh4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3cdb2a757332bdcdc2a2f3362913fe8f06003b0597de20b72627f9badd5e30af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z2gh4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:11:16Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-nm54f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:35Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:35 crc kubenswrapper[4988]: I1008 18:11:35.354886 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:11:35 crc kubenswrapper[4988]: I1008 18:11:35.354968 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:11:35 crc kubenswrapper[4988]: I1008 18:11:35.354989 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:11:35 crc kubenswrapper[4988]: I1008 18:11:35.355018 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:11:35 crc kubenswrapper[4988]: I1008 18:11:35.355038 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:11:35Z","lastTransitionTime":"2025-10-08T18:11:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:11:35 crc kubenswrapper[4988]: I1008 18:11:35.369066 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a35a5e989ba69c8ef36ce4cea272aadf67ce20473fc257adfcfa86a2faba6ae0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a522e937cfe03fb9f2e60cb1bea07214cf63c69bdecd5ba65a2dfdba882d28dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:35Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:35 crc kubenswrapper[4988]: I1008 18:11:35.384256 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9590ca0059f0ebbe86343cccb3475338c317382c3231e082655b61e67eece375\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:35Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:35 crc kubenswrapper[4988]: I1008 18:11:35.401090 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:35Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:35 crc kubenswrapper[4988]: I1008 18:11:35.432857 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-nsc65" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f010a463-69f2-4082-acc3-161899a939c3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://48a516f6c153743bd1a21ec6945fb368a37f4d2b9d6a326b68edc096a192ed05\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dnwgr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://03406e726f817bcedb841f81959ab12808ba35a530240f9abc5fa5c608b9a6ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dnwgr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9fbbeae0409d2e47e8de9a4306f63b0d13acbfd73cd86f83303fec14352ed718\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dnwgr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://55791327a370c0009c523b86b2234e82d3794985f2eaccf1922eb90f46b2b970\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dnwgr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cee5de93bc68345f9b737857f02832f56582d58757c50a8798eec25d19b2c20d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dnwgr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d314eb248e90371fed068b1173e26d4c9f9765417a70b341e1a5b136ec17353a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dnwgr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://08c1ae9c6aaca1379bd4b347fa26858dc50cbee1ae9cb9751a21351448257e11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://08c1ae9c6aaca1379bd4b347fa26858dc50cbee1ae9cb9751a21351448257e11\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-08T18:11:27Z\\\",\\\"message\\\":\\\"ntil: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1008 18:11:27.764809 6411 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1008 18:11:27.764817 6411 handler.go:208] Removed *v1.Node event handler 2\\\\nI1008 18:11:27.764773 6411 loadbalancer.go:304] Deleted 0 stale LBs for map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-ingress-operator/metrics\\\\\\\"}\\\\nI1008 18:11:27.765603 6411 transact.go:42] Configuring OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-machine-api/control-plane-machine-set-operator]} name:Service_openshift-machine-api/control-plane-machine-set-operator_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.41:9443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {589f95f7-f3e2-4140-80ed-9a0717201481}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1008 18:11:27.765621 6411 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1008 18:11:27.765753 6411 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T18:11:26Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-nsc65_openshift-ovn-kubernetes(f010a463-69f2-4082-acc3-161899a939c3)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dnwgr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://74fcb812b84291763cc6e5ccb20c12aaad580ad968dda4734a0bbfe33d2770ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dnwgr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://befe684cea9bd6dfda3ec8e85bca431c26bd8f82823b0a69dc0d04319a3ca385\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://befe684cea9bd6dfda3ec8e85bca431c26bd8f82823b0a69dc0d04319a3ca385\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dnwgr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:11:16Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-nsc65\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:35Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:35 crc kubenswrapper[4988]: I1008 18:11:35.451066 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-7wl8v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e433488d-db52-435f-b89c-f63cd477c5d3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76b30d183684740b647c1af11dc49712232cc8b0c9a392b491580e5ddb75ea7f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dbz6l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e55016c3e1b97d1b3a857130a5c8e2ec6da2333f786032118a61e3e2ad4c5651\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dbz6l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:11:29Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-7wl8v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:35Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:35 crc kubenswrapper[4988]: I1008 18:11:35.458599 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:11:35 crc kubenswrapper[4988]: I1008 18:11:35.458637 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:11:35 crc kubenswrapper[4988]: I1008 18:11:35.458651 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:11:35 crc kubenswrapper[4988]: I1008 18:11:35.458674 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:11:35 crc kubenswrapper[4988]: I1008 18:11:35.458688 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:11:35Z","lastTransitionTime":"2025-10-08T18:11:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:11:35 crc kubenswrapper[4988]: I1008 18:11:35.465688 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-jn2sx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"171e0902-95ee-4a39-8b1b-a58704c30574\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9f4q2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9f4q2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:11:31Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-jn2sx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:35Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:35 crc kubenswrapper[4988]: I1008 18:11:35.480424 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:35Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:35 crc kubenswrapper[4988]: I1008 18:11:35.501274 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:35Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:35 crc kubenswrapper[4988]: I1008 18:11:35.521539 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-h9jwp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4490d846-ca32-4cd5-9202-87cc6d499fb4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://00b7f6db043c28efda3a09fdc77a6ccded3eae8b6a3d2852aa500789d5f01194\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7k6rk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:11:16Z\\\"}}\" for pod \"openshift-multus\"/\"multus-h9jwp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:35Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:35 crc kubenswrapper[4988]: I1008 18:11:35.536816 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-s4f2k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ba5e11d7-7bde-4c3d-983b-884232219652\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://824bc34a2807068d3fceb10651803ae9ae5651e940f7aed4877c544d76dae4d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hn5jf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:11:15Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-s4f2k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:35Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:35 crc kubenswrapper[4988]: I1008 18:11:35.563109 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:11:35 crc kubenswrapper[4988]: I1008 18:11:35.563184 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:11:35 crc kubenswrapper[4988]: I1008 18:11:35.563202 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:11:35 crc kubenswrapper[4988]: I1008 18:11:35.563234 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:11:35 crc kubenswrapper[4988]: I1008 18:11:35.563253 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:11:35Z","lastTransitionTime":"2025-10-08T18:11:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:11:35 crc kubenswrapper[4988]: I1008 18:11:35.666737 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:11:35 crc kubenswrapper[4988]: I1008 18:11:35.666802 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:11:35 crc kubenswrapper[4988]: I1008 18:11:35.666818 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:11:35 crc kubenswrapper[4988]: I1008 18:11:35.666846 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:11:35 crc kubenswrapper[4988]: I1008 18:11:35.666864 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:11:35Z","lastTransitionTime":"2025-10-08T18:11:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:11:35 crc kubenswrapper[4988]: I1008 18:11:35.770515 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:11:35 crc kubenswrapper[4988]: I1008 18:11:35.770900 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:11:35 crc kubenswrapper[4988]: I1008 18:11:35.771572 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:11:35 crc kubenswrapper[4988]: I1008 18:11:35.771813 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:11:35 crc kubenswrapper[4988]: I1008 18:11:35.771985 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:11:35Z","lastTransitionTime":"2025-10-08T18:11:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:11:35 crc kubenswrapper[4988]: I1008 18:11:35.876628 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:11:35 crc kubenswrapper[4988]: I1008 18:11:35.876699 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:11:35 crc kubenswrapper[4988]: I1008 18:11:35.876718 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:11:35 crc kubenswrapper[4988]: I1008 18:11:35.876747 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:11:35 crc kubenswrapper[4988]: I1008 18:11:35.876766 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:11:35Z","lastTransitionTime":"2025-10-08T18:11:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:11:35 crc kubenswrapper[4988]: I1008 18:11:35.979505 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:11:35 crc kubenswrapper[4988]: I1008 18:11:35.979548 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:11:35 crc kubenswrapper[4988]: I1008 18:11:35.979559 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:11:35 crc kubenswrapper[4988]: I1008 18:11:35.979580 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:11:35 crc kubenswrapper[4988]: I1008 18:11:35.979592 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:11:35Z","lastTransitionTime":"2025-10-08T18:11:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:11:36 crc kubenswrapper[4988]: I1008 18:11:36.082703 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:11:36 crc kubenswrapper[4988]: I1008 18:11:36.082768 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:11:36 crc kubenswrapper[4988]: I1008 18:11:36.082780 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:11:36 crc kubenswrapper[4988]: I1008 18:11:36.082815 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:11:36 crc kubenswrapper[4988]: I1008 18:11:36.082828 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:11:36Z","lastTransitionTime":"2025-10-08T18:11:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:11:36 crc kubenswrapper[4988]: I1008 18:11:36.186542 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:11:36 crc kubenswrapper[4988]: I1008 18:11:36.186621 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:11:36 crc kubenswrapper[4988]: I1008 18:11:36.186639 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:11:36 crc kubenswrapper[4988]: I1008 18:11:36.186666 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:11:36 crc kubenswrapper[4988]: I1008 18:11:36.186685 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:11:36Z","lastTransitionTime":"2025-10-08T18:11:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:11:36 crc kubenswrapper[4988]: I1008 18:11:36.290312 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:11:36 crc kubenswrapper[4988]: I1008 18:11:36.290423 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:11:36 crc kubenswrapper[4988]: I1008 18:11:36.290441 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:11:36 crc kubenswrapper[4988]: I1008 18:11:36.290466 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:11:36 crc kubenswrapper[4988]: I1008 18:11:36.290487 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:11:36Z","lastTransitionTime":"2025-10-08T18:11:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:11:36 crc kubenswrapper[4988]: I1008 18:11:36.394096 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:11:36 crc kubenswrapper[4988]: I1008 18:11:36.394172 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:11:36 crc kubenswrapper[4988]: I1008 18:11:36.394191 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:11:36 crc kubenswrapper[4988]: I1008 18:11:36.394221 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:11:36 crc kubenswrapper[4988]: I1008 18:11:36.394241 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:11:36Z","lastTransitionTime":"2025-10-08T18:11:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:11:36 crc kubenswrapper[4988]: I1008 18:11:36.497411 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:11:36 crc kubenswrapper[4988]: I1008 18:11:36.497473 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:11:36 crc kubenswrapper[4988]: I1008 18:11:36.497488 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:11:36 crc kubenswrapper[4988]: I1008 18:11:36.497511 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:11:36 crc kubenswrapper[4988]: I1008 18:11:36.497528 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:11:36Z","lastTransitionTime":"2025-10-08T18:11:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:11:36 crc kubenswrapper[4988]: I1008 18:11:36.601332 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:11:36 crc kubenswrapper[4988]: I1008 18:11:36.601450 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:11:36 crc kubenswrapper[4988]: I1008 18:11:36.601470 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:11:36 crc kubenswrapper[4988]: I1008 18:11:36.601497 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:11:36 crc kubenswrapper[4988]: I1008 18:11:36.601514 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:11:36Z","lastTransitionTime":"2025-10-08T18:11:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:11:36 crc kubenswrapper[4988]: I1008 18:11:36.705262 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:11:36 crc kubenswrapper[4988]: I1008 18:11:36.705313 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:11:36 crc kubenswrapper[4988]: I1008 18:11:36.705327 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:11:36 crc kubenswrapper[4988]: I1008 18:11:36.705346 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:11:36 crc kubenswrapper[4988]: I1008 18:11:36.705363 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:11:36Z","lastTransitionTime":"2025-10-08T18:11:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:11:36 crc kubenswrapper[4988]: I1008 18:11:36.808875 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:11:36 crc kubenswrapper[4988]: I1008 18:11:36.808937 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:11:36 crc kubenswrapper[4988]: I1008 18:11:36.808957 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:11:36 crc kubenswrapper[4988]: I1008 18:11:36.808984 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:11:36 crc kubenswrapper[4988]: I1008 18:11:36.809002 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:11:36Z","lastTransitionTime":"2025-10-08T18:11:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:11:36 crc kubenswrapper[4988]: I1008 18:11:36.912109 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:11:36 crc kubenswrapper[4988]: I1008 18:11:36.912254 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:11:36 crc kubenswrapper[4988]: I1008 18:11:36.912277 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:11:36 crc kubenswrapper[4988]: I1008 18:11:36.912305 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:11:36 crc kubenswrapper[4988]: I1008 18:11:36.912325 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:11:36Z","lastTransitionTime":"2025-10-08T18:11:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:11:37 crc kubenswrapper[4988]: I1008 18:11:37.014939 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:11:37 crc kubenswrapper[4988]: I1008 18:11:37.015018 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:11:37 crc kubenswrapper[4988]: I1008 18:11:37.015045 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:11:37 crc kubenswrapper[4988]: I1008 18:11:37.015081 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:11:37 crc kubenswrapper[4988]: I1008 18:11:37.015108 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:11:37Z","lastTransitionTime":"2025-10-08T18:11:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:11:37 crc kubenswrapper[4988]: I1008 18:11:37.120295 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:11:37 crc kubenswrapper[4988]: I1008 18:11:37.120426 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:11:37 crc kubenswrapper[4988]: I1008 18:11:37.120451 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:11:37 crc kubenswrapper[4988]: I1008 18:11:37.120486 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:11:37 crc kubenswrapper[4988]: I1008 18:11:37.120510 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:11:37Z","lastTransitionTime":"2025-10-08T18:11:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:11:37 crc kubenswrapper[4988]: I1008 18:11:37.223036 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:11:37 crc kubenswrapper[4988]: I1008 18:11:37.223089 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:11:37 crc kubenswrapper[4988]: I1008 18:11:37.223101 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:11:37 crc kubenswrapper[4988]: I1008 18:11:37.223119 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:11:37 crc kubenswrapper[4988]: I1008 18:11:37.223133 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:11:37Z","lastTransitionTime":"2025-10-08T18:11:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:11:37 crc kubenswrapper[4988]: I1008 18:11:37.237052 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 18:11:37 crc kubenswrapper[4988]: I1008 18:11:37.237113 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 18:11:37 crc kubenswrapper[4988]: I1008 18:11:37.237053 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 18:11:37 crc kubenswrapper[4988]: I1008 18:11:37.237066 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jn2sx" Oct 08 18:11:37 crc kubenswrapper[4988]: E1008 18:11:37.237214 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 08 18:11:37 crc kubenswrapper[4988]: E1008 18:11:37.237319 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 08 18:11:37 crc kubenswrapper[4988]: E1008 18:11:37.237491 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jn2sx" podUID="171e0902-95ee-4a39-8b1b-a58704c30574" Oct 08 18:11:37 crc kubenswrapper[4988]: E1008 18:11:37.237682 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 08 18:11:37 crc kubenswrapper[4988]: I1008 18:11:37.326104 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:11:37 crc kubenswrapper[4988]: I1008 18:11:37.326195 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:11:37 crc kubenswrapper[4988]: I1008 18:11:37.326220 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:11:37 crc kubenswrapper[4988]: I1008 18:11:37.326253 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:11:37 crc kubenswrapper[4988]: I1008 18:11:37.326276 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:11:37Z","lastTransitionTime":"2025-10-08T18:11:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:11:37 crc kubenswrapper[4988]: I1008 18:11:37.429522 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:11:37 crc kubenswrapper[4988]: I1008 18:11:37.429583 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:11:37 crc kubenswrapper[4988]: I1008 18:11:37.429600 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:11:37 crc kubenswrapper[4988]: I1008 18:11:37.429626 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:11:37 crc kubenswrapper[4988]: I1008 18:11:37.429663 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:11:37Z","lastTransitionTime":"2025-10-08T18:11:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:11:37 crc kubenswrapper[4988]: I1008 18:11:37.532886 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:11:37 crc kubenswrapper[4988]: I1008 18:11:37.532970 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:11:37 crc kubenswrapper[4988]: I1008 18:11:37.532994 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:11:37 crc kubenswrapper[4988]: I1008 18:11:37.533028 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:11:37 crc kubenswrapper[4988]: I1008 18:11:37.533051 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:11:37Z","lastTransitionTime":"2025-10-08T18:11:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:11:37 crc kubenswrapper[4988]: I1008 18:11:37.635373 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:11:37 crc kubenswrapper[4988]: I1008 18:11:37.636413 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:11:37 crc kubenswrapper[4988]: I1008 18:11:37.636638 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:11:37 crc kubenswrapper[4988]: I1008 18:11:37.636878 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:11:37 crc kubenswrapper[4988]: I1008 18:11:37.637058 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:11:37Z","lastTransitionTime":"2025-10-08T18:11:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:11:37 crc kubenswrapper[4988]: I1008 18:11:37.740823 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:11:37 crc kubenswrapper[4988]: I1008 18:11:37.740880 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:11:37 crc kubenswrapper[4988]: I1008 18:11:37.740929 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:11:37 crc kubenswrapper[4988]: I1008 18:11:37.740956 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:11:37 crc kubenswrapper[4988]: I1008 18:11:37.740975 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:11:37Z","lastTransitionTime":"2025-10-08T18:11:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:11:37 crc kubenswrapper[4988]: I1008 18:11:37.844094 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:11:37 crc kubenswrapper[4988]: I1008 18:11:37.844134 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:11:37 crc kubenswrapper[4988]: I1008 18:11:37.844144 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:11:37 crc kubenswrapper[4988]: I1008 18:11:37.844159 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:11:37 crc kubenswrapper[4988]: I1008 18:11:37.844693 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:11:37Z","lastTransitionTime":"2025-10-08T18:11:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:11:37 crc kubenswrapper[4988]: I1008 18:11:37.951992 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:11:37 crc kubenswrapper[4988]: I1008 18:11:37.952072 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:11:37 crc kubenswrapper[4988]: I1008 18:11:37.952090 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:11:37 crc kubenswrapper[4988]: I1008 18:11:37.952117 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:11:37 crc kubenswrapper[4988]: I1008 18:11:37.952137 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:11:37Z","lastTransitionTime":"2025-10-08T18:11:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:11:38 crc kubenswrapper[4988]: I1008 18:11:38.056023 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:11:38 crc kubenswrapper[4988]: I1008 18:11:38.056087 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:11:38 crc kubenswrapper[4988]: I1008 18:11:38.056111 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:11:38 crc kubenswrapper[4988]: I1008 18:11:38.056146 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:11:38 crc kubenswrapper[4988]: I1008 18:11:38.056171 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:11:38Z","lastTransitionTime":"2025-10-08T18:11:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:11:38 crc kubenswrapper[4988]: I1008 18:11:38.159759 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:11:38 crc kubenswrapper[4988]: I1008 18:11:38.159833 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:11:38 crc kubenswrapper[4988]: I1008 18:11:38.159859 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:11:38 crc kubenswrapper[4988]: I1008 18:11:38.159893 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:11:38 crc kubenswrapper[4988]: I1008 18:11:38.159917 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:11:38Z","lastTransitionTime":"2025-10-08T18:11:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:11:38 crc kubenswrapper[4988]: I1008 18:11:38.262735 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:11:38 crc kubenswrapper[4988]: I1008 18:11:38.263026 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:11:38 crc kubenswrapper[4988]: I1008 18:11:38.263227 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:11:38 crc kubenswrapper[4988]: I1008 18:11:38.263546 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:11:38 crc kubenswrapper[4988]: I1008 18:11:38.263780 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:11:38Z","lastTransitionTime":"2025-10-08T18:11:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:11:38 crc kubenswrapper[4988]: I1008 18:11:38.367559 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:11:38 crc kubenswrapper[4988]: I1008 18:11:38.367944 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:11:38 crc kubenswrapper[4988]: I1008 18:11:38.368135 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:11:38 crc kubenswrapper[4988]: I1008 18:11:38.368288 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:11:38 crc kubenswrapper[4988]: I1008 18:11:38.368455 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:11:38Z","lastTransitionTime":"2025-10-08T18:11:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:11:38 crc kubenswrapper[4988]: I1008 18:11:38.472313 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:11:38 crc kubenswrapper[4988]: I1008 18:11:38.472381 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:11:38 crc kubenswrapper[4988]: I1008 18:11:38.472423 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:11:38 crc kubenswrapper[4988]: I1008 18:11:38.472446 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:11:38 crc kubenswrapper[4988]: I1008 18:11:38.472463 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:11:38Z","lastTransitionTime":"2025-10-08T18:11:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:11:38 crc kubenswrapper[4988]: I1008 18:11:38.575583 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:11:38 crc kubenswrapper[4988]: I1008 18:11:38.575953 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:11:38 crc kubenswrapper[4988]: I1008 18:11:38.576059 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:11:38 crc kubenswrapper[4988]: I1008 18:11:38.576148 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:11:38 crc kubenswrapper[4988]: I1008 18:11:38.576268 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:11:38Z","lastTransitionTime":"2025-10-08T18:11:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:11:38 crc kubenswrapper[4988]: I1008 18:11:38.679555 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:11:38 crc kubenswrapper[4988]: I1008 18:11:38.679640 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:11:38 crc kubenswrapper[4988]: I1008 18:11:38.679655 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:11:38 crc kubenswrapper[4988]: I1008 18:11:38.679675 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:11:38 crc kubenswrapper[4988]: I1008 18:11:38.679689 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:11:38Z","lastTransitionTime":"2025-10-08T18:11:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:11:38 crc kubenswrapper[4988]: I1008 18:11:38.783481 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:11:38 crc kubenswrapper[4988]: I1008 18:11:38.783535 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:11:38 crc kubenswrapper[4988]: I1008 18:11:38.783547 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:11:38 crc kubenswrapper[4988]: I1008 18:11:38.783570 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:11:38 crc kubenswrapper[4988]: I1008 18:11:38.783586 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:11:38Z","lastTransitionTime":"2025-10-08T18:11:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:11:38 crc kubenswrapper[4988]: I1008 18:11:38.886561 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:11:38 crc kubenswrapper[4988]: I1008 18:11:38.887002 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:11:38 crc kubenswrapper[4988]: I1008 18:11:38.887152 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:11:38 crc kubenswrapper[4988]: I1008 18:11:38.887308 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:11:38 crc kubenswrapper[4988]: I1008 18:11:38.887489 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:11:38Z","lastTransitionTime":"2025-10-08T18:11:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:11:38 crc kubenswrapper[4988]: I1008 18:11:38.921492 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/171e0902-95ee-4a39-8b1b-a58704c30574-metrics-certs\") pod \"network-metrics-daemon-jn2sx\" (UID: \"171e0902-95ee-4a39-8b1b-a58704c30574\") " pod="openshift-multus/network-metrics-daemon-jn2sx" Oct 08 18:11:38 crc kubenswrapper[4988]: E1008 18:11:38.921752 4988 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 08 18:11:38 crc kubenswrapper[4988]: E1008 18:11:38.922140 4988 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/171e0902-95ee-4a39-8b1b-a58704c30574-metrics-certs podName:171e0902-95ee-4a39-8b1b-a58704c30574 nodeName:}" failed. No retries permitted until 2025-10-08 18:11:46.922113834 +0000 UTC m=+52.371956604 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/171e0902-95ee-4a39-8b1b-a58704c30574-metrics-certs") pod "network-metrics-daemon-jn2sx" (UID: "171e0902-95ee-4a39-8b1b-a58704c30574") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 08 18:11:38 crc kubenswrapper[4988]: I1008 18:11:38.990007 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:11:38 crc kubenswrapper[4988]: I1008 18:11:38.990149 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:11:38 crc kubenswrapper[4988]: I1008 18:11:38.990174 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:11:38 crc kubenswrapper[4988]: I1008 18:11:38.990209 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:11:38 crc kubenswrapper[4988]: I1008 18:11:38.990233 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:11:38Z","lastTransitionTime":"2025-10-08T18:11:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:11:39 crc kubenswrapper[4988]: I1008 18:11:39.093849 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:11:39 crc kubenswrapper[4988]: I1008 18:11:39.094311 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:11:39 crc kubenswrapper[4988]: I1008 18:11:39.094449 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:11:39 crc kubenswrapper[4988]: I1008 18:11:39.094575 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:11:39 crc kubenswrapper[4988]: I1008 18:11:39.094666 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:11:39Z","lastTransitionTime":"2025-10-08T18:11:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:11:39 crc kubenswrapper[4988]: I1008 18:11:39.198921 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:11:39 crc kubenswrapper[4988]: I1008 18:11:39.199102 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:11:39 crc kubenswrapper[4988]: I1008 18:11:39.199121 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:11:39 crc kubenswrapper[4988]: I1008 18:11:39.199153 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:11:39 crc kubenswrapper[4988]: I1008 18:11:39.199174 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:11:39Z","lastTransitionTime":"2025-10-08T18:11:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:11:39 crc kubenswrapper[4988]: I1008 18:11:39.237346 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 18:11:39 crc kubenswrapper[4988]: I1008 18:11:39.237955 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 18:11:39 crc kubenswrapper[4988]: E1008 18:11:39.238374 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 08 18:11:39 crc kubenswrapper[4988]: I1008 18:11:39.238531 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 18:11:39 crc kubenswrapper[4988]: E1008 18:11:39.238651 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 08 18:11:39 crc kubenswrapper[4988]: E1008 18:11:39.238687 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 08 18:11:39 crc kubenswrapper[4988]: I1008 18:11:39.238882 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jn2sx" Oct 08 18:11:39 crc kubenswrapper[4988]: E1008 18:11:39.239131 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jn2sx" podUID="171e0902-95ee-4a39-8b1b-a58704c30574" Oct 08 18:11:39 crc kubenswrapper[4988]: I1008 18:11:39.303106 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:11:39 crc kubenswrapper[4988]: I1008 18:11:39.303579 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:11:39 crc kubenswrapper[4988]: I1008 18:11:39.303770 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:11:39 crc kubenswrapper[4988]: I1008 18:11:39.303994 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:11:39 crc kubenswrapper[4988]: I1008 18:11:39.304197 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:11:39Z","lastTransitionTime":"2025-10-08T18:11:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:11:39 crc kubenswrapper[4988]: I1008 18:11:39.408776 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:11:39 crc kubenswrapper[4988]: I1008 18:11:39.408848 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:11:39 crc kubenswrapper[4988]: I1008 18:11:39.408865 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:11:39 crc kubenswrapper[4988]: I1008 18:11:39.408889 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:11:39 crc kubenswrapper[4988]: I1008 18:11:39.408907 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:11:39Z","lastTransitionTime":"2025-10-08T18:11:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:11:39 crc kubenswrapper[4988]: I1008 18:11:39.512586 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:11:39 crc kubenswrapper[4988]: I1008 18:11:39.512688 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:11:39 crc kubenswrapper[4988]: I1008 18:11:39.512699 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:11:39 crc kubenswrapper[4988]: I1008 18:11:39.512723 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:11:39 crc kubenswrapper[4988]: I1008 18:11:39.512738 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:11:39Z","lastTransitionTime":"2025-10-08T18:11:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:11:39 crc kubenswrapper[4988]: I1008 18:11:39.616069 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:11:39 crc kubenswrapper[4988]: I1008 18:11:39.616559 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:11:39 crc kubenswrapper[4988]: I1008 18:11:39.616708 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:11:39 crc kubenswrapper[4988]: I1008 18:11:39.616839 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:11:39 crc kubenswrapper[4988]: I1008 18:11:39.616954 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:11:39Z","lastTransitionTime":"2025-10-08T18:11:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:11:39 crc kubenswrapper[4988]: I1008 18:11:39.719929 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:11:39 crc kubenswrapper[4988]: I1008 18:11:39.720607 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:11:39 crc kubenswrapper[4988]: I1008 18:11:39.720661 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:11:39 crc kubenswrapper[4988]: I1008 18:11:39.720683 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:11:39 crc kubenswrapper[4988]: I1008 18:11:39.720696 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:11:39Z","lastTransitionTime":"2025-10-08T18:11:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:11:39 crc kubenswrapper[4988]: I1008 18:11:39.824068 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:11:39 crc kubenswrapper[4988]: I1008 18:11:39.824127 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:11:39 crc kubenswrapper[4988]: I1008 18:11:39.824150 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:11:39 crc kubenswrapper[4988]: I1008 18:11:39.824170 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:11:39 crc kubenswrapper[4988]: I1008 18:11:39.824183 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:11:39Z","lastTransitionTime":"2025-10-08T18:11:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:11:39 crc kubenswrapper[4988]: I1008 18:11:39.927110 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:11:39 crc kubenswrapper[4988]: I1008 18:11:39.927145 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:11:39 crc kubenswrapper[4988]: I1008 18:11:39.927153 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:11:39 crc kubenswrapper[4988]: I1008 18:11:39.927169 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:11:39 crc kubenswrapper[4988]: I1008 18:11:39.927181 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:11:39Z","lastTransitionTime":"2025-10-08T18:11:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:11:40 crc kubenswrapper[4988]: I1008 18:11:40.030278 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:11:40 crc kubenswrapper[4988]: I1008 18:11:40.030338 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:11:40 crc kubenswrapper[4988]: I1008 18:11:40.030347 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:11:40 crc kubenswrapper[4988]: I1008 18:11:40.030365 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:11:40 crc kubenswrapper[4988]: I1008 18:11:40.030377 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:11:40Z","lastTransitionTime":"2025-10-08T18:11:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:11:40 crc kubenswrapper[4988]: I1008 18:11:40.133768 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:11:40 crc kubenswrapper[4988]: I1008 18:11:40.133845 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:11:40 crc kubenswrapper[4988]: I1008 18:11:40.133868 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:11:40 crc kubenswrapper[4988]: I1008 18:11:40.133904 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:11:40 crc kubenswrapper[4988]: I1008 18:11:40.133933 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:11:40Z","lastTransitionTime":"2025-10-08T18:11:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:11:40 crc kubenswrapper[4988]: I1008 18:11:40.238582 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:11:40 crc kubenswrapper[4988]: I1008 18:11:40.238667 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:11:40 crc kubenswrapper[4988]: I1008 18:11:40.238686 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:11:40 crc kubenswrapper[4988]: I1008 18:11:40.238715 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:11:40 crc kubenswrapper[4988]: I1008 18:11:40.238739 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:11:40Z","lastTransitionTime":"2025-10-08T18:11:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:11:40 crc kubenswrapper[4988]: I1008 18:11:40.341465 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:11:40 crc kubenswrapper[4988]: I1008 18:11:40.341562 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:11:40 crc kubenswrapper[4988]: I1008 18:11:40.341586 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:11:40 crc kubenswrapper[4988]: I1008 18:11:40.341622 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:11:40 crc kubenswrapper[4988]: I1008 18:11:40.341642 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:11:40Z","lastTransitionTime":"2025-10-08T18:11:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:11:40 crc kubenswrapper[4988]: I1008 18:11:40.445239 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:11:40 crc kubenswrapper[4988]: I1008 18:11:40.445677 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:11:40 crc kubenswrapper[4988]: I1008 18:11:40.445847 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:11:40 crc kubenswrapper[4988]: I1008 18:11:40.446055 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:11:40 crc kubenswrapper[4988]: I1008 18:11:40.446193 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:11:40Z","lastTransitionTime":"2025-10-08T18:11:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:11:40 crc kubenswrapper[4988]: I1008 18:11:40.549567 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:11:40 crc kubenswrapper[4988]: I1008 18:11:40.549616 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:11:40 crc kubenswrapper[4988]: I1008 18:11:40.549626 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:11:40 crc kubenswrapper[4988]: I1008 18:11:40.549642 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:11:40 crc kubenswrapper[4988]: I1008 18:11:40.549650 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:11:40Z","lastTransitionTime":"2025-10-08T18:11:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:11:40 crc kubenswrapper[4988]: I1008 18:11:40.652690 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:11:40 crc kubenswrapper[4988]: I1008 18:11:40.652763 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:11:40 crc kubenswrapper[4988]: I1008 18:11:40.652782 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:11:40 crc kubenswrapper[4988]: I1008 18:11:40.652812 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:11:40 crc kubenswrapper[4988]: I1008 18:11:40.652832 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:11:40Z","lastTransitionTime":"2025-10-08T18:11:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:11:40 crc kubenswrapper[4988]: I1008 18:11:40.756578 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:11:40 crc kubenswrapper[4988]: I1008 18:11:40.756654 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:11:40 crc kubenswrapper[4988]: I1008 18:11:40.756673 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:11:40 crc kubenswrapper[4988]: I1008 18:11:40.756703 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:11:40 crc kubenswrapper[4988]: I1008 18:11:40.756726 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:11:40Z","lastTransitionTime":"2025-10-08T18:11:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:11:40 crc kubenswrapper[4988]: I1008 18:11:40.859982 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:11:40 crc kubenswrapper[4988]: I1008 18:11:40.860032 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:11:40 crc kubenswrapper[4988]: I1008 18:11:40.860045 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:11:40 crc kubenswrapper[4988]: I1008 18:11:40.860062 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:11:40 crc kubenswrapper[4988]: I1008 18:11:40.860075 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:11:40Z","lastTransitionTime":"2025-10-08T18:11:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:11:40 crc kubenswrapper[4988]: I1008 18:11:40.963010 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:11:40 crc kubenswrapper[4988]: I1008 18:11:40.963060 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:11:40 crc kubenswrapper[4988]: I1008 18:11:40.963087 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:11:40 crc kubenswrapper[4988]: I1008 18:11:40.963109 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:11:40 crc kubenswrapper[4988]: I1008 18:11:40.963123 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:11:40Z","lastTransitionTime":"2025-10-08T18:11:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:11:41 crc kubenswrapper[4988]: I1008 18:11:41.066585 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:11:41 crc kubenswrapper[4988]: I1008 18:11:41.066938 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:11:41 crc kubenswrapper[4988]: I1008 18:11:41.067123 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:11:41 crc kubenswrapper[4988]: I1008 18:11:41.067250 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:11:41 crc kubenswrapper[4988]: I1008 18:11:41.067375 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:11:41Z","lastTransitionTime":"2025-10-08T18:11:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:11:41 crc kubenswrapper[4988]: I1008 18:11:41.170797 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:11:41 crc kubenswrapper[4988]: I1008 18:11:41.170861 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:11:41 crc kubenswrapper[4988]: I1008 18:11:41.170879 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:11:41 crc kubenswrapper[4988]: I1008 18:11:41.170918 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:11:41 crc kubenswrapper[4988]: I1008 18:11:41.170942 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:11:41Z","lastTransitionTime":"2025-10-08T18:11:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:11:41 crc kubenswrapper[4988]: I1008 18:11:41.236948 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 18:11:41 crc kubenswrapper[4988]: E1008 18:11:41.237139 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 08 18:11:41 crc kubenswrapper[4988]: I1008 18:11:41.238017 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 18:11:41 crc kubenswrapper[4988]: I1008 18:11:41.238143 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 18:11:41 crc kubenswrapper[4988]: I1008 18:11:41.238202 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jn2sx" Oct 08 18:11:41 crc kubenswrapper[4988]: E1008 18:11:41.238301 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 08 18:11:41 crc kubenswrapper[4988]: E1008 18:11:41.238494 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 08 18:11:41 crc kubenswrapper[4988]: E1008 18:11:41.238769 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jn2sx" podUID="171e0902-95ee-4a39-8b1b-a58704c30574" Oct 08 18:11:41 crc kubenswrapper[4988]: I1008 18:11:41.239691 4988 scope.go:117] "RemoveContainer" containerID="08c1ae9c6aaca1379bd4b347fa26858dc50cbee1ae9cb9751a21351448257e11" Oct 08 18:11:41 crc kubenswrapper[4988]: I1008 18:11:41.273911 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:11:41 crc kubenswrapper[4988]: I1008 18:11:41.274254 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:11:41 crc kubenswrapper[4988]: I1008 18:11:41.278066 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:11:41 crc kubenswrapper[4988]: I1008 18:11:41.278219 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:11:41 crc kubenswrapper[4988]: I1008 18:11:41.278422 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:11:41Z","lastTransitionTime":"2025-10-08T18:11:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:11:41 crc kubenswrapper[4988]: I1008 18:11:41.381182 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:11:41 crc kubenswrapper[4988]: I1008 18:11:41.381224 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:11:41 crc kubenswrapper[4988]: I1008 18:11:41.381233 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:11:41 crc kubenswrapper[4988]: I1008 18:11:41.381247 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:11:41 crc kubenswrapper[4988]: I1008 18:11:41.381260 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:11:41Z","lastTransitionTime":"2025-10-08T18:11:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:11:41 crc kubenswrapper[4988]: I1008 18:11:41.483932 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:11:41 crc kubenswrapper[4988]: I1008 18:11:41.483984 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:11:41 crc kubenswrapper[4988]: I1008 18:11:41.483996 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:11:41 crc kubenswrapper[4988]: I1008 18:11:41.484014 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:11:41 crc kubenswrapper[4988]: I1008 18:11:41.484026 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:11:41Z","lastTransitionTime":"2025-10-08T18:11:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:11:41 crc kubenswrapper[4988]: I1008 18:11:41.586932 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:11:41 crc kubenswrapper[4988]: I1008 18:11:41.586972 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:11:41 crc kubenswrapper[4988]: I1008 18:11:41.586985 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:11:41 crc kubenswrapper[4988]: I1008 18:11:41.587000 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:11:41 crc kubenswrapper[4988]: I1008 18:11:41.587010 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:11:41Z","lastTransitionTime":"2025-10-08T18:11:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:11:41 crc kubenswrapper[4988]: I1008 18:11:41.676737 4988 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-nsc65_f010a463-69f2-4082-acc3-161899a939c3/ovnkube-controller/1.log" Oct 08 18:11:41 crc kubenswrapper[4988]: I1008 18:11:41.680412 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-nsc65" event={"ID":"f010a463-69f2-4082-acc3-161899a939c3","Type":"ContainerStarted","Data":"7bf11dd48efebb9da2b836bfad6a54b274247fa60902dab2f084793e4ad51a89"} Oct 08 18:11:41 crc kubenswrapper[4988]: I1008 18:11:41.681448 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-nsc65" Oct 08 18:11:41 crc kubenswrapper[4988]: I1008 18:11:41.690038 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:11:41 crc kubenswrapper[4988]: I1008 18:11:41.690102 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:11:41 crc kubenswrapper[4988]: I1008 18:11:41.690121 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:11:41 crc kubenswrapper[4988]: I1008 18:11:41.690164 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:11:41 crc kubenswrapper[4988]: I1008 18:11:41.690189 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:11:41Z","lastTransitionTime":"2025-10-08T18:11:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:11:41 crc kubenswrapper[4988]: I1008 18:11:41.705733 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-dzxs2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8f307b20-5e33-4725-99cb-210e25113532\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72ee803c6ef88524e0ce0391476f8e5a7dfbc858b095a3c2ae31bad10991ce3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q65f7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:11:16Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-dzxs2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:41Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:41 crc kubenswrapper[4988]: I1008 18:11:41.730806 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aef71be1-1d12-46e5-95b2-bd399040fd4d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:10:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:10:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:10:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77c68f3ecb8a0724499e495cb26ef3db5beecb140f2558d8c0b25f41928b84ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:10:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a560ee84c2405227de71fda127566213cd4f017e73d5581442ba267781b1c539\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:10:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6100accc3bd08928bc1ff88310f29fa5da37fd2d5c315c94279801c9778c4745\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:10:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b0e596d7ce4e9341c44499af79d3b3caf0e5e2c156902ea40cf2a3dca672be9e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5bd35aef24be034a0b305a40bb53ca171a4f76db3b93b4f78ae95099b0a4cde3\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-08T18:11:14Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1008 18:11:09.091034 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1008 18:11:09.091984 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3472374062/tls.crt::/tmp/serving-cert-3472374062/tls.key\\\\\\\"\\\\nI1008 18:11:14.619696 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1008 18:11:14.628837 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1008 18:11:14.628898 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1008 18:11:14.628994 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1008 18:11:14.629036 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1008 18:11:14.641318 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1008 18:11:14.641363 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1008 18:11:14.641372 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1008 18:11:14.641382 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1008 18:11:14.641415 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1008 18:11:14.641423 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1008 18:11:14.641431 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1008 18:11:14.641484 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1008 18:11:14.646451 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T18:10:58Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c79ef704fda1bbcc486f4c7afa3039a3ab38669833f0f66bafe24dd4cc679e60\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:10:58Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://28d21731af0f16d6a8ae6c6e9d0bebf0560d4a533db4c78c10bf1837890394ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://28d21731af0f16d6a8ae6c6e9d0bebf0560d4a533db4c78c10bf1837890394ef\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:10:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:10:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:10:55Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:41Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:41 crc kubenswrapper[4988]: I1008 18:11:41.745900 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7e0ecefc-359f-4785-bb79-0343a62a25da\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:10:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:10:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:10:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://30bec350fefee3d19758b3ed735ede4e9aa085edb57aa3f3ec24249a67dc7ca2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:10:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2d850fd27d4ed65891ff69b9add65fb78535357f54819c2a8d196f5a340c58e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:10:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://90b1a9e14c82bf0f75f08fdb342eed35a503020f9db641e5e9ceff8b122cb3bd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:10:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://16f23c4c7163d32dc0df8d6a7ee17e2864fb261f8a40ecf05556774b173fc868\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:10:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:10:55Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:41Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:41 crc kubenswrapper[4988]: I1008 18:11:41.763509 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e8906182d55c285b3cf095e2d9e1dac397af4618a514f11e26b44c172dd573f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:41Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:41 crc kubenswrapper[4988]: I1008 18:11:41.769722 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:11:41 crc kubenswrapper[4988]: I1008 18:11:41.769762 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:11:41 crc kubenswrapper[4988]: I1008 18:11:41.769780 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:11:41 crc kubenswrapper[4988]: I1008 18:11:41.769806 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:11:41 crc kubenswrapper[4988]: I1008 18:11:41.769825 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:11:41Z","lastTransitionTime":"2025-10-08T18:11:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:11:41 crc kubenswrapper[4988]: I1008 18:11:41.782910 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:41Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:41 crc kubenswrapper[4988]: E1008 18:11:41.791135 4988 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T18:11:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:41Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T18:11:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:41Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T18:11:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:41Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T18:11:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:41Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"95ee9fd0-8bc9-4ca5-9d79-5ee94d58bfeb\\\",\\\"systemUUID\\\":\\\"bd9489c4-fb7d-486d-b7ce-a13b6d7de36a\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:41Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:41 crc kubenswrapper[4988]: I1008 18:11:41.795427 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:11:41 crc kubenswrapper[4988]: I1008 18:11:41.795469 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:11:41 crc kubenswrapper[4988]: I1008 18:11:41.795477 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:11:41 crc kubenswrapper[4988]: I1008 18:11:41.795494 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:11:41 crc kubenswrapper[4988]: I1008 18:11:41.795503 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:11:41Z","lastTransitionTime":"2025-10-08T18:11:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:11:41 crc kubenswrapper[4988]: I1008 18:11:41.807716 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-jnh7m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"556fb048-8afb-44a5-9ea0-2b7c89e6cf4b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bfa2f3382b23c3f788a1b31e80e8a90b3b6868870dccc46d86c3fc11ad027f52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmlt2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1a825886f105ccf9a7cc6f28a0cd012fb859f4990dade0e6aae932482146028\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1a825886f105ccf9a7cc6f28a0cd012fb859f4990dade0e6aae932482146028\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmlt2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ab2fe6e65418efa75bae72259157a2323bbb4b7d11d1d012053524e14135888\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0ab2fe6e65418efa75bae72259157a2323bbb4b7d11d1d012053524e14135888\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:11:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:11:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmlt2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://368a303309be112d9c0b7ccd6538fbaca7085d97a8ec8259f36092f07147e7cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://368a303309be112d9c0b7ccd6538fbaca7085d97a8ec8259f36092f07147e7cb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:11:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:11:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmlt2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a8ca69001dddef643f982a97c74c291e30cb9fac5b0e894e80066b885ea9151\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9a8ca69001dddef643f982a97c74c291e30cb9fac5b0e894e80066b885ea9151\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:11:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:11:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmlt2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://145303c46b48603ed1aae4b9bfe06332a3ffad28f206f8394574011a87b58cd3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://145303c46b48603ed1aae4b9bfe06332a3ffad28f206f8394574011a87b58cd3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:11:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:11:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmlt2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d08e5cfc7cbe5ec1a7a0265ed7c466363c3abeec492263170860689b46c7ff95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d08e5cfc7cbe5ec1a7a0265ed7c466363c3abeec492263170860689b46c7ff95\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:11:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:11:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmlt2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:11:16Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-jnh7m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:41Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:41 crc kubenswrapper[4988]: E1008 18:11:41.813597 4988 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T18:11:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:41Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T18:11:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:41Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T18:11:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:41Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T18:11:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:41Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"95ee9fd0-8bc9-4ca5-9d79-5ee94d58bfeb\\\",\\\"systemUUID\\\":\\\"bd9489c4-fb7d-486d-b7ce-a13b6d7de36a\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:41Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:41 crc kubenswrapper[4988]: I1008 18:11:41.819481 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:11:41 crc kubenswrapper[4988]: I1008 18:11:41.819510 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:11:41 crc kubenswrapper[4988]: I1008 18:11:41.819519 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:11:41 crc kubenswrapper[4988]: I1008 18:11:41.819535 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:11:41 crc kubenswrapper[4988]: I1008 18:11:41.819546 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:11:41Z","lastTransitionTime":"2025-10-08T18:11:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:11:41 crc kubenswrapper[4988]: I1008 18:11:41.829480 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f328a20b-5283-407e-8a93-2ab2da599f18\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8622d56fb198dadee31fb0ca8fc80d56828b6710648055f264533e2d5bdb3037\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z2gh4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3cdb2a757332bdcdc2a2f3362913fe8f06003b0597de20b72627f9badd5e30af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z2gh4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:11:16Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-nm54f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:41Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:41 crc kubenswrapper[4988]: E1008 18:11:41.837887 4988 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T18:11:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:41Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T18:11:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:41Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T18:11:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:41Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T18:11:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:41Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"95ee9fd0-8bc9-4ca5-9d79-5ee94d58bfeb\\\",\\\"systemUUID\\\":\\\"bd9489c4-fb7d-486d-b7ce-a13b6d7de36a\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:41Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:41 crc kubenswrapper[4988]: I1008 18:11:41.842004 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:11:41 crc kubenswrapper[4988]: I1008 18:11:41.842066 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:11:41 crc kubenswrapper[4988]: I1008 18:11:41.842077 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:11:41 crc kubenswrapper[4988]: I1008 18:11:41.842099 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:11:41 crc kubenswrapper[4988]: I1008 18:11:41.842113 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:11:41Z","lastTransitionTime":"2025-10-08T18:11:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:11:41 crc kubenswrapper[4988]: I1008 18:11:41.842754 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a35a5e989ba69c8ef36ce4cea272aadf67ce20473fc257adfcfa86a2faba6ae0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a522e937cfe03fb9f2e60cb1bea07214cf63c69bdecd5ba65a2dfdba882d28dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:41Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:41 crc kubenswrapper[4988]: E1008 18:11:41.854752 4988 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T18:11:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:41Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T18:11:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:41Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T18:11:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:41Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T18:11:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:41Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"95ee9fd0-8bc9-4ca5-9d79-5ee94d58bfeb\\\",\\\"systemUUID\\\":\\\"bd9489c4-fb7d-486d-b7ce-a13b6d7de36a\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:41Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:41 crc kubenswrapper[4988]: I1008 18:11:41.857296 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9590ca0059f0ebbe86343cccb3475338c317382c3231e082655b61e67eece375\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:41Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:41 crc kubenswrapper[4988]: I1008 18:11:41.859078 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:11:41 crc kubenswrapper[4988]: I1008 18:11:41.859120 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:11:41 crc kubenswrapper[4988]: I1008 18:11:41.859131 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:11:41 crc kubenswrapper[4988]: I1008 18:11:41.859151 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:11:41 crc kubenswrapper[4988]: I1008 18:11:41.859168 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:11:41Z","lastTransitionTime":"2025-10-08T18:11:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:11:41 crc kubenswrapper[4988]: E1008 18:11:41.872630 4988 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T18:11:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:41Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T18:11:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:41Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T18:11:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:41Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T18:11:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:41Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"95ee9fd0-8bc9-4ca5-9d79-5ee94d58bfeb\\\",\\\"systemUUID\\\":\\\"bd9489c4-fb7d-486d-b7ce-a13b6d7de36a\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:41Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:41 crc kubenswrapper[4988]: E1008 18:11:41.873066 4988 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Oct 08 18:11:41 crc kubenswrapper[4988]: I1008 18:11:41.874840 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:11:41 crc kubenswrapper[4988]: I1008 18:11:41.874968 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:11:41 crc kubenswrapper[4988]: I1008 18:11:41.875059 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:11:41 crc kubenswrapper[4988]: I1008 18:11:41.875154 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:11:41 crc kubenswrapper[4988]: I1008 18:11:41.875216 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:11:41Z","lastTransitionTime":"2025-10-08T18:11:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:11:41 crc kubenswrapper[4988]: I1008 18:11:41.875727 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-h9jwp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4490d846-ca32-4cd5-9202-87cc6d499fb4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://00b7f6db043c28efda3a09fdc77a6ccded3eae8b6a3d2852aa500789d5f01194\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7k6rk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:11:16Z\\\"}}\" for pod \"openshift-multus\"/\"multus-h9jwp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:41Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:41 crc kubenswrapper[4988]: I1008 18:11:41.897932 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-nsc65" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f010a463-69f2-4082-acc3-161899a939c3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://48a516f6c153743bd1a21ec6945fb368a37f4d2b9d6a326b68edc096a192ed05\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dnwgr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://03406e726f817bcedb841f81959ab12808ba35a530240f9abc5fa5c608b9a6ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dnwgr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9fbbeae0409d2e47e8de9a4306f63b0d13acbfd73cd86f83303fec14352ed718\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dnwgr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://55791327a370c0009c523b86b2234e82d3794985f2eaccf1922eb90f46b2b970\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dnwgr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cee5de93bc68345f9b737857f02832f56582d58757c50a8798eec25d19b2c20d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dnwgr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d314eb248e90371fed068b1173e26d4c9f9765417a70b341e1a5b136ec17353a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dnwgr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7bf11dd48efebb9da2b836bfad6a54b274247fa60902dab2f084793e4ad51a89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://08c1ae9c6aaca1379bd4b347fa26858dc50cbee1ae9cb9751a21351448257e11\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-08T18:11:27Z\\\",\\\"message\\\":\\\"ntil: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1008 18:11:27.764809 6411 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1008 18:11:27.764817 6411 handler.go:208] Removed *v1.Node event handler 2\\\\nI1008 18:11:27.764773 6411 loadbalancer.go:304] Deleted 0 stale LBs for map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-ingress-operator/metrics\\\\\\\"}\\\\nI1008 18:11:27.765603 6411 transact.go:42] Configuring OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-machine-api/control-plane-machine-set-operator]} name:Service_openshift-machine-api/control-plane-machine-set-operator_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.41:9443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {589f95f7-f3e2-4140-80ed-9a0717201481}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1008 18:11:27.765621 6411 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1008 18:11:27.765753 6411 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T18:11:26Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dnwgr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://74fcb812b84291763cc6e5ccb20c12aaad580ad968dda4734a0bbfe33d2770ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dnwgr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://befe684cea9bd6dfda3ec8e85bca431c26bd8f82823b0a69dc0d04319a3ca385\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://befe684cea9bd6dfda3ec8e85bca431c26bd8f82823b0a69dc0d04319a3ca385\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dnwgr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:11:16Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-nsc65\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:41Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:41 crc kubenswrapper[4988]: I1008 18:11:41.912252 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-7wl8v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e433488d-db52-435f-b89c-f63cd477c5d3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76b30d183684740b647c1af11dc49712232cc8b0c9a392b491580e5ddb75ea7f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dbz6l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e55016c3e1b97d1b3a857130a5c8e2ec6da2333f786032118a61e3e2ad4c5651\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dbz6l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:11:29Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-7wl8v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:41Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:41 crc kubenswrapper[4988]: I1008 18:11:41.924257 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-jn2sx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"171e0902-95ee-4a39-8b1b-a58704c30574\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9f4q2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9f4q2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:11:31Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-jn2sx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:41Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:41 crc kubenswrapper[4988]: I1008 18:11:41.940374 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:41Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:41 crc kubenswrapper[4988]: I1008 18:11:41.957932 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:41Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:41 crc kubenswrapper[4988]: I1008 18:11:41.972953 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-s4f2k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ba5e11d7-7bde-4c3d-983b-884232219652\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://824bc34a2807068d3fceb10651803ae9ae5651e940f7aed4877c544d76dae4d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hn5jf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:11:15Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-s4f2k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:41Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:41 crc kubenswrapper[4988]: I1008 18:11:41.978022 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:11:41 crc kubenswrapper[4988]: I1008 18:11:41.978095 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:11:41 crc kubenswrapper[4988]: I1008 18:11:41.978110 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:11:41 crc kubenswrapper[4988]: I1008 18:11:41.978180 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:11:41 crc kubenswrapper[4988]: I1008 18:11:41.978202 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:11:41Z","lastTransitionTime":"2025-10-08T18:11:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:11:42 crc kubenswrapper[4988]: I1008 18:11:42.081159 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:11:42 crc kubenswrapper[4988]: I1008 18:11:42.081195 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:11:42 crc kubenswrapper[4988]: I1008 18:11:42.081205 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:11:42 crc kubenswrapper[4988]: I1008 18:11:42.081221 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:11:42 crc kubenswrapper[4988]: I1008 18:11:42.081232 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:11:42Z","lastTransitionTime":"2025-10-08T18:11:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:11:42 crc kubenswrapper[4988]: I1008 18:11:42.184147 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:11:42 crc kubenswrapper[4988]: I1008 18:11:42.184580 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:11:42 crc kubenswrapper[4988]: I1008 18:11:42.184593 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:11:42 crc kubenswrapper[4988]: I1008 18:11:42.184620 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:11:42 crc kubenswrapper[4988]: I1008 18:11:42.184634 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:11:42Z","lastTransitionTime":"2025-10-08T18:11:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:11:42 crc kubenswrapper[4988]: I1008 18:11:42.287978 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:11:42 crc kubenswrapper[4988]: I1008 18:11:42.288028 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:11:42 crc kubenswrapper[4988]: I1008 18:11:42.288040 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:11:42 crc kubenswrapper[4988]: I1008 18:11:42.288062 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:11:42 crc kubenswrapper[4988]: I1008 18:11:42.288080 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:11:42Z","lastTransitionTime":"2025-10-08T18:11:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:11:42 crc kubenswrapper[4988]: I1008 18:11:42.392021 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:11:42 crc kubenswrapper[4988]: I1008 18:11:42.392108 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:11:42 crc kubenswrapper[4988]: I1008 18:11:42.392129 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:11:42 crc kubenswrapper[4988]: I1008 18:11:42.392162 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:11:42 crc kubenswrapper[4988]: I1008 18:11:42.392186 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:11:42Z","lastTransitionTime":"2025-10-08T18:11:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:11:42 crc kubenswrapper[4988]: I1008 18:11:42.495802 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:11:42 crc kubenswrapper[4988]: I1008 18:11:42.495868 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:11:42 crc kubenswrapper[4988]: I1008 18:11:42.495885 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:11:42 crc kubenswrapper[4988]: I1008 18:11:42.495914 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:11:42 crc kubenswrapper[4988]: I1008 18:11:42.495936 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:11:42Z","lastTransitionTime":"2025-10-08T18:11:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:11:42 crc kubenswrapper[4988]: I1008 18:11:42.598996 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:11:42 crc kubenswrapper[4988]: I1008 18:11:42.599065 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:11:42 crc kubenswrapper[4988]: I1008 18:11:42.599087 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:11:42 crc kubenswrapper[4988]: I1008 18:11:42.599117 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:11:42 crc kubenswrapper[4988]: I1008 18:11:42.599137 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:11:42Z","lastTransitionTime":"2025-10-08T18:11:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:11:42 crc kubenswrapper[4988]: I1008 18:11:42.687514 4988 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-nsc65_f010a463-69f2-4082-acc3-161899a939c3/ovnkube-controller/2.log" Oct 08 18:11:42 crc kubenswrapper[4988]: I1008 18:11:42.688464 4988 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-nsc65_f010a463-69f2-4082-acc3-161899a939c3/ovnkube-controller/1.log" Oct 08 18:11:42 crc kubenswrapper[4988]: I1008 18:11:42.692715 4988 generic.go:334] "Generic (PLEG): container finished" podID="f010a463-69f2-4082-acc3-161899a939c3" containerID="7bf11dd48efebb9da2b836bfad6a54b274247fa60902dab2f084793e4ad51a89" exitCode=1 Oct 08 18:11:42 crc kubenswrapper[4988]: I1008 18:11:42.692764 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-nsc65" event={"ID":"f010a463-69f2-4082-acc3-161899a939c3","Type":"ContainerDied","Data":"7bf11dd48efebb9da2b836bfad6a54b274247fa60902dab2f084793e4ad51a89"} Oct 08 18:11:42 crc kubenswrapper[4988]: I1008 18:11:42.692807 4988 scope.go:117] "RemoveContainer" containerID="08c1ae9c6aaca1379bd4b347fa26858dc50cbee1ae9cb9751a21351448257e11" Oct 08 18:11:42 crc kubenswrapper[4988]: I1008 18:11:42.694449 4988 scope.go:117] "RemoveContainer" containerID="7bf11dd48efebb9da2b836bfad6a54b274247fa60902dab2f084793e4ad51a89" Oct 08 18:11:42 crc kubenswrapper[4988]: E1008 18:11:42.694816 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-nsc65_openshift-ovn-kubernetes(f010a463-69f2-4082-acc3-161899a939c3)\"" pod="openshift-ovn-kubernetes/ovnkube-node-nsc65" podUID="f010a463-69f2-4082-acc3-161899a939c3" Oct 08 18:11:42 crc kubenswrapper[4988]: I1008 18:11:42.702755 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:11:42 crc kubenswrapper[4988]: I1008 18:11:42.702796 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:11:42 crc kubenswrapper[4988]: I1008 18:11:42.702811 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:11:42 crc kubenswrapper[4988]: I1008 18:11:42.702827 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:11:42 crc kubenswrapper[4988]: I1008 18:11:42.702840 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:11:42Z","lastTransitionTime":"2025-10-08T18:11:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:11:42 crc kubenswrapper[4988]: I1008 18:11:42.719726 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aef71be1-1d12-46e5-95b2-bd399040fd4d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:10:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:10:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:10:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77c68f3ecb8a0724499e495cb26ef3db5beecb140f2558d8c0b25f41928b84ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:10:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a560ee84c2405227de71fda127566213cd4f017e73d5581442ba267781b1c539\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:10:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6100accc3bd08928bc1ff88310f29fa5da37fd2d5c315c94279801c9778c4745\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:10:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b0e596d7ce4e9341c44499af79d3b3caf0e5e2c156902ea40cf2a3dca672be9e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5bd35aef24be034a0b305a40bb53ca171a4f76db3b93b4f78ae95099b0a4cde3\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-08T18:11:14Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1008 18:11:09.091034 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1008 18:11:09.091984 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3472374062/tls.crt::/tmp/serving-cert-3472374062/tls.key\\\\\\\"\\\\nI1008 18:11:14.619696 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1008 18:11:14.628837 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1008 18:11:14.628898 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1008 18:11:14.628994 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1008 18:11:14.629036 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1008 18:11:14.641318 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1008 18:11:14.641363 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1008 18:11:14.641372 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1008 18:11:14.641382 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1008 18:11:14.641415 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1008 18:11:14.641423 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1008 18:11:14.641431 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1008 18:11:14.641484 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1008 18:11:14.646451 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T18:10:58Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c79ef704fda1bbcc486f4c7afa3039a3ab38669833f0f66bafe24dd4cc679e60\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:10:58Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://28d21731af0f16d6a8ae6c6e9d0bebf0560d4a533db4c78c10bf1837890394ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://28d21731af0f16d6a8ae6c6e9d0bebf0560d4a533db4c78c10bf1837890394ef\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:10:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:10:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:10:55Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:42Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:42 crc kubenswrapper[4988]: I1008 18:11:42.739128 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7e0ecefc-359f-4785-bb79-0343a62a25da\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:10:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:10:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:10:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://30bec350fefee3d19758b3ed735ede4e9aa085edb57aa3f3ec24249a67dc7ca2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:10:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2d850fd27d4ed65891ff69b9add65fb78535357f54819c2a8d196f5a340c58e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:10:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://90b1a9e14c82bf0f75f08fdb342eed35a503020f9db641e5e9ceff8b122cb3bd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:10:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://16f23c4c7163d32dc0df8d6a7ee17e2864fb261f8a40ecf05556774b173fc868\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:10:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:10:55Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:42Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:42 crc kubenswrapper[4988]: I1008 18:11:42.760150 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e8906182d55c285b3cf095e2d9e1dac397af4618a514f11e26b44c172dd573f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:42Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:42 crc kubenswrapper[4988]: I1008 18:11:42.775042 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-dzxs2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8f307b20-5e33-4725-99cb-210e25113532\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72ee803c6ef88524e0ce0391476f8e5a7dfbc858b095a3c2ae31bad10991ce3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q65f7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:11:16Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-dzxs2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:42Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:42 crc kubenswrapper[4988]: I1008 18:11:42.792968 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-jnh7m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"556fb048-8afb-44a5-9ea0-2b7c89e6cf4b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bfa2f3382b23c3f788a1b31e80e8a90b3b6868870dccc46d86c3fc11ad027f52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmlt2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1a825886f105ccf9a7cc6f28a0cd012fb859f4990dade0e6aae932482146028\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1a825886f105ccf9a7cc6f28a0cd012fb859f4990dade0e6aae932482146028\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmlt2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ab2fe6e65418efa75bae72259157a2323bbb4b7d11d1d012053524e14135888\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0ab2fe6e65418efa75bae72259157a2323bbb4b7d11d1d012053524e14135888\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:11:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:11:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmlt2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://368a303309be112d9c0b7ccd6538fbaca7085d97a8ec8259f36092f07147e7cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://368a303309be112d9c0b7ccd6538fbaca7085d97a8ec8259f36092f07147e7cb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:11:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:11:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmlt2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a8ca69001dddef643f982a97c74c291e30cb9fac5b0e894e80066b885ea9151\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9a8ca69001dddef643f982a97c74c291e30cb9fac5b0e894e80066b885ea9151\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:11:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:11:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmlt2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://145303c46b48603ed1aae4b9bfe06332a3ffad28f206f8394574011a87b58cd3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://145303c46b48603ed1aae4b9bfe06332a3ffad28f206f8394574011a87b58cd3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:11:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:11:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmlt2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d08e5cfc7cbe5ec1a7a0265ed7c466363c3abeec492263170860689b46c7ff95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d08e5cfc7cbe5ec1a7a0265ed7c466363c3abeec492263170860689b46c7ff95\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:11:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:11:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmlt2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:11:16Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-jnh7m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:42Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:42 crc kubenswrapper[4988]: I1008 18:11:42.805367 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:11:42 crc kubenswrapper[4988]: I1008 18:11:42.805446 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:11:42 crc kubenswrapper[4988]: I1008 18:11:42.805468 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:11:42 crc kubenswrapper[4988]: I1008 18:11:42.805496 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:11:42 crc kubenswrapper[4988]: I1008 18:11:42.805516 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:11:42Z","lastTransitionTime":"2025-10-08T18:11:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:11:42 crc kubenswrapper[4988]: I1008 18:11:42.810217 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f328a20b-5283-407e-8a93-2ab2da599f18\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8622d56fb198dadee31fb0ca8fc80d56828b6710648055f264533e2d5bdb3037\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z2gh4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3cdb2a757332bdcdc2a2f3362913fe8f06003b0597de20b72627f9badd5e30af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z2gh4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:11:16Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-nm54f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:42Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:42 crc kubenswrapper[4988]: I1008 18:11:42.831933 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a35a5e989ba69c8ef36ce4cea272aadf67ce20473fc257adfcfa86a2faba6ae0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a522e937cfe03fb9f2e60cb1bea07214cf63c69bdecd5ba65a2dfdba882d28dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:42Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:42 crc kubenswrapper[4988]: I1008 18:11:42.847593 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9590ca0059f0ebbe86343cccb3475338c317382c3231e082655b61e67eece375\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:42Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:42 crc kubenswrapper[4988]: I1008 18:11:42.863195 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:42Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:42 crc kubenswrapper[4988]: I1008 18:11:42.886128 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-nsc65" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f010a463-69f2-4082-acc3-161899a939c3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://48a516f6c153743bd1a21ec6945fb368a37f4d2b9d6a326b68edc096a192ed05\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dnwgr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://03406e726f817bcedb841f81959ab12808ba35a530240f9abc5fa5c608b9a6ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dnwgr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9fbbeae0409d2e47e8de9a4306f63b0d13acbfd73cd86f83303fec14352ed718\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dnwgr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://55791327a370c0009c523b86b2234e82d3794985f2eaccf1922eb90f46b2b970\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dnwgr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cee5de93bc68345f9b737857f02832f56582d58757c50a8798eec25d19b2c20d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dnwgr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d314eb248e90371fed068b1173e26d4c9f9765417a70b341e1a5b136ec17353a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dnwgr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7bf11dd48efebb9da2b836bfad6a54b274247fa60902dab2f084793e4ad51a89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://08c1ae9c6aaca1379bd4b347fa26858dc50cbee1ae9cb9751a21351448257e11\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-08T18:11:27Z\\\",\\\"message\\\":\\\"ntil: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1008 18:11:27.764809 6411 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1008 18:11:27.764817 6411 handler.go:208] Removed *v1.Node event handler 2\\\\nI1008 18:11:27.764773 6411 loadbalancer.go:304] Deleted 0 stale LBs for map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-ingress-operator/metrics\\\\\\\"}\\\\nI1008 18:11:27.765603 6411 transact.go:42] Configuring OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-machine-api/control-plane-machine-set-operator]} name:Service_openshift-machine-api/control-plane-machine-set-operator_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.41:9443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {589f95f7-f3e2-4140-80ed-9a0717201481}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1008 18:11:27.765621 6411 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1008 18:11:27.765753 6411 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T18:11:26Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7bf11dd48efebb9da2b836bfad6a54b274247fa60902dab2f084793e4ad51a89\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-08T18:11:42Z\\\",\\\"message\\\":\\\" V4IPs:[]string(nil), V6IPs:[]string(nil)}, nodeEndpoints:map[string]services.lbEndpoints{}, externalTrafficLocal:false, internalTrafficLocal:false, hasNodePort:false}}\\\\nI1008 18:11:42.211985 6628 services_controller.go:444] Built service openshift-authentication-operator/metrics LB per-node configs for network=default: []services.lbConfig(nil)\\\\nI1008 18:11:42.211998 6628 services_controller.go:445] Built service openshift-authentication-operator/metrics LB template configs for network=default: []services.lbConfig(nil)\\\\nF1008 18:11:42.212009 6628 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:42Z is after 2025-08-\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T18:11:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dnwgr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://74fcb812b84291763cc6e5ccb20c12aaad580ad968dda4734a0bbfe33d2770ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dnwgr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://befe684cea9bd6dfda3ec8e85bca431c26bd8f82823b0a69dc0d04319a3ca385\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://befe684cea9bd6dfda3ec8e85bca431c26bd8f82823b0a69dc0d04319a3ca385\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dnwgr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:11:16Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-nsc65\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:42Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:42 crc kubenswrapper[4988]: I1008 18:11:42.905941 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-7wl8v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e433488d-db52-435f-b89c-f63cd477c5d3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76b30d183684740b647c1af11dc49712232cc8b0c9a392b491580e5ddb75ea7f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dbz6l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e55016c3e1b97d1b3a857130a5c8e2ec6da2333f786032118a61e3e2ad4c5651\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dbz6l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:11:29Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-7wl8v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:42Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:42 crc kubenswrapper[4988]: I1008 18:11:42.909100 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:11:42 crc kubenswrapper[4988]: I1008 18:11:42.909154 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:11:42 crc kubenswrapper[4988]: I1008 18:11:42.909174 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:11:42 crc kubenswrapper[4988]: I1008 18:11:42.909208 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:11:42 crc kubenswrapper[4988]: I1008 18:11:42.909227 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:11:42Z","lastTransitionTime":"2025-10-08T18:11:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:11:42 crc kubenswrapper[4988]: I1008 18:11:42.923671 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-jn2sx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"171e0902-95ee-4a39-8b1b-a58704c30574\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9f4q2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9f4q2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:11:31Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-jn2sx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:42Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:42 crc kubenswrapper[4988]: I1008 18:11:42.938508 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:42Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:42 crc kubenswrapper[4988]: I1008 18:11:42.953527 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:42Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:42 crc kubenswrapper[4988]: I1008 18:11:42.976300 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-h9jwp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4490d846-ca32-4cd5-9202-87cc6d499fb4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://00b7f6db043c28efda3a09fdc77a6ccded3eae8b6a3d2852aa500789d5f01194\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7k6rk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:11:16Z\\\"}}\" for pod \"openshift-multus\"/\"multus-h9jwp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:42Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:42 crc kubenswrapper[4988]: I1008 18:11:42.989597 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-s4f2k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ba5e11d7-7bde-4c3d-983b-884232219652\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://824bc34a2807068d3fceb10651803ae9ae5651e940f7aed4877c544d76dae4d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hn5jf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:11:15Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-s4f2k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:42Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:43 crc kubenswrapper[4988]: I1008 18:11:43.012751 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:11:43 crc kubenswrapper[4988]: I1008 18:11:43.012831 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:11:43 crc kubenswrapper[4988]: I1008 18:11:43.012855 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:11:43 crc kubenswrapper[4988]: I1008 18:11:43.012887 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:11:43 crc kubenswrapper[4988]: I1008 18:11:43.012910 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:11:43Z","lastTransitionTime":"2025-10-08T18:11:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:11:43 crc kubenswrapper[4988]: I1008 18:11:43.116500 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:11:43 crc kubenswrapper[4988]: I1008 18:11:43.116564 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:11:43 crc kubenswrapper[4988]: I1008 18:11:43.116580 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:11:43 crc kubenswrapper[4988]: I1008 18:11:43.116609 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:11:43 crc kubenswrapper[4988]: I1008 18:11:43.116629 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:11:43Z","lastTransitionTime":"2025-10-08T18:11:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:11:43 crc kubenswrapper[4988]: I1008 18:11:43.219669 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:11:43 crc kubenswrapper[4988]: I1008 18:11:43.219742 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:11:43 crc kubenswrapper[4988]: I1008 18:11:43.219761 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:11:43 crc kubenswrapper[4988]: I1008 18:11:43.219809 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:11:43 crc kubenswrapper[4988]: I1008 18:11:43.219836 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:11:43Z","lastTransitionTime":"2025-10-08T18:11:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:11:43 crc kubenswrapper[4988]: I1008 18:11:43.237828 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 18:11:43 crc kubenswrapper[4988]: I1008 18:11:43.237874 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 18:11:43 crc kubenswrapper[4988]: E1008 18:11:43.238042 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 08 18:11:43 crc kubenswrapper[4988]: I1008 18:11:43.238161 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 18:11:43 crc kubenswrapper[4988]: E1008 18:11:43.238343 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 08 18:11:43 crc kubenswrapper[4988]: I1008 18:11:43.238483 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jn2sx" Oct 08 18:11:43 crc kubenswrapper[4988]: E1008 18:11:43.238622 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 08 18:11:43 crc kubenswrapper[4988]: E1008 18:11:43.238764 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jn2sx" podUID="171e0902-95ee-4a39-8b1b-a58704c30574" Oct 08 18:11:43 crc kubenswrapper[4988]: I1008 18:11:43.323423 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:11:43 crc kubenswrapper[4988]: I1008 18:11:43.323492 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:11:43 crc kubenswrapper[4988]: I1008 18:11:43.323513 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:11:43 crc kubenswrapper[4988]: I1008 18:11:43.323546 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:11:43 crc kubenswrapper[4988]: I1008 18:11:43.323566 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:11:43Z","lastTransitionTime":"2025-10-08T18:11:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:11:43 crc kubenswrapper[4988]: I1008 18:11:43.427532 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:11:43 crc kubenswrapper[4988]: I1008 18:11:43.427596 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:11:43 crc kubenswrapper[4988]: I1008 18:11:43.427612 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:11:43 crc kubenswrapper[4988]: I1008 18:11:43.427636 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:11:43 crc kubenswrapper[4988]: I1008 18:11:43.427650 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:11:43Z","lastTransitionTime":"2025-10-08T18:11:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:11:43 crc kubenswrapper[4988]: I1008 18:11:43.531752 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:11:43 crc kubenswrapper[4988]: I1008 18:11:43.531824 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:11:43 crc kubenswrapper[4988]: I1008 18:11:43.531844 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:11:43 crc kubenswrapper[4988]: I1008 18:11:43.531876 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:11:43 crc kubenswrapper[4988]: I1008 18:11:43.531896 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:11:43Z","lastTransitionTime":"2025-10-08T18:11:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:11:43 crc kubenswrapper[4988]: I1008 18:11:43.635144 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:11:43 crc kubenswrapper[4988]: I1008 18:11:43.635196 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:11:43 crc kubenswrapper[4988]: I1008 18:11:43.635210 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:11:43 crc kubenswrapper[4988]: I1008 18:11:43.635228 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:11:43 crc kubenswrapper[4988]: I1008 18:11:43.635241 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:11:43Z","lastTransitionTime":"2025-10-08T18:11:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:11:43 crc kubenswrapper[4988]: I1008 18:11:43.701117 4988 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-nsc65_f010a463-69f2-4082-acc3-161899a939c3/ovnkube-controller/2.log" Oct 08 18:11:43 crc kubenswrapper[4988]: I1008 18:11:43.709282 4988 scope.go:117] "RemoveContainer" containerID="7bf11dd48efebb9da2b836bfad6a54b274247fa60902dab2f084793e4ad51a89" Oct 08 18:11:43 crc kubenswrapper[4988]: E1008 18:11:43.709615 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-nsc65_openshift-ovn-kubernetes(f010a463-69f2-4082-acc3-161899a939c3)\"" pod="openshift-ovn-kubernetes/ovnkube-node-nsc65" podUID="f010a463-69f2-4082-acc3-161899a939c3" Oct 08 18:11:43 crc kubenswrapper[4988]: I1008 18:11:43.728230 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-s4f2k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ba5e11d7-7bde-4c3d-983b-884232219652\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://824bc34a2807068d3fceb10651803ae9ae5651e940f7aed4877c544d76dae4d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hn5jf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:11:15Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-s4f2k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:43Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:43 crc kubenswrapper[4988]: I1008 18:11:43.738070 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:11:43 crc kubenswrapper[4988]: I1008 18:11:43.738553 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:11:43 crc kubenswrapper[4988]: I1008 18:11:43.738865 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:11:43 crc kubenswrapper[4988]: I1008 18:11:43.739085 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:11:43 crc kubenswrapper[4988]: I1008 18:11:43.739285 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:11:43Z","lastTransitionTime":"2025-10-08T18:11:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:11:43 crc kubenswrapper[4988]: I1008 18:11:43.754682 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e8906182d55c285b3cf095e2d9e1dac397af4618a514f11e26b44c172dd573f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:43Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:43 crc kubenswrapper[4988]: I1008 18:11:43.773524 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-dzxs2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8f307b20-5e33-4725-99cb-210e25113532\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72ee803c6ef88524e0ce0391476f8e5a7dfbc858b095a3c2ae31bad10991ce3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q65f7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:11:16Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-dzxs2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:43Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:43 crc kubenswrapper[4988]: I1008 18:11:43.797526 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aef71be1-1d12-46e5-95b2-bd399040fd4d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:10:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:10:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:10:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77c68f3ecb8a0724499e495cb26ef3db5beecb140f2558d8c0b25f41928b84ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:10:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a560ee84c2405227de71fda127566213cd4f017e73d5581442ba267781b1c539\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:10:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6100accc3bd08928bc1ff88310f29fa5da37fd2d5c315c94279801c9778c4745\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:10:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b0e596d7ce4e9341c44499af79d3b3caf0e5e2c156902ea40cf2a3dca672be9e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5bd35aef24be034a0b305a40bb53ca171a4f76db3b93b4f78ae95099b0a4cde3\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-08T18:11:14Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1008 18:11:09.091034 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1008 18:11:09.091984 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3472374062/tls.crt::/tmp/serving-cert-3472374062/tls.key\\\\\\\"\\\\nI1008 18:11:14.619696 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1008 18:11:14.628837 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1008 18:11:14.628898 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1008 18:11:14.628994 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1008 18:11:14.629036 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1008 18:11:14.641318 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1008 18:11:14.641363 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1008 18:11:14.641372 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1008 18:11:14.641382 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1008 18:11:14.641415 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1008 18:11:14.641423 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1008 18:11:14.641431 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1008 18:11:14.641484 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1008 18:11:14.646451 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T18:10:58Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c79ef704fda1bbcc486f4c7afa3039a3ab38669833f0f66bafe24dd4cc679e60\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:10:58Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://28d21731af0f16d6a8ae6c6e9d0bebf0560d4a533db4c78c10bf1837890394ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://28d21731af0f16d6a8ae6c6e9d0bebf0560d4a533db4c78c10bf1837890394ef\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:10:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:10:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:10:55Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:43Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:43 crc kubenswrapper[4988]: I1008 18:11:43.823105 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7e0ecefc-359f-4785-bb79-0343a62a25da\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:10:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:10:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:10:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://30bec350fefee3d19758b3ed735ede4e9aa085edb57aa3f3ec24249a67dc7ca2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:10:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2d850fd27d4ed65891ff69b9add65fb78535357f54819c2a8d196f5a340c58e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:10:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://90b1a9e14c82bf0f75f08fdb342eed35a503020f9db641e5e9ceff8b122cb3bd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:10:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://16f23c4c7163d32dc0df8d6a7ee17e2864fb261f8a40ecf05556774b173fc868\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:10:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:10:55Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:43Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:43 crc kubenswrapper[4988]: I1008 18:11:43.844154 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:11:43 crc kubenswrapper[4988]: I1008 18:11:43.844228 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:11:43 crc kubenswrapper[4988]: I1008 18:11:43.844244 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:11:43 crc kubenswrapper[4988]: I1008 18:11:43.844276 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:11:43 crc kubenswrapper[4988]: I1008 18:11:43.844294 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:11:43Z","lastTransitionTime":"2025-10-08T18:11:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:11:43 crc kubenswrapper[4988]: I1008 18:11:43.844294 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9590ca0059f0ebbe86343cccb3475338c317382c3231e082655b61e67eece375\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:43Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:43 crc kubenswrapper[4988]: I1008 18:11:43.866842 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:43Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:43 crc kubenswrapper[4988]: I1008 18:11:43.891848 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-jnh7m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"556fb048-8afb-44a5-9ea0-2b7c89e6cf4b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bfa2f3382b23c3f788a1b31e80e8a90b3b6868870dccc46d86c3fc11ad027f52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmlt2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1a825886f105ccf9a7cc6f28a0cd012fb859f4990dade0e6aae932482146028\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1a825886f105ccf9a7cc6f28a0cd012fb859f4990dade0e6aae932482146028\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmlt2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ab2fe6e65418efa75bae72259157a2323bbb4b7d11d1d012053524e14135888\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0ab2fe6e65418efa75bae72259157a2323bbb4b7d11d1d012053524e14135888\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:11:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:11:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmlt2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://368a303309be112d9c0b7ccd6538fbaca7085d97a8ec8259f36092f07147e7cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://368a303309be112d9c0b7ccd6538fbaca7085d97a8ec8259f36092f07147e7cb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:11:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:11:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmlt2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a8ca69001dddef643f982a97c74c291e30cb9fac5b0e894e80066b885ea9151\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9a8ca69001dddef643f982a97c74c291e30cb9fac5b0e894e80066b885ea9151\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:11:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:11:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmlt2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://145303c46b48603ed1aae4b9bfe06332a3ffad28f206f8394574011a87b58cd3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://145303c46b48603ed1aae4b9bfe06332a3ffad28f206f8394574011a87b58cd3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:11:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:11:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmlt2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d08e5cfc7cbe5ec1a7a0265ed7c466363c3abeec492263170860689b46c7ff95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d08e5cfc7cbe5ec1a7a0265ed7c466363c3abeec492263170860689b46c7ff95\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:11:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:11:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmlt2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:11:16Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-jnh7m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:43Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:43 crc kubenswrapper[4988]: I1008 18:11:43.916966 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f328a20b-5283-407e-8a93-2ab2da599f18\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8622d56fb198dadee31fb0ca8fc80d56828b6710648055f264533e2d5bdb3037\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z2gh4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3cdb2a757332bdcdc2a2f3362913fe8f06003b0597de20b72627f9badd5e30af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z2gh4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:11:16Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-nm54f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:43Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:43 crc kubenswrapper[4988]: I1008 18:11:43.939520 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a35a5e989ba69c8ef36ce4cea272aadf67ce20473fc257adfcfa86a2faba6ae0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a522e937cfe03fb9f2e60cb1bea07214cf63c69bdecd5ba65a2dfdba882d28dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:43Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:43 crc kubenswrapper[4988]: I1008 18:11:43.947551 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:11:43 crc kubenswrapper[4988]: I1008 18:11:43.947656 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:11:43 crc kubenswrapper[4988]: I1008 18:11:43.947678 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:11:43 crc kubenswrapper[4988]: I1008 18:11:43.947705 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:11:43 crc kubenswrapper[4988]: I1008 18:11:43.947725 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:11:43Z","lastTransitionTime":"2025-10-08T18:11:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:11:43 crc kubenswrapper[4988]: I1008 18:11:43.963192 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:43Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:43 crc kubenswrapper[4988]: I1008 18:11:43.984048 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-h9jwp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4490d846-ca32-4cd5-9202-87cc6d499fb4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://00b7f6db043c28efda3a09fdc77a6ccded3eae8b6a3d2852aa500789d5f01194\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7k6rk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:11:16Z\\\"}}\" for pod \"openshift-multus\"/\"multus-h9jwp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:43Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:44 crc kubenswrapper[4988]: I1008 18:11:44.015512 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-nsc65" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f010a463-69f2-4082-acc3-161899a939c3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://48a516f6c153743bd1a21ec6945fb368a37f4d2b9d6a326b68edc096a192ed05\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dnwgr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://03406e726f817bcedb841f81959ab12808ba35a530240f9abc5fa5c608b9a6ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dnwgr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9fbbeae0409d2e47e8de9a4306f63b0d13acbfd73cd86f83303fec14352ed718\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dnwgr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://55791327a370c0009c523b86b2234e82d3794985f2eaccf1922eb90f46b2b970\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dnwgr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cee5de93bc68345f9b737857f02832f56582d58757c50a8798eec25d19b2c20d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dnwgr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d314eb248e90371fed068b1173e26d4c9f9765417a70b341e1a5b136ec17353a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dnwgr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7bf11dd48efebb9da2b836bfad6a54b274247fa60902dab2f084793e4ad51a89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7bf11dd48efebb9da2b836bfad6a54b274247fa60902dab2f084793e4ad51a89\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-08T18:11:42Z\\\",\\\"message\\\":\\\" V4IPs:[]string(nil), V6IPs:[]string(nil)}, nodeEndpoints:map[string]services.lbEndpoints{}, externalTrafficLocal:false, internalTrafficLocal:false, hasNodePort:false}}\\\\nI1008 18:11:42.211985 6628 services_controller.go:444] Built service openshift-authentication-operator/metrics LB per-node configs for network=default: []services.lbConfig(nil)\\\\nI1008 18:11:42.211998 6628 services_controller.go:445] Built service openshift-authentication-operator/metrics LB template configs for network=default: []services.lbConfig(nil)\\\\nF1008 18:11:42.212009 6628 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:42Z is after 2025-08-\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T18:11:41Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-nsc65_openshift-ovn-kubernetes(f010a463-69f2-4082-acc3-161899a939c3)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dnwgr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://74fcb812b84291763cc6e5ccb20c12aaad580ad968dda4734a0bbfe33d2770ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dnwgr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://befe684cea9bd6dfda3ec8e85bca431c26bd8f82823b0a69dc0d04319a3ca385\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://befe684cea9bd6dfda3ec8e85bca431c26bd8f82823b0a69dc0d04319a3ca385\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dnwgr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:11:16Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-nsc65\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:44Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:44 crc kubenswrapper[4988]: I1008 18:11:44.036129 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-7wl8v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e433488d-db52-435f-b89c-f63cd477c5d3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76b30d183684740b647c1af11dc49712232cc8b0c9a392b491580e5ddb75ea7f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dbz6l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e55016c3e1b97d1b3a857130a5c8e2ec6da2333f786032118a61e3e2ad4c5651\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dbz6l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:11:29Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-7wl8v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:44Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:44 crc kubenswrapper[4988]: I1008 18:11:44.051067 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:11:44 crc kubenswrapper[4988]: I1008 18:11:44.051119 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:11:44 crc kubenswrapper[4988]: I1008 18:11:44.051129 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:11:44 crc kubenswrapper[4988]: I1008 18:11:44.051153 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:11:44 crc kubenswrapper[4988]: I1008 18:11:44.051165 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:11:44Z","lastTransitionTime":"2025-10-08T18:11:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:11:44 crc kubenswrapper[4988]: I1008 18:11:44.056782 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-jn2sx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"171e0902-95ee-4a39-8b1b-a58704c30574\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9f4q2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9f4q2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:11:31Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-jn2sx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:44Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:44 crc kubenswrapper[4988]: I1008 18:11:44.077910 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:44Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:44 crc kubenswrapper[4988]: I1008 18:11:44.155248 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:11:44 crc kubenswrapper[4988]: I1008 18:11:44.155682 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:11:44 crc kubenswrapper[4988]: I1008 18:11:44.155833 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:11:44 crc kubenswrapper[4988]: I1008 18:11:44.155987 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:11:44 crc kubenswrapper[4988]: I1008 18:11:44.156123 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:11:44Z","lastTransitionTime":"2025-10-08T18:11:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:11:44 crc kubenswrapper[4988]: I1008 18:11:44.259463 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:11:44 crc kubenswrapper[4988]: I1008 18:11:44.259550 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:11:44 crc kubenswrapper[4988]: I1008 18:11:44.259572 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:11:44 crc kubenswrapper[4988]: I1008 18:11:44.259603 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:11:44 crc kubenswrapper[4988]: I1008 18:11:44.259626 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:11:44Z","lastTransitionTime":"2025-10-08T18:11:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:11:44 crc kubenswrapper[4988]: I1008 18:11:44.389862 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:11:44 crc kubenswrapper[4988]: I1008 18:11:44.389936 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:11:44 crc kubenswrapper[4988]: I1008 18:11:44.389963 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:11:44 crc kubenswrapper[4988]: I1008 18:11:44.389998 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:11:44 crc kubenswrapper[4988]: I1008 18:11:44.390021 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:11:44Z","lastTransitionTime":"2025-10-08T18:11:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:11:44 crc kubenswrapper[4988]: I1008 18:11:44.493289 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:11:44 crc kubenswrapper[4988]: I1008 18:11:44.493334 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:11:44 crc kubenswrapper[4988]: I1008 18:11:44.493345 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:11:44 crc kubenswrapper[4988]: I1008 18:11:44.493367 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:11:44 crc kubenswrapper[4988]: I1008 18:11:44.493399 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:11:44Z","lastTransitionTime":"2025-10-08T18:11:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:11:44 crc kubenswrapper[4988]: I1008 18:11:44.596691 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:11:44 crc kubenswrapper[4988]: I1008 18:11:44.596758 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:11:44 crc kubenswrapper[4988]: I1008 18:11:44.596776 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:11:44 crc kubenswrapper[4988]: I1008 18:11:44.596800 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:11:44 crc kubenswrapper[4988]: I1008 18:11:44.596817 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:11:44Z","lastTransitionTime":"2025-10-08T18:11:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:11:44 crc kubenswrapper[4988]: I1008 18:11:44.700250 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:11:44 crc kubenswrapper[4988]: I1008 18:11:44.700300 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:11:44 crc kubenswrapper[4988]: I1008 18:11:44.700311 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:11:44 crc kubenswrapper[4988]: I1008 18:11:44.700330 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:11:44 crc kubenswrapper[4988]: I1008 18:11:44.700343 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:11:44Z","lastTransitionTime":"2025-10-08T18:11:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:11:44 crc kubenswrapper[4988]: I1008 18:11:44.802841 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:11:44 crc kubenswrapper[4988]: I1008 18:11:44.802892 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:11:44 crc kubenswrapper[4988]: I1008 18:11:44.802900 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:11:44 crc kubenswrapper[4988]: I1008 18:11:44.802913 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:11:44 crc kubenswrapper[4988]: I1008 18:11:44.802924 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:11:44Z","lastTransitionTime":"2025-10-08T18:11:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:11:44 crc kubenswrapper[4988]: I1008 18:11:44.907492 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:11:44 crc kubenswrapper[4988]: I1008 18:11:44.907549 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:11:44 crc kubenswrapper[4988]: I1008 18:11:44.907561 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:11:44 crc kubenswrapper[4988]: I1008 18:11:44.907583 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:11:44 crc kubenswrapper[4988]: I1008 18:11:44.907598 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:11:44Z","lastTransitionTime":"2025-10-08T18:11:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:11:45 crc kubenswrapper[4988]: I1008 18:11:45.010763 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:11:45 crc kubenswrapper[4988]: I1008 18:11:45.010809 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:11:45 crc kubenswrapper[4988]: I1008 18:11:45.010822 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:11:45 crc kubenswrapper[4988]: I1008 18:11:45.010839 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:11:45 crc kubenswrapper[4988]: I1008 18:11:45.010853 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:11:45Z","lastTransitionTime":"2025-10-08T18:11:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:11:45 crc kubenswrapper[4988]: I1008 18:11:45.114244 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:11:45 crc kubenswrapper[4988]: I1008 18:11:45.114298 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:11:45 crc kubenswrapper[4988]: I1008 18:11:45.114309 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:11:45 crc kubenswrapper[4988]: I1008 18:11:45.114331 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:11:45 crc kubenswrapper[4988]: I1008 18:11:45.114345 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:11:45Z","lastTransitionTime":"2025-10-08T18:11:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:11:45 crc kubenswrapper[4988]: I1008 18:11:45.217361 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:11:45 crc kubenswrapper[4988]: I1008 18:11:45.217420 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:11:45 crc kubenswrapper[4988]: I1008 18:11:45.217432 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:11:45 crc kubenswrapper[4988]: I1008 18:11:45.217451 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:11:45 crc kubenswrapper[4988]: I1008 18:11:45.217461 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:11:45Z","lastTransitionTime":"2025-10-08T18:11:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:11:45 crc kubenswrapper[4988]: I1008 18:11:45.237017 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jn2sx" Oct 08 18:11:45 crc kubenswrapper[4988]: I1008 18:11:45.237079 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 18:11:45 crc kubenswrapper[4988]: I1008 18:11:45.237087 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 18:11:45 crc kubenswrapper[4988]: I1008 18:11:45.237096 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 18:11:45 crc kubenswrapper[4988]: E1008 18:11:45.237275 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jn2sx" podUID="171e0902-95ee-4a39-8b1b-a58704c30574" Oct 08 18:11:45 crc kubenswrapper[4988]: E1008 18:11:45.237477 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 08 18:11:45 crc kubenswrapper[4988]: E1008 18:11:45.237556 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 08 18:11:45 crc kubenswrapper[4988]: E1008 18:11:45.237643 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 08 18:11:45 crc kubenswrapper[4988]: I1008 18:11:45.254978 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-s4f2k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ba5e11d7-7bde-4c3d-983b-884232219652\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://824bc34a2807068d3fceb10651803ae9ae5651e940f7aed4877c544d76dae4d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hn5jf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:11:15Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-s4f2k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:45Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:45 crc kubenswrapper[4988]: I1008 18:11:45.270551 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-dzxs2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8f307b20-5e33-4725-99cb-210e25113532\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72ee803c6ef88524e0ce0391476f8e5a7dfbc858b095a3c2ae31bad10991ce3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q65f7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:11:16Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-dzxs2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:45Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:45 crc kubenswrapper[4988]: I1008 18:11:45.291873 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aef71be1-1d12-46e5-95b2-bd399040fd4d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:10:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:10:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:10:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77c68f3ecb8a0724499e495cb26ef3db5beecb140f2558d8c0b25f41928b84ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:10:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a560ee84c2405227de71fda127566213cd4f017e73d5581442ba267781b1c539\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:10:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6100accc3bd08928bc1ff88310f29fa5da37fd2d5c315c94279801c9778c4745\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:10:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b0e596d7ce4e9341c44499af79d3b3caf0e5e2c156902ea40cf2a3dca672be9e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5bd35aef24be034a0b305a40bb53ca171a4f76db3b93b4f78ae95099b0a4cde3\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-08T18:11:14Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1008 18:11:09.091034 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1008 18:11:09.091984 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3472374062/tls.crt::/tmp/serving-cert-3472374062/tls.key\\\\\\\"\\\\nI1008 18:11:14.619696 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1008 18:11:14.628837 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1008 18:11:14.628898 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1008 18:11:14.628994 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1008 18:11:14.629036 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1008 18:11:14.641318 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1008 18:11:14.641363 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1008 18:11:14.641372 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1008 18:11:14.641382 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1008 18:11:14.641415 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1008 18:11:14.641423 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1008 18:11:14.641431 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1008 18:11:14.641484 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1008 18:11:14.646451 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T18:10:58Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c79ef704fda1bbcc486f4c7afa3039a3ab38669833f0f66bafe24dd4cc679e60\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:10:58Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://28d21731af0f16d6a8ae6c6e9d0bebf0560d4a533db4c78c10bf1837890394ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://28d21731af0f16d6a8ae6c6e9d0bebf0560d4a533db4c78c10bf1837890394ef\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:10:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:10:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:10:55Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:45Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:45 crc kubenswrapper[4988]: I1008 18:11:45.308960 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7e0ecefc-359f-4785-bb79-0343a62a25da\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:10:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:10:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:10:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://30bec350fefee3d19758b3ed735ede4e9aa085edb57aa3f3ec24249a67dc7ca2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:10:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2d850fd27d4ed65891ff69b9add65fb78535357f54819c2a8d196f5a340c58e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:10:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://90b1a9e14c82bf0f75f08fdb342eed35a503020f9db641e5e9ceff8b122cb3bd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:10:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://16f23c4c7163d32dc0df8d6a7ee17e2864fb261f8a40ecf05556774b173fc868\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:10:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:10:55Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:45Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:45 crc kubenswrapper[4988]: I1008 18:11:45.320743 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:11:45 crc kubenswrapper[4988]: I1008 18:11:45.320805 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:11:45 crc kubenswrapper[4988]: I1008 18:11:45.320823 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:11:45 crc kubenswrapper[4988]: I1008 18:11:45.320851 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:11:45 crc kubenswrapper[4988]: I1008 18:11:45.320869 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:11:45Z","lastTransitionTime":"2025-10-08T18:11:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:11:45 crc kubenswrapper[4988]: I1008 18:11:45.325911 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e8906182d55c285b3cf095e2d9e1dac397af4618a514f11e26b44c172dd573f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:45Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:45 crc kubenswrapper[4988]: I1008 18:11:45.342441 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:45Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:45 crc kubenswrapper[4988]: I1008 18:11:45.362800 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-jnh7m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"556fb048-8afb-44a5-9ea0-2b7c89e6cf4b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bfa2f3382b23c3f788a1b31e80e8a90b3b6868870dccc46d86c3fc11ad027f52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmlt2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1a825886f105ccf9a7cc6f28a0cd012fb859f4990dade0e6aae932482146028\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1a825886f105ccf9a7cc6f28a0cd012fb859f4990dade0e6aae932482146028\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmlt2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ab2fe6e65418efa75bae72259157a2323bbb4b7d11d1d012053524e14135888\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0ab2fe6e65418efa75bae72259157a2323bbb4b7d11d1d012053524e14135888\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:11:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:11:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmlt2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://368a303309be112d9c0b7ccd6538fbaca7085d97a8ec8259f36092f07147e7cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://368a303309be112d9c0b7ccd6538fbaca7085d97a8ec8259f36092f07147e7cb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:11:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:11:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmlt2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a8ca69001dddef643f982a97c74c291e30cb9fac5b0e894e80066b885ea9151\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9a8ca69001dddef643f982a97c74c291e30cb9fac5b0e894e80066b885ea9151\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:11:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:11:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmlt2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://145303c46b48603ed1aae4b9bfe06332a3ffad28f206f8394574011a87b58cd3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://145303c46b48603ed1aae4b9bfe06332a3ffad28f206f8394574011a87b58cd3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:11:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:11:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmlt2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d08e5cfc7cbe5ec1a7a0265ed7c466363c3abeec492263170860689b46c7ff95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d08e5cfc7cbe5ec1a7a0265ed7c466363c3abeec492263170860689b46c7ff95\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:11:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:11:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmlt2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:11:16Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-jnh7m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:45Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:45 crc kubenswrapper[4988]: I1008 18:11:45.379117 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f328a20b-5283-407e-8a93-2ab2da599f18\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8622d56fb198dadee31fb0ca8fc80d56828b6710648055f264533e2d5bdb3037\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z2gh4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3cdb2a757332bdcdc2a2f3362913fe8f06003b0597de20b72627f9badd5e30af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z2gh4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:11:16Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-nm54f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:45Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:45 crc kubenswrapper[4988]: I1008 18:11:45.393990 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a35a5e989ba69c8ef36ce4cea272aadf67ce20473fc257adfcfa86a2faba6ae0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a522e937cfe03fb9f2e60cb1bea07214cf63c69bdecd5ba65a2dfdba882d28dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:45Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:45 crc kubenswrapper[4988]: I1008 18:11:45.410662 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9590ca0059f0ebbe86343cccb3475338c317382c3231e082655b61e67eece375\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:45Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:45 crc kubenswrapper[4988]: I1008 18:11:45.423549 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:11:45 crc kubenswrapper[4988]: I1008 18:11:45.423597 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:11:45 crc kubenswrapper[4988]: I1008 18:11:45.423611 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:11:45 crc kubenswrapper[4988]: I1008 18:11:45.423637 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:11:45 crc kubenswrapper[4988]: I1008 18:11:45.423658 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:11:45Z","lastTransitionTime":"2025-10-08T18:11:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:11:45 crc kubenswrapper[4988]: I1008 18:11:45.430611 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-h9jwp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4490d846-ca32-4cd5-9202-87cc6d499fb4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://00b7f6db043c28efda3a09fdc77a6ccded3eae8b6a3d2852aa500789d5f01194\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7k6rk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:11:16Z\\\"}}\" for pod \"openshift-multus\"/\"multus-h9jwp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:45Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:45 crc kubenswrapper[4988]: I1008 18:11:45.480252 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-nsc65" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f010a463-69f2-4082-acc3-161899a939c3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://48a516f6c153743bd1a21ec6945fb368a37f4d2b9d6a326b68edc096a192ed05\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dnwgr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://03406e726f817bcedb841f81959ab12808ba35a530240f9abc5fa5c608b9a6ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dnwgr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9fbbeae0409d2e47e8de9a4306f63b0d13acbfd73cd86f83303fec14352ed718\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dnwgr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://55791327a370c0009c523b86b2234e82d3794985f2eaccf1922eb90f46b2b970\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dnwgr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cee5de93bc68345f9b737857f02832f56582d58757c50a8798eec25d19b2c20d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dnwgr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d314eb248e90371fed068b1173e26d4c9f9765417a70b341e1a5b136ec17353a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dnwgr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7bf11dd48efebb9da2b836bfad6a54b274247fa60902dab2f084793e4ad51a89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7bf11dd48efebb9da2b836bfad6a54b274247fa60902dab2f084793e4ad51a89\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-08T18:11:42Z\\\",\\\"message\\\":\\\" V4IPs:[]string(nil), V6IPs:[]string(nil)}, nodeEndpoints:map[string]services.lbEndpoints{}, externalTrafficLocal:false, internalTrafficLocal:false, hasNodePort:false}}\\\\nI1008 18:11:42.211985 6628 services_controller.go:444] Built service openshift-authentication-operator/metrics LB per-node configs for network=default: []services.lbConfig(nil)\\\\nI1008 18:11:42.211998 6628 services_controller.go:445] Built service openshift-authentication-operator/metrics LB template configs for network=default: []services.lbConfig(nil)\\\\nF1008 18:11:42.212009 6628 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:42Z is after 2025-08-\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T18:11:41Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-nsc65_openshift-ovn-kubernetes(f010a463-69f2-4082-acc3-161899a939c3)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dnwgr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://74fcb812b84291763cc6e5ccb20c12aaad580ad968dda4734a0bbfe33d2770ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dnwgr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://befe684cea9bd6dfda3ec8e85bca431c26bd8f82823b0a69dc0d04319a3ca385\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://befe684cea9bd6dfda3ec8e85bca431c26bd8f82823b0a69dc0d04319a3ca385\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dnwgr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:11:16Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-nsc65\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:45Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:45 crc kubenswrapper[4988]: I1008 18:11:45.504765 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-7wl8v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e433488d-db52-435f-b89c-f63cd477c5d3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76b30d183684740b647c1af11dc49712232cc8b0c9a392b491580e5ddb75ea7f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dbz6l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e55016c3e1b97d1b3a857130a5c8e2ec6da2333f786032118a61e3e2ad4c5651\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dbz6l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:11:29Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-7wl8v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:45Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:45 crc kubenswrapper[4988]: I1008 18:11:45.525743 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-jn2sx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"171e0902-95ee-4a39-8b1b-a58704c30574\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9f4q2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9f4q2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:11:31Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-jn2sx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:45Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:45 crc kubenswrapper[4988]: I1008 18:11:45.526499 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:11:45 crc kubenswrapper[4988]: I1008 18:11:45.526526 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:11:45 crc kubenswrapper[4988]: I1008 18:11:45.526536 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:11:45 crc kubenswrapper[4988]: I1008 18:11:45.526553 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:11:45 crc kubenswrapper[4988]: I1008 18:11:45.526565 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:11:45Z","lastTransitionTime":"2025-10-08T18:11:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:11:45 crc kubenswrapper[4988]: I1008 18:11:45.550603 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:45Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:45 crc kubenswrapper[4988]: I1008 18:11:45.565495 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:45Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:45 crc kubenswrapper[4988]: I1008 18:11:45.629201 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:11:45 crc kubenswrapper[4988]: I1008 18:11:45.629256 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:11:45 crc kubenswrapper[4988]: I1008 18:11:45.629269 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:11:45 crc kubenswrapper[4988]: I1008 18:11:45.629290 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:11:45 crc kubenswrapper[4988]: I1008 18:11:45.629304 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:11:45Z","lastTransitionTime":"2025-10-08T18:11:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:11:45 crc kubenswrapper[4988]: I1008 18:11:45.731646 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:11:45 crc kubenswrapper[4988]: I1008 18:11:45.731725 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:11:45 crc kubenswrapper[4988]: I1008 18:11:45.731755 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:11:45 crc kubenswrapper[4988]: I1008 18:11:45.731792 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:11:45 crc kubenswrapper[4988]: I1008 18:11:45.731819 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:11:45Z","lastTransitionTime":"2025-10-08T18:11:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:11:45 crc kubenswrapper[4988]: I1008 18:11:45.840734 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:11:45 crc kubenswrapper[4988]: I1008 18:11:45.840810 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:11:45 crc kubenswrapper[4988]: I1008 18:11:45.840830 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:11:45 crc kubenswrapper[4988]: I1008 18:11:45.840861 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:11:45 crc kubenswrapper[4988]: I1008 18:11:45.840881 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:11:45Z","lastTransitionTime":"2025-10-08T18:11:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:11:45 crc kubenswrapper[4988]: I1008 18:11:45.943943 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:11:45 crc kubenswrapper[4988]: I1008 18:11:45.944009 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:11:45 crc kubenswrapper[4988]: I1008 18:11:45.944026 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:11:45 crc kubenswrapper[4988]: I1008 18:11:45.944053 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:11:45 crc kubenswrapper[4988]: I1008 18:11:45.944074 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:11:45Z","lastTransitionTime":"2025-10-08T18:11:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:11:46 crc kubenswrapper[4988]: I1008 18:11:46.047124 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:11:46 crc kubenswrapper[4988]: I1008 18:11:46.047175 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:11:46 crc kubenswrapper[4988]: I1008 18:11:46.047192 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:11:46 crc kubenswrapper[4988]: I1008 18:11:46.047218 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:11:46 crc kubenswrapper[4988]: I1008 18:11:46.047236 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:11:46Z","lastTransitionTime":"2025-10-08T18:11:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:11:46 crc kubenswrapper[4988]: I1008 18:11:46.150195 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:11:46 crc kubenswrapper[4988]: I1008 18:11:46.150258 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:11:46 crc kubenswrapper[4988]: I1008 18:11:46.150277 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:11:46 crc kubenswrapper[4988]: I1008 18:11:46.150302 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:11:46 crc kubenswrapper[4988]: I1008 18:11:46.150321 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:11:46Z","lastTransitionTime":"2025-10-08T18:11:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:11:46 crc kubenswrapper[4988]: I1008 18:11:46.253109 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:11:46 crc kubenswrapper[4988]: I1008 18:11:46.253171 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:11:46 crc kubenswrapper[4988]: I1008 18:11:46.253181 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:11:46 crc kubenswrapper[4988]: I1008 18:11:46.253203 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:11:46 crc kubenswrapper[4988]: I1008 18:11:46.253215 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:11:46Z","lastTransitionTime":"2025-10-08T18:11:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:11:46 crc kubenswrapper[4988]: I1008 18:11:46.357302 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:11:46 crc kubenswrapper[4988]: I1008 18:11:46.357378 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:11:46 crc kubenswrapper[4988]: I1008 18:11:46.357437 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:11:46 crc kubenswrapper[4988]: I1008 18:11:46.357471 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:11:46 crc kubenswrapper[4988]: I1008 18:11:46.357492 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:11:46Z","lastTransitionTime":"2025-10-08T18:11:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:11:46 crc kubenswrapper[4988]: I1008 18:11:46.461055 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:11:46 crc kubenswrapper[4988]: I1008 18:11:46.461123 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:11:46 crc kubenswrapper[4988]: I1008 18:11:46.461142 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:11:46 crc kubenswrapper[4988]: I1008 18:11:46.461167 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:11:46 crc kubenswrapper[4988]: I1008 18:11:46.461185 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:11:46Z","lastTransitionTime":"2025-10-08T18:11:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:11:46 crc kubenswrapper[4988]: I1008 18:11:46.564703 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:11:46 crc kubenswrapper[4988]: I1008 18:11:46.564756 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:11:46 crc kubenswrapper[4988]: I1008 18:11:46.564766 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:11:46 crc kubenswrapper[4988]: I1008 18:11:46.564781 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:11:46 crc kubenswrapper[4988]: I1008 18:11:46.564794 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:11:46Z","lastTransitionTime":"2025-10-08T18:11:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:11:46 crc kubenswrapper[4988]: I1008 18:11:46.668875 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:11:46 crc kubenswrapper[4988]: I1008 18:11:46.668922 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:11:46 crc kubenswrapper[4988]: I1008 18:11:46.668935 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:11:46 crc kubenswrapper[4988]: I1008 18:11:46.668955 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:11:46 crc kubenswrapper[4988]: I1008 18:11:46.668968 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:11:46Z","lastTransitionTime":"2025-10-08T18:11:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:11:46 crc kubenswrapper[4988]: I1008 18:11:46.771933 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:11:46 crc kubenswrapper[4988]: I1008 18:11:46.771988 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:11:46 crc kubenswrapper[4988]: I1008 18:11:46.772002 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:11:46 crc kubenswrapper[4988]: I1008 18:11:46.772024 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:11:46 crc kubenswrapper[4988]: I1008 18:11:46.772039 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:11:46Z","lastTransitionTime":"2025-10-08T18:11:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:11:46 crc kubenswrapper[4988]: I1008 18:11:46.874577 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:11:46 crc kubenswrapper[4988]: I1008 18:11:46.874651 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:11:46 crc kubenswrapper[4988]: I1008 18:11:46.874673 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:11:46 crc kubenswrapper[4988]: I1008 18:11:46.874704 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:11:46 crc kubenswrapper[4988]: I1008 18:11:46.874728 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:11:46Z","lastTransitionTime":"2025-10-08T18:11:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:11:46 crc kubenswrapper[4988]: I1008 18:11:46.927498 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 18:11:46 crc kubenswrapper[4988]: I1008 18:11:46.927689 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 18:11:46 crc kubenswrapper[4988]: E1008 18:11:46.927806 4988 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 18:12:18.927758204 +0000 UTC m=+84.377601014 (durationBeforeRetry 32s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 18:11:46 crc kubenswrapper[4988]: E1008 18:11:46.927838 4988 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 08 18:11:46 crc kubenswrapper[4988]: E1008 18:11:46.927903 4988 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-08 18:12:18.927891397 +0000 UTC m=+84.377734167 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 08 18:11:46 crc kubenswrapper[4988]: I1008 18:11:46.927943 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/171e0902-95ee-4a39-8b1b-a58704c30574-metrics-certs\") pod \"network-metrics-daemon-jn2sx\" (UID: \"171e0902-95ee-4a39-8b1b-a58704c30574\") " pod="openshift-multus/network-metrics-daemon-jn2sx" Oct 08 18:11:46 crc kubenswrapper[4988]: I1008 18:11:46.928103 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 18:11:46 crc kubenswrapper[4988]: E1008 18:11:46.928204 4988 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 08 18:11:46 crc kubenswrapper[4988]: E1008 18:11:46.928194 4988 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 08 18:11:46 crc kubenswrapper[4988]: E1008 18:11:46.928250 4988 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-08 18:12:18.928239487 +0000 UTC m=+84.378082257 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 08 18:11:46 crc kubenswrapper[4988]: E1008 18:11:46.928335 4988 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/171e0902-95ee-4a39-8b1b-a58704c30574-metrics-certs podName:171e0902-95ee-4a39-8b1b-a58704c30574 nodeName:}" failed. No retries permitted until 2025-10-08 18:12:02.928298159 +0000 UTC m=+68.378140969 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/171e0902-95ee-4a39-8b1b-a58704c30574-metrics-certs") pod "network-metrics-daemon-jn2sx" (UID: "171e0902-95ee-4a39-8b1b-a58704c30574") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 08 18:11:46 crc kubenswrapper[4988]: I1008 18:11:46.978066 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:11:46 crc kubenswrapper[4988]: I1008 18:11:46.978136 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:11:46 crc kubenswrapper[4988]: I1008 18:11:46.978159 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:11:46 crc kubenswrapper[4988]: I1008 18:11:46.978190 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:11:46 crc kubenswrapper[4988]: I1008 18:11:46.978212 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:11:46Z","lastTransitionTime":"2025-10-08T18:11:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:11:47 crc kubenswrapper[4988]: I1008 18:11:47.029163 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 18:11:47 crc kubenswrapper[4988]: I1008 18:11:47.029322 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 18:11:47 crc kubenswrapper[4988]: E1008 18:11:47.029462 4988 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 08 18:11:47 crc kubenswrapper[4988]: E1008 18:11:47.029514 4988 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 08 18:11:47 crc kubenswrapper[4988]: E1008 18:11:47.029541 4988 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 08 18:11:47 crc kubenswrapper[4988]: E1008 18:11:47.029534 4988 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 08 18:11:47 crc kubenswrapper[4988]: E1008 18:11:47.029612 4988 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 08 18:11:47 crc kubenswrapper[4988]: E1008 18:11:47.029635 4988 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 08 18:11:47 crc kubenswrapper[4988]: E1008 18:11:47.029641 4988 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-08 18:12:19.029603201 +0000 UTC m=+84.479446011 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 08 18:11:47 crc kubenswrapper[4988]: E1008 18:11:47.029701 4988 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-08 18:12:19.029678203 +0000 UTC m=+84.479521013 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 08 18:11:47 crc kubenswrapper[4988]: I1008 18:11:47.081163 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:11:47 crc kubenswrapper[4988]: I1008 18:11:47.081209 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:11:47 crc kubenswrapper[4988]: I1008 18:11:47.081222 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:11:47 crc kubenswrapper[4988]: I1008 18:11:47.081241 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:11:47 crc kubenswrapper[4988]: I1008 18:11:47.081253 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:11:47Z","lastTransitionTime":"2025-10-08T18:11:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:11:47 crc kubenswrapper[4988]: I1008 18:11:47.184321 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:11:47 crc kubenswrapper[4988]: I1008 18:11:47.184430 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:11:47 crc kubenswrapper[4988]: I1008 18:11:47.184458 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:11:47 crc kubenswrapper[4988]: I1008 18:11:47.184491 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:11:47 crc kubenswrapper[4988]: I1008 18:11:47.184516 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:11:47Z","lastTransitionTime":"2025-10-08T18:11:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:11:47 crc kubenswrapper[4988]: I1008 18:11:47.237665 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 18:11:47 crc kubenswrapper[4988]: I1008 18:11:47.237704 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 18:11:47 crc kubenswrapper[4988]: I1008 18:11:47.237749 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jn2sx" Oct 08 18:11:47 crc kubenswrapper[4988]: I1008 18:11:47.237871 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 18:11:47 crc kubenswrapper[4988]: E1008 18:11:47.237867 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 08 18:11:47 crc kubenswrapper[4988]: E1008 18:11:47.237972 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 08 18:11:47 crc kubenswrapper[4988]: E1008 18:11:47.238139 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 08 18:11:47 crc kubenswrapper[4988]: E1008 18:11:47.238316 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jn2sx" podUID="171e0902-95ee-4a39-8b1b-a58704c30574" Oct 08 18:11:47 crc kubenswrapper[4988]: I1008 18:11:47.288080 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:11:47 crc kubenswrapper[4988]: I1008 18:11:47.288164 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:11:47 crc kubenswrapper[4988]: I1008 18:11:47.288187 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:11:47 crc kubenswrapper[4988]: I1008 18:11:47.288222 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:11:47 crc kubenswrapper[4988]: I1008 18:11:47.288244 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:11:47Z","lastTransitionTime":"2025-10-08T18:11:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:11:47 crc kubenswrapper[4988]: I1008 18:11:47.392004 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:11:47 crc kubenswrapper[4988]: I1008 18:11:47.392065 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:11:47 crc kubenswrapper[4988]: I1008 18:11:47.392082 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:11:47 crc kubenswrapper[4988]: I1008 18:11:47.392110 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:11:47 crc kubenswrapper[4988]: I1008 18:11:47.392129 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:11:47Z","lastTransitionTime":"2025-10-08T18:11:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:11:47 crc kubenswrapper[4988]: I1008 18:11:47.495697 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:11:47 crc kubenswrapper[4988]: I1008 18:11:47.495784 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:11:47 crc kubenswrapper[4988]: I1008 18:11:47.495804 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:11:47 crc kubenswrapper[4988]: I1008 18:11:47.495836 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:11:47 crc kubenswrapper[4988]: I1008 18:11:47.495859 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:11:47Z","lastTransitionTime":"2025-10-08T18:11:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:11:47 crc kubenswrapper[4988]: I1008 18:11:47.599864 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:11:47 crc kubenswrapper[4988]: I1008 18:11:47.600139 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:11:47 crc kubenswrapper[4988]: I1008 18:11:47.600159 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:11:47 crc kubenswrapper[4988]: I1008 18:11:47.600183 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:11:47 crc kubenswrapper[4988]: I1008 18:11:47.600202 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:11:47Z","lastTransitionTime":"2025-10-08T18:11:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:11:47 crc kubenswrapper[4988]: I1008 18:11:47.703504 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:11:47 crc kubenswrapper[4988]: I1008 18:11:47.703565 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:11:47 crc kubenswrapper[4988]: I1008 18:11:47.703593 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:11:47 crc kubenswrapper[4988]: I1008 18:11:47.703619 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:11:47 crc kubenswrapper[4988]: I1008 18:11:47.703636 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:11:47Z","lastTransitionTime":"2025-10-08T18:11:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:11:47 crc kubenswrapper[4988]: I1008 18:11:47.806827 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:11:47 crc kubenswrapper[4988]: I1008 18:11:47.806896 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:11:47 crc kubenswrapper[4988]: I1008 18:11:47.806912 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:11:47 crc kubenswrapper[4988]: I1008 18:11:47.806938 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:11:47 crc kubenswrapper[4988]: I1008 18:11:47.806957 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:11:47Z","lastTransitionTime":"2025-10-08T18:11:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:11:47 crc kubenswrapper[4988]: I1008 18:11:47.910991 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:11:47 crc kubenswrapper[4988]: I1008 18:11:47.911070 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:11:47 crc kubenswrapper[4988]: I1008 18:11:47.911093 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:11:47 crc kubenswrapper[4988]: I1008 18:11:47.911122 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:11:47 crc kubenswrapper[4988]: I1008 18:11:47.911142 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:11:47Z","lastTransitionTime":"2025-10-08T18:11:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:11:48 crc kubenswrapper[4988]: I1008 18:11:48.014900 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:11:48 crc kubenswrapper[4988]: I1008 18:11:48.014985 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:11:48 crc kubenswrapper[4988]: I1008 18:11:48.015005 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:11:48 crc kubenswrapper[4988]: I1008 18:11:48.015102 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:11:48 crc kubenswrapper[4988]: I1008 18:11:48.015126 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:11:48Z","lastTransitionTime":"2025-10-08T18:11:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:11:48 crc kubenswrapper[4988]: I1008 18:11:48.119895 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:11:48 crc kubenswrapper[4988]: I1008 18:11:48.119979 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:11:48 crc kubenswrapper[4988]: I1008 18:11:48.119998 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:11:48 crc kubenswrapper[4988]: I1008 18:11:48.120029 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:11:48 crc kubenswrapper[4988]: I1008 18:11:48.120050 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:11:48Z","lastTransitionTime":"2025-10-08T18:11:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:11:48 crc kubenswrapper[4988]: I1008 18:11:48.223587 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:11:48 crc kubenswrapper[4988]: I1008 18:11:48.223666 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:11:48 crc kubenswrapper[4988]: I1008 18:11:48.223684 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:11:48 crc kubenswrapper[4988]: I1008 18:11:48.223711 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:11:48 crc kubenswrapper[4988]: I1008 18:11:48.223733 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:11:48Z","lastTransitionTime":"2025-10-08T18:11:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:11:48 crc kubenswrapper[4988]: I1008 18:11:48.330103 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:11:48 crc kubenswrapper[4988]: I1008 18:11:48.330179 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:11:48 crc kubenswrapper[4988]: I1008 18:11:48.330192 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:11:48 crc kubenswrapper[4988]: I1008 18:11:48.330216 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:11:48 crc kubenswrapper[4988]: I1008 18:11:48.330230 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:11:48Z","lastTransitionTime":"2025-10-08T18:11:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:11:48 crc kubenswrapper[4988]: I1008 18:11:48.434545 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:11:48 crc kubenswrapper[4988]: I1008 18:11:48.434642 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:11:48 crc kubenswrapper[4988]: I1008 18:11:48.434668 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:11:48 crc kubenswrapper[4988]: I1008 18:11:48.434705 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:11:48 crc kubenswrapper[4988]: I1008 18:11:48.434729 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:11:48Z","lastTransitionTime":"2025-10-08T18:11:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:11:48 crc kubenswrapper[4988]: I1008 18:11:48.537911 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:11:48 crc kubenswrapper[4988]: I1008 18:11:48.537951 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:11:48 crc kubenswrapper[4988]: I1008 18:11:48.537964 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:11:48 crc kubenswrapper[4988]: I1008 18:11:48.537986 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:11:48 crc kubenswrapper[4988]: I1008 18:11:48.538001 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:11:48Z","lastTransitionTime":"2025-10-08T18:11:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:11:48 crc kubenswrapper[4988]: I1008 18:11:48.642043 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:11:48 crc kubenswrapper[4988]: I1008 18:11:48.642093 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:11:48 crc kubenswrapper[4988]: I1008 18:11:48.642107 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:11:48 crc kubenswrapper[4988]: I1008 18:11:48.642131 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:11:48 crc kubenswrapper[4988]: I1008 18:11:48.642149 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:11:48Z","lastTransitionTime":"2025-10-08T18:11:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:11:48 crc kubenswrapper[4988]: I1008 18:11:48.746028 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:11:48 crc kubenswrapper[4988]: I1008 18:11:48.746096 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:11:48 crc kubenswrapper[4988]: I1008 18:11:48.746114 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:11:48 crc kubenswrapper[4988]: I1008 18:11:48.746142 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:11:48 crc kubenswrapper[4988]: I1008 18:11:48.746164 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:11:48Z","lastTransitionTime":"2025-10-08T18:11:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:11:48 crc kubenswrapper[4988]: I1008 18:11:48.849113 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:11:48 crc kubenswrapper[4988]: I1008 18:11:48.849191 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:11:48 crc kubenswrapper[4988]: I1008 18:11:48.849212 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:11:48 crc kubenswrapper[4988]: I1008 18:11:48.849257 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:11:48 crc kubenswrapper[4988]: I1008 18:11:48.849279 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:11:48Z","lastTransitionTime":"2025-10-08T18:11:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:11:48 crc kubenswrapper[4988]: I1008 18:11:48.952255 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:11:48 crc kubenswrapper[4988]: I1008 18:11:48.952303 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:11:48 crc kubenswrapper[4988]: I1008 18:11:48.952319 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:11:48 crc kubenswrapper[4988]: I1008 18:11:48.952344 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:11:48 crc kubenswrapper[4988]: I1008 18:11:48.952363 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:11:48Z","lastTransitionTime":"2025-10-08T18:11:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:11:49 crc kubenswrapper[4988]: I1008 18:11:49.055968 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:11:49 crc kubenswrapper[4988]: I1008 18:11:49.056020 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:11:49 crc kubenswrapper[4988]: I1008 18:11:49.056032 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:11:49 crc kubenswrapper[4988]: I1008 18:11:49.056054 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:11:49 crc kubenswrapper[4988]: I1008 18:11:49.056066 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:11:49Z","lastTransitionTime":"2025-10-08T18:11:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:11:49 crc kubenswrapper[4988]: I1008 18:11:49.158992 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:11:49 crc kubenswrapper[4988]: I1008 18:11:49.159051 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:11:49 crc kubenswrapper[4988]: I1008 18:11:49.159070 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:11:49 crc kubenswrapper[4988]: I1008 18:11:49.159098 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:11:49 crc kubenswrapper[4988]: I1008 18:11:49.159121 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:11:49Z","lastTransitionTime":"2025-10-08T18:11:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:11:49 crc kubenswrapper[4988]: I1008 18:11:49.237752 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 18:11:49 crc kubenswrapper[4988]: I1008 18:11:49.237855 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 18:11:49 crc kubenswrapper[4988]: E1008 18:11:49.237931 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 08 18:11:49 crc kubenswrapper[4988]: E1008 18:11:49.238056 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 08 18:11:49 crc kubenswrapper[4988]: I1008 18:11:49.238208 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 18:11:49 crc kubenswrapper[4988]: E1008 18:11:49.238303 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 08 18:11:49 crc kubenswrapper[4988]: I1008 18:11:49.238429 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jn2sx" Oct 08 18:11:49 crc kubenswrapper[4988]: E1008 18:11:49.238568 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jn2sx" podUID="171e0902-95ee-4a39-8b1b-a58704c30574" Oct 08 18:11:49 crc kubenswrapper[4988]: I1008 18:11:49.262693 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:11:49 crc kubenswrapper[4988]: I1008 18:11:49.262782 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:11:49 crc kubenswrapper[4988]: I1008 18:11:49.262809 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:11:49 crc kubenswrapper[4988]: I1008 18:11:49.262847 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:11:49 crc kubenswrapper[4988]: I1008 18:11:49.262870 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:11:49Z","lastTransitionTime":"2025-10-08T18:11:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:11:49 crc kubenswrapper[4988]: I1008 18:11:49.365943 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:11:49 crc kubenswrapper[4988]: I1008 18:11:49.366004 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:11:49 crc kubenswrapper[4988]: I1008 18:11:49.366016 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:11:49 crc kubenswrapper[4988]: I1008 18:11:49.366037 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:11:49 crc kubenswrapper[4988]: I1008 18:11:49.366049 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:11:49Z","lastTransitionTime":"2025-10-08T18:11:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:11:49 crc kubenswrapper[4988]: I1008 18:11:49.469836 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:11:49 crc kubenswrapper[4988]: I1008 18:11:49.469901 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:11:49 crc kubenswrapper[4988]: I1008 18:11:49.469921 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:11:49 crc kubenswrapper[4988]: I1008 18:11:49.469947 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:11:49 crc kubenswrapper[4988]: I1008 18:11:49.469966 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:11:49Z","lastTransitionTime":"2025-10-08T18:11:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:11:49 crc kubenswrapper[4988]: I1008 18:11:49.573483 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:11:49 crc kubenswrapper[4988]: I1008 18:11:49.573575 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:11:49 crc kubenswrapper[4988]: I1008 18:11:49.573612 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:11:49 crc kubenswrapper[4988]: I1008 18:11:49.573649 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:11:49 crc kubenswrapper[4988]: I1008 18:11:49.573672 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:11:49Z","lastTransitionTime":"2025-10-08T18:11:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:11:49 crc kubenswrapper[4988]: I1008 18:11:49.677165 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:11:49 crc kubenswrapper[4988]: I1008 18:11:49.677230 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:11:49 crc kubenswrapper[4988]: I1008 18:11:49.677251 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:11:49 crc kubenswrapper[4988]: I1008 18:11:49.677277 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:11:49 crc kubenswrapper[4988]: I1008 18:11:49.677299 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:11:49Z","lastTransitionTime":"2025-10-08T18:11:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:11:49 crc kubenswrapper[4988]: I1008 18:11:49.780066 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:11:49 crc kubenswrapper[4988]: I1008 18:11:49.780120 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:11:49 crc kubenswrapper[4988]: I1008 18:11:49.780134 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:11:49 crc kubenswrapper[4988]: I1008 18:11:49.780154 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:11:49 crc kubenswrapper[4988]: I1008 18:11:49.780167 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:11:49Z","lastTransitionTime":"2025-10-08T18:11:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:11:49 crc kubenswrapper[4988]: I1008 18:11:49.882966 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:11:49 crc kubenswrapper[4988]: I1008 18:11:49.883009 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:11:49 crc kubenswrapper[4988]: I1008 18:11:49.883023 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:11:49 crc kubenswrapper[4988]: I1008 18:11:49.883040 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:11:49 crc kubenswrapper[4988]: I1008 18:11:49.883052 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:11:49Z","lastTransitionTime":"2025-10-08T18:11:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:11:49 crc kubenswrapper[4988]: I1008 18:11:49.985574 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:11:49 crc kubenswrapper[4988]: I1008 18:11:49.985652 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:11:49 crc kubenswrapper[4988]: I1008 18:11:49.985679 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:11:49 crc kubenswrapper[4988]: I1008 18:11:49.985712 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:11:49 crc kubenswrapper[4988]: I1008 18:11:49.985736 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:11:49Z","lastTransitionTime":"2025-10-08T18:11:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:11:50 crc kubenswrapper[4988]: I1008 18:11:50.088707 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:11:50 crc kubenswrapper[4988]: I1008 18:11:50.088784 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:11:50 crc kubenswrapper[4988]: I1008 18:11:50.088813 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:11:50 crc kubenswrapper[4988]: I1008 18:11:50.088846 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:11:50 crc kubenswrapper[4988]: I1008 18:11:50.088870 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:11:50Z","lastTransitionTime":"2025-10-08T18:11:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:11:50 crc kubenswrapper[4988]: I1008 18:11:50.192341 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:11:50 crc kubenswrapper[4988]: I1008 18:11:50.192378 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:11:50 crc kubenswrapper[4988]: I1008 18:11:50.192407 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:11:50 crc kubenswrapper[4988]: I1008 18:11:50.192422 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:11:50 crc kubenswrapper[4988]: I1008 18:11:50.192434 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:11:50Z","lastTransitionTime":"2025-10-08T18:11:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:11:50 crc kubenswrapper[4988]: I1008 18:11:50.208805 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 08 18:11:50 crc kubenswrapper[4988]: I1008 18:11:50.220875 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-s4f2k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ba5e11d7-7bde-4c3d-983b-884232219652\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://824bc34a2807068d3fceb10651803ae9ae5651e940f7aed4877c544d76dae4d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hn5jf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:11:15Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-s4f2k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:50Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:50 crc kubenswrapper[4988]: I1008 18:11:50.221886 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler/openshift-kube-scheduler-crc"] Oct 08 18:11:50 crc kubenswrapper[4988]: I1008 18:11:50.244751 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aef71be1-1d12-46e5-95b2-bd399040fd4d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:10:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:10:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:10:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77c68f3ecb8a0724499e495cb26ef3db5beecb140f2558d8c0b25f41928b84ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:10:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a560ee84c2405227de71fda127566213cd4f017e73d5581442ba267781b1c539\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:10:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6100accc3bd08928bc1ff88310f29fa5da37fd2d5c315c94279801c9778c4745\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:10:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b0e596d7ce4e9341c44499af79d3b3caf0e5e2c156902ea40cf2a3dca672be9e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5bd35aef24be034a0b305a40bb53ca171a4f76db3b93b4f78ae95099b0a4cde3\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-08T18:11:14Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1008 18:11:09.091034 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1008 18:11:09.091984 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3472374062/tls.crt::/tmp/serving-cert-3472374062/tls.key\\\\\\\"\\\\nI1008 18:11:14.619696 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1008 18:11:14.628837 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1008 18:11:14.628898 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1008 18:11:14.628994 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1008 18:11:14.629036 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1008 18:11:14.641318 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1008 18:11:14.641363 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1008 18:11:14.641372 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1008 18:11:14.641382 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1008 18:11:14.641415 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1008 18:11:14.641423 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1008 18:11:14.641431 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1008 18:11:14.641484 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1008 18:11:14.646451 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T18:10:58Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c79ef704fda1bbcc486f4c7afa3039a3ab38669833f0f66bafe24dd4cc679e60\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:10:58Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://28d21731af0f16d6a8ae6c6e9d0bebf0560d4a533db4c78c10bf1837890394ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://28d21731af0f16d6a8ae6c6e9d0bebf0560d4a533db4c78c10bf1837890394ef\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:10:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:10:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:10:55Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:50Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:50 crc kubenswrapper[4988]: I1008 18:11:50.263094 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7e0ecefc-359f-4785-bb79-0343a62a25da\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:10:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:10:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:10:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://30bec350fefee3d19758b3ed735ede4e9aa085edb57aa3f3ec24249a67dc7ca2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:10:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2d850fd27d4ed65891ff69b9add65fb78535357f54819c2a8d196f5a340c58e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:10:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://90b1a9e14c82bf0f75f08fdb342eed35a503020f9db641e5e9ceff8b122cb3bd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:10:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://16f23c4c7163d32dc0df8d6a7ee17e2864fb261f8a40ecf05556774b173fc868\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:10:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:10:55Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:50Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:50 crc kubenswrapper[4988]: I1008 18:11:50.280574 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e8906182d55c285b3cf095e2d9e1dac397af4618a514f11e26b44c172dd573f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:50Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:50 crc kubenswrapper[4988]: I1008 18:11:50.300988 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:11:50 crc kubenswrapper[4988]: I1008 18:11:50.301108 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:11:50 crc kubenswrapper[4988]: I1008 18:11:50.301136 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:11:50 crc kubenswrapper[4988]: I1008 18:11:50.301169 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:11:50 crc kubenswrapper[4988]: I1008 18:11:50.301194 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:11:50Z","lastTransitionTime":"2025-10-08T18:11:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:11:50 crc kubenswrapper[4988]: I1008 18:11:50.301594 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-dzxs2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8f307b20-5e33-4725-99cb-210e25113532\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72ee803c6ef88524e0ce0391476f8e5a7dfbc858b095a3c2ae31bad10991ce3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q65f7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:11:16Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-dzxs2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:50Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:50 crc kubenswrapper[4988]: I1008 18:11:50.317463 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a35a5e989ba69c8ef36ce4cea272aadf67ce20473fc257adfcfa86a2faba6ae0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a522e937cfe03fb9f2e60cb1bea07214cf63c69bdecd5ba65a2dfdba882d28dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:50Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:50 crc kubenswrapper[4988]: I1008 18:11:50.329026 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9590ca0059f0ebbe86343cccb3475338c317382c3231e082655b61e67eece375\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:50Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:50 crc kubenswrapper[4988]: I1008 18:11:50.341083 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:50Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:50 crc kubenswrapper[4988]: I1008 18:11:50.363115 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-jnh7m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"556fb048-8afb-44a5-9ea0-2b7c89e6cf4b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bfa2f3382b23c3f788a1b31e80e8a90b3b6868870dccc46d86c3fc11ad027f52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmlt2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1a825886f105ccf9a7cc6f28a0cd012fb859f4990dade0e6aae932482146028\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1a825886f105ccf9a7cc6f28a0cd012fb859f4990dade0e6aae932482146028\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmlt2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ab2fe6e65418efa75bae72259157a2323bbb4b7d11d1d012053524e14135888\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0ab2fe6e65418efa75bae72259157a2323bbb4b7d11d1d012053524e14135888\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:11:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:11:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmlt2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://368a303309be112d9c0b7ccd6538fbaca7085d97a8ec8259f36092f07147e7cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://368a303309be112d9c0b7ccd6538fbaca7085d97a8ec8259f36092f07147e7cb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:11:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:11:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmlt2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a8ca69001dddef643f982a97c74c291e30cb9fac5b0e894e80066b885ea9151\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9a8ca69001dddef643f982a97c74c291e30cb9fac5b0e894e80066b885ea9151\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:11:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:11:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmlt2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://145303c46b48603ed1aae4b9bfe06332a3ffad28f206f8394574011a87b58cd3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://145303c46b48603ed1aae4b9bfe06332a3ffad28f206f8394574011a87b58cd3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:11:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:11:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmlt2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d08e5cfc7cbe5ec1a7a0265ed7c466363c3abeec492263170860689b46c7ff95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d08e5cfc7cbe5ec1a7a0265ed7c466363c3abeec492263170860689b46c7ff95\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:11:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:11:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmlt2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:11:16Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-jnh7m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:50Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:50 crc kubenswrapper[4988]: I1008 18:11:50.376664 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f328a20b-5283-407e-8a93-2ab2da599f18\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8622d56fb198dadee31fb0ca8fc80d56828b6710648055f264533e2d5bdb3037\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z2gh4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3cdb2a757332bdcdc2a2f3362913fe8f06003b0597de20b72627f9badd5e30af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z2gh4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:11:16Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-nm54f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:50Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:50 crc kubenswrapper[4988]: I1008 18:11:50.392355 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:50Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:50 crc kubenswrapper[4988]: I1008 18:11:50.404512 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:11:50 crc kubenswrapper[4988]: I1008 18:11:50.404563 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:11:50 crc kubenswrapper[4988]: I1008 18:11:50.404575 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:11:50 crc kubenswrapper[4988]: I1008 18:11:50.404598 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:11:50 crc kubenswrapper[4988]: I1008 18:11:50.404614 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:11:50Z","lastTransitionTime":"2025-10-08T18:11:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:11:50 crc kubenswrapper[4988]: I1008 18:11:50.408426 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:50Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:50 crc kubenswrapper[4988]: I1008 18:11:50.424553 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-h9jwp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4490d846-ca32-4cd5-9202-87cc6d499fb4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://00b7f6db043c28efda3a09fdc77a6ccded3eae8b6a3d2852aa500789d5f01194\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7k6rk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:11:16Z\\\"}}\" for pod \"openshift-multus\"/\"multus-h9jwp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:50Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:50 crc kubenswrapper[4988]: I1008 18:11:50.443877 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-nsc65" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f010a463-69f2-4082-acc3-161899a939c3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://48a516f6c153743bd1a21ec6945fb368a37f4d2b9d6a326b68edc096a192ed05\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dnwgr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://03406e726f817bcedb841f81959ab12808ba35a530240f9abc5fa5c608b9a6ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dnwgr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9fbbeae0409d2e47e8de9a4306f63b0d13acbfd73cd86f83303fec14352ed718\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dnwgr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://55791327a370c0009c523b86b2234e82d3794985f2eaccf1922eb90f46b2b970\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dnwgr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cee5de93bc68345f9b737857f02832f56582d58757c50a8798eec25d19b2c20d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dnwgr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d314eb248e90371fed068b1173e26d4c9f9765417a70b341e1a5b136ec17353a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dnwgr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7bf11dd48efebb9da2b836bfad6a54b274247fa60902dab2f084793e4ad51a89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7bf11dd48efebb9da2b836bfad6a54b274247fa60902dab2f084793e4ad51a89\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-08T18:11:42Z\\\",\\\"message\\\":\\\" V4IPs:[]string(nil), V6IPs:[]string(nil)}, nodeEndpoints:map[string]services.lbEndpoints{}, externalTrafficLocal:false, internalTrafficLocal:false, hasNodePort:false}}\\\\nI1008 18:11:42.211985 6628 services_controller.go:444] Built service openshift-authentication-operator/metrics LB per-node configs for network=default: []services.lbConfig(nil)\\\\nI1008 18:11:42.211998 6628 services_controller.go:445] Built service openshift-authentication-operator/metrics LB template configs for network=default: []services.lbConfig(nil)\\\\nF1008 18:11:42.212009 6628 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:42Z is after 2025-08-\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T18:11:41Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-nsc65_openshift-ovn-kubernetes(f010a463-69f2-4082-acc3-161899a939c3)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dnwgr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://74fcb812b84291763cc6e5ccb20c12aaad580ad968dda4734a0bbfe33d2770ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dnwgr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://befe684cea9bd6dfda3ec8e85bca431c26bd8f82823b0a69dc0d04319a3ca385\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://befe684cea9bd6dfda3ec8e85bca431c26bd8f82823b0a69dc0d04319a3ca385\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dnwgr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:11:16Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-nsc65\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:50Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:50 crc kubenswrapper[4988]: I1008 18:11:50.455685 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-7wl8v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e433488d-db52-435f-b89c-f63cd477c5d3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76b30d183684740b647c1af11dc49712232cc8b0c9a392b491580e5ddb75ea7f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dbz6l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e55016c3e1b97d1b3a857130a5c8e2ec6da2333f786032118a61e3e2ad4c5651\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dbz6l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:11:29Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-7wl8v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:50Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:50 crc kubenswrapper[4988]: I1008 18:11:50.468708 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-jn2sx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"171e0902-95ee-4a39-8b1b-a58704c30574\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9f4q2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9f4q2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:11:31Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-jn2sx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:50Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:50 crc kubenswrapper[4988]: I1008 18:11:50.507228 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:11:50 crc kubenswrapper[4988]: I1008 18:11:50.507286 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:11:50 crc kubenswrapper[4988]: I1008 18:11:50.507296 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:11:50 crc kubenswrapper[4988]: I1008 18:11:50.507315 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:11:50 crc kubenswrapper[4988]: I1008 18:11:50.507327 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:11:50Z","lastTransitionTime":"2025-10-08T18:11:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:11:50 crc kubenswrapper[4988]: I1008 18:11:50.611553 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:11:50 crc kubenswrapper[4988]: I1008 18:11:50.611604 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:11:50 crc kubenswrapper[4988]: I1008 18:11:50.611614 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:11:50 crc kubenswrapper[4988]: I1008 18:11:50.611632 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:11:50 crc kubenswrapper[4988]: I1008 18:11:50.611644 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:11:50Z","lastTransitionTime":"2025-10-08T18:11:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:11:50 crc kubenswrapper[4988]: I1008 18:11:50.715598 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:11:50 crc kubenswrapper[4988]: I1008 18:11:50.715661 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:11:50 crc kubenswrapper[4988]: I1008 18:11:50.715681 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:11:50 crc kubenswrapper[4988]: I1008 18:11:50.715709 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:11:50 crc kubenswrapper[4988]: I1008 18:11:50.715732 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:11:50Z","lastTransitionTime":"2025-10-08T18:11:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:11:50 crc kubenswrapper[4988]: I1008 18:11:50.819445 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:11:50 crc kubenswrapper[4988]: I1008 18:11:50.819491 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:11:50 crc kubenswrapper[4988]: I1008 18:11:50.819509 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:11:50 crc kubenswrapper[4988]: I1008 18:11:50.819535 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:11:50 crc kubenswrapper[4988]: I1008 18:11:50.819554 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:11:50Z","lastTransitionTime":"2025-10-08T18:11:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:11:50 crc kubenswrapper[4988]: I1008 18:11:50.923117 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:11:50 crc kubenswrapper[4988]: I1008 18:11:50.923190 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:11:50 crc kubenswrapper[4988]: I1008 18:11:50.923209 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:11:50 crc kubenswrapper[4988]: I1008 18:11:50.923236 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:11:50 crc kubenswrapper[4988]: I1008 18:11:50.923292 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:11:50Z","lastTransitionTime":"2025-10-08T18:11:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:11:51 crc kubenswrapper[4988]: I1008 18:11:51.026553 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:11:51 crc kubenswrapper[4988]: I1008 18:11:51.026615 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:11:51 crc kubenswrapper[4988]: I1008 18:11:51.026633 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:11:51 crc kubenswrapper[4988]: I1008 18:11:51.026658 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:11:51 crc kubenswrapper[4988]: I1008 18:11:51.026687 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:11:51Z","lastTransitionTime":"2025-10-08T18:11:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:11:51 crc kubenswrapper[4988]: I1008 18:11:51.129893 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:11:51 crc kubenswrapper[4988]: I1008 18:11:51.129947 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:11:51 crc kubenswrapper[4988]: I1008 18:11:51.129965 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:11:51 crc kubenswrapper[4988]: I1008 18:11:51.129990 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:11:51 crc kubenswrapper[4988]: I1008 18:11:51.130009 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:11:51Z","lastTransitionTime":"2025-10-08T18:11:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:11:51 crc kubenswrapper[4988]: I1008 18:11:51.233361 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:11:51 crc kubenswrapper[4988]: I1008 18:11:51.233499 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:11:51 crc kubenswrapper[4988]: I1008 18:11:51.233532 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:11:51 crc kubenswrapper[4988]: I1008 18:11:51.233580 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:11:51 crc kubenswrapper[4988]: I1008 18:11:51.233603 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:11:51Z","lastTransitionTime":"2025-10-08T18:11:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:11:51 crc kubenswrapper[4988]: I1008 18:11:51.236807 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 18:11:51 crc kubenswrapper[4988]: I1008 18:11:51.236899 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 18:11:51 crc kubenswrapper[4988]: I1008 18:11:51.237020 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jn2sx" Oct 08 18:11:51 crc kubenswrapper[4988]: I1008 18:11:51.236807 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 18:11:51 crc kubenswrapper[4988]: E1008 18:11:51.237014 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 08 18:11:51 crc kubenswrapper[4988]: E1008 18:11:51.237164 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 08 18:11:51 crc kubenswrapper[4988]: E1008 18:11:51.237529 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 08 18:11:51 crc kubenswrapper[4988]: E1008 18:11:51.237604 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jn2sx" podUID="171e0902-95ee-4a39-8b1b-a58704c30574" Oct 08 18:11:51 crc kubenswrapper[4988]: I1008 18:11:51.336902 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:11:51 crc kubenswrapper[4988]: I1008 18:11:51.336984 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:11:51 crc kubenswrapper[4988]: I1008 18:11:51.337011 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:11:51 crc kubenswrapper[4988]: I1008 18:11:51.337043 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:11:51 crc kubenswrapper[4988]: I1008 18:11:51.337069 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:11:51Z","lastTransitionTime":"2025-10-08T18:11:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:11:51 crc kubenswrapper[4988]: I1008 18:11:51.440727 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:11:51 crc kubenswrapper[4988]: I1008 18:11:51.440802 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:11:51 crc kubenswrapper[4988]: I1008 18:11:51.440826 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:11:51 crc kubenswrapper[4988]: I1008 18:11:51.440860 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:11:51 crc kubenswrapper[4988]: I1008 18:11:51.440882 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:11:51Z","lastTransitionTime":"2025-10-08T18:11:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:11:51 crc kubenswrapper[4988]: I1008 18:11:51.544888 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:11:51 crc kubenswrapper[4988]: I1008 18:11:51.544977 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:11:51 crc kubenswrapper[4988]: I1008 18:11:51.545006 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:11:51 crc kubenswrapper[4988]: I1008 18:11:51.545039 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:11:51 crc kubenswrapper[4988]: I1008 18:11:51.545064 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:11:51Z","lastTransitionTime":"2025-10-08T18:11:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:11:51 crc kubenswrapper[4988]: I1008 18:11:51.649055 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:11:51 crc kubenswrapper[4988]: I1008 18:11:51.649107 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:11:51 crc kubenswrapper[4988]: I1008 18:11:51.649121 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:11:51 crc kubenswrapper[4988]: I1008 18:11:51.649138 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:11:51 crc kubenswrapper[4988]: I1008 18:11:51.649149 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:11:51Z","lastTransitionTime":"2025-10-08T18:11:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:11:51 crc kubenswrapper[4988]: I1008 18:11:51.752270 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:11:51 crc kubenswrapper[4988]: I1008 18:11:51.752355 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:11:51 crc kubenswrapper[4988]: I1008 18:11:51.752370 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:11:51 crc kubenswrapper[4988]: I1008 18:11:51.752418 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:11:51 crc kubenswrapper[4988]: I1008 18:11:51.752434 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:11:51Z","lastTransitionTime":"2025-10-08T18:11:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:11:51 crc kubenswrapper[4988]: I1008 18:11:51.855638 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:11:51 crc kubenswrapper[4988]: I1008 18:11:51.855712 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:11:51 crc kubenswrapper[4988]: I1008 18:11:51.855731 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:11:51 crc kubenswrapper[4988]: I1008 18:11:51.855762 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:11:51 crc kubenswrapper[4988]: I1008 18:11:51.855781 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:11:51Z","lastTransitionTime":"2025-10-08T18:11:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:11:51 crc kubenswrapper[4988]: I1008 18:11:51.958857 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:11:51 crc kubenswrapper[4988]: I1008 18:11:51.958943 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:11:51 crc kubenswrapper[4988]: I1008 18:11:51.958968 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:11:51 crc kubenswrapper[4988]: I1008 18:11:51.958999 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:11:51 crc kubenswrapper[4988]: I1008 18:11:51.959026 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:11:51Z","lastTransitionTime":"2025-10-08T18:11:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:11:52 crc kubenswrapper[4988]: I1008 18:11:52.001700 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:11:52 crc kubenswrapper[4988]: I1008 18:11:52.001789 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:11:52 crc kubenswrapper[4988]: I1008 18:11:52.001816 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:11:52 crc kubenswrapper[4988]: I1008 18:11:52.001849 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:11:52 crc kubenswrapper[4988]: I1008 18:11:52.001881 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:11:52Z","lastTransitionTime":"2025-10-08T18:11:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:11:52 crc kubenswrapper[4988]: E1008 18:11:52.027488 4988 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T18:11:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:52Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T18:11:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:52Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T18:11:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:52Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T18:11:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:52Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"95ee9fd0-8bc9-4ca5-9d79-5ee94d58bfeb\\\",\\\"systemUUID\\\":\\\"bd9489c4-fb7d-486d-b7ce-a13b6d7de36a\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:52Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:52 crc kubenswrapper[4988]: I1008 18:11:52.032563 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:11:52 crc kubenswrapper[4988]: I1008 18:11:52.032650 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:11:52 crc kubenswrapper[4988]: I1008 18:11:52.032668 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:11:52 crc kubenswrapper[4988]: I1008 18:11:52.032688 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:11:52 crc kubenswrapper[4988]: I1008 18:11:52.032705 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:11:52Z","lastTransitionTime":"2025-10-08T18:11:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:11:52 crc kubenswrapper[4988]: E1008 18:11:52.051519 4988 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T18:11:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:52Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T18:11:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:52Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T18:11:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:52Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T18:11:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:52Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"95ee9fd0-8bc9-4ca5-9d79-5ee94d58bfeb\\\",\\\"systemUUID\\\":\\\"bd9489c4-fb7d-486d-b7ce-a13b6d7de36a\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:52Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:52 crc kubenswrapper[4988]: I1008 18:11:52.056485 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:11:52 crc kubenswrapper[4988]: I1008 18:11:52.056546 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:11:52 crc kubenswrapper[4988]: I1008 18:11:52.056572 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:11:52 crc kubenswrapper[4988]: I1008 18:11:52.056606 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:11:52 crc kubenswrapper[4988]: I1008 18:11:52.056632 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:11:52Z","lastTransitionTime":"2025-10-08T18:11:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:11:52 crc kubenswrapper[4988]: E1008 18:11:52.079438 4988 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T18:11:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:52Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T18:11:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:52Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T18:11:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:52Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T18:11:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:52Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"95ee9fd0-8bc9-4ca5-9d79-5ee94d58bfeb\\\",\\\"systemUUID\\\":\\\"bd9489c4-fb7d-486d-b7ce-a13b6d7de36a\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:52Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:52 crc kubenswrapper[4988]: I1008 18:11:52.085512 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:11:52 crc kubenswrapper[4988]: I1008 18:11:52.085564 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:11:52 crc kubenswrapper[4988]: I1008 18:11:52.085582 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:11:52 crc kubenswrapper[4988]: I1008 18:11:52.085605 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:11:52 crc kubenswrapper[4988]: I1008 18:11:52.085623 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:11:52Z","lastTransitionTime":"2025-10-08T18:11:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:11:52 crc kubenswrapper[4988]: E1008 18:11:52.108177 4988 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T18:11:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:52Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T18:11:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:52Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T18:11:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:52Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T18:11:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:52Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"95ee9fd0-8bc9-4ca5-9d79-5ee94d58bfeb\\\",\\\"systemUUID\\\":\\\"bd9489c4-fb7d-486d-b7ce-a13b6d7de36a\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:52Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:52 crc kubenswrapper[4988]: I1008 18:11:52.114205 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:11:52 crc kubenswrapper[4988]: I1008 18:11:52.114278 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:11:52 crc kubenswrapper[4988]: I1008 18:11:52.114297 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:11:52 crc kubenswrapper[4988]: I1008 18:11:52.114324 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:11:52 crc kubenswrapper[4988]: I1008 18:11:52.114344 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:11:52Z","lastTransitionTime":"2025-10-08T18:11:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:11:52 crc kubenswrapper[4988]: E1008 18:11:52.136575 4988 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T18:11:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:52Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T18:11:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:52Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T18:11:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:52Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T18:11:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:52Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"95ee9fd0-8bc9-4ca5-9d79-5ee94d58bfeb\\\",\\\"systemUUID\\\":\\\"bd9489c4-fb7d-486d-b7ce-a13b6d7de36a\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:52Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:52 crc kubenswrapper[4988]: E1008 18:11:52.136716 4988 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Oct 08 18:11:52 crc kubenswrapper[4988]: I1008 18:11:52.138655 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:11:52 crc kubenswrapper[4988]: I1008 18:11:52.138731 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:11:52 crc kubenswrapper[4988]: I1008 18:11:52.138748 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:11:52 crc kubenswrapper[4988]: I1008 18:11:52.138773 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:11:52 crc kubenswrapper[4988]: I1008 18:11:52.138793 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:11:52Z","lastTransitionTime":"2025-10-08T18:11:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:11:52 crc kubenswrapper[4988]: I1008 18:11:52.242033 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:11:52 crc kubenswrapper[4988]: I1008 18:11:52.242078 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:11:52 crc kubenswrapper[4988]: I1008 18:11:52.242093 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:11:52 crc kubenswrapper[4988]: I1008 18:11:52.242110 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:11:52 crc kubenswrapper[4988]: I1008 18:11:52.242123 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:11:52Z","lastTransitionTime":"2025-10-08T18:11:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:11:52 crc kubenswrapper[4988]: I1008 18:11:52.345372 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:11:52 crc kubenswrapper[4988]: I1008 18:11:52.345477 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:11:52 crc kubenswrapper[4988]: I1008 18:11:52.345488 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:11:52 crc kubenswrapper[4988]: I1008 18:11:52.345506 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:11:52 crc kubenswrapper[4988]: I1008 18:11:52.345519 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:11:52Z","lastTransitionTime":"2025-10-08T18:11:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:11:52 crc kubenswrapper[4988]: I1008 18:11:52.449127 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:11:52 crc kubenswrapper[4988]: I1008 18:11:52.449169 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:11:52 crc kubenswrapper[4988]: I1008 18:11:52.449178 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:11:52 crc kubenswrapper[4988]: I1008 18:11:52.449191 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:11:52 crc kubenswrapper[4988]: I1008 18:11:52.449201 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:11:52Z","lastTransitionTime":"2025-10-08T18:11:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:11:52 crc kubenswrapper[4988]: I1008 18:11:52.553069 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:11:52 crc kubenswrapper[4988]: I1008 18:11:52.553124 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:11:52 crc kubenswrapper[4988]: I1008 18:11:52.553139 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:11:52 crc kubenswrapper[4988]: I1008 18:11:52.553161 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:11:52 crc kubenswrapper[4988]: I1008 18:11:52.553178 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:11:52Z","lastTransitionTime":"2025-10-08T18:11:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:11:52 crc kubenswrapper[4988]: I1008 18:11:52.656855 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:11:52 crc kubenswrapper[4988]: I1008 18:11:52.656928 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:11:52 crc kubenswrapper[4988]: I1008 18:11:52.656946 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:11:52 crc kubenswrapper[4988]: I1008 18:11:52.656971 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:11:52 crc kubenswrapper[4988]: I1008 18:11:52.656997 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:11:52Z","lastTransitionTime":"2025-10-08T18:11:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:11:52 crc kubenswrapper[4988]: I1008 18:11:52.759863 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:11:52 crc kubenswrapper[4988]: I1008 18:11:52.759933 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:11:52 crc kubenswrapper[4988]: I1008 18:11:52.759946 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:11:52 crc kubenswrapper[4988]: I1008 18:11:52.759966 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:11:52 crc kubenswrapper[4988]: I1008 18:11:52.759982 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:11:52Z","lastTransitionTime":"2025-10-08T18:11:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:11:52 crc kubenswrapper[4988]: I1008 18:11:52.863799 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:11:52 crc kubenswrapper[4988]: I1008 18:11:52.863877 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:11:52 crc kubenswrapper[4988]: I1008 18:11:52.863902 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:11:52 crc kubenswrapper[4988]: I1008 18:11:52.863936 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:11:52 crc kubenswrapper[4988]: I1008 18:11:52.863960 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:11:52Z","lastTransitionTime":"2025-10-08T18:11:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:11:52 crc kubenswrapper[4988]: I1008 18:11:52.968243 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:11:52 crc kubenswrapper[4988]: I1008 18:11:52.968331 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:11:52 crc kubenswrapper[4988]: I1008 18:11:52.968358 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:11:52 crc kubenswrapper[4988]: I1008 18:11:52.968418 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:11:52 crc kubenswrapper[4988]: I1008 18:11:52.968440 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:11:52Z","lastTransitionTime":"2025-10-08T18:11:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:11:53 crc kubenswrapper[4988]: I1008 18:11:53.073813 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:11:53 crc kubenswrapper[4988]: I1008 18:11:53.073868 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:11:53 crc kubenswrapper[4988]: I1008 18:11:53.073881 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:11:53 crc kubenswrapper[4988]: I1008 18:11:53.073902 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:11:53 crc kubenswrapper[4988]: I1008 18:11:53.073915 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:11:53Z","lastTransitionTime":"2025-10-08T18:11:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:11:53 crc kubenswrapper[4988]: I1008 18:11:53.176723 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:11:53 crc kubenswrapper[4988]: I1008 18:11:53.176845 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:11:53 crc kubenswrapper[4988]: I1008 18:11:53.176875 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:11:53 crc kubenswrapper[4988]: I1008 18:11:53.176905 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:11:53 crc kubenswrapper[4988]: I1008 18:11:53.176924 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:11:53Z","lastTransitionTime":"2025-10-08T18:11:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:11:53 crc kubenswrapper[4988]: I1008 18:11:53.237858 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jn2sx" Oct 08 18:11:53 crc kubenswrapper[4988]: I1008 18:11:53.237959 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 18:11:53 crc kubenswrapper[4988]: I1008 18:11:53.237961 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 18:11:53 crc kubenswrapper[4988]: E1008 18:11:53.238067 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jn2sx" podUID="171e0902-95ee-4a39-8b1b-a58704c30574" Oct 08 18:11:53 crc kubenswrapper[4988]: I1008 18:11:53.238097 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 18:11:53 crc kubenswrapper[4988]: E1008 18:11:53.238333 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 08 18:11:53 crc kubenswrapper[4988]: E1008 18:11:53.238419 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 08 18:11:53 crc kubenswrapper[4988]: E1008 18:11:53.238569 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 08 18:11:53 crc kubenswrapper[4988]: I1008 18:11:53.280143 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:11:53 crc kubenswrapper[4988]: I1008 18:11:53.280222 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:11:53 crc kubenswrapper[4988]: I1008 18:11:53.280244 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:11:53 crc kubenswrapper[4988]: I1008 18:11:53.280277 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:11:53 crc kubenswrapper[4988]: I1008 18:11:53.280301 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:11:53Z","lastTransitionTime":"2025-10-08T18:11:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:11:53 crc kubenswrapper[4988]: I1008 18:11:53.384854 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:11:53 crc kubenswrapper[4988]: I1008 18:11:53.384963 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:11:53 crc kubenswrapper[4988]: I1008 18:11:53.384984 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:11:53 crc kubenswrapper[4988]: I1008 18:11:53.385014 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:11:53 crc kubenswrapper[4988]: I1008 18:11:53.385045 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:11:53Z","lastTransitionTime":"2025-10-08T18:11:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:11:53 crc kubenswrapper[4988]: I1008 18:11:53.488369 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:11:53 crc kubenswrapper[4988]: I1008 18:11:53.488446 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:11:53 crc kubenswrapper[4988]: I1008 18:11:53.488457 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:11:53 crc kubenswrapper[4988]: I1008 18:11:53.488477 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:11:53 crc kubenswrapper[4988]: I1008 18:11:53.488493 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:11:53Z","lastTransitionTime":"2025-10-08T18:11:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:11:53 crc kubenswrapper[4988]: I1008 18:11:53.592222 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:11:53 crc kubenswrapper[4988]: I1008 18:11:53.592296 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:11:53 crc kubenswrapper[4988]: I1008 18:11:53.592314 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:11:53 crc kubenswrapper[4988]: I1008 18:11:53.592343 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:11:53 crc kubenswrapper[4988]: I1008 18:11:53.592364 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:11:53Z","lastTransitionTime":"2025-10-08T18:11:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:11:53 crc kubenswrapper[4988]: I1008 18:11:53.695539 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:11:53 crc kubenswrapper[4988]: I1008 18:11:53.695600 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:11:53 crc kubenswrapper[4988]: I1008 18:11:53.695615 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:11:53 crc kubenswrapper[4988]: I1008 18:11:53.695636 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:11:53 crc kubenswrapper[4988]: I1008 18:11:53.695650 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:11:53Z","lastTransitionTime":"2025-10-08T18:11:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:11:53 crc kubenswrapper[4988]: I1008 18:11:53.799316 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:11:53 crc kubenswrapper[4988]: I1008 18:11:53.799424 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:11:53 crc kubenswrapper[4988]: I1008 18:11:53.799450 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:11:53 crc kubenswrapper[4988]: I1008 18:11:53.799482 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:11:53 crc kubenswrapper[4988]: I1008 18:11:53.799509 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:11:53Z","lastTransitionTime":"2025-10-08T18:11:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:11:53 crc kubenswrapper[4988]: I1008 18:11:53.905533 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:11:53 crc kubenswrapper[4988]: I1008 18:11:53.905593 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:11:53 crc kubenswrapper[4988]: I1008 18:11:53.905613 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:11:53 crc kubenswrapper[4988]: I1008 18:11:53.905641 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:11:53 crc kubenswrapper[4988]: I1008 18:11:53.905662 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:11:53Z","lastTransitionTime":"2025-10-08T18:11:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:11:54 crc kubenswrapper[4988]: I1008 18:11:54.009189 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:11:54 crc kubenswrapper[4988]: I1008 18:11:54.009262 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:11:54 crc kubenswrapper[4988]: I1008 18:11:54.009280 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:11:54 crc kubenswrapper[4988]: I1008 18:11:54.009314 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:11:54 crc kubenswrapper[4988]: I1008 18:11:54.009333 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:11:54Z","lastTransitionTime":"2025-10-08T18:11:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:11:54 crc kubenswrapper[4988]: I1008 18:11:54.112845 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:11:54 crc kubenswrapper[4988]: I1008 18:11:54.112906 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:11:54 crc kubenswrapper[4988]: I1008 18:11:54.112923 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:11:54 crc kubenswrapper[4988]: I1008 18:11:54.112951 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:11:54 crc kubenswrapper[4988]: I1008 18:11:54.112967 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:11:54Z","lastTransitionTime":"2025-10-08T18:11:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:11:54 crc kubenswrapper[4988]: I1008 18:11:54.215738 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:11:54 crc kubenswrapper[4988]: I1008 18:11:54.215786 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:11:54 crc kubenswrapper[4988]: I1008 18:11:54.215799 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:11:54 crc kubenswrapper[4988]: I1008 18:11:54.215817 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:11:54 crc kubenswrapper[4988]: I1008 18:11:54.215829 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:11:54Z","lastTransitionTime":"2025-10-08T18:11:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:11:54 crc kubenswrapper[4988]: I1008 18:11:54.319949 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:11:54 crc kubenswrapper[4988]: I1008 18:11:54.320011 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:11:54 crc kubenswrapper[4988]: I1008 18:11:54.320028 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:11:54 crc kubenswrapper[4988]: I1008 18:11:54.320054 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:11:54 crc kubenswrapper[4988]: I1008 18:11:54.320071 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:11:54Z","lastTransitionTime":"2025-10-08T18:11:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:11:54 crc kubenswrapper[4988]: I1008 18:11:54.423368 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:11:54 crc kubenswrapper[4988]: I1008 18:11:54.423471 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:11:54 crc kubenswrapper[4988]: I1008 18:11:54.423489 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:11:54 crc kubenswrapper[4988]: I1008 18:11:54.423521 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:11:54 crc kubenswrapper[4988]: I1008 18:11:54.423540 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:11:54Z","lastTransitionTime":"2025-10-08T18:11:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:11:54 crc kubenswrapper[4988]: I1008 18:11:54.527127 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:11:54 crc kubenswrapper[4988]: I1008 18:11:54.527381 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:11:54 crc kubenswrapper[4988]: I1008 18:11:54.527438 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:11:54 crc kubenswrapper[4988]: I1008 18:11:54.527464 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:11:54 crc kubenswrapper[4988]: I1008 18:11:54.527488 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:11:54Z","lastTransitionTime":"2025-10-08T18:11:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:11:54 crc kubenswrapper[4988]: I1008 18:11:54.631427 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:11:54 crc kubenswrapper[4988]: I1008 18:11:54.631498 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:11:54 crc kubenswrapper[4988]: I1008 18:11:54.631517 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:11:54 crc kubenswrapper[4988]: I1008 18:11:54.631544 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:11:54 crc kubenswrapper[4988]: I1008 18:11:54.631564 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:11:54Z","lastTransitionTime":"2025-10-08T18:11:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:11:54 crc kubenswrapper[4988]: I1008 18:11:54.735781 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:11:54 crc kubenswrapper[4988]: I1008 18:11:54.737007 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:11:54 crc kubenswrapper[4988]: I1008 18:11:54.737244 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:11:54 crc kubenswrapper[4988]: I1008 18:11:54.737491 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:11:54 crc kubenswrapper[4988]: I1008 18:11:54.737689 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:11:54Z","lastTransitionTime":"2025-10-08T18:11:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:11:54 crc kubenswrapper[4988]: I1008 18:11:54.844248 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:11:54 crc kubenswrapper[4988]: I1008 18:11:54.844303 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:11:54 crc kubenswrapper[4988]: I1008 18:11:54.844316 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:11:54 crc kubenswrapper[4988]: I1008 18:11:54.844337 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:11:54 crc kubenswrapper[4988]: I1008 18:11:54.844352 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:11:54Z","lastTransitionTime":"2025-10-08T18:11:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:11:54 crc kubenswrapper[4988]: I1008 18:11:54.948158 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:11:54 crc kubenswrapper[4988]: I1008 18:11:54.948209 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:11:54 crc kubenswrapper[4988]: I1008 18:11:54.948222 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:11:54 crc kubenswrapper[4988]: I1008 18:11:54.948242 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:11:54 crc kubenswrapper[4988]: I1008 18:11:54.948256 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:11:54Z","lastTransitionTime":"2025-10-08T18:11:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:11:55 crc kubenswrapper[4988]: I1008 18:11:55.051931 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:11:55 crc kubenswrapper[4988]: I1008 18:11:55.051986 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:11:55 crc kubenswrapper[4988]: I1008 18:11:55.052004 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:11:55 crc kubenswrapper[4988]: I1008 18:11:55.052033 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:11:55 crc kubenswrapper[4988]: I1008 18:11:55.052053 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:11:55Z","lastTransitionTime":"2025-10-08T18:11:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:11:55 crc kubenswrapper[4988]: I1008 18:11:55.156205 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:11:55 crc kubenswrapper[4988]: I1008 18:11:55.156270 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:11:55 crc kubenswrapper[4988]: I1008 18:11:55.156284 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:11:55 crc kubenswrapper[4988]: I1008 18:11:55.156309 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:11:55 crc kubenswrapper[4988]: I1008 18:11:55.156326 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:11:55Z","lastTransitionTime":"2025-10-08T18:11:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:11:55 crc kubenswrapper[4988]: I1008 18:11:55.237521 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 18:11:55 crc kubenswrapper[4988]: I1008 18:11:55.237581 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 18:11:55 crc kubenswrapper[4988]: I1008 18:11:55.237633 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 18:11:55 crc kubenswrapper[4988]: E1008 18:11:55.237744 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 08 18:11:55 crc kubenswrapper[4988]: E1008 18:11:55.237980 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 08 18:11:55 crc kubenswrapper[4988]: E1008 18:11:55.238194 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 08 18:11:55 crc kubenswrapper[4988]: I1008 18:11:55.239304 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jn2sx" Oct 08 18:11:55 crc kubenswrapper[4988]: E1008 18:11:55.239781 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jn2sx" podUID="171e0902-95ee-4a39-8b1b-a58704c30574" Oct 08 18:11:55 crc kubenswrapper[4988]: I1008 18:11:55.254113 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-s4f2k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ba5e11d7-7bde-4c3d-983b-884232219652\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://824bc34a2807068d3fceb10651803ae9ae5651e940f7aed4877c544d76dae4d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hn5jf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:11:15Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-s4f2k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:55Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:55 crc kubenswrapper[4988]: I1008 18:11:55.261914 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:11:55 crc kubenswrapper[4988]: I1008 18:11:55.261948 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:11:55 crc kubenswrapper[4988]: I1008 18:11:55.261957 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:11:55 crc kubenswrapper[4988]: I1008 18:11:55.261973 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:11:55 crc kubenswrapper[4988]: I1008 18:11:55.261983 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:11:55Z","lastTransitionTime":"2025-10-08T18:11:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:11:55 crc kubenswrapper[4988]: I1008 18:11:55.271668 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aef71be1-1d12-46e5-95b2-bd399040fd4d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:10:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:10:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:10:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77c68f3ecb8a0724499e495cb26ef3db5beecb140f2558d8c0b25f41928b84ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:10:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a560ee84c2405227de71fda127566213cd4f017e73d5581442ba267781b1c539\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:10:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6100accc3bd08928bc1ff88310f29fa5da37fd2d5c315c94279801c9778c4745\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:10:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b0e596d7ce4e9341c44499af79d3b3caf0e5e2c156902ea40cf2a3dca672be9e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5bd35aef24be034a0b305a40bb53ca171a4f76db3b93b4f78ae95099b0a4cde3\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-08T18:11:14Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1008 18:11:09.091034 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1008 18:11:09.091984 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3472374062/tls.crt::/tmp/serving-cert-3472374062/tls.key\\\\\\\"\\\\nI1008 18:11:14.619696 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1008 18:11:14.628837 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1008 18:11:14.628898 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1008 18:11:14.628994 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1008 18:11:14.629036 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1008 18:11:14.641318 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1008 18:11:14.641363 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1008 18:11:14.641372 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1008 18:11:14.641382 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1008 18:11:14.641415 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1008 18:11:14.641423 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1008 18:11:14.641431 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1008 18:11:14.641484 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1008 18:11:14.646451 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T18:10:58Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c79ef704fda1bbcc486f4c7afa3039a3ab38669833f0f66bafe24dd4cc679e60\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:10:58Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://28d21731af0f16d6a8ae6c6e9d0bebf0560d4a533db4c78c10bf1837890394ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://28d21731af0f16d6a8ae6c6e9d0bebf0560d4a533db4c78c10bf1837890394ef\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:10:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:10:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:10:55Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:55Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:55 crc kubenswrapper[4988]: I1008 18:11:55.285920 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7e0ecefc-359f-4785-bb79-0343a62a25da\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:10:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:10:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:10:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://30bec350fefee3d19758b3ed735ede4e9aa085edb57aa3f3ec24249a67dc7ca2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:10:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2d850fd27d4ed65891ff69b9add65fb78535357f54819c2a8d196f5a340c58e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:10:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://90b1a9e14c82bf0f75f08fdb342eed35a503020f9db641e5e9ceff8b122cb3bd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:10:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://16f23c4c7163d32dc0df8d6a7ee17e2864fb261f8a40ecf05556774b173fc868\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:10:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:10:55Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:55Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:55 crc kubenswrapper[4988]: I1008 18:11:55.303057 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e8906182d55c285b3cf095e2d9e1dac397af4618a514f11e26b44c172dd573f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:55Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:55 crc kubenswrapper[4988]: I1008 18:11:55.313059 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-dzxs2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8f307b20-5e33-4725-99cb-210e25113532\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72ee803c6ef88524e0ce0391476f8e5a7dfbc858b095a3c2ae31bad10991ce3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q65f7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:11:16Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-dzxs2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:55Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:55 crc kubenswrapper[4988]: I1008 18:11:55.327497 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-jnh7m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"556fb048-8afb-44a5-9ea0-2b7c89e6cf4b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bfa2f3382b23c3f788a1b31e80e8a90b3b6868870dccc46d86c3fc11ad027f52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmlt2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1a825886f105ccf9a7cc6f28a0cd012fb859f4990dade0e6aae932482146028\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1a825886f105ccf9a7cc6f28a0cd012fb859f4990dade0e6aae932482146028\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmlt2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ab2fe6e65418efa75bae72259157a2323bbb4b7d11d1d012053524e14135888\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0ab2fe6e65418efa75bae72259157a2323bbb4b7d11d1d012053524e14135888\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:11:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:11:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmlt2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://368a303309be112d9c0b7ccd6538fbaca7085d97a8ec8259f36092f07147e7cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://368a303309be112d9c0b7ccd6538fbaca7085d97a8ec8259f36092f07147e7cb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:11:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:11:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmlt2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a8ca69001dddef643f982a97c74c291e30cb9fac5b0e894e80066b885ea9151\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9a8ca69001dddef643f982a97c74c291e30cb9fac5b0e894e80066b885ea9151\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:11:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:11:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmlt2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://145303c46b48603ed1aae4b9bfe06332a3ffad28f206f8394574011a87b58cd3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://145303c46b48603ed1aae4b9bfe06332a3ffad28f206f8394574011a87b58cd3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:11:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:11:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmlt2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d08e5cfc7cbe5ec1a7a0265ed7c466363c3abeec492263170860689b46c7ff95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d08e5cfc7cbe5ec1a7a0265ed7c466363c3abeec492263170860689b46c7ff95\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:11:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:11:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmlt2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:11:16Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-jnh7m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:55Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:55 crc kubenswrapper[4988]: I1008 18:11:55.343198 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f328a20b-5283-407e-8a93-2ab2da599f18\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8622d56fb198dadee31fb0ca8fc80d56828b6710648055f264533e2d5bdb3037\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z2gh4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3cdb2a757332bdcdc2a2f3362913fe8f06003b0597de20b72627f9badd5e30af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z2gh4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:11:16Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-nm54f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:55Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:55 crc kubenswrapper[4988]: I1008 18:11:55.360716 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a35a5e989ba69c8ef36ce4cea272aadf67ce20473fc257adfcfa86a2faba6ae0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a522e937cfe03fb9f2e60cb1bea07214cf63c69bdecd5ba65a2dfdba882d28dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:55Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:55 crc kubenswrapper[4988]: I1008 18:11:55.364677 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:11:55 crc kubenswrapper[4988]: I1008 18:11:55.364736 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:11:55 crc kubenswrapper[4988]: I1008 18:11:55.364746 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:11:55 crc kubenswrapper[4988]: I1008 18:11:55.364765 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:11:55 crc kubenswrapper[4988]: I1008 18:11:55.364777 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:11:55Z","lastTransitionTime":"2025-10-08T18:11:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:11:55 crc kubenswrapper[4988]: I1008 18:11:55.374158 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9590ca0059f0ebbe86343cccb3475338c317382c3231e082655b61e67eece375\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:55Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:55 crc kubenswrapper[4988]: I1008 18:11:55.390191 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:55Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:55 crc kubenswrapper[4988]: I1008 18:11:55.421725 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-nsc65" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f010a463-69f2-4082-acc3-161899a939c3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://48a516f6c153743bd1a21ec6945fb368a37f4d2b9d6a326b68edc096a192ed05\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dnwgr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://03406e726f817bcedb841f81959ab12808ba35a530240f9abc5fa5c608b9a6ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dnwgr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9fbbeae0409d2e47e8de9a4306f63b0d13acbfd73cd86f83303fec14352ed718\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dnwgr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://55791327a370c0009c523b86b2234e82d3794985f2eaccf1922eb90f46b2b970\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dnwgr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cee5de93bc68345f9b737857f02832f56582d58757c50a8798eec25d19b2c20d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dnwgr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d314eb248e90371fed068b1173e26d4c9f9765417a70b341e1a5b136ec17353a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dnwgr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7bf11dd48efebb9da2b836bfad6a54b274247fa60902dab2f084793e4ad51a89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7bf11dd48efebb9da2b836bfad6a54b274247fa60902dab2f084793e4ad51a89\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-08T18:11:42Z\\\",\\\"message\\\":\\\" V4IPs:[]string(nil), V6IPs:[]string(nil)}, nodeEndpoints:map[string]services.lbEndpoints{}, externalTrafficLocal:false, internalTrafficLocal:false, hasNodePort:false}}\\\\nI1008 18:11:42.211985 6628 services_controller.go:444] Built service openshift-authentication-operator/metrics LB per-node configs for network=default: []services.lbConfig(nil)\\\\nI1008 18:11:42.211998 6628 services_controller.go:445] Built service openshift-authentication-operator/metrics LB template configs for network=default: []services.lbConfig(nil)\\\\nF1008 18:11:42.212009 6628 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:42Z is after 2025-08-\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T18:11:41Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-nsc65_openshift-ovn-kubernetes(f010a463-69f2-4082-acc3-161899a939c3)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dnwgr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://74fcb812b84291763cc6e5ccb20c12aaad580ad968dda4734a0bbfe33d2770ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dnwgr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://befe684cea9bd6dfda3ec8e85bca431c26bd8f82823b0a69dc0d04319a3ca385\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://befe684cea9bd6dfda3ec8e85bca431c26bd8f82823b0a69dc0d04319a3ca385\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dnwgr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:11:16Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-nsc65\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:55Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:55 crc kubenswrapper[4988]: I1008 18:11:55.436814 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-7wl8v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e433488d-db52-435f-b89c-f63cd477c5d3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76b30d183684740b647c1af11dc49712232cc8b0c9a392b491580e5ddb75ea7f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dbz6l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e55016c3e1b97d1b3a857130a5c8e2ec6da2333f786032118a61e3e2ad4c5651\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dbz6l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:11:29Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-7wl8v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:55Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:55 crc kubenswrapper[4988]: I1008 18:11:55.452984 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-jn2sx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"171e0902-95ee-4a39-8b1b-a58704c30574\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9f4q2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9f4q2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:11:31Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-jn2sx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:55Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:55 crc kubenswrapper[4988]: I1008 18:11:55.468079 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:11:55 crc kubenswrapper[4988]: I1008 18:11:55.468153 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:11:55 crc kubenswrapper[4988]: I1008 18:11:55.468179 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:11:55 crc kubenswrapper[4988]: I1008 18:11:55.468211 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:11:55 crc kubenswrapper[4988]: I1008 18:11:55.468234 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:11:55Z","lastTransitionTime":"2025-10-08T18:11:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:11:55 crc kubenswrapper[4988]: I1008 18:11:55.474422 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d36f620-0f26-41c7-af6e-7fb5721a0e2d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:10:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:10:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:10:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3fd3b2ca80233400d2798b035a65b833f735632efd76d1160f6bb906e07b80a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:10:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://75a2bd2757133e79c0c7f2939eb433dc45291f14808e3c3f1daaf90a239c6a71\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:10:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5cb03de583c14ad6751227c21166ac5ac4c6e11b67d5b709361ced8a0877e6d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:10:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6eb46d76cf7cfb1d4ecc4fcee398dad8904732c1664c317865864b65f5703589\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6eb46d76cf7cfb1d4ecc4fcee398dad8904732c1664c317865864b65f5703589\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:10:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:10:56Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:10:55Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:55Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:55 crc kubenswrapper[4988]: I1008 18:11:55.501095 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:55Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:55 crc kubenswrapper[4988]: I1008 18:11:55.519992 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:55Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:55 crc kubenswrapper[4988]: I1008 18:11:55.541594 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-h9jwp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4490d846-ca32-4cd5-9202-87cc6d499fb4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://00b7f6db043c28efda3a09fdc77a6ccded3eae8b6a3d2852aa500789d5f01194\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7k6rk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:11:16Z\\\"}}\" for pod \"openshift-multus\"/\"multus-h9jwp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:55Z is after 2025-08-24T17:21:41Z" Oct 08 18:11:55 crc kubenswrapper[4988]: I1008 18:11:55.573115 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:11:55 crc kubenswrapper[4988]: I1008 18:11:55.573191 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:11:55 crc kubenswrapper[4988]: I1008 18:11:55.573260 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:11:55 crc kubenswrapper[4988]: I1008 18:11:55.573291 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:11:55 crc kubenswrapper[4988]: I1008 18:11:55.573308 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:11:55Z","lastTransitionTime":"2025-10-08T18:11:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:11:55 crc kubenswrapper[4988]: I1008 18:11:55.677190 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:11:55 crc kubenswrapper[4988]: I1008 18:11:55.677251 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:11:55 crc kubenswrapper[4988]: I1008 18:11:55.677269 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:11:55 crc kubenswrapper[4988]: I1008 18:11:55.677302 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:11:55 crc kubenswrapper[4988]: I1008 18:11:55.677324 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:11:55Z","lastTransitionTime":"2025-10-08T18:11:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:11:55 crc kubenswrapper[4988]: I1008 18:11:55.781470 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:11:55 crc kubenswrapper[4988]: I1008 18:11:55.781545 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:11:55 crc kubenswrapper[4988]: I1008 18:11:55.781570 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:11:55 crc kubenswrapper[4988]: I1008 18:11:55.781602 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:11:55 crc kubenswrapper[4988]: I1008 18:11:55.781625 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:11:55Z","lastTransitionTime":"2025-10-08T18:11:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:11:55 crc kubenswrapper[4988]: I1008 18:11:55.885690 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:11:55 crc kubenswrapper[4988]: I1008 18:11:55.886048 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:11:55 crc kubenswrapper[4988]: I1008 18:11:55.886069 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:11:55 crc kubenswrapper[4988]: I1008 18:11:55.886099 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:11:55 crc kubenswrapper[4988]: I1008 18:11:55.886117 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:11:55Z","lastTransitionTime":"2025-10-08T18:11:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:11:55 crc kubenswrapper[4988]: I1008 18:11:55.990230 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:11:55 crc kubenswrapper[4988]: I1008 18:11:55.990300 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:11:55 crc kubenswrapper[4988]: I1008 18:11:55.990319 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:11:55 crc kubenswrapper[4988]: I1008 18:11:55.990343 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:11:55 crc kubenswrapper[4988]: I1008 18:11:55.990363 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:11:55Z","lastTransitionTime":"2025-10-08T18:11:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:11:56 crc kubenswrapper[4988]: I1008 18:11:56.094890 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:11:56 crc kubenswrapper[4988]: I1008 18:11:56.094965 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:11:56 crc kubenswrapper[4988]: I1008 18:11:56.094982 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:11:56 crc kubenswrapper[4988]: I1008 18:11:56.095008 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:11:56 crc kubenswrapper[4988]: I1008 18:11:56.095026 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:11:56Z","lastTransitionTime":"2025-10-08T18:11:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:11:56 crc kubenswrapper[4988]: I1008 18:11:56.197819 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:11:56 crc kubenswrapper[4988]: I1008 18:11:56.197891 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:11:56 crc kubenswrapper[4988]: I1008 18:11:56.197910 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:11:56 crc kubenswrapper[4988]: I1008 18:11:56.197939 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:11:56 crc kubenswrapper[4988]: I1008 18:11:56.197960 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:11:56Z","lastTransitionTime":"2025-10-08T18:11:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:11:56 crc kubenswrapper[4988]: I1008 18:11:56.238174 4988 scope.go:117] "RemoveContainer" containerID="7bf11dd48efebb9da2b836bfad6a54b274247fa60902dab2f084793e4ad51a89" Oct 08 18:11:56 crc kubenswrapper[4988]: E1008 18:11:56.238706 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-nsc65_openshift-ovn-kubernetes(f010a463-69f2-4082-acc3-161899a939c3)\"" pod="openshift-ovn-kubernetes/ovnkube-node-nsc65" podUID="f010a463-69f2-4082-acc3-161899a939c3" Oct 08 18:11:56 crc kubenswrapper[4988]: I1008 18:11:56.301192 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:11:56 crc kubenswrapper[4988]: I1008 18:11:56.301245 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:11:56 crc kubenswrapper[4988]: I1008 18:11:56.301258 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:11:56 crc kubenswrapper[4988]: I1008 18:11:56.301280 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:11:56 crc kubenswrapper[4988]: I1008 18:11:56.301296 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:11:56Z","lastTransitionTime":"2025-10-08T18:11:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:11:56 crc kubenswrapper[4988]: I1008 18:11:56.405168 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:11:56 crc kubenswrapper[4988]: I1008 18:11:56.405265 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:11:56 crc kubenswrapper[4988]: I1008 18:11:56.405292 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:11:56 crc kubenswrapper[4988]: I1008 18:11:56.405328 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:11:56 crc kubenswrapper[4988]: I1008 18:11:56.405352 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:11:56Z","lastTransitionTime":"2025-10-08T18:11:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:11:56 crc kubenswrapper[4988]: I1008 18:11:56.509571 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:11:56 crc kubenswrapper[4988]: I1008 18:11:56.509670 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:11:56 crc kubenswrapper[4988]: I1008 18:11:56.509697 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:11:56 crc kubenswrapper[4988]: I1008 18:11:56.509738 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:11:56 crc kubenswrapper[4988]: I1008 18:11:56.509765 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:11:56Z","lastTransitionTime":"2025-10-08T18:11:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:11:56 crc kubenswrapper[4988]: I1008 18:11:56.613067 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:11:56 crc kubenswrapper[4988]: I1008 18:11:56.613127 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:11:56 crc kubenswrapper[4988]: I1008 18:11:56.613146 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:11:56 crc kubenswrapper[4988]: I1008 18:11:56.613172 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:11:56 crc kubenswrapper[4988]: I1008 18:11:56.613191 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:11:56Z","lastTransitionTime":"2025-10-08T18:11:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:11:56 crc kubenswrapper[4988]: I1008 18:11:56.717375 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:11:56 crc kubenswrapper[4988]: I1008 18:11:56.717461 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:11:56 crc kubenswrapper[4988]: I1008 18:11:56.717477 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:11:56 crc kubenswrapper[4988]: I1008 18:11:56.717502 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:11:56 crc kubenswrapper[4988]: I1008 18:11:56.717517 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:11:56Z","lastTransitionTime":"2025-10-08T18:11:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:11:56 crc kubenswrapper[4988]: I1008 18:11:56.821031 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:11:56 crc kubenswrapper[4988]: I1008 18:11:56.821187 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:11:56 crc kubenswrapper[4988]: I1008 18:11:56.821211 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:11:56 crc kubenswrapper[4988]: I1008 18:11:56.821282 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:11:56 crc kubenswrapper[4988]: I1008 18:11:56.821305 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:11:56Z","lastTransitionTime":"2025-10-08T18:11:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:11:56 crc kubenswrapper[4988]: I1008 18:11:56.924332 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:11:56 crc kubenswrapper[4988]: I1008 18:11:56.924412 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:11:56 crc kubenswrapper[4988]: I1008 18:11:56.924425 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:11:56 crc kubenswrapper[4988]: I1008 18:11:56.924445 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:11:56 crc kubenswrapper[4988]: I1008 18:11:56.924459 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:11:56Z","lastTransitionTime":"2025-10-08T18:11:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:11:57 crc kubenswrapper[4988]: I1008 18:11:57.027709 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:11:57 crc kubenswrapper[4988]: I1008 18:11:57.027785 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:11:57 crc kubenswrapper[4988]: I1008 18:11:57.027806 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:11:57 crc kubenswrapper[4988]: I1008 18:11:57.027834 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:11:57 crc kubenswrapper[4988]: I1008 18:11:57.027851 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:11:57Z","lastTransitionTime":"2025-10-08T18:11:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:11:57 crc kubenswrapper[4988]: I1008 18:11:57.131980 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:11:57 crc kubenswrapper[4988]: I1008 18:11:57.132032 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:11:57 crc kubenswrapper[4988]: I1008 18:11:57.132045 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:11:57 crc kubenswrapper[4988]: I1008 18:11:57.132068 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:11:57 crc kubenswrapper[4988]: I1008 18:11:57.132082 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:11:57Z","lastTransitionTime":"2025-10-08T18:11:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:11:57 crc kubenswrapper[4988]: I1008 18:11:57.237099 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 18:11:57 crc kubenswrapper[4988]: I1008 18:11:57.237154 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 18:11:57 crc kubenswrapper[4988]: I1008 18:11:57.237204 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 18:11:57 crc kubenswrapper[4988]: E1008 18:11:57.237227 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 08 18:11:57 crc kubenswrapper[4988]: E1008 18:11:57.237363 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 08 18:11:57 crc kubenswrapper[4988]: I1008 18:11:57.237427 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jn2sx" Oct 08 18:11:57 crc kubenswrapper[4988]: E1008 18:11:57.237450 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 08 18:11:57 crc kubenswrapper[4988]: E1008 18:11:57.237619 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jn2sx" podUID="171e0902-95ee-4a39-8b1b-a58704c30574" Oct 08 18:11:57 crc kubenswrapper[4988]: I1008 18:11:57.238465 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:11:57 crc kubenswrapper[4988]: I1008 18:11:57.238502 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:11:57 crc kubenswrapper[4988]: I1008 18:11:57.238518 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:11:57 crc kubenswrapper[4988]: I1008 18:11:57.238544 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:11:57 crc kubenswrapper[4988]: I1008 18:11:57.238560 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:11:57Z","lastTransitionTime":"2025-10-08T18:11:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:11:57 crc kubenswrapper[4988]: I1008 18:11:57.342344 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:11:57 crc kubenswrapper[4988]: I1008 18:11:57.342451 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:11:57 crc kubenswrapper[4988]: I1008 18:11:57.342470 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:11:57 crc kubenswrapper[4988]: I1008 18:11:57.342520 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:11:57 crc kubenswrapper[4988]: I1008 18:11:57.342546 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:11:57Z","lastTransitionTime":"2025-10-08T18:11:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:11:57 crc kubenswrapper[4988]: I1008 18:11:57.445597 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:11:57 crc kubenswrapper[4988]: I1008 18:11:57.445673 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:11:57 crc kubenswrapper[4988]: I1008 18:11:57.445692 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:11:57 crc kubenswrapper[4988]: I1008 18:11:57.445720 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:11:57 crc kubenswrapper[4988]: I1008 18:11:57.445738 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:11:57Z","lastTransitionTime":"2025-10-08T18:11:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:11:57 crc kubenswrapper[4988]: I1008 18:11:57.549761 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:11:57 crc kubenswrapper[4988]: I1008 18:11:57.549820 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:11:57 crc kubenswrapper[4988]: I1008 18:11:57.549844 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:11:57 crc kubenswrapper[4988]: I1008 18:11:57.549873 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:11:57 crc kubenswrapper[4988]: I1008 18:11:57.549892 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:11:57Z","lastTransitionTime":"2025-10-08T18:11:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:11:57 crc kubenswrapper[4988]: I1008 18:11:57.654219 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:11:57 crc kubenswrapper[4988]: I1008 18:11:57.654280 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:11:57 crc kubenswrapper[4988]: I1008 18:11:57.654297 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:11:57 crc kubenswrapper[4988]: I1008 18:11:57.654323 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:11:57 crc kubenswrapper[4988]: I1008 18:11:57.654342 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:11:57Z","lastTransitionTime":"2025-10-08T18:11:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:11:57 crc kubenswrapper[4988]: I1008 18:11:57.758887 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:11:57 crc kubenswrapper[4988]: I1008 18:11:57.758951 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:11:57 crc kubenswrapper[4988]: I1008 18:11:57.758969 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:11:57 crc kubenswrapper[4988]: I1008 18:11:57.758993 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:11:57 crc kubenswrapper[4988]: I1008 18:11:57.759013 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:11:57Z","lastTransitionTime":"2025-10-08T18:11:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:11:57 crc kubenswrapper[4988]: I1008 18:11:57.863583 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:11:57 crc kubenswrapper[4988]: I1008 18:11:57.863699 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:11:57 crc kubenswrapper[4988]: I1008 18:11:57.863726 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:11:57 crc kubenswrapper[4988]: I1008 18:11:57.863761 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:11:57 crc kubenswrapper[4988]: I1008 18:11:57.863927 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:11:57Z","lastTransitionTime":"2025-10-08T18:11:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:11:57 crc kubenswrapper[4988]: I1008 18:11:57.969063 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:11:57 crc kubenswrapper[4988]: I1008 18:11:57.969107 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:11:57 crc kubenswrapper[4988]: I1008 18:11:57.969118 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:11:57 crc kubenswrapper[4988]: I1008 18:11:57.969137 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:11:57 crc kubenswrapper[4988]: I1008 18:11:57.969149 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:11:57Z","lastTransitionTime":"2025-10-08T18:11:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:11:58 crc kubenswrapper[4988]: I1008 18:11:58.072793 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:11:58 crc kubenswrapper[4988]: I1008 18:11:58.072834 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:11:58 crc kubenswrapper[4988]: I1008 18:11:58.072847 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:11:58 crc kubenswrapper[4988]: I1008 18:11:58.072864 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:11:58 crc kubenswrapper[4988]: I1008 18:11:58.072877 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:11:58Z","lastTransitionTime":"2025-10-08T18:11:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:11:58 crc kubenswrapper[4988]: I1008 18:11:58.175465 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:11:58 crc kubenswrapper[4988]: I1008 18:11:58.175500 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:11:58 crc kubenswrapper[4988]: I1008 18:11:58.175511 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:11:58 crc kubenswrapper[4988]: I1008 18:11:58.175527 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:11:58 crc kubenswrapper[4988]: I1008 18:11:58.175541 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:11:58Z","lastTransitionTime":"2025-10-08T18:11:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:11:58 crc kubenswrapper[4988]: I1008 18:11:58.278917 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:11:58 crc kubenswrapper[4988]: I1008 18:11:58.278950 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:11:58 crc kubenswrapper[4988]: I1008 18:11:58.278963 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:11:58 crc kubenswrapper[4988]: I1008 18:11:58.278984 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:11:58 crc kubenswrapper[4988]: I1008 18:11:58.278997 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:11:58Z","lastTransitionTime":"2025-10-08T18:11:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:11:58 crc kubenswrapper[4988]: I1008 18:11:58.382070 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:11:58 crc kubenswrapper[4988]: I1008 18:11:58.382108 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:11:58 crc kubenswrapper[4988]: I1008 18:11:58.382119 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:11:58 crc kubenswrapper[4988]: I1008 18:11:58.382137 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:11:58 crc kubenswrapper[4988]: I1008 18:11:58.382147 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:11:58Z","lastTransitionTime":"2025-10-08T18:11:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:11:58 crc kubenswrapper[4988]: I1008 18:11:58.486314 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:11:58 crc kubenswrapper[4988]: I1008 18:11:58.486358 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:11:58 crc kubenswrapper[4988]: I1008 18:11:58.486368 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:11:58 crc kubenswrapper[4988]: I1008 18:11:58.486400 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:11:58 crc kubenswrapper[4988]: I1008 18:11:58.486412 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:11:58Z","lastTransitionTime":"2025-10-08T18:11:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:11:58 crc kubenswrapper[4988]: I1008 18:11:58.588535 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:11:58 crc kubenswrapper[4988]: I1008 18:11:58.588566 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:11:58 crc kubenswrapper[4988]: I1008 18:11:58.588576 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:11:58 crc kubenswrapper[4988]: I1008 18:11:58.588589 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:11:58 crc kubenswrapper[4988]: I1008 18:11:58.588598 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:11:58Z","lastTransitionTime":"2025-10-08T18:11:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:11:58 crc kubenswrapper[4988]: I1008 18:11:58.691841 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:11:58 crc kubenswrapper[4988]: I1008 18:11:58.691951 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:11:58 crc kubenswrapper[4988]: I1008 18:11:58.691981 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:11:58 crc kubenswrapper[4988]: I1008 18:11:58.692007 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:11:58 crc kubenswrapper[4988]: I1008 18:11:58.692025 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:11:58Z","lastTransitionTime":"2025-10-08T18:11:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:11:58 crc kubenswrapper[4988]: I1008 18:11:58.794448 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:11:58 crc kubenswrapper[4988]: I1008 18:11:58.794489 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:11:58 crc kubenswrapper[4988]: I1008 18:11:58.794499 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:11:58 crc kubenswrapper[4988]: I1008 18:11:58.794516 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:11:58 crc kubenswrapper[4988]: I1008 18:11:58.794526 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:11:58Z","lastTransitionTime":"2025-10-08T18:11:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:11:58 crc kubenswrapper[4988]: I1008 18:11:58.898357 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:11:58 crc kubenswrapper[4988]: I1008 18:11:58.898460 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:11:58 crc kubenswrapper[4988]: I1008 18:11:58.898479 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:11:58 crc kubenswrapper[4988]: I1008 18:11:58.898504 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:11:58 crc kubenswrapper[4988]: I1008 18:11:58.898521 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:11:58Z","lastTransitionTime":"2025-10-08T18:11:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:11:59 crc kubenswrapper[4988]: I1008 18:11:59.003168 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:11:59 crc kubenswrapper[4988]: I1008 18:11:59.003309 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:11:59 crc kubenswrapper[4988]: I1008 18:11:59.003338 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:11:59 crc kubenswrapper[4988]: I1008 18:11:59.003369 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:11:59 crc kubenswrapper[4988]: I1008 18:11:59.003430 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:11:59Z","lastTransitionTime":"2025-10-08T18:11:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:11:59 crc kubenswrapper[4988]: I1008 18:11:59.107000 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:11:59 crc kubenswrapper[4988]: I1008 18:11:59.107048 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:11:59 crc kubenswrapper[4988]: I1008 18:11:59.107193 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:11:59 crc kubenswrapper[4988]: I1008 18:11:59.107213 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:11:59 crc kubenswrapper[4988]: I1008 18:11:59.107226 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:11:59Z","lastTransitionTime":"2025-10-08T18:11:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:11:59 crc kubenswrapper[4988]: I1008 18:11:59.211220 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:11:59 crc kubenswrapper[4988]: I1008 18:11:59.211744 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:11:59 crc kubenswrapper[4988]: I1008 18:11:59.211847 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:11:59 crc kubenswrapper[4988]: I1008 18:11:59.211968 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:11:59 crc kubenswrapper[4988]: I1008 18:11:59.212106 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:11:59Z","lastTransitionTime":"2025-10-08T18:11:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:11:59 crc kubenswrapper[4988]: I1008 18:11:59.236888 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 18:11:59 crc kubenswrapper[4988]: I1008 18:11:59.236978 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 18:11:59 crc kubenswrapper[4988]: E1008 18:11:59.237265 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 08 18:11:59 crc kubenswrapper[4988]: I1008 18:11:59.237021 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jn2sx" Oct 08 18:11:59 crc kubenswrapper[4988]: I1008 18:11:59.236986 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 18:11:59 crc kubenswrapper[4988]: E1008 18:11:59.237356 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jn2sx" podUID="171e0902-95ee-4a39-8b1b-a58704c30574" Oct 08 18:11:59 crc kubenswrapper[4988]: E1008 18:11:59.237516 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 08 18:11:59 crc kubenswrapper[4988]: E1008 18:11:59.237191 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 08 18:11:59 crc kubenswrapper[4988]: I1008 18:11:59.315815 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:11:59 crc kubenswrapper[4988]: I1008 18:11:59.315867 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:11:59 crc kubenswrapper[4988]: I1008 18:11:59.315879 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:11:59 crc kubenswrapper[4988]: I1008 18:11:59.315898 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:11:59 crc kubenswrapper[4988]: I1008 18:11:59.315909 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:11:59Z","lastTransitionTime":"2025-10-08T18:11:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:11:59 crc kubenswrapper[4988]: I1008 18:11:59.418494 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:11:59 crc kubenswrapper[4988]: I1008 18:11:59.418548 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:11:59 crc kubenswrapper[4988]: I1008 18:11:59.418561 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:11:59 crc kubenswrapper[4988]: I1008 18:11:59.418583 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:11:59 crc kubenswrapper[4988]: I1008 18:11:59.418600 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:11:59Z","lastTransitionTime":"2025-10-08T18:11:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:11:59 crc kubenswrapper[4988]: I1008 18:11:59.523096 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:11:59 crc kubenswrapper[4988]: I1008 18:11:59.523147 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:11:59 crc kubenswrapper[4988]: I1008 18:11:59.523156 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:11:59 crc kubenswrapper[4988]: I1008 18:11:59.523174 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:11:59 crc kubenswrapper[4988]: I1008 18:11:59.523185 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:11:59Z","lastTransitionTime":"2025-10-08T18:11:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:11:59 crc kubenswrapper[4988]: I1008 18:11:59.626863 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:11:59 crc kubenswrapper[4988]: I1008 18:11:59.626922 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:11:59 crc kubenswrapper[4988]: I1008 18:11:59.626935 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:11:59 crc kubenswrapper[4988]: I1008 18:11:59.626962 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:11:59 crc kubenswrapper[4988]: I1008 18:11:59.626973 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:11:59Z","lastTransitionTime":"2025-10-08T18:11:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:11:59 crc kubenswrapper[4988]: I1008 18:11:59.730322 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:11:59 crc kubenswrapper[4988]: I1008 18:11:59.730730 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:11:59 crc kubenswrapper[4988]: I1008 18:11:59.730842 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:11:59 crc kubenswrapper[4988]: I1008 18:11:59.730949 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:11:59 crc kubenswrapper[4988]: I1008 18:11:59.731064 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:11:59Z","lastTransitionTime":"2025-10-08T18:11:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:11:59 crc kubenswrapper[4988]: I1008 18:11:59.835299 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:11:59 crc kubenswrapper[4988]: I1008 18:11:59.835358 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:11:59 crc kubenswrapper[4988]: I1008 18:11:59.835370 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:11:59 crc kubenswrapper[4988]: I1008 18:11:59.835411 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:11:59 crc kubenswrapper[4988]: I1008 18:11:59.835425 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:11:59Z","lastTransitionTime":"2025-10-08T18:11:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:11:59 crc kubenswrapper[4988]: I1008 18:11:59.938712 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:11:59 crc kubenswrapper[4988]: I1008 18:11:59.938784 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:11:59 crc kubenswrapper[4988]: I1008 18:11:59.938807 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:11:59 crc kubenswrapper[4988]: I1008 18:11:59.938836 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:11:59 crc kubenswrapper[4988]: I1008 18:11:59.938857 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:11:59Z","lastTransitionTime":"2025-10-08T18:11:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:12:00 crc kubenswrapper[4988]: I1008 18:12:00.042867 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:12:00 crc kubenswrapper[4988]: I1008 18:12:00.043306 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:12:00 crc kubenswrapper[4988]: I1008 18:12:00.043477 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:12:00 crc kubenswrapper[4988]: I1008 18:12:00.043634 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:12:00 crc kubenswrapper[4988]: I1008 18:12:00.043762 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:12:00Z","lastTransitionTime":"2025-10-08T18:12:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:12:00 crc kubenswrapper[4988]: I1008 18:12:00.147637 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:12:00 crc kubenswrapper[4988]: I1008 18:12:00.147699 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:12:00 crc kubenswrapper[4988]: I1008 18:12:00.147718 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:12:00 crc kubenswrapper[4988]: I1008 18:12:00.147747 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:12:00 crc kubenswrapper[4988]: I1008 18:12:00.147765 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:12:00Z","lastTransitionTime":"2025-10-08T18:12:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:12:00 crc kubenswrapper[4988]: I1008 18:12:00.260256 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:12:00 crc kubenswrapper[4988]: I1008 18:12:00.260321 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:12:00 crc kubenswrapper[4988]: I1008 18:12:00.260336 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:12:00 crc kubenswrapper[4988]: I1008 18:12:00.260358 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:12:00 crc kubenswrapper[4988]: I1008 18:12:00.260375 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:12:00Z","lastTransitionTime":"2025-10-08T18:12:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:12:00 crc kubenswrapper[4988]: I1008 18:12:00.363650 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:12:00 crc kubenswrapper[4988]: I1008 18:12:00.363690 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:12:00 crc kubenswrapper[4988]: I1008 18:12:00.363699 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:12:00 crc kubenswrapper[4988]: I1008 18:12:00.363715 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:12:00 crc kubenswrapper[4988]: I1008 18:12:00.363728 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:12:00Z","lastTransitionTime":"2025-10-08T18:12:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:12:00 crc kubenswrapper[4988]: I1008 18:12:00.466445 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:12:00 crc kubenswrapper[4988]: I1008 18:12:00.466497 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:12:00 crc kubenswrapper[4988]: I1008 18:12:00.466509 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:12:00 crc kubenswrapper[4988]: I1008 18:12:00.466527 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:12:00 crc kubenswrapper[4988]: I1008 18:12:00.466540 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:12:00Z","lastTransitionTime":"2025-10-08T18:12:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:12:00 crc kubenswrapper[4988]: I1008 18:12:00.570248 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:12:00 crc kubenswrapper[4988]: I1008 18:12:00.570652 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:12:00 crc kubenswrapper[4988]: I1008 18:12:00.570780 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:12:00 crc kubenswrapper[4988]: I1008 18:12:00.570865 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:12:00 crc kubenswrapper[4988]: I1008 18:12:00.570949 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:12:00Z","lastTransitionTime":"2025-10-08T18:12:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:12:00 crc kubenswrapper[4988]: I1008 18:12:00.673685 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:12:00 crc kubenswrapper[4988]: I1008 18:12:00.673754 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:12:00 crc kubenswrapper[4988]: I1008 18:12:00.673770 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:12:00 crc kubenswrapper[4988]: I1008 18:12:00.673795 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:12:00 crc kubenswrapper[4988]: I1008 18:12:00.673814 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:12:00Z","lastTransitionTime":"2025-10-08T18:12:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:12:00 crc kubenswrapper[4988]: I1008 18:12:00.775949 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:12:00 crc kubenswrapper[4988]: I1008 18:12:00.776330 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:12:00 crc kubenswrapper[4988]: I1008 18:12:00.776416 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:12:00 crc kubenswrapper[4988]: I1008 18:12:00.776518 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:12:00 crc kubenswrapper[4988]: I1008 18:12:00.776588 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:12:00Z","lastTransitionTime":"2025-10-08T18:12:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:12:00 crc kubenswrapper[4988]: I1008 18:12:00.879536 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:12:00 crc kubenswrapper[4988]: I1008 18:12:00.879878 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:12:00 crc kubenswrapper[4988]: I1008 18:12:00.879943 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:12:00 crc kubenswrapper[4988]: I1008 18:12:00.880020 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:12:00 crc kubenswrapper[4988]: I1008 18:12:00.880090 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:12:00Z","lastTransitionTime":"2025-10-08T18:12:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:12:00 crc kubenswrapper[4988]: I1008 18:12:00.982682 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:12:00 crc kubenswrapper[4988]: I1008 18:12:00.983011 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:12:00 crc kubenswrapper[4988]: I1008 18:12:00.983070 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:12:00 crc kubenswrapper[4988]: I1008 18:12:00.983135 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:12:00 crc kubenswrapper[4988]: I1008 18:12:00.983195 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:12:00Z","lastTransitionTime":"2025-10-08T18:12:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:12:01 crc kubenswrapper[4988]: I1008 18:12:01.085424 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:12:01 crc kubenswrapper[4988]: I1008 18:12:01.085471 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:12:01 crc kubenswrapper[4988]: I1008 18:12:01.085480 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:12:01 crc kubenswrapper[4988]: I1008 18:12:01.085497 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:12:01 crc kubenswrapper[4988]: I1008 18:12:01.085507 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:12:01Z","lastTransitionTime":"2025-10-08T18:12:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:12:01 crc kubenswrapper[4988]: I1008 18:12:01.188449 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:12:01 crc kubenswrapper[4988]: I1008 18:12:01.188503 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:12:01 crc kubenswrapper[4988]: I1008 18:12:01.188512 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:12:01 crc kubenswrapper[4988]: I1008 18:12:01.188529 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:12:01 crc kubenswrapper[4988]: I1008 18:12:01.188541 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:12:01Z","lastTransitionTime":"2025-10-08T18:12:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:12:01 crc kubenswrapper[4988]: I1008 18:12:01.237496 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 18:12:01 crc kubenswrapper[4988]: I1008 18:12:01.237622 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jn2sx" Oct 08 18:12:01 crc kubenswrapper[4988]: I1008 18:12:01.237703 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 18:12:01 crc kubenswrapper[4988]: I1008 18:12:01.237632 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 18:12:01 crc kubenswrapper[4988]: E1008 18:12:01.237640 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 08 18:12:01 crc kubenswrapper[4988]: E1008 18:12:01.237835 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jn2sx" podUID="171e0902-95ee-4a39-8b1b-a58704c30574" Oct 08 18:12:01 crc kubenswrapper[4988]: E1008 18:12:01.237899 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 08 18:12:01 crc kubenswrapper[4988]: E1008 18:12:01.237972 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 08 18:12:01 crc kubenswrapper[4988]: I1008 18:12:01.291506 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:12:01 crc kubenswrapper[4988]: I1008 18:12:01.291547 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:12:01 crc kubenswrapper[4988]: I1008 18:12:01.291558 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:12:01 crc kubenswrapper[4988]: I1008 18:12:01.291572 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:12:01 crc kubenswrapper[4988]: I1008 18:12:01.291584 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:12:01Z","lastTransitionTime":"2025-10-08T18:12:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:12:01 crc kubenswrapper[4988]: I1008 18:12:01.394461 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:12:01 crc kubenswrapper[4988]: I1008 18:12:01.394506 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:12:01 crc kubenswrapper[4988]: I1008 18:12:01.394518 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:12:01 crc kubenswrapper[4988]: I1008 18:12:01.394538 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:12:01 crc kubenswrapper[4988]: I1008 18:12:01.394553 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:12:01Z","lastTransitionTime":"2025-10-08T18:12:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:12:01 crc kubenswrapper[4988]: I1008 18:12:01.498235 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:12:01 crc kubenswrapper[4988]: I1008 18:12:01.498298 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:12:01 crc kubenswrapper[4988]: I1008 18:12:01.498314 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:12:01 crc kubenswrapper[4988]: I1008 18:12:01.498337 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:12:01 crc kubenswrapper[4988]: I1008 18:12:01.498360 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:12:01Z","lastTransitionTime":"2025-10-08T18:12:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:12:01 crc kubenswrapper[4988]: I1008 18:12:01.602582 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:12:01 crc kubenswrapper[4988]: I1008 18:12:01.603519 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:12:01 crc kubenswrapper[4988]: I1008 18:12:01.603565 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:12:01 crc kubenswrapper[4988]: I1008 18:12:01.603593 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:12:01 crc kubenswrapper[4988]: I1008 18:12:01.603613 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:12:01Z","lastTransitionTime":"2025-10-08T18:12:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:12:01 crc kubenswrapper[4988]: I1008 18:12:01.706371 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:12:01 crc kubenswrapper[4988]: I1008 18:12:01.706463 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:12:01 crc kubenswrapper[4988]: I1008 18:12:01.706488 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:12:01 crc kubenswrapper[4988]: I1008 18:12:01.706521 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:12:01 crc kubenswrapper[4988]: I1008 18:12:01.706546 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:12:01Z","lastTransitionTime":"2025-10-08T18:12:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:12:01 crc kubenswrapper[4988]: I1008 18:12:01.809056 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:12:01 crc kubenswrapper[4988]: I1008 18:12:01.809125 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:12:01 crc kubenswrapper[4988]: I1008 18:12:01.809152 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:12:01 crc kubenswrapper[4988]: I1008 18:12:01.809184 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:12:01 crc kubenswrapper[4988]: I1008 18:12:01.809204 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:12:01Z","lastTransitionTime":"2025-10-08T18:12:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:12:01 crc kubenswrapper[4988]: I1008 18:12:01.911852 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:12:01 crc kubenswrapper[4988]: I1008 18:12:01.911910 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:12:01 crc kubenswrapper[4988]: I1008 18:12:01.911927 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:12:01 crc kubenswrapper[4988]: I1008 18:12:01.911951 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:12:01 crc kubenswrapper[4988]: I1008 18:12:01.911970 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:12:01Z","lastTransitionTime":"2025-10-08T18:12:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:12:02 crc kubenswrapper[4988]: I1008 18:12:02.016127 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:12:02 crc kubenswrapper[4988]: I1008 18:12:02.016173 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:12:02 crc kubenswrapper[4988]: I1008 18:12:02.016191 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:12:02 crc kubenswrapper[4988]: I1008 18:12:02.016214 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:12:02 crc kubenswrapper[4988]: I1008 18:12:02.016228 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:12:02Z","lastTransitionTime":"2025-10-08T18:12:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:12:02 crc kubenswrapper[4988]: I1008 18:12:02.118942 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:12:02 crc kubenswrapper[4988]: I1008 18:12:02.119019 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:12:02 crc kubenswrapper[4988]: I1008 18:12:02.119038 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:12:02 crc kubenswrapper[4988]: I1008 18:12:02.119068 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:12:02 crc kubenswrapper[4988]: I1008 18:12:02.119091 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:12:02Z","lastTransitionTime":"2025-10-08T18:12:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:12:02 crc kubenswrapper[4988]: I1008 18:12:02.222884 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:12:02 crc kubenswrapper[4988]: I1008 18:12:02.222962 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:12:02 crc kubenswrapper[4988]: I1008 18:12:02.222981 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:12:02 crc kubenswrapper[4988]: I1008 18:12:02.223010 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:12:02 crc kubenswrapper[4988]: I1008 18:12:02.223029 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:12:02Z","lastTransitionTime":"2025-10-08T18:12:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:12:02 crc kubenswrapper[4988]: I1008 18:12:02.327227 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:12:02 crc kubenswrapper[4988]: I1008 18:12:02.327313 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:12:02 crc kubenswrapper[4988]: I1008 18:12:02.327339 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:12:02 crc kubenswrapper[4988]: I1008 18:12:02.327368 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:12:02 crc kubenswrapper[4988]: I1008 18:12:02.327418 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:12:02Z","lastTransitionTime":"2025-10-08T18:12:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:12:02 crc kubenswrapper[4988]: I1008 18:12:02.428866 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:12:02 crc kubenswrapper[4988]: I1008 18:12:02.428918 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:12:02 crc kubenswrapper[4988]: I1008 18:12:02.428929 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:12:02 crc kubenswrapper[4988]: I1008 18:12:02.428948 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:12:02 crc kubenswrapper[4988]: I1008 18:12:02.428961 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:12:02Z","lastTransitionTime":"2025-10-08T18:12:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:12:02 crc kubenswrapper[4988]: E1008 18:12:02.445279 4988 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T18:12:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T18:12:02Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T18:12:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T18:12:02Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T18:12:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T18:12:02Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T18:12:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T18:12:02Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"95ee9fd0-8bc9-4ca5-9d79-5ee94d58bfeb\\\",\\\"systemUUID\\\":\\\"bd9489c4-fb7d-486d-b7ce-a13b6d7de36a\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:12:02Z is after 2025-08-24T17:21:41Z" Oct 08 18:12:02 crc kubenswrapper[4988]: I1008 18:12:02.451433 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:12:02 crc kubenswrapper[4988]: I1008 18:12:02.451466 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:12:02 crc kubenswrapper[4988]: I1008 18:12:02.451479 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:12:02 crc kubenswrapper[4988]: I1008 18:12:02.451499 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:12:02 crc kubenswrapper[4988]: I1008 18:12:02.451511 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:12:02Z","lastTransitionTime":"2025-10-08T18:12:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:12:02 crc kubenswrapper[4988]: E1008 18:12:02.476477 4988 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T18:12:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T18:12:02Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T18:12:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T18:12:02Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T18:12:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T18:12:02Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T18:12:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T18:12:02Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"95ee9fd0-8bc9-4ca5-9d79-5ee94d58bfeb\\\",\\\"systemUUID\\\":\\\"bd9489c4-fb7d-486d-b7ce-a13b6d7de36a\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:12:02Z is after 2025-08-24T17:21:41Z" Oct 08 18:12:02 crc kubenswrapper[4988]: I1008 18:12:02.481441 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:12:02 crc kubenswrapper[4988]: I1008 18:12:02.481512 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:12:02 crc kubenswrapper[4988]: I1008 18:12:02.481527 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:12:02 crc kubenswrapper[4988]: I1008 18:12:02.481552 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:12:02 crc kubenswrapper[4988]: I1008 18:12:02.481570 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:12:02Z","lastTransitionTime":"2025-10-08T18:12:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:12:02 crc kubenswrapper[4988]: E1008 18:12:02.503373 4988 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T18:12:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T18:12:02Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T18:12:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T18:12:02Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T18:12:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T18:12:02Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T18:12:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T18:12:02Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"95ee9fd0-8bc9-4ca5-9d79-5ee94d58bfeb\\\",\\\"systemUUID\\\":\\\"bd9489c4-fb7d-486d-b7ce-a13b6d7de36a\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:12:02Z is after 2025-08-24T17:21:41Z" Oct 08 18:12:02 crc kubenswrapper[4988]: I1008 18:12:02.507705 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:12:02 crc kubenswrapper[4988]: I1008 18:12:02.507740 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:12:02 crc kubenswrapper[4988]: I1008 18:12:02.507750 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:12:02 crc kubenswrapper[4988]: I1008 18:12:02.507767 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:12:02 crc kubenswrapper[4988]: I1008 18:12:02.507781 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:12:02Z","lastTransitionTime":"2025-10-08T18:12:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:12:02 crc kubenswrapper[4988]: E1008 18:12:02.519926 4988 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T18:12:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T18:12:02Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T18:12:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T18:12:02Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T18:12:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T18:12:02Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T18:12:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T18:12:02Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"95ee9fd0-8bc9-4ca5-9d79-5ee94d58bfeb\\\",\\\"systemUUID\\\":\\\"bd9489c4-fb7d-486d-b7ce-a13b6d7de36a\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:12:02Z is after 2025-08-24T17:21:41Z" Oct 08 18:12:02 crc kubenswrapper[4988]: I1008 18:12:02.523408 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:12:02 crc kubenswrapper[4988]: I1008 18:12:02.523463 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:12:02 crc kubenswrapper[4988]: I1008 18:12:02.523480 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:12:02 crc kubenswrapper[4988]: I1008 18:12:02.523505 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:12:02 crc kubenswrapper[4988]: I1008 18:12:02.523522 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:12:02Z","lastTransitionTime":"2025-10-08T18:12:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:12:02 crc kubenswrapper[4988]: E1008 18:12:02.535777 4988 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T18:12:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T18:12:02Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T18:12:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T18:12:02Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T18:12:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T18:12:02Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T18:12:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T18:12:02Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"95ee9fd0-8bc9-4ca5-9d79-5ee94d58bfeb\\\",\\\"systemUUID\\\":\\\"bd9489c4-fb7d-486d-b7ce-a13b6d7de36a\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:12:02Z is after 2025-08-24T17:21:41Z" Oct 08 18:12:02 crc kubenswrapper[4988]: E1008 18:12:02.535939 4988 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Oct 08 18:12:02 crc kubenswrapper[4988]: I1008 18:12:02.537587 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:12:02 crc kubenswrapper[4988]: I1008 18:12:02.537655 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:12:02 crc kubenswrapper[4988]: I1008 18:12:02.537673 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:12:02 crc kubenswrapper[4988]: I1008 18:12:02.537696 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:12:02 crc kubenswrapper[4988]: I1008 18:12:02.537711 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:12:02Z","lastTransitionTime":"2025-10-08T18:12:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:12:02 crc kubenswrapper[4988]: I1008 18:12:02.640449 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:12:02 crc kubenswrapper[4988]: I1008 18:12:02.640488 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:12:02 crc kubenswrapper[4988]: I1008 18:12:02.640497 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:12:02 crc kubenswrapper[4988]: I1008 18:12:02.640513 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:12:02 crc kubenswrapper[4988]: I1008 18:12:02.640523 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:12:02Z","lastTransitionTime":"2025-10-08T18:12:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:12:02 crc kubenswrapper[4988]: I1008 18:12:02.744004 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:12:02 crc kubenswrapper[4988]: I1008 18:12:02.744071 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:12:02 crc kubenswrapper[4988]: I1008 18:12:02.744090 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:12:02 crc kubenswrapper[4988]: I1008 18:12:02.744123 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:12:02 crc kubenswrapper[4988]: I1008 18:12:02.744147 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:12:02Z","lastTransitionTime":"2025-10-08T18:12:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:12:02 crc kubenswrapper[4988]: I1008 18:12:02.847756 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:12:02 crc kubenswrapper[4988]: I1008 18:12:02.847792 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:12:02 crc kubenswrapper[4988]: I1008 18:12:02.847801 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:12:02 crc kubenswrapper[4988]: I1008 18:12:02.847816 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:12:02 crc kubenswrapper[4988]: I1008 18:12:02.847827 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:12:02Z","lastTransitionTime":"2025-10-08T18:12:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:12:02 crc kubenswrapper[4988]: I1008 18:12:02.941898 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/171e0902-95ee-4a39-8b1b-a58704c30574-metrics-certs\") pod \"network-metrics-daemon-jn2sx\" (UID: \"171e0902-95ee-4a39-8b1b-a58704c30574\") " pod="openshift-multus/network-metrics-daemon-jn2sx" Oct 08 18:12:02 crc kubenswrapper[4988]: E1008 18:12:02.942150 4988 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 08 18:12:02 crc kubenswrapper[4988]: E1008 18:12:02.942307 4988 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/171e0902-95ee-4a39-8b1b-a58704c30574-metrics-certs podName:171e0902-95ee-4a39-8b1b-a58704c30574 nodeName:}" failed. No retries permitted until 2025-10-08 18:12:34.942271203 +0000 UTC m=+100.392114013 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/171e0902-95ee-4a39-8b1b-a58704c30574-metrics-certs") pod "network-metrics-daemon-jn2sx" (UID: "171e0902-95ee-4a39-8b1b-a58704c30574") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 08 18:12:02 crc kubenswrapper[4988]: I1008 18:12:02.950448 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:12:02 crc kubenswrapper[4988]: I1008 18:12:02.950529 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:12:02 crc kubenswrapper[4988]: I1008 18:12:02.950547 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:12:02 crc kubenswrapper[4988]: I1008 18:12:02.950751 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:12:02 crc kubenswrapper[4988]: I1008 18:12:02.950766 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:12:02Z","lastTransitionTime":"2025-10-08T18:12:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:12:03 crc kubenswrapper[4988]: I1008 18:12:03.054256 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:12:03 crc kubenswrapper[4988]: I1008 18:12:03.054302 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:12:03 crc kubenswrapper[4988]: I1008 18:12:03.054312 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:12:03 crc kubenswrapper[4988]: I1008 18:12:03.054331 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:12:03 crc kubenswrapper[4988]: I1008 18:12:03.054345 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:12:03Z","lastTransitionTime":"2025-10-08T18:12:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:12:03 crc kubenswrapper[4988]: I1008 18:12:03.156949 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:12:03 crc kubenswrapper[4988]: I1008 18:12:03.156996 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:12:03 crc kubenswrapper[4988]: I1008 18:12:03.157010 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:12:03 crc kubenswrapper[4988]: I1008 18:12:03.157030 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:12:03 crc kubenswrapper[4988]: I1008 18:12:03.157042 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:12:03Z","lastTransitionTime":"2025-10-08T18:12:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:12:03 crc kubenswrapper[4988]: I1008 18:12:03.237040 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 18:12:03 crc kubenswrapper[4988]: I1008 18:12:03.237092 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 18:12:03 crc kubenswrapper[4988]: I1008 18:12:03.237047 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 18:12:03 crc kubenswrapper[4988]: I1008 18:12:03.237047 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jn2sx" Oct 08 18:12:03 crc kubenswrapper[4988]: E1008 18:12:03.237251 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 08 18:12:03 crc kubenswrapper[4988]: E1008 18:12:03.237542 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jn2sx" podUID="171e0902-95ee-4a39-8b1b-a58704c30574" Oct 08 18:12:03 crc kubenswrapper[4988]: E1008 18:12:03.237640 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 08 18:12:03 crc kubenswrapper[4988]: E1008 18:12:03.237770 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 08 18:12:03 crc kubenswrapper[4988]: I1008 18:12:03.259497 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:12:03 crc kubenswrapper[4988]: I1008 18:12:03.259560 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:12:03 crc kubenswrapper[4988]: I1008 18:12:03.259578 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:12:03 crc kubenswrapper[4988]: I1008 18:12:03.259604 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:12:03 crc kubenswrapper[4988]: I1008 18:12:03.259624 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:12:03Z","lastTransitionTime":"2025-10-08T18:12:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:12:03 crc kubenswrapper[4988]: I1008 18:12:03.363031 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:12:03 crc kubenswrapper[4988]: I1008 18:12:03.363141 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:12:03 crc kubenswrapper[4988]: I1008 18:12:03.363167 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:12:03 crc kubenswrapper[4988]: I1008 18:12:03.363203 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:12:03 crc kubenswrapper[4988]: I1008 18:12:03.363262 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:12:03Z","lastTransitionTime":"2025-10-08T18:12:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:12:03 crc kubenswrapper[4988]: I1008 18:12:03.466550 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:12:03 crc kubenswrapper[4988]: I1008 18:12:03.466645 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:12:03 crc kubenswrapper[4988]: I1008 18:12:03.466663 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:12:03 crc kubenswrapper[4988]: I1008 18:12:03.466689 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:12:03 crc kubenswrapper[4988]: I1008 18:12:03.466707 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:12:03Z","lastTransitionTime":"2025-10-08T18:12:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:12:03 crc kubenswrapper[4988]: I1008 18:12:03.569243 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:12:03 crc kubenswrapper[4988]: I1008 18:12:03.569326 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:12:03 crc kubenswrapper[4988]: I1008 18:12:03.569345 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:12:03 crc kubenswrapper[4988]: I1008 18:12:03.569369 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:12:03 crc kubenswrapper[4988]: I1008 18:12:03.569406 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:12:03Z","lastTransitionTime":"2025-10-08T18:12:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:12:03 crc kubenswrapper[4988]: I1008 18:12:03.673074 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:12:03 crc kubenswrapper[4988]: I1008 18:12:03.673140 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:12:03 crc kubenswrapper[4988]: I1008 18:12:03.673158 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:12:03 crc kubenswrapper[4988]: I1008 18:12:03.673187 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:12:03 crc kubenswrapper[4988]: I1008 18:12:03.673205 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:12:03Z","lastTransitionTime":"2025-10-08T18:12:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:12:03 crc kubenswrapper[4988]: I1008 18:12:03.776500 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:12:03 crc kubenswrapper[4988]: I1008 18:12:03.776564 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:12:03 crc kubenswrapper[4988]: I1008 18:12:03.776585 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:12:03 crc kubenswrapper[4988]: I1008 18:12:03.776611 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:12:03 crc kubenswrapper[4988]: I1008 18:12:03.776632 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:12:03Z","lastTransitionTime":"2025-10-08T18:12:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:12:03 crc kubenswrapper[4988]: I1008 18:12:03.880410 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:12:03 crc kubenswrapper[4988]: I1008 18:12:03.880498 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:12:03 crc kubenswrapper[4988]: I1008 18:12:03.880517 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:12:03 crc kubenswrapper[4988]: I1008 18:12:03.880607 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:12:03 crc kubenswrapper[4988]: I1008 18:12:03.880629 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:12:03Z","lastTransitionTime":"2025-10-08T18:12:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:12:03 crc kubenswrapper[4988]: I1008 18:12:03.983293 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:12:03 crc kubenswrapper[4988]: I1008 18:12:03.983345 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:12:03 crc kubenswrapper[4988]: I1008 18:12:03.983356 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:12:03 crc kubenswrapper[4988]: I1008 18:12:03.983373 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:12:03 crc kubenswrapper[4988]: I1008 18:12:03.983406 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:12:03Z","lastTransitionTime":"2025-10-08T18:12:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:12:04 crc kubenswrapper[4988]: I1008 18:12:04.086467 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:12:04 crc kubenswrapper[4988]: I1008 18:12:04.086520 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:12:04 crc kubenswrapper[4988]: I1008 18:12:04.086532 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:12:04 crc kubenswrapper[4988]: I1008 18:12:04.086550 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:12:04 crc kubenswrapper[4988]: I1008 18:12:04.086562 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:12:04Z","lastTransitionTime":"2025-10-08T18:12:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:12:04 crc kubenswrapper[4988]: I1008 18:12:04.190562 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:12:04 crc kubenswrapper[4988]: I1008 18:12:04.190647 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:12:04 crc kubenswrapper[4988]: I1008 18:12:04.190668 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:12:04 crc kubenswrapper[4988]: I1008 18:12:04.190692 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:12:04 crc kubenswrapper[4988]: I1008 18:12:04.190712 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:12:04Z","lastTransitionTime":"2025-10-08T18:12:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:12:04 crc kubenswrapper[4988]: I1008 18:12:04.294122 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:12:04 crc kubenswrapper[4988]: I1008 18:12:04.294199 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:12:04 crc kubenswrapper[4988]: I1008 18:12:04.294221 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:12:04 crc kubenswrapper[4988]: I1008 18:12:04.294255 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:12:04 crc kubenswrapper[4988]: I1008 18:12:04.294285 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:12:04Z","lastTransitionTime":"2025-10-08T18:12:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:12:04 crc kubenswrapper[4988]: I1008 18:12:04.397289 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:12:04 crc kubenswrapper[4988]: I1008 18:12:04.397348 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:12:04 crc kubenswrapper[4988]: I1008 18:12:04.397359 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:12:04 crc kubenswrapper[4988]: I1008 18:12:04.397378 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:12:04 crc kubenswrapper[4988]: I1008 18:12:04.397411 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:12:04Z","lastTransitionTime":"2025-10-08T18:12:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:12:04 crc kubenswrapper[4988]: I1008 18:12:04.500334 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:12:04 crc kubenswrapper[4988]: I1008 18:12:04.500412 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:12:04 crc kubenswrapper[4988]: I1008 18:12:04.500426 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:12:04 crc kubenswrapper[4988]: I1008 18:12:04.500442 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:12:04 crc kubenswrapper[4988]: I1008 18:12:04.500456 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:12:04Z","lastTransitionTime":"2025-10-08T18:12:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:12:04 crc kubenswrapper[4988]: I1008 18:12:04.603824 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:12:04 crc kubenswrapper[4988]: I1008 18:12:04.603902 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:12:04 crc kubenswrapper[4988]: I1008 18:12:04.603928 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:12:04 crc kubenswrapper[4988]: I1008 18:12:04.603987 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:12:04 crc kubenswrapper[4988]: I1008 18:12:04.604010 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:12:04Z","lastTransitionTime":"2025-10-08T18:12:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:12:04 crc kubenswrapper[4988]: I1008 18:12:04.706735 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:12:04 crc kubenswrapper[4988]: I1008 18:12:04.706790 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:12:04 crc kubenswrapper[4988]: I1008 18:12:04.706801 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:12:04 crc kubenswrapper[4988]: I1008 18:12:04.706823 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:12:04 crc kubenswrapper[4988]: I1008 18:12:04.706836 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:12:04Z","lastTransitionTime":"2025-10-08T18:12:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:12:04 crc kubenswrapper[4988]: I1008 18:12:04.791613 4988 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-h9jwp_4490d846-ca32-4cd5-9202-87cc6d499fb4/kube-multus/0.log" Oct 08 18:12:04 crc kubenswrapper[4988]: I1008 18:12:04.791690 4988 generic.go:334] "Generic (PLEG): container finished" podID="4490d846-ca32-4cd5-9202-87cc6d499fb4" containerID="00b7f6db043c28efda3a09fdc77a6ccded3eae8b6a3d2852aa500789d5f01194" exitCode=1 Oct 08 18:12:04 crc kubenswrapper[4988]: I1008 18:12:04.791737 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-h9jwp" event={"ID":"4490d846-ca32-4cd5-9202-87cc6d499fb4","Type":"ContainerDied","Data":"00b7f6db043c28efda3a09fdc77a6ccded3eae8b6a3d2852aa500789d5f01194"} Oct 08 18:12:04 crc kubenswrapper[4988]: I1008 18:12:04.792220 4988 scope.go:117] "RemoveContainer" containerID="00b7f6db043c28efda3a09fdc77a6ccded3eae8b6a3d2852aa500789d5f01194" Oct 08 18:12:04 crc kubenswrapper[4988]: I1008 18:12:04.808533 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e8906182d55c285b3cf095e2d9e1dac397af4618a514f11e26b44c172dd573f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:12:04Z is after 2025-08-24T17:21:41Z" Oct 08 18:12:04 crc kubenswrapper[4988]: I1008 18:12:04.809173 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:12:04 crc kubenswrapper[4988]: I1008 18:12:04.809214 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:12:04 crc kubenswrapper[4988]: I1008 18:12:04.809227 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:12:04 crc kubenswrapper[4988]: I1008 18:12:04.809247 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:12:04 crc kubenswrapper[4988]: I1008 18:12:04.809262 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:12:04Z","lastTransitionTime":"2025-10-08T18:12:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:12:04 crc kubenswrapper[4988]: I1008 18:12:04.823103 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-dzxs2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8f307b20-5e33-4725-99cb-210e25113532\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72ee803c6ef88524e0ce0391476f8e5a7dfbc858b095a3c2ae31bad10991ce3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q65f7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:11:16Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-dzxs2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:12:04Z is after 2025-08-24T17:21:41Z" Oct 08 18:12:04 crc kubenswrapper[4988]: I1008 18:12:04.839594 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aef71be1-1d12-46e5-95b2-bd399040fd4d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:10:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:10:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:10:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77c68f3ecb8a0724499e495cb26ef3db5beecb140f2558d8c0b25f41928b84ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:10:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a560ee84c2405227de71fda127566213cd4f017e73d5581442ba267781b1c539\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:10:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6100accc3bd08928bc1ff88310f29fa5da37fd2d5c315c94279801c9778c4745\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:10:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b0e596d7ce4e9341c44499af79d3b3caf0e5e2c156902ea40cf2a3dca672be9e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5bd35aef24be034a0b305a40bb53ca171a4f76db3b93b4f78ae95099b0a4cde3\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-08T18:11:14Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1008 18:11:09.091034 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1008 18:11:09.091984 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3472374062/tls.crt::/tmp/serving-cert-3472374062/tls.key\\\\\\\"\\\\nI1008 18:11:14.619696 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1008 18:11:14.628837 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1008 18:11:14.628898 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1008 18:11:14.628994 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1008 18:11:14.629036 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1008 18:11:14.641318 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1008 18:11:14.641363 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1008 18:11:14.641372 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1008 18:11:14.641382 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1008 18:11:14.641415 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1008 18:11:14.641423 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1008 18:11:14.641431 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1008 18:11:14.641484 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1008 18:11:14.646451 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T18:10:58Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c79ef704fda1bbcc486f4c7afa3039a3ab38669833f0f66bafe24dd4cc679e60\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:10:58Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://28d21731af0f16d6a8ae6c6e9d0bebf0560d4a533db4c78c10bf1837890394ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://28d21731af0f16d6a8ae6c6e9d0bebf0560d4a533db4c78c10bf1837890394ef\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:10:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:10:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:10:55Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:12:04Z is after 2025-08-24T17:21:41Z" Oct 08 18:12:04 crc kubenswrapper[4988]: I1008 18:12:04.854534 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7e0ecefc-359f-4785-bb79-0343a62a25da\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:10:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:10:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:10:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://30bec350fefee3d19758b3ed735ede4e9aa085edb57aa3f3ec24249a67dc7ca2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:10:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2d850fd27d4ed65891ff69b9add65fb78535357f54819c2a8d196f5a340c58e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:10:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://90b1a9e14c82bf0f75f08fdb342eed35a503020f9db641e5e9ceff8b122cb3bd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:10:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://16f23c4c7163d32dc0df8d6a7ee17e2864fb261f8a40ecf05556774b173fc868\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:10:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:10:55Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:12:04Z is after 2025-08-24T17:21:41Z" Oct 08 18:12:04 crc kubenswrapper[4988]: I1008 18:12:04.868513 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9590ca0059f0ebbe86343cccb3475338c317382c3231e082655b61e67eece375\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:12:04Z is after 2025-08-24T17:21:41Z" Oct 08 18:12:04 crc kubenswrapper[4988]: I1008 18:12:04.881492 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:12:04Z is after 2025-08-24T17:21:41Z" Oct 08 18:12:04 crc kubenswrapper[4988]: I1008 18:12:04.897321 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-jnh7m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"556fb048-8afb-44a5-9ea0-2b7c89e6cf4b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bfa2f3382b23c3f788a1b31e80e8a90b3b6868870dccc46d86c3fc11ad027f52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmlt2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1a825886f105ccf9a7cc6f28a0cd012fb859f4990dade0e6aae932482146028\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1a825886f105ccf9a7cc6f28a0cd012fb859f4990dade0e6aae932482146028\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmlt2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ab2fe6e65418efa75bae72259157a2323bbb4b7d11d1d012053524e14135888\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0ab2fe6e65418efa75bae72259157a2323bbb4b7d11d1d012053524e14135888\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:11:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:11:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmlt2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://368a303309be112d9c0b7ccd6538fbaca7085d97a8ec8259f36092f07147e7cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://368a303309be112d9c0b7ccd6538fbaca7085d97a8ec8259f36092f07147e7cb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:11:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:11:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmlt2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a8ca69001dddef643f982a97c74c291e30cb9fac5b0e894e80066b885ea9151\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9a8ca69001dddef643f982a97c74c291e30cb9fac5b0e894e80066b885ea9151\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:11:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:11:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmlt2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://145303c46b48603ed1aae4b9bfe06332a3ffad28f206f8394574011a87b58cd3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://145303c46b48603ed1aae4b9bfe06332a3ffad28f206f8394574011a87b58cd3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:11:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:11:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmlt2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d08e5cfc7cbe5ec1a7a0265ed7c466363c3abeec492263170860689b46c7ff95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d08e5cfc7cbe5ec1a7a0265ed7c466363c3abeec492263170860689b46c7ff95\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:11:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:11:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmlt2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:11:16Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-jnh7m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:12:04Z is after 2025-08-24T17:21:41Z" Oct 08 18:12:04 crc kubenswrapper[4988]: I1008 18:12:04.912073 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:12:04 crc kubenswrapper[4988]: I1008 18:12:04.912123 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:12:04 crc kubenswrapper[4988]: I1008 18:12:04.912156 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:12:04 crc kubenswrapper[4988]: I1008 18:12:04.912178 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:12:04 crc kubenswrapper[4988]: I1008 18:12:04.912196 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:12:04Z","lastTransitionTime":"2025-10-08T18:12:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:12:04 crc kubenswrapper[4988]: I1008 18:12:04.912919 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f328a20b-5283-407e-8a93-2ab2da599f18\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8622d56fb198dadee31fb0ca8fc80d56828b6710648055f264533e2d5bdb3037\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z2gh4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3cdb2a757332bdcdc2a2f3362913fe8f06003b0597de20b72627f9badd5e30af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z2gh4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:11:16Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-nm54f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:12:04Z is after 2025-08-24T17:21:41Z" Oct 08 18:12:04 crc kubenswrapper[4988]: I1008 18:12:04.928823 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a35a5e989ba69c8ef36ce4cea272aadf67ce20473fc257adfcfa86a2faba6ae0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a522e937cfe03fb9f2e60cb1bea07214cf63c69bdecd5ba65a2dfdba882d28dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:12:04Z is after 2025-08-24T17:21:41Z" Oct 08 18:12:04 crc kubenswrapper[4988]: I1008 18:12:04.942868 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:12:04Z is after 2025-08-24T17:21:41Z" Oct 08 18:12:04 crc kubenswrapper[4988]: I1008 18:12:04.954252 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-h9jwp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4490d846-ca32-4cd5-9202-87cc6d499fb4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:12:04Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:12:04Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://00b7f6db043c28efda3a09fdc77a6ccded3eae8b6a3d2852aa500789d5f01194\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://00b7f6db043c28efda3a09fdc77a6ccded3eae8b6a3d2852aa500789d5f01194\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-08T18:12:04Z\\\",\\\"message\\\":\\\"2025-10-08T18:11:18+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_bee291bc-08e6-4092-b05c-4207d1f7fe40\\\\n2025-10-08T18:11:18+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_bee291bc-08e6-4092-b05c-4207d1f7fe40 to /host/opt/cni/bin/\\\\n2025-10-08T18:11:18Z [verbose] multus-daemon started\\\\n2025-10-08T18:11:18Z [verbose] Readiness Indicator file check\\\\n2025-10-08T18:12:03Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T18:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7k6rk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:11:16Z\\\"}}\" for pod \"openshift-multus\"/\"multus-h9jwp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:12:04Z is after 2025-08-24T17:21:41Z" Oct 08 18:12:04 crc kubenswrapper[4988]: I1008 18:12:04.971800 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-nsc65" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f010a463-69f2-4082-acc3-161899a939c3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://48a516f6c153743bd1a21ec6945fb368a37f4d2b9d6a326b68edc096a192ed05\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dnwgr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://03406e726f817bcedb841f81959ab12808ba35a530240f9abc5fa5c608b9a6ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dnwgr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9fbbeae0409d2e47e8de9a4306f63b0d13acbfd73cd86f83303fec14352ed718\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dnwgr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://55791327a370c0009c523b86b2234e82d3794985f2eaccf1922eb90f46b2b970\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dnwgr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cee5de93bc68345f9b737857f02832f56582d58757c50a8798eec25d19b2c20d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dnwgr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d314eb248e90371fed068b1173e26d4c9f9765417a70b341e1a5b136ec17353a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dnwgr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7bf11dd48efebb9da2b836bfad6a54b274247fa60902dab2f084793e4ad51a89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7bf11dd48efebb9da2b836bfad6a54b274247fa60902dab2f084793e4ad51a89\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-08T18:11:42Z\\\",\\\"message\\\":\\\" V4IPs:[]string(nil), V6IPs:[]string(nil)}, nodeEndpoints:map[string]services.lbEndpoints{}, externalTrafficLocal:false, internalTrafficLocal:false, hasNodePort:false}}\\\\nI1008 18:11:42.211985 6628 services_controller.go:444] Built service openshift-authentication-operator/metrics LB per-node configs for network=default: []services.lbConfig(nil)\\\\nI1008 18:11:42.211998 6628 services_controller.go:445] Built service openshift-authentication-operator/metrics LB template configs for network=default: []services.lbConfig(nil)\\\\nF1008 18:11:42.212009 6628 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:42Z is after 2025-08-\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T18:11:41Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-nsc65_openshift-ovn-kubernetes(f010a463-69f2-4082-acc3-161899a939c3)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dnwgr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://74fcb812b84291763cc6e5ccb20c12aaad580ad968dda4734a0bbfe33d2770ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dnwgr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://befe684cea9bd6dfda3ec8e85bca431c26bd8f82823b0a69dc0d04319a3ca385\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://befe684cea9bd6dfda3ec8e85bca431c26bd8f82823b0a69dc0d04319a3ca385\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dnwgr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:11:16Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-nsc65\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:12:04Z is after 2025-08-24T17:21:41Z" Oct 08 18:12:04 crc kubenswrapper[4988]: I1008 18:12:04.988086 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-7wl8v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e433488d-db52-435f-b89c-f63cd477c5d3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76b30d183684740b647c1af11dc49712232cc8b0c9a392b491580e5ddb75ea7f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dbz6l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e55016c3e1b97d1b3a857130a5c8e2ec6da2333f786032118a61e3e2ad4c5651\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dbz6l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:11:29Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-7wl8v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:12:04Z is after 2025-08-24T17:21:41Z" Oct 08 18:12:05 crc kubenswrapper[4988]: I1008 18:12:05.001890 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-jn2sx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"171e0902-95ee-4a39-8b1b-a58704c30574\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9f4q2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9f4q2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:11:31Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-jn2sx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:12:04Z is after 2025-08-24T17:21:41Z" Oct 08 18:12:05 crc kubenswrapper[4988]: I1008 18:12:05.014819 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:12:05 crc kubenswrapper[4988]: I1008 18:12:05.014862 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:12:05 crc kubenswrapper[4988]: I1008 18:12:05.014878 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:12:05 crc kubenswrapper[4988]: I1008 18:12:05.014897 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:12:05 crc kubenswrapper[4988]: I1008 18:12:05.014908 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:12:05Z","lastTransitionTime":"2025-10-08T18:12:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:12:05 crc kubenswrapper[4988]: I1008 18:12:05.016874 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d36f620-0f26-41c7-af6e-7fb5721a0e2d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:10:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:10:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:10:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3fd3b2ca80233400d2798b035a65b833f735632efd76d1160f6bb906e07b80a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:10:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://75a2bd2757133e79c0c7f2939eb433dc45291f14808e3c3f1daaf90a239c6a71\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:10:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5cb03de583c14ad6751227c21166ac5ac4c6e11b67d5b709361ced8a0877e6d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:10:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6eb46d76cf7cfb1d4ecc4fcee398dad8904732c1664c317865864b65f5703589\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6eb46d76cf7cfb1d4ecc4fcee398dad8904732c1664c317865864b65f5703589\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:10:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:10:56Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:10:55Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:12:05Z is after 2025-08-24T17:21:41Z" Oct 08 18:12:05 crc kubenswrapper[4988]: I1008 18:12:05.030216 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:12:05Z is after 2025-08-24T17:21:41Z" Oct 08 18:12:05 crc kubenswrapper[4988]: I1008 18:12:05.042442 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-s4f2k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ba5e11d7-7bde-4c3d-983b-884232219652\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://824bc34a2807068d3fceb10651803ae9ae5651e940f7aed4877c544d76dae4d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hn5jf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:11:15Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-s4f2k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:12:05Z is after 2025-08-24T17:21:41Z" Oct 08 18:12:05 crc kubenswrapper[4988]: I1008 18:12:05.118027 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:12:05 crc kubenswrapper[4988]: I1008 18:12:05.118085 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:12:05 crc kubenswrapper[4988]: I1008 18:12:05.118098 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:12:05 crc kubenswrapper[4988]: I1008 18:12:05.118116 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:12:05 crc kubenswrapper[4988]: I1008 18:12:05.118129 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:12:05Z","lastTransitionTime":"2025-10-08T18:12:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:12:05 crc kubenswrapper[4988]: I1008 18:12:05.221401 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:12:05 crc kubenswrapper[4988]: I1008 18:12:05.221474 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:12:05 crc kubenswrapper[4988]: I1008 18:12:05.221487 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:12:05 crc kubenswrapper[4988]: I1008 18:12:05.221511 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:12:05 crc kubenswrapper[4988]: I1008 18:12:05.221527 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:12:05Z","lastTransitionTime":"2025-10-08T18:12:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:12:05 crc kubenswrapper[4988]: I1008 18:12:05.237346 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 18:12:05 crc kubenswrapper[4988]: I1008 18:12:05.237432 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 18:12:05 crc kubenswrapper[4988]: I1008 18:12:05.237568 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jn2sx" Oct 08 18:12:05 crc kubenswrapper[4988]: I1008 18:12:05.237763 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 18:12:05 crc kubenswrapper[4988]: E1008 18:12:05.237979 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 08 18:12:05 crc kubenswrapper[4988]: E1008 18:12:05.238135 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 08 18:12:05 crc kubenswrapper[4988]: E1008 18:12:05.238331 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jn2sx" podUID="171e0902-95ee-4a39-8b1b-a58704c30574" Oct 08 18:12:05 crc kubenswrapper[4988]: E1008 18:12:05.238464 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 08 18:12:05 crc kubenswrapper[4988]: I1008 18:12:05.252112 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/kube-rbac-proxy-crio-crc"] Oct 08 18:12:05 crc kubenswrapper[4988]: I1008 18:12:05.255364 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-dzxs2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8f307b20-5e33-4725-99cb-210e25113532\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72ee803c6ef88524e0ce0391476f8e5a7dfbc858b095a3c2ae31bad10991ce3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q65f7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:11:16Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-dzxs2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:12:05Z is after 2025-08-24T17:21:41Z" Oct 08 18:12:05 crc kubenswrapper[4988]: I1008 18:12:05.272754 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aef71be1-1d12-46e5-95b2-bd399040fd4d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:10:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:10:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:10:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77c68f3ecb8a0724499e495cb26ef3db5beecb140f2558d8c0b25f41928b84ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:10:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a560ee84c2405227de71fda127566213cd4f017e73d5581442ba267781b1c539\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:10:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6100accc3bd08928bc1ff88310f29fa5da37fd2d5c315c94279801c9778c4745\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:10:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b0e596d7ce4e9341c44499af79d3b3caf0e5e2c156902ea40cf2a3dca672be9e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5bd35aef24be034a0b305a40bb53ca171a4f76db3b93b4f78ae95099b0a4cde3\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-08T18:11:14Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1008 18:11:09.091034 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1008 18:11:09.091984 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3472374062/tls.crt::/tmp/serving-cert-3472374062/tls.key\\\\\\\"\\\\nI1008 18:11:14.619696 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1008 18:11:14.628837 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1008 18:11:14.628898 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1008 18:11:14.628994 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1008 18:11:14.629036 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1008 18:11:14.641318 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1008 18:11:14.641363 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1008 18:11:14.641372 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1008 18:11:14.641382 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1008 18:11:14.641415 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1008 18:11:14.641423 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1008 18:11:14.641431 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1008 18:11:14.641484 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1008 18:11:14.646451 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T18:10:58Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c79ef704fda1bbcc486f4c7afa3039a3ab38669833f0f66bafe24dd4cc679e60\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:10:58Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://28d21731af0f16d6a8ae6c6e9d0bebf0560d4a533db4c78c10bf1837890394ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://28d21731af0f16d6a8ae6c6e9d0bebf0560d4a533db4c78c10bf1837890394ef\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:10:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:10:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:10:55Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:12:05Z is after 2025-08-24T17:21:41Z" Oct 08 18:12:05 crc kubenswrapper[4988]: I1008 18:12:05.289581 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7e0ecefc-359f-4785-bb79-0343a62a25da\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:10:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:10:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:10:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://30bec350fefee3d19758b3ed735ede4e9aa085edb57aa3f3ec24249a67dc7ca2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:10:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2d850fd27d4ed65891ff69b9add65fb78535357f54819c2a8d196f5a340c58e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:10:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://90b1a9e14c82bf0f75f08fdb342eed35a503020f9db641e5e9ceff8b122cb3bd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:10:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://16f23c4c7163d32dc0df8d6a7ee17e2864fb261f8a40ecf05556774b173fc868\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:10:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:10:55Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:12:05Z is after 2025-08-24T17:21:41Z" Oct 08 18:12:05 crc kubenswrapper[4988]: I1008 18:12:05.309027 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e8906182d55c285b3cf095e2d9e1dac397af4618a514f11e26b44c172dd573f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:12:05Z is after 2025-08-24T17:21:41Z" Oct 08 18:12:05 crc kubenswrapper[4988]: I1008 18:12:05.324500 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:12:05 crc kubenswrapper[4988]: I1008 18:12:05.324538 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:12:05 crc kubenswrapper[4988]: I1008 18:12:05.324551 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:12:05 crc kubenswrapper[4988]: I1008 18:12:05.324568 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:12:05 crc kubenswrapper[4988]: I1008 18:12:05.324580 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:12:05Z","lastTransitionTime":"2025-10-08T18:12:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:12:05 crc kubenswrapper[4988]: I1008 18:12:05.328445 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:12:05Z is after 2025-08-24T17:21:41Z" Oct 08 18:12:05 crc kubenswrapper[4988]: I1008 18:12:05.348776 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-jnh7m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"556fb048-8afb-44a5-9ea0-2b7c89e6cf4b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bfa2f3382b23c3f788a1b31e80e8a90b3b6868870dccc46d86c3fc11ad027f52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmlt2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1a825886f105ccf9a7cc6f28a0cd012fb859f4990dade0e6aae932482146028\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1a825886f105ccf9a7cc6f28a0cd012fb859f4990dade0e6aae932482146028\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmlt2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ab2fe6e65418efa75bae72259157a2323bbb4b7d11d1d012053524e14135888\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0ab2fe6e65418efa75bae72259157a2323bbb4b7d11d1d012053524e14135888\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:11:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:11:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmlt2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://368a303309be112d9c0b7ccd6538fbaca7085d97a8ec8259f36092f07147e7cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://368a303309be112d9c0b7ccd6538fbaca7085d97a8ec8259f36092f07147e7cb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:11:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:11:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmlt2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a8ca69001dddef643f982a97c74c291e30cb9fac5b0e894e80066b885ea9151\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9a8ca69001dddef643f982a97c74c291e30cb9fac5b0e894e80066b885ea9151\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:11:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:11:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmlt2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://145303c46b48603ed1aae4b9bfe06332a3ffad28f206f8394574011a87b58cd3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://145303c46b48603ed1aae4b9bfe06332a3ffad28f206f8394574011a87b58cd3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:11:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:11:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmlt2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d08e5cfc7cbe5ec1a7a0265ed7c466363c3abeec492263170860689b46c7ff95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d08e5cfc7cbe5ec1a7a0265ed7c466363c3abeec492263170860689b46c7ff95\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:11:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:11:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmlt2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:11:16Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-jnh7m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:12:05Z is after 2025-08-24T17:21:41Z" Oct 08 18:12:05 crc kubenswrapper[4988]: I1008 18:12:05.361095 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f328a20b-5283-407e-8a93-2ab2da599f18\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8622d56fb198dadee31fb0ca8fc80d56828b6710648055f264533e2d5bdb3037\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z2gh4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3cdb2a757332bdcdc2a2f3362913fe8f06003b0597de20b72627f9badd5e30af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z2gh4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:11:16Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-nm54f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:12:05Z is after 2025-08-24T17:21:41Z" Oct 08 18:12:05 crc kubenswrapper[4988]: I1008 18:12:05.374721 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a35a5e989ba69c8ef36ce4cea272aadf67ce20473fc257adfcfa86a2faba6ae0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a522e937cfe03fb9f2e60cb1bea07214cf63c69bdecd5ba65a2dfdba882d28dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:12:05Z is after 2025-08-24T17:21:41Z" Oct 08 18:12:05 crc kubenswrapper[4988]: I1008 18:12:05.391957 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9590ca0059f0ebbe86343cccb3475338c317382c3231e082655b61e67eece375\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:12:05Z is after 2025-08-24T17:21:41Z" Oct 08 18:12:05 crc kubenswrapper[4988]: I1008 18:12:05.408885 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-h9jwp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4490d846-ca32-4cd5-9202-87cc6d499fb4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:12:04Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:12:04Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://00b7f6db043c28efda3a09fdc77a6ccded3eae8b6a3d2852aa500789d5f01194\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://00b7f6db043c28efda3a09fdc77a6ccded3eae8b6a3d2852aa500789d5f01194\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-08T18:12:04Z\\\",\\\"message\\\":\\\"2025-10-08T18:11:18+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_bee291bc-08e6-4092-b05c-4207d1f7fe40\\\\n2025-10-08T18:11:18+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_bee291bc-08e6-4092-b05c-4207d1f7fe40 to /host/opt/cni/bin/\\\\n2025-10-08T18:11:18Z [verbose] multus-daemon started\\\\n2025-10-08T18:11:18Z [verbose] Readiness Indicator file check\\\\n2025-10-08T18:12:03Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T18:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7k6rk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:11:16Z\\\"}}\" for pod \"openshift-multus\"/\"multus-h9jwp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:12:05Z is after 2025-08-24T17:21:41Z" Oct 08 18:12:05 crc kubenswrapper[4988]: I1008 18:12:05.427668 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:12:05 crc kubenswrapper[4988]: I1008 18:12:05.427801 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:12:05 crc kubenswrapper[4988]: I1008 18:12:05.427814 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:12:05 crc kubenswrapper[4988]: I1008 18:12:05.427847 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:12:05 crc kubenswrapper[4988]: I1008 18:12:05.427860 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:12:05Z","lastTransitionTime":"2025-10-08T18:12:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:12:05 crc kubenswrapper[4988]: I1008 18:12:05.438654 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-nsc65" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f010a463-69f2-4082-acc3-161899a939c3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://48a516f6c153743bd1a21ec6945fb368a37f4d2b9d6a326b68edc096a192ed05\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dnwgr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://03406e726f817bcedb841f81959ab12808ba35a530240f9abc5fa5c608b9a6ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dnwgr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9fbbeae0409d2e47e8de9a4306f63b0d13acbfd73cd86f83303fec14352ed718\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dnwgr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://55791327a370c0009c523b86b2234e82d3794985f2eaccf1922eb90f46b2b970\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dnwgr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cee5de93bc68345f9b737857f02832f56582d58757c50a8798eec25d19b2c20d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dnwgr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d314eb248e90371fed068b1173e26d4c9f9765417a70b341e1a5b136ec17353a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dnwgr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7bf11dd48efebb9da2b836bfad6a54b274247fa60902dab2f084793e4ad51a89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7bf11dd48efebb9da2b836bfad6a54b274247fa60902dab2f084793e4ad51a89\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-08T18:11:42Z\\\",\\\"message\\\":\\\" V4IPs:[]string(nil), V6IPs:[]string(nil)}, nodeEndpoints:map[string]services.lbEndpoints{}, externalTrafficLocal:false, internalTrafficLocal:false, hasNodePort:false}}\\\\nI1008 18:11:42.211985 6628 services_controller.go:444] Built service openshift-authentication-operator/metrics LB per-node configs for network=default: []services.lbConfig(nil)\\\\nI1008 18:11:42.211998 6628 services_controller.go:445] Built service openshift-authentication-operator/metrics LB template configs for network=default: []services.lbConfig(nil)\\\\nF1008 18:11:42.212009 6628 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:42Z is after 2025-08-\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T18:11:41Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-nsc65_openshift-ovn-kubernetes(f010a463-69f2-4082-acc3-161899a939c3)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dnwgr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://74fcb812b84291763cc6e5ccb20c12aaad580ad968dda4734a0bbfe33d2770ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dnwgr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://befe684cea9bd6dfda3ec8e85bca431c26bd8f82823b0a69dc0d04319a3ca385\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://befe684cea9bd6dfda3ec8e85bca431c26bd8f82823b0a69dc0d04319a3ca385\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dnwgr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:11:16Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-nsc65\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:12:05Z is after 2025-08-24T17:21:41Z" Oct 08 18:12:05 crc kubenswrapper[4988]: I1008 18:12:05.456328 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-7wl8v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e433488d-db52-435f-b89c-f63cd477c5d3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76b30d183684740b647c1af11dc49712232cc8b0c9a392b491580e5ddb75ea7f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dbz6l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e55016c3e1b97d1b3a857130a5c8e2ec6da2333f786032118a61e3e2ad4c5651\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dbz6l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:11:29Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-7wl8v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:12:05Z is after 2025-08-24T17:21:41Z" Oct 08 18:12:05 crc kubenswrapper[4988]: I1008 18:12:05.469219 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-jn2sx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"171e0902-95ee-4a39-8b1b-a58704c30574\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9f4q2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9f4q2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:11:31Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-jn2sx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:12:05Z is after 2025-08-24T17:21:41Z" Oct 08 18:12:05 crc kubenswrapper[4988]: I1008 18:12:05.485410 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d36f620-0f26-41c7-af6e-7fb5721a0e2d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:10:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:10:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:10:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3fd3b2ca80233400d2798b035a65b833f735632efd76d1160f6bb906e07b80a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:10:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://75a2bd2757133e79c0c7f2939eb433dc45291f14808e3c3f1daaf90a239c6a71\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:10:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5cb03de583c14ad6751227c21166ac5ac4c6e11b67d5b709361ced8a0877e6d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:10:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6eb46d76cf7cfb1d4ecc4fcee398dad8904732c1664c317865864b65f5703589\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6eb46d76cf7cfb1d4ecc4fcee398dad8904732c1664c317865864b65f5703589\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:10:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:10:56Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:10:55Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:12:05Z is after 2025-08-24T17:21:41Z" Oct 08 18:12:05 crc kubenswrapper[4988]: I1008 18:12:05.499636 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:12:05Z is after 2025-08-24T17:21:41Z" Oct 08 18:12:05 crc kubenswrapper[4988]: I1008 18:12:05.515768 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:12:05Z is after 2025-08-24T17:21:41Z" Oct 08 18:12:05 crc kubenswrapper[4988]: I1008 18:12:05.531690 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:12:05 crc kubenswrapper[4988]: I1008 18:12:05.531738 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:12:05 crc kubenswrapper[4988]: I1008 18:12:05.531752 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:12:05 crc kubenswrapper[4988]: I1008 18:12:05.531775 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:12:05 crc kubenswrapper[4988]: I1008 18:12:05.531788 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:12:05Z","lastTransitionTime":"2025-10-08T18:12:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:12:05 crc kubenswrapper[4988]: I1008 18:12:05.532067 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-s4f2k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ba5e11d7-7bde-4c3d-983b-884232219652\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://824bc34a2807068d3fceb10651803ae9ae5651e940f7aed4877c544d76dae4d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hn5jf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:11:15Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-s4f2k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:12:05Z is after 2025-08-24T17:21:41Z" Oct 08 18:12:05 crc kubenswrapper[4988]: I1008 18:12:05.635257 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:12:05 crc kubenswrapper[4988]: I1008 18:12:05.635318 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:12:05 crc kubenswrapper[4988]: I1008 18:12:05.635332 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:12:05 crc kubenswrapper[4988]: I1008 18:12:05.635358 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:12:05 crc kubenswrapper[4988]: I1008 18:12:05.635372 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:12:05Z","lastTransitionTime":"2025-10-08T18:12:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:12:05 crc kubenswrapper[4988]: I1008 18:12:05.738890 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:12:05 crc kubenswrapper[4988]: I1008 18:12:05.738930 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:12:05 crc kubenswrapper[4988]: I1008 18:12:05.738940 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:12:05 crc kubenswrapper[4988]: I1008 18:12:05.738956 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:12:05 crc kubenswrapper[4988]: I1008 18:12:05.738967 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:12:05Z","lastTransitionTime":"2025-10-08T18:12:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:12:05 crc kubenswrapper[4988]: I1008 18:12:05.798550 4988 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-h9jwp_4490d846-ca32-4cd5-9202-87cc6d499fb4/kube-multus/0.log" Oct 08 18:12:05 crc kubenswrapper[4988]: I1008 18:12:05.798683 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-h9jwp" event={"ID":"4490d846-ca32-4cd5-9202-87cc6d499fb4","Type":"ContainerStarted","Data":"c38c1b52127a970f792ef09ef1b217c979d94f7f74cc3c27e38949a204ad831f"} Oct 08 18:12:05 crc kubenswrapper[4988]: I1008 18:12:05.818746 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aef71be1-1d12-46e5-95b2-bd399040fd4d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:10:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:10:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:10:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77c68f3ecb8a0724499e495cb26ef3db5beecb140f2558d8c0b25f41928b84ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:10:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a560ee84c2405227de71fda127566213cd4f017e73d5581442ba267781b1c539\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:10:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6100accc3bd08928bc1ff88310f29fa5da37fd2d5c315c94279801c9778c4745\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:10:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b0e596d7ce4e9341c44499af79d3b3caf0e5e2c156902ea40cf2a3dca672be9e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5bd35aef24be034a0b305a40bb53ca171a4f76db3b93b4f78ae95099b0a4cde3\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-08T18:11:14Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1008 18:11:09.091034 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1008 18:11:09.091984 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3472374062/tls.crt::/tmp/serving-cert-3472374062/tls.key\\\\\\\"\\\\nI1008 18:11:14.619696 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1008 18:11:14.628837 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1008 18:11:14.628898 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1008 18:11:14.628994 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1008 18:11:14.629036 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1008 18:11:14.641318 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1008 18:11:14.641363 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1008 18:11:14.641372 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1008 18:11:14.641382 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1008 18:11:14.641415 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1008 18:11:14.641423 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1008 18:11:14.641431 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1008 18:11:14.641484 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1008 18:11:14.646451 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T18:10:58Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c79ef704fda1bbcc486f4c7afa3039a3ab38669833f0f66bafe24dd4cc679e60\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:10:58Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://28d21731af0f16d6a8ae6c6e9d0bebf0560d4a533db4c78c10bf1837890394ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://28d21731af0f16d6a8ae6c6e9d0bebf0560d4a533db4c78c10bf1837890394ef\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:10:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:10:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:10:55Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:12:05Z is after 2025-08-24T17:21:41Z" Oct 08 18:12:05 crc kubenswrapper[4988]: I1008 18:12:05.831525 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7e0ecefc-359f-4785-bb79-0343a62a25da\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:10:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:10:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:10:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://30bec350fefee3d19758b3ed735ede4e9aa085edb57aa3f3ec24249a67dc7ca2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:10:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2d850fd27d4ed65891ff69b9add65fb78535357f54819c2a8d196f5a340c58e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:10:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://90b1a9e14c82bf0f75f08fdb342eed35a503020f9db641e5e9ceff8b122cb3bd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:10:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://16f23c4c7163d32dc0df8d6a7ee17e2864fb261f8a40ecf05556774b173fc868\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:10:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:10:55Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:12:05Z is after 2025-08-24T17:21:41Z" Oct 08 18:12:05 crc kubenswrapper[4988]: I1008 18:12:05.843327 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:12:05 crc kubenswrapper[4988]: I1008 18:12:05.843371 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:12:05 crc kubenswrapper[4988]: I1008 18:12:05.843560 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:12:05 crc kubenswrapper[4988]: I1008 18:12:05.843601 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:12:05 crc kubenswrapper[4988]: I1008 18:12:05.843620 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:12:05Z","lastTransitionTime":"2025-10-08T18:12:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:12:05 crc kubenswrapper[4988]: I1008 18:12:05.844914 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e8906182d55c285b3cf095e2d9e1dac397af4618a514f11e26b44c172dd573f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:12:05Z is after 2025-08-24T17:21:41Z" Oct 08 18:12:05 crc kubenswrapper[4988]: I1008 18:12:05.854606 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-dzxs2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8f307b20-5e33-4725-99cb-210e25113532\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72ee803c6ef88524e0ce0391476f8e5a7dfbc858b095a3c2ae31bad10991ce3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q65f7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:11:16Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-dzxs2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:12:05Z is after 2025-08-24T17:21:41Z" Oct 08 18:12:05 crc kubenswrapper[4988]: I1008 18:12:05.869091 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-jnh7m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"556fb048-8afb-44a5-9ea0-2b7c89e6cf4b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bfa2f3382b23c3f788a1b31e80e8a90b3b6868870dccc46d86c3fc11ad027f52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmlt2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1a825886f105ccf9a7cc6f28a0cd012fb859f4990dade0e6aae932482146028\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1a825886f105ccf9a7cc6f28a0cd012fb859f4990dade0e6aae932482146028\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmlt2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ab2fe6e65418efa75bae72259157a2323bbb4b7d11d1d012053524e14135888\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0ab2fe6e65418efa75bae72259157a2323bbb4b7d11d1d012053524e14135888\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:11:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:11:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmlt2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://368a303309be112d9c0b7ccd6538fbaca7085d97a8ec8259f36092f07147e7cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://368a303309be112d9c0b7ccd6538fbaca7085d97a8ec8259f36092f07147e7cb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:11:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:11:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmlt2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a8ca69001dddef643f982a97c74c291e30cb9fac5b0e894e80066b885ea9151\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9a8ca69001dddef643f982a97c74c291e30cb9fac5b0e894e80066b885ea9151\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:11:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:11:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmlt2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://145303c46b48603ed1aae4b9bfe06332a3ffad28f206f8394574011a87b58cd3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://145303c46b48603ed1aae4b9bfe06332a3ffad28f206f8394574011a87b58cd3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:11:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:11:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmlt2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d08e5cfc7cbe5ec1a7a0265ed7c466363c3abeec492263170860689b46c7ff95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d08e5cfc7cbe5ec1a7a0265ed7c466363c3abeec492263170860689b46c7ff95\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:11:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:11:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmlt2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:11:16Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-jnh7m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:12:05Z is after 2025-08-24T17:21:41Z" Oct 08 18:12:05 crc kubenswrapper[4988]: I1008 18:12:05.884317 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f328a20b-5283-407e-8a93-2ab2da599f18\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8622d56fb198dadee31fb0ca8fc80d56828b6710648055f264533e2d5bdb3037\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z2gh4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3cdb2a757332bdcdc2a2f3362913fe8f06003b0597de20b72627f9badd5e30af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z2gh4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:11:16Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-nm54f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:12:05Z is after 2025-08-24T17:21:41Z" Oct 08 18:12:05 crc kubenswrapper[4988]: I1008 18:12:05.900443 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a35a5e989ba69c8ef36ce4cea272aadf67ce20473fc257adfcfa86a2faba6ae0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a522e937cfe03fb9f2e60cb1bea07214cf63c69bdecd5ba65a2dfdba882d28dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:12:05Z is after 2025-08-24T17:21:41Z" Oct 08 18:12:05 crc kubenswrapper[4988]: I1008 18:12:05.913929 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9590ca0059f0ebbe86343cccb3475338c317382c3231e082655b61e67eece375\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:12:05Z is after 2025-08-24T17:21:41Z" Oct 08 18:12:05 crc kubenswrapper[4988]: I1008 18:12:05.931885 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:12:05Z is after 2025-08-24T17:21:41Z" Oct 08 18:12:05 crc kubenswrapper[4988]: I1008 18:12:05.946368 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:12:05 crc kubenswrapper[4988]: I1008 18:12:05.946436 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:12:05 crc kubenswrapper[4988]: I1008 18:12:05.946447 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:12:05 crc kubenswrapper[4988]: I1008 18:12:05.946465 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:12:05 crc kubenswrapper[4988]: I1008 18:12:05.946479 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:12:05Z","lastTransitionTime":"2025-10-08T18:12:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:12:05 crc kubenswrapper[4988]: I1008 18:12:05.964645 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-nsc65" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f010a463-69f2-4082-acc3-161899a939c3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://48a516f6c153743bd1a21ec6945fb368a37f4d2b9d6a326b68edc096a192ed05\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dnwgr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://03406e726f817bcedb841f81959ab12808ba35a530240f9abc5fa5c608b9a6ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dnwgr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9fbbeae0409d2e47e8de9a4306f63b0d13acbfd73cd86f83303fec14352ed718\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dnwgr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://55791327a370c0009c523b86b2234e82d3794985f2eaccf1922eb90f46b2b970\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dnwgr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cee5de93bc68345f9b737857f02832f56582d58757c50a8798eec25d19b2c20d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dnwgr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d314eb248e90371fed068b1173e26d4c9f9765417a70b341e1a5b136ec17353a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dnwgr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7bf11dd48efebb9da2b836bfad6a54b274247fa60902dab2f084793e4ad51a89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7bf11dd48efebb9da2b836bfad6a54b274247fa60902dab2f084793e4ad51a89\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-08T18:11:42Z\\\",\\\"message\\\":\\\" V4IPs:[]string(nil), V6IPs:[]string(nil)}, nodeEndpoints:map[string]services.lbEndpoints{}, externalTrafficLocal:false, internalTrafficLocal:false, hasNodePort:false}}\\\\nI1008 18:11:42.211985 6628 services_controller.go:444] Built service openshift-authentication-operator/metrics LB per-node configs for network=default: []services.lbConfig(nil)\\\\nI1008 18:11:42.211998 6628 services_controller.go:445] Built service openshift-authentication-operator/metrics LB template configs for network=default: []services.lbConfig(nil)\\\\nF1008 18:11:42.212009 6628 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:42Z is after 2025-08-\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T18:11:41Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-nsc65_openshift-ovn-kubernetes(f010a463-69f2-4082-acc3-161899a939c3)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dnwgr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://74fcb812b84291763cc6e5ccb20c12aaad580ad968dda4734a0bbfe33d2770ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dnwgr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://befe684cea9bd6dfda3ec8e85bca431c26bd8f82823b0a69dc0d04319a3ca385\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://befe684cea9bd6dfda3ec8e85bca431c26bd8f82823b0a69dc0d04319a3ca385\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dnwgr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:11:16Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-nsc65\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:12:05Z is after 2025-08-24T17:21:41Z" Oct 08 18:12:05 crc kubenswrapper[4988]: I1008 18:12:05.978831 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-7wl8v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e433488d-db52-435f-b89c-f63cd477c5d3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76b30d183684740b647c1af11dc49712232cc8b0c9a392b491580e5ddb75ea7f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dbz6l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e55016c3e1b97d1b3a857130a5c8e2ec6da2333f786032118a61e3e2ad4c5651\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dbz6l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:11:29Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-7wl8v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:12:05Z is after 2025-08-24T17:21:41Z" Oct 08 18:12:05 crc kubenswrapper[4988]: I1008 18:12:05.990367 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-jn2sx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"171e0902-95ee-4a39-8b1b-a58704c30574\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9f4q2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9f4q2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:11:31Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-jn2sx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:12:05Z is after 2025-08-24T17:21:41Z" Oct 08 18:12:06 crc kubenswrapper[4988]: I1008 18:12:06.003332 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d36f620-0f26-41c7-af6e-7fb5721a0e2d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:10:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:10:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:10:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3fd3b2ca80233400d2798b035a65b833f735632efd76d1160f6bb906e07b80a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:10:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://75a2bd2757133e79c0c7f2939eb433dc45291f14808e3c3f1daaf90a239c6a71\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:10:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5cb03de583c14ad6751227c21166ac5ac4c6e11b67d5b709361ced8a0877e6d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:10:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6eb46d76cf7cfb1d4ecc4fcee398dad8904732c1664c317865864b65f5703589\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6eb46d76cf7cfb1d4ecc4fcee398dad8904732c1664c317865864b65f5703589\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:10:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:10:56Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:10:55Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:12:06Z is after 2025-08-24T17:21:41Z" Oct 08 18:12:06 crc kubenswrapper[4988]: I1008 18:12:06.014372 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:12:06Z is after 2025-08-24T17:21:41Z" Oct 08 18:12:06 crc kubenswrapper[4988]: I1008 18:12:06.025766 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:12:06Z is after 2025-08-24T17:21:41Z" Oct 08 18:12:06 crc kubenswrapper[4988]: I1008 18:12:06.038604 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-h9jwp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4490d846-ca32-4cd5-9202-87cc6d499fb4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:12:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:12:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c38c1b52127a970f792ef09ef1b217c979d94f7f74cc3c27e38949a204ad831f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://00b7f6db043c28efda3a09fdc77a6ccded3eae8b6a3d2852aa500789d5f01194\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-08T18:12:04Z\\\",\\\"message\\\":\\\"2025-10-08T18:11:18+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_bee291bc-08e6-4092-b05c-4207d1f7fe40\\\\n2025-10-08T18:11:18+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_bee291bc-08e6-4092-b05c-4207d1f7fe40 to /host/opt/cni/bin/\\\\n2025-10-08T18:11:18Z [verbose] multus-daemon started\\\\n2025-10-08T18:11:18Z [verbose] Readiness Indicator file check\\\\n2025-10-08T18:12:03Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T18:11:17Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:12:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7k6rk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:11:16Z\\\"}}\" for pod \"openshift-multus\"/\"multus-h9jwp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:12:06Z is after 2025-08-24T17:21:41Z" Oct 08 18:12:06 crc kubenswrapper[4988]: I1008 18:12:06.055560 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:12:06 crc kubenswrapper[4988]: I1008 18:12:06.055602 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:12:06 crc kubenswrapper[4988]: I1008 18:12:06.055616 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:12:06 crc kubenswrapper[4988]: I1008 18:12:06.055636 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:12:06 crc kubenswrapper[4988]: I1008 18:12:06.055651 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:12:06Z","lastTransitionTime":"2025-10-08T18:12:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:12:06 crc kubenswrapper[4988]: I1008 18:12:06.056607 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"63edba53-2181-4227-9a72-a21793ad7cd0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:10:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:10:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:10:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://adbdfab7dbc007ce91c26fa7a6535ef27949ad2567f1bc6b4c52b6a3f62789f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:10:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7b7185463f2a962f3214a2524d783a54b8fbc76f385ea27001087bfb7d023561\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7b7185463f2a962f3214a2524d783a54b8fbc76f385ea27001087bfb7d023561\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:10:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:10:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:10:55Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:12:06Z is after 2025-08-24T17:21:41Z" Oct 08 18:12:06 crc kubenswrapper[4988]: I1008 18:12:06.067811 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-s4f2k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ba5e11d7-7bde-4c3d-983b-884232219652\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://824bc34a2807068d3fceb10651803ae9ae5651e940f7aed4877c544d76dae4d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hn5jf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:11:15Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-s4f2k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:12:06Z is after 2025-08-24T17:21:41Z" Oct 08 18:12:06 crc kubenswrapper[4988]: I1008 18:12:06.158651 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:12:06 crc kubenswrapper[4988]: I1008 18:12:06.158707 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:12:06 crc kubenswrapper[4988]: I1008 18:12:06.158722 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:12:06 crc kubenswrapper[4988]: I1008 18:12:06.158745 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:12:06 crc kubenswrapper[4988]: I1008 18:12:06.158758 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:12:06Z","lastTransitionTime":"2025-10-08T18:12:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:12:06 crc kubenswrapper[4988]: I1008 18:12:06.261664 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:12:06 crc kubenswrapper[4988]: I1008 18:12:06.261728 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:12:06 crc kubenswrapper[4988]: I1008 18:12:06.261742 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:12:06 crc kubenswrapper[4988]: I1008 18:12:06.261766 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:12:06 crc kubenswrapper[4988]: I1008 18:12:06.261779 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:12:06Z","lastTransitionTime":"2025-10-08T18:12:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:12:06 crc kubenswrapper[4988]: I1008 18:12:06.365022 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:12:06 crc kubenswrapper[4988]: I1008 18:12:06.365105 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:12:06 crc kubenswrapper[4988]: I1008 18:12:06.365131 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:12:06 crc kubenswrapper[4988]: I1008 18:12:06.365163 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:12:06 crc kubenswrapper[4988]: I1008 18:12:06.365187 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:12:06Z","lastTransitionTime":"2025-10-08T18:12:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:12:06 crc kubenswrapper[4988]: I1008 18:12:06.468692 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:12:06 crc kubenswrapper[4988]: I1008 18:12:06.468764 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:12:06 crc kubenswrapper[4988]: I1008 18:12:06.468779 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:12:06 crc kubenswrapper[4988]: I1008 18:12:06.468801 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:12:06 crc kubenswrapper[4988]: I1008 18:12:06.468816 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:12:06Z","lastTransitionTime":"2025-10-08T18:12:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:12:06 crc kubenswrapper[4988]: I1008 18:12:06.571150 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:12:06 crc kubenswrapper[4988]: I1008 18:12:06.571200 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:12:06 crc kubenswrapper[4988]: I1008 18:12:06.571212 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:12:06 crc kubenswrapper[4988]: I1008 18:12:06.571238 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:12:06 crc kubenswrapper[4988]: I1008 18:12:06.571252 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:12:06Z","lastTransitionTime":"2025-10-08T18:12:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:12:06 crc kubenswrapper[4988]: I1008 18:12:06.674204 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:12:06 crc kubenswrapper[4988]: I1008 18:12:06.674278 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:12:06 crc kubenswrapper[4988]: I1008 18:12:06.674301 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:12:06 crc kubenswrapper[4988]: I1008 18:12:06.674330 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:12:06 crc kubenswrapper[4988]: I1008 18:12:06.674351 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:12:06Z","lastTransitionTime":"2025-10-08T18:12:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:12:06 crc kubenswrapper[4988]: I1008 18:12:06.776926 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:12:06 crc kubenswrapper[4988]: I1008 18:12:06.776991 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:12:06 crc kubenswrapper[4988]: I1008 18:12:06.777004 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:12:06 crc kubenswrapper[4988]: I1008 18:12:06.777029 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:12:06 crc kubenswrapper[4988]: I1008 18:12:06.777047 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:12:06Z","lastTransitionTime":"2025-10-08T18:12:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:12:06 crc kubenswrapper[4988]: I1008 18:12:06.879110 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:12:06 crc kubenswrapper[4988]: I1008 18:12:06.879157 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:12:06 crc kubenswrapper[4988]: I1008 18:12:06.879169 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:12:06 crc kubenswrapper[4988]: I1008 18:12:06.879189 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:12:06 crc kubenswrapper[4988]: I1008 18:12:06.879203 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:12:06Z","lastTransitionTime":"2025-10-08T18:12:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:12:06 crc kubenswrapper[4988]: I1008 18:12:06.981947 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:12:06 crc kubenswrapper[4988]: I1008 18:12:06.981999 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:12:06 crc kubenswrapper[4988]: I1008 18:12:06.982015 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:12:06 crc kubenswrapper[4988]: I1008 18:12:06.982039 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:12:06 crc kubenswrapper[4988]: I1008 18:12:06.982056 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:12:06Z","lastTransitionTime":"2025-10-08T18:12:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:12:07 crc kubenswrapper[4988]: I1008 18:12:07.084677 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:12:07 crc kubenswrapper[4988]: I1008 18:12:07.084733 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:12:07 crc kubenswrapper[4988]: I1008 18:12:07.084746 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:12:07 crc kubenswrapper[4988]: I1008 18:12:07.084768 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:12:07 crc kubenswrapper[4988]: I1008 18:12:07.084779 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:12:07Z","lastTransitionTime":"2025-10-08T18:12:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:12:07 crc kubenswrapper[4988]: I1008 18:12:07.187884 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:12:07 crc kubenswrapper[4988]: I1008 18:12:07.187956 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:12:07 crc kubenswrapper[4988]: I1008 18:12:07.187975 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:12:07 crc kubenswrapper[4988]: I1008 18:12:07.188001 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:12:07 crc kubenswrapper[4988]: I1008 18:12:07.188020 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:12:07Z","lastTransitionTime":"2025-10-08T18:12:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:12:07 crc kubenswrapper[4988]: I1008 18:12:07.236790 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 18:12:07 crc kubenswrapper[4988]: E1008 18:12:07.236978 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 08 18:12:07 crc kubenswrapper[4988]: I1008 18:12:07.237255 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 18:12:07 crc kubenswrapper[4988]: E1008 18:12:07.237336 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 08 18:12:07 crc kubenswrapper[4988]: I1008 18:12:07.237561 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jn2sx" Oct 08 18:12:07 crc kubenswrapper[4988]: E1008 18:12:07.237647 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jn2sx" podUID="171e0902-95ee-4a39-8b1b-a58704c30574" Oct 08 18:12:07 crc kubenswrapper[4988]: I1008 18:12:07.237894 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 18:12:07 crc kubenswrapper[4988]: E1008 18:12:07.237968 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 08 18:12:07 crc kubenswrapper[4988]: I1008 18:12:07.291448 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:12:07 crc kubenswrapper[4988]: I1008 18:12:07.291490 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:12:07 crc kubenswrapper[4988]: I1008 18:12:07.291500 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:12:07 crc kubenswrapper[4988]: I1008 18:12:07.291519 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:12:07 crc kubenswrapper[4988]: I1008 18:12:07.291529 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:12:07Z","lastTransitionTime":"2025-10-08T18:12:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:12:07 crc kubenswrapper[4988]: I1008 18:12:07.394206 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:12:07 crc kubenswrapper[4988]: I1008 18:12:07.394249 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:12:07 crc kubenswrapper[4988]: I1008 18:12:07.394259 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:12:07 crc kubenswrapper[4988]: I1008 18:12:07.394276 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:12:07 crc kubenswrapper[4988]: I1008 18:12:07.394288 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:12:07Z","lastTransitionTime":"2025-10-08T18:12:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:12:07 crc kubenswrapper[4988]: I1008 18:12:07.502040 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:12:07 crc kubenswrapper[4988]: I1008 18:12:07.502102 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:12:07 crc kubenswrapper[4988]: I1008 18:12:07.502121 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:12:07 crc kubenswrapper[4988]: I1008 18:12:07.502146 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:12:07 crc kubenswrapper[4988]: I1008 18:12:07.502166 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:12:07Z","lastTransitionTime":"2025-10-08T18:12:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:12:07 crc kubenswrapper[4988]: I1008 18:12:07.605561 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:12:07 crc kubenswrapper[4988]: I1008 18:12:07.605606 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:12:07 crc kubenswrapper[4988]: I1008 18:12:07.605617 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:12:07 crc kubenswrapper[4988]: I1008 18:12:07.605637 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:12:07 crc kubenswrapper[4988]: I1008 18:12:07.605655 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:12:07Z","lastTransitionTime":"2025-10-08T18:12:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:12:07 crc kubenswrapper[4988]: I1008 18:12:07.708484 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:12:07 crc kubenswrapper[4988]: I1008 18:12:07.708526 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:12:07 crc kubenswrapper[4988]: I1008 18:12:07.708535 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:12:07 crc kubenswrapper[4988]: I1008 18:12:07.708551 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:12:07 crc kubenswrapper[4988]: I1008 18:12:07.708562 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:12:07Z","lastTransitionTime":"2025-10-08T18:12:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:12:07 crc kubenswrapper[4988]: I1008 18:12:07.810463 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:12:07 crc kubenswrapper[4988]: I1008 18:12:07.810861 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:12:07 crc kubenswrapper[4988]: I1008 18:12:07.810998 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:12:07 crc kubenswrapper[4988]: I1008 18:12:07.811153 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:12:07 crc kubenswrapper[4988]: I1008 18:12:07.811309 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:12:07Z","lastTransitionTime":"2025-10-08T18:12:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:12:07 crc kubenswrapper[4988]: I1008 18:12:07.914608 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:12:07 crc kubenswrapper[4988]: I1008 18:12:07.914658 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:12:07 crc kubenswrapper[4988]: I1008 18:12:07.914670 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:12:07 crc kubenswrapper[4988]: I1008 18:12:07.914689 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:12:07 crc kubenswrapper[4988]: I1008 18:12:07.914701 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:12:07Z","lastTransitionTime":"2025-10-08T18:12:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:12:08 crc kubenswrapper[4988]: I1008 18:12:08.018592 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:12:08 crc kubenswrapper[4988]: I1008 18:12:08.018648 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:12:08 crc kubenswrapper[4988]: I1008 18:12:08.018657 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:12:08 crc kubenswrapper[4988]: I1008 18:12:08.018674 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:12:08 crc kubenswrapper[4988]: I1008 18:12:08.018693 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:12:08Z","lastTransitionTime":"2025-10-08T18:12:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:12:08 crc kubenswrapper[4988]: I1008 18:12:08.122677 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:12:08 crc kubenswrapper[4988]: I1008 18:12:08.122761 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:12:08 crc kubenswrapper[4988]: I1008 18:12:08.122784 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:12:08 crc kubenswrapper[4988]: I1008 18:12:08.122815 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:12:08 crc kubenswrapper[4988]: I1008 18:12:08.122836 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:12:08Z","lastTransitionTime":"2025-10-08T18:12:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:12:08 crc kubenswrapper[4988]: I1008 18:12:08.225212 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:12:08 crc kubenswrapper[4988]: I1008 18:12:08.225268 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:12:08 crc kubenswrapper[4988]: I1008 18:12:08.225280 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:12:08 crc kubenswrapper[4988]: I1008 18:12:08.225300 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:12:08 crc kubenswrapper[4988]: I1008 18:12:08.225316 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:12:08Z","lastTransitionTime":"2025-10-08T18:12:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:12:08 crc kubenswrapper[4988]: I1008 18:12:08.238494 4988 scope.go:117] "RemoveContainer" containerID="7bf11dd48efebb9da2b836bfad6a54b274247fa60902dab2f084793e4ad51a89" Oct 08 18:12:08 crc kubenswrapper[4988]: I1008 18:12:08.328571 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:12:08 crc kubenswrapper[4988]: I1008 18:12:08.328855 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:12:08 crc kubenswrapper[4988]: I1008 18:12:08.329068 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:12:08 crc kubenswrapper[4988]: I1008 18:12:08.329214 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:12:08 crc kubenswrapper[4988]: I1008 18:12:08.329331 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:12:08Z","lastTransitionTime":"2025-10-08T18:12:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:12:08 crc kubenswrapper[4988]: I1008 18:12:08.432090 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:12:08 crc kubenswrapper[4988]: I1008 18:12:08.432165 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:12:08 crc kubenswrapper[4988]: I1008 18:12:08.432176 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:12:08 crc kubenswrapper[4988]: I1008 18:12:08.432191 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:12:08 crc kubenswrapper[4988]: I1008 18:12:08.432201 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:12:08Z","lastTransitionTime":"2025-10-08T18:12:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:12:08 crc kubenswrapper[4988]: I1008 18:12:08.535342 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:12:08 crc kubenswrapper[4988]: I1008 18:12:08.535410 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:12:08 crc kubenswrapper[4988]: I1008 18:12:08.535423 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:12:08 crc kubenswrapper[4988]: I1008 18:12:08.535442 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:12:08 crc kubenswrapper[4988]: I1008 18:12:08.535454 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:12:08Z","lastTransitionTime":"2025-10-08T18:12:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:12:08 crc kubenswrapper[4988]: I1008 18:12:08.638446 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:12:08 crc kubenswrapper[4988]: I1008 18:12:08.638531 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:12:08 crc kubenswrapper[4988]: I1008 18:12:08.638601 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:12:08 crc kubenswrapper[4988]: I1008 18:12:08.638694 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:12:08 crc kubenswrapper[4988]: I1008 18:12:08.638781 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:12:08Z","lastTransitionTime":"2025-10-08T18:12:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:12:08 crc kubenswrapper[4988]: I1008 18:12:08.741436 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:12:08 crc kubenswrapper[4988]: I1008 18:12:08.741492 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:12:08 crc kubenswrapper[4988]: I1008 18:12:08.741504 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:12:08 crc kubenswrapper[4988]: I1008 18:12:08.741525 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:12:08 crc kubenswrapper[4988]: I1008 18:12:08.741538 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:12:08Z","lastTransitionTime":"2025-10-08T18:12:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:12:08 crc kubenswrapper[4988]: I1008 18:12:08.810948 4988 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-nsc65_f010a463-69f2-4082-acc3-161899a939c3/ovnkube-controller/2.log" Oct 08 18:12:08 crc kubenswrapper[4988]: I1008 18:12:08.813468 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-nsc65" event={"ID":"f010a463-69f2-4082-acc3-161899a939c3","Type":"ContainerStarted","Data":"763353df23d26dc093a685af435811240b5ee772d7279ee9a6d2c7a48c19cd30"} Oct 08 18:12:08 crc kubenswrapper[4988]: I1008 18:12:08.814580 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-nsc65" Oct 08 18:12:08 crc kubenswrapper[4988]: I1008 18:12:08.833061 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-jnh7m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"556fb048-8afb-44a5-9ea0-2b7c89e6cf4b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bfa2f3382b23c3f788a1b31e80e8a90b3b6868870dccc46d86c3fc11ad027f52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmlt2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1a825886f105ccf9a7cc6f28a0cd012fb859f4990dade0e6aae932482146028\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1a825886f105ccf9a7cc6f28a0cd012fb859f4990dade0e6aae932482146028\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmlt2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ab2fe6e65418efa75bae72259157a2323bbb4b7d11d1d012053524e14135888\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0ab2fe6e65418efa75bae72259157a2323bbb4b7d11d1d012053524e14135888\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:11:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:11:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmlt2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://368a303309be112d9c0b7ccd6538fbaca7085d97a8ec8259f36092f07147e7cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://368a303309be112d9c0b7ccd6538fbaca7085d97a8ec8259f36092f07147e7cb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:11:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:11:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmlt2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a8ca69001dddef643f982a97c74c291e30cb9fac5b0e894e80066b885ea9151\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9a8ca69001dddef643f982a97c74c291e30cb9fac5b0e894e80066b885ea9151\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:11:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:11:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmlt2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://145303c46b48603ed1aae4b9bfe06332a3ffad28f206f8394574011a87b58cd3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://145303c46b48603ed1aae4b9bfe06332a3ffad28f206f8394574011a87b58cd3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:11:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:11:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmlt2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d08e5cfc7cbe5ec1a7a0265ed7c466363c3abeec492263170860689b46c7ff95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d08e5cfc7cbe5ec1a7a0265ed7c466363c3abeec492263170860689b46c7ff95\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:11:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:11:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmlt2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:11:16Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-jnh7m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:12:08Z is after 2025-08-24T17:21:41Z" Oct 08 18:12:08 crc kubenswrapper[4988]: I1008 18:12:08.843959 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:12:08 crc kubenswrapper[4988]: I1008 18:12:08.843999 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:12:08 crc kubenswrapper[4988]: I1008 18:12:08.844008 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:12:08 crc kubenswrapper[4988]: I1008 18:12:08.844024 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:12:08 crc kubenswrapper[4988]: I1008 18:12:08.844036 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:12:08Z","lastTransitionTime":"2025-10-08T18:12:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:12:08 crc kubenswrapper[4988]: I1008 18:12:08.846318 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f328a20b-5283-407e-8a93-2ab2da599f18\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8622d56fb198dadee31fb0ca8fc80d56828b6710648055f264533e2d5bdb3037\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z2gh4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3cdb2a757332bdcdc2a2f3362913fe8f06003b0597de20b72627f9badd5e30af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z2gh4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:11:16Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-nm54f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:12:08Z is after 2025-08-24T17:21:41Z" Oct 08 18:12:08 crc kubenswrapper[4988]: I1008 18:12:08.860156 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a35a5e989ba69c8ef36ce4cea272aadf67ce20473fc257adfcfa86a2faba6ae0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a522e937cfe03fb9f2e60cb1bea07214cf63c69bdecd5ba65a2dfdba882d28dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:12:08Z is after 2025-08-24T17:21:41Z" Oct 08 18:12:08 crc kubenswrapper[4988]: I1008 18:12:08.874070 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9590ca0059f0ebbe86343cccb3475338c317382c3231e082655b61e67eece375\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:12:08Z is after 2025-08-24T17:21:41Z" Oct 08 18:12:08 crc kubenswrapper[4988]: I1008 18:12:08.894112 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:12:08Z is after 2025-08-24T17:21:41Z" Oct 08 18:12:08 crc kubenswrapper[4988]: I1008 18:12:08.917601 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-nsc65" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f010a463-69f2-4082-acc3-161899a939c3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://48a516f6c153743bd1a21ec6945fb368a37f4d2b9d6a326b68edc096a192ed05\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dnwgr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://03406e726f817bcedb841f81959ab12808ba35a530240f9abc5fa5c608b9a6ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dnwgr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9fbbeae0409d2e47e8de9a4306f63b0d13acbfd73cd86f83303fec14352ed718\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dnwgr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://55791327a370c0009c523b86b2234e82d3794985f2eaccf1922eb90f46b2b970\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dnwgr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cee5de93bc68345f9b737857f02832f56582d58757c50a8798eec25d19b2c20d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dnwgr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d314eb248e90371fed068b1173e26d4c9f9765417a70b341e1a5b136ec17353a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dnwgr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://763353df23d26dc093a685af435811240b5ee772d7279ee9a6d2c7a48c19cd30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7bf11dd48efebb9da2b836bfad6a54b274247fa60902dab2f084793e4ad51a89\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-08T18:11:42Z\\\",\\\"message\\\":\\\" V4IPs:[]string(nil), V6IPs:[]string(nil)}, nodeEndpoints:map[string]services.lbEndpoints{}, externalTrafficLocal:false, internalTrafficLocal:false, hasNodePort:false}}\\\\nI1008 18:11:42.211985 6628 services_controller.go:444] Built service openshift-authentication-operator/metrics LB per-node configs for network=default: []services.lbConfig(nil)\\\\nI1008 18:11:42.211998 6628 services_controller.go:445] Built service openshift-authentication-operator/metrics LB template configs for network=default: []services.lbConfig(nil)\\\\nF1008 18:11:42.212009 6628 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:42Z is after 2025-08-\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T18:11:41Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:12:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dnwgr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://74fcb812b84291763cc6e5ccb20c12aaad580ad968dda4734a0bbfe33d2770ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dnwgr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://befe684cea9bd6dfda3ec8e85bca431c26bd8f82823b0a69dc0d04319a3ca385\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://befe684cea9bd6dfda3ec8e85bca431c26bd8f82823b0a69dc0d04319a3ca385\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dnwgr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:11:16Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-nsc65\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:12:08Z is after 2025-08-24T17:21:41Z" Oct 08 18:12:08 crc kubenswrapper[4988]: I1008 18:12:08.930379 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-7wl8v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e433488d-db52-435f-b89c-f63cd477c5d3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76b30d183684740b647c1af11dc49712232cc8b0c9a392b491580e5ddb75ea7f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dbz6l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e55016c3e1b97d1b3a857130a5c8e2ec6da2333f786032118a61e3e2ad4c5651\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dbz6l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:11:29Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-7wl8v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:12:08Z is after 2025-08-24T17:21:41Z" Oct 08 18:12:08 crc kubenswrapper[4988]: I1008 18:12:08.941668 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-jn2sx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"171e0902-95ee-4a39-8b1b-a58704c30574\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9f4q2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9f4q2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:11:31Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-jn2sx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:12:08Z is after 2025-08-24T17:21:41Z" Oct 08 18:12:08 crc kubenswrapper[4988]: I1008 18:12:08.946579 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:12:08 crc kubenswrapper[4988]: I1008 18:12:08.946626 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:12:08 crc kubenswrapper[4988]: I1008 18:12:08.946641 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:12:08 crc kubenswrapper[4988]: I1008 18:12:08.946665 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:12:08 crc kubenswrapper[4988]: I1008 18:12:08.946680 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:12:08Z","lastTransitionTime":"2025-10-08T18:12:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:12:08 crc kubenswrapper[4988]: I1008 18:12:08.952077 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d36f620-0f26-41c7-af6e-7fb5721a0e2d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:10:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:10:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:10:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3fd3b2ca80233400d2798b035a65b833f735632efd76d1160f6bb906e07b80a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:10:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://75a2bd2757133e79c0c7f2939eb433dc45291f14808e3c3f1daaf90a239c6a71\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:10:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5cb03de583c14ad6751227c21166ac5ac4c6e11b67d5b709361ced8a0877e6d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:10:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6eb46d76cf7cfb1d4ecc4fcee398dad8904732c1664c317865864b65f5703589\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6eb46d76cf7cfb1d4ecc4fcee398dad8904732c1664c317865864b65f5703589\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:10:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:10:56Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:10:55Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:12:08Z is after 2025-08-24T17:21:41Z" Oct 08 18:12:08 crc kubenswrapper[4988]: I1008 18:12:08.966574 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:12:08Z is after 2025-08-24T17:21:41Z" Oct 08 18:12:08 crc kubenswrapper[4988]: I1008 18:12:08.988913 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:12:08Z is after 2025-08-24T17:21:41Z" Oct 08 18:12:09 crc kubenswrapper[4988]: I1008 18:12:09.005542 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-h9jwp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4490d846-ca32-4cd5-9202-87cc6d499fb4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:12:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:12:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c38c1b52127a970f792ef09ef1b217c979d94f7f74cc3c27e38949a204ad831f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://00b7f6db043c28efda3a09fdc77a6ccded3eae8b6a3d2852aa500789d5f01194\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-08T18:12:04Z\\\",\\\"message\\\":\\\"2025-10-08T18:11:18+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_bee291bc-08e6-4092-b05c-4207d1f7fe40\\\\n2025-10-08T18:11:18+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_bee291bc-08e6-4092-b05c-4207d1f7fe40 to /host/opt/cni/bin/\\\\n2025-10-08T18:11:18Z [verbose] multus-daemon started\\\\n2025-10-08T18:11:18Z [verbose] Readiness Indicator file check\\\\n2025-10-08T18:12:03Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T18:11:17Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:12:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7k6rk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:11:16Z\\\"}}\" for pod \"openshift-multus\"/\"multus-h9jwp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:12:09Z is after 2025-08-24T17:21:41Z" Oct 08 18:12:09 crc kubenswrapper[4988]: I1008 18:12:09.021921 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"63edba53-2181-4227-9a72-a21793ad7cd0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:10:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:10:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:10:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://adbdfab7dbc007ce91c26fa7a6535ef27949ad2567f1bc6b4c52b6a3f62789f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:10:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7b7185463f2a962f3214a2524d783a54b8fbc76f385ea27001087bfb7d023561\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7b7185463f2a962f3214a2524d783a54b8fbc76f385ea27001087bfb7d023561\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:10:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:10:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:10:55Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:12:09Z is after 2025-08-24T17:21:41Z" Oct 08 18:12:09 crc kubenswrapper[4988]: I1008 18:12:09.036375 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-s4f2k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ba5e11d7-7bde-4c3d-983b-884232219652\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://824bc34a2807068d3fceb10651803ae9ae5651e940f7aed4877c544d76dae4d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hn5jf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:11:15Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-s4f2k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:12:09Z is after 2025-08-24T17:21:41Z" Oct 08 18:12:09 crc kubenswrapper[4988]: I1008 18:12:09.049896 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:12:09 crc kubenswrapper[4988]: I1008 18:12:09.049946 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:12:09 crc kubenswrapper[4988]: I1008 18:12:09.049964 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:12:09 crc kubenswrapper[4988]: I1008 18:12:09.049991 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:12:09 crc kubenswrapper[4988]: I1008 18:12:09.050009 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:12:09Z","lastTransitionTime":"2025-10-08T18:12:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:12:09 crc kubenswrapper[4988]: I1008 18:12:09.052811 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aef71be1-1d12-46e5-95b2-bd399040fd4d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:10:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:10:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:10:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77c68f3ecb8a0724499e495cb26ef3db5beecb140f2558d8c0b25f41928b84ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:10:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a560ee84c2405227de71fda127566213cd4f017e73d5581442ba267781b1c539\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:10:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6100accc3bd08928bc1ff88310f29fa5da37fd2d5c315c94279801c9778c4745\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:10:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b0e596d7ce4e9341c44499af79d3b3caf0e5e2c156902ea40cf2a3dca672be9e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5bd35aef24be034a0b305a40bb53ca171a4f76db3b93b4f78ae95099b0a4cde3\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-08T18:11:14Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1008 18:11:09.091034 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1008 18:11:09.091984 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3472374062/tls.crt::/tmp/serving-cert-3472374062/tls.key\\\\\\\"\\\\nI1008 18:11:14.619696 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1008 18:11:14.628837 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1008 18:11:14.628898 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1008 18:11:14.628994 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1008 18:11:14.629036 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1008 18:11:14.641318 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1008 18:11:14.641363 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1008 18:11:14.641372 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1008 18:11:14.641382 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1008 18:11:14.641415 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1008 18:11:14.641423 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1008 18:11:14.641431 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1008 18:11:14.641484 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1008 18:11:14.646451 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T18:10:58Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c79ef704fda1bbcc486f4c7afa3039a3ab38669833f0f66bafe24dd4cc679e60\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:10:58Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://28d21731af0f16d6a8ae6c6e9d0bebf0560d4a533db4c78c10bf1837890394ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://28d21731af0f16d6a8ae6c6e9d0bebf0560d4a533db4c78c10bf1837890394ef\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:10:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:10:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:10:55Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:12:09Z is after 2025-08-24T17:21:41Z" Oct 08 18:12:09 crc kubenswrapper[4988]: I1008 18:12:09.076360 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7e0ecefc-359f-4785-bb79-0343a62a25da\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:10:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:10:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:10:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://30bec350fefee3d19758b3ed735ede4e9aa085edb57aa3f3ec24249a67dc7ca2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:10:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2d850fd27d4ed65891ff69b9add65fb78535357f54819c2a8d196f5a340c58e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:10:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://90b1a9e14c82bf0f75f08fdb342eed35a503020f9db641e5e9ceff8b122cb3bd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:10:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://16f23c4c7163d32dc0df8d6a7ee17e2864fb261f8a40ecf05556774b173fc868\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:10:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:10:55Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:12:09Z is after 2025-08-24T17:21:41Z" Oct 08 18:12:09 crc kubenswrapper[4988]: I1008 18:12:09.101121 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e8906182d55c285b3cf095e2d9e1dac397af4618a514f11e26b44c172dd573f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:12:09Z is after 2025-08-24T17:21:41Z" Oct 08 18:12:09 crc kubenswrapper[4988]: I1008 18:12:09.121462 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-dzxs2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8f307b20-5e33-4725-99cb-210e25113532\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72ee803c6ef88524e0ce0391476f8e5a7dfbc858b095a3c2ae31bad10991ce3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q65f7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:11:16Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-dzxs2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:12:09Z is after 2025-08-24T17:21:41Z" Oct 08 18:12:09 crc kubenswrapper[4988]: I1008 18:12:09.152617 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:12:09 crc kubenswrapper[4988]: I1008 18:12:09.152668 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:12:09 crc kubenswrapper[4988]: I1008 18:12:09.152685 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:12:09 crc kubenswrapper[4988]: I1008 18:12:09.152705 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:12:09 crc kubenswrapper[4988]: I1008 18:12:09.152719 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:12:09Z","lastTransitionTime":"2025-10-08T18:12:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:12:09 crc kubenswrapper[4988]: I1008 18:12:09.237886 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 18:12:09 crc kubenswrapper[4988]: I1008 18:12:09.238005 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 18:12:09 crc kubenswrapper[4988]: I1008 18:12:09.238055 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jn2sx" Oct 08 18:12:09 crc kubenswrapper[4988]: E1008 18:12:09.238214 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 08 18:12:09 crc kubenswrapper[4988]: I1008 18:12:09.238238 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 18:12:09 crc kubenswrapper[4988]: E1008 18:12:09.238518 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jn2sx" podUID="171e0902-95ee-4a39-8b1b-a58704c30574" Oct 08 18:12:09 crc kubenswrapper[4988]: E1008 18:12:09.238746 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 08 18:12:09 crc kubenswrapper[4988]: E1008 18:12:09.238853 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 08 18:12:09 crc kubenswrapper[4988]: I1008 18:12:09.256060 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:12:09 crc kubenswrapper[4988]: I1008 18:12:09.256090 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:12:09 crc kubenswrapper[4988]: I1008 18:12:09.256100 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:12:09 crc kubenswrapper[4988]: I1008 18:12:09.256116 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:12:09 crc kubenswrapper[4988]: I1008 18:12:09.256131 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:12:09Z","lastTransitionTime":"2025-10-08T18:12:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:12:09 crc kubenswrapper[4988]: I1008 18:12:09.359537 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:12:09 crc kubenswrapper[4988]: I1008 18:12:09.359624 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:12:09 crc kubenswrapper[4988]: I1008 18:12:09.359652 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:12:09 crc kubenswrapper[4988]: I1008 18:12:09.359688 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:12:09 crc kubenswrapper[4988]: I1008 18:12:09.359715 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:12:09Z","lastTransitionTime":"2025-10-08T18:12:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:12:09 crc kubenswrapper[4988]: I1008 18:12:09.462978 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:12:09 crc kubenswrapper[4988]: I1008 18:12:09.463051 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:12:09 crc kubenswrapper[4988]: I1008 18:12:09.463075 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:12:09 crc kubenswrapper[4988]: I1008 18:12:09.463103 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:12:09 crc kubenswrapper[4988]: I1008 18:12:09.463123 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:12:09Z","lastTransitionTime":"2025-10-08T18:12:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:12:09 crc kubenswrapper[4988]: I1008 18:12:09.567073 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:12:09 crc kubenswrapper[4988]: I1008 18:12:09.567139 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:12:09 crc kubenswrapper[4988]: I1008 18:12:09.567152 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:12:09 crc kubenswrapper[4988]: I1008 18:12:09.567171 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:12:09 crc kubenswrapper[4988]: I1008 18:12:09.567182 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:12:09Z","lastTransitionTime":"2025-10-08T18:12:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:12:09 crc kubenswrapper[4988]: I1008 18:12:09.670072 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:12:09 crc kubenswrapper[4988]: I1008 18:12:09.670129 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:12:09 crc kubenswrapper[4988]: I1008 18:12:09.670141 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:12:09 crc kubenswrapper[4988]: I1008 18:12:09.670162 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:12:09 crc kubenswrapper[4988]: I1008 18:12:09.670177 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:12:09Z","lastTransitionTime":"2025-10-08T18:12:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:12:09 crc kubenswrapper[4988]: I1008 18:12:09.772574 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:12:09 crc kubenswrapper[4988]: I1008 18:12:09.772659 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:12:09 crc kubenswrapper[4988]: I1008 18:12:09.772675 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:12:09 crc kubenswrapper[4988]: I1008 18:12:09.772698 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:12:09 crc kubenswrapper[4988]: I1008 18:12:09.772731 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:12:09Z","lastTransitionTime":"2025-10-08T18:12:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:12:09 crc kubenswrapper[4988]: I1008 18:12:09.818137 4988 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-nsc65_f010a463-69f2-4082-acc3-161899a939c3/ovnkube-controller/3.log" Oct 08 18:12:09 crc kubenswrapper[4988]: I1008 18:12:09.819047 4988 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-nsc65_f010a463-69f2-4082-acc3-161899a939c3/ovnkube-controller/2.log" Oct 08 18:12:09 crc kubenswrapper[4988]: I1008 18:12:09.822067 4988 generic.go:334] "Generic (PLEG): container finished" podID="f010a463-69f2-4082-acc3-161899a939c3" containerID="763353df23d26dc093a685af435811240b5ee772d7279ee9a6d2c7a48c19cd30" exitCode=1 Oct 08 18:12:09 crc kubenswrapper[4988]: I1008 18:12:09.822118 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-nsc65" event={"ID":"f010a463-69f2-4082-acc3-161899a939c3","Type":"ContainerDied","Data":"763353df23d26dc093a685af435811240b5ee772d7279ee9a6d2c7a48c19cd30"} Oct 08 18:12:09 crc kubenswrapper[4988]: I1008 18:12:09.822200 4988 scope.go:117] "RemoveContainer" containerID="7bf11dd48efebb9da2b836bfad6a54b274247fa60902dab2f084793e4ad51a89" Oct 08 18:12:09 crc kubenswrapper[4988]: I1008 18:12:09.823026 4988 scope.go:117] "RemoveContainer" containerID="763353df23d26dc093a685af435811240b5ee772d7279ee9a6d2c7a48c19cd30" Oct 08 18:12:09 crc kubenswrapper[4988]: E1008 18:12:09.823245 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-nsc65_openshift-ovn-kubernetes(f010a463-69f2-4082-acc3-161899a939c3)\"" pod="openshift-ovn-kubernetes/ovnkube-node-nsc65" podUID="f010a463-69f2-4082-acc3-161899a939c3" Oct 08 18:12:09 crc kubenswrapper[4988]: I1008 18:12:09.836280 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9590ca0059f0ebbe86343cccb3475338c317382c3231e082655b61e67eece375\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:12:09Z is after 2025-08-24T17:21:41Z" Oct 08 18:12:09 crc kubenswrapper[4988]: I1008 18:12:09.849354 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:12:09Z is after 2025-08-24T17:21:41Z" Oct 08 18:12:09 crc kubenswrapper[4988]: I1008 18:12:09.866595 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-jnh7m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"556fb048-8afb-44a5-9ea0-2b7c89e6cf4b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bfa2f3382b23c3f788a1b31e80e8a90b3b6868870dccc46d86c3fc11ad027f52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmlt2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1a825886f105ccf9a7cc6f28a0cd012fb859f4990dade0e6aae932482146028\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1a825886f105ccf9a7cc6f28a0cd012fb859f4990dade0e6aae932482146028\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmlt2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ab2fe6e65418efa75bae72259157a2323bbb4b7d11d1d012053524e14135888\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0ab2fe6e65418efa75bae72259157a2323bbb4b7d11d1d012053524e14135888\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:11:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:11:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmlt2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://368a303309be112d9c0b7ccd6538fbaca7085d97a8ec8259f36092f07147e7cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://368a303309be112d9c0b7ccd6538fbaca7085d97a8ec8259f36092f07147e7cb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:11:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:11:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmlt2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a8ca69001dddef643f982a97c74c291e30cb9fac5b0e894e80066b885ea9151\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9a8ca69001dddef643f982a97c74c291e30cb9fac5b0e894e80066b885ea9151\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:11:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:11:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmlt2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://145303c46b48603ed1aae4b9bfe06332a3ffad28f206f8394574011a87b58cd3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://145303c46b48603ed1aae4b9bfe06332a3ffad28f206f8394574011a87b58cd3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:11:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:11:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmlt2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d08e5cfc7cbe5ec1a7a0265ed7c466363c3abeec492263170860689b46c7ff95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d08e5cfc7cbe5ec1a7a0265ed7c466363c3abeec492263170860689b46c7ff95\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:11:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:11:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmlt2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:11:16Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-jnh7m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:12:09Z is after 2025-08-24T17:21:41Z" Oct 08 18:12:09 crc kubenswrapper[4988]: I1008 18:12:09.877610 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:12:09 crc kubenswrapper[4988]: I1008 18:12:09.877650 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:12:09 crc kubenswrapper[4988]: I1008 18:12:09.877659 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:12:09 crc kubenswrapper[4988]: I1008 18:12:09.877676 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:12:09 crc kubenswrapper[4988]: I1008 18:12:09.877686 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:12:09Z","lastTransitionTime":"2025-10-08T18:12:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:12:09 crc kubenswrapper[4988]: I1008 18:12:09.882856 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f328a20b-5283-407e-8a93-2ab2da599f18\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8622d56fb198dadee31fb0ca8fc80d56828b6710648055f264533e2d5bdb3037\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z2gh4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3cdb2a757332bdcdc2a2f3362913fe8f06003b0597de20b72627f9badd5e30af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z2gh4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:11:16Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-nm54f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:12:09Z is after 2025-08-24T17:21:41Z" Oct 08 18:12:09 crc kubenswrapper[4988]: I1008 18:12:09.897721 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a35a5e989ba69c8ef36ce4cea272aadf67ce20473fc257adfcfa86a2faba6ae0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a522e937cfe03fb9f2e60cb1bea07214cf63c69bdecd5ba65a2dfdba882d28dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:12:09Z is after 2025-08-24T17:21:41Z" Oct 08 18:12:09 crc kubenswrapper[4988]: I1008 18:12:09.918922 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:12:09Z is after 2025-08-24T17:21:41Z" Oct 08 18:12:09 crc kubenswrapper[4988]: I1008 18:12:09.940592 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-h9jwp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4490d846-ca32-4cd5-9202-87cc6d499fb4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:12:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:12:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c38c1b52127a970f792ef09ef1b217c979d94f7f74cc3c27e38949a204ad831f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://00b7f6db043c28efda3a09fdc77a6ccded3eae8b6a3d2852aa500789d5f01194\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-08T18:12:04Z\\\",\\\"message\\\":\\\"2025-10-08T18:11:18+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_bee291bc-08e6-4092-b05c-4207d1f7fe40\\\\n2025-10-08T18:11:18+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_bee291bc-08e6-4092-b05c-4207d1f7fe40 to /host/opt/cni/bin/\\\\n2025-10-08T18:11:18Z [verbose] multus-daemon started\\\\n2025-10-08T18:11:18Z [verbose] Readiness Indicator file check\\\\n2025-10-08T18:12:03Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T18:11:17Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:12:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7k6rk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:11:16Z\\\"}}\" for pod \"openshift-multus\"/\"multus-h9jwp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:12:09Z is after 2025-08-24T17:21:41Z" Oct 08 18:12:09 crc kubenswrapper[4988]: I1008 18:12:09.969356 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-nsc65" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f010a463-69f2-4082-acc3-161899a939c3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://48a516f6c153743bd1a21ec6945fb368a37f4d2b9d6a326b68edc096a192ed05\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dnwgr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://03406e726f817bcedb841f81959ab12808ba35a530240f9abc5fa5c608b9a6ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dnwgr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9fbbeae0409d2e47e8de9a4306f63b0d13acbfd73cd86f83303fec14352ed718\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dnwgr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://55791327a370c0009c523b86b2234e82d3794985f2eaccf1922eb90f46b2b970\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dnwgr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cee5de93bc68345f9b737857f02832f56582d58757c50a8798eec25d19b2c20d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dnwgr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d314eb248e90371fed068b1173e26d4c9f9765417a70b341e1a5b136ec17353a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dnwgr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://763353df23d26dc093a685af435811240b5ee772d7279ee9a6d2c7a48c19cd30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7bf11dd48efebb9da2b836bfad6a54b274247fa60902dab2f084793e4ad51a89\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-08T18:11:42Z\\\",\\\"message\\\":\\\" V4IPs:[]string(nil), V6IPs:[]string(nil)}, nodeEndpoints:map[string]services.lbEndpoints{}, externalTrafficLocal:false, internalTrafficLocal:false, hasNodePort:false}}\\\\nI1008 18:11:42.211985 6628 services_controller.go:444] Built service openshift-authentication-operator/metrics LB per-node configs for network=default: []services.lbConfig(nil)\\\\nI1008 18:11:42.211998 6628 services_controller.go:445] Built service openshift-authentication-operator/metrics LB template configs for network=default: []services.lbConfig(nil)\\\\nF1008 18:11:42.212009 6628 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:11:42Z is after 2025-08-\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T18:11:41Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://763353df23d26dc093a685af435811240b5ee772d7279ee9a6d2c7a48c19cd30\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-08T18:12:09Z\\\",\\\"message\\\":\\\"ine-config-controller\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.5.16\\\\\\\", Port:9001, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI1008 18:12:09.109877 6981 model_client.go:382] Update operations generated as: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-machine-config-operator/machine-config-controller]} name:Service_openshift-machine-config-operator/machine-config-controller_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.5.16:9001:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {3f1b9878-e751-4e46-a226-ce007d2c4aa7}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1008 18:12:09.109906 6981 transact.go:42] Configuring OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-machine-config-operator/machine-config-c\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T18:12:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dnwgr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://74fcb812b84291763cc6e5ccb20c12aaad580ad968dda4734a0bbfe33d2770ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dnwgr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://befe684cea9bd6dfda3ec8e85bca431c26bd8f82823b0a69dc0d04319a3ca385\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://befe684cea9bd6dfda3ec8e85bca431c26bd8f82823b0a69dc0d04319a3ca385\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dnwgr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:11:16Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-nsc65\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:12:09Z is after 2025-08-24T17:21:41Z" Oct 08 18:12:09 crc kubenswrapper[4988]: I1008 18:12:09.980740 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:12:09 crc kubenswrapper[4988]: I1008 18:12:09.980788 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:12:09 crc kubenswrapper[4988]: I1008 18:12:09.980798 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:12:09 crc kubenswrapper[4988]: I1008 18:12:09.980817 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:12:09 crc kubenswrapper[4988]: I1008 18:12:09.980828 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:12:09Z","lastTransitionTime":"2025-10-08T18:12:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:12:09 crc kubenswrapper[4988]: I1008 18:12:09.984454 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-7wl8v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e433488d-db52-435f-b89c-f63cd477c5d3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76b30d183684740b647c1af11dc49712232cc8b0c9a392b491580e5ddb75ea7f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dbz6l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e55016c3e1b97d1b3a857130a5c8e2ec6da2333f786032118a61e3e2ad4c5651\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dbz6l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:11:29Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-7wl8v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:12:09Z is after 2025-08-24T17:21:41Z" Oct 08 18:12:09 crc kubenswrapper[4988]: I1008 18:12:09.998750 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-jn2sx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"171e0902-95ee-4a39-8b1b-a58704c30574\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9f4q2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9f4q2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:11:31Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-jn2sx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:12:09Z is after 2025-08-24T17:21:41Z" Oct 08 18:12:10 crc kubenswrapper[4988]: I1008 18:12:10.014352 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d36f620-0f26-41c7-af6e-7fb5721a0e2d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:10:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:10:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:10:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3fd3b2ca80233400d2798b035a65b833f735632efd76d1160f6bb906e07b80a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:10:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://75a2bd2757133e79c0c7f2939eb433dc45291f14808e3c3f1daaf90a239c6a71\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:10:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5cb03de583c14ad6751227c21166ac5ac4c6e11b67d5b709361ced8a0877e6d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:10:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6eb46d76cf7cfb1d4ecc4fcee398dad8904732c1664c317865864b65f5703589\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6eb46d76cf7cfb1d4ecc4fcee398dad8904732c1664c317865864b65f5703589\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:10:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:10:56Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:10:55Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:12:10Z is after 2025-08-24T17:21:41Z" Oct 08 18:12:10 crc kubenswrapper[4988]: I1008 18:12:10.029419 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:12:10Z is after 2025-08-24T17:21:41Z" Oct 08 18:12:10 crc kubenswrapper[4988]: I1008 18:12:10.041547 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"63edba53-2181-4227-9a72-a21793ad7cd0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:10:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:10:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:10:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://adbdfab7dbc007ce91c26fa7a6535ef27949ad2567f1bc6b4c52b6a3f62789f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:10:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7b7185463f2a962f3214a2524d783a54b8fbc76f385ea27001087bfb7d023561\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7b7185463f2a962f3214a2524d783a54b8fbc76f385ea27001087bfb7d023561\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:10:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:10:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:10:55Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:12:10Z is after 2025-08-24T17:21:41Z" Oct 08 18:12:10 crc kubenswrapper[4988]: I1008 18:12:10.055760 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-s4f2k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ba5e11d7-7bde-4c3d-983b-884232219652\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://824bc34a2807068d3fceb10651803ae9ae5651e940f7aed4877c544d76dae4d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hn5jf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:11:15Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-s4f2k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:12:10Z is after 2025-08-24T17:21:41Z" Oct 08 18:12:10 crc kubenswrapper[4988]: I1008 18:12:10.076704 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e8906182d55c285b3cf095e2d9e1dac397af4618a514f11e26b44c172dd573f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:12:10Z is after 2025-08-24T17:21:41Z" Oct 08 18:12:10 crc kubenswrapper[4988]: I1008 18:12:10.084224 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:12:10 crc kubenswrapper[4988]: I1008 18:12:10.084276 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:12:10 crc kubenswrapper[4988]: I1008 18:12:10.084304 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:12:10 crc kubenswrapper[4988]: I1008 18:12:10.084333 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:12:10 crc kubenswrapper[4988]: I1008 18:12:10.084349 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:12:10Z","lastTransitionTime":"2025-10-08T18:12:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:12:10 crc kubenswrapper[4988]: I1008 18:12:10.093576 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-dzxs2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8f307b20-5e33-4725-99cb-210e25113532\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72ee803c6ef88524e0ce0391476f8e5a7dfbc858b095a3c2ae31bad10991ce3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q65f7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:11:16Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-dzxs2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:12:10Z is after 2025-08-24T17:21:41Z" Oct 08 18:12:10 crc kubenswrapper[4988]: I1008 18:12:10.115463 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aef71be1-1d12-46e5-95b2-bd399040fd4d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:10:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:10:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:10:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77c68f3ecb8a0724499e495cb26ef3db5beecb140f2558d8c0b25f41928b84ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:10:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a560ee84c2405227de71fda127566213cd4f017e73d5581442ba267781b1c539\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:10:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6100accc3bd08928bc1ff88310f29fa5da37fd2d5c315c94279801c9778c4745\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:10:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b0e596d7ce4e9341c44499af79d3b3caf0e5e2c156902ea40cf2a3dca672be9e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5bd35aef24be034a0b305a40bb53ca171a4f76db3b93b4f78ae95099b0a4cde3\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-08T18:11:14Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1008 18:11:09.091034 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1008 18:11:09.091984 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3472374062/tls.crt::/tmp/serving-cert-3472374062/tls.key\\\\\\\"\\\\nI1008 18:11:14.619696 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1008 18:11:14.628837 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1008 18:11:14.628898 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1008 18:11:14.628994 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1008 18:11:14.629036 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1008 18:11:14.641318 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1008 18:11:14.641363 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1008 18:11:14.641372 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1008 18:11:14.641382 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1008 18:11:14.641415 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1008 18:11:14.641423 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1008 18:11:14.641431 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1008 18:11:14.641484 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1008 18:11:14.646451 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T18:10:58Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c79ef704fda1bbcc486f4c7afa3039a3ab38669833f0f66bafe24dd4cc679e60\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:10:58Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://28d21731af0f16d6a8ae6c6e9d0bebf0560d4a533db4c78c10bf1837890394ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://28d21731af0f16d6a8ae6c6e9d0bebf0560d4a533db4c78c10bf1837890394ef\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:10:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:10:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:10:55Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:12:10Z is after 2025-08-24T17:21:41Z" Oct 08 18:12:10 crc kubenswrapper[4988]: I1008 18:12:10.136028 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7e0ecefc-359f-4785-bb79-0343a62a25da\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:10:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:10:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:10:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://30bec350fefee3d19758b3ed735ede4e9aa085edb57aa3f3ec24249a67dc7ca2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:10:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2d850fd27d4ed65891ff69b9add65fb78535357f54819c2a8d196f5a340c58e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:10:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://90b1a9e14c82bf0f75f08fdb342eed35a503020f9db641e5e9ceff8b122cb3bd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:10:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://16f23c4c7163d32dc0df8d6a7ee17e2864fb261f8a40ecf05556774b173fc868\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:10:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:10:55Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:12:10Z is after 2025-08-24T17:21:41Z" Oct 08 18:12:10 crc kubenswrapper[4988]: I1008 18:12:10.188174 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:12:10 crc kubenswrapper[4988]: I1008 18:12:10.188246 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:12:10 crc kubenswrapper[4988]: I1008 18:12:10.188267 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:12:10 crc kubenswrapper[4988]: I1008 18:12:10.188291 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:12:10 crc kubenswrapper[4988]: I1008 18:12:10.188306 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:12:10Z","lastTransitionTime":"2025-10-08T18:12:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:12:10 crc kubenswrapper[4988]: I1008 18:12:10.292078 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:12:10 crc kubenswrapper[4988]: I1008 18:12:10.292121 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:12:10 crc kubenswrapper[4988]: I1008 18:12:10.292131 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:12:10 crc kubenswrapper[4988]: I1008 18:12:10.292149 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:12:10 crc kubenswrapper[4988]: I1008 18:12:10.292160 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:12:10Z","lastTransitionTime":"2025-10-08T18:12:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:12:10 crc kubenswrapper[4988]: I1008 18:12:10.395935 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:12:10 crc kubenswrapper[4988]: I1008 18:12:10.396002 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:12:10 crc kubenswrapper[4988]: I1008 18:12:10.396020 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:12:10 crc kubenswrapper[4988]: I1008 18:12:10.396048 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:12:10 crc kubenswrapper[4988]: I1008 18:12:10.396069 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:12:10Z","lastTransitionTime":"2025-10-08T18:12:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:12:10 crc kubenswrapper[4988]: I1008 18:12:10.498832 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:12:10 crc kubenswrapper[4988]: I1008 18:12:10.498894 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:12:10 crc kubenswrapper[4988]: I1008 18:12:10.498915 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:12:10 crc kubenswrapper[4988]: I1008 18:12:10.498942 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:12:10 crc kubenswrapper[4988]: I1008 18:12:10.498961 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:12:10Z","lastTransitionTime":"2025-10-08T18:12:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:12:10 crc kubenswrapper[4988]: I1008 18:12:10.601920 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:12:10 crc kubenswrapper[4988]: I1008 18:12:10.602055 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:12:10 crc kubenswrapper[4988]: I1008 18:12:10.602081 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:12:10 crc kubenswrapper[4988]: I1008 18:12:10.602114 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:12:10 crc kubenswrapper[4988]: I1008 18:12:10.602140 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:12:10Z","lastTransitionTime":"2025-10-08T18:12:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:12:10 crc kubenswrapper[4988]: I1008 18:12:10.705508 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:12:10 crc kubenswrapper[4988]: I1008 18:12:10.705597 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:12:10 crc kubenswrapper[4988]: I1008 18:12:10.705611 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:12:10 crc kubenswrapper[4988]: I1008 18:12:10.705629 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:12:10 crc kubenswrapper[4988]: I1008 18:12:10.705641 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:12:10Z","lastTransitionTime":"2025-10-08T18:12:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:12:10 crc kubenswrapper[4988]: I1008 18:12:10.808547 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:12:10 crc kubenswrapper[4988]: I1008 18:12:10.808610 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:12:10 crc kubenswrapper[4988]: I1008 18:12:10.808626 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:12:10 crc kubenswrapper[4988]: I1008 18:12:10.808652 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:12:10 crc kubenswrapper[4988]: I1008 18:12:10.808671 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:12:10Z","lastTransitionTime":"2025-10-08T18:12:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:12:10 crc kubenswrapper[4988]: I1008 18:12:10.827359 4988 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-nsc65_f010a463-69f2-4082-acc3-161899a939c3/ovnkube-controller/3.log" Oct 08 18:12:10 crc kubenswrapper[4988]: I1008 18:12:10.831937 4988 scope.go:117] "RemoveContainer" containerID="763353df23d26dc093a685af435811240b5ee772d7279ee9a6d2c7a48c19cd30" Oct 08 18:12:10 crc kubenswrapper[4988]: E1008 18:12:10.832208 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-nsc65_openshift-ovn-kubernetes(f010a463-69f2-4082-acc3-161899a939c3)\"" pod="openshift-ovn-kubernetes/ovnkube-node-nsc65" podUID="f010a463-69f2-4082-acc3-161899a939c3" Oct 08 18:12:10 crc kubenswrapper[4988]: I1008 18:12:10.847730 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"63edba53-2181-4227-9a72-a21793ad7cd0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:10:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:10:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:10:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://adbdfab7dbc007ce91c26fa7a6535ef27949ad2567f1bc6b4c52b6a3f62789f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:10:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7b7185463f2a962f3214a2524d783a54b8fbc76f385ea27001087bfb7d023561\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7b7185463f2a962f3214a2524d783a54b8fbc76f385ea27001087bfb7d023561\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:10:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:10:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:10:55Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:12:10Z is after 2025-08-24T17:21:41Z" Oct 08 18:12:10 crc kubenswrapper[4988]: I1008 18:12:10.864515 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-s4f2k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ba5e11d7-7bde-4c3d-983b-884232219652\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://824bc34a2807068d3fceb10651803ae9ae5651e940f7aed4877c544d76dae4d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hn5jf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:11:15Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-s4f2k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:12:10Z is after 2025-08-24T17:21:41Z" Oct 08 18:12:10 crc kubenswrapper[4988]: I1008 18:12:10.888215 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aef71be1-1d12-46e5-95b2-bd399040fd4d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:10:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:10:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:10:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77c68f3ecb8a0724499e495cb26ef3db5beecb140f2558d8c0b25f41928b84ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:10:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a560ee84c2405227de71fda127566213cd4f017e73d5581442ba267781b1c539\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:10:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6100accc3bd08928bc1ff88310f29fa5da37fd2d5c315c94279801c9778c4745\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:10:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b0e596d7ce4e9341c44499af79d3b3caf0e5e2c156902ea40cf2a3dca672be9e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5bd35aef24be034a0b305a40bb53ca171a4f76db3b93b4f78ae95099b0a4cde3\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-08T18:11:14Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1008 18:11:09.091034 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1008 18:11:09.091984 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3472374062/tls.crt::/tmp/serving-cert-3472374062/tls.key\\\\\\\"\\\\nI1008 18:11:14.619696 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1008 18:11:14.628837 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1008 18:11:14.628898 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1008 18:11:14.628994 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1008 18:11:14.629036 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1008 18:11:14.641318 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1008 18:11:14.641363 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1008 18:11:14.641372 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1008 18:11:14.641382 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1008 18:11:14.641415 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1008 18:11:14.641423 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1008 18:11:14.641431 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1008 18:11:14.641484 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1008 18:11:14.646451 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T18:10:58Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c79ef704fda1bbcc486f4c7afa3039a3ab38669833f0f66bafe24dd4cc679e60\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:10:58Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://28d21731af0f16d6a8ae6c6e9d0bebf0560d4a533db4c78c10bf1837890394ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://28d21731af0f16d6a8ae6c6e9d0bebf0560d4a533db4c78c10bf1837890394ef\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:10:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:10:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:10:55Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:12:10Z is after 2025-08-24T17:21:41Z" Oct 08 18:12:10 crc kubenswrapper[4988]: I1008 18:12:10.909264 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7e0ecefc-359f-4785-bb79-0343a62a25da\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:10:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:10:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:10:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://30bec350fefee3d19758b3ed735ede4e9aa085edb57aa3f3ec24249a67dc7ca2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:10:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2d850fd27d4ed65891ff69b9add65fb78535357f54819c2a8d196f5a340c58e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:10:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://90b1a9e14c82bf0f75f08fdb342eed35a503020f9db641e5e9ceff8b122cb3bd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:10:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://16f23c4c7163d32dc0df8d6a7ee17e2864fb261f8a40ecf05556774b173fc868\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:10:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:10:55Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:12:10Z is after 2025-08-24T17:21:41Z" Oct 08 18:12:10 crc kubenswrapper[4988]: I1008 18:12:10.911353 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:12:10 crc kubenswrapper[4988]: I1008 18:12:10.911425 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:12:10 crc kubenswrapper[4988]: I1008 18:12:10.911443 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:12:10 crc kubenswrapper[4988]: I1008 18:12:10.911470 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:12:10 crc kubenswrapper[4988]: I1008 18:12:10.911487 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:12:10Z","lastTransitionTime":"2025-10-08T18:12:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:12:10 crc kubenswrapper[4988]: I1008 18:12:10.930314 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e8906182d55c285b3cf095e2d9e1dac397af4618a514f11e26b44c172dd573f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:12:10Z is after 2025-08-24T17:21:41Z" Oct 08 18:12:10 crc kubenswrapper[4988]: I1008 18:12:10.941937 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-dzxs2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8f307b20-5e33-4725-99cb-210e25113532\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72ee803c6ef88524e0ce0391476f8e5a7dfbc858b095a3c2ae31bad10991ce3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q65f7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:11:16Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-dzxs2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:12:10Z is after 2025-08-24T17:21:41Z" Oct 08 18:12:10 crc kubenswrapper[4988]: I1008 18:12:10.963449 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a35a5e989ba69c8ef36ce4cea272aadf67ce20473fc257adfcfa86a2faba6ae0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a522e937cfe03fb9f2e60cb1bea07214cf63c69bdecd5ba65a2dfdba882d28dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:12:10Z is after 2025-08-24T17:21:41Z" Oct 08 18:12:10 crc kubenswrapper[4988]: I1008 18:12:10.978805 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9590ca0059f0ebbe86343cccb3475338c317382c3231e082655b61e67eece375\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:12:10Z is after 2025-08-24T17:21:41Z" Oct 08 18:12:11 crc kubenswrapper[4988]: I1008 18:12:11.000909 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:12:10Z is after 2025-08-24T17:21:41Z" Oct 08 18:12:11 crc kubenswrapper[4988]: I1008 18:12:11.015022 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:12:11 crc kubenswrapper[4988]: I1008 18:12:11.015285 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:12:11 crc kubenswrapper[4988]: I1008 18:12:11.015506 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:12:11 crc kubenswrapper[4988]: I1008 18:12:11.015732 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:12:11 crc kubenswrapper[4988]: I1008 18:12:11.015866 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:12:11Z","lastTransitionTime":"2025-10-08T18:12:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:12:11 crc kubenswrapper[4988]: I1008 18:12:11.018642 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-jnh7m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"556fb048-8afb-44a5-9ea0-2b7c89e6cf4b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bfa2f3382b23c3f788a1b31e80e8a90b3b6868870dccc46d86c3fc11ad027f52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmlt2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1a825886f105ccf9a7cc6f28a0cd012fb859f4990dade0e6aae932482146028\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1a825886f105ccf9a7cc6f28a0cd012fb859f4990dade0e6aae932482146028\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmlt2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ab2fe6e65418efa75bae72259157a2323bbb4b7d11d1d012053524e14135888\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0ab2fe6e65418efa75bae72259157a2323bbb4b7d11d1d012053524e14135888\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:11:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:11:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmlt2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://368a303309be112d9c0b7ccd6538fbaca7085d97a8ec8259f36092f07147e7cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://368a303309be112d9c0b7ccd6538fbaca7085d97a8ec8259f36092f07147e7cb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:11:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:11:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmlt2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a8ca69001dddef643f982a97c74c291e30cb9fac5b0e894e80066b885ea9151\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9a8ca69001dddef643f982a97c74c291e30cb9fac5b0e894e80066b885ea9151\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:11:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:11:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmlt2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://145303c46b48603ed1aae4b9bfe06332a3ffad28f206f8394574011a87b58cd3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://145303c46b48603ed1aae4b9bfe06332a3ffad28f206f8394574011a87b58cd3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:11:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:11:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmlt2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d08e5cfc7cbe5ec1a7a0265ed7c466363c3abeec492263170860689b46c7ff95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d08e5cfc7cbe5ec1a7a0265ed7c466363c3abeec492263170860689b46c7ff95\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:11:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:11:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmlt2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:11:16Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-jnh7m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:12:11Z is after 2025-08-24T17:21:41Z" Oct 08 18:12:11 crc kubenswrapper[4988]: I1008 18:12:11.036292 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f328a20b-5283-407e-8a93-2ab2da599f18\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8622d56fb198dadee31fb0ca8fc80d56828b6710648055f264533e2d5bdb3037\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z2gh4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3cdb2a757332bdcdc2a2f3362913fe8f06003b0597de20b72627f9badd5e30af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z2gh4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:11:16Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-nm54f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:12:11Z is after 2025-08-24T17:21:41Z" Oct 08 18:12:11 crc kubenswrapper[4988]: I1008 18:12:11.053504 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d36f620-0f26-41c7-af6e-7fb5721a0e2d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:10:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:10:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:10:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3fd3b2ca80233400d2798b035a65b833f735632efd76d1160f6bb906e07b80a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:10:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://75a2bd2757133e79c0c7f2939eb433dc45291f14808e3c3f1daaf90a239c6a71\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:10:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5cb03de583c14ad6751227c21166ac5ac4c6e11b67d5b709361ced8a0877e6d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:10:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6eb46d76cf7cfb1d4ecc4fcee398dad8904732c1664c317865864b65f5703589\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6eb46d76cf7cfb1d4ecc4fcee398dad8904732c1664c317865864b65f5703589\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:10:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:10:56Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:10:55Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:12:11Z is after 2025-08-24T17:21:41Z" Oct 08 18:12:11 crc kubenswrapper[4988]: I1008 18:12:11.068224 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:12:11Z is after 2025-08-24T17:21:41Z" Oct 08 18:12:11 crc kubenswrapper[4988]: I1008 18:12:11.083856 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:12:11Z is after 2025-08-24T17:21:41Z" Oct 08 18:12:11 crc kubenswrapper[4988]: I1008 18:12:11.099898 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-h9jwp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4490d846-ca32-4cd5-9202-87cc6d499fb4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:12:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:12:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c38c1b52127a970f792ef09ef1b217c979d94f7f74cc3c27e38949a204ad831f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://00b7f6db043c28efda3a09fdc77a6ccded3eae8b6a3d2852aa500789d5f01194\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-08T18:12:04Z\\\",\\\"message\\\":\\\"2025-10-08T18:11:18+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_bee291bc-08e6-4092-b05c-4207d1f7fe40\\\\n2025-10-08T18:11:18+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_bee291bc-08e6-4092-b05c-4207d1f7fe40 to /host/opt/cni/bin/\\\\n2025-10-08T18:11:18Z [verbose] multus-daemon started\\\\n2025-10-08T18:11:18Z [verbose] Readiness Indicator file check\\\\n2025-10-08T18:12:03Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T18:11:17Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:12:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7k6rk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:11:16Z\\\"}}\" for pod \"openshift-multus\"/\"multus-h9jwp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:12:11Z is after 2025-08-24T17:21:41Z" Oct 08 18:12:11 crc kubenswrapper[4988]: I1008 18:12:11.118293 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:12:11 crc kubenswrapper[4988]: I1008 18:12:11.118539 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:12:11 crc kubenswrapper[4988]: I1008 18:12:11.118682 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:12:11 crc kubenswrapper[4988]: I1008 18:12:11.118832 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:12:11 crc kubenswrapper[4988]: I1008 18:12:11.118970 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:12:11Z","lastTransitionTime":"2025-10-08T18:12:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:12:11 crc kubenswrapper[4988]: I1008 18:12:11.130844 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-nsc65" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f010a463-69f2-4082-acc3-161899a939c3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://48a516f6c153743bd1a21ec6945fb368a37f4d2b9d6a326b68edc096a192ed05\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dnwgr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://03406e726f817bcedb841f81959ab12808ba35a530240f9abc5fa5c608b9a6ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dnwgr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9fbbeae0409d2e47e8de9a4306f63b0d13acbfd73cd86f83303fec14352ed718\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dnwgr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://55791327a370c0009c523b86b2234e82d3794985f2eaccf1922eb90f46b2b970\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dnwgr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cee5de93bc68345f9b737857f02832f56582d58757c50a8798eec25d19b2c20d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dnwgr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d314eb248e90371fed068b1173e26d4c9f9765417a70b341e1a5b136ec17353a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dnwgr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://763353df23d26dc093a685af435811240b5ee772d7279ee9a6d2c7a48c19cd30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://763353df23d26dc093a685af435811240b5ee772d7279ee9a6d2c7a48c19cd30\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-08T18:12:09Z\\\",\\\"message\\\":\\\"ine-config-controller\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.5.16\\\\\\\", Port:9001, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI1008 18:12:09.109877 6981 model_client.go:382] Update operations generated as: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-machine-config-operator/machine-config-controller]} name:Service_openshift-machine-config-operator/machine-config-controller_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.5.16:9001:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {3f1b9878-e751-4e46-a226-ce007d2c4aa7}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1008 18:12:09.109906 6981 transact.go:42] Configuring OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-machine-config-operator/machine-config-c\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T18:12:08Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-nsc65_openshift-ovn-kubernetes(f010a463-69f2-4082-acc3-161899a939c3)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dnwgr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://74fcb812b84291763cc6e5ccb20c12aaad580ad968dda4734a0bbfe33d2770ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dnwgr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://befe684cea9bd6dfda3ec8e85bca431c26bd8f82823b0a69dc0d04319a3ca385\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://befe684cea9bd6dfda3ec8e85bca431c26bd8f82823b0a69dc0d04319a3ca385\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dnwgr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:11:16Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-nsc65\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:12:11Z is after 2025-08-24T17:21:41Z" Oct 08 18:12:11 crc kubenswrapper[4988]: I1008 18:12:11.149014 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-7wl8v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e433488d-db52-435f-b89c-f63cd477c5d3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76b30d183684740b647c1af11dc49712232cc8b0c9a392b491580e5ddb75ea7f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dbz6l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e55016c3e1b97d1b3a857130a5c8e2ec6da2333f786032118a61e3e2ad4c5651\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dbz6l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:11:29Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-7wl8v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:12:11Z is after 2025-08-24T17:21:41Z" Oct 08 18:12:11 crc kubenswrapper[4988]: I1008 18:12:11.167025 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-jn2sx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"171e0902-95ee-4a39-8b1b-a58704c30574\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9f4q2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9f4q2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:11:31Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-jn2sx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:12:11Z is after 2025-08-24T17:21:41Z" Oct 08 18:12:11 crc kubenswrapper[4988]: I1008 18:12:11.221976 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:12:11 crc kubenswrapper[4988]: I1008 18:12:11.222123 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:12:11 crc kubenswrapper[4988]: I1008 18:12:11.222222 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:12:11 crc kubenswrapper[4988]: I1008 18:12:11.222323 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:12:11 crc kubenswrapper[4988]: I1008 18:12:11.222430 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:12:11Z","lastTransitionTime":"2025-10-08T18:12:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:12:11 crc kubenswrapper[4988]: I1008 18:12:11.237791 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 18:12:11 crc kubenswrapper[4988]: I1008 18:12:11.237806 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 18:12:11 crc kubenswrapper[4988]: I1008 18:12:11.237832 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jn2sx" Oct 08 18:12:11 crc kubenswrapper[4988]: E1008 18:12:11.238222 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 08 18:12:11 crc kubenswrapper[4988]: E1008 18:12:11.238023 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 08 18:12:11 crc kubenswrapper[4988]: I1008 18:12:11.237885 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 18:12:11 crc kubenswrapper[4988]: E1008 18:12:11.238365 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jn2sx" podUID="171e0902-95ee-4a39-8b1b-a58704c30574" Oct 08 18:12:11 crc kubenswrapper[4988]: E1008 18:12:11.238509 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 08 18:12:11 crc kubenswrapper[4988]: I1008 18:12:11.325057 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:12:11 crc kubenswrapper[4988]: I1008 18:12:11.325110 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:12:11 crc kubenswrapper[4988]: I1008 18:12:11.325127 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:12:11 crc kubenswrapper[4988]: I1008 18:12:11.325147 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:12:11 crc kubenswrapper[4988]: I1008 18:12:11.325163 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:12:11Z","lastTransitionTime":"2025-10-08T18:12:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:12:11 crc kubenswrapper[4988]: I1008 18:12:11.428915 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:12:11 crc kubenswrapper[4988]: I1008 18:12:11.428997 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:12:11 crc kubenswrapper[4988]: I1008 18:12:11.429019 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:12:11 crc kubenswrapper[4988]: I1008 18:12:11.429040 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:12:11 crc kubenswrapper[4988]: I1008 18:12:11.429056 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:12:11Z","lastTransitionTime":"2025-10-08T18:12:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:12:11 crc kubenswrapper[4988]: I1008 18:12:11.532029 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:12:11 crc kubenswrapper[4988]: I1008 18:12:11.532084 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:12:11 crc kubenswrapper[4988]: I1008 18:12:11.532098 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:12:11 crc kubenswrapper[4988]: I1008 18:12:11.532116 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:12:11 crc kubenswrapper[4988]: I1008 18:12:11.532129 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:12:11Z","lastTransitionTime":"2025-10-08T18:12:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:12:11 crc kubenswrapper[4988]: I1008 18:12:11.635331 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:12:11 crc kubenswrapper[4988]: I1008 18:12:11.635418 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:12:11 crc kubenswrapper[4988]: I1008 18:12:11.635433 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:12:11 crc kubenswrapper[4988]: I1008 18:12:11.635453 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:12:11 crc kubenswrapper[4988]: I1008 18:12:11.635466 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:12:11Z","lastTransitionTime":"2025-10-08T18:12:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:12:11 crc kubenswrapper[4988]: I1008 18:12:11.738323 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:12:11 crc kubenswrapper[4988]: I1008 18:12:11.738373 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:12:11 crc kubenswrapper[4988]: I1008 18:12:11.738399 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:12:11 crc kubenswrapper[4988]: I1008 18:12:11.738415 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:12:11 crc kubenswrapper[4988]: I1008 18:12:11.738425 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:12:11Z","lastTransitionTime":"2025-10-08T18:12:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:12:11 crc kubenswrapper[4988]: I1008 18:12:11.841857 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:12:11 crc kubenswrapper[4988]: I1008 18:12:11.841906 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:12:11 crc kubenswrapper[4988]: I1008 18:12:11.841923 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:12:11 crc kubenswrapper[4988]: I1008 18:12:11.841943 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:12:11 crc kubenswrapper[4988]: I1008 18:12:11.841957 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:12:11Z","lastTransitionTime":"2025-10-08T18:12:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:12:11 crc kubenswrapper[4988]: I1008 18:12:11.946049 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:12:11 crc kubenswrapper[4988]: I1008 18:12:11.946125 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:12:11 crc kubenswrapper[4988]: I1008 18:12:11.946143 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:12:11 crc kubenswrapper[4988]: I1008 18:12:11.946170 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:12:11 crc kubenswrapper[4988]: I1008 18:12:11.946190 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:12:11Z","lastTransitionTime":"2025-10-08T18:12:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:12:12 crc kubenswrapper[4988]: I1008 18:12:12.048455 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:12:12 crc kubenswrapper[4988]: I1008 18:12:12.048542 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:12:12 crc kubenswrapper[4988]: I1008 18:12:12.048557 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:12:12 crc kubenswrapper[4988]: I1008 18:12:12.048579 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:12:12 crc kubenswrapper[4988]: I1008 18:12:12.048593 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:12:12Z","lastTransitionTime":"2025-10-08T18:12:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:12:12 crc kubenswrapper[4988]: I1008 18:12:12.151837 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:12:12 crc kubenswrapper[4988]: I1008 18:12:12.151904 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:12:12 crc kubenswrapper[4988]: I1008 18:12:12.151920 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:12:12 crc kubenswrapper[4988]: I1008 18:12:12.151946 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:12:12 crc kubenswrapper[4988]: I1008 18:12:12.151964 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:12:12Z","lastTransitionTime":"2025-10-08T18:12:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:12:12 crc kubenswrapper[4988]: I1008 18:12:12.255191 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:12:12 crc kubenswrapper[4988]: I1008 18:12:12.255293 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:12:12 crc kubenswrapper[4988]: I1008 18:12:12.255325 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:12:12 crc kubenswrapper[4988]: I1008 18:12:12.255361 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:12:12 crc kubenswrapper[4988]: I1008 18:12:12.255414 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:12:12Z","lastTransitionTime":"2025-10-08T18:12:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:12:12 crc kubenswrapper[4988]: I1008 18:12:12.359533 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:12:12 crc kubenswrapper[4988]: I1008 18:12:12.359633 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:12:12 crc kubenswrapper[4988]: I1008 18:12:12.359656 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:12:12 crc kubenswrapper[4988]: I1008 18:12:12.359690 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:12:12 crc kubenswrapper[4988]: I1008 18:12:12.359713 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:12:12Z","lastTransitionTime":"2025-10-08T18:12:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:12:12 crc kubenswrapper[4988]: I1008 18:12:12.462829 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:12:12 crc kubenswrapper[4988]: I1008 18:12:12.462906 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:12:12 crc kubenswrapper[4988]: I1008 18:12:12.462927 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:12:12 crc kubenswrapper[4988]: I1008 18:12:12.462958 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:12:12 crc kubenswrapper[4988]: I1008 18:12:12.462979 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:12:12Z","lastTransitionTime":"2025-10-08T18:12:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:12:12 crc kubenswrapper[4988]: I1008 18:12:12.566604 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:12:12 crc kubenswrapper[4988]: I1008 18:12:12.566656 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:12:12 crc kubenswrapper[4988]: I1008 18:12:12.566667 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:12:12 crc kubenswrapper[4988]: I1008 18:12:12.566686 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:12:12 crc kubenswrapper[4988]: I1008 18:12:12.566699 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:12:12Z","lastTransitionTime":"2025-10-08T18:12:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:12:12 crc kubenswrapper[4988]: I1008 18:12:12.589485 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:12:12 crc kubenswrapper[4988]: I1008 18:12:12.589546 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:12:12 crc kubenswrapper[4988]: I1008 18:12:12.589559 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:12:12 crc kubenswrapper[4988]: I1008 18:12:12.589960 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:12:12 crc kubenswrapper[4988]: I1008 18:12:12.593616 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:12:12Z","lastTransitionTime":"2025-10-08T18:12:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:12:12 crc kubenswrapper[4988]: E1008 18:12:12.614286 4988 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T18:12:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T18:12:12Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T18:12:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T18:12:12Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T18:12:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T18:12:12Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T18:12:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T18:12:12Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"95ee9fd0-8bc9-4ca5-9d79-5ee94d58bfeb\\\",\\\"systemUUID\\\":\\\"bd9489c4-fb7d-486d-b7ce-a13b6d7de36a\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:12:12Z is after 2025-08-24T17:21:41Z" Oct 08 18:12:12 crc kubenswrapper[4988]: I1008 18:12:12.620247 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:12:12 crc kubenswrapper[4988]: I1008 18:12:12.620315 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:12:12 crc kubenswrapper[4988]: I1008 18:12:12.620330 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:12:12 crc kubenswrapper[4988]: I1008 18:12:12.620354 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:12:12 crc kubenswrapper[4988]: I1008 18:12:12.620370 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:12:12Z","lastTransitionTime":"2025-10-08T18:12:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:12:12 crc kubenswrapper[4988]: E1008 18:12:12.640262 4988 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T18:12:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T18:12:12Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T18:12:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T18:12:12Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T18:12:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T18:12:12Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T18:12:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T18:12:12Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"95ee9fd0-8bc9-4ca5-9d79-5ee94d58bfeb\\\",\\\"systemUUID\\\":\\\"bd9489c4-fb7d-486d-b7ce-a13b6d7de36a\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:12:12Z is after 2025-08-24T17:21:41Z" Oct 08 18:12:12 crc kubenswrapper[4988]: I1008 18:12:12.646482 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:12:12 crc kubenswrapper[4988]: I1008 18:12:12.646649 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:12:12 crc kubenswrapper[4988]: I1008 18:12:12.646751 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:12:12 crc kubenswrapper[4988]: I1008 18:12:12.646868 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:12:12 crc kubenswrapper[4988]: I1008 18:12:12.646961 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:12:12Z","lastTransitionTime":"2025-10-08T18:12:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:12:12 crc kubenswrapper[4988]: E1008 18:12:12.667838 4988 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T18:12:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T18:12:12Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T18:12:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T18:12:12Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T18:12:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T18:12:12Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T18:12:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T18:12:12Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"95ee9fd0-8bc9-4ca5-9d79-5ee94d58bfeb\\\",\\\"systemUUID\\\":\\\"bd9489c4-fb7d-486d-b7ce-a13b6d7de36a\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:12:12Z is after 2025-08-24T17:21:41Z" Oct 08 18:12:12 crc kubenswrapper[4988]: I1008 18:12:12.673880 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:12:12 crc kubenswrapper[4988]: I1008 18:12:12.674061 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:12:12 crc kubenswrapper[4988]: I1008 18:12:12.674211 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:12:12 crc kubenswrapper[4988]: I1008 18:12:12.674369 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:12:12 crc kubenswrapper[4988]: I1008 18:12:12.674556 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:12:12Z","lastTransitionTime":"2025-10-08T18:12:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:12:12 crc kubenswrapper[4988]: E1008 18:12:12.694991 4988 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T18:12:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T18:12:12Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T18:12:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T18:12:12Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T18:12:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T18:12:12Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T18:12:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T18:12:12Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"95ee9fd0-8bc9-4ca5-9d79-5ee94d58bfeb\\\",\\\"systemUUID\\\":\\\"bd9489c4-fb7d-486d-b7ce-a13b6d7de36a\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:12:12Z is after 2025-08-24T17:21:41Z" Oct 08 18:12:12 crc kubenswrapper[4988]: I1008 18:12:12.700910 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:12:12 crc kubenswrapper[4988]: I1008 18:12:12.700981 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:12:12 crc kubenswrapper[4988]: I1008 18:12:12.701002 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:12:12 crc kubenswrapper[4988]: I1008 18:12:12.701033 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:12:12 crc kubenswrapper[4988]: I1008 18:12:12.701053 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:12:12Z","lastTransitionTime":"2025-10-08T18:12:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:12:12 crc kubenswrapper[4988]: E1008 18:12:12.718951 4988 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T18:12:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T18:12:12Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T18:12:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T18:12:12Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T18:12:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T18:12:12Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T18:12:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T18:12:12Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"95ee9fd0-8bc9-4ca5-9d79-5ee94d58bfeb\\\",\\\"systemUUID\\\":\\\"bd9489c4-fb7d-486d-b7ce-a13b6d7de36a\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:12:12Z is after 2025-08-24T17:21:41Z" Oct 08 18:12:12 crc kubenswrapper[4988]: E1008 18:12:12.719144 4988 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Oct 08 18:12:12 crc kubenswrapper[4988]: I1008 18:12:12.721765 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:12:12 crc kubenswrapper[4988]: I1008 18:12:12.721800 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:12:12 crc kubenswrapper[4988]: I1008 18:12:12.721819 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:12:12 crc kubenswrapper[4988]: I1008 18:12:12.721842 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:12:12 crc kubenswrapper[4988]: I1008 18:12:12.721857 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:12:12Z","lastTransitionTime":"2025-10-08T18:12:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:12:12 crc kubenswrapper[4988]: I1008 18:12:12.825252 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:12:12 crc kubenswrapper[4988]: I1008 18:12:12.826318 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:12:12 crc kubenswrapper[4988]: I1008 18:12:12.826577 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:12:12 crc kubenswrapper[4988]: I1008 18:12:12.826747 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:12:12 crc kubenswrapper[4988]: I1008 18:12:12.826946 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:12:12Z","lastTransitionTime":"2025-10-08T18:12:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:12:12 crc kubenswrapper[4988]: I1008 18:12:12.933801 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:12:12 crc kubenswrapper[4988]: I1008 18:12:12.934354 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:12:12 crc kubenswrapper[4988]: I1008 18:12:12.934635 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:12:12 crc kubenswrapper[4988]: I1008 18:12:12.934801 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:12:12 crc kubenswrapper[4988]: I1008 18:12:12.934943 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:12:12Z","lastTransitionTime":"2025-10-08T18:12:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:12:13 crc kubenswrapper[4988]: I1008 18:12:13.038441 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:12:13 crc kubenswrapper[4988]: I1008 18:12:13.038530 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:12:13 crc kubenswrapper[4988]: I1008 18:12:13.038548 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:12:13 crc kubenswrapper[4988]: I1008 18:12:13.038578 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:12:13 crc kubenswrapper[4988]: I1008 18:12:13.038599 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:12:13Z","lastTransitionTime":"2025-10-08T18:12:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:12:13 crc kubenswrapper[4988]: I1008 18:12:13.142351 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:12:13 crc kubenswrapper[4988]: I1008 18:12:13.142451 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:12:13 crc kubenswrapper[4988]: I1008 18:12:13.142472 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:12:13 crc kubenswrapper[4988]: I1008 18:12:13.142503 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:12:13 crc kubenswrapper[4988]: I1008 18:12:13.142526 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:12:13Z","lastTransitionTime":"2025-10-08T18:12:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:12:13 crc kubenswrapper[4988]: I1008 18:12:13.236926 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jn2sx" Oct 08 18:12:13 crc kubenswrapper[4988]: I1008 18:12:13.236977 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 18:12:13 crc kubenswrapper[4988]: I1008 18:12:13.237113 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 18:12:13 crc kubenswrapper[4988]: E1008 18:12:13.237344 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jn2sx" podUID="171e0902-95ee-4a39-8b1b-a58704c30574" Oct 08 18:12:13 crc kubenswrapper[4988]: I1008 18:12:13.237461 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 18:12:13 crc kubenswrapper[4988]: E1008 18:12:13.237670 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 08 18:12:13 crc kubenswrapper[4988]: E1008 18:12:13.237734 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 08 18:12:13 crc kubenswrapper[4988]: E1008 18:12:13.237836 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 08 18:12:13 crc kubenswrapper[4988]: I1008 18:12:13.246480 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:12:13 crc kubenswrapper[4988]: I1008 18:12:13.246540 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:12:13 crc kubenswrapper[4988]: I1008 18:12:13.246557 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:12:13 crc kubenswrapper[4988]: I1008 18:12:13.246585 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:12:13 crc kubenswrapper[4988]: I1008 18:12:13.246604 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:12:13Z","lastTransitionTime":"2025-10-08T18:12:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:12:13 crc kubenswrapper[4988]: I1008 18:12:13.349937 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:12:13 crc kubenswrapper[4988]: I1008 18:12:13.349993 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:12:13 crc kubenswrapper[4988]: I1008 18:12:13.350011 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:12:13 crc kubenswrapper[4988]: I1008 18:12:13.350036 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:12:13 crc kubenswrapper[4988]: I1008 18:12:13.350054 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:12:13Z","lastTransitionTime":"2025-10-08T18:12:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:12:13 crc kubenswrapper[4988]: I1008 18:12:13.453911 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:12:13 crc kubenswrapper[4988]: I1008 18:12:13.453977 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:12:13 crc kubenswrapper[4988]: I1008 18:12:13.454001 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:12:13 crc kubenswrapper[4988]: I1008 18:12:13.454036 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:12:13 crc kubenswrapper[4988]: I1008 18:12:13.454059 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:12:13Z","lastTransitionTime":"2025-10-08T18:12:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:12:13 crc kubenswrapper[4988]: I1008 18:12:13.556914 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:12:13 crc kubenswrapper[4988]: I1008 18:12:13.556975 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:12:13 crc kubenswrapper[4988]: I1008 18:12:13.556996 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:12:13 crc kubenswrapper[4988]: I1008 18:12:13.557021 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:12:13 crc kubenswrapper[4988]: I1008 18:12:13.557038 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:12:13Z","lastTransitionTime":"2025-10-08T18:12:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:12:13 crc kubenswrapper[4988]: I1008 18:12:13.660695 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:12:13 crc kubenswrapper[4988]: I1008 18:12:13.660757 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:12:13 crc kubenswrapper[4988]: I1008 18:12:13.660775 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:12:13 crc kubenswrapper[4988]: I1008 18:12:13.660800 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:12:13 crc kubenswrapper[4988]: I1008 18:12:13.660820 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:12:13Z","lastTransitionTime":"2025-10-08T18:12:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:12:13 crc kubenswrapper[4988]: I1008 18:12:13.764001 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:12:13 crc kubenswrapper[4988]: I1008 18:12:13.764068 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:12:13 crc kubenswrapper[4988]: I1008 18:12:13.764085 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:12:13 crc kubenswrapper[4988]: I1008 18:12:13.764111 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:12:13 crc kubenswrapper[4988]: I1008 18:12:13.764129 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:12:13Z","lastTransitionTime":"2025-10-08T18:12:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:12:13 crc kubenswrapper[4988]: I1008 18:12:13.867214 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:12:13 crc kubenswrapper[4988]: I1008 18:12:13.867276 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:12:13 crc kubenswrapper[4988]: I1008 18:12:13.867288 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:12:13 crc kubenswrapper[4988]: I1008 18:12:13.867313 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:12:13 crc kubenswrapper[4988]: I1008 18:12:13.867329 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:12:13Z","lastTransitionTime":"2025-10-08T18:12:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:12:13 crc kubenswrapper[4988]: I1008 18:12:13.971877 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:12:13 crc kubenswrapper[4988]: I1008 18:12:13.971950 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:12:13 crc kubenswrapper[4988]: I1008 18:12:13.971974 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:12:13 crc kubenswrapper[4988]: I1008 18:12:13.972002 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:12:13 crc kubenswrapper[4988]: I1008 18:12:13.972022 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:12:13Z","lastTransitionTime":"2025-10-08T18:12:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:12:14 crc kubenswrapper[4988]: I1008 18:12:14.075761 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:12:14 crc kubenswrapper[4988]: I1008 18:12:14.075850 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:12:14 crc kubenswrapper[4988]: I1008 18:12:14.075863 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:12:14 crc kubenswrapper[4988]: I1008 18:12:14.075880 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:12:14 crc kubenswrapper[4988]: I1008 18:12:14.075891 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:12:14Z","lastTransitionTime":"2025-10-08T18:12:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:12:14 crc kubenswrapper[4988]: I1008 18:12:14.178564 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:12:14 crc kubenswrapper[4988]: I1008 18:12:14.178638 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:12:14 crc kubenswrapper[4988]: I1008 18:12:14.178662 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:12:14 crc kubenswrapper[4988]: I1008 18:12:14.178692 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:12:14 crc kubenswrapper[4988]: I1008 18:12:14.178718 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:12:14Z","lastTransitionTime":"2025-10-08T18:12:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:12:14 crc kubenswrapper[4988]: I1008 18:12:14.283341 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:12:14 crc kubenswrapper[4988]: I1008 18:12:14.283433 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:12:14 crc kubenswrapper[4988]: I1008 18:12:14.283446 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:12:14 crc kubenswrapper[4988]: I1008 18:12:14.283474 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:12:14 crc kubenswrapper[4988]: I1008 18:12:14.283487 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:12:14Z","lastTransitionTime":"2025-10-08T18:12:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:12:14 crc kubenswrapper[4988]: I1008 18:12:14.386965 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:12:14 crc kubenswrapper[4988]: I1008 18:12:14.387045 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:12:14 crc kubenswrapper[4988]: I1008 18:12:14.387064 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:12:14 crc kubenswrapper[4988]: I1008 18:12:14.387091 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:12:14 crc kubenswrapper[4988]: I1008 18:12:14.387110 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:12:14Z","lastTransitionTime":"2025-10-08T18:12:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:12:14 crc kubenswrapper[4988]: I1008 18:12:14.489792 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:12:14 crc kubenswrapper[4988]: I1008 18:12:14.489857 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:12:14 crc kubenswrapper[4988]: I1008 18:12:14.489874 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:12:14 crc kubenswrapper[4988]: I1008 18:12:14.489896 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:12:14 crc kubenswrapper[4988]: I1008 18:12:14.489908 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:12:14Z","lastTransitionTime":"2025-10-08T18:12:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:12:14 crc kubenswrapper[4988]: I1008 18:12:14.592651 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:12:14 crc kubenswrapper[4988]: I1008 18:12:14.592724 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:12:14 crc kubenswrapper[4988]: I1008 18:12:14.592742 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:12:14 crc kubenswrapper[4988]: I1008 18:12:14.592768 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:12:14 crc kubenswrapper[4988]: I1008 18:12:14.592785 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:12:14Z","lastTransitionTime":"2025-10-08T18:12:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:12:14 crc kubenswrapper[4988]: I1008 18:12:14.696243 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:12:14 crc kubenswrapper[4988]: I1008 18:12:14.696306 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:12:14 crc kubenswrapper[4988]: I1008 18:12:14.696324 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:12:14 crc kubenswrapper[4988]: I1008 18:12:14.696350 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:12:14 crc kubenswrapper[4988]: I1008 18:12:14.696371 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:12:14Z","lastTransitionTime":"2025-10-08T18:12:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:12:14 crc kubenswrapper[4988]: I1008 18:12:14.799352 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:12:14 crc kubenswrapper[4988]: I1008 18:12:14.799436 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:12:14 crc kubenswrapper[4988]: I1008 18:12:14.799446 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:12:14 crc kubenswrapper[4988]: I1008 18:12:14.799465 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:12:14 crc kubenswrapper[4988]: I1008 18:12:14.799479 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:12:14Z","lastTransitionTime":"2025-10-08T18:12:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:12:14 crc kubenswrapper[4988]: I1008 18:12:14.903508 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:12:14 crc kubenswrapper[4988]: I1008 18:12:14.903570 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:12:14 crc kubenswrapper[4988]: I1008 18:12:14.903592 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:12:14 crc kubenswrapper[4988]: I1008 18:12:14.903620 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:12:14 crc kubenswrapper[4988]: I1008 18:12:14.903646 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:12:14Z","lastTransitionTime":"2025-10-08T18:12:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:12:15 crc kubenswrapper[4988]: I1008 18:12:15.006477 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:12:15 crc kubenswrapper[4988]: I1008 18:12:15.006556 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:12:15 crc kubenswrapper[4988]: I1008 18:12:15.006569 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:12:15 crc kubenswrapper[4988]: I1008 18:12:15.006590 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:12:15 crc kubenswrapper[4988]: I1008 18:12:15.006604 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:12:15Z","lastTransitionTime":"2025-10-08T18:12:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:12:15 crc kubenswrapper[4988]: I1008 18:12:15.110669 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:12:15 crc kubenswrapper[4988]: I1008 18:12:15.110733 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:12:15 crc kubenswrapper[4988]: I1008 18:12:15.110747 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:12:15 crc kubenswrapper[4988]: I1008 18:12:15.110767 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:12:15 crc kubenswrapper[4988]: I1008 18:12:15.110781 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:12:15Z","lastTransitionTime":"2025-10-08T18:12:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:12:15 crc kubenswrapper[4988]: I1008 18:12:15.214245 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:12:15 crc kubenswrapper[4988]: I1008 18:12:15.214321 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:12:15 crc kubenswrapper[4988]: I1008 18:12:15.214345 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:12:15 crc kubenswrapper[4988]: I1008 18:12:15.214380 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:12:15 crc kubenswrapper[4988]: I1008 18:12:15.214439 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:12:15Z","lastTransitionTime":"2025-10-08T18:12:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:12:15 crc kubenswrapper[4988]: I1008 18:12:15.237752 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jn2sx" Oct 08 18:12:15 crc kubenswrapper[4988]: I1008 18:12:15.237789 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 18:12:15 crc kubenswrapper[4988]: I1008 18:12:15.237823 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 18:12:15 crc kubenswrapper[4988]: I1008 18:12:15.237853 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 18:12:15 crc kubenswrapper[4988]: E1008 18:12:15.238012 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jn2sx" podUID="171e0902-95ee-4a39-8b1b-a58704c30574" Oct 08 18:12:15 crc kubenswrapper[4988]: E1008 18:12:15.238202 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 08 18:12:15 crc kubenswrapper[4988]: E1008 18:12:15.238450 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 08 18:12:15 crc kubenswrapper[4988]: E1008 18:12:15.238574 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 08 18:12:15 crc kubenswrapper[4988]: I1008 18:12:15.256085 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"63edba53-2181-4227-9a72-a21793ad7cd0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:10:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:10:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:10:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://adbdfab7dbc007ce91c26fa7a6535ef27949ad2567f1bc6b4c52b6a3f62789f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:10:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7b7185463f2a962f3214a2524d783a54b8fbc76f385ea27001087bfb7d023561\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7b7185463f2a962f3214a2524d783a54b8fbc76f385ea27001087bfb7d023561\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:10:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:10:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:10:55Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:12:15Z is after 2025-08-24T17:21:41Z" Oct 08 18:12:15 crc kubenswrapper[4988]: I1008 18:12:15.273698 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-s4f2k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ba5e11d7-7bde-4c3d-983b-884232219652\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://824bc34a2807068d3fceb10651803ae9ae5651e940f7aed4877c544d76dae4d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hn5jf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:11:15Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-s4f2k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:12:15Z is after 2025-08-24T17:21:41Z" Oct 08 18:12:15 crc kubenswrapper[4988]: I1008 18:12:15.297417 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aef71be1-1d12-46e5-95b2-bd399040fd4d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:10:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:10:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:10:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77c68f3ecb8a0724499e495cb26ef3db5beecb140f2558d8c0b25f41928b84ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:10:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a560ee84c2405227de71fda127566213cd4f017e73d5581442ba267781b1c539\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:10:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6100accc3bd08928bc1ff88310f29fa5da37fd2d5c315c94279801c9778c4745\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:10:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b0e596d7ce4e9341c44499af79d3b3caf0e5e2c156902ea40cf2a3dca672be9e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5bd35aef24be034a0b305a40bb53ca171a4f76db3b93b4f78ae95099b0a4cde3\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-08T18:11:14Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1008 18:11:09.091034 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1008 18:11:09.091984 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3472374062/tls.crt::/tmp/serving-cert-3472374062/tls.key\\\\\\\"\\\\nI1008 18:11:14.619696 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1008 18:11:14.628837 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1008 18:11:14.628898 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1008 18:11:14.628994 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1008 18:11:14.629036 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1008 18:11:14.641318 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1008 18:11:14.641363 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1008 18:11:14.641372 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1008 18:11:14.641382 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1008 18:11:14.641415 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1008 18:11:14.641423 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1008 18:11:14.641431 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1008 18:11:14.641484 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1008 18:11:14.646451 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T18:10:58Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c79ef704fda1bbcc486f4c7afa3039a3ab38669833f0f66bafe24dd4cc679e60\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:10:58Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://28d21731af0f16d6a8ae6c6e9d0bebf0560d4a533db4c78c10bf1837890394ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://28d21731af0f16d6a8ae6c6e9d0bebf0560d4a533db4c78c10bf1837890394ef\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:10:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:10:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:10:55Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:12:15Z is after 2025-08-24T17:21:41Z" Oct 08 18:12:15 crc kubenswrapper[4988]: I1008 18:12:15.315833 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7e0ecefc-359f-4785-bb79-0343a62a25da\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:10:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:10:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:10:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://30bec350fefee3d19758b3ed735ede4e9aa085edb57aa3f3ec24249a67dc7ca2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:10:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2d850fd27d4ed65891ff69b9add65fb78535357f54819c2a8d196f5a340c58e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:10:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://90b1a9e14c82bf0f75f08fdb342eed35a503020f9db641e5e9ceff8b122cb3bd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:10:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://16f23c4c7163d32dc0df8d6a7ee17e2864fb261f8a40ecf05556774b173fc868\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:10:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:10:55Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:12:15Z is after 2025-08-24T17:21:41Z" Oct 08 18:12:15 crc kubenswrapper[4988]: I1008 18:12:15.317196 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:12:15 crc kubenswrapper[4988]: I1008 18:12:15.317275 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:12:15 crc kubenswrapper[4988]: I1008 18:12:15.317298 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:12:15 crc kubenswrapper[4988]: I1008 18:12:15.317329 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:12:15 crc kubenswrapper[4988]: I1008 18:12:15.317350 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:12:15Z","lastTransitionTime":"2025-10-08T18:12:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:12:15 crc kubenswrapper[4988]: I1008 18:12:15.339876 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e8906182d55c285b3cf095e2d9e1dac397af4618a514f11e26b44c172dd573f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:12:15Z is after 2025-08-24T17:21:41Z" Oct 08 18:12:15 crc kubenswrapper[4988]: I1008 18:12:15.360942 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-dzxs2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8f307b20-5e33-4725-99cb-210e25113532\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72ee803c6ef88524e0ce0391476f8e5a7dfbc858b095a3c2ae31bad10991ce3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q65f7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:11:16Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-dzxs2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:12:15Z is after 2025-08-24T17:21:41Z" Oct 08 18:12:15 crc kubenswrapper[4988]: I1008 18:12:15.386139 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a35a5e989ba69c8ef36ce4cea272aadf67ce20473fc257adfcfa86a2faba6ae0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a522e937cfe03fb9f2e60cb1bea07214cf63c69bdecd5ba65a2dfdba882d28dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:12:15Z is after 2025-08-24T17:21:41Z" Oct 08 18:12:15 crc kubenswrapper[4988]: I1008 18:12:15.409356 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9590ca0059f0ebbe86343cccb3475338c317382c3231e082655b61e67eece375\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:12:15Z is after 2025-08-24T17:21:41Z" Oct 08 18:12:15 crc kubenswrapper[4988]: I1008 18:12:15.420353 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:12:15 crc kubenswrapper[4988]: I1008 18:12:15.420475 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:12:15 crc kubenswrapper[4988]: I1008 18:12:15.420501 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:12:15 crc kubenswrapper[4988]: I1008 18:12:15.420535 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:12:15 crc kubenswrapper[4988]: I1008 18:12:15.420563 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:12:15Z","lastTransitionTime":"2025-10-08T18:12:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:12:15 crc kubenswrapper[4988]: I1008 18:12:15.433083 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:12:15Z is after 2025-08-24T17:21:41Z" Oct 08 18:12:15 crc kubenswrapper[4988]: I1008 18:12:15.459338 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-jnh7m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"556fb048-8afb-44a5-9ea0-2b7c89e6cf4b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bfa2f3382b23c3f788a1b31e80e8a90b3b6868870dccc46d86c3fc11ad027f52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmlt2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1a825886f105ccf9a7cc6f28a0cd012fb859f4990dade0e6aae932482146028\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1a825886f105ccf9a7cc6f28a0cd012fb859f4990dade0e6aae932482146028\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmlt2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ab2fe6e65418efa75bae72259157a2323bbb4b7d11d1d012053524e14135888\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0ab2fe6e65418efa75bae72259157a2323bbb4b7d11d1d012053524e14135888\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:11:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:11:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmlt2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://368a303309be112d9c0b7ccd6538fbaca7085d97a8ec8259f36092f07147e7cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://368a303309be112d9c0b7ccd6538fbaca7085d97a8ec8259f36092f07147e7cb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:11:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:11:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmlt2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a8ca69001dddef643f982a97c74c291e30cb9fac5b0e894e80066b885ea9151\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9a8ca69001dddef643f982a97c74c291e30cb9fac5b0e894e80066b885ea9151\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:11:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:11:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmlt2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://145303c46b48603ed1aae4b9bfe06332a3ffad28f206f8394574011a87b58cd3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://145303c46b48603ed1aae4b9bfe06332a3ffad28f206f8394574011a87b58cd3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:11:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:11:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmlt2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d08e5cfc7cbe5ec1a7a0265ed7c466363c3abeec492263170860689b46c7ff95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d08e5cfc7cbe5ec1a7a0265ed7c466363c3abeec492263170860689b46c7ff95\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:11:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:11:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmlt2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:11:16Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-jnh7m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:12:15Z is after 2025-08-24T17:21:41Z" Oct 08 18:12:15 crc kubenswrapper[4988]: I1008 18:12:15.478110 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f328a20b-5283-407e-8a93-2ab2da599f18\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8622d56fb198dadee31fb0ca8fc80d56828b6710648055f264533e2d5bdb3037\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z2gh4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3cdb2a757332bdcdc2a2f3362913fe8f06003b0597de20b72627f9badd5e30af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z2gh4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:11:16Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-nm54f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:12:15Z is after 2025-08-24T17:21:41Z" Oct 08 18:12:15 crc kubenswrapper[4988]: I1008 18:12:15.499082 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d36f620-0f26-41c7-af6e-7fb5721a0e2d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:10:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:10:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:10:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3fd3b2ca80233400d2798b035a65b833f735632efd76d1160f6bb906e07b80a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:10:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://75a2bd2757133e79c0c7f2939eb433dc45291f14808e3c3f1daaf90a239c6a71\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:10:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5cb03de583c14ad6751227c21166ac5ac4c6e11b67d5b709361ced8a0877e6d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:10:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6eb46d76cf7cfb1d4ecc4fcee398dad8904732c1664c317865864b65f5703589\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6eb46d76cf7cfb1d4ecc4fcee398dad8904732c1664c317865864b65f5703589\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:10:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:10:56Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:10:55Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:12:15Z is after 2025-08-24T17:21:41Z" Oct 08 18:12:15 crc kubenswrapper[4988]: I1008 18:12:15.519855 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:12:15Z is after 2025-08-24T17:21:41Z" Oct 08 18:12:15 crc kubenswrapper[4988]: I1008 18:12:15.525169 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:12:15 crc kubenswrapper[4988]: I1008 18:12:15.525245 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:12:15 crc kubenswrapper[4988]: I1008 18:12:15.525266 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:12:15 crc kubenswrapper[4988]: I1008 18:12:15.525295 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:12:15 crc kubenswrapper[4988]: I1008 18:12:15.525315 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:12:15Z","lastTransitionTime":"2025-10-08T18:12:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:12:15 crc kubenswrapper[4988]: I1008 18:12:15.540678 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:12:15Z is after 2025-08-24T17:21:41Z" Oct 08 18:12:15 crc kubenswrapper[4988]: I1008 18:12:15.567170 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-h9jwp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4490d846-ca32-4cd5-9202-87cc6d499fb4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:12:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:12:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c38c1b52127a970f792ef09ef1b217c979d94f7f74cc3c27e38949a204ad831f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://00b7f6db043c28efda3a09fdc77a6ccded3eae8b6a3d2852aa500789d5f01194\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-08T18:12:04Z\\\",\\\"message\\\":\\\"2025-10-08T18:11:18+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_bee291bc-08e6-4092-b05c-4207d1f7fe40\\\\n2025-10-08T18:11:18+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_bee291bc-08e6-4092-b05c-4207d1f7fe40 to /host/opt/cni/bin/\\\\n2025-10-08T18:11:18Z [verbose] multus-daemon started\\\\n2025-10-08T18:11:18Z [verbose] Readiness Indicator file check\\\\n2025-10-08T18:12:03Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T18:11:17Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:12:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7k6rk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:11:16Z\\\"}}\" for pod \"openshift-multus\"/\"multus-h9jwp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:12:15Z is after 2025-08-24T17:21:41Z" Oct 08 18:12:15 crc kubenswrapper[4988]: I1008 18:12:15.613546 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-nsc65" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f010a463-69f2-4082-acc3-161899a939c3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://48a516f6c153743bd1a21ec6945fb368a37f4d2b9d6a326b68edc096a192ed05\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dnwgr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://03406e726f817bcedb841f81959ab12808ba35a530240f9abc5fa5c608b9a6ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dnwgr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9fbbeae0409d2e47e8de9a4306f63b0d13acbfd73cd86f83303fec14352ed718\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dnwgr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://55791327a370c0009c523b86b2234e82d3794985f2eaccf1922eb90f46b2b970\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dnwgr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cee5de93bc68345f9b737857f02832f56582d58757c50a8798eec25d19b2c20d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dnwgr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d314eb248e90371fed068b1173e26d4c9f9765417a70b341e1a5b136ec17353a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dnwgr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://763353df23d26dc093a685af435811240b5ee772d7279ee9a6d2c7a48c19cd30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://763353df23d26dc093a685af435811240b5ee772d7279ee9a6d2c7a48c19cd30\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-08T18:12:09Z\\\",\\\"message\\\":\\\"ine-config-controller\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.5.16\\\\\\\", Port:9001, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI1008 18:12:09.109877 6981 model_client.go:382] Update operations generated as: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-machine-config-operator/machine-config-controller]} name:Service_openshift-machine-config-operator/machine-config-controller_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.5.16:9001:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {3f1b9878-e751-4e46-a226-ce007d2c4aa7}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1008 18:12:09.109906 6981 transact.go:42] Configuring OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-machine-config-operator/machine-config-c\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T18:12:08Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-nsc65_openshift-ovn-kubernetes(f010a463-69f2-4082-acc3-161899a939c3)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dnwgr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://74fcb812b84291763cc6e5ccb20c12aaad580ad968dda4734a0bbfe33d2770ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dnwgr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://befe684cea9bd6dfda3ec8e85bca431c26bd8f82823b0a69dc0d04319a3ca385\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://befe684cea9bd6dfda3ec8e85bca431c26bd8f82823b0a69dc0d04319a3ca385\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dnwgr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:11:16Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-nsc65\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:12:15Z is after 2025-08-24T17:21:41Z" Oct 08 18:12:15 crc kubenswrapper[4988]: I1008 18:12:15.629750 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:12:15 crc kubenswrapper[4988]: I1008 18:12:15.629810 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:12:15 crc kubenswrapper[4988]: I1008 18:12:15.629820 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:12:15 crc kubenswrapper[4988]: I1008 18:12:15.629845 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:12:15 crc kubenswrapper[4988]: I1008 18:12:15.629858 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:12:15Z","lastTransitionTime":"2025-10-08T18:12:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:12:15 crc kubenswrapper[4988]: I1008 18:12:15.632481 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-7wl8v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e433488d-db52-435f-b89c-f63cd477c5d3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76b30d183684740b647c1af11dc49712232cc8b0c9a392b491580e5ddb75ea7f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dbz6l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e55016c3e1b97d1b3a857130a5c8e2ec6da2333f786032118a61e3e2ad4c5651\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dbz6l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:11:29Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-7wl8v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:12:15Z is after 2025-08-24T17:21:41Z" Oct 08 18:12:15 crc kubenswrapper[4988]: I1008 18:12:15.647812 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-jn2sx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"171e0902-95ee-4a39-8b1b-a58704c30574\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9f4q2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9f4q2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:11:31Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-jn2sx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:12:15Z is after 2025-08-24T17:21:41Z" Oct 08 18:12:15 crc kubenswrapper[4988]: I1008 18:12:15.731747 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:12:15 crc kubenswrapper[4988]: I1008 18:12:15.731792 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:12:15 crc kubenswrapper[4988]: I1008 18:12:15.731801 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:12:15 crc kubenswrapper[4988]: I1008 18:12:15.731816 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:12:15 crc kubenswrapper[4988]: I1008 18:12:15.731825 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:12:15Z","lastTransitionTime":"2025-10-08T18:12:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:12:15 crc kubenswrapper[4988]: I1008 18:12:15.835717 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:12:15 crc kubenswrapper[4988]: I1008 18:12:15.835786 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:12:15 crc kubenswrapper[4988]: I1008 18:12:15.835800 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:12:15 crc kubenswrapper[4988]: I1008 18:12:15.835821 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:12:15 crc kubenswrapper[4988]: I1008 18:12:15.835840 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:12:15Z","lastTransitionTime":"2025-10-08T18:12:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:12:15 crc kubenswrapper[4988]: I1008 18:12:15.939562 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:12:15 crc kubenswrapper[4988]: I1008 18:12:15.939629 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:12:15 crc kubenswrapper[4988]: I1008 18:12:15.939649 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:12:15 crc kubenswrapper[4988]: I1008 18:12:15.939678 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:12:15 crc kubenswrapper[4988]: I1008 18:12:15.939697 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:12:15Z","lastTransitionTime":"2025-10-08T18:12:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:12:16 crc kubenswrapper[4988]: I1008 18:12:16.043228 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:12:16 crc kubenswrapper[4988]: I1008 18:12:16.043323 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:12:16 crc kubenswrapper[4988]: I1008 18:12:16.043353 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:12:16 crc kubenswrapper[4988]: I1008 18:12:16.043428 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:12:16 crc kubenswrapper[4988]: I1008 18:12:16.043457 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:12:16Z","lastTransitionTime":"2025-10-08T18:12:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:12:16 crc kubenswrapper[4988]: I1008 18:12:16.146611 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:12:16 crc kubenswrapper[4988]: I1008 18:12:16.146686 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:12:16 crc kubenswrapper[4988]: I1008 18:12:16.146704 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:12:16 crc kubenswrapper[4988]: I1008 18:12:16.146736 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:12:16 crc kubenswrapper[4988]: I1008 18:12:16.146755 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:12:16Z","lastTransitionTime":"2025-10-08T18:12:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:12:16 crc kubenswrapper[4988]: I1008 18:12:16.251030 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:12:16 crc kubenswrapper[4988]: I1008 18:12:16.251095 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:12:16 crc kubenswrapper[4988]: I1008 18:12:16.251115 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:12:16 crc kubenswrapper[4988]: I1008 18:12:16.251140 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:12:16 crc kubenswrapper[4988]: I1008 18:12:16.251159 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:12:16Z","lastTransitionTime":"2025-10-08T18:12:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:12:16 crc kubenswrapper[4988]: I1008 18:12:16.355727 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:12:16 crc kubenswrapper[4988]: I1008 18:12:16.355806 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:12:16 crc kubenswrapper[4988]: I1008 18:12:16.355833 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:12:16 crc kubenswrapper[4988]: I1008 18:12:16.355870 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:12:16 crc kubenswrapper[4988]: I1008 18:12:16.355894 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:12:16Z","lastTransitionTime":"2025-10-08T18:12:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:12:16 crc kubenswrapper[4988]: I1008 18:12:16.459698 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:12:16 crc kubenswrapper[4988]: I1008 18:12:16.459774 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:12:16 crc kubenswrapper[4988]: I1008 18:12:16.459796 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:12:16 crc kubenswrapper[4988]: I1008 18:12:16.459826 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:12:16 crc kubenswrapper[4988]: I1008 18:12:16.459850 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:12:16Z","lastTransitionTime":"2025-10-08T18:12:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:12:16 crc kubenswrapper[4988]: I1008 18:12:16.563143 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:12:16 crc kubenswrapper[4988]: I1008 18:12:16.563217 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:12:16 crc kubenswrapper[4988]: I1008 18:12:16.563233 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:12:16 crc kubenswrapper[4988]: I1008 18:12:16.563261 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:12:16 crc kubenswrapper[4988]: I1008 18:12:16.563277 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:12:16Z","lastTransitionTime":"2025-10-08T18:12:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:12:16 crc kubenswrapper[4988]: I1008 18:12:16.666798 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:12:16 crc kubenswrapper[4988]: I1008 18:12:16.666960 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:12:16 crc kubenswrapper[4988]: I1008 18:12:16.666982 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:12:16 crc kubenswrapper[4988]: I1008 18:12:16.667012 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:12:16 crc kubenswrapper[4988]: I1008 18:12:16.667034 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:12:16Z","lastTransitionTime":"2025-10-08T18:12:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:12:16 crc kubenswrapper[4988]: I1008 18:12:16.770312 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:12:16 crc kubenswrapper[4988]: I1008 18:12:16.770421 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:12:16 crc kubenswrapper[4988]: I1008 18:12:16.770448 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:12:16 crc kubenswrapper[4988]: I1008 18:12:16.770480 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:12:16 crc kubenswrapper[4988]: I1008 18:12:16.770499 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:12:16Z","lastTransitionTime":"2025-10-08T18:12:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:12:16 crc kubenswrapper[4988]: I1008 18:12:16.874035 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:12:16 crc kubenswrapper[4988]: I1008 18:12:16.874094 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:12:16 crc kubenswrapper[4988]: I1008 18:12:16.874106 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:12:16 crc kubenswrapper[4988]: I1008 18:12:16.874128 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:12:16 crc kubenswrapper[4988]: I1008 18:12:16.874142 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:12:16Z","lastTransitionTime":"2025-10-08T18:12:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:12:16 crc kubenswrapper[4988]: I1008 18:12:16.977598 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:12:16 crc kubenswrapper[4988]: I1008 18:12:16.977659 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:12:16 crc kubenswrapper[4988]: I1008 18:12:16.977672 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:12:16 crc kubenswrapper[4988]: I1008 18:12:16.977692 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:12:16 crc kubenswrapper[4988]: I1008 18:12:16.977706 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:12:16Z","lastTransitionTime":"2025-10-08T18:12:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:12:17 crc kubenswrapper[4988]: I1008 18:12:17.080614 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:12:17 crc kubenswrapper[4988]: I1008 18:12:17.080701 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:12:17 crc kubenswrapper[4988]: I1008 18:12:17.080718 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:12:17 crc kubenswrapper[4988]: I1008 18:12:17.080746 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:12:17 crc kubenswrapper[4988]: I1008 18:12:17.080767 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:12:17Z","lastTransitionTime":"2025-10-08T18:12:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:12:17 crc kubenswrapper[4988]: I1008 18:12:17.184233 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:12:17 crc kubenswrapper[4988]: I1008 18:12:17.184298 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:12:17 crc kubenswrapper[4988]: I1008 18:12:17.184319 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:12:17 crc kubenswrapper[4988]: I1008 18:12:17.184349 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:12:17 crc kubenswrapper[4988]: I1008 18:12:17.184369 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:12:17Z","lastTransitionTime":"2025-10-08T18:12:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:12:17 crc kubenswrapper[4988]: I1008 18:12:17.237237 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 18:12:17 crc kubenswrapper[4988]: I1008 18:12:17.237432 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jn2sx" Oct 08 18:12:17 crc kubenswrapper[4988]: E1008 18:12:17.237508 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 08 18:12:17 crc kubenswrapper[4988]: I1008 18:12:17.237439 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 18:12:17 crc kubenswrapper[4988]: I1008 18:12:17.237604 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 18:12:17 crc kubenswrapper[4988]: E1008 18:12:17.237699 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jn2sx" podUID="171e0902-95ee-4a39-8b1b-a58704c30574" Oct 08 18:12:17 crc kubenswrapper[4988]: E1008 18:12:17.237890 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 08 18:12:17 crc kubenswrapper[4988]: E1008 18:12:17.238203 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 08 18:12:17 crc kubenswrapper[4988]: I1008 18:12:17.288205 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:12:17 crc kubenswrapper[4988]: I1008 18:12:17.288281 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:12:17 crc kubenswrapper[4988]: I1008 18:12:17.288300 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:12:17 crc kubenswrapper[4988]: I1008 18:12:17.288328 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:12:17 crc kubenswrapper[4988]: I1008 18:12:17.288346 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:12:17Z","lastTransitionTime":"2025-10-08T18:12:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:12:17 crc kubenswrapper[4988]: I1008 18:12:17.391469 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:12:17 crc kubenswrapper[4988]: I1008 18:12:17.391540 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:12:17 crc kubenswrapper[4988]: I1008 18:12:17.391558 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:12:17 crc kubenswrapper[4988]: I1008 18:12:17.391585 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:12:17 crc kubenswrapper[4988]: I1008 18:12:17.391604 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:12:17Z","lastTransitionTime":"2025-10-08T18:12:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:12:17 crc kubenswrapper[4988]: I1008 18:12:17.494731 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:12:17 crc kubenswrapper[4988]: I1008 18:12:17.494801 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:12:17 crc kubenswrapper[4988]: I1008 18:12:17.494818 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:12:17 crc kubenswrapper[4988]: I1008 18:12:17.494843 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:12:17 crc kubenswrapper[4988]: I1008 18:12:17.494860 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:12:17Z","lastTransitionTime":"2025-10-08T18:12:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:12:17 crc kubenswrapper[4988]: I1008 18:12:17.598705 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:12:17 crc kubenswrapper[4988]: I1008 18:12:17.598827 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:12:17 crc kubenswrapper[4988]: I1008 18:12:17.598855 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:12:17 crc kubenswrapper[4988]: I1008 18:12:17.598887 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:12:17 crc kubenswrapper[4988]: I1008 18:12:17.598910 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:12:17Z","lastTransitionTime":"2025-10-08T18:12:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:12:17 crc kubenswrapper[4988]: I1008 18:12:17.702809 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:12:17 crc kubenswrapper[4988]: I1008 18:12:17.702882 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:12:17 crc kubenswrapper[4988]: I1008 18:12:17.702901 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:12:17 crc kubenswrapper[4988]: I1008 18:12:17.702929 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:12:17 crc kubenswrapper[4988]: I1008 18:12:17.702951 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:12:17Z","lastTransitionTime":"2025-10-08T18:12:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:12:17 crc kubenswrapper[4988]: I1008 18:12:17.806797 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:12:17 crc kubenswrapper[4988]: I1008 18:12:17.806861 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:12:17 crc kubenswrapper[4988]: I1008 18:12:17.806877 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:12:17 crc kubenswrapper[4988]: I1008 18:12:17.806906 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:12:17 crc kubenswrapper[4988]: I1008 18:12:17.806923 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:12:17Z","lastTransitionTime":"2025-10-08T18:12:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:12:17 crc kubenswrapper[4988]: I1008 18:12:17.910897 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:12:17 crc kubenswrapper[4988]: I1008 18:12:17.910968 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:12:17 crc kubenswrapper[4988]: I1008 18:12:17.910987 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:12:17 crc kubenswrapper[4988]: I1008 18:12:17.911015 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:12:17 crc kubenswrapper[4988]: I1008 18:12:17.911037 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:12:17Z","lastTransitionTime":"2025-10-08T18:12:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:12:18 crc kubenswrapper[4988]: I1008 18:12:18.014294 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:12:18 crc kubenswrapper[4988]: I1008 18:12:18.014353 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:12:18 crc kubenswrapper[4988]: I1008 18:12:18.014375 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:12:18 crc kubenswrapper[4988]: I1008 18:12:18.014438 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:12:18 crc kubenswrapper[4988]: I1008 18:12:18.014459 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:12:18Z","lastTransitionTime":"2025-10-08T18:12:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:12:18 crc kubenswrapper[4988]: I1008 18:12:18.117929 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:12:18 crc kubenswrapper[4988]: I1008 18:12:18.117990 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:12:18 crc kubenswrapper[4988]: I1008 18:12:18.118007 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:12:18 crc kubenswrapper[4988]: I1008 18:12:18.118035 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:12:18 crc kubenswrapper[4988]: I1008 18:12:18.118053 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:12:18Z","lastTransitionTime":"2025-10-08T18:12:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:12:18 crc kubenswrapper[4988]: I1008 18:12:18.220376 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:12:18 crc kubenswrapper[4988]: I1008 18:12:18.220465 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:12:18 crc kubenswrapper[4988]: I1008 18:12:18.220481 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:12:18 crc kubenswrapper[4988]: I1008 18:12:18.220508 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:12:18 crc kubenswrapper[4988]: I1008 18:12:18.220525 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:12:18Z","lastTransitionTime":"2025-10-08T18:12:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:12:18 crc kubenswrapper[4988]: I1008 18:12:18.324868 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:12:18 crc kubenswrapper[4988]: I1008 18:12:18.324945 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:12:18 crc kubenswrapper[4988]: I1008 18:12:18.324965 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:12:18 crc kubenswrapper[4988]: I1008 18:12:18.324996 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:12:18 crc kubenswrapper[4988]: I1008 18:12:18.325017 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:12:18Z","lastTransitionTime":"2025-10-08T18:12:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:12:18 crc kubenswrapper[4988]: I1008 18:12:18.428054 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:12:18 crc kubenswrapper[4988]: I1008 18:12:18.428481 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:12:18 crc kubenswrapper[4988]: I1008 18:12:18.428643 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:12:18 crc kubenswrapper[4988]: I1008 18:12:18.428789 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:12:18 crc kubenswrapper[4988]: I1008 18:12:18.428965 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:12:18Z","lastTransitionTime":"2025-10-08T18:12:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:12:18 crc kubenswrapper[4988]: I1008 18:12:18.532556 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:12:18 crc kubenswrapper[4988]: I1008 18:12:18.532630 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:12:18 crc kubenswrapper[4988]: I1008 18:12:18.532652 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:12:18 crc kubenswrapper[4988]: I1008 18:12:18.532680 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:12:18 crc kubenswrapper[4988]: I1008 18:12:18.532698 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:12:18Z","lastTransitionTime":"2025-10-08T18:12:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:12:18 crc kubenswrapper[4988]: I1008 18:12:18.636653 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:12:18 crc kubenswrapper[4988]: I1008 18:12:18.636716 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:12:18 crc kubenswrapper[4988]: I1008 18:12:18.636734 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:12:18 crc kubenswrapper[4988]: I1008 18:12:18.636760 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:12:18 crc kubenswrapper[4988]: I1008 18:12:18.636781 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:12:18Z","lastTransitionTime":"2025-10-08T18:12:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:12:18 crc kubenswrapper[4988]: I1008 18:12:18.740650 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:12:18 crc kubenswrapper[4988]: I1008 18:12:18.740723 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:12:18 crc kubenswrapper[4988]: I1008 18:12:18.740745 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:12:18 crc kubenswrapper[4988]: I1008 18:12:18.740773 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:12:18 crc kubenswrapper[4988]: I1008 18:12:18.740791 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:12:18Z","lastTransitionTime":"2025-10-08T18:12:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:12:18 crc kubenswrapper[4988]: I1008 18:12:18.844846 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:12:18 crc kubenswrapper[4988]: I1008 18:12:18.844933 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:12:18 crc kubenswrapper[4988]: I1008 18:12:18.844957 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:12:18 crc kubenswrapper[4988]: I1008 18:12:18.844998 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:12:18 crc kubenswrapper[4988]: I1008 18:12:18.845025 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:12:18Z","lastTransitionTime":"2025-10-08T18:12:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:12:18 crc kubenswrapper[4988]: I1008 18:12:18.942012 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 18:12:18 crc kubenswrapper[4988]: E1008 18:12:18.942338 4988 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 18:13:22.942288364 +0000 UTC m=+148.392131174 (durationBeforeRetry 1m4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 18:12:18 crc kubenswrapper[4988]: I1008 18:12:18.942572 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 18:12:18 crc kubenswrapper[4988]: I1008 18:12:18.942632 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 18:12:18 crc kubenswrapper[4988]: E1008 18:12:18.942779 4988 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 08 18:12:18 crc kubenswrapper[4988]: E1008 18:12:18.942860 4988 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 08 18:12:18 crc kubenswrapper[4988]: E1008 18:12:18.942872 4988 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-08 18:13:22.942850466 +0000 UTC m=+148.392693276 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 08 18:12:18 crc kubenswrapper[4988]: E1008 18:12:18.942997 4988 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-08 18:13:22.942966639 +0000 UTC m=+148.392809449 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 08 18:12:18 crc kubenswrapper[4988]: I1008 18:12:18.948629 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:12:18 crc kubenswrapper[4988]: I1008 18:12:18.948684 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:12:18 crc kubenswrapper[4988]: I1008 18:12:18.948704 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:12:18 crc kubenswrapper[4988]: I1008 18:12:18.948730 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:12:18 crc kubenswrapper[4988]: I1008 18:12:18.948748 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:12:18Z","lastTransitionTime":"2025-10-08T18:12:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:12:19 crc kubenswrapper[4988]: I1008 18:12:19.044345 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 18:12:19 crc kubenswrapper[4988]: I1008 18:12:19.044555 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 18:12:19 crc kubenswrapper[4988]: E1008 18:12:19.044745 4988 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 08 18:12:19 crc kubenswrapper[4988]: E1008 18:12:19.044809 4988 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 08 18:12:19 crc kubenswrapper[4988]: E1008 18:12:19.044833 4988 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 08 18:12:19 crc kubenswrapper[4988]: E1008 18:12:19.044868 4988 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 08 18:12:19 crc kubenswrapper[4988]: E1008 18:12:19.044912 4988 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 08 18:12:19 crc kubenswrapper[4988]: E1008 18:12:19.044938 4988 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 08 18:12:19 crc kubenswrapper[4988]: E1008 18:12:19.044948 4988 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-08 18:13:23.044910649 +0000 UTC m=+148.494753459 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 08 18:12:19 crc kubenswrapper[4988]: E1008 18:12:19.045043 4988 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-08 18:13:23.045004361 +0000 UTC m=+148.494847311 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 08 18:12:19 crc kubenswrapper[4988]: I1008 18:12:19.051851 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:12:19 crc kubenswrapper[4988]: I1008 18:12:19.051904 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:12:19 crc kubenswrapper[4988]: I1008 18:12:19.051922 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:12:19 crc kubenswrapper[4988]: I1008 18:12:19.051949 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:12:19 crc kubenswrapper[4988]: I1008 18:12:19.051966 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:12:19Z","lastTransitionTime":"2025-10-08T18:12:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:12:19 crc kubenswrapper[4988]: I1008 18:12:19.155488 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:12:19 crc kubenswrapper[4988]: I1008 18:12:19.155599 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:12:19 crc kubenswrapper[4988]: I1008 18:12:19.155629 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:12:19 crc kubenswrapper[4988]: I1008 18:12:19.155661 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:12:19 crc kubenswrapper[4988]: I1008 18:12:19.155687 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:12:19Z","lastTransitionTime":"2025-10-08T18:12:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:12:19 crc kubenswrapper[4988]: I1008 18:12:19.236845 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jn2sx" Oct 08 18:12:19 crc kubenswrapper[4988]: I1008 18:12:19.236891 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 18:12:19 crc kubenswrapper[4988]: I1008 18:12:19.237084 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 18:12:19 crc kubenswrapper[4988]: E1008 18:12:19.237273 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jn2sx" podUID="171e0902-95ee-4a39-8b1b-a58704c30574" Oct 08 18:12:19 crc kubenswrapper[4988]: I1008 18:12:19.237347 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 18:12:19 crc kubenswrapper[4988]: E1008 18:12:19.237575 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 08 18:12:19 crc kubenswrapper[4988]: E1008 18:12:19.237631 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 08 18:12:19 crc kubenswrapper[4988]: E1008 18:12:19.237718 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 08 18:12:19 crc kubenswrapper[4988]: I1008 18:12:19.259795 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:12:19 crc kubenswrapper[4988]: I1008 18:12:19.259837 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:12:19 crc kubenswrapper[4988]: I1008 18:12:19.259853 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:12:19 crc kubenswrapper[4988]: I1008 18:12:19.259878 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:12:19 crc kubenswrapper[4988]: I1008 18:12:19.259895 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:12:19Z","lastTransitionTime":"2025-10-08T18:12:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:12:19 crc kubenswrapper[4988]: I1008 18:12:19.363167 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:12:19 crc kubenswrapper[4988]: I1008 18:12:19.363608 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:12:19 crc kubenswrapper[4988]: I1008 18:12:19.363818 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:12:19 crc kubenswrapper[4988]: I1008 18:12:19.364032 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:12:19 crc kubenswrapper[4988]: I1008 18:12:19.364188 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:12:19Z","lastTransitionTime":"2025-10-08T18:12:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:12:19 crc kubenswrapper[4988]: I1008 18:12:19.467513 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:12:19 crc kubenswrapper[4988]: I1008 18:12:19.467864 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:12:19 crc kubenswrapper[4988]: I1008 18:12:19.468379 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:12:19 crc kubenswrapper[4988]: I1008 18:12:19.468789 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:12:19 crc kubenswrapper[4988]: I1008 18:12:19.469002 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:12:19Z","lastTransitionTime":"2025-10-08T18:12:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:12:19 crc kubenswrapper[4988]: I1008 18:12:19.572118 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:12:19 crc kubenswrapper[4988]: I1008 18:12:19.572171 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:12:19 crc kubenswrapper[4988]: I1008 18:12:19.572181 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:12:19 crc kubenswrapper[4988]: I1008 18:12:19.572198 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:12:19 crc kubenswrapper[4988]: I1008 18:12:19.572209 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:12:19Z","lastTransitionTime":"2025-10-08T18:12:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:12:19 crc kubenswrapper[4988]: I1008 18:12:19.676611 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:12:19 crc kubenswrapper[4988]: I1008 18:12:19.676693 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:12:19 crc kubenswrapper[4988]: I1008 18:12:19.676712 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:12:19 crc kubenswrapper[4988]: I1008 18:12:19.676744 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:12:19 crc kubenswrapper[4988]: I1008 18:12:19.676762 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:12:19Z","lastTransitionTime":"2025-10-08T18:12:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:12:19 crc kubenswrapper[4988]: I1008 18:12:19.779954 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:12:19 crc kubenswrapper[4988]: I1008 18:12:19.780024 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:12:19 crc kubenswrapper[4988]: I1008 18:12:19.780042 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:12:19 crc kubenswrapper[4988]: I1008 18:12:19.780067 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:12:19 crc kubenswrapper[4988]: I1008 18:12:19.780083 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:12:19Z","lastTransitionTime":"2025-10-08T18:12:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:12:19 crc kubenswrapper[4988]: I1008 18:12:19.884244 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:12:19 crc kubenswrapper[4988]: I1008 18:12:19.884295 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:12:19 crc kubenswrapper[4988]: I1008 18:12:19.884307 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:12:19 crc kubenswrapper[4988]: I1008 18:12:19.884328 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:12:19 crc kubenswrapper[4988]: I1008 18:12:19.884341 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:12:19Z","lastTransitionTime":"2025-10-08T18:12:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:12:19 crc kubenswrapper[4988]: I1008 18:12:19.987829 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:12:19 crc kubenswrapper[4988]: I1008 18:12:19.987906 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:12:19 crc kubenswrapper[4988]: I1008 18:12:19.987924 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:12:19 crc kubenswrapper[4988]: I1008 18:12:19.987952 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:12:19 crc kubenswrapper[4988]: I1008 18:12:19.987970 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:12:19Z","lastTransitionTime":"2025-10-08T18:12:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:12:20 crc kubenswrapper[4988]: I1008 18:12:20.091377 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:12:20 crc kubenswrapper[4988]: I1008 18:12:20.091491 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:12:20 crc kubenswrapper[4988]: I1008 18:12:20.091515 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:12:20 crc kubenswrapper[4988]: I1008 18:12:20.091549 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:12:20 crc kubenswrapper[4988]: I1008 18:12:20.091571 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:12:20Z","lastTransitionTime":"2025-10-08T18:12:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:12:20 crc kubenswrapper[4988]: I1008 18:12:20.195239 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:12:20 crc kubenswrapper[4988]: I1008 18:12:20.195305 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:12:20 crc kubenswrapper[4988]: I1008 18:12:20.195326 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:12:20 crc kubenswrapper[4988]: I1008 18:12:20.195351 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:12:20 crc kubenswrapper[4988]: I1008 18:12:20.195368 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:12:20Z","lastTransitionTime":"2025-10-08T18:12:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:12:20 crc kubenswrapper[4988]: I1008 18:12:20.298030 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:12:20 crc kubenswrapper[4988]: I1008 18:12:20.298074 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:12:20 crc kubenswrapper[4988]: I1008 18:12:20.298087 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:12:20 crc kubenswrapper[4988]: I1008 18:12:20.298104 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:12:20 crc kubenswrapper[4988]: I1008 18:12:20.298113 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:12:20Z","lastTransitionTime":"2025-10-08T18:12:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:12:20 crc kubenswrapper[4988]: I1008 18:12:20.402269 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:12:20 crc kubenswrapper[4988]: I1008 18:12:20.402328 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:12:20 crc kubenswrapper[4988]: I1008 18:12:20.402345 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:12:20 crc kubenswrapper[4988]: I1008 18:12:20.402372 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:12:20 crc kubenswrapper[4988]: I1008 18:12:20.402416 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:12:20Z","lastTransitionTime":"2025-10-08T18:12:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:12:20 crc kubenswrapper[4988]: I1008 18:12:20.507642 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:12:20 crc kubenswrapper[4988]: I1008 18:12:20.507714 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:12:20 crc kubenswrapper[4988]: I1008 18:12:20.507738 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:12:20 crc kubenswrapper[4988]: I1008 18:12:20.507772 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:12:20 crc kubenswrapper[4988]: I1008 18:12:20.507797 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:12:20Z","lastTransitionTime":"2025-10-08T18:12:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:12:20 crc kubenswrapper[4988]: I1008 18:12:20.611908 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:12:20 crc kubenswrapper[4988]: I1008 18:12:20.611981 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:12:20 crc kubenswrapper[4988]: I1008 18:12:20.612005 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:12:20 crc kubenswrapper[4988]: I1008 18:12:20.612036 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:12:20 crc kubenswrapper[4988]: I1008 18:12:20.612062 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:12:20Z","lastTransitionTime":"2025-10-08T18:12:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:12:20 crc kubenswrapper[4988]: I1008 18:12:20.715535 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:12:20 crc kubenswrapper[4988]: I1008 18:12:20.715603 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:12:20 crc kubenswrapper[4988]: I1008 18:12:20.715622 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:12:20 crc kubenswrapper[4988]: I1008 18:12:20.715654 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:12:20 crc kubenswrapper[4988]: I1008 18:12:20.715671 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:12:20Z","lastTransitionTime":"2025-10-08T18:12:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:12:20 crc kubenswrapper[4988]: I1008 18:12:20.819128 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:12:20 crc kubenswrapper[4988]: I1008 18:12:20.819191 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:12:20 crc kubenswrapper[4988]: I1008 18:12:20.819249 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:12:20 crc kubenswrapper[4988]: I1008 18:12:20.819276 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:12:20 crc kubenswrapper[4988]: I1008 18:12:20.819289 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:12:20Z","lastTransitionTime":"2025-10-08T18:12:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:12:20 crc kubenswrapper[4988]: I1008 18:12:20.923094 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:12:20 crc kubenswrapper[4988]: I1008 18:12:20.923142 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:12:20 crc kubenswrapper[4988]: I1008 18:12:20.923154 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:12:20 crc kubenswrapper[4988]: I1008 18:12:20.923177 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:12:20 crc kubenswrapper[4988]: I1008 18:12:20.923191 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:12:20Z","lastTransitionTime":"2025-10-08T18:12:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:12:21 crc kubenswrapper[4988]: I1008 18:12:21.026290 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:12:21 crc kubenswrapper[4988]: I1008 18:12:21.026435 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:12:21 crc kubenswrapper[4988]: I1008 18:12:21.026451 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:12:21 crc kubenswrapper[4988]: I1008 18:12:21.026477 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:12:21 crc kubenswrapper[4988]: I1008 18:12:21.026489 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:12:21Z","lastTransitionTime":"2025-10-08T18:12:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:12:21 crc kubenswrapper[4988]: I1008 18:12:21.129536 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:12:21 crc kubenswrapper[4988]: I1008 18:12:21.129590 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:12:21 crc kubenswrapper[4988]: I1008 18:12:21.129604 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:12:21 crc kubenswrapper[4988]: I1008 18:12:21.129624 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:12:21 crc kubenswrapper[4988]: I1008 18:12:21.129638 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:12:21Z","lastTransitionTime":"2025-10-08T18:12:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:12:21 crc kubenswrapper[4988]: I1008 18:12:21.233287 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:12:21 crc kubenswrapper[4988]: I1008 18:12:21.233370 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:12:21 crc kubenswrapper[4988]: I1008 18:12:21.233418 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:12:21 crc kubenswrapper[4988]: I1008 18:12:21.233443 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:12:21 crc kubenswrapper[4988]: I1008 18:12:21.233456 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:12:21Z","lastTransitionTime":"2025-10-08T18:12:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:12:21 crc kubenswrapper[4988]: I1008 18:12:21.237555 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jn2sx" Oct 08 18:12:21 crc kubenswrapper[4988]: I1008 18:12:21.237627 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 18:12:21 crc kubenswrapper[4988]: I1008 18:12:21.237677 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 18:12:21 crc kubenswrapper[4988]: E1008 18:12:21.237729 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jn2sx" podUID="171e0902-95ee-4a39-8b1b-a58704c30574" Oct 08 18:12:21 crc kubenswrapper[4988]: I1008 18:12:21.237835 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 18:12:21 crc kubenswrapper[4988]: E1008 18:12:21.237933 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 08 18:12:21 crc kubenswrapper[4988]: E1008 18:12:21.238054 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 08 18:12:21 crc kubenswrapper[4988]: E1008 18:12:21.238146 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 08 18:12:21 crc kubenswrapper[4988]: I1008 18:12:21.336080 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:12:21 crc kubenswrapper[4988]: I1008 18:12:21.336168 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:12:21 crc kubenswrapper[4988]: I1008 18:12:21.336187 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:12:21 crc kubenswrapper[4988]: I1008 18:12:21.336216 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:12:21 crc kubenswrapper[4988]: I1008 18:12:21.336236 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:12:21Z","lastTransitionTime":"2025-10-08T18:12:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:12:21 crc kubenswrapper[4988]: I1008 18:12:21.440105 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:12:21 crc kubenswrapper[4988]: I1008 18:12:21.440170 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:12:21 crc kubenswrapper[4988]: I1008 18:12:21.440180 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:12:21 crc kubenswrapper[4988]: I1008 18:12:21.440200 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:12:21 crc kubenswrapper[4988]: I1008 18:12:21.440211 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:12:21Z","lastTransitionTime":"2025-10-08T18:12:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:12:21 crc kubenswrapper[4988]: I1008 18:12:21.543440 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:12:21 crc kubenswrapper[4988]: I1008 18:12:21.543530 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:12:21 crc kubenswrapper[4988]: I1008 18:12:21.543553 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:12:21 crc kubenswrapper[4988]: I1008 18:12:21.543576 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:12:21 crc kubenswrapper[4988]: I1008 18:12:21.543620 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:12:21Z","lastTransitionTime":"2025-10-08T18:12:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:12:21 crc kubenswrapper[4988]: I1008 18:12:21.647057 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:12:21 crc kubenswrapper[4988]: I1008 18:12:21.647126 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:12:21 crc kubenswrapper[4988]: I1008 18:12:21.647143 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:12:21 crc kubenswrapper[4988]: I1008 18:12:21.647167 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:12:21 crc kubenswrapper[4988]: I1008 18:12:21.647185 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:12:21Z","lastTransitionTime":"2025-10-08T18:12:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:12:21 crc kubenswrapper[4988]: I1008 18:12:21.750152 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:12:21 crc kubenswrapper[4988]: I1008 18:12:21.750225 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:12:21 crc kubenswrapper[4988]: I1008 18:12:21.750255 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:12:21 crc kubenswrapper[4988]: I1008 18:12:21.750272 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:12:21 crc kubenswrapper[4988]: I1008 18:12:21.750283 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:12:21Z","lastTransitionTime":"2025-10-08T18:12:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:12:21 crc kubenswrapper[4988]: I1008 18:12:21.853612 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:12:21 crc kubenswrapper[4988]: I1008 18:12:21.853684 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:12:21 crc kubenswrapper[4988]: I1008 18:12:21.853705 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:12:21 crc kubenswrapper[4988]: I1008 18:12:21.853733 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:12:21 crc kubenswrapper[4988]: I1008 18:12:21.853751 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:12:21Z","lastTransitionTime":"2025-10-08T18:12:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:12:21 crc kubenswrapper[4988]: I1008 18:12:21.958471 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:12:21 crc kubenswrapper[4988]: I1008 18:12:21.958543 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:12:21 crc kubenswrapper[4988]: I1008 18:12:21.958569 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:12:21 crc kubenswrapper[4988]: I1008 18:12:21.958604 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:12:21 crc kubenswrapper[4988]: I1008 18:12:21.958628 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:12:21Z","lastTransitionTime":"2025-10-08T18:12:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:12:22 crc kubenswrapper[4988]: I1008 18:12:22.062119 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:12:22 crc kubenswrapper[4988]: I1008 18:12:22.062185 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:12:22 crc kubenswrapper[4988]: I1008 18:12:22.062196 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:12:22 crc kubenswrapper[4988]: I1008 18:12:22.062214 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:12:22 crc kubenswrapper[4988]: I1008 18:12:22.062227 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:12:22Z","lastTransitionTime":"2025-10-08T18:12:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:12:22 crc kubenswrapper[4988]: I1008 18:12:22.164823 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:12:22 crc kubenswrapper[4988]: I1008 18:12:22.164915 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:12:22 crc kubenswrapper[4988]: I1008 18:12:22.164934 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:12:22 crc kubenswrapper[4988]: I1008 18:12:22.164964 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:12:22 crc kubenswrapper[4988]: I1008 18:12:22.164998 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:12:22Z","lastTransitionTime":"2025-10-08T18:12:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:12:22 crc kubenswrapper[4988]: I1008 18:12:22.239027 4988 scope.go:117] "RemoveContainer" containerID="763353df23d26dc093a685af435811240b5ee772d7279ee9a6d2c7a48c19cd30" Oct 08 18:12:22 crc kubenswrapper[4988]: E1008 18:12:22.239355 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-nsc65_openshift-ovn-kubernetes(f010a463-69f2-4082-acc3-161899a939c3)\"" pod="openshift-ovn-kubernetes/ovnkube-node-nsc65" podUID="f010a463-69f2-4082-acc3-161899a939c3" Oct 08 18:12:22 crc kubenswrapper[4988]: I1008 18:12:22.268620 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:12:22 crc kubenswrapper[4988]: I1008 18:12:22.268672 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:12:22 crc kubenswrapper[4988]: I1008 18:12:22.268691 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:12:22 crc kubenswrapper[4988]: I1008 18:12:22.268714 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:12:22 crc kubenswrapper[4988]: I1008 18:12:22.268733 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:12:22Z","lastTransitionTime":"2025-10-08T18:12:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:12:22 crc kubenswrapper[4988]: I1008 18:12:22.372732 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:12:22 crc kubenswrapper[4988]: I1008 18:12:22.372834 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:12:22 crc kubenswrapper[4988]: I1008 18:12:22.372865 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:12:22 crc kubenswrapper[4988]: I1008 18:12:22.372887 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:12:22 crc kubenswrapper[4988]: I1008 18:12:22.372900 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:12:22Z","lastTransitionTime":"2025-10-08T18:12:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:12:22 crc kubenswrapper[4988]: I1008 18:12:22.476794 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:12:22 crc kubenswrapper[4988]: I1008 18:12:22.476871 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:12:22 crc kubenswrapper[4988]: I1008 18:12:22.476897 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:12:22 crc kubenswrapper[4988]: I1008 18:12:22.476927 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:12:22 crc kubenswrapper[4988]: I1008 18:12:22.476948 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:12:22Z","lastTransitionTime":"2025-10-08T18:12:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:12:22 crc kubenswrapper[4988]: I1008 18:12:22.579766 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:12:22 crc kubenswrapper[4988]: I1008 18:12:22.579846 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:12:22 crc kubenswrapper[4988]: I1008 18:12:22.579869 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:12:22 crc kubenswrapper[4988]: I1008 18:12:22.579900 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:12:22 crc kubenswrapper[4988]: I1008 18:12:22.579923 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:12:22Z","lastTransitionTime":"2025-10-08T18:12:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:12:22 crc kubenswrapper[4988]: I1008 18:12:22.683709 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:12:22 crc kubenswrapper[4988]: I1008 18:12:22.683772 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:12:22 crc kubenswrapper[4988]: I1008 18:12:22.683794 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:12:22 crc kubenswrapper[4988]: I1008 18:12:22.683824 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:12:22 crc kubenswrapper[4988]: I1008 18:12:22.683848 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:12:22Z","lastTransitionTime":"2025-10-08T18:12:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:12:22 crc kubenswrapper[4988]: I1008 18:12:22.787251 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:12:22 crc kubenswrapper[4988]: I1008 18:12:22.787307 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:12:22 crc kubenswrapper[4988]: I1008 18:12:22.787325 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:12:22 crc kubenswrapper[4988]: I1008 18:12:22.787348 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:12:22 crc kubenswrapper[4988]: I1008 18:12:22.787368 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:12:22Z","lastTransitionTime":"2025-10-08T18:12:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:12:22 crc kubenswrapper[4988]: I1008 18:12:22.889985 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:12:22 crc kubenswrapper[4988]: I1008 18:12:22.890041 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:12:22 crc kubenswrapper[4988]: I1008 18:12:22.890057 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:12:22 crc kubenswrapper[4988]: I1008 18:12:22.890113 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:12:22 crc kubenswrapper[4988]: I1008 18:12:22.890133 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:12:22Z","lastTransitionTime":"2025-10-08T18:12:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:12:22 crc kubenswrapper[4988]: I1008 18:12:22.992592 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:12:22 crc kubenswrapper[4988]: I1008 18:12:22.992667 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:12:22 crc kubenswrapper[4988]: I1008 18:12:22.992705 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:12:22 crc kubenswrapper[4988]: I1008 18:12:22.992749 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:12:22 crc kubenswrapper[4988]: I1008 18:12:22.992775 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:12:22Z","lastTransitionTime":"2025-10-08T18:12:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:12:23 crc kubenswrapper[4988]: I1008 18:12:23.095919 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:12:23 crc kubenswrapper[4988]: I1008 18:12:23.095994 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:12:23 crc kubenswrapper[4988]: I1008 18:12:23.096010 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:12:23 crc kubenswrapper[4988]: I1008 18:12:23.096036 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:12:23 crc kubenswrapper[4988]: I1008 18:12:23.096053 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:12:23Z","lastTransitionTime":"2025-10-08T18:12:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:12:23 crc kubenswrapper[4988]: I1008 18:12:23.105312 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:12:23 crc kubenswrapper[4988]: I1008 18:12:23.105372 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:12:23 crc kubenswrapper[4988]: I1008 18:12:23.105404 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:12:23 crc kubenswrapper[4988]: I1008 18:12:23.105426 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:12:23 crc kubenswrapper[4988]: I1008 18:12:23.105441 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:12:23Z","lastTransitionTime":"2025-10-08T18:12:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:12:23 crc kubenswrapper[4988]: E1008 18:12:23.126207 4988 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T18:12:23Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T18:12:23Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T18:12:23Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T18:12:23Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T18:12:23Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T18:12:23Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T18:12:23Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T18:12:23Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"95ee9fd0-8bc9-4ca5-9d79-5ee94d58bfeb\\\",\\\"systemUUID\\\":\\\"bd9489c4-fb7d-486d-b7ce-a13b6d7de36a\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:12:23Z is after 2025-08-24T17:21:41Z" Oct 08 18:12:23 crc kubenswrapper[4988]: I1008 18:12:23.131126 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:12:23 crc kubenswrapper[4988]: I1008 18:12:23.131171 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:12:23 crc kubenswrapper[4988]: I1008 18:12:23.131189 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:12:23 crc kubenswrapper[4988]: I1008 18:12:23.131212 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:12:23 crc kubenswrapper[4988]: I1008 18:12:23.131233 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:12:23Z","lastTransitionTime":"2025-10-08T18:12:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:12:23 crc kubenswrapper[4988]: E1008 18:12:23.154793 4988 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T18:12:23Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T18:12:23Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T18:12:23Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T18:12:23Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T18:12:23Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T18:12:23Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T18:12:23Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T18:12:23Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"95ee9fd0-8bc9-4ca5-9d79-5ee94d58bfeb\\\",\\\"systemUUID\\\":\\\"bd9489c4-fb7d-486d-b7ce-a13b6d7de36a\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:12:23Z is after 2025-08-24T17:21:41Z" Oct 08 18:12:23 crc kubenswrapper[4988]: I1008 18:12:23.160724 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:12:23 crc kubenswrapper[4988]: I1008 18:12:23.160836 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:12:23 crc kubenswrapper[4988]: I1008 18:12:23.160856 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:12:23 crc kubenswrapper[4988]: I1008 18:12:23.160917 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:12:23 crc kubenswrapper[4988]: I1008 18:12:23.160940 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:12:23Z","lastTransitionTime":"2025-10-08T18:12:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:12:23 crc kubenswrapper[4988]: E1008 18:12:23.186257 4988 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T18:12:23Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T18:12:23Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T18:12:23Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T18:12:23Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T18:12:23Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T18:12:23Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T18:12:23Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T18:12:23Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"95ee9fd0-8bc9-4ca5-9d79-5ee94d58bfeb\\\",\\\"systemUUID\\\":\\\"bd9489c4-fb7d-486d-b7ce-a13b6d7de36a\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:12:23Z is after 2025-08-24T17:21:41Z" Oct 08 18:12:23 crc kubenswrapper[4988]: I1008 18:12:23.192908 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:12:23 crc kubenswrapper[4988]: I1008 18:12:23.192982 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:12:23 crc kubenswrapper[4988]: I1008 18:12:23.193008 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:12:23 crc kubenswrapper[4988]: I1008 18:12:23.193046 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:12:23 crc kubenswrapper[4988]: I1008 18:12:23.193071 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:12:23Z","lastTransitionTime":"2025-10-08T18:12:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:12:23 crc kubenswrapper[4988]: E1008 18:12:23.215577 4988 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T18:12:23Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T18:12:23Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T18:12:23Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T18:12:23Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T18:12:23Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T18:12:23Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T18:12:23Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T18:12:23Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"95ee9fd0-8bc9-4ca5-9d79-5ee94d58bfeb\\\",\\\"systemUUID\\\":\\\"bd9489c4-fb7d-486d-b7ce-a13b6d7de36a\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:12:23Z is after 2025-08-24T17:21:41Z" Oct 08 18:12:23 crc kubenswrapper[4988]: I1008 18:12:23.220460 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:12:23 crc kubenswrapper[4988]: I1008 18:12:23.220533 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:12:23 crc kubenswrapper[4988]: I1008 18:12:23.220557 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:12:23 crc kubenswrapper[4988]: I1008 18:12:23.220589 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:12:23 crc kubenswrapper[4988]: I1008 18:12:23.220615 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:12:23Z","lastTransitionTime":"2025-10-08T18:12:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:12:23 crc kubenswrapper[4988]: I1008 18:12:23.236993 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 18:12:23 crc kubenswrapper[4988]: I1008 18:12:23.237007 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 18:12:23 crc kubenswrapper[4988]: E1008 18:12:23.237170 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 08 18:12:23 crc kubenswrapper[4988]: I1008 18:12:23.237196 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 18:12:23 crc kubenswrapper[4988]: E1008 18:12:23.237322 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 08 18:12:23 crc kubenswrapper[4988]: E1008 18:12:23.237425 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 08 18:12:23 crc kubenswrapper[4988]: I1008 18:12:23.237695 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jn2sx" Oct 08 18:12:23 crc kubenswrapper[4988]: E1008 18:12:23.237971 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jn2sx" podUID="171e0902-95ee-4a39-8b1b-a58704c30574" Oct 08 18:12:23 crc kubenswrapper[4988]: E1008 18:12:23.243267 4988 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T18:12:23Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T18:12:23Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T18:12:23Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T18:12:23Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T18:12:23Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T18:12:23Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T18:12:23Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T18:12:23Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"95ee9fd0-8bc9-4ca5-9d79-5ee94d58bfeb\\\",\\\"systemUUID\\\":\\\"bd9489c4-fb7d-486d-b7ce-a13b6d7de36a\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:12:23Z is after 2025-08-24T17:21:41Z" Oct 08 18:12:23 crc kubenswrapper[4988]: E1008 18:12:23.243823 4988 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Oct 08 18:12:23 crc kubenswrapper[4988]: I1008 18:12:23.246467 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:12:23 crc kubenswrapper[4988]: I1008 18:12:23.246537 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:12:23 crc kubenswrapper[4988]: I1008 18:12:23.246581 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:12:23 crc kubenswrapper[4988]: I1008 18:12:23.246615 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:12:23 crc kubenswrapper[4988]: I1008 18:12:23.246655 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:12:23Z","lastTransitionTime":"2025-10-08T18:12:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:12:23 crc kubenswrapper[4988]: I1008 18:12:23.350511 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:12:23 crc kubenswrapper[4988]: I1008 18:12:23.350584 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:12:23 crc kubenswrapper[4988]: I1008 18:12:23.350606 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:12:23 crc kubenswrapper[4988]: I1008 18:12:23.350642 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:12:23 crc kubenswrapper[4988]: I1008 18:12:23.350664 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:12:23Z","lastTransitionTime":"2025-10-08T18:12:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:12:23 crc kubenswrapper[4988]: I1008 18:12:23.453373 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:12:23 crc kubenswrapper[4988]: I1008 18:12:23.453471 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:12:23 crc kubenswrapper[4988]: I1008 18:12:23.453490 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:12:23 crc kubenswrapper[4988]: I1008 18:12:23.453521 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:12:23 crc kubenswrapper[4988]: I1008 18:12:23.453539 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:12:23Z","lastTransitionTime":"2025-10-08T18:12:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:12:23 crc kubenswrapper[4988]: I1008 18:12:23.556937 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:12:23 crc kubenswrapper[4988]: I1008 18:12:23.557003 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:12:23 crc kubenswrapper[4988]: I1008 18:12:23.557024 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:12:23 crc kubenswrapper[4988]: I1008 18:12:23.557052 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:12:23 crc kubenswrapper[4988]: I1008 18:12:23.557070 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:12:23Z","lastTransitionTime":"2025-10-08T18:12:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:12:23 crc kubenswrapper[4988]: I1008 18:12:23.660008 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:12:23 crc kubenswrapper[4988]: I1008 18:12:23.660077 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:12:23 crc kubenswrapper[4988]: I1008 18:12:23.660097 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:12:23 crc kubenswrapper[4988]: I1008 18:12:23.660126 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:12:23 crc kubenswrapper[4988]: I1008 18:12:23.660146 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:12:23Z","lastTransitionTime":"2025-10-08T18:12:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:12:23 crc kubenswrapper[4988]: I1008 18:12:23.763200 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:12:23 crc kubenswrapper[4988]: I1008 18:12:23.763273 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:12:23 crc kubenswrapper[4988]: I1008 18:12:23.763293 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:12:23 crc kubenswrapper[4988]: I1008 18:12:23.763323 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:12:23 crc kubenswrapper[4988]: I1008 18:12:23.763342 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:12:23Z","lastTransitionTime":"2025-10-08T18:12:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:12:23 crc kubenswrapper[4988]: I1008 18:12:23.866966 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:12:23 crc kubenswrapper[4988]: I1008 18:12:23.867019 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:12:23 crc kubenswrapper[4988]: I1008 18:12:23.867035 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:12:23 crc kubenswrapper[4988]: I1008 18:12:23.867056 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:12:23 crc kubenswrapper[4988]: I1008 18:12:23.867073 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:12:23Z","lastTransitionTime":"2025-10-08T18:12:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:12:23 crc kubenswrapper[4988]: I1008 18:12:23.970624 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:12:23 crc kubenswrapper[4988]: I1008 18:12:23.970750 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:12:23 crc kubenswrapper[4988]: I1008 18:12:23.970773 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:12:23 crc kubenswrapper[4988]: I1008 18:12:23.970805 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:12:23 crc kubenswrapper[4988]: I1008 18:12:23.970827 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:12:23Z","lastTransitionTime":"2025-10-08T18:12:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:12:24 crc kubenswrapper[4988]: I1008 18:12:24.074578 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:12:24 crc kubenswrapper[4988]: I1008 18:12:24.074627 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:12:24 crc kubenswrapper[4988]: I1008 18:12:24.074646 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:12:24 crc kubenswrapper[4988]: I1008 18:12:24.074669 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:12:24 crc kubenswrapper[4988]: I1008 18:12:24.074722 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:12:24Z","lastTransitionTime":"2025-10-08T18:12:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:12:24 crc kubenswrapper[4988]: I1008 18:12:24.177683 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:12:24 crc kubenswrapper[4988]: I1008 18:12:24.177751 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:12:24 crc kubenswrapper[4988]: I1008 18:12:24.177768 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:12:24 crc kubenswrapper[4988]: I1008 18:12:24.177793 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:12:24 crc kubenswrapper[4988]: I1008 18:12:24.177812 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:12:24Z","lastTransitionTime":"2025-10-08T18:12:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:12:24 crc kubenswrapper[4988]: I1008 18:12:24.281167 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:12:24 crc kubenswrapper[4988]: I1008 18:12:24.281242 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:12:24 crc kubenswrapper[4988]: I1008 18:12:24.281260 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:12:24 crc kubenswrapper[4988]: I1008 18:12:24.281287 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:12:24 crc kubenswrapper[4988]: I1008 18:12:24.281305 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:12:24Z","lastTransitionTime":"2025-10-08T18:12:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:12:24 crc kubenswrapper[4988]: I1008 18:12:24.384096 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:12:24 crc kubenswrapper[4988]: I1008 18:12:24.384149 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:12:24 crc kubenswrapper[4988]: I1008 18:12:24.384161 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:12:24 crc kubenswrapper[4988]: I1008 18:12:24.384180 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:12:24 crc kubenswrapper[4988]: I1008 18:12:24.384192 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:12:24Z","lastTransitionTime":"2025-10-08T18:12:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:12:24 crc kubenswrapper[4988]: I1008 18:12:24.487903 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:12:24 crc kubenswrapper[4988]: I1008 18:12:24.487967 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:12:24 crc kubenswrapper[4988]: I1008 18:12:24.487982 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:12:24 crc kubenswrapper[4988]: I1008 18:12:24.488005 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:12:24 crc kubenswrapper[4988]: I1008 18:12:24.488020 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:12:24Z","lastTransitionTime":"2025-10-08T18:12:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:12:24 crc kubenswrapper[4988]: I1008 18:12:24.591806 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:12:24 crc kubenswrapper[4988]: I1008 18:12:24.591879 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:12:24 crc kubenswrapper[4988]: I1008 18:12:24.591899 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:12:24 crc kubenswrapper[4988]: I1008 18:12:24.591933 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:12:24 crc kubenswrapper[4988]: I1008 18:12:24.591953 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:12:24Z","lastTransitionTime":"2025-10-08T18:12:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:12:24 crc kubenswrapper[4988]: I1008 18:12:24.694599 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:12:24 crc kubenswrapper[4988]: I1008 18:12:24.694670 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:12:24 crc kubenswrapper[4988]: I1008 18:12:24.694696 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:12:24 crc kubenswrapper[4988]: I1008 18:12:24.694728 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:12:24 crc kubenswrapper[4988]: I1008 18:12:24.694749 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:12:24Z","lastTransitionTime":"2025-10-08T18:12:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:12:24 crc kubenswrapper[4988]: I1008 18:12:24.797736 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:12:24 crc kubenswrapper[4988]: I1008 18:12:24.797799 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:12:24 crc kubenswrapper[4988]: I1008 18:12:24.797819 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:12:24 crc kubenswrapper[4988]: I1008 18:12:24.797844 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:12:24 crc kubenswrapper[4988]: I1008 18:12:24.797864 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:12:24Z","lastTransitionTime":"2025-10-08T18:12:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:12:24 crc kubenswrapper[4988]: I1008 18:12:24.900504 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:12:24 crc kubenswrapper[4988]: I1008 18:12:24.900627 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:12:24 crc kubenswrapper[4988]: I1008 18:12:24.900655 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:12:24 crc kubenswrapper[4988]: I1008 18:12:24.900687 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:12:24 crc kubenswrapper[4988]: I1008 18:12:24.900710 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:12:24Z","lastTransitionTime":"2025-10-08T18:12:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:12:25 crc kubenswrapper[4988]: I1008 18:12:25.004841 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:12:25 crc kubenswrapper[4988]: I1008 18:12:25.005043 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:12:25 crc kubenswrapper[4988]: I1008 18:12:25.005075 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:12:25 crc kubenswrapper[4988]: I1008 18:12:25.005102 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:12:25 crc kubenswrapper[4988]: I1008 18:12:25.005121 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:12:25Z","lastTransitionTime":"2025-10-08T18:12:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:12:25 crc kubenswrapper[4988]: I1008 18:12:25.108274 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:12:25 crc kubenswrapper[4988]: I1008 18:12:25.108346 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:12:25 crc kubenswrapper[4988]: I1008 18:12:25.108369 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:12:25 crc kubenswrapper[4988]: I1008 18:12:25.108444 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:12:25 crc kubenswrapper[4988]: I1008 18:12:25.108480 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:12:25Z","lastTransitionTime":"2025-10-08T18:12:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:12:25 crc kubenswrapper[4988]: I1008 18:12:25.211600 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:12:25 crc kubenswrapper[4988]: I1008 18:12:25.211649 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:12:25 crc kubenswrapper[4988]: I1008 18:12:25.211668 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:12:25 crc kubenswrapper[4988]: I1008 18:12:25.211691 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:12:25 crc kubenswrapper[4988]: I1008 18:12:25.211709 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:12:25Z","lastTransitionTime":"2025-10-08T18:12:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:12:25 crc kubenswrapper[4988]: I1008 18:12:25.237153 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 18:12:25 crc kubenswrapper[4988]: E1008 18:12:25.237344 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 08 18:12:25 crc kubenswrapper[4988]: I1008 18:12:25.237710 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 18:12:25 crc kubenswrapper[4988]: E1008 18:12:25.237858 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 08 18:12:25 crc kubenswrapper[4988]: I1008 18:12:25.238185 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jn2sx" Oct 08 18:12:25 crc kubenswrapper[4988]: E1008 18:12:25.238342 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jn2sx" podUID="171e0902-95ee-4a39-8b1b-a58704c30574" Oct 08 18:12:25 crc kubenswrapper[4988]: I1008 18:12:25.238810 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 18:12:25 crc kubenswrapper[4988]: E1008 18:12:25.238955 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 08 18:12:25 crc kubenswrapper[4988]: I1008 18:12:25.258183 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a35a5e989ba69c8ef36ce4cea272aadf67ce20473fc257adfcfa86a2faba6ae0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a522e937cfe03fb9f2e60cb1bea07214cf63c69bdecd5ba65a2dfdba882d28dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:12:25Z is after 2025-08-24T17:21:41Z" Oct 08 18:12:25 crc kubenswrapper[4988]: I1008 18:12:25.276926 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9590ca0059f0ebbe86343cccb3475338c317382c3231e082655b61e67eece375\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:12:25Z is after 2025-08-24T17:21:41Z" Oct 08 18:12:25 crc kubenswrapper[4988]: I1008 18:12:25.297974 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:12:25Z is after 2025-08-24T17:21:41Z" Oct 08 18:12:25 crc kubenswrapper[4988]: I1008 18:12:25.314584 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:12:25 crc kubenswrapper[4988]: I1008 18:12:25.314624 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:12:25 crc kubenswrapper[4988]: I1008 18:12:25.314635 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:12:25 crc kubenswrapper[4988]: I1008 18:12:25.314652 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:12:25 crc kubenswrapper[4988]: I1008 18:12:25.314663 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:12:25Z","lastTransitionTime":"2025-10-08T18:12:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:12:25 crc kubenswrapper[4988]: I1008 18:12:25.316918 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-jnh7m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"556fb048-8afb-44a5-9ea0-2b7c89e6cf4b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bfa2f3382b23c3f788a1b31e80e8a90b3b6868870dccc46d86c3fc11ad027f52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmlt2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1a825886f105ccf9a7cc6f28a0cd012fb859f4990dade0e6aae932482146028\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1a825886f105ccf9a7cc6f28a0cd012fb859f4990dade0e6aae932482146028\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmlt2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ab2fe6e65418efa75bae72259157a2323bbb4b7d11d1d012053524e14135888\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0ab2fe6e65418efa75bae72259157a2323bbb4b7d11d1d012053524e14135888\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:11:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:11:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmlt2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://368a303309be112d9c0b7ccd6538fbaca7085d97a8ec8259f36092f07147e7cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://368a303309be112d9c0b7ccd6538fbaca7085d97a8ec8259f36092f07147e7cb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:11:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:11:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmlt2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a8ca69001dddef643f982a97c74c291e30cb9fac5b0e894e80066b885ea9151\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9a8ca69001dddef643f982a97c74c291e30cb9fac5b0e894e80066b885ea9151\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:11:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:11:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmlt2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://145303c46b48603ed1aae4b9bfe06332a3ffad28f206f8394574011a87b58cd3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://145303c46b48603ed1aae4b9bfe06332a3ffad28f206f8394574011a87b58cd3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:11:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:11:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmlt2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d08e5cfc7cbe5ec1a7a0265ed7c466363c3abeec492263170860689b46c7ff95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d08e5cfc7cbe5ec1a7a0265ed7c466363c3abeec492263170860689b46c7ff95\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:11:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:11:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmlt2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:11:16Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-jnh7m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:12:25Z is after 2025-08-24T17:21:41Z" Oct 08 18:12:25 crc kubenswrapper[4988]: I1008 18:12:25.332252 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f328a20b-5283-407e-8a93-2ab2da599f18\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8622d56fb198dadee31fb0ca8fc80d56828b6710648055f264533e2d5bdb3037\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z2gh4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3cdb2a757332bdcdc2a2f3362913fe8f06003b0597de20b72627f9badd5e30af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z2gh4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:11:16Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-nm54f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:12:25Z is after 2025-08-24T17:21:41Z" Oct 08 18:12:25 crc kubenswrapper[4988]: I1008 18:12:25.349064 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:12:25Z is after 2025-08-24T17:21:41Z" Oct 08 18:12:25 crc kubenswrapper[4988]: I1008 18:12:25.368724 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:12:25Z is after 2025-08-24T17:21:41Z" Oct 08 18:12:25 crc kubenswrapper[4988]: I1008 18:12:25.385613 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-h9jwp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4490d846-ca32-4cd5-9202-87cc6d499fb4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:12:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:12:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c38c1b52127a970f792ef09ef1b217c979d94f7f74cc3c27e38949a204ad831f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://00b7f6db043c28efda3a09fdc77a6ccded3eae8b6a3d2852aa500789d5f01194\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-08T18:12:04Z\\\",\\\"message\\\":\\\"2025-10-08T18:11:18+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_bee291bc-08e6-4092-b05c-4207d1f7fe40\\\\n2025-10-08T18:11:18+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_bee291bc-08e6-4092-b05c-4207d1f7fe40 to /host/opt/cni/bin/\\\\n2025-10-08T18:11:18Z [verbose] multus-daemon started\\\\n2025-10-08T18:11:18Z [verbose] Readiness Indicator file check\\\\n2025-10-08T18:12:03Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T18:11:17Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:12:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7k6rk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:11:16Z\\\"}}\" for pod \"openshift-multus\"/\"multus-h9jwp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:12:25Z is after 2025-08-24T17:21:41Z" Oct 08 18:12:25 crc kubenswrapper[4988]: I1008 18:12:25.407182 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-nsc65" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f010a463-69f2-4082-acc3-161899a939c3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://48a516f6c153743bd1a21ec6945fb368a37f4d2b9d6a326b68edc096a192ed05\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dnwgr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://03406e726f817bcedb841f81959ab12808ba35a530240f9abc5fa5c608b9a6ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dnwgr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9fbbeae0409d2e47e8de9a4306f63b0d13acbfd73cd86f83303fec14352ed718\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dnwgr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://55791327a370c0009c523b86b2234e82d3794985f2eaccf1922eb90f46b2b970\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dnwgr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cee5de93bc68345f9b737857f02832f56582d58757c50a8798eec25d19b2c20d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dnwgr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d314eb248e90371fed068b1173e26d4c9f9765417a70b341e1a5b136ec17353a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dnwgr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://763353df23d26dc093a685af435811240b5ee772d7279ee9a6d2c7a48c19cd30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://763353df23d26dc093a685af435811240b5ee772d7279ee9a6d2c7a48c19cd30\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-08T18:12:09Z\\\",\\\"message\\\":\\\"ine-config-controller\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.5.16\\\\\\\", Port:9001, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI1008 18:12:09.109877 6981 model_client.go:382] Update operations generated as: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-machine-config-operator/machine-config-controller]} name:Service_openshift-machine-config-operator/machine-config-controller_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.5.16:9001:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {3f1b9878-e751-4e46-a226-ce007d2c4aa7}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1008 18:12:09.109906 6981 transact.go:42] Configuring OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-machine-config-operator/machine-config-c\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T18:12:08Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-nsc65_openshift-ovn-kubernetes(f010a463-69f2-4082-acc3-161899a939c3)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dnwgr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://74fcb812b84291763cc6e5ccb20c12aaad580ad968dda4734a0bbfe33d2770ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dnwgr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://befe684cea9bd6dfda3ec8e85bca431c26bd8f82823b0a69dc0d04319a3ca385\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://befe684cea9bd6dfda3ec8e85bca431c26bd8f82823b0a69dc0d04319a3ca385\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dnwgr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:11:16Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-nsc65\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:12:25Z is after 2025-08-24T17:21:41Z" Oct 08 18:12:25 crc kubenswrapper[4988]: I1008 18:12:25.416592 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:12:25 crc kubenswrapper[4988]: I1008 18:12:25.416643 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:12:25 crc kubenswrapper[4988]: I1008 18:12:25.416726 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:12:25 crc kubenswrapper[4988]: I1008 18:12:25.416784 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:12:25 crc kubenswrapper[4988]: I1008 18:12:25.416804 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:12:25Z","lastTransitionTime":"2025-10-08T18:12:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:12:25 crc kubenswrapper[4988]: I1008 18:12:25.422709 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-7wl8v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e433488d-db52-435f-b89c-f63cd477c5d3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76b30d183684740b647c1af11dc49712232cc8b0c9a392b491580e5ddb75ea7f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dbz6l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e55016c3e1b97d1b3a857130a5c8e2ec6da2333f786032118a61e3e2ad4c5651\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dbz6l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:11:29Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-7wl8v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:12:25Z is after 2025-08-24T17:21:41Z" Oct 08 18:12:25 crc kubenswrapper[4988]: I1008 18:12:25.438769 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-jn2sx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"171e0902-95ee-4a39-8b1b-a58704c30574\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9f4q2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9f4q2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:11:31Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-jn2sx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:12:25Z is after 2025-08-24T17:21:41Z" Oct 08 18:12:25 crc kubenswrapper[4988]: I1008 18:12:25.455352 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d36f620-0f26-41c7-af6e-7fb5721a0e2d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:10:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:10:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:10:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3fd3b2ca80233400d2798b035a65b833f735632efd76d1160f6bb906e07b80a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:10:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://75a2bd2757133e79c0c7f2939eb433dc45291f14808e3c3f1daaf90a239c6a71\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:10:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5cb03de583c14ad6751227c21166ac5ac4c6e11b67d5b709361ced8a0877e6d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:10:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6eb46d76cf7cfb1d4ecc4fcee398dad8904732c1664c317865864b65f5703589\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6eb46d76cf7cfb1d4ecc4fcee398dad8904732c1664c317865864b65f5703589\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:10:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:10:56Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:10:55Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:12:25Z is after 2025-08-24T17:21:41Z" Oct 08 18:12:25 crc kubenswrapper[4988]: I1008 18:12:25.468695 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-s4f2k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ba5e11d7-7bde-4c3d-983b-884232219652\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://824bc34a2807068d3fceb10651803ae9ae5651e940f7aed4877c544d76dae4d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hn5jf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:11:15Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-s4f2k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:12:25Z is after 2025-08-24T17:21:41Z" Oct 08 18:12:25 crc kubenswrapper[4988]: I1008 18:12:25.482016 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"63edba53-2181-4227-9a72-a21793ad7cd0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:10:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:10:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:10:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://adbdfab7dbc007ce91c26fa7a6535ef27949ad2567f1bc6b4c52b6a3f62789f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:10:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7b7185463f2a962f3214a2524d783a54b8fbc76f385ea27001087bfb7d023561\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7b7185463f2a962f3214a2524d783a54b8fbc76f385ea27001087bfb7d023561\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:10:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:10:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:10:55Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:12:25Z is after 2025-08-24T17:21:41Z" Oct 08 18:12:25 crc kubenswrapper[4988]: I1008 18:12:25.498456 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7e0ecefc-359f-4785-bb79-0343a62a25da\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:10:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:10:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:10:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://30bec350fefee3d19758b3ed735ede4e9aa085edb57aa3f3ec24249a67dc7ca2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:10:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2d850fd27d4ed65891ff69b9add65fb78535357f54819c2a8d196f5a340c58e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:10:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://90b1a9e14c82bf0f75f08fdb342eed35a503020f9db641e5e9ceff8b122cb3bd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:10:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://16f23c4c7163d32dc0df8d6a7ee17e2864fb261f8a40ecf05556774b173fc868\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:10:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:10:55Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:12:25Z is after 2025-08-24T17:21:41Z" Oct 08 18:12:25 crc kubenswrapper[4988]: I1008 18:12:25.520640 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:12:25 crc kubenswrapper[4988]: I1008 18:12:25.520763 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:12:25 crc kubenswrapper[4988]: I1008 18:12:25.520837 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:12:25 crc kubenswrapper[4988]: I1008 18:12:25.520878 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:12:25 crc kubenswrapper[4988]: I1008 18:12:25.520951 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:12:25Z","lastTransitionTime":"2025-10-08T18:12:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:12:25 crc kubenswrapper[4988]: I1008 18:12:25.521170 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e8906182d55c285b3cf095e2d9e1dac397af4618a514f11e26b44c172dd573f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:12:25Z is after 2025-08-24T17:21:41Z" Oct 08 18:12:25 crc kubenswrapper[4988]: I1008 18:12:25.539636 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-dzxs2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8f307b20-5e33-4725-99cb-210e25113532\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72ee803c6ef88524e0ce0391476f8e5a7dfbc858b095a3c2ae31bad10991ce3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q65f7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:11:16Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-dzxs2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:12:25Z is after 2025-08-24T17:21:41Z" Oct 08 18:12:25 crc kubenswrapper[4988]: I1008 18:12:25.557145 4988 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aef71be1-1d12-46e5-95b2-bd399040fd4d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:10:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:10:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:11:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:10:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77c68f3ecb8a0724499e495cb26ef3db5beecb140f2558d8c0b25f41928b84ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:10:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a560ee84c2405227de71fda127566213cd4f017e73d5581442ba267781b1c539\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:10:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6100accc3bd08928bc1ff88310f29fa5da37fd2d5c315c94279801c9778c4745\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:10:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b0e596d7ce4e9341c44499af79d3b3caf0e5e2c156902ea40cf2a3dca672be9e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5bd35aef24be034a0b305a40bb53ca171a4f76db3b93b4f78ae95099b0a4cde3\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-08T18:11:14Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1008 18:11:09.091034 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1008 18:11:09.091984 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3472374062/tls.crt::/tmp/serving-cert-3472374062/tls.key\\\\\\\"\\\\nI1008 18:11:14.619696 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1008 18:11:14.628837 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1008 18:11:14.628898 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1008 18:11:14.628994 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1008 18:11:14.629036 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1008 18:11:14.641318 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1008 18:11:14.641363 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1008 18:11:14.641372 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1008 18:11:14.641382 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1008 18:11:14.641415 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1008 18:11:14.641423 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1008 18:11:14.641431 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1008 18:11:14.641484 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1008 18:11:14.646451 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T18:10:58Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c79ef704fda1bbcc486f4c7afa3039a3ab38669833f0f66bafe24dd4cc679e60\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:10:58Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://28d21731af0f16d6a8ae6c6e9d0bebf0560d4a533db4c78c10bf1837890394ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://28d21731af0f16d6a8ae6c6e9d0bebf0560d4a533db4c78c10bf1837890394ef\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:10:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:10:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:10:55Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:12:25Z is after 2025-08-24T17:21:41Z" Oct 08 18:12:25 crc kubenswrapper[4988]: I1008 18:12:25.623966 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:12:25 crc kubenswrapper[4988]: I1008 18:12:25.624034 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:12:25 crc kubenswrapper[4988]: I1008 18:12:25.624047 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:12:25 crc kubenswrapper[4988]: I1008 18:12:25.624070 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:12:25 crc kubenswrapper[4988]: I1008 18:12:25.624083 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:12:25Z","lastTransitionTime":"2025-10-08T18:12:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:12:25 crc kubenswrapper[4988]: I1008 18:12:25.727933 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:12:25 crc kubenswrapper[4988]: I1008 18:12:25.728011 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:12:25 crc kubenswrapper[4988]: I1008 18:12:25.728047 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:12:25 crc kubenswrapper[4988]: I1008 18:12:25.728077 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:12:25 crc kubenswrapper[4988]: I1008 18:12:25.728099 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:12:25Z","lastTransitionTime":"2025-10-08T18:12:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:12:25 crc kubenswrapper[4988]: I1008 18:12:25.830704 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:12:25 crc kubenswrapper[4988]: I1008 18:12:25.830746 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:12:25 crc kubenswrapper[4988]: I1008 18:12:25.830755 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:12:25 crc kubenswrapper[4988]: I1008 18:12:25.830773 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:12:25 crc kubenswrapper[4988]: I1008 18:12:25.830787 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:12:25Z","lastTransitionTime":"2025-10-08T18:12:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:12:25 crc kubenswrapper[4988]: I1008 18:12:25.933415 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:12:25 crc kubenswrapper[4988]: I1008 18:12:25.933482 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:12:25 crc kubenswrapper[4988]: I1008 18:12:25.933501 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:12:25 crc kubenswrapper[4988]: I1008 18:12:25.933530 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:12:25 crc kubenswrapper[4988]: I1008 18:12:25.933549 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:12:25Z","lastTransitionTime":"2025-10-08T18:12:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:12:26 crc kubenswrapper[4988]: I1008 18:12:26.036946 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:12:26 crc kubenswrapper[4988]: I1008 18:12:26.037006 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:12:26 crc kubenswrapper[4988]: I1008 18:12:26.037020 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:12:26 crc kubenswrapper[4988]: I1008 18:12:26.037039 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:12:26 crc kubenswrapper[4988]: I1008 18:12:26.037052 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:12:26Z","lastTransitionTime":"2025-10-08T18:12:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:12:26 crc kubenswrapper[4988]: I1008 18:12:26.140361 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:12:26 crc kubenswrapper[4988]: I1008 18:12:26.140455 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:12:26 crc kubenswrapper[4988]: I1008 18:12:26.140473 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:12:26 crc kubenswrapper[4988]: I1008 18:12:26.140501 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:12:26 crc kubenswrapper[4988]: I1008 18:12:26.140520 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:12:26Z","lastTransitionTime":"2025-10-08T18:12:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:12:26 crc kubenswrapper[4988]: I1008 18:12:26.243878 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:12:26 crc kubenswrapper[4988]: I1008 18:12:26.243925 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:12:26 crc kubenswrapper[4988]: I1008 18:12:26.243942 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:12:26 crc kubenswrapper[4988]: I1008 18:12:26.243962 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:12:26 crc kubenswrapper[4988]: I1008 18:12:26.243977 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:12:26Z","lastTransitionTime":"2025-10-08T18:12:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:12:26 crc kubenswrapper[4988]: I1008 18:12:26.347203 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:12:26 crc kubenswrapper[4988]: I1008 18:12:26.347263 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:12:26 crc kubenswrapper[4988]: I1008 18:12:26.347283 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:12:26 crc kubenswrapper[4988]: I1008 18:12:26.347309 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:12:26 crc kubenswrapper[4988]: I1008 18:12:26.347331 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:12:26Z","lastTransitionTime":"2025-10-08T18:12:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:12:26 crc kubenswrapper[4988]: I1008 18:12:26.450799 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:12:26 crc kubenswrapper[4988]: I1008 18:12:26.450856 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:12:26 crc kubenswrapper[4988]: I1008 18:12:26.450874 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:12:26 crc kubenswrapper[4988]: I1008 18:12:26.450900 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:12:26 crc kubenswrapper[4988]: I1008 18:12:26.450917 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:12:26Z","lastTransitionTime":"2025-10-08T18:12:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:12:26 crc kubenswrapper[4988]: I1008 18:12:26.553617 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:12:26 crc kubenswrapper[4988]: I1008 18:12:26.553652 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:12:26 crc kubenswrapper[4988]: I1008 18:12:26.553663 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:12:26 crc kubenswrapper[4988]: I1008 18:12:26.553683 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:12:26 crc kubenswrapper[4988]: I1008 18:12:26.553694 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:12:26Z","lastTransitionTime":"2025-10-08T18:12:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:12:26 crc kubenswrapper[4988]: I1008 18:12:26.656748 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:12:26 crc kubenswrapper[4988]: I1008 18:12:26.656822 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:12:26 crc kubenswrapper[4988]: I1008 18:12:26.656843 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:12:26 crc kubenswrapper[4988]: I1008 18:12:26.656869 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:12:26 crc kubenswrapper[4988]: I1008 18:12:26.656887 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:12:26Z","lastTransitionTime":"2025-10-08T18:12:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:12:26 crc kubenswrapper[4988]: I1008 18:12:26.760613 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:12:26 crc kubenswrapper[4988]: I1008 18:12:26.760694 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:12:26 crc kubenswrapper[4988]: I1008 18:12:26.760717 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:12:26 crc kubenswrapper[4988]: I1008 18:12:26.760744 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:12:26 crc kubenswrapper[4988]: I1008 18:12:26.760762 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:12:26Z","lastTransitionTime":"2025-10-08T18:12:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:12:26 crc kubenswrapper[4988]: I1008 18:12:26.864091 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:12:26 crc kubenswrapper[4988]: I1008 18:12:26.864263 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:12:26 crc kubenswrapper[4988]: I1008 18:12:26.864290 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:12:26 crc kubenswrapper[4988]: I1008 18:12:26.864326 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:12:26 crc kubenswrapper[4988]: I1008 18:12:26.864349 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:12:26Z","lastTransitionTime":"2025-10-08T18:12:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:12:26 crc kubenswrapper[4988]: I1008 18:12:26.968044 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:12:26 crc kubenswrapper[4988]: I1008 18:12:26.968109 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:12:26 crc kubenswrapper[4988]: I1008 18:12:26.968132 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:12:26 crc kubenswrapper[4988]: I1008 18:12:26.968162 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:12:26 crc kubenswrapper[4988]: I1008 18:12:26.968182 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:12:26Z","lastTransitionTime":"2025-10-08T18:12:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:12:27 crc kubenswrapper[4988]: I1008 18:12:27.071812 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:12:27 crc kubenswrapper[4988]: I1008 18:12:27.071879 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:12:27 crc kubenswrapper[4988]: I1008 18:12:27.071899 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:12:27 crc kubenswrapper[4988]: I1008 18:12:27.071926 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:12:27 crc kubenswrapper[4988]: I1008 18:12:27.071946 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:12:27Z","lastTransitionTime":"2025-10-08T18:12:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:12:27 crc kubenswrapper[4988]: I1008 18:12:27.175637 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:12:27 crc kubenswrapper[4988]: I1008 18:12:27.175709 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:12:27 crc kubenswrapper[4988]: I1008 18:12:27.175723 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:12:27 crc kubenswrapper[4988]: I1008 18:12:27.175746 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:12:27 crc kubenswrapper[4988]: I1008 18:12:27.175760 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:12:27Z","lastTransitionTime":"2025-10-08T18:12:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:12:27 crc kubenswrapper[4988]: I1008 18:12:27.237618 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 18:12:27 crc kubenswrapper[4988]: I1008 18:12:27.237658 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jn2sx" Oct 08 18:12:27 crc kubenswrapper[4988]: E1008 18:12:27.237805 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 08 18:12:27 crc kubenswrapper[4988]: I1008 18:12:27.237897 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 18:12:27 crc kubenswrapper[4988]: E1008 18:12:27.238032 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jn2sx" podUID="171e0902-95ee-4a39-8b1b-a58704c30574" Oct 08 18:12:27 crc kubenswrapper[4988]: E1008 18:12:27.238157 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 08 18:12:27 crc kubenswrapper[4988]: I1008 18:12:27.241522 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 18:12:27 crc kubenswrapper[4988]: E1008 18:12:27.241727 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 08 18:12:27 crc kubenswrapper[4988]: I1008 18:12:27.278352 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:12:27 crc kubenswrapper[4988]: I1008 18:12:27.278432 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:12:27 crc kubenswrapper[4988]: I1008 18:12:27.278445 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:12:27 crc kubenswrapper[4988]: I1008 18:12:27.278462 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:12:27 crc kubenswrapper[4988]: I1008 18:12:27.278473 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:12:27Z","lastTransitionTime":"2025-10-08T18:12:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:12:27 crc kubenswrapper[4988]: I1008 18:12:27.382431 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:12:27 crc kubenswrapper[4988]: I1008 18:12:27.382478 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:12:27 crc kubenswrapper[4988]: I1008 18:12:27.382491 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:12:27 crc kubenswrapper[4988]: I1008 18:12:27.382512 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:12:27 crc kubenswrapper[4988]: I1008 18:12:27.382527 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:12:27Z","lastTransitionTime":"2025-10-08T18:12:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:12:27 crc kubenswrapper[4988]: I1008 18:12:27.485606 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:12:27 crc kubenswrapper[4988]: I1008 18:12:27.485672 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:12:27 crc kubenswrapper[4988]: I1008 18:12:27.485689 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:12:27 crc kubenswrapper[4988]: I1008 18:12:27.485714 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:12:27 crc kubenswrapper[4988]: I1008 18:12:27.485732 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:12:27Z","lastTransitionTime":"2025-10-08T18:12:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:12:27 crc kubenswrapper[4988]: I1008 18:12:27.588715 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:12:27 crc kubenswrapper[4988]: I1008 18:12:27.588796 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:12:27 crc kubenswrapper[4988]: I1008 18:12:27.588816 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:12:27 crc kubenswrapper[4988]: I1008 18:12:27.588849 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:12:27 crc kubenswrapper[4988]: I1008 18:12:27.588867 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:12:27Z","lastTransitionTime":"2025-10-08T18:12:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:12:27 crc kubenswrapper[4988]: I1008 18:12:27.692728 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:12:27 crc kubenswrapper[4988]: I1008 18:12:27.692792 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:12:27 crc kubenswrapper[4988]: I1008 18:12:27.692814 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:12:27 crc kubenswrapper[4988]: I1008 18:12:27.692839 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:12:27 crc kubenswrapper[4988]: I1008 18:12:27.692858 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:12:27Z","lastTransitionTime":"2025-10-08T18:12:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:12:27 crc kubenswrapper[4988]: I1008 18:12:27.795906 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:12:27 crc kubenswrapper[4988]: I1008 18:12:27.795996 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:12:27 crc kubenswrapper[4988]: I1008 18:12:27.796028 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:12:27 crc kubenswrapper[4988]: I1008 18:12:27.796061 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:12:27 crc kubenswrapper[4988]: I1008 18:12:27.796082 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:12:27Z","lastTransitionTime":"2025-10-08T18:12:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:12:27 crc kubenswrapper[4988]: I1008 18:12:27.898724 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:12:27 crc kubenswrapper[4988]: I1008 18:12:27.898815 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:12:27 crc kubenswrapper[4988]: I1008 18:12:27.898844 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:12:27 crc kubenswrapper[4988]: I1008 18:12:27.898876 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:12:27 crc kubenswrapper[4988]: I1008 18:12:27.898900 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:12:27Z","lastTransitionTime":"2025-10-08T18:12:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:12:28 crc kubenswrapper[4988]: I1008 18:12:28.002350 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:12:28 crc kubenswrapper[4988]: I1008 18:12:28.002445 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:12:28 crc kubenswrapper[4988]: I1008 18:12:28.002465 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:12:28 crc kubenswrapper[4988]: I1008 18:12:28.002640 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:12:28 crc kubenswrapper[4988]: I1008 18:12:28.002668 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:12:28Z","lastTransitionTime":"2025-10-08T18:12:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:12:28 crc kubenswrapper[4988]: I1008 18:12:28.105574 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:12:28 crc kubenswrapper[4988]: I1008 18:12:28.105639 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:12:28 crc kubenswrapper[4988]: I1008 18:12:28.105658 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:12:28 crc kubenswrapper[4988]: I1008 18:12:28.105683 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:12:28 crc kubenswrapper[4988]: I1008 18:12:28.105702 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:12:28Z","lastTransitionTime":"2025-10-08T18:12:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:12:28 crc kubenswrapper[4988]: I1008 18:12:28.209596 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:12:28 crc kubenswrapper[4988]: I1008 18:12:28.209684 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:12:28 crc kubenswrapper[4988]: I1008 18:12:28.209702 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:12:28 crc kubenswrapper[4988]: I1008 18:12:28.209733 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:12:28 crc kubenswrapper[4988]: I1008 18:12:28.209751 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:12:28Z","lastTransitionTime":"2025-10-08T18:12:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:12:28 crc kubenswrapper[4988]: I1008 18:12:28.313840 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:12:28 crc kubenswrapper[4988]: I1008 18:12:28.313917 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:12:28 crc kubenswrapper[4988]: I1008 18:12:28.313940 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:12:28 crc kubenswrapper[4988]: I1008 18:12:28.313978 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:12:28 crc kubenswrapper[4988]: I1008 18:12:28.314004 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:12:28Z","lastTransitionTime":"2025-10-08T18:12:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:12:28 crc kubenswrapper[4988]: I1008 18:12:28.417672 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:12:28 crc kubenswrapper[4988]: I1008 18:12:28.417718 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:12:28 crc kubenswrapper[4988]: I1008 18:12:28.417730 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:12:28 crc kubenswrapper[4988]: I1008 18:12:28.417751 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:12:28 crc kubenswrapper[4988]: I1008 18:12:28.417764 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:12:28Z","lastTransitionTime":"2025-10-08T18:12:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:12:28 crc kubenswrapper[4988]: I1008 18:12:28.522563 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:12:28 crc kubenswrapper[4988]: I1008 18:12:28.522636 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:12:28 crc kubenswrapper[4988]: I1008 18:12:28.522654 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:12:28 crc kubenswrapper[4988]: I1008 18:12:28.522683 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:12:28 crc kubenswrapper[4988]: I1008 18:12:28.522703 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:12:28Z","lastTransitionTime":"2025-10-08T18:12:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:12:28 crc kubenswrapper[4988]: I1008 18:12:28.626817 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:12:28 crc kubenswrapper[4988]: I1008 18:12:28.626907 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:12:28 crc kubenswrapper[4988]: I1008 18:12:28.626933 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:12:28 crc kubenswrapper[4988]: I1008 18:12:28.626968 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:12:28 crc kubenswrapper[4988]: I1008 18:12:28.626986 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:12:28Z","lastTransitionTime":"2025-10-08T18:12:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:12:28 crc kubenswrapper[4988]: I1008 18:12:28.729958 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:12:28 crc kubenswrapper[4988]: I1008 18:12:28.730164 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:12:28 crc kubenswrapper[4988]: I1008 18:12:28.730204 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:12:28 crc kubenswrapper[4988]: I1008 18:12:28.730244 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:12:28 crc kubenswrapper[4988]: I1008 18:12:28.730271 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:12:28Z","lastTransitionTime":"2025-10-08T18:12:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:12:28 crc kubenswrapper[4988]: I1008 18:12:28.833483 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:12:28 crc kubenswrapper[4988]: I1008 18:12:28.833547 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:12:28 crc kubenswrapper[4988]: I1008 18:12:28.833567 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:12:28 crc kubenswrapper[4988]: I1008 18:12:28.833595 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:12:28 crc kubenswrapper[4988]: I1008 18:12:28.833616 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:12:28Z","lastTransitionTime":"2025-10-08T18:12:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:12:28 crc kubenswrapper[4988]: I1008 18:12:28.938041 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:12:28 crc kubenswrapper[4988]: I1008 18:12:28.938126 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:12:28 crc kubenswrapper[4988]: I1008 18:12:28.938145 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:12:28 crc kubenswrapper[4988]: I1008 18:12:28.938171 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:12:28 crc kubenswrapper[4988]: I1008 18:12:28.938189 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:12:28Z","lastTransitionTime":"2025-10-08T18:12:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:12:29 crc kubenswrapper[4988]: I1008 18:12:29.041675 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:12:29 crc kubenswrapper[4988]: I1008 18:12:29.041736 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:12:29 crc kubenswrapper[4988]: I1008 18:12:29.041753 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:12:29 crc kubenswrapper[4988]: I1008 18:12:29.041777 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:12:29 crc kubenswrapper[4988]: I1008 18:12:29.041793 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:12:29Z","lastTransitionTime":"2025-10-08T18:12:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:12:29 crc kubenswrapper[4988]: I1008 18:12:29.145774 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:12:29 crc kubenswrapper[4988]: I1008 18:12:29.145834 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:12:29 crc kubenswrapper[4988]: I1008 18:12:29.145848 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:12:29 crc kubenswrapper[4988]: I1008 18:12:29.145871 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:12:29 crc kubenswrapper[4988]: I1008 18:12:29.145882 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:12:29Z","lastTransitionTime":"2025-10-08T18:12:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:12:29 crc kubenswrapper[4988]: I1008 18:12:29.237691 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 18:12:29 crc kubenswrapper[4988]: I1008 18:12:29.237761 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jn2sx" Oct 08 18:12:29 crc kubenswrapper[4988]: I1008 18:12:29.237773 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 18:12:29 crc kubenswrapper[4988]: I1008 18:12:29.237829 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 18:12:29 crc kubenswrapper[4988]: E1008 18:12:29.237921 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 08 18:12:29 crc kubenswrapper[4988]: E1008 18:12:29.238749 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jn2sx" podUID="171e0902-95ee-4a39-8b1b-a58704c30574" Oct 08 18:12:29 crc kubenswrapper[4988]: E1008 18:12:29.238951 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 08 18:12:29 crc kubenswrapper[4988]: E1008 18:12:29.239030 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 08 18:12:29 crc kubenswrapper[4988]: I1008 18:12:29.248168 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:12:29 crc kubenswrapper[4988]: I1008 18:12:29.248206 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:12:29 crc kubenswrapper[4988]: I1008 18:12:29.248217 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:12:29 crc kubenswrapper[4988]: I1008 18:12:29.248234 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:12:29 crc kubenswrapper[4988]: I1008 18:12:29.248248 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:12:29Z","lastTransitionTime":"2025-10-08T18:12:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:12:29 crc kubenswrapper[4988]: I1008 18:12:29.351679 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:12:29 crc kubenswrapper[4988]: I1008 18:12:29.351756 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:12:29 crc kubenswrapper[4988]: I1008 18:12:29.351795 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:12:29 crc kubenswrapper[4988]: I1008 18:12:29.351838 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:12:29 crc kubenswrapper[4988]: I1008 18:12:29.351863 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:12:29Z","lastTransitionTime":"2025-10-08T18:12:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:12:29 crc kubenswrapper[4988]: I1008 18:12:29.455725 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:12:29 crc kubenswrapper[4988]: I1008 18:12:29.455772 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:12:29 crc kubenswrapper[4988]: I1008 18:12:29.455781 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:12:29 crc kubenswrapper[4988]: I1008 18:12:29.455796 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:12:29 crc kubenswrapper[4988]: I1008 18:12:29.455805 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:12:29Z","lastTransitionTime":"2025-10-08T18:12:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:12:29 crc kubenswrapper[4988]: I1008 18:12:29.558325 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:12:29 crc kubenswrapper[4988]: I1008 18:12:29.558375 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:12:29 crc kubenswrapper[4988]: I1008 18:12:29.558409 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:12:29 crc kubenswrapper[4988]: I1008 18:12:29.558428 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:12:29 crc kubenswrapper[4988]: I1008 18:12:29.558441 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:12:29Z","lastTransitionTime":"2025-10-08T18:12:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:12:29 crc kubenswrapper[4988]: I1008 18:12:29.661446 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:12:29 crc kubenswrapper[4988]: I1008 18:12:29.661520 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:12:29 crc kubenswrapper[4988]: I1008 18:12:29.661544 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:12:29 crc kubenswrapper[4988]: I1008 18:12:29.661574 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:12:29 crc kubenswrapper[4988]: I1008 18:12:29.661591 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:12:29Z","lastTransitionTime":"2025-10-08T18:12:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:12:29 crc kubenswrapper[4988]: I1008 18:12:29.764887 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:12:29 crc kubenswrapper[4988]: I1008 18:12:29.764941 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:12:29 crc kubenswrapper[4988]: I1008 18:12:29.764961 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:12:29 crc kubenswrapper[4988]: I1008 18:12:29.764986 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:12:29 crc kubenswrapper[4988]: I1008 18:12:29.765002 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:12:29Z","lastTransitionTime":"2025-10-08T18:12:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:12:29 crc kubenswrapper[4988]: I1008 18:12:29.867608 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:12:29 crc kubenswrapper[4988]: I1008 18:12:29.867695 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:12:29 crc kubenswrapper[4988]: I1008 18:12:29.867718 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:12:29 crc kubenswrapper[4988]: I1008 18:12:29.867748 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:12:29 crc kubenswrapper[4988]: I1008 18:12:29.867769 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:12:29Z","lastTransitionTime":"2025-10-08T18:12:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:12:29 crc kubenswrapper[4988]: I1008 18:12:29.970714 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:12:29 crc kubenswrapper[4988]: I1008 18:12:29.970767 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:12:29 crc kubenswrapper[4988]: I1008 18:12:29.970784 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:12:29 crc kubenswrapper[4988]: I1008 18:12:29.970809 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:12:29 crc kubenswrapper[4988]: I1008 18:12:29.970826 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:12:29Z","lastTransitionTime":"2025-10-08T18:12:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:12:30 crc kubenswrapper[4988]: I1008 18:12:30.073203 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:12:30 crc kubenswrapper[4988]: I1008 18:12:30.073249 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:12:30 crc kubenswrapper[4988]: I1008 18:12:30.073258 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:12:30 crc kubenswrapper[4988]: I1008 18:12:30.073275 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:12:30 crc kubenswrapper[4988]: I1008 18:12:30.073285 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:12:30Z","lastTransitionTime":"2025-10-08T18:12:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:12:30 crc kubenswrapper[4988]: I1008 18:12:30.181658 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:12:30 crc kubenswrapper[4988]: I1008 18:12:30.181732 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:12:30 crc kubenswrapper[4988]: I1008 18:12:30.181750 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:12:30 crc kubenswrapper[4988]: I1008 18:12:30.181774 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:12:30 crc kubenswrapper[4988]: I1008 18:12:30.181792 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:12:30Z","lastTransitionTime":"2025-10-08T18:12:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:12:30 crc kubenswrapper[4988]: I1008 18:12:30.285438 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:12:30 crc kubenswrapper[4988]: I1008 18:12:30.285494 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:12:30 crc kubenswrapper[4988]: I1008 18:12:30.285507 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:12:30 crc kubenswrapper[4988]: I1008 18:12:30.285526 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:12:30 crc kubenswrapper[4988]: I1008 18:12:30.285539 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:12:30Z","lastTransitionTime":"2025-10-08T18:12:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:12:30 crc kubenswrapper[4988]: I1008 18:12:30.389237 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:12:30 crc kubenswrapper[4988]: I1008 18:12:30.389303 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:12:30 crc kubenswrapper[4988]: I1008 18:12:30.389318 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:12:30 crc kubenswrapper[4988]: I1008 18:12:30.389345 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:12:30 crc kubenswrapper[4988]: I1008 18:12:30.389362 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:12:30Z","lastTransitionTime":"2025-10-08T18:12:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:12:30 crc kubenswrapper[4988]: I1008 18:12:30.492430 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:12:30 crc kubenswrapper[4988]: I1008 18:12:30.492479 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:12:30 crc kubenswrapper[4988]: I1008 18:12:30.492487 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:12:30 crc kubenswrapper[4988]: I1008 18:12:30.492504 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:12:30 crc kubenswrapper[4988]: I1008 18:12:30.492515 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:12:30Z","lastTransitionTime":"2025-10-08T18:12:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:12:30 crc kubenswrapper[4988]: I1008 18:12:30.595166 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:12:30 crc kubenswrapper[4988]: I1008 18:12:30.595233 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:12:30 crc kubenswrapper[4988]: I1008 18:12:30.595252 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:12:30 crc kubenswrapper[4988]: I1008 18:12:30.595277 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:12:30 crc kubenswrapper[4988]: I1008 18:12:30.595299 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:12:30Z","lastTransitionTime":"2025-10-08T18:12:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:12:30 crc kubenswrapper[4988]: I1008 18:12:30.698994 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:12:30 crc kubenswrapper[4988]: I1008 18:12:30.699062 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:12:30 crc kubenswrapper[4988]: I1008 18:12:30.699085 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:12:30 crc kubenswrapper[4988]: I1008 18:12:30.699117 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:12:30 crc kubenswrapper[4988]: I1008 18:12:30.699142 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:12:30Z","lastTransitionTime":"2025-10-08T18:12:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:12:30 crc kubenswrapper[4988]: I1008 18:12:30.802743 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:12:30 crc kubenswrapper[4988]: I1008 18:12:30.802818 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:12:30 crc kubenswrapper[4988]: I1008 18:12:30.802842 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:12:30 crc kubenswrapper[4988]: I1008 18:12:30.802875 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:12:30 crc kubenswrapper[4988]: I1008 18:12:30.802900 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:12:30Z","lastTransitionTime":"2025-10-08T18:12:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:12:30 crc kubenswrapper[4988]: I1008 18:12:30.907362 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:12:30 crc kubenswrapper[4988]: I1008 18:12:30.907453 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:12:30 crc kubenswrapper[4988]: I1008 18:12:30.907478 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:12:30 crc kubenswrapper[4988]: I1008 18:12:30.907512 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:12:30 crc kubenswrapper[4988]: I1008 18:12:30.907533 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:12:30Z","lastTransitionTime":"2025-10-08T18:12:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:12:31 crc kubenswrapper[4988]: I1008 18:12:31.011309 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:12:31 crc kubenswrapper[4988]: I1008 18:12:31.011437 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:12:31 crc kubenswrapper[4988]: I1008 18:12:31.011480 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:12:31 crc kubenswrapper[4988]: I1008 18:12:31.011514 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:12:31 crc kubenswrapper[4988]: I1008 18:12:31.011540 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:12:31Z","lastTransitionTime":"2025-10-08T18:12:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:12:31 crc kubenswrapper[4988]: I1008 18:12:31.114716 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:12:31 crc kubenswrapper[4988]: I1008 18:12:31.114759 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:12:31 crc kubenswrapper[4988]: I1008 18:12:31.114769 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:12:31 crc kubenswrapper[4988]: I1008 18:12:31.114786 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:12:31 crc kubenswrapper[4988]: I1008 18:12:31.114794 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:12:31Z","lastTransitionTime":"2025-10-08T18:12:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:12:31 crc kubenswrapper[4988]: I1008 18:12:31.217820 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:12:31 crc kubenswrapper[4988]: I1008 18:12:31.217873 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:12:31 crc kubenswrapper[4988]: I1008 18:12:31.217886 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:12:31 crc kubenswrapper[4988]: I1008 18:12:31.217904 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:12:31 crc kubenswrapper[4988]: I1008 18:12:31.218252 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:12:31Z","lastTransitionTime":"2025-10-08T18:12:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:12:31 crc kubenswrapper[4988]: I1008 18:12:31.237456 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jn2sx" Oct 08 18:12:31 crc kubenswrapper[4988]: E1008 18:12:31.237664 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jn2sx" podUID="171e0902-95ee-4a39-8b1b-a58704c30574" Oct 08 18:12:31 crc kubenswrapper[4988]: I1008 18:12:31.237782 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 18:12:31 crc kubenswrapper[4988]: I1008 18:12:31.237846 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 18:12:31 crc kubenswrapper[4988]: I1008 18:12:31.237790 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 18:12:31 crc kubenswrapper[4988]: E1008 18:12:31.238008 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 08 18:12:31 crc kubenswrapper[4988]: E1008 18:12:31.238099 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 08 18:12:31 crc kubenswrapper[4988]: E1008 18:12:31.238220 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 08 18:12:31 crc kubenswrapper[4988]: I1008 18:12:31.321627 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:12:31 crc kubenswrapper[4988]: I1008 18:12:31.321690 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:12:31 crc kubenswrapper[4988]: I1008 18:12:31.321708 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:12:31 crc kubenswrapper[4988]: I1008 18:12:31.321735 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:12:31 crc kubenswrapper[4988]: I1008 18:12:31.321753 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:12:31Z","lastTransitionTime":"2025-10-08T18:12:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:12:31 crc kubenswrapper[4988]: I1008 18:12:31.425370 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:12:31 crc kubenswrapper[4988]: I1008 18:12:31.425495 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:12:31 crc kubenswrapper[4988]: I1008 18:12:31.425516 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:12:31 crc kubenswrapper[4988]: I1008 18:12:31.425548 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:12:31 crc kubenswrapper[4988]: I1008 18:12:31.425570 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:12:31Z","lastTransitionTime":"2025-10-08T18:12:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:12:31 crc kubenswrapper[4988]: I1008 18:12:31.528579 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:12:31 crc kubenswrapper[4988]: I1008 18:12:31.529480 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:12:31 crc kubenswrapper[4988]: I1008 18:12:31.529612 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:12:31 crc kubenswrapper[4988]: I1008 18:12:31.529728 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:12:31 crc kubenswrapper[4988]: I1008 18:12:31.529832 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:12:31Z","lastTransitionTime":"2025-10-08T18:12:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:12:31 crc kubenswrapper[4988]: I1008 18:12:31.632554 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:12:31 crc kubenswrapper[4988]: I1008 18:12:31.632594 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:12:31 crc kubenswrapper[4988]: I1008 18:12:31.632605 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:12:31 crc kubenswrapper[4988]: I1008 18:12:31.632626 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:12:31 crc kubenswrapper[4988]: I1008 18:12:31.632637 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:12:31Z","lastTransitionTime":"2025-10-08T18:12:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:12:31 crc kubenswrapper[4988]: I1008 18:12:31.735978 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:12:31 crc kubenswrapper[4988]: I1008 18:12:31.736034 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:12:31 crc kubenswrapper[4988]: I1008 18:12:31.736053 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:12:31 crc kubenswrapper[4988]: I1008 18:12:31.736083 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:12:31 crc kubenswrapper[4988]: I1008 18:12:31.736103 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:12:31Z","lastTransitionTime":"2025-10-08T18:12:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:12:31 crc kubenswrapper[4988]: I1008 18:12:31.839332 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:12:31 crc kubenswrapper[4988]: I1008 18:12:31.839447 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:12:31 crc kubenswrapper[4988]: I1008 18:12:31.839470 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:12:31 crc kubenswrapper[4988]: I1008 18:12:31.839505 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:12:31 crc kubenswrapper[4988]: I1008 18:12:31.839528 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:12:31Z","lastTransitionTime":"2025-10-08T18:12:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:12:31 crc kubenswrapper[4988]: I1008 18:12:31.942452 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:12:31 crc kubenswrapper[4988]: I1008 18:12:31.942512 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:12:31 crc kubenswrapper[4988]: I1008 18:12:31.942528 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:12:31 crc kubenswrapper[4988]: I1008 18:12:31.942554 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:12:31 crc kubenswrapper[4988]: I1008 18:12:31.942575 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:12:31Z","lastTransitionTime":"2025-10-08T18:12:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:12:32 crc kubenswrapper[4988]: I1008 18:12:32.045437 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:12:32 crc kubenswrapper[4988]: I1008 18:12:32.045496 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:12:32 crc kubenswrapper[4988]: I1008 18:12:32.045509 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:12:32 crc kubenswrapper[4988]: I1008 18:12:32.045533 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:12:32 crc kubenswrapper[4988]: I1008 18:12:32.045547 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:12:32Z","lastTransitionTime":"2025-10-08T18:12:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:12:32 crc kubenswrapper[4988]: I1008 18:12:32.149772 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:12:32 crc kubenswrapper[4988]: I1008 18:12:32.149840 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:12:32 crc kubenswrapper[4988]: I1008 18:12:32.149858 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:12:32 crc kubenswrapper[4988]: I1008 18:12:32.149885 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:12:32 crc kubenswrapper[4988]: I1008 18:12:32.149901 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:12:32Z","lastTransitionTime":"2025-10-08T18:12:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:12:32 crc kubenswrapper[4988]: I1008 18:12:32.254272 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:12:32 crc kubenswrapper[4988]: I1008 18:12:32.254333 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:12:32 crc kubenswrapper[4988]: I1008 18:12:32.254352 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:12:32 crc kubenswrapper[4988]: I1008 18:12:32.254414 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:12:32 crc kubenswrapper[4988]: I1008 18:12:32.254434 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:12:32Z","lastTransitionTime":"2025-10-08T18:12:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:12:32 crc kubenswrapper[4988]: I1008 18:12:32.357132 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:12:32 crc kubenswrapper[4988]: I1008 18:12:32.357195 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:12:32 crc kubenswrapper[4988]: I1008 18:12:32.357214 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:12:32 crc kubenswrapper[4988]: I1008 18:12:32.357246 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:12:32 crc kubenswrapper[4988]: I1008 18:12:32.357268 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:12:32Z","lastTransitionTime":"2025-10-08T18:12:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:12:32 crc kubenswrapper[4988]: I1008 18:12:32.460548 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:12:32 crc kubenswrapper[4988]: I1008 18:12:32.460596 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:12:32 crc kubenswrapper[4988]: I1008 18:12:32.460614 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:12:32 crc kubenswrapper[4988]: I1008 18:12:32.460639 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:12:32 crc kubenswrapper[4988]: I1008 18:12:32.460656 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:12:32Z","lastTransitionTime":"2025-10-08T18:12:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:12:32 crc kubenswrapper[4988]: I1008 18:12:32.563708 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:12:32 crc kubenswrapper[4988]: I1008 18:12:32.563802 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:12:32 crc kubenswrapper[4988]: I1008 18:12:32.563824 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:12:32 crc kubenswrapper[4988]: I1008 18:12:32.563860 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:12:32 crc kubenswrapper[4988]: I1008 18:12:32.563883 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:12:32Z","lastTransitionTime":"2025-10-08T18:12:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:12:32 crc kubenswrapper[4988]: I1008 18:12:32.667811 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:12:32 crc kubenswrapper[4988]: I1008 18:12:32.667863 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:12:32 crc kubenswrapper[4988]: I1008 18:12:32.667876 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:12:32 crc kubenswrapper[4988]: I1008 18:12:32.667899 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:12:32 crc kubenswrapper[4988]: I1008 18:12:32.667913 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:12:32Z","lastTransitionTime":"2025-10-08T18:12:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:12:32 crc kubenswrapper[4988]: I1008 18:12:32.772049 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:12:32 crc kubenswrapper[4988]: I1008 18:12:32.772134 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:12:32 crc kubenswrapper[4988]: I1008 18:12:32.772156 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:12:32 crc kubenswrapper[4988]: I1008 18:12:32.772186 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:12:32 crc kubenswrapper[4988]: I1008 18:12:32.772207 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:12:32Z","lastTransitionTime":"2025-10-08T18:12:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:12:32 crc kubenswrapper[4988]: I1008 18:12:32.875704 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:12:32 crc kubenswrapper[4988]: I1008 18:12:32.875768 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:12:32 crc kubenswrapper[4988]: I1008 18:12:32.875786 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:12:32 crc kubenswrapper[4988]: I1008 18:12:32.875812 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:12:32 crc kubenswrapper[4988]: I1008 18:12:32.875834 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:12:32Z","lastTransitionTime":"2025-10-08T18:12:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:12:32 crc kubenswrapper[4988]: I1008 18:12:32.980251 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:12:32 crc kubenswrapper[4988]: I1008 18:12:32.980326 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:12:32 crc kubenswrapper[4988]: I1008 18:12:32.980345 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:12:32 crc kubenswrapper[4988]: I1008 18:12:32.980374 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:12:32 crc kubenswrapper[4988]: I1008 18:12:32.980430 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:12:32Z","lastTransitionTime":"2025-10-08T18:12:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:12:33 crc kubenswrapper[4988]: I1008 18:12:33.084167 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:12:33 crc kubenswrapper[4988]: I1008 18:12:33.084234 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:12:33 crc kubenswrapper[4988]: I1008 18:12:33.084251 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:12:33 crc kubenswrapper[4988]: I1008 18:12:33.084280 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:12:33 crc kubenswrapper[4988]: I1008 18:12:33.084300 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:12:33Z","lastTransitionTime":"2025-10-08T18:12:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:12:33 crc kubenswrapper[4988]: I1008 18:12:33.186513 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:12:33 crc kubenswrapper[4988]: I1008 18:12:33.186537 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:12:33 crc kubenswrapper[4988]: I1008 18:12:33.186544 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:12:33 crc kubenswrapper[4988]: I1008 18:12:33.186556 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:12:33 crc kubenswrapper[4988]: I1008 18:12:33.186564 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:12:33Z","lastTransitionTime":"2025-10-08T18:12:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:12:33 crc kubenswrapper[4988]: I1008 18:12:33.237201 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 18:12:33 crc kubenswrapper[4988]: E1008 18:12:33.237367 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 08 18:12:33 crc kubenswrapper[4988]: I1008 18:12:33.237666 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 18:12:33 crc kubenswrapper[4988]: E1008 18:12:33.237785 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 08 18:12:33 crc kubenswrapper[4988]: I1008 18:12:33.238000 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jn2sx" Oct 08 18:12:33 crc kubenswrapper[4988]: E1008 18:12:33.238157 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jn2sx" podUID="171e0902-95ee-4a39-8b1b-a58704c30574" Oct 08 18:12:33 crc kubenswrapper[4988]: I1008 18:12:33.238186 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 18:12:33 crc kubenswrapper[4988]: E1008 18:12:33.238307 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 08 18:12:33 crc kubenswrapper[4988]: I1008 18:12:33.289720 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:12:33 crc kubenswrapper[4988]: I1008 18:12:33.289793 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:12:33 crc kubenswrapper[4988]: I1008 18:12:33.289814 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:12:33 crc kubenswrapper[4988]: I1008 18:12:33.289837 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:12:33 crc kubenswrapper[4988]: I1008 18:12:33.289856 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:12:33Z","lastTransitionTime":"2025-10-08T18:12:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:12:33 crc kubenswrapper[4988]: I1008 18:12:33.393058 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:12:33 crc kubenswrapper[4988]: I1008 18:12:33.393098 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:12:33 crc kubenswrapper[4988]: I1008 18:12:33.393110 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:12:33 crc kubenswrapper[4988]: I1008 18:12:33.393130 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:12:33 crc kubenswrapper[4988]: I1008 18:12:33.393143 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:12:33Z","lastTransitionTime":"2025-10-08T18:12:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:12:33 crc kubenswrapper[4988]: I1008 18:12:33.496227 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:12:33 crc kubenswrapper[4988]: I1008 18:12:33.496280 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:12:33 crc kubenswrapper[4988]: I1008 18:12:33.496295 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:12:33 crc kubenswrapper[4988]: I1008 18:12:33.496315 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:12:33 crc kubenswrapper[4988]: I1008 18:12:33.496332 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:12:33Z","lastTransitionTime":"2025-10-08T18:12:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:12:33 crc kubenswrapper[4988]: I1008 18:12:33.568478 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:12:33 crc kubenswrapper[4988]: I1008 18:12:33.568546 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:12:33 crc kubenswrapper[4988]: I1008 18:12:33.568561 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:12:33 crc kubenswrapper[4988]: I1008 18:12:33.568580 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:12:33 crc kubenswrapper[4988]: I1008 18:12:33.568591 4988 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:12:33Z","lastTransitionTime":"2025-10-08T18:12:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:12:33 crc kubenswrapper[4988]: I1008 18:12:33.620563 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-version/cluster-version-operator-5c965bbfc6-54xxf"] Oct 08 18:12:33 crc kubenswrapper[4988]: I1008 18:12:33.621132 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-54xxf" Oct 08 18:12:33 crc kubenswrapper[4988]: I1008 18:12:33.623843 4988 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"openshift-service-ca.crt" Oct 08 18:12:33 crc kubenswrapper[4988]: I1008 18:12:33.624845 4988 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"kube-root-ca.crt" Oct 08 18:12:33 crc kubenswrapper[4988]: I1008 18:12:33.626281 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"cluster-version-operator-serving-cert" Oct 08 18:12:33 crc kubenswrapper[4988]: I1008 18:12:33.629113 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/249d3444-fd0d-49f1-9ebc-9ddabf5957b7-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-54xxf\" (UID: \"249d3444-fd0d-49f1-9ebc-9ddabf5957b7\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-54xxf" Oct 08 18:12:33 crc kubenswrapper[4988]: I1008 18:12:33.629163 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/249d3444-fd0d-49f1-9ebc-9ddabf5957b7-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-54xxf\" (UID: \"249d3444-fd0d-49f1-9ebc-9ddabf5957b7\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-54xxf" Oct 08 18:12:33 crc kubenswrapper[4988]: I1008 18:12:33.629226 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/249d3444-fd0d-49f1-9ebc-9ddabf5957b7-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-54xxf\" (UID: \"249d3444-fd0d-49f1-9ebc-9ddabf5957b7\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-54xxf" Oct 08 18:12:33 crc kubenswrapper[4988]: I1008 18:12:33.629260 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/249d3444-fd0d-49f1-9ebc-9ddabf5957b7-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-54xxf\" (UID: \"249d3444-fd0d-49f1-9ebc-9ddabf5957b7\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-54xxf" Oct 08 18:12:33 crc kubenswrapper[4988]: I1008 18:12:33.629287 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/249d3444-fd0d-49f1-9ebc-9ddabf5957b7-service-ca\") pod \"cluster-version-operator-5c965bbfc6-54xxf\" (UID: \"249d3444-fd0d-49f1-9ebc-9ddabf5957b7\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-54xxf" Oct 08 18:12:33 crc kubenswrapper[4988]: I1008 18:12:33.631022 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"default-dockercfg-gxtc4" Oct 08 18:12:33 crc kubenswrapper[4988]: I1008 18:12:33.707217 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-h9jwp" podStartSLOduration=77.707192406 podStartE2EDuration="1m17.707192406s" podCreationTimestamp="2025-10-08 18:11:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 18:12:33.669026017 +0000 UTC m=+99.118868817" watchObservedRunningTime="2025-10-08 18:12:33.707192406 +0000 UTC m=+99.157035186" Oct 08 18:12:33 crc kubenswrapper[4988]: I1008 18:12:33.726831 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-7wl8v" podStartSLOduration=77.726793891 podStartE2EDuration="1m17.726793891s" podCreationTimestamp="2025-10-08 18:11:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 18:12:33.725940263 +0000 UTC m=+99.175783043" watchObservedRunningTime="2025-10-08 18:12:33.726793891 +0000 UTC m=+99.176636701" Oct 08 18:12:33 crc kubenswrapper[4988]: I1008 18:12:33.730175 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/249d3444-fd0d-49f1-9ebc-9ddabf5957b7-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-54xxf\" (UID: \"249d3444-fd0d-49f1-9ebc-9ddabf5957b7\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-54xxf" Oct 08 18:12:33 crc kubenswrapper[4988]: I1008 18:12:33.730249 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/249d3444-fd0d-49f1-9ebc-9ddabf5957b7-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-54xxf\" (UID: \"249d3444-fd0d-49f1-9ebc-9ddabf5957b7\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-54xxf" Oct 08 18:12:33 crc kubenswrapper[4988]: I1008 18:12:33.730301 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/249d3444-fd0d-49f1-9ebc-9ddabf5957b7-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-54xxf\" (UID: \"249d3444-fd0d-49f1-9ebc-9ddabf5957b7\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-54xxf" Oct 08 18:12:33 crc kubenswrapper[4988]: I1008 18:12:33.730308 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/249d3444-fd0d-49f1-9ebc-9ddabf5957b7-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-54xxf\" (UID: \"249d3444-fd0d-49f1-9ebc-9ddabf5957b7\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-54xxf" Oct 08 18:12:33 crc kubenswrapper[4988]: I1008 18:12:33.730328 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/249d3444-fd0d-49f1-9ebc-9ddabf5957b7-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-54xxf\" (UID: \"249d3444-fd0d-49f1-9ebc-9ddabf5957b7\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-54xxf" Oct 08 18:12:33 crc kubenswrapper[4988]: I1008 18:12:33.730364 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/249d3444-fd0d-49f1-9ebc-9ddabf5957b7-service-ca\") pod \"cluster-version-operator-5c965bbfc6-54xxf\" (UID: \"249d3444-fd0d-49f1-9ebc-9ddabf5957b7\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-54xxf" Oct 08 18:12:33 crc kubenswrapper[4988]: I1008 18:12:33.730518 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/249d3444-fd0d-49f1-9ebc-9ddabf5957b7-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-54xxf\" (UID: \"249d3444-fd0d-49f1-9ebc-9ddabf5957b7\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-54xxf" Oct 08 18:12:33 crc kubenswrapper[4988]: I1008 18:12:33.731708 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/249d3444-fd0d-49f1-9ebc-9ddabf5957b7-service-ca\") pod \"cluster-version-operator-5c965bbfc6-54xxf\" (UID: \"249d3444-fd0d-49f1-9ebc-9ddabf5957b7\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-54xxf" Oct 08 18:12:33 crc kubenswrapper[4988]: I1008 18:12:33.747081 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/249d3444-fd0d-49f1-9ebc-9ddabf5957b7-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-54xxf\" (UID: \"249d3444-fd0d-49f1-9ebc-9ddabf5957b7\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-54xxf" Oct 08 18:12:33 crc kubenswrapper[4988]: I1008 18:12:33.750714 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/249d3444-fd0d-49f1-9ebc-9ddabf5957b7-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-54xxf\" (UID: \"249d3444-fd0d-49f1-9ebc-9ddabf5957b7\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-54xxf" Oct 08 18:12:33 crc kubenswrapper[4988]: I1008 18:12:33.761915 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" podStartSLOduration=43.761899625 podStartE2EDuration="43.761899625s" podCreationTimestamp="2025-10-08 18:11:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 18:12:33.761625939 +0000 UTC m=+99.211468719" watchObservedRunningTime="2025-10-08 18:12:33.761899625 +0000 UTC m=+99.211742405" Oct 08 18:12:33 crc kubenswrapper[4988]: I1008 18:12:33.787899 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" podStartSLOduration=28.787873126 podStartE2EDuration="28.787873126s" podCreationTimestamp="2025-10-08 18:12:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 18:12:33.786478067 +0000 UTC m=+99.236320857" watchObservedRunningTime="2025-10-08 18:12:33.787873126 +0000 UTC m=+99.237715906" Oct 08 18:12:33 crc kubenswrapper[4988]: I1008 18:12:33.811478 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/node-resolver-s4f2k" podStartSLOduration=78.811459846 podStartE2EDuration="1m18.811459846s" podCreationTimestamp="2025-10-08 18:11:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 18:12:33.795790214 +0000 UTC m=+99.245633014" watchObservedRunningTime="2025-10-08 18:12:33.811459846 +0000 UTC m=+99.261302626" Oct 08 18:12:33 crc kubenswrapper[4988]: I1008 18:12:33.841702 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/node-ca-dzxs2" podStartSLOduration=78.841672537 podStartE2EDuration="1m18.841672537s" podCreationTimestamp="2025-10-08 18:11:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 18:12:33.823504761 +0000 UTC m=+99.273347571" watchObservedRunningTime="2025-10-08 18:12:33.841672537 +0000 UTC m=+99.291515327" Oct 08 18:12:33 crc kubenswrapper[4988]: I1008 18:12:33.858011 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=78.857989332 podStartE2EDuration="1m18.857989332s" podCreationTimestamp="2025-10-08 18:11:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 18:12:33.857548763 +0000 UTC m=+99.307391563" watchObservedRunningTime="2025-10-08 18:12:33.857989332 +0000 UTC m=+99.307832122" Oct 08 18:12:33 crc kubenswrapper[4988]: I1008 18:12:33.877890 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podStartSLOduration=72.877861574 podStartE2EDuration="1m12.877861574s" podCreationTimestamp="2025-10-08 18:11:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 18:12:33.876968464 +0000 UTC m=+99.326811304" watchObservedRunningTime="2025-10-08 18:12:33.877861574 +0000 UTC m=+99.327704364" Oct 08 18:12:33 crc kubenswrapper[4988]: I1008 18:12:33.924757 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-additional-cni-plugins-jnh7m" podStartSLOduration=77.924728177 podStartE2EDuration="1m17.924728177s" podCreationTimestamp="2025-10-08 18:11:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 18:12:33.923643984 +0000 UTC m=+99.373486764" watchObservedRunningTime="2025-10-08 18:12:33.924728177 +0000 UTC m=+99.374570997" Oct 08 18:12:33 crc kubenswrapper[4988]: I1008 18:12:33.937037 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-54xxf" Oct 08 18:12:33 crc kubenswrapper[4988]: I1008 18:12:33.938443 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podStartSLOduration=77.938377246 podStartE2EDuration="1m17.938377246s" podCreationTimestamp="2025-10-08 18:11:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 18:12:33.937659791 +0000 UTC m=+99.387502611" watchObservedRunningTime="2025-10-08 18:12:33.938377246 +0000 UTC m=+99.388220056" Oct 08 18:12:34 crc kubenswrapper[4988]: I1008 18:12:34.237955 4988 scope.go:117] "RemoveContainer" containerID="763353df23d26dc093a685af435811240b5ee772d7279ee9a6d2c7a48c19cd30" Oct 08 18:12:34 crc kubenswrapper[4988]: E1008 18:12:34.238197 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-nsc65_openshift-ovn-kubernetes(f010a463-69f2-4082-acc3-161899a939c3)\"" pod="openshift-ovn-kubernetes/ovnkube-node-nsc65" podUID="f010a463-69f2-4082-acc3-161899a939c3" Oct 08 18:12:34 crc kubenswrapper[4988]: I1008 18:12:34.925274 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-54xxf" event={"ID":"249d3444-fd0d-49f1-9ebc-9ddabf5957b7","Type":"ContainerStarted","Data":"a24ff5f8bfa9df72f57d0ed53076222735b8c24a252a0c9e88f414e3fd95b818"} Oct 08 18:12:34 crc kubenswrapper[4988]: I1008 18:12:34.925327 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-54xxf" event={"ID":"249d3444-fd0d-49f1-9ebc-9ddabf5957b7","Type":"ContainerStarted","Data":"54646dc5aebaa097d1689ffd22e1d1abdb6482086fae3b6821f10af27bba0945"} Oct 08 18:12:34 crc kubenswrapper[4988]: I1008 18:12:34.942735 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/171e0902-95ee-4a39-8b1b-a58704c30574-metrics-certs\") pod \"network-metrics-daemon-jn2sx\" (UID: \"171e0902-95ee-4a39-8b1b-a58704c30574\") " pod="openshift-multus/network-metrics-daemon-jn2sx" Oct 08 18:12:34 crc kubenswrapper[4988]: E1008 18:12:34.942890 4988 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 08 18:12:34 crc kubenswrapper[4988]: E1008 18:12:34.942962 4988 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/171e0902-95ee-4a39-8b1b-a58704c30574-metrics-certs podName:171e0902-95ee-4a39-8b1b-a58704c30574 nodeName:}" failed. No retries permitted until 2025-10-08 18:13:38.94294281 +0000 UTC m=+164.392785580 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/171e0902-95ee-4a39-8b1b-a58704c30574-metrics-certs") pod "network-metrics-daemon-jn2sx" (UID: "171e0902-95ee-4a39-8b1b-a58704c30574") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 08 18:12:35 crc kubenswrapper[4988]: I1008 18:12:35.237719 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 18:12:35 crc kubenswrapper[4988]: I1008 18:12:35.237815 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 18:12:35 crc kubenswrapper[4988]: E1008 18:12:35.239671 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 08 18:12:35 crc kubenswrapper[4988]: I1008 18:12:35.239742 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jn2sx" Oct 08 18:12:35 crc kubenswrapper[4988]: I1008 18:12:35.239723 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 18:12:35 crc kubenswrapper[4988]: E1008 18:12:35.239904 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 08 18:12:35 crc kubenswrapper[4988]: E1008 18:12:35.240014 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jn2sx" podUID="171e0902-95ee-4a39-8b1b-a58704c30574" Oct 08 18:12:35 crc kubenswrapper[4988]: E1008 18:12:35.240294 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 08 18:12:37 crc kubenswrapper[4988]: I1008 18:12:37.237789 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 18:12:37 crc kubenswrapper[4988]: I1008 18:12:37.237864 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jn2sx" Oct 08 18:12:37 crc kubenswrapper[4988]: I1008 18:12:37.237923 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 18:12:37 crc kubenswrapper[4988]: E1008 18:12:37.237989 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 08 18:12:37 crc kubenswrapper[4988]: I1008 18:12:37.238058 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 18:12:37 crc kubenswrapper[4988]: E1008 18:12:37.238463 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 08 18:12:37 crc kubenswrapper[4988]: E1008 18:12:37.238464 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jn2sx" podUID="171e0902-95ee-4a39-8b1b-a58704c30574" Oct 08 18:12:37 crc kubenswrapper[4988]: E1008 18:12:37.238890 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 08 18:12:38 crc kubenswrapper[4988]: I1008 18:12:38.261871 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-54xxf" podStartSLOduration=82.261845521 podStartE2EDuration="1m22.261845521s" podCreationTimestamp="2025-10-08 18:11:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 18:12:34.944556814 +0000 UTC m=+100.394399604" watchObservedRunningTime="2025-10-08 18:12:38.261845521 +0000 UTC m=+103.711688321" Oct 08 18:12:38 crc kubenswrapper[4988]: I1008 18:12:38.262468 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd/etcd-crc"] Oct 08 18:12:39 crc kubenswrapper[4988]: I1008 18:12:39.237783 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 18:12:39 crc kubenswrapper[4988]: I1008 18:12:39.237810 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jn2sx" Oct 08 18:12:39 crc kubenswrapper[4988]: I1008 18:12:39.237832 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 18:12:39 crc kubenswrapper[4988]: I1008 18:12:39.237836 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 18:12:39 crc kubenswrapper[4988]: E1008 18:12:39.238523 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jn2sx" podUID="171e0902-95ee-4a39-8b1b-a58704c30574" Oct 08 18:12:39 crc kubenswrapper[4988]: E1008 18:12:39.238675 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 08 18:12:39 crc kubenswrapper[4988]: E1008 18:12:39.238792 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 08 18:12:39 crc kubenswrapper[4988]: E1008 18:12:39.239085 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 08 18:12:41 crc kubenswrapper[4988]: I1008 18:12:41.237083 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 18:12:41 crc kubenswrapper[4988]: E1008 18:12:41.237280 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 08 18:12:41 crc kubenswrapper[4988]: I1008 18:12:41.237678 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 18:12:41 crc kubenswrapper[4988]: I1008 18:12:41.237764 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jn2sx" Oct 08 18:12:41 crc kubenswrapper[4988]: I1008 18:12:41.237764 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 18:12:41 crc kubenswrapper[4988]: E1008 18:12:41.237888 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 08 18:12:41 crc kubenswrapper[4988]: E1008 18:12:41.238042 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jn2sx" podUID="171e0902-95ee-4a39-8b1b-a58704c30574" Oct 08 18:12:41 crc kubenswrapper[4988]: E1008 18:12:41.238178 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 08 18:12:43 crc kubenswrapper[4988]: I1008 18:12:43.237511 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 18:12:43 crc kubenswrapper[4988]: I1008 18:12:43.237544 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 18:12:43 crc kubenswrapper[4988]: I1008 18:12:43.238770 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 18:12:43 crc kubenswrapper[4988]: I1008 18:12:43.238778 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jn2sx" Oct 08 18:12:43 crc kubenswrapper[4988]: E1008 18:12:43.239007 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 08 18:12:43 crc kubenswrapper[4988]: E1008 18:12:43.239074 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 08 18:12:43 crc kubenswrapper[4988]: E1008 18:12:43.239276 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 08 18:12:43 crc kubenswrapper[4988]: E1008 18:12:43.239480 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jn2sx" podUID="171e0902-95ee-4a39-8b1b-a58704c30574" Oct 08 18:12:45 crc kubenswrapper[4988]: I1008 18:12:45.236792 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 18:12:45 crc kubenswrapper[4988]: E1008 18:12:45.239534 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 08 18:12:45 crc kubenswrapper[4988]: I1008 18:12:45.239618 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 18:12:45 crc kubenswrapper[4988]: I1008 18:12:45.239655 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 18:12:45 crc kubenswrapper[4988]: E1008 18:12:45.239760 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 08 18:12:45 crc kubenswrapper[4988]: I1008 18:12:45.239673 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jn2sx" Oct 08 18:12:45 crc kubenswrapper[4988]: E1008 18:12:45.239917 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 08 18:12:45 crc kubenswrapper[4988]: E1008 18:12:45.240069 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jn2sx" podUID="171e0902-95ee-4a39-8b1b-a58704c30574" Oct 08 18:12:45 crc kubenswrapper[4988]: I1008 18:12:45.285806 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd/etcd-crc" podStartSLOduration=7.285778206 podStartE2EDuration="7.285778206s" podCreationTimestamp="2025-10-08 18:12:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 18:12:45.282627979 +0000 UTC m=+110.732470839" watchObservedRunningTime="2025-10-08 18:12:45.285778206 +0000 UTC m=+110.735621006" Oct 08 18:12:47 crc kubenswrapper[4988]: I1008 18:12:47.237234 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 18:12:47 crc kubenswrapper[4988]: I1008 18:12:47.237234 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 18:12:47 crc kubenswrapper[4988]: E1008 18:12:47.237732 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 08 18:12:47 crc kubenswrapper[4988]: I1008 18:12:47.237358 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jn2sx" Oct 08 18:12:47 crc kubenswrapper[4988]: I1008 18:12:47.237352 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 18:12:47 crc kubenswrapper[4988]: E1008 18:12:47.237928 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 08 18:12:47 crc kubenswrapper[4988]: E1008 18:12:47.237982 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 08 18:12:47 crc kubenswrapper[4988]: E1008 18:12:47.238109 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jn2sx" podUID="171e0902-95ee-4a39-8b1b-a58704c30574" Oct 08 18:12:48 crc kubenswrapper[4988]: I1008 18:12:48.238597 4988 scope.go:117] "RemoveContainer" containerID="763353df23d26dc093a685af435811240b5ee772d7279ee9a6d2c7a48c19cd30" Oct 08 18:12:48 crc kubenswrapper[4988]: E1008 18:12:48.238885 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-nsc65_openshift-ovn-kubernetes(f010a463-69f2-4082-acc3-161899a939c3)\"" pod="openshift-ovn-kubernetes/ovnkube-node-nsc65" podUID="f010a463-69f2-4082-acc3-161899a939c3" Oct 08 18:12:49 crc kubenswrapper[4988]: I1008 18:12:49.237883 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 18:12:49 crc kubenswrapper[4988]: I1008 18:12:49.238354 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jn2sx" Oct 08 18:12:49 crc kubenswrapper[4988]: I1008 18:12:49.238279 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 18:12:49 crc kubenswrapper[4988]: I1008 18:12:49.237884 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 18:12:49 crc kubenswrapper[4988]: E1008 18:12:49.238988 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 08 18:12:49 crc kubenswrapper[4988]: E1008 18:12:49.239137 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jn2sx" podUID="171e0902-95ee-4a39-8b1b-a58704c30574" Oct 08 18:12:49 crc kubenswrapper[4988]: E1008 18:12:49.239278 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 08 18:12:49 crc kubenswrapper[4988]: E1008 18:12:49.239654 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 08 18:12:50 crc kubenswrapper[4988]: I1008 18:12:50.984874 4988 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-h9jwp_4490d846-ca32-4cd5-9202-87cc6d499fb4/kube-multus/1.log" Oct 08 18:12:50 crc kubenswrapper[4988]: I1008 18:12:50.985656 4988 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-h9jwp_4490d846-ca32-4cd5-9202-87cc6d499fb4/kube-multus/0.log" Oct 08 18:12:50 crc kubenswrapper[4988]: I1008 18:12:50.985710 4988 generic.go:334] "Generic (PLEG): container finished" podID="4490d846-ca32-4cd5-9202-87cc6d499fb4" containerID="c38c1b52127a970f792ef09ef1b217c979d94f7f74cc3c27e38949a204ad831f" exitCode=1 Oct 08 18:12:50 crc kubenswrapper[4988]: I1008 18:12:50.985756 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-h9jwp" event={"ID":"4490d846-ca32-4cd5-9202-87cc6d499fb4","Type":"ContainerDied","Data":"c38c1b52127a970f792ef09ef1b217c979d94f7f74cc3c27e38949a204ad831f"} Oct 08 18:12:50 crc kubenswrapper[4988]: I1008 18:12:50.985817 4988 scope.go:117] "RemoveContainer" containerID="00b7f6db043c28efda3a09fdc77a6ccded3eae8b6a3d2852aa500789d5f01194" Oct 08 18:12:50 crc kubenswrapper[4988]: I1008 18:12:50.986450 4988 scope.go:117] "RemoveContainer" containerID="c38c1b52127a970f792ef09ef1b217c979d94f7f74cc3c27e38949a204ad831f" Oct 08 18:12:50 crc kubenswrapper[4988]: E1008 18:12:50.986706 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-multus pod=multus-h9jwp_openshift-multus(4490d846-ca32-4cd5-9202-87cc6d499fb4)\"" pod="openshift-multus/multus-h9jwp" podUID="4490d846-ca32-4cd5-9202-87cc6d499fb4" Oct 08 18:12:51 crc kubenswrapper[4988]: I1008 18:12:51.236954 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 18:12:51 crc kubenswrapper[4988]: I1008 18:12:51.236955 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 18:12:51 crc kubenswrapper[4988]: E1008 18:12:51.237413 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 08 18:12:51 crc kubenswrapper[4988]: I1008 18:12:51.236981 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 18:12:51 crc kubenswrapper[4988]: E1008 18:12:51.237502 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 08 18:12:51 crc kubenswrapper[4988]: I1008 18:12:51.236971 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jn2sx" Oct 08 18:12:51 crc kubenswrapper[4988]: E1008 18:12:51.237770 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 08 18:12:51 crc kubenswrapper[4988]: E1008 18:12:51.238536 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jn2sx" podUID="171e0902-95ee-4a39-8b1b-a58704c30574" Oct 08 18:12:51 crc kubenswrapper[4988]: I1008 18:12:51.991216 4988 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-h9jwp_4490d846-ca32-4cd5-9202-87cc6d499fb4/kube-multus/1.log" Oct 08 18:12:53 crc kubenswrapper[4988]: I1008 18:12:53.237445 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jn2sx" Oct 08 18:12:53 crc kubenswrapper[4988]: I1008 18:12:53.237462 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 18:12:53 crc kubenswrapper[4988]: I1008 18:12:53.237553 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 18:12:53 crc kubenswrapper[4988]: I1008 18:12:53.237723 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 18:12:53 crc kubenswrapper[4988]: E1008 18:12:53.237794 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jn2sx" podUID="171e0902-95ee-4a39-8b1b-a58704c30574" Oct 08 18:12:53 crc kubenswrapper[4988]: E1008 18:12:53.238047 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 08 18:12:53 crc kubenswrapper[4988]: E1008 18:12:53.238142 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 08 18:12:53 crc kubenswrapper[4988]: E1008 18:12:53.238932 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 08 18:12:55 crc kubenswrapper[4988]: E1008 18:12:55.233315 4988 kubelet_node_status.go:497] "Node not becoming ready in time after startup" Oct 08 18:12:55 crc kubenswrapper[4988]: I1008 18:12:55.236947 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jn2sx" Oct 08 18:12:55 crc kubenswrapper[4988]: I1008 18:12:55.237040 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 18:12:55 crc kubenswrapper[4988]: I1008 18:12:55.237094 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 18:12:55 crc kubenswrapper[4988]: I1008 18:12:55.237675 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 18:12:55 crc kubenswrapper[4988]: E1008 18:12:55.239977 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jn2sx" podUID="171e0902-95ee-4a39-8b1b-a58704c30574" Oct 08 18:12:55 crc kubenswrapper[4988]: E1008 18:12:55.240126 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 08 18:12:55 crc kubenswrapper[4988]: E1008 18:12:55.240311 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 08 18:12:55 crc kubenswrapper[4988]: E1008 18:12:55.240505 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 08 18:12:55 crc kubenswrapper[4988]: E1008 18:12:55.342320 4988 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Oct 08 18:12:57 crc kubenswrapper[4988]: I1008 18:12:57.237360 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 18:12:57 crc kubenswrapper[4988]: I1008 18:12:57.237416 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 18:12:57 crc kubenswrapper[4988]: I1008 18:12:57.237402 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jn2sx" Oct 08 18:12:57 crc kubenswrapper[4988]: I1008 18:12:57.237497 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 18:12:57 crc kubenswrapper[4988]: E1008 18:12:57.237637 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 08 18:12:57 crc kubenswrapper[4988]: E1008 18:12:57.237736 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 08 18:12:57 crc kubenswrapper[4988]: E1008 18:12:57.238105 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jn2sx" podUID="171e0902-95ee-4a39-8b1b-a58704c30574" Oct 08 18:12:57 crc kubenswrapper[4988]: E1008 18:12:57.238170 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 08 18:12:59 crc kubenswrapper[4988]: I1008 18:12:59.237050 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 18:12:59 crc kubenswrapper[4988]: I1008 18:12:59.237123 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 18:12:59 crc kubenswrapper[4988]: I1008 18:12:59.237216 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jn2sx" Oct 08 18:12:59 crc kubenswrapper[4988]: E1008 18:12:59.237268 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 08 18:12:59 crc kubenswrapper[4988]: I1008 18:12:59.237286 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 18:12:59 crc kubenswrapper[4988]: E1008 18:12:59.237522 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jn2sx" podUID="171e0902-95ee-4a39-8b1b-a58704c30574" Oct 08 18:12:59 crc kubenswrapper[4988]: E1008 18:12:59.237632 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 08 18:12:59 crc kubenswrapper[4988]: E1008 18:12:59.237809 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 08 18:13:00 crc kubenswrapper[4988]: E1008 18:13:00.344340 4988 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Oct 08 18:13:01 crc kubenswrapper[4988]: I1008 18:13:01.236827 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 18:13:01 crc kubenswrapper[4988]: I1008 18:13:01.236887 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 18:13:01 crc kubenswrapper[4988]: I1008 18:13:01.236827 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jn2sx" Oct 08 18:13:01 crc kubenswrapper[4988]: I1008 18:13:01.237044 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 18:13:01 crc kubenswrapper[4988]: E1008 18:13:01.237043 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 08 18:13:01 crc kubenswrapper[4988]: E1008 18:13:01.237171 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jn2sx" podUID="171e0902-95ee-4a39-8b1b-a58704c30574" Oct 08 18:13:01 crc kubenswrapper[4988]: E1008 18:13:01.237275 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 08 18:13:01 crc kubenswrapper[4988]: E1008 18:13:01.237338 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 08 18:13:02 crc kubenswrapper[4988]: I1008 18:13:02.238511 4988 scope.go:117] "RemoveContainer" containerID="763353df23d26dc093a685af435811240b5ee772d7279ee9a6d2c7a48c19cd30" Oct 08 18:13:03 crc kubenswrapper[4988]: I1008 18:13:03.036456 4988 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-nsc65_f010a463-69f2-4082-acc3-161899a939c3/ovnkube-controller/3.log" Oct 08 18:13:03 crc kubenswrapper[4988]: I1008 18:13:03.040538 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-nsc65" event={"ID":"f010a463-69f2-4082-acc3-161899a939c3","Type":"ContainerStarted","Data":"4987245ff838ca8d4b8c31a08147c052d23bdad26126fcb46367d4b5d4197fb2"} Oct 08 18:13:03 crc kubenswrapper[4988]: I1008 18:13:03.041286 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-nsc65" Oct 08 18:13:03 crc kubenswrapper[4988]: I1008 18:13:03.091736 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-nsc65" podStartSLOduration=107.091702403 podStartE2EDuration="1m47.091702403s" podCreationTimestamp="2025-10-08 18:11:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 18:13:03.088142108 +0000 UTC m=+128.537984968" watchObservedRunningTime="2025-10-08 18:13:03.091702403 +0000 UTC m=+128.541545213" Oct 08 18:13:03 crc kubenswrapper[4988]: I1008 18:13:03.237818 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 18:13:03 crc kubenswrapper[4988]: I1008 18:13:03.237857 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 18:13:03 crc kubenswrapper[4988]: I1008 18:13:03.237967 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 18:13:03 crc kubenswrapper[4988]: I1008 18:13:03.237967 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jn2sx" Oct 08 18:13:03 crc kubenswrapper[4988]: E1008 18:13:03.238071 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 08 18:13:03 crc kubenswrapper[4988]: E1008 18:13:03.238263 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 08 18:13:03 crc kubenswrapper[4988]: E1008 18:13:03.238373 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 08 18:13:03 crc kubenswrapper[4988]: E1008 18:13:03.238583 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jn2sx" podUID="171e0902-95ee-4a39-8b1b-a58704c30574" Oct 08 18:13:03 crc kubenswrapper[4988]: I1008 18:13:03.395875 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-jn2sx"] Oct 08 18:13:04 crc kubenswrapper[4988]: I1008 18:13:04.044661 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jn2sx" Oct 08 18:13:04 crc kubenswrapper[4988]: E1008 18:13:04.045218 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jn2sx" podUID="171e0902-95ee-4a39-8b1b-a58704c30574" Oct 08 18:13:05 crc kubenswrapper[4988]: I1008 18:13:05.237313 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 18:13:05 crc kubenswrapper[4988]: I1008 18:13:05.237512 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 18:13:05 crc kubenswrapper[4988]: E1008 18:13:05.239966 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 08 18:13:05 crc kubenswrapper[4988]: I1008 18:13:05.239995 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 18:13:05 crc kubenswrapper[4988]: E1008 18:13:05.240220 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 08 18:13:05 crc kubenswrapper[4988]: E1008 18:13:05.240660 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 08 18:13:05 crc kubenswrapper[4988]: I1008 18:13:05.240783 4988 scope.go:117] "RemoveContainer" containerID="c38c1b52127a970f792ef09ef1b217c979d94f7f74cc3c27e38949a204ad831f" Oct 08 18:13:05 crc kubenswrapper[4988]: E1008 18:13:05.345725 4988 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Oct 08 18:13:06 crc kubenswrapper[4988]: I1008 18:13:06.057676 4988 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-h9jwp_4490d846-ca32-4cd5-9202-87cc6d499fb4/kube-multus/1.log" Oct 08 18:13:06 crc kubenswrapper[4988]: I1008 18:13:06.057756 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-h9jwp" event={"ID":"4490d846-ca32-4cd5-9202-87cc6d499fb4","Type":"ContainerStarted","Data":"f62030d24f69a85e1807099969f0021ab1198002c76e11ef3dffadfabff73b49"} Oct 08 18:13:06 crc kubenswrapper[4988]: I1008 18:13:06.237163 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jn2sx" Oct 08 18:13:06 crc kubenswrapper[4988]: E1008 18:13:06.237369 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jn2sx" podUID="171e0902-95ee-4a39-8b1b-a58704c30574" Oct 08 18:13:07 crc kubenswrapper[4988]: I1008 18:13:07.237593 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 18:13:07 crc kubenswrapper[4988]: E1008 18:13:07.237820 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 08 18:13:07 crc kubenswrapper[4988]: I1008 18:13:07.238164 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 18:13:07 crc kubenswrapper[4988]: I1008 18:13:07.238167 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 18:13:07 crc kubenswrapper[4988]: E1008 18:13:07.238260 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 08 18:13:07 crc kubenswrapper[4988]: E1008 18:13:07.238475 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 08 18:13:08 crc kubenswrapper[4988]: I1008 18:13:08.237061 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jn2sx" Oct 08 18:13:08 crc kubenswrapper[4988]: E1008 18:13:08.237300 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jn2sx" podUID="171e0902-95ee-4a39-8b1b-a58704c30574" Oct 08 18:13:09 crc kubenswrapper[4988]: I1008 18:13:09.237195 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 18:13:09 crc kubenswrapper[4988]: I1008 18:13:09.237288 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 18:13:09 crc kubenswrapper[4988]: I1008 18:13:09.237221 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 18:13:09 crc kubenswrapper[4988]: E1008 18:13:09.237366 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 08 18:13:09 crc kubenswrapper[4988]: E1008 18:13:09.237560 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 08 18:13:09 crc kubenswrapper[4988]: E1008 18:13:09.237642 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 08 18:13:10 crc kubenswrapper[4988]: I1008 18:13:10.237733 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jn2sx" Oct 08 18:13:10 crc kubenswrapper[4988]: E1008 18:13:10.237955 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jn2sx" podUID="171e0902-95ee-4a39-8b1b-a58704c30574" Oct 08 18:13:11 crc kubenswrapper[4988]: I1008 18:13:11.237270 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 18:13:11 crc kubenswrapper[4988]: I1008 18:13:11.237347 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 18:13:11 crc kubenswrapper[4988]: I1008 18:13:11.237280 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 18:13:11 crc kubenswrapper[4988]: I1008 18:13:11.246532 4988 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Oct 08 18:13:11 crc kubenswrapper[4988]: I1008 18:13:11.246754 4988 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Oct 08 18:13:11 crc kubenswrapper[4988]: I1008 18:13:11.246974 4988 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Oct 08 18:13:11 crc kubenswrapper[4988]: I1008 18:13:11.250037 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Oct 08 18:13:12 crc kubenswrapper[4988]: I1008 18:13:12.237908 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jn2sx" Oct 08 18:13:12 crc kubenswrapper[4988]: I1008 18:13:12.242084 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Oct 08 18:13:12 crc kubenswrapper[4988]: I1008 18:13:12.242097 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.333903 4988 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeReady" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.396583 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-hwtzl"] Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.397336 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-7ctzs"] Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.398209 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-7ctzs" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.399129 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-hwtzl" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.400270 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/downloads-7954f5f757-qcgmb"] Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.400749 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-jt7fv"] Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.401376 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-jt7fv" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.402425 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-qcgmb" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.411711 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.413245 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.413710 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.414611 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-zjfnq"] Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.415078 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.415584 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.415763 4988 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.415926 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.416207 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-zjfnq" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.416286 4988 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.416364 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.424163 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-machine-approver/machine-approver-56656f9798-zf8t5"] Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.424989 4988 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.425715 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.426126 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-zf8t5" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.426356 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-6k9kl"] Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.427250 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-6k9kl" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.431322 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console-operator/console-operator-58897d9998-rt48k"] Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.433076 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-xttx4"] Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.433472 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-rt48k" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.433911 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-xttx4" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.456577 4988 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.456714 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-rgqjs"] Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.457444 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-rgqjs" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.457941 4988 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.458025 4988 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.458132 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.458288 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.458326 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.458749 4988 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.458957 4988 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.459158 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.459371 4988 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.460197 4988 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.461611 4988 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.461930 4988 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.462208 4988 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.466974 4988 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.468969 4988 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.470565 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-gqm5m"] Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.471231 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-gqm5m" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.471274 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-gd7f9"] Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.472823 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-gd7f9" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.510198 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-85kp6"] Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.510953 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-85kp6" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.526350 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.526742 4988 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.526896 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.534013 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6r9cj\" (UniqueName: \"kubernetes.io/projected/e294e9c8-16f2-46d4-91f8-5964b1e8ea5a-kube-api-access-6r9cj\") pod \"oauth-openshift-558db77b4-hwtzl\" (UID: \"e294e9c8-16f2-46d4-91f8-5964b1e8ea5a\") " pod="openshift-authentication/oauth-openshift-558db77b4-hwtzl" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.534070 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b2497303-9eb5-4f27-8595-763cb5a851a6-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-6k9kl\" (UID: \"b2497303-9eb5-4f27-8595-763cb5a851a6\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-6k9kl" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.534102 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xj2qd\" (UniqueName: \"kubernetes.io/projected/8deeca47-161b-43bd-a7c4-7f9a41f931d8-kube-api-access-xj2qd\") pod \"authentication-operator-69f744f599-zjfnq\" (UID: \"8deeca47-161b-43bd-a7c4-7f9a41f931d8\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-zjfnq" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.534130 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/8deeca47-161b-43bd-a7c4-7f9a41f931d8-service-ca-bundle\") pod \"authentication-operator-69f744f599-zjfnq\" (UID: \"8deeca47-161b-43bd-a7c4-7f9a41f931d8\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-zjfnq" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.534147 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/e294e9c8-16f2-46d4-91f8-5964b1e8ea5a-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-hwtzl\" (UID: \"e294e9c8-16f2-46d4-91f8-5964b1e8ea5a\") " pod="openshift-authentication/oauth-openshift-558db77b4-hwtzl" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.534168 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/658aaf4b-d6ae-4ebf-bc88-16bd718df0b5-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-jt7fv\" (UID: \"658aaf4b-d6ae-4ebf-bc88-16bd718df0b5\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-jt7fv" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.534187 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/eeee83f5-fe8f-47bf-be7b-cde1ea9ac356-trusted-ca\") pod \"console-operator-58897d9998-rt48k\" (UID: \"eeee83f5-fe8f-47bf-be7b-cde1ea9ac356\") " pod="openshift-console-operator/console-operator-58897d9998-rt48k" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.534207 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/e294e9c8-16f2-46d4-91f8-5964b1e8ea5a-audit-policies\") pod \"oauth-openshift-558db77b4-hwtzl\" (UID: \"e294e9c8-16f2-46d4-91f8-5964b1e8ea5a\") " pod="openshift-authentication/oauth-openshift-558db77b4-hwtzl" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.534223 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/8deeca47-161b-43bd-a7c4-7f9a41f931d8-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-zjfnq\" (UID: \"8deeca47-161b-43bd-a7c4-7f9a41f931d8\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-zjfnq" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.534240 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/63fe20f8-74c8-48ba-8c65-7f284d959254-available-featuregates\") pod \"openshift-config-operator-7777fb866f-7ctzs\" (UID: \"63fe20f8-74c8-48ba-8c65-7f284d959254\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-7ctzs" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.534257 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zprbm\" (UniqueName: \"kubernetes.io/projected/b2497303-9eb5-4f27-8595-763cb5a851a6-kube-api-access-zprbm\") pod \"kube-storage-version-migrator-operator-b67b599dd-6k9kl\" (UID: \"b2497303-9eb5-4f27-8595-763cb5a851a6\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-6k9kl" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.534278 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/5ec13023-d870-43fc-aa01-9cdb2b3015c4-machine-approver-tls\") pod \"machine-approver-56656f9798-zf8t5\" (UID: \"5ec13023-d870-43fc-aa01-9cdb2b3015c4\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-zf8t5" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.534294 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l795v\" (UniqueName: \"kubernetes.io/projected/658aaf4b-d6ae-4ebf-bc88-16bd718df0b5-kube-api-access-l795v\") pod \"cluster-image-registry-operator-dc59b4c8b-jt7fv\" (UID: \"658aaf4b-d6ae-4ebf-bc88-16bd718df0b5\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-jt7fv" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.534314 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/e294e9c8-16f2-46d4-91f8-5964b1e8ea5a-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-hwtzl\" (UID: \"e294e9c8-16f2-46d4-91f8-5964b1e8ea5a\") " pod="openshift-authentication/oauth-openshift-558db77b4-hwtzl" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.534338 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8deeca47-161b-43bd-a7c4-7f9a41f931d8-config\") pod \"authentication-operator-69f744f599-zjfnq\" (UID: \"8deeca47-161b-43bd-a7c4-7f9a41f931d8\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-zjfnq" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.534359 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d8dmg\" (UniqueName: \"kubernetes.io/projected/0bc2144f-4dbe-4dc1-a07a-b764d8582324-kube-api-access-d8dmg\") pod \"downloads-7954f5f757-qcgmb\" (UID: \"0bc2144f-4dbe-4dc1-a07a-b764d8582324\") " pod="openshift-console/downloads-7954f5f757-qcgmb" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.534378 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0ccdb64f-445e-4b91-a4b9-437b22d22853-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-xttx4\" (UID: \"0ccdb64f-445e-4b91-a4b9-437b22d22853\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-xttx4" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.534411 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e294e9c8-16f2-46d4-91f8-5964b1e8ea5a-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-hwtzl\" (UID: \"e294e9c8-16f2-46d4-91f8-5964b1e8ea5a\") " pod="openshift-authentication/oauth-openshift-558db77b4-hwtzl" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.534433 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b2497303-9eb5-4f27-8595-763cb5a851a6-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-6k9kl\" (UID: \"b2497303-9eb5-4f27-8595-763cb5a851a6\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-6k9kl" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.534455 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/e294e9c8-16f2-46d4-91f8-5964b1e8ea5a-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-hwtzl\" (UID: \"e294e9c8-16f2-46d4-91f8-5964b1e8ea5a\") " pod="openshift-authentication/oauth-openshift-558db77b4-hwtzl" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.534474 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/658aaf4b-d6ae-4ebf-bc88-16bd718df0b5-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-jt7fv\" (UID: \"658aaf4b-d6ae-4ebf-bc88-16bd718df0b5\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-jt7fv" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.534494 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/63fe20f8-74c8-48ba-8c65-7f284d959254-serving-cert\") pod \"openshift-config-operator-7777fb866f-7ctzs\" (UID: \"63fe20f8-74c8-48ba-8c65-7f284d959254\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-7ctzs" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.534529 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jvw4s\" (UniqueName: \"kubernetes.io/projected/5ec13023-d870-43fc-aa01-9cdb2b3015c4-kube-api-access-jvw4s\") pod \"machine-approver-56656f9798-zf8t5\" (UID: \"5ec13023-d870-43fc-aa01-9cdb2b3015c4\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-zf8t5" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.534549 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hf5s6\" (UniqueName: \"kubernetes.io/projected/0ccdb64f-445e-4b91-a4b9-437b22d22853-kube-api-access-hf5s6\") pod \"openshift-apiserver-operator-796bbdcf4f-xttx4\" (UID: \"0ccdb64f-445e-4b91-a4b9-437b22d22853\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-xttx4" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.534569 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/e294e9c8-16f2-46d4-91f8-5964b1e8ea5a-audit-dir\") pod \"oauth-openshift-558db77b4-hwtzl\" (UID: \"e294e9c8-16f2-46d4-91f8-5964b1e8ea5a\") " pod="openshift-authentication/oauth-openshift-558db77b4-hwtzl" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.534589 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/e294e9c8-16f2-46d4-91f8-5964b1e8ea5a-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-hwtzl\" (UID: \"e294e9c8-16f2-46d4-91f8-5964b1e8ea5a\") " pod="openshift-authentication/oauth-openshift-558db77b4-hwtzl" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.534608 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/eeee83f5-fe8f-47bf-be7b-cde1ea9ac356-config\") pod \"console-operator-58897d9998-rt48k\" (UID: \"eeee83f5-fe8f-47bf-be7b-cde1ea9ac356\") " pod="openshift-console-operator/console-operator-58897d9998-rt48k" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.534633 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/5ec13023-d870-43fc-aa01-9cdb2b3015c4-auth-proxy-config\") pod \"machine-approver-56656f9798-zf8t5\" (UID: \"5ec13023-d870-43fc-aa01-9cdb2b3015c4\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-zf8t5" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.534652 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pszq7\" (UniqueName: \"kubernetes.io/projected/eeee83f5-fe8f-47bf-be7b-cde1ea9ac356-kube-api-access-pszq7\") pod \"console-operator-58897d9998-rt48k\" (UID: \"eeee83f5-fe8f-47bf-be7b-cde1ea9ac356\") " pod="openshift-console-operator/console-operator-58897d9998-rt48k" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.534668 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/e294e9c8-16f2-46d4-91f8-5964b1e8ea5a-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-hwtzl\" (UID: \"e294e9c8-16f2-46d4-91f8-5964b1e8ea5a\") " pod="openshift-authentication/oauth-openshift-558db77b4-hwtzl" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.534688 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8deeca47-161b-43bd-a7c4-7f9a41f931d8-serving-cert\") pod \"authentication-operator-69f744f599-zjfnq\" (UID: \"8deeca47-161b-43bd-a7c4-7f9a41f931d8\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-zjfnq" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.534714 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/e294e9c8-16f2-46d4-91f8-5964b1e8ea5a-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-hwtzl\" (UID: \"e294e9c8-16f2-46d4-91f8-5964b1e8ea5a\") " pod="openshift-authentication/oauth-openshift-558db77b4-hwtzl" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.534736 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/e294e9c8-16f2-46d4-91f8-5964b1e8ea5a-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-hwtzl\" (UID: \"e294e9c8-16f2-46d4-91f8-5964b1e8ea5a\") " pod="openshift-authentication/oauth-openshift-558db77b4-hwtzl" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.534758 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/e294e9c8-16f2-46d4-91f8-5964b1e8ea5a-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-hwtzl\" (UID: \"e294e9c8-16f2-46d4-91f8-5964b1e8ea5a\") " pod="openshift-authentication/oauth-openshift-558db77b4-hwtzl" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.534779 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/e294e9c8-16f2-46d4-91f8-5964b1e8ea5a-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-hwtzl\" (UID: \"e294e9c8-16f2-46d4-91f8-5964b1e8ea5a\") " pod="openshift-authentication/oauth-openshift-558db77b4-hwtzl" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.534799 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sj2zk\" (UniqueName: \"kubernetes.io/projected/63fe20f8-74c8-48ba-8c65-7f284d959254-kube-api-access-sj2zk\") pod \"openshift-config-operator-7777fb866f-7ctzs\" (UID: \"63fe20f8-74c8-48ba-8c65-7f284d959254\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-7ctzs" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.534819 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0ccdb64f-445e-4b91-a4b9-437b22d22853-config\") pod \"openshift-apiserver-operator-796bbdcf4f-xttx4\" (UID: \"0ccdb64f-445e-4b91-a4b9-437b22d22853\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-xttx4" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.534835 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/eeee83f5-fe8f-47bf-be7b-cde1ea9ac356-serving-cert\") pod \"console-operator-58897d9998-rt48k\" (UID: \"eeee83f5-fe8f-47bf-be7b-cde1ea9ac356\") " pod="openshift-console-operator/console-operator-58897d9998-rt48k" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.534854 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5ec13023-d870-43fc-aa01-9cdb2b3015c4-config\") pod \"machine-approver-56656f9798-zf8t5\" (UID: \"5ec13023-d870-43fc-aa01-9cdb2b3015c4\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-zf8t5" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.534872 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/e294e9c8-16f2-46d4-91f8-5964b1e8ea5a-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-hwtzl\" (UID: \"e294e9c8-16f2-46d4-91f8-5964b1e8ea5a\") " pod="openshift-authentication/oauth-openshift-558db77b4-hwtzl" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.534890 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/658aaf4b-d6ae-4ebf-bc88-16bd718df0b5-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-jt7fv\" (UID: \"658aaf4b-d6ae-4ebf-bc88-16bd718df0b5\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-jt7fv" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.546915 4988 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.547233 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.547492 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.547705 4988 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config" Oct 08 18:13:14 crc kubenswrapper[4988]: W1008 18:13:14.547779 4988 reflector.go:561] object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt": failed to list *v1.ConfigMap: configmaps "kube-root-ca.crt" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "openshift-kube-apiserver-operator": no relationship found between node 'crc' and this object Oct 08 18:13:14 crc kubenswrapper[4988]: E1008 18:13:14.547836 4988 reflector.go:158] "Unhandled Error" err="object-\"openshift-kube-apiserver-operator\"/\"kube-root-ca.crt\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"kube-root-ca.crt\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"openshift-kube-apiserver-operator\": no relationship found between node 'crc' and this object" logger="UnhandledError" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.547889 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"kube-storage-version-migrator-operator-dockercfg-2bh8d" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.547929 4988 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.547942 4988 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.548220 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.550774 4988 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.550778 4988 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.550826 4988 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.550875 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.552103 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-l86nd"] Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.552848 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-l86nd" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.557379 4988 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.557756 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.558049 4988 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.558429 4988 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.558631 4988 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.558833 4988 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.559019 4988 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.559259 4988 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.560175 4988 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.560429 4988 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.561021 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.562579 4988 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"etcd-serving-ca" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.562581 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.562700 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"encryption-config-1" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.562906 4988 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.563896 4988 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.564083 4988 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.564953 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.565102 4988 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.565167 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"etcd-client" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.565513 4988 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"audit-1" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.565571 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.565709 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-dockercfg-r9srn" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.565737 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.565773 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.565836 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.565888 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.565915 4988 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.565933 4988 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.565987 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"serving-cert" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.566017 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.566062 4988 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"trusted-ca-bundle" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.566022 4988 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.565955 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.566113 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.567858 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-k4ghh"] Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.568406 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-k4ghh" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.568843 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-q2w8f"] Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.569239 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-q2w8f" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.584759 4988 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.586851 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-jrt4j"] Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.587743 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-sndzf"] Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.588531 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-sndzf" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.590971 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-jrt4j" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.603980 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.613728 4988 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.614260 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-f9d7485db-7xdnc"] Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.615908 4988 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.616467 4988 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.617046 4988 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.617048 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.618125 4988 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.618331 4988 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.618604 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.618952 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.619182 4988 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.619417 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.619442 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-m55d6"] Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.619805 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"catalog-operator-serving-cert" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.619888 4988 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.619974 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rq7zk" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.619895 4988 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.620290 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.620550 4988 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.620915 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.620998 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.621548 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-7xdnc" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.622562 4988 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.623073 4988 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.623670 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-m55d6" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.632052 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-q2mxf"] Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.634095 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-qmwln"] Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.635621 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-qmwln" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.636170 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-q2mxf" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.636871 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-serving-cert" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.637489 4988 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"console-config" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.648968 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/658aaf4b-d6ae-4ebf-bc88-16bd718df0b5-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-jt7fv\" (UID: \"658aaf4b-d6ae-4ebf-bc88-16bd718df0b5\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-jt7fv" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.649019 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/e294e9c8-16f2-46d4-91f8-5964b1e8ea5a-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-hwtzl\" (UID: \"e294e9c8-16f2-46d4-91f8-5964b1e8ea5a\") " pod="openshift-authentication/oauth-openshift-558db77b4-hwtzl" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.649080 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/63fe20f8-74c8-48ba-8c65-7f284d959254-serving-cert\") pod \"openshift-config-operator-7777fb866f-7ctzs\" (UID: \"63fe20f8-74c8-48ba-8c65-7f284d959254\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-7ctzs" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.649104 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jvw4s\" (UniqueName: \"kubernetes.io/projected/5ec13023-d870-43fc-aa01-9cdb2b3015c4-kube-api-access-jvw4s\") pod \"machine-approver-56656f9798-zf8t5\" (UID: \"5ec13023-d870-43fc-aa01-9cdb2b3015c4\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-zf8t5" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.649147 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hf5s6\" (UniqueName: \"kubernetes.io/projected/0ccdb64f-445e-4b91-a4b9-437b22d22853-kube-api-access-hf5s6\") pod \"openshift-apiserver-operator-796bbdcf4f-xttx4\" (UID: \"0ccdb64f-445e-4b91-a4b9-437b22d22853\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-xttx4" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.649171 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/eeee83f5-fe8f-47bf-be7b-cde1ea9ac356-config\") pod \"console-operator-58897d9998-rt48k\" (UID: \"eeee83f5-fe8f-47bf-be7b-cde1ea9ac356\") " pod="openshift-console-operator/console-operator-58897d9998-rt48k" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.649189 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/e294e9c8-16f2-46d4-91f8-5964b1e8ea5a-audit-dir\") pod \"oauth-openshift-558db77b4-hwtzl\" (UID: \"e294e9c8-16f2-46d4-91f8-5964b1e8ea5a\") " pod="openshift-authentication/oauth-openshift-558db77b4-hwtzl" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.649206 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/e294e9c8-16f2-46d4-91f8-5964b1e8ea5a-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-hwtzl\" (UID: \"e294e9c8-16f2-46d4-91f8-5964b1e8ea5a\") " pod="openshift-authentication/oauth-openshift-558db77b4-hwtzl" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.649231 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5df90a93-b572-4439-b3aa-2666b2873c1d-serving-cert\") pod \"route-controller-manager-6576b87f9c-gd7f9\" (UID: \"5df90a93-b572-4439-b3aa-2666b2873c1d\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-gd7f9" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.649254 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/ba81cb6f-e6fb-49e1-aa4c-5a9e6f8769fb-etcd-client\") pod \"apiserver-7bbb656c7d-85kp6\" (UID: \"ba81cb6f-e6fb-49e1-aa4c-5a9e6f8769fb\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-85kp6" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.649314 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/e294e9c8-16f2-46d4-91f8-5964b1e8ea5a-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-hwtzl\" (UID: \"e294e9c8-16f2-46d4-91f8-5964b1e8ea5a\") " pod="openshift-authentication/oauth-openshift-558db77b4-hwtzl" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.649333 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/ba81cb6f-e6fb-49e1-aa4c-5a9e6f8769fb-audit-dir\") pod \"apiserver-7bbb656c7d-85kp6\" (UID: \"ba81cb6f-e6fb-49e1-aa4c-5a9e6f8769fb\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-85kp6" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.649354 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/5ec13023-d870-43fc-aa01-9cdb2b3015c4-auth-proxy-config\") pod \"machine-approver-56656f9798-zf8t5\" (UID: \"5ec13023-d870-43fc-aa01-9cdb2b3015c4\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-zf8t5" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.649372 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pszq7\" (UniqueName: \"kubernetes.io/projected/eeee83f5-fe8f-47bf-be7b-cde1ea9ac356-kube-api-access-pszq7\") pod \"console-operator-58897d9998-rt48k\" (UID: \"eeee83f5-fe8f-47bf-be7b-cde1ea9ac356\") " pod="openshift-console-operator/console-operator-58897d9998-rt48k" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.649428 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8deeca47-161b-43bd-a7c4-7f9a41f931d8-serving-cert\") pod \"authentication-operator-69f744f599-zjfnq\" (UID: \"8deeca47-161b-43bd-a7c4-7f9a41f931d8\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-zjfnq" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.649444 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/e294e9c8-16f2-46d4-91f8-5964b1e8ea5a-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-hwtzl\" (UID: \"e294e9c8-16f2-46d4-91f8-5964b1e8ea5a\") " pod="openshift-authentication/oauth-openshift-558db77b4-hwtzl" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.649460 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/e294e9c8-16f2-46d4-91f8-5964b1e8ea5a-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-hwtzl\" (UID: \"e294e9c8-16f2-46d4-91f8-5964b1e8ea5a\") " pod="openshift-authentication/oauth-openshift-558db77b4-hwtzl" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.649477 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3021fdc0-7e51-4506-ac3c-09ac3a00c77f-config\") pod \"etcd-operator-b45778765-gqm5m\" (UID: \"3021fdc0-7e51-4506-ac3c-09ac3a00c77f\") " pod="openshift-etcd-operator/etcd-operator-b45778765-gqm5m" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.649503 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/e294e9c8-16f2-46d4-91f8-5964b1e8ea5a-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-hwtzl\" (UID: \"e294e9c8-16f2-46d4-91f8-5964b1e8ea5a\") " pod="openshift-authentication/oauth-openshift-558db77b4-hwtzl" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.649521 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/e294e9c8-16f2-46d4-91f8-5964b1e8ea5a-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-hwtzl\" (UID: \"e294e9c8-16f2-46d4-91f8-5964b1e8ea5a\") " pod="openshift-authentication/oauth-openshift-558db77b4-hwtzl" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.649545 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sj2zk\" (UniqueName: \"kubernetes.io/projected/63fe20f8-74c8-48ba-8c65-7f284d959254-kube-api-access-sj2zk\") pod \"openshift-config-operator-7777fb866f-7ctzs\" (UID: \"63fe20f8-74c8-48ba-8c65-7f284d959254\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-7ctzs" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.649564 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0ccdb64f-445e-4b91-a4b9-437b22d22853-config\") pod \"openshift-apiserver-operator-796bbdcf4f-xttx4\" (UID: \"0ccdb64f-445e-4b91-a4b9-437b22d22853\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-xttx4" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.649580 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/eeee83f5-fe8f-47bf-be7b-cde1ea9ac356-serving-cert\") pod \"console-operator-58897d9998-rt48k\" (UID: \"eeee83f5-fe8f-47bf-be7b-cde1ea9ac356\") " pod="openshift-console-operator/console-operator-58897d9998-rt48k" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.649624 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/ba81cb6f-e6fb-49e1-aa4c-5a9e6f8769fb-encryption-config\") pod \"apiserver-7bbb656c7d-85kp6\" (UID: \"ba81cb6f-e6fb-49e1-aa4c-5a9e6f8769fb\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-85kp6" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.649644 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5df90a93-b572-4439-b3aa-2666b2873c1d-client-ca\") pod \"route-controller-manager-6576b87f9c-gd7f9\" (UID: \"5df90a93-b572-4439-b3aa-2666b2873c1d\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-gd7f9" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.649664 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/ba81cb6f-e6fb-49e1-aa4c-5a9e6f8769fb-audit-policies\") pod \"apiserver-7bbb656c7d-85kp6\" (UID: \"ba81cb6f-e6fb-49e1-aa4c-5a9e6f8769fb\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-85kp6" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.649685 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/e294e9c8-16f2-46d4-91f8-5964b1e8ea5a-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-hwtzl\" (UID: \"e294e9c8-16f2-46d4-91f8-5964b1e8ea5a\") " pod="openshift-authentication/oauth-openshift-558db77b4-hwtzl" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.649711 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5ec13023-d870-43fc-aa01-9cdb2b3015c4-config\") pod \"machine-approver-56656f9798-zf8t5\" (UID: \"5ec13023-d870-43fc-aa01-9cdb2b3015c4\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-zf8t5" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.649738 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/3021fdc0-7e51-4506-ac3c-09ac3a00c77f-etcd-client\") pod \"etcd-operator-b45778765-gqm5m\" (UID: \"3021fdc0-7e51-4506-ac3c-09ac3a00c77f\") " pod="openshift-etcd-operator/etcd-operator-b45778765-gqm5m" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.649780 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/3021fdc0-7e51-4506-ac3c-09ac3a00c77f-etcd-service-ca\") pod \"etcd-operator-b45778765-gqm5m\" (UID: \"3021fdc0-7e51-4506-ac3c-09ac3a00c77f\") " pod="openshift-etcd-operator/etcd-operator-b45778765-gqm5m" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.649806 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/ba81cb6f-e6fb-49e1-aa4c-5a9e6f8769fb-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-85kp6\" (UID: \"ba81cb6f-e6fb-49e1-aa4c-5a9e6f8769fb\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-85kp6" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.649833 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w2vx4\" (UniqueName: \"kubernetes.io/projected/ba81cb6f-e6fb-49e1-aa4c-5a9e6f8769fb-kube-api-access-w2vx4\") pod \"apiserver-7bbb656c7d-85kp6\" (UID: \"ba81cb6f-e6fb-49e1-aa4c-5a9e6f8769fb\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-85kp6" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.649865 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/658aaf4b-d6ae-4ebf-bc88-16bd718df0b5-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-jt7fv\" (UID: \"658aaf4b-d6ae-4ebf-bc88-16bd718df0b5\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-jt7fv" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.649905 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s5hnj\" (UniqueName: \"kubernetes.io/projected/5df90a93-b572-4439-b3aa-2666b2873c1d-kube-api-access-s5hnj\") pod \"route-controller-manager-6576b87f9c-gd7f9\" (UID: \"5df90a93-b572-4439-b3aa-2666b2873c1d\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-gd7f9" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.649944 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3021fdc0-7e51-4506-ac3c-09ac3a00c77f-serving-cert\") pod \"etcd-operator-b45778765-gqm5m\" (UID: \"3021fdc0-7e51-4506-ac3c-09ac3a00c77f\") " pod="openshift-etcd-operator/etcd-operator-b45778765-gqm5m" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.650339 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-4wh9p"] Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.651508 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-4wh9p" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.649983 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6r9cj\" (UniqueName: \"kubernetes.io/projected/e294e9c8-16f2-46d4-91f8-5964b1e8ea5a-kube-api-access-6r9cj\") pod \"oauth-openshift-558db77b4-hwtzl\" (UID: \"e294e9c8-16f2-46d4-91f8-5964b1e8ea5a\") " pod="openshift-authentication/oauth-openshift-558db77b4-hwtzl" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.652208 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b2497303-9eb5-4f27-8595-763cb5a851a6-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-6k9kl\" (UID: \"b2497303-9eb5-4f27-8595-763cb5a851a6\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-6k9kl" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.652240 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5df90a93-b572-4439-b3aa-2666b2873c1d-config\") pod \"route-controller-manager-6576b87f9c-gd7f9\" (UID: \"5df90a93-b572-4439-b3aa-2666b2873c1d\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-gd7f9" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.652274 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xj2qd\" (UniqueName: \"kubernetes.io/projected/8deeca47-161b-43bd-a7c4-7f9a41f931d8-kube-api-access-xj2qd\") pod \"authentication-operator-69f744f599-zjfnq\" (UID: \"8deeca47-161b-43bd-a7c4-7f9a41f931d8\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-zjfnq" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.652311 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/8deeca47-161b-43bd-a7c4-7f9a41f931d8-service-ca-bundle\") pod \"authentication-operator-69f744f599-zjfnq\" (UID: \"8deeca47-161b-43bd-a7c4-7f9a41f931d8\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-zjfnq" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.652330 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/e294e9c8-16f2-46d4-91f8-5964b1e8ea5a-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-hwtzl\" (UID: \"e294e9c8-16f2-46d4-91f8-5964b1e8ea5a\") " pod="openshift-authentication/oauth-openshift-558db77b4-hwtzl" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.652353 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/e294e9c8-16f2-46d4-91f8-5964b1e8ea5a-audit-policies\") pod \"oauth-openshift-558db77b4-hwtzl\" (UID: \"e294e9c8-16f2-46d4-91f8-5964b1e8ea5a\") " pod="openshift-authentication/oauth-openshift-558db77b4-hwtzl" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.652372 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/ba81cb6f-e6fb-49e1-aa4c-5a9e6f8769fb-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-85kp6\" (UID: \"ba81cb6f-e6fb-49e1-aa4c-5a9e6f8769fb\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-85kp6" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.653506 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b2497303-9eb5-4f27-8595-763cb5a851a6-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-6k9kl\" (UID: \"b2497303-9eb5-4f27-8595-763cb5a851a6\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-6k9kl" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.653704 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-oauth-config" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.654055 4988 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"service-ca" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.653586 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/658aaf4b-d6ae-4ebf-bc88-16bd718df0b5-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-jt7fv\" (UID: \"658aaf4b-d6ae-4ebf-bc88-16bd718df0b5\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-jt7fv" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.654561 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/8deeca47-161b-43bd-a7c4-7f9a41f931d8-service-ca-bundle\") pod \"authentication-operator-69f744f599-zjfnq\" (UID: \"8deeca47-161b-43bd-a7c4-7f9a41f931d8\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-zjfnq" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.655252 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/e294e9c8-16f2-46d4-91f8-5964b1e8ea5a-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-hwtzl\" (UID: \"e294e9c8-16f2-46d4-91f8-5964b1e8ea5a\") " pod="openshift-authentication/oauth-openshift-558db77b4-hwtzl" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.655822 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/e294e9c8-16f2-46d4-91f8-5964b1e8ea5a-audit-policies\") pod \"oauth-openshift-558db77b4-hwtzl\" (UID: \"e294e9c8-16f2-46d4-91f8-5964b1e8ea5a\") " pod="openshift-authentication/oauth-openshift-558db77b4-hwtzl" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.655838 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-5mrrc"] Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.655917 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/eeee83f5-fe8f-47bf-be7b-cde1ea9ac356-trusted-ca\") pod \"console-operator-58897d9998-rt48k\" (UID: \"eeee83f5-fe8f-47bf-be7b-cde1ea9ac356\") " pod="openshift-console-operator/console-operator-58897d9998-rt48k" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.655978 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/8deeca47-161b-43bd-a7c4-7f9a41f931d8-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-zjfnq\" (UID: \"8deeca47-161b-43bd-a7c4-7f9a41f931d8\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-zjfnq" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.656018 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/63fe20f8-74c8-48ba-8c65-7f284d959254-available-featuregates\") pod \"openshift-config-operator-7777fb866f-7ctzs\" (UID: \"63fe20f8-74c8-48ba-8c65-7f284d959254\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-7ctzs" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.656873 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-5mrrc" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.656910 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-984jv"] Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.657662 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/eeee83f5-fe8f-47bf-be7b-cde1ea9ac356-trusted-ca\") pod \"console-operator-58897d9998-rt48k\" (UID: \"eeee83f5-fe8f-47bf-be7b-cde1ea9ac356\") " pod="openshift-console-operator/console-operator-58897d9998-rt48k" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.657865 4988 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.658287 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/63fe20f8-74c8-48ba-8c65-7f284d959254-available-featuregates\") pod \"openshift-config-operator-7777fb866f-7ctzs\" (UID: \"63fe20f8-74c8-48ba-8c65-7f284d959254\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-7ctzs" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.658466 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zprbm\" (UniqueName: \"kubernetes.io/projected/b2497303-9eb5-4f27-8595-763cb5a851a6-kube-api-access-zprbm\") pod \"kube-storage-version-migrator-operator-b67b599dd-6k9kl\" (UID: \"b2497303-9eb5-4f27-8595-763cb5a851a6\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-6k9kl" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.658522 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1ca33297-d58c-4493-b1d3-45dd83983be4-config\") pod \"kube-apiserver-operator-766d6c64bb-rgqjs\" (UID: \"1ca33297-d58c-4493-b1d3-45dd83983be4\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-rgqjs" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.658637 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ba81cb6f-e6fb-49e1-aa4c-5a9e6f8769fb-serving-cert\") pod \"apiserver-7bbb656c7d-85kp6\" (UID: \"ba81cb6f-e6fb-49e1-aa4c-5a9e6f8769fb\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-85kp6" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.658657 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/5ec13023-d870-43fc-aa01-9cdb2b3015c4-machine-approver-tls\") pod \"machine-approver-56656f9798-zf8t5\" (UID: \"5ec13023-d870-43fc-aa01-9cdb2b3015c4\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-zf8t5" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.658688 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l795v\" (UniqueName: \"kubernetes.io/projected/658aaf4b-d6ae-4ebf-bc88-16bd718df0b5-kube-api-access-l795v\") pod \"cluster-image-registry-operator-dc59b4c8b-jt7fv\" (UID: \"658aaf4b-d6ae-4ebf-bc88-16bd718df0b5\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-jt7fv" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.658728 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/e294e9c8-16f2-46d4-91f8-5964b1e8ea5a-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-hwtzl\" (UID: \"e294e9c8-16f2-46d4-91f8-5964b1e8ea5a\") " pod="openshift-authentication/oauth-openshift-558db77b4-hwtzl" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.658749 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1ca33297-d58c-4493-b1d3-45dd83983be4-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-rgqjs\" (UID: \"1ca33297-d58c-4493-b1d3-45dd83983be4\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-rgqjs" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.658807 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8deeca47-161b-43bd-a7c4-7f9a41f931d8-config\") pod \"authentication-operator-69f744f599-zjfnq\" (UID: \"8deeca47-161b-43bd-a7c4-7f9a41f931d8\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-zjfnq" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.659525 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/8deeca47-161b-43bd-a7c4-7f9a41f931d8-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-zjfnq\" (UID: \"8deeca47-161b-43bd-a7c4-7f9a41f931d8\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-zjfnq" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.659769 4988 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.659785 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d9zcl\" (UniqueName: \"kubernetes.io/projected/3021fdc0-7e51-4506-ac3c-09ac3a00c77f-kube-api-access-d9zcl\") pod \"etcd-operator-b45778765-gqm5m\" (UID: \"3021fdc0-7e51-4506-ac3c-09ac3a00c77f\") " pod="openshift-etcd-operator/etcd-operator-b45778765-gqm5m" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.659829 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d8dmg\" (UniqueName: \"kubernetes.io/projected/0bc2144f-4dbe-4dc1-a07a-b764d8582324-kube-api-access-d8dmg\") pod \"downloads-7954f5f757-qcgmb\" (UID: \"0bc2144f-4dbe-4dc1-a07a-b764d8582324\") " pod="openshift-console/downloads-7954f5f757-qcgmb" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.660442 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-984jv" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.661019 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8deeca47-161b-43bd-a7c4-7f9a41f931d8-config\") pod \"authentication-operator-69f744f599-zjfnq\" (UID: \"8deeca47-161b-43bd-a7c4-7f9a41f931d8\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-zjfnq" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.661115 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0ccdb64f-445e-4b91-a4b9-437b22d22853-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-xttx4\" (UID: \"0ccdb64f-445e-4b91-a4b9-437b22d22853\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-xttx4" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.661297 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5ec13023-d870-43fc-aa01-9cdb2b3015c4-config\") pod \"machine-approver-56656f9798-zf8t5\" (UID: \"5ec13023-d870-43fc-aa01-9cdb2b3015c4\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-zf8t5" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.663741 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/e294e9c8-16f2-46d4-91f8-5964b1e8ea5a-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-hwtzl\" (UID: \"e294e9c8-16f2-46d4-91f8-5964b1e8ea5a\") " pod="openshift-authentication/oauth-openshift-558db77b4-hwtzl" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.664726 4988 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"trusted-ca-bundle" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.665793 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress/router-default-5444994796-jjnkn"] Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.667545 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0ccdb64f-445e-4b91-a4b9-437b22d22853-config\") pod \"openshift-apiserver-operator-796bbdcf4f-xttx4\" (UID: \"0ccdb64f-445e-4b91-a4b9-437b22d22853\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-xttx4" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.666147 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/e294e9c8-16f2-46d4-91f8-5964b1e8ea5a-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-hwtzl\" (UID: \"e294e9c8-16f2-46d4-91f8-5964b1e8ea5a\") " pod="openshift-authentication/oauth-openshift-558db77b4-hwtzl" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.667579 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0ccdb64f-445e-4b91-a4b9-437b22d22853-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-xttx4\" (UID: \"0ccdb64f-445e-4b91-a4b9-437b22d22853\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-xttx4" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.668992 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/e294e9c8-16f2-46d4-91f8-5964b1e8ea5a-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-hwtzl\" (UID: \"e294e9c8-16f2-46d4-91f8-5964b1e8ea5a\") " pod="openshift-authentication/oauth-openshift-558db77b4-hwtzl" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.669570 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/e294e9c8-16f2-46d4-91f8-5964b1e8ea5a-audit-dir\") pod \"oauth-openshift-558db77b4-hwtzl\" (UID: \"e294e9c8-16f2-46d4-91f8-5964b1e8ea5a\") " pod="openshift-authentication/oauth-openshift-558db77b4-hwtzl" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.669710 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e294e9c8-16f2-46d4-91f8-5964b1e8ea5a-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-hwtzl\" (UID: \"e294e9c8-16f2-46d4-91f8-5964b1e8ea5a\") " pod="openshift-authentication/oauth-openshift-558db77b4-hwtzl" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.669759 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1ca33297-d58c-4493-b1d3-45dd83983be4-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-rgqjs\" (UID: \"1ca33297-d58c-4493-b1d3-45dd83983be4\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-rgqjs" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.669791 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/3021fdc0-7e51-4506-ac3c-09ac3a00c77f-etcd-ca\") pod \"etcd-operator-b45778765-gqm5m\" (UID: \"3021fdc0-7e51-4506-ac3c-09ac3a00c77f\") " pod="openshift-etcd-operator/etcd-operator-b45778765-gqm5m" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.669814 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/5ec13023-d870-43fc-aa01-9cdb2b3015c4-auth-proxy-config\") pod \"machine-approver-56656f9798-zf8t5\" (UID: \"5ec13023-d870-43fc-aa01-9cdb2b3015c4\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-zf8t5" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.669763 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/63fe20f8-74c8-48ba-8c65-7f284d959254-serving-cert\") pod \"openshift-config-operator-7777fb866f-7ctzs\" (UID: \"63fe20f8-74c8-48ba-8c65-7f284d959254\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-7ctzs" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.670268 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/e294e9c8-16f2-46d4-91f8-5964b1e8ea5a-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-hwtzl\" (UID: \"e294e9c8-16f2-46d4-91f8-5964b1e8ea5a\") " pod="openshift-authentication/oauth-openshift-558db77b4-hwtzl" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.670645 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e294e9c8-16f2-46d4-91f8-5964b1e8ea5a-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-hwtzl\" (UID: \"e294e9c8-16f2-46d4-91f8-5964b1e8ea5a\") " pod="openshift-authentication/oauth-openshift-558db77b4-hwtzl" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.670671 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/5ec13023-d870-43fc-aa01-9cdb2b3015c4-machine-approver-tls\") pod \"machine-approver-56656f9798-zf8t5\" (UID: \"5ec13023-d870-43fc-aa01-9cdb2b3015c4\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-zf8t5" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.670708 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b2497303-9eb5-4f27-8595-763cb5a851a6-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-6k9kl\" (UID: \"b2497303-9eb5-4f27-8595-763cb5a851a6\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-6k9kl" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.670774 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/eeee83f5-fe8f-47bf-be7b-cde1ea9ac356-config\") pod \"console-operator-58897d9998-rt48k\" (UID: \"eeee83f5-fe8f-47bf-be7b-cde1ea9ac356\") " pod="openshift-console-operator/console-operator-58897d9998-rt48k" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.671280 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-jjnkn" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.665896 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.672226 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/658aaf4b-d6ae-4ebf-bc88-16bd718df0b5-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-jt7fv\" (UID: \"658aaf4b-d6ae-4ebf-bc88-16bd718df0b5\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-jt7fv" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.673224 4988 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"oauth-serving-cert" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.673283 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-bbpzz"] Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.673358 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/e294e9c8-16f2-46d4-91f8-5964b1e8ea5a-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-hwtzl\" (UID: \"e294e9c8-16f2-46d4-91f8-5964b1e8ea5a\") " pod="openshift-authentication/oauth-openshift-558db77b4-hwtzl" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.673956 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/e294e9c8-16f2-46d4-91f8-5964b1e8ea5a-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-hwtzl\" (UID: \"e294e9c8-16f2-46d4-91f8-5964b1e8ea5a\") " pod="openshift-authentication/oauth-openshift-558db77b4-hwtzl" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.674151 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-wnkxn"] Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.674370 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-bbpzz" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.674785 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-wnkxn" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.674997 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-2d794"] Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.675715 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-2d794" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.676751 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/e294e9c8-16f2-46d4-91f8-5964b1e8ea5a-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-hwtzl\" (UID: \"e294e9c8-16f2-46d4-91f8-5964b1e8ea5a\") " pod="openshift-authentication/oauth-openshift-558db77b4-hwtzl" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.676938 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-h5cw8"] Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.677600 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/658aaf4b-d6ae-4ebf-bc88-16bd718df0b5-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-jt7fv\" (UID: \"658aaf4b-d6ae-4ebf-bc88-16bd718df0b5\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-jt7fv" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.677936 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-h5cw8" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.678007 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-g7fbv"] Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.678552 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/e294e9c8-16f2-46d4-91f8-5964b1e8ea5a-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-hwtzl\" (UID: \"e294e9c8-16f2-46d4-91f8-5964b1e8ea5a\") " pod="openshift-authentication/oauth-openshift-558db77b4-hwtzl" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.678700 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-g7fbv" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.678915 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-d9ks8"] Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.679073 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b2497303-9eb5-4f27-8595-763cb5a851a6-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-6k9kl\" (UID: \"b2497303-9eb5-4f27-8595-763cb5a851a6\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-6k9kl" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.679343 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-d9ks8" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.679734 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8deeca47-161b-43bd-a7c4-7f9a41f931d8-serving-cert\") pod \"authentication-operator-69f744f599-zjfnq\" (UID: \"8deeca47-161b-43bd-a7c4-7f9a41f931d8\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-zjfnq" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.680574 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-hgnvq"] Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.681676 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-wxv6d"] Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.681785 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-hgnvq" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.682188 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/e294e9c8-16f2-46d4-91f8-5964b1e8ea5a-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-hwtzl\" (UID: \"e294e9c8-16f2-46d4-91f8-5964b1e8ea5a\") " pod="openshift-authentication/oauth-openshift-558db77b4-hwtzl" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.682561 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-g775v"] Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.682624 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-wxv6d" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.683050 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-g775v" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.683341 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-4phs4"] Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.684114 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/eeee83f5-fe8f-47bf-be7b-cde1ea9ac356-serving-cert\") pod \"console-operator-58897d9998-rt48k\" (UID: \"eeee83f5-fe8f-47bf-be7b-cde1ea9ac356\") " pod="openshift-console-operator/console-operator-58897d9998-rt48k" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.684166 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-4phs4" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.684635 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29332440-qxt8r"] Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.685273 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29332440-qxt8r" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.686118 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.686404 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-294dt"] Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.687979 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-hwtzl"] Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.688103 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-294dt" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.688558 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-6k9kl"] Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.691771 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-zjfnq"] Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.691835 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-7ctzs"] Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.691847 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-jt7fv"] Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.696312 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-rt48k"] Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.696337 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-gqm5m"] Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.696349 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-rgqjs"] Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.697592 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-qcgmb"] Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.698506 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-canary/ingress-canary-s4r2k"] Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.699297 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-s4r2k" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.699935 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-85kp6"] Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.701241 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-sndzf"] Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.702489 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-q2w8f"] Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.703447 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-7xdnc"] Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.704455 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-l86nd"] Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.711515 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-gd7f9"] Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.711657 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-xttx4"] Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.718596 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-qmwln"] Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.718915 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-984jv"] Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.721944 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-4wh9p"] Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.723945 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-server-5hj6s"] Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.725792 4988 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.728351 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-5hj6s" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.730985 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-hgnvq"] Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.733721 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-m55d6"] Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.735944 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-bbpzz"] Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.736002 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-h5cw8"] Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.738063 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-2d794"] Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.738862 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-294dt"] Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.740219 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-wxv6d"] Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.741987 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-5mrrc"] Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.742294 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-jrt4j"] Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.742483 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-dockercfg-qt55r" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.744637 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-4phs4"] Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.746201 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-q2mxf"] Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.747522 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-wnkxn"] Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.750997 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29332440-qxt8r"] Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.754336 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-g7fbv"] Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.759003 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-d9ks8"] Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.760528 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-g775v"] Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.762149 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-k4ghh"] Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.763666 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/dns-default-kwcts"] Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.764626 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-kwcts" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.764663 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.765366 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-s4r2k"] Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.765910 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-kwcts"] Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.770713 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-gl7lr"] Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.771410 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5df90a93-b572-4439-b3aa-2666b2873c1d-serving-cert\") pod \"route-controller-manager-6576b87f9c-gd7f9\" (UID: \"5df90a93-b572-4439-b3aa-2666b2873c1d\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-gd7f9" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.771443 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c83b25d6-a1c1-4394-9954-697156fe53b7-config\") pod \"kube-controller-manager-operator-78b949d7b-l86nd\" (UID: \"c83b25d6-a1c1-4394-9954-697156fe53b7\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-l86nd" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.771468 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/ba81cb6f-e6fb-49e1-aa4c-5a9e6f8769fb-etcd-client\") pod \"apiserver-7bbb656c7d-85kp6\" (UID: \"ba81cb6f-e6fb-49e1-aa4c-5a9e6f8769fb\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-85kp6" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.771487 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/b5cd6079-031c-4cc8-8161-88f0e6f84064-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-m55d6\" (UID: \"b5cd6079-031c-4cc8-8161-88f0e6f84064\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-m55d6" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.771514 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/ba81cb6f-e6fb-49e1-aa4c-5a9e6f8769fb-audit-dir\") pod \"apiserver-7bbb656c7d-85kp6\" (UID: \"ba81cb6f-e6fb-49e1-aa4c-5a9e6f8769fb\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-85kp6" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.771529 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/aa86616c-479c-45be-aaa3-d3c4994721bf-metrics-tls\") pod \"ingress-operator-5b745b69d9-q2mxf\" (UID: \"aa86616c-479c-45be-aaa3-d3c4994721bf\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-q2mxf" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.771543 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/aa86616c-479c-45be-aaa3-d3c4994721bf-bound-sa-token\") pod \"ingress-operator-5b745b69d9-q2mxf\" (UID: \"aa86616c-479c-45be-aaa3-d3c4994721bf\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-q2mxf" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.771561 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/6c651396-5e8f-4920-a2b9-75f2c40eeea8-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-qmwln\" (UID: \"6c651396-5e8f-4920-a2b9-75f2c40eeea8\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-qmwln" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.771576 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/9c8874ed-3f70-45b6-9f10-d8737b6a7a8f-console-config\") pod \"console-f9d7485db-7xdnc\" (UID: \"9c8874ed-3f70-45b6-9f10-d8737b6a7a8f\") " pod="openshift-console/console-f9d7485db-7xdnc" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.771591 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l5tq9\" (UniqueName: \"kubernetes.io/projected/4719309b-a3b7-4910-a389-635acc09d01d-kube-api-access-l5tq9\") pod \"openshift-controller-manager-operator-756b6f6bc6-k4ghh\" (UID: \"4719309b-a3b7-4910-a389-635acc09d01d\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-k4ghh" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.771611 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3021fdc0-7e51-4506-ac3c-09ac3a00c77f-config\") pod \"etcd-operator-b45778765-gqm5m\" (UID: \"3021fdc0-7e51-4506-ac3c-09ac3a00c77f\") " pod="openshift-etcd-operator/etcd-operator-b45778765-gqm5m" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.771634 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/d0df68ec-1481-4992-bc83-16b4a6b85f0b-profile-collector-cert\") pod \"catalog-operator-68c6474976-jrt4j\" (UID: \"d0df68ec-1481-4992-bc83-16b4a6b85f0b\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-jrt4j" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.771654 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/ba81cb6f-e6fb-49e1-aa4c-5a9e6f8769fb-audit-policies\") pod \"apiserver-7bbb656c7d-85kp6\" (UID: \"ba81cb6f-e6fb-49e1-aa4c-5a9e6f8769fb\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-85kp6" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.771673 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/ba81cb6f-e6fb-49e1-aa4c-5a9e6f8769fb-encryption-config\") pod \"apiserver-7bbb656c7d-85kp6\" (UID: \"ba81cb6f-e6fb-49e1-aa4c-5a9e6f8769fb\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-85kp6" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.771692 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5df90a93-b572-4439-b3aa-2666b2873c1d-client-ca\") pod \"route-controller-manager-6576b87f9c-gd7f9\" (UID: \"5df90a93-b572-4439-b3aa-2666b2873c1d\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-gd7f9" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.771710 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b8nbn\" (UniqueName: \"kubernetes.io/projected/d0df68ec-1481-4992-bc83-16b4a6b85f0b-kube-api-access-b8nbn\") pod \"catalog-operator-68c6474976-jrt4j\" (UID: \"d0df68ec-1481-4992-bc83-16b4a6b85f0b\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-jrt4j" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.771728 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/aa86616c-479c-45be-aaa3-d3c4994721bf-trusted-ca\") pod \"ingress-operator-5b745b69d9-q2mxf\" (UID: \"aa86616c-479c-45be-aaa3-d3c4994721bf\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-q2mxf" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.771748 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/a818c760-46b8-4f91-89b3-18e4b0c9a649-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-sndzf\" (UID: \"a818c760-46b8-4f91-89b3-18e4b0c9a649\") " pod="openshift-controller-manager/controller-manager-879f6c89f-sndzf" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.771767 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4719309b-a3b7-4910-a389-635acc09d01d-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-k4ghh\" (UID: \"4719309b-a3b7-4910-a389-635acc09d01d\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-k4ghh" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.771787 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/3021fdc0-7e51-4506-ac3c-09ac3a00c77f-etcd-client\") pod \"etcd-operator-b45778765-gqm5m\" (UID: \"3021fdc0-7e51-4506-ac3c-09ac3a00c77f\") " pod="openshift-etcd-operator/etcd-operator-b45778765-gqm5m" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.771807 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/3021fdc0-7e51-4506-ac3c-09ac3a00c77f-etcd-service-ca\") pod \"etcd-operator-b45778765-gqm5m\" (UID: \"3021fdc0-7e51-4506-ac3c-09ac3a00c77f\") " pod="openshift-etcd-operator/etcd-operator-b45778765-gqm5m" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.771830 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/ba81cb6f-e6fb-49e1-aa4c-5a9e6f8769fb-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-85kp6\" (UID: \"ba81cb6f-e6fb-49e1-aa4c-5a9e6f8769fb\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-85kp6" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.771848 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w2vx4\" (UniqueName: \"kubernetes.io/projected/ba81cb6f-e6fb-49e1-aa4c-5a9e6f8769fb-kube-api-access-w2vx4\") pod \"apiserver-7bbb656c7d-85kp6\" (UID: \"ba81cb6f-e6fb-49e1-aa4c-5a9e6f8769fb\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-85kp6" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.771865 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b5cd6079-031c-4cc8-8161-88f0e6f84064-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-m55d6\" (UID: \"b5cd6079-031c-4cc8-8161-88f0e6f84064\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-m55d6" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.771888 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s5hnj\" (UniqueName: \"kubernetes.io/projected/5df90a93-b572-4439-b3aa-2666b2873c1d-kube-api-access-s5hnj\") pod \"route-controller-manager-6576b87f9c-gd7f9\" (UID: \"5df90a93-b572-4439-b3aa-2666b2873c1d\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-gd7f9" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.771904 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3021fdc0-7e51-4506-ac3c-09ac3a00c77f-serving-cert\") pod \"etcd-operator-b45778765-gqm5m\" (UID: \"3021fdc0-7e51-4506-ac3c-09ac3a00c77f\") " pod="openshift-etcd-operator/etcd-operator-b45778765-gqm5m" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.771929 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5df90a93-b572-4439-b3aa-2666b2873c1d-config\") pod \"route-controller-manager-6576b87f9c-gd7f9\" (UID: \"5df90a93-b572-4439-b3aa-2666b2873c1d\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-gd7f9" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.771946 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/6c651396-5e8f-4920-a2b9-75f2c40eeea8-proxy-tls\") pod \"machine-config-controller-84d6567774-qmwln\" (UID: \"6c651396-5e8f-4920-a2b9-75f2c40eeea8\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-qmwln" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.771984 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4kxgh\" (UniqueName: \"kubernetes.io/projected/a818c760-46b8-4f91-89b3-18e4b0c9a649-kube-api-access-4kxgh\") pod \"controller-manager-879f6c89f-sndzf\" (UID: \"a818c760-46b8-4f91-89b3-18e4b0c9a649\") " pod="openshift-controller-manager/controller-manager-879f6c89f-sndzf" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.772001 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tmtkc\" (UniqueName: \"kubernetes.io/projected/6c651396-5e8f-4920-a2b9-75f2c40eeea8-kube-api-access-tmtkc\") pod \"machine-config-controller-84d6567774-qmwln\" (UID: \"6c651396-5e8f-4920-a2b9-75f2c40eeea8\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-qmwln" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.772019 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/ba81cb6f-e6fb-49e1-aa4c-5a9e6f8769fb-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-85kp6\" (UID: \"ba81cb6f-e6fb-49e1-aa4c-5a9e6f8769fb\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-85kp6" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.772035 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/9c8874ed-3f70-45b6-9f10-d8737b6a7a8f-oauth-serving-cert\") pod \"console-f9d7485db-7xdnc\" (UID: \"9c8874ed-3f70-45b6-9f10-d8737b6a7a8f\") " pod="openshift-console/console-f9d7485db-7xdnc" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.772053 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/d0df68ec-1481-4992-bc83-16b4a6b85f0b-srv-cert\") pod \"catalog-operator-68c6474976-jrt4j\" (UID: \"d0df68ec-1481-4992-bc83-16b4a6b85f0b\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-jrt4j" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.772068 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/9c8874ed-3f70-45b6-9f10-d8737b6a7a8f-trusted-ca-bundle\") pod \"console-f9d7485db-7xdnc\" (UID: \"9c8874ed-3f70-45b6-9f10-d8737b6a7a8f\") " pod="openshift-console/console-f9d7485db-7xdnc" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.772083 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b5cd6079-031c-4cc8-8161-88f0e6f84064-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-m55d6\" (UID: \"b5cd6079-031c-4cc8-8161-88f0e6f84064\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-m55d6" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.772099 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/a818c760-46b8-4f91-89b3-18e4b0c9a649-client-ca\") pod \"controller-manager-879f6c89f-sndzf\" (UID: \"a818c760-46b8-4f91-89b3-18e4b0c9a649\") " pod="openshift-controller-manager/controller-manager-879f6c89f-sndzf" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.772122 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ba81cb6f-e6fb-49e1-aa4c-5a9e6f8769fb-serving-cert\") pod \"apiserver-7bbb656c7d-85kp6\" (UID: \"ba81cb6f-e6fb-49e1-aa4c-5a9e6f8769fb\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-85kp6" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.772139 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1ca33297-d58c-4493-b1d3-45dd83983be4-config\") pod \"kube-apiserver-operator-766d6c64bb-rgqjs\" (UID: \"1ca33297-d58c-4493-b1d3-45dd83983be4\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-rgqjs" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.772154 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a818c760-46b8-4f91-89b3-18e4b0c9a649-config\") pod \"controller-manager-879f6c89f-sndzf\" (UID: \"a818c760-46b8-4f91-89b3-18e4b0c9a649\") " pod="openshift-controller-manager/controller-manager-879f6c89f-sndzf" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.772183 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1ca33297-d58c-4493-b1d3-45dd83983be4-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-rgqjs\" (UID: \"1ca33297-d58c-4493-b1d3-45dd83983be4\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-rgqjs" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.772200 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d9zcl\" (UniqueName: \"kubernetes.io/projected/3021fdc0-7e51-4506-ac3c-09ac3a00c77f-kube-api-access-d9zcl\") pod \"etcd-operator-b45778765-gqm5m\" (UID: \"3021fdc0-7e51-4506-ac3c-09ac3a00c77f\") " pod="openshift-etcd-operator/etcd-operator-b45778765-gqm5m" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.772219 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/9c8874ed-3f70-45b6-9f10-d8737b6a7a8f-console-oauth-config\") pod \"console-f9d7485db-7xdnc\" (UID: \"9c8874ed-3f70-45b6-9f10-d8737b6a7a8f\") " pod="openshift-console/console-f9d7485db-7xdnc" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.772234 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/9c8874ed-3f70-45b6-9f10-d8737b6a7a8f-service-ca\") pod \"console-f9d7485db-7xdnc\" (UID: \"9c8874ed-3f70-45b6-9f10-d8737b6a7a8f\") " pod="openshift-console/console-f9d7485db-7xdnc" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.772250 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/0769dc07-8f56-42a5-8eb5-b900eefaaf1e-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-q2w8f\" (UID: \"0769dc07-8f56-42a5-8eb5-b900eefaaf1e\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-q2w8f" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.772267 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4719309b-a3b7-4910-a389-635acc09d01d-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-k4ghh\" (UID: \"4719309b-a3b7-4910-a389-635acc09d01d\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-k4ghh" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.772282 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c83b25d6-a1c1-4394-9954-697156fe53b7-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-l86nd\" (UID: \"c83b25d6-a1c1-4394-9954-697156fe53b7\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-l86nd" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.772298 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/c83b25d6-a1c1-4394-9954-697156fe53b7-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-l86nd\" (UID: \"c83b25d6-a1c1-4394-9954-697156fe53b7\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-l86nd" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.772323 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1ca33297-d58c-4493-b1d3-45dd83983be4-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-rgqjs\" (UID: \"1ca33297-d58c-4493-b1d3-45dd83983be4\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-rgqjs" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.772339 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/3021fdc0-7e51-4506-ac3c-09ac3a00c77f-etcd-ca\") pod \"etcd-operator-b45778765-gqm5m\" (UID: \"3021fdc0-7e51-4506-ac3c-09ac3a00c77f\") " pod="openshift-etcd-operator/etcd-operator-b45778765-gqm5m" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.772356 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t45mv\" (UniqueName: \"kubernetes.io/projected/0769dc07-8f56-42a5-8eb5-b900eefaaf1e-kube-api-access-t45mv\") pod \"multus-admission-controller-857f4d67dd-q2w8f\" (UID: \"0769dc07-8f56-42a5-8eb5-b900eefaaf1e\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-q2w8f" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.772380 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a818c760-46b8-4f91-89b3-18e4b0c9a649-serving-cert\") pod \"controller-manager-879f6c89f-sndzf\" (UID: \"a818c760-46b8-4f91-89b3-18e4b0c9a649\") " pod="openshift-controller-manager/controller-manager-879f6c89f-sndzf" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.772419 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/9c8874ed-3f70-45b6-9f10-d8737b6a7a8f-console-serving-cert\") pod \"console-f9d7485db-7xdnc\" (UID: \"9c8874ed-3f70-45b6-9f10-d8737b6a7a8f\") " pod="openshift-console/console-f9d7485db-7xdnc" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.772437 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qhgxr\" (UniqueName: \"kubernetes.io/projected/9c8874ed-3f70-45b6-9f10-d8737b6a7a8f-kube-api-access-qhgxr\") pod \"console-f9d7485db-7xdnc\" (UID: \"9c8874ed-3f70-45b6-9f10-d8737b6a7a8f\") " pod="openshift-console/console-f9d7485db-7xdnc" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.772453 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bqs24\" (UniqueName: \"kubernetes.io/projected/aa86616c-479c-45be-aaa3-d3c4994721bf-kube-api-access-bqs24\") pod \"ingress-operator-5b745b69d9-q2mxf\" (UID: \"aa86616c-479c-45be-aaa3-d3c4994721bf\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-q2mxf" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.772576 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3021fdc0-7e51-4506-ac3c-09ac3a00c77f-config\") pod \"etcd-operator-b45778765-gqm5m\" (UID: \"3021fdc0-7e51-4506-ac3c-09ac3a00c77f\") " pod="openshift-etcd-operator/etcd-operator-b45778765-gqm5m" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.773783 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/ba81cb6f-e6fb-49e1-aa4c-5a9e6f8769fb-audit-policies\") pod \"apiserver-7bbb656c7d-85kp6\" (UID: \"ba81cb6f-e6fb-49e1-aa4c-5a9e6f8769fb\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-85kp6" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.773929 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/ba81cb6f-e6fb-49e1-aa4c-5a9e6f8769fb-audit-dir\") pod \"apiserver-7bbb656c7d-85kp6\" (UID: \"ba81cb6f-e6fb-49e1-aa4c-5a9e6f8769fb\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-85kp6" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.773960 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-gl7lr" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.774408 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/3021fdc0-7e51-4506-ac3c-09ac3a00c77f-etcd-service-ca\") pod \"etcd-operator-b45778765-gqm5m\" (UID: \"3021fdc0-7e51-4506-ac3c-09ac3a00c77f\") " pod="openshift-etcd-operator/etcd-operator-b45778765-gqm5m" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.774866 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/ba81cb6f-e6fb-49e1-aa4c-5a9e6f8769fb-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-85kp6\" (UID: \"ba81cb6f-e6fb-49e1-aa4c-5a9e6f8769fb\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-85kp6" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.774911 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/ba81cb6f-e6fb-49e1-aa4c-5a9e6f8769fb-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-85kp6\" (UID: \"ba81cb6f-e6fb-49e1-aa4c-5a9e6f8769fb\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-85kp6" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.774960 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/3021fdc0-7e51-4506-ac3c-09ac3a00c77f-etcd-ca\") pod \"etcd-operator-b45778765-gqm5m\" (UID: \"3021fdc0-7e51-4506-ac3c-09ac3a00c77f\") " pod="openshift-etcd-operator/etcd-operator-b45778765-gqm5m" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.775112 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1ca33297-d58c-4493-b1d3-45dd83983be4-config\") pod \"kube-apiserver-operator-766d6c64bb-rgqjs\" (UID: \"1ca33297-d58c-4493-b1d3-45dd83983be4\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-rgqjs" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.776884 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5df90a93-b572-4439-b3aa-2666b2873c1d-client-ca\") pod \"route-controller-manager-6576b87f9c-gd7f9\" (UID: \"5df90a93-b572-4439-b3aa-2666b2873c1d\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-gd7f9" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.777175 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5df90a93-b572-4439-b3aa-2666b2873c1d-serving-cert\") pod \"route-controller-manager-6576b87f9c-gd7f9\" (UID: \"5df90a93-b572-4439-b3aa-2666b2873c1d\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-gd7f9" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.777267 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5df90a93-b572-4439-b3aa-2666b2873c1d-config\") pod \"route-controller-manager-6576b87f9c-gd7f9\" (UID: \"5df90a93-b572-4439-b3aa-2666b2873c1d\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-gd7f9" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.777568 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-gl7lr"] Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.777812 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ba81cb6f-e6fb-49e1-aa4c-5a9e6f8769fb-serving-cert\") pod \"apiserver-7bbb656c7d-85kp6\" (UID: \"ba81cb6f-e6fb-49e1-aa4c-5a9e6f8769fb\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-85kp6" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.778472 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/ba81cb6f-e6fb-49e1-aa4c-5a9e6f8769fb-encryption-config\") pod \"apiserver-7bbb656c7d-85kp6\" (UID: \"ba81cb6f-e6fb-49e1-aa4c-5a9e6f8769fb\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-85kp6" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.778737 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3021fdc0-7e51-4506-ac3c-09ac3a00c77f-serving-cert\") pod \"etcd-operator-b45778765-gqm5m\" (UID: \"3021fdc0-7e51-4506-ac3c-09ac3a00c77f\") " pod="openshift-etcd-operator/etcd-operator-b45778765-gqm5m" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.779038 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1ca33297-d58c-4493-b1d3-45dd83983be4-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-rgqjs\" (UID: \"1ca33297-d58c-4493-b1d3-45dd83983be4\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-rgqjs" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.779957 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/3021fdc0-7e51-4506-ac3c-09ac3a00c77f-etcd-client\") pod \"etcd-operator-b45778765-gqm5m\" (UID: \"3021fdc0-7e51-4506-ac3c-09ac3a00c77f\") " pod="openshift-etcd-operator/etcd-operator-b45778765-gqm5m" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.781719 4988 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.782627 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/ba81cb6f-e6fb-49e1-aa4c-5a9e6f8769fb-etcd-client\") pod \"apiserver-7bbb656c7d-85kp6\" (UID: \"ba81cb6f-e6fb-49e1-aa4c-5a9e6f8769fb\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-85kp6" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.801824 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.822152 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.841633 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.860767 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.874585 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b5cd6079-031c-4cc8-8161-88f0e6f84064-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-m55d6\" (UID: \"b5cd6079-031c-4cc8-8161-88f0e6f84064\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-m55d6" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.874683 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/6c651396-5e8f-4920-a2b9-75f2c40eeea8-proxy-tls\") pod \"machine-config-controller-84d6567774-qmwln\" (UID: \"6c651396-5e8f-4920-a2b9-75f2c40eeea8\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-qmwln" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.874749 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4kxgh\" (UniqueName: \"kubernetes.io/projected/a818c760-46b8-4f91-89b3-18e4b0c9a649-kube-api-access-4kxgh\") pod \"controller-manager-879f6c89f-sndzf\" (UID: \"a818c760-46b8-4f91-89b3-18e4b0c9a649\") " pod="openshift-controller-manager/controller-manager-879f6c89f-sndzf" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.874774 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tmtkc\" (UniqueName: \"kubernetes.io/projected/6c651396-5e8f-4920-a2b9-75f2c40eeea8-kube-api-access-tmtkc\") pod \"machine-config-controller-84d6567774-qmwln\" (UID: \"6c651396-5e8f-4920-a2b9-75f2c40eeea8\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-qmwln" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.874802 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/9c8874ed-3f70-45b6-9f10-d8737b6a7a8f-oauth-serving-cert\") pod \"console-f9d7485db-7xdnc\" (UID: \"9c8874ed-3f70-45b6-9f10-d8737b6a7a8f\") " pod="openshift-console/console-f9d7485db-7xdnc" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.874831 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/d0df68ec-1481-4992-bc83-16b4a6b85f0b-srv-cert\") pod \"catalog-operator-68c6474976-jrt4j\" (UID: \"d0df68ec-1481-4992-bc83-16b4a6b85f0b\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-jrt4j" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.874853 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/9c8874ed-3f70-45b6-9f10-d8737b6a7a8f-trusted-ca-bundle\") pod \"console-f9d7485db-7xdnc\" (UID: \"9c8874ed-3f70-45b6-9f10-d8737b6a7a8f\") " pod="openshift-console/console-f9d7485db-7xdnc" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.875288 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b5cd6079-031c-4cc8-8161-88f0e6f84064-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-m55d6\" (UID: \"b5cd6079-031c-4cc8-8161-88f0e6f84064\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-m55d6" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.875343 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/a818c760-46b8-4f91-89b3-18e4b0c9a649-client-ca\") pod \"controller-manager-879f6c89f-sndzf\" (UID: \"a818c760-46b8-4f91-89b3-18e4b0c9a649\") " pod="openshift-controller-manager/controller-manager-879f6c89f-sndzf" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.875394 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a818c760-46b8-4f91-89b3-18e4b0c9a649-config\") pod \"controller-manager-879f6c89f-sndzf\" (UID: \"a818c760-46b8-4f91-89b3-18e4b0c9a649\") " pod="openshift-controller-manager/controller-manager-879f6c89f-sndzf" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.875442 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4719309b-a3b7-4910-a389-635acc09d01d-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-k4ghh\" (UID: \"4719309b-a3b7-4910-a389-635acc09d01d\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-k4ghh" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.875467 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c83b25d6-a1c1-4394-9954-697156fe53b7-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-l86nd\" (UID: \"c83b25d6-a1c1-4394-9954-697156fe53b7\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-l86nd" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.875500 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/9c8874ed-3f70-45b6-9f10-d8737b6a7a8f-console-oauth-config\") pod \"console-f9d7485db-7xdnc\" (UID: \"9c8874ed-3f70-45b6-9f10-d8737b6a7a8f\") " pod="openshift-console/console-f9d7485db-7xdnc" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.875523 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/9c8874ed-3f70-45b6-9f10-d8737b6a7a8f-service-ca\") pod \"console-f9d7485db-7xdnc\" (UID: \"9c8874ed-3f70-45b6-9f10-d8737b6a7a8f\") " pod="openshift-console/console-f9d7485db-7xdnc" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.875548 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/0769dc07-8f56-42a5-8eb5-b900eefaaf1e-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-q2w8f\" (UID: \"0769dc07-8f56-42a5-8eb5-b900eefaaf1e\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-q2w8f" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.875574 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/c83b25d6-a1c1-4394-9954-697156fe53b7-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-l86nd\" (UID: \"c83b25d6-a1c1-4394-9954-697156fe53b7\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-l86nd" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.875615 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t45mv\" (UniqueName: \"kubernetes.io/projected/0769dc07-8f56-42a5-8eb5-b900eefaaf1e-kube-api-access-t45mv\") pod \"multus-admission-controller-857f4d67dd-q2w8f\" (UID: \"0769dc07-8f56-42a5-8eb5-b900eefaaf1e\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-q2w8f" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.875623 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b5cd6079-031c-4cc8-8161-88f0e6f84064-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-m55d6\" (UID: \"b5cd6079-031c-4cc8-8161-88f0e6f84064\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-m55d6" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.875653 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a818c760-46b8-4f91-89b3-18e4b0c9a649-serving-cert\") pod \"controller-manager-879f6c89f-sndzf\" (UID: \"a818c760-46b8-4f91-89b3-18e4b0c9a649\") " pod="openshift-controller-manager/controller-manager-879f6c89f-sndzf" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.875697 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/9c8874ed-3f70-45b6-9f10-d8737b6a7a8f-console-serving-cert\") pod \"console-f9d7485db-7xdnc\" (UID: \"9c8874ed-3f70-45b6-9f10-d8737b6a7a8f\") " pod="openshift-console/console-f9d7485db-7xdnc" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.875720 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qhgxr\" (UniqueName: \"kubernetes.io/projected/9c8874ed-3f70-45b6-9f10-d8737b6a7a8f-kube-api-access-qhgxr\") pod \"console-f9d7485db-7xdnc\" (UID: \"9c8874ed-3f70-45b6-9f10-d8737b6a7a8f\") " pod="openshift-console/console-f9d7485db-7xdnc" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.875746 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bqs24\" (UniqueName: \"kubernetes.io/projected/aa86616c-479c-45be-aaa3-d3c4994721bf-kube-api-access-bqs24\") pod \"ingress-operator-5b745b69d9-q2mxf\" (UID: \"aa86616c-479c-45be-aaa3-d3c4994721bf\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-q2mxf" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.875770 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c83b25d6-a1c1-4394-9954-697156fe53b7-config\") pod \"kube-controller-manager-operator-78b949d7b-l86nd\" (UID: \"c83b25d6-a1c1-4394-9954-697156fe53b7\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-l86nd" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.875796 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/b5cd6079-031c-4cc8-8161-88f0e6f84064-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-m55d6\" (UID: \"b5cd6079-031c-4cc8-8161-88f0e6f84064\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-m55d6" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.875819 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/aa86616c-479c-45be-aaa3-d3c4994721bf-metrics-tls\") pod \"ingress-operator-5b745b69d9-q2mxf\" (UID: \"aa86616c-479c-45be-aaa3-d3c4994721bf\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-q2mxf" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.875840 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/aa86616c-479c-45be-aaa3-d3c4994721bf-bound-sa-token\") pod \"ingress-operator-5b745b69d9-q2mxf\" (UID: \"aa86616c-479c-45be-aaa3-d3c4994721bf\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-q2mxf" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.875873 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/6c651396-5e8f-4920-a2b9-75f2c40eeea8-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-qmwln\" (UID: \"6c651396-5e8f-4920-a2b9-75f2c40eeea8\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-qmwln" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.875897 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/9c8874ed-3f70-45b6-9f10-d8737b6a7a8f-console-config\") pod \"console-f9d7485db-7xdnc\" (UID: \"9c8874ed-3f70-45b6-9f10-d8737b6a7a8f\") " pod="openshift-console/console-f9d7485db-7xdnc" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.875924 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l5tq9\" (UniqueName: \"kubernetes.io/projected/4719309b-a3b7-4910-a389-635acc09d01d-kube-api-access-l5tq9\") pod \"openshift-controller-manager-operator-756b6f6bc6-k4ghh\" (UID: \"4719309b-a3b7-4910-a389-635acc09d01d\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-k4ghh" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.875950 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/9c8874ed-3f70-45b6-9f10-d8737b6a7a8f-oauth-serving-cert\") pod \"console-f9d7485db-7xdnc\" (UID: \"9c8874ed-3f70-45b6-9f10-d8737b6a7a8f\") " pod="openshift-console/console-f9d7485db-7xdnc" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.875955 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/d0df68ec-1481-4992-bc83-16b4a6b85f0b-profile-collector-cert\") pod \"catalog-operator-68c6474976-jrt4j\" (UID: \"d0df68ec-1481-4992-bc83-16b4a6b85f0b\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-jrt4j" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.876013 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b8nbn\" (UniqueName: \"kubernetes.io/projected/d0df68ec-1481-4992-bc83-16b4a6b85f0b-kube-api-access-b8nbn\") pod \"catalog-operator-68c6474976-jrt4j\" (UID: \"d0df68ec-1481-4992-bc83-16b4a6b85f0b\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-jrt4j" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.876042 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/aa86616c-479c-45be-aaa3-d3c4994721bf-trusted-ca\") pod \"ingress-operator-5b745b69d9-q2mxf\" (UID: \"aa86616c-479c-45be-aaa3-d3c4994721bf\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-q2mxf" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.876067 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/a818c760-46b8-4f91-89b3-18e4b0c9a649-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-sndzf\" (UID: \"a818c760-46b8-4f91-89b3-18e4b0c9a649\") " pod="openshift-controller-manager/controller-manager-879f6c89f-sndzf" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.876090 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4719309b-a3b7-4910-a389-635acc09d01d-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-k4ghh\" (UID: \"4719309b-a3b7-4910-a389-635acc09d01d\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-k4ghh" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.876738 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4719309b-a3b7-4910-a389-635acc09d01d-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-k4ghh\" (UID: \"4719309b-a3b7-4910-a389-635acc09d01d\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-k4ghh" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.876964 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/9c8874ed-3f70-45b6-9f10-d8737b6a7a8f-trusted-ca-bundle\") pod \"console-f9d7485db-7xdnc\" (UID: \"9c8874ed-3f70-45b6-9f10-d8737b6a7a8f\") " pod="openshift-console/console-f9d7485db-7xdnc" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.878497 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/a818c760-46b8-4f91-89b3-18e4b0c9a649-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-sndzf\" (UID: \"a818c760-46b8-4f91-89b3-18e4b0c9a649\") " pod="openshift-controller-manager/controller-manager-879f6c89f-sndzf" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.878755 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/6c651396-5e8f-4920-a2b9-75f2c40eeea8-proxy-tls\") pod \"machine-config-controller-84d6567774-qmwln\" (UID: \"6c651396-5e8f-4920-a2b9-75f2c40eeea8\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-qmwln" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.879045 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c83b25d6-a1c1-4394-9954-697156fe53b7-config\") pod \"kube-controller-manager-operator-78b949d7b-l86nd\" (UID: \"c83b25d6-a1c1-4394-9954-697156fe53b7\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-l86nd" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.879347 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/d0df68ec-1481-4992-bc83-16b4a6b85f0b-srv-cert\") pod \"catalog-operator-68c6474976-jrt4j\" (UID: \"d0df68ec-1481-4992-bc83-16b4a6b85f0b\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-jrt4j" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.879590 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/a818c760-46b8-4f91-89b3-18e4b0c9a649-client-ca\") pod \"controller-manager-879f6c89f-sndzf\" (UID: \"a818c760-46b8-4f91-89b3-18e4b0c9a649\") " pod="openshift-controller-manager/controller-manager-879f6c89f-sndzf" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.879824 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/aa86616c-479c-45be-aaa3-d3c4994721bf-metrics-tls\") pod \"ingress-operator-5b745b69d9-q2mxf\" (UID: \"aa86616c-479c-45be-aaa3-d3c4994721bf\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-q2mxf" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.879925 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/6c651396-5e8f-4920-a2b9-75f2c40eeea8-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-qmwln\" (UID: \"6c651396-5e8f-4920-a2b9-75f2c40eeea8\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-qmwln" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.880003 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c83b25d6-a1c1-4394-9954-697156fe53b7-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-l86nd\" (UID: \"c83b25d6-a1c1-4394-9954-697156fe53b7\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-l86nd" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.880990 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/9c8874ed-3f70-45b6-9f10-d8737b6a7a8f-service-ca\") pod \"console-f9d7485db-7xdnc\" (UID: \"9c8874ed-3f70-45b6-9f10-d8737b6a7a8f\") " pod="openshift-console/console-f9d7485db-7xdnc" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.880998 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/9c8874ed-3f70-45b6-9f10-d8737b6a7a8f-console-config\") pod \"console-f9d7485db-7xdnc\" (UID: \"9c8874ed-3f70-45b6-9f10-d8737b6a7a8f\") " pod="openshift-console/console-f9d7485db-7xdnc" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.881161 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a818c760-46b8-4f91-89b3-18e4b0c9a649-config\") pod \"controller-manager-879f6c89f-sndzf\" (UID: \"a818c760-46b8-4f91-89b3-18e4b0c9a649\") " pod="openshift-controller-manager/controller-manager-879f6c89f-sndzf" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.882433 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b5cd6079-031c-4cc8-8161-88f0e6f84064-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-m55d6\" (UID: \"b5cd6079-031c-4cc8-8161-88f0e6f84064\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-m55d6" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.882954 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4719309b-a3b7-4910-a389-635acc09d01d-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-k4ghh\" (UID: \"4719309b-a3b7-4910-a389-635acc09d01d\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-k4ghh" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.882979 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/9c8874ed-3f70-45b6-9f10-d8737b6a7a8f-console-serving-cert\") pod \"console-f9d7485db-7xdnc\" (UID: \"9c8874ed-3f70-45b6-9f10-d8737b6a7a8f\") " pod="openshift-console/console-f9d7485db-7xdnc" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.883129 4988 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.883342 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/d0df68ec-1481-4992-bc83-16b4a6b85f0b-profile-collector-cert\") pod \"catalog-operator-68c6474976-jrt4j\" (UID: \"d0df68ec-1481-4992-bc83-16b4a6b85f0b\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-jrt4j" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.884469 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a818c760-46b8-4f91-89b3-18e4b0c9a649-serving-cert\") pod \"controller-manager-879f6c89f-sndzf\" (UID: \"a818c760-46b8-4f91-89b3-18e4b0c9a649\") " pod="openshift-controller-manager/controller-manager-879f6c89f-sndzf" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.885027 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/9c8874ed-3f70-45b6-9f10-d8737b6a7a8f-console-oauth-config\") pod \"console-f9d7485db-7xdnc\" (UID: \"9c8874ed-3f70-45b6-9f10-d8737b6a7a8f\") " pod="openshift-console/console-f9d7485db-7xdnc" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.886893 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/0769dc07-8f56-42a5-8eb5-b900eefaaf1e-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-q2w8f\" (UID: \"0769dc07-8f56-42a5-8eb5-b900eefaaf1e\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-q2w8f" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.912146 4988 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.917811 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/aa86616c-479c-45be-aaa3-d3c4994721bf-trusted-ca\") pod \"ingress-operator-5b745b69d9-q2mxf\" (UID: \"aa86616c-479c-45be-aaa3-d3c4994721bf\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-q2mxf" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.921950 4988 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.957973 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6r9cj\" (UniqueName: \"kubernetes.io/projected/e294e9c8-16f2-46d4-91f8-5964b1e8ea5a-kube-api-access-6r9cj\") pod \"oauth-openshift-558db77b4-hwtzl\" (UID: \"e294e9c8-16f2-46d4-91f8-5964b1e8ea5a\") " pod="openshift-authentication/oauth-openshift-558db77b4-hwtzl" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.961277 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets" Oct 08 18:13:14 crc kubenswrapper[4988]: I1008 18:13:14.981549 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Oct 08 18:13:15 crc kubenswrapper[4988]: I1008 18:13:15.002959 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls" Oct 08 18:13:15 crc kubenswrapper[4988]: I1008 18:13:15.039945 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xj2qd\" (UniqueName: \"kubernetes.io/projected/8deeca47-161b-43bd-a7c4-7f9a41f931d8-kube-api-access-xj2qd\") pod \"authentication-operator-69f744f599-zjfnq\" (UID: \"8deeca47-161b-43bd-a7c4-7f9a41f931d8\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-zjfnq" Oct 08 18:13:15 crc kubenswrapper[4988]: I1008 18:13:15.042315 4988 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt" Oct 08 18:13:15 crc kubenswrapper[4988]: I1008 18:13:15.058764 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-hwtzl" Oct 08 18:13:15 crc kubenswrapper[4988]: I1008 18:13:15.063241 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7" Oct 08 18:13:15 crc kubenswrapper[4988]: I1008 18:13:15.082787 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-tls" Oct 08 18:13:15 crc kubenswrapper[4988]: I1008 18:13:15.103399 4988 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy" Oct 08 18:13:15 crc kubenswrapper[4988]: I1008 18:13:15.114980 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-zjfnq" Oct 08 18:13:15 crc kubenswrapper[4988]: I1008 18:13:15.123284 4988 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"machine-api-operator-images" Oct 08 18:13:15 crc kubenswrapper[4988]: I1008 18:13:15.143283 4988 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt" Oct 08 18:13:15 crc kubenswrapper[4988]: I1008 18:13:15.197922 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zprbm\" (UniqueName: \"kubernetes.io/projected/b2497303-9eb5-4f27-8595-763cb5a851a6-kube-api-access-zprbm\") pod \"kube-storage-version-migrator-operator-b67b599dd-6k9kl\" (UID: \"b2497303-9eb5-4f27-8595-763cb5a851a6\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-6k9kl" Oct 08 18:13:15 crc kubenswrapper[4988]: I1008 18:13:15.203750 4988 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Oct 08 18:13:15 crc kubenswrapper[4988]: I1008 18:13:15.207187 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d8dmg\" (UniqueName: \"kubernetes.io/projected/0bc2144f-4dbe-4dc1-a07a-b764d8582324-kube-api-access-d8dmg\") pod \"downloads-7954f5f757-qcgmb\" (UID: \"0bc2144f-4dbe-4dc1-a07a-b764d8582324\") " pod="openshift-console/downloads-7954f5f757-qcgmb" Oct 08 18:13:15 crc kubenswrapper[4988]: I1008 18:13:15.245953 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"samples-operator-tls" Oct 08 18:13:15 crc kubenswrapper[4988]: I1008 18:13:15.254615 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jvw4s\" (UniqueName: \"kubernetes.io/projected/5ec13023-d870-43fc-aa01-9cdb2b3015c4-kube-api-access-jvw4s\") pod \"machine-approver-56656f9798-zf8t5\" (UID: \"5ec13023-d870-43fc-aa01-9cdb2b3015c4\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-zf8t5" Oct 08 18:13:15 crc kubenswrapper[4988]: I1008 18:13:15.263564 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w" Oct 08 18:13:15 crc kubenswrapper[4988]: I1008 18:13:15.303313 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/658aaf4b-d6ae-4ebf-bc88-16bd718df0b5-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-jt7fv\" (UID: \"658aaf4b-d6ae-4ebf-bc88-16bd718df0b5\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-jt7fv" Oct 08 18:13:15 crc kubenswrapper[4988]: I1008 18:13:15.319397 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sj2zk\" (UniqueName: \"kubernetes.io/projected/63fe20f8-74c8-48ba-8c65-7f284d959254-kube-api-access-sj2zk\") pod \"openshift-config-operator-7777fb866f-7ctzs\" (UID: \"63fe20f8-74c8-48ba-8c65-7f284d959254\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-7ctzs" Oct 08 18:13:15 crc kubenswrapper[4988]: I1008 18:13:15.322602 4988 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Oct 08 18:13:15 crc kubenswrapper[4988]: I1008 18:13:15.325994 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-hwtzl"] Oct 08 18:13:15 crc kubenswrapper[4988]: I1008 18:13:15.335330 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-7ctzs" Oct 08 18:13:15 crc kubenswrapper[4988]: I1008 18:13:15.350806 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-zjfnq"] Oct 08 18:13:15 crc kubenswrapper[4988]: I1008 18:13:15.357425 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l795v\" (UniqueName: \"kubernetes.io/projected/658aaf4b-d6ae-4ebf-bc88-16bd718df0b5-kube-api-access-l795v\") pod \"cluster-image-registry-operator-dc59b4c8b-jt7fv\" (UID: \"658aaf4b-d6ae-4ebf-bc88-16bd718df0b5\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-jt7fv" Oct 08 18:13:15 crc kubenswrapper[4988]: W1008 18:13:15.367017 4988 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8deeca47_161b_43bd_a7c4_7f9a41f931d8.slice/crio-14f5b3db4b1e7d0c1269a31a403d2bf0d5fbc4a2f9f3b7f89d65892a93c93aad WatchSource:0}: Error finding container 14f5b3db4b1e7d0c1269a31a403d2bf0d5fbc4a2f9f3b7f89d65892a93c93aad: Status 404 returned error can't find the container with id 14f5b3db4b1e7d0c1269a31a403d2bf0d5fbc4a2f9f3b7f89d65892a93c93aad Oct 08 18:13:15 crc kubenswrapper[4988]: I1008 18:13:15.377211 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pszq7\" (UniqueName: \"kubernetes.io/projected/eeee83f5-fe8f-47bf-be7b-cde1ea9ac356-kube-api-access-pszq7\") pod \"console-operator-58897d9998-rt48k\" (UID: \"eeee83f5-fe8f-47bf-be7b-cde1ea9ac356\") " pod="openshift-console-operator/console-operator-58897d9998-rt48k" Oct 08 18:13:15 crc kubenswrapper[4988]: I1008 18:13:15.391893 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-jt7fv" Oct 08 18:13:15 crc kubenswrapper[4988]: I1008 18:13:15.399941 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-qcgmb" Oct 08 18:13:15 crc kubenswrapper[4988]: I1008 18:13:15.401268 4988 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle" Oct 08 18:13:15 crc kubenswrapper[4988]: I1008 18:13:15.405284 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hf5s6\" (UniqueName: \"kubernetes.io/projected/0ccdb64f-445e-4b91-a4b9-437b22d22853-kube-api-access-hf5s6\") pod \"openshift-apiserver-operator-796bbdcf4f-xttx4\" (UID: \"0ccdb64f-445e-4b91-a4b9-437b22d22853\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-xttx4" Oct 08 18:13:15 crc kubenswrapper[4988]: I1008 18:13:15.421303 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-zf8t5" Oct 08 18:13:15 crc kubenswrapper[4988]: I1008 18:13:15.422313 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-dockercfg-zdk86" Oct 08 18:13:15 crc kubenswrapper[4988]: I1008 18:13:15.441752 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-6k9kl" Oct 08 18:13:15 crc kubenswrapper[4988]: I1008 18:13:15.443946 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default" Oct 08 18:13:15 crc kubenswrapper[4988]: I1008 18:13:15.447817 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-xttx4" Oct 08 18:13:15 crc kubenswrapper[4988]: I1008 18:13:15.461013 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-rt48k" Oct 08 18:13:15 crc kubenswrapper[4988]: I1008 18:13:15.461804 4988 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt" Oct 08 18:13:15 crc kubenswrapper[4988]: I1008 18:13:15.482447 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default" Oct 08 18:13:15 crc kubenswrapper[4988]: I1008 18:13:15.501814 4988 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt" Oct 08 18:13:15 crc kubenswrapper[4988]: I1008 18:13:15.521199 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default" Oct 08 18:13:15 crc kubenswrapper[4988]: I1008 18:13:15.539401 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-7ctzs"] Oct 08 18:13:15 crc kubenswrapper[4988]: I1008 18:13:15.543111 4988 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Oct 08 18:13:15 crc kubenswrapper[4988]: W1008 18:13:15.564415 4988 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod63fe20f8_74c8_48ba_8c65_7f284d959254.slice/crio-433334d8ecb916a7e040143954a19db43376cb5d8960f40e80270cfb875ec833 WatchSource:0}: Error finding container 433334d8ecb916a7e040143954a19db43376cb5d8960f40e80270cfb875ec833: Status 404 returned error can't find the container with id 433334d8ecb916a7e040143954a19db43376cb5d8960f40e80270cfb875ec833 Oct 08 18:13:15 crc kubenswrapper[4988]: I1008 18:13:15.564880 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87" Oct 08 18:13:15 crc kubenswrapper[4988]: I1008 18:13:15.582909 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Oct 08 18:13:15 crc kubenswrapper[4988]: I1008 18:13:15.593187 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-jt7fv"] Oct 08 18:13:15 crc kubenswrapper[4988]: I1008 18:13:15.603154 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert" Oct 08 18:13:15 crc kubenswrapper[4988]: I1008 18:13:15.621788 4988 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Oct 08 18:13:15 crc kubenswrapper[4988]: W1008 18:13:15.636672 4988 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod658aaf4b_d6ae_4ebf_bc88_16bd718df0b5.slice/crio-969574c9fc8d653c7ccdc616421e2c60c2d656b7f0593a68bc4667492c86a860 WatchSource:0}: Error finding container 969574c9fc8d653c7ccdc616421e2c60c2d656b7f0593a68bc4667492c86a860: Status 404 returned error can't find the container with id 969574c9fc8d653c7ccdc616421e2c60c2d656b7f0593a68bc4667492c86a860 Oct 08 18:13:15 crc kubenswrapper[4988]: I1008 18:13:15.641583 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls" Oct 08 18:13:15 crc kubenswrapper[4988]: I1008 18:13:15.652520 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-qcgmb"] Oct 08 18:13:15 crc kubenswrapper[4988]: I1008 18:13:15.661800 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"dns-operator-dockercfg-9mqw5" Oct 08 18:13:15 crc kubenswrapper[4988]: I1008 18:13:15.683452 4988 request.go:700] Waited for 1.007083371s due to client-side throttling, not priority and fairness, request: GET:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-dns-operator/configmaps?fieldSelector=metadata.name%3Dkube-root-ca.crt&limit=500&resourceVersion=0 Oct 08 18:13:15 crc kubenswrapper[4988]: I1008 18:13:15.686419 4988 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Oct 08 18:13:15 crc kubenswrapper[4988]: I1008 18:13:15.701967 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert" Oct 08 18:13:15 crc kubenswrapper[4988]: I1008 18:13:15.721665 4988 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Oct 08 18:13:15 crc kubenswrapper[4988]: I1008 18:13:15.722071 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-xttx4"] Oct 08 18:13:15 crc kubenswrapper[4988]: I1008 18:13:15.742227 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg" Oct 08 18:13:15 crc kubenswrapper[4988]: W1008 18:13:15.753908 4988 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0ccdb64f_445e_4b91_a4b9_437b22d22853.slice/crio-c368e0952ed44014b68173efe2d6346c9e4381547a3c9bd3ae1138920328972e WatchSource:0}: Error finding container c368e0952ed44014b68173efe2d6346c9e4381547a3c9bd3ae1138920328972e: Status 404 returned error can't find the container with id c368e0952ed44014b68173efe2d6346c9e4381547a3c9bd3ae1138920328972e Oct 08 18:13:15 crc kubenswrapper[4988]: I1008 18:13:15.757433 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-6k9kl"] Oct 08 18:13:15 crc kubenswrapper[4988]: I1008 18:13:15.763958 4988 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Oct 08 18:13:15 crc kubenswrapper[4988]: I1008 18:13:15.781280 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-rt48k"] Oct 08 18:13:15 crc kubenswrapper[4988]: I1008 18:13:15.781326 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"packageserver-service-cert" Oct 08 18:13:15 crc kubenswrapper[4988]: W1008 18:13:15.793904 4988 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podeeee83f5_fe8f_47bf_be7b_cde1ea9ac356.slice/crio-d7d094bcc843e5f8a240a996b72e4c702e05f59c11d1b9a7d1470922a535b885 WatchSource:0}: Error finding container d7d094bcc843e5f8a240a996b72e4c702e05f59c11d1b9a7d1470922a535b885: Status 404 returned error can't find the container with id d7d094bcc843e5f8a240a996b72e4c702e05f59c11d1b9a7d1470922a535b885 Oct 08 18:13:15 crc kubenswrapper[4988]: I1008 18:13:15.802200 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-k9rxt" Oct 08 18:13:15 crc kubenswrapper[4988]: I1008 18:13:15.822649 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Oct 08 18:13:15 crc kubenswrapper[4988]: I1008 18:13:15.842302 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"signing-key" Oct 08 18:13:15 crc kubenswrapper[4988]: I1008 18:13:15.861791 4988 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"signing-cabundle" Oct 08 18:13:15 crc kubenswrapper[4988]: I1008 18:13:15.881895 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"service-ca-dockercfg-pn86c" Oct 08 18:13:15 crc kubenswrapper[4988]: I1008 18:13:15.902422 4988 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt" Oct 08 18:13:15 crc kubenswrapper[4988]: I1008 18:13:15.922404 4988 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt" Oct 08 18:13:15 crc kubenswrapper[4988]: I1008 18:13:15.942419 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"service-ca-operator-dockercfg-rg9jl" Oct 08 18:13:15 crc kubenswrapper[4988]: I1008 18:13:15.961247 4988 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt" Oct 08 18:13:15 crc kubenswrapper[4988]: I1008 18:13:15.985903 4988 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt" Oct 08 18:13:16 crc kubenswrapper[4988]: I1008 18:13:16.001919 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert" Oct 08 18:13:16 crc kubenswrapper[4988]: I1008 18:13:16.023493 4988 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config" Oct 08 18:13:16 crc kubenswrapper[4988]: I1008 18:13:16.041918 4988 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca" Oct 08 18:13:16 crc kubenswrapper[4988]: I1008 18:13:16.061823 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff" Oct 08 18:13:16 crc kubenswrapper[4988]: I1008 18:13:16.081472 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client" Oct 08 18:13:16 crc kubenswrapper[4988]: I1008 18:13:16.101969 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert" Oct 08 18:13:16 crc kubenswrapper[4988]: I1008 18:13:16.104955 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-jt7fv" event={"ID":"658aaf4b-d6ae-4ebf-bc88-16bd718df0b5","Type":"ContainerStarted","Data":"45da939fc179970b262c9699aa373a2698af448b3516cce3d0a7dbb1ecaa9044"} Oct 08 18:13:16 crc kubenswrapper[4988]: I1008 18:13:16.105005 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-jt7fv" event={"ID":"658aaf4b-d6ae-4ebf-bc88-16bd718df0b5","Type":"ContainerStarted","Data":"969574c9fc8d653c7ccdc616421e2c60c2d656b7f0593a68bc4667492c86a860"} Oct 08 18:13:16 crc kubenswrapper[4988]: I1008 18:13:16.106216 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-6k9kl" event={"ID":"b2497303-9eb5-4f27-8595-763cb5a851a6","Type":"ContainerStarted","Data":"3a5318cb3681be324b66a7df0d6a4459c2be9b8924a9969f67b62e39f3b720f7"} Oct 08 18:13:16 crc kubenswrapper[4988]: I1008 18:13:16.106255 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-6k9kl" event={"ID":"b2497303-9eb5-4f27-8595-763cb5a851a6","Type":"ContainerStarted","Data":"b450bfbc7edbc6fdfc5f98521acf5898bffe9444a30410c1fdc3b56e23cb4207"} Oct 08 18:13:16 crc kubenswrapper[4988]: I1008 18:13:16.108304 4988 generic.go:334] "Generic (PLEG): container finished" podID="63fe20f8-74c8-48ba-8c65-7f284d959254" containerID="086edefa6132520bb27aff424118accb83eaa0bde2df9ab0f36a59d5dba15bd6" exitCode=0 Oct 08 18:13:16 crc kubenswrapper[4988]: I1008 18:13:16.108411 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-7ctzs" event={"ID":"63fe20f8-74c8-48ba-8c65-7f284d959254","Type":"ContainerDied","Data":"086edefa6132520bb27aff424118accb83eaa0bde2df9ab0f36a59d5dba15bd6"} Oct 08 18:13:16 crc kubenswrapper[4988]: I1008 18:13:16.108444 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-7ctzs" event={"ID":"63fe20f8-74c8-48ba-8c65-7f284d959254","Type":"ContainerStarted","Data":"433334d8ecb916a7e040143954a19db43376cb5d8960f40e80270cfb875ec833"} Oct 08 18:13:16 crc kubenswrapper[4988]: I1008 18:13:16.110570 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-zjfnq" event={"ID":"8deeca47-161b-43bd-a7c4-7f9a41f931d8","Type":"ContainerStarted","Data":"5961b5ee0ce3568cb1d430a39fd454b6965ffdfac3b3b2712f6aaf2924bac7a6"} Oct 08 18:13:16 crc kubenswrapper[4988]: I1008 18:13:16.110613 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-zjfnq" event={"ID":"8deeca47-161b-43bd-a7c4-7f9a41f931d8","Type":"ContainerStarted","Data":"14f5b3db4b1e7d0c1269a31a403d2bf0d5fbc4a2f9f3b7f89d65892a93c93aad"} Oct 08 18:13:16 crc kubenswrapper[4988]: I1008 18:13:16.112616 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-xttx4" event={"ID":"0ccdb64f-445e-4b91-a4b9-437b22d22853","Type":"ContainerStarted","Data":"aee194abecddb5d137e9207a86801c80c712d077edab9655450cb95ec6487404"} Oct 08 18:13:16 crc kubenswrapper[4988]: I1008 18:13:16.112642 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-xttx4" event={"ID":"0ccdb64f-445e-4b91-a4b9-437b22d22853","Type":"ContainerStarted","Data":"c368e0952ed44014b68173efe2d6346c9e4381547a3c9bd3ae1138920328972e"} Oct 08 18:13:16 crc kubenswrapper[4988]: I1008 18:13:16.116889 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-zf8t5" event={"ID":"5ec13023-d870-43fc-aa01-9cdb2b3015c4","Type":"ContainerStarted","Data":"5bce1a04245d0339df5b757cc7ba0c5dced1d33ba4f8ebd6baa1a7d033f4701b"} Oct 08 18:13:16 crc kubenswrapper[4988]: I1008 18:13:16.116946 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-zf8t5" event={"ID":"5ec13023-d870-43fc-aa01-9cdb2b3015c4","Type":"ContainerStarted","Data":"7b372ad72438a1c4c452ad7defaf64cd05fa0c0f6b9c3f62a8bac872ba9b2d74"} Oct 08 18:13:16 crc kubenswrapper[4988]: I1008 18:13:16.116963 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-zf8t5" event={"ID":"5ec13023-d870-43fc-aa01-9cdb2b3015c4","Type":"ContainerStarted","Data":"bb701297a9e15f93dc99a2355cf45b259edeea9f3df2adecdf711820b4d0d72e"} Oct 08 18:13:16 crc kubenswrapper[4988]: I1008 18:13:16.118681 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-qcgmb" event={"ID":"0bc2144f-4dbe-4dc1-a07a-b764d8582324","Type":"ContainerStarted","Data":"b921d8e906e15632f71f6351e0bbdce6314e228cb2faf3bcb6a9e211ddd25987"} Oct 08 18:13:16 crc kubenswrapper[4988]: I1008 18:13:16.118709 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-qcgmb" event={"ID":"0bc2144f-4dbe-4dc1-a07a-b764d8582324","Type":"ContainerStarted","Data":"760cec11d1927dcb0327a7c797e8c9535605ff74fdd29f0192f87503d16440a7"} Oct 08 18:13:16 crc kubenswrapper[4988]: I1008 18:13:16.118942 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/downloads-7954f5f757-qcgmb" Oct 08 18:13:16 crc kubenswrapper[4988]: I1008 18:13:16.120363 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-rt48k" event={"ID":"eeee83f5-fe8f-47bf-be7b-cde1ea9ac356","Type":"ContainerStarted","Data":"7d2564c88e15dc93654cc79ffbcb13bdd7d6fef5374fbff03937cd31cc3128b3"} Oct 08 18:13:16 crc kubenswrapper[4988]: I1008 18:13:16.120409 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-rt48k" event={"ID":"eeee83f5-fe8f-47bf-be7b-cde1ea9ac356","Type":"ContainerStarted","Data":"d7d094bcc843e5f8a240a996b72e4c702e05f59c11d1b9a7d1470922a535b885"} Oct 08 18:13:16 crc kubenswrapper[4988]: I1008 18:13:16.120783 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console-operator/console-operator-58897d9998-rt48k" Oct 08 18:13:16 crc kubenswrapper[4988]: I1008 18:13:16.121570 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-1" Oct 08 18:13:16 crc kubenswrapper[4988]: I1008 18:13:16.121567 4988 patch_prober.go:28] interesting pod/downloads-7954f5f757-qcgmb container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.13:8080/\": dial tcp 10.217.0.13:8080: connect: connection refused" start-of-body= Oct 08 18:13:16 crc kubenswrapper[4988]: I1008 18:13:16.121733 4988 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-qcgmb" podUID="0bc2144f-4dbe-4dc1-a07a-b764d8582324" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.13:8080/\": dial tcp 10.217.0.13:8080: connect: connection refused" Oct 08 18:13:16 crc kubenswrapper[4988]: I1008 18:13:16.122436 4988 patch_prober.go:28] interesting pod/console-operator-58897d9998-rt48k container/console-operator namespace/openshift-console-operator: Readiness probe status=failure output="Get \"https://10.217.0.10:8443/readyz\": dial tcp 10.217.0.10:8443: connect: connection refused" start-of-body= Oct 08 18:13:16 crc kubenswrapper[4988]: I1008 18:13:16.122505 4988 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console-operator/console-operator-58897d9998-rt48k" podUID="eeee83f5-fe8f-47bf-be7b-cde1ea9ac356" containerName="console-operator" probeResult="failure" output="Get \"https://10.217.0.10:8443/readyz\": dial tcp 10.217.0.10:8443: connect: connection refused" Oct 08 18:13:16 crc kubenswrapper[4988]: I1008 18:13:16.123846 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-hwtzl" event={"ID":"e294e9c8-16f2-46d4-91f8-5964b1e8ea5a","Type":"ContainerStarted","Data":"5bf6c97c2a92106c5d151cc199928afbbcc2d70b4025fb7cd24058e7da3b8157"} Oct 08 18:13:16 crc kubenswrapper[4988]: I1008 18:13:16.123886 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-hwtzl" event={"ID":"e294e9c8-16f2-46d4-91f8-5964b1e8ea5a","Type":"ContainerStarted","Data":"918f45165b7980541dbf929e550b866d9b90c6015d6ef15b1e4f7ceca9097eca"} Oct 08 18:13:16 crc kubenswrapper[4988]: I1008 18:13:16.124254 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-558db77b4-hwtzl" Oct 08 18:13:16 crc kubenswrapper[4988]: I1008 18:13:16.126628 4988 patch_prober.go:28] interesting pod/oauth-openshift-558db77b4-hwtzl container/oauth-openshift namespace/openshift-authentication: Readiness probe status=failure output="Get \"https://10.217.0.5:6443/healthz\": dial tcp 10.217.0.5:6443: connect: connection refused" start-of-body= Oct 08 18:13:16 crc kubenswrapper[4988]: I1008 18:13:16.126727 4988 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-authentication/oauth-openshift-558db77b4-hwtzl" podUID="e294e9c8-16f2-46d4-91f8-5964b1e8ea5a" containerName="oauth-openshift" probeResult="failure" output="Get \"https://10.217.0.5:6443/healthz\": dial tcp 10.217.0.5:6443: connect: connection refused" Oct 08 18:13:16 crc kubenswrapper[4988]: I1008 18:13:16.142704 4988 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca" Oct 08 18:13:16 crc kubenswrapper[4988]: I1008 18:13:16.169214 4988 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle" Oct 08 18:13:16 crc kubenswrapper[4988]: I1008 18:13:16.181279 4988 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Oct 08 18:13:16 crc kubenswrapper[4988]: I1008 18:13:16.202305 4988 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Oct 08 18:13:16 crc kubenswrapper[4988]: I1008 18:13:16.221729 4988 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config" Oct 08 18:13:16 crc kubenswrapper[4988]: I1008 18:13:16.242436 4988 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-1" Oct 08 18:13:16 crc kubenswrapper[4988]: I1008 18:13:16.263059 4988 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 08 18:13:16 crc kubenswrapper[4988]: I1008 18:13:16.282237 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 08 18:13:16 crc kubenswrapper[4988]: I1008 18:13:16.302693 4988 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Oct 08 18:13:16 crc kubenswrapper[4988]: I1008 18:13:16.322900 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Oct 08 18:13:16 crc kubenswrapper[4988]: I1008 18:13:16.342064 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Oct 08 18:13:16 crc kubenswrapper[4988]: I1008 18:13:16.370510 4988 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Oct 08 18:13:16 crc kubenswrapper[4988]: I1008 18:13:16.382938 4988 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Oct 08 18:13:16 crc kubenswrapper[4988]: I1008 18:13:16.402264 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-2llfx" Oct 08 18:13:16 crc kubenswrapper[4988]: I1008 18:13:16.422801 4988 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt" Oct 08 18:13:16 crc kubenswrapper[4988]: I1008 18:13:16.442430 4988 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt" Oct 08 18:13:16 crc kubenswrapper[4988]: I1008 18:13:16.462023 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"canary-serving-cert" Oct 08 18:13:16 crc kubenswrapper[4988]: I1008 18:13:16.503302 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-qx5rd" Oct 08 18:13:16 crc kubenswrapper[4988]: I1008 18:13:16.521584 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"node-bootstrapper-token" Oct 08 18:13:16 crc kubenswrapper[4988]: I1008 18:13:16.542422 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-tls" Oct 08 18:13:16 crc kubenswrapper[4988]: I1008 18:13:16.583004 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-dockercfg-jwfmh" Oct 08 18:13:16 crc kubenswrapper[4988]: I1008 18:13:16.602234 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-default-metrics-tls" Oct 08 18:13:16 crc kubenswrapper[4988]: I1008 18:13:16.622113 4988 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"dns-default" Oct 08 18:13:16 crc kubenswrapper[4988]: I1008 18:13:16.669893 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s5hnj\" (UniqueName: \"kubernetes.io/projected/5df90a93-b572-4439-b3aa-2666b2873c1d-kube-api-access-s5hnj\") pod \"route-controller-manager-6576b87f9c-gd7f9\" (UID: \"5df90a93-b572-4439-b3aa-2666b2873c1d\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-gd7f9" Oct 08 18:13:16 crc kubenswrapper[4988]: I1008 18:13:16.699578 4988 request.go:700] Waited for 1.925168147s due to client-side throttling, not priority and fairness, request: GET:https://api-int.crc.testing:6443/api/v1/namespaces/hostpath-provisioner/configmaps?fieldSelector=metadata.name%3Dopenshift-service-ca.crt&limit=500&resourceVersion=0 Oct 08 18:13:16 crc kubenswrapper[4988]: I1008 18:13:16.702498 4988 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"openshift-service-ca.crt" Oct 08 18:13:16 crc kubenswrapper[4988]: I1008 18:13:16.703023 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w2vx4\" (UniqueName: \"kubernetes.io/projected/ba81cb6f-e6fb-49e1-aa4c-5a9e6f8769fb-kube-api-access-w2vx4\") pod \"apiserver-7bbb656c7d-85kp6\" (UID: \"ba81cb6f-e6fb-49e1-aa4c-5a9e6f8769fb\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-85kp6" Oct 08 18:13:16 crc kubenswrapper[4988]: I1008 18:13:16.722376 4988 reflector.go:368] Caches populated for *v1.Secret from object-"hostpath-provisioner"/"csi-hostpath-provisioner-sa-dockercfg-qd74k" Oct 08 18:13:16 crc kubenswrapper[4988]: I1008 18:13:16.741678 4988 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"kube-root-ca.crt" Oct 08 18:13:16 crc kubenswrapper[4988]: I1008 18:13:16.765218 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-85kp6" Oct 08 18:13:16 crc kubenswrapper[4988]: I1008 18:13:16.766265 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-gd7f9" Oct 08 18:13:16 crc kubenswrapper[4988]: I1008 18:13:16.774945 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d9zcl\" (UniqueName: \"kubernetes.io/projected/3021fdc0-7e51-4506-ac3c-09ac3a00c77f-kube-api-access-d9zcl\") pod \"etcd-operator-b45778765-gqm5m\" (UID: \"3021fdc0-7e51-4506-ac3c-09ac3a00c77f\") " pod="openshift-etcd-operator/etcd-operator-b45778765-gqm5m" Oct 08 18:13:16 crc kubenswrapper[4988]: I1008 18:13:16.807315 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4kxgh\" (UniqueName: \"kubernetes.io/projected/a818c760-46b8-4f91-89b3-18e4b0c9a649-kube-api-access-4kxgh\") pod \"controller-manager-879f6c89f-sndzf\" (UID: \"a818c760-46b8-4f91-89b3-18e4b0c9a649\") " pod="openshift-controller-manager/controller-manager-879f6c89f-sndzf" Oct 08 18:13:16 crc kubenswrapper[4988]: I1008 18:13:16.821174 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tmtkc\" (UniqueName: \"kubernetes.io/projected/6c651396-5e8f-4920-a2b9-75f2c40eeea8-kube-api-access-tmtkc\") pod \"machine-config-controller-84d6567774-qmwln\" (UID: \"6c651396-5e8f-4920-a2b9-75f2c40eeea8\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-qmwln" Oct 08 18:13:16 crc kubenswrapper[4988]: I1008 18:13:16.839462 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qhgxr\" (UniqueName: \"kubernetes.io/projected/9c8874ed-3f70-45b6-9f10-d8737b6a7a8f-kube-api-access-qhgxr\") pod \"console-f9d7485db-7xdnc\" (UID: \"9c8874ed-3f70-45b6-9f10-d8737b6a7a8f\") " pod="openshift-console/console-f9d7485db-7xdnc" Oct 08 18:13:16 crc kubenswrapper[4988]: I1008 18:13:16.844478 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-sndzf" Oct 08 18:13:16 crc kubenswrapper[4988]: I1008 18:13:16.859339 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b8nbn\" (UniqueName: \"kubernetes.io/projected/d0df68ec-1481-4992-bc83-16b4a6b85f0b-kube-api-access-b8nbn\") pod \"catalog-operator-68c6474976-jrt4j\" (UID: \"d0df68ec-1481-4992-bc83-16b4a6b85f0b\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-jrt4j" Oct 08 18:13:16 crc kubenswrapper[4988]: I1008 18:13:16.861061 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-7xdnc" Oct 08 18:13:16 crc kubenswrapper[4988]: I1008 18:13:16.883985 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/b5cd6079-031c-4cc8-8161-88f0e6f84064-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-m55d6\" (UID: \"b5cd6079-031c-4cc8-8161-88f0e6f84064\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-m55d6" Oct 08 18:13:16 crc kubenswrapper[4988]: I1008 18:13:16.898914 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-qmwln" Oct 08 18:13:16 crc kubenswrapper[4988]: I1008 18:13:16.904199 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l5tq9\" (UniqueName: \"kubernetes.io/projected/4719309b-a3b7-4910-a389-635acc09d01d-kube-api-access-l5tq9\") pod \"openshift-controller-manager-operator-756b6f6bc6-k4ghh\" (UID: \"4719309b-a3b7-4910-a389-635acc09d01d\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-k4ghh" Oct 08 18:13:16 crc kubenswrapper[4988]: I1008 18:13:16.925623 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/c83b25d6-a1c1-4394-9954-697156fe53b7-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-l86nd\" (UID: \"c83b25d6-a1c1-4394-9954-697156fe53b7\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-l86nd" Oct 08 18:13:16 crc kubenswrapper[4988]: I1008 18:13:16.942257 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/aa86616c-479c-45be-aaa3-d3c4994721bf-bound-sa-token\") pod \"ingress-operator-5b745b69d9-q2mxf\" (UID: \"aa86616c-479c-45be-aaa3-d3c4994721bf\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-q2mxf" Oct 08 18:13:16 crc kubenswrapper[4988]: I1008 18:13:16.963168 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t45mv\" (UniqueName: \"kubernetes.io/projected/0769dc07-8f56-42a5-8eb5-b900eefaaf1e-kube-api-access-t45mv\") pod \"multus-admission-controller-857f4d67dd-q2w8f\" (UID: \"0769dc07-8f56-42a5-8eb5-b900eefaaf1e\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-q2w8f" Oct 08 18:13:16 crc kubenswrapper[4988]: I1008 18:13:16.981443 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bqs24\" (UniqueName: \"kubernetes.io/projected/aa86616c-479c-45be-aaa3-d3c4994721bf-kube-api-access-bqs24\") pod \"ingress-operator-5b745b69d9-q2mxf\" (UID: \"aa86616c-479c-45be-aaa3-d3c4994721bf\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-q2mxf" Oct 08 18:13:16 crc kubenswrapper[4988]: I1008 18:13:16.983508 4988 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Oct 08 18:13:17 crc kubenswrapper[4988]: I1008 18:13:17.004853 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1ca33297-d58c-4493-b1d3-45dd83983be4-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-rgqjs\" (UID: \"1ca33297-d58c-4493-b1d3-45dd83983be4\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-rgqjs" Oct 08 18:13:17 crc kubenswrapper[4988]: I1008 18:13:17.014787 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-gqm5m" Oct 08 18:13:17 crc kubenswrapper[4988]: I1008 18:13:17.034873 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-gd7f9"] Oct 08 18:13:17 crc kubenswrapper[4988]: I1008 18:13:17.100988 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-85kp6"] Oct 08 18:13:17 crc kubenswrapper[4988]: I1008 18:13:17.122547 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-l86nd" Oct 08 18:13:17 crc kubenswrapper[4988]: I1008 18:13:17.124465 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-727n7\" (UniqueName: \"kubernetes.io/projected/db1043d1-7232-4fcd-9ce5-50b590798d36-kube-api-access-727n7\") pod \"ingress-canary-s4r2k\" (UID: \"db1043d1-7232-4fcd-9ce5-50b590798d36\") " pod="openshift-ingress-canary/ingress-canary-s4r2k" Oct 08 18:13:17 crc kubenswrapper[4988]: I1008 18:13:17.124532 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/31a216f1-4766-4836-b3c5-a245ad42d709-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-hgnvq\" (UID: \"31a216f1-4766-4836-b3c5-a245ad42d709\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-hgnvq" Oct 08 18:13:17 crc kubenswrapper[4988]: I1008 18:13:17.124559 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/7ee72f71-70e8-49ff-8407-96531bc67543-signing-cabundle\") pod \"service-ca-9c57cc56f-wxv6d\" (UID: \"7ee72f71-70e8-49ff-8407-96531bc67543\") " pod="openshift-service-ca/service-ca-9c57cc56f-wxv6d" Oct 08 18:13:17 crc kubenswrapper[4988]: I1008 18:13:17.124612 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9778cf39-b76d-4de6-b7ce-6cd9f3a4e224-serving-cert\") pod \"service-ca-operator-777779d784-g775v\" (UID: \"9778cf39-b76d-4de6-b7ce-6cd9f3a4e224\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-g775v" Oct 08 18:13:17 crc kubenswrapper[4988]: I1008 18:13:17.124640 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/05152e1e-3bc1-48b0-818d-41bc0d50fbf5-encryption-config\") pod \"apiserver-76f77b778f-4phs4\" (UID: \"05152e1e-3bc1-48b0-818d-41bc0d50fbf5\") " pod="openshift-apiserver/apiserver-76f77b778f-4phs4" Oct 08 18:13:17 crc kubenswrapper[4988]: I1008 18:13:17.124663 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7kspv\" (UniqueName: \"kubernetes.io/projected/66bebf42-9ec1-48ae-ae89-7586af40d06f-kube-api-access-7kspv\") pod \"cluster-samples-operator-665b6dd947-984jv\" (UID: \"66bebf42-9ec1-48ae-ae89-7586af40d06f\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-984jv" Oct 08 18:13:17 crc kubenswrapper[4988]: I1008 18:13:17.124692 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/ed79f1d0-1282-4be1-8e71-68a5320cdab1-service-ca-bundle\") pod \"router-default-5444994796-jjnkn\" (UID: \"ed79f1d0-1282-4be1-8e71-68a5320cdab1\") " pod="openshift-ingress/router-default-5444994796-jjnkn" Oct 08 18:13:17 crc kubenswrapper[4988]: I1008 18:13:17.124723 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8a33d7bd-26bd-409b-93f1-5f151ebca347-ca-trust-extracted\") pod \"image-registry-697d97f7c8-4wh9p\" (UID: \"8a33d7bd-26bd-409b-93f1-5f151ebca347\") " pod="openshift-image-registry/image-registry-697d97f7c8-4wh9p" Oct 08 18:13:17 crc kubenswrapper[4988]: I1008 18:13:17.124749 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d4psg\" (UniqueName: \"kubernetes.io/projected/8bf9175c-5101-4fcd-8cba-e113be11dde4-kube-api-access-d4psg\") pod \"marketplace-operator-79b997595-294dt\" (UID: \"8bf9175c-5101-4fcd-8cba-e113be11dde4\") " pod="openshift-marketplace/marketplace-operator-79b997595-294dt" Oct 08 18:13:17 crc kubenswrapper[4988]: I1008 18:13:17.124795 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8a33d7bd-26bd-409b-93f1-5f151ebca347-installation-pull-secrets\") pod \"image-registry-697d97f7c8-4wh9p\" (UID: \"8a33d7bd-26bd-409b-93f1-5f151ebca347\") " pod="openshift-image-registry/image-registry-697d97f7c8-4wh9p" Oct 08 18:13:17 crc kubenswrapper[4988]: I1008 18:13:17.124820 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5qpkc\" (UniqueName: \"kubernetes.io/projected/176096f3-5341-4aa9-a706-5e96cfd8eca1-kube-api-access-5qpkc\") pod \"olm-operator-6b444d44fb-wnkxn\" (UID: \"176096f3-5341-4aa9-a706-5e96cfd8eca1\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-wnkxn" Oct 08 18:13:17 crc kubenswrapper[4988]: I1008 18:13:17.124875 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/05152e1e-3bc1-48b0-818d-41bc0d50fbf5-etcd-serving-ca\") pod \"apiserver-76f77b778f-4phs4\" (UID: \"05152e1e-3bc1-48b0-818d-41bc0d50fbf5\") " pod="openshift-apiserver/apiserver-76f77b778f-4phs4" Oct 08 18:13:17 crc kubenswrapper[4988]: I1008 18:13:17.124898 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/176096f3-5341-4aa9-a706-5e96cfd8eca1-profile-collector-cert\") pod \"olm-operator-6b444d44fb-wnkxn\" (UID: \"176096f3-5341-4aa9-a706-5e96cfd8eca1\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-wnkxn" Oct 08 18:13:17 crc kubenswrapper[4988]: I1008 18:13:17.124935 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cz4bv\" (UniqueName: \"kubernetes.io/projected/870431bf-3dfc-4ac5-a68a-39de04683899-kube-api-access-cz4bv\") pod \"machine-config-server-5hj6s\" (UID: \"870431bf-3dfc-4ac5-a68a-39de04683899\") " pod="openshift-machine-config-operator/machine-config-server-5hj6s" Oct 08 18:13:17 crc kubenswrapper[4988]: I1008 18:13:17.125103 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-46d7g\" (UniqueName: \"kubernetes.io/projected/47f38446-8f8d-4d84-8a16-f74e6f2836b7-kube-api-access-46d7g\") pod \"migrator-59844c95c7-g7fbv\" (UID: \"47f38446-8f8d-4d84-8a16-f74e6f2836b7\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-g7fbv" Oct 08 18:13:17 crc kubenswrapper[4988]: I1008 18:13:17.125149 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5hgmh\" (UniqueName: \"kubernetes.io/projected/e6f22dd3-65b7-4fe3-8265-8fb96095b6c4-kube-api-access-5hgmh\") pod \"dns-operator-744455d44c-2d794\" (UID: \"e6f22dd3-65b7-4fe3-8265-8fb96095b6c4\") " pod="openshift-dns-operator/dns-operator-744455d44c-2d794" Oct 08 18:13:17 crc kubenswrapper[4988]: I1008 18:13:17.125212 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8a33d7bd-26bd-409b-93f1-5f151ebca347-trusted-ca\") pod \"image-registry-697d97f7c8-4wh9p\" (UID: \"8a33d7bd-26bd-409b-93f1-5f151ebca347\") " pod="openshift-image-registry/image-registry-697d97f7c8-4wh9p" Oct 08 18:13:17 crc kubenswrapper[4988]: I1008 18:13:17.125246 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c08e9af6-fc25-4c32-bd25-ac82aebb828e-config-volume\") pod \"collect-profiles-29332440-qxt8r\" (UID: \"c08e9af6-fc25-4c32-bd25-ac82aebb828e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29332440-qxt8r" Oct 08 18:13:17 crc kubenswrapper[4988]: I1008 18:13:17.125270 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/a760b68f-a1f6-4741-a9e9-51f6d3968748-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-5mrrc\" (UID: \"a760b68f-a1f6-4741-a9e9-51f6d3968748\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-5mrrc" Oct 08 18:13:17 crc kubenswrapper[4988]: I1008 18:13:17.125327 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/05152e1e-3bc1-48b0-818d-41bc0d50fbf5-etcd-client\") pod \"apiserver-76f77b778f-4phs4\" (UID: \"05152e1e-3bc1-48b0-818d-41bc0d50fbf5\") " pod="openshift-apiserver/apiserver-76f77b778f-4phs4" Oct 08 18:13:17 crc kubenswrapper[4988]: I1008 18:13:17.125350 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8a33d7bd-26bd-409b-93f1-5f151ebca347-registry-tls\") pod \"image-registry-697d97f7c8-4wh9p\" (UID: \"8a33d7bd-26bd-409b-93f1-5f151ebca347\") " pod="openshift-image-registry/image-registry-697d97f7c8-4wh9p" Oct 08 18:13:17 crc kubenswrapper[4988]: I1008 18:13:17.125370 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/ed79f1d0-1282-4be1-8e71-68a5320cdab1-stats-auth\") pod \"router-default-5444994796-jjnkn\" (UID: \"ed79f1d0-1282-4be1-8e71-68a5320cdab1\") " pod="openshift-ingress/router-default-5444994796-jjnkn" Oct 08 18:13:17 crc kubenswrapper[4988]: I1008 18:13:17.126060 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/05152e1e-3bc1-48b0-818d-41bc0d50fbf5-audit-dir\") pod \"apiserver-76f77b778f-4phs4\" (UID: \"05152e1e-3bc1-48b0-818d-41bc0d50fbf5\") " pod="openshift-apiserver/apiserver-76f77b778f-4phs4" Oct 08 18:13:17 crc kubenswrapper[4988]: I1008 18:13:17.126109 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/176096f3-5341-4aa9-a706-5e96cfd8eca1-srv-cert\") pod \"olm-operator-6b444d44fb-wnkxn\" (UID: \"176096f3-5341-4aa9-a706-5e96cfd8eca1\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-wnkxn" Oct 08 18:13:17 crc kubenswrapper[4988]: I1008 18:13:17.126133 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/4ddacc00-9f9b-44ec-8356-cd932ce26e5b-images\") pod \"machine-config-operator-74547568cd-bbpzz\" (UID: \"4ddacc00-9f9b-44ec-8356-cd932ce26e5b\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-bbpzz" Oct 08 18:13:17 crc kubenswrapper[4988]: I1008 18:13:17.126184 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/4ddacc00-9f9b-44ec-8356-cd932ce26e5b-auth-proxy-config\") pod \"machine-config-operator-74547568cd-bbpzz\" (UID: \"4ddacc00-9f9b-44ec-8356-cd932ce26e5b\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-bbpzz" Oct 08 18:13:17 crc kubenswrapper[4988]: I1008 18:13:17.126225 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/ba2264e5-af3b-4400-9c20-04905b075dfa-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-h5cw8\" (UID: \"ba2264e5-af3b-4400-9c20-04905b075dfa\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-h5cw8" Oct 08 18:13:17 crc kubenswrapper[4988]: I1008 18:13:17.127560 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/0a5e45fa-e12a-4199-aa81-96468c383594-apiservice-cert\") pod \"packageserver-d55dfcdfc-d9ks8\" (UID: \"0a5e45fa-e12a-4199-aa81-96468c383594\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-d9ks8" Oct 08 18:13:17 crc kubenswrapper[4988]: I1008 18:13:17.127608 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/0a5e45fa-e12a-4199-aa81-96468c383594-webhook-cert\") pod \"packageserver-d55dfcdfc-d9ks8\" (UID: \"0a5e45fa-e12a-4199-aa81-96468c383594\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-d9ks8" Oct 08 18:13:17 crc kubenswrapper[4988]: I1008 18:13:17.127708 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a760b68f-a1f6-4741-a9e9-51f6d3968748-config\") pod \"machine-api-operator-5694c8668f-5mrrc\" (UID: \"a760b68f-a1f6-4741-a9e9-51f6d3968748\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-5mrrc" Oct 08 18:13:17 crc kubenswrapper[4988]: I1008 18:13:17.127800 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/05152e1e-3bc1-48b0-818d-41bc0d50fbf5-trusted-ca-bundle\") pod \"apiserver-76f77b778f-4phs4\" (UID: \"05152e1e-3bc1-48b0-818d-41bc0d50fbf5\") " pod="openshift-apiserver/apiserver-76f77b778f-4phs4" Oct 08 18:13:17 crc kubenswrapper[4988]: I1008 18:13:17.127821 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/0a5e45fa-e12a-4199-aa81-96468c383594-tmpfs\") pod \"packageserver-d55dfcdfc-d9ks8\" (UID: \"0a5e45fa-e12a-4199-aa81-96468c383594\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-d9ks8" Oct 08 18:13:17 crc kubenswrapper[4988]: I1008 18:13:17.128472 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/ed79f1d0-1282-4be1-8e71-68a5320cdab1-default-certificate\") pod \"router-default-5444994796-jjnkn\" (UID: \"ed79f1d0-1282-4be1-8e71-68a5320cdab1\") " pod="openshift-ingress/router-default-5444994796-jjnkn" Oct 08 18:13:17 crc kubenswrapper[4988]: I1008 18:13:17.128524 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8bf9175c-5101-4fcd-8cba-e113be11dde4-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-294dt\" (UID: \"8bf9175c-5101-4fcd-8cba-e113be11dde4\") " pod="openshift-marketplace/marketplace-operator-79b997595-294dt" Oct 08 18:13:17 crc kubenswrapper[4988]: I1008 18:13:17.129180 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z9pjn\" (UniqueName: \"kubernetes.io/projected/05152e1e-3bc1-48b0-818d-41bc0d50fbf5-kube-api-access-z9pjn\") pod \"apiserver-76f77b778f-4phs4\" (UID: \"05152e1e-3bc1-48b0-818d-41bc0d50fbf5\") " pod="openshift-apiserver/apiserver-76f77b778f-4phs4" Oct 08 18:13:17 crc kubenswrapper[4988]: I1008 18:13:17.129311 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kq442\" (UniqueName: \"kubernetes.io/projected/ed79f1d0-1282-4be1-8e71-68a5320cdab1-kube-api-access-kq442\") pod \"router-default-5444994796-jjnkn\" (UID: \"ed79f1d0-1282-4be1-8e71-68a5320cdab1\") " pod="openshift-ingress/router-default-5444994796-jjnkn" Oct 08 18:13:17 crc kubenswrapper[4988]: I1008 18:13:17.129337 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/4ddacc00-9f9b-44ec-8356-cd932ce26e5b-proxy-tls\") pod \"machine-config-operator-74547568cd-bbpzz\" (UID: \"4ddacc00-9f9b-44ec-8356-cd932ce26e5b\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-bbpzz" Oct 08 18:13:17 crc kubenswrapper[4988]: I1008 18:13:17.129355 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/870431bf-3dfc-4ac5-a68a-39de04683899-certs\") pod \"machine-config-server-5hj6s\" (UID: \"870431bf-3dfc-4ac5-a68a-39de04683899\") " pod="openshift-machine-config-operator/machine-config-server-5hj6s" Oct 08 18:13:17 crc kubenswrapper[4988]: I1008 18:13:17.148832 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-k4ghh" Oct 08 18:13:17 crc kubenswrapper[4988]: I1008 18:13:17.150805 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-q2w8f" Oct 08 18:13:17 crc kubenswrapper[4988]: I1008 18:13:17.151039 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9778cf39-b76d-4de6-b7ce-6cd9f3a4e224-config\") pod \"service-ca-operator-777779d784-g775v\" (UID: \"9778cf39-b76d-4de6-b7ce-6cd9f3a4e224\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-g775v" Oct 08 18:13:17 crc kubenswrapper[4988]: I1008 18:13:17.151586 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-jrt4j" Oct 08 18:13:17 crc kubenswrapper[4988]: I1008 18:13:17.151814 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/05152e1e-3bc1-48b0-818d-41bc0d50fbf5-serving-cert\") pod \"apiserver-76f77b778f-4phs4\" (UID: \"05152e1e-3bc1-48b0-818d-41bc0d50fbf5\") " pod="openshift-apiserver/apiserver-76f77b778f-4phs4" Oct 08 18:13:17 crc kubenswrapper[4988]: I1008 18:13:17.151892 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-4wh9p\" (UID: \"8a33d7bd-26bd-409b-93f1-5f151ebca347\") " pod="openshift-image-registry/image-registry-697d97f7c8-4wh9p" Oct 08 18:13:17 crc kubenswrapper[4988]: I1008 18:13:17.151948 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bthq4\" (UniqueName: \"kubernetes.io/projected/ba2264e5-af3b-4400-9c20-04905b075dfa-kube-api-access-bthq4\") pod \"package-server-manager-789f6589d5-h5cw8\" (UID: \"ba2264e5-af3b-4400-9c20-04905b075dfa\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-h5cw8" Oct 08 18:13:17 crc kubenswrapper[4988]: I1008 18:13:17.152092 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-572bk\" (UniqueName: \"kubernetes.io/projected/a760b68f-a1f6-4741-a9e9-51f6d3968748-kube-api-access-572bk\") pod \"machine-api-operator-5694c8668f-5mrrc\" (UID: \"a760b68f-a1f6-4741-a9e9-51f6d3968748\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-5mrrc" Oct 08 18:13:17 crc kubenswrapper[4988]: I1008 18:13:17.152123 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/7ee72f71-70e8-49ff-8407-96531bc67543-signing-key\") pod \"service-ca-9c57cc56f-wxv6d\" (UID: \"7ee72f71-70e8-49ff-8407-96531bc67543\") " pod="openshift-service-ca/service-ca-9c57cc56f-wxv6d" Oct 08 18:13:17 crc kubenswrapper[4988]: I1008 18:13:17.152164 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9gmqs\" (UniqueName: \"kubernetes.io/projected/7ee72f71-70e8-49ff-8407-96531bc67543-kube-api-access-9gmqs\") pod \"service-ca-9c57cc56f-wxv6d\" (UID: \"7ee72f71-70e8-49ff-8407-96531bc67543\") " pod="openshift-service-ca/service-ca-9c57cc56f-wxv6d" Oct 08 18:13:17 crc kubenswrapper[4988]: I1008 18:13:17.152232 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/05152e1e-3bc1-48b0-818d-41bc0d50fbf5-audit\") pod \"apiserver-76f77b778f-4phs4\" (UID: \"05152e1e-3bc1-48b0-818d-41bc0d50fbf5\") " pod="openshift-apiserver/apiserver-76f77b778f-4phs4" Oct 08 18:13:17 crc kubenswrapper[4988]: I1008 18:13:17.152272 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6jts7\" (UniqueName: \"kubernetes.io/projected/c08e9af6-fc25-4c32-bd25-ac82aebb828e-kube-api-access-6jts7\") pod \"collect-profiles-29332440-qxt8r\" (UID: \"c08e9af6-fc25-4c32-bd25-ac82aebb828e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29332440-qxt8r" Oct 08 18:13:17 crc kubenswrapper[4988]: I1008 18:13:17.152306 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/66bebf42-9ec1-48ae-ae89-7586af40d06f-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-984jv\" (UID: \"66bebf42-9ec1-48ae-ae89-7586af40d06f\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-984jv" Oct 08 18:13:17 crc kubenswrapper[4988]: I1008 18:13:17.152344 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/ed79f1d0-1282-4be1-8e71-68a5320cdab1-metrics-certs\") pod \"router-default-5444994796-jjnkn\" (UID: \"ed79f1d0-1282-4be1-8e71-68a5320cdab1\") " pod="openshift-ingress/router-default-5444994796-jjnkn" Oct 08 18:13:17 crc kubenswrapper[4988]: I1008 18:13:17.152943 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/e6f22dd3-65b7-4fe3-8265-8fb96095b6c4-metrics-tls\") pod \"dns-operator-744455d44c-2d794\" (UID: \"e6f22dd3-65b7-4fe3-8265-8fb96095b6c4\") " pod="openshift-dns-operator/dns-operator-744455d44c-2d794" Oct 08 18:13:17 crc kubenswrapper[4988]: I1008 18:13:17.153009 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jp6mk\" (UniqueName: \"kubernetes.io/projected/9778cf39-b76d-4de6-b7ce-6cd9f3a4e224-kube-api-access-jp6mk\") pod \"service-ca-operator-777779d784-g775v\" (UID: \"9778cf39-b76d-4de6-b7ce-6cd9f3a4e224\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-g775v" Oct 08 18:13:17 crc kubenswrapper[4988]: I1008 18:13:17.153093 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kxfl7\" (UniqueName: \"kubernetes.io/projected/8a33d7bd-26bd-409b-93f1-5f151ebca347-kube-api-access-kxfl7\") pod \"image-registry-697d97f7c8-4wh9p\" (UID: \"8a33d7bd-26bd-409b-93f1-5f151ebca347\") " pod="openshift-image-registry/image-registry-697d97f7c8-4wh9p" Oct 08 18:13:17 crc kubenswrapper[4988]: I1008 18:13:17.153131 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/a760b68f-a1f6-4741-a9e9-51f6d3968748-images\") pod \"machine-api-operator-5694c8668f-5mrrc\" (UID: \"a760b68f-a1f6-4741-a9e9-51f6d3968748\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-5mrrc" Oct 08 18:13:17 crc kubenswrapper[4988]: I1008 18:13:17.153161 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/05152e1e-3bc1-48b0-818d-41bc0d50fbf5-image-import-ca\") pod \"apiserver-76f77b778f-4phs4\" (UID: \"05152e1e-3bc1-48b0-818d-41bc0d50fbf5\") " pod="openshift-apiserver/apiserver-76f77b778f-4phs4" Oct 08 18:13:17 crc kubenswrapper[4988]: I1008 18:13:17.153227 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vgkbx\" (UniqueName: \"kubernetes.io/projected/4ddacc00-9f9b-44ec-8356-cd932ce26e5b-kube-api-access-vgkbx\") pod \"machine-config-operator-74547568cd-bbpzz\" (UID: \"4ddacc00-9f9b-44ec-8356-cd932ce26e5b\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-bbpzz" Oct 08 18:13:17 crc kubenswrapper[4988]: I1008 18:13:17.153312 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v68bb\" (UniqueName: \"kubernetes.io/projected/0a5e45fa-e12a-4199-aa81-96468c383594-kube-api-access-v68bb\") pod \"packageserver-d55dfcdfc-d9ks8\" (UID: \"0a5e45fa-e12a-4199-aa81-96468c383594\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-d9ks8" Oct 08 18:13:17 crc kubenswrapper[4988]: I1008 18:13:17.153617 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8a33d7bd-26bd-409b-93f1-5f151ebca347-registry-certificates\") pod \"image-registry-697d97f7c8-4wh9p\" (UID: \"8a33d7bd-26bd-409b-93f1-5f151ebca347\") " pod="openshift-image-registry/image-registry-697d97f7c8-4wh9p" Oct 08 18:13:17 crc kubenswrapper[4988]: E1008 18:13:17.153830 4988 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-08 18:13:17.653811093 +0000 UTC m=+143.103653863 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-4wh9p" (UID: "8a33d7bd-26bd-409b-93f1-5f151ebca347") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 18:13:17 crc kubenswrapper[4988]: I1008 18:13:17.154828 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/05152e1e-3bc1-48b0-818d-41bc0d50fbf5-node-pullsecrets\") pod \"apiserver-76f77b778f-4phs4\" (UID: \"05152e1e-3bc1-48b0-818d-41bc0d50fbf5\") " pod="openshift-apiserver/apiserver-76f77b778f-4phs4" Oct 08 18:13:17 crc kubenswrapper[4988]: I1008 18:13:17.154929 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/db1043d1-7232-4fcd-9ce5-50b590798d36-cert\") pod \"ingress-canary-s4r2k\" (UID: \"db1043d1-7232-4fcd-9ce5-50b590798d36\") " pod="openshift-ingress-canary/ingress-canary-s4r2k" Oct 08 18:13:17 crc kubenswrapper[4988]: I1008 18:13:17.154998 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/8bf9175c-5101-4fcd-8cba-e113be11dde4-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-294dt\" (UID: \"8bf9175c-5101-4fcd-8cba-e113be11dde4\") " pod="openshift-marketplace/marketplace-operator-79b997595-294dt" Oct 08 18:13:17 crc kubenswrapper[4988]: I1008 18:13:17.156920 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/c08e9af6-fc25-4c32-bd25-ac82aebb828e-secret-volume\") pod \"collect-profiles-29332440-qxt8r\" (UID: \"c08e9af6-fc25-4c32-bd25-ac82aebb828e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29332440-qxt8r" Oct 08 18:13:17 crc kubenswrapper[4988]: I1008 18:13:17.156949 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8a33d7bd-26bd-409b-93f1-5f151ebca347-bound-sa-token\") pod \"image-registry-697d97f7c8-4wh9p\" (UID: \"8a33d7bd-26bd-409b-93f1-5f151ebca347\") " pod="openshift-image-registry/image-registry-697d97f7c8-4wh9p" Oct 08 18:13:17 crc kubenswrapper[4988]: I1008 18:13:17.157981 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/05152e1e-3bc1-48b0-818d-41bc0d50fbf5-config\") pod \"apiserver-76f77b778f-4phs4\" (UID: \"05152e1e-3bc1-48b0-818d-41bc0d50fbf5\") " pod="openshift-apiserver/apiserver-76f77b778f-4phs4" Oct 08 18:13:17 crc kubenswrapper[4988]: I1008 18:13:17.158974 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/870431bf-3dfc-4ac5-a68a-39de04683899-node-bootstrap-token\") pod \"machine-config-server-5hj6s\" (UID: \"870431bf-3dfc-4ac5-a68a-39de04683899\") " pod="openshift-machine-config-operator/machine-config-server-5hj6s" Oct 08 18:13:17 crc kubenswrapper[4988]: I1008 18:13:17.159176 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wg4r7\" (UniqueName: \"kubernetes.io/projected/31a216f1-4766-4836-b3c5-a245ad42d709-kube-api-access-wg4r7\") pod \"control-plane-machine-set-operator-78cbb6b69f-hgnvq\" (UID: \"31a216f1-4766-4836-b3c5-a245ad42d709\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-hgnvq" Oct 08 18:13:17 crc kubenswrapper[4988]: I1008 18:13:17.166679 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-7ctzs" event={"ID":"63fe20f8-74c8-48ba-8c65-7f284d959254","Type":"ContainerStarted","Data":"0675e820e0952a216a0329fdd40282df85d5f1ae68600976871b49795bb50d5b"} Oct 08 18:13:17 crc kubenswrapper[4988]: I1008 18:13:17.167195 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-config-operator/openshift-config-operator-7777fb866f-7ctzs" Oct 08 18:13:17 crc kubenswrapper[4988]: I1008 18:13:17.171235 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-m55d6" Oct 08 18:13:17 crc kubenswrapper[4988]: I1008 18:13:17.180213 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-gd7f9" event={"ID":"5df90a93-b572-4439-b3aa-2666b2873c1d","Type":"ContainerStarted","Data":"b20de68b1bde58d505848ab392d577b41545089a287c6b13a4dab4070e060506"} Oct 08 18:13:17 crc kubenswrapper[4988]: I1008 18:13:17.199364 4988 patch_prober.go:28] interesting pod/downloads-7954f5f757-qcgmb container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.13:8080/\": dial tcp 10.217.0.13:8080: connect: connection refused" start-of-body= Oct 08 18:13:17 crc kubenswrapper[4988]: I1008 18:13:17.199458 4988 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-qcgmb" podUID="0bc2144f-4dbe-4dc1-a07a-b764d8582324" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.13:8080/\": dial tcp 10.217.0.13:8080: connect: connection refused" Oct 08 18:13:17 crc kubenswrapper[4988]: I1008 18:13:17.200139 4988 patch_prober.go:28] interesting pod/console-operator-58897d9998-rt48k container/console-operator namespace/openshift-console-operator: Readiness probe status=failure output="Get \"https://10.217.0.10:8443/readyz\": dial tcp 10.217.0.10:8443: connect: connection refused" start-of-body= Oct 08 18:13:17 crc kubenswrapper[4988]: I1008 18:13:17.200232 4988 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console-operator/console-operator-58897d9998-rt48k" podUID="eeee83f5-fe8f-47bf-be7b-cde1ea9ac356" containerName="console-operator" probeResult="failure" output="Get \"https://10.217.0.10:8443/readyz\": dial tcp 10.217.0.10:8443: connect: connection refused" Oct 08 18:13:17 crc kubenswrapper[4988]: I1008 18:13:17.203176 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-558db77b4-hwtzl" Oct 08 18:13:17 crc kubenswrapper[4988]: I1008 18:13:17.203336 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-sndzf"] Oct 08 18:13:17 crc kubenswrapper[4988]: I1008 18:13:17.209901 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-q2mxf" Oct 08 18:13:17 crc kubenswrapper[4988]: I1008 18:13:17.252691 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-qmwln"] Oct 08 18:13:17 crc kubenswrapper[4988]: I1008 18:13:17.260265 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 18:13:17 crc kubenswrapper[4988]: I1008 18:13:17.260456 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wg4r7\" (UniqueName: \"kubernetes.io/projected/31a216f1-4766-4836-b3c5-a245ad42d709-kube-api-access-wg4r7\") pod \"control-plane-machine-set-operator-78cbb6b69f-hgnvq\" (UID: \"31a216f1-4766-4836-b3c5-a245ad42d709\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-hgnvq" Oct 08 18:13:17 crc kubenswrapper[4988]: I1008 18:13:17.260507 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-727n7\" (UniqueName: \"kubernetes.io/projected/db1043d1-7232-4fcd-9ce5-50b590798d36-kube-api-access-727n7\") pod \"ingress-canary-s4r2k\" (UID: \"db1043d1-7232-4fcd-9ce5-50b590798d36\") " pod="openshift-ingress-canary/ingress-canary-s4r2k" Oct 08 18:13:17 crc kubenswrapper[4988]: I1008 18:13:17.260531 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/31a216f1-4766-4836-b3c5-a245ad42d709-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-hgnvq\" (UID: \"31a216f1-4766-4836-b3c5-a245ad42d709\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-hgnvq" Oct 08 18:13:17 crc kubenswrapper[4988]: I1008 18:13:17.260549 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/7ee72f71-70e8-49ff-8407-96531bc67543-signing-cabundle\") pod \"service-ca-9c57cc56f-wxv6d\" (UID: \"7ee72f71-70e8-49ff-8407-96531bc67543\") " pod="openshift-service-ca/service-ca-9c57cc56f-wxv6d" Oct 08 18:13:17 crc kubenswrapper[4988]: I1008 18:13:17.260587 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9778cf39-b76d-4de6-b7ce-6cd9f3a4e224-serving-cert\") pod \"service-ca-operator-777779d784-g775v\" (UID: \"9778cf39-b76d-4de6-b7ce-6cd9f3a4e224\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-g775v" Oct 08 18:13:17 crc kubenswrapper[4988]: I1008 18:13:17.260607 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/05152e1e-3bc1-48b0-818d-41bc0d50fbf5-encryption-config\") pod \"apiserver-76f77b778f-4phs4\" (UID: \"05152e1e-3bc1-48b0-818d-41bc0d50fbf5\") " pod="openshift-apiserver/apiserver-76f77b778f-4phs4" Oct 08 18:13:17 crc kubenswrapper[4988]: I1008 18:13:17.260624 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7kspv\" (UniqueName: \"kubernetes.io/projected/66bebf42-9ec1-48ae-ae89-7586af40d06f-kube-api-access-7kspv\") pod \"cluster-samples-operator-665b6dd947-984jv\" (UID: \"66bebf42-9ec1-48ae-ae89-7586af40d06f\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-984jv" Oct 08 18:13:17 crc kubenswrapper[4988]: I1008 18:13:17.260641 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/ed79f1d0-1282-4be1-8e71-68a5320cdab1-service-ca-bundle\") pod \"router-default-5444994796-jjnkn\" (UID: \"ed79f1d0-1282-4be1-8e71-68a5320cdab1\") " pod="openshift-ingress/router-default-5444994796-jjnkn" Oct 08 18:13:17 crc kubenswrapper[4988]: I1008 18:13:17.260656 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8a33d7bd-26bd-409b-93f1-5f151ebca347-ca-trust-extracted\") pod \"image-registry-697d97f7c8-4wh9p\" (UID: \"8a33d7bd-26bd-409b-93f1-5f151ebca347\") " pod="openshift-image-registry/image-registry-697d97f7c8-4wh9p" Oct 08 18:13:17 crc kubenswrapper[4988]: I1008 18:13:17.260674 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d4psg\" (UniqueName: \"kubernetes.io/projected/8bf9175c-5101-4fcd-8cba-e113be11dde4-kube-api-access-d4psg\") pod \"marketplace-operator-79b997595-294dt\" (UID: \"8bf9175c-5101-4fcd-8cba-e113be11dde4\") " pod="openshift-marketplace/marketplace-operator-79b997595-294dt" Oct 08 18:13:17 crc kubenswrapper[4988]: I1008 18:13:17.260692 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5qpkc\" (UniqueName: \"kubernetes.io/projected/176096f3-5341-4aa9-a706-5e96cfd8eca1-kube-api-access-5qpkc\") pod \"olm-operator-6b444d44fb-wnkxn\" (UID: \"176096f3-5341-4aa9-a706-5e96cfd8eca1\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-wnkxn" Oct 08 18:13:17 crc kubenswrapper[4988]: I1008 18:13:17.260729 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8a33d7bd-26bd-409b-93f1-5f151ebca347-installation-pull-secrets\") pod \"image-registry-697d97f7c8-4wh9p\" (UID: \"8a33d7bd-26bd-409b-93f1-5f151ebca347\") " pod="openshift-image-registry/image-registry-697d97f7c8-4wh9p" Oct 08 18:13:17 crc kubenswrapper[4988]: I1008 18:13:17.260745 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/05152e1e-3bc1-48b0-818d-41bc0d50fbf5-etcd-serving-ca\") pod \"apiserver-76f77b778f-4phs4\" (UID: \"05152e1e-3bc1-48b0-818d-41bc0d50fbf5\") " pod="openshift-apiserver/apiserver-76f77b778f-4phs4" Oct 08 18:13:17 crc kubenswrapper[4988]: I1008 18:13:17.260764 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/176096f3-5341-4aa9-a706-5e96cfd8eca1-profile-collector-cert\") pod \"olm-operator-6b444d44fb-wnkxn\" (UID: \"176096f3-5341-4aa9-a706-5e96cfd8eca1\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-wnkxn" Oct 08 18:13:17 crc kubenswrapper[4988]: I1008 18:13:17.260785 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/7f0973f4-25d3-48a6-ac02-b84f9bb4f947-config-volume\") pod \"dns-default-kwcts\" (UID: \"7f0973f4-25d3-48a6-ac02-b84f9bb4f947\") " pod="openshift-dns/dns-default-kwcts" Oct 08 18:13:17 crc kubenswrapper[4988]: I1008 18:13:17.260820 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cz4bv\" (UniqueName: \"kubernetes.io/projected/870431bf-3dfc-4ac5-a68a-39de04683899-kube-api-access-cz4bv\") pod \"machine-config-server-5hj6s\" (UID: \"870431bf-3dfc-4ac5-a68a-39de04683899\") " pod="openshift-machine-config-operator/machine-config-server-5hj6s" Oct 08 18:13:17 crc kubenswrapper[4988]: I1008 18:13:17.260840 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5hgmh\" (UniqueName: \"kubernetes.io/projected/e6f22dd3-65b7-4fe3-8265-8fb96095b6c4-kube-api-access-5hgmh\") pod \"dns-operator-744455d44c-2d794\" (UID: \"e6f22dd3-65b7-4fe3-8265-8fb96095b6c4\") " pod="openshift-dns-operator/dns-operator-744455d44c-2d794" Oct 08 18:13:17 crc kubenswrapper[4988]: I1008 18:13:17.260862 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-46d7g\" (UniqueName: \"kubernetes.io/projected/47f38446-8f8d-4d84-8a16-f74e6f2836b7-kube-api-access-46d7g\") pod \"migrator-59844c95c7-g7fbv\" (UID: \"47f38446-8f8d-4d84-8a16-f74e6f2836b7\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-g7fbv" Oct 08 18:13:17 crc kubenswrapper[4988]: I1008 18:13:17.260878 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8a33d7bd-26bd-409b-93f1-5f151ebca347-trusted-ca\") pod \"image-registry-697d97f7c8-4wh9p\" (UID: \"8a33d7bd-26bd-409b-93f1-5f151ebca347\") " pod="openshift-image-registry/image-registry-697d97f7c8-4wh9p" Oct 08 18:13:17 crc kubenswrapper[4988]: I1008 18:13:17.260897 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/9cb89712-c3bf-4ed9-87f8-8124a55f5f7e-plugins-dir\") pod \"csi-hostpathplugin-gl7lr\" (UID: \"9cb89712-c3bf-4ed9-87f8-8124a55f5f7e\") " pod="hostpath-provisioner/csi-hostpathplugin-gl7lr" Oct 08 18:13:17 crc kubenswrapper[4988]: I1008 18:13:17.260916 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c08e9af6-fc25-4c32-bd25-ac82aebb828e-config-volume\") pod \"collect-profiles-29332440-qxt8r\" (UID: \"c08e9af6-fc25-4c32-bd25-ac82aebb828e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29332440-qxt8r" Oct 08 18:13:17 crc kubenswrapper[4988]: I1008 18:13:17.260957 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/a760b68f-a1f6-4741-a9e9-51f6d3968748-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-5mrrc\" (UID: \"a760b68f-a1f6-4741-a9e9-51f6d3968748\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-5mrrc" Oct 08 18:13:17 crc kubenswrapper[4988]: I1008 18:13:17.260974 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8a33d7bd-26bd-409b-93f1-5f151ebca347-registry-tls\") pod \"image-registry-697d97f7c8-4wh9p\" (UID: \"8a33d7bd-26bd-409b-93f1-5f151ebca347\") " pod="openshift-image-registry/image-registry-697d97f7c8-4wh9p" Oct 08 18:13:17 crc kubenswrapper[4988]: I1008 18:13:17.260990 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/ed79f1d0-1282-4be1-8e71-68a5320cdab1-stats-auth\") pod \"router-default-5444994796-jjnkn\" (UID: \"ed79f1d0-1282-4be1-8e71-68a5320cdab1\") " pod="openshift-ingress/router-default-5444994796-jjnkn" Oct 08 18:13:17 crc kubenswrapper[4988]: I1008 18:13:17.261006 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/05152e1e-3bc1-48b0-818d-41bc0d50fbf5-etcd-client\") pod \"apiserver-76f77b778f-4phs4\" (UID: \"05152e1e-3bc1-48b0-818d-41bc0d50fbf5\") " pod="openshift-apiserver/apiserver-76f77b778f-4phs4" Oct 08 18:13:17 crc kubenswrapper[4988]: I1008 18:13:17.261031 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/4ddacc00-9f9b-44ec-8356-cd932ce26e5b-images\") pod \"machine-config-operator-74547568cd-bbpzz\" (UID: \"4ddacc00-9f9b-44ec-8356-cd932ce26e5b\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-bbpzz" Oct 08 18:13:17 crc kubenswrapper[4988]: I1008 18:13:17.261049 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/05152e1e-3bc1-48b0-818d-41bc0d50fbf5-audit-dir\") pod \"apiserver-76f77b778f-4phs4\" (UID: \"05152e1e-3bc1-48b0-818d-41bc0d50fbf5\") " pod="openshift-apiserver/apiserver-76f77b778f-4phs4" Oct 08 18:13:17 crc kubenswrapper[4988]: I1008 18:13:17.261138 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/176096f3-5341-4aa9-a706-5e96cfd8eca1-srv-cert\") pod \"olm-operator-6b444d44fb-wnkxn\" (UID: \"176096f3-5341-4aa9-a706-5e96cfd8eca1\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-wnkxn" Oct 08 18:13:17 crc kubenswrapper[4988]: I1008 18:13:17.261330 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/05152e1e-3bc1-48b0-818d-41bc0d50fbf5-audit-dir\") pod \"apiserver-76f77b778f-4phs4\" (UID: \"05152e1e-3bc1-48b0-818d-41bc0d50fbf5\") " pod="openshift-apiserver/apiserver-76f77b778f-4phs4" Oct 08 18:13:17 crc kubenswrapper[4988]: I1008 18:13:17.261731 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/05152e1e-3bc1-48b0-818d-41bc0d50fbf5-etcd-serving-ca\") pod \"apiserver-76f77b778f-4phs4\" (UID: \"05152e1e-3bc1-48b0-818d-41bc0d50fbf5\") " pod="openshift-apiserver/apiserver-76f77b778f-4phs4" Oct 08 18:13:17 crc kubenswrapper[4988]: I1008 18:13:17.262119 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/ed79f1d0-1282-4be1-8e71-68a5320cdab1-service-ca-bundle\") pod \"router-default-5444994796-jjnkn\" (UID: \"ed79f1d0-1282-4be1-8e71-68a5320cdab1\") " pod="openshift-ingress/router-default-5444994796-jjnkn" Oct 08 18:13:17 crc kubenswrapper[4988]: I1008 18:13:17.262151 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c08e9af6-fc25-4c32-bd25-ac82aebb828e-config-volume\") pod \"collect-profiles-29332440-qxt8r\" (UID: \"c08e9af6-fc25-4c32-bd25-ac82aebb828e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29332440-qxt8r" Oct 08 18:13:17 crc kubenswrapper[4988]: I1008 18:13:17.262191 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/7ee72f71-70e8-49ff-8407-96531bc67543-signing-cabundle\") pod \"service-ca-9c57cc56f-wxv6d\" (UID: \"7ee72f71-70e8-49ff-8407-96531bc67543\") " pod="openshift-service-ca/service-ca-9c57cc56f-wxv6d" Oct 08 18:13:17 crc kubenswrapper[4988]: I1008 18:13:17.262527 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8a33d7bd-26bd-409b-93f1-5f151ebca347-ca-trust-extracted\") pod \"image-registry-697d97f7c8-4wh9p\" (UID: \"8a33d7bd-26bd-409b-93f1-5f151ebca347\") " pod="openshift-image-registry/image-registry-697d97f7c8-4wh9p" Oct 08 18:13:17 crc kubenswrapper[4988]: I1008 18:13:17.262906 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8a33d7bd-26bd-409b-93f1-5f151ebca347-trusted-ca\") pod \"image-registry-697d97f7c8-4wh9p\" (UID: \"8a33d7bd-26bd-409b-93f1-5f151ebca347\") " pod="openshift-image-registry/image-registry-697d97f7c8-4wh9p" Oct 08 18:13:17 crc kubenswrapper[4988]: I1008 18:13:17.262975 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/4ddacc00-9f9b-44ec-8356-cd932ce26e5b-auth-proxy-config\") pod \"machine-config-operator-74547568cd-bbpzz\" (UID: \"4ddacc00-9f9b-44ec-8356-cd932ce26e5b\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-bbpzz" Oct 08 18:13:17 crc kubenswrapper[4988]: E1008 18:13:17.263034 4988 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 18:13:17.763014958 +0000 UTC m=+143.212857728 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 18:13:17 crc kubenswrapper[4988]: I1008 18:13:17.264377 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8a33d7bd-26bd-409b-93f1-5f151ebca347-installation-pull-secrets\") pod \"image-registry-697d97f7c8-4wh9p\" (UID: \"8a33d7bd-26bd-409b-93f1-5f151ebca347\") " pod="openshift-image-registry/image-registry-697d97f7c8-4wh9p" Oct 08 18:13:17 crc kubenswrapper[4988]: I1008 18:13:17.264438 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/05152e1e-3bc1-48b0-818d-41bc0d50fbf5-encryption-config\") pod \"apiserver-76f77b778f-4phs4\" (UID: \"05152e1e-3bc1-48b0-818d-41bc0d50fbf5\") " pod="openshift-apiserver/apiserver-76f77b778f-4phs4" Oct 08 18:13:17 crc kubenswrapper[4988]: I1008 18:13:17.264549 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/ba2264e5-af3b-4400-9c20-04905b075dfa-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-h5cw8\" (UID: \"ba2264e5-af3b-4400-9c20-04905b075dfa\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-h5cw8" Oct 08 18:13:17 crc kubenswrapper[4988]: I1008 18:13:17.264755 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/31a216f1-4766-4836-b3c5-a245ad42d709-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-hgnvq\" (UID: \"31a216f1-4766-4836-b3c5-a245ad42d709\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-hgnvq" Oct 08 18:13:17 crc kubenswrapper[4988]: I1008 18:13:17.264972 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9778cf39-b76d-4de6-b7ce-6cd9f3a4e224-serving-cert\") pod \"service-ca-operator-777779d784-g775v\" (UID: \"9778cf39-b76d-4de6-b7ce-6cd9f3a4e224\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-g775v" Oct 08 18:13:17 crc kubenswrapper[4988]: I1008 18:13:17.265252 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/176096f3-5341-4aa9-a706-5e96cfd8eca1-srv-cert\") pod \"olm-operator-6b444d44fb-wnkxn\" (UID: \"176096f3-5341-4aa9-a706-5e96cfd8eca1\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-wnkxn" Oct 08 18:13:17 crc kubenswrapper[4988]: I1008 18:13:17.267557 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/ed79f1d0-1282-4be1-8e71-68a5320cdab1-stats-auth\") pod \"router-default-5444994796-jjnkn\" (UID: \"ed79f1d0-1282-4be1-8e71-68a5320cdab1\") " pod="openshift-ingress/router-default-5444994796-jjnkn" Oct 08 18:13:17 crc kubenswrapper[4988]: I1008 18:13:17.267802 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/05152e1e-3bc1-48b0-818d-41bc0d50fbf5-etcd-client\") pod \"apiserver-76f77b778f-4phs4\" (UID: \"05152e1e-3bc1-48b0-818d-41bc0d50fbf5\") " pod="openshift-apiserver/apiserver-76f77b778f-4phs4" Oct 08 18:13:17 crc kubenswrapper[4988]: I1008 18:13:17.273686 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/176096f3-5341-4aa9-a706-5e96cfd8eca1-profile-collector-cert\") pod \"olm-operator-6b444d44fb-wnkxn\" (UID: \"176096f3-5341-4aa9-a706-5e96cfd8eca1\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-wnkxn" Oct 08 18:13:17 crc kubenswrapper[4988]: I1008 18:13:17.274194 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/4ddacc00-9f9b-44ec-8356-cd932ce26e5b-images\") pod \"machine-config-operator-74547568cd-bbpzz\" (UID: \"4ddacc00-9f9b-44ec-8356-cd932ce26e5b\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-bbpzz" Oct 08 18:13:17 crc kubenswrapper[4988]: I1008 18:13:17.274505 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/a760b68f-a1f6-4741-a9e9-51f6d3968748-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-5mrrc\" (UID: \"a760b68f-a1f6-4741-a9e9-51f6d3968748\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-5mrrc" Oct 08 18:13:17 crc kubenswrapper[4988]: I1008 18:13:17.274085 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/4ddacc00-9f9b-44ec-8356-cd932ce26e5b-auth-proxy-config\") pod \"machine-config-operator-74547568cd-bbpzz\" (UID: \"4ddacc00-9f9b-44ec-8356-cd932ce26e5b\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-bbpzz" Oct 08 18:13:17 crc kubenswrapper[4988]: I1008 18:13:17.277180 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/0a5e45fa-e12a-4199-aa81-96468c383594-apiservice-cert\") pod \"packageserver-d55dfcdfc-d9ks8\" (UID: \"0a5e45fa-e12a-4199-aa81-96468c383594\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-d9ks8" Oct 08 18:13:17 crc kubenswrapper[4988]: I1008 18:13:17.277226 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/9cb89712-c3bf-4ed9-87f8-8124a55f5f7e-socket-dir\") pod \"csi-hostpathplugin-gl7lr\" (UID: \"9cb89712-c3bf-4ed9-87f8-8124a55f5f7e\") " pod="hostpath-provisioner/csi-hostpathplugin-gl7lr" Oct 08 18:13:17 crc kubenswrapper[4988]: I1008 18:13:17.277464 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/0a5e45fa-e12a-4199-aa81-96468c383594-webhook-cert\") pod \"packageserver-d55dfcdfc-d9ks8\" (UID: \"0a5e45fa-e12a-4199-aa81-96468c383594\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-d9ks8" Oct 08 18:13:17 crc kubenswrapper[4988]: I1008 18:13:17.277498 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/7f0973f4-25d3-48a6-ac02-b84f9bb4f947-metrics-tls\") pod \"dns-default-kwcts\" (UID: \"7f0973f4-25d3-48a6-ac02-b84f9bb4f947\") " pod="openshift-dns/dns-default-kwcts" Oct 08 18:13:17 crc kubenswrapper[4988]: I1008 18:13:17.277522 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a760b68f-a1f6-4741-a9e9-51f6d3968748-config\") pod \"machine-api-operator-5694c8668f-5mrrc\" (UID: \"a760b68f-a1f6-4741-a9e9-51f6d3968748\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-5mrrc" Oct 08 18:13:17 crc kubenswrapper[4988]: I1008 18:13:17.277732 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/05152e1e-3bc1-48b0-818d-41bc0d50fbf5-trusted-ca-bundle\") pod \"apiserver-76f77b778f-4phs4\" (UID: \"05152e1e-3bc1-48b0-818d-41bc0d50fbf5\") " pod="openshift-apiserver/apiserver-76f77b778f-4phs4" Oct 08 18:13:17 crc kubenswrapper[4988]: I1008 18:13:17.277759 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/0a5e45fa-e12a-4199-aa81-96468c383594-tmpfs\") pod \"packageserver-d55dfcdfc-d9ks8\" (UID: \"0a5e45fa-e12a-4199-aa81-96468c383594\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-d9ks8" Oct 08 18:13:17 crc kubenswrapper[4988]: I1008 18:13:17.277822 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/ed79f1d0-1282-4be1-8e71-68a5320cdab1-default-certificate\") pod \"router-default-5444994796-jjnkn\" (UID: \"ed79f1d0-1282-4be1-8e71-68a5320cdab1\") " pod="openshift-ingress/router-default-5444994796-jjnkn" Oct 08 18:13:17 crc kubenswrapper[4988]: I1008 18:13:17.277843 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8bf9175c-5101-4fcd-8cba-e113be11dde4-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-294dt\" (UID: \"8bf9175c-5101-4fcd-8cba-e113be11dde4\") " pod="openshift-marketplace/marketplace-operator-79b997595-294dt" Oct 08 18:13:17 crc kubenswrapper[4988]: I1008 18:13:17.277863 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z9pjn\" (UniqueName: \"kubernetes.io/projected/05152e1e-3bc1-48b0-818d-41bc0d50fbf5-kube-api-access-z9pjn\") pod \"apiserver-76f77b778f-4phs4\" (UID: \"05152e1e-3bc1-48b0-818d-41bc0d50fbf5\") " pod="openshift-apiserver/apiserver-76f77b778f-4phs4" Oct 08 18:13:17 crc kubenswrapper[4988]: I1008 18:13:17.278222 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/0a5e45fa-e12a-4199-aa81-96468c383594-tmpfs\") pod \"packageserver-d55dfcdfc-d9ks8\" (UID: \"0a5e45fa-e12a-4199-aa81-96468c383594\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-d9ks8" Oct 08 18:13:17 crc kubenswrapper[4988]: I1008 18:13:17.280099 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/05152e1e-3bc1-48b0-818d-41bc0d50fbf5-trusted-ca-bundle\") pod \"apiserver-76f77b778f-4phs4\" (UID: \"05152e1e-3bc1-48b0-818d-41bc0d50fbf5\") " pod="openshift-apiserver/apiserver-76f77b778f-4phs4" Oct 08 18:13:17 crc kubenswrapper[4988]: I1008 18:13:17.283987 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/0a5e45fa-e12a-4199-aa81-96468c383594-apiservice-cert\") pod \"packageserver-d55dfcdfc-d9ks8\" (UID: \"0a5e45fa-e12a-4199-aa81-96468c383594\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-d9ks8" Oct 08 18:13:17 crc kubenswrapper[4988]: I1008 18:13:17.285166 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8bf9175c-5101-4fcd-8cba-e113be11dde4-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-294dt\" (UID: \"8bf9175c-5101-4fcd-8cba-e113be11dde4\") " pod="openshift-marketplace/marketplace-operator-79b997595-294dt" Oct 08 18:13:17 crc kubenswrapper[4988]: I1008 18:13:17.285255 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kq442\" (UniqueName: \"kubernetes.io/projected/ed79f1d0-1282-4be1-8e71-68a5320cdab1-kube-api-access-kq442\") pod \"router-default-5444994796-jjnkn\" (UID: \"ed79f1d0-1282-4be1-8e71-68a5320cdab1\") " pod="openshift-ingress/router-default-5444994796-jjnkn" Oct 08 18:13:17 crc kubenswrapper[4988]: I1008 18:13:17.285292 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/4ddacc00-9f9b-44ec-8356-cd932ce26e5b-proxy-tls\") pod \"machine-config-operator-74547568cd-bbpzz\" (UID: \"4ddacc00-9f9b-44ec-8356-cd932ce26e5b\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-bbpzz" Oct 08 18:13:17 crc kubenswrapper[4988]: I1008 18:13:17.285317 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/870431bf-3dfc-4ac5-a68a-39de04683899-certs\") pod \"machine-config-server-5hj6s\" (UID: \"870431bf-3dfc-4ac5-a68a-39de04683899\") " pod="openshift-machine-config-operator/machine-config-server-5hj6s" Oct 08 18:13:17 crc kubenswrapper[4988]: I1008 18:13:17.285363 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/9cb89712-c3bf-4ed9-87f8-8124a55f5f7e-registration-dir\") pod \"csi-hostpathplugin-gl7lr\" (UID: \"9cb89712-c3bf-4ed9-87f8-8124a55f5f7e\") " pod="hostpath-provisioner/csi-hostpathplugin-gl7lr" Oct 08 18:13:17 crc kubenswrapper[4988]: I1008 18:13:17.285982 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/ba2264e5-af3b-4400-9c20-04905b075dfa-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-h5cw8\" (UID: \"ba2264e5-af3b-4400-9c20-04905b075dfa\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-h5cw8" Oct 08 18:13:17 crc kubenswrapper[4988]: I1008 18:13:17.286397 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/0a5e45fa-e12a-4199-aa81-96468c383594-webhook-cert\") pod \"packageserver-d55dfcdfc-d9ks8\" (UID: \"0a5e45fa-e12a-4199-aa81-96468c383594\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-d9ks8" Oct 08 18:13:17 crc kubenswrapper[4988]: I1008 18:13:17.286588 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9778cf39-b76d-4de6-b7ce-6cd9f3a4e224-config\") pod \"service-ca-operator-777779d784-g775v\" (UID: \"9778cf39-b76d-4de6-b7ce-6cd9f3a4e224\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-g775v" Oct 08 18:13:17 crc kubenswrapper[4988]: I1008 18:13:17.287046 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/05152e1e-3bc1-48b0-818d-41bc0d50fbf5-serving-cert\") pod \"apiserver-76f77b778f-4phs4\" (UID: \"05152e1e-3bc1-48b0-818d-41bc0d50fbf5\") " pod="openshift-apiserver/apiserver-76f77b778f-4phs4" Oct 08 18:13:17 crc kubenswrapper[4988]: I1008 18:13:17.287280 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9778cf39-b76d-4de6-b7ce-6cd9f3a4e224-config\") pod \"service-ca-operator-777779d784-g775v\" (UID: \"9778cf39-b76d-4de6-b7ce-6cd9f3a4e224\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-g775v" Oct 08 18:13:17 crc kubenswrapper[4988]: I1008 18:13:17.287353 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-4wh9p\" (UID: \"8a33d7bd-26bd-409b-93f1-5f151ebca347\") " pod="openshift-image-registry/image-registry-697d97f7c8-4wh9p" Oct 08 18:13:17 crc kubenswrapper[4988]: I1008 18:13:17.287470 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bthq4\" (UniqueName: \"kubernetes.io/projected/ba2264e5-af3b-4400-9c20-04905b075dfa-kube-api-access-bthq4\") pod \"package-server-manager-789f6589d5-h5cw8\" (UID: \"ba2264e5-af3b-4400-9c20-04905b075dfa\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-h5cw8" Oct 08 18:13:17 crc kubenswrapper[4988]: I1008 18:13:17.287500 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-572bk\" (UniqueName: \"kubernetes.io/projected/a760b68f-a1f6-4741-a9e9-51f6d3968748-kube-api-access-572bk\") pod \"machine-api-operator-5694c8668f-5mrrc\" (UID: \"a760b68f-a1f6-4741-a9e9-51f6d3968748\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-5mrrc" Oct 08 18:13:17 crc kubenswrapper[4988]: I1008 18:13:17.287543 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/7ee72f71-70e8-49ff-8407-96531bc67543-signing-key\") pod \"service-ca-9c57cc56f-wxv6d\" (UID: \"7ee72f71-70e8-49ff-8407-96531bc67543\") " pod="openshift-service-ca/service-ca-9c57cc56f-wxv6d" Oct 08 18:13:17 crc kubenswrapper[4988]: I1008 18:13:17.287564 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9gmqs\" (UniqueName: \"kubernetes.io/projected/7ee72f71-70e8-49ff-8407-96531bc67543-kube-api-access-9gmqs\") pod \"service-ca-9c57cc56f-wxv6d\" (UID: \"7ee72f71-70e8-49ff-8407-96531bc67543\") " pod="openshift-service-ca/service-ca-9c57cc56f-wxv6d" Oct 08 18:13:17 crc kubenswrapper[4988]: I1008 18:13:17.287648 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-rgqjs" Oct 08 18:13:17 crc kubenswrapper[4988]: E1008 18:13:17.287922 4988 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-08 18:13:17.787906413 +0000 UTC m=+143.237749183 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-4wh9p" (UID: "8a33d7bd-26bd-409b-93f1-5f151ebca347") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 18:13:17 crc kubenswrapper[4988]: I1008 18:13:17.288060 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/05152e1e-3bc1-48b0-818d-41bc0d50fbf5-audit\") pod \"apiserver-76f77b778f-4phs4\" (UID: \"05152e1e-3bc1-48b0-818d-41bc0d50fbf5\") " pod="openshift-apiserver/apiserver-76f77b778f-4phs4" Oct 08 18:13:17 crc kubenswrapper[4988]: I1008 18:13:17.288209 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6jts7\" (UniqueName: \"kubernetes.io/projected/c08e9af6-fc25-4c32-bd25-ac82aebb828e-kube-api-access-6jts7\") pod \"collect-profiles-29332440-qxt8r\" (UID: \"c08e9af6-fc25-4c32-bd25-ac82aebb828e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29332440-qxt8r" Oct 08 18:13:17 crc kubenswrapper[4988]: I1008 18:13:17.288310 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/66bebf42-9ec1-48ae-ae89-7586af40d06f-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-984jv\" (UID: \"66bebf42-9ec1-48ae-ae89-7586af40d06f\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-984jv" Oct 08 18:13:17 crc kubenswrapper[4988]: I1008 18:13:17.288428 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/ed79f1d0-1282-4be1-8e71-68a5320cdab1-metrics-certs\") pod \"router-default-5444994796-jjnkn\" (UID: \"ed79f1d0-1282-4be1-8e71-68a5320cdab1\") " pod="openshift-ingress/router-default-5444994796-jjnkn" Oct 08 18:13:17 crc kubenswrapper[4988]: I1008 18:13:17.288520 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/9cb89712-c3bf-4ed9-87f8-8124a55f5f7e-mountpoint-dir\") pod \"csi-hostpathplugin-gl7lr\" (UID: \"9cb89712-c3bf-4ed9-87f8-8124a55f5f7e\") " pod="hostpath-provisioner/csi-hostpathplugin-gl7lr" Oct 08 18:13:17 crc kubenswrapper[4988]: I1008 18:13:17.288771 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/e6f22dd3-65b7-4fe3-8265-8fb96095b6c4-metrics-tls\") pod \"dns-operator-744455d44c-2d794\" (UID: \"e6f22dd3-65b7-4fe3-8265-8fb96095b6c4\") " pod="openshift-dns-operator/dns-operator-744455d44c-2d794" Oct 08 18:13:17 crc kubenswrapper[4988]: I1008 18:13:17.288988 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/05152e1e-3bc1-48b0-818d-41bc0d50fbf5-audit\") pod \"apiserver-76f77b778f-4phs4\" (UID: \"05152e1e-3bc1-48b0-818d-41bc0d50fbf5\") " pod="openshift-apiserver/apiserver-76f77b778f-4phs4" Oct 08 18:13:17 crc kubenswrapper[4988]: I1008 18:13:17.289460 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-znm4b\" (UniqueName: \"kubernetes.io/projected/7f0973f4-25d3-48a6-ac02-b84f9bb4f947-kube-api-access-znm4b\") pod \"dns-default-kwcts\" (UID: \"7f0973f4-25d3-48a6-ac02-b84f9bb4f947\") " pod="openshift-dns/dns-default-kwcts" Oct 08 18:13:17 crc kubenswrapper[4988]: I1008 18:13:17.292001 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/7ee72f71-70e8-49ff-8407-96531bc67543-signing-key\") pod \"service-ca-9c57cc56f-wxv6d\" (UID: \"7ee72f71-70e8-49ff-8407-96531bc67543\") " pod="openshift-service-ca/service-ca-9c57cc56f-wxv6d" Oct 08 18:13:17 crc kubenswrapper[4988]: I1008 18:13:17.292683 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8a33d7bd-26bd-409b-93f1-5f151ebca347-registry-tls\") pod \"image-registry-697d97f7c8-4wh9p\" (UID: \"8a33d7bd-26bd-409b-93f1-5f151ebca347\") " pod="openshift-image-registry/image-registry-697d97f7c8-4wh9p" Oct 08 18:13:17 crc kubenswrapper[4988]: I1008 18:13:17.292696 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"certs\" (UniqueName: \"kubernetes.io/secret/870431bf-3dfc-4ac5-a68a-39de04683899-certs\") pod \"machine-config-server-5hj6s\" (UID: \"870431bf-3dfc-4ac5-a68a-39de04683899\") " pod="openshift-machine-config-operator/machine-config-server-5hj6s" Oct 08 18:13:17 crc kubenswrapper[4988]: I1008 18:13:17.293510 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/05152e1e-3bc1-48b0-818d-41bc0d50fbf5-serving-cert\") pod \"apiserver-76f77b778f-4phs4\" (UID: \"05152e1e-3bc1-48b0-818d-41bc0d50fbf5\") " pod="openshift-apiserver/apiserver-76f77b778f-4phs4" Oct 08 18:13:17 crc kubenswrapper[4988]: I1008 18:13:17.294095 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/66bebf42-9ec1-48ae-ae89-7586af40d06f-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-984jv\" (UID: \"66bebf42-9ec1-48ae-ae89-7586af40d06f\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-984jv" Oct 08 18:13:17 crc kubenswrapper[4988]: I1008 18:13:17.299172 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-gqm5m"] Oct 08 18:13:17 crc kubenswrapper[4988]: I1008 18:13:17.300294 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/e6f22dd3-65b7-4fe3-8265-8fb96095b6c4-metrics-tls\") pod \"dns-operator-744455d44c-2d794\" (UID: \"e6f22dd3-65b7-4fe3-8265-8fb96095b6c4\") " pod="openshift-dns-operator/dns-operator-744455d44c-2d794" Oct 08 18:13:17 crc kubenswrapper[4988]: I1008 18:13:17.300814 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jp6mk\" (UniqueName: \"kubernetes.io/projected/9778cf39-b76d-4de6-b7ce-6cd9f3a4e224-kube-api-access-jp6mk\") pod \"service-ca-operator-777779d784-g775v\" (UID: \"9778cf39-b76d-4de6-b7ce-6cd9f3a4e224\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-g775v" Oct 08 18:13:17 crc kubenswrapper[4988]: I1008 18:13:17.300868 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kxfl7\" (UniqueName: \"kubernetes.io/projected/8a33d7bd-26bd-409b-93f1-5f151ebca347-kube-api-access-kxfl7\") pod \"image-registry-697d97f7c8-4wh9p\" (UID: \"8a33d7bd-26bd-409b-93f1-5f151ebca347\") " pod="openshift-image-registry/image-registry-697d97f7c8-4wh9p" Oct 08 18:13:17 crc kubenswrapper[4988]: I1008 18:13:17.301040 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/a760b68f-a1f6-4741-a9e9-51f6d3968748-images\") pod \"machine-api-operator-5694c8668f-5mrrc\" (UID: \"a760b68f-a1f6-4741-a9e9-51f6d3968748\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-5mrrc" Oct 08 18:13:17 crc kubenswrapper[4988]: I1008 18:13:17.301190 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/05152e1e-3bc1-48b0-818d-41bc0d50fbf5-image-import-ca\") pod \"apiserver-76f77b778f-4phs4\" (UID: \"05152e1e-3bc1-48b0-818d-41bc0d50fbf5\") " pod="openshift-apiserver/apiserver-76f77b778f-4phs4" Oct 08 18:13:17 crc kubenswrapper[4988]: I1008 18:13:17.301229 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vgkbx\" (UniqueName: \"kubernetes.io/projected/4ddacc00-9f9b-44ec-8356-cd932ce26e5b-kube-api-access-vgkbx\") pod \"machine-config-operator-74547568cd-bbpzz\" (UID: \"4ddacc00-9f9b-44ec-8356-cd932ce26e5b\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-bbpzz" Oct 08 18:13:17 crc kubenswrapper[4988]: I1008 18:13:17.303183 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/a760b68f-a1f6-4741-a9e9-51f6d3968748-images\") pod \"machine-api-operator-5694c8668f-5mrrc\" (UID: \"a760b68f-a1f6-4741-a9e9-51f6d3968748\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-5mrrc" Oct 08 18:13:17 crc kubenswrapper[4988]: I1008 18:13:17.304863 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/05152e1e-3bc1-48b0-818d-41bc0d50fbf5-image-import-ca\") pod \"apiserver-76f77b778f-4phs4\" (UID: \"05152e1e-3bc1-48b0-818d-41bc0d50fbf5\") " pod="openshift-apiserver/apiserver-76f77b778f-4phs4" Oct 08 18:13:17 crc kubenswrapper[4988]: I1008 18:13:17.304944 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v68bb\" (UniqueName: \"kubernetes.io/projected/0a5e45fa-e12a-4199-aa81-96468c383594-kube-api-access-v68bb\") pod \"packageserver-d55dfcdfc-d9ks8\" (UID: \"0a5e45fa-e12a-4199-aa81-96468c383594\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-d9ks8" Oct 08 18:13:17 crc kubenswrapper[4988]: I1008 18:13:17.304997 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8a33d7bd-26bd-409b-93f1-5f151ebca347-registry-certificates\") pod \"image-registry-697d97f7c8-4wh9p\" (UID: \"8a33d7bd-26bd-409b-93f1-5f151ebca347\") " pod="openshift-image-registry/image-registry-697d97f7c8-4wh9p" Oct 08 18:13:17 crc kubenswrapper[4988]: I1008 18:13:17.305424 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/05152e1e-3bc1-48b0-818d-41bc0d50fbf5-node-pullsecrets\") pod \"apiserver-76f77b778f-4phs4\" (UID: \"05152e1e-3bc1-48b0-818d-41bc0d50fbf5\") " pod="openshift-apiserver/apiserver-76f77b778f-4phs4" Oct 08 18:13:17 crc kubenswrapper[4988]: I1008 18:13:17.305579 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/05152e1e-3bc1-48b0-818d-41bc0d50fbf5-node-pullsecrets\") pod \"apiserver-76f77b778f-4phs4\" (UID: \"05152e1e-3bc1-48b0-818d-41bc0d50fbf5\") " pod="openshift-apiserver/apiserver-76f77b778f-4phs4" Oct 08 18:13:17 crc kubenswrapper[4988]: I1008 18:13:17.306169 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/db1043d1-7232-4fcd-9ce5-50b590798d36-cert\") pod \"ingress-canary-s4r2k\" (UID: \"db1043d1-7232-4fcd-9ce5-50b590798d36\") " pod="openshift-ingress-canary/ingress-canary-s4r2k" Oct 08 18:13:17 crc kubenswrapper[4988]: I1008 18:13:17.306221 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/8bf9175c-5101-4fcd-8cba-e113be11dde4-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-294dt\" (UID: \"8bf9175c-5101-4fcd-8cba-e113be11dde4\") " pod="openshift-marketplace/marketplace-operator-79b997595-294dt" Oct 08 18:13:17 crc kubenswrapper[4988]: I1008 18:13:17.306249 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/c08e9af6-fc25-4c32-bd25-ac82aebb828e-secret-volume\") pod \"collect-profiles-29332440-qxt8r\" (UID: \"c08e9af6-fc25-4c32-bd25-ac82aebb828e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29332440-qxt8r" Oct 08 18:13:17 crc kubenswrapper[4988]: I1008 18:13:17.306313 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8a33d7bd-26bd-409b-93f1-5f151ebca347-bound-sa-token\") pod \"image-registry-697d97f7c8-4wh9p\" (UID: \"8a33d7bd-26bd-409b-93f1-5f151ebca347\") " pod="openshift-image-registry/image-registry-697d97f7c8-4wh9p" Oct 08 18:13:17 crc kubenswrapper[4988]: I1008 18:13:17.307173 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8a33d7bd-26bd-409b-93f1-5f151ebca347-registry-certificates\") pod \"image-registry-697d97f7c8-4wh9p\" (UID: \"8a33d7bd-26bd-409b-93f1-5f151ebca347\") " pod="openshift-image-registry/image-registry-697d97f7c8-4wh9p" Oct 08 18:13:17 crc kubenswrapper[4988]: I1008 18:13:17.309343 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/05152e1e-3bc1-48b0-818d-41bc0d50fbf5-config\") pod \"apiserver-76f77b778f-4phs4\" (UID: \"05152e1e-3bc1-48b0-818d-41bc0d50fbf5\") " pod="openshift-apiserver/apiserver-76f77b778f-4phs4" Oct 08 18:13:17 crc kubenswrapper[4988]: I1008 18:13:17.309515 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/870431bf-3dfc-4ac5-a68a-39de04683899-node-bootstrap-token\") pod \"machine-config-server-5hj6s\" (UID: \"870431bf-3dfc-4ac5-a68a-39de04683899\") " pod="openshift-machine-config-operator/machine-config-server-5hj6s" Oct 08 18:13:17 crc kubenswrapper[4988]: I1008 18:13:17.309569 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/9cb89712-c3bf-4ed9-87f8-8124a55f5f7e-csi-data-dir\") pod \"csi-hostpathplugin-gl7lr\" (UID: \"9cb89712-c3bf-4ed9-87f8-8124a55f5f7e\") " pod="hostpath-provisioner/csi-hostpathplugin-gl7lr" Oct 08 18:13:17 crc kubenswrapper[4988]: I1008 18:13:17.309608 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cgjkq\" (UniqueName: \"kubernetes.io/projected/9cb89712-c3bf-4ed9-87f8-8124a55f5f7e-kube-api-access-cgjkq\") pod \"csi-hostpathplugin-gl7lr\" (UID: \"9cb89712-c3bf-4ed9-87f8-8124a55f5f7e\") " pod="hostpath-provisioner/csi-hostpathplugin-gl7lr" Oct 08 18:13:17 crc kubenswrapper[4988]: I1008 18:13:17.310116 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/05152e1e-3bc1-48b0-818d-41bc0d50fbf5-config\") pod \"apiserver-76f77b778f-4phs4\" (UID: \"05152e1e-3bc1-48b0-818d-41bc0d50fbf5\") " pod="openshift-apiserver/apiserver-76f77b778f-4phs4" Oct 08 18:13:17 crc kubenswrapper[4988]: I1008 18:13:17.310486 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/8bf9175c-5101-4fcd-8cba-e113be11dde4-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-294dt\" (UID: \"8bf9175c-5101-4fcd-8cba-e113be11dde4\") " pod="openshift-marketplace/marketplace-operator-79b997595-294dt" Oct 08 18:13:17 crc kubenswrapper[4988]: I1008 18:13:17.312185 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/4ddacc00-9f9b-44ec-8356-cd932ce26e5b-proxy-tls\") pod \"machine-config-operator-74547568cd-bbpzz\" (UID: \"4ddacc00-9f9b-44ec-8356-cd932ce26e5b\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-bbpzz" Oct 08 18:13:17 crc kubenswrapper[4988]: I1008 18:13:17.313570 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/c08e9af6-fc25-4c32-bd25-ac82aebb828e-secret-volume\") pod \"collect-profiles-29332440-qxt8r\" (UID: \"c08e9af6-fc25-4c32-bd25-ac82aebb828e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29332440-qxt8r" Oct 08 18:13:17 crc kubenswrapper[4988]: I1008 18:13:17.314105 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/870431bf-3dfc-4ac5-a68a-39de04683899-node-bootstrap-token\") pod \"machine-config-server-5hj6s\" (UID: \"870431bf-3dfc-4ac5-a68a-39de04683899\") " pod="openshift-machine-config-operator/machine-config-server-5hj6s" Oct 08 18:13:17 crc kubenswrapper[4988]: I1008 18:13:17.315922 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/ed79f1d0-1282-4be1-8e71-68a5320cdab1-metrics-certs\") pod \"router-default-5444994796-jjnkn\" (UID: \"ed79f1d0-1282-4be1-8e71-68a5320cdab1\") " pod="openshift-ingress/router-default-5444994796-jjnkn" Oct 08 18:13:17 crc kubenswrapper[4988]: I1008 18:13:17.316237 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/db1043d1-7232-4fcd-9ce5-50b590798d36-cert\") pod \"ingress-canary-s4r2k\" (UID: \"db1043d1-7232-4fcd-9ce5-50b590798d36\") " pod="openshift-ingress-canary/ingress-canary-s4r2k" Oct 08 18:13:17 crc kubenswrapper[4988]: I1008 18:13:17.316592 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wg4r7\" (UniqueName: \"kubernetes.io/projected/31a216f1-4766-4836-b3c5-a245ad42d709-kube-api-access-wg4r7\") pod \"control-plane-machine-set-operator-78cbb6b69f-hgnvq\" (UID: \"31a216f1-4766-4836-b3c5-a245ad42d709\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-hgnvq" Oct 08 18:13:17 crc kubenswrapper[4988]: I1008 18:13:17.317208 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/ed79f1d0-1282-4be1-8e71-68a5320cdab1-default-certificate\") pod \"router-default-5444994796-jjnkn\" (UID: \"ed79f1d0-1282-4be1-8e71-68a5320cdab1\") " pod="openshift-ingress/router-default-5444994796-jjnkn" Oct 08 18:13:17 crc kubenswrapper[4988]: I1008 18:13:17.325903 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a760b68f-a1f6-4741-a9e9-51f6d3968748-config\") pod \"machine-api-operator-5694c8668f-5mrrc\" (UID: \"a760b68f-a1f6-4741-a9e9-51f6d3968748\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-5mrrc" Oct 08 18:13:17 crc kubenswrapper[4988]: I1008 18:13:17.332369 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d4psg\" (UniqueName: \"kubernetes.io/projected/8bf9175c-5101-4fcd-8cba-e113be11dde4-kube-api-access-d4psg\") pod \"marketplace-operator-79b997595-294dt\" (UID: \"8bf9175c-5101-4fcd-8cba-e113be11dde4\") " pod="openshift-marketplace/marketplace-operator-79b997595-294dt" Oct 08 18:13:17 crc kubenswrapper[4988]: I1008 18:13:17.332649 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-294dt" Oct 08 18:13:17 crc kubenswrapper[4988]: I1008 18:13:17.345361 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7kspv\" (UniqueName: \"kubernetes.io/projected/66bebf42-9ec1-48ae-ae89-7586af40d06f-kube-api-access-7kspv\") pod \"cluster-samples-operator-665b6dd947-984jv\" (UID: \"66bebf42-9ec1-48ae-ae89-7586af40d06f\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-984jv" Oct 08 18:13:17 crc kubenswrapper[4988]: I1008 18:13:17.346489 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-7xdnc"] Oct 08 18:13:17 crc kubenswrapper[4988]: I1008 18:13:17.376177 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5qpkc\" (UniqueName: \"kubernetes.io/projected/176096f3-5341-4aa9-a706-5e96cfd8eca1-kube-api-access-5qpkc\") pod \"olm-operator-6b444d44fb-wnkxn\" (UID: \"176096f3-5341-4aa9-a706-5e96cfd8eca1\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-wnkxn" Oct 08 18:13:17 crc kubenswrapper[4988]: I1008 18:13:17.383942 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-727n7\" (UniqueName: \"kubernetes.io/projected/db1043d1-7232-4fcd-9ce5-50b590798d36-kube-api-access-727n7\") pod \"ingress-canary-s4r2k\" (UID: \"db1043d1-7232-4fcd-9ce5-50b590798d36\") " pod="openshift-ingress-canary/ingress-canary-s4r2k" Oct 08 18:13:17 crc kubenswrapper[4988]: W1008 18:13:17.390304 4988 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9c8874ed_3f70_45b6_9f10_d8737b6a7a8f.slice/crio-04d2296a013043a794ce84eebbf03997f04316d51b2603d40464866e8c4e87cd WatchSource:0}: Error finding container 04d2296a013043a794ce84eebbf03997f04316d51b2603d40464866e8c4e87cd: Status 404 returned error can't find the container with id 04d2296a013043a794ce84eebbf03997f04316d51b2603d40464866e8c4e87cd Oct 08 18:13:17 crc kubenswrapper[4988]: I1008 18:13:17.396910 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cz4bv\" (UniqueName: \"kubernetes.io/projected/870431bf-3dfc-4ac5-a68a-39de04683899-kube-api-access-cz4bv\") pod \"machine-config-server-5hj6s\" (UID: \"870431bf-3dfc-4ac5-a68a-39de04683899\") " pod="openshift-machine-config-operator/machine-config-server-5hj6s" Oct 08 18:13:17 crc kubenswrapper[4988]: I1008 18:13:17.411324 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 18:13:17 crc kubenswrapper[4988]: I1008 18:13:17.411614 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/7f0973f4-25d3-48a6-ac02-b84f9bb4f947-config-volume\") pod \"dns-default-kwcts\" (UID: \"7f0973f4-25d3-48a6-ac02-b84f9bb4f947\") " pod="openshift-dns/dns-default-kwcts" Oct 08 18:13:17 crc kubenswrapper[4988]: I1008 18:13:17.411698 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/9cb89712-c3bf-4ed9-87f8-8124a55f5f7e-plugins-dir\") pod \"csi-hostpathplugin-gl7lr\" (UID: \"9cb89712-c3bf-4ed9-87f8-8124a55f5f7e\") " pod="hostpath-provisioner/csi-hostpathplugin-gl7lr" Oct 08 18:13:17 crc kubenswrapper[4988]: I1008 18:13:17.411803 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/9cb89712-c3bf-4ed9-87f8-8124a55f5f7e-socket-dir\") pod \"csi-hostpathplugin-gl7lr\" (UID: \"9cb89712-c3bf-4ed9-87f8-8124a55f5f7e\") " pod="hostpath-provisioner/csi-hostpathplugin-gl7lr" Oct 08 18:13:17 crc kubenswrapper[4988]: I1008 18:13:17.411822 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/7f0973f4-25d3-48a6-ac02-b84f9bb4f947-metrics-tls\") pod \"dns-default-kwcts\" (UID: \"7f0973f4-25d3-48a6-ac02-b84f9bb4f947\") " pod="openshift-dns/dns-default-kwcts" Oct 08 18:13:17 crc kubenswrapper[4988]: I1008 18:13:17.411925 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/9cb89712-c3bf-4ed9-87f8-8124a55f5f7e-registration-dir\") pod \"csi-hostpathplugin-gl7lr\" (UID: \"9cb89712-c3bf-4ed9-87f8-8124a55f5f7e\") " pod="hostpath-provisioner/csi-hostpathplugin-gl7lr" Oct 08 18:13:17 crc kubenswrapper[4988]: I1008 18:13:17.412002 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/9cb89712-c3bf-4ed9-87f8-8124a55f5f7e-mountpoint-dir\") pod \"csi-hostpathplugin-gl7lr\" (UID: \"9cb89712-c3bf-4ed9-87f8-8124a55f5f7e\") " pod="hostpath-provisioner/csi-hostpathplugin-gl7lr" Oct 08 18:13:17 crc kubenswrapper[4988]: I1008 18:13:17.412020 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-znm4b\" (UniqueName: \"kubernetes.io/projected/7f0973f4-25d3-48a6-ac02-b84f9bb4f947-kube-api-access-znm4b\") pod \"dns-default-kwcts\" (UID: \"7f0973f4-25d3-48a6-ac02-b84f9bb4f947\") " pod="openshift-dns/dns-default-kwcts" Oct 08 18:13:17 crc kubenswrapper[4988]: I1008 18:13:17.412111 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/9cb89712-c3bf-4ed9-87f8-8124a55f5f7e-csi-data-dir\") pod \"csi-hostpathplugin-gl7lr\" (UID: \"9cb89712-c3bf-4ed9-87f8-8124a55f5f7e\") " pod="hostpath-provisioner/csi-hostpathplugin-gl7lr" Oct 08 18:13:17 crc kubenswrapper[4988]: I1008 18:13:17.412126 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cgjkq\" (UniqueName: \"kubernetes.io/projected/9cb89712-c3bf-4ed9-87f8-8124a55f5f7e-kube-api-access-cgjkq\") pod \"csi-hostpathplugin-gl7lr\" (UID: \"9cb89712-c3bf-4ed9-87f8-8124a55f5f7e\") " pod="hostpath-provisioner/csi-hostpathplugin-gl7lr" Oct 08 18:13:17 crc kubenswrapper[4988]: E1008 18:13:17.412356 4988 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 18:13:17.912339422 +0000 UTC m=+143.362182192 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 18:13:17 crc kubenswrapper[4988]: I1008 18:13:17.420105 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/9cb89712-c3bf-4ed9-87f8-8124a55f5f7e-socket-dir\") pod \"csi-hostpathplugin-gl7lr\" (UID: \"9cb89712-c3bf-4ed9-87f8-8124a55f5f7e\") " pod="hostpath-provisioner/csi-hostpathplugin-gl7lr" Oct 08 18:13:17 crc kubenswrapper[4988]: I1008 18:13:17.420860 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-46d7g\" (UniqueName: \"kubernetes.io/projected/47f38446-8f8d-4d84-8a16-f74e6f2836b7-kube-api-access-46d7g\") pod \"migrator-59844c95c7-g7fbv\" (UID: \"47f38446-8f8d-4d84-8a16-f74e6f2836b7\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-g7fbv" Oct 08 18:13:17 crc kubenswrapper[4988]: I1008 18:13:17.422468 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/7f0973f4-25d3-48a6-ac02-b84f9bb4f947-config-volume\") pod \"dns-default-kwcts\" (UID: \"7f0973f4-25d3-48a6-ac02-b84f9bb4f947\") " pod="openshift-dns/dns-default-kwcts" Oct 08 18:13:17 crc kubenswrapper[4988]: I1008 18:13:17.422990 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/9cb89712-c3bf-4ed9-87f8-8124a55f5f7e-mountpoint-dir\") pod \"csi-hostpathplugin-gl7lr\" (UID: \"9cb89712-c3bf-4ed9-87f8-8124a55f5f7e\") " pod="hostpath-provisioner/csi-hostpathplugin-gl7lr" Oct 08 18:13:17 crc kubenswrapper[4988]: I1008 18:13:17.423047 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/9cb89712-c3bf-4ed9-87f8-8124a55f5f7e-registration-dir\") pod \"csi-hostpathplugin-gl7lr\" (UID: \"9cb89712-c3bf-4ed9-87f8-8124a55f5f7e\") " pod="hostpath-provisioner/csi-hostpathplugin-gl7lr" Oct 08 18:13:17 crc kubenswrapper[4988]: I1008 18:13:17.423747 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/9cb89712-c3bf-4ed9-87f8-8124a55f5f7e-plugins-dir\") pod \"csi-hostpathplugin-gl7lr\" (UID: \"9cb89712-c3bf-4ed9-87f8-8124a55f5f7e\") " pod="hostpath-provisioner/csi-hostpathplugin-gl7lr" Oct 08 18:13:17 crc kubenswrapper[4988]: I1008 18:13:17.423841 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/9cb89712-c3bf-4ed9-87f8-8124a55f5f7e-csi-data-dir\") pod \"csi-hostpathplugin-gl7lr\" (UID: \"9cb89712-c3bf-4ed9-87f8-8124a55f5f7e\") " pod="hostpath-provisioner/csi-hostpathplugin-gl7lr" Oct 08 18:13:17 crc kubenswrapper[4988]: I1008 18:13:17.433832 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/7f0973f4-25d3-48a6-ac02-b84f9bb4f947-metrics-tls\") pod \"dns-default-kwcts\" (UID: \"7f0973f4-25d3-48a6-ac02-b84f9bb4f947\") " pod="openshift-dns/dns-default-kwcts" Oct 08 18:13:17 crc kubenswrapper[4988]: I1008 18:13:17.507619 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5hgmh\" (UniqueName: \"kubernetes.io/projected/e6f22dd3-65b7-4fe3-8265-8fb96095b6c4-kube-api-access-5hgmh\") pod \"dns-operator-744455d44c-2d794\" (UID: \"e6f22dd3-65b7-4fe3-8265-8fb96095b6c4\") " pod="openshift-dns-operator/dns-operator-744455d44c-2d794" Oct 08 18:13:17 crc kubenswrapper[4988]: I1008 18:13:17.524087 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-4wh9p\" (UID: \"8a33d7bd-26bd-409b-93f1-5f151ebca347\") " pod="openshift-image-registry/image-registry-697d97f7c8-4wh9p" Oct 08 18:13:17 crc kubenswrapper[4988]: E1008 18:13:17.524714 4988 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-08 18:13:18.024699919 +0000 UTC m=+143.474542689 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-4wh9p" (UID: "8a33d7bd-26bd-409b-93f1-5f151ebca347") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 18:13:17 crc kubenswrapper[4988]: I1008 18:13:17.527526 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-984jv" Oct 08 18:13:17 crc kubenswrapper[4988]: I1008 18:13:17.530916 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9gmqs\" (UniqueName: \"kubernetes.io/projected/7ee72f71-70e8-49ff-8407-96531bc67543-kube-api-access-9gmqs\") pod \"service-ca-9c57cc56f-wxv6d\" (UID: \"7ee72f71-70e8-49ff-8407-96531bc67543\") " pod="openshift-service-ca/service-ca-9c57cc56f-wxv6d" Oct 08 18:13:17 crc kubenswrapper[4988]: I1008 18:13:17.534071 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z9pjn\" (UniqueName: \"kubernetes.io/projected/05152e1e-3bc1-48b0-818d-41bc0d50fbf5-kube-api-access-z9pjn\") pod \"apiserver-76f77b778f-4phs4\" (UID: \"05152e1e-3bc1-48b0-818d-41bc0d50fbf5\") " pod="openshift-apiserver/apiserver-76f77b778f-4phs4" Oct 08 18:13:17 crc kubenswrapper[4988]: I1008 18:13:17.545851 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kq442\" (UniqueName: \"kubernetes.io/projected/ed79f1d0-1282-4be1-8e71-68a5320cdab1-kube-api-access-kq442\") pod \"router-default-5444994796-jjnkn\" (UID: \"ed79f1d0-1282-4be1-8e71-68a5320cdab1\") " pod="openshift-ingress/router-default-5444994796-jjnkn" Oct 08 18:13:17 crc kubenswrapper[4988]: I1008 18:13:17.546993 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bthq4\" (UniqueName: \"kubernetes.io/projected/ba2264e5-af3b-4400-9c20-04905b075dfa-kube-api-access-bthq4\") pod \"package-server-manager-789f6589d5-h5cw8\" (UID: \"ba2264e5-af3b-4400-9c20-04905b075dfa\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-h5cw8" Oct 08 18:13:17 crc kubenswrapper[4988]: I1008 18:13:17.564208 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-2d794" Oct 08 18:13:17 crc kubenswrapper[4988]: I1008 18:13:17.565499 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-572bk\" (UniqueName: \"kubernetes.io/projected/a760b68f-a1f6-4741-a9e9-51f6d3968748-kube-api-access-572bk\") pod \"machine-api-operator-5694c8668f-5mrrc\" (UID: \"a760b68f-a1f6-4741-a9e9-51f6d3968748\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-5mrrc" Oct 08 18:13:17 crc kubenswrapper[4988]: I1008 18:13:17.565649 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-wnkxn" Oct 08 18:13:17 crc kubenswrapper[4988]: I1008 18:13:17.567525 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-h5cw8" Oct 08 18:13:17 crc kubenswrapper[4988]: I1008 18:13:17.574820 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-g7fbv" Oct 08 18:13:17 crc kubenswrapper[4988]: I1008 18:13:17.579084 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6jts7\" (UniqueName: \"kubernetes.io/projected/c08e9af6-fc25-4c32-bd25-ac82aebb828e-kube-api-access-6jts7\") pod \"collect-profiles-29332440-qxt8r\" (UID: \"c08e9af6-fc25-4c32-bd25-ac82aebb828e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29332440-qxt8r" Oct 08 18:13:17 crc kubenswrapper[4988]: I1008 18:13:17.599179 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-hgnvq" Oct 08 18:13:17 crc kubenswrapper[4988]: I1008 18:13:17.604654 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kxfl7\" (UniqueName: \"kubernetes.io/projected/8a33d7bd-26bd-409b-93f1-5f151ebca347-kube-api-access-kxfl7\") pod \"image-registry-697d97f7c8-4wh9p\" (UID: \"8a33d7bd-26bd-409b-93f1-5f151ebca347\") " pod="openshift-image-registry/image-registry-697d97f7c8-4wh9p" Oct 08 18:13:17 crc kubenswrapper[4988]: I1008 18:13:17.605333 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-wxv6d" Oct 08 18:13:17 crc kubenswrapper[4988]: I1008 18:13:17.618562 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-4phs4" Oct 08 18:13:17 crc kubenswrapper[4988]: I1008 18:13:17.619999 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jp6mk\" (UniqueName: \"kubernetes.io/projected/9778cf39-b76d-4de6-b7ce-6cd9f3a4e224-kube-api-access-jp6mk\") pod \"service-ca-operator-777779d784-g775v\" (UID: \"9778cf39-b76d-4de6-b7ce-6cd9f3a4e224\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-g775v" Oct 08 18:13:17 crc kubenswrapper[4988]: I1008 18:13:17.625190 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 18:13:17 crc kubenswrapper[4988]: E1008 18:13:17.625790 4988 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 18:13:18.125774117 +0000 UTC m=+143.575616887 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 18:13:17 crc kubenswrapper[4988]: I1008 18:13:17.631634 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29332440-qxt8r" Oct 08 18:13:17 crc kubenswrapper[4988]: I1008 18:13:17.640325 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-s4r2k" Oct 08 18:13:17 crc kubenswrapper[4988]: I1008 18:13:17.651723 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-5hj6s" Oct 08 18:13:17 crc kubenswrapper[4988]: I1008 18:13:17.654705 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vgkbx\" (UniqueName: \"kubernetes.io/projected/4ddacc00-9f9b-44ec-8356-cd932ce26e5b-kube-api-access-vgkbx\") pod \"machine-config-operator-74547568cd-bbpzz\" (UID: \"4ddacc00-9f9b-44ec-8356-cd932ce26e5b\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-bbpzz" Oct 08 18:13:17 crc kubenswrapper[4988]: I1008 18:13:17.671330 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v68bb\" (UniqueName: \"kubernetes.io/projected/0a5e45fa-e12a-4199-aa81-96468c383594-kube-api-access-v68bb\") pod \"packageserver-d55dfcdfc-d9ks8\" (UID: \"0a5e45fa-e12a-4199-aa81-96468c383594\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-d9ks8" Oct 08 18:13:17 crc kubenswrapper[4988]: I1008 18:13:17.682293 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8a33d7bd-26bd-409b-93f1-5f151ebca347-bound-sa-token\") pod \"image-registry-697d97f7c8-4wh9p\" (UID: \"8a33d7bd-26bd-409b-93f1-5f151ebca347\") " pod="openshift-image-registry/image-registry-697d97f7c8-4wh9p" Oct 08 18:13:17 crc kubenswrapper[4988]: I1008 18:13:17.706661 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cgjkq\" (UniqueName: \"kubernetes.io/projected/9cb89712-c3bf-4ed9-87f8-8124a55f5f7e-kube-api-access-cgjkq\") pod \"csi-hostpathplugin-gl7lr\" (UID: \"9cb89712-c3bf-4ed9-87f8-8124a55f5f7e\") " pod="hostpath-provisioner/csi-hostpathplugin-gl7lr" Oct 08 18:13:17 crc kubenswrapper[4988]: I1008 18:13:17.727973 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-znm4b\" (UniqueName: \"kubernetes.io/projected/7f0973f4-25d3-48a6-ac02-b84f9bb4f947-kube-api-access-znm4b\") pod \"dns-default-kwcts\" (UID: \"7f0973f4-25d3-48a6-ac02-b84f9bb4f947\") " pod="openshift-dns/dns-default-kwcts" Oct 08 18:13:17 crc kubenswrapper[4988]: I1008 18:13:17.729745 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-4wh9p\" (UID: \"8a33d7bd-26bd-409b-93f1-5f151ebca347\") " pod="openshift-image-registry/image-registry-697d97f7c8-4wh9p" Oct 08 18:13:17 crc kubenswrapper[4988]: E1008 18:13:17.731488 4988 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-08 18:13:18.231464589 +0000 UTC m=+143.681307359 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-4wh9p" (UID: "8a33d7bd-26bd-409b-93f1-5f151ebca347") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 18:13:17 crc kubenswrapper[4988]: I1008 18:13:17.791366 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-jrt4j"] Oct 08 18:13:17 crc kubenswrapper[4988]: I1008 18:13:17.832257 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-l86nd"] Oct 08 18:13:17 crc kubenswrapper[4988]: I1008 18:13:17.832802 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-5mrrc" Oct 08 18:13:17 crc kubenswrapper[4988]: I1008 18:13:17.834151 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 18:13:17 crc kubenswrapper[4988]: I1008 18:13:17.837144 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-jjnkn" Oct 08 18:13:17 crc kubenswrapper[4988]: I1008 18:13:17.846201 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-bbpzz" Oct 08 18:13:17 crc kubenswrapper[4988]: E1008 18:13:17.843221 4988 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 18:13:18.343192027 +0000 UTC m=+143.793034797 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 18:13:17 crc kubenswrapper[4988]: I1008 18:13:17.880358 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-rgqjs"] Oct 08 18:13:17 crc kubenswrapper[4988]: I1008 18:13:17.886102 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-d9ks8" Oct 08 18:13:17 crc kubenswrapper[4988]: I1008 18:13:17.897131 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-m55d6"] Oct 08 18:13:17 crc kubenswrapper[4988]: I1008 18:13:17.916559 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-g775v" Oct 08 18:13:17 crc kubenswrapper[4988]: I1008 18:13:17.943091 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-4wh9p\" (UID: \"8a33d7bd-26bd-409b-93f1-5f151ebca347\") " pod="openshift-image-registry/image-registry-697d97f7c8-4wh9p" Oct 08 18:13:17 crc kubenswrapper[4988]: E1008 18:13:17.943502 4988 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-08 18:13:18.443486933 +0000 UTC m=+143.893329703 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-4wh9p" (UID: "8a33d7bd-26bd-409b-93f1-5f151ebca347") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 18:13:17 crc kubenswrapper[4988]: I1008 18:13:17.960814 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-kwcts" Oct 08 18:13:17 crc kubenswrapper[4988]: I1008 18:13:17.979564 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-gl7lr" Oct 08 18:13:18 crc kubenswrapper[4988]: W1008 18:13:18.044986 4988 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc83b25d6_a1c1_4394_9954_697156fe53b7.slice/crio-1b807e07e0797ce0760d1f577ed4b80a928d2df8e04c27e07d8633d18c6ff2b3 WatchSource:0}: Error finding container 1b807e07e0797ce0760d1f577ed4b80a928d2df8e04c27e07d8633d18c6ff2b3: Status 404 returned error can't find the container with id 1b807e07e0797ce0760d1f577ed4b80a928d2df8e04c27e07d8633d18c6ff2b3 Oct 08 18:13:18 crc kubenswrapper[4988]: I1008 18:13:18.046668 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 18:13:18 crc kubenswrapper[4988]: E1008 18:13:18.047091 4988 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 18:13:18.547066854 +0000 UTC m=+143.996909624 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 18:13:18 crc kubenswrapper[4988]: W1008 18:13:18.107125 4988 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1ca33297_d58c_4493_b1d3_45dd83983be4.slice/crio-127b0a033e01d22bd0bdec68a1c5aa9bb041f9d6f79eb1ddab0f9effb7c76bbf WatchSource:0}: Error finding container 127b0a033e01d22bd0bdec68a1c5aa9bb041f9d6f79eb1ddab0f9effb7c76bbf: Status 404 returned error can't find the container with id 127b0a033e01d22bd0bdec68a1c5aa9bb041f9d6f79eb1ddab0f9effb7c76bbf Oct 08 18:13:18 crc kubenswrapper[4988]: I1008 18:13:18.149275 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-4wh9p\" (UID: \"8a33d7bd-26bd-409b-93f1-5f151ebca347\") " pod="openshift-image-registry/image-registry-697d97f7c8-4wh9p" Oct 08 18:13:18 crc kubenswrapper[4988]: E1008 18:13:18.149688 4988 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-08 18:13:18.649675457 +0000 UTC m=+144.099518227 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-4wh9p" (UID: "8a33d7bd-26bd-409b-93f1-5f151ebca347") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 18:13:18 crc kubenswrapper[4988]: I1008 18:13:18.217190 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-85kp6" event={"ID":"ba81cb6f-e6fb-49e1-aa4c-5a9e6f8769fb","Type":"ContainerStarted","Data":"78c4e98a7072ccbc61150082159dcb378b8354b62ab803439e7a889c07757c73"} Oct 08 18:13:18 crc kubenswrapper[4988]: I1008 18:13:18.218866 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-gqm5m" event={"ID":"3021fdc0-7e51-4506-ac3c-09ac3a00c77f","Type":"ContainerStarted","Data":"1ebe357cf0462bbeee00a61d607b99c5bfba2eecbb8a21c9294a04c1e8ead14a"} Oct 08 18:13:18 crc kubenswrapper[4988]: I1008 18:13:18.222869 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-sndzf" event={"ID":"a818c760-46b8-4f91-89b3-18e4b0c9a649","Type":"ContainerStarted","Data":"4fa62de363968513075deaaa17c2182ce80008223020f5c529b510a21a17100a"} Oct 08 18:13:18 crc kubenswrapper[4988]: I1008 18:13:18.222936 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-sndzf" event={"ID":"a818c760-46b8-4f91-89b3-18e4b0c9a649","Type":"ContainerStarted","Data":"cbe9e4880bb77fc7c479bd686e48cd1f92910ef87dc500bbd134a928ad1c55fb"} Oct 08 18:13:18 crc kubenswrapper[4988]: I1008 18:13:18.223510 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-879f6c89f-sndzf" Oct 08 18:13:18 crc kubenswrapper[4988]: I1008 18:13:18.233610 4988 patch_prober.go:28] interesting pod/controller-manager-879f6c89f-sndzf container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.17:8443/healthz\": dial tcp 10.217.0.17:8443: connect: connection refused" start-of-body= Oct 08 18:13:18 crc kubenswrapper[4988]: I1008 18:13:18.233659 4988 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-879f6c89f-sndzf" podUID="a818c760-46b8-4f91-89b3-18e4b0c9a649" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.17:8443/healthz\": dial tcp 10.217.0.17:8443: connect: connection refused" Oct 08 18:13:18 crc kubenswrapper[4988]: I1008 18:13:18.253369 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-q2w8f"] Oct 08 18:13:18 crc kubenswrapper[4988]: I1008 18:13:18.253883 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 18:13:18 crc kubenswrapper[4988]: E1008 18:13:18.255439 4988 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 18:13:18.75541921 +0000 UTC m=+144.205261980 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 18:13:18 crc kubenswrapper[4988]: I1008 18:13:18.259226 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-7xdnc" event={"ID":"9c8874ed-3f70-45b6-9f10-d8737b6a7a8f","Type":"ContainerStarted","Data":"8decb88999dd5ec9d72091c408c71df63d4adcf20acd387e7c499f07211eb8fc"} Oct 08 18:13:18 crc kubenswrapper[4988]: I1008 18:13:18.259263 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-7xdnc" event={"ID":"9c8874ed-3f70-45b6-9f10-d8737b6a7a8f","Type":"ContainerStarted","Data":"04d2296a013043a794ce84eebbf03997f04316d51b2603d40464866e8c4e87cd"} Oct 08 18:13:18 crc kubenswrapper[4988]: I1008 18:13:18.312365 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29332440-qxt8r"] Oct 08 18:13:18 crc kubenswrapper[4988]: I1008 18:13:18.316816 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-5hj6s" event={"ID":"870431bf-3dfc-4ac5-a68a-39de04683899","Type":"ContainerStarted","Data":"d8722eb229b9153a0a755d52169b0133b224c2235810a3bd2539ced51585b277"} Oct 08 18:13:18 crc kubenswrapper[4988]: I1008 18:13:18.326220 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-q2mxf"] Oct 08 18:13:18 crc kubenswrapper[4988]: I1008 18:13:18.355453 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-294dt"] Oct 08 18:13:18 crc kubenswrapper[4988]: I1008 18:13:18.357991 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-4wh9p\" (UID: \"8a33d7bd-26bd-409b-93f1-5f151ebca347\") " pod="openshift-image-registry/image-registry-697d97f7c8-4wh9p" Oct 08 18:13:18 crc kubenswrapper[4988]: E1008 18:13:18.360082 4988 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-08 18:13:18.860064823 +0000 UTC m=+144.309907593 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-4wh9p" (UID: "8a33d7bd-26bd-409b-93f1-5f151ebca347") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 18:13:18 crc kubenswrapper[4988]: I1008 18:13:18.364137 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-k4ghh"] Oct 08 18:13:18 crc kubenswrapper[4988]: I1008 18:13:18.373878 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-m55d6" event={"ID":"b5cd6079-031c-4cc8-8161-88f0e6f84064","Type":"ContainerStarted","Data":"0b7043cdf0cce43b92949b1d5070528af36293d1357f3923c0e46363019b91d8"} Oct 08 18:13:18 crc kubenswrapper[4988]: I1008 18:13:18.376243 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-rgqjs" event={"ID":"1ca33297-d58c-4493-b1d3-45dd83983be4","Type":"ContainerStarted","Data":"127b0a033e01d22bd0bdec68a1c5aa9bb041f9d6f79eb1ddab0f9effb7c76bbf"} Oct 08 18:13:18 crc kubenswrapper[4988]: I1008 18:13:18.377839 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-l86nd" event={"ID":"c83b25d6-a1c1-4394-9954-697156fe53b7","Type":"ContainerStarted","Data":"1b807e07e0797ce0760d1f577ed4b80a928d2df8e04c27e07d8633d18c6ff2b3"} Oct 08 18:13:18 crc kubenswrapper[4988]: I1008 18:13:18.391992 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-jrt4j" event={"ID":"d0df68ec-1481-4992-bc83-16b4a6b85f0b","Type":"ContainerStarted","Data":"a8292cf4938100272f7ff51eb29f60604dad2533d772d9fab0c061c1c6863ff5"} Oct 08 18:13:18 crc kubenswrapper[4988]: I1008 18:13:18.396709 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-qmwln" event={"ID":"6c651396-5e8f-4920-a2b9-75f2c40eeea8","Type":"ContainerStarted","Data":"1ee19f700443e58203d0ff8652353162d562ee085bdd7bc1a92ba2db5c489719"} Oct 08 18:13:18 crc kubenswrapper[4988]: I1008 18:13:18.396773 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-qmwln" event={"ID":"6c651396-5e8f-4920-a2b9-75f2c40eeea8","Type":"ContainerStarted","Data":"7bdb55b1bfd49c829a43246d33157bcfa1844c6e9a1b349af9690dc11d0473f0"} Oct 08 18:13:18 crc kubenswrapper[4988]: I1008 18:13:18.408958 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-2d794"] Oct 08 18:13:18 crc kubenswrapper[4988]: I1008 18:13:18.415309 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-gd7f9" event={"ID":"5df90a93-b572-4439-b3aa-2666b2873c1d","Type":"ContainerStarted","Data":"e59fc598f1c05950932c9d2b3dc8cd319c7d660adde066d68b723ab1df6201a7"} Oct 08 18:13:18 crc kubenswrapper[4988]: I1008 18:13:18.415445 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-984jv"] Oct 08 18:13:18 crc kubenswrapper[4988]: I1008 18:13:18.415581 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-gd7f9" Oct 08 18:13:18 crc kubenswrapper[4988]: I1008 18:13:18.416484 4988 patch_prober.go:28] interesting pod/route-controller-manager-6576b87f9c-gd7f9 container/route-controller-manager namespace/openshift-route-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.21:8443/healthz\": dial tcp 10.217.0.21:8443: connect: connection refused" start-of-body= Oct 08 18:13:18 crc kubenswrapper[4988]: I1008 18:13:18.416540 4988 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-gd7f9" podUID="5df90a93-b572-4439-b3aa-2666b2873c1d" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.21:8443/healthz\": dial tcp 10.217.0.21:8443: connect: connection refused" Oct 08 18:13:18 crc kubenswrapper[4988]: I1008 18:13:18.458649 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 18:13:18 crc kubenswrapper[4988]: E1008 18:13:18.458843 4988 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 18:13:18.958794942 +0000 UTC m=+144.408637712 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 18:13:18 crc kubenswrapper[4988]: I1008 18:13:18.459002 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-4wh9p\" (UID: \"8a33d7bd-26bd-409b-93f1-5f151ebca347\") " pod="openshift-image-registry/image-registry-697d97f7c8-4wh9p" Oct 08 18:13:18 crc kubenswrapper[4988]: I1008 18:13:18.482640 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-g7fbv"] Oct 08 18:13:18 crc kubenswrapper[4988]: E1008 18:13:18.499136 4988 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-08 18:13:18.977492787 +0000 UTC m=+144.427335567 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-4wh9p" (UID: "8a33d7bd-26bd-409b-93f1-5f151ebca347") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 18:13:18 crc kubenswrapper[4988]: I1008 18:13:18.609994 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 18:13:18 crc kubenswrapper[4988]: E1008 18:13:18.610619 4988 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 18:13:19.110577698 +0000 UTC m=+144.560420458 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 18:13:18 crc kubenswrapper[4988]: I1008 18:13:18.610904 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-4wh9p\" (UID: \"8a33d7bd-26bd-409b-93f1-5f151ebca347\") " pod="openshift-image-registry/image-registry-697d97f7c8-4wh9p" Oct 08 18:13:18 crc kubenswrapper[4988]: E1008 18:13:18.611635 4988 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-08 18:13:19.111627019 +0000 UTC m=+144.561469779 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-4wh9p" (UID: "8a33d7bd-26bd-409b-93f1-5f151ebca347") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 18:13:18 crc kubenswrapper[4988]: I1008 18:13:18.713464 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 18:13:18 crc kubenswrapper[4988]: E1008 18:13:18.714117 4988 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 18:13:19.214089267 +0000 UTC m=+144.663932037 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 18:13:18 crc kubenswrapper[4988]: I1008 18:13:18.815259 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-4wh9p\" (UID: \"8a33d7bd-26bd-409b-93f1-5f151ebca347\") " pod="openshift-image-registry/image-registry-697d97f7c8-4wh9p" Oct 08 18:13:18 crc kubenswrapper[4988]: E1008 18:13:18.815682 4988 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-08 18:13:19.31567079 +0000 UTC m=+144.765513560 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-4wh9p" (UID: "8a33d7bd-26bd-409b-93f1-5f151ebca347") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 18:13:18 crc kubenswrapper[4988]: I1008 18:13:18.916141 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 18:13:18 crc kubenswrapper[4988]: E1008 18:13:18.916801 4988 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 18:13:19.416777859 +0000 UTC m=+144.866620619 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 18:13:18 crc kubenswrapper[4988]: I1008 18:13:18.952169 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-s4r2k"] Oct 08 18:13:18 crc kubenswrapper[4988]: I1008 18:13:18.964854 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console-operator/console-operator-58897d9998-rt48k" podStartSLOduration=122.96482215 podStartE2EDuration="2m2.96482215s" podCreationTimestamp="2025-10-08 18:11:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 18:13:18.962228555 +0000 UTC m=+144.412071345" watchObservedRunningTime="2025-10-08 18:13:18.96482215 +0000 UTC m=+144.414664940" Oct 08 18:13:19 crc kubenswrapper[4988]: I1008 18:13:19.032034 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-4wh9p\" (UID: \"8a33d7bd-26bd-409b-93f1-5f151ebca347\") " pod="openshift-image-registry/image-registry-697d97f7c8-4wh9p" Oct 08 18:13:19 crc kubenswrapper[4988]: E1008 18:13:19.032613 4988 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-08 18:13:19.532597587 +0000 UTC m=+144.982440357 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-4wh9p" (UID: "8a33d7bd-26bd-409b-93f1-5f151ebca347") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 18:13:19 crc kubenswrapper[4988]: I1008 18:13:19.064065 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-558db77b4-hwtzl" podStartSLOduration=123.064046944 podStartE2EDuration="2m3.064046944s" podCreationTimestamp="2025-10-08 18:11:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 18:13:19.062127848 +0000 UTC m=+144.511970618" watchObservedRunningTime="2025-10-08 18:13:19.064046944 +0000 UTC m=+144.513889714" Oct 08 18:13:19 crc kubenswrapper[4988]: I1008 18:13:19.086785 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-h5cw8"] Oct 08 18:13:19 crc kubenswrapper[4988]: I1008 18:13:19.088098 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-d9ks8"] Oct 08 18:13:19 crc kubenswrapper[4988]: I1008 18:13:19.103536 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-jt7fv" podStartSLOduration=123.103507094 podStartE2EDuration="2m3.103507094s" podCreationTimestamp="2025-10-08 18:11:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 18:13:19.096639035 +0000 UTC m=+144.546481805" watchObservedRunningTime="2025-10-08 18:13:19.103507094 +0000 UTC m=+144.553349864" Oct 08 18:13:19 crc kubenswrapper[4988]: I1008 18:13:19.107182 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-wnkxn"] Oct 08 18:13:19 crc kubenswrapper[4988]: I1008 18:13:19.134398 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 18:13:19 crc kubenswrapper[4988]: E1008 18:13:19.134846 4988 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 18:13:19.634828568 +0000 UTC m=+145.084671338 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 18:13:19 crc kubenswrapper[4988]: W1008 18:13:19.159048 4988 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0a5e45fa_e12a_4199_aa81_96468c383594.slice/crio-2c38e8cfd3c4de9374271aff360b5f9c6175df3a2c985d1f62a9e5fc34fbadcc WatchSource:0}: Error finding container 2c38e8cfd3c4de9374271aff360b5f9c6175df3a2c985d1f62a9e5fc34fbadcc: Status 404 returned error can't find the container with id 2c38e8cfd3c4de9374271aff360b5f9c6175df3a2c985d1f62a9e5fc34fbadcc Oct 08 18:13:19 crc kubenswrapper[4988]: I1008 18:13:19.164018 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-gd7f9" podStartSLOduration=123.163996359 podStartE2EDuration="2m3.163996359s" podCreationTimestamp="2025-10-08 18:11:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 18:13:19.16163307 +0000 UTC m=+144.611475840" watchObservedRunningTime="2025-10-08 18:13:19.163996359 +0000 UTC m=+144.613839129" Oct 08 18:13:19 crc kubenswrapper[4988]: I1008 18:13:19.237086 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-4wh9p\" (UID: \"8a33d7bd-26bd-409b-93f1-5f151ebca347\") " pod="openshift-image-registry/image-registry-697d97f7c8-4wh9p" Oct 08 18:13:19 crc kubenswrapper[4988]: E1008 18:13:19.238058 4988 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-08 18:13:19.738024888 +0000 UTC m=+145.187867658 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-4wh9p" (UID: "8a33d7bd-26bd-409b-93f1-5f151ebca347") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 18:13:19 crc kubenswrapper[4988]: I1008 18:13:19.336178 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication-operator/authentication-operator-69f744f599-zjfnq" podStartSLOduration=123.336144169 podStartE2EDuration="2m3.336144169s" podCreationTimestamp="2025-10-08 18:11:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 18:13:19.322952135 +0000 UTC m=+144.772794905" watchObservedRunningTime="2025-10-08 18:13:19.336144169 +0000 UTC m=+144.785986939" Oct 08 18:13:19 crc kubenswrapper[4988]: I1008 18:13:19.341140 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-4phs4"] Oct 08 18:13:19 crc kubenswrapper[4988]: I1008 18:13:19.341969 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 18:13:19 crc kubenswrapper[4988]: E1008 18:13:19.342436 4988 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 18:13:19.842411762 +0000 UTC m=+145.292254532 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 18:13:19 crc kubenswrapper[4988]: I1008 18:13:19.358466 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-xttx4" podStartSLOduration=123.35844276 podStartE2EDuration="2m3.35844276s" podCreationTimestamp="2025-10-08 18:11:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 18:13:19.35569999 +0000 UTC m=+144.805542770" watchObservedRunningTime="2025-10-08 18:13:19.35844276 +0000 UTC m=+144.808285550" Oct 08 18:13:19 crc kubenswrapper[4988]: I1008 18:13:19.422952 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-wxv6d"] Oct 08 18:13:19 crc kubenswrapper[4988]: I1008 18:13:19.444356 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-4wh9p\" (UID: \"8a33d7bd-26bd-409b-93f1-5f151ebca347\") " pod="openshift-image-registry/image-registry-697d97f7c8-4wh9p" Oct 08 18:13:19 crc kubenswrapper[4988]: E1008 18:13:19.444706 4988 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-08 18:13:19.944692795 +0000 UTC m=+145.394535565 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-4wh9p" (UID: "8a33d7bd-26bd-409b-93f1-5f151ebca347") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 18:13:19 crc kubenswrapper[4988]: W1008 18:13:19.451257 4988 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod05152e1e_3bc1_48b0_818d_41bc0d50fbf5.slice/crio-f9808f90b1da24ea3faa88a547674e28633679371159124db1e661aa25cbe17f WatchSource:0}: Error finding container f9808f90b1da24ea3faa88a547674e28633679371159124db1e661aa25cbe17f: Status 404 returned error can't find the container with id f9808f90b1da24ea3faa88a547674e28633679371159124db1e661aa25cbe17f Oct 08 18:13:19 crc kubenswrapper[4988]: I1008 18:13:19.452742 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-2d794" event={"ID":"e6f22dd3-65b7-4fe3-8265-8fb96095b6c4","Type":"ContainerStarted","Data":"fc0e159d767b0cf972ca4ac5312647887ef1e34a3149585ccf3b67f49262de37"} Oct 08 18:13:19 crc kubenswrapper[4988]: I1008 18:13:19.455722 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-g7fbv" event={"ID":"47f38446-8f8d-4d84-8a16-f74e6f2836b7","Type":"ContainerStarted","Data":"7b35e7ab472e1b7ace59c90563ced940e0c59d4b92cadba23b262d6777942da8"} Oct 08 18:13:19 crc kubenswrapper[4988]: I1008 18:13:19.468735 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-hgnvq"] Oct 08 18:13:19 crc kubenswrapper[4988]: I1008 18:13:19.475929 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-kwcts"] Oct 08 18:13:19 crc kubenswrapper[4988]: I1008 18:13:19.490497 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-q2mxf" event={"ID":"aa86616c-479c-45be-aaa3-d3c4994721bf","Type":"ContainerStarted","Data":"8edf6d1ad9f1bd5f15d5b353551767b8108afd620ca48c577c97c397c0d6390f"} Oct 08 18:13:19 crc kubenswrapper[4988]: I1008 18:13:19.492043 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-zf8t5" podStartSLOduration=123.492018586 podStartE2EDuration="2m3.492018586s" podCreationTimestamp="2025-10-08 18:11:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 18:13:19.490838051 +0000 UTC m=+144.940680821" watchObservedRunningTime="2025-10-08 18:13:19.492018586 +0000 UTC m=+144.941861356" Oct 08 18:13:19 crc kubenswrapper[4988]: I1008 18:13:19.513744 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-294dt" event={"ID":"8bf9175c-5101-4fcd-8cba-e113be11dde4","Type":"ContainerStarted","Data":"20dfd87103b1c34d1d77781ce5e72b28df28ecc2efa06881c31118ab87686314"} Oct 08 18:13:19 crc kubenswrapper[4988]: I1008 18:13:19.513804 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-294dt" event={"ID":"8bf9175c-5101-4fcd-8cba-e113be11dde4","Type":"ContainerStarted","Data":"9ca6901123f580bf717fd66335996ba54cbe84d4a491dea9de46c48bb519c2eb"} Oct 08 18:13:19 crc kubenswrapper[4988]: I1008 18:13:19.515068 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-294dt" Oct 08 18:13:19 crc kubenswrapper[4988]: I1008 18:13:19.529466 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-bbpzz"] Oct 08 18:13:19 crc kubenswrapper[4988]: I1008 18:13:19.532871 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-g775v"] Oct 08 18:13:19 crc kubenswrapper[4988]: I1008 18:13:19.538803 4988 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-294dt container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.29:8080/healthz\": dial tcp 10.217.0.29:8080: connect: connection refused" start-of-body= Oct 08 18:13:19 crc kubenswrapper[4988]: I1008 18:13:19.538885 4988 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-294dt" podUID="8bf9175c-5101-4fcd-8cba-e113be11dde4" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.29:8080/healthz\": dial tcp 10.217.0.29:8080: connect: connection refused" Oct 08 18:13:19 crc kubenswrapper[4988]: I1008 18:13:19.544241 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-h5cw8" event={"ID":"ba2264e5-af3b-4400-9c20-04905b075dfa","Type":"ContainerStarted","Data":"dfd8fb6cff8953b1aff158f4317bdf288358908cdd0d92c30f0013d9fda15075"} Oct 08 18:13:19 crc kubenswrapper[4988]: I1008 18:13:19.545238 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 18:13:19 crc kubenswrapper[4988]: E1008 18:13:19.545777 4988 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 18:13:20.045754832 +0000 UTC m=+145.495597602 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 18:13:19 crc kubenswrapper[4988]: I1008 18:13:19.549836 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-gqm5m" event={"ID":"3021fdc0-7e51-4506-ac3c-09ac3a00c77f","Type":"ContainerStarted","Data":"bf6418e10ee0d253c4bd37c09ceed2ef658b5762105e467ca58bae81a3e9c0fa"} Oct 08 18:13:19 crc kubenswrapper[4988]: I1008 18:13:19.580258 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-wnkxn" event={"ID":"176096f3-5341-4aa9-a706-5e96cfd8eca1","Type":"ContainerStarted","Data":"c6aa3e1662989ef830fd9a716b43ac8ec4407d61467a3136be907bc5ab0e6349"} Oct 08 18:13:19 crc kubenswrapper[4988]: I1008 18:13:19.635933 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-jjnkn" event={"ID":"ed79f1d0-1282-4be1-8e71-68a5320cdab1","Type":"ContainerStarted","Data":"9ee2253db1d342bbf9c4ca3cb4536b6402426778bd2b0d15636cc42d02c9bd5b"} Oct 08 18:13:19 crc kubenswrapper[4988]: I1008 18:13:19.635990 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-jjnkn" event={"ID":"ed79f1d0-1282-4be1-8e71-68a5320cdab1","Type":"ContainerStarted","Data":"252e144936f8a015f75ea05cd5c8d047399394549e9fe3619a0723edad773a01"} Oct 08 18:13:19 crc kubenswrapper[4988]: I1008 18:13:19.648269 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-4wh9p\" (UID: \"8a33d7bd-26bd-409b-93f1-5f151ebca347\") " pod="openshift-image-registry/image-registry-697d97f7c8-4wh9p" Oct 08 18:13:19 crc kubenswrapper[4988]: I1008 18:13:19.653378 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-gl7lr"] Oct 08 18:13:19 crc kubenswrapper[4988]: I1008 18:13:19.657696 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-5hj6s" event={"ID":"870431bf-3dfc-4ac5-a68a-39de04683899","Type":"ContainerStarted","Data":"dcb9d344c7421762a0287e8095b7ba5c959d7f1f9474c891ad9db48ea0df8add"} Oct 08 18:13:19 crc kubenswrapper[4988]: E1008 18:13:19.658224 4988 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-08 18:13:20.158181281 +0000 UTC m=+145.608024261 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-4wh9p" (UID: "8a33d7bd-26bd-409b-93f1-5f151ebca347") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 18:13:19 crc kubenswrapper[4988]: I1008 18:13:19.734644 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-5mrrc"] Oct 08 18:13:19 crc kubenswrapper[4988]: I1008 18:13:19.734658 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/downloads-7954f5f757-qcgmb" podStartSLOduration=123.734627961 podStartE2EDuration="2m3.734627961s" podCreationTimestamp="2025-10-08 18:11:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 18:13:19.725291219 +0000 UTC m=+145.175133989" watchObservedRunningTime="2025-10-08 18:13:19.734627961 +0000 UTC m=+145.184470731" Oct 08 18:13:19 crc kubenswrapper[4988]: I1008 18:13:19.751191 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 18:13:19 crc kubenswrapper[4988]: I1008 18:13:19.755687 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-f9d7485db-7xdnc" podStartSLOduration=123.755660554 podStartE2EDuration="2m3.755660554s" podCreationTimestamp="2025-10-08 18:11:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 18:13:19.754150001 +0000 UTC m=+145.203992791" watchObservedRunningTime="2025-10-08 18:13:19.755660554 +0000 UTC m=+145.205503324" Oct 08 18:13:19 crc kubenswrapper[4988]: E1008 18:13:19.759284 4988 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 18:13:20.259252059 +0000 UTC m=+145.709094829 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 18:13:19 crc kubenswrapper[4988]: I1008 18:13:19.759490 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-4wh9p\" (UID: \"8a33d7bd-26bd-409b-93f1-5f151ebca347\") " pod="openshift-image-registry/image-registry-697d97f7c8-4wh9p" Oct 08 18:13:19 crc kubenswrapper[4988]: E1008 18:13:19.761378 4988 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-08 18:13:20.26135605 +0000 UTC m=+145.711198820 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-4wh9p" (UID: "8a33d7bd-26bd-409b-93f1-5f151ebca347") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 18:13:19 crc kubenswrapper[4988]: I1008 18:13:19.771668 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-jrt4j" event={"ID":"d0df68ec-1481-4992-bc83-16b4a6b85f0b","Type":"ContainerStarted","Data":"0c53d49c930ea8910af27301e206a241909d3e691c72220be9e2f10978078f4e"} Oct 08 18:13:19 crc kubenswrapper[4988]: I1008 18:13:19.772326 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-jrt4j" Oct 08 18:13:19 crc kubenswrapper[4988]: I1008 18:13:19.791022 4988 patch_prober.go:28] interesting pod/catalog-operator-68c6474976-jrt4j container/catalog-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.34:8443/healthz\": dial tcp 10.217.0.34:8443: connect: connection refused" start-of-body= Oct 08 18:13:19 crc kubenswrapper[4988]: I1008 18:13:19.791336 4988 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-jrt4j" podUID="d0df68ec-1481-4992-bc83-16b4a6b85f0b" containerName="catalog-operator" probeResult="failure" output="Get \"https://10.217.0.34:8443/healthz\": dial tcp 10.217.0.34:8443: connect: connection refused" Oct 08 18:13:19 crc kubenswrapper[4988]: I1008 18:13:19.822976 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-s4r2k" event={"ID":"db1043d1-7232-4fcd-9ce5-50b590798d36","Type":"ContainerStarted","Data":"d53b33d7fa85ad5eae386d29f73fc5e4897f8927b988abb9cd03d52b1efc5fec"} Oct 08 18:13:19 crc kubenswrapper[4988]: I1008 18:13:19.834598 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-qmwln" event={"ID":"6c651396-5e8f-4920-a2b9-75f2c40eeea8","Type":"ContainerStarted","Data":"f08058697861402c881c2a6a168c06529cae11f073a73bbb212fcfbf1b193515"} Oct 08 18:13:19 crc kubenswrapper[4988]: I1008 18:13:19.839430 4988 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-ingress/router-default-5444994796-jjnkn" Oct 08 18:13:19 crc kubenswrapper[4988]: I1008 18:13:19.839675 4988 patch_prober.go:28] interesting pod/router-default-5444994796-jjnkn container/router namespace/openshift-ingress: Startup probe status=failure output="Get \"http://localhost:1936/healthz/ready\": dial tcp [::1]:1936: connect: connection refused" start-of-body= Oct 08 18:13:19 crc kubenswrapper[4988]: I1008 18:13:19.839744 4988 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-jjnkn" podUID="ed79f1d0-1282-4be1-8e71-68a5320cdab1" containerName="router" probeResult="failure" output="Get \"http://localhost:1936/healthz/ready\": dial tcp [::1]:1936: connect: connection refused" Oct 08 18:13:19 crc kubenswrapper[4988]: I1008 18:13:19.841145 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-k4ghh" event={"ID":"4719309b-a3b7-4910-a389-635acc09d01d","Type":"ContainerStarted","Data":"bf6f9f554ddf535a58a491fc3843b55c801377b1e096c62a7cae76ced0d2a6d5"} Oct 08 18:13:19 crc kubenswrapper[4988]: I1008 18:13:19.845144 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-config-operator/openshift-config-operator-7777fb866f-7ctzs" podStartSLOduration=123.845114953 podStartE2EDuration="2m3.845114953s" podCreationTimestamp="2025-10-08 18:11:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 18:13:19.843574249 +0000 UTC m=+145.293417019" watchObservedRunningTime="2025-10-08 18:13:19.845114953 +0000 UTC m=+145.294957713" Oct 08 18:13:19 crc kubenswrapper[4988]: I1008 18:13:19.863813 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 18:13:19 crc kubenswrapper[4988]: E1008 18:13:19.864289 4988 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 18:13:20.364264862 +0000 UTC m=+145.814107632 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 18:13:19 crc kubenswrapper[4988]: I1008 18:13:19.870201 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-q2w8f" event={"ID":"0769dc07-8f56-42a5-8eb5-b900eefaaf1e","Type":"ContainerStarted","Data":"51d8c191dc6ed25d3ce50d36533d1c649ab4d4f8ceb94ac1f574c0fe0159ffd5"} Oct 08 18:13:19 crc kubenswrapper[4988]: I1008 18:13:19.883771 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-6k9kl" podStartSLOduration=123.883752611 podStartE2EDuration="2m3.883752611s" podCreationTimestamp="2025-10-08 18:11:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 18:13:19.88136983 +0000 UTC m=+145.331212590" watchObservedRunningTime="2025-10-08 18:13:19.883752611 +0000 UTC m=+145.333595381" Oct 08 18:13:19 crc kubenswrapper[4988]: I1008 18:13:19.893229 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-4wh9p\" (UID: \"8a33d7bd-26bd-409b-93f1-5f151ebca347\") " pod="openshift-image-registry/image-registry-697d97f7c8-4wh9p" Oct 08 18:13:19 crc kubenswrapper[4988]: E1008 18:13:19.893759 4988 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-08 18:13:20.393744352 +0000 UTC m=+145.843587112 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-4wh9p" (UID: "8a33d7bd-26bd-409b-93f1-5f151ebca347") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 18:13:19 crc kubenswrapper[4988]: I1008 18:13:19.897353 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-d9ks8" event={"ID":"0a5e45fa-e12a-4199-aa81-96468c383594","Type":"ContainerStarted","Data":"2c38e8cfd3c4de9374271aff360b5f9c6175df3a2c985d1f62a9e5fc34fbadcc"} Oct 08 18:13:19 crc kubenswrapper[4988]: I1008 18:13:19.898018 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-d9ks8" Oct 08 18:13:19 crc kubenswrapper[4988]: I1008 18:13:19.909418 4988 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-d9ks8 container/packageserver namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.38:5443/healthz\": dial tcp 10.217.0.38:5443: connect: connection refused" start-of-body= Oct 08 18:13:19 crc kubenswrapper[4988]: I1008 18:13:19.909477 4988 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-d9ks8" podUID="0a5e45fa-e12a-4199-aa81-96468c383594" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.38:5443/healthz\": dial tcp 10.217.0.38:5443: connect: connection refused" Oct 08 18:13:19 crc kubenswrapper[4988]: I1008 18:13:19.919869 4988 generic.go:334] "Generic (PLEG): container finished" podID="ba81cb6f-e6fb-49e1-aa4c-5a9e6f8769fb" containerID="cba96854838de415a8a4dc92f03c243d94e2d2ff1b6850d8d825b783f9dd71d9" exitCode=0 Oct 08 18:13:19 crc kubenswrapper[4988]: I1008 18:13:19.920020 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-85kp6" event={"ID":"ba81cb6f-e6fb-49e1-aa4c-5a9e6f8769fb","Type":"ContainerDied","Data":"cba96854838de415a8a4dc92f03c243d94e2d2ff1b6850d8d825b783f9dd71d9"} Oct 08 18:13:19 crc kubenswrapper[4988]: I1008 18:13:19.960876 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29332440-qxt8r" event={"ID":"c08e9af6-fc25-4c32-bd25-ac82aebb828e","Type":"ContainerStarted","Data":"313c57df12ab0a4da3188d7fc8368593246ef417dbcc4914e60937f26494b1b0"} Oct 08 18:13:19 crc kubenswrapper[4988]: I1008 18:13:19.961056 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29332440-qxt8r" event={"ID":"c08e9af6-fc25-4c32-bd25-ac82aebb828e","Type":"ContainerStarted","Data":"e00740e7d2a3e5e6adb6acba4611aa1779dc5931cb73f118f64d0e3152caad9c"} Oct 08 18:13:19 crc kubenswrapper[4988]: I1008 18:13:19.975821 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-gd7f9" Oct 08 18:13:19 crc kubenswrapper[4988]: I1008 18:13:19.979086 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-879f6c89f-sndzf" podStartSLOduration=123.97905532 podStartE2EDuration="2m3.97905532s" podCreationTimestamp="2025-10-08 18:11:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 18:13:19.924285092 +0000 UTC m=+145.374127862" watchObservedRunningTime="2025-10-08 18:13:19.97905532 +0000 UTC m=+145.428898090" Oct 08 18:13:19 crc kubenswrapper[4988]: I1008 18:13:19.985889 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-879f6c89f-sndzf" Oct 08 18:13:19 crc kubenswrapper[4988]: I1008 18:13:19.980437 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-server-5hj6s" podStartSLOduration=5.98043182 podStartE2EDuration="5.98043182s" podCreationTimestamp="2025-10-08 18:13:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 18:13:19.971334374 +0000 UTC m=+145.421177144" watchObservedRunningTime="2025-10-08 18:13:19.98043182 +0000 UTC m=+145.430274590" Oct 08 18:13:19 crc kubenswrapper[4988]: I1008 18:13:19.993947 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 18:13:19 crc kubenswrapper[4988]: E1008 18:13:19.994534 4988 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 18:13:20.49450452 +0000 UTC m=+145.944347460 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 18:13:19 crc kubenswrapper[4988]: I1008 18:13:19.995064 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-4wh9p\" (UID: \"8a33d7bd-26bd-409b-93f1-5f151ebca347\") " pod="openshift-image-registry/image-registry-697d97f7c8-4wh9p" Oct 08 18:13:19 crc kubenswrapper[4988]: E1008 18:13:19.995581 4988 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-08 18:13:20.495572672 +0000 UTC m=+145.945415442 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-4wh9p" (UID: "8a33d7bd-26bd-409b-93f1-5f151ebca347") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 18:13:20 crc kubenswrapper[4988]: I1008 18:13:20.001733 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress/router-default-5444994796-jjnkn" podStartSLOduration=124.00171023 podStartE2EDuration="2m4.00171023s" podCreationTimestamp="2025-10-08 18:11:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 18:13:20.000957159 +0000 UTC m=+145.450799929" watchObservedRunningTime="2025-10-08 18:13:20.00171023 +0000 UTC m=+145.451553000" Oct 08 18:13:20 crc kubenswrapper[4988]: I1008 18:13:20.078430 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-l86nd" podStartSLOduration=124.078402057 podStartE2EDuration="2m4.078402057s" podCreationTimestamp="2025-10-08 18:11:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 18:13:20.076032678 +0000 UTC m=+145.525875458" watchObservedRunningTime="2025-10-08 18:13:20.078402057 +0000 UTC m=+145.528244827" Oct 08 18:13:20 crc kubenswrapper[4988]: I1008 18:13:20.098420 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 18:13:20 crc kubenswrapper[4988]: E1008 18:13:20.099666 4988 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 18:13:20.599646636 +0000 UTC m=+146.049489406 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 18:13:20 crc kubenswrapper[4988]: I1008 18:13:20.106672 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-d9ks8" podStartSLOduration=124.106653701 podStartE2EDuration="2m4.106653701s" podCreationTimestamp="2025-10-08 18:11:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 18:13:20.105701553 +0000 UTC m=+145.555544323" watchObservedRunningTime="2025-10-08 18:13:20.106653701 +0000 UTC m=+145.556496461" Oct 08 18:13:20 crc kubenswrapper[4988]: I1008 18:13:20.202404 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-4wh9p\" (UID: \"8a33d7bd-26bd-409b-93f1-5f151ebca347\") " pod="openshift-image-registry/image-registry-697d97f7c8-4wh9p" Oct 08 18:13:20 crc kubenswrapper[4988]: E1008 18:13:20.203064 4988 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-08 18:13:20.703049923 +0000 UTC m=+146.152892693 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-4wh9p" (UID: "8a33d7bd-26bd-409b-93f1-5f151ebca347") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 18:13:20 crc kubenswrapper[4988]: I1008 18:13:20.232483 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-k4ghh" podStartSLOduration=124.23245765 podStartE2EDuration="2m4.23245765s" podCreationTimestamp="2025-10-08 18:11:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 18:13:20.168311279 +0000 UTC m=+145.618154059" watchObservedRunningTime="2025-10-08 18:13:20.23245765 +0000 UTC m=+145.682300420" Oct 08 18:13:20 crc kubenswrapper[4988]: I1008 18:13:20.294796 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-jrt4j" podStartSLOduration=124.294773478 podStartE2EDuration="2m4.294773478s" podCreationTimestamp="2025-10-08 18:11:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 18:13:20.248132197 +0000 UTC m=+145.697974967" watchObservedRunningTime="2025-10-08 18:13:20.294773478 +0000 UTC m=+145.744616248" Oct 08 18:13:20 crc kubenswrapper[4988]: I1008 18:13:20.295102 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-294dt" podStartSLOduration=124.295097877 podStartE2EDuration="2m4.295097877s" podCreationTimestamp="2025-10-08 18:11:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 18:13:20.29384378 +0000 UTC m=+145.743686550" watchObservedRunningTime="2025-10-08 18:13:20.295097877 +0000 UTC m=+145.744940647" Oct 08 18:13:20 crc kubenswrapper[4988]: I1008 18:13:20.306055 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 18:13:20 crc kubenswrapper[4988]: E1008 18:13:20.306521 4988 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 18:13:20.80649987 +0000 UTC m=+146.256342640 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 18:13:20 crc kubenswrapper[4988]: I1008 18:13:20.394252 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-qmwln" podStartSLOduration=124.394228827 podStartE2EDuration="2m4.394228827s" podCreationTimestamp="2025-10-08 18:11:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 18:13:20.368443455 +0000 UTC m=+145.818286225" watchObservedRunningTime="2025-10-08 18:13:20.394228827 +0000 UTC m=+145.844071597" Oct 08 18:13:20 crc kubenswrapper[4988]: I1008 18:13:20.407852 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-4wh9p\" (UID: \"8a33d7bd-26bd-409b-93f1-5f151ebca347\") " pod="openshift-image-registry/image-registry-697d97f7c8-4wh9p" Oct 08 18:13:20 crc kubenswrapper[4988]: E1008 18:13:20.408825 4988 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-08 18:13:20.908810223 +0000 UTC m=+146.358652993 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-4wh9p" (UID: "8a33d7bd-26bd-409b-93f1-5f151ebca347") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 18:13:20 crc kubenswrapper[4988]: I1008 18:13:20.446209 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-canary/ingress-canary-s4r2k" podStartSLOduration=6.446182623 podStartE2EDuration="6.446182623s" podCreationTimestamp="2025-10-08 18:13:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 18:13:20.444753661 +0000 UTC m=+145.894596441" watchObservedRunningTime="2025-10-08 18:13:20.446182623 +0000 UTC m=+145.896025393" Oct 08 18:13:20 crc kubenswrapper[4988]: I1008 18:13:20.446806 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd-operator/etcd-operator-b45778765-gqm5m" podStartSLOduration=124.446800511 podStartE2EDuration="2m4.446800511s" podCreationTimestamp="2025-10-08 18:11:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 18:13:20.404678813 +0000 UTC m=+145.854521583" watchObservedRunningTime="2025-10-08 18:13:20.446800511 +0000 UTC m=+145.896643281" Oct 08 18:13:20 crc kubenswrapper[4988]: I1008 18:13:20.524561 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 18:13:20 crc kubenswrapper[4988]: E1008 18:13:20.525885 4988 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 18:13:21.025863497 +0000 UTC m=+146.475706267 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 18:13:20 crc kubenswrapper[4988]: I1008 18:13:20.564894 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29332440-qxt8r" podStartSLOduration=124.564863064 podStartE2EDuration="2m4.564863064s" podCreationTimestamp="2025-10-08 18:11:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 18:13:20.519321017 +0000 UTC m=+145.969163787" watchObservedRunningTime="2025-10-08 18:13:20.564863064 +0000 UTC m=+146.014705844" Oct 08 18:13:20 crc kubenswrapper[4988]: I1008 18:13:20.636348 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-4wh9p\" (UID: \"8a33d7bd-26bd-409b-93f1-5f151ebca347\") " pod="openshift-image-registry/image-registry-697d97f7c8-4wh9p" Oct 08 18:13:20 crc kubenswrapper[4988]: E1008 18:13:20.636879 4988 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-08 18:13:21.136855725 +0000 UTC m=+146.586698495 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-4wh9p" (UID: "8a33d7bd-26bd-409b-93f1-5f151ebca347") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 18:13:20 crc kubenswrapper[4988]: I1008 18:13:20.737919 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 18:13:20 crc kubenswrapper[4988]: E1008 18:13:20.738560 4988 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 18:13:21.238512799 +0000 UTC m=+146.688355569 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 18:13:20 crc kubenswrapper[4988]: I1008 18:13:20.738746 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-4wh9p\" (UID: \"8a33d7bd-26bd-409b-93f1-5f151ebca347\") " pod="openshift-image-registry/image-registry-697d97f7c8-4wh9p" Oct 08 18:13:20 crc kubenswrapper[4988]: E1008 18:13:20.739148 4988 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-08 18:13:21.239130447 +0000 UTC m=+146.688973217 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-4wh9p" (UID: "8a33d7bd-26bd-409b-93f1-5f151ebca347") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 18:13:20 crc kubenswrapper[4988]: I1008 18:13:20.845570 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 18:13:20 crc kubenswrapper[4988]: E1008 18:13:20.845874 4988 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 18:13:21.345836979 +0000 UTC m=+146.795679749 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 18:13:20 crc kubenswrapper[4988]: I1008 18:13:20.846207 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-4wh9p\" (UID: \"8a33d7bd-26bd-409b-93f1-5f151ebca347\") " pod="openshift-image-registry/image-registry-697d97f7c8-4wh9p" Oct 08 18:13:20 crc kubenswrapper[4988]: E1008 18:13:20.846632 4988 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-08 18:13:21.346617142 +0000 UTC m=+146.796459912 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-4wh9p" (UID: "8a33d7bd-26bd-409b-93f1-5f151ebca347") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 18:13:20 crc kubenswrapper[4988]: I1008 18:13:20.862045 4988 patch_prober.go:28] interesting pod/router-default-5444994796-jjnkn container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 08 18:13:20 crc kubenswrapper[4988]: [-]has-synced failed: reason withheld Oct 08 18:13:20 crc kubenswrapper[4988]: [+]process-running ok Oct 08 18:13:20 crc kubenswrapper[4988]: healthz check failed Oct 08 18:13:20 crc kubenswrapper[4988]: I1008 18:13:20.862097 4988 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-jjnkn" podUID="ed79f1d0-1282-4be1-8e71-68a5320cdab1" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 08 18:13:20 crc kubenswrapper[4988]: I1008 18:13:20.948870 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 18:13:20 crc kubenswrapper[4988]: E1008 18:13:20.949305 4988 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 18:13:21.449284376 +0000 UTC m=+146.899127146 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 18:13:20 crc kubenswrapper[4988]: I1008 18:13:20.982991 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-hgnvq" event={"ID":"31a216f1-4766-4836-b3c5-a245ad42d709","Type":"ContainerStarted","Data":"950c8a0bb4d19581824d37e9fb897aae51b4d0f7a353d62324c3e811f64f72a4"} Oct 08 18:13:20 crc kubenswrapper[4988]: I1008 18:13:20.983223 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-hgnvq" event={"ID":"31a216f1-4766-4836-b3c5-a245ad42d709","Type":"ContainerStarted","Data":"ec32889fe9560444ef0713d0e32be5386ec2e7e8ab312744f0d55d70d98d679a"} Oct 08 18:13:20 crc kubenswrapper[4988]: I1008 18:13:20.998745 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-g775v" event={"ID":"9778cf39-b76d-4de6-b7ce-6cd9f3a4e224","Type":"ContainerStarted","Data":"10e5ac2b065edbfabf966378310a70be5e0c03eeca9594a454e6964bff4a21d0"} Oct 08 18:13:20 crc kubenswrapper[4988]: I1008 18:13:20.998805 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-g775v" event={"ID":"9778cf39-b76d-4de6-b7ce-6cd9f3a4e224","Type":"ContainerStarted","Data":"8f887b309a40778d1ce62c9a48db79ca22f23416056bbce766eb56c0e5ebfe06"} Oct 08 18:13:21 crc kubenswrapper[4988]: I1008 18:13:21.015367 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-d9ks8" event={"ID":"0a5e45fa-e12a-4199-aa81-96468c383594","Type":"ContainerStarted","Data":"34292f87925da7c2eba62f027819bb5083e80a77aaa9e63ef3b7a522fd8a1e3f"} Oct 08 18:13:21 crc kubenswrapper[4988]: I1008 18:13:21.016344 4988 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-d9ks8 container/packageserver namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.38:5443/healthz\": dial tcp 10.217.0.38:5443: connect: connection refused" start-of-body= Oct 08 18:13:21 crc kubenswrapper[4988]: I1008 18:13:21.016426 4988 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-d9ks8" podUID="0a5e45fa-e12a-4199-aa81-96468c383594" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.38:5443/healthz\": dial tcp 10.217.0.38:5443: connect: connection refused" Oct 08 18:13:21 crc kubenswrapper[4988]: I1008 18:13:21.031757 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-wxv6d" event={"ID":"7ee72f71-70e8-49ff-8407-96531bc67543","Type":"ContainerStarted","Data":"407e045cf7426ab3f4c00e627e28cc415ecceb1db508b3345f379a4eba265305"} Oct 08 18:13:21 crc kubenswrapper[4988]: I1008 18:13:21.031813 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-wxv6d" event={"ID":"7ee72f71-70e8-49ff-8407-96531bc67543","Type":"ContainerStarted","Data":"dba6352d88479fc2e9adeba3846b4e9b12a04c082a29562489784b823dd71be8"} Oct 08 18:13:21 crc kubenswrapper[4988]: I1008 18:13:21.056053 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-4wh9p\" (UID: \"8a33d7bd-26bd-409b-93f1-5f151ebca347\") " pod="openshift-image-registry/image-registry-697d97f7c8-4wh9p" Oct 08 18:13:21 crc kubenswrapper[4988]: E1008 18:13:21.057001 4988 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-08 18:13:21.556535344 +0000 UTC m=+147.006378114 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-4wh9p" (UID: "8a33d7bd-26bd-409b-93f1-5f151ebca347") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 18:13:21 crc kubenswrapper[4988]: I1008 18:13:21.058676 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-g7fbv" event={"ID":"47f38446-8f8d-4d84-8a16-f74e6f2836b7","Type":"ContainerStarted","Data":"f76bbe998d87ab347812701171e5a415d91debf0d2777d061b964d70ee73e3b4"} Oct 08 18:13:21 crc kubenswrapper[4988]: I1008 18:13:21.058744 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-g7fbv" event={"ID":"47f38446-8f8d-4d84-8a16-f74e6f2836b7","Type":"ContainerStarted","Data":"9537708bf604ae41e4664b7f042c0e8cd6f691baaad78a82fc31eb205982c295"} Oct 08 18:13:21 crc kubenswrapper[4988]: I1008 18:13:21.063223 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-hgnvq" podStartSLOduration=125.063203578 podStartE2EDuration="2m5.063203578s" podCreationTimestamp="2025-10-08 18:11:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 18:13:21.012048427 +0000 UTC m=+146.461891197" watchObservedRunningTime="2025-10-08 18:13:21.063203578 +0000 UTC m=+146.513046338" Oct 08 18:13:21 crc kubenswrapper[4988]: I1008 18:13:21.064179 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca-operator/service-ca-operator-777779d784-g775v" podStartSLOduration=125.064175547 podStartE2EDuration="2m5.064175547s" podCreationTimestamp="2025-10-08 18:11:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 18:13:21.0618894 +0000 UTC m=+146.511732170" watchObservedRunningTime="2025-10-08 18:13:21.064175547 +0000 UTC m=+146.514018317" Oct 08 18:13:21 crc kubenswrapper[4988]: I1008 18:13:21.081704 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-s4r2k" event={"ID":"db1043d1-7232-4fcd-9ce5-50b590798d36","Type":"ContainerStarted","Data":"351748b742ebbdf5f751870599ca97174350909e996304fa467ffff6138a6e61"} Oct 08 18:13:21 crc kubenswrapper[4988]: I1008 18:13:21.095351 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-gl7lr" event={"ID":"9cb89712-c3bf-4ed9-87f8-8124a55f5f7e","Type":"ContainerStarted","Data":"ebbf2bee11f4524e24ca45ebf9248185c9286deb7029a7bdc8c1af0ab5cb0be3"} Oct 08 18:13:21 crc kubenswrapper[4988]: I1008 18:13:21.104978 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca/service-ca-9c57cc56f-wxv6d" podStartSLOduration=125.104941505 podStartE2EDuration="2m5.104941505s" podCreationTimestamp="2025-10-08 18:11:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 18:13:21.098834678 +0000 UTC m=+146.548677448" watchObservedRunningTime="2025-10-08 18:13:21.104941505 +0000 UTC m=+146.554784275" Oct 08 18:13:21 crc kubenswrapper[4988]: I1008 18:13:21.105650 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-q2mxf" event={"ID":"aa86616c-479c-45be-aaa3-d3c4994721bf","Type":"ContainerStarted","Data":"aa4b503101443d60577da11609430b76823285b1bbd0b07f2e8bd0c95a30cbe0"} Oct 08 18:13:21 crc kubenswrapper[4988]: I1008 18:13:21.107754 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-q2mxf" event={"ID":"aa86616c-479c-45be-aaa3-d3c4994721bf","Type":"ContainerStarted","Data":"fcfe89d5dfe7d2aba2e8114879df69b7c9d6e41b80e2e665f877ccd76023a1a3"} Oct 08 18:13:21 crc kubenswrapper[4988]: I1008 18:13:21.138027 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-h5cw8" event={"ID":"ba2264e5-af3b-4400-9c20-04905b075dfa","Type":"ContainerStarted","Data":"8ade6bb6d58ae08f15efa818bc2f846d250dc75a076b51673edaf205ec3fb619"} Oct 08 18:13:21 crc kubenswrapper[4988]: I1008 18:13:21.138079 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-h5cw8" event={"ID":"ba2264e5-af3b-4400-9c20-04905b075dfa","Type":"ContainerStarted","Data":"f29fbb88da2daf7545a61f69a84bcaaa89dad60eb5df995cfb7ca78651505d08"} Oct 08 18:13:21 crc kubenswrapper[4988]: I1008 18:13:21.138788 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-h5cw8" Oct 08 18:13:21 crc kubenswrapper[4988]: I1008 18:13:21.145438 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-2d794" event={"ID":"e6f22dd3-65b7-4fe3-8265-8fb96095b6c4","Type":"ContainerStarted","Data":"cf290e3d3ac0cd8aad201ec15b5e749998e3ea12254d22b6aae91d36974a723f"} Oct 08 18:13:21 crc kubenswrapper[4988]: I1008 18:13:21.162161 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 18:13:21 crc kubenswrapper[4988]: E1008 18:13:21.163010 4988 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 18:13:21.662982839 +0000 UTC m=+147.112825609 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 18:13:21 crc kubenswrapper[4988]: I1008 18:13:21.173670 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-g7fbv" podStartSLOduration=125.17364533 podStartE2EDuration="2m5.17364533s" podCreationTimestamp="2025-10-08 18:11:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 18:13:21.135117656 +0000 UTC m=+146.584960426" watchObservedRunningTime="2025-10-08 18:13:21.17364533 +0000 UTC m=+146.623488100" Oct 08 18:13:21 crc kubenswrapper[4988]: I1008 18:13:21.174589 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-q2mxf" podStartSLOduration=125.174584167 podStartE2EDuration="2m5.174584167s" podCreationTimestamp="2025-10-08 18:11:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 18:13:21.172431534 +0000 UTC m=+146.622274314" watchObservedRunningTime="2025-10-08 18:13:21.174584167 +0000 UTC m=+146.624426937" Oct 08 18:13:21 crc kubenswrapper[4988]: I1008 18:13:21.182048 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-bbpzz" event={"ID":"4ddacc00-9f9b-44ec-8356-cd932ce26e5b","Type":"ContainerStarted","Data":"2cc02de6a44285491513249d2a4879d789dbaacf46516fd186629d6502a9788a"} Oct 08 18:13:21 crc kubenswrapper[4988]: I1008 18:13:21.182095 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-bbpzz" event={"ID":"4ddacc00-9f9b-44ec-8356-cd932ce26e5b","Type":"ContainerStarted","Data":"a23f50c7a31cb8c56c030d7ca1443a4c4f10195cc2f24d5af2ab6e097c25266e"} Oct 08 18:13:21 crc kubenswrapper[4988]: I1008 18:13:21.212519 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-5mrrc" event={"ID":"a760b68f-a1f6-4741-a9e9-51f6d3968748","Type":"ContainerStarted","Data":"0ff6adb1bba5f3d5cb4d56b19f39024a5711695e75d3a593bf7eb4aca697fbb3"} Oct 08 18:13:21 crc kubenswrapper[4988]: I1008 18:13:21.212571 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-5mrrc" event={"ID":"a760b68f-a1f6-4741-a9e9-51f6d3968748","Type":"ContainerStarted","Data":"a8813f2ca4273369f618572c457bac3a94f622d7536c308aa184ebe8ed6bf7fa"} Oct 08 18:13:21 crc kubenswrapper[4988]: I1008 18:13:21.255113 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-m55d6" event={"ID":"b5cd6079-031c-4cc8-8161-88f0e6f84064","Type":"ContainerStarted","Data":"f806b26545d577be2b769fcfdb797ce9f103d25936bcfd62f5db307854c68c15"} Oct 08 18:13:21 crc kubenswrapper[4988]: I1008 18:13:21.265516 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-4wh9p\" (UID: \"8a33d7bd-26bd-409b-93f1-5f151ebca347\") " pod="openshift-image-registry/image-registry-697d97f7c8-4wh9p" Oct 08 18:13:21 crc kubenswrapper[4988]: E1008 18:13:21.267415 4988 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-08 18:13:21.767399403 +0000 UTC m=+147.217242173 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-4wh9p" (UID: "8a33d7bd-26bd-409b-93f1-5f151ebca347") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 18:13:21 crc kubenswrapper[4988]: I1008 18:13:21.273503 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-k4ghh" event={"ID":"4719309b-a3b7-4910-a389-635acc09d01d","Type":"ContainerStarted","Data":"fbb4d4b7d4ac8662dd2010b479f553750f622c7083a0a4b21e994604f487ac52"} Oct 08 18:13:21 crc kubenswrapper[4988]: I1008 18:13:21.279652 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns-operator/dns-operator-744455d44c-2d794" podStartSLOduration=125.279640411 podStartE2EDuration="2m5.279640411s" podCreationTimestamp="2025-10-08 18:11:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 18:13:21.225014928 +0000 UTC m=+146.674857718" watchObservedRunningTime="2025-10-08 18:13:21.279640411 +0000 UTC m=+146.729483181" Oct 08 18:13:21 crc kubenswrapper[4988]: I1008 18:13:21.289558 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-l86nd" event={"ID":"c83b25d6-a1c1-4394-9954-697156fe53b7","Type":"ContainerStarted","Data":"a7986b346f89ec2cc3fa17e4b6f02809a3841e4d70b8855bd11fb00502ecf0d2"} Oct 08 18:13:21 crc kubenswrapper[4988]: I1008 18:13:21.310301 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-m55d6" podStartSLOduration=125.310283184 podStartE2EDuration="2m5.310283184s" podCreationTimestamp="2025-10-08 18:11:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 18:13:21.309681807 +0000 UTC m=+146.759524577" watchObservedRunningTime="2025-10-08 18:13:21.310283184 +0000 UTC m=+146.760125944" Oct 08 18:13:21 crc kubenswrapper[4988]: I1008 18:13:21.311211 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-h5cw8" podStartSLOduration=125.311201142 podStartE2EDuration="2m5.311201142s" podCreationTimestamp="2025-10-08 18:11:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 18:13:21.279065214 +0000 UTC m=+146.728907984" watchObservedRunningTime="2025-10-08 18:13:21.311201142 +0000 UTC m=+146.761043912" Oct 08 18:13:21 crc kubenswrapper[4988]: I1008 18:13:21.310360 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-4phs4" event={"ID":"05152e1e-3bc1-48b0-818d-41bc0d50fbf5","Type":"ContainerDied","Data":"77267ece2b1dac47544e94c2fd21cb4b1d440cc7bc77948a9e61a07ff1fc7dd0"} Oct 08 18:13:21 crc kubenswrapper[4988]: I1008 18:13:21.310322 4988 generic.go:334] "Generic (PLEG): container finished" podID="05152e1e-3bc1-48b0-818d-41bc0d50fbf5" containerID="77267ece2b1dac47544e94c2fd21cb4b1d440cc7bc77948a9e61a07ff1fc7dd0" exitCode=0 Oct 08 18:13:21 crc kubenswrapper[4988]: I1008 18:13:21.315604 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-4phs4" event={"ID":"05152e1e-3bc1-48b0-818d-41bc0d50fbf5","Type":"ContainerStarted","Data":"f9808f90b1da24ea3faa88a547674e28633679371159124db1e661aa25cbe17f"} Oct 08 18:13:21 crc kubenswrapper[4988]: I1008 18:13:21.330052 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-q2w8f" event={"ID":"0769dc07-8f56-42a5-8eb5-b900eefaaf1e","Type":"ContainerStarted","Data":"fdd1f8f67ffbd2cfdefaba27610bd686500e5ac5f0ec42d960ff087fa58eb766"} Oct 08 18:13:21 crc kubenswrapper[4988]: I1008 18:13:21.330424 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-q2w8f" event={"ID":"0769dc07-8f56-42a5-8eb5-b900eefaaf1e","Type":"ContainerStarted","Data":"f5d2944f36cd16eb181d6ece3ff5185f01c8a2d8fd08209df2130d8dc2446c96"} Oct 08 18:13:21 crc kubenswrapper[4988]: I1008 18:13:21.332280 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-rgqjs" event={"ID":"1ca33297-d58c-4493-b1d3-45dd83983be4","Type":"ContainerStarted","Data":"72549f2c79de6f7ea37db11b9c5d099d659df5c98d84348dd7691d535597aaf6"} Oct 08 18:13:21 crc kubenswrapper[4988]: I1008 18:13:21.334355 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-kwcts" event={"ID":"7f0973f4-25d3-48a6-ac02-b84f9bb4f947","Type":"ContainerStarted","Data":"db868f41baaf12980f8c9fabda9525bdf26fd1401324543eb720a2bd7c3d0066"} Oct 08 18:13:21 crc kubenswrapper[4988]: I1008 18:13:21.334377 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-kwcts" event={"ID":"7f0973f4-25d3-48a6-ac02-b84f9bb4f947","Type":"ContainerStarted","Data":"24ec6fc420393280cd221e4c4128aaef5222d6e99632cf813176a47d2ca26d69"} Oct 08 18:13:21 crc kubenswrapper[4988]: I1008 18:13:21.345722 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/machine-api-operator-5694c8668f-5mrrc" podStartSLOduration=125.345694297 podStartE2EDuration="2m5.345694297s" podCreationTimestamp="2025-10-08 18:11:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 18:13:21.336793878 +0000 UTC m=+146.786636678" watchObservedRunningTime="2025-10-08 18:13:21.345694297 +0000 UTC m=+146.795537067" Oct 08 18:13:21 crc kubenswrapper[4988]: I1008 18:13:21.367608 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 18:13:21 crc kubenswrapper[4988]: E1008 18:13:21.369856 4988 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 18:13:21.869829101 +0000 UTC m=+147.319671871 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 18:13:21 crc kubenswrapper[4988]: I1008 18:13:21.372794 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-984jv" event={"ID":"66bebf42-9ec1-48ae-ae89-7586af40d06f","Type":"ContainerStarted","Data":"9e3fd48193bcd27544850ca6521ac22746810fc58d13a11b228cfc44a0ae2a30"} Oct 08 18:13:21 crc kubenswrapper[4988]: I1008 18:13:21.372927 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-984jv" event={"ID":"66bebf42-9ec1-48ae-ae89-7586af40d06f","Type":"ContainerStarted","Data":"ff932b8aca1074701b284a2870475907fbb3ebd7d56c94f6a2eae51d7442dc56"} Oct 08 18:13:21 crc kubenswrapper[4988]: I1008 18:13:21.372997 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-984jv" event={"ID":"66bebf42-9ec1-48ae-ae89-7586af40d06f","Type":"ContainerStarted","Data":"260c4d3504de19ee2f0581301115411763ad7337b08140cc2baa7aaae06dba1e"} Oct 08 18:13:21 crc kubenswrapper[4988]: I1008 18:13:21.375098 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-config-operator/openshift-config-operator-7777fb866f-7ctzs" Oct 08 18:13:21 crc kubenswrapper[4988]: I1008 18:13:21.403074 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-85kp6" event={"ID":"ba81cb6f-e6fb-49e1-aa4c-5a9e6f8769fb","Type":"ContainerStarted","Data":"3b3630b60bd0e2d2f21e4404fcb74a852dafe9118df00154c39eeddfb8d37375"} Oct 08 18:13:21 crc kubenswrapper[4988]: I1008 18:13:21.408344 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-rgqjs" podStartSLOduration=125.408318944 podStartE2EDuration="2m5.408318944s" podCreationTimestamp="2025-10-08 18:11:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 18:13:21.407975673 +0000 UTC m=+146.857818443" watchObservedRunningTime="2025-10-08 18:13:21.408318944 +0000 UTC m=+146.858161714" Oct 08 18:13:21 crc kubenswrapper[4988]: I1008 18:13:21.409964 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-bbpzz" podStartSLOduration=125.409958021 podStartE2EDuration="2m5.409958021s" podCreationTimestamp="2025-10-08 18:11:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 18:13:21.374002373 +0000 UTC m=+146.823845153" watchObservedRunningTime="2025-10-08 18:13:21.409958021 +0000 UTC m=+146.859800791" Oct 08 18:13:21 crc kubenswrapper[4988]: I1008 18:13:21.459781 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-wnkxn" event={"ID":"176096f3-5341-4aa9-a706-5e96cfd8eca1","Type":"ContainerStarted","Data":"e3ad3dd5f6e31df81476ffc99042dfd05ffcf7faf7f88edc79c1363419da0474"} Oct 08 18:13:21 crc kubenswrapper[4988]: I1008 18:13:21.461923 4988 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-294dt container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.29:8080/healthz\": dial tcp 10.217.0.29:8080: connect: connection refused" start-of-body= Oct 08 18:13:21 crc kubenswrapper[4988]: I1008 18:13:21.462000 4988 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-294dt" podUID="8bf9175c-5101-4fcd-8cba-e113be11dde4" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.29:8080/healthz\": dial tcp 10.217.0.29:8080: connect: connection refused" Oct 08 18:13:21 crc kubenswrapper[4988]: I1008 18:13:21.469869 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-4wh9p\" (UID: \"8a33d7bd-26bd-409b-93f1-5f151ebca347\") " pod="openshift-image-registry/image-registry-697d97f7c8-4wh9p" Oct 08 18:13:21 crc kubenswrapper[4988]: E1008 18:13:21.475879 4988 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-08 18:13:21.975854744 +0000 UTC m=+147.425697514 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-4wh9p" (UID: "8a33d7bd-26bd-409b-93f1-5f151ebca347") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 18:13:21 crc kubenswrapper[4988]: I1008 18:13:21.499019 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-jrt4j" Oct 08 18:13:21 crc kubenswrapper[4988]: I1008 18:13:21.572128 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 18:13:21 crc kubenswrapper[4988]: E1008 18:13:21.573539 4988 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 18:13:22.073509721 +0000 UTC m=+147.523352491 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 18:13:21 crc kubenswrapper[4988]: I1008 18:13:21.601550 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-admission-controller-857f4d67dd-q2w8f" podStartSLOduration=125.601522258 podStartE2EDuration="2m5.601522258s" podCreationTimestamp="2025-10-08 18:11:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 18:13:21.51996163 +0000 UTC m=+146.969804390" watchObservedRunningTime="2025-10-08 18:13:21.601522258 +0000 UTC m=+147.051365028" Oct 08 18:13:21 crc kubenswrapper[4988]: I1008 18:13:21.650297 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-984jv" podStartSLOduration=125.65026695 podStartE2EDuration="2m5.65026695s" podCreationTimestamp="2025-10-08 18:11:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 18:13:21.646997025 +0000 UTC m=+147.096839805" watchObservedRunningTime="2025-10-08 18:13:21.65026695 +0000 UTC m=+147.100109720" Oct 08 18:13:21 crc kubenswrapper[4988]: I1008 18:13:21.675046 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-4wh9p\" (UID: \"8a33d7bd-26bd-409b-93f1-5f151ebca347\") " pod="openshift-image-registry/image-registry-697d97f7c8-4wh9p" Oct 08 18:13:21 crc kubenswrapper[4988]: E1008 18:13:21.675966 4988 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-08 18:13:22.175932928 +0000 UTC m=+147.625775868 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-4wh9p" (UID: "8a33d7bd-26bd-409b-93f1-5f151ebca347") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 18:13:21 crc kubenswrapper[4988]: I1008 18:13:21.770232 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-85kp6" Oct 08 18:13:21 crc kubenswrapper[4988]: I1008 18:13:21.770306 4988 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-85kp6" Oct 08 18:13:21 crc kubenswrapper[4988]: I1008 18:13:21.776517 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 18:13:21 crc kubenswrapper[4988]: E1008 18:13:21.776774 4988 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 18:13:22.276722968 +0000 UTC m=+147.726565748 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 18:13:21 crc kubenswrapper[4988]: I1008 18:13:21.777016 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-4wh9p\" (UID: \"8a33d7bd-26bd-409b-93f1-5f151ebca347\") " pod="openshift-image-registry/image-registry-697d97f7c8-4wh9p" Oct 08 18:13:21 crc kubenswrapper[4988]: E1008 18:13:21.777498 4988 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-08 18:13:22.27748347 +0000 UTC m=+147.727326250 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-4wh9p" (UID: "8a33d7bd-26bd-409b-93f1-5f151ebca347") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 18:13:21 crc kubenswrapper[4988]: I1008 18:13:21.814529 4988 patch_prober.go:28] interesting pod/apiserver-7bbb656c7d-85kp6 container/oauth-apiserver namespace/openshift-oauth-apiserver: Startup probe status=failure output="Get \"https://10.217.0.14:8443/livez\": dial tcp 10.217.0.14:8443: connect: connection refused" start-of-body= Oct 08 18:13:21 crc kubenswrapper[4988]: I1008 18:13:21.814595 4988 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-85kp6" podUID="ba81cb6f-e6fb-49e1-aa4c-5a9e6f8769fb" containerName="oauth-apiserver" probeResult="failure" output="Get \"https://10.217.0.14:8443/livez\": dial tcp 10.217.0.14:8443: connect: connection refused" Oct 08 18:13:21 crc kubenswrapper[4988]: I1008 18:13:21.846321 4988 patch_prober.go:28] interesting pod/router-default-5444994796-jjnkn container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 08 18:13:21 crc kubenswrapper[4988]: [-]has-synced failed: reason withheld Oct 08 18:13:21 crc kubenswrapper[4988]: [+]process-running ok Oct 08 18:13:21 crc kubenswrapper[4988]: healthz check failed Oct 08 18:13:21 crc kubenswrapper[4988]: I1008 18:13:21.846395 4988 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-jjnkn" podUID="ed79f1d0-1282-4be1-8e71-68a5320cdab1" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 08 18:13:21 crc kubenswrapper[4988]: I1008 18:13:21.872074 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-85kp6" podStartSLOduration=125.872051188 podStartE2EDuration="2m5.872051188s" podCreationTimestamp="2025-10-08 18:11:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 18:13:21.808564447 +0000 UTC m=+147.258407217" watchObservedRunningTime="2025-10-08 18:13:21.872051188 +0000 UTC m=+147.321893958" Oct 08 18:13:21 crc kubenswrapper[4988]: I1008 18:13:21.877864 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 18:13:21 crc kubenswrapper[4988]: E1008 18:13:21.878297 4988 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 18:13:22.37827569 +0000 UTC m=+147.828118460 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 18:13:21 crc kubenswrapper[4988]: I1008 18:13:21.979948 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-4wh9p\" (UID: \"8a33d7bd-26bd-409b-93f1-5f151ebca347\") " pod="openshift-image-registry/image-registry-697d97f7c8-4wh9p" Oct 08 18:13:21 crc kubenswrapper[4988]: E1008 18:13:21.980486 4988 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-08 18:13:22.48046298 +0000 UTC m=+147.930305750 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-4wh9p" (UID: "8a33d7bd-26bd-409b-93f1-5f151ebca347") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 18:13:22 crc kubenswrapper[4988]: I1008 18:13:22.081646 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 18:13:22 crc kubenswrapper[4988]: E1008 18:13:22.081958 4988 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 18:13:22.581921389 +0000 UTC m=+148.031764159 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 18:13:22 crc kubenswrapper[4988]: I1008 18:13:22.082016 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-4wh9p\" (UID: \"8a33d7bd-26bd-409b-93f1-5f151ebca347\") " pod="openshift-image-registry/image-registry-697d97f7c8-4wh9p" Oct 08 18:13:22 crc kubenswrapper[4988]: E1008 18:13:22.082484 4988 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-08 18:13:22.582463335 +0000 UTC m=+148.032306315 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-4wh9p" (UID: "8a33d7bd-26bd-409b-93f1-5f151ebca347") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 18:13:22 crc kubenswrapper[4988]: I1008 18:13:22.183450 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 18:13:22 crc kubenswrapper[4988]: E1008 18:13:22.183675 4988 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 18:13:22.683641086 +0000 UTC m=+148.133483856 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 18:13:22 crc kubenswrapper[4988]: I1008 18:13:22.183886 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-4wh9p\" (UID: \"8a33d7bd-26bd-409b-93f1-5f151ebca347\") " pod="openshift-image-registry/image-registry-697d97f7c8-4wh9p" Oct 08 18:13:22 crc kubenswrapper[4988]: E1008 18:13:22.184325 4988 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-08 18:13:22.684306515 +0000 UTC m=+148.134149285 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-4wh9p" (UID: "8a33d7bd-26bd-409b-93f1-5f151ebca347") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 18:13:22 crc kubenswrapper[4988]: I1008 18:13:22.284812 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 18:13:22 crc kubenswrapper[4988]: E1008 18:13:22.284974 4988 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 18:13:22.784954731 +0000 UTC m=+148.234797501 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 18:13:22 crc kubenswrapper[4988]: I1008 18:13:22.285210 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-4wh9p\" (UID: \"8a33d7bd-26bd-409b-93f1-5f151ebca347\") " pod="openshift-image-registry/image-registry-697d97f7c8-4wh9p" Oct 08 18:13:22 crc kubenswrapper[4988]: E1008 18:13:22.285576 4988 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-08 18:13:22.785565909 +0000 UTC m=+148.235408679 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-4wh9p" (UID: "8a33d7bd-26bd-409b-93f1-5f151ebca347") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 18:13:22 crc kubenswrapper[4988]: I1008 18:13:22.386347 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 18:13:22 crc kubenswrapper[4988]: E1008 18:13:22.386532 4988 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 18:13:22.886484032 +0000 UTC m=+148.336326802 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 18:13:22 crc kubenswrapper[4988]: I1008 18:13:22.386951 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-4wh9p\" (UID: \"8a33d7bd-26bd-409b-93f1-5f151ebca347\") " pod="openshift-image-registry/image-registry-697d97f7c8-4wh9p" Oct 08 18:13:22 crc kubenswrapper[4988]: E1008 18:13:22.387278 4988 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-08 18:13:22.887270574 +0000 UTC m=+148.337113344 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-4wh9p" (UID: "8a33d7bd-26bd-409b-93f1-5f151ebca347") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 18:13:22 crc kubenswrapper[4988]: I1008 18:13:22.470244 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-kwcts" event={"ID":"7f0973f4-25d3-48a6-ac02-b84f9bb4f947","Type":"ContainerStarted","Data":"3290351f4af6e5eca4399ab01738321d801bf6603a50923e7c530b1674ea16ad"} Oct 08 18:13:22 crc kubenswrapper[4988]: I1008 18:13:22.471449 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-dns/dns-default-kwcts" Oct 08 18:13:22 crc kubenswrapper[4988]: I1008 18:13:22.473340 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-4phs4" event={"ID":"05152e1e-3bc1-48b0-818d-41bc0d50fbf5","Type":"ContainerStarted","Data":"5169e65ba0a2c900ebacaddcb69bada5f5e1d131c669dec009d17eecb4cbf138"} Oct 08 18:13:22 crc kubenswrapper[4988]: I1008 18:13:22.473371 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-4phs4" event={"ID":"05152e1e-3bc1-48b0-818d-41bc0d50fbf5","Type":"ContainerStarted","Data":"692fb4787c293aed0a14e58f75249af782d75c27d87b688499d2c3e5b211a356"} Oct 08 18:13:22 crc kubenswrapper[4988]: I1008 18:13:22.474937 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-2d794" event={"ID":"e6f22dd3-65b7-4fe3-8265-8fb96095b6c4","Type":"ContainerStarted","Data":"9ac1bf916b7b9384c207dfdc7d36da4137dc0cfb62a24af677dcd270946c761c"} Oct 08 18:13:22 crc kubenswrapper[4988]: I1008 18:13:22.477137 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-bbpzz" event={"ID":"4ddacc00-9f9b-44ec-8356-cd932ce26e5b","Type":"ContainerStarted","Data":"4576b3c419353409c62b1da40fef31015c31287a8ef97e97cb1ee0df33e00017"} Oct 08 18:13:22 crc kubenswrapper[4988]: I1008 18:13:22.478627 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-5mrrc" event={"ID":"a760b68f-a1f6-4741-a9e9-51f6d3968748","Type":"ContainerStarted","Data":"09adda0e5eeccbef30535d0b1d2a2a7689192b3538f612fee42d3e21904cbb50"} Oct 08 18:13:22 crc kubenswrapper[4988]: I1008 18:13:22.481306 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-gl7lr" event={"ID":"9cb89712-c3bf-4ed9-87f8-8124a55f5f7e","Type":"ContainerStarted","Data":"e5c67e3fc8e9ed170e22b8bfc68773c6ad698c56da1b51b5d33585fc91ac00da"} Oct 08 18:13:22 crc kubenswrapper[4988]: I1008 18:13:22.482999 4988 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-294dt container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.29:8080/healthz\": dial tcp 10.217.0.29:8080: connect: connection refused" start-of-body= Oct 08 18:13:22 crc kubenswrapper[4988]: I1008 18:13:22.483073 4988 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-294dt" podUID="8bf9175c-5101-4fcd-8cba-e113be11dde4" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.29:8080/healthz\": dial tcp 10.217.0.29:8080: connect: connection refused" Oct 08 18:13:22 crc kubenswrapper[4988]: I1008 18:13:22.486246 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-wnkxn" Oct 08 18:13:22 crc kubenswrapper[4988]: I1008 18:13:22.487701 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 18:13:22 crc kubenswrapper[4988]: E1008 18:13:22.487860 4988 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 18:13:22.987834007 +0000 UTC m=+148.437676767 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 18:13:22 crc kubenswrapper[4988]: I1008 18:13:22.488340 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-4wh9p\" (UID: \"8a33d7bd-26bd-409b-93f1-5f151ebca347\") " pod="openshift-image-registry/image-registry-697d97f7c8-4wh9p" Oct 08 18:13:22 crc kubenswrapper[4988]: E1008 18:13:22.498624 4988 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-08 18:13:22.998602542 +0000 UTC m=+148.448445302 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-4wh9p" (UID: "8a33d7bd-26bd-409b-93f1-5f151ebca347") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 18:13:22 crc kubenswrapper[4988]: I1008 18:13:22.520073 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-wnkxn" podStartSLOduration=126.520049497 podStartE2EDuration="2m6.520049497s" podCreationTimestamp="2025-10-08 18:11:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 18:13:21.969328406 +0000 UTC m=+147.419171176" watchObservedRunningTime="2025-10-08 18:13:22.520049497 +0000 UTC m=+147.969892267" Oct 08 18:13:22 crc kubenswrapper[4988]: I1008 18:13:22.525130 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-wnkxn" Oct 08 18:13:22 crc kubenswrapper[4988]: I1008 18:13:22.553249 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/dns-default-kwcts" podStartSLOduration=8.553222734 podStartE2EDuration="8.553222734s" podCreationTimestamp="2025-10-08 18:13:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 18:13:22.525908518 +0000 UTC m=+147.975751288" watchObservedRunningTime="2025-10-08 18:13:22.553222734 +0000 UTC m=+148.003065504" Oct 08 18:13:22 crc kubenswrapper[4988]: I1008 18:13:22.556469 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-77xtx"] Oct 08 18:13:22 crc kubenswrapper[4988]: I1008 18:13:22.558124 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-77xtx" Oct 08 18:13:22 crc kubenswrapper[4988]: I1008 18:13:22.565741 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver/apiserver-76f77b778f-4phs4" podStartSLOduration=126.565621496 podStartE2EDuration="2m6.565621496s" podCreationTimestamp="2025-10-08 18:11:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 18:13:22.554544353 +0000 UTC m=+148.004387123" watchObservedRunningTime="2025-10-08 18:13:22.565621496 +0000 UTC m=+148.015464266" Oct 08 18:13:22 crc kubenswrapper[4988]: I1008 18:13:22.566367 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Oct 08 18:13:22 crc kubenswrapper[4988]: I1008 18:13:22.573358 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-77xtx"] Oct 08 18:13:22 crc kubenswrapper[4988]: I1008 18:13:22.594019 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 18:13:22 crc kubenswrapper[4988]: E1008 18:13:22.596485 4988 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 18:13:23.096446475 +0000 UTC m=+148.546289245 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 18:13:22 crc kubenswrapper[4988]: I1008 18:13:22.622458 4988 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-apiserver/apiserver-76f77b778f-4phs4" Oct 08 18:13:22 crc kubenswrapper[4988]: I1008 18:13:22.622577 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-apiserver/apiserver-76f77b778f-4phs4" Oct 08 18:13:22 crc kubenswrapper[4988]: I1008 18:13:22.623373 4988 patch_prober.go:28] interesting pod/apiserver-76f77b778f-4phs4 container/openshift-apiserver namespace/openshift-apiserver: Startup probe status=failure output="Get \"https://10.217.0.31:8443/livez\": dial tcp 10.217.0.31:8443: connect: connection refused" start-of-body= Oct 08 18:13:22 crc kubenswrapper[4988]: I1008 18:13:22.623433 4988 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-apiserver/apiserver-76f77b778f-4phs4" podUID="05152e1e-3bc1-48b0-818d-41bc0d50fbf5" containerName="openshift-apiserver" probeResult="failure" output="Get \"https://10.217.0.31:8443/livez\": dial tcp 10.217.0.31:8443: connect: connection refused" Oct 08 18:13:22 crc kubenswrapper[4988]: I1008 18:13:22.703600 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kcc4w\" (UniqueName: \"kubernetes.io/projected/7c0e2c11-212a-46ed-b72a-911bde29199a-kube-api-access-kcc4w\") pod \"community-operators-77xtx\" (UID: \"7c0e2c11-212a-46ed-b72a-911bde29199a\") " pod="openshift-marketplace/community-operators-77xtx" Oct 08 18:13:22 crc kubenswrapper[4988]: I1008 18:13:22.703837 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-4wh9p\" (UID: \"8a33d7bd-26bd-409b-93f1-5f151ebca347\") " pod="openshift-image-registry/image-registry-697d97f7c8-4wh9p" Oct 08 18:13:22 crc kubenswrapper[4988]: I1008 18:13:22.703914 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7c0e2c11-212a-46ed-b72a-911bde29199a-catalog-content\") pod \"community-operators-77xtx\" (UID: \"7c0e2c11-212a-46ed-b72a-911bde29199a\") " pod="openshift-marketplace/community-operators-77xtx" Oct 08 18:13:22 crc kubenswrapper[4988]: I1008 18:13:22.703970 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7c0e2c11-212a-46ed-b72a-911bde29199a-utilities\") pod \"community-operators-77xtx\" (UID: \"7c0e2c11-212a-46ed-b72a-911bde29199a\") " pod="openshift-marketplace/community-operators-77xtx" Oct 08 18:13:22 crc kubenswrapper[4988]: E1008 18:13:22.705591 4988 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-08 18:13:23.205572578 +0000 UTC m=+148.655415558 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-4wh9p" (UID: "8a33d7bd-26bd-409b-93f1-5f151ebca347") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 18:13:22 crc kubenswrapper[4988]: I1008 18:13:22.729544 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-dmdqd"] Oct 08 18:13:22 crc kubenswrapper[4988]: I1008 18:13:22.730779 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-dmdqd" Oct 08 18:13:22 crc kubenswrapper[4988]: I1008 18:13:22.736452 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Oct 08 18:13:22 crc kubenswrapper[4988]: I1008 18:13:22.747626 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-dmdqd"] Oct 08 18:13:22 crc kubenswrapper[4988]: I1008 18:13:22.804562 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 18:13:22 crc kubenswrapper[4988]: I1008 18:13:22.804685 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7c0e2c11-212a-46ed-b72a-911bde29199a-utilities\") pod \"community-operators-77xtx\" (UID: \"7c0e2c11-212a-46ed-b72a-911bde29199a\") " pod="openshift-marketplace/community-operators-77xtx" Oct 08 18:13:22 crc kubenswrapper[4988]: I1008 18:13:22.804744 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kcc4w\" (UniqueName: \"kubernetes.io/projected/7c0e2c11-212a-46ed-b72a-911bde29199a-kube-api-access-kcc4w\") pod \"community-operators-77xtx\" (UID: \"7c0e2c11-212a-46ed-b72a-911bde29199a\") " pod="openshift-marketplace/community-operators-77xtx" Oct 08 18:13:22 crc kubenswrapper[4988]: I1008 18:13:22.804805 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/15a52185-0755-45a6-b665-c1185d1996b4-utilities\") pod \"certified-operators-dmdqd\" (UID: \"15a52185-0755-45a6-b665-c1185d1996b4\") " pod="openshift-marketplace/certified-operators-dmdqd" Oct 08 18:13:22 crc kubenswrapper[4988]: I1008 18:13:22.804824 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n95d9\" (UniqueName: \"kubernetes.io/projected/15a52185-0755-45a6-b665-c1185d1996b4-kube-api-access-n95d9\") pod \"certified-operators-dmdqd\" (UID: \"15a52185-0755-45a6-b665-c1185d1996b4\") " pod="openshift-marketplace/certified-operators-dmdqd" Oct 08 18:13:22 crc kubenswrapper[4988]: I1008 18:13:22.804846 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7c0e2c11-212a-46ed-b72a-911bde29199a-catalog-content\") pod \"community-operators-77xtx\" (UID: \"7c0e2c11-212a-46ed-b72a-911bde29199a\") " pod="openshift-marketplace/community-operators-77xtx" Oct 08 18:13:22 crc kubenswrapper[4988]: I1008 18:13:22.804878 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/15a52185-0755-45a6-b665-c1185d1996b4-catalog-content\") pod \"certified-operators-dmdqd\" (UID: \"15a52185-0755-45a6-b665-c1185d1996b4\") " pod="openshift-marketplace/certified-operators-dmdqd" Oct 08 18:13:22 crc kubenswrapper[4988]: I1008 18:13:22.805346 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7c0e2c11-212a-46ed-b72a-911bde29199a-utilities\") pod \"community-operators-77xtx\" (UID: \"7c0e2c11-212a-46ed-b72a-911bde29199a\") " pod="openshift-marketplace/community-operators-77xtx" Oct 08 18:13:22 crc kubenswrapper[4988]: E1008 18:13:22.805481 4988 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 18:13:23.30544078 +0000 UTC m=+148.755283550 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 18:13:22 crc kubenswrapper[4988]: I1008 18:13:22.805596 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7c0e2c11-212a-46ed-b72a-911bde29199a-catalog-content\") pod \"community-operators-77xtx\" (UID: \"7c0e2c11-212a-46ed-b72a-911bde29199a\") " pod="openshift-marketplace/community-operators-77xtx" Oct 08 18:13:22 crc kubenswrapper[4988]: I1008 18:13:22.838282 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kcc4w\" (UniqueName: \"kubernetes.io/projected/7c0e2c11-212a-46ed-b72a-911bde29199a-kube-api-access-kcc4w\") pod \"community-operators-77xtx\" (UID: \"7c0e2c11-212a-46ed-b72a-911bde29199a\") " pod="openshift-marketplace/community-operators-77xtx" Oct 08 18:13:22 crc kubenswrapper[4988]: I1008 18:13:22.845659 4988 patch_prober.go:28] interesting pod/router-default-5444994796-jjnkn container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 08 18:13:22 crc kubenswrapper[4988]: [-]has-synced failed: reason withheld Oct 08 18:13:22 crc kubenswrapper[4988]: [+]process-running ok Oct 08 18:13:22 crc kubenswrapper[4988]: healthz check failed Oct 08 18:13:22 crc kubenswrapper[4988]: I1008 18:13:22.845723 4988 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-jjnkn" podUID="ed79f1d0-1282-4be1-8e71-68a5320cdab1" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 08 18:13:22 crc kubenswrapper[4988]: I1008 18:13:22.883803 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-77xtx" Oct 08 18:13:22 crc kubenswrapper[4988]: I1008 18:13:22.901896 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-d9ks8" Oct 08 18:13:22 crc kubenswrapper[4988]: I1008 18:13:22.906587 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-4wh9p\" (UID: \"8a33d7bd-26bd-409b-93f1-5f151ebca347\") " pod="openshift-image-registry/image-registry-697d97f7c8-4wh9p" Oct 08 18:13:22 crc kubenswrapper[4988]: I1008 18:13:22.906616 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/15a52185-0755-45a6-b665-c1185d1996b4-utilities\") pod \"certified-operators-dmdqd\" (UID: \"15a52185-0755-45a6-b665-c1185d1996b4\") " pod="openshift-marketplace/certified-operators-dmdqd" Oct 08 18:13:22 crc kubenswrapper[4988]: I1008 18:13:22.906642 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n95d9\" (UniqueName: \"kubernetes.io/projected/15a52185-0755-45a6-b665-c1185d1996b4-kube-api-access-n95d9\") pod \"certified-operators-dmdqd\" (UID: \"15a52185-0755-45a6-b665-c1185d1996b4\") " pod="openshift-marketplace/certified-operators-dmdqd" Oct 08 18:13:22 crc kubenswrapper[4988]: I1008 18:13:22.906687 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/15a52185-0755-45a6-b665-c1185d1996b4-catalog-content\") pod \"certified-operators-dmdqd\" (UID: \"15a52185-0755-45a6-b665-c1185d1996b4\") " pod="openshift-marketplace/certified-operators-dmdqd" Oct 08 18:13:22 crc kubenswrapper[4988]: E1008 18:13:22.907077 4988 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-08 18:13:23.407053114 +0000 UTC m=+148.856895894 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-4wh9p" (UID: "8a33d7bd-26bd-409b-93f1-5f151ebca347") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 18:13:22 crc kubenswrapper[4988]: I1008 18:13:22.907310 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/15a52185-0755-45a6-b665-c1185d1996b4-utilities\") pod \"certified-operators-dmdqd\" (UID: \"15a52185-0755-45a6-b665-c1185d1996b4\") " pod="openshift-marketplace/certified-operators-dmdqd" Oct 08 18:13:22 crc kubenswrapper[4988]: I1008 18:13:22.907341 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/15a52185-0755-45a6-b665-c1185d1996b4-catalog-content\") pod \"certified-operators-dmdqd\" (UID: \"15a52185-0755-45a6-b665-c1185d1996b4\") " pod="openshift-marketplace/certified-operators-dmdqd" Oct 08 18:13:22 crc kubenswrapper[4988]: I1008 18:13:22.947169 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n95d9\" (UniqueName: \"kubernetes.io/projected/15a52185-0755-45a6-b665-c1185d1996b4-kube-api-access-n95d9\") pod \"certified-operators-dmdqd\" (UID: \"15a52185-0755-45a6-b665-c1185d1996b4\") " pod="openshift-marketplace/certified-operators-dmdqd" Oct 08 18:13:22 crc kubenswrapper[4988]: I1008 18:13:22.953993 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-lm95x"] Oct 08 18:13:22 crc kubenswrapper[4988]: I1008 18:13:22.955215 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-lm95x" Oct 08 18:13:22 crc kubenswrapper[4988]: I1008 18:13:22.992618 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-lm95x"] Oct 08 18:13:23 crc kubenswrapper[4988]: I1008 18:13:23.007502 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 18:13:23 crc kubenswrapper[4988]: I1008 18:13:23.007635 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 18:13:23 crc kubenswrapper[4988]: I1008 18:13:23.007668 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 18:13:23 crc kubenswrapper[4988]: I1008 18:13:23.007706 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-484hv\" (UniqueName: \"kubernetes.io/projected/619228a7-ec1c-4880-a3d5-9d26ac51afd9-kube-api-access-484hv\") pod \"community-operators-lm95x\" (UID: \"619228a7-ec1c-4880-a3d5-9d26ac51afd9\") " pod="openshift-marketplace/community-operators-lm95x" Oct 08 18:13:23 crc kubenswrapper[4988]: I1008 18:13:23.007731 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/619228a7-ec1c-4880-a3d5-9d26ac51afd9-utilities\") pod \"community-operators-lm95x\" (UID: \"619228a7-ec1c-4880-a3d5-9d26ac51afd9\") " pod="openshift-marketplace/community-operators-lm95x" Oct 08 18:13:23 crc kubenswrapper[4988]: I1008 18:13:23.007773 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/619228a7-ec1c-4880-a3d5-9d26ac51afd9-catalog-content\") pod \"community-operators-lm95x\" (UID: \"619228a7-ec1c-4880-a3d5-9d26ac51afd9\") " pod="openshift-marketplace/community-operators-lm95x" Oct 08 18:13:23 crc kubenswrapper[4988]: E1008 18:13:23.007927 4988 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 18:13:23.507907496 +0000 UTC m=+148.957750266 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 18:13:23 crc kubenswrapper[4988]: I1008 18:13:23.014082 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 18:13:23 crc kubenswrapper[4988]: I1008 18:13:23.024998 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 18:13:23 crc kubenswrapper[4988]: I1008 18:13:23.060707 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-dmdqd" Oct 08 18:13:23 crc kubenswrapper[4988]: I1008 18:13:23.109327 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/619228a7-ec1c-4880-a3d5-9d26ac51afd9-utilities\") pod \"community-operators-lm95x\" (UID: \"619228a7-ec1c-4880-a3d5-9d26ac51afd9\") " pod="openshift-marketplace/community-operators-lm95x" Oct 08 18:13:23 crc kubenswrapper[4988]: I1008 18:13:23.109418 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-4wh9p\" (UID: \"8a33d7bd-26bd-409b-93f1-5f151ebca347\") " pod="openshift-image-registry/image-registry-697d97f7c8-4wh9p" Oct 08 18:13:23 crc kubenswrapper[4988]: I1008 18:13:23.109451 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/619228a7-ec1c-4880-a3d5-9d26ac51afd9-catalog-content\") pod \"community-operators-lm95x\" (UID: \"619228a7-ec1c-4880-a3d5-9d26ac51afd9\") " pod="openshift-marketplace/community-operators-lm95x" Oct 08 18:13:23 crc kubenswrapper[4988]: I1008 18:13:23.109503 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 18:13:23 crc kubenswrapper[4988]: I1008 18:13:23.109542 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 18:13:23 crc kubenswrapper[4988]: I1008 18:13:23.109576 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-484hv\" (UniqueName: \"kubernetes.io/projected/619228a7-ec1c-4880-a3d5-9d26ac51afd9-kube-api-access-484hv\") pod \"community-operators-lm95x\" (UID: \"619228a7-ec1c-4880-a3d5-9d26ac51afd9\") " pod="openshift-marketplace/community-operators-lm95x" Oct 08 18:13:23 crc kubenswrapper[4988]: E1008 18:13:23.110950 4988 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-08 18:13:23.610927369 +0000 UTC m=+149.060770139 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-4wh9p" (UID: "8a33d7bd-26bd-409b-93f1-5f151ebca347") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 18:13:23 crc kubenswrapper[4988]: I1008 18:13:23.111248 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/619228a7-ec1c-4880-a3d5-9d26ac51afd9-catalog-content\") pod \"community-operators-lm95x\" (UID: \"619228a7-ec1c-4880-a3d5-9d26ac51afd9\") " pod="openshift-marketplace/community-operators-lm95x" Oct 08 18:13:23 crc kubenswrapper[4988]: I1008 18:13:23.111528 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/619228a7-ec1c-4880-a3d5-9d26ac51afd9-utilities\") pod \"community-operators-lm95x\" (UID: \"619228a7-ec1c-4880-a3d5-9d26ac51afd9\") " pod="openshift-marketplace/community-operators-lm95x" Oct 08 18:13:23 crc kubenswrapper[4988]: I1008 18:13:23.115717 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 18:13:23 crc kubenswrapper[4988]: I1008 18:13:23.115924 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 18:13:23 crc kubenswrapper[4988]: I1008 18:13:23.150161 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-hvx7q"] Oct 08 18:13:23 crc kubenswrapper[4988]: I1008 18:13:23.151186 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-hvx7q" Oct 08 18:13:23 crc kubenswrapper[4988]: I1008 18:13:23.158462 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-484hv\" (UniqueName: \"kubernetes.io/projected/619228a7-ec1c-4880-a3d5-9d26ac51afd9-kube-api-access-484hv\") pod \"community-operators-lm95x\" (UID: \"619228a7-ec1c-4880-a3d5-9d26ac51afd9\") " pod="openshift-marketplace/community-operators-lm95x" Oct 08 18:13:23 crc kubenswrapper[4988]: I1008 18:13:23.161332 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-hvx7q"] Oct 08 18:13:23 crc kubenswrapper[4988]: I1008 18:13:23.213698 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 18:13:23 crc kubenswrapper[4988]: I1008 18:13:23.213896 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4eb14139-7a10-4429-9834-378a89ea8255-catalog-content\") pod \"certified-operators-hvx7q\" (UID: \"4eb14139-7a10-4429-9834-378a89ea8255\") " pod="openshift-marketplace/certified-operators-hvx7q" Oct 08 18:13:23 crc kubenswrapper[4988]: I1008 18:13:23.213952 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4eb14139-7a10-4429-9834-378a89ea8255-utilities\") pod \"certified-operators-hvx7q\" (UID: \"4eb14139-7a10-4429-9834-378a89ea8255\") " pod="openshift-marketplace/certified-operators-hvx7q" Oct 08 18:13:23 crc kubenswrapper[4988]: I1008 18:13:23.213970 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rmkt7\" (UniqueName: \"kubernetes.io/projected/4eb14139-7a10-4429-9834-378a89ea8255-kube-api-access-rmkt7\") pod \"certified-operators-hvx7q\" (UID: \"4eb14139-7a10-4429-9834-378a89ea8255\") " pod="openshift-marketplace/certified-operators-hvx7q" Oct 08 18:13:23 crc kubenswrapper[4988]: E1008 18:13:23.214123 4988 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 18:13:23.714102298 +0000 UTC m=+149.163945068 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 18:13:23 crc kubenswrapper[4988]: I1008 18:13:23.269740 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 18:13:23 crc kubenswrapper[4988]: I1008 18:13:23.273744 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 18:13:23 crc kubenswrapper[4988]: I1008 18:13:23.294142 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 18:13:23 crc kubenswrapper[4988]: I1008 18:13:23.323034 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4eb14139-7a10-4429-9834-378a89ea8255-catalog-content\") pod \"certified-operators-hvx7q\" (UID: \"4eb14139-7a10-4429-9834-378a89ea8255\") " pod="openshift-marketplace/certified-operators-hvx7q" Oct 08 18:13:23 crc kubenswrapper[4988]: I1008 18:13:23.323091 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-4wh9p\" (UID: \"8a33d7bd-26bd-409b-93f1-5f151ebca347\") " pod="openshift-image-registry/image-registry-697d97f7c8-4wh9p" Oct 08 18:13:23 crc kubenswrapper[4988]: I1008 18:13:23.323128 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4eb14139-7a10-4429-9834-378a89ea8255-utilities\") pod \"certified-operators-hvx7q\" (UID: \"4eb14139-7a10-4429-9834-378a89ea8255\") " pod="openshift-marketplace/certified-operators-hvx7q" Oct 08 18:13:23 crc kubenswrapper[4988]: I1008 18:13:23.323147 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rmkt7\" (UniqueName: \"kubernetes.io/projected/4eb14139-7a10-4429-9834-378a89ea8255-kube-api-access-rmkt7\") pod \"certified-operators-hvx7q\" (UID: \"4eb14139-7a10-4429-9834-378a89ea8255\") " pod="openshift-marketplace/certified-operators-hvx7q" Oct 08 18:13:23 crc kubenswrapper[4988]: I1008 18:13:23.324350 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4eb14139-7a10-4429-9834-378a89ea8255-catalog-content\") pod \"certified-operators-hvx7q\" (UID: \"4eb14139-7a10-4429-9834-378a89ea8255\") " pod="openshift-marketplace/certified-operators-hvx7q" Oct 08 18:13:23 crc kubenswrapper[4988]: E1008 18:13:23.324649 4988 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-08 18:13:23.824636663 +0000 UTC m=+149.274479433 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-4wh9p" (UID: "8a33d7bd-26bd-409b-93f1-5f151ebca347") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 18:13:23 crc kubenswrapper[4988]: I1008 18:13:23.325037 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4eb14139-7a10-4429-9834-378a89ea8255-utilities\") pod \"certified-operators-hvx7q\" (UID: \"4eb14139-7a10-4429-9834-378a89ea8255\") " pod="openshift-marketplace/certified-operators-hvx7q" Oct 08 18:13:23 crc kubenswrapper[4988]: I1008 18:13:23.328638 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-lm95x" Oct 08 18:13:23 crc kubenswrapper[4988]: I1008 18:13:23.338996 4988 patch_prober.go:28] interesting pod/machine-config-daemon-nm54f container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 18:13:23 crc kubenswrapper[4988]: I1008 18:13:23.339056 4988 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 18:13:23 crc kubenswrapper[4988]: I1008 18:13:23.372412 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rmkt7\" (UniqueName: \"kubernetes.io/projected/4eb14139-7a10-4429-9834-378a89ea8255-kube-api-access-rmkt7\") pod \"certified-operators-hvx7q\" (UID: \"4eb14139-7a10-4429-9834-378a89ea8255\") " pod="openshift-marketplace/certified-operators-hvx7q" Oct 08 18:13:23 crc kubenswrapper[4988]: I1008 18:13:23.427217 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 18:13:23 crc kubenswrapper[4988]: E1008 18:13:23.427702 4988 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 18:13:23.927682498 +0000 UTC m=+149.377525268 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 18:13:23 crc kubenswrapper[4988]: I1008 18:13:23.504839 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-hvx7q" Oct 08 18:13:23 crc kubenswrapper[4988]: I1008 18:13:23.535824 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-4wh9p\" (UID: \"8a33d7bd-26bd-409b-93f1-5f151ebca347\") " pod="openshift-image-registry/image-registry-697d97f7c8-4wh9p" Oct 08 18:13:23 crc kubenswrapper[4988]: E1008 18:13:23.536246 4988 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-08 18:13:24.036230043 +0000 UTC m=+149.486072813 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-4wh9p" (UID: "8a33d7bd-26bd-409b-93f1-5f151ebca347") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 18:13:23 crc kubenswrapper[4988]: I1008 18:13:23.638143 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 18:13:23 crc kubenswrapper[4988]: E1008 18:13:23.638429 4988 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 18:13:24.138346771 +0000 UTC m=+149.588189541 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 18:13:23 crc kubenswrapper[4988]: I1008 18:13:23.638477 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-4wh9p\" (UID: \"8a33d7bd-26bd-409b-93f1-5f151ebca347\") " pod="openshift-image-registry/image-registry-697d97f7c8-4wh9p" Oct 08 18:13:23 crc kubenswrapper[4988]: E1008 18:13:23.638958 4988 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-08 18:13:24.138943919 +0000 UTC m=+149.588786889 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-4wh9p" (UID: "8a33d7bd-26bd-409b-93f1-5f151ebca347") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 18:13:23 crc kubenswrapper[4988]: I1008 18:13:23.661838 4988 generic.go:334] "Generic (PLEG): container finished" podID="c08e9af6-fc25-4c32-bd25-ac82aebb828e" containerID="313c57df12ab0a4da3188d7fc8368593246ef417dbcc4914e60937f26494b1b0" exitCode=0 Oct 08 18:13:23 crc kubenswrapper[4988]: I1008 18:13:23.662833 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29332440-qxt8r" event={"ID":"c08e9af6-fc25-4c32-bd25-ac82aebb828e","Type":"ContainerDied","Data":"313c57df12ab0a4da3188d7fc8368593246ef417dbcc4914e60937f26494b1b0"} Oct 08 18:13:23 crc kubenswrapper[4988]: I1008 18:13:23.742083 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 18:13:23 crc kubenswrapper[4988]: E1008 18:13:23.746620 4988 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 18:13:24.246580578 +0000 UTC m=+149.696423348 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 18:13:23 crc kubenswrapper[4988]: I1008 18:13:23.829258 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-77xtx"] Oct 08 18:13:23 crc kubenswrapper[4988]: I1008 18:13:23.844292 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-4wh9p\" (UID: \"8a33d7bd-26bd-409b-93f1-5f151ebca347\") " pod="openshift-image-registry/image-registry-697d97f7c8-4wh9p" Oct 08 18:13:23 crc kubenswrapper[4988]: E1008 18:13:23.844733 4988 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-08 18:13:24.34471665 +0000 UTC m=+149.794559420 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-4wh9p" (UID: "8a33d7bd-26bd-409b-93f1-5f151ebca347") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 18:13:23 crc kubenswrapper[4988]: I1008 18:13:23.897592 4988 patch_prober.go:28] interesting pod/router-default-5444994796-jjnkn container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 08 18:13:23 crc kubenswrapper[4988]: [-]has-synced failed: reason withheld Oct 08 18:13:23 crc kubenswrapper[4988]: [+]process-running ok Oct 08 18:13:23 crc kubenswrapper[4988]: healthz check failed Oct 08 18:13:23 crc kubenswrapper[4988]: I1008 18:13:23.897655 4988 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-jjnkn" podUID="ed79f1d0-1282-4be1-8e71-68a5320cdab1" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 08 18:13:23 crc kubenswrapper[4988]: I1008 18:13:23.951497 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 18:13:23 crc kubenswrapper[4988]: E1008 18:13:23.952092 4988 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 18:13:24.452069441 +0000 UTC m=+149.901912211 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 18:13:24 crc kubenswrapper[4988]: I1008 18:13:24.053187 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-4wh9p\" (UID: \"8a33d7bd-26bd-409b-93f1-5f151ebca347\") " pod="openshift-image-registry/image-registry-697d97f7c8-4wh9p" Oct 08 18:13:24 crc kubenswrapper[4988]: E1008 18:13:24.054039 4988 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-08 18:13:24.554020505 +0000 UTC m=+150.003863275 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-4wh9p" (UID: "8a33d7bd-26bd-409b-93f1-5f151ebca347") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 18:13:24 crc kubenswrapper[4988]: I1008 18:13:24.062035 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-dmdqd"] Oct 08 18:13:24 crc kubenswrapper[4988]: W1008 18:13:24.083643 4988 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod15a52185_0755_45a6_b665_c1185d1996b4.slice/crio-75e9f6bd560703b6df4bfd0f409c13086c917285458c72edaa4739dabcf9a986 WatchSource:0}: Error finding container 75e9f6bd560703b6df4bfd0f409c13086c917285458c72edaa4739dabcf9a986: Status 404 returned error can't find the container with id 75e9f6bd560703b6df4bfd0f409c13086c917285458c72edaa4739dabcf9a986 Oct 08 18:13:24 crc kubenswrapper[4988]: I1008 18:13:24.155428 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 18:13:24 crc kubenswrapper[4988]: E1008 18:13:24.155692 4988 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 18:13:24.655668709 +0000 UTC m=+150.105511479 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 18:13:24 crc kubenswrapper[4988]: I1008 18:13:24.257009 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-4wh9p\" (UID: \"8a33d7bd-26bd-409b-93f1-5f151ebca347\") " pod="openshift-image-registry/image-registry-697d97f7c8-4wh9p" Oct 08 18:13:24 crc kubenswrapper[4988]: E1008 18:13:24.257972 4988 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-08 18:13:24.757952582 +0000 UTC m=+150.207795352 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-4wh9p" (UID: "8a33d7bd-26bd-409b-93f1-5f151ebca347") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 18:13:24 crc kubenswrapper[4988]: I1008 18:13:24.288271 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-lm95x"] Oct 08 18:13:24 crc kubenswrapper[4988]: I1008 18:13:24.358351 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 18:13:24 crc kubenswrapper[4988]: E1008 18:13:24.358625 4988 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 18:13:24.858607608 +0000 UTC m=+150.308450378 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 18:13:24 crc kubenswrapper[4988]: I1008 18:13:24.459745 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-4wh9p\" (UID: \"8a33d7bd-26bd-409b-93f1-5f151ebca347\") " pod="openshift-image-registry/image-registry-697d97f7c8-4wh9p" Oct 08 18:13:24 crc kubenswrapper[4988]: E1008 18:13:24.460154 4988 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-08 18:13:24.960139959 +0000 UTC m=+150.409982729 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-4wh9p" (UID: "8a33d7bd-26bd-409b-93f1-5f151ebca347") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 18:13:24 crc kubenswrapper[4988]: I1008 18:13:24.535162 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-kmg58"] Oct 08 18:13:24 crc kubenswrapper[4988]: I1008 18:13:24.538597 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-kmg58" Oct 08 18:13:24 crc kubenswrapper[4988]: I1008 18:13:24.540853 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Oct 08 18:13:24 crc kubenswrapper[4988]: I1008 18:13:24.547520 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-kmg58"] Oct 08 18:13:24 crc kubenswrapper[4988]: I1008 18:13:24.562110 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 18:13:24 crc kubenswrapper[4988]: E1008 18:13:24.562318 4988 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 18:13:25.062279197 +0000 UTC m=+150.512121967 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 18:13:24 crc kubenswrapper[4988]: I1008 18:13:24.562420 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/38d2198b-2aab-47d2-aa2e-93436ff69625-utilities\") pod \"redhat-marketplace-kmg58\" (UID: \"38d2198b-2aab-47d2-aa2e-93436ff69625\") " pod="openshift-marketplace/redhat-marketplace-kmg58" Oct 08 18:13:24 crc kubenswrapper[4988]: I1008 18:13:24.562464 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/38d2198b-2aab-47d2-aa2e-93436ff69625-catalog-content\") pod \"redhat-marketplace-kmg58\" (UID: \"38d2198b-2aab-47d2-aa2e-93436ff69625\") " pod="openshift-marketplace/redhat-marketplace-kmg58" Oct 08 18:13:24 crc kubenswrapper[4988]: I1008 18:13:24.562524 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-4wh9p\" (UID: \"8a33d7bd-26bd-409b-93f1-5f151ebca347\") " pod="openshift-image-registry/image-registry-697d97f7c8-4wh9p" Oct 08 18:13:24 crc kubenswrapper[4988]: I1008 18:13:24.562601 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6d8gx\" (UniqueName: \"kubernetes.io/projected/38d2198b-2aab-47d2-aa2e-93436ff69625-kube-api-access-6d8gx\") pod \"redhat-marketplace-kmg58\" (UID: \"38d2198b-2aab-47d2-aa2e-93436ff69625\") " pod="openshift-marketplace/redhat-marketplace-kmg58" Oct 08 18:13:24 crc kubenswrapper[4988]: E1008 18:13:24.562908 4988 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-08 18:13:25.062897316 +0000 UTC m=+150.512740086 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-4wh9p" (UID: "8a33d7bd-26bd-409b-93f1-5f151ebca347") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 18:13:24 crc kubenswrapper[4988]: I1008 18:13:24.622724 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-hvx7q"] Oct 08 18:13:24 crc kubenswrapper[4988]: I1008 18:13:24.666044 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 18:13:24 crc kubenswrapper[4988]: E1008 18:13:24.670213 4988 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 18:13:25.170181845 +0000 UTC m=+150.620024615 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 18:13:24 crc kubenswrapper[4988]: I1008 18:13:24.674047 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/38d2198b-2aab-47d2-aa2e-93436ff69625-utilities\") pod \"redhat-marketplace-kmg58\" (UID: \"38d2198b-2aab-47d2-aa2e-93436ff69625\") " pod="openshift-marketplace/redhat-marketplace-kmg58" Oct 08 18:13:24 crc kubenswrapper[4988]: I1008 18:13:24.674100 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/38d2198b-2aab-47d2-aa2e-93436ff69625-catalog-content\") pod \"redhat-marketplace-kmg58\" (UID: \"38d2198b-2aab-47d2-aa2e-93436ff69625\") " pod="openshift-marketplace/redhat-marketplace-kmg58" Oct 08 18:13:24 crc kubenswrapper[4988]: I1008 18:13:24.674204 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-4wh9p\" (UID: \"8a33d7bd-26bd-409b-93f1-5f151ebca347\") " pod="openshift-image-registry/image-registry-697d97f7c8-4wh9p" Oct 08 18:13:24 crc kubenswrapper[4988]: I1008 18:13:24.674319 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6d8gx\" (UniqueName: \"kubernetes.io/projected/38d2198b-2aab-47d2-aa2e-93436ff69625-kube-api-access-6d8gx\") pod \"redhat-marketplace-kmg58\" (UID: \"38d2198b-2aab-47d2-aa2e-93436ff69625\") " pod="openshift-marketplace/redhat-marketplace-kmg58" Oct 08 18:13:24 crc kubenswrapper[4988]: I1008 18:13:24.674991 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/38d2198b-2aab-47d2-aa2e-93436ff69625-utilities\") pod \"redhat-marketplace-kmg58\" (UID: \"38d2198b-2aab-47d2-aa2e-93436ff69625\") " pod="openshift-marketplace/redhat-marketplace-kmg58" Oct 08 18:13:24 crc kubenswrapper[4988]: I1008 18:13:24.675258 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/38d2198b-2aab-47d2-aa2e-93436ff69625-catalog-content\") pod \"redhat-marketplace-kmg58\" (UID: \"38d2198b-2aab-47d2-aa2e-93436ff69625\") " pod="openshift-marketplace/redhat-marketplace-kmg58" Oct 08 18:13:24 crc kubenswrapper[4988]: E1008 18:13:24.675666 4988 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-08 18:13:25.175647414 +0000 UTC m=+150.625490184 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-4wh9p" (UID: "8a33d7bd-26bd-409b-93f1-5f151ebca347") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 18:13:24 crc kubenswrapper[4988]: I1008 18:13:24.707587 4988 generic.go:334] "Generic (PLEG): container finished" podID="15a52185-0755-45a6-b665-c1185d1996b4" containerID="3713f5d1afcd9846c811ff79adf171b4d51e6c8706b9d87fc2b59eb4d8f3f964" exitCode=0 Oct 08 18:13:24 crc kubenswrapper[4988]: I1008 18:13:24.707995 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dmdqd" event={"ID":"15a52185-0755-45a6-b665-c1185d1996b4","Type":"ContainerDied","Data":"3713f5d1afcd9846c811ff79adf171b4d51e6c8706b9d87fc2b59eb4d8f3f964"} Oct 08 18:13:24 crc kubenswrapper[4988]: I1008 18:13:24.708037 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dmdqd" event={"ID":"15a52185-0755-45a6-b665-c1185d1996b4","Type":"ContainerStarted","Data":"75e9f6bd560703b6df4bfd0f409c13086c917285458c72edaa4739dabcf9a986"} Oct 08 18:13:24 crc kubenswrapper[4988]: I1008 18:13:24.709329 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6d8gx\" (UniqueName: \"kubernetes.io/projected/38d2198b-2aab-47d2-aa2e-93436ff69625-kube-api-access-6d8gx\") pod \"redhat-marketplace-kmg58\" (UID: \"38d2198b-2aab-47d2-aa2e-93436ff69625\") " pod="openshift-marketplace/redhat-marketplace-kmg58" Oct 08 18:13:24 crc kubenswrapper[4988]: I1008 18:13:24.728243 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-gl7lr" event={"ID":"9cb89712-c3bf-4ed9-87f8-8124a55f5f7e","Type":"ContainerStarted","Data":"60f2aeb07c4ef094912d5fb43775e99cf0da178ad659a633579bdd9c200f79be"} Oct 08 18:13:24 crc kubenswrapper[4988]: I1008 18:13:24.730585 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-hvx7q" event={"ID":"4eb14139-7a10-4429-9834-378a89ea8255","Type":"ContainerStarted","Data":"20bd3fe0a116cbade50d4021346ff01b51b7443d1d530257122b6892173cbd0e"} Oct 08 18:13:24 crc kubenswrapper[4988]: I1008 18:13:24.733028 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"f27cf677fe2c3c95b83233893f215447f4025fc697eed1b2e920c12f75d89d94"} Oct 08 18:13:24 crc kubenswrapper[4988]: I1008 18:13:24.736767 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-lm95x" event={"ID":"619228a7-ec1c-4880-a3d5-9d26ac51afd9","Type":"ContainerStarted","Data":"17c3fd14341d22847cb79f4347c7968d7f1d1bb4b7299641d10e5782790b0e7a"} Oct 08 18:13:24 crc kubenswrapper[4988]: I1008 18:13:24.742137 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"15548cddcdc609ec8ee2d5e8b0d38eb145e9aca3da1ac6250f810b9ab076de2d"} Oct 08 18:13:24 crc kubenswrapper[4988]: I1008 18:13:24.747753 4988 generic.go:334] "Generic (PLEG): container finished" podID="7c0e2c11-212a-46ed-b72a-911bde29199a" containerID="aaf1ecf558697a87bf3a10117a2784a64dff7aae5eb9a2cf38599f88ef41a71b" exitCode=0 Oct 08 18:13:24 crc kubenswrapper[4988]: I1008 18:13:24.747831 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-77xtx" event={"ID":"7c0e2c11-212a-46ed-b72a-911bde29199a","Type":"ContainerDied","Data":"aaf1ecf558697a87bf3a10117a2784a64dff7aae5eb9a2cf38599f88ef41a71b"} Oct 08 18:13:24 crc kubenswrapper[4988]: I1008 18:13:24.747908 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-77xtx" event={"ID":"7c0e2c11-212a-46ed-b72a-911bde29199a","Type":"ContainerStarted","Data":"220ad062d243aa03d816f4a27f313ef1cae071aab188d8f04985110691f8023f"} Oct 08 18:13:24 crc kubenswrapper[4988]: I1008 18:13:24.751413 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"406179bd62991b7c7818295785a090c53a60ce8b8883dce2baac2f42b48b0408"} Oct 08 18:13:24 crc kubenswrapper[4988]: I1008 18:13:24.752994 4988 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 08 18:13:24 crc kubenswrapper[4988]: I1008 18:13:24.775360 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 18:13:24 crc kubenswrapper[4988]: E1008 18:13:24.775675 4988 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 18:13:25.2756369 +0000 UTC m=+150.725479670 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 18:13:24 crc kubenswrapper[4988]: I1008 18:13:24.775945 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-4wh9p\" (UID: \"8a33d7bd-26bd-409b-93f1-5f151ebca347\") " pod="openshift-image-registry/image-registry-697d97f7c8-4wh9p" Oct 08 18:13:24 crc kubenswrapper[4988]: E1008 18:13:24.777176 4988 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-08 18:13:25.277155615 +0000 UTC m=+150.726998395 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-4wh9p" (UID: "8a33d7bd-26bd-409b-93f1-5f151ebca347") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 18:13:24 crc kubenswrapper[4988]: I1008 18:13:24.827728 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-nsc65" Oct 08 18:13:24 crc kubenswrapper[4988]: I1008 18:13:24.849647 4988 patch_prober.go:28] interesting pod/router-default-5444994796-jjnkn container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 08 18:13:24 crc kubenswrapper[4988]: [-]has-synced failed: reason withheld Oct 08 18:13:24 crc kubenswrapper[4988]: [+]process-running ok Oct 08 18:13:24 crc kubenswrapper[4988]: healthz check failed Oct 08 18:13:24 crc kubenswrapper[4988]: I1008 18:13:24.849733 4988 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-jjnkn" podUID="ed79f1d0-1282-4be1-8e71-68a5320cdab1" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 08 18:13:24 crc kubenswrapper[4988]: I1008 18:13:24.881960 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 18:13:24 crc kubenswrapper[4988]: E1008 18:13:24.883299 4988 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 18:13:25.383270459 +0000 UTC m=+150.833113239 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 18:13:24 crc kubenswrapper[4988]: I1008 18:13:24.924404 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-w4qxm"] Oct 08 18:13:24 crc kubenswrapper[4988]: I1008 18:13:24.925417 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-w4qxm" Oct 08 18:13:24 crc kubenswrapper[4988]: I1008 18:13:24.929877 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-kmg58" Oct 08 18:13:24 crc kubenswrapper[4988]: I1008 18:13:24.960808 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-w4qxm"] Oct 08 18:13:24 crc kubenswrapper[4988]: I1008 18:13:24.987812 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4jxl8\" (UniqueName: \"kubernetes.io/projected/56027983-5f79-4555-abb9-4ab9f13cd39f-kube-api-access-4jxl8\") pod \"redhat-marketplace-w4qxm\" (UID: \"56027983-5f79-4555-abb9-4ab9f13cd39f\") " pod="openshift-marketplace/redhat-marketplace-w4qxm" Oct 08 18:13:24 crc kubenswrapper[4988]: I1008 18:13:24.987885 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-4wh9p\" (UID: \"8a33d7bd-26bd-409b-93f1-5f151ebca347\") " pod="openshift-image-registry/image-registry-697d97f7c8-4wh9p" Oct 08 18:13:24 crc kubenswrapper[4988]: I1008 18:13:24.987935 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/56027983-5f79-4555-abb9-4ab9f13cd39f-utilities\") pod \"redhat-marketplace-w4qxm\" (UID: \"56027983-5f79-4555-abb9-4ab9f13cd39f\") " pod="openshift-marketplace/redhat-marketplace-w4qxm" Oct 08 18:13:24 crc kubenswrapper[4988]: I1008 18:13:24.987969 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/56027983-5f79-4555-abb9-4ab9f13cd39f-catalog-content\") pod \"redhat-marketplace-w4qxm\" (UID: \"56027983-5f79-4555-abb9-4ab9f13cd39f\") " pod="openshift-marketplace/redhat-marketplace-w4qxm" Oct 08 18:13:24 crc kubenswrapper[4988]: E1008 18:13:24.988497 4988 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-08 18:13:25.488472467 +0000 UTC m=+150.938315237 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-4wh9p" (UID: "8a33d7bd-26bd-409b-93f1-5f151ebca347") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 18:13:25 crc kubenswrapper[4988]: I1008 18:13:25.089206 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 18:13:25 crc kubenswrapper[4988]: E1008 18:13:25.089462 4988 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 18:13:25.589421321 +0000 UTC m=+151.039264091 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 18:13:25 crc kubenswrapper[4988]: I1008 18:13:25.089858 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/56027983-5f79-4555-abb9-4ab9f13cd39f-catalog-content\") pod \"redhat-marketplace-w4qxm\" (UID: \"56027983-5f79-4555-abb9-4ab9f13cd39f\") " pod="openshift-marketplace/redhat-marketplace-w4qxm" Oct 08 18:13:25 crc kubenswrapper[4988]: I1008 18:13:25.089907 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4jxl8\" (UniqueName: \"kubernetes.io/projected/56027983-5f79-4555-abb9-4ab9f13cd39f-kube-api-access-4jxl8\") pod \"redhat-marketplace-w4qxm\" (UID: \"56027983-5f79-4555-abb9-4ab9f13cd39f\") " pod="openshift-marketplace/redhat-marketplace-w4qxm" Oct 08 18:13:25 crc kubenswrapper[4988]: I1008 18:13:25.089941 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-4wh9p\" (UID: \"8a33d7bd-26bd-409b-93f1-5f151ebca347\") " pod="openshift-image-registry/image-registry-697d97f7c8-4wh9p" Oct 08 18:13:25 crc kubenswrapper[4988]: I1008 18:13:25.089988 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/56027983-5f79-4555-abb9-4ab9f13cd39f-utilities\") pod \"redhat-marketplace-w4qxm\" (UID: \"56027983-5f79-4555-abb9-4ab9f13cd39f\") " pod="openshift-marketplace/redhat-marketplace-w4qxm" Oct 08 18:13:25 crc kubenswrapper[4988]: I1008 18:13:25.090464 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/56027983-5f79-4555-abb9-4ab9f13cd39f-utilities\") pod \"redhat-marketplace-w4qxm\" (UID: \"56027983-5f79-4555-abb9-4ab9f13cd39f\") " pod="openshift-marketplace/redhat-marketplace-w4qxm" Oct 08 18:13:25 crc kubenswrapper[4988]: I1008 18:13:25.090693 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/56027983-5f79-4555-abb9-4ab9f13cd39f-catalog-content\") pod \"redhat-marketplace-w4qxm\" (UID: \"56027983-5f79-4555-abb9-4ab9f13cd39f\") " pod="openshift-marketplace/redhat-marketplace-w4qxm" Oct 08 18:13:25 crc kubenswrapper[4988]: E1008 18:13:25.090809 4988 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-08 18:13:25.590802092 +0000 UTC m=+151.040644862 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-4wh9p" (UID: "8a33d7bd-26bd-409b-93f1-5f151ebca347") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 18:13:25 crc kubenswrapper[4988]: I1008 18:13:25.129772 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29332440-qxt8r" Oct 08 18:13:25 crc kubenswrapper[4988]: I1008 18:13:25.169501 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4jxl8\" (UniqueName: \"kubernetes.io/projected/56027983-5f79-4555-abb9-4ab9f13cd39f-kube-api-access-4jxl8\") pod \"redhat-marketplace-w4qxm\" (UID: \"56027983-5f79-4555-abb9-4ab9f13cd39f\") " pod="openshift-marketplace/redhat-marketplace-w4qxm" Oct 08 18:13:25 crc kubenswrapper[4988]: I1008 18:13:25.191658 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6jts7\" (UniqueName: \"kubernetes.io/projected/c08e9af6-fc25-4c32-bd25-ac82aebb828e-kube-api-access-6jts7\") pod \"c08e9af6-fc25-4c32-bd25-ac82aebb828e\" (UID: \"c08e9af6-fc25-4c32-bd25-ac82aebb828e\") " Oct 08 18:13:25 crc kubenswrapper[4988]: I1008 18:13:25.191742 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c08e9af6-fc25-4c32-bd25-ac82aebb828e-config-volume\") pod \"c08e9af6-fc25-4c32-bd25-ac82aebb828e\" (UID: \"c08e9af6-fc25-4c32-bd25-ac82aebb828e\") " Oct 08 18:13:25 crc kubenswrapper[4988]: I1008 18:13:25.191892 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 18:13:25 crc kubenswrapper[4988]: I1008 18:13:25.192061 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/c08e9af6-fc25-4c32-bd25-ac82aebb828e-secret-volume\") pod \"c08e9af6-fc25-4c32-bd25-ac82aebb828e\" (UID: \"c08e9af6-fc25-4c32-bd25-ac82aebb828e\") " Oct 08 18:13:25 crc kubenswrapper[4988]: E1008 18:13:25.192351 4988 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 18:13:25.692312432 +0000 UTC m=+151.142155202 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 18:13:25 crc kubenswrapper[4988]: I1008 18:13:25.192976 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c08e9af6-fc25-4c32-bd25-ac82aebb828e-config-volume" (OuterVolumeSpecName: "config-volume") pod "c08e9af6-fc25-4c32-bd25-ac82aebb828e" (UID: "c08e9af6-fc25-4c32-bd25-ac82aebb828e"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 18:13:25 crc kubenswrapper[4988]: I1008 18:13:25.203987 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c08e9af6-fc25-4c32-bd25-ac82aebb828e-kube-api-access-6jts7" (OuterVolumeSpecName: "kube-api-access-6jts7") pod "c08e9af6-fc25-4c32-bd25-ac82aebb828e" (UID: "c08e9af6-fc25-4c32-bd25-ac82aebb828e"). InnerVolumeSpecName "kube-api-access-6jts7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:13:25 crc kubenswrapper[4988]: I1008 18:13:25.207179 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-kmg58"] Oct 08 18:13:25 crc kubenswrapper[4988]: I1008 18:13:25.211749 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c08e9af6-fc25-4c32-bd25-ac82aebb828e-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "c08e9af6-fc25-4c32-bd25-ac82aebb828e" (UID: "c08e9af6-fc25-4c32-bd25-ac82aebb828e"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:13:25 crc kubenswrapper[4988]: W1008 18:13:25.214665 4988 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod38d2198b_2aab_47d2_aa2e_93436ff69625.slice/crio-1b82e7dfedfd0b2564a4226267eebbda9537214b654b88ea144c09b1cc87d46c WatchSource:0}: Error finding container 1b82e7dfedfd0b2564a4226267eebbda9537214b654b88ea144c09b1cc87d46c: Status 404 returned error can't find the container with id 1b82e7dfedfd0b2564a4226267eebbda9537214b654b88ea144c09b1cc87d46c Oct 08 18:13:25 crc kubenswrapper[4988]: I1008 18:13:25.257254 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-w4qxm" Oct 08 18:13:25 crc kubenswrapper[4988]: I1008 18:13:25.296315 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-4wh9p\" (UID: \"8a33d7bd-26bd-409b-93f1-5f151ebca347\") " pod="openshift-image-registry/image-registry-697d97f7c8-4wh9p" Oct 08 18:13:25 crc kubenswrapper[4988]: E1008 18:13:25.296538 4988 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-08 18:13:25.796509391 +0000 UTC m=+151.246352161 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-4wh9p" (UID: "8a33d7bd-26bd-409b-93f1-5f151ebca347") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 18:13:25 crc kubenswrapper[4988]: I1008 18:13:25.297164 4988 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/c08e9af6-fc25-4c32-bd25-ac82aebb828e-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 08 18:13:25 crc kubenswrapper[4988]: I1008 18:13:25.297183 4988 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6jts7\" (UniqueName: \"kubernetes.io/projected/c08e9af6-fc25-4c32-bd25-ac82aebb828e-kube-api-access-6jts7\") on node \"crc\" DevicePath \"\"" Oct 08 18:13:25 crc kubenswrapper[4988]: I1008 18:13:25.297213 4988 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c08e9af6-fc25-4c32-bd25-ac82aebb828e-config-volume\") on node \"crc\" DevicePath \"\"" Oct 08 18:13:25 crc kubenswrapper[4988]: I1008 18:13:25.345511 4988 plugin_watcher.go:194] "Adding socket path or updating timestamp to desired state cache" path="/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock" Oct 08 18:13:25 crc kubenswrapper[4988]: I1008 18:13:25.398072 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 18:13:25 crc kubenswrapper[4988]: E1008 18:13:25.398573 4988 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 18:13:25.898549487 +0000 UTC m=+151.348392257 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 18:13:25 crc kubenswrapper[4988]: I1008 18:13:25.402072 4988 patch_prober.go:28] interesting pod/downloads-7954f5f757-qcgmb container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.13:8080/\": dial tcp 10.217.0.13:8080: connect: connection refused" start-of-body= Oct 08 18:13:25 crc kubenswrapper[4988]: I1008 18:13:25.402133 4988 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-qcgmb" podUID="0bc2144f-4dbe-4dc1-a07a-b764d8582324" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.13:8080/\": dial tcp 10.217.0.13:8080: connect: connection refused" Oct 08 18:13:25 crc kubenswrapper[4988]: I1008 18:13:25.404982 4988 patch_prober.go:28] interesting pod/downloads-7954f5f757-qcgmb container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.13:8080/\": dial tcp 10.217.0.13:8080: connect: connection refused" start-of-body= Oct 08 18:13:25 crc kubenswrapper[4988]: I1008 18:13:25.405037 4988 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-qcgmb" podUID="0bc2144f-4dbe-4dc1-a07a-b764d8582324" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.13:8080/\": dial tcp 10.217.0.13:8080: connect: connection refused" Oct 08 18:13:25 crc kubenswrapper[4988]: I1008 18:13:25.411933 4988 reconciler.go:161] "OperationExecutor.RegisterPlugin started" plugin={"SocketPath":"/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock","Timestamp":"2025-10-08T18:13:25.345540671Z","Handler":null,"Name":""} Oct 08 18:13:25 crc kubenswrapper[4988]: I1008 18:13:25.446475 4988 csi_plugin.go:100] kubernetes.io/csi: Trying to validate a new CSI Driver with name: kubevirt.io.hostpath-provisioner endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock versions: 1.0.0 Oct 08 18:13:25 crc kubenswrapper[4988]: I1008 18:13:25.446898 4988 csi_plugin.go:113] kubernetes.io/csi: Register new plugin with name: kubevirt.io.hostpath-provisioner at endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock Oct 08 18:13:25 crc kubenswrapper[4988]: I1008 18:13:25.472104 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console-operator/console-operator-58897d9998-rt48k" Oct 08 18:13:25 crc kubenswrapper[4988]: I1008 18:13:25.502865 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-4wh9p\" (UID: \"8a33d7bd-26bd-409b-93f1-5f151ebca347\") " pod="openshift-image-registry/image-registry-697d97f7c8-4wh9p" Oct 08 18:13:25 crc kubenswrapper[4988]: I1008 18:13:25.510982 4988 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Oct 08 18:13:25 crc kubenswrapper[4988]: I1008 18:13:25.511028 4988 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-4wh9p\" (UID: \"8a33d7bd-26bd-409b-93f1-5f151ebca347\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount\"" pod="openshift-image-registry/image-registry-697d97f7c8-4wh9p" Oct 08 18:13:25 crc kubenswrapper[4988]: I1008 18:13:25.581742 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-w4qxm"] Oct 08 18:13:25 crc kubenswrapper[4988]: I1008 18:13:25.642804 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-4wh9p\" (UID: \"8a33d7bd-26bd-409b-93f1-5f151ebca347\") " pod="openshift-image-registry/image-registry-697d97f7c8-4wh9p" Oct 08 18:13:25 crc kubenswrapper[4988]: W1008 18:13:25.655059 4988 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod56027983_5f79_4555_abb9_4ab9f13cd39f.slice/crio-52b79f26119c72dd510a0f961e5d09437cc12d0a9fb8fc02d8339db6f24ec76b WatchSource:0}: Error finding container 52b79f26119c72dd510a0f961e5d09437cc12d0a9fb8fc02d8339db6f24ec76b: Status 404 returned error can't find the container with id 52b79f26119c72dd510a0f961e5d09437cc12d0a9fb8fc02d8339db6f24ec76b Oct 08 18:13:25 crc kubenswrapper[4988]: I1008 18:13:25.706453 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 18:13:25 crc kubenswrapper[4988]: I1008 18:13:25.716497 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Oct 08 18:13:25 crc kubenswrapper[4988]: I1008 18:13:25.764813 4988 generic.go:334] "Generic (PLEG): container finished" podID="38d2198b-2aab-47d2-aa2e-93436ff69625" containerID="33110da129142b21ee73fce202569b578ade65662103418e94513a5f6a8b5cfc" exitCode=0 Oct 08 18:13:25 crc kubenswrapper[4988]: I1008 18:13:25.765840 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-kmg58" event={"ID":"38d2198b-2aab-47d2-aa2e-93436ff69625","Type":"ContainerDied","Data":"33110da129142b21ee73fce202569b578ade65662103418e94513a5f6a8b5cfc"} Oct 08 18:13:25 crc kubenswrapper[4988]: I1008 18:13:25.765880 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-kmg58" event={"ID":"38d2198b-2aab-47d2-aa2e-93436ff69625","Type":"ContainerStarted","Data":"1b82e7dfedfd0b2564a4226267eebbda9537214b654b88ea144c09b1cc87d46c"} Oct 08 18:13:25 crc kubenswrapper[4988]: I1008 18:13:25.769771 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"9034635d4bef384e56e5e070c841313da89988911a756c4547dacfa59f3f2cda"} Oct 08 18:13:25 crc kubenswrapper[4988]: I1008 18:13:25.770482 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 18:13:25 crc kubenswrapper[4988]: I1008 18:13:25.772904 4988 generic.go:334] "Generic (PLEG): container finished" podID="4eb14139-7a10-4429-9834-378a89ea8255" containerID="b08fb6dd2b021eb06be9c6729d730ae01f09f7e84a01463406306bae12ccee0d" exitCode=0 Oct 08 18:13:25 crc kubenswrapper[4988]: I1008 18:13:25.772992 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-hvx7q" event={"ID":"4eb14139-7a10-4429-9834-378a89ea8255","Type":"ContainerDied","Data":"b08fb6dd2b021eb06be9c6729d730ae01f09f7e84a01463406306bae12ccee0d"} Oct 08 18:13:25 crc kubenswrapper[4988]: I1008 18:13:25.779999 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-gl7lr" event={"ID":"9cb89712-c3bf-4ed9-87f8-8124a55f5f7e","Type":"ContainerStarted","Data":"8c8eb671bcd9accf61f909cc238bda63279662abf16a7786373f962fc8ed9e92"} Oct 08 18:13:25 crc kubenswrapper[4988]: I1008 18:13:25.780045 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-gl7lr" event={"ID":"9cb89712-c3bf-4ed9-87f8-8124a55f5f7e","Type":"ContainerStarted","Data":"9b7dd41a9f0706e6a834fe40a194695f16f5166e2ad22b117c1336c6dca62c7d"} Oct 08 18:13:25 crc kubenswrapper[4988]: I1008 18:13:25.781667 4988 generic.go:334] "Generic (PLEG): container finished" podID="619228a7-ec1c-4880-a3d5-9d26ac51afd9" containerID="2e4478d7c154278d486e98dbbd0498fd59742508545a14684e692ef91d6c0c0b" exitCode=0 Oct 08 18:13:25 crc kubenswrapper[4988]: I1008 18:13:25.781750 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-lm95x" event={"ID":"619228a7-ec1c-4880-a3d5-9d26ac51afd9","Type":"ContainerDied","Data":"2e4478d7c154278d486e98dbbd0498fd59742508545a14684e692ef91d6c0c0b"} Oct 08 18:13:25 crc kubenswrapper[4988]: I1008 18:13:25.797184 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"d0edd574d637729c9f038525c5d470a0d59a9d4f69edce1078d68edc29cf4164"} Oct 08 18:13:25 crc kubenswrapper[4988]: I1008 18:13:25.799856 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29332440-qxt8r" event={"ID":"c08e9af6-fc25-4c32-bd25-ac82aebb828e","Type":"ContainerDied","Data":"e00740e7d2a3e5e6adb6acba4611aa1779dc5931cb73f118f64d0e3152caad9c"} Oct 08 18:13:25 crc kubenswrapper[4988]: I1008 18:13:25.799900 4988 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e00740e7d2a3e5e6adb6acba4611aa1779dc5931cb73f118f64d0e3152caad9c" Oct 08 18:13:25 crc kubenswrapper[4988]: I1008 18:13:25.799995 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29332440-qxt8r" Oct 08 18:13:25 crc kubenswrapper[4988]: I1008 18:13:25.804084 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"2b535d578269fc9cfa0c3befadc603d5b9acaa1fd8c24218b471a957aaccc2ed"} Oct 08 18:13:25 crc kubenswrapper[4988]: I1008 18:13:25.805717 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-w4qxm" event={"ID":"56027983-5f79-4555-abb9-4ab9f13cd39f","Type":"ContainerStarted","Data":"52b79f26119c72dd510a0f961e5d09437cc12d0a9fb8fc02d8339db6f24ec76b"} Oct 08 18:13:25 crc kubenswrapper[4988]: I1008 18:13:25.837322 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Oct 08 18:13:25 crc kubenswrapper[4988]: E1008 18:13:25.837629 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c08e9af6-fc25-4c32-bd25-ac82aebb828e" containerName="collect-profiles" Oct 08 18:13:25 crc kubenswrapper[4988]: I1008 18:13:25.837666 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="c08e9af6-fc25-4c32-bd25-ac82aebb828e" containerName="collect-profiles" Oct 08 18:13:25 crc kubenswrapper[4988]: I1008 18:13:25.837821 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="c08e9af6-fc25-4c32-bd25-ac82aebb828e" containerName="collect-profiles" Oct 08 18:13:25 crc kubenswrapper[4988]: I1008 18:13:25.838402 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 08 18:13:25 crc kubenswrapper[4988]: I1008 18:13:25.842255 4988 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager"/"kube-root-ca.crt" Oct 08 18:13:25 crc kubenswrapper[4988]: I1008 18:13:25.842543 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager"/"installer-sa-dockercfg-kjl2n" Oct 08 18:13:25 crc kubenswrapper[4988]: I1008 18:13:25.849011 4988 patch_prober.go:28] interesting pod/router-default-5444994796-jjnkn container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 08 18:13:25 crc kubenswrapper[4988]: [-]has-synced failed: reason withheld Oct 08 18:13:25 crc kubenswrapper[4988]: [+]process-running ok Oct 08 18:13:25 crc kubenswrapper[4988]: healthz check failed Oct 08 18:13:25 crc kubenswrapper[4988]: I1008 18:13:25.849079 4988 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-jjnkn" podUID="ed79f1d0-1282-4be1-8e71-68a5320cdab1" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 08 18:13:25 crc kubenswrapper[4988]: I1008 18:13:25.856235 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Oct 08 18:13:25 crc kubenswrapper[4988]: I1008 18:13:25.898424 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="hostpath-provisioner/csi-hostpathplugin-gl7lr" podStartSLOduration=11.898402275 podStartE2EDuration="11.898402275s" podCreationTimestamp="2025-10-08 18:13:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 18:13:25.871399828 +0000 UTC m=+151.321242618" watchObservedRunningTime="2025-10-08 18:13:25.898402275 +0000 UTC m=+151.348245045" Oct 08 18:13:25 crc kubenswrapper[4988]: I1008 18:13:25.910638 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/7f4c229e-0791-4953-aa04-5928f5b1480f-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"7f4c229e-0791-4953-aa04-5928f5b1480f\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 08 18:13:25 crc kubenswrapper[4988]: I1008 18:13:25.910920 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/7f4c229e-0791-4953-aa04-5928f5b1480f-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"7f4c229e-0791-4953-aa04-5928f5b1480f\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 08 18:13:25 crc kubenswrapper[4988]: I1008 18:13:25.913707 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-4wh9p" Oct 08 18:13:25 crc kubenswrapper[4988]: I1008 18:13:25.930843 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-7ck2j"] Oct 08 18:13:25 crc kubenswrapper[4988]: I1008 18:13:25.932317 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-7ck2j" Oct 08 18:13:25 crc kubenswrapper[4988]: I1008 18:13:25.936964 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Oct 08 18:13:25 crc kubenswrapper[4988]: I1008 18:13:25.960143 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-7ck2j"] Oct 08 18:13:25 crc kubenswrapper[4988]: I1008 18:13:25.982513 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Oct 08 18:13:25 crc kubenswrapper[4988]: I1008 18:13:25.983563 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 08 18:13:25 crc kubenswrapper[4988]: I1008 18:13:25.988870 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Oct 08 18:13:25 crc kubenswrapper[4988]: I1008 18:13:25.996674 4988 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Oct 08 18:13:26 crc kubenswrapper[4988]: I1008 18:13:26.012679 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-js74s\" (UniqueName: \"kubernetes.io/projected/b94d3b86-2fd1-4045-9a5e-dc6c296cfba4-kube-api-access-js74s\") pod \"redhat-operators-7ck2j\" (UID: \"b94d3b86-2fd1-4045-9a5e-dc6c296cfba4\") " pod="openshift-marketplace/redhat-operators-7ck2j" Oct 08 18:13:26 crc kubenswrapper[4988]: I1008 18:13:26.012756 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/d24735bb-2d72-44e0-9a5b-b220a787e116-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"d24735bb-2d72-44e0-9a5b-b220a787e116\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 08 18:13:26 crc kubenswrapper[4988]: I1008 18:13:26.012823 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/7f4c229e-0791-4953-aa04-5928f5b1480f-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"7f4c229e-0791-4953-aa04-5928f5b1480f\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 08 18:13:26 crc kubenswrapper[4988]: I1008 18:13:26.012853 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/d24735bb-2d72-44e0-9a5b-b220a787e116-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"d24735bb-2d72-44e0-9a5b-b220a787e116\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 08 18:13:26 crc kubenswrapper[4988]: I1008 18:13:26.012872 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/7f4c229e-0791-4953-aa04-5928f5b1480f-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"7f4c229e-0791-4953-aa04-5928f5b1480f\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 08 18:13:26 crc kubenswrapper[4988]: I1008 18:13:26.012889 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b94d3b86-2fd1-4045-9a5e-dc6c296cfba4-catalog-content\") pod \"redhat-operators-7ck2j\" (UID: \"b94d3b86-2fd1-4045-9a5e-dc6c296cfba4\") " pod="openshift-marketplace/redhat-operators-7ck2j" Oct 08 18:13:26 crc kubenswrapper[4988]: I1008 18:13:26.012907 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b94d3b86-2fd1-4045-9a5e-dc6c296cfba4-utilities\") pod \"redhat-operators-7ck2j\" (UID: \"b94d3b86-2fd1-4045-9a5e-dc6c296cfba4\") " pod="openshift-marketplace/redhat-operators-7ck2j" Oct 08 18:13:26 crc kubenswrapper[4988]: I1008 18:13:26.013017 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/7f4c229e-0791-4953-aa04-5928f5b1480f-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"7f4c229e-0791-4953-aa04-5928f5b1480f\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 08 18:13:26 crc kubenswrapper[4988]: I1008 18:13:26.026116 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Oct 08 18:13:26 crc kubenswrapper[4988]: I1008 18:13:26.033102 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/7f4c229e-0791-4953-aa04-5928f5b1480f-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"7f4c229e-0791-4953-aa04-5928f5b1480f\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 08 18:13:26 crc kubenswrapper[4988]: I1008 18:13:26.114512 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/d24735bb-2d72-44e0-9a5b-b220a787e116-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"d24735bb-2d72-44e0-9a5b-b220a787e116\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 08 18:13:26 crc kubenswrapper[4988]: I1008 18:13:26.114616 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/d24735bb-2d72-44e0-9a5b-b220a787e116-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"d24735bb-2d72-44e0-9a5b-b220a787e116\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 08 18:13:26 crc kubenswrapper[4988]: I1008 18:13:26.114639 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b94d3b86-2fd1-4045-9a5e-dc6c296cfba4-catalog-content\") pod \"redhat-operators-7ck2j\" (UID: \"b94d3b86-2fd1-4045-9a5e-dc6c296cfba4\") " pod="openshift-marketplace/redhat-operators-7ck2j" Oct 08 18:13:26 crc kubenswrapper[4988]: I1008 18:13:26.114658 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b94d3b86-2fd1-4045-9a5e-dc6c296cfba4-utilities\") pod \"redhat-operators-7ck2j\" (UID: \"b94d3b86-2fd1-4045-9a5e-dc6c296cfba4\") " pod="openshift-marketplace/redhat-operators-7ck2j" Oct 08 18:13:26 crc kubenswrapper[4988]: I1008 18:13:26.114682 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-js74s\" (UniqueName: \"kubernetes.io/projected/b94d3b86-2fd1-4045-9a5e-dc6c296cfba4-kube-api-access-js74s\") pod \"redhat-operators-7ck2j\" (UID: \"b94d3b86-2fd1-4045-9a5e-dc6c296cfba4\") " pod="openshift-marketplace/redhat-operators-7ck2j" Oct 08 18:13:26 crc kubenswrapper[4988]: I1008 18:13:26.115222 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/d24735bb-2d72-44e0-9a5b-b220a787e116-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"d24735bb-2d72-44e0-9a5b-b220a787e116\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 08 18:13:26 crc kubenswrapper[4988]: I1008 18:13:26.115686 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b94d3b86-2fd1-4045-9a5e-dc6c296cfba4-utilities\") pod \"redhat-operators-7ck2j\" (UID: \"b94d3b86-2fd1-4045-9a5e-dc6c296cfba4\") " pod="openshift-marketplace/redhat-operators-7ck2j" Oct 08 18:13:26 crc kubenswrapper[4988]: I1008 18:13:26.116181 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b94d3b86-2fd1-4045-9a5e-dc6c296cfba4-catalog-content\") pod \"redhat-operators-7ck2j\" (UID: \"b94d3b86-2fd1-4045-9a5e-dc6c296cfba4\") " pod="openshift-marketplace/redhat-operators-7ck2j" Oct 08 18:13:26 crc kubenswrapper[4988]: I1008 18:13:26.133827 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-js74s\" (UniqueName: \"kubernetes.io/projected/b94d3b86-2fd1-4045-9a5e-dc6c296cfba4-kube-api-access-js74s\") pod \"redhat-operators-7ck2j\" (UID: \"b94d3b86-2fd1-4045-9a5e-dc6c296cfba4\") " pod="openshift-marketplace/redhat-operators-7ck2j" Oct 08 18:13:26 crc kubenswrapper[4988]: I1008 18:13:26.135254 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/d24735bb-2d72-44e0-9a5b-b220a787e116-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"d24735bb-2d72-44e0-9a5b-b220a787e116\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 08 18:13:26 crc kubenswrapper[4988]: I1008 18:13:26.161641 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 08 18:13:26 crc kubenswrapper[4988]: I1008 18:13:26.254979 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-7ck2j" Oct 08 18:13:26 crc kubenswrapper[4988]: I1008 18:13:26.329531 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 08 18:13:26 crc kubenswrapper[4988]: I1008 18:13:26.336801 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-pk7fz"] Oct 08 18:13:26 crc kubenswrapper[4988]: I1008 18:13:26.339250 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-pk7fz" Oct 08 18:13:26 crc kubenswrapper[4988]: I1008 18:13:26.357401 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-pk7fz"] Oct 08 18:13:26 crc kubenswrapper[4988]: I1008 18:13:26.420181 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e6bfa362-455b-422e-9c92-dc81aaa1e0ec-utilities\") pod \"redhat-operators-pk7fz\" (UID: \"e6bfa362-455b-422e-9c92-dc81aaa1e0ec\") " pod="openshift-marketplace/redhat-operators-pk7fz" Oct 08 18:13:26 crc kubenswrapper[4988]: I1008 18:13:26.420440 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r4rld\" (UniqueName: \"kubernetes.io/projected/e6bfa362-455b-422e-9c92-dc81aaa1e0ec-kube-api-access-r4rld\") pod \"redhat-operators-pk7fz\" (UID: \"e6bfa362-455b-422e-9c92-dc81aaa1e0ec\") " pod="openshift-marketplace/redhat-operators-pk7fz" Oct 08 18:13:26 crc kubenswrapper[4988]: I1008 18:13:26.420478 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e6bfa362-455b-422e-9c92-dc81aaa1e0ec-catalog-content\") pod \"redhat-operators-pk7fz\" (UID: \"e6bfa362-455b-422e-9c92-dc81aaa1e0ec\") " pod="openshift-marketplace/redhat-operators-pk7fz" Oct 08 18:13:26 crc kubenswrapper[4988]: I1008 18:13:26.428953 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-4wh9p"] Oct 08 18:13:26 crc kubenswrapper[4988]: W1008 18:13:26.483595 4988 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8a33d7bd_26bd_409b_93f1_5f151ebca347.slice/crio-1e0deb69d6e71818c0dc06067684a7c6e51c0f0ed7ef628282ba777c56ff7d00 WatchSource:0}: Error finding container 1e0deb69d6e71818c0dc06067684a7c6e51c0f0ed7ef628282ba777c56ff7d00: Status 404 returned error can't find the container with id 1e0deb69d6e71818c0dc06067684a7c6e51c0f0ed7ef628282ba777c56ff7d00 Oct 08 18:13:26 crc kubenswrapper[4988]: I1008 18:13:26.523277 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r4rld\" (UniqueName: \"kubernetes.io/projected/e6bfa362-455b-422e-9c92-dc81aaa1e0ec-kube-api-access-r4rld\") pod \"redhat-operators-pk7fz\" (UID: \"e6bfa362-455b-422e-9c92-dc81aaa1e0ec\") " pod="openshift-marketplace/redhat-operators-pk7fz" Oct 08 18:13:26 crc kubenswrapper[4988]: I1008 18:13:26.523694 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e6bfa362-455b-422e-9c92-dc81aaa1e0ec-catalog-content\") pod \"redhat-operators-pk7fz\" (UID: \"e6bfa362-455b-422e-9c92-dc81aaa1e0ec\") " pod="openshift-marketplace/redhat-operators-pk7fz" Oct 08 18:13:26 crc kubenswrapper[4988]: I1008 18:13:26.523727 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e6bfa362-455b-422e-9c92-dc81aaa1e0ec-utilities\") pod \"redhat-operators-pk7fz\" (UID: \"e6bfa362-455b-422e-9c92-dc81aaa1e0ec\") " pod="openshift-marketplace/redhat-operators-pk7fz" Oct 08 18:13:26 crc kubenswrapper[4988]: I1008 18:13:26.524216 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e6bfa362-455b-422e-9c92-dc81aaa1e0ec-utilities\") pod \"redhat-operators-pk7fz\" (UID: \"e6bfa362-455b-422e-9c92-dc81aaa1e0ec\") " pod="openshift-marketplace/redhat-operators-pk7fz" Oct 08 18:13:26 crc kubenswrapper[4988]: I1008 18:13:26.524678 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e6bfa362-455b-422e-9c92-dc81aaa1e0ec-catalog-content\") pod \"redhat-operators-pk7fz\" (UID: \"e6bfa362-455b-422e-9c92-dc81aaa1e0ec\") " pod="openshift-marketplace/redhat-operators-pk7fz" Oct 08 18:13:26 crc kubenswrapper[4988]: I1008 18:13:26.548906 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r4rld\" (UniqueName: \"kubernetes.io/projected/e6bfa362-455b-422e-9c92-dc81aaa1e0ec-kube-api-access-r4rld\") pod \"redhat-operators-pk7fz\" (UID: \"e6bfa362-455b-422e-9c92-dc81aaa1e0ec\") " pod="openshift-marketplace/redhat-operators-pk7fz" Oct 08 18:13:26 crc kubenswrapper[4988]: I1008 18:13:26.620860 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Oct 08 18:13:26 crc kubenswrapper[4988]: W1008 18:13:26.638587 4988 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-pod7f4c229e_0791_4953_aa04_5928f5b1480f.slice/crio-e9e5c2289941a151c6333ff745249e30cce48a6643e0996fef5bba5fb5a6c49d WatchSource:0}: Error finding container e9e5c2289941a151c6333ff745249e30cce48a6643e0996fef5bba5fb5a6c49d: Status 404 returned error can't find the container with id e9e5c2289941a151c6333ff745249e30cce48a6643e0996fef5bba5fb5a6c49d Oct 08 18:13:26 crc kubenswrapper[4988]: I1008 18:13:26.692271 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-pk7fz" Oct 08 18:13:26 crc kubenswrapper[4988]: I1008 18:13:26.733396 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-7ck2j"] Oct 08 18:13:26 crc kubenswrapper[4988]: I1008 18:13:26.774124 4988 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-85kp6" Oct 08 18:13:26 crc kubenswrapper[4988]: I1008 18:13:26.782533 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-85kp6" Oct 08 18:13:26 crc kubenswrapper[4988]: I1008 18:13:26.855931 4988 patch_prober.go:28] interesting pod/router-default-5444994796-jjnkn container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 08 18:13:26 crc kubenswrapper[4988]: [-]has-synced failed: reason withheld Oct 08 18:13:26 crc kubenswrapper[4988]: [+]process-running ok Oct 08 18:13:26 crc kubenswrapper[4988]: healthz check failed Oct 08 18:13:26 crc kubenswrapper[4988]: I1008 18:13:26.856021 4988 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-jjnkn" podUID="ed79f1d0-1282-4be1-8e71-68a5320cdab1" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 08 18:13:26 crc kubenswrapper[4988]: I1008 18:13:26.861544 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-f9d7485db-7xdnc" Oct 08 18:13:26 crc kubenswrapper[4988]: I1008 18:13:26.861616 4988 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-f9d7485db-7xdnc" Oct 08 18:13:26 crc kubenswrapper[4988]: I1008 18:13:26.881611 4988 patch_prober.go:28] interesting pod/console-f9d7485db-7xdnc container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.18:8443/health\": dial tcp 10.217.0.18:8443: connect: connection refused" start-of-body= Oct 08 18:13:26 crc kubenswrapper[4988]: I1008 18:13:26.881676 4988 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-f9d7485db-7xdnc" podUID="9c8874ed-3f70-45b6-9f10-d8737b6a7a8f" containerName="console" probeResult="failure" output="Get \"https://10.217.0.18:8443/health\": dial tcp 10.217.0.18:8443: connect: connection refused" Oct 08 18:13:26 crc kubenswrapper[4988]: I1008 18:13:26.897019 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-4wh9p" event={"ID":"8a33d7bd-26bd-409b-93f1-5f151ebca347","Type":"ContainerStarted","Data":"6506e3bcb57c2dd9a9720948dbce35710215bd409e5a7999e73d3ecb72f926a7"} Oct 08 18:13:26 crc kubenswrapper[4988]: I1008 18:13:26.897117 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-697d97f7c8-4wh9p" Oct 08 18:13:26 crc kubenswrapper[4988]: I1008 18:13:26.897136 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-4wh9p" event={"ID":"8a33d7bd-26bd-409b-93f1-5f151ebca347","Type":"ContainerStarted","Data":"1e0deb69d6e71818c0dc06067684a7c6e51c0f0ed7ef628282ba777c56ff7d00"} Oct 08 18:13:26 crc kubenswrapper[4988]: I1008 18:13:26.953187 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7ck2j" event={"ID":"b94d3b86-2fd1-4045-9a5e-dc6c296cfba4","Type":"ContainerStarted","Data":"de0cc2017cd41432a9aacd5339de40ccb247809b0b5b2a0762f144de451c0ac4"} Oct 08 18:13:26 crc kubenswrapper[4988]: I1008 18:13:26.965672 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-697d97f7c8-4wh9p" podStartSLOduration=130.965652211 podStartE2EDuration="2m10.965652211s" podCreationTimestamp="2025-10-08 18:11:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 18:13:26.964058195 +0000 UTC m=+152.413900965" watchObservedRunningTime="2025-10-08 18:13:26.965652211 +0000 UTC m=+152.415494981" Oct 08 18:13:26 crc kubenswrapper[4988]: I1008 18:13:26.986858 4988 generic.go:334] "Generic (PLEG): container finished" podID="56027983-5f79-4555-abb9-4ab9f13cd39f" containerID="f97986ccfa87d32adc71e1c583cb4e49eeb9013ae85bff9c98dc432d1bc57d60" exitCode=0 Oct 08 18:13:26 crc kubenswrapper[4988]: I1008 18:13:26.987021 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-w4qxm" event={"ID":"56027983-5f79-4555-abb9-4ab9f13cd39f","Type":"ContainerDied","Data":"f97986ccfa87d32adc71e1c583cb4e49eeb9013ae85bff9c98dc432d1bc57d60"} Oct 08 18:13:27 crc kubenswrapper[4988]: I1008 18:13:27.012157 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"7f4c229e-0791-4953-aa04-5928f5b1480f","Type":"ContainerStarted","Data":"e9e5c2289941a151c6333ff745249e30cce48a6643e0996fef5bba5fb5a6c49d"} Oct 08 18:13:27 crc kubenswrapper[4988]: I1008 18:13:27.136369 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Oct 08 18:13:27 crc kubenswrapper[4988]: I1008 18:13:27.263554 4988 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8f668bae-612b-4b75-9490-919e737c6a3b" path="/var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes" Oct 08 18:13:27 crc kubenswrapper[4988]: I1008 18:13:27.337722 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-294dt" Oct 08 18:13:27 crc kubenswrapper[4988]: I1008 18:13:27.586201 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-pk7fz"] Oct 08 18:13:27 crc kubenswrapper[4988]: I1008 18:13:27.626523 4988 patch_prober.go:28] interesting pod/apiserver-76f77b778f-4phs4 container/openshift-apiserver namespace/openshift-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[+]ping ok Oct 08 18:13:27 crc kubenswrapper[4988]: [+]log ok Oct 08 18:13:27 crc kubenswrapper[4988]: [+]etcd ok Oct 08 18:13:27 crc kubenswrapper[4988]: [+]poststarthook/start-apiserver-admission-initializer ok Oct 08 18:13:27 crc kubenswrapper[4988]: [+]poststarthook/generic-apiserver-start-informers ok Oct 08 18:13:27 crc kubenswrapper[4988]: [+]poststarthook/max-in-flight-filter ok Oct 08 18:13:27 crc kubenswrapper[4988]: [+]poststarthook/storage-object-count-tracker-hook ok Oct 08 18:13:27 crc kubenswrapper[4988]: [+]poststarthook/image.openshift.io-apiserver-caches ok Oct 08 18:13:27 crc kubenswrapper[4988]: [-]poststarthook/authorization.openshift.io-bootstrapclusterroles failed: reason withheld Oct 08 18:13:27 crc kubenswrapper[4988]: [+]poststarthook/authorization.openshift.io-ensurenodebootstrap-sa ok Oct 08 18:13:27 crc kubenswrapper[4988]: [+]poststarthook/project.openshift.io-projectcache ok Oct 08 18:13:27 crc kubenswrapper[4988]: [+]poststarthook/project.openshift.io-projectauthorizationcache ok Oct 08 18:13:27 crc kubenswrapper[4988]: [+]poststarthook/openshift.io-startinformers ok Oct 08 18:13:27 crc kubenswrapper[4988]: [+]poststarthook/openshift.io-restmapperupdater ok Oct 08 18:13:27 crc kubenswrapper[4988]: [+]poststarthook/quota.openshift.io-clusterquotamapping ok Oct 08 18:13:27 crc kubenswrapper[4988]: livez check failed Oct 08 18:13:27 crc kubenswrapper[4988]: I1008 18:13:27.626701 4988 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-apiserver/apiserver-76f77b778f-4phs4" podUID="05152e1e-3bc1-48b0-818d-41bc0d50fbf5" containerName="openshift-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 08 18:13:27 crc kubenswrapper[4988]: I1008 18:13:27.838045 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ingress/router-default-5444994796-jjnkn" Oct 08 18:13:27 crc kubenswrapper[4988]: I1008 18:13:27.844717 4988 patch_prober.go:28] interesting pod/router-default-5444994796-jjnkn container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 08 18:13:27 crc kubenswrapper[4988]: [-]has-synced failed: reason withheld Oct 08 18:13:27 crc kubenswrapper[4988]: [+]process-running ok Oct 08 18:13:27 crc kubenswrapper[4988]: healthz check failed Oct 08 18:13:27 crc kubenswrapper[4988]: I1008 18:13:27.844815 4988 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-jjnkn" podUID="ed79f1d0-1282-4be1-8e71-68a5320cdab1" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 08 18:13:28 crc kubenswrapper[4988]: I1008 18:13:28.027665 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"7f4c229e-0791-4953-aa04-5928f5b1480f","Type":"ContainerStarted","Data":"a8ec668f1e048ed2a08cda33297421658c1c2e68ca4ea3e7536184da3de898da"} Oct 08 18:13:28 crc kubenswrapper[4988]: I1008 18:13:28.030594 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-pk7fz" event={"ID":"e6bfa362-455b-422e-9c92-dc81aaa1e0ec","Type":"ContainerStarted","Data":"2fd015d5d9120268a081acbca954f7e8f24e879b1c3d61a8d0b89b492bccc7a4"} Oct 08 18:13:28 crc kubenswrapper[4988]: I1008 18:13:28.030617 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-pk7fz" event={"ID":"e6bfa362-455b-422e-9c92-dc81aaa1e0ec","Type":"ContainerStarted","Data":"d22a816c0113f7c9d6ada62d387385068dcbac632ddac86cee376be6a3a61910"} Oct 08 18:13:28 crc kubenswrapper[4988]: I1008 18:13:28.034177 4988 generic.go:334] "Generic (PLEG): container finished" podID="b94d3b86-2fd1-4045-9a5e-dc6c296cfba4" containerID="7295a0e0d87f1abcaabd86b013caddd1d3952a984ac114b32b7d628ca250dfed" exitCode=0 Oct 08 18:13:28 crc kubenswrapper[4988]: I1008 18:13:28.034298 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7ck2j" event={"ID":"b94d3b86-2fd1-4045-9a5e-dc6c296cfba4","Type":"ContainerDied","Data":"7295a0e0d87f1abcaabd86b013caddd1d3952a984ac114b32b7d628ca250dfed"} Oct 08 18:13:28 crc kubenswrapper[4988]: I1008 18:13:28.041587 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"d24735bb-2d72-44e0-9a5b-b220a787e116","Type":"ContainerStarted","Data":"85860d7e7dac441718868ca96563b0edabdb913441deaf94e21ec9c2a9d2967b"} Oct 08 18:13:28 crc kubenswrapper[4988]: I1008 18:13:28.041620 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"d24735bb-2d72-44e0-9a5b-b220a787e116","Type":"ContainerStarted","Data":"f7c22e1544f818511aa7f410b4c31f5a8e3194373cfea1f346d60859358fefc1"} Oct 08 18:13:28 crc kubenswrapper[4988]: I1008 18:13:28.843653 4988 patch_prober.go:28] interesting pod/router-default-5444994796-jjnkn container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 08 18:13:28 crc kubenswrapper[4988]: [-]has-synced failed: reason withheld Oct 08 18:13:28 crc kubenswrapper[4988]: [+]process-running ok Oct 08 18:13:28 crc kubenswrapper[4988]: healthz check failed Oct 08 18:13:28 crc kubenswrapper[4988]: I1008 18:13:28.843737 4988 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-jjnkn" podUID="ed79f1d0-1282-4be1-8e71-68a5320cdab1" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 08 18:13:29 crc kubenswrapper[4988]: I1008 18:13:29.059662 4988 generic.go:334] "Generic (PLEG): container finished" podID="7f4c229e-0791-4953-aa04-5928f5b1480f" containerID="a8ec668f1e048ed2a08cda33297421658c1c2e68ca4ea3e7536184da3de898da" exitCode=0 Oct 08 18:13:29 crc kubenswrapper[4988]: I1008 18:13:29.059752 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"7f4c229e-0791-4953-aa04-5928f5b1480f","Type":"ContainerDied","Data":"a8ec668f1e048ed2a08cda33297421658c1c2e68ca4ea3e7536184da3de898da"} Oct 08 18:13:29 crc kubenswrapper[4988]: I1008 18:13:29.078025 4988 generic.go:334] "Generic (PLEG): container finished" podID="e6bfa362-455b-422e-9c92-dc81aaa1e0ec" containerID="2fd015d5d9120268a081acbca954f7e8f24e879b1c3d61a8d0b89b492bccc7a4" exitCode=0 Oct 08 18:13:29 crc kubenswrapper[4988]: I1008 18:13:29.078160 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-pk7fz" event={"ID":"e6bfa362-455b-422e-9c92-dc81aaa1e0ec","Type":"ContainerDied","Data":"2fd015d5d9120268a081acbca954f7e8f24e879b1c3d61a8d0b89b492bccc7a4"} Oct 08 18:13:29 crc kubenswrapper[4988]: I1008 18:13:29.097054 4988 generic.go:334] "Generic (PLEG): container finished" podID="d24735bb-2d72-44e0-9a5b-b220a787e116" containerID="85860d7e7dac441718868ca96563b0edabdb913441deaf94e21ec9c2a9d2967b" exitCode=0 Oct 08 18:13:29 crc kubenswrapper[4988]: I1008 18:13:29.097165 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"d24735bb-2d72-44e0-9a5b-b220a787e116","Type":"ContainerDied","Data":"85860d7e7dac441718868ca96563b0edabdb913441deaf94e21ec9c2a9d2967b"} Oct 08 18:13:29 crc kubenswrapper[4988]: I1008 18:13:29.843549 4988 patch_prober.go:28] interesting pod/router-default-5444994796-jjnkn container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 08 18:13:29 crc kubenswrapper[4988]: [-]has-synced failed: reason withheld Oct 08 18:13:29 crc kubenswrapper[4988]: [+]process-running ok Oct 08 18:13:29 crc kubenswrapper[4988]: healthz check failed Oct 08 18:13:29 crc kubenswrapper[4988]: I1008 18:13:29.843644 4988 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-jjnkn" podUID="ed79f1d0-1282-4be1-8e71-68a5320cdab1" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 08 18:13:30 crc kubenswrapper[4988]: I1008 18:13:30.871362 4988 patch_prober.go:28] interesting pod/router-default-5444994796-jjnkn container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 08 18:13:30 crc kubenswrapper[4988]: [-]has-synced failed: reason withheld Oct 08 18:13:30 crc kubenswrapper[4988]: [+]process-running ok Oct 08 18:13:30 crc kubenswrapper[4988]: healthz check failed Oct 08 18:13:30 crc kubenswrapper[4988]: I1008 18:13:30.871508 4988 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-jjnkn" podUID="ed79f1d0-1282-4be1-8e71-68a5320cdab1" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 08 18:13:31 crc kubenswrapper[4988]: I1008 18:13:31.840581 4988 patch_prober.go:28] interesting pod/router-default-5444994796-jjnkn container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 08 18:13:31 crc kubenswrapper[4988]: [-]has-synced failed: reason withheld Oct 08 18:13:31 crc kubenswrapper[4988]: [+]process-running ok Oct 08 18:13:31 crc kubenswrapper[4988]: healthz check failed Oct 08 18:13:31 crc kubenswrapper[4988]: I1008 18:13:31.840843 4988 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-jjnkn" podUID="ed79f1d0-1282-4be1-8e71-68a5320cdab1" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 08 18:13:32 crc kubenswrapper[4988]: I1008 18:13:32.628101 4988 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-apiserver/apiserver-76f77b778f-4phs4" Oct 08 18:13:32 crc kubenswrapper[4988]: I1008 18:13:32.633487 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-apiserver/apiserver-76f77b778f-4phs4" Oct 08 18:13:32 crc kubenswrapper[4988]: I1008 18:13:32.842786 4988 patch_prober.go:28] interesting pod/router-default-5444994796-jjnkn container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 08 18:13:32 crc kubenswrapper[4988]: [-]has-synced failed: reason withheld Oct 08 18:13:32 crc kubenswrapper[4988]: [+]process-running ok Oct 08 18:13:32 crc kubenswrapper[4988]: healthz check failed Oct 08 18:13:32 crc kubenswrapper[4988]: I1008 18:13:32.842958 4988 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-jjnkn" podUID="ed79f1d0-1282-4be1-8e71-68a5320cdab1" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 08 18:13:32 crc kubenswrapper[4988]: I1008 18:13:32.964562 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-dns/dns-default-kwcts" Oct 08 18:13:33 crc kubenswrapper[4988]: I1008 18:13:33.840886 4988 patch_prober.go:28] interesting pod/router-default-5444994796-jjnkn container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 08 18:13:33 crc kubenswrapper[4988]: [-]has-synced failed: reason withheld Oct 08 18:13:33 crc kubenswrapper[4988]: [+]process-running ok Oct 08 18:13:33 crc kubenswrapper[4988]: healthz check failed Oct 08 18:13:33 crc kubenswrapper[4988]: I1008 18:13:33.840947 4988 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-jjnkn" podUID="ed79f1d0-1282-4be1-8e71-68a5320cdab1" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 08 18:13:34 crc kubenswrapper[4988]: I1008 18:13:34.839863 4988 patch_prober.go:28] interesting pod/router-default-5444994796-jjnkn container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 08 18:13:34 crc kubenswrapper[4988]: [-]has-synced failed: reason withheld Oct 08 18:13:34 crc kubenswrapper[4988]: [+]process-running ok Oct 08 18:13:34 crc kubenswrapper[4988]: healthz check failed Oct 08 18:13:34 crc kubenswrapper[4988]: I1008 18:13:34.840245 4988 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-jjnkn" podUID="ed79f1d0-1282-4be1-8e71-68a5320cdab1" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 08 18:13:35 crc kubenswrapper[4988]: I1008 18:13:35.400644 4988 patch_prober.go:28] interesting pod/downloads-7954f5f757-qcgmb container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.13:8080/\": dial tcp 10.217.0.13:8080: connect: connection refused" start-of-body= Oct 08 18:13:35 crc kubenswrapper[4988]: I1008 18:13:35.400700 4988 patch_prober.go:28] interesting pod/downloads-7954f5f757-qcgmb container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.13:8080/\": dial tcp 10.217.0.13:8080: connect: connection refused" start-of-body= Oct 08 18:13:35 crc kubenswrapper[4988]: I1008 18:13:35.400732 4988 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-qcgmb" podUID="0bc2144f-4dbe-4dc1-a07a-b764d8582324" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.13:8080/\": dial tcp 10.217.0.13:8080: connect: connection refused" Oct 08 18:13:35 crc kubenswrapper[4988]: I1008 18:13:35.400759 4988 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-qcgmb" podUID="0bc2144f-4dbe-4dc1-a07a-b764d8582324" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.13:8080/\": dial tcp 10.217.0.13:8080: connect: connection refused" Oct 08 18:13:35 crc kubenswrapper[4988]: I1008 18:13:35.840436 4988 patch_prober.go:28] interesting pod/router-default-5444994796-jjnkn container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 08 18:13:35 crc kubenswrapper[4988]: [-]has-synced failed: reason withheld Oct 08 18:13:35 crc kubenswrapper[4988]: [+]process-running ok Oct 08 18:13:35 crc kubenswrapper[4988]: healthz check failed Oct 08 18:13:35 crc kubenswrapper[4988]: I1008 18:13:35.840509 4988 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-jjnkn" podUID="ed79f1d0-1282-4be1-8e71-68a5320cdab1" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 08 18:13:36 crc kubenswrapper[4988]: I1008 18:13:36.840605 4988 patch_prober.go:28] interesting pod/router-default-5444994796-jjnkn container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 08 18:13:36 crc kubenswrapper[4988]: [-]has-synced failed: reason withheld Oct 08 18:13:36 crc kubenswrapper[4988]: [+]process-running ok Oct 08 18:13:36 crc kubenswrapper[4988]: healthz check failed Oct 08 18:13:36 crc kubenswrapper[4988]: I1008 18:13:36.840693 4988 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-jjnkn" podUID="ed79f1d0-1282-4be1-8e71-68a5320cdab1" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 08 18:13:36 crc kubenswrapper[4988]: I1008 18:13:36.861927 4988 patch_prober.go:28] interesting pod/console-f9d7485db-7xdnc container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.18:8443/health\": dial tcp 10.217.0.18:8443: connect: connection refused" start-of-body= Oct 08 18:13:36 crc kubenswrapper[4988]: I1008 18:13:36.862024 4988 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-f9d7485db-7xdnc" podUID="9c8874ed-3f70-45b6-9f10-d8737b6a7a8f" containerName="console" probeResult="failure" output="Get \"https://10.217.0.18:8443/health\": dial tcp 10.217.0.18:8443: connect: connection refused" Oct 08 18:13:37 crc kubenswrapper[4988]: I1008 18:13:37.530396 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 08 18:13:37 crc kubenswrapper[4988]: I1008 18:13:37.535297 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 08 18:13:37 crc kubenswrapper[4988]: I1008 18:13:37.649170 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/d24735bb-2d72-44e0-9a5b-b220a787e116-kubelet-dir\") pod \"d24735bb-2d72-44e0-9a5b-b220a787e116\" (UID: \"d24735bb-2d72-44e0-9a5b-b220a787e116\") " Oct 08 18:13:37 crc kubenswrapper[4988]: I1008 18:13:37.649295 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/7f4c229e-0791-4953-aa04-5928f5b1480f-kube-api-access\") pod \"7f4c229e-0791-4953-aa04-5928f5b1480f\" (UID: \"7f4c229e-0791-4953-aa04-5928f5b1480f\") " Oct 08 18:13:37 crc kubenswrapper[4988]: I1008 18:13:37.649373 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/d24735bb-2d72-44e0-9a5b-b220a787e116-kube-api-access\") pod \"d24735bb-2d72-44e0-9a5b-b220a787e116\" (UID: \"d24735bb-2d72-44e0-9a5b-b220a787e116\") " Oct 08 18:13:37 crc kubenswrapper[4988]: I1008 18:13:37.649432 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/7f4c229e-0791-4953-aa04-5928f5b1480f-kubelet-dir\") pod \"7f4c229e-0791-4953-aa04-5928f5b1480f\" (UID: \"7f4c229e-0791-4953-aa04-5928f5b1480f\") " Oct 08 18:13:37 crc kubenswrapper[4988]: I1008 18:13:37.649348 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d24735bb-2d72-44e0-9a5b-b220a787e116-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "d24735bb-2d72-44e0-9a5b-b220a787e116" (UID: "d24735bb-2d72-44e0-9a5b-b220a787e116"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 08 18:13:37 crc kubenswrapper[4988]: I1008 18:13:37.649650 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/7f4c229e-0791-4953-aa04-5928f5b1480f-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "7f4c229e-0791-4953-aa04-5928f5b1480f" (UID: "7f4c229e-0791-4953-aa04-5928f5b1480f"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 08 18:13:37 crc kubenswrapper[4988]: I1008 18:13:37.649724 4988 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/d24735bb-2d72-44e0-9a5b-b220a787e116-kubelet-dir\") on node \"crc\" DevicePath \"\"" Oct 08 18:13:37 crc kubenswrapper[4988]: I1008 18:13:37.656918 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7f4c229e-0791-4953-aa04-5928f5b1480f-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "7f4c229e-0791-4953-aa04-5928f5b1480f" (UID: "7f4c229e-0791-4953-aa04-5928f5b1480f"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:13:37 crc kubenswrapper[4988]: I1008 18:13:37.656977 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d24735bb-2d72-44e0-9a5b-b220a787e116-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "d24735bb-2d72-44e0-9a5b-b220a787e116" (UID: "d24735bb-2d72-44e0-9a5b-b220a787e116"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:13:37 crc kubenswrapper[4988]: I1008 18:13:37.751891 4988 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/d24735bb-2d72-44e0-9a5b-b220a787e116-kube-api-access\") on node \"crc\" DevicePath \"\"" Oct 08 18:13:37 crc kubenswrapper[4988]: I1008 18:13:37.751933 4988 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/7f4c229e-0791-4953-aa04-5928f5b1480f-kubelet-dir\") on node \"crc\" DevicePath \"\"" Oct 08 18:13:37 crc kubenswrapper[4988]: I1008 18:13:37.751943 4988 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/7f4c229e-0791-4953-aa04-5928f5b1480f-kube-api-access\") on node \"crc\" DevicePath \"\"" Oct 08 18:13:37 crc kubenswrapper[4988]: I1008 18:13:37.842271 4988 patch_prober.go:28] interesting pod/router-default-5444994796-jjnkn container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 08 18:13:37 crc kubenswrapper[4988]: [+]has-synced ok Oct 08 18:13:37 crc kubenswrapper[4988]: [+]process-running ok Oct 08 18:13:37 crc kubenswrapper[4988]: healthz check failed Oct 08 18:13:37 crc kubenswrapper[4988]: I1008 18:13:37.842370 4988 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-jjnkn" podUID="ed79f1d0-1282-4be1-8e71-68a5320cdab1" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 08 18:13:38 crc kubenswrapper[4988]: I1008 18:13:38.189122 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"7f4c229e-0791-4953-aa04-5928f5b1480f","Type":"ContainerDied","Data":"e9e5c2289941a151c6333ff745249e30cce48a6643e0996fef5bba5fb5a6c49d"} Oct 08 18:13:38 crc kubenswrapper[4988]: I1008 18:13:38.189171 4988 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e9e5c2289941a151c6333ff745249e30cce48a6643e0996fef5bba5fb5a6c49d" Oct 08 18:13:38 crc kubenswrapper[4988]: I1008 18:13:38.189164 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 08 18:13:38 crc kubenswrapper[4988]: I1008 18:13:38.191361 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"d24735bb-2d72-44e0-9a5b-b220a787e116","Type":"ContainerDied","Data":"f7c22e1544f818511aa7f410b4c31f5a8e3194373cfea1f346d60859358fefc1"} Oct 08 18:13:38 crc kubenswrapper[4988]: I1008 18:13:38.191415 4988 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f7c22e1544f818511aa7f410b4c31f5a8e3194373cfea1f346d60859358fefc1" Oct 08 18:13:38 crc kubenswrapper[4988]: I1008 18:13:38.191419 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 08 18:13:38 crc kubenswrapper[4988]: I1008 18:13:38.842065 4988 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-ingress/router-default-5444994796-jjnkn" Oct 08 18:13:38 crc kubenswrapper[4988]: I1008 18:13:38.845375 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ingress/router-default-5444994796-jjnkn" Oct 08 18:13:38 crc kubenswrapper[4988]: I1008 18:13:38.969305 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/171e0902-95ee-4a39-8b1b-a58704c30574-metrics-certs\") pod \"network-metrics-daemon-jn2sx\" (UID: \"171e0902-95ee-4a39-8b1b-a58704c30574\") " pod="openshift-multus/network-metrics-daemon-jn2sx" Oct 08 18:13:38 crc kubenswrapper[4988]: I1008 18:13:38.975016 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/171e0902-95ee-4a39-8b1b-a58704c30574-metrics-certs\") pod \"network-metrics-daemon-jn2sx\" (UID: \"171e0902-95ee-4a39-8b1b-a58704c30574\") " pod="openshift-multus/network-metrics-daemon-jn2sx" Oct 08 18:13:39 crc kubenswrapper[4988]: I1008 18:13:39.260990 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jn2sx" Oct 08 18:13:45 crc kubenswrapper[4988]: I1008 18:13:45.408590 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/downloads-7954f5f757-qcgmb" Oct 08 18:13:45 crc kubenswrapper[4988]: I1008 18:13:45.923483 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-697d97f7c8-4wh9p" Oct 08 18:13:46 crc kubenswrapper[4988]: I1008 18:13:46.867564 4988 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-f9d7485db-7xdnc" Oct 08 18:13:46 crc kubenswrapper[4988]: I1008 18:13:46.872301 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-f9d7485db-7xdnc" Oct 08 18:13:53 crc kubenswrapper[4988]: I1008 18:13:53.338680 4988 patch_prober.go:28] interesting pod/machine-config-daemon-nm54f container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 18:13:53 crc kubenswrapper[4988]: I1008 18:13:53.339144 4988 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 18:13:57 crc kubenswrapper[4988]: I1008 18:13:57.573703 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-h5cw8" Oct 08 18:13:58 crc kubenswrapper[4988]: E1008 18:13:58.411329 4988 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Oct 08 18:13:58 crc kubenswrapper[4988]: E1008 18:13:58.412007 4988 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-6d8gx,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-kmg58_openshift-marketplace(38d2198b-2aab-47d2-aa2e-93436ff69625): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Oct 08 18:13:58 crc kubenswrapper[4988]: E1008 18:13:58.413311 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-marketplace-kmg58" podUID="38d2198b-2aab-47d2-aa2e-93436ff69625" Oct 08 18:13:58 crc kubenswrapper[4988]: E1008 18:13:58.649212 4988 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/community-operator-index:v4.18" Oct 08 18:13:58 crc kubenswrapper[4988]: E1008 18:13:58.649410 4988 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-484hv,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-lm95x_openshift-marketplace(619228a7-ec1c-4880-a3d5-9d26ac51afd9): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Oct 08 18:13:58 crc kubenswrapper[4988]: E1008 18:13:58.650603 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/community-operators-lm95x" podUID="619228a7-ec1c-4880-a3d5-9d26ac51afd9" Oct 08 18:14:01 crc kubenswrapper[4988]: E1008 18:14:01.697850 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-kmg58" podUID="38d2198b-2aab-47d2-aa2e-93436ff69625" Oct 08 18:14:01 crc kubenswrapper[4988]: E1008 18:14:01.697932 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-lm95x" podUID="619228a7-ec1c-4880-a3d5-9d26ac51afd9" Oct 08 18:14:01 crc kubenswrapper[4988]: E1008 18:14:01.738676 4988 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-operator-index:v4.18" Oct 08 18:14:01 crc kubenswrapper[4988]: E1008 18:14:01.739056 4988 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-r4rld,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-operators-pk7fz_openshift-marketplace(e6bfa362-455b-422e-9c92-dc81aaa1e0ec): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Oct 08 18:14:01 crc kubenswrapper[4988]: E1008 18:14:01.740340 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-operators-pk7fz" podUID="e6bfa362-455b-422e-9c92-dc81aaa1e0ec" Oct 08 18:14:01 crc kubenswrapper[4988]: E1008 18:14:01.790790 4988 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Oct 08 18:14:01 crc kubenswrapper[4988]: E1008 18:14:01.790953 4988 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-rmkt7,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-hvx7q_openshift-marketplace(4eb14139-7a10-4429-9834-378a89ea8255): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Oct 08 18:14:01 crc kubenswrapper[4988]: E1008 18:14:01.792195 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/certified-operators-hvx7q" podUID="4eb14139-7a10-4429-9834-378a89ea8255" Oct 08 18:14:01 crc kubenswrapper[4988]: E1008 18:14:01.806476 4988 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-operator-index:v4.18" Oct 08 18:14:01 crc kubenswrapper[4988]: E1008 18:14:01.806659 4988 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-js74s,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-operators-7ck2j_openshift-marketplace(b94d3b86-2fd1-4045-9a5e-dc6c296cfba4): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Oct 08 18:14:01 crc kubenswrapper[4988]: E1008 18:14:01.807863 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-operators-7ck2j" podUID="b94d3b86-2fd1-4045-9a5e-dc6c296cfba4" Oct 08 18:14:01 crc kubenswrapper[4988]: E1008 18:14:01.826889 4988 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Oct 08 18:14:01 crc kubenswrapper[4988]: E1008 18:14:01.827253 4988 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-4jxl8,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-w4qxm_openshift-marketplace(56027983-5f79-4555-abb9-4ab9f13cd39f): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Oct 08 18:14:01 crc kubenswrapper[4988]: E1008 18:14:01.828637 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-marketplace-w4qxm" podUID="56027983-5f79-4555-abb9-4ab9f13cd39f" Oct 08 18:14:01 crc kubenswrapper[4988]: I1008 18:14:01.979905 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-jn2sx"] Oct 08 18:14:01 crc kubenswrapper[4988]: W1008 18:14:01.990701 4988 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod171e0902_95ee_4a39_8b1b_a58704c30574.slice/crio-d2f50ccc27c02eb196f6505e67ca238e09563063ae226bbe98f6713772bbc157 WatchSource:0}: Error finding container d2f50ccc27c02eb196f6505e67ca238e09563063ae226bbe98f6713772bbc157: Status 404 returned error can't find the container with id d2f50ccc27c02eb196f6505e67ca238e09563063ae226bbe98f6713772bbc157 Oct 08 18:14:02 crc kubenswrapper[4988]: I1008 18:14:02.377941 4988 generic.go:334] "Generic (PLEG): container finished" podID="7c0e2c11-212a-46ed-b72a-911bde29199a" containerID="7285e62439fa2c40d6fb7ad0da89aaa7edc907b20ba48be35aae53b4d742c498" exitCode=0 Oct 08 18:14:02 crc kubenswrapper[4988]: I1008 18:14:02.377999 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-77xtx" event={"ID":"7c0e2c11-212a-46ed-b72a-911bde29199a","Type":"ContainerDied","Data":"7285e62439fa2c40d6fb7ad0da89aaa7edc907b20ba48be35aae53b4d742c498"} Oct 08 18:14:02 crc kubenswrapper[4988]: I1008 18:14:02.381940 4988 generic.go:334] "Generic (PLEG): container finished" podID="15a52185-0755-45a6-b665-c1185d1996b4" containerID="31cf1e847f46eb07b87ffc654f5981503665b3b75fd7897c9b7e31e08d1e91e7" exitCode=0 Oct 08 18:14:02 crc kubenswrapper[4988]: I1008 18:14:02.382024 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dmdqd" event={"ID":"15a52185-0755-45a6-b665-c1185d1996b4","Type":"ContainerDied","Data":"31cf1e847f46eb07b87ffc654f5981503665b3b75fd7897c9b7e31e08d1e91e7"} Oct 08 18:14:02 crc kubenswrapper[4988]: I1008 18:14:02.407493 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-jn2sx" event={"ID":"171e0902-95ee-4a39-8b1b-a58704c30574","Type":"ContainerStarted","Data":"1a78d6a8982533d0a8779c00424906efb79e158a3f6ee005818eba88940ff14b"} Oct 08 18:14:02 crc kubenswrapper[4988]: I1008 18:14:02.407554 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-jn2sx" event={"ID":"171e0902-95ee-4a39-8b1b-a58704c30574","Type":"ContainerStarted","Data":"d2f50ccc27c02eb196f6505e67ca238e09563063ae226bbe98f6713772bbc157"} Oct 08 18:14:02 crc kubenswrapper[4988]: E1008 18:14:02.409919 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-7ck2j" podUID="b94d3b86-2fd1-4045-9a5e-dc6c296cfba4" Oct 08 18:14:02 crc kubenswrapper[4988]: E1008 18:14:02.410756 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-pk7fz" podUID="e6bfa362-455b-422e-9c92-dc81aaa1e0ec" Oct 08 18:14:02 crc kubenswrapper[4988]: E1008 18:14:02.410801 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-w4qxm" podUID="56027983-5f79-4555-abb9-4ab9f13cd39f" Oct 08 18:14:02 crc kubenswrapper[4988]: E1008 18:14:02.415309 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-hvx7q" podUID="4eb14139-7a10-4429-9834-378a89ea8255" Oct 08 18:14:03 crc kubenswrapper[4988]: I1008 18:14:03.275023 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 18:14:03 crc kubenswrapper[4988]: I1008 18:14:03.415563 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dmdqd" event={"ID":"15a52185-0755-45a6-b665-c1185d1996b4","Type":"ContainerStarted","Data":"376064a31a6403d1a62039ad1b6d1d240e5e224930f9b781cffd57e988576c51"} Oct 08 18:14:03 crc kubenswrapper[4988]: I1008 18:14:03.417730 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-jn2sx" event={"ID":"171e0902-95ee-4a39-8b1b-a58704c30574","Type":"ContainerStarted","Data":"493fd8d9f8df72e816c0ae25b5faa3fdf5946586421f193f7649e9303e71a5f2"} Oct 08 18:14:03 crc kubenswrapper[4988]: I1008 18:14:03.420215 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-77xtx" event={"ID":"7c0e2c11-212a-46ed-b72a-911bde29199a","Type":"ContainerStarted","Data":"615d0e89e4190c703d515e7df5d5b3b444dd61b73916cb4b40cc5c9ce105c872"} Oct 08 18:14:03 crc kubenswrapper[4988]: I1008 18:14:03.432897 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-dmdqd" podStartSLOduration=4.408632326 podStartE2EDuration="41.432874524s" podCreationTimestamp="2025-10-08 18:13:22 +0000 UTC" firstStartedPulling="2025-10-08 18:13:25.806872166 +0000 UTC m=+151.256714926" lastFinishedPulling="2025-10-08 18:14:02.831114294 +0000 UTC m=+188.280957124" observedRunningTime="2025-10-08 18:14:03.431331528 +0000 UTC m=+188.881174308" watchObservedRunningTime="2025-10-08 18:14:03.432874524 +0000 UTC m=+188.882717304" Oct 08 18:14:03 crc kubenswrapper[4988]: I1008 18:14:03.456921 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-77xtx" podStartSLOduration=3.154685314 podStartE2EDuration="41.456899184s" podCreationTimestamp="2025-10-08 18:13:22 +0000 UTC" firstStartedPulling="2025-10-08 18:13:24.75267473 +0000 UTC m=+150.202517500" lastFinishedPulling="2025-10-08 18:14:03.0548886 +0000 UTC m=+188.504731370" observedRunningTime="2025-10-08 18:14:03.453268759 +0000 UTC m=+188.903111589" watchObservedRunningTime="2025-10-08 18:14:03.456899184 +0000 UTC m=+188.906741954" Oct 08 18:14:03 crc kubenswrapper[4988]: I1008 18:14:03.477456 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/network-metrics-daemon-jn2sx" podStartSLOduration=167.477432263 podStartE2EDuration="2m47.477432263s" podCreationTimestamp="2025-10-08 18:11:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 18:14:03.475727184 +0000 UTC m=+188.925569964" watchObservedRunningTime="2025-10-08 18:14:03.477432263 +0000 UTC m=+188.927275033" Oct 08 18:14:12 crc kubenswrapper[4988]: I1008 18:14:12.884839 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-77xtx" Oct 08 18:14:12 crc kubenswrapper[4988]: I1008 18:14:12.885779 4988 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-77xtx" Oct 08 18:14:13 crc kubenswrapper[4988]: I1008 18:14:13.062193 4988 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-dmdqd" Oct 08 18:14:13 crc kubenswrapper[4988]: I1008 18:14:13.063173 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-dmdqd" Oct 08 18:14:13 crc kubenswrapper[4988]: I1008 18:14:13.461159 4988 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-77xtx" Oct 08 18:14:13 crc kubenswrapper[4988]: I1008 18:14:13.465886 4988 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-dmdqd" Oct 08 18:14:13 crc kubenswrapper[4988]: I1008 18:14:13.526250 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-dmdqd" Oct 08 18:14:13 crc kubenswrapper[4988]: I1008 18:14:13.573946 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-77xtx" Oct 08 18:14:16 crc kubenswrapper[4988]: I1008 18:14:16.499902 4988 generic.go:334] "Generic (PLEG): container finished" podID="38d2198b-2aab-47d2-aa2e-93436ff69625" containerID="b54e8a500850c375abfec8de4ae46eaae5f4c989bc6e31bcc77229321643e998" exitCode=0 Oct 08 18:14:16 crc kubenswrapper[4988]: I1008 18:14:16.499973 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-kmg58" event={"ID":"38d2198b-2aab-47d2-aa2e-93436ff69625","Type":"ContainerDied","Data":"b54e8a500850c375abfec8de4ae46eaae5f4c989bc6e31bcc77229321643e998"} Oct 08 18:14:16 crc kubenswrapper[4988]: I1008 18:14:16.506459 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-pk7fz" event={"ID":"e6bfa362-455b-422e-9c92-dc81aaa1e0ec","Type":"ContainerStarted","Data":"1c352caeee323da70f697f14e91dcb1dacba6a2f54dc6e212cda84c95e3ecb24"} Oct 08 18:14:16 crc kubenswrapper[4988]: I1008 18:14:16.512083 4988 generic.go:334] "Generic (PLEG): container finished" podID="b94d3b86-2fd1-4045-9a5e-dc6c296cfba4" containerID="7fe16147afcdef2e218e05e65b8ff5119b856be161a400f38a5400ea66ea2f8a" exitCode=0 Oct 08 18:14:16 crc kubenswrapper[4988]: I1008 18:14:16.512234 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7ck2j" event={"ID":"b94d3b86-2fd1-4045-9a5e-dc6c296cfba4","Type":"ContainerDied","Data":"7fe16147afcdef2e218e05e65b8ff5119b856be161a400f38a5400ea66ea2f8a"} Oct 08 18:14:16 crc kubenswrapper[4988]: I1008 18:14:16.515361 4988 generic.go:334] "Generic (PLEG): container finished" podID="4eb14139-7a10-4429-9834-378a89ea8255" containerID="bc9d8017bf8452b17dcbc32859ed1bf714b54116c673f5a0c667fb167664cd8f" exitCode=0 Oct 08 18:14:16 crc kubenswrapper[4988]: I1008 18:14:16.515464 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-hvx7q" event={"ID":"4eb14139-7a10-4429-9834-378a89ea8255","Type":"ContainerDied","Data":"bc9d8017bf8452b17dcbc32859ed1bf714b54116c673f5a0c667fb167664cd8f"} Oct 08 18:14:16 crc kubenswrapper[4988]: I1008 18:14:16.522208 4988 generic.go:334] "Generic (PLEG): container finished" podID="619228a7-ec1c-4880-a3d5-9d26ac51afd9" containerID="deb583d21b801e995436d011c5a5e5a148cc267d5e298f602b094b0bb42df73f" exitCode=0 Oct 08 18:14:16 crc kubenswrapper[4988]: I1008 18:14:16.522260 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-lm95x" event={"ID":"619228a7-ec1c-4880-a3d5-9d26ac51afd9","Type":"ContainerDied","Data":"deb583d21b801e995436d011c5a5e5a148cc267d5e298f602b094b0bb42df73f"} Oct 08 18:14:17 crc kubenswrapper[4988]: I1008 18:14:17.531349 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-kmg58" event={"ID":"38d2198b-2aab-47d2-aa2e-93436ff69625","Type":"ContainerStarted","Data":"80113d52537f531d6bba38b5509f7a18c44a61e0f36a58742df56936d485d0f9"} Oct 08 18:14:17 crc kubenswrapper[4988]: I1008 18:14:17.535564 4988 generic.go:334] "Generic (PLEG): container finished" podID="e6bfa362-455b-422e-9c92-dc81aaa1e0ec" containerID="1c352caeee323da70f697f14e91dcb1dacba6a2f54dc6e212cda84c95e3ecb24" exitCode=0 Oct 08 18:14:17 crc kubenswrapper[4988]: I1008 18:14:17.535642 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-pk7fz" event={"ID":"e6bfa362-455b-422e-9c92-dc81aaa1e0ec","Type":"ContainerDied","Data":"1c352caeee323da70f697f14e91dcb1dacba6a2f54dc6e212cda84c95e3ecb24"} Oct 08 18:14:17 crc kubenswrapper[4988]: I1008 18:14:17.538426 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7ck2j" event={"ID":"b94d3b86-2fd1-4045-9a5e-dc6c296cfba4","Type":"ContainerStarted","Data":"021595f654ab0cc3eebfd44bb31805da9eef8f2dbc83c40b88a030a1e8fe3c54"} Oct 08 18:14:17 crc kubenswrapper[4988]: I1008 18:14:17.543437 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-hvx7q" event={"ID":"4eb14139-7a10-4429-9834-378a89ea8255","Type":"ContainerStarted","Data":"1400ee3c44221a939e385a6beb47acd9e923650615de9c03152ae663fa37c3ac"} Oct 08 18:14:17 crc kubenswrapper[4988]: I1008 18:14:17.546499 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-lm95x" event={"ID":"619228a7-ec1c-4880-a3d5-9d26ac51afd9","Type":"ContainerStarted","Data":"a72d0e017b4c6aca1bf84f74a538fe323e19c34d76f05250b2bbf4a504ff9f39"} Oct 08 18:14:17 crc kubenswrapper[4988]: I1008 18:14:17.562114 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-kmg58" podStartSLOduration=2.224837789 podStartE2EDuration="53.562096024s" podCreationTimestamp="2025-10-08 18:13:24 +0000 UTC" firstStartedPulling="2025-10-08 18:13:25.767166587 +0000 UTC m=+151.217009357" lastFinishedPulling="2025-10-08 18:14:17.104424822 +0000 UTC m=+202.554267592" observedRunningTime="2025-10-08 18:14:17.559475 +0000 UTC m=+203.009317780" watchObservedRunningTime="2025-10-08 18:14:17.562096024 +0000 UTC m=+203.011938794" Oct 08 18:14:17 crc kubenswrapper[4988]: I1008 18:14:17.584529 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-lm95x" podStartSLOduration=4.340206691 podStartE2EDuration="55.584509806s" podCreationTimestamp="2025-10-08 18:13:22 +0000 UTC" firstStartedPulling="2025-10-08 18:13:25.784616057 +0000 UTC m=+151.234458837" lastFinishedPulling="2025-10-08 18:14:17.028919182 +0000 UTC m=+202.478761952" observedRunningTime="2025-10-08 18:14:17.581617773 +0000 UTC m=+203.031460543" watchObservedRunningTime="2025-10-08 18:14:17.584509806 +0000 UTC m=+203.034352576" Oct 08 18:14:17 crc kubenswrapper[4988]: I1008 18:14:17.636689 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-hvx7q" podStartSLOduration=3.418264942 podStartE2EDuration="54.636659944s" podCreationTimestamp="2025-10-08 18:13:23 +0000 UTC" firstStartedPulling="2025-10-08 18:13:25.774495971 +0000 UTC m=+151.224338741" lastFinishedPulling="2025-10-08 18:14:16.992890943 +0000 UTC m=+202.442733743" observedRunningTime="2025-10-08 18:14:17.632411258 +0000 UTC m=+203.082254038" watchObservedRunningTime="2025-10-08 18:14:17.636659944 +0000 UTC m=+203.086502704" Oct 08 18:14:17 crc kubenswrapper[4988]: I1008 18:14:17.707084 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-7ck2j" podStartSLOduration=3.790570798 podStartE2EDuration="52.7070599s" podCreationTimestamp="2025-10-08 18:13:25 +0000 UTC" firstStartedPulling="2025-10-08 18:13:28.041697564 +0000 UTC m=+153.491540334" lastFinishedPulling="2025-10-08 18:14:16.958186656 +0000 UTC m=+202.408029436" observedRunningTime="2025-10-08 18:14:17.697970947 +0000 UTC m=+203.147813727" watchObservedRunningTime="2025-10-08 18:14:17.7070599 +0000 UTC m=+203.156902670" Oct 08 18:14:18 crc kubenswrapper[4988]: I1008 18:14:18.554139 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-pk7fz" event={"ID":"e6bfa362-455b-422e-9c92-dc81aaa1e0ec","Type":"ContainerStarted","Data":"59bc1e445020368043ac5535d5c5e63a263cabb59e7f76532191cbdc959bdc70"} Oct 08 18:14:18 crc kubenswrapper[4988]: I1008 18:14:18.558121 4988 generic.go:334] "Generic (PLEG): container finished" podID="56027983-5f79-4555-abb9-4ab9f13cd39f" containerID="47761230f30576c443259c7106cd6271733fa977159af6c4ef87a9f6c583107c" exitCode=0 Oct 08 18:14:18 crc kubenswrapper[4988]: I1008 18:14:18.558169 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-w4qxm" event={"ID":"56027983-5f79-4555-abb9-4ab9f13cd39f","Type":"ContainerDied","Data":"47761230f30576c443259c7106cd6271733fa977159af6c4ef87a9f6c583107c"} Oct 08 18:14:18 crc kubenswrapper[4988]: I1008 18:14:18.600238 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-pk7fz" podStartSLOduration=3.581552595 podStartE2EDuration="52.600206939s" podCreationTimestamp="2025-10-08 18:13:26 +0000 UTC" firstStartedPulling="2025-10-08 18:13:29.081178889 +0000 UTC m=+154.531021659" lastFinishedPulling="2025-10-08 18:14:18.099833233 +0000 UTC m=+203.549676003" observedRunningTime="2025-10-08 18:14:18.575235365 +0000 UTC m=+204.025078135" watchObservedRunningTime="2025-10-08 18:14:18.600206939 +0000 UTC m=+204.050049709" Oct 08 18:14:19 crc kubenswrapper[4988]: I1008 18:14:19.577313 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-w4qxm" event={"ID":"56027983-5f79-4555-abb9-4ab9f13cd39f","Type":"ContainerStarted","Data":"73ce37d8d4e252e6d55f78380d048b1ce9aeea2857a615a8fbbbb1f5ce221565"} Oct 08 18:14:19 crc kubenswrapper[4988]: I1008 18:14:19.597252 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-w4qxm" podStartSLOduration=2.367212699 podStartE2EDuration="55.59722141s" podCreationTimestamp="2025-10-08 18:13:24 +0000 UTC" firstStartedPulling="2025-10-08 18:13:25.807535015 +0000 UTC m=+151.257377785" lastFinishedPulling="2025-10-08 18:14:19.037543726 +0000 UTC m=+204.487386496" observedRunningTime="2025-10-08 18:14:19.594533514 +0000 UTC m=+205.044376314" watchObservedRunningTime="2025-10-08 18:14:19.59722141 +0000 UTC m=+205.047064200" Oct 08 18:14:22 crc kubenswrapper[4988]: I1008 18:14:22.595122 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-hwtzl"] Oct 08 18:14:23 crc kubenswrapper[4988]: I1008 18:14:23.329380 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-lm95x" Oct 08 18:14:23 crc kubenswrapper[4988]: I1008 18:14:23.329826 4988 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-lm95x" Oct 08 18:14:23 crc kubenswrapper[4988]: I1008 18:14:23.338048 4988 patch_prober.go:28] interesting pod/machine-config-daemon-nm54f container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 18:14:23 crc kubenswrapper[4988]: I1008 18:14:23.338126 4988 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 18:14:23 crc kubenswrapper[4988]: I1008 18:14:23.338184 4988 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" Oct 08 18:14:23 crc kubenswrapper[4988]: I1008 18:14:23.338858 4988 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"3cdb2a757332bdcdc2a2f3362913fe8f06003b0597de20b72627f9badd5e30af"} pod="openshift-machine-config-operator/machine-config-daemon-nm54f" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 08 18:14:23 crc kubenswrapper[4988]: I1008 18:14:23.339046 4988 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" containerName="machine-config-daemon" containerID="cri-o://3cdb2a757332bdcdc2a2f3362913fe8f06003b0597de20b72627f9badd5e30af" gracePeriod=600 Oct 08 18:14:23 crc kubenswrapper[4988]: I1008 18:14:23.387152 4988 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-lm95x" Oct 08 18:14:23 crc kubenswrapper[4988]: I1008 18:14:23.506720 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-hvx7q" Oct 08 18:14:23 crc kubenswrapper[4988]: I1008 18:14:23.506797 4988 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-hvx7q" Oct 08 18:14:23 crc kubenswrapper[4988]: I1008 18:14:23.550723 4988 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-hvx7q" Oct 08 18:14:23 crc kubenswrapper[4988]: I1008 18:14:23.652559 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-lm95x" Oct 08 18:14:23 crc kubenswrapper[4988]: I1008 18:14:23.666159 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-hvx7q" Oct 08 18:14:24 crc kubenswrapper[4988]: I1008 18:14:24.615241 4988 generic.go:334] "Generic (PLEG): container finished" podID="f328a20b-5283-407e-8a93-2ab2da599f18" containerID="3cdb2a757332bdcdc2a2f3362913fe8f06003b0597de20b72627f9badd5e30af" exitCode=0 Oct 08 18:14:24 crc kubenswrapper[4988]: I1008 18:14:24.615328 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" event={"ID":"f328a20b-5283-407e-8a93-2ab2da599f18","Type":"ContainerDied","Data":"3cdb2a757332bdcdc2a2f3362913fe8f06003b0597de20b72627f9badd5e30af"} Oct 08 18:14:24 crc kubenswrapper[4988]: I1008 18:14:24.930308 4988 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-kmg58" Oct 08 18:14:24 crc kubenswrapper[4988]: I1008 18:14:24.930362 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-kmg58" Oct 08 18:14:24 crc kubenswrapper[4988]: I1008 18:14:24.995075 4988 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-kmg58" Oct 08 18:14:25 crc kubenswrapper[4988]: I1008 18:14:25.258350 4988 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-w4qxm" Oct 08 18:14:25 crc kubenswrapper[4988]: I1008 18:14:25.260988 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-w4qxm" Oct 08 18:14:25 crc kubenswrapper[4988]: I1008 18:14:25.281177 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-lm95x"] Oct 08 18:14:25 crc kubenswrapper[4988]: I1008 18:14:25.331706 4988 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-w4qxm" Oct 08 18:14:25 crc kubenswrapper[4988]: I1008 18:14:25.623232 4988 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-lm95x" podUID="619228a7-ec1c-4880-a3d5-9d26ac51afd9" containerName="registry-server" containerID="cri-o://a72d0e017b4c6aca1bf84f74a538fe323e19c34d76f05250b2bbf4a504ff9f39" gracePeriod=2 Oct 08 18:14:25 crc kubenswrapper[4988]: I1008 18:14:25.623329 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" event={"ID":"f328a20b-5283-407e-8a93-2ab2da599f18","Type":"ContainerStarted","Data":"6f23f485d033acafdd8130a121cb37cf846dfb3c004e9a0d5e65009a35c2c0b0"} Oct 08 18:14:25 crc kubenswrapper[4988]: I1008 18:14:25.674759 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-w4qxm" Oct 08 18:14:25 crc kubenswrapper[4988]: I1008 18:14:25.678834 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-kmg58" Oct 08 18:14:25 crc kubenswrapper[4988]: I1008 18:14:25.878485 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-hvx7q"] Oct 08 18:14:25 crc kubenswrapper[4988]: I1008 18:14:25.879157 4988 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-hvx7q" podUID="4eb14139-7a10-4429-9834-378a89ea8255" containerName="registry-server" containerID="cri-o://1400ee3c44221a939e385a6beb47acd9e923650615de9c03152ae663fa37c3ac" gracePeriod=2 Oct 08 18:14:26 crc kubenswrapper[4988]: I1008 18:14:26.009563 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-lm95x" Oct 08 18:14:26 crc kubenswrapper[4988]: I1008 18:14:26.016856 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-484hv\" (UniqueName: \"kubernetes.io/projected/619228a7-ec1c-4880-a3d5-9d26ac51afd9-kube-api-access-484hv\") pod \"619228a7-ec1c-4880-a3d5-9d26ac51afd9\" (UID: \"619228a7-ec1c-4880-a3d5-9d26ac51afd9\") " Oct 08 18:14:26 crc kubenswrapper[4988]: I1008 18:14:26.016944 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/619228a7-ec1c-4880-a3d5-9d26ac51afd9-catalog-content\") pod \"619228a7-ec1c-4880-a3d5-9d26ac51afd9\" (UID: \"619228a7-ec1c-4880-a3d5-9d26ac51afd9\") " Oct 08 18:14:26 crc kubenswrapper[4988]: I1008 18:14:26.017000 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/619228a7-ec1c-4880-a3d5-9d26ac51afd9-utilities\") pod \"619228a7-ec1c-4880-a3d5-9d26ac51afd9\" (UID: \"619228a7-ec1c-4880-a3d5-9d26ac51afd9\") " Oct 08 18:14:26 crc kubenswrapper[4988]: I1008 18:14:26.017888 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/619228a7-ec1c-4880-a3d5-9d26ac51afd9-utilities" (OuterVolumeSpecName: "utilities") pod "619228a7-ec1c-4880-a3d5-9d26ac51afd9" (UID: "619228a7-ec1c-4880-a3d5-9d26ac51afd9"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 18:14:26 crc kubenswrapper[4988]: I1008 18:14:26.031864 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/619228a7-ec1c-4880-a3d5-9d26ac51afd9-kube-api-access-484hv" (OuterVolumeSpecName: "kube-api-access-484hv") pod "619228a7-ec1c-4880-a3d5-9d26ac51afd9" (UID: "619228a7-ec1c-4880-a3d5-9d26ac51afd9"). InnerVolumeSpecName "kube-api-access-484hv". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:14:26 crc kubenswrapper[4988]: I1008 18:14:26.113964 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/619228a7-ec1c-4880-a3d5-9d26ac51afd9-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "619228a7-ec1c-4880-a3d5-9d26ac51afd9" (UID: "619228a7-ec1c-4880-a3d5-9d26ac51afd9"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 18:14:26 crc kubenswrapper[4988]: I1008 18:14:26.118981 4988 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-484hv\" (UniqueName: \"kubernetes.io/projected/619228a7-ec1c-4880-a3d5-9d26ac51afd9-kube-api-access-484hv\") on node \"crc\" DevicePath \"\"" Oct 08 18:14:26 crc kubenswrapper[4988]: I1008 18:14:26.119042 4988 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/619228a7-ec1c-4880-a3d5-9d26ac51afd9-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 08 18:14:26 crc kubenswrapper[4988]: I1008 18:14:26.119058 4988 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/619228a7-ec1c-4880-a3d5-9d26ac51afd9-utilities\") on node \"crc\" DevicePath \"\"" Oct 08 18:14:26 crc kubenswrapper[4988]: I1008 18:14:26.256151 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-7ck2j" Oct 08 18:14:26 crc kubenswrapper[4988]: I1008 18:14:26.256204 4988 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-7ck2j" Oct 08 18:14:26 crc kubenswrapper[4988]: I1008 18:14:26.322438 4988 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-7ck2j" Oct 08 18:14:26 crc kubenswrapper[4988]: I1008 18:14:26.351956 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-hvx7q" Oct 08 18:14:26 crc kubenswrapper[4988]: I1008 18:14:26.421982 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4eb14139-7a10-4429-9834-378a89ea8255-utilities\") pod \"4eb14139-7a10-4429-9834-378a89ea8255\" (UID: \"4eb14139-7a10-4429-9834-378a89ea8255\") " Oct 08 18:14:26 crc kubenswrapper[4988]: I1008 18:14:26.422139 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rmkt7\" (UniqueName: \"kubernetes.io/projected/4eb14139-7a10-4429-9834-378a89ea8255-kube-api-access-rmkt7\") pod \"4eb14139-7a10-4429-9834-378a89ea8255\" (UID: \"4eb14139-7a10-4429-9834-378a89ea8255\") " Oct 08 18:14:26 crc kubenswrapper[4988]: I1008 18:14:26.422160 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4eb14139-7a10-4429-9834-378a89ea8255-catalog-content\") pod \"4eb14139-7a10-4429-9834-378a89ea8255\" (UID: \"4eb14139-7a10-4429-9834-378a89ea8255\") " Oct 08 18:14:26 crc kubenswrapper[4988]: I1008 18:14:26.422668 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4eb14139-7a10-4429-9834-378a89ea8255-utilities" (OuterVolumeSpecName: "utilities") pod "4eb14139-7a10-4429-9834-378a89ea8255" (UID: "4eb14139-7a10-4429-9834-378a89ea8255"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 18:14:26 crc kubenswrapper[4988]: I1008 18:14:26.425792 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4eb14139-7a10-4429-9834-378a89ea8255-kube-api-access-rmkt7" (OuterVolumeSpecName: "kube-api-access-rmkt7") pod "4eb14139-7a10-4429-9834-378a89ea8255" (UID: "4eb14139-7a10-4429-9834-378a89ea8255"). InnerVolumeSpecName "kube-api-access-rmkt7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:14:26 crc kubenswrapper[4988]: I1008 18:14:26.474051 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4eb14139-7a10-4429-9834-378a89ea8255-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "4eb14139-7a10-4429-9834-378a89ea8255" (UID: "4eb14139-7a10-4429-9834-378a89ea8255"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 18:14:26 crc kubenswrapper[4988]: I1008 18:14:26.523786 4988 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4eb14139-7a10-4429-9834-378a89ea8255-utilities\") on node \"crc\" DevicePath \"\"" Oct 08 18:14:26 crc kubenswrapper[4988]: I1008 18:14:26.524102 4988 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rmkt7\" (UniqueName: \"kubernetes.io/projected/4eb14139-7a10-4429-9834-378a89ea8255-kube-api-access-rmkt7\") on node \"crc\" DevicePath \"\"" Oct 08 18:14:26 crc kubenswrapper[4988]: I1008 18:14:26.524121 4988 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4eb14139-7a10-4429-9834-378a89ea8255-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 08 18:14:26 crc kubenswrapper[4988]: I1008 18:14:26.632695 4988 generic.go:334] "Generic (PLEG): container finished" podID="4eb14139-7a10-4429-9834-378a89ea8255" containerID="1400ee3c44221a939e385a6beb47acd9e923650615de9c03152ae663fa37c3ac" exitCode=0 Oct 08 18:14:26 crc kubenswrapper[4988]: I1008 18:14:26.632748 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-hvx7q" Oct 08 18:14:26 crc kubenswrapper[4988]: I1008 18:14:26.632771 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-hvx7q" event={"ID":"4eb14139-7a10-4429-9834-378a89ea8255","Type":"ContainerDied","Data":"1400ee3c44221a939e385a6beb47acd9e923650615de9c03152ae663fa37c3ac"} Oct 08 18:14:26 crc kubenswrapper[4988]: I1008 18:14:26.632812 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-hvx7q" event={"ID":"4eb14139-7a10-4429-9834-378a89ea8255","Type":"ContainerDied","Data":"20bd3fe0a116cbade50d4021346ff01b51b7443d1d530257122b6892173cbd0e"} Oct 08 18:14:26 crc kubenswrapper[4988]: I1008 18:14:26.632836 4988 scope.go:117] "RemoveContainer" containerID="1400ee3c44221a939e385a6beb47acd9e923650615de9c03152ae663fa37c3ac" Oct 08 18:14:26 crc kubenswrapper[4988]: I1008 18:14:26.638660 4988 generic.go:334] "Generic (PLEG): container finished" podID="619228a7-ec1c-4880-a3d5-9d26ac51afd9" containerID="a72d0e017b4c6aca1bf84f74a538fe323e19c34d76f05250b2bbf4a504ff9f39" exitCode=0 Oct 08 18:14:26 crc kubenswrapper[4988]: I1008 18:14:26.639690 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-lm95x" event={"ID":"619228a7-ec1c-4880-a3d5-9d26ac51afd9","Type":"ContainerDied","Data":"a72d0e017b4c6aca1bf84f74a538fe323e19c34d76f05250b2bbf4a504ff9f39"} Oct 08 18:14:26 crc kubenswrapper[4988]: I1008 18:14:26.639774 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-lm95x" event={"ID":"619228a7-ec1c-4880-a3d5-9d26ac51afd9","Type":"ContainerDied","Data":"17c3fd14341d22847cb79f4347c7968d7f1d1bb4b7299641d10e5782790b0e7a"} Oct 08 18:14:26 crc kubenswrapper[4988]: I1008 18:14:26.639730 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-lm95x" Oct 08 18:14:26 crc kubenswrapper[4988]: I1008 18:14:26.665096 4988 scope.go:117] "RemoveContainer" containerID="bc9d8017bf8452b17dcbc32859ed1bf714b54116c673f5a0c667fb167664cd8f" Oct 08 18:14:26 crc kubenswrapper[4988]: I1008 18:14:26.682923 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-lm95x"] Oct 08 18:14:26 crc kubenswrapper[4988]: I1008 18:14:26.689121 4988 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-lm95x"] Oct 08 18:14:26 crc kubenswrapper[4988]: I1008 18:14:26.693414 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-pk7fz" Oct 08 18:14:26 crc kubenswrapper[4988]: I1008 18:14:26.693468 4988 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-pk7fz" Oct 08 18:14:26 crc kubenswrapper[4988]: I1008 18:14:26.699195 4988 scope.go:117] "RemoveContainer" containerID="b08fb6dd2b021eb06be9c6729d730ae01f09f7e84a01463406306bae12ccee0d" Oct 08 18:14:26 crc kubenswrapper[4988]: I1008 18:14:26.707850 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-hvx7q"] Oct 08 18:14:26 crc kubenswrapper[4988]: I1008 18:14:26.714560 4988 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-hvx7q"] Oct 08 18:14:26 crc kubenswrapper[4988]: I1008 18:14:26.722339 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-7ck2j" Oct 08 18:14:26 crc kubenswrapper[4988]: I1008 18:14:26.723254 4988 scope.go:117] "RemoveContainer" containerID="1400ee3c44221a939e385a6beb47acd9e923650615de9c03152ae663fa37c3ac" Oct 08 18:14:26 crc kubenswrapper[4988]: E1008 18:14:26.723795 4988 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1400ee3c44221a939e385a6beb47acd9e923650615de9c03152ae663fa37c3ac\": container with ID starting with 1400ee3c44221a939e385a6beb47acd9e923650615de9c03152ae663fa37c3ac not found: ID does not exist" containerID="1400ee3c44221a939e385a6beb47acd9e923650615de9c03152ae663fa37c3ac" Oct 08 18:14:26 crc kubenswrapper[4988]: I1008 18:14:26.723843 4988 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1400ee3c44221a939e385a6beb47acd9e923650615de9c03152ae663fa37c3ac"} err="failed to get container status \"1400ee3c44221a939e385a6beb47acd9e923650615de9c03152ae663fa37c3ac\": rpc error: code = NotFound desc = could not find container \"1400ee3c44221a939e385a6beb47acd9e923650615de9c03152ae663fa37c3ac\": container with ID starting with 1400ee3c44221a939e385a6beb47acd9e923650615de9c03152ae663fa37c3ac not found: ID does not exist" Oct 08 18:14:26 crc kubenswrapper[4988]: I1008 18:14:26.723873 4988 scope.go:117] "RemoveContainer" containerID="bc9d8017bf8452b17dcbc32859ed1bf714b54116c673f5a0c667fb167664cd8f" Oct 08 18:14:26 crc kubenswrapper[4988]: E1008 18:14:26.724202 4988 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bc9d8017bf8452b17dcbc32859ed1bf714b54116c673f5a0c667fb167664cd8f\": container with ID starting with bc9d8017bf8452b17dcbc32859ed1bf714b54116c673f5a0c667fb167664cd8f not found: ID does not exist" containerID="bc9d8017bf8452b17dcbc32859ed1bf714b54116c673f5a0c667fb167664cd8f" Oct 08 18:14:26 crc kubenswrapper[4988]: I1008 18:14:26.724226 4988 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bc9d8017bf8452b17dcbc32859ed1bf714b54116c673f5a0c667fb167664cd8f"} err="failed to get container status \"bc9d8017bf8452b17dcbc32859ed1bf714b54116c673f5a0c667fb167664cd8f\": rpc error: code = NotFound desc = could not find container \"bc9d8017bf8452b17dcbc32859ed1bf714b54116c673f5a0c667fb167664cd8f\": container with ID starting with bc9d8017bf8452b17dcbc32859ed1bf714b54116c673f5a0c667fb167664cd8f not found: ID does not exist" Oct 08 18:14:26 crc kubenswrapper[4988]: I1008 18:14:26.724241 4988 scope.go:117] "RemoveContainer" containerID="b08fb6dd2b021eb06be9c6729d730ae01f09f7e84a01463406306bae12ccee0d" Oct 08 18:14:26 crc kubenswrapper[4988]: E1008 18:14:26.724711 4988 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b08fb6dd2b021eb06be9c6729d730ae01f09f7e84a01463406306bae12ccee0d\": container with ID starting with b08fb6dd2b021eb06be9c6729d730ae01f09f7e84a01463406306bae12ccee0d not found: ID does not exist" containerID="b08fb6dd2b021eb06be9c6729d730ae01f09f7e84a01463406306bae12ccee0d" Oct 08 18:14:26 crc kubenswrapper[4988]: I1008 18:14:26.724737 4988 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b08fb6dd2b021eb06be9c6729d730ae01f09f7e84a01463406306bae12ccee0d"} err="failed to get container status \"b08fb6dd2b021eb06be9c6729d730ae01f09f7e84a01463406306bae12ccee0d\": rpc error: code = NotFound desc = could not find container \"b08fb6dd2b021eb06be9c6729d730ae01f09f7e84a01463406306bae12ccee0d\": container with ID starting with b08fb6dd2b021eb06be9c6729d730ae01f09f7e84a01463406306bae12ccee0d not found: ID does not exist" Oct 08 18:14:26 crc kubenswrapper[4988]: I1008 18:14:26.724751 4988 scope.go:117] "RemoveContainer" containerID="a72d0e017b4c6aca1bf84f74a538fe323e19c34d76f05250b2bbf4a504ff9f39" Oct 08 18:14:26 crc kubenswrapper[4988]: I1008 18:14:26.749497 4988 scope.go:117] "RemoveContainer" containerID="deb583d21b801e995436d011c5a5e5a148cc267d5e298f602b094b0bb42df73f" Oct 08 18:14:26 crc kubenswrapper[4988]: I1008 18:14:26.806397 4988 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-pk7fz" Oct 08 18:14:26 crc kubenswrapper[4988]: I1008 18:14:26.816427 4988 scope.go:117] "RemoveContainer" containerID="2e4478d7c154278d486e98dbbd0498fd59742508545a14684e692ef91d6c0c0b" Oct 08 18:14:26 crc kubenswrapper[4988]: I1008 18:14:26.834324 4988 scope.go:117] "RemoveContainer" containerID="a72d0e017b4c6aca1bf84f74a538fe323e19c34d76f05250b2bbf4a504ff9f39" Oct 08 18:14:26 crc kubenswrapper[4988]: E1008 18:14:26.835006 4988 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a72d0e017b4c6aca1bf84f74a538fe323e19c34d76f05250b2bbf4a504ff9f39\": container with ID starting with a72d0e017b4c6aca1bf84f74a538fe323e19c34d76f05250b2bbf4a504ff9f39 not found: ID does not exist" containerID="a72d0e017b4c6aca1bf84f74a538fe323e19c34d76f05250b2bbf4a504ff9f39" Oct 08 18:14:26 crc kubenswrapper[4988]: I1008 18:14:26.835054 4988 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a72d0e017b4c6aca1bf84f74a538fe323e19c34d76f05250b2bbf4a504ff9f39"} err="failed to get container status \"a72d0e017b4c6aca1bf84f74a538fe323e19c34d76f05250b2bbf4a504ff9f39\": rpc error: code = NotFound desc = could not find container \"a72d0e017b4c6aca1bf84f74a538fe323e19c34d76f05250b2bbf4a504ff9f39\": container with ID starting with a72d0e017b4c6aca1bf84f74a538fe323e19c34d76f05250b2bbf4a504ff9f39 not found: ID does not exist" Oct 08 18:14:26 crc kubenswrapper[4988]: I1008 18:14:26.835088 4988 scope.go:117] "RemoveContainer" containerID="deb583d21b801e995436d011c5a5e5a148cc267d5e298f602b094b0bb42df73f" Oct 08 18:14:26 crc kubenswrapper[4988]: E1008 18:14:26.835779 4988 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"deb583d21b801e995436d011c5a5e5a148cc267d5e298f602b094b0bb42df73f\": container with ID starting with deb583d21b801e995436d011c5a5e5a148cc267d5e298f602b094b0bb42df73f not found: ID does not exist" containerID="deb583d21b801e995436d011c5a5e5a148cc267d5e298f602b094b0bb42df73f" Oct 08 18:14:26 crc kubenswrapper[4988]: I1008 18:14:26.835866 4988 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"deb583d21b801e995436d011c5a5e5a148cc267d5e298f602b094b0bb42df73f"} err="failed to get container status \"deb583d21b801e995436d011c5a5e5a148cc267d5e298f602b094b0bb42df73f\": rpc error: code = NotFound desc = could not find container \"deb583d21b801e995436d011c5a5e5a148cc267d5e298f602b094b0bb42df73f\": container with ID starting with deb583d21b801e995436d011c5a5e5a148cc267d5e298f602b094b0bb42df73f not found: ID does not exist" Oct 08 18:14:26 crc kubenswrapper[4988]: I1008 18:14:26.835914 4988 scope.go:117] "RemoveContainer" containerID="2e4478d7c154278d486e98dbbd0498fd59742508545a14684e692ef91d6c0c0b" Oct 08 18:14:26 crc kubenswrapper[4988]: E1008 18:14:26.836358 4988 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2e4478d7c154278d486e98dbbd0498fd59742508545a14684e692ef91d6c0c0b\": container with ID starting with 2e4478d7c154278d486e98dbbd0498fd59742508545a14684e692ef91d6c0c0b not found: ID does not exist" containerID="2e4478d7c154278d486e98dbbd0498fd59742508545a14684e692ef91d6c0c0b" Oct 08 18:14:26 crc kubenswrapper[4988]: I1008 18:14:26.836406 4988 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2e4478d7c154278d486e98dbbd0498fd59742508545a14684e692ef91d6c0c0b"} err="failed to get container status \"2e4478d7c154278d486e98dbbd0498fd59742508545a14684e692ef91d6c0c0b\": rpc error: code = NotFound desc = could not find container \"2e4478d7c154278d486e98dbbd0498fd59742508545a14684e692ef91d6c0c0b\": container with ID starting with 2e4478d7c154278d486e98dbbd0498fd59742508545a14684e692ef91d6c0c0b not found: ID does not exist" Oct 08 18:14:27 crc kubenswrapper[4988]: I1008 18:14:27.244428 4988 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4eb14139-7a10-4429-9834-378a89ea8255" path="/var/lib/kubelet/pods/4eb14139-7a10-4429-9834-378a89ea8255/volumes" Oct 08 18:14:27 crc kubenswrapper[4988]: I1008 18:14:27.245754 4988 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="619228a7-ec1c-4880-a3d5-9d26ac51afd9" path="/var/lib/kubelet/pods/619228a7-ec1c-4880-a3d5-9d26ac51afd9/volumes" Oct 08 18:14:27 crc kubenswrapper[4988]: I1008 18:14:27.680789 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-w4qxm"] Oct 08 18:14:27 crc kubenswrapper[4988]: I1008 18:14:27.697018 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-pk7fz" Oct 08 18:14:28 crc kubenswrapper[4988]: I1008 18:14:28.653584 4988 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-w4qxm" podUID="56027983-5f79-4555-abb9-4ab9f13cd39f" containerName="registry-server" containerID="cri-o://73ce37d8d4e252e6d55f78380d048b1ce9aeea2857a615a8fbbbb1f5ce221565" gracePeriod=2 Oct 08 18:14:30 crc kubenswrapper[4988]: I1008 18:14:30.279052 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-pk7fz"] Oct 08 18:14:30 crc kubenswrapper[4988]: I1008 18:14:30.279898 4988 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-pk7fz" podUID="e6bfa362-455b-422e-9c92-dc81aaa1e0ec" containerName="registry-server" containerID="cri-o://59bc1e445020368043ac5535d5c5e63a263cabb59e7f76532191cbdc959bdc70" gracePeriod=2 Oct 08 18:14:30 crc kubenswrapper[4988]: I1008 18:14:30.308356 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-w4qxm" Oct 08 18:14:30 crc kubenswrapper[4988]: I1008 18:14:30.383716 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4jxl8\" (UniqueName: \"kubernetes.io/projected/56027983-5f79-4555-abb9-4ab9f13cd39f-kube-api-access-4jxl8\") pod \"56027983-5f79-4555-abb9-4ab9f13cd39f\" (UID: \"56027983-5f79-4555-abb9-4ab9f13cd39f\") " Oct 08 18:14:30 crc kubenswrapper[4988]: I1008 18:14:30.384022 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/56027983-5f79-4555-abb9-4ab9f13cd39f-catalog-content\") pod \"56027983-5f79-4555-abb9-4ab9f13cd39f\" (UID: \"56027983-5f79-4555-abb9-4ab9f13cd39f\") " Oct 08 18:14:30 crc kubenswrapper[4988]: I1008 18:14:30.384053 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/56027983-5f79-4555-abb9-4ab9f13cd39f-utilities\") pod \"56027983-5f79-4555-abb9-4ab9f13cd39f\" (UID: \"56027983-5f79-4555-abb9-4ab9f13cd39f\") " Oct 08 18:14:30 crc kubenswrapper[4988]: I1008 18:14:30.384876 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/56027983-5f79-4555-abb9-4ab9f13cd39f-utilities" (OuterVolumeSpecName: "utilities") pod "56027983-5f79-4555-abb9-4ab9f13cd39f" (UID: "56027983-5f79-4555-abb9-4ab9f13cd39f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 18:14:30 crc kubenswrapper[4988]: I1008 18:14:30.393547 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/56027983-5f79-4555-abb9-4ab9f13cd39f-kube-api-access-4jxl8" (OuterVolumeSpecName: "kube-api-access-4jxl8") pod "56027983-5f79-4555-abb9-4ab9f13cd39f" (UID: "56027983-5f79-4555-abb9-4ab9f13cd39f"). InnerVolumeSpecName "kube-api-access-4jxl8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:14:30 crc kubenswrapper[4988]: I1008 18:14:30.401344 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/56027983-5f79-4555-abb9-4ab9f13cd39f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "56027983-5f79-4555-abb9-4ab9f13cd39f" (UID: "56027983-5f79-4555-abb9-4ab9f13cd39f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 18:14:30 crc kubenswrapper[4988]: I1008 18:14:30.485042 4988 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4jxl8\" (UniqueName: \"kubernetes.io/projected/56027983-5f79-4555-abb9-4ab9f13cd39f-kube-api-access-4jxl8\") on node \"crc\" DevicePath \"\"" Oct 08 18:14:30 crc kubenswrapper[4988]: I1008 18:14:30.485077 4988 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/56027983-5f79-4555-abb9-4ab9f13cd39f-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 08 18:14:30 crc kubenswrapper[4988]: I1008 18:14:30.485091 4988 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/56027983-5f79-4555-abb9-4ab9f13cd39f-utilities\") on node \"crc\" DevicePath \"\"" Oct 08 18:14:30 crc kubenswrapper[4988]: I1008 18:14:30.642305 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-pk7fz" Oct 08 18:14:30 crc kubenswrapper[4988]: I1008 18:14:30.671645 4988 generic.go:334] "Generic (PLEG): container finished" podID="56027983-5f79-4555-abb9-4ab9f13cd39f" containerID="73ce37d8d4e252e6d55f78380d048b1ce9aeea2857a615a8fbbbb1f5ce221565" exitCode=0 Oct 08 18:14:30 crc kubenswrapper[4988]: I1008 18:14:30.671726 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-w4qxm" Oct 08 18:14:30 crc kubenswrapper[4988]: I1008 18:14:30.671742 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-w4qxm" event={"ID":"56027983-5f79-4555-abb9-4ab9f13cd39f","Type":"ContainerDied","Data":"73ce37d8d4e252e6d55f78380d048b1ce9aeea2857a615a8fbbbb1f5ce221565"} Oct 08 18:14:30 crc kubenswrapper[4988]: I1008 18:14:30.672344 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-w4qxm" event={"ID":"56027983-5f79-4555-abb9-4ab9f13cd39f","Type":"ContainerDied","Data":"52b79f26119c72dd510a0f961e5d09437cc12d0a9fb8fc02d8339db6f24ec76b"} Oct 08 18:14:30 crc kubenswrapper[4988]: I1008 18:14:30.672433 4988 scope.go:117] "RemoveContainer" containerID="73ce37d8d4e252e6d55f78380d048b1ce9aeea2857a615a8fbbbb1f5ce221565" Oct 08 18:14:30 crc kubenswrapper[4988]: I1008 18:14:30.674208 4988 generic.go:334] "Generic (PLEG): container finished" podID="e6bfa362-455b-422e-9c92-dc81aaa1e0ec" containerID="59bc1e445020368043ac5535d5c5e63a263cabb59e7f76532191cbdc959bdc70" exitCode=0 Oct 08 18:14:30 crc kubenswrapper[4988]: I1008 18:14:30.674237 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-pk7fz" event={"ID":"e6bfa362-455b-422e-9c92-dc81aaa1e0ec","Type":"ContainerDied","Data":"59bc1e445020368043ac5535d5c5e63a263cabb59e7f76532191cbdc959bdc70"} Oct 08 18:14:30 crc kubenswrapper[4988]: I1008 18:14:30.674262 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-pk7fz" event={"ID":"e6bfa362-455b-422e-9c92-dc81aaa1e0ec","Type":"ContainerDied","Data":"d22a816c0113f7c9d6ada62d387385068dcbac632ddac86cee376be6a3a61910"} Oct 08 18:14:30 crc kubenswrapper[4988]: I1008 18:14:30.674328 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-pk7fz" Oct 08 18:14:30 crc kubenswrapper[4988]: I1008 18:14:30.687053 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e6bfa362-455b-422e-9c92-dc81aaa1e0ec-utilities\") pod \"e6bfa362-455b-422e-9c92-dc81aaa1e0ec\" (UID: \"e6bfa362-455b-422e-9c92-dc81aaa1e0ec\") " Oct 08 18:14:30 crc kubenswrapper[4988]: I1008 18:14:30.687132 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r4rld\" (UniqueName: \"kubernetes.io/projected/e6bfa362-455b-422e-9c92-dc81aaa1e0ec-kube-api-access-r4rld\") pod \"e6bfa362-455b-422e-9c92-dc81aaa1e0ec\" (UID: \"e6bfa362-455b-422e-9c92-dc81aaa1e0ec\") " Oct 08 18:14:30 crc kubenswrapper[4988]: I1008 18:14:30.687234 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e6bfa362-455b-422e-9c92-dc81aaa1e0ec-catalog-content\") pod \"e6bfa362-455b-422e-9c92-dc81aaa1e0ec\" (UID: \"e6bfa362-455b-422e-9c92-dc81aaa1e0ec\") " Oct 08 18:14:30 crc kubenswrapper[4988]: I1008 18:14:30.688036 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e6bfa362-455b-422e-9c92-dc81aaa1e0ec-utilities" (OuterVolumeSpecName: "utilities") pod "e6bfa362-455b-422e-9c92-dc81aaa1e0ec" (UID: "e6bfa362-455b-422e-9c92-dc81aaa1e0ec"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 18:14:30 crc kubenswrapper[4988]: I1008 18:14:30.690212 4988 scope.go:117] "RemoveContainer" containerID="47761230f30576c443259c7106cd6271733fa977159af6c4ef87a9f6c583107c" Oct 08 18:14:30 crc kubenswrapper[4988]: I1008 18:14:30.691649 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e6bfa362-455b-422e-9c92-dc81aaa1e0ec-kube-api-access-r4rld" (OuterVolumeSpecName: "kube-api-access-r4rld") pod "e6bfa362-455b-422e-9c92-dc81aaa1e0ec" (UID: "e6bfa362-455b-422e-9c92-dc81aaa1e0ec"). InnerVolumeSpecName "kube-api-access-r4rld". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:14:30 crc kubenswrapper[4988]: I1008 18:14:30.723368 4988 scope.go:117] "RemoveContainer" containerID="f97986ccfa87d32adc71e1c583cb4e49eeb9013ae85bff9c98dc432d1bc57d60" Oct 08 18:14:30 crc kubenswrapper[4988]: I1008 18:14:30.726792 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-w4qxm"] Oct 08 18:14:30 crc kubenswrapper[4988]: I1008 18:14:30.729492 4988 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-w4qxm"] Oct 08 18:14:30 crc kubenswrapper[4988]: I1008 18:14:30.755999 4988 scope.go:117] "RemoveContainer" containerID="73ce37d8d4e252e6d55f78380d048b1ce9aeea2857a615a8fbbbb1f5ce221565" Oct 08 18:14:30 crc kubenswrapper[4988]: E1008 18:14:30.756507 4988 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"73ce37d8d4e252e6d55f78380d048b1ce9aeea2857a615a8fbbbb1f5ce221565\": container with ID starting with 73ce37d8d4e252e6d55f78380d048b1ce9aeea2857a615a8fbbbb1f5ce221565 not found: ID does not exist" containerID="73ce37d8d4e252e6d55f78380d048b1ce9aeea2857a615a8fbbbb1f5ce221565" Oct 08 18:14:30 crc kubenswrapper[4988]: I1008 18:14:30.756564 4988 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"73ce37d8d4e252e6d55f78380d048b1ce9aeea2857a615a8fbbbb1f5ce221565"} err="failed to get container status \"73ce37d8d4e252e6d55f78380d048b1ce9aeea2857a615a8fbbbb1f5ce221565\": rpc error: code = NotFound desc = could not find container \"73ce37d8d4e252e6d55f78380d048b1ce9aeea2857a615a8fbbbb1f5ce221565\": container with ID starting with 73ce37d8d4e252e6d55f78380d048b1ce9aeea2857a615a8fbbbb1f5ce221565 not found: ID does not exist" Oct 08 18:14:30 crc kubenswrapper[4988]: I1008 18:14:30.756603 4988 scope.go:117] "RemoveContainer" containerID="47761230f30576c443259c7106cd6271733fa977159af6c4ef87a9f6c583107c" Oct 08 18:14:30 crc kubenswrapper[4988]: E1008 18:14:30.756925 4988 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"47761230f30576c443259c7106cd6271733fa977159af6c4ef87a9f6c583107c\": container with ID starting with 47761230f30576c443259c7106cd6271733fa977159af6c4ef87a9f6c583107c not found: ID does not exist" containerID="47761230f30576c443259c7106cd6271733fa977159af6c4ef87a9f6c583107c" Oct 08 18:14:30 crc kubenswrapper[4988]: I1008 18:14:30.756946 4988 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"47761230f30576c443259c7106cd6271733fa977159af6c4ef87a9f6c583107c"} err="failed to get container status \"47761230f30576c443259c7106cd6271733fa977159af6c4ef87a9f6c583107c\": rpc error: code = NotFound desc = could not find container \"47761230f30576c443259c7106cd6271733fa977159af6c4ef87a9f6c583107c\": container with ID starting with 47761230f30576c443259c7106cd6271733fa977159af6c4ef87a9f6c583107c not found: ID does not exist" Oct 08 18:14:30 crc kubenswrapper[4988]: I1008 18:14:30.756962 4988 scope.go:117] "RemoveContainer" containerID="f97986ccfa87d32adc71e1c583cb4e49eeb9013ae85bff9c98dc432d1bc57d60" Oct 08 18:14:30 crc kubenswrapper[4988]: E1008 18:14:30.757317 4988 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f97986ccfa87d32adc71e1c583cb4e49eeb9013ae85bff9c98dc432d1bc57d60\": container with ID starting with f97986ccfa87d32adc71e1c583cb4e49eeb9013ae85bff9c98dc432d1bc57d60 not found: ID does not exist" containerID="f97986ccfa87d32adc71e1c583cb4e49eeb9013ae85bff9c98dc432d1bc57d60" Oct 08 18:14:30 crc kubenswrapper[4988]: I1008 18:14:30.757338 4988 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f97986ccfa87d32adc71e1c583cb4e49eeb9013ae85bff9c98dc432d1bc57d60"} err="failed to get container status \"f97986ccfa87d32adc71e1c583cb4e49eeb9013ae85bff9c98dc432d1bc57d60\": rpc error: code = NotFound desc = could not find container \"f97986ccfa87d32adc71e1c583cb4e49eeb9013ae85bff9c98dc432d1bc57d60\": container with ID starting with f97986ccfa87d32adc71e1c583cb4e49eeb9013ae85bff9c98dc432d1bc57d60 not found: ID does not exist" Oct 08 18:14:30 crc kubenswrapper[4988]: I1008 18:14:30.757355 4988 scope.go:117] "RemoveContainer" containerID="59bc1e445020368043ac5535d5c5e63a263cabb59e7f76532191cbdc959bdc70" Oct 08 18:14:30 crc kubenswrapper[4988]: I1008 18:14:30.770838 4988 scope.go:117] "RemoveContainer" containerID="1c352caeee323da70f697f14e91dcb1dacba6a2f54dc6e212cda84c95e3ecb24" Oct 08 18:14:30 crc kubenswrapper[4988]: I1008 18:14:30.795023 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e6bfa362-455b-422e-9c92-dc81aaa1e0ec-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "e6bfa362-455b-422e-9c92-dc81aaa1e0ec" (UID: "e6bfa362-455b-422e-9c92-dc81aaa1e0ec"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 18:14:30 crc kubenswrapper[4988]: I1008 18:14:30.795682 4988 scope.go:117] "RemoveContainer" containerID="2fd015d5d9120268a081acbca954f7e8f24e879b1c3d61a8d0b89b492bccc7a4" Oct 08 18:14:30 crc kubenswrapper[4988]: I1008 18:14:30.796107 4988 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e6bfa362-455b-422e-9c92-dc81aaa1e0ec-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 08 18:14:30 crc kubenswrapper[4988]: I1008 18:14:30.796134 4988 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e6bfa362-455b-422e-9c92-dc81aaa1e0ec-utilities\") on node \"crc\" DevicePath \"\"" Oct 08 18:14:30 crc kubenswrapper[4988]: I1008 18:14:30.796146 4988 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r4rld\" (UniqueName: \"kubernetes.io/projected/e6bfa362-455b-422e-9c92-dc81aaa1e0ec-kube-api-access-r4rld\") on node \"crc\" DevicePath \"\"" Oct 08 18:14:30 crc kubenswrapper[4988]: I1008 18:14:30.829290 4988 scope.go:117] "RemoveContainer" containerID="59bc1e445020368043ac5535d5c5e63a263cabb59e7f76532191cbdc959bdc70" Oct 08 18:14:30 crc kubenswrapper[4988]: E1008 18:14:30.831044 4988 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"59bc1e445020368043ac5535d5c5e63a263cabb59e7f76532191cbdc959bdc70\": container with ID starting with 59bc1e445020368043ac5535d5c5e63a263cabb59e7f76532191cbdc959bdc70 not found: ID does not exist" containerID="59bc1e445020368043ac5535d5c5e63a263cabb59e7f76532191cbdc959bdc70" Oct 08 18:14:30 crc kubenswrapper[4988]: I1008 18:14:30.831094 4988 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"59bc1e445020368043ac5535d5c5e63a263cabb59e7f76532191cbdc959bdc70"} err="failed to get container status \"59bc1e445020368043ac5535d5c5e63a263cabb59e7f76532191cbdc959bdc70\": rpc error: code = NotFound desc = could not find container \"59bc1e445020368043ac5535d5c5e63a263cabb59e7f76532191cbdc959bdc70\": container with ID starting with 59bc1e445020368043ac5535d5c5e63a263cabb59e7f76532191cbdc959bdc70 not found: ID does not exist" Oct 08 18:14:30 crc kubenswrapper[4988]: I1008 18:14:30.831129 4988 scope.go:117] "RemoveContainer" containerID="1c352caeee323da70f697f14e91dcb1dacba6a2f54dc6e212cda84c95e3ecb24" Oct 08 18:14:30 crc kubenswrapper[4988]: E1008 18:14:30.831895 4988 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1c352caeee323da70f697f14e91dcb1dacba6a2f54dc6e212cda84c95e3ecb24\": container with ID starting with 1c352caeee323da70f697f14e91dcb1dacba6a2f54dc6e212cda84c95e3ecb24 not found: ID does not exist" containerID="1c352caeee323da70f697f14e91dcb1dacba6a2f54dc6e212cda84c95e3ecb24" Oct 08 18:14:30 crc kubenswrapper[4988]: I1008 18:14:30.831920 4988 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1c352caeee323da70f697f14e91dcb1dacba6a2f54dc6e212cda84c95e3ecb24"} err="failed to get container status \"1c352caeee323da70f697f14e91dcb1dacba6a2f54dc6e212cda84c95e3ecb24\": rpc error: code = NotFound desc = could not find container \"1c352caeee323da70f697f14e91dcb1dacba6a2f54dc6e212cda84c95e3ecb24\": container with ID starting with 1c352caeee323da70f697f14e91dcb1dacba6a2f54dc6e212cda84c95e3ecb24 not found: ID does not exist" Oct 08 18:14:30 crc kubenswrapper[4988]: I1008 18:14:30.831936 4988 scope.go:117] "RemoveContainer" containerID="2fd015d5d9120268a081acbca954f7e8f24e879b1c3d61a8d0b89b492bccc7a4" Oct 08 18:14:30 crc kubenswrapper[4988]: E1008 18:14:30.832197 4988 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2fd015d5d9120268a081acbca954f7e8f24e879b1c3d61a8d0b89b492bccc7a4\": container with ID starting with 2fd015d5d9120268a081acbca954f7e8f24e879b1c3d61a8d0b89b492bccc7a4 not found: ID does not exist" containerID="2fd015d5d9120268a081acbca954f7e8f24e879b1c3d61a8d0b89b492bccc7a4" Oct 08 18:14:30 crc kubenswrapper[4988]: I1008 18:14:30.832224 4988 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2fd015d5d9120268a081acbca954f7e8f24e879b1c3d61a8d0b89b492bccc7a4"} err="failed to get container status \"2fd015d5d9120268a081acbca954f7e8f24e879b1c3d61a8d0b89b492bccc7a4\": rpc error: code = NotFound desc = could not find container \"2fd015d5d9120268a081acbca954f7e8f24e879b1c3d61a8d0b89b492bccc7a4\": container with ID starting with 2fd015d5d9120268a081acbca954f7e8f24e879b1c3d61a8d0b89b492bccc7a4 not found: ID does not exist" Oct 08 18:14:31 crc kubenswrapper[4988]: I1008 18:14:31.003301 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-pk7fz"] Oct 08 18:14:31 crc kubenswrapper[4988]: I1008 18:14:31.006191 4988 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-pk7fz"] Oct 08 18:14:31 crc kubenswrapper[4988]: I1008 18:14:31.245051 4988 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="56027983-5f79-4555-abb9-4ab9f13cd39f" path="/var/lib/kubelet/pods/56027983-5f79-4555-abb9-4ab9f13cd39f/volumes" Oct 08 18:14:31 crc kubenswrapper[4988]: I1008 18:14:31.245958 4988 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e6bfa362-455b-422e-9c92-dc81aaa1e0ec" path="/var/lib/kubelet/pods/e6bfa362-455b-422e-9c92-dc81aaa1e0ec/volumes" Oct 08 18:14:47 crc kubenswrapper[4988]: I1008 18:14:47.634597 4988 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-authentication/oauth-openshift-558db77b4-hwtzl" podUID="e294e9c8-16f2-46d4-91f8-5964b1e8ea5a" containerName="oauth-openshift" containerID="cri-o://5bf6c97c2a92106c5d151cc199928afbbcc2d70b4025fb7cd24058e7da3b8157" gracePeriod=15 Oct 08 18:14:47 crc kubenswrapper[4988]: I1008 18:14:47.790262 4988 generic.go:334] "Generic (PLEG): container finished" podID="e294e9c8-16f2-46d4-91f8-5964b1e8ea5a" containerID="5bf6c97c2a92106c5d151cc199928afbbcc2d70b4025fb7cd24058e7da3b8157" exitCode=0 Oct 08 18:14:47 crc kubenswrapper[4988]: I1008 18:14:47.790330 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-hwtzl" event={"ID":"e294e9c8-16f2-46d4-91f8-5964b1e8ea5a","Type":"ContainerDied","Data":"5bf6c97c2a92106c5d151cc199928afbbcc2d70b4025fb7cd24058e7da3b8157"} Oct 08 18:14:48 crc kubenswrapper[4988]: I1008 18:14:48.071761 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-hwtzl" Oct 08 18:14:48 crc kubenswrapper[4988]: I1008 18:14:48.116743 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-79656f7ff7-kckxs"] Oct 08 18:14:48 crc kubenswrapper[4988]: E1008 18:14:48.117036 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4eb14139-7a10-4429-9834-378a89ea8255" containerName="extract-utilities" Oct 08 18:14:48 crc kubenswrapper[4988]: I1008 18:14:48.117051 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="4eb14139-7a10-4429-9834-378a89ea8255" containerName="extract-utilities" Oct 08 18:14:48 crc kubenswrapper[4988]: E1008 18:14:48.117061 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e6bfa362-455b-422e-9c92-dc81aaa1e0ec" containerName="extract-content" Oct 08 18:14:48 crc kubenswrapper[4988]: I1008 18:14:48.117071 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="e6bfa362-455b-422e-9c92-dc81aaa1e0ec" containerName="extract-content" Oct 08 18:14:48 crc kubenswrapper[4988]: E1008 18:14:48.117128 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4eb14139-7a10-4429-9834-378a89ea8255" containerName="registry-server" Oct 08 18:14:48 crc kubenswrapper[4988]: I1008 18:14:48.117137 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="4eb14139-7a10-4429-9834-378a89ea8255" containerName="registry-server" Oct 08 18:14:48 crc kubenswrapper[4988]: E1008 18:14:48.117151 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d24735bb-2d72-44e0-9a5b-b220a787e116" containerName="pruner" Oct 08 18:14:48 crc kubenswrapper[4988]: I1008 18:14:48.117160 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="d24735bb-2d72-44e0-9a5b-b220a787e116" containerName="pruner" Oct 08 18:14:48 crc kubenswrapper[4988]: E1008 18:14:48.117181 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="56027983-5f79-4555-abb9-4ab9f13cd39f" containerName="registry-server" Oct 08 18:14:48 crc kubenswrapper[4988]: I1008 18:14:48.117190 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="56027983-5f79-4555-abb9-4ab9f13cd39f" containerName="registry-server" Oct 08 18:14:48 crc kubenswrapper[4988]: E1008 18:14:48.117202 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e6bfa362-455b-422e-9c92-dc81aaa1e0ec" containerName="extract-utilities" Oct 08 18:14:48 crc kubenswrapper[4988]: I1008 18:14:48.117211 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="e6bfa362-455b-422e-9c92-dc81aaa1e0ec" containerName="extract-utilities" Oct 08 18:14:48 crc kubenswrapper[4988]: E1008 18:14:48.117224 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7f4c229e-0791-4953-aa04-5928f5b1480f" containerName="pruner" Oct 08 18:14:48 crc kubenswrapper[4988]: I1008 18:14:48.117232 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="7f4c229e-0791-4953-aa04-5928f5b1480f" containerName="pruner" Oct 08 18:14:48 crc kubenswrapper[4988]: E1008 18:14:48.117243 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4eb14139-7a10-4429-9834-378a89ea8255" containerName="extract-content" Oct 08 18:14:48 crc kubenswrapper[4988]: I1008 18:14:48.117251 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="4eb14139-7a10-4429-9834-378a89ea8255" containerName="extract-content" Oct 08 18:14:48 crc kubenswrapper[4988]: E1008 18:14:48.117261 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e6bfa362-455b-422e-9c92-dc81aaa1e0ec" containerName="registry-server" Oct 08 18:14:48 crc kubenswrapper[4988]: I1008 18:14:48.117269 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="e6bfa362-455b-422e-9c92-dc81aaa1e0ec" containerName="registry-server" Oct 08 18:14:48 crc kubenswrapper[4988]: E1008 18:14:48.117279 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="56027983-5f79-4555-abb9-4ab9f13cd39f" containerName="extract-content" Oct 08 18:14:48 crc kubenswrapper[4988]: I1008 18:14:48.117315 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="56027983-5f79-4555-abb9-4ab9f13cd39f" containerName="extract-content" Oct 08 18:14:48 crc kubenswrapper[4988]: E1008 18:14:48.117330 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="619228a7-ec1c-4880-a3d5-9d26ac51afd9" containerName="registry-server" Oct 08 18:14:48 crc kubenswrapper[4988]: I1008 18:14:48.117338 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="619228a7-ec1c-4880-a3d5-9d26ac51afd9" containerName="registry-server" Oct 08 18:14:48 crc kubenswrapper[4988]: E1008 18:14:48.117352 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e294e9c8-16f2-46d4-91f8-5964b1e8ea5a" containerName="oauth-openshift" Oct 08 18:14:48 crc kubenswrapper[4988]: I1008 18:14:48.117360 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="e294e9c8-16f2-46d4-91f8-5964b1e8ea5a" containerName="oauth-openshift" Oct 08 18:14:48 crc kubenswrapper[4988]: E1008 18:14:48.117374 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="619228a7-ec1c-4880-a3d5-9d26ac51afd9" containerName="extract-utilities" Oct 08 18:14:48 crc kubenswrapper[4988]: I1008 18:14:48.117382 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="619228a7-ec1c-4880-a3d5-9d26ac51afd9" containerName="extract-utilities" Oct 08 18:14:48 crc kubenswrapper[4988]: E1008 18:14:48.117418 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="56027983-5f79-4555-abb9-4ab9f13cd39f" containerName="extract-utilities" Oct 08 18:14:48 crc kubenswrapper[4988]: I1008 18:14:48.117426 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="56027983-5f79-4555-abb9-4ab9f13cd39f" containerName="extract-utilities" Oct 08 18:14:48 crc kubenswrapper[4988]: E1008 18:14:48.117439 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="619228a7-ec1c-4880-a3d5-9d26ac51afd9" containerName="extract-content" Oct 08 18:14:48 crc kubenswrapper[4988]: I1008 18:14:48.117447 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="619228a7-ec1c-4880-a3d5-9d26ac51afd9" containerName="extract-content" Oct 08 18:14:48 crc kubenswrapper[4988]: I1008 18:14:48.117614 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="e294e9c8-16f2-46d4-91f8-5964b1e8ea5a" containerName="oauth-openshift" Oct 08 18:14:48 crc kubenswrapper[4988]: I1008 18:14:48.117631 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="7f4c229e-0791-4953-aa04-5928f5b1480f" containerName="pruner" Oct 08 18:14:48 crc kubenswrapper[4988]: I1008 18:14:48.117644 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="619228a7-ec1c-4880-a3d5-9d26ac51afd9" containerName="registry-server" Oct 08 18:14:48 crc kubenswrapper[4988]: I1008 18:14:48.117656 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="4eb14139-7a10-4429-9834-378a89ea8255" containerName="registry-server" Oct 08 18:14:48 crc kubenswrapper[4988]: I1008 18:14:48.117670 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="56027983-5f79-4555-abb9-4ab9f13cd39f" containerName="registry-server" Oct 08 18:14:48 crc kubenswrapper[4988]: I1008 18:14:48.117685 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="e6bfa362-455b-422e-9c92-dc81aaa1e0ec" containerName="registry-server" Oct 08 18:14:48 crc kubenswrapper[4988]: I1008 18:14:48.117696 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="d24735bb-2d72-44e0-9a5b-b220a787e116" containerName="pruner" Oct 08 18:14:48 crc kubenswrapper[4988]: I1008 18:14:48.118187 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-79656f7ff7-kckxs" Oct 08 18:14:48 crc kubenswrapper[4988]: I1008 18:14:48.131020 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-79656f7ff7-kckxs"] Oct 08 18:14:48 crc kubenswrapper[4988]: I1008 18:14:48.134074 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/e294e9c8-16f2-46d4-91f8-5964b1e8ea5a-audit-policies\") pod \"e294e9c8-16f2-46d4-91f8-5964b1e8ea5a\" (UID: \"e294e9c8-16f2-46d4-91f8-5964b1e8ea5a\") " Oct 08 18:14:48 crc kubenswrapper[4988]: I1008 18:14:48.134589 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/e294e9c8-16f2-46d4-91f8-5964b1e8ea5a-v4-0-config-system-session\") pod \"e294e9c8-16f2-46d4-91f8-5964b1e8ea5a\" (UID: \"e294e9c8-16f2-46d4-91f8-5964b1e8ea5a\") " Oct 08 18:14:48 crc kubenswrapper[4988]: I1008 18:14:48.134694 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/e294e9c8-16f2-46d4-91f8-5964b1e8ea5a-v4-0-config-user-template-provider-selection\") pod \"e294e9c8-16f2-46d4-91f8-5964b1e8ea5a\" (UID: \"e294e9c8-16f2-46d4-91f8-5964b1e8ea5a\") " Oct 08 18:14:48 crc kubenswrapper[4988]: I1008 18:14:48.134807 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/e294e9c8-16f2-46d4-91f8-5964b1e8ea5a-v4-0-config-system-serving-cert\") pod \"e294e9c8-16f2-46d4-91f8-5964b1e8ea5a\" (UID: \"e294e9c8-16f2-46d4-91f8-5964b1e8ea5a\") " Oct 08 18:14:48 crc kubenswrapper[4988]: I1008 18:14:48.134910 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/e294e9c8-16f2-46d4-91f8-5964b1e8ea5a-v4-0-config-user-template-error\") pod \"e294e9c8-16f2-46d4-91f8-5964b1e8ea5a\" (UID: \"e294e9c8-16f2-46d4-91f8-5964b1e8ea5a\") " Oct 08 18:14:48 crc kubenswrapper[4988]: I1008 18:14:48.135009 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/e294e9c8-16f2-46d4-91f8-5964b1e8ea5a-v4-0-config-system-router-certs\") pod \"e294e9c8-16f2-46d4-91f8-5964b1e8ea5a\" (UID: \"e294e9c8-16f2-46d4-91f8-5964b1e8ea5a\") " Oct 08 18:14:48 crc kubenswrapper[4988]: I1008 18:14:48.135194 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/e294e9c8-16f2-46d4-91f8-5964b1e8ea5a-v4-0-config-system-cliconfig\") pod \"e294e9c8-16f2-46d4-91f8-5964b1e8ea5a\" (UID: \"e294e9c8-16f2-46d4-91f8-5964b1e8ea5a\") " Oct 08 18:14:48 crc kubenswrapper[4988]: I1008 18:14:48.135324 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6r9cj\" (UniqueName: \"kubernetes.io/projected/e294e9c8-16f2-46d4-91f8-5964b1e8ea5a-kube-api-access-6r9cj\") pod \"e294e9c8-16f2-46d4-91f8-5964b1e8ea5a\" (UID: \"e294e9c8-16f2-46d4-91f8-5964b1e8ea5a\") " Oct 08 18:14:48 crc kubenswrapper[4988]: I1008 18:14:48.135489 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/e294e9c8-16f2-46d4-91f8-5964b1e8ea5a-v4-0-config-user-template-login\") pod \"e294e9c8-16f2-46d4-91f8-5964b1e8ea5a\" (UID: \"e294e9c8-16f2-46d4-91f8-5964b1e8ea5a\") " Oct 08 18:14:48 crc kubenswrapper[4988]: I1008 18:14:48.135589 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/e294e9c8-16f2-46d4-91f8-5964b1e8ea5a-v4-0-config-system-service-ca\") pod \"e294e9c8-16f2-46d4-91f8-5964b1e8ea5a\" (UID: \"e294e9c8-16f2-46d4-91f8-5964b1e8ea5a\") " Oct 08 18:14:48 crc kubenswrapper[4988]: I1008 18:14:48.135702 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/e294e9c8-16f2-46d4-91f8-5964b1e8ea5a-v4-0-config-system-ocp-branding-template\") pod \"e294e9c8-16f2-46d4-91f8-5964b1e8ea5a\" (UID: \"e294e9c8-16f2-46d4-91f8-5964b1e8ea5a\") " Oct 08 18:14:48 crc kubenswrapper[4988]: I1008 18:14:48.135804 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/e294e9c8-16f2-46d4-91f8-5964b1e8ea5a-v4-0-config-user-idp-0-file-data\") pod \"e294e9c8-16f2-46d4-91f8-5964b1e8ea5a\" (UID: \"e294e9c8-16f2-46d4-91f8-5964b1e8ea5a\") " Oct 08 18:14:48 crc kubenswrapper[4988]: I1008 18:14:48.135893 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e294e9c8-16f2-46d4-91f8-5964b1e8ea5a-v4-0-config-system-trusted-ca-bundle\") pod \"e294e9c8-16f2-46d4-91f8-5964b1e8ea5a\" (UID: \"e294e9c8-16f2-46d4-91f8-5964b1e8ea5a\") " Oct 08 18:14:48 crc kubenswrapper[4988]: I1008 18:14:48.137524 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/e294e9c8-16f2-46d4-91f8-5964b1e8ea5a-audit-dir\") pod \"e294e9c8-16f2-46d4-91f8-5964b1e8ea5a\" (UID: \"e294e9c8-16f2-46d4-91f8-5964b1e8ea5a\") " Oct 08 18:14:48 crc kubenswrapper[4988]: I1008 18:14:48.136967 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e294e9c8-16f2-46d4-91f8-5964b1e8ea5a-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "e294e9c8-16f2-46d4-91f8-5964b1e8ea5a" (UID: "e294e9c8-16f2-46d4-91f8-5964b1e8ea5a"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 18:14:48 crc kubenswrapper[4988]: I1008 18:14:48.136976 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e294e9c8-16f2-46d4-91f8-5964b1e8ea5a-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "e294e9c8-16f2-46d4-91f8-5964b1e8ea5a" (UID: "e294e9c8-16f2-46d4-91f8-5964b1e8ea5a"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 18:14:48 crc kubenswrapper[4988]: I1008 18:14:48.137555 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e294e9c8-16f2-46d4-91f8-5964b1e8ea5a-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "e294e9c8-16f2-46d4-91f8-5964b1e8ea5a" (UID: "e294e9c8-16f2-46d4-91f8-5964b1e8ea5a"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 18:14:48 crc kubenswrapper[4988]: I1008 18:14:48.137931 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/74701ff2-016a-4cb3-814a-e3f81b62572e-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-79656f7ff7-kckxs\" (UID: \"74701ff2-016a-4cb3-814a-e3f81b62572e\") " pod="openshift-authentication/oauth-openshift-79656f7ff7-kckxs" Oct 08 18:14:48 crc kubenswrapper[4988]: I1008 18:14:48.138214 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/74701ff2-016a-4cb3-814a-e3f81b62572e-v4-0-config-system-cliconfig\") pod \"oauth-openshift-79656f7ff7-kckxs\" (UID: \"74701ff2-016a-4cb3-814a-e3f81b62572e\") " pod="openshift-authentication/oauth-openshift-79656f7ff7-kckxs" Oct 08 18:14:48 crc kubenswrapper[4988]: I1008 18:14:48.138317 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/74701ff2-016a-4cb3-814a-e3f81b62572e-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-79656f7ff7-kckxs\" (UID: \"74701ff2-016a-4cb3-814a-e3f81b62572e\") " pod="openshift-authentication/oauth-openshift-79656f7ff7-kckxs" Oct 08 18:14:48 crc kubenswrapper[4988]: I1008 18:14:48.138556 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/74701ff2-016a-4cb3-814a-e3f81b62572e-v4-0-config-system-serving-cert\") pod \"oauth-openshift-79656f7ff7-kckxs\" (UID: \"74701ff2-016a-4cb3-814a-e3f81b62572e\") " pod="openshift-authentication/oauth-openshift-79656f7ff7-kckxs" Oct 08 18:14:48 crc kubenswrapper[4988]: I1008 18:14:48.138729 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e294e9c8-16f2-46d4-91f8-5964b1e8ea5a-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "e294e9c8-16f2-46d4-91f8-5964b1e8ea5a" (UID: "e294e9c8-16f2-46d4-91f8-5964b1e8ea5a"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 18:14:48 crc kubenswrapper[4988]: I1008 18:14:48.138720 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/74701ff2-016a-4cb3-814a-e3f81b62572e-v4-0-config-user-template-login\") pod \"oauth-openshift-79656f7ff7-kckxs\" (UID: \"74701ff2-016a-4cb3-814a-e3f81b62572e\") " pod="openshift-authentication/oauth-openshift-79656f7ff7-kckxs" Oct 08 18:14:48 crc kubenswrapper[4988]: I1008 18:14:48.138895 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/74701ff2-016a-4cb3-814a-e3f81b62572e-v4-0-config-system-session\") pod \"oauth-openshift-79656f7ff7-kckxs\" (UID: \"74701ff2-016a-4cb3-814a-e3f81b62572e\") " pod="openshift-authentication/oauth-openshift-79656f7ff7-kckxs" Oct 08 18:14:48 crc kubenswrapper[4988]: I1008 18:14:48.138985 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/74701ff2-016a-4cb3-814a-e3f81b62572e-audit-policies\") pod \"oauth-openshift-79656f7ff7-kckxs\" (UID: \"74701ff2-016a-4cb3-814a-e3f81b62572e\") " pod="openshift-authentication/oauth-openshift-79656f7ff7-kckxs" Oct 08 18:14:48 crc kubenswrapper[4988]: I1008 18:14:48.139048 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/74701ff2-016a-4cb3-814a-e3f81b62572e-v4-0-config-system-router-certs\") pod \"oauth-openshift-79656f7ff7-kckxs\" (UID: \"74701ff2-016a-4cb3-814a-e3f81b62572e\") " pod="openshift-authentication/oauth-openshift-79656f7ff7-kckxs" Oct 08 18:14:48 crc kubenswrapper[4988]: I1008 18:14:48.139115 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/74701ff2-016a-4cb3-814a-e3f81b62572e-v4-0-config-user-template-error\") pod \"oauth-openshift-79656f7ff7-kckxs\" (UID: \"74701ff2-016a-4cb3-814a-e3f81b62572e\") " pod="openshift-authentication/oauth-openshift-79656f7ff7-kckxs" Oct 08 18:14:48 crc kubenswrapper[4988]: I1008 18:14:48.139166 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mnm9s\" (UniqueName: \"kubernetes.io/projected/74701ff2-016a-4cb3-814a-e3f81b62572e-kube-api-access-mnm9s\") pod \"oauth-openshift-79656f7ff7-kckxs\" (UID: \"74701ff2-016a-4cb3-814a-e3f81b62572e\") " pod="openshift-authentication/oauth-openshift-79656f7ff7-kckxs" Oct 08 18:14:48 crc kubenswrapper[4988]: I1008 18:14:48.139273 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/74701ff2-016a-4cb3-814a-e3f81b62572e-v4-0-config-system-service-ca\") pod \"oauth-openshift-79656f7ff7-kckxs\" (UID: \"74701ff2-016a-4cb3-814a-e3f81b62572e\") " pod="openshift-authentication/oauth-openshift-79656f7ff7-kckxs" Oct 08 18:14:48 crc kubenswrapper[4988]: I1008 18:14:48.139335 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/74701ff2-016a-4cb3-814a-e3f81b62572e-audit-dir\") pod \"oauth-openshift-79656f7ff7-kckxs\" (UID: \"74701ff2-016a-4cb3-814a-e3f81b62572e\") " pod="openshift-authentication/oauth-openshift-79656f7ff7-kckxs" Oct 08 18:14:48 crc kubenswrapper[4988]: I1008 18:14:48.139380 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/74701ff2-016a-4cb3-814a-e3f81b62572e-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-79656f7ff7-kckxs\" (UID: \"74701ff2-016a-4cb3-814a-e3f81b62572e\") " pod="openshift-authentication/oauth-openshift-79656f7ff7-kckxs" Oct 08 18:14:48 crc kubenswrapper[4988]: I1008 18:14:48.139467 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/74701ff2-016a-4cb3-814a-e3f81b62572e-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-79656f7ff7-kckxs\" (UID: \"74701ff2-016a-4cb3-814a-e3f81b62572e\") " pod="openshift-authentication/oauth-openshift-79656f7ff7-kckxs" Oct 08 18:14:48 crc kubenswrapper[4988]: I1008 18:14:48.139692 4988 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e294e9c8-16f2-46d4-91f8-5964b1e8ea5a-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 18:14:48 crc kubenswrapper[4988]: I1008 18:14:48.139721 4988 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/e294e9c8-16f2-46d4-91f8-5964b1e8ea5a-audit-policies\") on node \"crc\" DevicePath \"\"" Oct 08 18:14:48 crc kubenswrapper[4988]: I1008 18:14:48.139750 4988 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/e294e9c8-16f2-46d4-91f8-5964b1e8ea5a-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Oct 08 18:14:48 crc kubenswrapper[4988]: I1008 18:14:48.139775 4988 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/e294e9c8-16f2-46d4-91f8-5964b1e8ea5a-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Oct 08 18:14:48 crc kubenswrapper[4988]: I1008 18:14:48.139972 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/e294e9c8-16f2-46d4-91f8-5964b1e8ea5a-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "e294e9c8-16f2-46d4-91f8-5964b1e8ea5a" (UID: "e294e9c8-16f2-46d4-91f8-5964b1e8ea5a"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 08 18:14:48 crc kubenswrapper[4988]: I1008 18:14:48.147014 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e294e9c8-16f2-46d4-91f8-5964b1e8ea5a-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "e294e9c8-16f2-46d4-91f8-5964b1e8ea5a" (UID: "e294e9c8-16f2-46d4-91f8-5964b1e8ea5a"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:14:48 crc kubenswrapper[4988]: I1008 18:14:48.147452 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e294e9c8-16f2-46d4-91f8-5964b1e8ea5a-kube-api-access-6r9cj" (OuterVolumeSpecName: "kube-api-access-6r9cj") pod "e294e9c8-16f2-46d4-91f8-5964b1e8ea5a" (UID: "e294e9c8-16f2-46d4-91f8-5964b1e8ea5a"). InnerVolumeSpecName "kube-api-access-6r9cj". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:14:48 crc kubenswrapper[4988]: I1008 18:14:48.147535 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e294e9c8-16f2-46d4-91f8-5964b1e8ea5a-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "e294e9c8-16f2-46d4-91f8-5964b1e8ea5a" (UID: "e294e9c8-16f2-46d4-91f8-5964b1e8ea5a"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:14:48 crc kubenswrapper[4988]: I1008 18:14:48.148602 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e294e9c8-16f2-46d4-91f8-5964b1e8ea5a-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "e294e9c8-16f2-46d4-91f8-5964b1e8ea5a" (UID: "e294e9c8-16f2-46d4-91f8-5964b1e8ea5a"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:14:48 crc kubenswrapper[4988]: I1008 18:14:48.154696 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e294e9c8-16f2-46d4-91f8-5964b1e8ea5a-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "e294e9c8-16f2-46d4-91f8-5964b1e8ea5a" (UID: "e294e9c8-16f2-46d4-91f8-5964b1e8ea5a"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:14:48 crc kubenswrapper[4988]: I1008 18:14:48.155182 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e294e9c8-16f2-46d4-91f8-5964b1e8ea5a-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "e294e9c8-16f2-46d4-91f8-5964b1e8ea5a" (UID: "e294e9c8-16f2-46d4-91f8-5964b1e8ea5a"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:14:48 crc kubenswrapper[4988]: I1008 18:14:48.155569 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e294e9c8-16f2-46d4-91f8-5964b1e8ea5a-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "e294e9c8-16f2-46d4-91f8-5964b1e8ea5a" (UID: "e294e9c8-16f2-46d4-91f8-5964b1e8ea5a"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:14:48 crc kubenswrapper[4988]: I1008 18:14:48.157678 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e294e9c8-16f2-46d4-91f8-5964b1e8ea5a-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "e294e9c8-16f2-46d4-91f8-5964b1e8ea5a" (UID: "e294e9c8-16f2-46d4-91f8-5964b1e8ea5a"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:14:48 crc kubenswrapper[4988]: I1008 18:14:48.162598 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e294e9c8-16f2-46d4-91f8-5964b1e8ea5a-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "e294e9c8-16f2-46d4-91f8-5964b1e8ea5a" (UID: "e294e9c8-16f2-46d4-91f8-5964b1e8ea5a"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:14:48 crc kubenswrapper[4988]: I1008 18:14:48.240872 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/74701ff2-016a-4cb3-814a-e3f81b62572e-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-79656f7ff7-kckxs\" (UID: \"74701ff2-016a-4cb3-814a-e3f81b62572e\") " pod="openshift-authentication/oauth-openshift-79656f7ff7-kckxs" Oct 08 18:14:48 crc kubenswrapper[4988]: I1008 18:14:48.240944 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/74701ff2-016a-4cb3-814a-e3f81b62572e-v4-0-config-system-cliconfig\") pod \"oauth-openshift-79656f7ff7-kckxs\" (UID: \"74701ff2-016a-4cb3-814a-e3f81b62572e\") " pod="openshift-authentication/oauth-openshift-79656f7ff7-kckxs" Oct 08 18:14:48 crc kubenswrapper[4988]: I1008 18:14:48.240970 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/74701ff2-016a-4cb3-814a-e3f81b62572e-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-79656f7ff7-kckxs\" (UID: \"74701ff2-016a-4cb3-814a-e3f81b62572e\") " pod="openshift-authentication/oauth-openshift-79656f7ff7-kckxs" Oct 08 18:14:48 crc kubenswrapper[4988]: I1008 18:14:48.240995 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/74701ff2-016a-4cb3-814a-e3f81b62572e-v4-0-config-system-serving-cert\") pod \"oauth-openshift-79656f7ff7-kckxs\" (UID: \"74701ff2-016a-4cb3-814a-e3f81b62572e\") " pod="openshift-authentication/oauth-openshift-79656f7ff7-kckxs" Oct 08 18:14:48 crc kubenswrapper[4988]: I1008 18:14:48.241022 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/74701ff2-016a-4cb3-814a-e3f81b62572e-v4-0-config-user-template-login\") pod \"oauth-openshift-79656f7ff7-kckxs\" (UID: \"74701ff2-016a-4cb3-814a-e3f81b62572e\") " pod="openshift-authentication/oauth-openshift-79656f7ff7-kckxs" Oct 08 18:14:48 crc kubenswrapper[4988]: I1008 18:14:48.241044 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/74701ff2-016a-4cb3-814a-e3f81b62572e-v4-0-config-system-session\") pod \"oauth-openshift-79656f7ff7-kckxs\" (UID: \"74701ff2-016a-4cb3-814a-e3f81b62572e\") " pod="openshift-authentication/oauth-openshift-79656f7ff7-kckxs" Oct 08 18:14:48 crc kubenswrapper[4988]: I1008 18:14:48.241073 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/74701ff2-016a-4cb3-814a-e3f81b62572e-audit-policies\") pod \"oauth-openshift-79656f7ff7-kckxs\" (UID: \"74701ff2-016a-4cb3-814a-e3f81b62572e\") " pod="openshift-authentication/oauth-openshift-79656f7ff7-kckxs" Oct 08 18:14:48 crc kubenswrapper[4988]: I1008 18:14:48.241096 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/74701ff2-016a-4cb3-814a-e3f81b62572e-v4-0-config-system-router-certs\") pod \"oauth-openshift-79656f7ff7-kckxs\" (UID: \"74701ff2-016a-4cb3-814a-e3f81b62572e\") " pod="openshift-authentication/oauth-openshift-79656f7ff7-kckxs" Oct 08 18:14:48 crc kubenswrapper[4988]: I1008 18:14:48.241117 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/74701ff2-016a-4cb3-814a-e3f81b62572e-v4-0-config-user-template-error\") pod \"oauth-openshift-79656f7ff7-kckxs\" (UID: \"74701ff2-016a-4cb3-814a-e3f81b62572e\") " pod="openshift-authentication/oauth-openshift-79656f7ff7-kckxs" Oct 08 18:14:48 crc kubenswrapper[4988]: I1008 18:14:48.241138 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mnm9s\" (UniqueName: \"kubernetes.io/projected/74701ff2-016a-4cb3-814a-e3f81b62572e-kube-api-access-mnm9s\") pod \"oauth-openshift-79656f7ff7-kckxs\" (UID: \"74701ff2-016a-4cb3-814a-e3f81b62572e\") " pod="openshift-authentication/oauth-openshift-79656f7ff7-kckxs" Oct 08 18:14:48 crc kubenswrapper[4988]: I1008 18:14:48.241171 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/74701ff2-016a-4cb3-814a-e3f81b62572e-v4-0-config-system-service-ca\") pod \"oauth-openshift-79656f7ff7-kckxs\" (UID: \"74701ff2-016a-4cb3-814a-e3f81b62572e\") " pod="openshift-authentication/oauth-openshift-79656f7ff7-kckxs" Oct 08 18:14:48 crc kubenswrapper[4988]: I1008 18:14:48.241196 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/74701ff2-016a-4cb3-814a-e3f81b62572e-audit-dir\") pod \"oauth-openshift-79656f7ff7-kckxs\" (UID: \"74701ff2-016a-4cb3-814a-e3f81b62572e\") " pod="openshift-authentication/oauth-openshift-79656f7ff7-kckxs" Oct 08 18:14:48 crc kubenswrapper[4988]: I1008 18:14:48.241218 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/74701ff2-016a-4cb3-814a-e3f81b62572e-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-79656f7ff7-kckxs\" (UID: \"74701ff2-016a-4cb3-814a-e3f81b62572e\") " pod="openshift-authentication/oauth-openshift-79656f7ff7-kckxs" Oct 08 18:14:48 crc kubenswrapper[4988]: I1008 18:14:48.241239 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/74701ff2-016a-4cb3-814a-e3f81b62572e-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-79656f7ff7-kckxs\" (UID: \"74701ff2-016a-4cb3-814a-e3f81b62572e\") " pod="openshift-authentication/oauth-openshift-79656f7ff7-kckxs" Oct 08 18:14:48 crc kubenswrapper[4988]: I1008 18:14:48.241296 4988 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/e294e9c8-16f2-46d4-91f8-5964b1e8ea5a-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Oct 08 18:14:48 crc kubenswrapper[4988]: I1008 18:14:48.241314 4988 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/e294e9c8-16f2-46d4-91f8-5964b1e8ea5a-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Oct 08 18:14:48 crc kubenswrapper[4988]: I1008 18:14:48.241329 4988 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/e294e9c8-16f2-46d4-91f8-5964b1e8ea5a-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 08 18:14:48 crc kubenswrapper[4988]: I1008 18:14:48.241343 4988 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/e294e9c8-16f2-46d4-91f8-5964b1e8ea5a-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Oct 08 18:14:48 crc kubenswrapper[4988]: I1008 18:14:48.241357 4988 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/e294e9c8-16f2-46d4-91f8-5964b1e8ea5a-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Oct 08 18:14:48 crc kubenswrapper[4988]: I1008 18:14:48.241375 4988 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6r9cj\" (UniqueName: \"kubernetes.io/projected/e294e9c8-16f2-46d4-91f8-5964b1e8ea5a-kube-api-access-6r9cj\") on node \"crc\" DevicePath \"\"" Oct 08 18:14:48 crc kubenswrapper[4988]: I1008 18:14:48.241428 4988 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/e294e9c8-16f2-46d4-91f8-5964b1e8ea5a-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Oct 08 18:14:48 crc kubenswrapper[4988]: I1008 18:14:48.241450 4988 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/e294e9c8-16f2-46d4-91f8-5964b1e8ea5a-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Oct 08 18:14:48 crc kubenswrapper[4988]: I1008 18:14:48.241470 4988 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/e294e9c8-16f2-46d4-91f8-5964b1e8ea5a-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Oct 08 18:14:48 crc kubenswrapper[4988]: I1008 18:14:48.241488 4988 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/e294e9c8-16f2-46d4-91f8-5964b1e8ea5a-audit-dir\") on node \"crc\" DevicePath \"\"" Oct 08 18:14:48 crc kubenswrapper[4988]: I1008 18:14:48.242208 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/74701ff2-016a-4cb3-814a-e3f81b62572e-v4-0-config-system-cliconfig\") pod \"oauth-openshift-79656f7ff7-kckxs\" (UID: \"74701ff2-016a-4cb3-814a-e3f81b62572e\") " pod="openshift-authentication/oauth-openshift-79656f7ff7-kckxs" Oct 08 18:14:48 crc kubenswrapper[4988]: I1008 18:14:48.242564 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/74701ff2-016a-4cb3-814a-e3f81b62572e-audit-dir\") pod \"oauth-openshift-79656f7ff7-kckxs\" (UID: \"74701ff2-016a-4cb3-814a-e3f81b62572e\") " pod="openshift-authentication/oauth-openshift-79656f7ff7-kckxs" Oct 08 18:14:48 crc kubenswrapper[4988]: I1008 18:14:48.242828 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/74701ff2-016a-4cb3-814a-e3f81b62572e-v4-0-config-system-service-ca\") pod \"oauth-openshift-79656f7ff7-kckxs\" (UID: \"74701ff2-016a-4cb3-814a-e3f81b62572e\") " pod="openshift-authentication/oauth-openshift-79656f7ff7-kckxs" Oct 08 18:14:48 crc kubenswrapper[4988]: I1008 18:14:48.243103 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/74701ff2-016a-4cb3-814a-e3f81b62572e-audit-policies\") pod \"oauth-openshift-79656f7ff7-kckxs\" (UID: \"74701ff2-016a-4cb3-814a-e3f81b62572e\") " pod="openshift-authentication/oauth-openshift-79656f7ff7-kckxs" Oct 08 18:14:48 crc kubenswrapper[4988]: I1008 18:14:48.243705 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/74701ff2-016a-4cb3-814a-e3f81b62572e-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-79656f7ff7-kckxs\" (UID: \"74701ff2-016a-4cb3-814a-e3f81b62572e\") " pod="openshift-authentication/oauth-openshift-79656f7ff7-kckxs" Oct 08 18:14:48 crc kubenswrapper[4988]: I1008 18:14:48.244883 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/74701ff2-016a-4cb3-814a-e3f81b62572e-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-79656f7ff7-kckxs\" (UID: \"74701ff2-016a-4cb3-814a-e3f81b62572e\") " pod="openshift-authentication/oauth-openshift-79656f7ff7-kckxs" Oct 08 18:14:48 crc kubenswrapper[4988]: I1008 18:14:48.245208 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/74701ff2-016a-4cb3-814a-e3f81b62572e-v4-0-config-user-template-login\") pod \"oauth-openshift-79656f7ff7-kckxs\" (UID: \"74701ff2-016a-4cb3-814a-e3f81b62572e\") " pod="openshift-authentication/oauth-openshift-79656f7ff7-kckxs" Oct 08 18:14:48 crc kubenswrapper[4988]: I1008 18:14:48.245309 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/74701ff2-016a-4cb3-814a-e3f81b62572e-v4-0-config-system-router-certs\") pod \"oauth-openshift-79656f7ff7-kckxs\" (UID: \"74701ff2-016a-4cb3-814a-e3f81b62572e\") " pod="openshift-authentication/oauth-openshift-79656f7ff7-kckxs" Oct 08 18:14:48 crc kubenswrapper[4988]: I1008 18:14:48.246043 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/74701ff2-016a-4cb3-814a-e3f81b62572e-v4-0-config-system-serving-cert\") pod \"oauth-openshift-79656f7ff7-kckxs\" (UID: \"74701ff2-016a-4cb3-814a-e3f81b62572e\") " pod="openshift-authentication/oauth-openshift-79656f7ff7-kckxs" Oct 08 18:14:48 crc kubenswrapper[4988]: I1008 18:14:48.246472 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/74701ff2-016a-4cb3-814a-e3f81b62572e-v4-0-config-user-template-error\") pod \"oauth-openshift-79656f7ff7-kckxs\" (UID: \"74701ff2-016a-4cb3-814a-e3f81b62572e\") " pod="openshift-authentication/oauth-openshift-79656f7ff7-kckxs" Oct 08 18:14:48 crc kubenswrapper[4988]: I1008 18:14:48.246507 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/74701ff2-016a-4cb3-814a-e3f81b62572e-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-79656f7ff7-kckxs\" (UID: \"74701ff2-016a-4cb3-814a-e3f81b62572e\") " pod="openshift-authentication/oauth-openshift-79656f7ff7-kckxs" Oct 08 18:14:48 crc kubenswrapper[4988]: I1008 18:14:48.246984 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/74701ff2-016a-4cb3-814a-e3f81b62572e-v4-0-config-system-session\") pod \"oauth-openshift-79656f7ff7-kckxs\" (UID: \"74701ff2-016a-4cb3-814a-e3f81b62572e\") " pod="openshift-authentication/oauth-openshift-79656f7ff7-kckxs" Oct 08 18:14:48 crc kubenswrapper[4988]: I1008 18:14:48.247614 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/74701ff2-016a-4cb3-814a-e3f81b62572e-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-79656f7ff7-kckxs\" (UID: \"74701ff2-016a-4cb3-814a-e3f81b62572e\") " pod="openshift-authentication/oauth-openshift-79656f7ff7-kckxs" Oct 08 18:14:48 crc kubenswrapper[4988]: I1008 18:14:48.267311 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mnm9s\" (UniqueName: \"kubernetes.io/projected/74701ff2-016a-4cb3-814a-e3f81b62572e-kube-api-access-mnm9s\") pod \"oauth-openshift-79656f7ff7-kckxs\" (UID: \"74701ff2-016a-4cb3-814a-e3f81b62572e\") " pod="openshift-authentication/oauth-openshift-79656f7ff7-kckxs" Oct 08 18:14:48 crc kubenswrapper[4988]: I1008 18:14:48.448334 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-79656f7ff7-kckxs" Oct 08 18:14:48 crc kubenswrapper[4988]: I1008 18:14:48.801896 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-hwtzl" event={"ID":"e294e9c8-16f2-46d4-91f8-5964b1e8ea5a","Type":"ContainerDied","Data":"918f45165b7980541dbf929e550b866d9b90c6015d6ef15b1e4f7ceca9097eca"} Oct 08 18:14:48 crc kubenswrapper[4988]: I1008 18:14:48.801976 4988 scope.go:117] "RemoveContainer" containerID="5bf6c97c2a92106c5d151cc199928afbbcc2d70b4025fb7cd24058e7da3b8157" Oct 08 18:14:48 crc kubenswrapper[4988]: I1008 18:14:48.802193 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-hwtzl" Oct 08 18:14:48 crc kubenswrapper[4988]: I1008 18:14:48.855580 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-hwtzl"] Oct 08 18:14:48 crc kubenswrapper[4988]: I1008 18:14:48.857511 4988 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-hwtzl"] Oct 08 18:14:48 crc kubenswrapper[4988]: I1008 18:14:48.913126 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-79656f7ff7-kckxs"] Oct 08 18:14:48 crc kubenswrapper[4988]: W1008 18:14:48.921113 4988 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod74701ff2_016a_4cb3_814a_e3f81b62572e.slice/crio-9e481f0b37ca5729a6b6e6e989477c90f342c1dacc342e0e64cb88a6f9eb7b3c WatchSource:0}: Error finding container 9e481f0b37ca5729a6b6e6e989477c90f342c1dacc342e0e64cb88a6f9eb7b3c: Status 404 returned error can't find the container with id 9e481f0b37ca5729a6b6e6e989477c90f342c1dacc342e0e64cb88a6f9eb7b3c Oct 08 18:14:49 crc kubenswrapper[4988]: I1008 18:14:49.245847 4988 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e294e9c8-16f2-46d4-91f8-5964b1e8ea5a" path="/var/lib/kubelet/pods/e294e9c8-16f2-46d4-91f8-5964b1e8ea5a/volumes" Oct 08 18:14:49 crc kubenswrapper[4988]: I1008 18:14:49.812090 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-79656f7ff7-kckxs" event={"ID":"74701ff2-016a-4cb3-814a-e3f81b62572e","Type":"ContainerStarted","Data":"b85c1f66efaafe7505f9aa69cdb34754b68c63f78a425f65c9a5def0a9c77c8d"} Oct 08 18:14:49 crc kubenswrapper[4988]: I1008 18:14:49.812175 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-79656f7ff7-kckxs" event={"ID":"74701ff2-016a-4cb3-814a-e3f81b62572e","Type":"ContainerStarted","Data":"9e481f0b37ca5729a6b6e6e989477c90f342c1dacc342e0e64cb88a6f9eb7b3c"} Oct 08 18:14:49 crc kubenswrapper[4988]: I1008 18:14:49.812572 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-79656f7ff7-kckxs" Oct 08 18:14:49 crc kubenswrapper[4988]: I1008 18:14:49.822357 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-79656f7ff7-kckxs" Oct 08 18:14:49 crc kubenswrapper[4988]: I1008 18:14:49.849176 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-79656f7ff7-kckxs" podStartSLOduration=27.84915439 podStartE2EDuration="27.84915439s" podCreationTimestamp="2025-10-08 18:14:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 18:14:49.841699541 +0000 UTC m=+235.291542311" watchObservedRunningTime="2025-10-08 18:14:49.84915439 +0000 UTC m=+235.298997160" Oct 08 18:15:00 crc kubenswrapper[4988]: I1008 18:15:00.147880 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29332455-v2b7c"] Oct 08 18:15:00 crc kubenswrapper[4988]: I1008 18:15:00.149087 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29332455-v2b7c" Oct 08 18:15:00 crc kubenswrapper[4988]: I1008 18:15:00.152184 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 08 18:15:00 crc kubenswrapper[4988]: I1008 18:15:00.152579 4988 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 08 18:15:00 crc kubenswrapper[4988]: I1008 18:15:00.159567 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29332455-v2b7c"] Oct 08 18:15:00 crc kubenswrapper[4988]: I1008 18:15:00.253525 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/3cd39683-3b34-4c0e-9fb1-f63668fe5e4a-config-volume\") pod \"collect-profiles-29332455-v2b7c\" (UID: \"3cd39683-3b34-4c0e-9fb1-f63668fe5e4a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29332455-v2b7c" Oct 08 18:15:00 crc kubenswrapper[4988]: I1008 18:15:00.253993 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q4464\" (UniqueName: \"kubernetes.io/projected/3cd39683-3b34-4c0e-9fb1-f63668fe5e4a-kube-api-access-q4464\") pod \"collect-profiles-29332455-v2b7c\" (UID: \"3cd39683-3b34-4c0e-9fb1-f63668fe5e4a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29332455-v2b7c" Oct 08 18:15:00 crc kubenswrapper[4988]: I1008 18:15:00.254249 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/3cd39683-3b34-4c0e-9fb1-f63668fe5e4a-secret-volume\") pod \"collect-profiles-29332455-v2b7c\" (UID: \"3cd39683-3b34-4c0e-9fb1-f63668fe5e4a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29332455-v2b7c" Oct 08 18:15:00 crc kubenswrapper[4988]: I1008 18:15:00.356135 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/3cd39683-3b34-4c0e-9fb1-f63668fe5e4a-secret-volume\") pod \"collect-profiles-29332455-v2b7c\" (UID: \"3cd39683-3b34-4c0e-9fb1-f63668fe5e4a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29332455-v2b7c" Oct 08 18:15:00 crc kubenswrapper[4988]: I1008 18:15:00.356250 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/3cd39683-3b34-4c0e-9fb1-f63668fe5e4a-config-volume\") pod \"collect-profiles-29332455-v2b7c\" (UID: \"3cd39683-3b34-4c0e-9fb1-f63668fe5e4a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29332455-v2b7c" Oct 08 18:15:00 crc kubenswrapper[4988]: I1008 18:15:00.356311 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q4464\" (UniqueName: \"kubernetes.io/projected/3cd39683-3b34-4c0e-9fb1-f63668fe5e4a-kube-api-access-q4464\") pod \"collect-profiles-29332455-v2b7c\" (UID: \"3cd39683-3b34-4c0e-9fb1-f63668fe5e4a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29332455-v2b7c" Oct 08 18:15:00 crc kubenswrapper[4988]: I1008 18:15:00.358244 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/3cd39683-3b34-4c0e-9fb1-f63668fe5e4a-config-volume\") pod \"collect-profiles-29332455-v2b7c\" (UID: \"3cd39683-3b34-4c0e-9fb1-f63668fe5e4a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29332455-v2b7c" Oct 08 18:15:00 crc kubenswrapper[4988]: I1008 18:15:00.369072 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/3cd39683-3b34-4c0e-9fb1-f63668fe5e4a-secret-volume\") pod \"collect-profiles-29332455-v2b7c\" (UID: \"3cd39683-3b34-4c0e-9fb1-f63668fe5e4a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29332455-v2b7c" Oct 08 18:15:00 crc kubenswrapper[4988]: I1008 18:15:00.375618 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q4464\" (UniqueName: \"kubernetes.io/projected/3cd39683-3b34-4c0e-9fb1-f63668fe5e4a-kube-api-access-q4464\") pod \"collect-profiles-29332455-v2b7c\" (UID: \"3cd39683-3b34-4c0e-9fb1-f63668fe5e4a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29332455-v2b7c" Oct 08 18:15:00 crc kubenswrapper[4988]: I1008 18:15:00.468972 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29332455-v2b7c" Oct 08 18:15:00 crc kubenswrapper[4988]: I1008 18:15:00.684819 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29332455-v2b7c"] Oct 08 18:15:00 crc kubenswrapper[4988]: I1008 18:15:00.883468 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29332455-v2b7c" event={"ID":"3cd39683-3b34-4c0e-9fb1-f63668fe5e4a","Type":"ContainerStarted","Data":"09aeb58123957fe3fad0f83bd19a7f72bda0b4f670d430867d190556638609ce"} Oct 08 18:15:00 crc kubenswrapper[4988]: I1008 18:15:00.883545 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29332455-v2b7c" event={"ID":"3cd39683-3b34-4c0e-9fb1-f63668fe5e4a","Type":"ContainerStarted","Data":"dc51bd7b14b3dece26c9a5c6dc50f94fa871667dfd20ea1def49b78ef4818679"} Oct 08 18:15:00 crc kubenswrapper[4988]: I1008 18:15:00.901448 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29332455-v2b7c" podStartSLOduration=0.901426377 podStartE2EDuration="901.426377ms" podCreationTimestamp="2025-10-08 18:15:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 18:15:00.900651333 +0000 UTC m=+246.350494103" watchObservedRunningTime="2025-10-08 18:15:00.901426377 +0000 UTC m=+246.351269157" Oct 08 18:15:01 crc kubenswrapper[4988]: I1008 18:15:01.890000 4988 generic.go:334] "Generic (PLEG): container finished" podID="3cd39683-3b34-4c0e-9fb1-f63668fe5e4a" containerID="09aeb58123957fe3fad0f83bd19a7f72bda0b4f670d430867d190556638609ce" exitCode=0 Oct 08 18:15:01 crc kubenswrapper[4988]: I1008 18:15:01.890108 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29332455-v2b7c" event={"ID":"3cd39683-3b34-4c0e-9fb1-f63668fe5e4a","Type":"ContainerDied","Data":"09aeb58123957fe3fad0f83bd19a7f72bda0b4f670d430867d190556638609ce"} Oct 08 18:15:03 crc kubenswrapper[4988]: I1008 18:15:03.165182 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29332455-v2b7c" Oct 08 18:15:03 crc kubenswrapper[4988]: I1008 18:15:03.303642 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/3cd39683-3b34-4c0e-9fb1-f63668fe5e4a-config-volume\") pod \"3cd39683-3b34-4c0e-9fb1-f63668fe5e4a\" (UID: \"3cd39683-3b34-4c0e-9fb1-f63668fe5e4a\") " Oct 08 18:15:03 crc kubenswrapper[4988]: I1008 18:15:03.304566 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/3cd39683-3b34-4c0e-9fb1-f63668fe5e4a-secret-volume\") pod \"3cd39683-3b34-4c0e-9fb1-f63668fe5e4a\" (UID: \"3cd39683-3b34-4c0e-9fb1-f63668fe5e4a\") " Oct 08 18:15:03 crc kubenswrapper[4988]: I1008 18:15:03.304809 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q4464\" (UniqueName: \"kubernetes.io/projected/3cd39683-3b34-4c0e-9fb1-f63668fe5e4a-kube-api-access-q4464\") pod \"3cd39683-3b34-4c0e-9fb1-f63668fe5e4a\" (UID: \"3cd39683-3b34-4c0e-9fb1-f63668fe5e4a\") " Oct 08 18:15:03 crc kubenswrapper[4988]: I1008 18:15:03.304932 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3cd39683-3b34-4c0e-9fb1-f63668fe5e4a-config-volume" (OuterVolumeSpecName: "config-volume") pod "3cd39683-3b34-4c0e-9fb1-f63668fe5e4a" (UID: "3cd39683-3b34-4c0e-9fb1-f63668fe5e4a"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 18:15:03 crc kubenswrapper[4988]: I1008 18:15:03.305610 4988 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/3cd39683-3b34-4c0e-9fb1-f63668fe5e4a-config-volume\") on node \"crc\" DevicePath \"\"" Oct 08 18:15:03 crc kubenswrapper[4988]: I1008 18:15:03.315956 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3cd39683-3b34-4c0e-9fb1-f63668fe5e4a-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "3cd39683-3b34-4c0e-9fb1-f63668fe5e4a" (UID: "3cd39683-3b34-4c0e-9fb1-f63668fe5e4a"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:15:03 crc kubenswrapper[4988]: I1008 18:15:03.316006 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3cd39683-3b34-4c0e-9fb1-f63668fe5e4a-kube-api-access-q4464" (OuterVolumeSpecName: "kube-api-access-q4464") pod "3cd39683-3b34-4c0e-9fb1-f63668fe5e4a" (UID: "3cd39683-3b34-4c0e-9fb1-f63668fe5e4a"). InnerVolumeSpecName "kube-api-access-q4464". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:15:03 crc kubenswrapper[4988]: I1008 18:15:03.407256 4988 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/3cd39683-3b34-4c0e-9fb1-f63668fe5e4a-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 08 18:15:03 crc kubenswrapper[4988]: I1008 18:15:03.407318 4988 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q4464\" (UniqueName: \"kubernetes.io/projected/3cd39683-3b34-4c0e-9fb1-f63668fe5e4a-kube-api-access-q4464\") on node \"crc\" DevicePath \"\"" Oct 08 18:15:03 crc kubenswrapper[4988]: I1008 18:15:03.907124 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29332455-v2b7c" event={"ID":"3cd39683-3b34-4c0e-9fb1-f63668fe5e4a","Type":"ContainerDied","Data":"dc51bd7b14b3dece26c9a5c6dc50f94fa871667dfd20ea1def49b78ef4818679"} Oct 08 18:15:03 crc kubenswrapper[4988]: I1008 18:15:03.907196 4988 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="dc51bd7b14b3dece26c9a5c6dc50f94fa871667dfd20ea1def49b78ef4818679" Oct 08 18:15:03 crc kubenswrapper[4988]: I1008 18:15:03.907228 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29332455-v2b7c" Oct 08 18:15:05 crc kubenswrapper[4988]: I1008 18:15:05.845299 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-dmdqd"] Oct 08 18:15:05 crc kubenswrapper[4988]: I1008 18:15:05.846009 4988 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-dmdqd" podUID="15a52185-0755-45a6-b665-c1185d1996b4" containerName="registry-server" containerID="cri-o://376064a31a6403d1a62039ad1b6d1d240e5e224930f9b781cffd57e988576c51" gracePeriod=30 Oct 08 18:15:05 crc kubenswrapper[4988]: I1008 18:15:05.855136 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-77xtx"] Oct 08 18:15:05 crc kubenswrapper[4988]: I1008 18:15:05.855573 4988 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-77xtx" podUID="7c0e2c11-212a-46ed-b72a-911bde29199a" containerName="registry-server" containerID="cri-o://615d0e89e4190c703d515e7df5d5b3b444dd61b73916cb4b40cc5c9ce105c872" gracePeriod=30 Oct 08 18:15:05 crc kubenswrapper[4988]: I1008 18:15:05.869857 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-294dt"] Oct 08 18:15:05 crc kubenswrapper[4988]: I1008 18:15:05.870195 4988 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/marketplace-operator-79b997595-294dt" podUID="8bf9175c-5101-4fcd-8cba-e113be11dde4" containerName="marketplace-operator" containerID="cri-o://20dfd87103b1c34d1d77781ce5e72b28df28ecc2efa06881c31118ab87686314" gracePeriod=30 Oct 08 18:15:05 crc kubenswrapper[4988]: I1008 18:15:05.879432 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-kmg58"] Oct 08 18:15:05 crc kubenswrapper[4988]: I1008 18:15:05.879743 4988 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-kmg58" podUID="38d2198b-2aab-47d2-aa2e-93436ff69625" containerName="registry-server" containerID="cri-o://80113d52537f531d6bba38b5509f7a18c44a61e0f36a58742df56936d485d0f9" gracePeriod=30 Oct 08 18:15:05 crc kubenswrapper[4988]: I1008 18:15:05.893958 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-7ck2j"] Oct 08 18:15:05 crc kubenswrapper[4988]: I1008 18:15:05.894225 4988 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-7ck2j" podUID="b94d3b86-2fd1-4045-9a5e-dc6c296cfba4" containerName="registry-server" containerID="cri-o://021595f654ab0cc3eebfd44bb31805da9eef8f2dbc83c40b88a030a1e8fe3c54" gracePeriod=30 Oct 08 18:15:05 crc kubenswrapper[4988]: I1008 18:15:05.904913 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-kb5xn"] Oct 08 18:15:05 crc kubenswrapper[4988]: E1008 18:15:05.905256 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3cd39683-3b34-4c0e-9fb1-f63668fe5e4a" containerName="collect-profiles" Oct 08 18:15:05 crc kubenswrapper[4988]: I1008 18:15:05.905284 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="3cd39683-3b34-4c0e-9fb1-f63668fe5e4a" containerName="collect-profiles" Oct 08 18:15:05 crc kubenswrapper[4988]: I1008 18:15:05.905454 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="3cd39683-3b34-4c0e-9fb1-f63668fe5e4a" containerName="collect-profiles" Oct 08 18:15:05 crc kubenswrapper[4988]: I1008 18:15:05.906007 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-kb5xn" Oct 08 18:15:05 crc kubenswrapper[4988]: I1008 18:15:05.922859 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-kb5xn"] Oct 08 18:15:06 crc kubenswrapper[4988]: I1008 18:15:06.045535 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-27fch\" (UniqueName: \"kubernetes.io/projected/590d218c-e870-4304-b2b6-3c0fa7cc7c7a-kube-api-access-27fch\") pod \"marketplace-operator-79b997595-kb5xn\" (UID: \"590d218c-e870-4304-b2b6-3c0fa7cc7c7a\") " pod="openshift-marketplace/marketplace-operator-79b997595-kb5xn" Oct 08 18:15:06 crc kubenswrapper[4988]: I1008 18:15:06.045613 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/590d218c-e870-4304-b2b6-3c0fa7cc7c7a-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-kb5xn\" (UID: \"590d218c-e870-4304-b2b6-3c0fa7cc7c7a\") " pod="openshift-marketplace/marketplace-operator-79b997595-kb5xn" Oct 08 18:15:06 crc kubenswrapper[4988]: I1008 18:15:06.045679 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/590d218c-e870-4304-b2b6-3c0fa7cc7c7a-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-kb5xn\" (UID: \"590d218c-e870-4304-b2b6-3c0fa7cc7c7a\") " pod="openshift-marketplace/marketplace-operator-79b997595-kb5xn" Oct 08 18:15:06 crc kubenswrapper[4988]: I1008 18:15:06.148074 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-27fch\" (UniqueName: \"kubernetes.io/projected/590d218c-e870-4304-b2b6-3c0fa7cc7c7a-kube-api-access-27fch\") pod \"marketplace-operator-79b997595-kb5xn\" (UID: \"590d218c-e870-4304-b2b6-3c0fa7cc7c7a\") " pod="openshift-marketplace/marketplace-operator-79b997595-kb5xn" Oct 08 18:15:06 crc kubenswrapper[4988]: I1008 18:15:06.148462 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/590d218c-e870-4304-b2b6-3c0fa7cc7c7a-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-kb5xn\" (UID: \"590d218c-e870-4304-b2b6-3c0fa7cc7c7a\") " pod="openshift-marketplace/marketplace-operator-79b997595-kb5xn" Oct 08 18:15:06 crc kubenswrapper[4988]: I1008 18:15:06.148543 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/590d218c-e870-4304-b2b6-3c0fa7cc7c7a-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-kb5xn\" (UID: \"590d218c-e870-4304-b2b6-3c0fa7cc7c7a\") " pod="openshift-marketplace/marketplace-operator-79b997595-kb5xn" Oct 08 18:15:06 crc kubenswrapper[4988]: I1008 18:15:06.150727 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/590d218c-e870-4304-b2b6-3c0fa7cc7c7a-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-kb5xn\" (UID: \"590d218c-e870-4304-b2b6-3c0fa7cc7c7a\") " pod="openshift-marketplace/marketplace-operator-79b997595-kb5xn" Oct 08 18:15:06 crc kubenswrapper[4988]: I1008 18:15:06.164356 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/590d218c-e870-4304-b2b6-3c0fa7cc7c7a-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-kb5xn\" (UID: \"590d218c-e870-4304-b2b6-3c0fa7cc7c7a\") " pod="openshift-marketplace/marketplace-operator-79b997595-kb5xn" Oct 08 18:15:06 crc kubenswrapper[4988]: I1008 18:15:06.168376 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-27fch\" (UniqueName: \"kubernetes.io/projected/590d218c-e870-4304-b2b6-3c0fa7cc7c7a-kube-api-access-27fch\") pod \"marketplace-operator-79b997595-kb5xn\" (UID: \"590d218c-e870-4304-b2b6-3c0fa7cc7c7a\") " pod="openshift-marketplace/marketplace-operator-79b997595-kb5xn" Oct 08 18:15:06 crc kubenswrapper[4988]: I1008 18:15:06.226997 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-77xtx" Oct 08 18:15:06 crc kubenswrapper[4988]: I1008 18:15:06.249076 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7c0e2c11-212a-46ed-b72a-911bde29199a-catalog-content\") pod \"7c0e2c11-212a-46ed-b72a-911bde29199a\" (UID: \"7c0e2c11-212a-46ed-b72a-911bde29199a\") " Oct 08 18:15:06 crc kubenswrapper[4988]: I1008 18:15:06.249291 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kcc4w\" (UniqueName: \"kubernetes.io/projected/7c0e2c11-212a-46ed-b72a-911bde29199a-kube-api-access-kcc4w\") pod \"7c0e2c11-212a-46ed-b72a-911bde29199a\" (UID: \"7c0e2c11-212a-46ed-b72a-911bde29199a\") " Oct 08 18:15:06 crc kubenswrapper[4988]: I1008 18:15:06.249326 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7c0e2c11-212a-46ed-b72a-911bde29199a-utilities\") pod \"7c0e2c11-212a-46ed-b72a-911bde29199a\" (UID: \"7c0e2c11-212a-46ed-b72a-911bde29199a\") " Oct 08 18:15:06 crc kubenswrapper[4988]: I1008 18:15:06.251948 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7c0e2c11-212a-46ed-b72a-911bde29199a-utilities" (OuterVolumeSpecName: "utilities") pod "7c0e2c11-212a-46ed-b72a-911bde29199a" (UID: "7c0e2c11-212a-46ed-b72a-911bde29199a"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 18:15:06 crc kubenswrapper[4988]: I1008 18:15:06.255033 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7c0e2c11-212a-46ed-b72a-911bde29199a-kube-api-access-kcc4w" (OuterVolumeSpecName: "kube-api-access-kcc4w") pod "7c0e2c11-212a-46ed-b72a-911bde29199a" (UID: "7c0e2c11-212a-46ed-b72a-911bde29199a"). InnerVolumeSpecName "kube-api-access-kcc4w". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:15:06 crc kubenswrapper[4988]: E1008 18:15:06.259558 4988 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 021595f654ab0cc3eebfd44bb31805da9eef8f2dbc83c40b88a030a1e8fe3c54 is running failed: container process not found" containerID="021595f654ab0cc3eebfd44bb31805da9eef8f2dbc83c40b88a030a1e8fe3c54" cmd=["grpc_health_probe","-addr=:50051"] Oct 08 18:15:06 crc kubenswrapper[4988]: E1008 18:15:06.260221 4988 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 021595f654ab0cc3eebfd44bb31805da9eef8f2dbc83c40b88a030a1e8fe3c54 is running failed: container process not found" containerID="021595f654ab0cc3eebfd44bb31805da9eef8f2dbc83c40b88a030a1e8fe3c54" cmd=["grpc_health_probe","-addr=:50051"] Oct 08 18:15:06 crc kubenswrapper[4988]: E1008 18:15:06.260717 4988 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 021595f654ab0cc3eebfd44bb31805da9eef8f2dbc83c40b88a030a1e8fe3c54 is running failed: container process not found" containerID="021595f654ab0cc3eebfd44bb31805da9eef8f2dbc83c40b88a030a1e8fe3c54" cmd=["grpc_health_probe","-addr=:50051"] Oct 08 18:15:06 crc kubenswrapper[4988]: E1008 18:15:06.260758 4988 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 021595f654ab0cc3eebfd44bb31805da9eef8f2dbc83c40b88a030a1e8fe3c54 is running failed: container process not found" probeType="Readiness" pod="openshift-marketplace/redhat-operators-7ck2j" podUID="b94d3b86-2fd1-4045-9a5e-dc6c296cfba4" containerName="registry-server" Oct 08 18:15:06 crc kubenswrapper[4988]: I1008 18:15:06.337574 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7c0e2c11-212a-46ed-b72a-911bde29199a-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "7c0e2c11-212a-46ed-b72a-911bde29199a" (UID: "7c0e2c11-212a-46ed-b72a-911bde29199a"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 18:15:06 crc kubenswrapper[4988]: I1008 18:15:06.355784 4988 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7c0e2c11-212a-46ed-b72a-911bde29199a-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 08 18:15:06 crc kubenswrapper[4988]: I1008 18:15:06.355842 4988 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kcc4w\" (UniqueName: \"kubernetes.io/projected/7c0e2c11-212a-46ed-b72a-911bde29199a-kube-api-access-kcc4w\") on node \"crc\" DevicePath \"\"" Oct 08 18:15:06 crc kubenswrapper[4988]: I1008 18:15:06.355874 4988 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7c0e2c11-212a-46ed-b72a-911bde29199a-utilities\") on node \"crc\" DevicePath \"\"" Oct 08 18:15:06 crc kubenswrapper[4988]: I1008 18:15:06.418664 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-294dt" Oct 08 18:15:06 crc kubenswrapper[4988]: I1008 18:15:06.419151 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-dmdqd" Oct 08 18:15:06 crc kubenswrapper[4988]: I1008 18:15:06.424233 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-7ck2j" Oct 08 18:15:06 crc kubenswrapper[4988]: I1008 18:15:06.429255 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-kmg58" Oct 08 18:15:06 crc kubenswrapper[4988]: I1008 18:15:06.449321 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-kb5xn" Oct 08 18:15:06 crc kubenswrapper[4988]: I1008 18:15:06.457346 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6d8gx\" (UniqueName: \"kubernetes.io/projected/38d2198b-2aab-47d2-aa2e-93436ff69625-kube-api-access-6d8gx\") pod \"38d2198b-2aab-47d2-aa2e-93436ff69625\" (UID: \"38d2198b-2aab-47d2-aa2e-93436ff69625\") " Oct 08 18:15:06 crc kubenswrapper[4988]: I1008 18:15:06.458470 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/15a52185-0755-45a6-b665-c1185d1996b4-utilities\") pod \"15a52185-0755-45a6-b665-c1185d1996b4\" (UID: \"15a52185-0755-45a6-b665-c1185d1996b4\") " Oct 08 18:15:06 crc kubenswrapper[4988]: I1008 18:15:06.458520 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-js74s\" (UniqueName: \"kubernetes.io/projected/b94d3b86-2fd1-4045-9a5e-dc6c296cfba4-kube-api-access-js74s\") pod \"b94d3b86-2fd1-4045-9a5e-dc6c296cfba4\" (UID: \"b94d3b86-2fd1-4045-9a5e-dc6c296cfba4\") " Oct 08 18:15:06 crc kubenswrapper[4988]: I1008 18:15:06.458584 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/38d2198b-2aab-47d2-aa2e-93436ff69625-utilities\") pod \"38d2198b-2aab-47d2-aa2e-93436ff69625\" (UID: \"38d2198b-2aab-47d2-aa2e-93436ff69625\") " Oct 08 18:15:06 crc kubenswrapper[4988]: I1008 18:15:06.458614 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/15a52185-0755-45a6-b665-c1185d1996b4-catalog-content\") pod \"15a52185-0755-45a6-b665-c1185d1996b4\" (UID: \"15a52185-0755-45a6-b665-c1185d1996b4\") " Oct 08 18:15:06 crc kubenswrapper[4988]: I1008 18:15:06.458656 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n95d9\" (UniqueName: \"kubernetes.io/projected/15a52185-0755-45a6-b665-c1185d1996b4-kube-api-access-n95d9\") pod \"15a52185-0755-45a6-b665-c1185d1996b4\" (UID: \"15a52185-0755-45a6-b665-c1185d1996b4\") " Oct 08 18:15:06 crc kubenswrapper[4988]: I1008 18:15:06.458688 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8bf9175c-5101-4fcd-8cba-e113be11dde4-marketplace-trusted-ca\") pod \"8bf9175c-5101-4fcd-8cba-e113be11dde4\" (UID: \"8bf9175c-5101-4fcd-8cba-e113be11dde4\") " Oct 08 18:15:06 crc kubenswrapper[4988]: I1008 18:15:06.458765 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/8bf9175c-5101-4fcd-8cba-e113be11dde4-marketplace-operator-metrics\") pod \"8bf9175c-5101-4fcd-8cba-e113be11dde4\" (UID: \"8bf9175c-5101-4fcd-8cba-e113be11dde4\") " Oct 08 18:15:06 crc kubenswrapper[4988]: I1008 18:15:06.458790 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d4psg\" (UniqueName: \"kubernetes.io/projected/8bf9175c-5101-4fcd-8cba-e113be11dde4-kube-api-access-d4psg\") pod \"8bf9175c-5101-4fcd-8cba-e113be11dde4\" (UID: \"8bf9175c-5101-4fcd-8cba-e113be11dde4\") " Oct 08 18:15:06 crc kubenswrapper[4988]: I1008 18:15:06.458816 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b94d3b86-2fd1-4045-9a5e-dc6c296cfba4-utilities\") pod \"b94d3b86-2fd1-4045-9a5e-dc6c296cfba4\" (UID: \"b94d3b86-2fd1-4045-9a5e-dc6c296cfba4\") " Oct 08 18:15:06 crc kubenswrapper[4988]: I1008 18:15:06.458848 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/38d2198b-2aab-47d2-aa2e-93436ff69625-catalog-content\") pod \"38d2198b-2aab-47d2-aa2e-93436ff69625\" (UID: \"38d2198b-2aab-47d2-aa2e-93436ff69625\") " Oct 08 18:15:06 crc kubenswrapper[4988]: I1008 18:15:06.458870 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b94d3b86-2fd1-4045-9a5e-dc6c296cfba4-catalog-content\") pod \"b94d3b86-2fd1-4045-9a5e-dc6c296cfba4\" (UID: \"b94d3b86-2fd1-4045-9a5e-dc6c296cfba4\") " Oct 08 18:15:06 crc kubenswrapper[4988]: I1008 18:15:06.461691 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/15a52185-0755-45a6-b665-c1185d1996b4-utilities" (OuterVolumeSpecName: "utilities") pod "15a52185-0755-45a6-b665-c1185d1996b4" (UID: "15a52185-0755-45a6-b665-c1185d1996b4"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 18:15:06 crc kubenswrapper[4988]: I1008 18:15:06.462331 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/38d2198b-2aab-47d2-aa2e-93436ff69625-kube-api-access-6d8gx" (OuterVolumeSpecName: "kube-api-access-6d8gx") pod "38d2198b-2aab-47d2-aa2e-93436ff69625" (UID: "38d2198b-2aab-47d2-aa2e-93436ff69625"). InnerVolumeSpecName "kube-api-access-6d8gx". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:15:06 crc kubenswrapper[4988]: I1008 18:15:06.463285 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/38d2198b-2aab-47d2-aa2e-93436ff69625-utilities" (OuterVolumeSpecName: "utilities") pod "38d2198b-2aab-47d2-aa2e-93436ff69625" (UID: "38d2198b-2aab-47d2-aa2e-93436ff69625"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 18:15:06 crc kubenswrapper[4988]: I1008 18:15:06.464250 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b94d3b86-2fd1-4045-9a5e-dc6c296cfba4-utilities" (OuterVolumeSpecName: "utilities") pod "b94d3b86-2fd1-4045-9a5e-dc6c296cfba4" (UID: "b94d3b86-2fd1-4045-9a5e-dc6c296cfba4"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 18:15:06 crc kubenswrapper[4988]: I1008 18:15:06.464351 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b94d3b86-2fd1-4045-9a5e-dc6c296cfba4-kube-api-access-js74s" (OuterVolumeSpecName: "kube-api-access-js74s") pod "b94d3b86-2fd1-4045-9a5e-dc6c296cfba4" (UID: "b94d3b86-2fd1-4045-9a5e-dc6c296cfba4"). InnerVolumeSpecName "kube-api-access-js74s". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:15:06 crc kubenswrapper[4988]: I1008 18:15:06.466246 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8bf9175c-5101-4fcd-8cba-e113be11dde4-kube-api-access-d4psg" (OuterVolumeSpecName: "kube-api-access-d4psg") pod "8bf9175c-5101-4fcd-8cba-e113be11dde4" (UID: "8bf9175c-5101-4fcd-8cba-e113be11dde4"). InnerVolumeSpecName "kube-api-access-d4psg". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:15:06 crc kubenswrapper[4988]: I1008 18:15:06.467101 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8bf9175c-5101-4fcd-8cba-e113be11dde4-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "8bf9175c-5101-4fcd-8cba-e113be11dde4" (UID: "8bf9175c-5101-4fcd-8cba-e113be11dde4"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 18:15:06 crc kubenswrapper[4988]: I1008 18:15:06.467234 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/15a52185-0755-45a6-b665-c1185d1996b4-kube-api-access-n95d9" (OuterVolumeSpecName: "kube-api-access-n95d9") pod "15a52185-0755-45a6-b665-c1185d1996b4" (UID: "15a52185-0755-45a6-b665-c1185d1996b4"). InnerVolumeSpecName "kube-api-access-n95d9". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:15:06 crc kubenswrapper[4988]: I1008 18:15:06.474192 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8bf9175c-5101-4fcd-8cba-e113be11dde4-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "8bf9175c-5101-4fcd-8cba-e113be11dde4" (UID: "8bf9175c-5101-4fcd-8cba-e113be11dde4"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:15:06 crc kubenswrapper[4988]: I1008 18:15:06.491854 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/38d2198b-2aab-47d2-aa2e-93436ff69625-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "38d2198b-2aab-47d2-aa2e-93436ff69625" (UID: "38d2198b-2aab-47d2-aa2e-93436ff69625"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 18:15:06 crc kubenswrapper[4988]: I1008 18:15:06.533441 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/15a52185-0755-45a6-b665-c1185d1996b4-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "15a52185-0755-45a6-b665-c1185d1996b4" (UID: "15a52185-0755-45a6-b665-c1185d1996b4"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 18:15:06 crc kubenswrapper[4988]: I1008 18:15:06.560792 4988 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/15a52185-0755-45a6-b665-c1185d1996b4-utilities\") on node \"crc\" DevicePath \"\"" Oct 08 18:15:06 crc kubenswrapper[4988]: I1008 18:15:06.560837 4988 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-js74s\" (UniqueName: \"kubernetes.io/projected/b94d3b86-2fd1-4045-9a5e-dc6c296cfba4-kube-api-access-js74s\") on node \"crc\" DevicePath \"\"" Oct 08 18:15:06 crc kubenswrapper[4988]: I1008 18:15:06.560854 4988 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/38d2198b-2aab-47d2-aa2e-93436ff69625-utilities\") on node \"crc\" DevicePath \"\"" Oct 08 18:15:06 crc kubenswrapper[4988]: I1008 18:15:06.560866 4988 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/15a52185-0755-45a6-b665-c1185d1996b4-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 08 18:15:06 crc kubenswrapper[4988]: I1008 18:15:06.560875 4988 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n95d9\" (UniqueName: \"kubernetes.io/projected/15a52185-0755-45a6-b665-c1185d1996b4-kube-api-access-n95d9\") on node \"crc\" DevicePath \"\"" Oct 08 18:15:06 crc kubenswrapper[4988]: I1008 18:15:06.560885 4988 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8bf9175c-5101-4fcd-8cba-e113be11dde4-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Oct 08 18:15:06 crc kubenswrapper[4988]: I1008 18:15:06.560895 4988 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/8bf9175c-5101-4fcd-8cba-e113be11dde4-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Oct 08 18:15:06 crc kubenswrapper[4988]: I1008 18:15:06.560907 4988 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d4psg\" (UniqueName: \"kubernetes.io/projected/8bf9175c-5101-4fcd-8cba-e113be11dde4-kube-api-access-d4psg\") on node \"crc\" DevicePath \"\"" Oct 08 18:15:06 crc kubenswrapper[4988]: I1008 18:15:06.560919 4988 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b94d3b86-2fd1-4045-9a5e-dc6c296cfba4-utilities\") on node \"crc\" DevicePath \"\"" Oct 08 18:15:06 crc kubenswrapper[4988]: I1008 18:15:06.560929 4988 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/38d2198b-2aab-47d2-aa2e-93436ff69625-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 08 18:15:06 crc kubenswrapper[4988]: I1008 18:15:06.560939 4988 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6d8gx\" (UniqueName: \"kubernetes.io/projected/38d2198b-2aab-47d2-aa2e-93436ff69625-kube-api-access-6d8gx\") on node \"crc\" DevicePath \"\"" Oct 08 18:15:06 crc kubenswrapper[4988]: I1008 18:15:06.582057 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b94d3b86-2fd1-4045-9a5e-dc6c296cfba4-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b94d3b86-2fd1-4045-9a5e-dc6c296cfba4" (UID: "b94d3b86-2fd1-4045-9a5e-dc6c296cfba4"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 18:15:06 crc kubenswrapper[4988]: I1008 18:15:06.662711 4988 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b94d3b86-2fd1-4045-9a5e-dc6c296cfba4-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 08 18:15:06 crc kubenswrapper[4988]: I1008 18:15:06.700714 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-kb5xn"] Oct 08 18:15:06 crc kubenswrapper[4988]: I1008 18:15:06.932372 4988 generic.go:334] "Generic (PLEG): container finished" podID="7c0e2c11-212a-46ed-b72a-911bde29199a" containerID="615d0e89e4190c703d515e7df5d5b3b444dd61b73916cb4b40cc5c9ce105c872" exitCode=0 Oct 08 18:15:06 crc kubenswrapper[4988]: I1008 18:15:06.932475 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-77xtx" event={"ID":"7c0e2c11-212a-46ed-b72a-911bde29199a","Type":"ContainerDied","Data":"615d0e89e4190c703d515e7df5d5b3b444dd61b73916cb4b40cc5c9ce105c872"} Oct 08 18:15:06 crc kubenswrapper[4988]: I1008 18:15:06.932508 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-77xtx" event={"ID":"7c0e2c11-212a-46ed-b72a-911bde29199a","Type":"ContainerDied","Data":"220ad062d243aa03d816f4a27f313ef1cae071aab188d8f04985110691f8023f"} Oct 08 18:15:06 crc kubenswrapper[4988]: I1008 18:15:06.932511 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-77xtx" Oct 08 18:15:06 crc kubenswrapper[4988]: I1008 18:15:06.932528 4988 scope.go:117] "RemoveContainer" containerID="615d0e89e4190c703d515e7df5d5b3b444dd61b73916cb4b40cc5c9ce105c872" Oct 08 18:15:06 crc kubenswrapper[4988]: I1008 18:15:06.935774 4988 generic.go:334] "Generic (PLEG): container finished" podID="38d2198b-2aab-47d2-aa2e-93436ff69625" containerID="80113d52537f531d6bba38b5509f7a18c44a61e0f36a58742df56936d485d0f9" exitCode=0 Oct 08 18:15:06 crc kubenswrapper[4988]: I1008 18:15:06.935884 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-kmg58" Oct 08 18:15:06 crc kubenswrapper[4988]: I1008 18:15:06.938585 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-kmg58" event={"ID":"38d2198b-2aab-47d2-aa2e-93436ff69625","Type":"ContainerDied","Data":"80113d52537f531d6bba38b5509f7a18c44a61e0f36a58742df56936d485d0f9"} Oct 08 18:15:06 crc kubenswrapper[4988]: I1008 18:15:06.938736 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-kmg58" event={"ID":"38d2198b-2aab-47d2-aa2e-93436ff69625","Type":"ContainerDied","Data":"1b82e7dfedfd0b2564a4226267eebbda9537214b654b88ea144c09b1cc87d46c"} Oct 08 18:15:06 crc kubenswrapper[4988]: I1008 18:15:06.940638 4988 generic.go:334] "Generic (PLEG): container finished" podID="b94d3b86-2fd1-4045-9a5e-dc6c296cfba4" containerID="021595f654ab0cc3eebfd44bb31805da9eef8f2dbc83c40b88a030a1e8fe3c54" exitCode=0 Oct 08 18:15:06 crc kubenswrapper[4988]: I1008 18:15:06.940759 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7ck2j" event={"ID":"b94d3b86-2fd1-4045-9a5e-dc6c296cfba4","Type":"ContainerDied","Data":"021595f654ab0cc3eebfd44bb31805da9eef8f2dbc83c40b88a030a1e8fe3c54"} Oct 08 18:15:06 crc kubenswrapper[4988]: I1008 18:15:06.940843 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7ck2j" event={"ID":"b94d3b86-2fd1-4045-9a5e-dc6c296cfba4","Type":"ContainerDied","Data":"de0cc2017cd41432a9aacd5339de40ccb247809b0b5b2a0762f144de451c0ac4"} Oct 08 18:15:06 crc kubenswrapper[4988]: I1008 18:15:06.940995 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-7ck2j" Oct 08 18:15:06 crc kubenswrapper[4988]: I1008 18:15:06.951986 4988 generic.go:334] "Generic (PLEG): container finished" podID="15a52185-0755-45a6-b665-c1185d1996b4" containerID="376064a31a6403d1a62039ad1b6d1d240e5e224930f9b781cffd57e988576c51" exitCode=0 Oct 08 18:15:06 crc kubenswrapper[4988]: I1008 18:15:06.952989 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dmdqd" event={"ID":"15a52185-0755-45a6-b665-c1185d1996b4","Type":"ContainerDied","Data":"376064a31a6403d1a62039ad1b6d1d240e5e224930f9b781cffd57e988576c51"} Oct 08 18:15:06 crc kubenswrapper[4988]: I1008 18:15:06.953111 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dmdqd" event={"ID":"15a52185-0755-45a6-b665-c1185d1996b4","Type":"ContainerDied","Data":"75e9f6bd560703b6df4bfd0f409c13086c917285458c72edaa4739dabcf9a986"} Oct 08 18:15:06 crc kubenswrapper[4988]: I1008 18:15:06.953655 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-dmdqd" Oct 08 18:15:06 crc kubenswrapper[4988]: I1008 18:15:06.959479 4988 generic.go:334] "Generic (PLEG): container finished" podID="8bf9175c-5101-4fcd-8cba-e113be11dde4" containerID="20dfd87103b1c34d1d77781ce5e72b28df28ecc2efa06881c31118ab87686314" exitCode=0 Oct 08 18:15:06 crc kubenswrapper[4988]: I1008 18:15:06.959593 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-294dt" event={"ID":"8bf9175c-5101-4fcd-8cba-e113be11dde4","Type":"ContainerDied","Data":"20dfd87103b1c34d1d77781ce5e72b28df28ecc2efa06881c31118ab87686314"} Oct 08 18:15:06 crc kubenswrapper[4988]: I1008 18:15:06.959658 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-294dt" event={"ID":"8bf9175c-5101-4fcd-8cba-e113be11dde4","Type":"ContainerDied","Data":"9ca6901123f580bf717fd66335996ba54cbe84d4a491dea9de46c48bb519c2eb"} Oct 08 18:15:06 crc kubenswrapper[4988]: I1008 18:15:06.959937 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-294dt" Oct 08 18:15:06 crc kubenswrapper[4988]: I1008 18:15:06.963024 4988 scope.go:117] "RemoveContainer" containerID="7285e62439fa2c40d6fb7ad0da89aaa7edc907b20ba48be35aae53b4d742c498" Oct 08 18:15:06 crc kubenswrapper[4988]: I1008 18:15:06.967808 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-kb5xn" event={"ID":"590d218c-e870-4304-b2b6-3c0fa7cc7c7a","Type":"ContainerStarted","Data":"b57a2b239ebbafafaa27b5025d43b5810f5e1226cc1788588e5ce88419cb29ab"} Oct 08 18:15:06 crc kubenswrapper[4988]: I1008 18:15:06.967890 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-kb5xn" event={"ID":"590d218c-e870-4304-b2b6-3c0fa7cc7c7a","Type":"ContainerStarted","Data":"ddc7688afc2a2d6c368ce817538741f50c524fe351fb24023d38b6c460a01114"} Oct 08 18:15:06 crc kubenswrapper[4988]: I1008 18:15:06.969586 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-kb5xn" Oct 08 18:15:06 crc kubenswrapper[4988]: I1008 18:15:06.980131 4988 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-kb5xn container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.56:8080/healthz\": dial tcp 10.217.0.56:8080: connect: connection refused" start-of-body= Oct 08 18:15:06 crc kubenswrapper[4988]: I1008 18:15:06.980376 4988 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-kb5xn" podUID="590d218c-e870-4304-b2b6-3c0fa7cc7c7a" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.56:8080/healthz\": dial tcp 10.217.0.56:8080: connect: connection refused" Oct 08 18:15:06 crc kubenswrapper[4988]: I1008 18:15:06.999236 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-77xtx"] Oct 08 18:15:07 crc kubenswrapper[4988]: I1008 18:15:07.009429 4988 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-77xtx"] Oct 08 18:15:07 crc kubenswrapper[4988]: I1008 18:15:07.016374 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-kmg58"] Oct 08 18:15:07 crc kubenswrapper[4988]: I1008 18:15:07.020669 4988 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-kmg58"] Oct 08 18:15:07 crc kubenswrapper[4988]: I1008 18:15:07.031550 4988 scope.go:117] "RemoveContainer" containerID="aaf1ecf558697a87bf3a10117a2784a64dff7aae5eb9a2cf38599f88ef41a71b" Oct 08 18:15:07 crc kubenswrapper[4988]: I1008 18:15:07.033088 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-kb5xn" podStartSLOduration=2.033058321 podStartE2EDuration="2.033058321s" podCreationTimestamp="2025-10-08 18:15:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 18:15:07.026819969 +0000 UTC m=+252.476662749" watchObservedRunningTime="2025-10-08 18:15:07.033058321 +0000 UTC m=+252.482901091" Oct 08 18:15:07 crc kubenswrapper[4988]: I1008 18:15:07.058886 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-7ck2j"] Oct 08 18:15:07 crc kubenswrapper[4988]: I1008 18:15:07.062143 4988 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-7ck2j"] Oct 08 18:15:07 crc kubenswrapper[4988]: I1008 18:15:07.069260 4988 scope.go:117] "RemoveContainer" containerID="615d0e89e4190c703d515e7df5d5b3b444dd61b73916cb4b40cc5c9ce105c872" Oct 08 18:15:07 crc kubenswrapper[4988]: E1008 18:15:07.070508 4988 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"615d0e89e4190c703d515e7df5d5b3b444dd61b73916cb4b40cc5c9ce105c872\": container with ID starting with 615d0e89e4190c703d515e7df5d5b3b444dd61b73916cb4b40cc5c9ce105c872 not found: ID does not exist" containerID="615d0e89e4190c703d515e7df5d5b3b444dd61b73916cb4b40cc5c9ce105c872" Oct 08 18:15:07 crc kubenswrapper[4988]: I1008 18:15:07.070571 4988 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"615d0e89e4190c703d515e7df5d5b3b444dd61b73916cb4b40cc5c9ce105c872"} err="failed to get container status \"615d0e89e4190c703d515e7df5d5b3b444dd61b73916cb4b40cc5c9ce105c872\": rpc error: code = NotFound desc = could not find container \"615d0e89e4190c703d515e7df5d5b3b444dd61b73916cb4b40cc5c9ce105c872\": container with ID starting with 615d0e89e4190c703d515e7df5d5b3b444dd61b73916cb4b40cc5c9ce105c872 not found: ID does not exist" Oct 08 18:15:07 crc kubenswrapper[4988]: I1008 18:15:07.070611 4988 scope.go:117] "RemoveContainer" containerID="7285e62439fa2c40d6fb7ad0da89aaa7edc907b20ba48be35aae53b4d742c498" Oct 08 18:15:07 crc kubenswrapper[4988]: E1008 18:15:07.071217 4988 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7285e62439fa2c40d6fb7ad0da89aaa7edc907b20ba48be35aae53b4d742c498\": container with ID starting with 7285e62439fa2c40d6fb7ad0da89aaa7edc907b20ba48be35aae53b4d742c498 not found: ID does not exist" containerID="7285e62439fa2c40d6fb7ad0da89aaa7edc907b20ba48be35aae53b4d742c498" Oct 08 18:15:07 crc kubenswrapper[4988]: I1008 18:15:07.071253 4988 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7285e62439fa2c40d6fb7ad0da89aaa7edc907b20ba48be35aae53b4d742c498"} err="failed to get container status \"7285e62439fa2c40d6fb7ad0da89aaa7edc907b20ba48be35aae53b4d742c498\": rpc error: code = NotFound desc = could not find container \"7285e62439fa2c40d6fb7ad0da89aaa7edc907b20ba48be35aae53b4d742c498\": container with ID starting with 7285e62439fa2c40d6fb7ad0da89aaa7edc907b20ba48be35aae53b4d742c498 not found: ID does not exist" Oct 08 18:15:07 crc kubenswrapper[4988]: I1008 18:15:07.071278 4988 scope.go:117] "RemoveContainer" containerID="aaf1ecf558697a87bf3a10117a2784a64dff7aae5eb9a2cf38599f88ef41a71b" Oct 08 18:15:07 crc kubenswrapper[4988]: E1008 18:15:07.071772 4988 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"aaf1ecf558697a87bf3a10117a2784a64dff7aae5eb9a2cf38599f88ef41a71b\": container with ID starting with aaf1ecf558697a87bf3a10117a2784a64dff7aae5eb9a2cf38599f88ef41a71b not found: ID does not exist" containerID="aaf1ecf558697a87bf3a10117a2784a64dff7aae5eb9a2cf38599f88ef41a71b" Oct 08 18:15:07 crc kubenswrapper[4988]: I1008 18:15:07.071873 4988 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"aaf1ecf558697a87bf3a10117a2784a64dff7aae5eb9a2cf38599f88ef41a71b"} err="failed to get container status \"aaf1ecf558697a87bf3a10117a2784a64dff7aae5eb9a2cf38599f88ef41a71b\": rpc error: code = NotFound desc = could not find container \"aaf1ecf558697a87bf3a10117a2784a64dff7aae5eb9a2cf38599f88ef41a71b\": container with ID starting with aaf1ecf558697a87bf3a10117a2784a64dff7aae5eb9a2cf38599f88ef41a71b not found: ID does not exist" Oct 08 18:15:07 crc kubenswrapper[4988]: I1008 18:15:07.071966 4988 scope.go:117] "RemoveContainer" containerID="80113d52537f531d6bba38b5509f7a18c44a61e0f36a58742df56936d485d0f9" Oct 08 18:15:07 crc kubenswrapper[4988]: I1008 18:15:07.087164 4988 scope.go:117] "RemoveContainer" containerID="b54e8a500850c375abfec8de4ae46eaae5f4c989bc6e31bcc77229321643e998" Oct 08 18:15:07 crc kubenswrapper[4988]: I1008 18:15:07.090096 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-294dt"] Oct 08 18:15:07 crc kubenswrapper[4988]: I1008 18:15:07.094836 4988 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-294dt"] Oct 08 18:15:07 crc kubenswrapper[4988]: I1008 18:15:07.107229 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-dmdqd"] Oct 08 18:15:07 crc kubenswrapper[4988]: I1008 18:15:07.110608 4988 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-dmdqd"] Oct 08 18:15:07 crc kubenswrapper[4988]: I1008 18:15:07.122664 4988 scope.go:117] "RemoveContainer" containerID="33110da129142b21ee73fce202569b578ade65662103418e94513a5f6a8b5cfc" Oct 08 18:15:07 crc kubenswrapper[4988]: I1008 18:15:07.148988 4988 scope.go:117] "RemoveContainer" containerID="80113d52537f531d6bba38b5509f7a18c44a61e0f36a58742df56936d485d0f9" Oct 08 18:15:07 crc kubenswrapper[4988]: E1008 18:15:07.149675 4988 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"80113d52537f531d6bba38b5509f7a18c44a61e0f36a58742df56936d485d0f9\": container with ID starting with 80113d52537f531d6bba38b5509f7a18c44a61e0f36a58742df56936d485d0f9 not found: ID does not exist" containerID="80113d52537f531d6bba38b5509f7a18c44a61e0f36a58742df56936d485d0f9" Oct 08 18:15:07 crc kubenswrapper[4988]: I1008 18:15:07.149740 4988 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"80113d52537f531d6bba38b5509f7a18c44a61e0f36a58742df56936d485d0f9"} err="failed to get container status \"80113d52537f531d6bba38b5509f7a18c44a61e0f36a58742df56936d485d0f9\": rpc error: code = NotFound desc = could not find container \"80113d52537f531d6bba38b5509f7a18c44a61e0f36a58742df56936d485d0f9\": container with ID starting with 80113d52537f531d6bba38b5509f7a18c44a61e0f36a58742df56936d485d0f9 not found: ID does not exist" Oct 08 18:15:07 crc kubenswrapper[4988]: I1008 18:15:07.149788 4988 scope.go:117] "RemoveContainer" containerID="b54e8a500850c375abfec8de4ae46eaae5f4c989bc6e31bcc77229321643e998" Oct 08 18:15:07 crc kubenswrapper[4988]: E1008 18:15:07.150251 4988 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b54e8a500850c375abfec8de4ae46eaae5f4c989bc6e31bcc77229321643e998\": container with ID starting with b54e8a500850c375abfec8de4ae46eaae5f4c989bc6e31bcc77229321643e998 not found: ID does not exist" containerID="b54e8a500850c375abfec8de4ae46eaae5f4c989bc6e31bcc77229321643e998" Oct 08 18:15:07 crc kubenswrapper[4988]: I1008 18:15:07.150302 4988 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b54e8a500850c375abfec8de4ae46eaae5f4c989bc6e31bcc77229321643e998"} err="failed to get container status \"b54e8a500850c375abfec8de4ae46eaae5f4c989bc6e31bcc77229321643e998\": rpc error: code = NotFound desc = could not find container \"b54e8a500850c375abfec8de4ae46eaae5f4c989bc6e31bcc77229321643e998\": container with ID starting with b54e8a500850c375abfec8de4ae46eaae5f4c989bc6e31bcc77229321643e998 not found: ID does not exist" Oct 08 18:15:07 crc kubenswrapper[4988]: I1008 18:15:07.150508 4988 scope.go:117] "RemoveContainer" containerID="33110da129142b21ee73fce202569b578ade65662103418e94513a5f6a8b5cfc" Oct 08 18:15:07 crc kubenswrapper[4988]: E1008 18:15:07.150923 4988 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"33110da129142b21ee73fce202569b578ade65662103418e94513a5f6a8b5cfc\": container with ID starting with 33110da129142b21ee73fce202569b578ade65662103418e94513a5f6a8b5cfc not found: ID does not exist" containerID="33110da129142b21ee73fce202569b578ade65662103418e94513a5f6a8b5cfc" Oct 08 18:15:07 crc kubenswrapper[4988]: I1008 18:15:07.150970 4988 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"33110da129142b21ee73fce202569b578ade65662103418e94513a5f6a8b5cfc"} err="failed to get container status \"33110da129142b21ee73fce202569b578ade65662103418e94513a5f6a8b5cfc\": rpc error: code = NotFound desc = could not find container \"33110da129142b21ee73fce202569b578ade65662103418e94513a5f6a8b5cfc\": container with ID starting with 33110da129142b21ee73fce202569b578ade65662103418e94513a5f6a8b5cfc not found: ID does not exist" Oct 08 18:15:07 crc kubenswrapper[4988]: I1008 18:15:07.151020 4988 scope.go:117] "RemoveContainer" containerID="021595f654ab0cc3eebfd44bb31805da9eef8f2dbc83c40b88a030a1e8fe3c54" Oct 08 18:15:07 crc kubenswrapper[4988]: I1008 18:15:07.174344 4988 scope.go:117] "RemoveContainer" containerID="7fe16147afcdef2e218e05e65b8ff5119b856be161a400f38a5400ea66ea2f8a" Oct 08 18:15:07 crc kubenswrapper[4988]: I1008 18:15:07.194962 4988 scope.go:117] "RemoveContainer" containerID="7295a0e0d87f1abcaabd86b013caddd1d3952a984ac114b32b7d628ca250dfed" Oct 08 18:15:07 crc kubenswrapper[4988]: I1008 18:15:07.214039 4988 scope.go:117] "RemoveContainer" containerID="021595f654ab0cc3eebfd44bb31805da9eef8f2dbc83c40b88a030a1e8fe3c54" Oct 08 18:15:07 crc kubenswrapper[4988]: E1008 18:15:07.216526 4988 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"021595f654ab0cc3eebfd44bb31805da9eef8f2dbc83c40b88a030a1e8fe3c54\": container with ID starting with 021595f654ab0cc3eebfd44bb31805da9eef8f2dbc83c40b88a030a1e8fe3c54 not found: ID does not exist" containerID="021595f654ab0cc3eebfd44bb31805da9eef8f2dbc83c40b88a030a1e8fe3c54" Oct 08 18:15:07 crc kubenswrapper[4988]: I1008 18:15:07.216594 4988 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"021595f654ab0cc3eebfd44bb31805da9eef8f2dbc83c40b88a030a1e8fe3c54"} err="failed to get container status \"021595f654ab0cc3eebfd44bb31805da9eef8f2dbc83c40b88a030a1e8fe3c54\": rpc error: code = NotFound desc = could not find container \"021595f654ab0cc3eebfd44bb31805da9eef8f2dbc83c40b88a030a1e8fe3c54\": container with ID starting with 021595f654ab0cc3eebfd44bb31805da9eef8f2dbc83c40b88a030a1e8fe3c54 not found: ID does not exist" Oct 08 18:15:07 crc kubenswrapper[4988]: I1008 18:15:07.216648 4988 scope.go:117] "RemoveContainer" containerID="7fe16147afcdef2e218e05e65b8ff5119b856be161a400f38a5400ea66ea2f8a" Oct 08 18:15:07 crc kubenswrapper[4988]: E1008 18:15:07.217430 4988 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7fe16147afcdef2e218e05e65b8ff5119b856be161a400f38a5400ea66ea2f8a\": container with ID starting with 7fe16147afcdef2e218e05e65b8ff5119b856be161a400f38a5400ea66ea2f8a not found: ID does not exist" containerID="7fe16147afcdef2e218e05e65b8ff5119b856be161a400f38a5400ea66ea2f8a" Oct 08 18:15:07 crc kubenswrapper[4988]: I1008 18:15:07.217476 4988 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7fe16147afcdef2e218e05e65b8ff5119b856be161a400f38a5400ea66ea2f8a"} err="failed to get container status \"7fe16147afcdef2e218e05e65b8ff5119b856be161a400f38a5400ea66ea2f8a\": rpc error: code = NotFound desc = could not find container \"7fe16147afcdef2e218e05e65b8ff5119b856be161a400f38a5400ea66ea2f8a\": container with ID starting with 7fe16147afcdef2e218e05e65b8ff5119b856be161a400f38a5400ea66ea2f8a not found: ID does not exist" Oct 08 18:15:07 crc kubenswrapper[4988]: I1008 18:15:07.217581 4988 scope.go:117] "RemoveContainer" containerID="7295a0e0d87f1abcaabd86b013caddd1d3952a984ac114b32b7d628ca250dfed" Oct 08 18:15:07 crc kubenswrapper[4988]: E1008 18:15:07.221274 4988 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7295a0e0d87f1abcaabd86b013caddd1d3952a984ac114b32b7d628ca250dfed\": container with ID starting with 7295a0e0d87f1abcaabd86b013caddd1d3952a984ac114b32b7d628ca250dfed not found: ID does not exist" containerID="7295a0e0d87f1abcaabd86b013caddd1d3952a984ac114b32b7d628ca250dfed" Oct 08 18:15:07 crc kubenswrapper[4988]: I1008 18:15:07.221312 4988 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7295a0e0d87f1abcaabd86b013caddd1d3952a984ac114b32b7d628ca250dfed"} err="failed to get container status \"7295a0e0d87f1abcaabd86b013caddd1d3952a984ac114b32b7d628ca250dfed\": rpc error: code = NotFound desc = could not find container \"7295a0e0d87f1abcaabd86b013caddd1d3952a984ac114b32b7d628ca250dfed\": container with ID starting with 7295a0e0d87f1abcaabd86b013caddd1d3952a984ac114b32b7d628ca250dfed not found: ID does not exist" Oct 08 18:15:07 crc kubenswrapper[4988]: I1008 18:15:07.221340 4988 scope.go:117] "RemoveContainer" containerID="376064a31a6403d1a62039ad1b6d1d240e5e224930f9b781cffd57e988576c51" Oct 08 18:15:07 crc kubenswrapper[4988]: I1008 18:15:07.238248 4988 scope.go:117] "RemoveContainer" containerID="31cf1e847f46eb07b87ffc654f5981503665b3b75fd7897c9b7e31e08d1e91e7" Oct 08 18:15:07 crc kubenswrapper[4988]: I1008 18:15:07.245610 4988 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="15a52185-0755-45a6-b665-c1185d1996b4" path="/var/lib/kubelet/pods/15a52185-0755-45a6-b665-c1185d1996b4/volumes" Oct 08 18:15:07 crc kubenswrapper[4988]: I1008 18:15:07.246647 4988 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="38d2198b-2aab-47d2-aa2e-93436ff69625" path="/var/lib/kubelet/pods/38d2198b-2aab-47d2-aa2e-93436ff69625/volumes" Oct 08 18:15:07 crc kubenswrapper[4988]: I1008 18:15:07.250587 4988 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7c0e2c11-212a-46ed-b72a-911bde29199a" path="/var/lib/kubelet/pods/7c0e2c11-212a-46ed-b72a-911bde29199a/volumes" Oct 08 18:15:07 crc kubenswrapper[4988]: I1008 18:15:07.251434 4988 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8bf9175c-5101-4fcd-8cba-e113be11dde4" path="/var/lib/kubelet/pods/8bf9175c-5101-4fcd-8cba-e113be11dde4/volumes" Oct 08 18:15:07 crc kubenswrapper[4988]: I1008 18:15:07.252535 4988 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b94d3b86-2fd1-4045-9a5e-dc6c296cfba4" path="/var/lib/kubelet/pods/b94d3b86-2fd1-4045-9a5e-dc6c296cfba4/volumes" Oct 08 18:15:07 crc kubenswrapper[4988]: I1008 18:15:07.270214 4988 scope.go:117] "RemoveContainer" containerID="3713f5d1afcd9846c811ff79adf171b4d51e6c8706b9d87fc2b59eb4d8f3f964" Oct 08 18:15:07 crc kubenswrapper[4988]: I1008 18:15:07.284198 4988 scope.go:117] "RemoveContainer" containerID="376064a31a6403d1a62039ad1b6d1d240e5e224930f9b781cffd57e988576c51" Oct 08 18:15:07 crc kubenswrapper[4988]: E1008 18:15:07.284910 4988 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"376064a31a6403d1a62039ad1b6d1d240e5e224930f9b781cffd57e988576c51\": container with ID starting with 376064a31a6403d1a62039ad1b6d1d240e5e224930f9b781cffd57e988576c51 not found: ID does not exist" containerID="376064a31a6403d1a62039ad1b6d1d240e5e224930f9b781cffd57e988576c51" Oct 08 18:15:07 crc kubenswrapper[4988]: I1008 18:15:07.284941 4988 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"376064a31a6403d1a62039ad1b6d1d240e5e224930f9b781cffd57e988576c51"} err="failed to get container status \"376064a31a6403d1a62039ad1b6d1d240e5e224930f9b781cffd57e988576c51\": rpc error: code = NotFound desc = could not find container \"376064a31a6403d1a62039ad1b6d1d240e5e224930f9b781cffd57e988576c51\": container with ID starting with 376064a31a6403d1a62039ad1b6d1d240e5e224930f9b781cffd57e988576c51 not found: ID does not exist" Oct 08 18:15:07 crc kubenswrapper[4988]: I1008 18:15:07.284973 4988 scope.go:117] "RemoveContainer" containerID="31cf1e847f46eb07b87ffc654f5981503665b3b75fd7897c9b7e31e08d1e91e7" Oct 08 18:15:07 crc kubenswrapper[4988]: E1008 18:15:07.285326 4988 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"31cf1e847f46eb07b87ffc654f5981503665b3b75fd7897c9b7e31e08d1e91e7\": container with ID starting with 31cf1e847f46eb07b87ffc654f5981503665b3b75fd7897c9b7e31e08d1e91e7 not found: ID does not exist" containerID="31cf1e847f46eb07b87ffc654f5981503665b3b75fd7897c9b7e31e08d1e91e7" Oct 08 18:15:07 crc kubenswrapper[4988]: I1008 18:15:07.285374 4988 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"31cf1e847f46eb07b87ffc654f5981503665b3b75fd7897c9b7e31e08d1e91e7"} err="failed to get container status \"31cf1e847f46eb07b87ffc654f5981503665b3b75fd7897c9b7e31e08d1e91e7\": rpc error: code = NotFound desc = could not find container \"31cf1e847f46eb07b87ffc654f5981503665b3b75fd7897c9b7e31e08d1e91e7\": container with ID starting with 31cf1e847f46eb07b87ffc654f5981503665b3b75fd7897c9b7e31e08d1e91e7 not found: ID does not exist" Oct 08 18:15:07 crc kubenswrapper[4988]: I1008 18:15:07.285439 4988 scope.go:117] "RemoveContainer" containerID="3713f5d1afcd9846c811ff79adf171b4d51e6c8706b9d87fc2b59eb4d8f3f964" Oct 08 18:15:07 crc kubenswrapper[4988]: E1008 18:15:07.285874 4988 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3713f5d1afcd9846c811ff79adf171b4d51e6c8706b9d87fc2b59eb4d8f3f964\": container with ID starting with 3713f5d1afcd9846c811ff79adf171b4d51e6c8706b9d87fc2b59eb4d8f3f964 not found: ID does not exist" containerID="3713f5d1afcd9846c811ff79adf171b4d51e6c8706b9d87fc2b59eb4d8f3f964" Oct 08 18:15:07 crc kubenswrapper[4988]: I1008 18:15:07.285913 4988 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3713f5d1afcd9846c811ff79adf171b4d51e6c8706b9d87fc2b59eb4d8f3f964"} err="failed to get container status \"3713f5d1afcd9846c811ff79adf171b4d51e6c8706b9d87fc2b59eb4d8f3f964\": rpc error: code = NotFound desc = could not find container \"3713f5d1afcd9846c811ff79adf171b4d51e6c8706b9d87fc2b59eb4d8f3f964\": container with ID starting with 3713f5d1afcd9846c811ff79adf171b4d51e6c8706b9d87fc2b59eb4d8f3f964 not found: ID does not exist" Oct 08 18:15:07 crc kubenswrapper[4988]: I1008 18:15:07.285929 4988 scope.go:117] "RemoveContainer" containerID="20dfd87103b1c34d1d77781ce5e72b28df28ecc2efa06881c31118ab87686314" Oct 08 18:15:07 crc kubenswrapper[4988]: I1008 18:15:07.302810 4988 scope.go:117] "RemoveContainer" containerID="20dfd87103b1c34d1d77781ce5e72b28df28ecc2efa06881c31118ab87686314" Oct 08 18:15:07 crc kubenswrapper[4988]: E1008 18:15:07.304395 4988 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"20dfd87103b1c34d1d77781ce5e72b28df28ecc2efa06881c31118ab87686314\": container with ID starting with 20dfd87103b1c34d1d77781ce5e72b28df28ecc2efa06881c31118ab87686314 not found: ID does not exist" containerID="20dfd87103b1c34d1d77781ce5e72b28df28ecc2efa06881c31118ab87686314" Oct 08 18:15:07 crc kubenswrapper[4988]: I1008 18:15:07.304433 4988 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"20dfd87103b1c34d1d77781ce5e72b28df28ecc2efa06881c31118ab87686314"} err="failed to get container status \"20dfd87103b1c34d1d77781ce5e72b28df28ecc2efa06881c31118ab87686314\": rpc error: code = NotFound desc = could not find container \"20dfd87103b1c34d1d77781ce5e72b28df28ecc2efa06881c31118ab87686314\": container with ID starting with 20dfd87103b1c34d1d77781ce5e72b28df28ecc2efa06881c31118ab87686314 not found: ID does not exist" Oct 08 18:15:07 crc kubenswrapper[4988]: I1008 18:15:07.985127 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-kb5xn" Oct 08 18:15:08 crc kubenswrapper[4988]: I1008 18:15:08.060864 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-gl4lh"] Oct 08 18:15:08 crc kubenswrapper[4988]: E1008 18:15:08.061125 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b94d3b86-2fd1-4045-9a5e-dc6c296cfba4" containerName="registry-server" Oct 08 18:15:08 crc kubenswrapper[4988]: I1008 18:15:08.061140 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="b94d3b86-2fd1-4045-9a5e-dc6c296cfba4" containerName="registry-server" Oct 08 18:15:08 crc kubenswrapper[4988]: E1008 18:15:08.061155 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="38d2198b-2aab-47d2-aa2e-93436ff69625" containerName="extract-utilities" Oct 08 18:15:08 crc kubenswrapper[4988]: I1008 18:15:08.061162 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="38d2198b-2aab-47d2-aa2e-93436ff69625" containerName="extract-utilities" Oct 08 18:15:08 crc kubenswrapper[4988]: E1008 18:15:08.061172 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7c0e2c11-212a-46ed-b72a-911bde29199a" containerName="extract-content" Oct 08 18:15:08 crc kubenswrapper[4988]: I1008 18:15:08.061181 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="7c0e2c11-212a-46ed-b72a-911bde29199a" containerName="extract-content" Oct 08 18:15:08 crc kubenswrapper[4988]: E1008 18:15:08.061189 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="15a52185-0755-45a6-b665-c1185d1996b4" containerName="extract-utilities" Oct 08 18:15:08 crc kubenswrapper[4988]: I1008 18:15:08.061197 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="15a52185-0755-45a6-b665-c1185d1996b4" containerName="extract-utilities" Oct 08 18:15:08 crc kubenswrapper[4988]: E1008 18:15:08.061208 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7c0e2c11-212a-46ed-b72a-911bde29199a" containerName="extract-utilities" Oct 08 18:15:08 crc kubenswrapper[4988]: I1008 18:15:08.061214 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="7c0e2c11-212a-46ed-b72a-911bde29199a" containerName="extract-utilities" Oct 08 18:15:08 crc kubenswrapper[4988]: E1008 18:15:08.061224 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="38d2198b-2aab-47d2-aa2e-93436ff69625" containerName="extract-content" Oct 08 18:15:08 crc kubenswrapper[4988]: I1008 18:15:08.061230 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="38d2198b-2aab-47d2-aa2e-93436ff69625" containerName="extract-content" Oct 08 18:15:08 crc kubenswrapper[4988]: E1008 18:15:08.061245 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b94d3b86-2fd1-4045-9a5e-dc6c296cfba4" containerName="extract-content" Oct 08 18:15:08 crc kubenswrapper[4988]: I1008 18:15:08.061253 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="b94d3b86-2fd1-4045-9a5e-dc6c296cfba4" containerName="extract-content" Oct 08 18:15:08 crc kubenswrapper[4988]: E1008 18:15:08.061262 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="15a52185-0755-45a6-b665-c1185d1996b4" containerName="registry-server" Oct 08 18:15:08 crc kubenswrapper[4988]: I1008 18:15:08.061269 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="15a52185-0755-45a6-b665-c1185d1996b4" containerName="registry-server" Oct 08 18:15:08 crc kubenswrapper[4988]: E1008 18:15:08.061279 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="15a52185-0755-45a6-b665-c1185d1996b4" containerName="extract-content" Oct 08 18:15:08 crc kubenswrapper[4988]: I1008 18:15:08.061286 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="15a52185-0755-45a6-b665-c1185d1996b4" containerName="extract-content" Oct 08 18:15:08 crc kubenswrapper[4988]: E1008 18:15:08.061294 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="38d2198b-2aab-47d2-aa2e-93436ff69625" containerName="registry-server" Oct 08 18:15:08 crc kubenswrapper[4988]: I1008 18:15:08.061301 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="38d2198b-2aab-47d2-aa2e-93436ff69625" containerName="registry-server" Oct 08 18:15:08 crc kubenswrapper[4988]: E1008 18:15:08.061313 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b94d3b86-2fd1-4045-9a5e-dc6c296cfba4" containerName="extract-utilities" Oct 08 18:15:08 crc kubenswrapper[4988]: I1008 18:15:08.061320 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="b94d3b86-2fd1-4045-9a5e-dc6c296cfba4" containerName="extract-utilities" Oct 08 18:15:08 crc kubenswrapper[4988]: E1008 18:15:08.061329 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7c0e2c11-212a-46ed-b72a-911bde29199a" containerName="registry-server" Oct 08 18:15:08 crc kubenswrapper[4988]: I1008 18:15:08.061338 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="7c0e2c11-212a-46ed-b72a-911bde29199a" containerName="registry-server" Oct 08 18:15:08 crc kubenswrapper[4988]: E1008 18:15:08.061347 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8bf9175c-5101-4fcd-8cba-e113be11dde4" containerName="marketplace-operator" Oct 08 18:15:08 crc kubenswrapper[4988]: I1008 18:15:08.061354 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="8bf9175c-5101-4fcd-8cba-e113be11dde4" containerName="marketplace-operator" Oct 08 18:15:08 crc kubenswrapper[4988]: I1008 18:15:08.061480 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="7c0e2c11-212a-46ed-b72a-911bde29199a" containerName="registry-server" Oct 08 18:15:08 crc kubenswrapper[4988]: I1008 18:15:08.061496 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="38d2198b-2aab-47d2-aa2e-93436ff69625" containerName="registry-server" Oct 08 18:15:08 crc kubenswrapper[4988]: I1008 18:15:08.061509 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="8bf9175c-5101-4fcd-8cba-e113be11dde4" containerName="marketplace-operator" Oct 08 18:15:08 crc kubenswrapper[4988]: I1008 18:15:08.061520 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="b94d3b86-2fd1-4045-9a5e-dc6c296cfba4" containerName="registry-server" Oct 08 18:15:08 crc kubenswrapper[4988]: I1008 18:15:08.061528 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="15a52185-0755-45a6-b665-c1185d1996b4" containerName="registry-server" Oct 08 18:15:08 crc kubenswrapper[4988]: I1008 18:15:08.062440 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-gl4lh" Oct 08 18:15:08 crc kubenswrapper[4988]: I1008 18:15:08.065489 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Oct 08 18:15:08 crc kubenswrapper[4988]: I1008 18:15:08.072798 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-gl4lh"] Oct 08 18:15:08 crc kubenswrapper[4988]: I1008 18:15:08.088926 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-74v6f\" (UniqueName: \"kubernetes.io/projected/6936c4e6-f11e-47a2-870a-0df3df4ed33c-kube-api-access-74v6f\") pod \"redhat-marketplace-gl4lh\" (UID: \"6936c4e6-f11e-47a2-870a-0df3df4ed33c\") " pod="openshift-marketplace/redhat-marketplace-gl4lh" Oct 08 18:15:08 crc kubenswrapper[4988]: I1008 18:15:08.089025 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6936c4e6-f11e-47a2-870a-0df3df4ed33c-utilities\") pod \"redhat-marketplace-gl4lh\" (UID: \"6936c4e6-f11e-47a2-870a-0df3df4ed33c\") " pod="openshift-marketplace/redhat-marketplace-gl4lh" Oct 08 18:15:08 crc kubenswrapper[4988]: I1008 18:15:08.089149 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6936c4e6-f11e-47a2-870a-0df3df4ed33c-catalog-content\") pod \"redhat-marketplace-gl4lh\" (UID: \"6936c4e6-f11e-47a2-870a-0df3df4ed33c\") " pod="openshift-marketplace/redhat-marketplace-gl4lh" Oct 08 18:15:08 crc kubenswrapper[4988]: I1008 18:15:08.190008 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-74v6f\" (UniqueName: \"kubernetes.io/projected/6936c4e6-f11e-47a2-870a-0df3df4ed33c-kube-api-access-74v6f\") pod \"redhat-marketplace-gl4lh\" (UID: \"6936c4e6-f11e-47a2-870a-0df3df4ed33c\") " pod="openshift-marketplace/redhat-marketplace-gl4lh" Oct 08 18:15:08 crc kubenswrapper[4988]: I1008 18:15:08.190142 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6936c4e6-f11e-47a2-870a-0df3df4ed33c-utilities\") pod \"redhat-marketplace-gl4lh\" (UID: \"6936c4e6-f11e-47a2-870a-0df3df4ed33c\") " pod="openshift-marketplace/redhat-marketplace-gl4lh" Oct 08 18:15:08 crc kubenswrapper[4988]: I1008 18:15:08.190176 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6936c4e6-f11e-47a2-870a-0df3df4ed33c-catalog-content\") pod \"redhat-marketplace-gl4lh\" (UID: \"6936c4e6-f11e-47a2-870a-0df3df4ed33c\") " pod="openshift-marketplace/redhat-marketplace-gl4lh" Oct 08 18:15:08 crc kubenswrapper[4988]: I1008 18:15:08.190869 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6936c4e6-f11e-47a2-870a-0df3df4ed33c-catalog-content\") pod \"redhat-marketplace-gl4lh\" (UID: \"6936c4e6-f11e-47a2-870a-0df3df4ed33c\") " pod="openshift-marketplace/redhat-marketplace-gl4lh" Oct 08 18:15:08 crc kubenswrapper[4988]: I1008 18:15:08.190886 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6936c4e6-f11e-47a2-870a-0df3df4ed33c-utilities\") pod \"redhat-marketplace-gl4lh\" (UID: \"6936c4e6-f11e-47a2-870a-0df3df4ed33c\") " pod="openshift-marketplace/redhat-marketplace-gl4lh" Oct 08 18:15:08 crc kubenswrapper[4988]: I1008 18:15:08.213885 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-74v6f\" (UniqueName: \"kubernetes.io/projected/6936c4e6-f11e-47a2-870a-0df3df4ed33c-kube-api-access-74v6f\") pod \"redhat-marketplace-gl4lh\" (UID: \"6936c4e6-f11e-47a2-870a-0df3df4ed33c\") " pod="openshift-marketplace/redhat-marketplace-gl4lh" Oct 08 18:15:08 crc kubenswrapper[4988]: I1008 18:15:08.270505 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-vwv6l"] Oct 08 18:15:08 crc kubenswrapper[4988]: I1008 18:15:08.271779 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-vwv6l" Oct 08 18:15:08 crc kubenswrapper[4988]: I1008 18:15:08.275289 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Oct 08 18:15:08 crc kubenswrapper[4988]: I1008 18:15:08.305558 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-vwv6l"] Oct 08 18:15:08 crc kubenswrapper[4988]: I1008 18:15:08.383299 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-gl4lh" Oct 08 18:15:08 crc kubenswrapper[4988]: I1008 18:15:08.393175 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t6mhl\" (UniqueName: \"kubernetes.io/projected/4b10186d-3d69-4998-9162-89df44da5521-kube-api-access-t6mhl\") pod \"redhat-operators-vwv6l\" (UID: \"4b10186d-3d69-4998-9162-89df44da5521\") " pod="openshift-marketplace/redhat-operators-vwv6l" Oct 08 18:15:08 crc kubenswrapper[4988]: I1008 18:15:08.393257 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4b10186d-3d69-4998-9162-89df44da5521-catalog-content\") pod \"redhat-operators-vwv6l\" (UID: \"4b10186d-3d69-4998-9162-89df44da5521\") " pod="openshift-marketplace/redhat-operators-vwv6l" Oct 08 18:15:08 crc kubenswrapper[4988]: I1008 18:15:08.393306 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4b10186d-3d69-4998-9162-89df44da5521-utilities\") pod \"redhat-operators-vwv6l\" (UID: \"4b10186d-3d69-4998-9162-89df44da5521\") " pod="openshift-marketplace/redhat-operators-vwv6l" Oct 08 18:15:08 crc kubenswrapper[4988]: I1008 18:15:08.494559 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4b10186d-3d69-4998-9162-89df44da5521-utilities\") pod \"redhat-operators-vwv6l\" (UID: \"4b10186d-3d69-4998-9162-89df44da5521\") " pod="openshift-marketplace/redhat-operators-vwv6l" Oct 08 18:15:08 crc kubenswrapper[4988]: I1008 18:15:08.495081 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t6mhl\" (UniqueName: \"kubernetes.io/projected/4b10186d-3d69-4998-9162-89df44da5521-kube-api-access-t6mhl\") pod \"redhat-operators-vwv6l\" (UID: \"4b10186d-3d69-4998-9162-89df44da5521\") " pod="openshift-marketplace/redhat-operators-vwv6l" Oct 08 18:15:08 crc kubenswrapper[4988]: I1008 18:15:08.495145 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4b10186d-3d69-4998-9162-89df44da5521-catalog-content\") pod \"redhat-operators-vwv6l\" (UID: \"4b10186d-3d69-4998-9162-89df44da5521\") " pod="openshift-marketplace/redhat-operators-vwv6l" Oct 08 18:15:08 crc kubenswrapper[4988]: I1008 18:15:08.495408 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4b10186d-3d69-4998-9162-89df44da5521-utilities\") pod \"redhat-operators-vwv6l\" (UID: \"4b10186d-3d69-4998-9162-89df44da5521\") " pod="openshift-marketplace/redhat-operators-vwv6l" Oct 08 18:15:08 crc kubenswrapper[4988]: I1008 18:15:08.495622 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4b10186d-3d69-4998-9162-89df44da5521-catalog-content\") pod \"redhat-operators-vwv6l\" (UID: \"4b10186d-3d69-4998-9162-89df44da5521\") " pod="openshift-marketplace/redhat-operators-vwv6l" Oct 08 18:15:08 crc kubenswrapper[4988]: I1008 18:15:08.517997 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t6mhl\" (UniqueName: \"kubernetes.io/projected/4b10186d-3d69-4998-9162-89df44da5521-kube-api-access-t6mhl\") pod \"redhat-operators-vwv6l\" (UID: \"4b10186d-3d69-4998-9162-89df44da5521\") " pod="openshift-marketplace/redhat-operators-vwv6l" Oct 08 18:15:08 crc kubenswrapper[4988]: I1008 18:15:08.592196 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-vwv6l" Oct 08 18:15:08 crc kubenswrapper[4988]: I1008 18:15:08.793483 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-gl4lh"] Oct 08 18:15:08 crc kubenswrapper[4988]: W1008 18:15:08.799911 4988 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6936c4e6_f11e_47a2_870a_0df3df4ed33c.slice/crio-5af82df24bae24e5397be528a3f9cdff95fc4f7b055c439eedca0de3a493cb75 WatchSource:0}: Error finding container 5af82df24bae24e5397be528a3f9cdff95fc4f7b055c439eedca0de3a493cb75: Status 404 returned error can't find the container with id 5af82df24bae24e5397be528a3f9cdff95fc4f7b055c439eedca0de3a493cb75 Oct 08 18:15:08 crc kubenswrapper[4988]: I1008 18:15:08.983245 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-vwv6l"] Oct 08 18:15:08 crc kubenswrapper[4988]: I1008 18:15:08.989557 4988 generic.go:334] "Generic (PLEG): container finished" podID="6936c4e6-f11e-47a2-870a-0df3df4ed33c" containerID="6d25e07c48b1d43896c981b6952f0dae75667438a8ad94f324fba2ea80a89440" exitCode=0 Oct 08 18:15:08 crc kubenswrapper[4988]: I1008 18:15:08.989613 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-gl4lh" event={"ID":"6936c4e6-f11e-47a2-870a-0df3df4ed33c","Type":"ContainerDied","Data":"6d25e07c48b1d43896c981b6952f0dae75667438a8ad94f324fba2ea80a89440"} Oct 08 18:15:08 crc kubenswrapper[4988]: I1008 18:15:08.989651 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-gl4lh" event={"ID":"6936c4e6-f11e-47a2-870a-0df3df4ed33c","Type":"ContainerStarted","Data":"5af82df24bae24e5397be528a3f9cdff95fc4f7b055c439eedca0de3a493cb75"} Oct 08 18:15:09 crc kubenswrapper[4988]: W1008 18:15:09.019894 4988 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4b10186d_3d69_4998_9162_89df44da5521.slice/crio-17d21191fde6133b6543cf09625669f85f65a836e78a5e1c7a3bd079a161076c WatchSource:0}: Error finding container 17d21191fde6133b6543cf09625669f85f65a836e78a5e1c7a3bd079a161076c: Status 404 returned error can't find the container with id 17d21191fde6133b6543cf09625669f85f65a836e78a5e1c7a3bd079a161076c Oct 08 18:15:10 crc kubenswrapper[4988]: I1008 18:15:10.003811 4988 generic.go:334] "Generic (PLEG): container finished" podID="4b10186d-3d69-4998-9162-89df44da5521" containerID="197c8465314669a6e2d6e2fd416db05b3828e44cbcefdbea90919838e2b8669a" exitCode=0 Oct 08 18:15:10 crc kubenswrapper[4988]: I1008 18:15:10.003882 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vwv6l" event={"ID":"4b10186d-3d69-4998-9162-89df44da5521","Type":"ContainerDied","Data":"197c8465314669a6e2d6e2fd416db05b3828e44cbcefdbea90919838e2b8669a"} Oct 08 18:15:10 crc kubenswrapper[4988]: I1008 18:15:10.004306 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vwv6l" event={"ID":"4b10186d-3d69-4998-9162-89df44da5521","Type":"ContainerStarted","Data":"17d21191fde6133b6543cf09625669f85f65a836e78a5e1c7a3bd079a161076c"} Oct 08 18:15:10 crc kubenswrapper[4988]: I1008 18:15:10.468156 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-zmb2k"] Oct 08 18:15:10 crc kubenswrapper[4988]: I1008 18:15:10.470350 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-zmb2k" Oct 08 18:15:10 crc kubenswrapper[4988]: I1008 18:15:10.474502 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Oct 08 18:15:10 crc kubenswrapper[4988]: I1008 18:15:10.485207 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-zmb2k"] Oct 08 18:15:10 crc kubenswrapper[4988]: I1008 18:15:10.627053 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/83e5c288-f6e0-4ce6-a888-c7adbdb3e7e1-utilities\") pod \"community-operators-zmb2k\" (UID: \"83e5c288-f6e0-4ce6-a888-c7adbdb3e7e1\") " pod="openshift-marketplace/community-operators-zmb2k" Oct 08 18:15:10 crc kubenswrapper[4988]: I1008 18:15:10.627118 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/83e5c288-f6e0-4ce6-a888-c7adbdb3e7e1-catalog-content\") pod \"community-operators-zmb2k\" (UID: \"83e5c288-f6e0-4ce6-a888-c7adbdb3e7e1\") " pod="openshift-marketplace/community-operators-zmb2k" Oct 08 18:15:10 crc kubenswrapper[4988]: I1008 18:15:10.627155 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g6db2\" (UniqueName: \"kubernetes.io/projected/83e5c288-f6e0-4ce6-a888-c7adbdb3e7e1-kube-api-access-g6db2\") pod \"community-operators-zmb2k\" (UID: \"83e5c288-f6e0-4ce6-a888-c7adbdb3e7e1\") " pod="openshift-marketplace/community-operators-zmb2k" Oct 08 18:15:10 crc kubenswrapper[4988]: I1008 18:15:10.667587 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-c8g26"] Oct 08 18:15:10 crc kubenswrapper[4988]: I1008 18:15:10.669243 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-c8g26" Oct 08 18:15:10 crc kubenswrapper[4988]: I1008 18:15:10.678165 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Oct 08 18:15:10 crc kubenswrapper[4988]: I1008 18:15:10.683607 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-c8g26"] Oct 08 18:15:10 crc kubenswrapper[4988]: I1008 18:15:10.728634 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/83e5c288-f6e0-4ce6-a888-c7adbdb3e7e1-utilities\") pod \"community-operators-zmb2k\" (UID: \"83e5c288-f6e0-4ce6-a888-c7adbdb3e7e1\") " pod="openshift-marketplace/community-operators-zmb2k" Oct 08 18:15:10 crc kubenswrapper[4988]: I1008 18:15:10.728896 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/83e5c288-f6e0-4ce6-a888-c7adbdb3e7e1-catalog-content\") pod \"community-operators-zmb2k\" (UID: \"83e5c288-f6e0-4ce6-a888-c7adbdb3e7e1\") " pod="openshift-marketplace/community-operators-zmb2k" Oct 08 18:15:10 crc kubenswrapper[4988]: I1008 18:15:10.728920 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g6db2\" (UniqueName: \"kubernetes.io/projected/83e5c288-f6e0-4ce6-a888-c7adbdb3e7e1-kube-api-access-g6db2\") pod \"community-operators-zmb2k\" (UID: \"83e5c288-f6e0-4ce6-a888-c7adbdb3e7e1\") " pod="openshift-marketplace/community-operators-zmb2k" Oct 08 18:15:10 crc kubenswrapper[4988]: I1008 18:15:10.729135 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/83e5c288-f6e0-4ce6-a888-c7adbdb3e7e1-utilities\") pod \"community-operators-zmb2k\" (UID: \"83e5c288-f6e0-4ce6-a888-c7adbdb3e7e1\") " pod="openshift-marketplace/community-operators-zmb2k" Oct 08 18:15:10 crc kubenswrapper[4988]: I1008 18:15:10.729353 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/83e5c288-f6e0-4ce6-a888-c7adbdb3e7e1-catalog-content\") pod \"community-operators-zmb2k\" (UID: \"83e5c288-f6e0-4ce6-a888-c7adbdb3e7e1\") " pod="openshift-marketplace/community-operators-zmb2k" Oct 08 18:15:10 crc kubenswrapper[4988]: I1008 18:15:10.753628 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g6db2\" (UniqueName: \"kubernetes.io/projected/83e5c288-f6e0-4ce6-a888-c7adbdb3e7e1-kube-api-access-g6db2\") pod \"community-operators-zmb2k\" (UID: \"83e5c288-f6e0-4ce6-a888-c7adbdb3e7e1\") " pod="openshift-marketplace/community-operators-zmb2k" Oct 08 18:15:10 crc kubenswrapper[4988]: I1008 18:15:10.793051 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-zmb2k" Oct 08 18:15:10 crc kubenswrapper[4988]: I1008 18:15:10.830574 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7fnzv\" (UniqueName: \"kubernetes.io/projected/6d4ae11d-cd07-4151-ac5d-70100ef241ec-kube-api-access-7fnzv\") pod \"certified-operators-c8g26\" (UID: \"6d4ae11d-cd07-4151-ac5d-70100ef241ec\") " pod="openshift-marketplace/certified-operators-c8g26" Oct 08 18:15:10 crc kubenswrapper[4988]: I1008 18:15:10.830632 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6d4ae11d-cd07-4151-ac5d-70100ef241ec-utilities\") pod \"certified-operators-c8g26\" (UID: \"6d4ae11d-cd07-4151-ac5d-70100ef241ec\") " pod="openshift-marketplace/certified-operators-c8g26" Oct 08 18:15:10 crc kubenswrapper[4988]: I1008 18:15:10.830687 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6d4ae11d-cd07-4151-ac5d-70100ef241ec-catalog-content\") pod \"certified-operators-c8g26\" (UID: \"6d4ae11d-cd07-4151-ac5d-70100ef241ec\") " pod="openshift-marketplace/certified-operators-c8g26" Oct 08 18:15:10 crc kubenswrapper[4988]: I1008 18:15:10.931613 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7fnzv\" (UniqueName: \"kubernetes.io/projected/6d4ae11d-cd07-4151-ac5d-70100ef241ec-kube-api-access-7fnzv\") pod \"certified-operators-c8g26\" (UID: \"6d4ae11d-cd07-4151-ac5d-70100ef241ec\") " pod="openshift-marketplace/certified-operators-c8g26" Oct 08 18:15:10 crc kubenswrapper[4988]: I1008 18:15:10.932217 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6d4ae11d-cd07-4151-ac5d-70100ef241ec-utilities\") pod \"certified-operators-c8g26\" (UID: \"6d4ae11d-cd07-4151-ac5d-70100ef241ec\") " pod="openshift-marketplace/certified-operators-c8g26" Oct 08 18:15:10 crc kubenswrapper[4988]: I1008 18:15:10.932350 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6d4ae11d-cd07-4151-ac5d-70100ef241ec-catalog-content\") pod \"certified-operators-c8g26\" (UID: \"6d4ae11d-cd07-4151-ac5d-70100ef241ec\") " pod="openshift-marketplace/certified-operators-c8g26" Oct 08 18:15:10 crc kubenswrapper[4988]: I1008 18:15:10.932827 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6d4ae11d-cd07-4151-ac5d-70100ef241ec-utilities\") pod \"certified-operators-c8g26\" (UID: \"6d4ae11d-cd07-4151-ac5d-70100ef241ec\") " pod="openshift-marketplace/certified-operators-c8g26" Oct 08 18:15:10 crc kubenswrapper[4988]: I1008 18:15:10.933007 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6d4ae11d-cd07-4151-ac5d-70100ef241ec-catalog-content\") pod \"certified-operators-c8g26\" (UID: \"6d4ae11d-cd07-4151-ac5d-70100ef241ec\") " pod="openshift-marketplace/certified-operators-c8g26" Oct 08 18:15:10 crc kubenswrapper[4988]: I1008 18:15:10.963614 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7fnzv\" (UniqueName: \"kubernetes.io/projected/6d4ae11d-cd07-4151-ac5d-70100ef241ec-kube-api-access-7fnzv\") pod \"certified-operators-c8g26\" (UID: \"6d4ae11d-cd07-4151-ac5d-70100ef241ec\") " pod="openshift-marketplace/certified-operators-c8g26" Oct 08 18:15:11 crc kubenswrapper[4988]: I1008 18:15:11.001469 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-c8g26" Oct 08 18:15:11 crc kubenswrapper[4988]: I1008 18:15:11.014689 4988 generic.go:334] "Generic (PLEG): container finished" podID="6936c4e6-f11e-47a2-870a-0df3df4ed33c" containerID="6b7dd961cca0ded29adbb27754195774f61956ce2231c755e305d9c195a0c42d" exitCode=0 Oct 08 18:15:11 crc kubenswrapper[4988]: I1008 18:15:11.014790 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-gl4lh" event={"ID":"6936c4e6-f11e-47a2-870a-0df3df4ed33c","Type":"ContainerDied","Data":"6b7dd961cca0ded29adbb27754195774f61956ce2231c755e305d9c195a0c42d"} Oct 08 18:15:11 crc kubenswrapper[4988]: I1008 18:15:11.016246 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-zmb2k"] Oct 08 18:15:11 crc kubenswrapper[4988]: I1008 18:15:11.017866 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vwv6l" event={"ID":"4b10186d-3d69-4998-9162-89df44da5521","Type":"ContainerStarted","Data":"76310705f3485066e73e16fb999684e4fb4adc95b5b059127dcae7a532c5df81"} Oct 08 18:15:11 crc kubenswrapper[4988]: W1008 18:15:11.038095 4988 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod83e5c288_f6e0_4ce6_a888_c7adbdb3e7e1.slice/crio-285ced78d1e9d43cda13bdce696814d2d55b1da1bb9161a94738d489979d7d8d WatchSource:0}: Error finding container 285ced78d1e9d43cda13bdce696814d2d55b1da1bb9161a94738d489979d7d8d: Status 404 returned error can't find the container with id 285ced78d1e9d43cda13bdce696814d2d55b1da1bb9161a94738d489979d7d8d Oct 08 18:15:11 crc kubenswrapper[4988]: I1008 18:15:11.252953 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-c8g26"] Oct 08 18:15:11 crc kubenswrapper[4988]: W1008 18:15:11.328443 4988 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6d4ae11d_cd07_4151_ac5d_70100ef241ec.slice/crio-037d5acb6691418e1f3d665560d78f0db1ffb8de60c10516bacb5ef96b46a230 WatchSource:0}: Error finding container 037d5acb6691418e1f3d665560d78f0db1ffb8de60c10516bacb5ef96b46a230: Status 404 returned error can't find the container with id 037d5acb6691418e1f3d665560d78f0db1ffb8de60c10516bacb5ef96b46a230 Oct 08 18:15:12 crc kubenswrapper[4988]: I1008 18:15:12.043926 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-gl4lh" event={"ID":"6936c4e6-f11e-47a2-870a-0df3df4ed33c","Type":"ContainerStarted","Data":"e6aaf1e44521b67a3a9b367670d88ca76e5d5246c6ab91fc430f5646ac3951f2"} Oct 08 18:15:12 crc kubenswrapper[4988]: I1008 18:15:12.050558 4988 generic.go:334] "Generic (PLEG): container finished" podID="6d4ae11d-cd07-4151-ac5d-70100ef241ec" containerID="8626b25025233ea160859d26817cf1f326fd1a1fad847ef88028914bb97d1fbc" exitCode=0 Oct 08 18:15:12 crc kubenswrapper[4988]: I1008 18:15:12.050672 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-c8g26" event={"ID":"6d4ae11d-cd07-4151-ac5d-70100ef241ec","Type":"ContainerDied","Data":"8626b25025233ea160859d26817cf1f326fd1a1fad847ef88028914bb97d1fbc"} Oct 08 18:15:12 crc kubenswrapper[4988]: I1008 18:15:12.050749 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-c8g26" event={"ID":"6d4ae11d-cd07-4151-ac5d-70100ef241ec","Type":"ContainerStarted","Data":"037d5acb6691418e1f3d665560d78f0db1ffb8de60c10516bacb5ef96b46a230"} Oct 08 18:15:12 crc kubenswrapper[4988]: I1008 18:15:12.054694 4988 generic.go:334] "Generic (PLEG): container finished" podID="83e5c288-f6e0-4ce6-a888-c7adbdb3e7e1" containerID="f483fbe3b4e1253bd0cafb496fb26cf08a4e61d46d73d7bb7346fa370a4710d8" exitCode=0 Oct 08 18:15:12 crc kubenswrapper[4988]: I1008 18:15:12.054980 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zmb2k" event={"ID":"83e5c288-f6e0-4ce6-a888-c7adbdb3e7e1","Type":"ContainerDied","Data":"f483fbe3b4e1253bd0cafb496fb26cf08a4e61d46d73d7bb7346fa370a4710d8"} Oct 08 18:15:12 crc kubenswrapper[4988]: I1008 18:15:12.055037 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zmb2k" event={"ID":"83e5c288-f6e0-4ce6-a888-c7adbdb3e7e1","Type":"ContainerStarted","Data":"285ced78d1e9d43cda13bdce696814d2d55b1da1bb9161a94738d489979d7d8d"} Oct 08 18:15:12 crc kubenswrapper[4988]: I1008 18:15:12.056927 4988 generic.go:334] "Generic (PLEG): container finished" podID="4b10186d-3d69-4998-9162-89df44da5521" containerID="76310705f3485066e73e16fb999684e4fb4adc95b5b059127dcae7a532c5df81" exitCode=0 Oct 08 18:15:12 crc kubenswrapper[4988]: I1008 18:15:12.056951 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vwv6l" event={"ID":"4b10186d-3d69-4998-9162-89df44da5521","Type":"ContainerDied","Data":"76310705f3485066e73e16fb999684e4fb4adc95b5b059127dcae7a532c5df81"} Oct 08 18:15:12 crc kubenswrapper[4988]: I1008 18:15:12.071792 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-gl4lh" podStartSLOduration=1.578012097 podStartE2EDuration="4.071773855s" podCreationTimestamp="2025-10-08 18:15:08 +0000 UTC" firstStartedPulling="2025-10-08 18:15:08.993427021 +0000 UTC m=+254.443269791" lastFinishedPulling="2025-10-08 18:15:11.487188779 +0000 UTC m=+256.937031549" observedRunningTime="2025-10-08 18:15:12.070115592 +0000 UTC m=+257.519958362" watchObservedRunningTime="2025-10-08 18:15:12.071773855 +0000 UTC m=+257.521616615" Oct 08 18:15:13 crc kubenswrapper[4988]: I1008 18:15:13.069380 4988 generic.go:334] "Generic (PLEG): container finished" podID="83e5c288-f6e0-4ce6-a888-c7adbdb3e7e1" containerID="8e301b6d8f5852d31eda919618cfe13d22b037994cc589817faee8c42fb24519" exitCode=0 Oct 08 18:15:13 crc kubenswrapper[4988]: I1008 18:15:13.069516 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zmb2k" event={"ID":"83e5c288-f6e0-4ce6-a888-c7adbdb3e7e1","Type":"ContainerDied","Data":"8e301b6d8f5852d31eda919618cfe13d22b037994cc589817faee8c42fb24519"} Oct 08 18:15:13 crc kubenswrapper[4988]: I1008 18:15:13.087076 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vwv6l" event={"ID":"4b10186d-3d69-4998-9162-89df44da5521","Type":"ContainerStarted","Data":"5cd42894d0e8ee8f0c6230add8f9b4513ca9f2fa7caaa2bafb2de87a92d823ea"} Oct 08 18:15:13 crc kubenswrapper[4988]: I1008 18:15:13.107827 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-vwv6l" podStartSLOduration=2.603449183 podStartE2EDuration="5.107807683s" podCreationTimestamp="2025-10-08 18:15:08 +0000 UTC" firstStartedPulling="2025-10-08 18:15:10.008684269 +0000 UTC m=+255.458527049" lastFinishedPulling="2025-10-08 18:15:12.513042779 +0000 UTC m=+257.962885549" observedRunningTime="2025-10-08 18:15:13.107019438 +0000 UTC m=+258.556862208" watchObservedRunningTime="2025-10-08 18:15:13.107807683 +0000 UTC m=+258.557650453" Oct 08 18:15:15 crc kubenswrapper[4988]: I1008 18:15:15.099960 4988 generic.go:334] "Generic (PLEG): container finished" podID="6d4ae11d-cd07-4151-ac5d-70100ef241ec" containerID="8263c18b0f98554d6a548fb0cb0a999619832065b17969d0a11620fcd12222af" exitCode=0 Oct 08 18:15:15 crc kubenswrapper[4988]: I1008 18:15:15.100410 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-c8g26" event={"ID":"6d4ae11d-cd07-4151-ac5d-70100ef241ec","Type":"ContainerDied","Data":"8263c18b0f98554d6a548fb0cb0a999619832065b17969d0a11620fcd12222af"} Oct 08 18:15:15 crc kubenswrapper[4988]: I1008 18:15:15.107768 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zmb2k" event={"ID":"83e5c288-f6e0-4ce6-a888-c7adbdb3e7e1","Type":"ContainerStarted","Data":"a860c3ddf4622c3d16089c37ff60bea0f404a66fa5f108d923b282d43d8d93e9"} Oct 08 18:15:15 crc kubenswrapper[4988]: I1008 18:15:15.141079 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-zmb2k" podStartSLOduration=3.611187025 podStartE2EDuration="5.141051988s" podCreationTimestamp="2025-10-08 18:15:10 +0000 UTC" firstStartedPulling="2025-10-08 18:15:12.056451482 +0000 UTC m=+257.506294252" lastFinishedPulling="2025-10-08 18:15:13.586316425 +0000 UTC m=+259.036159215" observedRunningTime="2025-10-08 18:15:15.14048882 +0000 UTC m=+260.590331590" watchObservedRunningTime="2025-10-08 18:15:15.141051988 +0000 UTC m=+260.590894758" Oct 08 18:15:17 crc kubenswrapper[4988]: I1008 18:15:17.123136 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-c8g26" event={"ID":"6d4ae11d-cd07-4151-ac5d-70100ef241ec","Type":"ContainerStarted","Data":"d6557063485758b9991c491d4b9eae9dac650d15dc25dd3baaf86fc4432f3164"} Oct 08 18:15:17 crc kubenswrapper[4988]: I1008 18:15:17.141979 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-c8g26" podStartSLOduration=3.369601913 podStartE2EDuration="7.141956637s" podCreationTimestamp="2025-10-08 18:15:10 +0000 UTC" firstStartedPulling="2025-10-08 18:15:12.0523441 +0000 UTC m=+257.502186870" lastFinishedPulling="2025-10-08 18:15:15.824698824 +0000 UTC m=+261.274541594" observedRunningTime="2025-10-08 18:15:17.140782593 +0000 UTC m=+262.590625363" watchObservedRunningTime="2025-10-08 18:15:17.141956637 +0000 UTC m=+262.591799407" Oct 08 18:15:18 crc kubenswrapper[4988]: I1008 18:15:18.384027 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-gl4lh" Oct 08 18:15:18 crc kubenswrapper[4988]: I1008 18:15:18.384367 4988 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-gl4lh" Oct 08 18:15:18 crc kubenswrapper[4988]: I1008 18:15:18.421125 4988 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-gl4lh" Oct 08 18:15:18 crc kubenswrapper[4988]: I1008 18:15:18.593435 4988 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-vwv6l" Oct 08 18:15:18 crc kubenswrapper[4988]: I1008 18:15:18.593529 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-vwv6l" Oct 08 18:15:18 crc kubenswrapper[4988]: I1008 18:15:18.656204 4988 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-vwv6l" Oct 08 18:15:19 crc kubenswrapper[4988]: I1008 18:15:19.178346 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-gl4lh" Oct 08 18:15:19 crc kubenswrapper[4988]: I1008 18:15:19.182845 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-vwv6l" Oct 08 18:15:20 crc kubenswrapper[4988]: I1008 18:15:20.793728 4988 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-zmb2k" Oct 08 18:15:20 crc kubenswrapper[4988]: I1008 18:15:20.794358 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-zmb2k" Oct 08 18:15:20 crc kubenswrapper[4988]: I1008 18:15:20.838180 4988 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-zmb2k" Oct 08 18:15:21 crc kubenswrapper[4988]: I1008 18:15:21.001726 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-c8g26" Oct 08 18:15:21 crc kubenswrapper[4988]: I1008 18:15:21.001777 4988 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-c8g26" Oct 08 18:15:21 crc kubenswrapper[4988]: I1008 18:15:21.060982 4988 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-c8g26" Oct 08 18:15:21 crc kubenswrapper[4988]: I1008 18:15:21.183745 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-c8g26" Oct 08 18:15:21 crc kubenswrapper[4988]: I1008 18:15:21.203125 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-zmb2k" Oct 08 18:16:53 crc kubenswrapper[4988]: I1008 18:16:53.338486 4988 patch_prober.go:28] interesting pod/machine-config-daemon-nm54f container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 18:16:53 crc kubenswrapper[4988]: I1008 18:16:53.339338 4988 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 18:17:23 crc kubenswrapper[4988]: I1008 18:17:23.338671 4988 patch_prober.go:28] interesting pod/machine-config-daemon-nm54f container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 18:17:23 crc kubenswrapper[4988]: I1008 18:17:23.339566 4988 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 18:17:47 crc kubenswrapper[4988]: I1008 18:17:47.403251 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-29mh9"] Oct 08 18:17:47 crc kubenswrapper[4988]: I1008 18:17:47.404682 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-29mh9" Oct 08 18:17:47 crc kubenswrapper[4988]: I1008 18:17:47.412790 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-29mh9"] Oct 08 18:17:47 crc kubenswrapper[4988]: I1008 18:17:47.525831 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/c0b0de88-4cc3-4e45-9c58-a75367365ff4-bound-sa-token\") pod \"image-registry-66df7c8f76-29mh9\" (UID: \"c0b0de88-4cc3-4e45-9c58-a75367365ff4\") " pod="openshift-image-registry/image-registry-66df7c8f76-29mh9" Oct 08 18:17:47 crc kubenswrapper[4988]: I1008 18:17:47.525951 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/c0b0de88-4cc3-4e45-9c58-a75367365ff4-ca-trust-extracted\") pod \"image-registry-66df7c8f76-29mh9\" (UID: \"c0b0de88-4cc3-4e45-9c58-a75367365ff4\") " pod="openshift-image-registry/image-registry-66df7c8f76-29mh9" Oct 08 18:17:47 crc kubenswrapper[4988]: I1008 18:17:47.526007 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/c0b0de88-4cc3-4e45-9c58-a75367365ff4-registry-certificates\") pod \"image-registry-66df7c8f76-29mh9\" (UID: \"c0b0de88-4cc3-4e45-9c58-a75367365ff4\") " pod="openshift-image-registry/image-registry-66df7c8f76-29mh9" Oct 08 18:17:47 crc kubenswrapper[4988]: I1008 18:17:47.526150 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/c0b0de88-4cc3-4e45-9c58-a75367365ff4-registry-tls\") pod \"image-registry-66df7c8f76-29mh9\" (UID: \"c0b0de88-4cc3-4e45-9c58-a75367365ff4\") " pod="openshift-image-registry/image-registry-66df7c8f76-29mh9" Oct 08 18:17:47 crc kubenswrapper[4988]: I1008 18:17:47.526209 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-29mh9\" (UID: \"c0b0de88-4cc3-4e45-9c58-a75367365ff4\") " pod="openshift-image-registry/image-registry-66df7c8f76-29mh9" Oct 08 18:17:47 crc kubenswrapper[4988]: I1008 18:17:47.526248 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/c0b0de88-4cc3-4e45-9c58-a75367365ff4-trusted-ca\") pod \"image-registry-66df7c8f76-29mh9\" (UID: \"c0b0de88-4cc3-4e45-9c58-a75367365ff4\") " pod="openshift-image-registry/image-registry-66df7c8f76-29mh9" Oct 08 18:17:47 crc kubenswrapper[4988]: I1008 18:17:47.526283 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8tzgf\" (UniqueName: \"kubernetes.io/projected/c0b0de88-4cc3-4e45-9c58-a75367365ff4-kube-api-access-8tzgf\") pod \"image-registry-66df7c8f76-29mh9\" (UID: \"c0b0de88-4cc3-4e45-9c58-a75367365ff4\") " pod="openshift-image-registry/image-registry-66df7c8f76-29mh9" Oct 08 18:17:47 crc kubenswrapper[4988]: I1008 18:17:47.526325 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/c0b0de88-4cc3-4e45-9c58-a75367365ff4-installation-pull-secrets\") pod \"image-registry-66df7c8f76-29mh9\" (UID: \"c0b0de88-4cc3-4e45-9c58-a75367365ff4\") " pod="openshift-image-registry/image-registry-66df7c8f76-29mh9" Oct 08 18:17:47 crc kubenswrapper[4988]: I1008 18:17:47.556774 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-29mh9\" (UID: \"c0b0de88-4cc3-4e45-9c58-a75367365ff4\") " pod="openshift-image-registry/image-registry-66df7c8f76-29mh9" Oct 08 18:17:47 crc kubenswrapper[4988]: I1008 18:17:47.627270 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/c0b0de88-4cc3-4e45-9c58-a75367365ff4-trusted-ca\") pod \"image-registry-66df7c8f76-29mh9\" (UID: \"c0b0de88-4cc3-4e45-9c58-a75367365ff4\") " pod="openshift-image-registry/image-registry-66df7c8f76-29mh9" Oct 08 18:17:47 crc kubenswrapper[4988]: I1008 18:17:47.627336 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8tzgf\" (UniqueName: \"kubernetes.io/projected/c0b0de88-4cc3-4e45-9c58-a75367365ff4-kube-api-access-8tzgf\") pod \"image-registry-66df7c8f76-29mh9\" (UID: \"c0b0de88-4cc3-4e45-9c58-a75367365ff4\") " pod="openshift-image-registry/image-registry-66df7c8f76-29mh9" Oct 08 18:17:47 crc kubenswrapper[4988]: I1008 18:17:47.627427 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/c0b0de88-4cc3-4e45-9c58-a75367365ff4-installation-pull-secrets\") pod \"image-registry-66df7c8f76-29mh9\" (UID: \"c0b0de88-4cc3-4e45-9c58-a75367365ff4\") " pod="openshift-image-registry/image-registry-66df7c8f76-29mh9" Oct 08 18:17:47 crc kubenswrapper[4988]: I1008 18:17:47.627502 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/c0b0de88-4cc3-4e45-9c58-a75367365ff4-bound-sa-token\") pod \"image-registry-66df7c8f76-29mh9\" (UID: \"c0b0de88-4cc3-4e45-9c58-a75367365ff4\") " pod="openshift-image-registry/image-registry-66df7c8f76-29mh9" Oct 08 18:17:47 crc kubenswrapper[4988]: I1008 18:17:47.627557 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/c0b0de88-4cc3-4e45-9c58-a75367365ff4-ca-trust-extracted\") pod \"image-registry-66df7c8f76-29mh9\" (UID: \"c0b0de88-4cc3-4e45-9c58-a75367365ff4\") " pod="openshift-image-registry/image-registry-66df7c8f76-29mh9" Oct 08 18:17:47 crc kubenswrapper[4988]: I1008 18:17:47.627604 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/c0b0de88-4cc3-4e45-9c58-a75367365ff4-registry-certificates\") pod \"image-registry-66df7c8f76-29mh9\" (UID: \"c0b0de88-4cc3-4e45-9c58-a75367365ff4\") " pod="openshift-image-registry/image-registry-66df7c8f76-29mh9" Oct 08 18:17:47 crc kubenswrapper[4988]: I1008 18:17:47.627692 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/c0b0de88-4cc3-4e45-9c58-a75367365ff4-registry-tls\") pod \"image-registry-66df7c8f76-29mh9\" (UID: \"c0b0de88-4cc3-4e45-9c58-a75367365ff4\") " pod="openshift-image-registry/image-registry-66df7c8f76-29mh9" Oct 08 18:17:47 crc kubenswrapper[4988]: I1008 18:17:47.628670 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/c0b0de88-4cc3-4e45-9c58-a75367365ff4-ca-trust-extracted\") pod \"image-registry-66df7c8f76-29mh9\" (UID: \"c0b0de88-4cc3-4e45-9c58-a75367365ff4\") " pod="openshift-image-registry/image-registry-66df7c8f76-29mh9" Oct 08 18:17:47 crc kubenswrapper[4988]: I1008 18:17:47.629340 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/c0b0de88-4cc3-4e45-9c58-a75367365ff4-trusted-ca\") pod \"image-registry-66df7c8f76-29mh9\" (UID: \"c0b0de88-4cc3-4e45-9c58-a75367365ff4\") " pod="openshift-image-registry/image-registry-66df7c8f76-29mh9" Oct 08 18:17:47 crc kubenswrapper[4988]: I1008 18:17:47.629430 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/c0b0de88-4cc3-4e45-9c58-a75367365ff4-registry-certificates\") pod \"image-registry-66df7c8f76-29mh9\" (UID: \"c0b0de88-4cc3-4e45-9c58-a75367365ff4\") " pod="openshift-image-registry/image-registry-66df7c8f76-29mh9" Oct 08 18:17:47 crc kubenswrapper[4988]: I1008 18:17:47.636604 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/c0b0de88-4cc3-4e45-9c58-a75367365ff4-registry-tls\") pod \"image-registry-66df7c8f76-29mh9\" (UID: \"c0b0de88-4cc3-4e45-9c58-a75367365ff4\") " pod="openshift-image-registry/image-registry-66df7c8f76-29mh9" Oct 08 18:17:47 crc kubenswrapper[4988]: I1008 18:17:47.645415 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8tzgf\" (UniqueName: \"kubernetes.io/projected/c0b0de88-4cc3-4e45-9c58-a75367365ff4-kube-api-access-8tzgf\") pod \"image-registry-66df7c8f76-29mh9\" (UID: \"c0b0de88-4cc3-4e45-9c58-a75367365ff4\") " pod="openshift-image-registry/image-registry-66df7c8f76-29mh9" Oct 08 18:17:47 crc kubenswrapper[4988]: I1008 18:17:47.647162 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/c0b0de88-4cc3-4e45-9c58-a75367365ff4-installation-pull-secrets\") pod \"image-registry-66df7c8f76-29mh9\" (UID: \"c0b0de88-4cc3-4e45-9c58-a75367365ff4\") " pod="openshift-image-registry/image-registry-66df7c8f76-29mh9" Oct 08 18:17:47 crc kubenswrapper[4988]: I1008 18:17:47.657048 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/c0b0de88-4cc3-4e45-9c58-a75367365ff4-bound-sa-token\") pod \"image-registry-66df7c8f76-29mh9\" (UID: \"c0b0de88-4cc3-4e45-9c58-a75367365ff4\") " pod="openshift-image-registry/image-registry-66df7c8f76-29mh9" Oct 08 18:17:47 crc kubenswrapper[4988]: I1008 18:17:47.724499 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-29mh9" Oct 08 18:17:48 crc kubenswrapper[4988]: I1008 18:17:48.150137 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-29mh9"] Oct 08 18:17:49 crc kubenswrapper[4988]: I1008 18:17:49.175331 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-29mh9" event={"ID":"c0b0de88-4cc3-4e45-9c58-a75367365ff4","Type":"ContainerStarted","Data":"ca3c0b56fb3dda1782a222a1080c4a2a0108744dd2ece9efe3dbc267e272ae21"} Oct 08 18:17:49 crc kubenswrapper[4988]: I1008 18:17:49.175745 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-29mh9" event={"ID":"c0b0de88-4cc3-4e45-9c58-a75367365ff4","Type":"ContainerStarted","Data":"991d819abf8152107d2cb2b68c9943741f103be8a1fb4c8611d6adaf2c88da06"} Oct 08 18:17:49 crc kubenswrapper[4988]: I1008 18:17:49.175781 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-66df7c8f76-29mh9" Oct 08 18:17:49 crc kubenswrapper[4988]: I1008 18:17:49.191469 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-66df7c8f76-29mh9" podStartSLOduration=2.19144948 podStartE2EDuration="2.19144948s" podCreationTimestamp="2025-10-08 18:17:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 18:17:49.190864988 +0000 UTC m=+414.640707788" watchObservedRunningTime="2025-10-08 18:17:49.19144948 +0000 UTC m=+414.641292260" Oct 08 18:17:53 crc kubenswrapper[4988]: I1008 18:17:53.338010 4988 patch_prober.go:28] interesting pod/machine-config-daemon-nm54f container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 18:17:53 crc kubenswrapper[4988]: I1008 18:17:53.340161 4988 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 18:17:53 crc kubenswrapper[4988]: I1008 18:17:53.340461 4988 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" Oct 08 18:17:53 crc kubenswrapper[4988]: I1008 18:17:53.341555 4988 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"6f23f485d033acafdd8130a121cb37cf846dfb3c004e9a0d5e65009a35c2c0b0"} pod="openshift-machine-config-operator/machine-config-daemon-nm54f" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 08 18:17:53 crc kubenswrapper[4988]: I1008 18:17:53.341845 4988 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" containerName="machine-config-daemon" containerID="cri-o://6f23f485d033acafdd8130a121cb37cf846dfb3c004e9a0d5e65009a35c2c0b0" gracePeriod=600 Oct 08 18:17:54 crc kubenswrapper[4988]: I1008 18:17:54.234338 4988 generic.go:334] "Generic (PLEG): container finished" podID="f328a20b-5283-407e-8a93-2ab2da599f18" containerID="6f23f485d033acafdd8130a121cb37cf846dfb3c004e9a0d5e65009a35c2c0b0" exitCode=0 Oct 08 18:17:54 crc kubenswrapper[4988]: I1008 18:17:54.234429 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" event={"ID":"f328a20b-5283-407e-8a93-2ab2da599f18","Type":"ContainerDied","Data":"6f23f485d033acafdd8130a121cb37cf846dfb3c004e9a0d5e65009a35c2c0b0"} Oct 08 18:17:54 crc kubenswrapper[4988]: I1008 18:17:54.234719 4988 scope.go:117] "RemoveContainer" containerID="3cdb2a757332bdcdc2a2f3362913fe8f06003b0597de20b72627f9badd5e30af" Oct 08 18:17:55 crc kubenswrapper[4988]: I1008 18:17:55.244918 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" event={"ID":"f328a20b-5283-407e-8a93-2ab2da599f18","Type":"ContainerStarted","Data":"3134be7a9e11eb7efeae4470b4af55e4a033a7a3717b6d3502d9e3bd3007b110"} Oct 08 18:18:07 crc kubenswrapper[4988]: I1008 18:18:07.732801 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-66df7c8f76-29mh9" Oct 08 18:18:07 crc kubenswrapper[4988]: I1008 18:18:07.818495 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-4wh9p"] Oct 08 18:18:32 crc kubenswrapper[4988]: I1008 18:18:32.864630 4988 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-image-registry/image-registry-697d97f7c8-4wh9p" podUID="8a33d7bd-26bd-409b-93f1-5f151ebca347" containerName="registry" containerID="cri-o://6506e3bcb57c2dd9a9720948dbce35710215bd409e5a7999e73d3ecb72f926a7" gracePeriod=30 Oct 08 18:18:33 crc kubenswrapper[4988]: I1008 18:18:33.218874 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-4wh9p" Oct 08 18:18:33 crc kubenswrapper[4988]: I1008 18:18:33.372667 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8a33d7bd-26bd-409b-93f1-5f151ebca347-registry-tls\") pod \"8a33d7bd-26bd-409b-93f1-5f151ebca347\" (UID: \"8a33d7bd-26bd-409b-93f1-5f151ebca347\") " Oct 08 18:18:33 crc kubenswrapper[4988]: I1008 18:18:33.372750 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8a33d7bd-26bd-409b-93f1-5f151ebca347-bound-sa-token\") pod \"8a33d7bd-26bd-409b-93f1-5f151ebca347\" (UID: \"8a33d7bd-26bd-409b-93f1-5f151ebca347\") " Oct 08 18:18:33 crc kubenswrapper[4988]: I1008 18:18:33.372801 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8a33d7bd-26bd-409b-93f1-5f151ebca347-ca-trust-extracted\") pod \"8a33d7bd-26bd-409b-93f1-5f151ebca347\" (UID: \"8a33d7bd-26bd-409b-93f1-5f151ebca347\") " Oct 08 18:18:33 crc kubenswrapper[4988]: I1008 18:18:33.373048 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-storage\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8a33d7bd-26bd-409b-93f1-5f151ebca347\" (UID: \"8a33d7bd-26bd-409b-93f1-5f151ebca347\") " Oct 08 18:18:33 crc kubenswrapper[4988]: I1008 18:18:33.373088 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8a33d7bd-26bd-409b-93f1-5f151ebca347-installation-pull-secrets\") pod \"8a33d7bd-26bd-409b-93f1-5f151ebca347\" (UID: \"8a33d7bd-26bd-409b-93f1-5f151ebca347\") " Oct 08 18:18:33 crc kubenswrapper[4988]: I1008 18:18:33.373119 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8a33d7bd-26bd-409b-93f1-5f151ebca347-registry-certificates\") pod \"8a33d7bd-26bd-409b-93f1-5f151ebca347\" (UID: \"8a33d7bd-26bd-409b-93f1-5f151ebca347\") " Oct 08 18:18:33 crc kubenswrapper[4988]: I1008 18:18:33.373143 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8a33d7bd-26bd-409b-93f1-5f151ebca347-trusted-ca\") pod \"8a33d7bd-26bd-409b-93f1-5f151ebca347\" (UID: \"8a33d7bd-26bd-409b-93f1-5f151ebca347\") " Oct 08 18:18:33 crc kubenswrapper[4988]: I1008 18:18:33.373158 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kxfl7\" (UniqueName: \"kubernetes.io/projected/8a33d7bd-26bd-409b-93f1-5f151ebca347-kube-api-access-kxfl7\") pod \"8a33d7bd-26bd-409b-93f1-5f151ebca347\" (UID: \"8a33d7bd-26bd-409b-93f1-5f151ebca347\") " Oct 08 18:18:33 crc kubenswrapper[4988]: I1008 18:18:33.374664 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8a33d7bd-26bd-409b-93f1-5f151ebca347-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "8a33d7bd-26bd-409b-93f1-5f151ebca347" (UID: "8a33d7bd-26bd-409b-93f1-5f151ebca347"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 18:18:33 crc kubenswrapper[4988]: I1008 18:18:33.374838 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8a33d7bd-26bd-409b-93f1-5f151ebca347-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "8a33d7bd-26bd-409b-93f1-5f151ebca347" (UID: "8a33d7bd-26bd-409b-93f1-5f151ebca347"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 18:18:33 crc kubenswrapper[4988]: I1008 18:18:33.380589 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8a33d7bd-26bd-409b-93f1-5f151ebca347-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "8a33d7bd-26bd-409b-93f1-5f151ebca347" (UID: "8a33d7bd-26bd-409b-93f1-5f151ebca347"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:18:33 crc kubenswrapper[4988]: I1008 18:18:33.382826 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8a33d7bd-26bd-409b-93f1-5f151ebca347-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "8a33d7bd-26bd-409b-93f1-5f151ebca347" (UID: "8a33d7bd-26bd-409b-93f1-5f151ebca347"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:18:33 crc kubenswrapper[4988]: I1008 18:18:33.386758 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8a33d7bd-26bd-409b-93f1-5f151ebca347-kube-api-access-kxfl7" (OuterVolumeSpecName: "kube-api-access-kxfl7") pod "8a33d7bd-26bd-409b-93f1-5f151ebca347" (UID: "8a33d7bd-26bd-409b-93f1-5f151ebca347"). InnerVolumeSpecName "kube-api-access-kxfl7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:18:33 crc kubenswrapper[4988]: I1008 18:18:33.387107 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8a33d7bd-26bd-409b-93f1-5f151ebca347-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "8a33d7bd-26bd-409b-93f1-5f151ebca347" (UID: "8a33d7bd-26bd-409b-93f1-5f151ebca347"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:18:33 crc kubenswrapper[4988]: I1008 18:18:33.387792 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "registry-storage") pod "8a33d7bd-26bd-409b-93f1-5f151ebca347" (UID: "8a33d7bd-26bd-409b-93f1-5f151ebca347"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Oct 08 18:18:33 crc kubenswrapper[4988]: I1008 18:18:33.408947 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8a33d7bd-26bd-409b-93f1-5f151ebca347-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "8a33d7bd-26bd-409b-93f1-5f151ebca347" (UID: "8a33d7bd-26bd-409b-93f1-5f151ebca347"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 18:18:33 crc kubenswrapper[4988]: I1008 18:18:33.475927 4988 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8a33d7bd-26bd-409b-93f1-5f151ebca347-registry-tls\") on node \"crc\" DevicePath \"\"" Oct 08 18:18:33 crc kubenswrapper[4988]: I1008 18:18:33.475991 4988 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8a33d7bd-26bd-409b-93f1-5f151ebca347-bound-sa-token\") on node \"crc\" DevicePath \"\"" Oct 08 18:18:33 crc kubenswrapper[4988]: I1008 18:18:33.476012 4988 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8a33d7bd-26bd-409b-93f1-5f151ebca347-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Oct 08 18:18:33 crc kubenswrapper[4988]: I1008 18:18:33.476033 4988 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8a33d7bd-26bd-409b-93f1-5f151ebca347-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Oct 08 18:18:33 crc kubenswrapper[4988]: I1008 18:18:33.476056 4988 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8a33d7bd-26bd-409b-93f1-5f151ebca347-registry-certificates\") on node \"crc\" DevicePath \"\"" Oct 08 18:18:33 crc kubenswrapper[4988]: I1008 18:18:33.476075 4988 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kxfl7\" (UniqueName: \"kubernetes.io/projected/8a33d7bd-26bd-409b-93f1-5f151ebca347-kube-api-access-kxfl7\") on node \"crc\" DevicePath \"\"" Oct 08 18:18:33 crc kubenswrapper[4988]: I1008 18:18:33.476094 4988 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8a33d7bd-26bd-409b-93f1-5f151ebca347-trusted-ca\") on node \"crc\" DevicePath \"\"" Oct 08 18:18:33 crc kubenswrapper[4988]: I1008 18:18:33.531875 4988 generic.go:334] "Generic (PLEG): container finished" podID="8a33d7bd-26bd-409b-93f1-5f151ebca347" containerID="6506e3bcb57c2dd9a9720948dbce35710215bd409e5a7999e73d3ecb72f926a7" exitCode=0 Oct 08 18:18:33 crc kubenswrapper[4988]: I1008 18:18:33.531937 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-4wh9p" event={"ID":"8a33d7bd-26bd-409b-93f1-5f151ebca347","Type":"ContainerDied","Data":"6506e3bcb57c2dd9a9720948dbce35710215bd409e5a7999e73d3ecb72f926a7"} Oct 08 18:18:33 crc kubenswrapper[4988]: I1008 18:18:33.531983 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-4wh9p" event={"ID":"8a33d7bd-26bd-409b-93f1-5f151ebca347","Type":"ContainerDied","Data":"1e0deb69d6e71818c0dc06067684a7c6e51c0f0ed7ef628282ba777c56ff7d00"} Oct 08 18:18:33 crc kubenswrapper[4988]: I1008 18:18:33.532010 4988 scope.go:117] "RemoveContainer" containerID="6506e3bcb57c2dd9a9720948dbce35710215bd409e5a7999e73d3ecb72f926a7" Oct 08 18:18:33 crc kubenswrapper[4988]: I1008 18:18:33.532035 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-4wh9p" Oct 08 18:18:33 crc kubenswrapper[4988]: I1008 18:18:33.554245 4988 scope.go:117] "RemoveContainer" containerID="6506e3bcb57c2dd9a9720948dbce35710215bd409e5a7999e73d3ecb72f926a7" Oct 08 18:18:33 crc kubenswrapper[4988]: E1008 18:18:33.555013 4988 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6506e3bcb57c2dd9a9720948dbce35710215bd409e5a7999e73d3ecb72f926a7\": container with ID starting with 6506e3bcb57c2dd9a9720948dbce35710215bd409e5a7999e73d3ecb72f926a7 not found: ID does not exist" containerID="6506e3bcb57c2dd9a9720948dbce35710215bd409e5a7999e73d3ecb72f926a7" Oct 08 18:18:33 crc kubenswrapper[4988]: I1008 18:18:33.555074 4988 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6506e3bcb57c2dd9a9720948dbce35710215bd409e5a7999e73d3ecb72f926a7"} err="failed to get container status \"6506e3bcb57c2dd9a9720948dbce35710215bd409e5a7999e73d3ecb72f926a7\": rpc error: code = NotFound desc = could not find container \"6506e3bcb57c2dd9a9720948dbce35710215bd409e5a7999e73d3ecb72f926a7\": container with ID starting with 6506e3bcb57c2dd9a9720948dbce35710215bd409e5a7999e73d3ecb72f926a7 not found: ID does not exist" Oct 08 18:18:33 crc kubenswrapper[4988]: I1008 18:18:33.583124 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-4wh9p"] Oct 08 18:18:33 crc kubenswrapper[4988]: I1008 18:18:33.586267 4988 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-4wh9p"] Oct 08 18:18:35 crc kubenswrapper[4988]: I1008 18:18:35.249650 4988 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8a33d7bd-26bd-409b-93f1-5f151ebca347" path="/var/lib/kubelet/pods/8a33d7bd-26bd-409b-93f1-5f151ebca347/volumes" Oct 08 18:20:23 crc kubenswrapper[4988]: I1008 18:20:23.337956 4988 patch_prober.go:28] interesting pod/machine-config-daemon-nm54f container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 18:20:23 crc kubenswrapper[4988]: I1008 18:20:23.340509 4988 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 18:20:53 crc kubenswrapper[4988]: I1008 18:20:53.338252 4988 patch_prober.go:28] interesting pod/machine-config-daemon-nm54f container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 18:20:53 crc kubenswrapper[4988]: I1008 18:20:53.338893 4988 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 18:21:23 crc kubenswrapper[4988]: I1008 18:21:23.338122 4988 patch_prober.go:28] interesting pod/machine-config-daemon-nm54f container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 18:21:23 crc kubenswrapper[4988]: I1008 18:21:23.338729 4988 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 18:21:23 crc kubenswrapper[4988]: I1008 18:21:23.338778 4988 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" Oct 08 18:21:23 crc kubenswrapper[4988]: I1008 18:21:23.339340 4988 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"3134be7a9e11eb7efeae4470b4af55e4a033a7a3717b6d3502d9e3bd3007b110"} pod="openshift-machine-config-operator/machine-config-daemon-nm54f" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 08 18:21:23 crc kubenswrapper[4988]: I1008 18:21:23.339416 4988 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" containerName="machine-config-daemon" containerID="cri-o://3134be7a9e11eb7efeae4470b4af55e4a033a7a3717b6d3502d9e3bd3007b110" gracePeriod=600 Oct 08 18:21:23 crc kubenswrapper[4988]: I1008 18:21:23.678577 4988 generic.go:334] "Generic (PLEG): container finished" podID="f328a20b-5283-407e-8a93-2ab2da599f18" containerID="3134be7a9e11eb7efeae4470b4af55e4a033a7a3717b6d3502d9e3bd3007b110" exitCode=0 Oct 08 18:21:23 crc kubenswrapper[4988]: I1008 18:21:23.678652 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" event={"ID":"f328a20b-5283-407e-8a93-2ab2da599f18","Type":"ContainerDied","Data":"3134be7a9e11eb7efeae4470b4af55e4a033a7a3717b6d3502d9e3bd3007b110"} Oct 08 18:21:23 crc kubenswrapper[4988]: I1008 18:21:23.679332 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" event={"ID":"f328a20b-5283-407e-8a93-2ab2da599f18","Type":"ContainerStarted","Data":"0b7eb1da8fbdc295b99305bf18316c88ee5b435b3bee717d217e3efa0977c5a9"} Oct 08 18:21:23 crc kubenswrapper[4988]: I1008 18:21:23.679424 4988 scope.go:117] "RemoveContainer" containerID="6f23f485d033acafdd8130a121cb37cf846dfb3c004e9a0d5e65009a35c2c0b0" Oct 08 18:21:50 crc kubenswrapper[4988]: I1008 18:21:50.363939 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-nsc65"] Oct 08 18:21:50 crc kubenswrapper[4988]: I1008 18:21:50.365215 4988 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-nsc65" podUID="f010a463-69f2-4082-acc3-161899a939c3" containerName="ovn-controller" containerID="cri-o://d314eb248e90371fed068b1173e26d4c9f9765417a70b341e1a5b136ec17353a" gracePeriod=30 Oct 08 18:21:50 crc kubenswrapper[4988]: I1008 18:21:50.365276 4988 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-nsc65" podUID="f010a463-69f2-4082-acc3-161899a939c3" containerName="nbdb" containerID="cri-o://9fbbeae0409d2e47e8de9a4306f63b0d13acbfd73cd86f83303fec14352ed718" gracePeriod=30 Oct 08 18:21:50 crc kubenswrapper[4988]: I1008 18:21:50.365364 4988 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-nsc65" podUID="f010a463-69f2-4082-acc3-161899a939c3" containerName="sbdb" containerID="cri-o://74fcb812b84291763cc6e5ccb20c12aaad580ad968dda4734a0bbfe33d2770ca" gracePeriod=30 Oct 08 18:21:50 crc kubenswrapper[4988]: I1008 18:21:50.365346 4988 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-nsc65" podUID="f010a463-69f2-4082-acc3-161899a939c3" containerName="kube-rbac-proxy-ovn-metrics" containerID="cri-o://03406e726f817bcedb841f81959ab12808ba35a530240f9abc5fa5c608b9a6ae" gracePeriod=30 Oct 08 18:21:50 crc kubenswrapper[4988]: I1008 18:21:50.365457 4988 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-nsc65" podUID="f010a463-69f2-4082-acc3-161899a939c3" containerName="ovn-acl-logging" containerID="cri-o://cee5de93bc68345f9b737857f02832f56582d58757c50a8798eec25d19b2c20d" gracePeriod=30 Oct 08 18:21:50 crc kubenswrapper[4988]: I1008 18:21:50.365435 4988 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-nsc65" podUID="f010a463-69f2-4082-acc3-161899a939c3" containerName="kube-rbac-proxy-node" containerID="cri-o://48a516f6c153743bd1a21ec6945fb368a37f4d2b9d6a326b68edc096a192ed05" gracePeriod=30 Oct 08 18:21:50 crc kubenswrapper[4988]: I1008 18:21:50.365460 4988 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-nsc65" podUID="f010a463-69f2-4082-acc3-161899a939c3" containerName="northd" containerID="cri-o://55791327a370c0009c523b86b2234e82d3794985f2eaccf1922eb90f46b2b970" gracePeriod=30 Oct 08 18:21:50 crc kubenswrapper[4988]: I1008 18:21:50.431180 4988 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-nsc65" podUID="f010a463-69f2-4082-acc3-161899a939c3" containerName="ovnkube-controller" containerID="cri-o://4987245ff838ca8d4b8c31a08147c052d23bdad26126fcb46367d4b5d4197fb2" gracePeriod=30 Oct 08 18:21:50 crc kubenswrapper[4988]: I1008 18:21:50.722529 4988 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-nsc65_f010a463-69f2-4082-acc3-161899a939c3/ovnkube-controller/3.log" Oct 08 18:21:50 crc kubenswrapper[4988]: I1008 18:21:50.724403 4988 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-nsc65_f010a463-69f2-4082-acc3-161899a939c3/ovn-acl-logging/0.log" Oct 08 18:21:50 crc kubenswrapper[4988]: I1008 18:21:50.724858 4988 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-nsc65_f010a463-69f2-4082-acc3-161899a939c3/ovn-controller/0.log" Oct 08 18:21:50 crc kubenswrapper[4988]: I1008 18:21:50.725209 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-nsc65" Oct 08 18:21:50 crc kubenswrapper[4988]: I1008 18:21:50.755535 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/f010a463-69f2-4082-acc3-161899a939c3-etc-openvswitch\") pod \"f010a463-69f2-4082-acc3-161899a939c3\" (UID: \"f010a463-69f2-4082-acc3-161899a939c3\") " Oct 08 18:21:50 crc kubenswrapper[4988]: I1008 18:21:50.755607 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/f010a463-69f2-4082-acc3-161899a939c3-host-run-ovn-kubernetes\") pod \"f010a463-69f2-4082-acc3-161899a939c3\" (UID: \"f010a463-69f2-4082-acc3-161899a939c3\") " Oct 08 18:21:50 crc kubenswrapper[4988]: I1008 18:21:50.755628 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/f010a463-69f2-4082-acc3-161899a939c3-host-cni-netd\") pod \"f010a463-69f2-4082-acc3-161899a939c3\" (UID: \"f010a463-69f2-4082-acc3-161899a939c3\") " Oct 08 18:21:50 crc kubenswrapper[4988]: I1008 18:21:50.755644 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/f010a463-69f2-4082-acc3-161899a939c3-run-ovn\") pod \"f010a463-69f2-4082-acc3-161899a939c3\" (UID: \"f010a463-69f2-4082-acc3-161899a939c3\") " Oct 08 18:21:50 crc kubenswrapper[4988]: I1008 18:21:50.755667 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/f010a463-69f2-4082-acc3-161899a939c3-run-systemd\") pod \"f010a463-69f2-4082-acc3-161899a939c3\" (UID: \"f010a463-69f2-4082-acc3-161899a939c3\") " Oct 08 18:21:50 crc kubenswrapper[4988]: I1008 18:21:50.755687 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/f010a463-69f2-4082-acc3-161899a939c3-env-overrides\") pod \"f010a463-69f2-4082-acc3-161899a939c3\" (UID: \"f010a463-69f2-4082-acc3-161899a939c3\") " Oct 08 18:21:50 crc kubenswrapper[4988]: I1008 18:21:50.755707 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/f010a463-69f2-4082-acc3-161899a939c3-host-cni-bin\") pod \"f010a463-69f2-4082-acc3-161899a939c3\" (UID: \"f010a463-69f2-4082-acc3-161899a939c3\") " Oct 08 18:21:50 crc kubenswrapper[4988]: I1008 18:21:50.755706 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f010a463-69f2-4082-acc3-161899a939c3-host-run-ovn-kubernetes" (OuterVolumeSpecName: "host-run-ovn-kubernetes") pod "f010a463-69f2-4082-acc3-161899a939c3" (UID: "f010a463-69f2-4082-acc3-161899a939c3"). InnerVolumeSpecName "host-run-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 08 18:21:50 crc kubenswrapper[4988]: I1008 18:21:50.755726 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/f010a463-69f2-4082-acc3-161899a939c3-systemd-units\") pod \"f010a463-69f2-4082-acc3-161899a939c3\" (UID: \"f010a463-69f2-4082-acc3-161899a939c3\") " Oct 08 18:21:50 crc kubenswrapper[4988]: I1008 18:21:50.755718 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f010a463-69f2-4082-acc3-161899a939c3-etc-openvswitch" (OuterVolumeSpecName: "etc-openvswitch") pod "f010a463-69f2-4082-acc3-161899a939c3" (UID: "f010a463-69f2-4082-acc3-161899a939c3"). InnerVolumeSpecName "etc-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 08 18:21:50 crc kubenswrapper[4988]: I1008 18:21:50.755743 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f010a463-69f2-4082-acc3-161899a939c3-host-cni-netd" (OuterVolumeSpecName: "host-cni-netd") pod "f010a463-69f2-4082-acc3-161899a939c3" (UID: "f010a463-69f2-4082-acc3-161899a939c3"). InnerVolumeSpecName "host-cni-netd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 08 18:21:50 crc kubenswrapper[4988]: I1008 18:21:50.755743 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dnwgr\" (UniqueName: \"kubernetes.io/projected/f010a463-69f2-4082-acc3-161899a939c3-kube-api-access-dnwgr\") pod \"f010a463-69f2-4082-acc3-161899a939c3\" (UID: \"f010a463-69f2-4082-acc3-161899a939c3\") " Oct 08 18:21:50 crc kubenswrapper[4988]: I1008 18:21:50.755795 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f010a463-69f2-4082-acc3-161899a939c3-host-cni-bin" (OuterVolumeSpecName: "host-cni-bin") pod "f010a463-69f2-4082-acc3-161899a939c3" (UID: "f010a463-69f2-4082-acc3-161899a939c3"). InnerVolumeSpecName "host-cni-bin". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 08 18:21:50 crc kubenswrapper[4988]: I1008 18:21:50.755836 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/f010a463-69f2-4082-acc3-161899a939c3-log-socket\") pod \"f010a463-69f2-4082-acc3-161899a939c3\" (UID: \"f010a463-69f2-4082-acc3-161899a939c3\") " Oct 08 18:21:50 crc kubenswrapper[4988]: I1008 18:21:50.755863 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/f010a463-69f2-4082-acc3-161899a939c3-var-lib-openvswitch\") pod \"f010a463-69f2-4082-acc3-161899a939c3\" (UID: \"f010a463-69f2-4082-acc3-161899a939c3\") " Oct 08 18:21:50 crc kubenswrapper[4988]: I1008 18:21:50.755888 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/f010a463-69f2-4082-acc3-161899a939c3-node-log\") pod \"f010a463-69f2-4082-acc3-161899a939c3\" (UID: \"f010a463-69f2-4082-acc3-161899a939c3\") " Oct 08 18:21:50 crc kubenswrapper[4988]: I1008 18:21:50.755912 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/f010a463-69f2-4082-acc3-161899a939c3-ovn-node-metrics-cert\") pod \"f010a463-69f2-4082-acc3-161899a939c3\" (UID: \"f010a463-69f2-4082-acc3-161899a939c3\") " Oct 08 18:21:50 crc kubenswrapper[4988]: I1008 18:21:50.755934 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/f010a463-69f2-4082-acc3-161899a939c3-ovnkube-config\") pod \"f010a463-69f2-4082-acc3-161899a939c3\" (UID: \"f010a463-69f2-4082-acc3-161899a939c3\") " Oct 08 18:21:50 crc kubenswrapper[4988]: I1008 18:21:50.755949 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/f010a463-69f2-4082-acc3-161899a939c3-run-openvswitch\") pod \"f010a463-69f2-4082-acc3-161899a939c3\" (UID: \"f010a463-69f2-4082-acc3-161899a939c3\") " Oct 08 18:21:50 crc kubenswrapper[4988]: I1008 18:21:50.755970 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/f010a463-69f2-4082-acc3-161899a939c3-ovnkube-script-lib\") pod \"f010a463-69f2-4082-acc3-161899a939c3\" (UID: \"f010a463-69f2-4082-acc3-161899a939c3\") " Oct 08 18:21:50 crc kubenswrapper[4988]: I1008 18:21:50.755994 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/f010a463-69f2-4082-acc3-161899a939c3-host-run-netns\") pod \"f010a463-69f2-4082-acc3-161899a939c3\" (UID: \"f010a463-69f2-4082-acc3-161899a939c3\") " Oct 08 18:21:50 crc kubenswrapper[4988]: I1008 18:21:50.756021 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/f010a463-69f2-4082-acc3-161899a939c3-host-kubelet\") pod \"f010a463-69f2-4082-acc3-161899a939c3\" (UID: \"f010a463-69f2-4082-acc3-161899a939c3\") " Oct 08 18:21:50 crc kubenswrapper[4988]: I1008 18:21:50.756044 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/f010a463-69f2-4082-acc3-161899a939c3-host-var-lib-cni-networks-ovn-kubernetes\") pod \"f010a463-69f2-4082-acc3-161899a939c3\" (UID: \"f010a463-69f2-4082-acc3-161899a939c3\") " Oct 08 18:21:50 crc kubenswrapper[4988]: I1008 18:21:50.756066 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/f010a463-69f2-4082-acc3-161899a939c3-host-slash\") pod \"f010a463-69f2-4082-acc3-161899a939c3\" (UID: \"f010a463-69f2-4082-acc3-161899a939c3\") " Oct 08 18:21:50 crc kubenswrapper[4988]: I1008 18:21:50.756294 4988 reconciler_common.go:293] "Volume detached for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/f010a463-69f2-4082-acc3-161899a939c3-host-cni-bin\") on node \"crc\" DevicePath \"\"" Oct 08 18:21:50 crc kubenswrapper[4988]: I1008 18:21:50.756309 4988 reconciler_common.go:293] "Volume detached for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/f010a463-69f2-4082-acc3-161899a939c3-etc-openvswitch\") on node \"crc\" DevicePath \"\"" Oct 08 18:21:50 crc kubenswrapper[4988]: I1008 18:21:50.756308 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f010a463-69f2-4082-acc3-161899a939c3-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "f010a463-69f2-4082-acc3-161899a939c3" (UID: "f010a463-69f2-4082-acc3-161899a939c3"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 18:21:50 crc kubenswrapper[4988]: I1008 18:21:50.756321 4988 reconciler_common.go:293] "Volume detached for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/f010a463-69f2-4082-acc3-161899a939c3-host-run-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Oct 08 18:21:50 crc kubenswrapper[4988]: I1008 18:21:50.756333 4988 reconciler_common.go:293] "Volume detached for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/f010a463-69f2-4082-acc3-161899a939c3-host-cni-netd\") on node \"crc\" DevicePath \"\"" Oct 08 18:21:50 crc kubenswrapper[4988]: I1008 18:21:50.756352 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f010a463-69f2-4082-acc3-161899a939c3-systemd-units" (OuterVolumeSpecName: "systemd-units") pod "f010a463-69f2-4082-acc3-161899a939c3" (UID: "f010a463-69f2-4082-acc3-161899a939c3"). InnerVolumeSpecName "systemd-units". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 08 18:21:50 crc kubenswrapper[4988]: I1008 18:21:50.756373 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f010a463-69f2-4082-acc3-161899a939c3-host-slash" (OuterVolumeSpecName: "host-slash") pod "f010a463-69f2-4082-acc3-161899a939c3" (UID: "f010a463-69f2-4082-acc3-161899a939c3"). InnerVolumeSpecName "host-slash". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 08 18:21:50 crc kubenswrapper[4988]: I1008 18:21:50.756400 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f010a463-69f2-4082-acc3-161899a939c3-run-ovn" (OuterVolumeSpecName: "run-ovn") pod "f010a463-69f2-4082-acc3-161899a939c3" (UID: "f010a463-69f2-4082-acc3-161899a939c3"). InnerVolumeSpecName "run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 08 18:21:50 crc kubenswrapper[4988]: I1008 18:21:50.756420 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f010a463-69f2-4082-acc3-161899a939c3-log-socket" (OuterVolumeSpecName: "log-socket") pod "f010a463-69f2-4082-acc3-161899a939c3" (UID: "f010a463-69f2-4082-acc3-161899a939c3"). InnerVolumeSpecName "log-socket". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 08 18:21:50 crc kubenswrapper[4988]: I1008 18:21:50.756434 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f010a463-69f2-4082-acc3-161899a939c3-host-run-netns" (OuterVolumeSpecName: "host-run-netns") pod "f010a463-69f2-4082-acc3-161899a939c3" (UID: "f010a463-69f2-4082-acc3-161899a939c3"). InnerVolumeSpecName "host-run-netns". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 08 18:21:50 crc kubenswrapper[4988]: I1008 18:21:50.756447 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f010a463-69f2-4082-acc3-161899a939c3-host-kubelet" (OuterVolumeSpecName: "host-kubelet") pod "f010a463-69f2-4082-acc3-161899a939c3" (UID: "f010a463-69f2-4082-acc3-161899a939c3"). InnerVolumeSpecName "host-kubelet". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 08 18:21:50 crc kubenswrapper[4988]: I1008 18:21:50.756445 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f010a463-69f2-4082-acc3-161899a939c3-run-openvswitch" (OuterVolumeSpecName: "run-openvswitch") pod "f010a463-69f2-4082-acc3-161899a939c3" (UID: "f010a463-69f2-4082-acc3-161899a939c3"). InnerVolumeSpecName "run-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 08 18:21:50 crc kubenswrapper[4988]: I1008 18:21:50.756476 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f010a463-69f2-4082-acc3-161899a939c3-host-var-lib-cni-networks-ovn-kubernetes" (OuterVolumeSpecName: "host-var-lib-cni-networks-ovn-kubernetes") pod "f010a463-69f2-4082-acc3-161899a939c3" (UID: "f010a463-69f2-4082-acc3-161899a939c3"). InnerVolumeSpecName "host-var-lib-cni-networks-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 08 18:21:50 crc kubenswrapper[4988]: I1008 18:21:50.756490 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f010a463-69f2-4082-acc3-161899a939c3-var-lib-openvswitch" (OuterVolumeSpecName: "var-lib-openvswitch") pod "f010a463-69f2-4082-acc3-161899a939c3" (UID: "f010a463-69f2-4082-acc3-161899a939c3"). InnerVolumeSpecName "var-lib-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 08 18:21:50 crc kubenswrapper[4988]: I1008 18:21:50.756808 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f010a463-69f2-4082-acc3-161899a939c3-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "f010a463-69f2-4082-acc3-161899a939c3" (UID: "f010a463-69f2-4082-acc3-161899a939c3"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 18:21:50 crc kubenswrapper[4988]: I1008 18:21:50.756852 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f010a463-69f2-4082-acc3-161899a939c3-node-log" (OuterVolumeSpecName: "node-log") pod "f010a463-69f2-4082-acc3-161899a939c3" (UID: "f010a463-69f2-4082-acc3-161899a939c3"). InnerVolumeSpecName "node-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 08 18:21:50 crc kubenswrapper[4988]: I1008 18:21:50.756975 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f010a463-69f2-4082-acc3-161899a939c3-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "f010a463-69f2-4082-acc3-161899a939c3" (UID: "f010a463-69f2-4082-acc3-161899a939c3"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 18:21:50 crc kubenswrapper[4988]: I1008 18:21:50.766044 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f010a463-69f2-4082-acc3-161899a939c3-kube-api-access-dnwgr" (OuterVolumeSpecName: "kube-api-access-dnwgr") pod "f010a463-69f2-4082-acc3-161899a939c3" (UID: "f010a463-69f2-4082-acc3-161899a939c3"). InnerVolumeSpecName "kube-api-access-dnwgr". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:21:50 crc kubenswrapper[4988]: I1008 18:21:50.766232 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f010a463-69f2-4082-acc3-161899a939c3-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "f010a463-69f2-4082-acc3-161899a939c3" (UID: "f010a463-69f2-4082-acc3-161899a939c3"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:21:50 crc kubenswrapper[4988]: I1008 18:21:50.769169 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f010a463-69f2-4082-acc3-161899a939c3-run-systemd" (OuterVolumeSpecName: "run-systemd") pod "f010a463-69f2-4082-acc3-161899a939c3" (UID: "f010a463-69f2-4082-acc3-161899a939c3"). InnerVolumeSpecName "run-systemd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 08 18:21:50 crc kubenswrapper[4988]: I1008 18:21:50.771639 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-74ftp"] Oct 08 18:21:50 crc kubenswrapper[4988]: E1008 18:21:50.771894 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f010a463-69f2-4082-acc3-161899a939c3" containerName="ovnkube-controller" Oct 08 18:21:50 crc kubenswrapper[4988]: I1008 18:21:50.771954 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="f010a463-69f2-4082-acc3-161899a939c3" containerName="ovnkube-controller" Oct 08 18:21:50 crc kubenswrapper[4988]: E1008 18:21:50.772006 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8a33d7bd-26bd-409b-93f1-5f151ebca347" containerName="registry" Oct 08 18:21:50 crc kubenswrapper[4988]: I1008 18:21:50.772054 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="8a33d7bd-26bd-409b-93f1-5f151ebca347" containerName="registry" Oct 08 18:21:50 crc kubenswrapper[4988]: E1008 18:21:50.772108 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f010a463-69f2-4082-acc3-161899a939c3" containerName="kube-rbac-proxy-node" Oct 08 18:21:50 crc kubenswrapper[4988]: I1008 18:21:50.772153 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="f010a463-69f2-4082-acc3-161899a939c3" containerName="kube-rbac-proxy-node" Oct 08 18:21:50 crc kubenswrapper[4988]: E1008 18:21:50.772206 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f010a463-69f2-4082-acc3-161899a939c3" containerName="ovn-controller" Oct 08 18:21:50 crc kubenswrapper[4988]: I1008 18:21:50.772252 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="f010a463-69f2-4082-acc3-161899a939c3" containerName="ovn-controller" Oct 08 18:21:50 crc kubenswrapper[4988]: E1008 18:21:50.772301 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f010a463-69f2-4082-acc3-161899a939c3" containerName="northd" Oct 08 18:21:50 crc kubenswrapper[4988]: I1008 18:21:50.772347 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="f010a463-69f2-4082-acc3-161899a939c3" containerName="northd" Oct 08 18:21:50 crc kubenswrapper[4988]: E1008 18:21:50.772414 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f010a463-69f2-4082-acc3-161899a939c3" containerName="ovnkube-controller" Oct 08 18:21:50 crc kubenswrapper[4988]: I1008 18:21:50.772468 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="f010a463-69f2-4082-acc3-161899a939c3" containerName="ovnkube-controller" Oct 08 18:21:50 crc kubenswrapper[4988]: E1008 18:21:50.772517 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f010a463-69f2-4082-acc3-161899a939c3" containerName="sbdb" Oct 08 18:21:50 crc kubenswrapper[4988]: I1008 18:21:50.772560 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="f010a463-69f2-4082-acc3-161899a939c3" containerName="sbdb" Oct 08 18:21:50 crc kubenswrapper[4988]: E1008 18:21:50.772606 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f010a463-69f2-4082-acc3-161899a939c3" containerName="ovnkube-controller" Oct 08 18:21:50 crc kubenswrapper[4988]: I1008 18:21:50.772651 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="f010a463-69f2-4082-acc3-161899a939c3" containerName="ovnkube-controller" Oct 08 18:21:50 crc kubenswrapper[4988]: E1008 18:21:50.772702 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f010a463-69f2-4082-acc3-161899a939c3" containerName="kubecfg-setup" Oct 08 18:21:50 crc kubenswrapper[4988]: I1008 18:21:50.772748 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="f010a463-69f2-4082-acc3-161899a939c3" containerName="kubecfg-setup" Oct 08 18:21:50 crc kubenswrapper[4988]: E1008 18:21:50.772794 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f010a463-69f2-4082-acc3-161899a939c3" containerName="ovnkube-controller" Oct 08 18:21:50 crc kubenswrapper[4988]: I1008 18:21:50.772843 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="f010a463-69f2-4082-acc3-161899a939c3" containerName="ovnkube-controller" Oct 08 18:21:50 crc kubenswrapper[4988]: E1008 18:21:50.772892 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f010a463-69f2-4082-acc3-161899a939c3" containerName="kube-rbac-proxy-ovn-metrics" Oct 08 18:21:50 crc kubenswrapper[4988]: I1008 18:21:50.772942 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="f010a463-69f2-4082-acc3-161899a939c3" containerName="kube-rbac-proxy-ovn-metrics" Oct 08 18:21:50 crc kubenswrapper[4988]: E1008 18:21:50.772991 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f010a463-69f2-4082-acc3-161899a939c3" containerName="ovn-acl-logging" Oct 08 18:21:50 crc kubenswrapper[4988]: I1008 18:21:50.773041 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="f010a463-69f2-4082-acc3-161899a939c3" containerName="ovn-acl-logging" Oct 08 18:21:50 crc kubenswrapper[4988]: E1008 18:21:50.773089 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f010a463-69f2-4082-acc3-161899a939c3" containerName="nbdb" Oct 08 18:21:50 crc kubenswrapper[4988]: I1008 18:21:50.773135 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="f010a463-69f2-4082-acc3-161899a939c3" containerName="nbdb" Oct 08 18:21:50 crc kubenswrapper[4988]: I1008 18:21:50.773270 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="f010a463-69f2-4082-acc3-161899a939c3" containerName="ovn-acl-logging" Oct 08 18:21:50 crc kubenswrapper[4988]: I1008 18:21:50.773327 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="f010a463-69f2-4082-acc3-161899a939c3" containerName="ovn-controller" Oct 08 18:21:50 crc kubenswrapper[4988]: I1008 18:21:50.773379 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="8a33d7bd-26bd-409b-93f1-5f151ebca347" containerName="registry" Oct 08 18:21:50 crc kubenswrapper[4988]: I1008 18:21:50.773445 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="f010a463-69f2-4082-acc3-161899a939c3" containerName="ovnkube-controller" Oct 08 18:21:50 crc kubenswrapper[4988]: I1008 18:21:50.773494 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="f010a463-69f2-4082-acc3-161899a939c3" containerName="sbdb" Oct 08 18:21:50 crc kubenswrapper[4988]: I1008 18:21:50.773541 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="f010a463-69f2-4082-acc3-161899a939c3" containerName="kube-rbac-proxy-ovn-metrics" Oct 08 18:21:50 crc kubenswrapper[4988]: I1008 18:21:50.773587 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="f010a463-69f2-4082-acc3-161899a939c3" containerName="ovnkube-controller" Oct 08 18:21:50 crc kubenswrapper[4988]: I1008 18:21:50.773634 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="f010a463-69f2-4082-acc3-161899a939c3" containerName="northd" Oct 08 18:21:50 crc kubenswrapper[4988]: I1008 18:21:50.773678 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="f010a463-69f2-4082-acc3-161899a939c3" containerName="kube-rbac-proxy-node" Oct 08 18:21:50 crc kubenswrapper[4988]: I1008 18:21:50.773729 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="f010a463-69f2-4082-acc3-161899a939c3" containerName="nbdb" Oct 08 18:21:50 crc kubenswrapper[4988]: I1008 18:21:50.773775 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="f010a463-69f2-4082-acc3-161899a939c3" containerName="ovnkube-controller" Oct 08 18:21:50 crc kubenswrapper[4988]: I1008 18:21:50.773821 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="f010a463-69f2-4082-acc3-161899a939c3" containerName="ovnkube-controller" Oct 08 18:21:50 crc kubenswrapper[4988]: E1008 18:21:50.773951 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f010a463-69f2-4082-acc3-161899a939c3" containerName="ovnkube-controller" Oct 08 18:21:50 crc kubenswrapper[4988]: I1008 18:21:50.774000 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="f010a463-69f2-4082-acc3-161899a939c3" containerName="ovnkube-controller" Oct 08 18:21:50 crc kubenswrapper[4988]: I1008 18:21:50.774120 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="f010a463-69f2-4082-acc3-161899a939c3" containerName="ovnkube-controller" Oct 08 18:21:50 crc kubenswrapper[4988]: I1008 18:21:50.775589 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-74ftp" Oct 08 18:21:50 crc kubenswrapper[4988]: I1008 18:21:50.857830 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/8eee6858-0c8b-47a7-847b-ef454c951f6d-systemd-units\") pod \"ovnkube-node-74ftp\" (UID: \"8eee6858-0c8b-47a7-847b-ef454c951f6d\") " pod="openshift-ovn-kubernetes/ovnkube-node-74ftp" Oct 08 18:21:50 crc kubenswrapper[4988]: I1008 18:21:50.858118 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xt76f\" (UniqueName: \"kubernetes.io/projected/8eee6858-0c8b-47a7-847b-ef454c951f6d-kube-api-access-xt76f\") pod \"ovnkube-node-74ftp\" (UID: \"8eee6858-0c8b-47a7-847b-ef454c951f6d\") " pod="openshift-ovn-kubernetes/ovnkube-node-74ftp" Oct 08 18:21:50 crc kubenswrapper[4988]: I1008 18:21:50.858140 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/8eee6858-0c8b-47a7-847b-ef454c951f6d-env-overrides\") pod \"ovnkube-node-74ftp\" (UID: \"8eee6858-0c8b-47a7-847b-ef454c951f6d\") " pod="openshift-ovn-kubernetes/ovnkube-node-74ftp" Oct 08 18:21:50 crc kubenswrapper[4988]: I1008 18:21:50.858160 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/8eee6858-0c8b-47a7-847b-ef454c951f6d-run-ovn\") pod \"ovnkube-node-74ftp\" (UID: \"8eee6858-0c8b-47a7-847b-ef454c951f6d\") " pod="openshift-ovn-kubernetes/ovnkube-node-74ftp" Oct 08 18:21:50 crc kubenswrapper[4988]: I1008 18:21:50.858174 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/8eee6858-0c8b-47a7-847b-ef454c951f6d-host-run-netns\") pod \"ovnkube-node-74ftp\" (UID: \"8eee6858-0c8b-47a7-847b-ef454c951f6d\") " pod="openshift-ovn-kubernetes/ovnkube-node-74ftp" Oct 08 18:21:50 crc kubenswrapper[4988]: I1008 18:21:50.858194 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/8eee6858-0c8b-47a7-847b-ef454c951f6d-log-socket\") pod \"ovnkube-node-74ftp\" (UID: \"8eee6858-0c8b-47a7-847b-ef454c951f6d\") " pod="openshift-ovn-kubernetes/ovnkube-node-74ftp" Oct 08 18:21:50 crc kubenswrapper[4988]: I1008 18:21:50.858210 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/8eee6858-0c8b-47a7-847b-ef454c951f6d-host-run-ovn-kubernetes\") pod \"ovnkube-node-74ftp\" (UID: \"8eee6858-0c8b-47a7-847b-ef454c951f6d\") " pod="openshift-ovn-kubernetes/ovnkube-node-74ftp" Oct 08 18:21:50 crc kubenswrapper[4988]: I1008 18:21:50.858228 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/8eee6858-0c8b-47a7-847b-ef454c951f6d-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-74ftp\" (UID: \"8eee6858-0c8b-47a7-847b-ef454c951f6d\") " pod="openshift-ovn-kubernetes/ovnkube-node-74ftp" Oct 08 18:21:50 crc kubenswrapper[4988]: I1008 18:21:50.858341 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/8eee6858-0c8b-47a7-847b-ef454c951f6d-run-openvswitch\") pod \"ovnkube-node-74ftp\" (UID: \"8eee6858-0c8b-47a7-847b-ef454c951f6d\") " pod="openshift-ovn-kubernetes/ovnkube-node-74ftp" Oct 08 18:21:50 crc kubenswrapper[4988]: I1008 18:21:50.858414 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/8eee6858-0c8b-47a7-847b-ef454c951f6d-var-lib-openvswitch\") pod \"ovnkube-node-74ftp\" (UID: \"8eee6858-0c8b-47a7-847b-ef454c951f6d\") " pod="openshift-ovn-kubernetes/ovnkube-node-74ftp" Oct 08 18:21:50 crc kubenswrapper[4988]: I1008 18:21:50.858452 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/8eee6858-0c8b-47a7-847b-ef454c951f6d-host-slash\") pod \"ovnkube-node-74ftp\" (UID: \"8eee6858-0c8b-47a7-847b-ef454c951f6d\") " pod="openshift-ovn-kubernetes/ovnkube-node-74ftp" Oct 08 18:21:50 crc kubenswrapper[4988]: I1008 18:21:50.858513 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/8eee6858-0c8b-47a7-847b-ef454c951f6d-run-systemd\") pod \"ovnkube-node-74ftp\" (UID: \"8eee6858-0c8b-47a7-847b-ef454c951f6d\") " pod="openshift-ovn-kubernetes/ovnkube-node-74ftp" Oct 08 18:21:50 crc kubenswrapper[4988]: I1008 18:21:50.858610 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/8eee6858-0c8b-47a7-847b-ef454c951f6d-ovn-node-metrics-cert\") pod \"ovnkube-node-74ftp\" (UID: \"8eee6858-0c8b-47a7-847b-ef454c951f6d\") " pod="openshift-ovn-kubernetes/ovnkube-node-74ftp" Oct 08 18:21:50 crc kubenswrapper[4988]: I1008 18:21:50.858644 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/8eee6858-0c8b-47a7-847b-ef454c951f6d-host-kubelet\") pod \"ovnkube-node-74ftp\" (UID: \"8eee6858-0c8b-47a7-847b-ef454c951f6d\") " pod="openshift-ovn-kubernetes/ovnkube-node-74ftp" Oct 08 18:21:50 crc kubenswrapper[4988]: I1008 18:21:50.858667 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/8eee6858-0c8b-47a7-847b-ef454c951f6d-host-cni-bin\") pod \"ovnkube-node-74ftp\" (UID: \"8eee6858-0c8b-47a7-847b-ef454c951f6d\") " pod="openshift-ovn-kubernetes/ovnkube-node-74ftp" Oct 08 18:21:50 crc kubenswrapper[4988]: I1008 18:21:50.858711 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/8eee6858-0c8b-47a7-847b-ef454c951f6d-node-log\") pod \"ovnkube-node-74ftp\" (UID: \"8eee6858-0c8b-47a7-847b-ef454c951f6d\") " pod="openshift-ovn-kubernetes/ovnkube-node-74ftp" Oct 08 18:21:50 crc kubenswrapper[4988]: I1008 18:21:50.858730 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/8eee6858-0c8b-47a7-847b-ef454c951f6d-ovnkube-script-lib\") pod \"ovnkube-node-74ftp\" (UID: \"8eee6858-0c8b-47a7-847b-ef454c951f6d\") " pod="openshift-ovn-kubernetes/ovnkube-node-74ftp" Oct 08 18:21:50 crc kubenswrapper[4988]: I1008 18:21:50.858785 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/8eee6858-0c8b-47a7-847b-ef454c951f6d-host-cni-netd\") pod \"ovnkube-node-74ftp\" (UID: \"8eee6858-0c8b-47a7-847b-ef454c951f6d\") " pod="openshift-ovn-kubernetes/ovnkube-node-74ftp" Oct 08 18:21:50 crc kubenswrapper[4988]: I1008 18:21:50.858804 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/8eee6858-0c8b-47a7-847b-ef454c951f6d-ovnkube-config\") pod \"ovnkube-node-74ftp\" (UID: \"8eee6858-0c8b-47a7-847b-ef454c951f6d\") " pod="openshift-ovn-kubernetes/ovnkube-node-74ftp" Oct 08 18:21:50 crc kubenswrapper[4988]: I1008 18:21:50.858840 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/8eee6858-0c8b-47a7-847b-ef454c951f6d-etc-openvswitch\") pod \"ovnkube-node-74ftp\" (UID: \"8eee6858-0c8b-47a7-847b-ef454c951f6d\") " pod="openshift-ovn-kubernetes/ovnkube-node-74ftp" Oct 08 18:21:50 crc kubenswrapper[4988]: I1008 18:21:50.858922 4988 reconciler_common.go:293] "Volume detached for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/f010a463-69f2-4082-acc3-161899a939c3-host-kubelet\") on node \"crc\" DevicePath \"\"" Oct 08 18:21:50 crc kubenswrapper[4988]: I1008 18:21:50.858941 4988 reconciler_common.go:293] "Volume detached for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/f010a463-69f2-4082-acc3-161899a939c3-host-var-lib-cni-networks-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Oct 08 18:21:50 crc kubenswrapper[4988]: I1008 18:21:50.858952 4988 reconciler_common.go:293] "Volume detached for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/f010a463-69f2-4082-acc3-161899a939c3-host-slash\") on node \"crc\" DevicePath \"\"" Oct 08 18:21:50 crc kubenswrapper[4988]: I1008 18:21:50.858980 4988 reconciler_common.go:293] "Volume detached for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/f010a463-69f2-4082-acc3-161899a939c3-run-ovn\") on node \"crc\" DevicePath \"\"" Oct 08 18:21:50 crc kubenswrapper[4988]: I1008 18:21:50.858999 4988 reconciler_common.go:293] "Volume detached for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/f010a463-69f2-4082-acc3-161899a939c3-run-systemd\") on node \"crc\" DevicePath \"\"" Oct 08 18:21:50 crc kubenswrapper[4988]: I1008 18:21:50.859008 4988 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/f010a463-69f2-4082-acc3-161899a939c3-env-overrides\") on node \"crc\" DevicePath \"\"" Oct 08 18:21:50 crc kubenswrapper[4988]: I1008 18:21:50.859018 4988 reconciler_common.go:293] "Volume detached for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/f010a463-69f2-4082-acc3-161899a939c3-systemd-units\") on node \"crc\" DevicePath \"\"" Oct 08 18:21:50 crc kubenswrapper[4988]: I1008 18:21:50.859027 4988 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dnwgr\" (UniqueName: \"kubernetes.io/projected/f010a463-69f2-4082-acc3-161899a939c3-kube-api-access-dnwgr\") on node \"crc\" DevicePath \"\"" Oct 08 18:21:50 crc kubenswrapper[4988]: I1008 18:21:50.859035 4988 reconciler_common.go:293] "Volume detached for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/f010a463-69f2-4082-acc3-161899a939c3-log-socket\") on node \"crc\" DevicePath \"\"" Oct 08 18:21:50 crc kubenswrapper[4988]: I1008 18:21:50.859043 4988 reconciler_common.go:293] "Volume detached for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/f010a463-69f2-4082-acc3-161899a939c3-var-lib-openvswitch\") on node \"crc\" DevicePath \"\"" Oct 08 18:21:50 crc kubenswrapper[4988]: I1008 18:21:50.859052 4988 reconciler_common.go:293] "Volume detached for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/f010a463-69f2-4082-acc3-161899a939c3-node-log\") on node \"crc\" DevicePath \"\"" Oct 08 18:21:50 crc kubenswrapper[4988]: I1008 18:21:50.859062 4988 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/f010a463-69f2-4082-acc3-161899a939c3-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Oct 08 18:21:50 crc kubenswrapper[4988]: I1008 18:21:50.859071 4988 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/f010a463-69f2-4082-acc3-161899a939c3-ovnkube-config\") on node \"crc\" DevicePath \"\"" Oct 08 18:21:50 crc kubenswrapper[4988]: I1008 18:21:50.859080 4988 reconciler_common.go:293] "Volume detached for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/f010a463-69f2-4082-acc3-161899a939c3-run-openvswitch\") on node \"crc\" DevicePath \"\"" Oct 08 18:21:50 crc kubenswrapper[4988]: I1008 18:21:50.859091 4988 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/f010a463-69f2-4082-acc3-161899a939c3-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Oct 08 18:21:50 crc kubenswrapper[4988]: I1008 18:21:50.859099 4988 reconciler_common.go:293] "Volume detached for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/f010a463-69f2-4082-acc3-161899a939c3-host-run-netns\") on node \"crc\" DevicePath \"\"" Oct 08 18:21:50 crc kubenswrapper[4988]: I1008 18:21:50.878521 4988 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-nsc65_f010a463-69f2-4082-acc3-161899a939c3/ovnkube-controller/3.log" Oct 08 18:21:50 crc kubenswrapper[4988]: I1008 18:21:50.881178 4988 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-nsc65_f010a463-69f2-4082-acc3-161899a939c3/ovn-acl-logging/0.log" Oct 08 18:21:50 crc kubenswrapper[4988]: I1008 18:21:50.881738 4988 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-nsc65_f010a463-69f2-4082-acc3-161899a939c3/ovn-controller/0.log" Oct 08 18:21:50 crc kubenswrapper[4988]: I1008 18:21:50.882256 4988 generic.go:334] "Generic (PLEG): container finished" podID="f010a463-69f2-4082-acc3-161899a939c3" containerID="4987245ff838ca8d4b8c31a08147c052d23bdad26126fcb46367d4b5d4197fb2" exitCode=0 Oct 08 18:21:50 crc kubenswrapper[4988]: I1008 18:21:50.882307 4988 generic.go:334] "Generic (PLEG): container finished" podID="f010a463-69f2-4082-acc3-161899a939c3" containerID="74fcb812b84291763cc6e5ccb20c12aaad580ad968dda4734a0bbfe33d2770ca" exitCode=0 Oct 08 18:21:50 crc kubenswrapper[4988]: I1008 18:21:50.882316 4988 generic.go:334] "Generic (PLEG): container finished" podID="f010a463-69f2-4082-acc3-161899a939c3" containerID="9fbbeae0409d2e47e8de9a4306f63b0d13acbfd73cd86f83303fec14352ed718" exitCode=0 Oct 08 18:21:50 crc kubenswrapper[4988]: I1008 18:21:50.882325 4988 generic.go:334] "Generic (PLEG): container finished" podID="f010a463-69f2-4082-acc3-161899a939c3" containerID="55791327a370c0009c523b86b2234e82d3794985f2eaccf1922eb90f46b2b970" exitCode=0 Oct 08 18:21:50 crc kubenswrapper[4988]: I1008 18:21:50.882331 4988 generic.go:334] "Generic (PLEG): container finished" podID="f010a463-69f2-4082-acc3-161899a939c3" containerID="03406e726f817bcedb841f81959ab12808ba35a530240f9abc5fa5c608b9a6ae" exitCode=0 Oct 08 18:21:50 crc kubenswrapper[4988]: I1008 18:21:50.882340 4988 generic.go:334] "Generic (PLEG): container finished" podID="f010a463-69f2-4082-acc3-161899a939c3" containerID="48a516f6c153743bd1a21ec6945fb368a37f4d2b9d6a326b68edc096a192ed05" exitCode=0 Oct 08 18:21:50 crc kubenswrapper[4988]: I1008 18:21:50.882371 4988 generic.go:334] "Generic (PLEG): container finished" podID="f010a463-69f2-4082-acc3-161899a939c3" containerID="cee5de93bc68345f9b737857f02832f56582d58757c50a8798eec25d19b2c20d" exitCode=143 Oct 08 18:21:50 crc kubenswrapper[4988]: I1008 18:21:50.882430 4988 generic.go:334] "Generic (PLEG): container finished" podID="f010a463-69f2-4082-acc3-161899a939c3" containerID="d314eb248e90371fed068b1173e26d4c9f9765417a70b341e1a5b136ec17353a" exitCode=143 Oct 08 18:21:50 crc kubenswrapper[4988]: I1008 18:21:50.882409 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-nsc65" event={"ID":"f010a463-69f2-4082-acc3-161899a939c3","Type":"ContainerDied","Data":"4987245ff838ca8d4b8c31a08147c052d23bdad26126fcb46367d4b5d4197fb2"} Oct 08 18:21:50 crc kubenswrapper[4988]: I1008 18:21:50.882352 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-nsc65" Oct 08 18:21:50 crc kubenswrapper[4988]: I1008 18:21:50.882475 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-nsc65" event={"ID":"f010a463-69f2-4082-acc3-161899a939c3","Type":"ContainerDied","Data":"74fcb812b84291763cc6e5ccb20c12aaad580ad968dda4734a0bbfe33d2770ca"} Oct 08 18:21:50 crc kubenswrapper[4988]: I1008 18:21:50.882533 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-nsc65" event={"ID":"f010a463-69f2-4082-acc3-161899a939c3","Type":"ContainerDied","Data":"9fbbeae0409d2e47e8de9a4306f63b0d13acbfd73cd86f83303fec14352ed718"} Oct 08 18:21:50 crc kubenswrapper[4988]: I1008 18:21:50.882549 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-nsc65" event={"ID":"f010a463-69f2-4082-acc3-161899a939c3","Type":"ContainerDied","Data":"55791327a370c0009c523b86b2234e82d3794985f2eaccf1922eb90f46b2b970"} Oct 08 18:21:50 crc kubenswrapper[4988]: I1008 18:21:50.882563 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-nsc65" event={"ID":"f010a463-69f2-4082-acc3-161899a939c3","Type":"ContainerDied","Data":"03406e726f817bcedb841f81959ab12808ba35a530240f9abc5fa5c608b9a6ae"} Oct 08 18:21:50 crc kubenswrapper[4988]: I1008 18:21:50.882576 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-nsc65" event={"ID":"f010a463-69f2-4082-acc3-161899a939c3","Type":"ContainerDied","Data":"48a516f6c153743bd1a21ec6945fb368a37f4d2b9d6a326b68edc096a192ed05"} Oct 08 18:21:50 crc kubenswrapper[4988]: I1008 18:21:50.882589 4988 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"763353df23d26dc093a685af435811240b5ee772d7279ee9a6d2c7a48c19cd30"} Oct 08 18:21:50 crc kubenswrapper[4988]: I1008 18:21:50.882602 4988 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"74fcb812b84291763cc6e5ccb20c12aaad580ad968dda4734a0bbfe33d2770ca"} Oct 08 18:21:50 crc kubenswrapper[4988]: I1008 18:21:50.882610 4988 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"9fbbeae0409d2e47e8de9a4306f63b0d13acbfd73cd86f83303fec14352ed718"} Oct 08 18:21:50 crc kubenswrapper[4988]: I1008 18:21:50.882618 4988 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"55791327a370c0009c523b86b2234e82d3794985f2eaccf1922eb90f46b2b970"} Oct 08 18:21:50 crc kubenswrapper[4988]: I1008 18:21:50.882625 4988 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"03406e726f817bcedb841f81959ab12808ba35a530240f9abc5fa5c608b9a6ae"} Oct 08 18:21:50 crc kubenswrapper[4988]: I1008 18:21:50.882632 4988 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"48a516f6c153743bd1a21ec6945fb368a37f4d2b9d6a326b68edc096a192ed05"} Oct 08 18:21:50 crc kubenswrapper[4988]: I1008 18:21:50.882639 4988 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"cee5de93bc68345f9b737857f02832f56582d58757c50a8798eec25d19b2c20d"} Oct 08 18:21:50 crc kubenswrapper[4988]: I1008 18:21:50.882646 4988 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"d314eb248e90371fed068b1173e26d4c9f9765417a70b341e1a5b136ec17353a"} Oct 08 18:21:50 crc kubenswrapper[4988]: I1008 18:21:50.882654 4988 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"befe684cea9bd6dfda3ec8e85bca431c26bd8f82823b0a69dc0d04319a3ca385"} Oct 08 18:21:50 crc kubenswrapper[4988]: I1008 18:21:50.882664 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-nsc65" event={"ID":"f010a463-69f2-4082-acc3-161899a939c3","Type":"ContainerDied","Data":"cee5de93bc68345f9b737857f02832f56582d58757c50a8798eec25d19b2c20d"} Oct 08 18:21:50 crc kubenswrapper[4988]: I1008 18:21:50.882676 4988 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"4987245ff838ca8d4b8c31a08147c052d23bdad26126fcb46367d4b5d4197fb2"} Oct 08 18:21:50 crc kubenswrapper[4988]: I1008 18:21:50.882685 4988 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"763353df23d26dc093a685af435811240b5ee772d7279ee9a6d2c7a48c19cd30"} Oct 08 18:21:50 crc kubenswrapper[4988]: I1008 18:21:50.882692 4988 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"74fcb812b84291763cc6e5ccb20c12aaad580ad968dda4734a0bbfe33d2770ca"} Oct 08 18:21:50 crc kubenswrapper[4988]: I1008 18:21:50.882647 4988 scope.go:117] "RemoveContainer" containerID="4987245ff838ca8d4b8c31a08147c052d23bdad26126fcb46367d4b5d4197fb2" Oct 08 18:21:50 crc kubenswrapper[4988]: I1008 18:21:50.882791 4988 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"9fbbeae0409d2e47e8de9a4306f63b0d13acbfd73cd86f83303fec14352ed718"} Oct 08 18:21:50 crc kubenswrapper[4988]: I1008 18:21:50.882805 4988 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"55791327a370c0009c523b86b2234e82d3794985f2eaccf1922eb90f46b2b970"} Oct 08 18:21:50 crc kubenswrapper[4988]: I1008 18:21:50.882813 4988 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"03406e726f817bcedb841f81959ab12808ba35a530240f9abc5fa5c608b9a6ae"} Oct 08 18:21:50 crc kubenswrapper[4988]: I1008 18:21:50.882821 4988 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"48a516f6c153743bd1a21ec6945fb368a37f4d2b9d6a326b68edc096a192ed05"} Oct 08 18:21:50 crc kubenswrapper[4988]: I1008 18:21:50.882828 4988 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"cee5de93bc68345f9b737857f02832f56582d58757c50a8798eec25d19b2c20d"} Oct 08 18:21:50 crc kubenswrapper[4988]: I1008 18:21:50.882836 4988 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"d314eb248e90371fed068b1173e26d4c9f9765417a70b341e1a5b136ec17353a"} Oct 08 18:21:50 crc kubenswrapper[4988]: I1008 18:21:50.882842 4988 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"befe684cea9bd6dfda3ec8e85bca431c26bd8f82823b0a69dc0d04319a3ca385"} Oct 08 18:21:50 crc kubenswrapper[4988]: I1008 18:21:50.882854 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-nsc65" event={"ID":"f010a463-69f2-4082-acc3-161899a939c3","Type":"ContainerDied","Data":"d314eb248e90371fed068b1173e26d4c9f9765417a70b341e1a5b136ec17353a"} Oct 08 18:21:50 crc kubenswrapper[4988]: I1008 18:21:50.882868 4988 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"4987245ff838ca8d4b8c31a08147c052d23bdad26126fcb46367d4b5d4197fb2"} Oct 08 18:21:50 crc kubenswrapper[4988]: I1008 18:21:50.882877 4988 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"763353df23d26dc093a685af435811240b5ee772d7279ee9a6d2c7a48c19cd30"} Oct 08 18:21:50 crc kubenswrapper[4988]: I1008 18:21:50.882884 4988 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"74fcb812b84291763cc6e5ccb20c12aaad580ad968dda4734a0bbfe33d2770ca"} Oct 08 18:21:50 crc kubenswrapper[4988]: I1008 18:21:50.882891 4988 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"9fbbeae0409d2e47e8de9a4306f63b0d13acbfd73cd86f83303fec14352ed718"} Oct 08 18:21:50 crc kubenswrapper[4988]: I1008 18:21:50.882919 4988 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"55791327a370c0009c523b86b2234e82d3794985f2eaccf1922eb90f46b2b970"} Oct 08 18:21:50 crc kubenswrapper[4988]: I1008 18:21:50.882927 4988 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"03406e726f817bcedb841f81959ab12808ba35a530240f9abc5fa5c608b9a6ae"} Oct 08 18:21:50 crc kubenswrapper[4988]: I1008 18:21:50.882934 4988 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"48a516f6c153743bd1a21ec6945fb368a37f4d2b9d6a326b68edc096a192ed05"} Oct 08 18:21:50 crc kubenswrapper[4988]: I1008 18:21:50.882940 4988 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"cee5de93bc68345f9b737857f02832f56582d58757c50a8798eec25d19b2c20d"} Oct 08 18:21:50 crc kubenswrapper[4988]: I1008 18:21:50.882947 4988 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"d314eb248e90371fed068b1173e26d4c9f9765417a70b341e1a5b136ec17353a"} Oct 08 18:21:50 crc kubenswrapper[4988]: I1008 18:21:50.882953 4988 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"befe684cea9bd6dfda3ec8e85bca431c26bd8f82823b0a69dc0d04319a3ca385"} Oct 08 18:21:50 crc kubenswrapper[4988]: I1008 18:21:50.882964 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-nsc65" event={"ID":"f010a463-69f2-4082-acc3-161899a939c3","Type":"ContainerDied","Data":"937461074d800725cfae8d0f2f3c1a96fcb0473c00b305eb9c50b400d5f92549"} Oct 08 18:21:50 crc kubenswrapper[4988]: I1008 18:21:50.882976 4988 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"4987245ff838ca8d4b8c31a08147c052d23bdad26126fcb46367d4b5d4197fb2"} Oct 08 18:21:50 crc kubenswrapper[4988]: I1008 18:21:50.882985 4988 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"763353df23d26dc093a685af435811240b5ee772d7279ee9a6d2c7a48c19cd30"} Oct 08 18:21:50 crc kubenswrapper[4988]: I1008 18:21:50.882992 4988 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"74fcb812b84291763cc6e5ccb20c12aaad580ad968dda4734a0bbfe33d2770ca"} Oct 08 18:21:50 crc kubenswrapper[4988]: I1008 18:21:50.882999 4988 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"9fbbeae0409d2e47e8de9a4306f63b0d13acbfd73cd86f83303fec14352ed718"} Oct 08 18:21:50 crc kubenswrapper[4988]: I1008 18:21:50.883006 4988 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"55791327a370c0009c523b86b2234e82d3794985f2eaccf1922eb90f46b2b970"} Oct 08 18:21:50 crc kubenswrapper[4988]: I1008 18:21:50.883014 4988 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"03406e726f817bcedb841f81959ab12808ba35a530240f9abc5fa5c608b9a6ae"} Oct 08 18:21:50 crc kubenswrapper[4988]: I1008 18:21:50.883021 4988 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"48a516f6c153743bd1a21ec6945fb368a37f4d2b9d6a326b68edc096a192ed05"} Oct 08 18:21:50 crc kubenswrapper[4988]: I1008 18:21:50.883027 4988 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"cee5de93bc68345f9b737857f02832f56582d58757c50a8798eec25d19b2c20d"} Oct 08 18:21:50 crc kubenswrapper[4988]: I1008 18:21:50.883034 4988 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"d314eb248e90371fed068b1173e26d4c9f9765417a70b341e1a5b136ec17353a"} Oct 08 18:21:50 crc kubenswrapper[4988]: I1008 18:21:50.883042 4988 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"befe684cea9bd6dfda3ec8e85bca431c26bd8f82823b0a69dc0d04319a3ca385"} Oct 08 18:21:50 crc kubenswrapper[4988]: I1008 18:21:50.885671 4988 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-h9jwp_4490d846-ca32-4cd5-9202-87cc6d499fb4/kube-multus/2.log" Oct 08 18:21:50 crc kubenswrapper[4988]: I1008 18:21:50.886244 4988 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-h9jwp_4490d846-ca32-4cd5-9202-87cc6d499fb4/kube-multus/1.log" Oct 08 18:21:50 crc kubenswrapper[4988]: I1008 18:21:50.886280 4988 generic.go:334] "Generic (PLEG): container finished" podID="4490d846-ca32-4cd5-9202-87cc6d499fb4" containerID="f62030d24f69a85e1807099969f0021ab1198002c76e11ef3dffadfabff73b49" exitCode=2 Oct 08 18:21:50 crc kubenswrapper[4988]: I1008 18:21:50.886303 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-h9jwp" event={"ID":"4490d846-ca32-4cd5-9202-87cc6d499fb4","Type":"ContainerDied","Data":"f62030d24f69a85e1807099969f0021ab1198002c76e11ef3dffadfabff73b49"} Oct 08 18:21:50 crc kubenswrapper[4988]: I1008 18:21:50.886320 4988 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"c38c1b52127a970f792ef09ef1b217c979d94f7f74cc3c27e38949a204ad831f"} Oct 08 18:21:50 crc kubenswrapper[4988]: I1008 18:21:50.886797 4988 scope.go:117] "RemoveContainer" containerID="f62030d24f69a85e1807099969f0021ab1198002c76e11ef3dffadfabff73b49" Oct 08 18:21:50 crc kubenswrapper[4988]: E1008 18:21:50.887032 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-multus pod=multus-h9jwp_openshift-multus(4490d846-ca32-4cd5-9202-87cc6d499fb4)\"" pod="openshift-multus/multus-h9jwp" podUID="4490d846-ca32-4cd5-9202-87cc6d499fb4" Oct 08 18:21:50 crc kubenswrapper[4988]: I1008 18:21:50.897981 4988 scope.go:117] "RemoveContainer" containerID="763353df23d26dc093a685af435811240b5ee772d7279ee9a6d2c7a48c19cd30" Oct 08 18:21:50 crc kubenswrapper[4988]: I1008 18:21:50.918661 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-nsc65"] Oct 08 18:21:50 crc kubenswrapper[4988]: I1008 18:21:50.926666 4988 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-nsc65"] Oct 08 18:21:50 crc kubenswrapper[4988]: I1008 18:21:50.933335 4988 scope.go:117] "RemoveContainer" containerID="74fcb812b84291763cc6e5ccb20c12aaad580ad968dda4734a0bbfe33d2770ca" Oct 08 18:21:50 crc kubenswrapper[4988]: I1008 18:21:50.947654 4988 scope.go:117] "RemoveContainer" containerID="9fbbeae0409d2e47e8de9a4306f63b0d13acbfd73cd86f83303fec14352ed718" Oct 08 18:21:50 crc kubenswrapper[4988]: I1008 18:21:50.959573 4988 scope.go:117] "RemoveContainer" containerID="55791327a370c0009c523b86b2234e82d3794985f2eaccf1922eb90f46b2b970" Oct 08 18:21:50 crc kubenswrapper[4988]: I1008 18:21:50.960140 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/8eee6858-0c8b-47a7-847b-ef454c951f6d-ovn-node-metrics-cert\") pod \"ovnkube-node-74ftp\" (UID: \"8eee6858-0c8b-47a7-847b-ef454c951f6d\") " pod="openshift-ovn-kubernetes/ovnkube-node-74ftp" Oct 08 18:21:50 crc kubenswrapper[4988]: I1008 18:21:50.960190 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/8eee6858-0c8b-47a7-847b-ef454c951f6d-host-kubelet\") pod \"ovnkube-node-74ftp\" (UID: \"8eee6858-0c8b-47a7-847b-ef454c951f6d\") " pod="openshift-ovn-kubernetes/ovnkube-node-74ftp" Oct 08 18:21:50 crc kubenswrapper[4988]: I1008 18:21:50.960219 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/8eee6858-0c8b-47a7-847b-ef454c951f6d-host-cni-bin\") pod \"ovnkube-node-74ftp\" (UID: \"8eee6858-0c8b-47a7-847b-ef454c951f6d\") " pod="openshift-ovn-kubernetes/ovnkube-node-74ftp" Oct 08 18:21:50 crc kubenswrapper[4988]: I1008 18:21:50.960251 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/8eee6858-0c8b-47a7-847b-ef454c951f6d-node-log\") pod \"ovnkube-node-74ftp\" (UID: \"8eee6858-0c8b-47a7-847b-ef454c951f6d\") " pod="openshift-ovn-kubernetes/ovnkube-node-74ftp" Oct 08 18:21:50 crc kubenswrapper[4988]: I1008 18:21:50.960275 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/8eee6858-0c8b-47a7-847b-ef454c951f6d-ovnkube-script-lib\") pod \"ovnkube-node-74ftp\" (UID: \"8eee6858-0c8b-47a7-847b-ef454c951f6d\") " pod="openshift-ovn-kubernetes/ovnkube-node-74ftp" Oct 08 18:21:50 crc kubenswrapper[4988]: I1008 18:21:50.960314 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/8eee6858-0c8b-47a7-847b-ef454c951f6d-host-cni-bin\") pod \"ovnkube-node-74ftp\" (UID: \"8eee6858-0c8b-47a7-847b-ef454c951f6d\") " pod="openshift-ovn-kubernetes/ovnkube-node-74ftp" Oct 08 18:21:50 crc kubenswrapper[4988]: I1008 18:21:50.960316 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/8eee6858-0c8b-47a7-847b-ef454c951f6d-node-log\") pod \"ovnkube-node-74ftp\" (UID: \"8eee6858-0c8b-47a7-847b-ef454c951f6d\") " pod="openshift-ovn-kubernetes/ovnkube-node-74ftp" Oct 08 18:21:50 crc kubenswrapper[4988]: I1008 18:21:50.960321 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/8eee6858-0c8b-47a7-847b-ef454c951f6d-host-cni-netd\") pod \"ovnkube-node-74ftp\" (UID: \"8eee6858-0c8b-47a7-847b-ef454c951f6d\") " pod="openshift-ovn-kubernetes/ovnkube-node-74ftp" Oct 08 18:21:50 crc kubenswrapper[4988]: I1008 18:21:50.960429 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/8eee6858-0c8b-47a7-847b-ef454c951f6d-ovnkube-config\") pod \"ovnkube-node-74ftp\" (UID: \"8eee6858-0c8b-47a7-847b-ef454c951f6d\") " pod="openshift-ovn-kubernetes/ovnkube-node-74ftp" Oct 08 18:21:50 crc kubenswrapper[4988]: I1008 18:21:50.960351 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/8eee6858-0c8b-47a7-847b-ef454c951f6d-host-cni-netd\") pod \"ovnkube-node-74ftp\" (UID: \"8eee6858-0c8b-47a7-847b-ef454c951f6d\") " pod="openshift-ovn-kubernetes/ovnkube-node-74ftp" Oct 08 18:21:50 crc kubenswrapper[4988]: I1008 18:21:50.960477 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/8eee6858-0c8b-47a7-847b-ef454c951f6d-etc-openvswitch\") pod \"ovnkube-node-74ftp\" (UID: \"8eee6858-0c8b-47a7-847b-ef454c951f6d\") " pod="openshift-ovn-kubernetes/ovnkube-node-74ftp" Oct 08 18:21:50 crc kubenswrapper[4988]: I1008 18:21:50.960521 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/8eee6858-0c8b-47a7-847b-ef454c951f6d-systemd-units\") pod \"ovnkube-node-74ftp\" (UID: \"8eee6858-0c8b-47a7-847b-ef454c951f6d\") " pod="openshift-ovn-kubernetes/ovnkube-node-74ftp" Oct 08 18:21:50 crc kubenswrapper[4988]: I1008 18:21:50.960537 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/8eee6858-0c8b-47a7-847b-ef454c951f6d-etc-openvswitch\") pod \"ovnkube-node-74ftp\" (UID: \"8eee6858-0c8b-47a7-847b-ef454c951f6d\") " pod="openshift-ovn-kubernetes/ovnkube-node-74ftp" Oct 08 18:21:50 crc kubenswrapper[4988]: I1008 18:21:50.960548 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xt76f\" (UniqueName: \"kubernetes.io/projected/8eee6858-0c8b-47a7-847b-ef454c951f6d-kube-api-access-xt76f\") pod \"ovnkube-node-74ftp\" (UID: \"8eee6858-0c8b-47a7-847b-ef454c951f6d\") " pod="openshift-ovn-kubernetes/ovnkube-node-74ftp" Oct 08 18:21:50 crc kubenswrapper[4988]: I1008 18:21:50.960593 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/8eee6858-0c8b-47a7-847b-ef454c951f6d-env-overrides\") pod \"ovnkube-node-74ftp\" (UID: \"8eee6858-0c8b-47a7-847b-ef454c951f6d\") " pod="openshift-ovn-kubernetes/ovnkube-node-74ftp" Oct 08 18:21:50 crc kubenswrapper[4988]: I1008 18:21:50.961022 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/8eee6858-0c8b-47a7-847b-ef454c951f6d-run-ovn\") pod \"ovnkube-node-74ftp\" (UID: \"8eee6858-0c8b-47a7-847b-ef454c951f6d\") " pod="openshift-ovn-kubernetes/ovnkube-node-74ftp" Oct 08 18:21:50 crc kubenswrapper[4988]: I1008 18:21:50.961099 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/8eee6858-0c8b-47a7-847b-ef454c951f6d-run-ovn\") pod \"ovnkube-node-74ftp\" (UID: \"8eee6858-0c8b-47a7-847b-ef454c951f6d\") " pod="openshift-ovn-kubernetes/ovnkube-node-74ftp" Oct 08 18:21:50 crc kubenswrapper[4988]: I1008 18:21:50.960655 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/8eee6858-0c8b-47a7-847b-ef454c951f6d-systemd-units\") pod \"ovnkube-node-74ftp\" (UID: \"8eee6858-0c8b-47a7-847b-ef454c951f6d\") " pod="openshift-ovn-kubernetes/ovnkube-node-74ftp" Oct 08 18:21:50 crc kubenswrapper[4988]: I1008 18:21:50.961136 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/8eee6858-0c8b-47a7-847b-ef454c951f6d-host-run-netns\") pod \"ovnkube-node-74ftp\" (UID: \"8eee6858-0c8b-47a7-847b-ef454c951f6d\") " pod="openshift-ovn-kubernetes/ovnkube-node-74ftp" Oct 08 18:21:50 crc kubenswrapper[4988]: I1008 18:21:50.961108 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/8eee6858-0c8b-47a7-847b-ef454c951f6d-host-run-netns\") pod \"ovnkube-node-74ftp\" (UID: \"8eee6858-0c8b-47a7-847b-ef454c951f6d\") " pod="openshift-ovn-kubernetes/ovnkube-node-74ftp" Oct 08 18:21:50 crc kubenswrapper[4988]: I1008 18:21:50.960703 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/8eee6858-0c8b-47a7-847b-ef454c951f6d-host-kubelet\") pod \"ovnkube-node-74ftp\" (UID: \"8eee6858-0c8b-47a7-847b-ef454c951f6d\") " pod="openshift-ovn-kubernetes/ovnkube-node-74ftp" Oct 08 18:21:50 crc kubenswrapper[4988]: I1008 18:21:50.961187 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/8eee6858-0c8b-47a7-847b-ef454c951f6d-log-socket\") pod \"ovnkube-node-74ftp\" (UID: \"8eee6858-0c8b-47a7-847b-ef454c951f6d\") " pod="openshift-ovn-kubernetes/ovnkube-node-74ftp" Oct 08 18:21:50 crc kubenswrapper[4988]: I1008 18:21:50.961211 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/8eee6858-0c8b-47a7-847b-ef454c951f6d-host-run-ovn-kubernetes\") pod \"ovnkube-node-74ftp\" (UID: \"8eee6858-0c8b-47a7-847b-ef454c951f6d\") " pod="openshift-ovn-kubernetes/ovnkube-node-74ftp" Oct 08 18:21:50 crc kubenswrapper[4988]: I1008 18:21:50.961236 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/8eee6858-0c8b-47a7-847b-ef454c951f6d-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-74ftp\" (UID: \"8eee6858-0c8b-47a7-847b-ef454c951f6d\") " pod="openshift-ovn-kubernetes/ovnkube-node-74ftp" Oct 08 18:21:50 crc kubenswrapper[4988]: I1008 18:21:50.961267 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/8eee6858-0c8b-47a7-847b-ef454c951f6d-run-openvswitch\") pod \"ovnkube-node-74ftp\" (UID: \"8eee6858-0c8b-47a7-847b-ef454c951f6d\") " pod="openshift-ovn-kubernetes/ovnkube-node-74ftp" Oct 08 18:21:50 crc kubenswrapper[4988]: I1008 18:21:50.961291 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/8eee6858-0c8b-47a7-847b-ef454c951f6d-var-lib-openvswitch\") pod \"ovnkube-node-74ftp\" (UID: \"8eee6858-0c8b-47a7-847b-ef454c951f6d\") " pod="openshift-ovn-kubernetes/ovnkube-node-74ftp" Oct 08 18:21:50 crc kubenswrapper[4988]: I1008 18:21:50.961298 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/8eee6858-0c8b-47a7-847b-ef454c951f6d-host-run-ovn-kubernetes\") pod \"ovnkube-node-74ftp\" (UID: \"8eee6858-0c8b-47a7-847b-ef454c951f6d\") " pod="openshift-ovn-kubernetes/ovnkube-node-74ftp" Oct 08 18:21:50 crc kubenswrapper[4988]: I1008 18:21:50.961317 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/8eee6858-0c8b-47a7-847b-ef454c951f6d-host-slash\") pod \"ovnkube-node-74ftp\" (UID: \"8eee6858-0c8b-47a7-847b-ef454c951f6d\") " pod="openshift-ovn-kubernetes/ovnkube-node-74ftp" Oct 08 18:21:50 crc kubenswrapper[4988]: I1008 18:21:50.961343 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/8eee6858-0c8b-47a7-847b-ef454c951f6d-log-socket\") pod \"ovnkube-node-74ftp\" (UID: \"8eee6858-0c8b-47a7-847b-ef454c951f6d\") " pod="openshift-ovn-kubernetes/ovnkube-node-74ftp" Oct 08 18:21:50 crc kubenswrapper[4988]: I1008 18:21:50.961346 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/8eee6858-0c8b-47a7-847b-ef454c951f6d-run-systemd\") pod \"ovnkube-node-74ftp\" (UID: \"8eee6858-0c8b-47a7-847b-ef454c951f6d\") " pod="openshift-ovn-kubernetes/ovnkube-node-74ftp" Oct 08 18:21:50 crc kubenswrapper[4988]: I1008 18:21:50.961370 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/8eee6858-0c8b-47a7-847b-ef454c951f6d-run-systemd\") pod \"ovnkube-node-74ftp\" (UID: \"8eee6858-0c8b-47a7-847b-ef454c951f6d\") " pod="openshift-ovn-kubernetes/ovnkube-node-74ftp" Oct 08 18:21:50 crc kubenswrapper[4988]: I1008 18:21:50.961451 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/8eee6858-0c8b-47a7-847b-ef454c951f6d-host-slash\") pod \"ovnkube-node-74ftp\" (UID: \"8eee6858-0c8b-47a7-847b-ef454c951f6d\") " pod="openshift-ovn-kubernetes/ovnkube-node-74ftp" Oct 08 18:21:50 crc kubenswrapper[4988]: I1008 18:21:50.961471 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/8eee6858-0c8b-47a7-847b-ef454c951f6d-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-74ftp\" (UID: \"8eee6858-0c8b-47a7-847b-ef454c951f6d\") " pod="openshift-ovn-kubernetes/ovnkube-node-74ftp" Oct 08 18:21:50 crc kubenswrapper[4988]: I1008 18:21:50.961487 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/8eee6858-0c8b-47a7-847b-ef454c951f6d-run-openvswitch\") pod \"ovnkube-node-74ftp\" (UID: \"8eee6858-0c8b-47a7-847b-ef454c951f6d\") " pod="openshift-ovn-kubernetes/ovnkube-node-74ftp" Oct 08 18:21:50 crc kubenswrapper[4988]: I1008 18:21:50.961506 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/8eee6858-0c8b-47a7-847b-ef454c951f6d-var-lib-openvswitch\") pod \"ovnkube-node-74ftp\" (UID: \"8eee6858-0c8b-47a7-847b-ef454c951f6d\") " pod="openshift-ovn-kubernetes/ovnkube-node-74ftp" Oct 08 18:21:50 crc kubenswrapper[4988]: I1008 18:21:50.961766 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/8eee6858-0c8b-47a7-847b-ef454c951f6d-env-overrides\") pod \"ovnkube-node-74ftp\" (UID: \"8eee6858-0c8b-47a7-847b-ef454c951f6d\") " pod="openshift-ovn-kubernetes/ovnkube-node-74ftp" Oct 08 18:21:50 crc kubenswrapper[4988]: I1008 18:21:50.962237 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/8eee6858-0c8b-47a7-847b-ef454c951f6d-ovnkube-script-lib\") pod \"ovnkube-node-74ftp\" (UID: \"8eee6858-0c8b-47a7-847b-ef454c951f6d\") " pod="openshift-ovn-kubernetes/ovnkube-node-74ftp" Oct 08 18:21:50 crc kubenswrapper[4988]: I1008 18:21:50.962706 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/8eee6858-0c8b-47a7-847b-ef454c951f6d-ovnkube-config\") pod \"ovnkube-node-74ftp\" (UID: \"8eee6858-0c8b-47a7-847b-ef454c951f6d\") " pod="openshift-ovn-kubernetes/ovnkube-node-74ftp" Oct 08 18:21:50 crc kubenswrapper[4988]: I1008 18:21:50.965154 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/8eee6858-0c8b-47a7-847b-ef454c951f6d-ovn-node-metrics-cert\") pod \"ovnkube-node-74ftp\" (UID: \"8eee6858-0c8b-47a7-847b-ef454c951f6d\") " pod="openshift-ovn-kubernetes/ovnkube-node-74ftp" Oct 08 18:21:50 crc kubenswrapper[4988]: I1008 18:21:50.974728 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xt76f\" (UniqueName: \"kubernetes.io/projected/8eee6858-0c8b-47a7-847b-ef454c951f6d-kube-api-access-xt76f\") pod \"ovnkube-node-74ftp\" (UID: \"8eee6858-0c8b-47a7-847b-ef454c951f6d\") " pod="openshift-ovn-kubernetes/ovnkube-node-74ftp" Oct 08 18:21:50 crc kubenswrapper[4988]: I1008 18:21:50.974912 4988 scope.go:117] "RemoveContainer" containerID="03406e726f817bcedb841f81959ab12808ba35a530240f9abc5fa5c608b9a6ae" Oct 08 18:21:50 crc kubenswrapper[4988]: I1008 18:21:50.987632 4988 scope.go:117] "RemoveContainer" containerID="48a516f6c153743bd1a21ec6945fb368a37f4d2b9d6a326b68edc096a192ed05" Oct 08 18:21:50 crc kubenswrapper[4988]: I1008 18:21:50.998966 4988 scope.go:117] "RemoveContainer" containerID="cee5de93bc68345f9b737857f02832f56582d58757c50a8798eec25d19b2c20d" Oct 08 18:21:51 crc kubenswrapper[4988]: I1008 18:21:51.011406 4988 scope.go:117] "RemoveContainer" containerID="d314eb248e90371fed068b1173e26d4c9f9765417a70b341e1a5b136ec17353a" Oct 08 18:21:51 crc kubenswrapper[4988]: I1008 18:21:51.031269 4988 scope.go:117] "RemoveContainer" containerID="befe684cea9bd6dfda3ec8e85bca431c26bd8f82823b0a69dc0d04319a3ca385" Oct 08 18:21:51 crc kubenswrapper[4988]: I1008 18:21:51.051867 4988 scope.go:117] "RemoveContainer" containerID="4987245ff838ca8d4b8c31a08147c052d23bdad26126fcb46367d4b5d4197fb2" Oct 08 18:21:51 crc kubenswrapper[4988]: E1008 18:21:51.052373 4988 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4987245ff838ca8d4b8c31a08147c052d23bdad26126fcb46367d4b5d4197fb2\": container with ID starting with 4987245ff838ca8d4b8c31a08147c052d23bdad26126fcb46367d4b5d4197fb2 not found: ID does not exist" containerID="4987245ff838ca8d4b8c31a08147c052d23bdad26126fcb46367d4b5d4197fb2" Oct 08 18:21:51 crc kubenswrapper[4988]: I1008 18:21:51.052441 4988 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4987245ff838ca8d4b8c31a08147c052d23bdad26126fcb46367d4b5d4197fb2"} err="failed to get container status \"4987245ff838ca8d4b8c31a08147c052d23bdad26126fcb46367d4b5d4197fb2\": rpc error: code = NotFound desc = could not find container \"4987245ff838ca8d4b8c31a08147c052d23bdad26126fcb46367d4b5d4197fb2\": container with ID starting with 4987245ff838ca8d4b8c31a08147c052d23bdad26126fcb46367d4b5d4197fb2 not found: ID does not exist" Oct 08 18:21:51 crc kubenswrapper[4988]: I1008 18:21:51.052479 4988 scope.go:117] "RemoveContainer" containerID="763353df23d26dc093a685af435811240b5ee772d7279ee9a6d2c7a48c19cd30" Oct 08 18:21:51 crc kubenswrapper[4988]: E1008 18:21:51.052870 4988 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"763353df23d26dc093a685af435811240b5ee772d7279ee9a6d2c7a48c19cd30\": container with ID starting with 763353df23d26dc093a685af435811240b5ee772d7279ee9a6d2c7a48c19cd30 not found: ID does not exist" containerID="763353df23d26dc093a685af435811240b5ee772d7279ee9a6d2c7a48c19cd30" Oct 08 18:21:51 crc kubenswrapper[4988]: I1008 18:21:51.052918 4988 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"763353df23d26dc093a685af435811240b5ee772d7279ee9a6d2c7a48c19cd30"} err="failed to get container status \"763353df23d26dc093a685af435811240b5ee772d7279ee9a6d2c7a48c19cd30\": rpc error: code = NotFound desc = could not find container \"763353df23d26dc093a685af435811240b5ee772d7279ee9a6d2c7a48c19cd30\": container with ID starting with 763353df23d26dc093a685af435811240b5ee772d7279ee9a6d2c7a48c19cd30 not found: ID does not exist" Oct 08 18:21:51 crc kubenswrapper[4988]: I1008 18:21:51.052945 4988 scope.go:117] "RemoveContainer" containerID="74fcb812b84291763cc6e5ccb20c12aaad580ad968dda4734a0bbfe33d2770ca" Oct 08 18:21:51 crc kubenswrapper[4988]: E1008 18:21:51.053350 4988 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"74fcb812b84291763cc6e5ccb20c12aaad580ad968dda4734a0bbfe33d2770ca\": container with ID starting with 74fcb812b84291763cc6e5ccb20c12aaad580ad968dda4734a0bbfe33d2770ca not found: ID does not exist" containerID="74fcb812b84291763cc6e5ccb20c12aaad580ad968dda4734a0bbfe33d2770ca" Oct 08 18:21:51 crc kubenswrapper[4988]: I1008 18:21:51.053376 4988 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"74fcb812b84291763cc6e5ccb20c12aaad580ad968dda4734a0bbfe33d2770ca"} err="failed to get container status \"74fcb812b84291763cc6e5ccb20c12aaad580ad968dda4734a0bbfe33d2770ca\": rpc error: code = NotFound desc = could not find container \"74fcb812b84291763cc6e5ccb20c12aaad580ad968dda4734a0bbfe33d2770ca\": container with ID starting with 74fcb812b84291763cc6e5ccb20c12aaad580ad968dda4734a0bbfe33d2770ca not found: ID does not exist" Oct 08 18:21:51 crc kubenswrapper[4988]: I1008 18:21:51.053410 4988 scope.go:117] "RemoveContainer" containerID="9fbbeae0409d2e47e8de9a4306f63b0d13acbfd73cd86f83303fec14352ed718" Oct 08 18:21:51 crc kubenswrapper[4988]: E1008 18:21:51.053728 4988 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9fbbeae0409d2e47e8de9a4306f63b0d13acbfd73cd86f83303fec14352ed718\": container with ID starting with 9fbbeae0409d2e47e8de9a4306f63b0d13acbfd73cd86f83303fec14352ed718 not found: ID does not exist" containerID="9fbbeae0409d2e47e8de9a4306f63b0d13acbfd73cd86f83303fec14352ed718" Oct 08 18:21:51 crc kubenswrapper[4988]: I1008 18:21:51.053762 4988 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9fbbeae0409d2e47e8de9a4306f63b0d13acbfd73cd86f83303fec14352ed718"} err="failed to get container status \"9fbbeae0409d2e47e8de9a4306f63b0d13acbfd73cd86f83303fec14352ed718\": rpc error: code = NotFound desc = could not find container \"9fbbeae0409d2e47e8de9a4306f63b0d13acbfd73cd86f83303fec14352ed718\": container with ID starting with 9fbbeae0409d2e47e8de9a4306f63b0d13acbfd73cd86f83303fec14352ed718 not found: ID does not exist" Oct 08 18:21:51 crc kubenswrapper[4988]: I1008 18:21:51.053783 4988 scope.go:117] "RemoveContainer" containerID="55791327a370c0009c523b86b2234e82d3794985f2eaccf1922eb90f46b2b970" Oct 08 18:21:51 crc kubenswrapper[4988]: E1008 18:21:51.054239 4988 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"55791327a370c0009c523b86b2234e82d3794985f2eaccf1922eb90f46b2b970\": container with ID starting with 55791327a370c0009c523b86b2234e82d3794985f2eaccf1922eb90f46b2b970 not found: ID does not exist" containerID="55791327a370c0009c523b86b2234e82d3794985f2eaccf1922eb90f46b2b970" Oct 08 18:21:51 crc kubenswrapper[4988]: I1008 18:21:51.054289 4988 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"55791327a370c0009c523b86b2234e82d3794985f2eaccf1922eb90f46b2b970"} err="failed to get container status \"55791327a370c0009c523b86b2234e82d3794985f2eaccf1922eb90f46b2b970\": rpc error: code = NotFound desc = could not find container \"55791327a370c0009c523b86b2234e82d3794985f2eaccf1922eb90f46b2b970\": container with ID starting with 55791327a370c0009c523b86b2234e82d3794985f2eaccf1922eb90f46b2b970 not found: ID does not exist" Oct 08 18:21:51 crc kubenswrapper[4988]: I1008 18:21:51.054324 4988 scope.go:117] "RemoveContainer" containerID="03406e726f817bcedb841f81959ab12808ba35a530240f9abc5fa5c608b9a6ae" Oct 08 18:21:51 crc kubenswrapper[4988]: E1008 18:21:51.054690 4988 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"03406e726f817bcedb841f81959ab12808ba35a530240f9abc5fa5c608b9a6ae\": container with ID starting with 03406e726f817bcedb841f81959ab12808ba35a530240f9abc5fa5c608b9a6ae not found: ID does not exist" containerID="03406e726f817bcedb841f81959ab12808ba35a530240f9abc5fa5c608b9a6ae" Oct 08 18:21:51 crc kubenswrapper[4988]: I1008 18:21:51.054718 4988 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"03406e726f817bcedb841f81959ab12808ba35a530240f9abc5fa5c608b9a6ae"} err="failed to get container status \"03406e726f817bcedb841f81959ab12808ba35a530240f9abc5fa5c608b9a6ae\": rpc error: code = NotFound desc = could not find container \"03406e726f817bcedb841f81959ab12808ba35a530240f9abc5fa5c608b9a6ae\": container with ID starting with 03406e726f817bcedb841f81959ab12808ba35a530240f9abc5fa5c608b9a6ae not found: ID does not exist" Oct 08 18:21:51 crc kubenswrapper[4988]: I1008 18:21:51.054736 4988 scope.go:117] "RemoveContainer" containerID="48a516f6c153743bd1a21ec6945fb368a37f4d2b9d6a326b68edc096a192ed05" Oct 08 18:21:51 crc kubenswrapper[4988]: E1008 18:21:51.055230 4988 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"48a516f6c153743bd1a21ec6945fb368a37f4d2b9d6a326b68edc096a192ed05\": container with ID starting with 48a516f6c153743bd1a21ec6945fb368a37f4d2b9d6a326b68edc096a192ed05 not found: ID does not exist" containerID="48a516f6c153743bd1a21ec6945fb368a37f4d2b9d6a326b68edc096a192ed05" Oct 08 18:21:51 crc kubenswrapper[4988]: I1008 18:21:51.055290 4988 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"48a516f6c153743bd1a21ec6945fb368a37f4d2b9d6a326b68edc096a192ed05"} err="failed to get container status \"48a516f6c153743bd1a21ec6945fb368a37f4d2b9d6a326b68edc096a192ed05\": rpc error: code = NotFound desc = could not find container \"48a516f6c153743bd1a21ec6945fb368a37f4d2b9d6a326b68edc096a192ed05\": container with ID starting with 48a516f6c153743bd1a21ec6945fb368a37f4d2b9d6a326b68edc096a192ed05 not found: ID does not exist" Oct 08 18:21:51 crc kubenswrapper[4988]: I1008 18:21:51.055340 4988 scope.go:117] "RemoveContainer" containerID="cee5de93bc68345f9b737857f02832f56582d58757c50a8798eec25d19b2c20d" Oct 08 18:21:51 crc kubenswrapper[4988]: E1008 18:21:51.055795 4988 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cee5de93bc68345f9b737857f02832f56582d58757c50a8798eec25d19b2c20d\": container with ID starting with cee5de93bc68345f9b737857f02832f56582d58757c50a8798eec25d19b2c20d not found: ID does not exist" containerID="cee5de93bc68345f9b737857f02832f56582d58757c50a8798eec25d19b2c20d" Oct 08 18:21:51 crc kubenswrapper[4988]: I1008 18:21:51.055832 4988 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cee5de93bc68345f9b737857f02832f56582d58757c50a8798eec25d19b2c20d"} err="failed to get container status \"cee5de93bc68345f9b737857f02832f56582d58757c50a8798eec25d19b2c20d\": rpc error: code = NotFound desc = could not find container \"cee5de93bc68345f9b737857f02832f56582d58757c50a8798eec25d19b2c20d\": container with ID starting with cee5de93bc68345f9b737857f02832f56582d58757c50a8798eec25d19b2c20d not found: ID does not exist" Oct 08 18:21:51 crc kubenswrapper[4988]: I1008 18:21:51.055851 4988 scope.go:117] "RemoveContainer" containerID="d314eb248e90371fed068b1173e26d4c9f9765417a70b341e1a5b136ec17353a" Oct 08 18:21:51 crc kubenswrapper[4988]: E1008 18:21:51.056151 4988 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d314eb248e90371fed068b1173e26d4c9f9765417a70b341e1a5b136ec17353a\": container with ID starting with d314eb248e90371fed068b1173e26d4c9f9765417a70b341e1a5b136ec17353a not found: ID does not exist" containerID="d314eb248e90371fed068b1173e26d4c9f9765417a70b341e1a5b136ec17353a" Oct 08 18:21:51 crc kubenswrapper[4988]: I1008 18:21:51.056196 4988 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d314eb248e90371fed068b1173e26d4c9f9765417a70b341e1a5b136ec17353a"} err="failed to get container status \"d314eb248e90371fed068b1173e26d4c9f9765417a70b341e1a5b136ec17353a\": rpc error: code = NotFound desc = could not find container \"d314eb248e90371fed068b1173e26d4c9f9765417a70b341e1a5b136ec17353a\": container with ID starting with d314eb248e90371fed068b1173e26d4c9f9765417a70b341e1a5b136ec17353a not found: ID does not exist" Oct 08 18:21:51 crc kubenswrapper[4988]: I1008 18:21:51.056229 4988 scope.go:117] "RemoveContainer" containerID="befe684cea9bd6dfda3ec8e85bca431c26bd8f82823b0a69dc0d04319a3ca385" Oct 08 18:21:51 crc kubenswrapper[4988]: E1008 18:21:51.056647 4988 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"befe684cea9bd6dfda3ec8e85bca431c26bd8f82823b0a69dc0d04319a3ca385\": container with ID starting with befe684cea9bd6dfda3ec8e85bca431c26bd8f82823b0a69dc0d04319a3ca385 not found: ID does not exist" containerID="befe684cea9bd6dfda3ec8e85bca431c26bd8f82823b0a69dc0d04319a3ca385" Oct 08 18:21:51 crc kubenswrapper[4988]: I1008 18:21:51.056682 4988 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"befe684cea9bd6dfda3ec8e85bca431c26bd8f82823b0a69dc0d04319a3ca385"} err="failed to get container status \"befe684cea9bd6dfda3ec8e85bca431c26bd8f82823b0a69dc0d04319a3ca385\": rpc error: code = NotFound desc = could not find container \"befe684cea9bd6dfda3ec8e85bca431c26bd8f82823b0a69dc0d04319a3ca385\": container with ID starting with befe684cea9bd6dfda3ec8e85bca431c26bd8f82823b0a69dc0d04319a3ca385 not found: ID does not exist" Oct 08 18:21:51 crc kubenswrapper[4988]: I1008 18:21:51.056706 4988 scope.go:117] "RemoveContainer" containerID="4987245ff838ca8d4b8c31a08147c052d23bdad26126fcb46367d4b5d4197fb2" Oct 08 18:21:51 crc kubenswrapper[4988]: I1008 18:21:51.057008 4988 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4987245ff838ca8d4b8c31a08147c052d23bdad26126fcb46367d4b5d4197fb2"} err="failed to get container status \"4987245ff838ca8d4b8c31a08147c052d23bdad26126fcb46367d4b5d4197fb2\": rpc error: code = NotFound desc = could not find container \"4987245ff838ca8d4b8c31a08147c052d23bdad26126fcb46367d4b5d4197fb2\": container with ID starting with 4987245ff838ca8d4b8c31a08147c052d23bdad26126fcb46367d4b5d4197fb2 not found: ID does not exist" Oct 08 18:21:51 crc kubenswrapper[4988]: I1008 18:21:51.057034 4988 scope.go:117] "RemoveContainer" containerID="763353df23d26dc093a685af435811240b5ee772d7279ee9a6d2c7a48c19cd30" Oct 08 18:21:51 crc kubenswrapper[4988]: I1008 18:21:51.057417 4988 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"763353df23d26dc093a685af435811240b5ee772d7279ee9a6d2c7a48c19cd30"} err="failed to get container status \"763353df23d26dc093a685af435811240b5ee772d7279ee9a6d2c7a48c19cd30\": rpc error: code = NotFound desc = could not find container \"763353df23d26dc093a685af435811240b5ee772d7279ee9a6d2c7a48c19cd30\": container with ID starting with 763353df23d26dc093a685af435811240b5ee772d7279ee9a6d2c7a48c19cd30 not found: ID does not exist" Oct 08 18:21:51 crc kubenswrapper[4988]: I1008 18:21:51.057445 4988 scope.go:117] "RemoveContainer" containerID="74fcb812b84291763cc6e5ccb20c12aaad580ad968dda4734a0bbfe33d2770ca" Oct 08 18:21:51 crc kubenswrapper[4988]: I1008 18:21:51.057823 4988 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"74fcb812b84291763cc6e5ccb20c12aaad580ad968dda4734a0bbfe33d2770ca"} err="failed to get container status \"74fcb812b84291763cc6e5ccb20c12aaad580ad968dda4734a0bbfe33d2770ca\": rpc error: code = NotFound desc = could not find container \"74fcb812b84291763cc6e5ccb20c12aaad580ad968dda4734a0bbfe33d2770ca\": container with ID starting with 74fcb812b84291763cc6e5ccb20c12aaad580ad968dda4734a0bbfe33d2770ca not found: ID does not exist" Oct 08 18:21:51 crc kubenswrapper[4988]: I1008 18:21:51.057847 4988 scope.go:117] "RemoveContainer" containerID="9fbbeae0409d2e47e8de9a4306f63b0d13acbfd73cd86f83303fec14352ed718" Oct 08 18:21:51 crc kubenswrapper[4988]: I1008 18:21:51.058134 4988 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9fbbeae0409d2e47e8de9a4306f63b0d13acbfd73cd86f83303fec14352ed718"} err="failed to get container status \"9fbbeae0409d2e47e8de9a4306f63b0d13acbfd73cd86f83303fec14352ed718\": rpc error: code = NotFound desc = could not find container \"9fbbeae0409d2e47e8de9a4306f63b0d13acbfd73cd86f83303fec14352ed718\": container with ID starting with 9fbbeae0409d2e47e8de9a4306f63b0d13acbfd73cd86f83303fec14352ed718 not found: ID does not exist" Oct 08 18:21:51 crc kubenswrapper[4988]: I1008 18:21:51.058172 4988 scope.go:117] "RemoveContainer" containerID="55791327a370c0009c523b86b2234e82d3794985f2eaccf1922eb90f46b2b970" Oct 08 18:21:51 crc kubenswrapper[4988]: I1008 18:21:51.058478 4988 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"55791327a370c0009c523b86b2234e82d3794985f2eaccf1922eb90f46b2b970"} err="failed to get container status \"55791327a370c0009c523b86b2234e82d3794985f2eaccf1922eb90f46b2b970\": rpc error: code = NotFound desc = could not find container \"55791327a370c0009c523b86b2234e82d3794985f2eaccf1922eb90f46b2b970\": container with ID starting with 55791327a370c0009c523b86b2234e82d3794985f2eaccf1922eb90f46b2b970 not found: ID does not exist" Oct 08 18:21:51 crc kubenswrapper[4988]: I1008 18:21:51.058507 4988 scope.go:117] "RemoveContainer" containerID="03406e726f817bcedb841f81959ab12808ba35a530240f9abc5fa5c608b9a6ae" Oct 08 18:21:51 crc kubenswrapper[4988]: I1008 18:21:51.058795 4988 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"03406e726f817bcedb841f81959ab12808ba35a530240f9abc5fa5c608b9a6ae"} err="failed to get container status \"03406e726f817bcedb841f81959ab12808ba35a530240f9abc5fa5c608b9a6ae\": rpc error: code = NotFound desc = could not find container \"03406e726f817bcedb841f81959ab12808ba35a530240f9abc5fa5c608b9a6ae\": container with ID starting with 03406e726f817bcedb841f81959ab12808ba35a530240f9abc5fa5c608b9a6ae not found: ID does not exist" Oct 08 18:21:51 crc kubenswrapper[4988]: I1008 18:21:51.058815 4988 scope.go:117] "RemoveContainer" containerID="48a516f6c153743bd1a21ec6945fb368a37f4d2b9d6a326b68edc096a192ed05" Oct 08 18:21:51 crc kubenswrapper[4988]: I1008 18:21:51.059085 4988 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"48a516f6c153743bd1a21ec6945fb368a37f4d2b9d6a326b68edc096a192ed05"} err="failed to get container status \"48a516f6c153743bd1a21ec6945fb368a37f4d2b9d6a326b68edc096a192ed05\": rpc error: code = NotFound desc = could not find container \"48a516f6c153743bd1a21ec6945fb368a37f4d2b9d6a326b68edc096a192ed05\": container with ID starting with 48a516f6c153743bd1a21ec6945fb368a37f4d2b9d6a326b68edc096a192ed05 not found: ID does not exist" Oct 08 18:21:51 crc kubenswrapper[4988]: I1008 18:21:51.059116 4988 scope.go:117] "RemoveContainer" containerID="cee5de93bc68345f9b737857f02832f56582d58757c50a8798eec25d19b2c20d" Oct 08 18:21:51 crc kubenswrapper[4988]: I1008 18:21:51.059369 4988 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cee5de93bc68345f9b737857f02832f56582d58757c50a8798eec25d19b2c20d"} err="failed to get container status \"cee5de93bc68345f9b737857f02832f56582d58757c50a8798eec25d19b2c20d\": rpc error: code = NotFound desc = could not find container \"cee5de93bc68345f9b737857f02832f56582d58757c50a8798eec25d19b2c20d\": container with ID starting with cee5de93bc68345f9b737857f02832f56582d58757c50a8798eec25d19b2c20d not found: ID does not exist" Oct 08 18:21:51 crc kubenswrapper[4988]: I1008 18:21:51.059412 4988 scope.go:117] "RemoveContainer" containerID="d314eb248e90371fed068b1173e26d4c9f9765417a70b341e1a5b136ec17353a" Oct 08 18:21:51 crc kubenswrapper[4988]: I1008 18:21:51.059662 4988 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d314eb248e90371fed068b1173e26d4c9f9765417a70b341e1a5b136ec17353a"} err="failed to get container status \"d314eb248e90371fed068b1173e26d4c9f9765417a70b341e1a5b136ec17353a\": rpc error: code = NotFound desc = could not find container \"d314eb248e90371fed068b1173e26d4c9f9765417a70b341e1a5b136ec17353a\": container with ID starting with d314eb248e90371fed068b1173e26d4c9f9765417a70b341e1a5b136ec17353a not found: ID does not exist" Oct 08 18:21:51 crc kubenswrapper[4988]: I1008 18:21:51.059687 4988 scope.go:117] "RemoveContainer" containerID="befe684cea9bd6dfda3ec8e85bca431c26bd8f82823b0a69dc0d04319a3ca385" Oct 08 18:21:51 crc kubenswrapper[4988]: I1008 18:21:51.059966 4988 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"befe684cea9bd6dfda3ec8e85bca431c26bd8f82823b0a69dc0d04319a3ca385"} err="failed to get container status \"befe684cea9bd6dfda3ec8e85bca431c26bd8f82823b0a69dc0d04319a3ca385\": rpc error: code = NotFound desc = could not find container \"befe684cea9bd6dfda3ec8e85bca431c26bd8f82823b0a69dc0d04319a3ca385\": container with ID starting with befe684cea9bd6dfda3ec8e85bca431c26bd8f82823b0a69dc0d04319a3ca385 not found: ID does not exist" Oct 08 18:21:51 crc kubenswrapper[4988]: I1008 18:21:51.059992 4988 scope.go:117] "RemoveContainer" containerID="4987245ff838ca8d4b8c31a08147c052d23bdad26126fcb46367d4b5d4197fb2" Oct 08 18:21:51 crc kubenswrapper[4988]: I1008 18:21:51.060276 4988 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4987245ff838ca8d4b8c31a08147c052d23bdad26126fcb46367d4b5d4197fb2"} err="failed to get container status \"4987245ff838ca8d4b8c31a08147c052d23bdad26126fcb46367d4b5d4197fb2\": rpc error: code = NotFound desc = could not find container \"4987245ff838ca8d4b8c31a08147c052d23bdad26126fcb46367d4b5d4197fb2\": container with ID starting with 4987245ff838ca8d4b8c31a08147c052d23bdad26126fcb46367d4b5d4197fb2 not found: ID does not exist" Oct 08 18:21:51 crc kubenswrapper[4988]: I1008 18:21:51.060313 4988 scope.go:117] "RemoveContainer" containerID="763353df23d26dc093a685af435811240b5ee772d7279ee9a6d2c7a48c19cd30" Oct 08 18:21:51 crc kubenswrapper[4988]: I1008 18:21:51.060673 4988 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"763353df23d26dc093a685af435811240b5ee772d7279ee9a6d2c7a48c19cd30"} err="failed to get container status \"763353df23d26dc093a685af435811240b5ee772d7279ee9a6d2c7a48c19cd30\": rpc error: code = NotFound desc = could not find container \"763353df23d26dc093a685af435811240b5ee772d7279ee9a6d2c7a48c19cd30\": container with ID starting with 763353df23d26dc093a685af435811240b5ee772d7279ee9a6d2c7a48c19cd30 not found: ID does not exist" Oct 08 18:21:51 crc kubenswrapper[4988]: I1008 18:21:51.060701 4988 scope.go:117] "RemoveContainer" containerID="74fcb812b84291763cc6e5ccb20c12aaad580ad968dda4734a0bbfe33d2770ca" Oct 08 18:21:51 crc kubenswrapper[4988]: I1008 18:21:51.060974 4988 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"74fcb812b84291763cc6e5ccb20c12aaad580ad968dda4734a0bbfe33d2770ca"} err="failed to get container status \"74fcb812b84291763cc6e5ccb20c12aaad580ad968dda4734a0bbfe33d2770ca\": rpc error: code = NotFound desc = could not find container \"74fcb812b84291763cc6e5ccb20c12aaad580ad968dda4734a0bbfe33d2770ca\": container with ID starting with 74fcb812b84291763cc6e5ccb20c12aaad580ad968dda4734a0bbfe33d2770ca not found: ID does not exist" Oct 08 18:21:51 crc kubenswrapper[4988]: I1008 18:21:51.061005 4988 scope.go:117] "RemoveContainer" containerID="9fbbeae0409d2e47e8de9a4306f63b0d13acbfd73cd86f83303fec14352ed718" Oct 08 18:21:51 crc kubenswrapper[4988]: I1008 18:21:51.061317 4988 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9fbbeae0409d2e47e8de9a4306f63b0d13acbfd73cd86f83303fec14352ed718"} err="failed to get container status \"9fbbeae0409d2e47e8de9a4306f63b0d13acbfd73cd86f83303fec14352ed718\": rpc error: code = NotFound desc = could not find container \"9fbbeae0409d2e47e8de9a4306f63b0d13acbfd73cd86f83303fec14352ed718\": container with ID starting with 9fbbeae0409d2e47e8de9a4306f63b0d13acbfd73cd86f83303fec14352ed718 not found: ID does not exist" Oct 08 18:21:51 crc kubenswrapper[4988]: I1008 18:21:51.061349 4988 scope.go:117] "RemoveContainer" containerID="55791327a370c0009c523b86b2234e82d3794985f2eaccf1922eb90f46b2b970" Oct 08 18:21:51 crc kubenswrapper[4988]: I1008 18:21:51.061719 4988 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"55791327a370c0009c523b86b2234e82d3794985f2eaccf1922eb90f46b2b970"} err="failed to get container status \"55791327a370c0009c523b86b2234e82d3794985f2eaccf1922eb90f46b2b970\": rpc error: code = NotFound desc = could not find container \"55791327a370c0009c523b86b2234e82d3794985f2eaccf1922eb90f46b2b970\": container with ID starting with 55791327a370c0009c523b86b2234e82d3794985f2eaccf1922eb90f46b2b970 not found: ID does not exist" Oct 08 18:21:51 crc kubenswrapper[4988]: I1008 18:21:51.061746 4988 scope.go:117] "RemoveContainer" containerID="03406e726f817bcedb841f81959ab12808ba35a530240f9abc5fa5c608b9a6ae" Oct 08 18:21:51 crc kubenswrapper[4988]: I1008 18:21:51.062023 4988 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"03406e726f817bcedb841f81959ab12808ba35a530240f9abc5fa5c608b9a6ae"} err="failed to get container status \"03406e726f817bcedb841f81959ab12808ba35a530240f9abc5fa5c608b9a6ae\": rpc error: code = NotFound desc = could not find container \"03406e726f817bcedb841f81959ab12808ba35a530240f9abc5fa5c608b9a6ae\": container with ID starting with 03406e726f817bcedb841f81959ab12808ba35a530240f9abc5fa5c608b9a6ae not found: ID does not exist" Oct 08 18:21:51 crc kubenswrapper[4988]: I1008 18:21:51.062055 4988 scope.go:117] "RemoveContainer" containerID="48a516f6c153743bd1a21ec6945fb368a37f4d2b9d6a326b68edc096a192ed05" Oct 08 18:21:51 crc kubenswrapper[4988]: I1008 18:21:51.062361 4988 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"48a516f6c153743bd1a21ec6945fb368a37f4d2b9d6a326b68edc096a192ed05"} err="failed to get container status \"48a516f6c153743bd1a21ec6945fb368a37f4d2b9d6a326b68edc096a192ed05\": rpc error: code = NotFound desc = could not find container \"48a516f6c153743bd1a21ec6945fb368a37f4d2b9d6a326b68edc096a192ed05\": container with ID starting with 48a516f6c153743bd1a21ec6945fb368a37f4d2b9d6a326b68edc096a192ed05 not found: ID does not exist" Oct 08 18:21:51 crc kubenswrapper[4988]: I1008 18:21:51.062407 4988 scope.go:117] "RemoveContainer" containerID="cee5de93bc68345f9b737857f02832f56582d58757c50a8798eec25d19b2c20d" Oct 08 18:21:51 crc kubenswrapper[4988]: I1008 18:21:51.062702 4988 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cee5de93bc68345f9b737857f02832f56582d58757c50a8798eec25d19b2c20d"} err="failed to get container status \"cee5de93bc68345f9b737857f02832f56582d58757c50a8798eec25d19b2c20d\": rpc error: code = NotFound desc = could not find container \"cee5de93bc68345f9b737857f02832f56582d58757c50a8798eec25d19b2c20d\": container with ID starting with cee5de93bc68345f9b737857f02832f56582d58757c50a8798eec25d19b2c20d not found: ID does not exist" Oct 08 18:21:51 crc kubenswrapper[4988]: I1008 18:21:51.062738 4988 scope.go:117] "RemoveContainer" containerID="d314eb248e90371fed068b1173e26d4c9f9765417a70b341e1a5b136ec17353a" Oct 08 18:21:51 crc kubenswrapper[4988]: I1008 18:21:51.063098 4988 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d314eb248e90371fed068b1173e26d4c9f9765417a70b341e1a5b136ec17353a"} err="failed to get container status \"d314eb248e90371fed068b1173e26d4c9f9765417a70b341e1a5b136ec17353a\": rpc error: code = NotFound desc = could not find container \"d314eb248e90371fed068b1173e26d4c9f9765417a70b341e1a5b136ec17353a\": container with ID starting with d314eb248e90371fed068b1173e26d4c9f9765417a70b341e1a5b136ec17353a not found: ID does not exist" Oct 08 18:21:51 crc kubenswrapper[4988]: I1008 18:21:51.063116 4988 scope.go:117] "RemoveContainer" containerID="befe684cea9bd6dfda3ec8e85bca431c26bd8f82823b0a69dc0d04319a3ca385" Oct 08 18:21:51 crc kubenswrapper[4988]: I1008 18:21:51.063352 4988 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"befe684cea9bd6dfda3ec8e85bca431c26bd8f82823b0a69dc0d04319a3ca385"} err="failed to get container status \"befe684cea9bd6dfda3ec8e85bca431c26bd8f82823b0a69dc0d04319a3ca385\": rpc error: code = NotFound desc = could not find container \"befe684cea9bd6dfda3ec8e85bca431c26bd8f82823b0a69dc0d04319a3ca385\": container with ID starting with befe684cea9bd6dfda3ec8e85bca431c26bd8f82823b0a69dc0d04319a3ca385 not found: ID does not exist" Oct 08 18:21:51 crc kubenswrapper[4988]: I1008 18:21:51.063418 4988 scope.go:117] "RemoveContainer" containerID="4987245ff838ca8d4b8c31a08147c052d23bdad26126fcb46367d4b5d4197fb2" Oct 08 18:21:51 crc kubenswrapper[4988]: I1008 18:21:51.063672 4988 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4987245ff838ca8d4b8c31a08147c052d23bdad26126fcb46367d4b5d4197fb2"} err="failed to get container status \"4987245ff838ca8d4b8c31a08147c052d23bdad26126fcb46367d4b5d4197fb2\": rpc error: code = NotFound desc = could not find container \"4987245ff838ca8d4b8c31a08147c052d23bdad26126fcb46367d4b5d4197fb2\": container with ID starting with 4987245ff838ca8d4b8c31a08147c052d23bdad26126fcb46367d4b5d4197fb2 not found: ID does not exist" Oct 08 18:21:51 crc kubenswrapper[4988]: I1008 18:21:51.063694 4988 scope.go:117] "RemoveContainer" containerID="763353df23d26dc093a685af435811240b5ee772d7279ee9a6d2c7a48c19cd30" Oct 08 18:21:51 crc kubenswrapper[4988]: I1008 18:21:51.064092 4988 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"763353df23d26dc093a685af435811240b5ee772d7279ee9a6d2c7a48c19cd30"} err="failed to get container status \"763353df23d26dc093a685af435811240b5ee772d7279ee9a6d2c7a48c19cd30\": rpc error: code = NotFound desc = could not find container \"763353df23d26dc093a685af435811240b5ee772d7279ee9a6d2c7a48c19cd30\": container with ID starting with 763353df23d26dc093a685af435811240b5ee772d7279ee9a6d2c7a48c19cd30 not found: ID does not exist" Oct 08 18:21:51 crc kubenswrapper[4988]: I1008 18:21:51.064126 4988 scope.go:117] "RemoveContainer" containerID="74fcb812b84291763cc6e5ccb20c12aaad580ad968dda4734a0bbfe33d2770ca" Oct 08 18:21:51 crc kubenswrapper[4988]: I1008 18:21:51.064446 4988 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"74fcb812b84291763cc6e5ccb20c12aaad580ad968dda4734a0bbfe33d2770ca"} err="failed to get container status \"74fcb812b84291763cc6e5ccb20c12aaad580ad968dda4734a0bbfe33d2770ca\": rpc error: code = NotFound desc = could not find container \"74fcb812b84291763cc6e5ccb20c12aaad580ad968dda4734a0bbfe33d2770ca\": container with ID starting with 74fcb812b84291763cc6e5ccb20c12aaad580ad968dda4734a0bbfe33d2770ca not found: ID does not exist" Oct 08 18:21:51 crc kubenswrapper[4988]: I1008 18:21:51.064467 4988 scope.go:117] "RemoveContainer" containerID="9fbbeae0409d2e47e8de9a4306f63b0d13acbfd73cd86f83303fec14352ed718" Oct 08 18:21:51 crc kubenswrapper[4988]: I1008 18:21:51.064746 4988 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9fbbeae0409d2e47e8de9a4306f63b0d13acbfd73cd86f83303fec14352ed718"} err="failed to get container status \"9fbbeae0409d2e47e8de9a4306f63b0d13acbfd73cd86f83303fec14352ed718\": rpc error: code = NotFound desc = could not find container \"9fbbeae0409d2e47e8de9a4306f63b0d13acbfd73cd86f83303fec14352ed718\": container with ID starting with 9fbbeae0409d2e47e8de9a4306f63b0d13acbfd73cd86f83303fec14352ed718 not found: ID does not exist" Oct 08 18:21:51 crc kubenswrapper[4988]: I1008 18:21:51.064763 4988 scope.go:117] "RemoveContainer" containerID="55791327a370c0009c523b86b2234e82d3794985f2eaccf1922eb90f46b2b970" Oct 08 18:21:51 crc kubenswrapper[4988]: I1008 18:21:51.065008 4988 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"55791327a370c0009c523b86b2234e82d3794985f2eaccf1922eb90f46b2b970"} err="failed to get container status \"55791327a370c0009c523b86b2234e82d3794985f2eaccf1922eb90f46b2b970\": rpc error: code = NotFound desc = could not find container \"55791327a370c0009c523b86b2234e82d3794985f2eaccf1922eb90f46b2b970\": container with ID starting with 55791327a370c0009c523b86b2234e82d3794985f2eaccf1922eb90f46b2b970 not found: ID does not exist" Oct 08 18:21:51 crc kubenswrapper[4988]: I1008 18:21:51.065039 4988 scope.go:117] "RemoveContainer" containerID="03406e726f817bcedb841f81959ab12808ba35a530240f9abc5fa5c608b9a6ae" Oct 08 18:21:51 crc kubenswrapper[4988]: I1008 18:21:51.065572 4988 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"03406e726f817bcedb841f81959ab12808ba35a530240f9abc5fa5c608b9a6ae"} err="failed to get container status \"03406e726f817bcedb841f81959ab12808ba35a530240f9abc5fa5c608b9a6ae\": rpc error: code = NotFound desc = could not find container \"03406e726f817bcedb841f81959ab12808ba35a530240f9abc5fa5c608b9a6ae\": container with ID starting with 03406e726f817bcedb841f81959ab12808ba35a530240f9abc5fa5c608b9a6ae not found: ID does not exist" Oct 08 18:21:51 crc kubenswrapper[4988]: I1008 18:21:51.065729 4988 scope.go:117] "RemoveContainer" containerID="48a516f6c153743bd1a21ec6945fb368a37f4d2b9d6a326b68edc096a192ed05" Oct 08 18:21:51 crc kubenswrapper[4988]: I1008 18:21:51.066015 4988 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"48a516f6c153743bd1a21ec6945fb368a37f4d2b9d6a326b68edc096a192ed05"} err="failed to get container status \"48a516f6c153743bd1a21ec6945fb368a37f4d2b9d6a326b68edc096a192ed05\": rpc error: code = NotFound desc = could not find container \"48a516f6c153743bd1a21ec6945fb368a37f4d2b9d6a326b68edc096a192ed05\": container with ID starting with 48a516f6c153743bd1a21ec6945fb368a37f4d2b9d6a326b68edc096a192ed05 not found: ID does not exist" Oct 08 18:21:51 crc kubenswrapper[4988]: I1008 18:21:51.066032 4988 scope.go:117] "RemoveContainer" containerID="cee5de93bc68345f9b737857f02832f56582d58757c50a8798eec25d19b2c20d" Oct 08 18:21:51 crc kubenswrapper[4988]: I1008 18:21:51.066309 4988 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cee5de93bc68345f9b737857f02832f56582d58757c50a8798eec25d19b2c20d"} err="failed to get container status \"cee5de93bc68345f9b737857f02832f56582d58757c50a8798eec25d19b2c20d\": rpc error: code = NotFound desc = could not find container \"cee5de93bc68345f9b737857f02832f56582d58757c50a8798eec25d19b2c20d\": container with ID starting with cee5de93bc68345f9b737857f02832f56582d58757c50a8798eec25d19b2c20d not found: ID does not exist" Oct 08 18:21:51 crc kubenswrapper[4988]: I1008 18:21:51.066350 4988 scope.go:117] "RemoveContainer" containerID="d314eb248e90371fed068b1173e26d4c9f9765417a70b341e1a5b136ec17353a" Oct 08 18:21:51 crc kubenswrapper[4988]: I1008 18:21:51.066772 4988 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d314eb248e90371fed068b1173e26d4c9f9765417a70b341e1a5b136ec17353a"} err="failed to get container status \"d314eb248e90371fed068b1173e26d4c9f9765417a70b341e1a5b136ec17353a\": rpc error: code = NotFound desc = could not find container \"d314eb248e90371fed068b1173e26d4c9f9765417a70b341e1a5b136ec17353a\": container with ID starting with d314eb248e90371fed068b1173e26d4c9f9765417a70b341e1a5b136ec17353a not found: ID does not exist" Oct 08 18:21:51 crc kubenswrapper[4988]: I1008 18:21:51.066792 4988 scope.go:117] "RemoveContainer" containerID="befe684cea9bd6dfda3ec8e85bca431c26bd8f82823b0a69dc0d04319a3ca385" Oct 08 18:21:51 crc kubenswrapper[4988]: I1008 18:21:51.067048 4988 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"befe684cea9bd6dfda3ec8e85bca431c26bd8f82823b0a69dc0d04319a3ca385"} err="failed to get container status \"befe684cea9bd6dfda3ec8e85bca431c26bd8f82823b0a69dc0d04319a3ca385\": rpc error: code = NotFound desc = could not find container \"befe684cea9bd6dfda3ec8e85bca431c26bd8f82823b0a69dc0d04319a3ca385\": container with ID starting with befe684cea9bd6dfda3ec8e85bca431c26bd8f82823b0a69dc0d04319a3ca385 not found: ID does not exist" Oct 08 18:21:51 crc kubenswrapper[4988]: I1008 18:21:51.067091 4988 scope.go:117] "RemoveContainer" containerID="4987245ff838ca8d4b8c31a08147c052d23bdad26126fcb46367d4b5d4197fb2" Oct 08 18:21:51 crc kubenswrapper[4988]: I1008 18:21:51.067370 4988 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4987245ff838ca8d4b8c31a08147c052d23bdad26126fcb46367d4b5d4197fb2"} err="failed to get container status \"4987245ff838ca8d4b8c31a08147c052d23bdad26126fcb46367d4b5d4197fb2\": rpc error: code = NotFound desc = could not find container \"4987245ff838ca8d4b8c31a08147c052d23bdad26126fcb46367d4b5d4197fb2\": container with ID starting with 4987245ff838ca8d4b8c31a08147c052d23bdad26126fcb46367d4b5d4197fb2 not found: ID does not exist" Oct 08 18:21:51 crc kubenswrapper[4988]: I1008 18:21:51.088808 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-74ftp" Oct 08 18:21:51 crc kubenswrapper[4988]: I1008 18:21:51.247600 4988 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f010a463-69f2-4082-acc3-161899a939c3" path="/var/lib/kubelet/pods/f010a463-69f2-4082-acc3-161899a939c3/volumes" Oct 08 18:21:51 crc kubenswrapper[4988]: I1008 18:21:51.896522 4988 generic.go:334] "Generic (PLEG): container finished" podID="8eee6858-0c8b-47a7-847b-ef454c951f6d" containerID="0a2ac9a049a60759294f063fd9ef7f75914896d942448237372e5d551e48c8db" exitCode=0 Oct 08 18:21:51 crc kubenswrapper[4988]: I1008 18:21:51.896650 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-74ftp" event={"ID":"8eee6858-0c8b-47a7-847b-ef454c951f6d","Type":"ContainerDied","Data":"0a2ac9a049a60759294f063fd9ef7f75914896d942448237372e5d551e48c8db"} Oct 08 18:21:51 crc kubenswrapper[4988]: I1008 18:21:51.897059 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-74ftp" event={"ID":"8eee6858-0c8b-47a7-847b-ef454c951f6d","Type":"ContainerStarted","Data":"7c8528f39ae6116352fd71879fb9583111cce5ac25be9e6a35c9b2da140e2713"} Oct 08 18:21:52 crc kubenswrapper[4988]: I1008 18:21:52.906549 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-74ftp" event={"ID":"8eee6858-0c8b-47a7-847b-ef454c951f6d","Type":"ContainerStarted","Data":"75f4b7d65e1a1bce7cf607f07523e09f78775f8d15a57189d90ad2c7acc52d4f"} Oct 08 18:21:52 crc kubenswrapper[4988]: I1008 18:21:52.907138 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-74ftp" event={"ID":"8eee6858-0c8b-47a7-847b-ef454c951f6d","Type":"ContainerStarted","Data":"8528f1f84a06066163a095f6b6b0307d64fe9df5785be3a30e549bc7fd2df375"} Oct 08 18:21:52 crc kubenswrapper[4988]: I1008 18:21:52.907161 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-74ftp" event={"ID":"8eee6858-0c8b-47a7-847b-ef454c951f6d","Type":"ContainerStarted","Data":"c60705c940ec085e9c757f32c475870bb371716e5c87591329b54e47376700a4"} Oct 08 18:21:52 crc kubenswrapper[4988]: I1008 18:21:52.907180 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-74ftp" event={"ID":"8eee6858-0c8b-47a7-847b-ef454c951f6d","Type":"ContainerStarted","Data":"3834c5b24928ba17e2065a40dd159622796e695ddc1ad9fa204f407d8952538b"} Oct 08 18:21:52 crc kubenswrapper[4988]: I1008 18:21:52.907197 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-74ftp" event={"ID":"8eee6858-0c8b-47a7-847b-ef454c951f6d","Type":"ContainerStarted","Data":"e489ddb90870de46c496fd48ade448756bbae271d0bb2c7f48a203984b25b644"} Oct 08 18:21:52 crc kubenswrapper[4988]: I1008 18:21:52.907214 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-74ftp" event={"ID":"8eee6858-0c8b-47a7-847b-ef454c951f6d","Type":"ContainerStarted","Data":"d3a920de7f95d5d703705d2da19f354b9f8cd4c17d1ee492e4108aaecad61427"} Oct 08 18:21:55 crc kubenswrapper[4988]: I1008 18:21:55.474983 4988 scope.go:117] "RemoveContainer" containerID="c38c1b52127a970f792ef09ef1b217c979d94f7f74cc3c27e38949a204ad831f" Oct 08 18:21:55 crc kubenswrapper[4988]: I1008 18:21:55.934246 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-74ftp" event={"ID":"8eee6858-0c8b-47a7-847b-ef454c951f6d","Type":"ContainerStarted","Data":"3a88a7714b7e0873ef98724ae58aaf67cdad403a31eed6677eca79b3684d16c6"} Oct 08 18:21:55 crc kubenswrapper[4988]: I1008 18:21:55.937309 4988 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-h9jwp_4490d846-ca32-4cd5-9202-87cc6d499fb4/kube-multus/2.log" Oct 08 18:21:57 crc kubenswrapper[4988]: I1008 18:21:57.959899 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-74ftp" event={"ID":"8eee6858-0c8b-47a7-847b-ef454c951f6d","Type":"ContainerStarted","Data":"35f4cbd4634c49c03fee60192ece1e10a5e4ade5f7f642946847515839fab4c4"} Oct 08 18:21:57 crc kubenswrapper[4988]: I1008 18:21:57.960721 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-74ftp" Oct 08 18:21:57 crc kubenswrapper[4988]: I1008 18:21:57.960757 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-74ftp" Oct 08 18:21:57 crc kubenswrapper[4988]: I1008 18:21:57.960782 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-74ftp" Oct 08 18:21:58 crc kubenswrapper[4988]: I1008 18:21:58.011368 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-74ftp" Oct 08 18:21:58 crc kubenswrapper[4988]: I1008 18:21:58.018916 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-74ftp" Oct 08 18:21:58 crc kubenswrapper[4988]: I1008 18:21:58.020184 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-74ftp" podStartSLOduration=8.020162361 podStartE2EDuration="8.020162361s" podCreationTimestamp="2025-10-08 18:21:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 18:21:58.01769405 +0000 UTC m=+663.467536840" watchObservedRunningTime="2025-10-08 18:21:58.020162361 +0000 UTC m=+663.470005181" Oct 08 18:21:59 crc kubenswrapper[4988]: I1008 18:21:59.427633 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["crc-storage/crc-storage-crc-4s445"] Oct 08 18:21:59 crc kubenswrapper[4988]: I1008 18:21:59.428430 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-4s445" Oct 08 18:21:59 crc kubenswrapper[4988]: I1008 18:21:59.430321 4988 reflector.go:368] Caches populated for *v1.Secret from object-"crc-storage"/"crc-storage-dockercfg-frtmh" Oct 08 18:21:59 crc kubenswrapper[4988]: I1008 18:21:59.430801 4988 reflector.go:368] Caches populated for *v1.ConfigMap from object-"crc-storage"/"crc-storage" Oct 08 18:21:59 crc kubenswrapper[4988]: I1008 18:21:59.432785 4988 reflector.go:368] Caches populated for *v1.ConfigMap from object-"crc-storage"/"openshift-service-ca.crt" Oct 08 18:21:59 crc kubenswrapper[4988]: I1008 18:21:59.436805 4988 reflector.go:368] Caches populated for *v1.ConfigMap from object-"crc-storage"/"kube-root-ca.crt" Oct 08 18:21:59 crc kubenswrapper[4988]: I1008 18:21:59.437309 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["crc-storage/crc-storage-crc-4s445"] Oct 08 18:21:59 crc kubenswrapper[4988]: I1008 18:21:59.484241 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/3cf2d054-6c91-4e4c-829b-1e038c477c41-node-mnt\") pod \"crc-storage-crc-4s445\" (UID: \"3cf2d054-6c91-4e4c-829b-1e038c477c41\") " pod="crc-storage/crc-storage-crc-4s445" Oct 08 18:21:59 crc kubenswrapper[4988]: I1008 18:21:59.484509 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7whnj\" (UniqueName: \"kubernetes.io/projected/3cf2d054-6c91-4e4c-829b-1e038c477c41-kube-api-access-7whnj\") pod \"crc-storage-crc-4s445\" (UID: \"3cf2d054-6c91-4e4c-829b-1e038c477c41\") " pod="crc-storage/crc-storage-crc-4s445" Oct 08 18:21:59 crc kubenswrapper[4988]: I1008 18:21:59.484627 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/3cf2d054-6c91-4e4c-829b-1e038c477c41-crc-storage\") pod \"crc-storage-crc-4s445\" (UID: \"3cf2d054-6c91-4e4c-829b-1e038c477c41\") " pod="crc-storage/crc-storage-crc-4s445" Oct 08 18:21:59 crc kubenswrapper[4988]: I1008 18:21:59.585603 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7whnj\" (UniqueName: \"kubernetes.io/projected/3cf2d054-6c91-4e4c-829b-1e038c477c41-kube-api-access-7whnj\") pod \"crc-storage-crc-4s445\" (UID: \"3cf2d054-6c91-4e4c-829b-1e038c477c41\") " pod="crc-storage/crc-storage-crc-4s445" Oct 08 18:21:59 crc kubenswrapper[4988]: I1008 18:21:59.585657 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/3cf2d054-6c91-4e4c-829b-1e038c477c41-crc-storage\") pod \"crc-storage-crc-4s445\" (UID: \"3cf2d054-6c91-4e4c-829b-1e038c477c41\") " pod="crc-storage/crc-storage-crc-4s445" Oct 08 18:21:59 crc kubenswrapper[4988]: I1008 18:21:59.585688 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/3cf2d054-6c91-4e4c-829b-1e038c477c41-node-mnt\") pod \"crc-storage-crc-4s445\" (UID: \"3cf2d054-6c91-4e4c-829b-1e038c477c41\") " pod="crc-storage/crc-storage-crc-4s445" Oct 08 18:21:59 crc kubenswrapper[4988]: I1008 18:21:59.585927 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/3cf2d054-6c91-4e4c-829b-1e038c477c41-node-mnt\") pod \"crc-storage-crc-4s445\" (UID: \"3cf2d054-6c91-4e4c-829b-1e038c477c41\") " pod="crc-storage/crc-storage-crc-4s445" Oct 08 18:21:59 crc kubenswrapper[4988]: I1008 18:21:59.587100 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/3cf2d054-6c91-4e4c-829b-1e038c477c41-crc-storage\") pod \"crc-storage-crc-4s445\" (UID: \"3cf2d054-6c91-4e4c-829b-1e038c477c41\") " pod="crc-storage/crc-storage-crc-4s445" Oct 08 18:21:59 crc kubenswrapper[4988]: I1008 18:21:59.607195 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7whnj\" (UniqueName: \"kubernetes.io/projected/3cf2d054-6c91-4e4c-829b-1e038c477c41-kube-api-access-7whnj\") pod \"crc-storage-crc-4s445\" (UID: \"3cf2d054-6c91-4e4c-829b-1e038c477c41\") " pod="crc-storage/crc-storage-crc-4s445" Oct 08 18:21:59 crc kubenswrapper[4988]: I1008 18:21:59.755645 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-4s445" Oct 08 18:21:59 crc kubenswrapper[4988]: E1008 18:21:59.791438 4988 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_crc-storage-crc-4s445_crc-storage_3cf2d054-6c91-4e4c-829b-1e038c477c41_0(184e231b37ad66ed2e824528f3af7a5edad1aabe8b852b66d185af62e8378de9): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Oct 08 18:21:59 crc kubenswrapper[4988]: E1008 18:21:59.791547 4988 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_crc-storage-crc-4s445_crc-storage_3cf2d054-6c91-4e4c-829b-1e038c477c41_0(184e231b37ad66ed2e824528f3af7a5edad1aabe8b852b66d185af62e8378de9): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="crc-storage/crc-storage-crc-4s445" Oct 08 18:21:59 crc kubenswrapper[4988]: E1008 18:21:59.791607 4988 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_crc-storage-crc-4s445_crc-storage_3cf2d054-6c91-4e4c-829b-1e038c477c41_0(184e231b37ad66ed2e824528f3af7a5edad1aabe8b852b66d185af62e8378de9): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="crc-storage/crc-storage-crc-4s445" Oct 08 18:21:59 crc kubenswrapper[4988]: E1008 18:21:59.791687 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"crc-storage-crc-4s445_crc-storage(3cf2d054-6c91-4e4c-829b-1e038c477c41)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"crc-storage-crc-4s445_crc-storage(3cf2d054-6c91-4e4c-829b-1e038c477c41)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_crc-storage-crc-4s445_crc-storage_3cf2d054-6c91-4e4c-829b-1e038c477c41_0(184e231b37ad66ed2e824528f3af7a5edad1aabe8b852b66d185af62e8378de9): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="crc-storage/crc-storage-crc-4s445" podUID="3cf2d054-6c91-4e4c-829b-1e038c477c41" Oct 08 18:21:59 crc kubenswrapper[4988]: I1008 18:21:59.970809 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-4s445" Oct 08 18:21:59 crc kubenswrapper[4988]: I1008 18:21:59.971365 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-4s445" Oct 08 18:22:00 crc kubenswrapper[4988]: E1008 18:22:00.001788 4988 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_crc-storage-crc-4s445_crc-storage_3cf2d054-6c91-4e4c-829b-1e038c477c41_0(e9c4d2d30764ca140e854949474469de3daa424e7cb166085465d797917d25be): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Oct 08 18:22:00 crc kubenswrapper[4988]: E1008 18:22:00.001911 4988 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_crc-storage-crc-4s445_crc-storage_3cf2d054-6c91-4e4c-829b-1e038c477c41_0(e9c4d2d30764ca140e854949474469de3daa424e7cb166085465d797917d25be): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="crc-storage/crc-storage-crc-4s445" Oct 08 18:22:00 crc kubenswrapper[4988]: E1008 18:22:00.001976 4988 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_crc-storage-crc-4s445_crc-storage_3cf2d054-6c91-4e4c-829b-1e038c477c41_0(e9c4d2d30764ca140e854949474469de3daa424e7cb166085465d797917d25be): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="crc-storage/crc-storage-crc-4s445" Oct 08 18:22:00 crc kubenswrapper[4988]: E1008 18:22:00.002080 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"crc-storage-crc-4s445_crc-storage(3cf2d054-6c91-4e4c-829b-1e038c477c41)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"crc-storage-crc-4s445_crc-storage(3cf2d054-6c91-4e4c-829b-1e038c477c41)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_crc-storage-crc-4s445_crc-storage_3cf2d054-6c91-4e4c-829b-1e038c477c41_0(e9c4d2d30764ca140e854949474469de3daa424e7cb166085465d797917d25be): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="crc-storage/crc-storage-crc-4s445" podUID="3cf2d054-6c91-4e4c-829b-1e038c477c41" Oct 08 18:22:02 crc kubenswrapper[4988]: I1008 18:22:02.238341 4988 scope.go:117] "RemoveContainer" containerID="f62030d24f69a85e1807099969f0021ab1198002c76e11ef3dffadfabff73b49" Oct 08 18:22:02 crc kubenswrapper[4988]: E1008 18:22:02.238667 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-multus pod=multus-h9jwp_openshift-multus(4490d846-ca32-4cd5-9202-87cc6d499fb4)\"" pod="openshift-multus/multus-h9jwp" podUID="4490d846-ca32-4cd5-9202-87cc6d499fb4" Oct 08 18:22:13 crc kubenswrapper[4988]: I1008 18:22:13.237896 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-4s445" Oct 08 18:22:13 crc kubenswrapper[4988]: I1008 18:22:13.239074 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-4s445" Oct 08 18:22:13 crc kubenswrapper[4988]: E1008 18:22:13.278818 4988 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_crc-storage-crc-4s445_crc-storage_3cf2d054-6c91-4e4c-829b-1e038c477c41_0(791e03c95d3507b4e46ca9047dfefbaaf1f2fd7377f2ce283380abea0e1dc2f1): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Oct 08 18:22:13 crc kubenswrapper[4988]: E1008 18:22:13.279204 4988 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_crc-storage-crc-4s445_crc-storage_3cf2d054-6c91-4e4c-829b-1e038c477c41_0(791e03c95d3507b4e46ca9047dfefbaaf1f2fd7377f2ce283380abea0e1dc2f1): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="crc-storage/crc-storage-crc-4s445" Oct 08 18:22:13 crc kubenswrapper[4988]: E1008 18:22:13.279233 4988 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_crc-storage-crc-4s445_crc-storage_3cf2d054-6c91-4e4c-829b-1e038c477c41_0(791e03c95d3507b4e46ca9047dfefbaaf1f2fd7377f2ce283380abea0e1dc2f1): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="crc-storage/crc-storage-crc-4s445" Oct 08 18:22:13 crc kubenswrapper[4988]: E1008 18:22:13.279283 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"crc-storage-crc-4s445_crc-storage(3cf2d054-6c91-4e4c-829b-1e038c477c41)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"crc-storage-crc-4s445_crc-storage(3cf2d054-6c91-4e4c-829b-1e038c477c41)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_crc-storage-crc-4s445_crc-storage_3cf2d054-6c91-4e4c-829b-1e038c477c41_0(791e03c95d3507b4e46ca9047dfefbaaf1f2fd7377f2ce283380abea0e1dc2f1): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="crc-storage/crc-storage-crc-4s445" podUID="3cf2d054-6c91-4e4c-829b-1e038c477c41" Oct 08 18:22:17 crc kubenswrapper[4988]: I1008 18:22:17.238269 4988 scope.go:117] "RemoveContainer" containerID="f62030d24f69a85e1807099969f0021ab1198002c76e11ef3dffadfabff73b49" Oct 08 18:22:18 crc kubenswrapper[4988]: I1008 18:22:18.097745 4988 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-h9jwp_4490d846-ca32-4cd5-9202-87cc6d499fb4/kube-multus/2.log" Oct 08 18:22:18 crc kubenswrapper[4988]: I1008 18:22:18.098761 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-h9jwp" event={"ID":"4490d846-ca32-4cd5-9202-87cc6d499fb4","Type":"ContainerStarted","Data":"ea01612007c7a72669bbe324d4127502fda88c4750c17474b9a5db01d71221ca"} Oct 08 18:22:21 crc kubenswrapper[4988]: I1008 18:22:21.120187 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-74ftp" Oct 08 18:22:25 crc kubenswrapper[4988]: I1008 18:22:25.237580 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-4s445" Oct 08 18:22:25 crc kubenswrapper[4988]: I1008 18:22:25.243752 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-4s445" Oct 08 18:22:25 crc kubenswrapper[4988]: I1008 18:22:25.474026 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["crc-storage/crc-storage-crc-4s445"] Oct 08 18:22:25 crc kubenswrapper[4988]: I1008 18:22:25.483330 4988 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 08 18:22:26 crc kubenswrapper[4988]: I1008 18:22:26.155605 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="crc-storage/crc-storage-crc-4s445" event={"ID":"3cf2d054-6c91-4e4c-829b-1e038c477c41","Type":"ContainerStarted","Data":"ba94e2169136e40eaeee39bf718fa11c4585877e4a3c9f35d01afc7108bfd523"} Oct 08 18:22:28 crc kubenswrapper[4988]: I1008 18:22:28.167952 4988 generic.go:334] "Generic (PLEG): container finished" podID="3cf2d054-6c91-4e4c-829b-1e038c477c41" containerID="2cc88615496f18e32e9c6b61e125c82fc20f3b526f8ead63c6e79da3b05461a8" exitCode=0 Oct 08 18:22:28 crc kubenswrapper[4988]: I1008 18:22:28.168036 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="crc-storage/crc-storage-crc-4s445" event={"ID":"3cf2d054-6c91-4e4c-829b-1e038c477c41","Type":"ContainerDied","Data":"2cc88615496f18e32e9c6b61e125c82fc20f3b526f8ead63c6e79da3b05461a8"} Oct 08 18:22:29 crc kubenswrapper[4988]: I1008 18:22:29.443859 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-4s445" Oct 08 18:22:29 crc kubenswrapper[4988]: I1008 18:22:29.520191 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/3cf2d054-6c91-4e4c-829b-1e038c477c41-node-mnt\") pod \"3cf2d054-6c91-4e4c-829b-1e038c477c41\" (UID: \"3cf2d054-6c91-4e4c-829b-1e038c477c41\") " Oct 08 18:22:29 crc kubenswrapper[4988]: I1008 18:22:29.520312 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/3cf2d054-6c91-4e4c-829b-1e038c477c41-crc-storage\") pod \"3cf2d054-6c91-4e4c-829b-1e038c477c41\" (UID: \"3cf2d054-6c91-4e4c-829b-1e038c477c41\") " Oct 08 18:22:29 crc kubenswrapper[4988]: I1008 18:22:29.520299 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/3cf2d054-6c91-4e4c-829b-1e038c477c41-node-mnt" (OuterVolumeSpecName: "node-mnt") pod "3cf2d054-6c91-4e4c-829b-1e038c477c41" (UID: "3cf2d054-6c91-4e4c-829b-1e038c477c41"). InnerVolumeSpecName "node-mnt". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 08 18:22:29 crc kubenswrapper[4988]: I1008 18:22:29.520443 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7whnj\" (UniqueName: \"kubernetes.io/projected/3cf2d054-6c91-4e4c-829b-1e038c477c41-kube-api-access-7whnj\") pod \"3cf2d054-6c91-4e4c-829b-1e038c477c41\" (UID: \"3cf2d054-6c91-4e4c-829b-1e038c477c41\") " Oct 08 18:22:29 crc kubenswrapper[4988]: I1008 18:22:29.520711 4988 reconciler_common.go:293] "Volume detached for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/3cf2d054-6c91-4e4c-829b-1e038c477c41-node-mnt\") on node \"crc\" DevicePath \"\"" Oct 08 18:22:29 crc kubenswrapper[4988]: I1008 18:22:29.527158 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3cf2d054-6c91-4e4c-829b-1e038c477c41-kube-api-access-7whnj" (OuterVolumeSpecName: "kube-api-access-7whnj") pod "3cf2d054-6c91-4e4c-829b-1e038c477c41" (UID: "3cf2d054-6c91-4e4c-829b-1e038c477c41"). InnerVolumeSpecName "kube-api-access-7whnj". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:22:29 crc kubenswrapper[4988]: I1008 18:22:29.546611 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3cf2d054-6c91-4e4c-829b-1e038c477c41-crc-storage" (OuterVolumeSpecName: "crc-storage") pod "3cf2d054-6c91-4e4c-829b-1e038c477c41" (UID: "3cf2d054-6c91-4e4c-829b-1e038c477c41"). InnerVolumeSpecName "crc-storage". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 18:22:29 crc kubenswrapper[4988]: I1008 18:22:29.622797 4988 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7whnj\" (UniqueName: \"kubernetes.io/projected/3cf2d054-6c91-4e4c-829b-1e038c477c41-kube-api-access-7whnj\") on node \"crc\" DevicePath \"\"" Oct 08 18:22:29 crc kubenswrapper[4988]: I1008 18:22:29.622928 4988 reconciler_common.go:293] "Volume detached for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/3cf2d054-6c91-4e4c-829b-1e038c477c41-crc-storage\") on node \"crc\" DevicePath \"\"" Oct 08 18:22:30 crc kubenswrapper[4988]: I1008 18:22:30.183170 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="crc-storage/crc-storage-crc-4s445" event={"ID":"3cf2d054-6c91-4e4c-829b-1e038c477c41","Type":"ContainerDied","Data":"ba94e2169136e40eaeee39bf718fa11c4585877e4a3c9f35d01afc7108bfd523"} Oct 08 18:22:30 crc kubenswrapper[4988]: I1008 18:22:30.183233 4988 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ba94e2169136e40eaeee39bf718fa11c4585877e4a3c9f35d01afc7108bfd523" Oct 08 18:22:30 crc kubenswrapper[4988]: I1008 18:22:30.183262 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-4s445" Oct 08 18:22:38 crc kubenswrapper[4988]: I1008 18:22:38.002775 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c55jx7"] Oct 08 18:22:38 crc kubenswrapper[4988]: E1008 18:22:38.003590 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3cf2d054-6c91-4e4c-829b-1e038c477c41" containerName="storage" Oct 08 18:22:38 crc kubenswrapper[4988]: I1008 18:22:38.003605 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="3cf2d054-6c91-4e4c-829b-1e038c477c41" containerName="storage" Oct 08 18:22:38 crc kubenswrapper[4988]: I1008 18:22:38.003757 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="3cf2d054-6c91-4e4c-829b-1e038c477c41" containerName="storage" Oct 08 18:22:38 crc kubenswrapper[4988]: I1008 18:22:38.004627 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c55jx7" Oct 08 18:22:38 crc kubenswrapper[4988]: I1008 18:22:38.007500 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Oct 08 18:22:38 crc kubenswrapper[4988]: I1008 18:22:38.016325 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c55jx7"] Oct 08 18:22:38 crc kubenswrapper[4988]: I1008 18:22:38.133305 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/ffc18ede-f7bc-450c-b37a-b5dcf72a1a50-util\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c55jx7\" (UID: \"ffc18ede-f7bc-450c-b37a-b5dcf72a1a50\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c55jx7" Oct 08 18:22:38 crc kubenswrapper[4988]: I1008 18:22:38.133379 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/ffc18ede-f7bc-450c-b37a-b5dcf72a1a50-bundle\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c55jx7\" (UID: \"ffc18ede-f7bc-450c-b37a-b5dcf72a1a50\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c55jx7" Oct 08 18:22:38 crc kubenswrapper[4988]: I1008 18:22:38.133652 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8tj69\" (UniqueName: \"kubernetes.io/projected/ffc18ede-f7bc-450c-b37a-b5dcf72a1a50-kube-api-access-8tj69\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c55jx7\" (UID: \"ffc18ede-f7bc-450c-b37a-b5dcf72a1a50\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c55jx7" Oct 08 18:22:38 crc kubenswrapper[4988]: I1008 18:22:38.234976 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/ffc18ede-f7bc-450c-b37a-b5dcf72a1a50-bundle\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c55jx7\" (UID: \"ffc18ede-f7bc-450c-b37a-b5dcf72a1a50\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c55jx7" Oct 08 18:22:38 crc kubenswrapper[4988]: I1008 18:22:38.235096 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8tj69\" (UniqueName: \"kubernetes.io/projected/ffc18ede-f7bc-450c-b37a-b5dcf72a1a50-kube-api-access-8tj69\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c55jx7\" (UID: \"ffc18ede-f7bc-450c-b37a-b5dcf72a1a50\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c55jx7" Oct 08 18:22:38 crc kubenswrapper[4988]: I1008 18:22:38.235233 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/ffc18ede-f7bc-450c-b37a-b5dcf72a1a50-util\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c55jx7\" (UID: \"ffc18ede-f7bc-450c-b37a-b5dcf72a1a50\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c55jx7" Oct 08 18:22:38 crc kubenswrapper[4988]: I1008 18:22:38.236180 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/ffc18ede-f7bc-450c-b37a-b5dcf72a1a50-util\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c55jx7\" (UID: \"ffc18ede-f7bc-450c-b37a-b5dcf72a1a50\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c55jx7" Oct 08 18:22:38 crc kubenswrapper[4988]: I1008 18:22:38.236190 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/ffc18ede-f7bc-450c-b37a-b5dcf72a1a50-bundle\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c55jx7\" (UID: \"ffc18ede-f7bc-450c-b37a-b5dcf72a1a50\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c55jx7" Oct 08 18:22:38 crc kubenswrapper[4988]: I1008 18:22:38.268740 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8tj69\" (UniqueName: \"kubernetes.io/projected/ffc18ede-f7bc-450c-b37a-b5dcf72a1a50-kube-api-access-8tj69\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c55jx7\" (UID: \"ffc18ede-f7bc-450c-b37a-b5dcf72a1a50\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c55jx7" Oct 08 18:22:38 crc kubenswrapper[4988]: I1008 18:22:38.333338 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c55jx7" Oct 08 18:22:38 crc kubenswrapper[4988]: I1008 18:22:38.846583 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c55jx7"] Oct 08 18:22:39 crc kubenswrapper[4988]: I1008 18:22:39.248513 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c55jx7" event={"ID":"ffc18ede-f7bc-450c-b37a-b5dcf72a1a50","Type":"ContainerStarted","Data":"a95e79e5a9e0a74d81e053229558673a7e21d398f3a4e8e9ea1c9f52221bd0c3"} Oct 08 18:22:39 crc kubenswrapper[4988]: I1008 18:22:39.248870 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c55jx7" event={"ID":"ffc18ede-f7bc-450c-b37a-b5dcf72a1a50","Type":"ContainerStarted","Data":"1a9032f2b5835bcfc0e312791c1444de7498c1a44f2a2a259da922bcd60b0b24"} Oct 08 18:22:40 crc kubenswrapper[4988]: I1008 18:22:40.253051 4988 generic.go:334] "Generic (PLEG): container finished" podID="ffc18ede-f7bc-450c-b37a-b5dcf72a1a50" containerID="a95e79e5a9e0a74d81e053229558673a7e21d398f3a4e8e9ea1c9f52221bd0c3" exitCode=0 Oct 08 18:22:40 crc kubenswrapper[4988]: I1008 18:22:40.253120 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c55jx7" event={"ID":"ffc18ede-f7bc-450c-b37a-b5dcf72a1a50","Type":"ContainerDied","Data":"a95e79e5a9e0a74d81e053229558673a7e21d398f3a4e8e9ea1c9f52221bd0c3"} Oct 08 18:22:42 crc kubenswrapper[4988]: I1008 18:22:42.271292 4988 generic.go:334] "Generic (PLEG): container finished" podID="ffc18ede-f7bc-450c-b37a-b5dcf72a1a50" containerID="77865ac5c351a45c759792e3fcefc34c70852b541fa7116ffd037c7ee56754d1" exitCode=0 Oct 08 18:22:42 crc kubenswrapper[4988]: I1008 18:22:42.271371 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c55jx7" event={"ID":"ffc18ede-f7bc-450c-b37a-b5dcf72a1a50","Type":"ContainerDied","Data":"77865ac5c351a45c759792e3fcefc34c70852b541fa7116ffd037c7ee56754d1"} Oct 08 18:22:43 crc kubenswrapper[4988]: I1008 18:22:43.281972 4988 generic.go:334] "Generic (PLEG): container finished" podID="ffc18ede-f7bc-450c-b37a-b5dcf72a1a50" containerID="57fcee777cd60887b7524e69ac0fe103f9bf18e96e67489afc5a41678063f988" exitCode=0 Oct 08 18:22:43 crc kubenswrapper[4988]: I1008 18:22:43.282028 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c55jx7" event={"ID":"ffc18ede-f7bc-450c-b37a-b5dcf72a1a50","Type":"ContainerDied","Data":"57fcee777cd60887b7524e69ac0fe103f9bf18e96e67489afc5a41678063f988"} Oct 08 18:22:44 crc kubenswrapper[4988]: I1008 18:22:44.537882 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c55jx7" Oct 08 18:22:44 crc kubenswrapper[4988]: I1008 18:22:44.624116 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/ffc18ede-f7bc-450c-b37a-b5dcf72a1a50-bundle\") pod \"ffc18ede-f7bc-450c-b37a-b5dcf72a1a50\" (UID: \"ffc18ede-f7bc-450c-b37a-b5dcf72a1a50\") " Oct 08 18:22:44 crc kubenswrapper[4988]: I1008 18:22:44.624199 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8tj69\" (UniqueName: \"kubernetes.io/projected/ffc18ede-f7bc-450c-b37a-b5dcf72a1a50-kube-api-access-8tj69\") pod \"ffc18ede-f7bc-450c-b37a-b5dcf72a1a50\" (UID: \"ffc18ede-f7bc-450c-b37a-b5dcf72a1a50\") " Oct 08 18:22:44 crc kubenswrapper[4988]: I1008 18:22:44.624326 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/ffc18ede-f7bc-450c-b37a-b5dcf72a1a50-util\") pod \"ffc18ede-f7bc-450c-b37a-b5dcf72a1a50\" (UID: \"ffc18ede-f7bc-450c-b37a-b5dcf72a1a50\") " Oct 08 18:22:44 crc kubenswrapper[4988]: I1008 18:22:44.625039 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ffc18ede-f7bc-450c-b37a-b5dcf72a1a50-bundle" (OuterVolumeSpecName: "bundle") pod "ffc18ede-f7bc-450c-b37a-b5dcf72a1a50" (UID: "ffc18ede-f7bc-450c-b37a-b5dcf72a1a50"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 18:22:44 crc kubenswrapper[4988]: I1008 18:22:44.632686 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ffc18ede-f7bc-450c-b37a-b5dcf72a1a50-kube-api-access-8tj69" (OuterVolumeSpecName: "kube-api-access-8tj69") pod "ffc18ede-f7bc-450c-b37a-b5dcf72a1a50" (UID: "ffc18ede-f7bc-450c-b37a-b5dcf72a1a50"). InnerVolumeSpecName "kube-api-access-8tj69". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:22:44 crc kubenswrapper[4988]: I1008 18:22:44.726121 4988 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/ffc18ede-f7bc-450c-b37a-b5dcf72a1a50-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 18:22:44 crc kubenswrapper[4988]: I1008 18:22:44.726207 4988 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8tj69\" (UniqueName: \"kubernetes.io/projected/ffc18ede-f7bc-450c-b37a-b5dcf72a1a50-kube-api-access-8tj69\") on node \"crc\" DevicePath \"\"" Oct 08 18:22:44 crc kubenswrapper[4988]: I1008 18:22:44.917118 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ffc18ede-f7bc-450c-b37a-b5dcf72a1a50-util" (OuterVolumeSpecName: "util") pod "ffc18ede-f7bc-450c-b37a-b5dcf72a1a50" (UID: "ffc18ede-f7bc-450c-b37a-b5dcf72a1a50"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 18:22:44 crc kubenswrapper[4988]: I1008 18:22:44.929703 4988 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/ffc18ede-f7bc-450c-b37a-b5dcf72a1a50-util\") on node \"crc\" DevicePath \"\"" Oct 08 18:22:45 crc kubenswrapper[4988]: I1008 18:22:45.296676 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c55jx7" event={"ID":"ffc18ede-f7bc-450c-b37a-b5dcf72a1a50","Type":"ContainerDied","Data":"1a9032f2b5835bcfc0e312791c1444de7498c1a44f2a2a259da922bcd60b0b24"} Oct 08 18:22:45 crc kubenswrapper[4988]: I1008 18:22:45.296711 4988 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1a9032f2b5835bcfc0e312791c1444de7498c1a44f2a2a259da922bcd60b0b24" Oct 08 18:22:45 crc kubenswrapper[4988]: I1008 18:22:45.296790 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c55jx7" Oct 08 18:22:49 crc kubenswrapper[4988]: I1008 18:22:49.374828 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-operator-858ddd8f98-hplwk"] Oct 08 18:22:49 crc kubenswrapper[4988]: E1008 18:22:49.375431 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ffc18ede-f7bc-450c-b37a-b5dcf72a1a50" containerName="util" Oct 08 18:22:49 crc kubenswrapper[4988]: I1008 18:22:49.375449 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="ffc18ede-f7bc-450c-b37a-b5dcf72a1a50" containerName="util" Oct 08 18:22:49 crc kubenswrapper[4988]: E1008 18:22:49.375473 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ffc18ede-f7bc-450c-b37a-b5dcf72a1a50" containerName="pull" Oct 08 18:22:49 crc kubenswrapper[4988]: I1008 18:22:49.375482 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="ffc18ede-f7bc-450c-b37a-b5dcf72a1a50" containerName="pull" Oct 08 18:22:49 crc kubenswrapper[4988]: E1008 18:22:49.375497 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ffc18ede-f7bc-450c-b37a-b5dcf72a1a50" containerName="extract" Oct 08 18:22:49 crc kubenswrapper[4988]: I1008 18:22:49.375507 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="ffc18ede-f7bc-450c-b37a-b5dcf72a1a50" containerName="extract" Oct 08 18:22:49 crc kubenswrapper[4988]: I1008 18:22:49.375619 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="ffc18ede-f7bc-450c-b37a-b5dcf72a1a50" containerName="extract" Oct 08 18:22:49 crc kubenswrapper[4988]: I1008 18:22:49.376053 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-858ddd8f98-hplwk" Oct 08 18:22:49 crc kubenswrapper[4988]: I1008 18:22:49.379294 4988 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"kube-root-ca.crt" Oct 08 18:22:49 crc kubenswrapper[4988]: I1008 18:22:49.379467 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"nmstate-operator-dockercfg-4rdnb" Oct 08 18:22:49 crc kubenswrapper[4988]: I1008 18:22:49.381456 4988 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"openshift-service-ca.crt" Oct 08 18:22:49 crc kubenswrapper[4988]: I1008 18:22:49.388735 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-operator-858ddd8f98-hplwk"] Oct 08 18:22:49 crc kubenswrapper[4988]: I1008 18:22:49.502828 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mwm6p\" (UniqueName: \"kubernetes.io/projected/cc122836-fc53-4258-8074-f0943ca9c946-kube-api-access-mwm6p\") pod \"nmstate-operator-858ddd8f98-hplwk\" (UID: \"cc122836-fc53-4258-8074-f0943ca9c946\") " pod="openshift-nmstate/nmstate-operator-858ddd8f98-hplwk" Oct 08 18:22:49 crc kubenswrapper[4988]: I1008 18:22:49.604585 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mwm6p\" (UniqueName: \"kubernetes.io/projected/cc122836-fc53-4258-8074-f0943ca9c946-kube-api-access-mwm6p\") pod \"nmstate-operator-858ddd8f98-hplwk\" (UID: \"cc122836-fc53-4258-8074-f0943ca9c946\") " pod="openshift-nmstate/nmstate-operator-858ddd8f98-hplwk" Oct 08 18:22:49 crc kubenswrapper[4988]: I1008 18:22:49.624119 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mwm6p\" (UniqueName: \"kubernetes.io/projected/cc122836-fc53-4258-8074-f0943ca9c946-kube-api-access-mwm6p\") pod \"nmstate-operator-858ddd8f98-hplwk\" (UID: \"cc122836-fc53-4258-8074-f0943ca9c946\") " pod="openshift-nmstate/nmstate-operator-858ddd8f98-hplwk" Oct 08 18:22:49 crc kubenswrapper[4988]: I1008 18:22:49.689156 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-858ddd8f98-hplwk" Oct 08 18:22:49 crc kubenswrapper[4988]: I1008 18:22:49.898030 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-operator-858ddd8f98-hplwk"] Oct 08 18:22:50 crc kubenswrapper[4988]: I1008 18:22:50.323609 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-operator-858ddd8f98-hplwk" event={"ID":"cc122836-fc53-4258-8074-f0943ca9c946","Type":"ContainerStarted","Data":"89d808ce4b78483ec7ee89b646bbfe651f4c45e7cb340fb17567a6c24ed403b8"} Oct 08 18:22:52 crc kubenswrapper[4988]: I1008 18:22:52.339284 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-operator-858ddd8f98-hplwk" event={"ID":"cc122836-fc53-4258-8074-f0943ca9c946","Type":"ContainerStarted","Data":"027195ea575743e78dbf28b3bced5e483daa0f8aa7af6ddd68ef0b4ad228c50c"} Oct 08 18:22:52 crc kubenswrapper[4988]: I1008 18:22:52.359627 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-operator-858ddd8f98-hplwk" podStartSLOduration=1.095493811 podStartE2EDuration="3.35960933s" podCreationTimestamp="2025-10-08 18:22:49 +0000 UTC" firstStartedPulling="2025-10-08 18:22:49.908547754 +0000 UTC m=+715.358390524" lastFinishedPulling="2025-10-08 18:22:52.172663283 +0000 UTC m=+717.622506043" observedRunningTime="2025-10-08 18:22:52.358055368 +0000 UTC m=+717.807898188" watchObservedRunningTime="2025-10-08 18:22:52.35960933 +0000 UTC m=+717.809452100" Oct 08 18:22:58 crc kubenswrapper[4988]: I1008 18:22:58.436309 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-metrics-fdff9cb8d-b574s"] Oct 08 18:22:58 crc kubenswrapper[4988]: I1008 18:22:58.437971 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-b574s" Oct 08 18:22:58 crc kubenswrapper[4988]: I1008 18:22:58.442563 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"nmstate-handler-dockercfg-zgf8l" Oct 08 18:22:58 crc kubenswrapper[4988]: I1008 18:22:58.453957 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-metrics-fdff9cb8d-b574s"] Oct 08 18:22:58 crc kubenswrapper[4988]: I1008 18:22:58.466971 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-webhook-6cdbc54649-xg4fg"] Oct 08 18:22:58 crc kubenswrapper[4988]: I1008 18:22:58.467580 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-webhook-6cdbc54649-xg4fg" Oct 08 18:22:58 crc kubenswrapper[4988]: I1008 18:22:58.472397 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"openshift-nmstate-webhook" Oct 08 18:22:58 crc kubenswrapper[4988]: I1008 18:22:58.494691 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-webhook-6cdbc54649-xg4fg"] Oct 08 18:22:58 crc kubenswrapper[4988]: I1008 18:22:58.515187 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-handler-dv92p"] Oct 08 18:22:58 crc kubenswrapper[4988]: I1008 18:22:58.516597 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-handler-dv92p" Oct 08 18:22:58 crc kubenswrapper[4988]: I1008 18:22:58.520269 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vthv2\" (UniqueName: \"kubernetes.io/projected/09b9ef2d-78cf-4bb7-a5ed-79f63f65fa8c-kube-api-access-vthv2\") pod \"nmstate-metrics-fdff9cb8d-b574s\" (UID: \"09b9ef2d-78cf-4bb7-a5ed-79f63f65fa8c\") " pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-b574s" Oct 08 18:22:58 crc kubenswrapper[4988]: I1008 18:22:58.613353 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-console-plugin-6b874cbd85-wh7bx"] Oct 08 18:22:58 crc kubenswrapper[4988]: I1008 18:22:58.615479 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-wh7bx" Oct 08 18:22:58 crc kubenswrapper[4988]: I1008 18:22:58.617248 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-6b874cbd85-wh7bx"] Oct 08 18:22:58 crc kubenswrapper[4988]: I1008 18:22:58.618104 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"default-dockercfg-jmkgf" Oct 08 18:22:58 crc kubenswrapper[4988]: I1008 18:22:58.618365 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"plugin-serving-cert" Oct 08 18:22:58 crc kubenswrapper[4988]: I1008 18:22:58.618559 4988 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"nginx-conf" Oct 08 18:22:58 crc kubenswrapper[4988]: I1008 18:22:58.623324 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5d8xm\" (UniqueName: \"kubernetes.io/projected/bb27ea19-b9c4-4ea6-b7b1-6de4ae603b43-kube-api-access-5d8xm\") pod \"nmstate-webhook-6cdbc54649-xg4fg\" (UID: \"bb27ea19-b9c4-4ea6-b7b1-6de4ae603b43\") " pod="openshift-nmstate/nmstate-webhook-6cdbc54649-xg4fg" Oct 08 18:22:58 crc kubenswrapper[4988]: I1008 18:22:58.623440 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/bb27ea19-b9c4-4ea6-b7b1-6de4ae603b43-tls-key-pair\") pod \"nmstate-webhook-6cdbc54649-xg4fg\" (UID: \"bb27ea19-b9c4-4ea6-b7b1-6de4ae603b43\") " pod="openshift-nmstate/nmstate-webhook-6cdbc54649-xg4fg" Oct 08 18:22:58 crc kubenswrapper[4988]: I1008 18:22:58.623510 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/b0d72824-9655-4a95-9ada-e7d7f8938213-dbus-socket\") pod \"nmstate-handler-dv92p\" (UID: \"b0d72824-9655-4a95-9ada-e7d7f8938213\") " pod="openshift-nmstate/nmstate-handler-dv92p" Oct 08 18:22:58 crc kubenswrapper[4988]: I1008 18:22:58.623528 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xkk6l\" (UniqueName: \"kubernetes.io/projected/b0d72824-9655-4a95-9ada-e7d7f8938213-kube-api-access-xkk6l\") pod \"nmstate-handler-dv92p\" (UID: \"b0d72824-9655-4a95-9ada-e7d7f8938213\") " pod="openshift-nmstate/nmstate-handler-dv92p" Oct 08 18:22:58 crc kubenswrapper[4988]: I1008 18:22:58.623562 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vthv2\" (UniqueName: \"kubernetes.io/projected/09b9ef2d-78cf-4bb7-a5ed-79f63f65fa8c-kube-api-access-vthv2\") pod \"nmstate-metrics-fdff9cb8d-b574s\" (UID: \"09b9ef2d-78cf-4bb7-a5ed-79f63f65fa8c\") " pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-b574s" Oct 08 18:22:58 crc kubenswrapper[4988]: I1008 18:22:58.623591 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/b0d72824-9655-4a95-9ada-e7d7f8938213-nmstate-lock\") pod \"nmstate-handler-dv92p\" (UID: \"b0d72824-9655-4a95-9ada-e7d7f8938213\") " pod="openshift-nmstate/nmstate-handler-dv92p" Oct 08 18:22:58 crc kubenswrapper[4988]: I1008 18:22:58.623635 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/b0d72824-9655-4a95-9ada-e7d7f8938213-ovs-socket\") pod \"nmstate-handler-dv92p\" (UID: \"b0d72824-9655-4a95-9ada-e7d7f8938213\") " pod="openshift-nmstate/nmstate-handler-dv92p" Oct 08 18:22:58 crc kubenswrapper[4988]: I1008 18:22:58.648932 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vthv2\" (UniqueName: \"kubernetes.io/projected/09b9ef2d-78cf-4bb7-a5ed-79f63f65fa8c-kube-api-access-vthv2\") pod \"nmstate-metrics-fdff9cb8d-b574s\" (UID: \"09b9ef2d-78cf-4bb7-a5ed-79f63f65fa8c\") " pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-b574s" Oct 08 18:22:58 crc kubenswrapper[4988]: I1008 18:22:58.725711 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/b0d72824-9655-4a95-9ada-e7d7f8938213-ovs-socket\") pod \"nmstate-handler-dv92p\" (UID: \"b0d72824-9655-4a95-9ada-e7d7f8938213\") " pod="openshift-nmstate/nmstate-handler-dv92p" Oct 08 18:22:58 crc kubenswrapper[4988]: I1008 18:22:58.726340 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jltp9\" (UniqueName: \"kubernetes.io/projected/670ba223-6bd1-429c-8739-61aea9074914-kube-api-access-jltp9\") pod \"nmstate-console-plugin-6b874cbd85-wh7bx\" (UID: \"670ba223-6bd1-429c-8739-61aea9074914\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-wh7bx" Oct 08 18:22:58 crc kubenswrapper[4988]: I1008 18:22:58.726375 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5d8xm\" (UniqueName: \"kubernetes.io/projected/bb27ea19-b9c4-4ea6-b7b1-6de4ae603b43-kube-api-access-5d8xm\") pod \"nmstate-webhook-6cdbc54649-xg4fg\" (UID: \"bb27ea19-b9c4-4ea6-b7b1-6de4ae603b43\") " pod="openshift-nmstate/nmstate-webhook-6cdbc54649-xg4fg" Oct 08 18:22:58 crc kubenswrapper[4988]: I1008 18:22:58.725871 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/b0d72824-9655-4a95-9ada-e7d7f8938213-ovs-socket\") pod \"nmstate-handler-dv92p\" (UID: \"b0d72824-9655-4a95-9ada-e7d7f8938213\") " pod="openshift-nmstate/nmstate-handler-dv92p" Oct 08 18:22:58 crc kubenswrapper[4988]: I1008 18:22:58.726414 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/bb27ea19-b9c4-4ea6-b7b1-6de4ae603b43-tls-key-pair\") pod \"nmstate-webhook-6cdbc54649-xg4fg\" (UID: \"bb27ea19-b9c4-4ea6-b7b1-6de4ae603b43\") " pod="openshift-nmstate/nmstate-webhook-6cdbc54649-xg4fg" Oct 08 18:22:58 crc kubenswrapper[4988]: I1008 18:22:58.726730 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/b0d72824-9655-4a95-9ada-e7d7f8938213-dbus-socket\") pod \"nmstate-handler-dv92p\" (UID: \"b0d72824-9655-4a95-9ada-e7d7f8938213\") " pod="openshift-nmstate/nmstate-handler-dv92p" Oct 08 18:22:58 crc kubenswrapper[4988]: I1008 18:22:58.726769 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xkk6l\" (UniqueName: \"kubernetes.io/projected/b0d72824-9655-4a95-9ada-e7d7f8938213-kube-api-access-xkk6l\") pod \"nmstate-handler-dv92p\" (UID: \"b0d72824-9655-4a95-9ada-e7d7f8938213\") " pod="openshift-nmstate/nmstate-handler-dv92p" Oct 08 18:22:58 crc kubenswrapper[4988]: I1008 18:22:58.726837 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/670ba223-6bd1-429c-8739-61aea9074914-nginx-conf\") pod \"nmstate-console-plugin-6b874cbd85-wh7bx\" (UID: \"670ba223-6bd1-429c-8739-61aea9074914\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-wh7bx" Oct 08 18:22:58 crc kubenswrapper[4988]: I1008 18:22:58.726935 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/670ba223-6bd1-429c-8739-61aea9074914-plugin-serving-cert\") pod \"nmstate-console-plugin-6b874cbd85-wh7bx\" (UID: \"670ba223-6bd1-429c-8739-61aea9074914\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-wh7bx" Oct 08 18:22:58 crc kubenswrapper[4988]: I1008 18:22:58.727067 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/b0d72824-9655-4a95-9ada-e7d7f8938213-nmstate-lock\") pod \"nmstate-handler-dv92p\" (UID: \"b0d72824-9655-4a95-9ada-e7d7f8938213\") " pod="openshift-nmstate/nmstate-handler-dv92p" Oct 08 18:22:58 crc kubenswrapper[4988]: I1008 18:22:58.727094 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/b0d72824-9655-4a95-9ada-e7d7f8938213-dbus-socket\") pod \"nmstate-handler-dv92p\" (UID: \"b0d72824-9655-4a95-9ada-e7d7f8938213\") " pod="openshift-nmstate/nmstate-handler-dv92p" Oct 08 18:22:58 crc kubenswrapper[4988]: I1008 18:22:58.727163 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/b0d72824-9655-4a95-9ada-e7d7f8938213-nmstate-lock\") pod \"nmstate-handler-dv92p\" (UID: \"b0d72824-9655-4a95-9ada-e7d7f8938213\") " pod="openshift-nmstate/nmstate-handler-dv92p" Oct 08 18:22:58 crc kubenswrapper[4988]: I1008 18:22:58.732230 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/bb27ea19-b9c4-4ea6-b7b1-6de4ae603b43-tls-key-pair\") pod \"nmstate-webhook-6cdbc54649-xg4fg\" (UID: \"bb27ea19-b9c4-4ea6-b7b1-6de4ae603b43\") " pod="openshift-nmstate/nmstate-webhook-6cdbc54649-xg4fg" Oct 08 18:22:58 crc kubenswrapper[4988]: I1008 18:22:58.755858 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-b574s" Oct 08 18:22:58 crc kubenswrapper[4988]: I1008 18:22:58.757953 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5d8xm\" (UniqueName: \"kubernetes.io/projected/bb27ea19-b9c4-4ea6-b7b1-6de4ae603b43-kube-api-access-5d8xm\") pod \"nmstate-webhook-6cdbc54649-xg4fg\" (UID: \"bb27ea19-b9c4-4ea6-b7b1-6de4ae603b43\") " pod="openshift-nmstate/nmstate-webhook-6cdbc54649-xg4fg" Oct 08 18:22:58 crc kubenswrapper[4988]: I1008 18:22:58.768528 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xkk6l\" (UniqueName: \"kubernetes.io/projected/b0d72824-9655-4a95-9ada-e7d7f8938213-kube-api-access-xkk6l\") pod \"nmstate-handler-dv92p\" (UID: \"b0d72824-9655-4a95-9ada-e7d7f8938213\") " pod="openshift-nmstate/nmstate-handler-dv92p" Oct 08 18:22:58 crc kubenswrapper[4988]: I1008 18:22:58.789845 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-webhook-6cdbc54649-xg4fg" Oct 08 18:22:58 crc kubenswrapper[4988]: I1008 18:22:58.808823 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-64ff57b4d9-96fmx"] Oct 08 18:22:58 crc kubenswrapper[4988]: I1008 18:22:58.810032 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-64ff57b4d9-96fmx" Oct 08 18:22:58 crc kubenswrapper[4988]: I1008 18:22:58.836864 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-handler-dv92p" Oct 08 18:22:58 crc kubenswrapper[4988]: I1008 18:22:58.843542 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jltp9\" (UniqueName: \"kubernetes.io/projected/670ba223-6bd1-429c-8739-61aea9074914-kube-api-access-jltp9\") pod \"nmstate-console-plugin-6b874cbd85-wh7bx\" (UID: \"670ba223-6bd1-429c-8739-61aea9074914\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-wh7bx" Oct 08 18:22:58 crc kubenswrapper[4988]: I1008 18:22:58.843666 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/670ba223-6bd1-429c-8739-61aea9074914-nginx-conf\") pod \"nmstate-console-plugin-6b874cbd85-wh7bx\" (UID: \"670ba223-6bd1-429c-8739-61aea9074914\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-wh7bx" Oct 08 18:22:58 crc kubenswrapper[4988]: I1008 18:22:58.843708 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/670ba223-6bd1-429c-8739-61aea9074914-plugin-serving-cert\") pod \"nmstate-console-plugin-6b874cbd85-wh7bx\" (UID: \"670ba223-6bd1-429c-8739-61aea9074914\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-wh7bx" Oct 08 18:22:58 crc kubenswrapper[4988]: I1008 18:22:58.845980 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/670ba223-6bd1-429c-8739-61aea9074914-nginx-conf\") pod \"nmstate-console-plugin-6b874cbd85-wh7bx\" (UID: \"670ba223-6bd1-429c-8739-61aea9074914\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-wh7bx" Oct 08 18:22:58 crc kubenswrapper[4988]: I1008 18:22:58.863049 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/670ba223-6bd1-429c-8739-61aea9074914-plugin-serving-cert\") pod \"nmstate-console-plugin-6b874cbd85-wh7bx\" (UID: \"670ba223-6bd1-429c-8739-61aea9074914\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-wh7bx" Oct 08 18:22:58 crc kubenswrapper[4988]: I1008 18:22:58.865724 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-64ff57b4d9-96fmx"] Oct 08 18:22:58 crc kubenswrapper[4988]: I1008 18:22:58.867709 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jltp9\" (UniqueName: \"kubernetes.io/projected/670ba223-6bd1-429c-8739-61aea9074914-kube-api-access-jltp9\") pod \"nmstate-console-plugin-6b874cbd85-wh7bx\" (UID: \"670ba223-6bd1-429c-8739-61aea9074914\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-wh7bx" Oct 08 18:22:58 crc kubenswrapper[4988]: I1008 18:22:58.931051 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-wh7bx" Oct 08 18:22:58 crc kubenswrapper[4988]: I1008 18:22:58.944800 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e6eee236-e4df-4df0-8e42-eec7d1b971dd-trusted-ca-bundle\") pod \"console-64ff57b4d9-96fmx\" (UID: \"e6eee236-e4df-4df0-8e42-eec7d1b971dd\") " pod="openshift-console/console-64ff57b4d9-96fmx" Oct 08 18:22:58 crc kubenswrapper[4988]: I1008 18:22:58.944837 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/e6eee236-e4df-4df0-8e42-eec7d1b971dd-oauth-serving-cert\") pod \"console-64ff57b4d9-96fmx\" (UID: \"e6eee236-e4df-4df0-8e42-eec7d1b971dd\") " pod="openshift-console/console-64ff57b4d9-96fmx" Oct 08 18:22:58 crc kubenswrapper[4988]: I1008 18:22:58.944868 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/e6eee236-e4df-4df0-8e42-eec7d1b971dd-console-config\") pod \"console-64ff57b4d9-96fmx\" (UID: \"e6eee236-e4df-4df0-8e42-eec7d1b971dd\") " pod="openshift-console/console-64ff57b4d9-96fmx" Oct 08 18:22:58 crc kubenswrapper[4988]: I1008 18:22:58.944903 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/e6eee236-e4df-4df0-8e42-eec7d1b971dd-console-serving-cert\") pod \"console-64ff57b4d9-96fmx\" (UID: \"e6eee236-e4df-4df0-8e42-eec7d1b971dd\") " pod="openshift-console/console-64ff57b4d9-96fmx" Oct 08 18:22:58 crc kubenswrapper[4988]: I1008 18:22:58.944921 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fqtrr\" (UniqueName: \"kubernetes.io/projected/e6eee236-e4df-4df0-8e42-eec7d1b971dd-kube-api-access-fqtrr\") pod \"console-64ff57b4d9-96fmx\" (UID: \"e6eee236-e4df-4df0-8e42-eec7d1b971dd\") " pod="openshift-console/console-64ff57b4d9-96fmx" Oct 08 18:22:58 crc kubenswrapper[4988]: I1008 18:22:58.944944 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/e6eee236-e4df-4df0-8e42-eec7d1b971dd-service-ca\") pod \"console-64ff57b4d9-96fmx\" (UID: \"e6eee236-e4df-4df0-8e42-eec7d1b971dd\") " pod="openshift-console/console-64ff57b4d9-96fmx" Oct 08 18:22:58 crc kubenswrapper[4988]: I1008 18:22:58.944967 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/e6eee236-e4df-4df0-8e42-eec7d1b971dd-console-oauth-config\") pod \"console-64ff57b4d9-96fmx\" (UID: \"e6eee236-e4df-4df0-8e42-eec7d1b971dd\") " pod="openshift-console/console-64ff57b4d9-96fmx" Oct 08 18:22:58 crc kubenswrapper[4988]: I1008 18:22:58.997618 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-metrics-fdff9cb8d-b574s"] Oct 08 18:22:59 crc kubenswrapper[4988]: W1008 18:22:59.005985 4988 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod09b9ef2d_78cf_4bb7_a5ed_79f63f65fa8c.slice/crio-b64a0076011188da8ee698f381246e818833e2ff97fa638742a5ec8861ed51f1 WatchSource:0}: Error finding container b64a0076011188da8ee698f381246e818833e2ff97fa638742a5ec8861ed51f1: Status 404 returned error can't find the container with id b64a0076011188da8ee698f381246e818833e2ff97fa638742a5ec8861ed51f1 Oct 08 18:22:59 crc kubenswrapper[4988]: I1008 18:22:59.046618 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-webhook-6cdbc54649-xg4fg"] Oct 08 18:22:59 crc kubenswrapper[4988]: I1008 18:22:59.046655 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e6eee236-e4df-4df0-8e42-eec7d1b971dd-trusted-ca-bundle\") pod \"console-64ff57b4d9-96fmx\" (UID: \"e6eee236-e4df-4df0-8e42-eec7d1b971dd\") " pod="openshift-console/console-64ff57b4d9-96fmx" Oct 08 18:22:59 crc kubenswrapper[4988]: I1008 18:22:59.046708 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/e6eee236-e4df-4df0-8e42-eec7d1b971dd-oauth-serving-cert\") pod \"console-64ff57b4d9-96fmx\" (UID: \"e6eee236-e4df-4df0-8e42-eec7d1b971dd\") " pod="openshift-console/console-64ff57b4d9-96fmx" Oct 08 18:22:59 crc kubenswrapper[4988]: I1008 18:22:59.046743 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/e6eee236-e4df-4df0-8e42-eec7d1b971dd-console-config\") pod \"console-64ff57b4d9-96fmx\" (UID: \"e6eee236-e4df-4df0-8e42-eec7d1b971dd\") " pod="openshift-console/console-64ff57b4d9-96fmx" Oct 08 18:22:59 crc kubenswrapper[4988]: I1008 18:22:59.046787 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/e6eee236-e4df-4df0-8e42-eec7d1b971dd-console-serving-cert\") pod \"console-64ff57b4d9-96fmx\" (UID: \"e6eee236-e4df-4df0-8e42-eec7d1b971dd\") " pod="openshift-console/console-64ff57b4d9-96fmx" Oct 08 18:22:59 crc kubenswrapper[4988]: I1008 18:22:59.046811 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fqtrr\" (UniqueName: \"kubernetes.io/projected/e6eee236-e4df-4df0-8e42-eec7d1b971dd-kube-api-access-fqtrr\") pod \"console-64ff57b4d9-96fmx\" (UID: \"e6eee236-e4df-4df0-8e42-eec7d1b971dd\") " pod="openshift-console/console-64ff57b4d9-96fmx" Oct 08 18:22:59 crc kubenswrapper[4988]: I1008 18:22:59.046846 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/e6eee236-e4df-4df0-8e42-eec7d1b971dd-service-ca\") pod \"console-64ff57b4d9-96fmx\" (UID: \"e6eee236-e4df-4df0-8e42-eec7d1b971dd\") " pod="openshift-console/console-64ff57b4d9-96fmx" Oct 08 18:22:59 crc kubenswrapper[4988]: I1008 18:22:59.046875 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/e6eee236-e4df-4df0-8e42-eec7d1b971dd-console-oauth-config\") pod \"console-64ff57b4d9-96fmx\" (UID: \"e6eee236-e4df-4df0-8e42-eec7d1b971dd\") " pod="openshift-console/console-64ff57b4d9-96fmx" Oct 08 18:22:59 crc kubenswrapper[4988]: I1008 18:22:59.048690 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e6eee236-e4df-4df0-8e42-eec7d1b971dd-trusted-ca-bundle\") pod \"console-64ff57b4d9-96fmx\" (UID: \"e6eee236-e4df-4df0-8e42-eec7d1b971dd\") " pod="openshift-console/console-64ff57b4d9-96fmx" Oct 08 18:22:59 crc kubenswrapper[4988]: I1008 18:22:59.050473 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/e6eee236-e4df-4df0-8e42-eec7d1b971dd-console-config\") pod \"console-64ff57b4d9-96fmx\" (UID: \"e6eee236-e4df-4df0-8e42-eec7d1b971dd\") " pod="openshift-console/console-64ff57b4d9-96fmx" Oct 08 18:22:59 crc kubenswrapper[4988]: I1008 18:22:59.050612 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/e6eee236-e4df-4df0-8e42-eec7d1b971dd-console-oauth-config\") pod \"console-64ff57b4d9-96fmx\" (UID: \"e6eee236-e4df-4df0-8e42-eec7d1b971dd\") " pod="openshift-console/console-64ff57b4d9-96fmx" Oct 08 18:22:59 crc kubenswrapper[4988]: I1008 18:22:59.052434 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/e6eee236-e4df-4df0-8e42-eec7d1b971dd-service-ca\") pod \"console-64ff57b4d9-96fmx\" (UID: \"e6eee236-e4df-4df0-8e42-eec7d1b971dd\") " pod="openshift-console/console-64ff57b4d9-96fmx" Oct 08 18:22:59 crc kubenswrapper[4988]: I1008 18:22:59.054194 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/e6eee236-e4df-4df0-8e42-eec7d1b971dd-console-serving-cert\") pod \"console-64ff57b4d9-96fmx\" (UID: \"e6eee236-e4df-4df0-8e42-eec7d1b971dd\") " pod="openshift-console/console-64ff57b4d9-96fmx" Oct 08 18:22:59 crc kubenswrapper[4988]: I1008 18:22:59.054220 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/e6eee236-e4df-4df0-8e42-eec7d1b971dd-oauth-serving-cert\") pod \"console-64ff57b4d9-96fmx\" (UID: \"e6eee236-e4df-4df0-8e42-eec7d1b971dd\") " pod="openshift-console/console-64ff57b4d9-96fmx" Oct 08 18:22:59 crc kubenswrapper[4988]: I1008 18:22:59.068271 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fqtrr\" (UniqueName: \"kubernetes.io/projected/e6eee236-e4df-4df0-8e42-eec7d1b971dd-kube-api-access-fqtrr\") pod \"console-64ff57b4d9-96fmx\" (UID: \"e6eee236-e4df-4df0-8e42-eec7d1b971dd\") " pod="openshift-console/console-64ff57b4d9-96fmx" Oct 08 18:22:59 crc kubenswrapper[4988]: I1008 18:22:59.170211 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-64ff57b4d9-96fmx" Oct 08 18:22:59 crc kubenswrapper[4988]: I1008 18:22:59.347161 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-64ff57b4d9-96fmx"] Oct 08 18:22:59 crc kubenswrapper[4988]: I1008 18:22:59.351841 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-6b874cbd85-wh7bx"] Oct 08 18:22:59 crc kubenswrapper[4988]: W1008 18:22:59.352617 4988 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode6eee236_e4df_4df0_8e42_eec7d1b971dd.slice/crio-fc463941862efc5e8ca9d3d840af40d294c51e6076d252ec7226870c053b43e8 WatchSource:0}: Error finding container fc463941862efc5e8ca9d3d840af40d294c51e6076d252ec7226870c053b43e8: Status 404 returned error can't find the container with id fc463941862efc5e8ca9d3d840af40d294c51e6076d252ec7226870c053b43e8 Oct 08 18:22:59 crc kubenswrapper[4988]: W1008 18:22:59.353673 4988 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod670ba223_6bd1_429c_8739_61aea9074914.slice/crio-d0d289ec4dc78a840c893633c094298268f0ac110319205dc1270412254ba206 WatchSource:0}: Error finding container d0d289ec4dc78a840c893633c094298268f0ac110319205dc1270412254ba206: Status 404 returned error can't find the container with id d0d289ec4dc78a840c893633c094298268f0ac110319205dc1270412254ba206 Oct 08 18:22:59 crc kubenswrapper[4988]: I1008 18:22:59.378576 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-b574s" event={"ID":"09b9ef2d-78cf-4bb7-a5ed-79f63f65fa8c","Type":"ContainerStarted","Data":"b64a0076011188da8ee698f381246e818833e2ff97fa638742a5ec8861ed51f1"} Oct 08 18:22:59 crc kubenswrapper[4988]: I1008 18:22:59.379459 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-64ff57b4d9-96fmx" event={"ID":"e6eee236-e4df-4df0-8e42-eec7d1b971dd","Type":"ContainerStarted","Data":"fc463941862efc5e8ca9d3d840af40d294c51e6076d252ec7226870c053b43e8"} Oct 08 18:22:59 crc kubenswrapper[4988]: I1008 18:22:59.380548 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-wh7bx" event={"ID":"670ba223-6bd1-429c-8739-61aea9074914","Type":"ContainerStarted","Data":"d0d289ec4dc78a840c893633c094298268f0ac110319205dc1270412254ba206"} Oct 08 18:22:59 crc kubenswrapper[4988]: I1008 18:22:59.381928 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-webhook-6cdbc54649-xg4fg" event={"ID":"bb27ea19-b9c4-4ea6-b7b1-6de4ae603b43","Type":"ContainerStarted","Data":"be034fb62af596579836cecdfb0802e69d9e28276cbcdf801ebaf3e286cd0e00"} Oct 08 18:22:59 crc kubenswrapper[4988]: I1008 18:22:59.382837 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-handler-dv92p" event={"ID":"b0d72824-9655-4a95-9ada-e7d7f8938213","Type":"ContainerStarted","Data":"50369ee8eb34f4c82ecbe81f33b6a5a875b9e3e58db24d2e394e2d6e1e81f354"} Oct 08 18:23:00 crc kubenswrapper[4988]: I1008 18:23:00.390360 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-64ff57b4d9-96fmx" event={"ID":"e6eee236-e4df-4df0-8e42-eec7d1b971dd","Type":"ContainerStarted","Data":"429392f4a02d41bb524f434cf33175424eddb939f1238b10d376e127be67d31e"} Oct 08 18:23:00 crc kubenswrapper[4988]: I1008 18:23:00.411484 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-64ff57b4d9-96fmx" podStartSLOduration=2.411460912 podStartE2EDuration="2.411460912s" podCreationTimestamp="2025-10-08 18:22:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 18:23:00.410090058 +0000 UTC m=+725.859932868" watchObservedRunningTime="2025-10-08 18:23:00.411460912 +0000 UTC m=+725.861303692" Oct 08 18:23:02 crc kubenswrapper[4988]: I1008 18:23:02.405850 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-b574s" event={"ID":"09b9ef2d-78cf-4bb7-a5ed-79f63f65fa8c","Type":"ContainerStarted","Data":"2aa5dd6f7f35b889fc217f6254342c43f0f6263e5f8ff8068a5f9154f6241bfa"} Oct 08 18:23:02 crc kubenswrapper[4988]: I1008 18:23:02.409817 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-webhook-6cdbc54649-xg4fg" event={"ID":"bb27ea19-b9c4-4ea6-b7b1-6de4ae603b43","Type":"ContainerStarted","Data":"6c9a3075f6ef7c8c6fd84f1bb673d961448259a3e8db720912459748fb29e14f"} Oct 08 18:23:02 crc kubenswrapper[4988]: I1008 18:23:02.409973 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-nmstate/nmstate-webhook-6cdbc54649-xg4fg" Oct 08 18:23:02 crc kubenswrapper[4988]: I1008 18:23:02.424030 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-wh7bx" event={"ID":"670ba223-6bd1-429c-8739-61aea9074914","Type":"ContainerStarted","Data":"019846ce06b90ebda7d63983e2171ef4ae419dc72130746aba03fb6c002a38c0"} Oct 08 18:23:02 crc kubenswrapper[4988]: I1008 18:23:02.426373 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-handler-dv92p" event={"ID":"b0d72824-9655-4a95-9ada-e7d7f8938213","Type":"ContainerStarted","Data":"8f5f7e0147646f6f9decaa9883a096b93f9bb41204a9dfbaee1c597f261e2df1"} Oct 08 18:23:02 crc kubenswrapper[4988]: I1008 18:23:02.426553 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-nmstate/nmstate-handler-dv92p" Oct 08 18:23:02 crc kubenswrapper[4988]: I1008 18:23:02.438979 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-webhook-6cdbc54649-xg4fg" podStartSLOduration=1.501053057 podStartE2EDuration="4.438952033s" podCreationTimestamp="2025-10-08 18:22:58 +0000 UTC" firstStartedPulling="2025-10-08 18:22:59.060769997 +0000 UTC m=+724.510612767" lastFinishedPulling="2025-10-08 18:23:01.998668973 +0000 UTC m=+727.448511743" observedRunningTime="2025-10-08 18:23:02.432712509 +0000 UTC m=+727.882555269" watchObservedRunningTime="2025-10-08 18:23:02.438952033 +0000 UTC m=+727.888794843" Oct 08 18:23:02 crc kubenswrapper[4988]: I1008 18:23:02.480447 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-handler-dv92p" podStartSLOduration=1.346614882 podStartE2EDuration="4.480418347s" podCreationTimestamp="2025-10-08 18:22:58 +0000 UTC" firstStartedPulling="2025-10-08 18:22:58.875693192 +0000 UTC m=+724.325535962" lastFinishedPulling="2025-10-08 18:23:02.009496627 +0000 UTC m=+727.459339427" observedRunningTime="2025-10-08 18:23:02.454680837 +0000 UTC m=+727.904523607" watchObservedRunningTime="2025-10-08 18:23:02.480418347 +0000 UTC m=+727.930261137" Oct 08 18:23:02 crc kubenswrapper[4988]: I1008 18:23:02.485161 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-wh7bx" podStartSLOduration=1.841390424 podStartE2EDuration="4.485139292s" podCreationTimestamp="2025-10-08 18:22:58 +0000 UTC" firstStartedPulling="2025-10-08 18:22:59.355361389 +0000 UTC m=+724.805204159" lastFinishedPulling="2025-10-08 18:23:01.999110257 +0000 UTC m=+727.448953027" observedRunningTime="2025-10-08 18:23:02.482407303 +0000 UTC m=+727.932250093" watchObservedRunningTime="2025-10-08 18:23:02.485139292 +0000 UTC m=+727.934982072" Oct 08 18:23:05 crc kubenswrapper[4988]: I1008 18:23:05.453106 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-b574s" event={"ID":"09b9ef2d-78cf-4bb7-a5ed-79f63f65fa8c","Type":"ContainerStarted","Data":"dcb54c8e58af8d180189fb7f4bbf8315f9145717b72ff611f7e5ec86150861f4"} Oct 08 18:23:05 crc kubenswrapper[4988]: I1008 18:23:05.480885 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-b574s" podStartSLOduration=2.102563854 podStartE2EDuration="7.480855196s" podCreationTimestamp="2025-10-08 18:22:58 +0000 UTC" firstStartedPulling="2025-10-08 18:22:59.009416 +0000 UTC m=+724.459258770" lastFinishedPulling="2025-10-08 18:23:04.387707342 +0000 UTC m=+729.837550112" observedRunningTime="2025-10-08 18:23:05.475144709 +0000 UTC m=+730.924987519" watchObservedRunningTime="2025-10-08 18:23:05.480855196 +0000 UTC m=+730.930697996" Oct 08 18:23:08 crc kubenswrapper[4988]: I1008 18:23:08.865689 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-nmstate/nmstate-handler-dv92p" Oct 08 18:23:09 crc kubenswrapper[4988]: I1008 18:23:09.171096 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-64ff57b4d9-96fmx" Oct 08 18:23:09 crc kubenswrapper[4988]: I1008 18:23:09.171493 4988 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-64ff57b4d9-96fmx" Oct 08 18:23:09 crc kubenswrapper[4988]: I1008 18:23:09.179214 4988 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-64ff57b4d9-96fmx" Oct 08 18:23:09 crc kubenswrapper[4988]: I1008 18:23:09.479707 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-64ff57b4d9-96fmx" Oct 08 18:23:09 crc kubenswrapper[4988]: I1008 18:23:09.538578 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-f9d7485db-7xdnc"] Oct 08 18:23:18 crc kubenswrapper[4988]: I1008 18:23:18.796472 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-nmstate/nmstate-webhook-6cdbc54649-xg4fg" Oct 08 18:23:22 crc kubenswrapper[4988]: I1008 18:23:22.833928 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-sndzf"] Oct 08 18:23:22 crc kubenswrapper[4988]: I1008 18:23:22.839343 4988 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-879f6c89f-sndzf" podUID="a818c760-46b8-4f91-89b3-18e4b0c9a649" containerName="controller-manager" containerID="cri-o://4fa62de363968513075deaaa17c2182ce80008223020f5c529b510a21a17100a" gracePeriod=30 Oct 08 18:23:22 crc kubenswrapper[4988]: I1008 18:23:22.911576 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-gd7f9"] Oct 08 18:23:22 crc kubenswrapper[4988]: I1008 18:23:22.911767 4988 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-gd7f9" podUID="5df90a93-b572-4439-b3aa-2666b2873c1d" containerName="route-controller-manager" containerID="cri-o://e59fc598f1c05950932c9d2b3dc8cd319c7d660adde066d68b723ab1df6201a7" gracePeriod=30 Oct 08 18:23:23 crc kubenswrapper[4988]: I1008 18:23:23.233420 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-sndzf" Oct 08 18:23:23 crc kubenswrapper[4988]: I1008 18:23:23.256126 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-gd7f9" Oct 08 18:23:23 crc kubenswrapper[4988]: I1008 18:23:23.338458 4988 patch_prober.go:28] interesting pod/machine-config-daemon-nm54f container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 18:23:23 crc kubenswrapper[4988]: I1008 18:23:23.338510 4988 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 18:23:23 crc kubenswrapper[4988]: I1008 18:23:23.388854 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/a818c760-46b8-4f91-89b3-18e4b0c9a649-proxy-ca-bundles\") pod \"a818c760-46b8-4f91-89b3-18e4b0c9a649\" (UID: \"a818c760-46b8-4f91-89b3-18e4b0c9a649\") " Oct 08 18:23:23 crc kubenswrapper[4988]: I1008 18:23:23.388931 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5df90a93-b572-4439-b3aa-2666b2873c1d-client-ca\") pod \"5df90a93-b572-4439-b3aa-2666b2873c1d\" (UID: \"5df90a93-b572-4439-b3aa-2666b2873c1d\") " Oct 08 18:23:23 crc kubenswrapper[4988]: I1008 18:23:23.388963 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a818c760-46b8-4f91-89b3-18e4b0c9a649-config\") pod \"a818c760-46b8-4f91-89b3-18e4b0c9a649\" (UID: \"a818c760-46b8-4f91-89b3-18e4b0c9a649\") " Oct 08 18:23:23 crc kubenswrapper[4988]: I1008 18:23:23.388996 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5df90a93-b572-4439-b3aa-2666b2873c1d-serving-cert\") pod \"5df90a93-b572-4439-b3aa-2666b2873c1d\" (UID: \"5df90a93-b572-4439-b3aa-2666b2873c1d\") " Oct 08 18:23:23 crc kubenswrapper[4988]: I1008 18:23:23.389019 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a818c760-46b8-4f91-89b3-18e4b0c9a649-serving-cert\") pod \"a818c760-46b8-4f91-89b3-18e4b0c9a649\" (UID: \"a818c760-46b8-4f91-89b3-18e4b0c9a649\") " Oct 08 18:23:23 crc kubenswrapper[4988]: I1008 18:23:23.389063 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/a818c760-46b8-4f91-89b3-18e4b0c9a649-client-ca\") pod \"a818c760-46b8-4f91-89b3-18e4b0c9a649\" (UID: \"a818c760-46b8-4f91-89b3-18e4b0c9a649\") " Oct 08 18:23:23 crc kubenswrapper[4988]: I1008 18:23:23.389101 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5df90a93-b572-4439-b3aa-2666b2873c1d-config\") pod \"5df90a93-b572-4439-b3aa-2666b2873c1d\" (UID: \"5df90a93-b572-4439-b3aa-2666b2873c1d\") " Oct 08 18:23:23 crc kubenswrapper[4988]: I1008 18:23:23.389127 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s5hnj\" (UniqueName: \"kubernetes.io/projected/5df90a93-b572-4439-b3aa-2666b2873c1d-kube-api-access-s5hnj\") pod \"5df90a93-b572-4439-b3aa-2666b2873c1d\" (UID: \"5df90a93-b572-4439-b3aa-2666b2873c1d\") " Oct 08 18:23:23 crc kubenswrapper[4988]: I1008 18:23:23.389152 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4kxgh\" (UniqueName: \"kubernetes.io/projected/a818c760-46b8-4f91-89b3-18e4b0c9a649-kube-api-access-4kxgh\") pod \"a818c760-46b8-4f91-89b3-18e4b0c9a649\" (UID: \"a818c760-46b8-4f91-89b3-18e4b0c9a649\") " Oct 08 18:23:23 crc kubenswrapper[4988]: I1008 18:23:23.390014 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a818c760-46b8-4f91-89b3-18e4b0c9a649-client-ca" (OuterVolumeSpecName: "client-ca") pod "a818c760-46b8-4f91-89b3-18e4b0c9a649" (UID: "a818c760-46b8-4f91-89b3-18e4b0c9a649"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 18:23:23 crc kubenswrapper[4988]: I1008 18:23:23.390036 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a818c760-46b8-4f91-89b3-18e4b0c9a649-config" (OuterVolumeSpecName: "config") pod "a818c760-46b8-4f91-89b3-18e4b0c9a649" (UID: "a818c760-46b8-4f91-89b3-18e4b0c9a649"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 18:23:23 crc kubenswrapper[4988]: I1008 18:23:23.390082 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5df90a93-b572-4439-b3aa-2666b2873c1d-config" (OuterVolumeSpecName: "config") pod "5df90a93-b572-4439-b3aa-2666b2873c1d" (UID: "5df90a93-b572-4439-b3aa-2666b2873c1d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 18:23:23 crc kubenswrapper[4988]: I1008 18:23:23.390205 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a818c760-46b8-4f91-89b3-18e4b0c9a649-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "a818c760-46b8-4f91-89b3-18e4b0c9a649" (UID: "a818c760-46b8-4f91-89b3-18e4b0c9a649"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 18:23:23 crc kubenswrapper[4988]: I1008 18:23:23.390254 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5df90a93-b572-4439-b3aa-2666b2873c1d-client-ca" (OuterVolumeSpecName: "client-ca") pod "5df90a93-b572-4439-b3aa-2666b2873c1d" (UID: "5df90a93-b572-4439-b3aa-2666b2873c1d"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 18:23:23 crc kubenswrapper[4988]: I1008 18:23:23.395220 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5df90a93-b572-4439-b3aa-2666b2873c1d-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "5df90a93-b572-4439-b3aa-2666b2873c1d" (UID: "5df90a93-b572-4439-b3aa-2666b2873c1d"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:23:23 crc kubenswrapper[4988]: I1008 18:23:23.395266 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a818c760-46b8-4f91-89b3-18e4b0c9a649-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "a818c760-46b8-4f91-89b3-18e4b0c9a649" (UID: "a818c760-46b8-4f91-89b3-18e4b0c9a649"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:23:23 crc kubenswrapper[4988]: I1008 18:23:23.395353 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a818c760-46b8-4f91-89b3-18e4b0c9a649-kube-api-access-4kxgh" (OuterVolumeSpecName: "kube-api-access-4kxgh") pod "a818c760-46b8-4f91-89b3-18e4b0c9a649" (UID: "a818c760-46b8-4f91-89b3-18e4b0c9a649"). InnerVolumeSpecName "kube-api-access-4kxgh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:23:23 crc kubenswrapper[4988]: I1008 18:23:23.395893 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5df90a93-b572-4439-b3aa-2666b2873c1d-kube-api-access-s5hnj" (OuterVolumeSpecName: "kube-api-access-s5hnj") pod "5df90a93-b572-4439-b3aa-2666b2873c1d" (UID: "5df90a93-b572-4439-b3aa-2666b2873c1d"). InnerVolumeSpecName "kube-api-access-s5hnj". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:23:23 crc kubenswrapper[4988]: I1008 18:23:23.490376 4988 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/a818c760-46b8-4f91-89b3-18e4b0c9a649-client-ca\") on node \"crc\" DevicePath \"\"" Oct 08 18:23:23 crc kubenswrapper[4988]: I1008 18:23:23.490428 4988 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5df90a93-b572-4439-b3aa-2666b2873c1d-config\") on node \"crc\" DevicePath \"\"" Oct 08 18:23:23 crc kubenswrapper[4988]: I1008 18:23:23.490441 4988 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s5hnj\" (UniqueName: \"kubernetes.io/projected/5df90a93-b572-4439-b3aa-2666b2873c1d-kube-api-access-s5hnj\") on node \"crc\" DevicePath \"\"" Oct 08 18:23:23 crc kubenswrapper[4988]: I1008 18:23:23.490455 4988 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4kxgh\" (UniqueName: \"kubernetes.io/projected/a818c760-46b8-4f91-89b3-18e4b0c9a649-kube-api-access-4kxgh\") on node \"crc\" DevicePath \"\"" Oct 08 18:23:23 crc kubenswrapper[4988]: I1008 18:23:23.490466 4988 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/a818c760-46b8-4f91-89b3-18e4b0c9a649-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Oct 08 18:23:23 crc kubenswrapper[4988]: I1008 18:23:23.490478 4988 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5df90a93-b572-4439-b3aa-2666b2873c1d-client-ca\") on node \"crc\" DevicePath \"\"" Oct 08 18:23:23 crc kubenswrapper[4988]: I1008 18:23:23.490489 4988 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a818c760-46b8-4f91-89b3-18e4b0c9a649-config\") on node \"crc\" DevicePath \"\"" Oct 08 18:23:23 crc kubenswrapper[4988]: I1008 18:23:23.490499 4988 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5df90a93-b572-4439-b3aa-2666b2873c1d-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 08 18:23:23 crc kubenswrapper[4988]: I1008 18:23:23.490510 4988 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a818c760-46b8-4f91-89b3-18e4b0c9a649-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 08 18:23:23 crc kubenswrapper[4988]: I1008 18:23:23.556949 4988 generic.go:334] "Generic (PLEG): container finished" podID="5df90a93-b572-4439-b3aa-2666b2873c1d" containerID="e59fc598f1c05950932c9d2b3dc8cd319c7d660adde066d68b723ab1df6201a7" exitCode=0 Oct 08 18:23:23 crc kubenswrapper[4988]: I1008 18:23:23.557020 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-gd7f9" event={"ID":"5df90a93-b572-4439-b3aa-2666b2873c1d","Type":"ContainerDied","Data":"e59fc598f1c05950932c9d2b3dc8cd319c7d660adde066d68b723ab1df6201a7"} Oct 08 18:23:23 crc kubenswrapper[4988]: I1008 18:23:23.557051 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-gd7f9" event={"ID":"5df90a93-b572-4439-b3aa-2666b2873c1d","Type":"ContainerDied","Data":"b20de68b1bde58d505848ab392d577b41545089a287c6b13a4dab4070e060506"} Oct 08 18:23:23 crc kubenswrapper[4988]: I1008 18:23:23.557071 4988 scope.go:117] "RemoveContainer" containerID="e59fc598f1c05950932c9d2b3dc8cd319c7d660adde066d68b723ab1df6201a7" Oct 08 18:23:23 crc kubenswrapper[4988]: I1008 18:23:23.557171 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-gd7f9" Oct 08 18:23:23 crc kubenswrapper[4988]: I1008 18:23:23.559810 4988 generic.go:334] "Generic (PLEG): container finished" podID="a818c760-46b8-4f91-89b3-18e4b0c9a649" containerID="4fa62de363968513075deaaa17c2182ce80008223020f5c529b510a21a17100a" exitCode=0 Oct 08 18:23:23 crc kubenswrapper[4988]: I1008 18:23:23.559848 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-sndzf" event={"ID":"a818c760-46b8-4f91-89b3-18e4b0c9a649","Type":"ContainerDied","Data":"4fa62de363968513075deaaa17c2182ce80008223020f5c529b510a21a17100a"} Oct 08 18:23:23 crc kubenswrapper[4988]: I1008 18:23:23.559876 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-sndzf" event={"ID":"a818c760-46b8-4f91-89b3-18e4b0c9a649","Type":"ContainerDied","Data":"cbe9e4880bb77fc7c479bd686e48cd1f92910ef87dc500bbd134a928ad1c55fb"} Oct 08 18:23:23 crc kubenswrapper[4988]: I1008 18:23:23.559928 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-sndzf" Oct 08 18:23:23 crc kubenswrapper[4988]: I1008 18:23:23.588656 4988 scope.go:117] "RemoveContainer" containerID="e59fc598f1c05950932c9d2b3dc8cd319c7d660adde066d68b723ab1df6201a7" Oct 08 18:23:23 crc kubenswrapper[4988]: E1008 18:23:23.589052 4988 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e59fc598f1c05950932c9d2b3dc8cd319c7d660adde066d68b723ab1df6201a7\": container with ID starting with e59fc598f1c05950932c9d2b3dc8cd319c7d660adde066d68b723ab1df6201a7 not found: ID does not exist" containerID="e59fc598f1c05950932c9d2b3dc8cd319c7d660adde066d68b723ab1df6201a7" Oct 08 18:23:23 crc kubenswrapper[4988]: I1008 18:23:23.589084 4988 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e59fc598f1c05950932c9d2b3dc8cd319c7d660adde066d68b723ab1df6201a7"} err="failed to get container status \"e59fc598f1c05950932c9d2b3dc8cd319c7d660adde066d68b723ab1df6201a7\": rpc error: code = NotFound desc = could not find container \"e59fc598f1c05950932c9d2b3dc8cd319c7d660adde066d68b723ab1df6201a7\": container with ID starting with e59fc598f1c05950932c9d2b3dc8cd319c7d660adde066d68b723ab1df6201a7 not found: ID does not exist" Oct 08 18:23:23 crc kubenswrapper[4988]: I1008 18:23:23.589104 4988 scope.go:117] "RemoveContainer" containerID="4fa62de363968513075deaaa17c2182ce80008223020f5c529b510a21a17100a" Oct 08 18:23:23 crc kubenswrapper[4988]: I1008 18:23:23.595737 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-sndzf"] Oct 08 18:23:23 crc kubenswrapper[4988]: I1008 18:23:23.602027 4988 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-sndzf"] Oct 08 18:23:23 crc kubenswrapper[4988]: I1008 18:23:23.606896 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-gd7f9"] Oct 08 18:23:23 crc kubenswrapper[4988]: I1008 18:23:23.609477 4988 scope.go:117] "RemoveContainer" containerID="4fa62de363968513075deaaa17c2182ce80008223020f5c529b510a21a17100a" Oct 08 18:23:23 crc kubenswrapper[4988]: I1008 18:23:23.610095 4988 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-gd7f9"] Oct 08 18:23:23 crc kubenswrapper[4988]: E1008 18:23:23.610148 4988 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4fa62de363968513075deaaa17c2182ce80008223020f5c529b510a21a17100a\": container with ID starting with 4fa62de363968513075deaaa17c2182ce80008223020f5c529b510a21a17100a not found: ID does not exist" containerID="4fa62de363968513075deaaa17c2182ce80008223020f5c529b510a21a17100a" Oct 08 18:23:23 crc kubenswrapper[4988]: I1008 18:23:23.610205 4988 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4fa62de363968513075deaaa17c2182ce80008223020f5c529b510a21a17100a"} err="failed to get container status \"4fa62de363968513075deaaa17c2182ce80008223020f5c529b510a21a17100a\": rpc error: code = NotFound desc = could not find container \"4fa62de363968513075deaaa17c2182ce80008223020f5c529b510a21a17100a\": container with ID starting with 4fa62de363968513075deaaa17c2182ce80008223020f5c529b510a21a17100a not found: ID does not exist" Oct 08 18:23:24 crc kubenswrapper[4988]: I1008 18:23:24.208013 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-5d45c7689b-kkxvv"] Oct 08 18:23:24 crc kubenswrapper[4988]: E1008 18:23:24.208474 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a818c760-46b8-4f91-89b3-18e4b0c9a649" containerName="controller-manager" Oct 08 18:23:24 crc kubenswrapper[4988]: I1008 18:23:24.208501 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="a818c760-46b8-4f91-89b3-18e4b0c9a649" containerName="controller-manager" Oct 08 18:23:24 crc kubenswrapper[4988]: E1008 18:23:24.208526 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5df90a93-b572-4439-b3aa-2666b2873c1d" containerName="route-controller-manager" Oct 08 18:23:24 crc kubenswrapper[4988]: I1008 18:23:24.208545 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="5df90a93-b572-4439-b3aa-2666b2873c1d" containerName="route-controller-manager" Oct 08 18:23:24 crc kubenswrapper[4988]: I1008 18:23:24.209082 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="a818c760-46b8-4f91-89b3-18e4b0c9a649" containerName="controller-manager" Oct 08 18:23:24 crc kubenswrapper[4988]: I1008 18:23:24.209134 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="5df90a93-b572-4439-b3aa-2666b2873c1d" containerName="route-controller-manager" Oct 08 18:23:24 crc kubenswrapper[4988]: I1008 18:23:24.209898 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-5d45c7689b-kkxvv" Oct 08 18:23:24 crc kubenswrapper[4988]: I1008 18:23:24.214773 4988 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Oct 08 18:23:24 crc kubenswrapper[4988]: I1008 18:23:24.215222 4988 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Oct 08 18:23:24 crc kubenswrapper[4988]: I1008 18:23:24.215261 4988 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Oct 08 18:23:24 crc kubenswrapper[4988]: I1008 18:23:24.216436 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Oct 08 18:23:24 crc kubenswrapper[4988]: I1008 18:23:24.216937 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-fc4d6cc7f-6hvkj"] Oct 08 18:23:24 crc kubenswrapper[4988]: I1008 18:23:24.218285 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-fc4d6cc7f-6hvkj" Oct 08 18:23:24 crc kubenswrapper[4988]: I1008 18:23:24.220870 4988 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Oct 08 18:23:24 crc kubenswrapper[4988]: I1008 18:23:24.221932 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Oct 08 18:23:24 crc kubenswrapper[4988]: I1008 18:23:24.224023 4988 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Oct 08 18:23:24 crc kubenswrapper[4988]: I1008 18:23:24.224580 4988 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Oct 08 18:23:24 crc kubenswrapper[4988]: I1008 18:23:24.224753 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Oct 08 18:23:24 crc kubenswrapper[4988]: I1008 18:23:24.224987 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Oct 08 18:23:24 crc kubenswrapper[4988]: I1008 18:23:24.225120 4988 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Oct 08 18:23:24 crc kubenswrapper[4988]: I1008 18:23:24.225309 4988 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Oct 08 18:23:24 crc kubenswrapper[4988]: I1008 18:23:24.227681 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-5d45c7689b-kkxvv"] Oct 08 18:23:24 crc kubenswrapper[4988]: I1008 18:23:24.229896 4988 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Oct 08 18:23:24 crc kubenswrapper[4988]: I1008 18:23:24.242864 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-fc4d6cc7f-6hvkj"] Oct 08 18:23:24 crc kubenswrapper[4988]: I1008 18:23:24.300849 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/9d4ca14f-7dd1-487d-8b08-12a66a0ac79f-proxy-ca-bundles\") pod \"controller-manager-5d45c7689b-kkxvv\" (UID: \"9d4ca14f-7dd1-487d-8b08-12a66a0ac79f\") " pod="openshift-controller-manager/controller-manager-5d45c7689b-kkxvv" Oct 08 18:23:24 crc kubenswrapper[4988]: I1008 18:23:24.300944 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/61878ed7-2593-41ca-b8e9-1746d5ea3527-config\") pod \"route-controller-manager-fc4d6cc7f-6hvkj\" (UID: \"61878ed7-2593-41ca-b8e9-1746d5ea3527\") " pod="openshift-route-controller-manager/route-controller-manager-fc4d6cc7f-6hvkj" Oct 08 18:23:24 crc kubenswrapper[4988]: I1008 18:23:24.300962 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/61878ed7-2593-41ca-b8e9-1746d5ea3527-serving-cert\") pod \"route-controller-manager-fc4d6cc7f-6hvkj\" (UID: \"61878ed7-2593-41ca-b8e9-1746d5ea3527\") " pod="openshift-route-controller-manager/route-controller-manager-fc4d6cc7f-6hvkj" Oct 08 18:23:24 crc kubenswrapper[4988]: I1008 18:23:24.300976 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4ca14f-7dd1-487d-8b08-12a66a0ac79f-serving-cert\") pod \"controller-manager-5d45c7689b-kkxvv\" (UID: \"9d4ca14f-7dd1-487d-8b08-12a66a0ac79f\") " pod="openshift-controller-manager/controller-manager-5d45c7689b-kkxvv" Oct 08 18:23:24 crc kubenswrapper[4988]: I1008 18:23:24.300999 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/61878ed7-2593-41ca-b8e9-1746d5ea3527-client-ca\") pod \"route-controller-manager-fc4d6cc7f-6hvkj\" (UID: \"61878ed7-2593-41ca-b8e9-1746d5ea3527\") " pod="openshift-route-controller-manager/route-controller-manager-fc4d6cc7f-6hvkj" Oct 08 18:23:24 crc kubenswrapper[4988]: I1008 18:23:24.301050 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gv279\" (UniqueName: \"kubernetes.io/projected/9d4ca14f-7dd1-487d-8b08-12a66a0ac79f-kube-api-access-gv279\") pod \"controller-manager-5d45c7689b-kkxvv\" (UID: \"9d4ca14f-7dd1-487d-8b08-12a66a0ac79f\") " pod="openshift-controller-manager/controller-manager-5d45c7689b-kkxvv" Oct 08 18:23:24 crc kubenswrapper[4988]: I1008 18:23:24.301079 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/9d4ca14f-7dd1-487d-8b08-12a66a0ac79f-client-ca\") pod \"controller-manager-5d45c7689b-kkxvv\" (UID: \"9d4ca14f-7dd1-487d-8b08-12a66a0ac79f\") " pod="openshift-controller-manager/controller-manager-5d45c7689b-kkxvv" Oct 08 18:23:24 crc kubenswrapper[4988]: I1008 18:23:24.301131 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4ca14f-7dd1-487d-8b08-12a66a0ac79f-config\") pod \"controller-manager-5d45c7689b-kkxvv\" (UID: \"9d4ca14f-7dd1-487d-8b08-12a66a0ac79f\") " pod="openshift-controller-manager/controller-manager-5d45c7689b-kkxvv" Oct 08 18:23:24 crc kubenswrapper[4988]: I1008 18:23:24.301156 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l8rx5\" (UniqueName: \"kubernetes.io/projected/61878ed7-2593-41ca-b8e9-1746d5ea3527-kube-api-access-l8rx5\") pod \"route-controller-manager-fc4d6cc7f-6hvkj\" (UID: \"61878ed7-2593-41ca-b8e9-1746d5ea3527\") " pod="openshift-route-controller-manager/route-controller-manager-fc4d6cc7f-6hvkj" Oct 08 18:23:24 crc kubenswrapper[4988]: I1008 18:23:24.402774 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gv279\" (UniqueName: \"kubernetes.io/projected/9d4ca14f-7dd1-487d-8b08-12a66a0ac79f-kube-api-access-gv279\") pod \"controller-manager-5d45c7689b-kkxvv\" (UID: \"9d4ca14f-7dd1-487d-8b08-12a66a0ac79f\") " pod="openshift-controller-manager/controller-manager-5d45c7689b-kkxvv" Oct 08 18:23:24 crc kubenswrapper[4988]: I1008 18:23:24.402844 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/9d4ca14f-7dd1-487d-8b08-12a66a0ac79f-client-ca\") pod \"controller-manager-5d45c7689b-kkxvv\" (UID: \"9d4ca14f-7dd1-487d-8b08-12a66a0ac79f\") " pod="openshift-controller-manager/controller-manager-5d45c7689b-kkxvv" Oct 08 18:23:24 crc kubenswrapper[4988]: I1008 18:23:24.402885 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4ca14f-7dd1-487d-8b08-12a66a0ac79f-config\") pod \"controller-manager-5d45c7689b-kkxvv\" (UID: \"9d4ca14f-7dd1-487d-8b08-12a66a0ac79f\") " pod="openshift-controller-manager/controller-manager-5d45c7689b-kkxvv" Oct 08 18:23:24 crc kubenswrapper[4988]: I1008 18:23:24.402920 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l8rx5\" (UniqueName: \"kubernetes.io/projected/61878ed7-2593-41ca-b8e9-1746d5ea3527-kube-api-access-l8rx5\") pod \"route-controller-manager-fc4d6cc7f-6hvkj\" (UID: \"61878ed7-2593-41ca-b8e9-1746d5ea3527\") " pod="openshift-route-controller-manager/route-controller-manager-fc4d6cc7f-6hvkj" Oct 08 18:23:24 crc kubenswrapper[4988]: I1008 18:23:24.402946 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/9d4ca14f-7dd1-487d-8b08-12a66a0ac79f-proxy-ca-bundles\") pod \"controller-manager-5d45c7689b-kkxvv\" (UID: \"9d4ca14f-7dd1-487d-8b08-12a66a0ac79f\") " pod="openshift-controller-manager/controller-manager-5d45c7689b-kkxvv" Oct 08 18:23:24 crc kubenswrapper[4988]: I1008 18:23:24.402976 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/61878ed7-2593-41ca-b8e9-1746d5ea3527-config\") pod \"route-controller-manager-fc4d6cc7f-6hvkj\" (UID: \"61878ed7-2593-41ca-b8e9-1746d5ea3527\") " pod="openshift-route-controller-manager/route-controller-manager-fc4d6cc7f-6hvkj" Oct 08 18:23:24 crc kubenswrapper[4988]: I1008 18:23:24.402996 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/61878ed7-2593-41ca-b8e9-1746d5ea3527-serving-cert\") pod \"route-controller-manager-fc4d6cc7f-6hvkj\" (UID: \"61878ed7-2593-41ca-b8e9-1746d5ea3527\") " pod="openshift-route-controller-manager/route-controller-manager-fc4d6cc7f-6hvkj" Oct 08 18:23:24 crc kubenswrapper[4988]: I1008 18:23:24.403014 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4ca14f-7dd1-487d-8b08-12a66a0ac79f-serving-cert\") pod \"controller-manager-5d45c7689b-kkxvv\" (UID: \"9d4ca14f-7dd1-487d-8b08-12a66a0ac79f\") " pod="openshift-controller-manager/controller-manager-5d45c7689b-kkxvv" Oct 08 18:23:24 crc kubenswrapper[4988]: I1008 18:23:24.403037 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/61878ed7-2593-41ca-b8e9-1746d5ea3527-client-ca\") pod \"route-controller-manager-fc4d6cc7f-6hvkj\" (UID: \"61878ed7-2593-41ca-b8e9-1746d5ea3527\") " pod="openshift-route-controller-manager/route-controller-manager-fc4d6cc7f-6hvkj" Oct 08 18:23:24 crc kubenswrapper[4988]: I1008 18:23:24.403947 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/61878ed7-2593-41ca-b8e9-1746d5ea3527-client-ca\") pod \"route-controller-manager-fc4d6cc7f-6hvkj\" (UID: \"61878ed7-2593-41ca-b8e9-1746d5ea3527\") " pod="openshift-route-controller-manager/route-controller-manager-fc4d6cc7f-6hvkj" Oct 08 18:23:24 crc kubenswrapper[4988]: I1008 18:23:24.404784 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/9d4ca14f-7dd1-487d-8b08-12a66a0ac79f-client-ca\") pod \"controller-manager-5d45c7689b-kkxvv\" (UID: \"9d4ca14f-7dd1-487d-8b08-12a66a0ac79f\") " pod="openshift-controller-manager/controller-manager-5d45c7689b-kkxvv" Oct 08 18:23:24 crc kubenswrapper[4988]: I1008 18:23:24.405106 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/9d4ca14f-7dd1-487d-8b08-12a66a0ac79f-proxy-ca-bundles\") pod \"controller-manager-5d45c7689b-kkxvv\" (UID: \"9d4ca14f-7dd1-487d-8b08-12a66a0ac79f\") " pod="openshift-controller-manager/controller-manager-5d45c7689b-kkxvv" Oct 08 18:23:24 crc kubenswrapper[4988]: I1008 18:23:24.407840 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/61878ed7-2593-41ca-b8e9-1746d5ea3527-config\") pod \"route-controller-manager-fc4d6cc7f-6hvkj\" (UID: \"61878ed7-2593-41ca-b8e9-1746d5ea3527\") " pod="openshift-route-controller-manager/route-controller-manager-fc4d6cc7f-6hvkj" Oct 08 18:23:24 crc kubenswrapper[4988]: I1008 18:23:24.409449 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4ca14f-7dd1-487d-8b08-12a66a0ac79f-config\") pod \"controller-manager-5d45c7689b-kkxvv\" (UID: \"9d4ca14f-7dd1-487d-8b08-12a66a0ac79f\") " pod="openshift-controller-manager/controller-manager-5d45c7689b-kkxvv" Oct 08 18:23:24 crc kubenswrapper[4988]: I1008 18:23:24.413420 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4ca14f-7dd1-487d-8b08-12a66a0ac79f-serving-cert\") pod \"controller-manager-5d45c7689b-kkxvv\" (UID: \"9d4ca14f-7dd1-487d-8b08-12a66a0ac79f\") " pod="openshift-controller-manager/controller-manager-5d45c7689b-kkxvv" Oct 08 18:23:24 crc kubenswrapper[4988]: I1008 18:23:24.414428 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/61878ed7-2593-41ca-b8e9-1746d5ea3527-serving-cert\") pod \"route-controller-manager-fc4d6cc7f-6hvkj\" (UID: \"61878ed7-2593-41ca-b8e9-1746d5ea3527\") " pod="openshift-route-controller-manager/route-controller-manager-fc4d6cc7f-6hvkj" Oct 08 18:23:24 crc kubenswrapper[4988]: I1008 18:23:24.424091 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gv279\" (UniqueName: \"kubernetes.io/projected/9d4ca14f-7dd1-487d-8b08-12a66a0ac79f-kube-api-access-gv279\") pod \"controller-manager-5d45c7689b-kkxvv\" (UID: \"9d4ca14f-7dd1-487d-8b08-12a66a0ac79f\") " pod="openshift-controller-manager/controller-manager-5d45c7689b-kkxvv" Oct 08 18:23:24 crc kubenswrapper[4988]: I1008 18:23:24.439704 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l8rx5\" (UniqueName: \"kubernetes.io/projected/61878ed7-2593-41ca-b8e9-1746d5ea3527-kube-api-access-l8rx5\") pod \"route-controller-manager-fc4d6cc7f-6hvkj\" (UID: \"61878ed7-2593-41ca-b8e9-1746d5ea3527\") " pod="openshift-route-controller-manager/route-controller-manager-fc4d6cc7f-6hvkj" Oct 08 18:23:24 crc kubenswrapper[4988]: I1008 18:23:24.538084 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-5d45c7689b-kkxvv" Oct 08 18:23:24 crc kubenswrapper[4988]: I1008 18:23:24.550809 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-fc4d6cc7f-6hvkj" Oct 08 18:23:24 crc kubenswrapper[4988]: I1008 18:23:24.764719 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-5d45c7689b-kkxvv"] Oct 08 18:23:24 crc kubenswrapper[4988]: W1008 18:23:24.780013 4988 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9d4ca14f_7dd1_487d_8b08_12a66a0ac79f.slice/crio-6b0f48e895916494c5d6ec8973a43130b341b90caf23d873ddddf59b20f0f128 WatchSource:0}: Error finding container 6b0f48e895916494c5d6ec8973a43130b341b90caf23d873ddddf59b20f0f128: Status 404 returned error can't find the container with id 6b0f48e895916494c5d6ec8973a43130b341b90caf23d873ddddf59b20f0f128 Oct 08 18:23:24 crc kubenswrapper[4988]: I1008 18:23:24.799646 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-fc4d6cc7f-6hvkj"] Oct 08 18:23:24 crc kubenswrapper[4988]: W1008 18:23:24.810583 4988 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod61878ed7_2593_41ca_b8e9_1746d5ea3527.slice/crio-b854183c27e3d3ffec66671c4c80e7eb878a2e720b28734ce45c421219960adb WatchSource:0}: Error finding container b854183c27e3d3ffec66671c4c80e7eb878a2e720b28734ce45c421219960adb: Status 404 returned error can't find the container with id b854183c27e3d3ffec66671c4c80e7eb878a2e720b28734ce45c421219960adb Oct 08 18:23:25 crc kubenswrapper[4988]: I1008 18:23:25.245592 4988 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5df90a93-b572-4439-b3aa-2666b2873c1d" path="/var/lib/kubelet/pods/5df90a93-b572-4439-b3aa-2666b2873c1d/volumes" Oct 08 18:23:25 crc kubenswrapper[4988]: I1008 18:23:25.246590 4988 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a818c760-46b8-4f91-89b3-18e4b0c9a649" path="/var/lib/kubelet/pods/a818c760-46b8-4f91-89b3-18e4b0c9a649/volumes" Oct 08 18:23:25 crc kubenswrapper[4988]: I1008 18:23:25.576019 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-5d45c7689b-kkxvv" event={"ID":"9d4ca14f-7dd1-487d-8b08-12a66a0ac79f","Type":"ContainerStarted","Data":"e1fe876f523a99df112865323398b51c8f18fc4aee29e6563957d848bfb0e433"} Oct 08 18:23:25 crc kubenswrapper[4988]: I1008 18:23:25.576061 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-5d45c7689b-kkxvv" event={"ID":"9d4ca14f-7dd1-487d-8b08-12a66a0ac79f","Type":"ContainerStarted","Data":"6b0f48e895916494c5d6ec8973a43130b341b90caf23d873ddddf59b20f0f128"} Oct 08 18:23:25 crc kubenswrapper[4988]: I1008 18:23:25.576173 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-5d45c7689b-kkxvv" Oct 08 18:23:25 crc kubenswrapper[4988]: I1008 18:23:25.577390 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-fc4d6cc7f-6hvkj" event={"ID":"61878ed7-2593-41ca-b8e9-1746d5ea3527","Type":"ContainerStarted","Data":"ef781b70d57256d4a64c6ea2b2f614109636e4c0d141569b3baba1f2b59dc256"} Oct 08 18:23:25 crc kubenswrapper[4988]: I1008 18:23:25.577413 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-fc4d6cc7f-6hvkj" event={"ID":"61878ed7-2593-41ca-b8e9-1746d5ea3527","Type":"ContainerStarted","Data":"b854183c27e3d3ffec66671c4c80e7eb878a2e720b28734ce45c421219960adb"} Oct 08 18:23:25 crc kubenswrapper[4988]: I1008 18:23:25.577559 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-fc4d6cc7f-6hvkj" Oct 08 18:23:25 crc kubenswrapper[4988]: I1008 18:23:25.582015 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-5d45c7689b-kkxvv" Oct 08 18:23:25 crc kubenswrapper[4988]: I1008 18:23:25.595356 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-5d45c7689b-kkxvv" podStartSLOduration=3.595342751 podStartE2EDuration="3.595342751s" podCreationTimestamp="2025-10-08 18:23:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 18:23:25.593189821 +0000 UTC m=+751.043032591" watchObservedRunningTime="2025-10-08 18:23:25.595342751 +0000 UTC m=+751.045185521" Oct 08 18:23:25 crc kubenswrapper[4988]: I1008 18:23:25.616188 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-fc4d6cc7f-6hvkj" podStartSLOduration=3.616173351 podStartE2EDuration="3.616173351s" podCreationTimestamp="2025-10-08 18:23:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 18:23:25.613968329 +0000 UTC m=+751.063811129" watchObservedRunningTime="2025-10-08 18:23:25.616173351 +0000 UTC m=+751.066016121" Oct 08 18:23:26 crc kubenswrapper[4988]: I1008 18:23:26.094837 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-fc4d6cc7f-6hvkj" Oct 08 18:23:30 crc kubenswrapper[4988]: I1008 18:23:30.528287 4988 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Oct 08 18:23:34 crc kubenswrapper[4988]: I1008 18:23:34.576989 4988 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-console/console-f9d7485db-7xdnc" podUID="9c8874ed-3f70-45b6-9f10-d8737b6a7a8f" containerName="console" containerID="cri-o://8decb88999dd5ec9d72091c408c71df63d4adcf20acd387e7c499f07211eb8fc" gracePeriod=15 Oct 08 18:23:34 crc kubenswrapper[4988]: I1008 18:23:34.774645 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d28pl68"] Oct 08 18:23:34 crc kubenswrapper[4988]: I1008 18:23:34.776029 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d28pl68" Oct 08 18:23:34 crc kubenswrapper[4988]: I1008 18:23:34.777768 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Oct 08 18:23:34 crc kubenswrapper[4988]: I1008 18:23:34.784592 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d28pl68"] Oct 08 18:23:34 crc kubenswrapper[4988]: I1008 18:23:34.841650 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/fdfe229e-0b1f-4af5-8c9b-2d1ed4c64fcb-util\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d28pl68\" (UID: \"fdfe229e-0b1f-4af5-8c9b-2d1ed4c64fcb\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d28pl68" Oct 08 18:23:34 crc kubenswrapper[4988]: I1008 18:23:34.841702 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/fdfe229e-0b1f-4af5-8c9b-2d1ed4c64fcb-bundle\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d28pl68\" (UID: \"fdfe229e-0b1f-4af5-8c9b-2d1ed4c64fcb\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d28pl68" Oct 08 18:23:34 crc kubenswrapper[4988]: I1008 18:23:34.841980 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lvtxw\" (UniqueName: \"kubernetes.io/projected/fdfe229e-0b1f-4af5-8c9b-2d1ed4c64fcb-kube-api-access-lvtxw\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d28pl68\" (UID: \"fdfe229e-0b1f-4af5-8c9b-2d1ed4c64fcb\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d28pl68" Oct 08 18:23:34 crc kubenswrapper[4988]: I1008 18:23:34.943332 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lvtxw\" (UniqueName: \"kubernetes.io/projected/fdfe229e-0b1f-4af5-8c9b-2d1ed4c64fcb-kube-api-access-lvtxw\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d28pl68\" (UID: \"fdfe229e-0b1f-4af5-8c9b-2d1ed4c64fcb\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d28pl68" Oct 08 18:23:34 crc kubenswrapper[4988]: I1008 18:23:34.943647 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/fdfe229e-0b1f-4af5-8c9b-2d1ed4c64fcb-util\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d28pl68\" (UID: \"fdfe229e-0b1f-4af5-8c9b-2d1ed4c64fcb\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d28pl68" Oct 08 18:23:34 crc kubenswrapper[4988]: I1008 18:23:34.943740 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/fdfe229e-0b1f-4af5-8c9b-2d1ed4c64fcb-bundle\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d28pl68\" (UID: \"fdfe229e-0b1f-4af5-8c9b-2d1ed4c64fcb\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d28pl68" Oct 08 18:23:34 crc kubenswrapper[4988]: I1008 18:23:34.944457 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/fdfe229e-0b1f-4af5-8c9b-2d1ed4c64fcb-bundle\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d28pl68\" (UID: \"fdfe229e-0b1f-4af5-8c9b-2d1ed4c64fcb\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d28pl68" Oct 08 18:23:34 crc kubenswrapper[4988]: I1008 18:23:34.944547 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/fdfe229e-0b1f-4af5-8c9b-2d1ed4c64fcb-util\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d28pl68\" (UID: \"fdfe229e-0b1f-4af5-8c9b-2d1ed4c64fcb\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d28pl68" Oct 08 18:23:34 crc kubenswrapper[4988]: I1008 18:23:34.963404 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lvtxw\" (UniqueName: \"kubernetes.io/projected/fdfe229e-0b1f-4af5-8c9b-2d1ed4c64fcb-kube-api-access-lvtxw\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d28pl68\" (UID: \"fdfe229e-0b1f-4af5-8c9b-2d1ed4c64fcb\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d28pl68" Oct 08 18:23:34 crc kubenswrapper[4988]: I1008 18:23:34.996355 4988 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-7xdnc_9c8874ed-3f70-45b6-9f10-d8737b6a7a8f/console/0.log" Oct 08 18:23:34 crc kubenswrapper[4988]: I1008 18:23:34.996455 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-7xdnc" Oct 08 18:23:35 crc kubenswrapper[4988]: I1008 18:23:35.092654 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d28pl68" Oct 08 18:23:35 crc kubenswrapper[4988]: I1008 18:23:35.145310 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/9c8874ed-3f70-45b6-9f10-d8737b6a7a8f-trusted-ca-bundle\") pod \"9c8874ed-3f70-45b6-9f10-d8737b6a7a8f\" (UID: \"9c8874ed-3f70-45b6-9f10-d8737b6a7a8f\") " Oct 08 18:23:35 crc kubenswrapper[4988]: I1008 18:23:35.145358 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/9c8874ed-3f70-45b6-9f10-d8737b6a7a8f-service-ca\") pod \"9c8874ed-3f70-45b6-9f10-d8737b6a7a8f\" (UID: \"9c8874ed-3f70-45b6-9f10-d8737b6a7a8f\") " Oct 08 18:23:35 crc kubenswrapper[4988]: I1008 18:23:35.145402 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/9c8874ed-3f70-45b6-9f10-d8737b6a7a8f-console-serving-cert\") pod \"9c8874ed-3f70-45b6-9f10-d8737b6a7a8f\" (UID: \"9c8874ed-3f70-45b6-9f10-d8737b6a7a8f\") " Oct 08 18:23:35 crc kubenswrapper[4988]: I1008 18:23:35.145426 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/9c8874ed-3f70-45b6-9f10-d8737b6a7a8f-console-config\") pod \"9c8874ed-3f70-45b6-9f10-d8737b6a7a8f\" (UID: \"9c8874ed-3f70-45b6-9f10-d8737b6a7a8f\") " Oct 08 18:23:35 crc kubenswrapper[4988]: I1008 18:23:35.145441 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/9c8874ed-3f70-45b6-9f10-d8737b6a7a8f-console-oauth-config\") pod \"9c8874ed-3f70-45b6-9f10-d8737b6a7a8f\" (UID: \"9c8874ed-3f70-45b6-9f10-d8737b6a7a8f\") " Oct 08 18:23:35 crc kubenswrapper[4988]: I1008 18:23:35.145467 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/9c8874ed-3f70-45b6-9f10-d8737b6a7a8f-oauth-serving-cert\") pod \"9c8874ed-3f70-45b6-9f10-d8737b6a7a8f\" (UID: \"9c8874ed-3f70-45b6-9f10-d8737b6a7a8f\") " Oct 08 18:23:35 crc kubenswrapper[4988]: I1008 18:23:35.145521 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qhgxr\" (UniqueName: \"kubernetes.io/projected/9c8874ed-3f70-45b6-9f10-d8737b6a7a8f-kube-api-access-qhgxr\") pod \"9c8874ed-3f70-45b6-9f10-d8737b6a7a8f\" (UID: \"9c8874ed-3f70-45b6-9f10-d8737b6a7a8f\") " Oct 08 18:23:35 crc kubenswrapper[4988]: I1008 18:23:35.146586 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9c8874ed-3f70-45b6-9f10-d8737b6a7a8f-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "9c8874ed-3f70-45b6-9f10-d8737b6a7a8f" (UID: "9c8874ed-3f70-45b6-9f10-d8737b6a7a8f"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 18:23:35 crc kubenswrapper[4988]: I1008 18:23:35.146592 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9c8874ed-3f70-45b6-9f10-d8737b6a7a8f-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "9c8874ed-3f70-45b6-9f10-d8737b6a7a8f" (UID: "9c8874ed-3f70-45b6-9f10-d8737b6a7a8f"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 18:23:35 crc kubenswrapper[4988]: I1008 18:23:35.146955 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9c8874ed-3f70-45b6-9f10-d8737b6a7a8f-service-ca" (OuterVolumeSpecName: "service-ca") pod "9c8874ed-3f70-45b6-9f10-d8737b6a7a8f" (UID: "9c8874ed-3f70-45b6-9f10-d8737b6a7a8f"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 18:23:35 crc kubenswrapper[4988]: I1008 18:23:35.147083 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9c8874ed-3f70-45b6-9f10-d8737b6a7a8f-console-config" (OuterVolumeSpecName: "console-config") pod "9c8874ed-3f70-45b6-9f10-d8737b6a7a8f" (UID: "9c8874ed-3f70-45b6-9f10-d8737b6a7a8f"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 18:23:35 crc kubenswrapper[4988]: I1008 18:23:35.149478 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9c8874ed-3f70-45b6-9f10-d8737b6a7a8f-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "9c8874ed-3f70-45b6-9f10-d8737b6a7a8f" (UID: "9c8874ed-3f70-45b6-9f10-d8737b6a7a8f"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:23:35 crc kubenswrapper[4988]: I1008 18:23:35.150730 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9c8874ed-3f70-45b6-9f10-d8737b6a7a8f-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "9c8874ed-3f70-45b6-9f10-d8737b6a7a8f" (UID: "9c8874ed-3f70-45b6-9f10-d8737b6a7a8f"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:23:35 crc kubenswrapper[4988]: I1008 18:23:35.150748 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9c8874ed-3f70-45b6-9f10-d8737b6a7a8f-kube-api-access-qhgxr" (OuterVolumeSpecName: "kube-api-access-qhgxr") pod "9c8874ed-3f70-45b6-9f10-d8737b6a7a8f" (UID: "9c8874ed-3f70-45b6-9f10-d8737b6a7a8f"). InnerVolumeSpecName "kube-api-access-qhgxr". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:23:35 crc kubenswrapper[4988]: I1008 18:23:35.247659 4988 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/9c8874ed-3f70-45b6-9f10-d8737b6a7a8f-console-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 08 18:23:35 crc kubenswrapper[4988]: I1008 18:23:35.247711 4988 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/9c8874ed-3f70-45b6-9f10-d8737b6a7a8f-console-config\") on node \"crc\" DevicePath \"\"" Oct 08 18:23:35 crc kubenswrapper[4988]: I1008 18:23:35.247728 4988 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/9c8874ed-3f70-45b6-9f10-d8737b6a7a8f-console-oauth-config\") on node \"crc\" DevicePath \"\"" Oct 08 18:23:35 crc kubenswrapper[4988]: I1008 18:23:35.247743 4988 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/9c8874ed-3f70-45b6-9f10-d8737b6a7a8f-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 08 18:23:35 crc kubenswrapper[4988]: I1008 18:23:35.247757 4988 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qhgxr\" (UniqueName: \"kubernetes.io/projected/9c8874ed-3f70-45b6-9f10-d8737b6a7a8f-kube-api-access-qhgxr\") on node \"crc\" DevicePath \"\"" Oct 08 18:23:35 crc kubenswrapper[4988]: I1008 18:23:35.247775 4988 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/9c8874ed-3f70-45b6-9f10-d8737b6a7a8f-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 18:23:35 crc kubenswrapper[4988]: I1008 18:23:35.247788 4988 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/9c8874ed-3f70-45b6-9f10-d8737b6a7a8f-service-ca\") on node \"crc\" DevicePath \"\"" Oct 08 18:23:35 crc kubenswrapper[4988]: I1008 18:23:35.538066 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d28pl68"] Oct 08 18:23:35 crc kubenswrapper[4988]: I1008 18:23:35.633221 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d28pl68" event={"ID":"fdfe229e-0b1f-4af5-8c9b-2d1ed4c64fcb","Type":"ContainerStarted","Data":"28e87bf93faa7e53eddf32d556a3f6ebf32f9fff57e558ae3d5d09e3b2f45470"} Oct 08 18:23:35 crc kubenswrapper[4988]: I1008 18:23:35.636776 4988 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-7xdnc_9c8874ed-3f70-45b6-9f10-d8737b6a7a8f/console/0.log" Oct 08 18:23:35 crc kubenswrapper[4988]: I1008 18:23:35.636823 4988 generic.go:334] "Generic (PLEG): container finished" podID="9c8874ed-3f70-45b6-9f10-d8737b6a7a8f" containerID="8decb88999dd5ec9d72091c408c71df63d4adcf20acd387e7c499f07211eb8fc" exitCode=2 Oct 08 18:23:35 crc kubenswrapper[4988]: I1008 18:23:35.636851 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-7xdnc" event={"ID":"9c8874ed-3f70-45b6-9f10-d8737b6a7a8f","Type":"ContainerDied","Data":"8decb88999dd5ec9d72091c408c71df63d4adcf20acd387e7c499f07211eb8fc"} Oct 08 18:23:35 crc kubenswrapper[4988]: I1008 18:23:35.636876 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-7xdnc" event={"ID":"9c8874ed-3f70-45b6-9f10-d8737b6a7a8f","Type":"ContainerDied","Data":"04d2296a013043a794ce84eebbf03997f04316d51b2603d40464866e8c4e87cd"} Oct 08 18:23:35 crc kubenswrapper[4988]: I1008 18:23:35.636892 4988 scope.go:117] "RemoveContainer" containerID="8decb88999dd5ec9d72091c408c71df63d4adcf20acd387e7c499f07211eb8fc" Oct 08 18:23:35 crc kubenswrapper[4988]: I1008 18:23:35.636964 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-7xdnc" Oct 08 18:23:35 crc kubenswrapper[4988]: I1008 18:23:35.654222 4988 scope.go:117] "RemoveContainer" containerID="8decb88999dd5ec9d72091c408c71df63d4adcf20acd387e7c499f07211eb8fc" Oct 08 18:23:35 crc kubenswrapper[4988]: E1008 18:23:35.656466 4988 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8decb88999dd5ec9d72091c408c71df63d4adcf20acd387e7c499f07211eb8fc\": container with ID starting with 8decb88999dd5ec9d72091c408c71df63d4adcf20acd387e7c499f07211eb8fc not found: ID does not exist" containerID="8decb88999dd5ec9d72091c408c71df63d4adcf20acd387e7c499f07211eb8fc" Oct 08 18:23:35 crc kubenswrapper[4988]: I1008 18:23:35.656532 4988 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8decb88999dd5ec9d72091c408c71df63d4adcf20acd387e7c499f07211eb8fc"} err="failed to get container status \"8decb88999dd5ec9d72091c408c71df63d4adcf20acd387e7c499f07211eb8fc\": rpc error: code = NotFound desc = could not find container \"8decb88999dd5ec9d72091c408c71df63d4adcf20acd387e7c499f07211eb8fc\": container with ID starting with 8decb88999dd5ec9d72091c408c71df63d4adcf20acd387e7c499f07211eb8fc not found: ID does not exist" Oct 08 18:23:35 crc kubenswrapper[4988]: I1008 18:23:35.666672 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-f9d7485db-7xdnc"] Oct 08 18:23:35 crc kubenswrapper[4988]: I1008 18:23:35.670257 4988 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-console/console-f9d7485db-7xdnc"] Oct 08 18:23:36 crc kubenswrapper[4988]: I1008 18:23:36.650593 4988 generic.go:334] "Generic (PLEG): container finished" podID="fdfe229e-0b1f-4af5-8c9b-2d1ed4c64fcb" containerID="a43b3b8486837565a57b0a434072c6f2c2faaacb69bdb818ba20f2fe3484bbb7" exitCode=0 Oct 08 18:23:36 crc kubenswrapper[4988]: I1008 18:23:36.650669 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d28pl68" event={"ID":"fdfe229e-0b1f-4af5-8c9b-2d1ed4c64fcb","Type":"ContainerDied","Data":"a43b3b8486837565a57b0a434072c6f2c2faaacb69bdb818ba20f2fe3484bbb7"} Oct 08 18:23:37 crc kubenswrapper[4988]: I1008 18:23:37.244344 4988 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9c8874ed-3f70-45b6-9f10-d8737b6a7a8f" path="/var/lib/kubelet/pods/9c8874ed-3f70-45b6-9f10-d8737b6a7a8f/volumes" Oct 08 18:23:38 crc kubenswrapper[4988]: I1008 18:23:38.141190 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-nxp8r"] Oct 08 18:23:38 crc kubenswrapper[4988]: E1008 18:23:38.141794 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9c8874ed-3f70-45b6-9f10-d8737b6a7a8f" containerName="console" Oct 08 18:23:38 crc kubenswrapper[4988]: I1008 18:23:38.141810 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="9c8874ed-3f70-45b6-9f10-d8737b6a7a8f" containerName="console" Oct 08 18:23:38 crc kubenswrapper[4988]: I1008 18:23:38.141909 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="9c8874ed-3f70-45b6-9f10-d8737b6a7a8f" containerName="console" Oct 08 18:23:38 crc kubenswrapper[4988]: I1008 18:23:38.142690 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-nxp8r" Oct 08 18:23:38 crc kubenswrapper[4988]: I1008 18:23:38.165727 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-nxp8r"] Oct 08 18:23:38 crc kubenswrapper[4988]: I1008 18:23:38.286533 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d47573f1-5191-4f1e-bb80-1dc976cd46db-catalog-content\") pod \"redhat-operators-nxp8r\" (UID: \"d47573f1-5191-4f1e-bb80-1dc976cd46db\") " pod="openshift-marketplace/redhat-operators-nxp8r" Oct 08 18:23:38 crc kubenswrapper[4988]: I1008 18:23:38.286633 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lnvm9\" (UniqueName: \"kubernetes.io/projected/d47573f1-5191-4f1e-bb80-1dc976cd46db-kube-api-access-lnvm9\") pod \"redhat-operators-nxp8r\" (UID: \"d47573f1-5191-4f1e-bb80-1dc976cd46db\") " pod="openshift-marketplace/redhat-operators-nxp8r" Oct 08 18:23:38 crc kubenswrapper[4988]: I1008 18:23:38.286668 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d47573f1-5191-4f1e-bb80-1dc976cd46db-utilities\") pod \"redhat-operators-nxp8r\" (UID: \"d47573f1-5191-4f1e-bb80-1dc976cd46db\") " pod="openshift-marketplace/redhat-operators-nxp8r" Oct 08 18:23:38 crc kubenswrapper[4988]: I1008 18:23:38.387404 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d47573f1-5191-4f1e-bb80-1dc976cd46db-catalog-content\") pod \"redhat-operators-nxp8r\" (UID: \"d47573f1-5191-4f1e-bb80-1dc976cd46db\") " pod="openshift-marketplace/redhat-operators-nxp8r" Oct 08 18:23:38 crc kubenswrapper[4988]: I1008 18:23:38.387497 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lnvm9\" (UniqueName: \"kubernetes.io/projected/d47573f1-5191-4f1e-bb80-1dc976cd46db-kube-api-access-lnvm9\") pod \"redhat-operators-nxp8r\" (UID: \"d47573f1-5191-4f1e-bb80-1dc976cd46db\") " pod="openshift-marketplace/redhat-operators-nxp8r" Oct 08 18:23:38 crc kubenswrapper[4988]: I1008 18:23:38.387522 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d47573f1-5191-4f1e-bb80-1dc976cd46db-utilities\") pod \"redhat-operators-nxp8r\" (UID: \"d47573f1-5191-4f1e-bb80-1dc976cd46db\") " pod="openshift-marketplace/redhat-operators-nxp8r" Oct 08 18:23:38 crc kubenswrapper[4988]: I1008 18:23:38.389338 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d47573f1-5191-4f1e-bb80-1dc976cd46db-catalog-content\") pod \"redhat-operators-nxp8r\" (UID: \"d47573f1-5191-4f1e-bb80-1dc976cd46db\") " pod="openshift-marketplace/redhat-operators-nxp8r" Oct 08 18:23:38 crc kubenswrapper[4988]: I1008 18:23:38.390368 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d47573f1-5191-4f1e-bb80-1dc976cd46db-utilities\") pod \"redhat-operators-nxp8r\" (UID: \"d47573f1-5191-4f1e-bb80-1dc976cd46db\") " pod="openshift-marketplace/redhat-operators-nxp8r" Oct 08 18:23:38 crc kubenswrapper[4988]: I1008 18:23:38.421811 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lnvm9\" (UniqueName: \"kubernetes.io/projected/d47573f1-5191-4f1e-bb80-1dc976cd46db-kube-api-access-lnvm9\") pod \"redhat-operators-nxp8r\" (UID: \"d47573f1-5191-4f1e-bb80-1dc976cd46db\") " pod="openshift-marketplace/redhat-operators-nxp8r" Oct 08 18:23:38 crc kubenswrapper[4988]: I1008 18:23:38.474731 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-nxp8r" Oct 08 18:23:38 crc kubenswrapper[4988]: I1008 18:23:38.887077 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-nxp8r"] Oct 08 18:23:39 crc kubenswrapper[4988]: I1008 18:23:39.671847 4988 generic.go:334] "Generic (PLEG): container finished" podID="fdfe229e-0b1f-4af5-8c9b-2d1ed4c64fcb" containerID="5c20ff6d7413bbecb40e1db63c13c1f13d96c296f74f849fa9f03591d213e87b" exitCode=0 Oct 08 18:23:39 crc kubenswrapper[4988]: I1008 18:23:39.671946 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d28pl68" event={"ID":"fdfe229e-0b1f-4af5-8c9b-2d1ed4c64fcb","Type":"ContainerDied","Data":"5c20ff6d7413bbecb40e1db63c13c1f13d96c296f74f849fa9f03591d213e87b"} Oct 08 18:23:39 crc kubenswrapper[4988]: I1008 18:23:39.674027 4988 generic.go:334] "Generic (PLEG): container finished" podID="d47573f1-5191-4f1e-bb80-1dc976cd46db" containerID="03f8abaaf4741d4bb9549fdb15668236513b8f63d01d7d4c17f28babb2149329" exitCode=0 Oct 08 18:23:39 crc kubenswrapper[4988]: I1008 18:23:39.674050 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-nxp8r" event={"ID":"d47573f1-5191-4f1e-bb80-1dc976cd46db","Type":"ContainerDied","Data":"03f8abaaf4741d4bb9549fdb15668236513b8f63d01d7d4c17f28babb2149329"} Oct 08 18:23:39 crc kubenswrapper[4988]: I1008 18:23:39.674082 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-nxp8r" event={"ID":"d47573f1-5191-4f1e-bb80-1dc976cd46db","Type":"ContainerStarted","Data":"ccb40612114c11b7452d1ed30e283df02c76f4aedd4330882baced1cd9d47979"} Oct 08 18:23:40 crc kubenswrapper[4988]: I1008 18:23:40.682335 4988 generic.go:334] "Generic (PLEG): container finished" podID="fdfe229e-0b1f-4af5-8c9b-2d1ed4c64fcb" containerID="d032e4b092583a3c8b3182d7e18e384367780f6b661b9c677a04922ee4d3312c" exitCode=0 Oct 08 18:23:40 crc kubenswrapper[4988]: I1008 18:23:40.682425 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d28pl68" event={"ID":"fdfe229e-0b1f-4af5-8c9b-2d1ed4c64fcb","Type":"ContainerDied","Data":"d032e4b092583a3c8b3182d7e18e384367780f6b661b9c677a04922ee4d3312c"} Oct 08 18:23:41 crc kubenswrapper[4988]: I1008 18:23:41.692700 4988 generic.go:334] "Generic (PLEG): container finished" podID="d47573f1-5191-4f1e-bb80-1dc976cd46db" containerID="7157c927c78ae5f79ebf1c69f810361e80162ea72db8f843f1d8f43cf4396b74" exitCode=0 Oct 08 18:23:41 crc kubenswrapper[4988]: I1008 18:23:41.692757 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-nxp8r" event={"ID":"d47573f1-5191-4f1e-bb80-1dc976cd46db","Type":"ContainerDied","Data":"7157c927c78ae5f79ebf1c69f810361e80162ea72db8f843f1d8f43cf4396b74"} Oct 08 18:23:42 crc kubenswrapper[4988]: I1008 18:23:42.077229 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d28pl68" Oct 08 18:23:42 crc kubenswrapper[4988]: I1008 18:23:42.242957 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/fdfe229e-0b1f-4af5-8c9b-2d1ed4c64fcb-util\") pod \"fdfe229e-0b1f-4af5-8c9b-2d1ed4c64fcb\" (UID: \"fdfe229e-0b1f-4af5-8c9b-2d1ed4c64fcb\") " Oct 08 18:23:42 crc kubenswrapper[4988]: I1008 18:23:42.243095 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/fdfe229e-0b1f-4af5-8c9b-2d1ed4c64fcb-bundle\") pod \"fdfe229e-0b1f-4af5-8c9b-2d1ed4c64fcb\" (UID: \"fdfe229e-0b1f-4af5-8c9b-2d1ed4c64fcb\") " Oct 08 18:23:42 crc kubenswrapper[4988]: I1008 18:23:42.243136 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lvtxw\" (UniqueName: \"kubernetes.io/projected/fdfe229e-0b1f-4af5-8c9b-2d1ed4c64fcb-kube-api-access-lvtxw\") pod \"fdfe229e-0b1f-4af5-8c9b-2d1ed4c64fcb\" (UID: \"fdfe229e-0b1f-4af5-8c9b-2d1ed4c64fcb\") " Oct 08 18:23:42 crc kubenswrapper[4988]: I1008 18:23:42.245571 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fdfe229e-0b1f-4af5-8c9b-2d1ed4c64fcb-bundle" (OuterVolumeSpecName: "bundle") pod "fdfe229e-0b1f-4af5-8c9b-2d1ed4c64fcb" (UID: "fdfe229e-0b1f-4af5-8c9b-2d1ed4c64fcb"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 18:23:42 crc kubenswrapper[4988]: I1008 18:23:42.253619 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fdfe229e-0b1f-4af5-8c9b-2d1ed4c64fcb-kube-api-access-lvtxw" (OuterVolumeSpecName: "kube-api-access-lvtxw") pod "fdfe229e-0b1f-4af5-8c9b-2d1ed4c64fcb" (UID: "fdfe229e-0b1f-4af5-8c9b-2d1ed4c64fcb"). InnerVolumeSpecName "kube-api-access-lvtxw". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:23:42 crc kubenswrapper[4988]: I1008 18:23:42.254204 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fdfe229e-0b1f-4af5-8c9b-2d1ed4c64fcb-util" (OuterVolumeSpecName: "util") pod "fdfe229e-0b1f-4af5-8c9b-2d1ed4c64fcb" (UID: "fdfe229e-0b1f-4af5-8c9b-2d1ed4c64fcb"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 18:23:42 crc kubenswrapper[4988]: I1008 18:23:42.344187 4988 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/fdfe229e-0b1f-4af5-8c9b-2d1ed4c64fcb-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 18:23:42 crc kubenswrapper[4988]: I1008 18:23:42.344211 4988 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lvtxw\" (UniqueName: \"kubernetes.io/projected/fdfe229e-0b1f-4af5-8c9b-2d1ed4c64fcb-kube-api-access-lvtxw\") on node \"crc\" DevicePath \"\"" Oct 08 18:23:42 crc kubenswrapper[4988]: I1008 18:23:42.344223 4988 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/fdfe229e-0b1f-4af5-8c9b-2d1ed4c64fcb-util\") on node \"crc\" DevicePath \"\"" Oct 08 18:23:42 crc kubenswrapper[4988]: I1008 18:23:42.702855 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d28pl68" event={"ID":"fdfe229e-0b1f-4af5-8c9b-2d1ed4c64fcb","Type":"ContainerDied","Data":"28e87bf93faa7e53eddf32d556a3f6ebf32f9fff57e558ae3d5d09e3b2f45470"} Oct 08 18:23:42 crc kubenswrapper[4988]: I1008 18:23:42.702905 4988 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="28e87bf93faa7e53eddf32d556a3f6ebf32f9fff57e558ae3d5d09e3b2f45470" Oct 08 18:23:42 crc kubenswrapper[4988]: I1008 18:23:42.702902 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d28pl68" Oct 08 18:23:42 crc kubenswrapper[4988]: I1008 18:23:42.705777 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-nxp8r" event={"ID":"d47573f1-5191-4f1e-bb80-1dc976cd46db","Type":"ContainerStarted","Data":"ef63e716f1c2484237b8476937f220d9c9d3cb3aa483614c618aaf1ed7a06064"} Oct 08 18:23:42 crc kubenswrapper[4988]: I1008 18:23:42.732182 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-nxp8r" podStartSLOduration=2.11736598 podStartE2EDuration="4.732164612s" podCreationTimestamp="2025-10-08 18:23:38 +0000 UTC" firstStartedPulling="2025-10-08 18:23:39.675236829 +0000 UTC m=+765.125079599" lastFinishedPulling="2025-10-08 18:23:42.290035461 +0000 UTC m=+767.739878231" observedRunningTime="2025-10-08 18:23:42.728880795 +0000 UTC m=+768.178723595" watchObservedRunningTime="2025-10-08 18:23:42.732164612 +0000 UTC m=+768.182007382" Oct 08 18:23:42 crc kubenswrapper[4988]: I1008 18:23:42.736431 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-rf6lj"] Oct 08 18:23:42 crc kubenswrapper[4988]: E1008 18:23:42.736629 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fdfe229e-0b1f-4af5-8c9b-2d1ed4c64fcb" containerName="pull" Oct 08 18:23:42 crc kubenswrapper[4988]: I1008 18:23:42.736642 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="fdfe229e-0b1f-4af5-8c9b-2d1ed4c64fcb" containerName="pull" Oct 08 18:23:42 crc kubenswrapper[4988]: E1008 18:23:42.736654 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fdfe229e-0b1f-4af5-8c9b-2d1ed4c64fcb" containerName="util" Oct 08 18:23:42 crc kubenswrapper[4988]: I1008 18:23:42.736660 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="fdfe229e-0b1f-4af5-8c9b-2d1ed4c64fcb" containerName="util" Oct 08 18:23:42 crc kubenswrapper[4988]: E1008 18:23:42.736671 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fdfe229e-0b1f-4af5-8c9b-2d1ed4c64fcb" containerName="extract" Oct 08 18:23:42 crc kubenswrapper[4988]: I1008 18:23:42.736677 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="fdfe229e-0b1f-4af5-8c9b-2d1ed4c64fcb" containerName="extract" Oct 08 18:23:42 crc kubenswrapper[4988]: I1008 18:23:42.736786 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="fdfe229e-0b1f-4af5-8c9b-2d1ed4c64fcb" containerName="extract" Oct 08 18:23:42 crc kubenswrapper[4988]: I1008 18:23:42.737636 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-rf6lj" Oct 08 18:23:42 crc kubenswrapper[4988]: I1008 18:23:42.755138 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-rf6lj"] Oct 08 18:23:42 crc kubenswrapper[4988]: I1008 18:23:42.852031 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f784fa35-8bdf-46c0-a73e-4eea4a4960fa-catalog-content\") pod \"certified-operators-rf6lj\" (UID: \"f784fa35-8bdf-46c0-a73e-4eea4a4960fa\") " pod="openshift-marketplace/certified-operators-rf6lj" Oct 08 18:23:42 crc kubenswrapper[4988]: I1008 18:23:42.852091 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j7crz\" (UniqueName: \"kubernetes.io/projected/f784fa35-8bdf-46c0-a73e-4eea4a4960fa-kube-api-access-j7crz\") pod \"certified-operators-rf6lj\" (UID: \"f784fa35-8bdf-46c0-a73e-4eea4a4960fa\") " pod="openshift-marketplace/certified-operators-rf6lj" Oct 08 18:23:42 crc kubenswrapper[4988]: I1008 18:23:42.852117 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f784fa35-8bdf-46c0-a73e-4eea4a4960fa-utilities\") pod \"certified-operators-rf6lj\" (UID: \"f784fa35-8bdf-46c0-a73e-4eea4a4960fa\") " pod="openshift-marketplace/certified-operators-rf6lj" Oct 08 18:23:42 crc kubenswrapper[4988]: I1008 18:23:42.953869 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f784fa35-8bdf-46c0-a73e-4eea4a4960fa-catalog-content\") pod \"certified-operators-rf6lj\" (UID: \"f784fa35-8bdf-46c0-a73e-4eea4a4960fa\") " pod="openshift-marketplace/certified-operators-rf6lj" Oct 08 18:23:42 crc kubenswrapper[4988]: I1008 18:23:42.953935 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j7crz\" (UniqueName: \"kubernetes.io/projected/f784fa35-8bdf-46c0-a73e-4eea4a4960fa-kube-api-access-j7crz\") pod \"certified-operators-rf6lj\" (UID: \"f784fa35-8bdf-46c0-a73e-4eea4a4960fa\") " pod="openshift-marketplace/certified-operators-rf6lj" Oct 08 18:23:42 crc kubenswrapper[4988]: I1008 18:23:42.953960 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f784fa35-8bdf-46c0-a73e-4eea4a4960fa-utilities\") pod \"certified-operators-rf6lj\" (UID: \"f784fa35-8bdf-46c0-a73e-4eea4a4960fa\") " pod="openshift-marketplace/certified-operators-rf6lj" Oct 08 18:23:42 crc kubenswrapper[4988]: I1008 18:23:42.954535 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f784fa35-8bdf-46c0-a73e-4eea4a4960fa-utilities\") pod \"certified-operators-rf6lj\" (UID: \"f784fa35-8bdf-46c0-a73e-4eea4a4960fa\") " pod="openshift-marketplace/certified-operators-rf6lj" Oct 08 18:23:42 crc kubenswrapper[4988]: I1008 18:23:42.954657 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f784fa35-8bdf-46c0-a73e-4eea4a4960fa-catalog-content\") pod \"certified-operators-rf6lj\" (UID: \"f784fa35-8bdf-46c0-a73e-4eea4a4960fa\") " pod="openshift-marketplace/certified-operators-rf6lj" Oct 08 18:23:42 crc kubenswrapper[4988]: I1008 18:23:42.983694 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j7crz\" (UniqueName: \"kubernetes.io/projected/f784fa35-8bdf-46c0-a73e-4eea4a4960fa-kube-api-access-j7crz\") pod \"certified-operators-rf6lj\" (UID: \"f784fa35-8bdf-46c0-a73e-4eea4a4960fa\") " pod="openshift-marketplace/certified-operators-rf6lj" Oct 08 18:23:43 crc kubenswrapper[4988]: I1008 18:23:43.054105 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-rf6lj" Oct 08 18:23:43 crc kubenswrapper[4988]: I1008 18:23:43.526590 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-rf6lj"] Oct 08 18:23:43 crc kubenswrapper[4988]: W1008 18:23:43.527573 4988 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf784fa35_8bdf_46c0_a73e_4eea4a4960fa.slice/crio-50805f49768460643e6806f606e09a6758546ff55c2fa1a9bcc21ad07201e1fc WatchSource:0}: Error finding container 50805f49768460643e6806f606e09a6758546ff55c2fa1a9bcc21ad07201e1fc: Status 404 returned error can't find the container with id 50805f49768460643e6806f606e09a6758546ff55c2fa1a9bcc21ad07201e1fc Oct 08 18:23:43 crc kubenswrapper[4988]: I1008 18:23:43.713080 4988 generic.go:334] "Generic (PLEG): container finished" podID="f784fa35-8bdf-46c0-a73e-4eea4a4960fa" containerID="881d0aa8a5abcf4653bade8c2cbb8147930c6942536a26628a2a4c5f4ea9cc7b" exitCode=0 Oct 08 18:23:43 crc kubenswrapper[4988]: I1008 18:23:43.713207 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rf6lj" event={"ID":"f784fa35-8bdf-46c0-a73e-4eea4a4960fa","Type":"ContainerDied","Data":"881d0aa8a5abcf4653bade8c2cbb8147930c6942536a26628a2a4c5f4ea9cc7b"} Oct 08 18:23:43 crc kubenswrapper[4988]: I1008 18:23:43.713236 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rf6lj" event={"ID":"f784fa35-8bdf-46c0-a73e-4eea4a4960fa","Type":"ContainerStarted","Data":"50805f49768460643e6806f606e09a6758546ff55c2fa1a9bcc21ad07201e1fc"} Oct 08 18:23:44 crc kubenswrapper[4988]: I1008 18:23:44.725161 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rf6lj" event={"ID":"f784fa35-8bdf-46c0-a73e-4eea4a4960fa","Type":"ContainerStarted","Data":"8682ac28a278c1c912a71239de0f05c947b183b2716cb24eb4ae980c51535983"} Oct 08 18:23:45 crc kubenswrapper[4988]: I1008 18:23:45.733201 4988 generic.go:334] "Generic (PLEG): container finished" podID="f784fa35-8bdf-46c0-a73e-4eea4a4960fa" containerID="8682ac28a278c1c912a71239de0f05c947b183b2716cb24eb4ae980c51535983" exitCode=0 Oct 08 18:23:45 crc kubenswrapper[4988]: I1008 18:23:45.733409 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rf6lj" event={"ID":"f784fa35-8bdf-46c0-a73e-4eea4a4960fa","Type":"ContainerDied","Data":"8682ac28a278c1c912a71239de0f05c947b183b2716cb24eb4ae980c51535983"} Oct 08 18:23:46 crc kubenswrapper[4988]: I1008 18:23:46.740794 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rf6lj" event={"ID":"f784fa35-8bdf-46c0-a73e-4eea4a4960fa","Type":"ContainerStarted","Data":"5820741c0ddc46179f354267811cdbbabe2fd943b108328ba5c987b9942b4d11"} Oct 08 18:23:46 crc kubenswrapper[4988]: I1008 18:23:46.771026 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-rf6lj" podStartSLOduration=2.065640005 podStartE2EDuration="4.771005555s" podCreationTimestamp="2025-10-08 18:23:42 +0000 UTC" firstStartedPulling="2025-10-08 18:23:43.714823237 +0000 UTC m=+769.164666017" lastFinishedPulling="2025-10-08 18:23:46.420188797 +0000 UTC m=+771.870031567" observedRunningTime="2025-10-08 18:23:46.767699917 +0000 UTC m=+772.217542687" watchObservedRunningTime="2025-10-08 18:23:46.771005555 +0000 UTC m=+772.220848335" Oct 08 18:23:48 crc kubenswrapper[4988]: I1008 18:23:48.475657 4988 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-nxp8r" Oct 08 18:23:48 crc kubenswrapper[4988]: I1008 18:23:48.476013 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-nxp8r" Oct 08 18:23:48 crc kubenswrapper[4988]: I1008 18:23:48.516410 4988 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-nxp8r" Oct 08 18:23:48 crc kubenswrapper[4988]: I1008 18:23:48.797687 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-nxp8r" Oct 08 18:23:50 crc kubenswrapper[4988]: I1008 18:23:50.728105 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-nxp8r"] Oct 08 18:23:50 crc kubenswrapper[4988]: I1008 18:23:50.762462 4988 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-nxp8r" podUID="d47573f1-5191-4f1e-bb80-1dc976cd46db" containerName="registry-server" containerID="cri-o://ef63e716f1c2484237b8476937f220d9c9d3cb3aa483614c618aaf1ed7a06064" gracePeriod=2 Oct 08 18:23:51 crc kubenswrapper[4988]: I1008 18:23:51.270759 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-nxp8r" Oct 08 18:23:51 crc kubenswrapper[4988]: I1008 18:23:51.369149 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d47573f1-5191-4f1e-bb80-1dc976cd46db-utilities\") pod \"d47573f1-5191-4f1e-bb80-1dc976cd46db\" (UID: \"d47573f1-5191-4f1e-bb80-1dc976cd46db\") " Oct 08 18:23:51 crc kubenswrapper[4988]: I1008 18:23:51.369231 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d47573f1-5191-4f1e-bb80-1dc976cd46db-catalog-content\") pod \"d47573f1-5191-4f1e-bb80-1dc976cd46db\" (UID: \"d47573f1-5191-4f1e-bb80-1dc976cd46db\") " Oct 08 18:23:51 crc kubenswrapper[4988]: I1008 18:23:51.369271 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lnvm9\" (UniqueName: \"kubernetes.io/projected/d47573f1-5191-4f1e-bb80-1dc976cd46db-kube-api-access-lnvm9\") pod \"d47573f1-5191-4f1e-bb80-1dc976cd46db\" (UID: \"d47573f1-5191-4f1e-bb80-1dc976cd46db\") " Oct 08 18:23:51 crc kubenswrapper[4988]: I1008 18:23:51.370698 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d47573f1-5191-4f1e-bb80-1dc976cd46db-utilities" (OuterVolumeSpecName: "utilities") pod "d47573f1-5191-4f1e-bb80-1dc976cd46db" (UID: "d47573f1-5191-4f1e-bb80-1dc976cd46db"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 18:23:51 crc kubenswrapper[4988]: I1008 18:23:51.384514 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d47573f1-5191-4f1e-bb80-1dc976cd46db-kube-api-access-lnvm9" (OuterVolumeSpecName: "kube-api-access-lnvm9") pod "d47573f1-5191-4f1e-bb80-1dc976cd46db" (UID: "d47573f1-5191-4f1e-bb80-1dc976cd46db"). InnerVolumeSpecName "kube-api-access-lnvm9". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:23:51 crc kubenswrapper[4988]: I1008 18:23:51.470558 4988 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lnvm9\" (UniqueName: \"kubernetes.io/projected/d47573f1-5191-4f1e-bb80-1dc976cd46db-kube-api-access-lnvm9\") on node \"crc\" DevicePath \"\"" Oct 08 18:23:51 crc kubenswrapper[4988]: I1008 18:23:51.470592 4988 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d47573f1-5191-4f1e-bb80-1dc976cd46db-utilities\") on node \"crc\" DevicePath \"\"" Oct 08 18:23:51 crc kubenswrapper[4988]: I1008 18:23:51.610906 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-controller-manager-55b79cb758-ld9ht"] Oct 08 18:23:51 crc kubenswrapper[4988]: E1008 18:23:51.611166 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d47573f1-5191-4f1e-bb80-1dc976cd46db" containerName="registry-server" Oct 08 18:23:51 crc kubenswrapper[4988]: I1008 18:23:51.611184 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="d47573f1-5191-4f1e-bb80-1dc976cd46db" containerName="registry-server" Oct 08 18:23:51 crc kubenswrapper[4988]: E1008 18:23:51.611203 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d47573f1-5191-4f1e-bb80-1dc976cd46db" containerName="extract-content" Oct 08 18:23:51 crc kubenswrapper[4988]: I1008 18:23:51.611210 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="d47573f1-5191-4f1e-bb80-1dc976cd46db" containerName="extract-content" Oct 08 18:23:51 crc kubenswrapper[4988]: E1008 18:23:51.611234 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d47573f1-5191-4f1e-bb80-1dc976cd46db" containerName="extract-utilities" Oct 08 18:23:51 crc kubenswrapper[4988]: I1008 18:23:51.611242 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="d47573f1-5191-4f1e-bb80-1dc976cd46db" containerName="extract-utilities" Oct 08 18:23:51 crc kubenswrapper[4988]: I1008 18:23:51.611365 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="d47573f1-5191-4f1e-bb80-1dc976cd46db" containerName="registry-server" Oct 08 18:23:51 crc kubenswrapper[4988]: I1008 18:23:51.611850 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-55b79cb758-ld9ht" Oct 08 18:23:51 crc kubenswrapper[4988]: I1008 18:23:51.613901 4988 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-controller-manager-service-cert" Oct 08 18:23:51 crc kubenswrapper[4988]: I1008 18:23:51.614313 4988 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"manager-account-dockercfg-swpdz" Oct 08 18:23:51 crc kubenswrapper[4988]: I1008 18:23:51.615956 4988 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"kube-root-ca.crt" Oct 08 18:23:51 crc kubenswrapper[4988]: I1008 18:23:51.616817 4988 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"openshift-service-ca.crt" Oct 08 18:23:51 crc kubenswrapper[4988]: I1008 18:23:51.617276 4988 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-cert" Oct 08 18:23:51 crc kubenswrapper[4988]: I1008 18:23:51.629282 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-55b79cb758-ld9ht"] Oct 08 18:23:51 crc kubenswrapper[4988]: I1008 18:23:51.769326 4988 generic.go:334] "Generic (PLEG): container finished" podID="d47573f1-5191-4f1e-bb80-1dc976cd46db" containerID="ef63e716f1c2484237b8476937f220d9c9d3cb3aa483614c618aaf1ed7a06064" exitCode=0 Oct 08 18:23:51 crc kubenswrapper[4988]: I1008 18:23:51.769405 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-nxp8r" Oct 08 18:23:51 crc kubenswrapper[4988]: I1008 18:23:51.769405 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-nxp8r" event={"ID":"d47573f1-5191-4f1e-bb80-1dc976cd46db","Type":"ContainerDied","Data":"ef63e716f1c2484237b8476937f220d9c9d3cb3aa483614c618aaf1ed7a06064"} Oct 08 18:23:51 crc kubenswrapper[4988]: I1008 18:23:51.769562 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-nxp8r" event={"ID":"d47573f1-5191-4f1e-bb80-1dc976cd46db","Type":"ContainerDied","Data":"ccb40612114c11b7452d1ed30e283df02c76f4aedd4330882baced1cd9d47979"} Oct 08 18:23:51 crc kubenswrapper[4988]: I1008 18:23:51.769594 4988 scope.go:117] "RemoveContainer" containerID="ef63e716f1c2484237b8476937f220d9c9d3cb3aa483614c618aaf1ed7a06064" Oct 08 18:23:51 crc kubenswrapper[4988]: I1008 18:23:51.773870 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/4cc757e2-0271-4785-830f-00aa42bd5a6d-webhook-cert\") pod \"metallb-operator-controller-manager-55b79cb758-ld9ht\" (UID: \"4cc757e2-0271-4785-830f-00aa42bd5a6d\") " pod="metallb-system/metallb-operator-controller-manager-55b79cb758-ld9ht" Oct 08 18:23:51 crc kubenswrapper[4988]: I1008 18:23:51.773904 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p9xdz\" (UniqueName: \"kubernetes.io/projected/4cc757e2-0271-4785-830f-00aa42bd5a6d-kube-api-access-p9xdz\") pod \"metallb-operator-controller-manager-55b79cb758-ld9ht\" (UID: \"4cc757e2-0271-4785-830f-00aa42bd5a6d\") " pod="metallb-system/metallb-operator-controller-manager-55b79cb758-ld9ht" Oct 08 18:23:51 crc kubenswrapper[4988]: I1008 18:23:51.773994 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/4cc757e2-0271-4785-830f-00aa42bd5a6d-apiservice-cert\") pod \"metallb-operator-controller-manager-55b79cb758-ld9ht\" (UID: \"4cc757e2-0271-4785-830f-00aa42bd5a6d\") " pod="metallb-system/metallb-operator-controller-manager-55b79cb758-ld9ht" Oct 08 18:23:51 crc kubenswrapper[4988]: I1008 18:23:51.790860 4988 scope.go:117] "RemoveContainer" containerID="7157c927c78ae5f79ebf1c69f810361e80162ea72db8f843f1d8f43cf4396b74" Oct 08 18:23:51 crc kubenswrapper[4988]: I1008 18:23:51.825345 4988 scope.go:117] "RemoveContainer" containerID="03f8abaaf4741d4bb9549fdb15668236513b8f63d01d7d4c17f28babb2149329" Oct 08 18:23:51 crc kubenswrapper[4988]: I1008 18:23:51.839509 4988 scope.go:117] "RemoveContainer" containerID="ef63e716f1c2484237b8476937f220d9c9d3cb3aa483614c618aaf1ed7a06064" Oct 08 18:23:51 crc kubenswrapper[4988]: E1008 18:23:51.840030 4988 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ef63e716f1c2484237b8476937f220d9c9d3cb3aa483614c618aaf1ed7a06064\": container with ID starting with ef63e716f1c2484237b8476937f220d9c9d3cb3aa483614c618aaf1ed7a06064 not found: ID does not exist" containerID="ef63e716f1c2484237b8476937f220d9c9d3cb3aa483614c618aaf1ed7a06064" Oct 08 18:23:51 crc kubenswrapper[4988]: I1008 18:23:51.840084 4988 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ef63e716f1c2484237b8476937f220d9c9d3cb3aa483614c618aaf1ed7a06064"} err="failed to get container status \"ef63e716f1c2484237b8476937f220d9c9d3cb3aa483614c618aaf1ed7a06064\": rpc error: code = NotFound desc = could not find container \"ef63e716f1c2484237b8476937f220d9c9d3cb3aa483614c618aaf1ed7a06064\": container with ID starting with ef63e716f1c2484237b8476937f220d9c9d3cb3aa483614c618aaf1ed7a06064 not found: ID does not exist" Oct 08 18:23:51 crc kubenswrapper[4988]: I1008 18:23:51.840132 4988 scope.go:117] "RemoveContainer" containerID="7157c927c78ae5f79ebf1c69f810361e80162ea72db8f843f1d8f43cf4396b74" Oct 08 18:23:51 crc kubenswrapper[4988]: E1008 18:23:51.851611 4988 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7157c927c78ae5f79ebf1c69f810361e80162ea72db8f843f1d8f43cf4396b74\": container with ID starting with 7157c927c78ae5f79ebf1c69f810361e80162ea72db8f843f1d8f43cf4396b74 not found: ID does not exist" containerID="7157c927c78ae5f79ebf1c69f810361e80162ea72db8f843f1d8f43cf4396b74" Oct 08 18:23:51 crc kubenswrapper[4988]: I1008 18:23:51.851659 4988 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7157c927c78ae5f79ebf1c69f810361e80162ea72db8f843f1d8f43cf4396b74"} err="failed to get container status \"7157c927c78ae5f79ebf1c69f810361e80162ea72db8f843f1d8f43cf4396b74\": rpc error: code = NotFound desc = could not find container \"7157c927c78ae5f79ebf1c69f810361e80162ea72db8f843f1d8f43cf4396b74\": container with ID starting with 7157c927c78ae5f79ebf1c69f810361e80162ea72db8f843f1d8f43cf4396b74 not found: ID does not exist" Oct 08 18:23:51 crc kubenswrapper[4988]: I1008 18:23:51.851689 4988 scope.go:117] "RemoveContainer" containerID="03f8abaaf4741d4bb9549fdb15668236513b8f63d01d7d4c17f28babb2149329" Oct 08 18:23:51 crc kubenswrapper[4988]: E1008 18:23:51.853938 4988 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"03f8abaaf4741d4bb9549fdb15668236513b8f63d01d7d4c17f28babb2149329\": container with ID starting with 03f8abaaf4741d4bb9549fdb15668236513b8f63d01d7d4c17f28babb2149329 not found: ID does not exist" containerID="03f8abaaf4741d4bb9549fdb15668236513b8f63d01d7d4c17f28babb2149329" Oct 08 18:23:51 crc kubenswrapper[4988]: I1008 18:23:51.854000 4988 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"03f8abaaf4741d4bb9549fdb15668236513b8f63d01d7d4c17f28babb2149329"} err="failed to get container status \"03f8abaaf4741d4bb9549fdb15668236513b8f63d01d7d4c17f28babb2149329\": rpc error: code = NotFound desc = could not find container \"03f8abaaf4741d4bb9549fdb15668236513b8f63d01d7d4c17f28babb2149329\": container with ID starting with 03f8abaaf4741d4bb9549fdb15668236513b8f63d01d7d4c17f28babb2149329 not found: ID does not exist" Oct 08 18:23:51 crc kubenswrapper[4988]: I1008 18:23:51.875117 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p9xdz\" (UniqueName: \"kubernetes.io/projected/4cc757e2-0271-4785-830f-00aa42bd5a6d-kube-api-access-p9xdz\") pod \"metallb-operator-controller-manager-55b79cb758-ld9ht\" (UID: \"4cc757e2-0271-4785-830f-00aa42bd5a6d\") " pod="metallb-system/metallb-operator-controller-manager-55b79cb758-ld9ht" Oct 08 18:23:51 crc kubenswrapper[4988]: I1008 18:23:51.875166 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/4cc757e2-0271-4785-830f-00aa42bd5a6d-webhook-cert\") pod \"metallb-operator-controller-manager-55b79cb758-ld9ht\" (UID: \"4cc757e2-0271-4785-830f-00aa42bd5a6d\") " pod="metallb-system/metallb-operator-controller-manager-55b79cb758-ld9ht" Oct 08 18:23:51 crc kubenswrapper[4988]: I1008 18:23:51.875222 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/4cc757e2-0271-4785-830f-00aa42bd5a6d-apiservice-cert\") pod \"metallb-operator-controller-manager-55b79cb758-ld9ht\" (UID: \"4cc757e2-0271-4785-830f-00aa42bd5a6d\") " pod="metallb-system/metallb-operator-controller-manager-55b79cb758-ld9ht" Oct 08 18:23:51 crc kubenswrapper[4988]: I1008 18:23:51.880035 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/4cc757e2-0271-4785-830f-00aa42bd5a6d-webhook-cert\") pod \"metallb-operator-controller-manager-55b79cb758-ld9ht\" (UID: \"4cc757e2-0271-4785-830f-00aa42bd5a6d\") " pod="metallb-system/metallb-operator-controller-manager-55b79cb758-ld9ht" Oct 08 18:23:51 crc kubenswrapper[4988]: I1008 18:23:51.880623 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/4cc757e2-0271-4785-830f-00aa42bd5a6d-apiservice-cert\") pod \"metallb-operator-controller-manager-55b79cb758-ld9ht\" (UID: \"4cc757e2-0271-4785-830f-00aa42bd5a6d\") " pod="metallb-system/metallb-operator-controller-manager-55b79cb758-ld9ht" Oct 08 18:23:51 crc kubenswrapper[4988]: I1008 18:23:51.895881 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p9xdz\" (UniqueName: \"kubernetes.io/projected/4cc757e2-0271-4785-830f-00aa42bd5a6d-kube-api-access-p9xdz\") pod \"metallb-operator-controller-manager-55b79cb758-ld9ht\" (UID: \"4cc757e2-0271-4785-830f-00aa42bd5a6d\") " pod="metallb-system/metallb-operator-controller-manager-55b79cb758-ld9ht" Oct 08 18:23:51 crc kubenswrapper[4988]: I1008 18:23:51.924605 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-55b79cb758-ld9ht" Oct 08 18:23:52 crc kubenswrapper[4988]: I1008 18:23:52.057060 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-webhook-server-6585b9b6b6-zpjq5"] Oct 08 18:23:52 crc kubenswrapper[4988]: I1008 18:23:52.063630 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-6585b9b6b6-zpjq5" Oct 08 18:23:52 crc kubenswrapper[4988]: I1008 18:23:52.066621 4988 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-dockercfg-bq2xc" Oct 08 18:23:52 crc kubenswrapper[4988]: I1008 18:23:52.067928 4988 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-webhook-cert" Oct 08 18:23:52 crc kubenswrapper[4988]: I1008 18:23:52.072042 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-6585b9b6b6-zpjq5"] Oct 08 18:23:52 crc kubenswrapper[4988]: I1008 18:23:52.072994 4988 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-service-cert" Oct 08 18:23:52 crc kubenswrapper[4988]: I1008 18:23:52.178948 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/11101830-713e-483b-bfae-699e80dbde14-webhook-cert\") pod \"metallb-operator-webhook-server-6585b9b6b6-zpjq5\" (UID: \"11101830-713e-483b-bfae-699e80dbde14\") " pod="metallb-system/metallb-operator-webhook-server-6585b9b6b6-zpjq5" Oct 08 18:23:52 crc kubenswrapper[4988]: I1008 18:23:52.179660 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bxdr7\" (UniqueName: \"kubernetes.io/projected/11101830-713e-483b-bfae-699e80dbde14-kube-api-access-bxdr7\") pod \"metallb-operator-webhook-server-6585b9b6b6-zpjq5\" (UID: \"11101830-713e-483b-bfae-699e80dbde14\") " pod="metallb-system/metallb-operator-webhook-server-6585b9b6b6-zpjq5" Oct 08 18:23:52 crc kubenswrapper[4988]: I1008 18:23:52.179724 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/11101830-713e-483b-bfae-699e80dbde14-apiservice-cert\") pod \"metallb-operator-webhook-server-6585b9b6b6-zpjq5\" (UID: \"11101830-713e-483b-bfae-699e80dbde14\") " pod="metallb-system/metallb-operator-webhook-server-6585b9b6b6-zpjq5" Oct 08 18:23:52 crc kubenswrapper[4988]: I1008 18:23:52.281530 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/11101830-713e-483b-bfae-699e80dbde14-webhook-cert\") pod \"metallb-operator-webhook-server-6585b9b6b6-zpjq5\" (UID: \"11101830-713e-483b-bfae-699e80dbde14\") " pod="metallb-system/metallb-operator-webhook-server-6585b9b6b6-zpjq5" Oct 08 18:23:52 crc kubenswrapper[4988]: I1008 18:23:52.281948 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bxdr7\" (UniqueName: \"kubernetes.io/projected/11101830-713e-483b-bfae-699e80dbde14-kube-api-access-bxdr7\") pod \"metallb-operator-webhook-server-6585b9b6b6-zpjq5\" (UID: \"11101830-713e-483b-bfae-699e80dbde14\") " pod="metallb-system/metallb-operator-webhook-server-6585b9b6b6-zpjq5" Oct 08 18:23:52 crc kubenswrapper[4988]: I1008 18:23:52.282013 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/11101830-713e-483b-bfae-699e80dbde14-apiservice-cert\") pod \"metallb-operator-webhook-server-6585b9b6b6-zpjq5\" (UID: \"11101830-713e-483b-bfae-699e80dbde14\") " pod="metallb-system/metallb-operator-webhook-server-6585b9b6b6-zpjq5" Oct 08 18:23:52 crc kubenswrapper[4988]: I1008 18:23:52.285637 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/11101830-713e-483b-bfae-699e80dbde14-apiservice-cert\") pod \"metallb-operator-webhook-server-6585b9b6b6-zpjq5\" (UID: \"11101830-713e-483b-bfae-699e80dbde14\") " pod="metallb-system/metallb-operator-webhook-server-6585b9b6b6-zpjq5" Oct 08 18:23:52 crc kubenswrapper[4988]: I1008 18:23:52.286956 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/11101830-713e-483b-bfae-699e80dbde14-webhook-cert\") pod \"metallb-operator-webhook-server-6585b9b6b6-zpjq5\" (UID: \"11101830-713e-483b-bfae-699e80dbde14\") " pod="metallb-system/metallb-operator-webhook-server-6585b9b6b6-zpjq5" Oct 08 18:23:52 crc kubenswrapper[4988]: I1008 18:23:52.310375 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bxdr7\" (UniqueName: \"kubernetes.io/projected/11101830-713e-483b-bfae-699e80dbde14-kube-api-access-bxdr7\") pod \"metallb-operator-webhook-server-6585b9b6b6-zpjq5\" (UID: \"11101830-713e-483b-bfae-699e80dbde14\") " pod="metallb-system/metallb-operator-webhook-server-6585b9b6b6-zpjq5" Oct 08 18:23:52 crc kubenswrapper[4988]: I1008 18:23:52.398098 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-55b79cb758-ld9ht"] Oct 08 18:23:52 crc kubenswrapper[4988]: I1008 18:23:52.401806 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-6585b9b6b6-zpjq5" Oct 08 18:23:52 crc kubenswrapper[4988]: W1008 18:23:52.403284 4988 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4cc757e2_0271_4785_830f_00aa42bd5a6d.slice/crio-86be2f669bc29048a7847c0cdd4f4c00373f7432d73c625501d0050d08823578 WatchSource:0}: Error finding container 86be2f669bc29048a7847c0cdd4f4c00373f7432d73c625501d0050d08823578: Status 404 returned error can't find the container with id 86be2f669bc29048a7847c0cdd4f4c00373f7432d73c625501d0050d08823578 Oct 08 18:23:52 crc kubenswrapper[4988]: I1008 18:23:52.419819 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d47573f1-5191-4f1e-bb80-1dc976cd46db-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "d47573f1-5191-4f1e-bb80-1dc976cd46db" (UID: "d47573f1-5191-4f1e-bb80-1dc976cd46db"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 18:23:52 crc kubenswrapper[4988]: I1008 18:23:52.484873 4988 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d47573f1-5191-4f1e-bb80-1dc976cd46db-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 08 18:23:52 crc kubenswrapper[4988]: I1008 18:23:52.707090 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-nxp8r"] Oct 08 18:23:52 crc kubenswrapper[4988]: I1008 18:23:52.710011 4988 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-nxp8r"] Oct 08 18:23:52 crc kubenswrapper[4988]: I1008 18:23:52.774369 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-55b79cb758-ld9ht" event={"ID":"4cc757e2-0271-4785-830f-00aa42bd5a6d","Type":"ContainerStarted","Data":"86be2f669bc29048a7847c0cdd4f4c00373f7432d73c625501d0050d08823578"} Oct 08 18:23:52 crc kubenswrapper[4988]: I1008 18:23:52.828553 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-6585b9b6b6-zpjq5"] Oct 08 18:23:52 crc kubenswrapper[4988]: W1008 18:23:52.832812 4988 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod11101830_713e_483b_bfae_699e80dbde14.slice/crio-19c3f788b2065e18c3cdc8b1b9e9ac56dc23dcd31d24e7bdcd9e347a934c6b08 WatchSource:0}: Error finding container 19c3f788b2065e18c3cdc8b1b9e9ac56dc23dcd31d24e7bdcd9e347a934c6b08: Status 404 returned error can't find the container with id 19c3f788b2065e18c3cdc8b1b9e9ac56dc23dcd31d24e7bdcd9e347a934c6b08 Oct 08 18:23:53 crc kubenswrapper[4988]: I1008 18:23:53.054941 4988 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-rf6lj" Oct 08 18:23:53 crc kubenswrapper[4988]: I1008 18:23:53.055312 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-rf6lj" Oct 08 18:23:53 crc kubenswrapper[4988]: I1008 18:23:53.117080 4988 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-rf6lj" Oct 08 18:23:53 crc kubenswrapper[4988]: I1008 18:23:53.247015 4988 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d47573f1-5191-4f1e-bb80-1dc976cd46db" path="/var/lib/kubelet/pods/d47573f1-5191-4f1e-bb80-1dc976cd46db/volumes" Oct 08 18:23:53 crc kubenswrapper[4988]: I1008 18:23:53.337822 4988 patch_prober.go:28] interesting pod/machine-config-daemon-nm54f container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 18:23:53 crc kubenswrapper[4988]: I1008 18:23:53.338046 4988 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 18:23:53 crc kubenswrapper[4988]: I1008 18:23:53.784183 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-6585b9b6b6-zpjq5" event={"ID":"11101830-713e-483b-bfae-699e80dbde14","Type":"ContainerStarted","Data":"19c3f788b2065e18c3cdc8b1b9e9ac56dc23dcd31d24e7bdcd9e347a934c6b08"} Oct 08 18:23:53 crc kubenswrapper[4988]: I1008 18:23:53.847071 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-rf6lj" Oct 08 18:23:55 crc kubenswrapper[4988]: I1008 18:23:55.326271 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-rf6lj"] Oct 08 18:23:55 crc kubenswrapper[4988]: I1008 18:23:55.802439 4988 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-rf6lj" podUID="f784fa35-8bdf-46c0-a73e-4eea4a4960fa" containerName="registry-server" containerID="cri-o://5820741c0ddc46179f354267811cdbbabe2fd943b108328ba5c987b9942b4d11" gracePeriod=2 Oct 08 18:23:56 crc kubenswrapper[4988]: I1008 18:23:56.813939 4988 generic.go:334] "Generic (PLEG): container finished" podID="f784fa35-8bdf-46c0-a73e-4eea4a4960fa" containerID="5820741c0ddc46179f354267811cdbbabe2fd943b108328ba5c987b9942b4d11" exitCode=0 Oct 08 18:23:56 crc kubenswrapper[4988]: I1008 18:23:56.814034 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rf6lj" event={"ID":"f784fa35-8bdf-46c0-a73e-4eea4a4960fa","Type":"ContainerDied","Data":"5820741c0ddc46179f354267811cdbbabe2fd943b108328ba5c987b9942b4d11"} Oct 08 18:23:58 crc kubenswrapper[4988]: I1008 18:23:58.481880 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-rf6lj" Oct 08 18:23:58 crc kubenswrapper[4988]: I1008 18:23:58.571330 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f784fa35-8bdf-46c0-a73e-4eea4a4960fa-utilities\") pod \"f784fa35-8bdf-46c0-a73e-4eea4a4960fa\" (UID: \"f784fa35-8bdf-46c0-a73e-4eea4a4960fa\") " Oct 08 18:23:58 crc kubenswrapper[4988]: I1008 18:23:58.571423 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-j7crz\" (UniqueName: \"kubernetes.io/projected/f784fa35-8bdf-46c0-a73e-4eea4a4960fa-kube-api-access-j7crz\") pod \"f784fa35-8bdf-46c0-a73e-4eea4a4960fa\" (UID: \"f784fa35-8bdf-46c0-a73e-4eea4a4960fa\") " Oct 08 18:23:58 crc kubenswrapper[4988]: I1008 18:23:58.571469 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f784fa35-8bdf-46c0-a73e-4eea4a4960fa-catalog-content\") pod \"f784fa35-8bdf-46c0-a73e-4eea4a4960fa\" (UID: \"f784fa35-8bdf-46c0-a73e-4eea4a4960fa\") " Oct 08 18:23:58 crc kubenswrapper[4988]: I1008 18:23:58.572100 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f784fa35-8bdf-46c0-a73e-4eea4a4960fa-utilities" (OuterVolumeSpecName: "utilities") pod "f784fa35-8bdf-46c0-a73e-4eea4a4960fa" (UID: "f784fa35-8bdf-46c0-a73e-4eea4a4960fa"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 18:23:58 crc kubenswrapper[4988]: I1008 18:23:58.576690 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f784fa35-8bdf-46c0-a73e-4eea4a4960fa-kube-api-access-j7crz" (OuterVolumeSpecName: "kube-api-access-j7crz") pod "f784fa35-8bdf-46c0-a73e-4eea4a4960fa" (UID: "f784fa35-8bdf-46c0-a73e-4eea4a4960fa"). InnerVolumeSpecName "kube-api-access-j7crz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:23:58 crc kubenswrapper[4988]: I1008 18:23:58.613146 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f784fa35-8bdf-46c0-a73e-4eea4a4960fa-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "f784fa35-8bdf-46c0-a73e-4eea4a4960fa" (UID: "f784fa35-8bdf-46c0-a73e-4eea4a4960fa"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 18:23:58 crc kubenswrapper[4988]: I1008 18:23:58.672646 4988 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-j7crz\" (UniqueName: \"kubernetes.io/projected/f784fa35-8bdf-46c0-a73e-4eea4a4960fa-kube-api-access-j7crz\") on node \"crc\" DevicePath \"\"" Oct 08 18:23:58 crc kubenswrapper[4988]: I1008 18:23:58.672687 4988 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f784fa35-8bdf-46c0-a73e-4eea4a4960fa-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 08 18:23:58 crc kubenswrapper[4988]: I1008 18:23:58.672700 4988 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f784fa35-8bdf-46c0-a73e-4eea4a4960fa-utilities\") on node \"crc\" DevicePath \"\"" Oct 08 18:23:58 crc kubenswrapper[4988]: I1008 18:23:58.828427 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-6585b9b6b6-zpjq5" event={"ID":"11101830-713e-483b-bfae-699e80dbde14","Type":"ContainerStarted","Data":"5bef9afcf7552d1043add348c6c9c04a1f96a0260300e3bf597476ff40f513b6"} Oct 08 18:23:58 crc kubenswrapper[4988]: I1008 18:23:58.828570 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-webhook-server-6585b9b6b6-zpjq5" Oct 08 18:23:58 crc kubenswrapper[4988]: I1008 18:23:58.830274 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-55b79cb758-ld9ht" event={"ID":"4cc757e2-0271-4785-830f-00aa42bd5a6d","Type":"ContainerStarted","Data":"2f4682cab0f68a4d070b0b057c647de01cb3c793d04d08655888211a691c6a5c"} Oct 08 18:23:58 crc kubenswrapper[4988]: I1008 18:23:58.830399 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-controller-manager-55b79cb758-ld9ht" Oct 08 18:23:58 crc kubenswrapper[4988]: I1008 18:23:58.832731 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rf6lj" event={"ID":"f784fa35-8bdf-46c0-a73e-4eea4a4960fa","Type":"ContainerDied","Data":"50805f49768460643e6806f606e09a6758546ff55c2fa1a9bcc21ad07201e1fc"} Oct 08 18:23:58 crc kubenswrapper[4988]: I1008 18:23:58.832769 4988 scope.go:117] "RemoveContainer" containerID="5820741c0ddc46179f354267811cdbbabe2fd943b108328ba5c987b9942b4d11" Oct 08 18:23:58 crc kubenswrapper[4988]: I1008 18:23:58.832786 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-rf6lj" Oct 08 18:23:58 crc kubenswrapper[4988]: I1008 18:23:58.859041 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-webhook-server-6585b9b6b6-zpjq5" podStartSLOduration=1.187673851 podStartE2EDuration="6.859014689s" podCreationTimestamp="2025-10-08 18:23:52 +0000 UTC" firstStartedPulling="2025-10-08 18:23:52.836882525 +0000 UTC m=+778.286725295" lastFinishedPulling="2025-10-08 18:23:58.508223353 +0000 UTC m=+783.958066133" observedRunningTime="2025-10-08 18:23:58.85280293 +0000 UTC m=+784.302645700" watchObservedRunningTime="2025-10-08 18:23:58.859014689 +0000 UTC m=+784.308857479" Oct 08 18:23:58 crc kubenswrapper[4988]: I1008 18:23:58.874935 4988 scope.go:117] "RemoveContainer" containerID="8682ac28a278c1c912a71239de0f05c947b183b2716cb24eb4ae980c51535983" Oct 08 18:23:58 crc kubenswrapper[4988]: I1008 18:23:58.878606 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-controller-manager-55b79cb758-ld9ht" podStartSLOduration=1.79348935 podStartE2EDuration="7.878558542s" podCreationTimestamp="2025-10-08 18:23:51 +0000 UTC" firstStartedPulling="2025-10-08 18:23:52.408929857 +0000 UTC m=+777.858772617" lastFinishedPulling="2025-10-08 18:23:58.493999039 +0000 UTC m=+783.943841809" observedRunningTime="2025-10-08 18:23:58.875455903 +0000 UTC m=+784.325298693" watchObservedRunningTime="2025-10-08 18:23:58.878558542 +0000 UTC m=+784.328401312" Oct 08 18:23:58 crc kubenswrapper[4988]: I1008 18:23:58.893731 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-rf6lj"] Oct 08 18:23:58 crc kubenswrapper[4988]: I1008 18:23:58.896436 4988 scope.go:117] "RemoveContainer" containerID="881d0aa8a5abcf4653bade8c2cbb8147930c6942536a26628a2a4c5f4ea9cc7b" Oct 08 18:23:58 crc kubenswrapper[4988]: I1008 18:23:58.899445 4988 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-rf6lj"] Oct 08 18:23:59 crc kubenswrapper[4988]: I1008 18:23:59.251073 4988 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f784fa35-8bdf-46c0-a73e-4eea4a4960fa" path="/var/lib/kubelet/pods/f784fa35-8bdf-46c0-a73e-4eea4a4960fa/volumes" Oct 08 18:24:03 crc kubenswrapper[4988]: I1008 18:24:03.739625 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-tp58r"] Oct 08 18:24:03 crc kubenswrapper[4988]: E1008 18:24:03.740349 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f784fa35-8bdf-46c0-a73e-4eea4a4960fa" containerName="registry-server" Oct 08 18:24:03 crc kubenswrapper[4988]: I1008 18:24:03.740365 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="f784fa35-8bdf-46c0-a73e-4eea4a4960fa" containerName="registry-server" Oct 08 18:24:03 crc kubenswrapper[4988]: E1008 18:24:03.740415 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f784fa35-8bdf-46c0-a73e-4eea4a4960fa" containerName="extract-content" Oct 08 18:24:03 crc kubenswrapper[4988]: I1008 18:24:03.740424 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="f784fa35-8bdf-46c0-a73e-4eea4a4960fa" containerName="extract-content" Oct 08 18:24:03 crc kubenswrapper[4988]: E1008 18:24:03.740435 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f784fa35-8bdf-46c0-a73e-4eea4a4960fa" containerName="extract-utilities" Oct 08 18:24:03 crc kubenswrapper[4988]: I1008 18:24:03.740444 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="f784fa35-8bdf-46c0-a73e-4eea4a4960fa" containerName="extract-utilities" Oct 08 18:24:03 crc kubenswrapper[4988]: I1008 18:24:03.740711 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="f784fa35-8bdf-46c0-a73e-4eea4a4960fa" containerName="registry-server" Oct 08 18:24:03 crc kubenswrapper[4988]: I1008 18:24:03.746562 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-tp58r" Oct 08 18:24:03 crc kubenswrapper[4988]: I1008 18:24:03.760636 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-tp58r"] Oct 08 18:24:03 crc kubenswrapper[4988]: I1008 18:24:03.840616 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/188fccde-d7a5-4d7b-ba19-999f6b67ffcd-catalog-content\") pod \"community-operators-tp58r\" (UID: \"188fccde-d7a5-4d7b-ba19-999f6b67ffcd\") " pod="openshift-marketplace/community-operators-tp58r" Oct 08 18:24:03 crc kubenswrapper[4988]: I1008 18:24:03.840668 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gktbr\" (UniqueName: \"kubernetes.io/projected/188fccde-d7a5-4d7b-ba19-999f6b67ffcd-kube-api-access-gktbr\") pod \"community-operators-tp58r\" (UID: \"188fccde-d7a5-4d7b-ba19-999f6b67ffcd\") " pod="openshift-marketplace/community-operators-tp58r" Oct 08 18:24:03 crc kubenswrapper[4988]: I1008 18:24:03.840732 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/188fccde-d7a5-4d7b-ba19-999f6b67ffcd-utilities\") pod \"community-operators-tp58r\" (UID: \"188fccde-d7a5-4d7b-ba19-999f6b67ffcd\") " pod="openshift-marketplace/community-operators-tp58r" Oct 08 18:24:03 crc kubenswrapper[4988]: I1008 18:24:03.941991 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/188fccde-d7a5-4d7b-ba19-999f6b67ffcd-catalog-content\") pod \"community-operators-tp58r\" (UID: \"188fccde-d7a5-4d7b-ba19-999f6b67ffcd\") " pod="openshift-marketplace/community-operators-tp58r" Oct 08 18:24:03 crc kubenswrapper[4988]: I1008 18:24:03.942069 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gktbr\" (UniqueName: \"kubernetes.io/projected/188fccde-d7a5-4d7b-ba19-999f6b67ffcd-kube-api-access-gktbr\") pod \"community-operators-tp58r\" (UID: \"188fccde-d7a5-4d7b-ba19-999f6b67ffcd\") " pod="openshift-marketplace/community-operators-tp58r" Oct 08 18:24:03 crc kubenswrapper[4988]: I1008 18:24:03.942120 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/188fccde-d7a5-4d7b-ba19-999f6b67ffcd-utilities\") pod \"community-operators-tp58r\" (UID: \"188fccde-d7a5-4d7b-ba19-999f6b67ffcd\") " pod="openshift-marketplace/community-operators-tp58r" Oct 08 18:24:03 crc kubenswrapper[4988]: I1008 18:24:03.942574 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/188fccde-d7a5-4d7b-ba19-999f6b67ffcd-catalog-content\") pod \"community-operators-tp58r\" (UID: \"188fccde-d7a5-4d7b-ba19-999f6b67ffcd\") " pod="openshift-marketplace/community-operators-tp58r" Oct 08 18:24:03 crc kubenswrapper[4988]: I1008 18:24:03.942694 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/188fccde-d7a5-4d7b-ba19-999f6b67ffcd-utilities\") pod \"community-operators-tp58r\" (UID: \"188fccde-d7a5-4d7b-ba19-999f6b67ffcd\") " pod="openshift-marketplace/community-operators-tp58r" Oct 08 18:24:03 crc kubenswrapper[4988]: I1008 18:24:03.962326 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gktbr\" (UniqueName: \"kubernetes.io/projected/188fccde-d7a5-4d7b-ba19-999f6b67ffcd-kube-api-access-gktbr\") pod \"community-operators-tp58r\" (UID: \"188fccde-d7a5-4d7b-ba19-999f6b67ffcd\") " pod="openshift-marketplace/community-operators-tp58r" Oct 08 18:24:04 crc kubenswrapper[4988]: I1008 18:24:04.086727 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-tp58r" Oct 08 18:24:04 crc kubenswrapper[4988]: I1008 18:24:04.648127 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-tp58r"] Oct 08 18:24:04 crc kubenswrapper[4988]: I1008 18:24:04.873502 4988 generic.go:334] "Generic (PLEG): container finished" podID="188fccde-d7a5-4d7b-ba19-999f6b67ffcd" containerID="10caac9d7d374762716e0b7f42d546348ce2b894faa79dd243755efe15e033ff" exitCode=0 Oct 08 18:24:04 crc kubenswrapper[4988]: I1008 18:24:04.873584 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-tp58r" event={"ID":"188fccde-d7a5-4d7b-ba19-999f6b67ffcd","Type":"ContainerDied","Data":"10caac9d7d374762716e0b7f42d546348ce2b894faa79dd243755efe15e033ff"} Oct 08 18:24:04 crc kubenswrapper[4988]: I1008 18:24:04.873887 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-tp58r" event={"ID":"188fccde-d7a5-4d7b-ba19-999f6b67ffcd","Type":"ContainerStarted","Data":"519cd2b7bbec4878608a9eb32138ab29aca48160ef19a7829fa5fe88d0cb2705"} Oct 08 18:24:05 crc kubenswrapper[4988]: I1008 18:24:05.882010 4988 generic.go:334] "Generic (PLEG): container finished" podID="188fccde-d7a5-4d7b-ba19-999f6b67ffcd" containerID="2853c203af0f75cfc3d109a28a19ea4733d76771a964411bcdd629230abb1a01" exitCode=0 Oct 08 18:24:05 crc kubenswrapper[4988]: I1008 18:24:05.882118 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-tp58r" event={"ID":"188fccde-d7a5-4d7b-ba19-999f6b67ffcd","Type":"ContainerDied","Data":"2853c203af0f75cfc3d109a28a19ea4733d76771a964411bcdd629230abb1a01"} Oct 08 18:24:06 crc kubenswrapper[4988]: I1008 18:24:06.894986 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-tp58r" event={"ID":"188fccde-d7a5-4d7b-ba19-999f6b67ffcd","Type":"ContainerStarted","Data":"54e237d87373be0e8cad37127613d73b569d9607b42278817d90f01c9c81396b"} Oct 08 18:24:06 crc kubenswrapper[4988]: I1008 18:24:06.927428 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-tp58r" podStartSLOduration=2.412034564 podStartE2EDuration="3.927411359s" podCreationTimestamp="2025-10-08 18:24:03 +0000 UTC" firstStartedPulling="2025-10-08 18:24:04.875968765 +0000 UTC m=+790.325811535" lastFinishedPulling="2025-10-08 18:24:06.39134556 +0000 UTC m=+791.841188330" observedRunningTime="2025-10-08 18:24:06.919679182 +0000 UTC m=+792.369521982" watchObservedRunningTime="2025-10-08 18:24:06.927411359 +0000 UTC m=+792.377254129" Oct 08 18:24:10 crc kubenswrapper[4988]: I1008 18:24:10.335134 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-x7w9s"] Oct 08 18:24:10 crc kubenswrapper[4988]: I1008 18:24:10.336763 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-x7w9s" Oct 08 18:24:10 crc kubenswrapper[4988]: I1008 18:24:10.347799 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-x7w9s"] Oct 08 18:24:10 crc kubenswrapper[4988]: I1008 18:24:10.429207 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qsn9z\" (UniqueName: \"kubernetes.io/projected/9f3ad494-ee2f-40d9-a0a3-ed2824add4b4-kube-api-access-qsn9z\") pod \"redhat-marketplace-x7w9s\" (UID: \"9f3ad494-ee2f-40d9-a0a3-ed2824add4b4\") " pod="openshift-marketplace/redhat-marketplace-x7w9s" Oct 08 18:24:10 crc kubenswrapper[4988]: I1008 18:24:10.429590 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9f3ad494-ee2f-40d9-a0a3-ed2824add4b4-utilities\") pod \"redhat-marketplace-x7w9s\" (UID: \"9f3ad494-ee2f-40d9-a0a3-ed2824add4b4\") " pod="openshift-marketplace/redhat-marketplace-x7w9s" Oct 08 18:24:10 crc kubenswrapper[4988]: I1008 18:24:10.429774 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9f3ad494-ee2f-40d9-a0a3-ed2824add4b4-catalog-content\") pod \"redhat-marketplace-x7w9s\" (UID: \"9f3ad494-ee2f-40d9-a0a3-ed2824add4b4\") " pod="openshift-marketplace/redhat-marketplace-x7w9s" Oct 08 18:24:10 crc kubenswrapper[4988]: I1008 18:24:10.530873 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qsn9z\" (UniqueName: \"kubernetes.io/projected/9f3ad494-ee2f-40d9-a0a3-ed2824add4b4-kube-api-access-qsn9z\") pod \"redhat-marketplace-x7w9s\" (UID: \"9f3ad494-ee2f-40d9-a0a3-ed2824add4b4\") " pod="openshift-marketplace/redhat-marketplace-x7w9s" Oct 08 18:24:10 crc kubenswrapper[4988]: I1008 18:24:10.531132 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9f3ad494-ee2f-40d9-a0a3-ed2824add4b4-utilities\") pod \"redhat-marketplace-x7w9s\" (UID: \"9f3ad494-ee2f-40d9-a0a3-ed2824add4b4\") " pod="openshift-marketplace/redhat-marketplace-x7w9s" Oct 08 18:24:10 crc kubenswrapper[4988]: I1008 18:24:10.531209 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9f3ad494-ee2f-40d9-a0a3-ed2824add4b4-catalog-content\") pod \"redhat-marketplace-x7w9s\" (UID: \"9f3ad494-ee2f-40d9-a0a3-ed2824add4b4\") " pod="openshift-marketplace/redhat-marketplace-x7w9s" Oct 08 18:24:10 crc kubenswrapper[4988]: I1008 18:24:10.531674 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9f3ad494-ee2f-40d9-a0a3-ed2824add4b4-utilities\") pod \"redhat-marketplace-x7w9s\" (UID: \"9f3ad494-ee2f-40d9-a0a3-ed2824add4b4\") " pod="openshift-marketplace/redhat-marketplace-x7w9s" Oct 08 18:24:10 crc kubenswrapper[4988]: I1008 18:24:10.531747 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9f3ad494-ee2f-40d9-a0a3-ed2824add4b4-catalog-content\") pod \"redhat-marketplace-x7w9s\" (UID: \"9f3ad494-ee2f-40d9-a0a3-ed2824add4b4\") " pod="openshift-marketplace/redhat-marketplace-x7w9s" Oct 08 18:24:10 crc kubenswrapper[4988]: I1008 18:24:10.550273 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qsn9z\" (UniqueName: \"kubernetes.io/projected/9f3ad494-ee2f-40d9-a0a3-ed2824add4b4-kube-api-access-qsn9z\") pod \"redhat-marketplace-x7w9s\" (UID: \"9f3ad494-ee2f-40d9-a0a3-ed2824add4b4\") " pod="openshift-marketplace/redhat-marketplace-x7w9s" Oct 08 18:24:10 crc kubenswrapper[4988]: I1008 18:24:10.653973 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-x7w9s" Oct 08 18:24:11 crc kubenswrapper[4988]: I1008 18:24:11.113227 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-x7w9s"] Oct 08 18:24:11 crc kubenswrapper[4988]: I1008 18:24:11.939145 4988 generic.go:334] "Generic (PLEG): container finished" podID="9f3ad494-ee2f-40d9-a0a3-ed2824add4b4" containerID="5a9ada12000c7c6e96ba7c496ea3c8adaa91865ff1f34b53fef89a2f9451dd5b" exitCode=0 Oct 08 18:24:11 crc kubenswrapper[4988]: I1008 18:24:11.939245 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-x7w9s" event={"ID":"9f3ad494-ee2f-40d9-a0a3-ed2824add4b4","Type":"ContainerDied","Data":"5a9ada12000c7c6e96ba7c496ea3c8adaa91865ff1f34b53fef89a2f9451dd5b"} Oct 08 18:24:11 crc kubenswrapper[4988]: I1008 18:24:11.950433 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-x7w9s" event={"ID":"9f3ad494-ee2f-40d9-a0a3-ed2824add4b4","Type":"ContainerStarted","Data":"2384a680cf1e57f4917f96d75c91b11a4eacba725c4c968239b50c9e020ccfdc"} Oct 08 18:24:12 crc kubenswrapper[4988]: I1008 18:24:12.411182 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-webhook-server-6585b9b6b6-zpjq5" Oct 08 18:24:12 crc kubenswrapper[4988]: I1008 18:24:12.960003 4988 generic.go:334] "Generic (PLEG): container finished" podID="9f3ad494-ee2f-40d9-a0a3-ed2824add4b4" containerID="dc415c5cf6a2b96cb7281c7740b90752caf851040939ebde2864c1d407db15ac" exitCode=0 Oct 08 18:24:12 crc kubenswrapper[4988]: I1008 18:24:12.960096 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-x7w9s" event={"ID":"9f3ad494-ee2f-40d9-a0a3-ed2824add4b4","Type":"ContainerDied","Data":"dc415c5cf6a2b96cb7281c7740b90752caf851040939ebde2864c1d407db15ac"} Oct 08 18:24:13 crc kubenswrapper[4988]: I1008 18:24:13.968946 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-x7w9s" event={"ID":"9f3ad494-ee2f-40d9-a0a3-ed2824add4b4","Type":"ContainerStarted","Data":"ebcaf8187a9337f6e8f70903e189fc2626be3a07ec9842cd523ff5b44b03d0dc"} Oct 08 18:24:13 crc kubenswrapper[4988]: I1008 18:24:13.985495 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-x7w9s" podStartSLOduration=2.569655326 podStartE2EDuration="3.985475383s" podCreationTimestamp="2025-10-08 18:24:10 +0000 UTC" firstStartedPulling="2025-10-08 18:24:11.942465149 +0000 UTC m=+797.392307919" lastFinishedPulling="2025-10-08 18:24:13.358285196 +0000 UTC m=+798.808127976" observedRunningTime="2025-10-08 18:24:13.982534699 +0000 UTC m=+799.432377479" watchObservedRunningTime="2025-10-08 18:24:13.985475383 +0000 UTC m=+799.435318153" Oct 08 18:24:14 crc kubenswrapper[4988]: I1008 18:24:14.087735 4988 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-tp58r" Oct 08 18:24:14 crc kubenswrapper[4988]: I1008 18:24:14.087792 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-tp58r" Oct 08 18:24:14 crc kubenswrapper[4988]: I1008 18:24:14.136757 4988 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-tp58r" Oct 08 18:24:15 crc kubenswrapper[4988]: I1008 18:24:15.014248 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-tp58r" Oct 08 18:24:19 crc kubenswrapper[4988]: I1008 18:24:19.530973 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-tp58r"] Oct 08 18:24:19 crc kubenswrapper[4988]: I1008 18:24:19.532127 4988 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-tp58r" podUID="188fccde-d7a5-4d7b-ba19-999f6b67ffcd" containerName="registry-server" containerID="cri-o://54e237d87373be0e8cad37127613d73b569d9607b42278817d90f01c9c81396b" gracePeriod=2 Oct 08 18:24:19 crc kubenswrapper[4988]: I1008 18:24:19.939175 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-tp58r" Oct 08 18:24:20 crc kubenswrapper[4988]: I1008 18:24:20.004481 4988 generic.go:334] "Generic (PLEG): container finished" podID="188fccde-d7a5-4d7b-ba19-999f6b67ffcd" containerID="54e237d87373be0e8cad37127613d73b569d9607b42278817d90f01c9c81396b" exitCode=0 Oct 08 18:24:20 crc kubenswrapper[4988]: I1008 18:24:20.004528 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-tp58r" event={"ID":"188fccde-d7a5-4d7b-ba19-999f6b67ffcd","Type":"ContainerDied","Data":"54e237d87373be0e8cad37127613d73b569d9607b42278817d90f01c9c81396b"} Oct 08 18:24:20 crc kubenswrapper[4988]: I1008 18:24:20.004574 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-tp58r" event={"ID":"188fccde-d7a5-4d7b-ba19-999f6b67ffcd","Type":"ContainerDied","Data":"519cd2b7bbec4878608a9eb32138ab29aca48160ef19a7829fa5fe88d0cb2705"} Oct 08 18:24:20 crc kubenswrapper[4988]: I1008 18:24:20.004570 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-tp58r" Oct 08 18:24:20 crc kubenswrapper[4988]: I1008 18:24:20.004592 4988 scope.go:117] "RemoveContainer" containerID="54e237d87373be0e8cad37127613d73b569d9607b42278817d90f01c9c81396b" Oct 08 18:24:20 crc kubenswrapper[4988]: I1008 18:24:20.022116 4988 scope.go:117] "RemoveContainer" containerID="2853c203af0f75cfc3d109a28a19ea4733d76771a964411bcdd629230abb1a01" Oct 08 18:24:20 crc kubenswrapper[4988]: I1008 18:24:20.039235 4988 scope.go:117] "RemoveContainer" containerID="10caac9d7d374762716e0b7f42d546348ce2b894faa79dd243755efe15e033ff" Oct 08 18:24:20 crc kubenswrapper[4988]: I1008 18:24:20.055615 4988 scope.go:117] "RemoveContainer" containerID="54e237d87373be0e8cad37127613d73b569d9607b42278817d90f01c9c81396b" Oct 08 18:24:20 crc kubenswrapper[4988]: E1008 18:24:20.056122 4988 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"54e237d87373be0e8cad37127613d73b569d9607b42278817d90f01c9c81396b\": container with ID starting with 54e237d87373be0e8cad37127613d73b569d9607b42278817d90f01c9c81396b not found: ID does not exist" containerID="54e237d87373be0e8cad37127613d73b569d9607b42278817d90f01c9c81396b" Oct 08 18:24:20 crc kubenswrapper[4988]: I1008 18:24:20.056164 4988 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"54e237d87373be0e8cad37127613d73b569d9607b42278817d90f01c9c81396b"} err="failed to get container status \"54e237d87373be0e8cad37127613d73b569d9607b42278817d90f01c9c81396b\": rpc error: code = NotFound desc = could not find container \"54e237d87373be0e8cad37127613d73b569d9607b42278817d90f01c9c81396b\": container with ID starting with 54e237d87373be0e8cad37127613d73b569d9607b42278817d90f01c9c81396b not found: ID does not exist" Oct 08 18:24:20 crc kubenswrapper[4988]: I1008 18:24:20.056191 4988 scope.go:117] "RemoveContainer" containerID="2853c203af0f75cfc3d109a28a19ea4733d76771a964411bcdd629230abb1a01" Oct 08 18:24:20 crc kubenswrapper[4988]: E1008 18:24:20.056512 4988 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2853c203af0f75cfc3d109a28a19ea4733d76771a964411bcdd629230abb1a01\": container with ID starting with 2853c203af0f75cfc3d109a28a19ea4733d76771a964411bcdd629230abb1a01 not found: ID does not exist" containerID="2853c203af0f75cfc3d109a28a19ea4733d76771a964411bcdd629230abb1a01" Oct 08 18:24:20 crc kubenswrapper[4988]: I1008 18:24:20.056534 4988 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2853c203af0f75cfc3d109a28a19ea4733d76771a964411bcdd629230abb1a01"} err="failed to get container status \"2853c203af0f75cfc3d109a28a19ea4733d76771a964411bcdd629230abb1a01\": rpc error: code = NotFound desc = could not find container \"2853c203af0f75cfc3d109a28a19ea4733d76771a964411bcdd629230abb1a01\": container with ID starting with 2853c203af0f75cfc3d109a28a19ea4733d76771a964411bcdd629230abb1a01 not found: ID does not exist" Oct 08 18:24:20 crc kubenswrapper[4988]: I1008 18:24:20.056547 4988 scope.go:117] "RemoveContainer" containerID="10caac9d7d374762716e0b7f42d546348ce2b894faa79dd243755efe15e033ff" Oct 08 18:24:20 crc kubenswrapper[4988]: E1008 18:24:20.056849 4988 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"10caac9d7d374762716e0b7f42d546348ce2b894faa79dd243755efe15e033ff\": container with ID starting with 10caac9d7d374762716e0b7f42d546348ce2b894faa79dd243755efe15e033ff not found: ID does not exist" containerID="10caac9d7d374762716e0b7f42d546348ce2b894faa79dd243755efe15e033ff" Oct 08 18:24:20 crc kubenswrapper[4988]: I1008 18:24:20.056871 4988 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"10caac9d7d374762716e0b7f42d546348ce2b894faa79dd243755efe15e033ff"} err="failed to get container status \"10caac9d7d374762716e0b7f42d546348ce2b894faa79dd243755efe15e033ff\": rpc error: code = NotFound desc = could not find container \"10caac9d7d374762716e0b7f42d546348ce2b894faa79dd243755efe15e033ff\": container with ID starting with 10caac9d7d374762716e0b7f42d546348ce2b894faa79dd243755efe15e033ff not found: ID does not exist" Oct 08 18:24:20 crc kubenswrapper[4988]: I1008 18:24:20.072230 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/188fccde-d7a5-4d7b-ba19-999f6b67ffcd-catalog-content\") pod \"188fccde-d7a5-4d7b-ba19-999f6b67ffcd\" (UID: \"188fccde-d7a5-4d7b-ba19-999f6b67ffcd\") " Oct 08 18:24:20 crc kubenswrapper[4988]: I1008 18:24:20.072358 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gktbr\" (UniqueName: \"kubernetes.io/projected/188fccde-d7a5-4d7b-ba19-999f6b67ffcd-kube-api-access-gktbr\") pod \"188fccde-d7a5-4d7b-ba19-999f6b67ffcd\" (UID: \"188fccde-d7a5-4d7b-ba19-999f6b67ffcd\") " Oct 08 18:24:20 crc kubenswrapper[4988]: I1008 18:24:20.072432 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/188fccde-d7a5-4d7b-ba19-999f6b67ffcd-utilities\") pod \"188fccde-d7a5-4d7b-ba19-999f6b67ffcd\" (UID: \"188fccde-d7a5-4d7b-ba19-999f6b67ffcd\") " Oct 08 18:24:20 crc kubenswrapper[4988]: I1008 18:24:20.074265 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/188fccde-d7a5-4d7b-ba19-999f6b67ffcd-utilities" (OuterVolumeSpecName: "utilities") pod "188fccde-d7a5-4d7b-ba19-999f6b67ffcd" (UID: "188fccde-d7a5-4d7b-ba19-999f6b67ffcd"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 18:24:20 crc kubenswrapper[4988]: I1008 18:24:20.080610 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/188fccde-d7a5-4d7b-ba19-999f6b67ffcd-kube-api-access-gktbr" (OuterVolumeSpecName: "kube-api-access-gktbr") pod "188fccde-d7a5-4d7b-ba19-999f6b67ffcd" (UID: "188fccde-d7a5-4d7b-ba19-999f6b67ffcd"). InnerVolumeSpecName "kube-api-access-gktbr". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:24:20 crc kubenswrapper[4988]: I1008 18:24:20.132444 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/188fccde-d7a5-4d7b-ba19-999f6b67ffcd-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "188fccde-d7a5-4d7b-ba19-999f6b67ffcd" (UID: "188fccde-d7a5-4d7b-ba19-999f6b67ffcd"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 18:24:20 crc kubenswrapper[4988]: I1008 18:24:20.173762 4988 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/188fccde-d7a5-4d7b-ba19-999f6b67ffcd-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 08 18:24:20 crc kubenswrapper[4988]: I1008 18:24:20.173802 4988 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gktbr\" (UniqueName: \"kubernetes.io/projected/188fccde-d7a5-4d7b-ba19-999f6b67ffcd-kube-api-access-gktbr\") on node \"crc\" DevicePath \"\"" Oct 08 18:24:20 crc kubenswrapper[4988]: I1008 18:24:20.173821 4988 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/188fccde-d7a5-4d7b-ba19-999f6b67ffcd-utilities\") on node \"crc\" DevicePath \"\"" Oct 08 18:24:20 crc kubenswrapper[4988]: I1008 18:24:20.346600 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-tp58r"] Oct 08 18:24:20 crc kubenswrapper[4988]: I1008 18:24:20.352836 4988 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-tp58r"] Oct 08 18:24:20 crc kubenswrapper[4988]: I1008 18:24:20.654850 4988 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-x7w9s" Oct 08 18:24:20 crc kubenswrapper[4988]: I1008 18:24:20.654901 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-x7w9s" Oct 08 18:24:20 crc kubenswrapper[4988]: I1008 18:24:20.700121 4988 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-x7w9s" Oct 08 18:24:21 crc kubenswrapper[4988]: I1008 18:24:21.048540 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-x7w9s" Oct 08 18:24:21 crc kubenswrapper[4988]: I1008 18:24:21.247762 4988 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="188fccde-d7a5-4d7b-ba19-999f6b67ffcd" path="/var/lib/kubelet/pods/188fccde-d7a5-4d7b-ba19-999f6b67ffcd/volumes" Oct 08 18:24:23 crc kubenswrapper[4988]: I1008 18:24:23.337641 4988 patch_prober.go:28] interesting pod/machine-config-daemon-nm54f container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 18:24:23 crc kubenswrapper[4988]: I1008 18:24:23.337963 4988 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 18:24:23 crc kubenswrapper[4988]: I1008 18:24:23.338015 4988 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" Oct 08 18:24:23 crc kubenswrapper[4988]: I1008 18:24:23.338674 4988 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"0b7eb1da8fbdc295b99305bf18316c88ee5b435b3bee717d217e3efa0977c5a9"} pod="openshift-machine-config-operator/machine-config-daemon-nm54f" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 08 18:24:23 crc kubenswrapper[4988]: I1008 18:24:23.338745 4988 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" containerName="machine-config-daemon" containerID="cri-o://0b7eb1da8fbdc295b99305bf18316c88ee5b435b3bee717d217e3efa0977c5a9" gracePeriod=600 Oct 08 18:24:24 crc kubenswrapper[4988]: I1008 18:24:24.032568 4988 generic.go:334] "Generic (PLEG): container finished" podID="f328a20b-5283-407e-8a93-2ab2da599f18" containerID="0b7eb1da8fbdc295b99305bf18316c88ee5b435b3bee717d217e3efa0977c5a9" exitCode=0 Oct 08 18:24:24 crc kubenswrapper[4988]: I1008 18:24:24.032594 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" event={"ID":"f328a20b-5283-407e-8a93-2ab2da599f18","Type":"ContainerDied","Data":"0b7eb1da8fbdc295b99305bf18316c88ee5b435b3bee717d217e3efa0977c5a9"} Oct 08 18:24:24 crc kubenswrapper[4988]: I1008 18:24:24.032892 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" event={"ID":"f328a20b-5283-407e-8a93-2ab2da599f18","Type":"ContainerStarted","Data":"a06add767e00bcfef86e27b752f06b6fb27241526ad4e59aaac22482ba9f2df4"} Oct 08 18:24:24 crc kubenswrapper[4988]: I1008 18:24:24.032913 4988 scope.go:117] "RemoveContainer" containerID="3134be7a9e11eb7efeae4470b4af55e4a033a7a3717b6d3502d9e3bd3007b110" Oct 08 18:24:24 crc kubenswrapper[4988]: I1008 18:24:24.124275 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-x7w9s"] Oct 08 18:24:24 crc kubenswrapper[4988]: I1008 18:24:24.124793 4988 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-x7w9s" podUID="9f3ad494-ee2f-40d9-a0a3-ed2824add4b4" containerName="registry-server" containerID="cri-o://ebcaf8187a9337f6e8f70903e189fc2626be3a07ec9842cd523ff5b44b03d0dc" gracePeriod=2 Oct 08 18:24:24 crc kubenswrapper[4988]: I1008 18:24:24.524944 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-x7w9s" Oct 08 18:24:24 crc kubenswrapper[4988]: I1008 18:24:24.629986 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9f3ad494-ee2f-40d9-a0a3-ed2824add4b4-catalog-content\") pod \"9f3ad494-ee2f-40d9-a0a3-ed2824add4b4\" (UID: \"9f3ad494-ee2f-40d9-a0a3-ed2824add4b4\") " Oct 08 18:24:24 crc kubenswrapper[4988]: I1008 18:24:24.630099 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qsn9z\" (UniqueName: \"kubernetes.io/projected/9f3ad494-ee2f-40d9-a0a3-ed2824add4b4-kube-api-access-qsn9z\") pod \"9f3ad494-ee2f-40d9-a0a3-ed2824add4b4\" (UID: \"9f3ad494-ee2f-40d9-a0a3-ed2824add4b4\") " Oct 08 18:24:24 crc kubenswrapper[4988]: I1008 18:24:24.630155 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9f3ad494-ee2f-40d9-a0a3-ed2824add4b4-utilities\") pod \"9f3ad494-ee2f-40d9-a0a3-ed2824add4b4\" (UID: \"9f3ad494-ee2f-40d9-a0a3-ed2824add4b4\") " Oct 08 18:24:24 crc kubenswrapper[4988]: I1008 18:24:24.631117 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9f3ad494-ee2f-40d9-a0a3-ed2824add4b4-utilities" (OuterVolumeSpecName: "utilities") pod "9f3ad494-ee2f-40d9-a0a3-ed2824add4b4" (UID: "9f3ad494-ee2f-40d9-a0a3-ed2824add4b4"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 18:24:24 crc kubenswrapper[4988]: I1008 18:24:24.636054 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9f3ad494-ee2f-40d9-a0a3-ed2824add4b4-kube-api-access-qsn9z" (OuterVolumeSpecName: "kube-api-access-qsn9z") pod "9f3ad494-ee2f-40d9-a0a3-ed2824add4b4" (UID: "9f3ad494-ee2f-40d9-a0a3-ed2824add4b4"). InnerVolumeSpecName "kube-api-access-qsn9z". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:24:24 crc kubenswrapper[4988]: I1008 18:24:24.645235 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9f3ad494-ee2f-40d9-a0a3-ed2824add4b4-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "9f3ad494-ee2f-40d9-a0a3-ed2824add4b4" (UID: "9f3ad494-ee2f-40d9-a0a3-ed2824add4b4"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 18:24:24 crc kubenswrapper[4988]: I1008 18:24:24.732265 4988 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9f3ad494-ee2f-40d9-a0a3-ed2824add4b4-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 08 18:24:24 crc kubenswrapper[4988]: I1008 18:24:24.732312 4988 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qsn9z\" (UniqueName: \"kubernetes.io/projected/9f3ad494-ee2f-40d9-a0a3-ed2824add4b4-kube-api-access-qsn9z\") on node \"crc\" DevicePath \"\"" Oct 08 18:24:24 crc kubenswrapper[4988]: I1008 18:24:24.732331 4988 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9f3ad494-ee2f-40d9-a0a3-ed2824add4b4-utilities\") on node \"crc\" DevicePath \"\"" Oct 08 18:24:25 crc kubenswrapper[4988]: I1008 18:24:25.046762 4988 generic.go:334] "Generic (PLEG): container finished" podID="9f3ad494-ee2f-40d9-a0a3-ed2824add4b4" containerID="ebcaf8187a9337f6e8f70903e189fc2626be3a07ec9842cd523ff5b44b03d0dc" exitCode=0 Oct 08 18:24:25 crc kubenswrapper[4988]: I1008 18:24:25.046841 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-x7w9s" Oct 08 18:24:25 crc kubenswrapper[4988]: I1008 18:24:25.046867 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-x7w9s" event={"ID":"9f3ad494-ee2f-40d9-a0a3-ed2824add4b4","Type":"ContainerDied","Data":"ebcaf8187a9337f6e8f70903e189fc2626be3a07ec9842cd523ff5b44b03d0dc"} Oct 08 18:24:25 crc kubenswrapper[4988]: I1008 18:24:25.047628 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-x7w9s" event={"ID":"9f3ad494-ee2f-40d9-a0a3-ed2824add4b4","Type":"ContainerDied","Data":"2384a680cf1e57f4917f96d75c91b11a4eacba725c4c968239b50c9e020ccfdc"} Oct 08 18:24:25 crc kubenswrapper[4988]: I1008 18:24:25.047688 4988 scope.go:117] "RemoveContainer" containerID="ebcaf8187a9337f6e8f70903e189fc2626be3a07ec9842cd523ff5b44b03d0dc" Oct 08 18:24:25 crc kubenswrapper[4988]: I1008 18:24:25.106487 4988 scope.go:117] "RemoveContainer" containerID="dc415c5cf6a2b96cb7281c7740b90752caf851040939ebde2864c1d407db15ac" Oct 08 18:24:25 crc kubenswrapper[4988]: I1008 18:24:25.141228 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-x7w9s"] Oct 08 18:24:25 crc kubenswrapper[4988]: I1008 18:24:25.144071 4988 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-x7w9s"] Oct 08 18:24:25 crc kubenswrapper[4988]: I1008 18:24:25.166611 4988 scope.go:117] "RemoveContainer" containerID="5a9ada12000c7c6e96ba7c496ea3c8adaa91865ff1f34b53fef89a2f9451dd5b" Oct 08 18:24:25 crc kubenswrapper[4988]: I1008 18:24:25.195363 4988 scope.go:117] "RemoveContainer" containerID="ebcaf8187a9337f6e8f70903e189fc2626be3a07ec9842cd523ff5b44b03d0dc" Oct 08 18:24:25 crc kubenswrapper[4988]: E1008 18:24:25.195759 4988 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ebcaf8187a9337f6e8f70903e189fc2626be3a07ec9842cd523ff5b44b03d0dc\": container with ID starting with ebcaf8187a9337f6e8f70903e189fc2626be3a07ec9842cd523ff5b44b03d0dc not found: ID does not exist" containerID="ebcaf8187a9337f6e8f70903e189fc2626be3a07ec9842cd523ff5b44b03d0dc" Oct 08 18:24:25 crc kubenswrapper[4988]: I1008 18:24:25.195786 4988 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ebcaf8187a9337f6e8f70903e189fc2626be3a07ec9842cd523ff5b44b03d0dc"} err="failed to get container status \"ebcaf8187a9337f6e8f70903e189fc2626be3a07ec9842cd523ff5b44b03d0dc\": rpc error: code = NotFound desc = could not find container \"ebcaf8187a9337f6e8f70903e189fc2626be3a07ec9842cd523ff5b44b03d0dc\": container with ID starting with ebcaf8187a9337f6e8f70903e189fc2626be3a07ec9842cd523ff5b44b03d0dc not found: ID does not exist" Oct 08 18:24:25 crc kubenswrapper[4988]: I1008 18:24:25.195805 4988 scope.go:117] "RemoveContainer" containerID="dc415c5cf6a2b96cb7281c7740b90752caf851040939ebde2864c1d407db15ac" Oct 08 18:24:25 crc kubenswrapper[4988]: E1008 18:24:25.196068 4988 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"dc415c5cf6a2b96cb7281c7740b90752caf851040939ebde2864c1d407db15ac\": container with ID starting with dc415c5cf6a2b96cb7281c7740b90752caf851040939ebde2864c1d407db15ac not found: ID does not exist" containerID="dc415c5cf6a2b96cb7281c7740b90752caf851040939ebde2864c1d407db15ac" Oct 08 18:24:25 crc kubenswrapper[4988]: I1008 18:24:25.196107 4988 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dc415c5cf6a2b96cb7281c7740b90752caf851040939ebde2864c1d407db15ac"} err="failed to get container status \"dc415c5cf6a2b96cb7281c7740b90752caf851040939ebde2864c1d407db15ac\": rpc error: code = NotFound desc = could not find container \"dc415c5cf6a2b96cb7281c7740b90752caf851040939ebde2864c1d407db15ac\": container with ID starting with dc415c5cf6a2b96cb7281c7740b90752caf851040939ebde2864c1d407db15ac not found: ID does not exist" Oct 08 18:24:25 crc kubenswrapper[4988]: I1008 18:24:25.196134 4988 scope.go:117] "RemoveContainer" containerID="5a9ada12000c7c6e96ba7c496ea3c8adaa91865ff1f34b53fef89a2f9451dd5b" Oct 08 18:24:25 crc kubenswrapper[4988]: E1008 18:24:25.196434 4988 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5a9ada12000c7c6e96ba7c496ea3c8adaa91865ff1f34b53fef89a2f9451dd5b\": container with ID starting with 5a9ada12000c7c6e96ba7c496ea3c8adaa91865ff1f34b53fef89a2f9451dd5b not found: ID does not exist" containerID="5a9ada12000c7c6e96ba7c496ea3c8adaa91865ff1f34b53fef89a2f9451dd5b" Oct 08 18:24:25 crc kubenswrapper[4988]: I1008 18:24:25.196466 4988 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5a9ada12000c7c6e96ba7c496ea3c8adaa91865ff1f34b53fef89a2f9451dd5b"} err="failed to get container status \"5a9ada12000c7c6e96ba7c496ea3c8adaa91865ff1f34b53fef89a2f9451dd5b\": rpc error: code = NotFound desc = could not find container \"5a9ada12000c7c6e96ba7c496ea3c8adaa91865ff1f34b53fef89a2f9451dd5b\": container with ID starting with 5a9ada12000c7c6e96ba7c496ea3c8adaa91865ff1f34b53fef89a2f9451dd5b not found: ID does not exist" Oct 08 18:24:25 crc kubenswrapper[4988]: I1008 18:24:25.256206 4988 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9f3ad494-ee2f-40d9-a0a3-ed2824add4b4" path="/var/lib/kubelet/pods/9f3ad494-ee2f-40d9-a0a3-ed2824add4b4/volumes" Oct 08 18:24:31 crc kubenswrapper[4988]: I1008 18:24:31.928872 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-controller-manager-55b79cb758-ld9ht" Oct 08 18:24:32 crc kubenswrapper[4988]: I1008 18:24:32.766287 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-xfwvv"] Oct 08 18:24:32 crc kubenswrapper[4988]: E1008 18:24:32.766900 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9f3ad494-ee2f-40d9-a0a3-ed2824add4b4" containerName="extract-content" Oct 08 18:24:32 crc kubenswrapper[4988]: I1008 18:24:32.766922 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="9f3ad494-ee2f-40d9-a0a3-ed2824add4b4" containerName="extract-content" Oct 08 18:24:32 crc kubenswrapper[4988]: E1008 18:24:32.766943 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="188fccde-d7a5-4d7b-ba19-999f6b67ffcd" containerName="extract-utilities" Oct 08 18:24:32 crc kubenswrapper[4988]: I1008 18:24:32.766952 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="188fccde-d7a5-4d7b-ba19-999f6b67ffcd" containerName="extract-utilities" Oct 08 18:24:32 crc kubenswrapper[4988]: E1008 18:24:32.766962 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="188fccde-d7a5-4d7b-ba19-999f6b67ffcd" containerName="extract-content" Oct 08 18:24:32 crc kubenswrapper[4988]: I1008 18:24:32.766971 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="188fccde-d7a5-4d7b-ba19-999f6b67ffcd" containerName="extract-content" Oct 08 18:24:32 crc kubenswrapper[4988]: E1008 18:24:32.766984 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="188fccde-d7a5-4d7b-ba19-999f6b67ffcd" containerName="registry-server" Oct 08 18:24:32 crc kubenswrapper[4988]: I1008 18:24:32.766991 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="188fccde-d7a5-4d7b-ba19-999f6b67ffcd" containerName="registry-server" Oct 08 18:24:32 crc kubenswrapper[4988]: E1008 18:24:32.767003 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9f3ad494-ee2f-40d9-a0a3-ed2824add4b4" containerName="registry-server" Oct 08 18:24:32 crc kubenswrapper[4988]: I1008 18:24:32.767011 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="9f3ad494-ee2f-40d9-a0a3-ed2824add4b4" containerName="registry-server" Oct 08 18:24:32 crc kubenswrapper[4988]: E1008 18:24:32.767026 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9f3ad494-ee2f-40d9-a0a3-ed2824add4b4" containerName="extract-utilities" Oct 08 18:24:32 crc kubenswrapper[4988]: I1008 18:24:32.767033 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="9f3ad494-ee2f-40d9-a0a3-ed2824add4b4" containerName="extract-utilities" Oct 08 18:24:32 crc kubenswrapper[4988]: I1008 18:24:32.767158 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="9f3ad494-ee2f-40d9-a0a3-ed2824add4b4" containerName="registry-server" Oct 08 18:24:32 crc kubenswrapper[4988]: I1008 18:24:32.767175 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="188fccde-d7a5-4d7b-ba19-999f6b67ffcd" containerName="registry-server" Oct 08 18:24:32 crc kubenswrapper[4988]: I1008 18:24:32.769815 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-xfwvv" Oct 08 18:24:32 crc kubenswrapper[4988]: I1008 18:24:32.771354 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-webhook-server-64bf5d555-66f55"] Oct 08 18:24:32 crc kubenswrapper[4988]: I1008 18:24:32.773493 4988 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-daemon-dockercfg-6bcgj" Oct 08 18:24:32 crc kubenswrapper[4988]: I1008 18:24:32.773865 4988 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-certs-secret" Oct 08 18:24:32 crc kubenswrapper[4988]: I1008 18:24:32.773957 4988 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"frr-startup" Oct 08 18:24:32 crc kubenswrapper[4988]: I1008 18:24:32.776572 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-64bf5d555-66f55" Oct 08 18:24:32 crc kubenswrapper[4988]: I1008 18:24:32.780451 4988 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-webhook-server-cert" Oct 08 18:24:32 crc kubenswrapper[4988]: I1008 18:24:32.788619 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-64bf5d555-66f55"] Oct 08 18:24:32 crc kubenswrapper[4988]: I1008 18:24:32.850079 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/speaker-nbhjx"] Oct 08 18:24:32 crc kubenswrapper[4988]: I1008 18:24:32.851286 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-nbhjx" Oct 08 18:24:32 crc kubenswrapper[4988]: I1008 18:24:32.853579 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/5e1113b9-5d57-427e-9c92-0da0866e8e4d-cert\") pod \"frr-k8s-webhook-server-64bf5d555-66f55\" (UID: \"5e1113b9-5d57-427e-9c92-0da0866e8e4d\") " pod="metallb-system/frr-k8s-webhook-server-64bf5d555-66f55" Oct 08 18:24:32 crc kubenswrapper[4988]: I1008 18:24:32.853643 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/f350c4ca-de05-4509-8c61-1941081b8b79-frr-conf\") pod \"frr-k8s-xfwvv\" (UID: \"f350c4ca-de05-4509-8c61-1941081b8b79\") " pod="metallb-system/frr-k8s-xfwvv" Oct 08 18:24:32 crc kubenswrapper[4988]: I1008 18:24:32.853675 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/645996a1-4e9d-471e-ad3f-866864472407-memberlist\") pod \"speaker-nbhjx\" (UID: \"645996a1-4e9d-471e-ad3f-866864472407\") " pod="metallb-system/speaker-nbhjx" Oct 08 18:24:32 crc kubenswrapper[4988]: I1008 18:24:32.853703 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-phs68\" (UniqueName: \"kubernetes.io/projected/645996a1-4e9d-471e-ad3f-866864472407-kube-api-access-phs68\") pod \"speaker-nbhjx\" (UID: \"645996a1-4e9d-471e-ad3f-866864472407\") " pod="metallb-system/speaker-nbhjx" Oct 08 18:24:32 crc kubenswrapper[4988]: I1008 18:24:32.853729 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/f350c4ca-de05-4509-8c61-1941081b8b79-frr-sockets\") pod \"frr-k8s-xfwvv\" (UID: \"f350c4ca-de05-4509-8c61-1941081b8b79\") " pod="metallb-system/frr-k8s-xfwvv" Oct 08 18:24:32 crc kubenswrapper[4988]: I1008 18:24:32.853749 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/f350c4ca-de05-4509-8c61-1941081b8b79-metrics\") pod \"frr-k8s-xfwvv\" (UID: \"f350c4ca-de05-4509-8c61-1941081b8b79\") " pod="metallb-system/frr-k8s-xfwvv" Oct 08 18:24:32 crc kubenswrapper[4988]: I1008 18:24:32.853766 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mdtxw\" (UniqueName: \"kubernetes.io/projected/5e1113b9-5d57-427e-9c92-0da0866e8e4d-kube-api-access-mdtxw\") pod \"frr-k8s-webhook-server-64bf5d555-66f55\" (UID: \"5e1113b9-5d57-427e-9c92-0da0866e8e4d\") " pod="metallb-system/frr-k8s-webhook-server-64bf5d555-66f55" Oct 08 18:24:32 crc kubenswrapper[4988]: I1008 18:24:32.853806 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/645996a1-4e9d-471e-ad3f-866864472407-metrics-certs\") pod \"speaker-nbhjx\" (UID: \"645996a1-4e9d-471e-ad3f-866864472407\") " pod="metallb-system/speaker-nbhjx" Oct 08 18:24:32 crc kubenswrapper[4988]: I1008 18:24:32.853835 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vj2pm\" (UniqueName: \"kubernetes.io/projected/f350c4ca-de05-4509-8c61-1941081b8b79-kube-api-access-vj2pm\") pod \"frr-k8s-xfwvv\" (UID: \"f350c4ca-de05-4509-8c61-1941081b8b79\") " pod="metallb-system/frr-k8s-xfwvv" Oct 08 18:24:32 crc kubenswrapper[4988]: I1008 18:24:32.853861 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/645996a1-4e9d-471e-ad3f-866864472407-metallb-excludel2\") pod \"speaker-nbhjx\" (UID: \"645996a1-4e9d-471e-ad3f-866864472407\") " pod="metallb-system/speaker-nbhjx" Oct 08 18:24:32 crc kubenswrapper[4988]: I1008 18:24:32.853889 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/f350c4ca-de05-4509-8c61-1941081b8b79-frr-startup\") pod \"frr-k8s-xfwvv\" (UID: \"f350c4ca-de05-4509-8c61-1941081b8b79\") " pod="metallb-system/frr-k8s-xfwvv" Oct 08 18:24:32 crc kubenswrapper[4988]: I1008 18:24:32.853911 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/f350c4ca-de05-4509-8c61-1941081b8b79-reloader\") pod \"frr-k8s-xfwvv\" (UID: \"f350c4ca-de05-4509-8c61-1941081b8b79\") " pod="metallb-system/frr-k8s-xfwvv" Oct 08 18:24:32 crc kubenswrapper[4988]: I1008 18:24:32.853935 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/f350c4ca-de05-4509-8c61-1941081b8b79-metrics-certs\") pod \"frr-k8s-xfwvv\" (UID: \"f350c4ca-de05-4509-8c61-1941081b8b79\") " pod="metallb-system/frr-k8s-xfwvv" Oct 08 18:24:32 crc kubenswrapper[4988]: I1008 18:24:32.854470 4988 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-memberlist" Oct 08 18:24:32 crc kubenswrapper[4988]: I1008 18:24:32.854674 4988 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"metallb-excludel2" Oct 08 18:24:32 crc kubenswrapper[4988]: I1008 18:24:32.854811 4988 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-dockercfg-drmwr" Oct 08 18:24:32 crc kubenswrapper[4988]: I1008 18:24:32.854952 4988 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-certs-secret" Oct 08 18:24:32 crc kubenswrapper[4988]: I1008 18:24:32.883835 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/controller-68d546b9d8-kxkcv"] Oct 08 18:24:32 crc kubenswrapper[4988]: I1008 18:24:32.885144 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-68d546b9d8-kxkcv" Oct 08 18:24:32 crc kubenswrapper[4988]: I1008 18:24:32.889952 4988 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-certs-secret" Oct 08 18:24:32 crc kubenswrapper[4988]: I1008 18:24:32.910662 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-68d546b9d8-kxkcv"] Oct 08 18:24:32 crc kubenswrapper[4988]: I1008 18:24:32.954672 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/645996a1-4e9d-471e-ad3f-866864472407-metrics-certs\") pod \"speaker-nbhjx\" (UID: \"645996a1-4e9d-471e-ad3f-866864472407\") " pod="metallb-system/speaker-nbhjx" Oct 08 18:24:32 crc kubenswrapper[4988]: I1008 18:24:32.954749 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vj2pm\" (UniqueName: \"kubernetes.io/projected/f350c4ca-de05-4509-8c61-1941081b8b79-kube-api-access-vj2pm\") pod \"frr-k8s-xfwvv\" (UID: \"f350c4ca-de05-4509-8c61-1941081b8b79\") " pod="metallb-system/frr-k8s-xfwvv" Oct 08 18:24:32 crc kubenswrapper[4988]: I1008 18:24:32.954778 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qrlff\" (UniqueName: \"kubernetes.io/projected/dcae2201-01d1-4f23-bf2d-134d6c01a1a4-kube-api-access-qrlff\") pod \"controller-68d546b9d8-kxkcv\" (UID: \"dcae2201-01d1-4f23-bf2d-134d6c01a1a4\") " pod="metallb-system/controller-68d546b9d8-kxkcv" Oct 08 18:24:32 crc kubenswrapper[4988]: I1008 18:24:32.954800 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/645996a1-4e9d-471e-ad3f-866864472407-metallb-excludel2\") pod \"speaker-nbhjx\" (UID: \"645996a1-4e9d-471e-ad3f-866864472407\") " pod="metallb-system/speaker-nbhjx" Oct 08 18:24:32 crc kubenswrapper[4988]: I1008 18:24:32.954818 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/f350c4ca-de05-4509-8c61-1941081b8b79-frr-startup\") pod \"frr-k8s-xfwvv\" (UID: \"f350c4ca-de05-4509-8c61-1941081b8b79\") " pod="metallb-system/frr-k8s-xfwvv" Oct 08 18:24:32 crc kubenswrapper[4988]: I1008 18:24:32.954835 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/f350c4ca-de05-4509-8c61-1941081b8b79-reloader\") pod \"frr-k8s-xfwvv\" (UID: \"f350c4ca-de05-4509-8c61-1941081b8b79\") " pod="metallb-system/frr-k8s-xfwvv" Oct 08 18:24:32 crc kubenswrapper[4988]: E1008 18:24:32.955149 4988 secret.go:188] Couldn't get secret metallb-system/speaker-certs-secret: secret "speaker-certs-secret" not found Oct 08 18:24:32 crc kubenswrapper[4988]: E1008 18:24:32.955211 4988 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/645996a1-4e9d-471e-ad3f-866864472407-metrics-certs podName:645996a1-4e9d-471e-ad3f-866864472407 nodeName:}" failed. No retries permitted until 2025-10-08 18:24:33.455193566 +0000 UTC m=+818.905036336 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/645996a1-4e9d-471e-ad3f-866864472407-metrics-certs") pod "speaker-nbhjx" (UID: "645996a1-4e9d-471e-ad3f-866864472407") : secret "speaker-certs-secret" not found Oct 08 18:24:32 crc kubenswrapper[4988]: I1008 18:24:32.955357 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/dcae2201-01d1-4f23-bf2d-134d6c01a1a4-cert\") pod \"controller-68d546b9d8-kxkcv\" (UID: \"dcae2201-01d1-4f23-bf2d-134d6c01a1a4\") " pod="metallb-system/controller-68d546b9d8-kxkcv" Oct 08 18:24:32 crc kubenswrapper[4988]: I1008 18:24:32.955401 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/dcae2201-01d1-4f23-bf2d-134d6c01a1a4-metrics-certs\") pod \"controller-68d546b9d8-kxkcv\" (UID: \"dcae2201-01d1-4f23-bf2d-134d6c01a1a4\") " pod="metallb-system/controller-68d546b9d8-kxkcv" Oct 08 18:24:32 crc kubenswrapper[4988]: I1008 18:24:32.955458 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/f350c4ca-de05-4509-8c61-1941081b8b79-metrics-certs\") pod \"frr-k8s-xfwvv\" (UID: \"f350c4ca-de05-4509-8c61-1941081b8b79\") " pod="metallb-system/frr-k8s-xfwvv" Oct 08 18:24:32 crc kubenswrapper[4988]: I1008 18:24:32.955596 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/5e1113b9-5d57-427e-9c92-0da0866e8e4d-cert\") pod \"frr-k8s-webhook-server-64bf5d555-66f55\" (UID: \"5e1113b9-5d57-427e-9c92-0da0866e8e4d\") " pod="metallb-system/frr-k8s-webhook-server-64bf5d555-66f55" Oct 08 18:24:32 crc kubenswrapper[4988]: I1008 18:24:32.955618 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/f350c4ca-de05-4509-8c61-1941081b8b79-frr-conf\") pod \"frr-k8s-xfwvv\" (UID: \"f350c4ca-de05-4509-8c61-1941081b8b79\") " pod="metallb-system/frr-k8s-xfwvv" Oct 08 18:24:32 crc kubenswrapper[4988]: I1008 18:24:32.955646 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/645996a1-4e9d-471e-ad3f-866864472407-memberlist\") pod \"speaker-nbhjx\" (UID: \"645996a1-4e9d-471e-ad3f-866864472407\") " pod="metallb-system/speaker-nbhjx" Oct 08 18:24:32 crc kubenswrapper[4988]: I1008 18:24:32.955678 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-phs68\" (UniqueName: \"kubernetes.io/projected/645996a1-4e9d-471e-ad3f-866864472407-kube-api-access-phs68\") pod \"speaker-nbhjx\" (UID: \"645996a1-4e9d-471e-ad3f-866864472407\") " pod="metallb-system/speaker-nbhjx" Oct 08 18:24:32 crc kubenswrapper[4988]: I1008 18:24:32.955709 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/f350c4ca-de05-4509-8c61-1941081b8b79-metrics\") pod \"frr-k8s-xfwvv\" (UID: \"f350c4ca-de05-4509-8c61-1941081b8b79\") " pod="metallb-system/frr-k8s-xfwvv" Oct 08 18:24:32 crc kubenswrapper[4988]: I1008 18:24:32.955730 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mdtxw\" (UniqueName: \"kubernetes.io/projected/5e1113b9-5d57-427e-9c92-0da0866e8e4d-kube-api-access-mdtxw\") pod \"frr-k8s-webhook-server-64bf5d555-66f55\" (UID: \"5e1113b9-5d57-427e-9c92-0da0866e8e4d\") " pod="metallb-system/frr-k8s-webhook-server-64bf5d555-66f55" Oct 08 18:24:32 crc kubenswrapper[4988]: I1008 18:24:32.955754 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/f350c4ca-de05-4509-8c61-1941081b8b79-frr-sockets\") pod \"frr-k8s-xfwvv\" (UID: \"f350c4ca-de05-4509-8c61-1941081b8b79\") " pod="metallb-system/frr-k8s-xfwvv" Oct 08 18:24:32 crc kubenswrapper[4988]: E1008 18:24:32.956260 4988 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Oct 08 18:24:32 crc kubenswrapper[4988]: E1008 18:24:32.956299 4988 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/645996a1-4e9d-471e-ad3f-866864472407-memberlist podName:645996a1-4e9d-471e-ad3f-866864472407 nodeName:}" failed. No retries permitted until 2025-10-08 18:24:33.456290232 +0000 UTC m=+818.906133002 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/645996a1-4e9d-471e-ad3f-866864472407-memberlist") pod "speaker-nbhjx" (UID: "645996a1-4e9d-471e-ad3f-866864472407") : secret "metallb-memberlist" not found Oct 08 18:24:32 crc kubenswrapper[4988]: E1008 18:24:32.956261 4988 secret.go:188] Couldn't get secret metallb-system/frr-k8s-certs-secret: secret "frr-k8s-certs-secret" not found Oct 08 18:24:32 crc kubenswrapper[4988]: I1008 18:24:32.956546 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/f350c4ca-de05-4509-8c61-1941081b8b79-frr-conf\") pod \"frr-k8s-xfwvv\" (UID: \"f350c4ca-de05-4509-8c61-1941081b8b79\") " pod="metallb-system/frr-k8s-xfwvv" Oct 08 18:24:32 crc kubenswrapper[4988]: E1008 18:24:32.956778 4988 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/f350c4ca-de05-4509-8c61-1941081b8b79-metrics-certs podName:f350c4ca-de05-4509-8c61-1941081b8b79 nodeName:}" failed. No retries permitted until 2025-10-08 18:24:33.456753247 +0000 UTC m=+818.906596087 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/f350c4ca-de05-4509-8c61-1941081b8b79-metrics-certs") pod "frr-k8s-xfwvv" (UID: "f350c4ca-de05-4509-8c61-1941081b8b79") : secret "frr-k8s-certs-secret" not found Oct 08 18:24:32 crc kubenswrapper[4988]: I1008 18:24:32.956923 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/f350c4ca-de05-4509-8c61-1941081b8b79-frr-startup\") pod \"frr-k8s-xfwvv\" (UID: \"f350c4ca-de05-4509-8c61-1941081b8b79\") " pod="metallb-system/frr-k8s-xfwvv" Oct 08 18:24:32 crc kubenswrapper[4988]: I1008 18:24:32.957002 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/645996a1-4e9d-471e-ad3f-866864472407-metallb-excludel2\") pod \"speaker-nbhjx\" (UID: \"645996a1-4e9d-471e-ad3f-866864472407\") " pod="metallb-system/speaker-nbhjx" Oct 08 18:24:32 crc kubenswrapper[4988]: I1008 18:24:32.957064 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/f350c4ca-de05-4509-8c61-1941081b8b79-frr-sockets\") pod \"frr-k8s-xfwvv\" (UID: \"f350c4ca-de05-4509-8c61-1941081b8b79\") " pod="metallb-system/frr-k8s-xfwvv" Oct 08 18:24:32 crc kubenswrapper[4988]: I1008 18:24:32.957322 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/f350c4ca-de05-4509-8c61-1941081b8b79-reloader\") pod \"frr-k8s-xfwvv\" (UID: \"f350c4ca-de05-4509-8c61-1941081b8b79\") " pod="metallb-system/frr-k8s-xfwvv" Oct 08 18:24:32 crc kubenswrapper[4988]: I1008 18:24:32.957421 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/f350c4ca-de05-4509-8c61-1941081b8b79-metrics\") pod \"frr-k8s-xfwvv\" (UID: \"f350c4ca-de05-4509-8c61-1941081b8b79\") " pod="metallb-system/frr-k8s-xfwvv" Oct 08 18:24:32 crc kubenswrapper[4988]: I1008 18:24:32.968213 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/5e1113b9-5d57-427e-9c92-0da0866e8e4d-cert\") pod \"frr-k8s-webhook-server-64bf5d555-66f55\" (UID: \"5e1113b9-5d57-427e-9c92-0da0866e8e4d\") " pod="metallb-system/frr-k8s-webhook-server-64bf5d555-66f55" Oct 08 18:24:32 crc kubenswrapper[4988]: I1008 18:24:32.990332 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-phs68\" (UniqueName: \"kubernetes.io/projected/645996a1-4e9d-471e-ad3f-866864472407-kube-api-access-phs68\") pod \"speaker-nbhjx\" (UID: \"645996a1-4e9d-471e-ad3f-866864472407\") " pod="metallb-system/speaker-nbhjx" Oct 08 18:24:32 crc kubenswrapper[4988]: I1008 18:24:32.990849 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vj2pm\" (UniqueName: \"kubernetes.io/projected/f350c4ca-de05-4509-8c61-1941081b8b79-kube-api-access-vj2pm\") pod \"frr-k8s-xfwvv\" (UID: \"f350c4ca-de05-4509-8c61-1941081b8b79\") " pod="metallb-system/frr-k8s-xfwvv" Oct 08 18:24:33 crc kubenswrapper[4988]: I1008 18:24:33.014111 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mdtxw\" (UniqueName: \"kubernetes.io/projected/5e1113b9-5d57-427e-9c92-0da0866e8e4d-kube-api-access-mdtxw\") pod \"frr-k8s-webhook-server-64bf5d555-66f55\" (UID: \"5e1113b9-5d57-427e-9c92-0da0866e8e4d\") " pod="metallb-system/frr-k8s-webhook-server-64bf5d555-66f55" Oct 08 18:24:33 crc kubenswrapper[4988]: I1008 18:24:33.057001 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qrlff\" (UniqueName: \"kubernetes.io/projected/dcae2201-01d1-4f23-bf2d-134d6c01a1a4-kube-api-access-qrlff\") pod \"controller-68d546b9d8-kxkcv\" (UID: \"dcae2201-01d1-4f23-bf2d-134d6c01a1a4\") " pod="metallb-system/controller-68d546b9d8-kxkcv" Oct 08 18:24:33 crc kubenswrapper[4988]: I1008 18:24:33.057049 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/dcae2201-01d1-4f23-bf2d-134d6c01a1a4-cert\") pod \"controller-68d546b9d8-kxkcv\" (UID: \"dcae2201-01d1-4f23-bf2d-134d6c01a1a4\") " pod="metallb-system/controller-68d546b9d8-kxkcv" Oct 08 18:24:33 crc kubenswrapper[4988]: I1008 18:24:33.057066 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/dcae2201-01d1-4f23-bf2d-134d6c01a1a4-metrics-certs\") pod \"controller-68d546b9d8-kxkcv\" (UID: \"dcae2201-01d1-4f23-bf2d-134d6c01a1a4\") " pod="metallb-system/controller-68d546b9d8-kxkcv" Oct 08 18:24:33 crc kubenswrapper[4988]: I1008 18:24:33.061013 4988 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-webhook-cert" Oct 08 18:24:33 crc kubenswrapper[4988]: I1008 18:24:33.063924 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/dcae2201-01d1-4f23-bf2d-134d6c01a1a4-metrics-certs\") pod \"controller-68d546b9d8-kxkcv\" (UID: \"dcae2201-01d1-4f23-bf2d-134d6c01a1a4\") " pod="metallb-system/controller-68d546b9d8-kxkcv" Oct 08 18:24:33 crc kubenswrapper[4988]: I1008 18:24:33.071783 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/dcae2201-01d1-4f23-bf2d-134d6c01a1a4-cert\") pod \"controller-68d546b9d8-kxkcv\" (UID: \"dcae2201-01d1-4f23-bf2d-134d6c01a1a4\") " pod="metallb-system/controller-68d546b9d8-kxkcv" Oct 08 18:24:33 crc kubenswrapper[4988]: I1008 18:24:33.075543 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qrlff\" (UniqueName: \"kubernetes.io/projected/dcae2201-01d1-4f23-bf2d-134d6c01a1a4-kube-api-access-qrlff\") pod \"controller-68d546b9d8-kxkcv\" (UID: \"dcae2201-01d1-4f23-bf2d-134d6c01a1a4\") " pod="metallb-system/controller-68d546b9d8-kxkcv" Oct 08 18:24:33 crc kubenswrapper[4988]: I1008 18:24:33.099290 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-64bf5d555-66f55" Oct 08 18:24:33 crc kubenswrapper[4988]: I1008 18:24:33.200715 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-68d546b9d8-kxkcv" Oct 08 18:24:33 crc kubenswrapper[4988]: I1008 18:24:33.467228 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/645996a1-4e9d-471e-ad3f-866864472407-memberlist\") pod \"speaker-nbhjx\" (UID: \"645996a1-4e9d-471e-ad3f-866864472407\") " pod="metallb-system/speaker-nbhjx" Oct 08 18:24:33 crc kubenswrapper[4988]: I1008 18:24:33.467317 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/645996a1-4e9d-471e-ad3f-866864472407-metrics-certs\") pod \"speaker-nbhjx\" (UID: \"645996a1-4e9d-471e-ad3f-866864472407\") " pod="metallb-system/speaker-nbhjx" Oct 08 18:24:33 crc kubenswrapper[4988]: I1008 18:24:33.467361 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/f350c4ca-de05-4509-8c61-1941081b8b79-metrics-certs\") pod \"frr-k8s-xfwvv\" (UID: \"f350c4ca-de05-4509-8c61-1941081b8b79\") " pod="metallb-system/frr-k8s-xfwvv" Oct 08 18:24:33 crc kubenswrapper[4988]: E1008 18:24:33.468557 4988 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Oct 08 18:24:33 crc kubenswrapper[4988]: E1008 18:24:33.468626 4988 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/645996a1-4e9d-471e-ad3f-866864472407-memberlist podName:645996a1-4e9d-471e-ad3f-866864472407 nodeName:}" failed. No retries permitted until 2025-10-08 18:24:34.468602783 +0000 UTC m=+819.918445553 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/645996a1-4e9d-471e-ad3f-866864472407-memberlist") pod "speaker-nbhjx" (UID: "645996a1-4e9d-471e-ad3f-866864472407") : secret "metallb-memberlist" not found Oct 08 18:24:33 crc kubenswrapper[4988]: I1008 18:24:33.472713 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/645996a1-4e9d-471e-ad3f-866864472407-metrics-certs\") pod \"speaker-nbhjx\" (UID: \"645996a1-4e9d-471e-ad3f-866864472407\") " pod="metallb-system/speaker-nbhjx" Oct 08 18:24:33 crc kubenswrapper[4988]: I1008 18:24:33.474092 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/f350c4ca-de05-4509-8c61-1941081b8b79-metrics-certs\") pod \"frr-k8s-xfwvv\" (UID: \"f350c4ca-de05-4509-8c61-1941081b8b79\") " pod="metallb-system/frr-k8s-xfwvv" Oct 08 18:24:33 crc kubenswrapper[4988]: I1008 18:24:33.599018 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-64bf5d555-66f55"] Oct 08 18:24:33 crc kubenswrapper[4988]: I1008 18:24:33.685544 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-68d546b9d8-kxkcv"] Oct 08 18:24:33 crc kubenswrapper[4988]: I1008 18:24:33.692489 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-xfwvv" Oct 08 18:24:33 crc kubenswrapper[4988]: W1008 18:24:33.694410 4988 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poddcae2201_01d1_4f23_bf2d_134d6c01a1a4.slice/crio-c28bba0a8dd6d9ac745dc6d01e82b0c7a68133ddffda8e514ae783cf4a64bee5 WatchSource:0}: Error finding container c28bba0a8dd6d9ac745dc6d01e82b0c7a68133ddffda8e514ae783cf4a64bee5: Status 404 returned error can't find the container with id c28bba0a8dd6d9ac745dc6d01e82b0c7a68133ddffda8e514ae783cf4a64bee5 Oct 08 18:24:34 crc kubenswrapper[4988]: I1008 18:24:34.101498 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-xfwvv" event={"ID":"f350c4ca-de05-4509-8c61-1941081b8b79","Type":"ContainerStarted","Data":"42e5f542941d64e6ffb76cc3676b51be20a415994ac17ceb0915ccb99ef2b47d"} Oct 08 18:24:34 crc kubenswrapper[4988]: I1008 18:24:34.104354 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-68d546b9d8-kxkcv" event={"ID":"dcae2201-01d1-4f23-bf2d-134d6c01a1a4","Type":"ContainerStarted","Data":"455740fa065a3d8a4f429232f25ca16d40ddb192c56b73892851cbbfebeb02a5"} Oct 08 18:24:34 crc kubenswrapper[4988]: I1008 18:24:34.104413 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-68d546b9d8-kxkcv" event={"ID":"dcae2201-01d1-4f23-bf2d-134d6c01a1a4","Type":"ContainerStarted","Data":"9715bf454dc76ea42af7b4d9102afb3746b3f05c04c8554952a3682abcb6a6a1"} Oct 08 18:24:34 crc kubenswrapper[4988]: I1008 18:24:34.104423 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-68d546b9d8-kxkcv" event={"ID":"dcae2201-01d1-4f23-bf2d-134d6c01a1a4","Type":"ContainerStarted","Data":"c28bba0a8dd6d9ac745dc6d01e82b0c7a68133ddffda8e514ae783cf4a64bee5"} Oct 08 18:24:34 crc kubenswrapper[4988]: I1008 18:24:34.104800 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/controller-68d546b9d8-kxkcv" Oct 08 18:24:34 crc kubenswrapper[4988]: I1008 18:24:34.106419 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-64bf5d555-66f55" event={"ID":"5e1113b9-5d57-427e-9c92-0da0866e8e4d","Type":"ContainerStarted","Data":"c86ae2437d1c35221fc92d582b2b09045c159b0b4c0e351d79912528f5bec078"} Oct 08 18:24:34 crc kubenswrapper[4988]: I1008 18:24:34.120776 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/controller-68d546b9d8-kxkcv" podStartSLOduration=2.120744406 podStartE2EDuration="2.120744406s" podCreationTimestamp="2025-10-08 18:24:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 18:24:34.119119304 +0000 UTC m=+819.568962094" watchObservedRunningTime="2025-10-08 18:24:34.120744406 +0000 UTC m=+819.570587177" Oct 08 18:24:34 crc kubenswrapper[4988]: I1008 18:24:34.480666 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/645996a1-4e9d-471e-ad3f-866864472407-memberlist\") pod \"speaker-nbhjx\" (UID: \"645996a1-4e9d-471e-ad3f-866864472407\") " pod="metallb-system/speaker-nbhjx" Oct 08 18:24:34 crc kubenswrapper[4988]: I1008 18:24:34.489252 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/645996a1-4e9d-471e-ad3f-866864472407-memberlist\") pod \"speaker-nbhjx\" (UID: \"645996a1-4e9d-471e-ad3f-866864472407\") " pod="metallb-system/speaker-nbhjx" Oct 08 18:24:34 crc kubenswrapper[4988]: I1008 18:24:34.665836 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-nbhjx" Oct 08 18:24:34 crc kubenswrapper[4988]: W1008 18:24:34.691289 4988 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod645996a1_4e9d_471e_ad3f_866864472407.slice/crio-f9bc8dd054b5fea278c716fce99b03e3c57b83b76ef6fec927d9005233d5dc11 WatchSource:0}: Error finding container f9bc8dd054b5fea278c716fce99b03e3c57b83b76ef6fec927d9005233d5dc11: Status 404 returned error can't find the container with id f9bc8dd054b5fea278c716fce99b03e3c57b83b76ef6fec927d9005233d5dc11 Oct 08 18:24:35 crc kubenswrapper[4988]: I1008 18:24:35.124759 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-nbhjx" event={"ID":"645996a1-4e9d-471e-ad3f-866864472407","Type":"ContainerStarted","Data":"950418832a4be073ec21ff5000c31e8c9edf291326f853cbc69962228706cda2"} Oct 08 18:24:35 crc kubenswrapper[4988]: I1008 18:24:35.124805 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-nbhjx" event={"ID":"645996a1-4e9d-471e-ad3f-866864472407","Type":"ContainerStarted","Data":"f9bc8dd054b5fea278c716fce99b03e3c57b83b76ef6fec927d9005233d5dc11"} Oct 08 18:24:36 crc kubenswrapper[4988]: I1008 18:24:36.135219 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-nbhjx" event={"ID":"645996a1-4e9d-471e-ad3f-866864472407","Type":"ContainerStarted","Data":"fa71f7faee3106c84acd75329b97c88f49493a737cf0cea0cce3a86c0bdba442"} Oct 08 18:24:36 crc kubenswrapper[4988]: I1008 18:24:36.135540 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/speaker-nbhjx" Oct 08 18:24:36 crc kubenswrapper[4988]: I1008 18:24:36.156776 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/speaker-nbhjx" podStartSLOduration=4.156758058 podStartE2EDuration="4.156758058s" podCreationTimestamp="2025-10-08 18:24:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 18:24:36.155928921 +0000 UTC m=+821.605771721" watchObservedRunningTime="2025-10-08 18:24:36.156758058 +0000 UTC m=+821.606600828" Oct 08 18:24:41 crc kubenswrapper[4988]: I1008 18:24:41.165180 4988 generic.go:334] "Generic (PLEG): container finished" podID="f350c4ca-de05-4509-8c61-1941081b8b79" containerID="097adf15cafe474f2bd34daecebedc377b25a3f02738a288572353b478fd88ea" exitCode=0 Oct 08 18:24:41 crc kubenswrapper[4988]: I1008 18:24:41.165242 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-xfwvv" event={"ID":"f350c4ca-de05-4509-8c61-1941081b8b79","Type":"ContainerDied","Data":"097adf15cafe474f2bd34daecebedc377b25a3f02738a288572353b478fd88ea"} Oct 08 18:24:41 crc kubenswrapper[4988]: I1008 18:24:41.167827 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-64bf5d555-66f55" event={"ID":"5e1113b9-5d57-427e-9c92-0da0866e8e4d","Type":"ContainerStarted","Data":"e881b7391821700d3c37611197905b4e75e020077927820dc3ac0fbe06d041df"} Oct 08 18:24:41 crc kubenswrapper[4988]: I1008 18:24:41.167965 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-webhook-server-64bf5d555-66f55" Oct 08 18:24:41 crc kubenswrapper[4988]: I1008 18:24:41.202265 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-webhook-server-64bf5d555-66f55" podStartSLOduration=2.136719167 podStartE2EDuration="9.202241399s" podCreationTimestamp="2025-10-08 18:24:32 +0000 UTC" firstStartedPulling="2025-10-08 18:24:33.609003074 +0000 UTC m=+819.058845884" lastFinishedPulling="2025-10-08 18:24:40.674525336 +0000 UTC m=+826.124368116" observedRunningTime="2025-10-08 18:24:41.200313337 +0000 UTC m=+826.650156127" watchObservedRunningTime="2025-10-08 18:24:41.202241399 +0000 UTC m=+826.652084189" Oct 08 18:24:42 crc kubenswrapper[4988]: I1008 18:24:42.176061 4988 generic.go:334] "Generic (PLEG): container finished" podID="f350c4ca-de05-4509-8c61-1941081b8b79" containerID="8fa938b9e74dd61fe2070f60fba99288ce82c7b630120bcc9363e50f615b5c67" exitCode=0 Oct 08 18:24:42 crc kubenswrapper[4988]: I1008 18:24:42.176178 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-xfwvv" event={"ID":"f350c4ca-de05-4509-8c61-1941081b8b79","Type":"ContainerDied","Data":"8fa938b9e74dd61fe2070f60fba99288ce82c7b630120bcc9363e50f615b5c67"} Oct 08 18:24:43 crc kubenswrapper[4988]: I1008 18:24:43.184360 4988 generic.go:334] "Generic (PLEG): container finished" podID="f350c4ca-de05-4509-8c61-1941081b8b79" containerID="c7ab65fe88f39c8fb8313875b255bfc72529380bd375920344313dc0387885aa" exitCode=0 Oct 08 18:24:43 crc kubenswrapper[4988]: I1008 18:24:43.184422 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-xfwvv" event={"ID":"f350c4ca-de05-4509-8c61-1941081b8b79","Type":"ContainerDied","Data":"c7ab65fe88f39c8fb8313875b255bfc72529380bd375920344313dc0387885aa"} Oct 08 18:24:43 crc kubenswrapper[4988]: I1008 18:24:43.205606 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/controller-68d546b9d8-kxkcv" Oct 08 18:24:44 crc kubenswrapper[4988]: I1008 18:24:44.195045 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-xfwvv" event={"ID":"f350c4ca-de05-4509-8c61-1941081b8b79","Type":"ContainerStarted","Data":"6cf65044fa54b0788714b86b2667e6276c6a9f1a5ec47a75dd25b0fe1b61bff2"} Oct 08 18:24:44 crc kubenswrapper[4988]: I1008 18:24:44.195628 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-xfwvv" Oct 08 18:24:44 crc kubenswrapper[4988]: I1008 18:24:44.195642 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-xfwvv" event={"ID":"f350c4ca-de05-4509-8c61-1941081b8b79","Type":"ContainerStarted","Data":"3c5f0f744bf4e69fdfe6b5365e20b0747e072e4d92d1670f72563c0e31b0ad80"} Oct 08 18:24:44 crc kubenswrapper[4988]: I1008 18:24:44.195656 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-xfwvv" event={"ID":"f350c4ca-de05-4509-8c61-1941081b8b79","Type":"ContainerStarted","Data":"f0a52bfafa0894e946cfafc50cf627747d809eadb30f7c8811b763a2f8bc0f37"} Oct 08 18:24:44 crc kubenswrapper[4988]: I1008 18:24:44.195667 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-xfwvv" event={"ID":"f350c4ca-de05-4509-8c61-1941081b8b79","Type":"ContainerStarted","Data":"62b0912e0a646d87ef6a9f67a06afd4d812225050fd7b8a632bac13cb1cc3cdb"} Oct 08 18:24:44 crc kubenswrapper[4988]: I1008 18:24:44.195678 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-xfwvv" event={"ID":"f350c4ca-de05-4509-8c61-1941081b8b79","Type":"ContainerStarted","Data":"1f35eda6dfba5847bfcd7a6f945abe390476a42e6f5c04cd5f365e66ca5e1a9c"} Oct 08 18:24:44 crc kubenswrapper[4988]: I1008 18:24:44.195688 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-xfwvv" event={"ID":"f350c4ca-de05-4509-8c61-1941081b8b79","Type":"ContainerStarted","Data":"78e902d91bff89ca6dad7bb2fe5745018545f3460ecd0ae8d66d75bf566b59b0"} Oct 08 18:24:44 crc kubenswrapper[4988]: I1008 18:24:44.670565 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/speaker-nbhjx" Oct 08 18:24:44 crc kubenswrapper[4988]: I1008 18:24:44.688028 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-xfwvv" podStartSLOduration=5.858605712 podStartE2EDuration="12.687998179s" podCreationTimestamp="2025-10-08 18:24:32 +0000 UTC" firstStartedPulling="2025-10-08 18:24:33.845700318 +0000 UTC m=+819.295543088" lastFinishedPulling="2025-10-08 18:24:40.675092785 +0000 UTC m=+826.124935555" observedRunningTime="2025-10-08 18:24:44.225936402 +0000 UTC m=+829.675779172" watchObservedRunningTime="2025-10-08 18:24:44.687998179 +0000 UTC m=+830.137840989" Oct 08 18:24:46 crc kubenswrapper[4988]: I1008 18:24:46.283033 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69rf72h"] Oct 08 18:24:46 crc kubenswrapper[4988]: I1008 18:24:46.284658 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69rf72h" Oct 08 18:24:46 crc kubenswrapper[4988]: I1008 18:24:46.286903 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Oct 08 18:24:46 crc kubenswrapper[4988]: I1008 18:24:46.303873 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69rf72h"] Oct 08 18:24:46 crc kubenswrapper[4988]: I1008 18:24:46.445979 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/86be9f4e-06b6-4024-a098-820172b202db-bundle\") pod \"695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69rf72h\" (UID: \"86be9f4e-06b6-4024-a098-820172b202db\") " pod="openshift-marketplace/695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69rf72h" Oct 08 18:24:46 crc kubenswrapper[4988]: I1008 18:24:46.446043 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/86be9f4e-06b6-4024-a098-820172b202db-util\") pod \"695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69rf72h\" (UID: \"86be9f4e-06b6-4024-a098-820172b202db\") " pod="openshift-marketplace/695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69rf72h" Oct 08 18:24:46 crc kubenswrapper[4988]: I1008 18:24:46.446122 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f2krj\" (UniqueName: \"kubernetes.io/projected/86be9f4e-06b6-4024-a098-820172b202db-kube-api-access-f2krj\") pod \"695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69rf72h\" (UID: \"86be9f4e-06b6-4024-a098-820172b202db\") " pod="openshift-marketplace/695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69rf72h" Oct 08 18:24:46 crc kubenswrapper[4988]: I1008 18:24:46.546971 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/86be9f4e-06b6-4024-a098-820172b202db-bundle\") pod \"695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69rf72h\" (UID: \"86be9f4e-06b6-4024-a098-820172b202db\") " pod="openshift-marketplace/695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69rf72h" Oct 08 18:24:46 crc kubenswrapper[4988]: I1008 18:24:46.547033 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/86be9f4e-06b6-4024-a098-820172b202db-util\") pod \"695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69rf72h\" (UID: \"86be9f4e-06b6-4024-a098-820172b202db\") " pod="openshift-marketplace/695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69rf72h" Oct 08 18:24:46 crc kubenswrapper[4988]: I1008 18:24:46.547081 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f2krj\" (UniqueName: \"kubernetes.io/projected/86be9f4e-06b6-4024-a098-820172b202db-kube-api-access-f2krj\") pod \"695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69rf72h\" (UID: \"86be9f4e-06b6-4024-a098-820172b202db\") " pod="openshift-marketplace/695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69rf72h" Oct 08 18:24:46 crc kubenswrapper[4988]: I1008 18:24:46.547376 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/86be9f4e-06b6-4024-a098-820172b202db-bundle\") pod \"695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69rf72h\" (UID: \"86be9f4e-06b6-4024-a098-820172b202db\") " pod="openshift-marketplace/695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69rf72h" Oct 08 18:24:46 crc kubenswrapper[4988]: I1008 18:24:46.547497 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/86be9f4e-06b6-4024-a098-820172b202db-util\") pod \"695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69rf72h\" (UID: \"86be9f4e-06b6-4024-a098-820172b202db\") " pod="openshift-marketplace/695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69rf72h" Oct 08 18:24:46 crc kubenswrapper[4988]: I1008 18:24:46.580308 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f2krj\" (UniqueName: \"kubernetes.io/projected/86be9f4e-06b6-4024-a098-820172b202db-kube-api-access-f2krj\") pod \"695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69rf72h\" (UID: \"86be9f4e-06b6-4024-a098-820172b202db\") " pod="openshift-marketplace/695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69rf72h" Oct 08 18:24:46 crc kubenswrapper[4988]: I1008 18:24:46.604058 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69rf72h" Oct 08 18:24:47 crc kubenswrapper[4988]: I1008 18:24:47.009268 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69rf72h"] Oct 08 18:24:47 crc kubenswrapper[4988]: I1008 18:24:47.216698 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69rf72h" event={"ID":"86be9f4e-06b6-4024-a098-820172b202db","Type":"ContainerStarted","Data":"4d58a21806551dc6c93955568a03ea09c6e5050149ed66ac74f81c0555b44706"} Oct 08 18:24:47 crc kubenswrapper[4988]: I1008 18:24:47.216753 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69rf72h" event={"ID":"86be9f4e-06b6-4024-a098-820172b202db","Type":"ContainerStarted","Data":"4223165cbc84b7f436c65efe062a6e02e032d4d6b451c94e796c84f85a706bec"} Oct 08 18:24:48 crc kubenswrapper[4988]: I1008 18:24:48.226422 4988 generic.go:334] "Generic (PLEG): container finished" podID="86be9f4e-06b6-4024-a098-820172b202db" containerID="4d58a21806551dc6c93955568a03ea09c6e5050149ed66ac74f81c0555b44706" exitCode=0 Oct 08 18:24:48 crc kubenswrapper[4988]: I1008 18:24:48.226507 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69rf72h" event={"ID":"86be9f4e-06b6-4024-a098-820172b202db","Type":"ContainerDied","Data":"4d58a21806551dc6c93955568a03ea09c6e5050149ed66ac74f81c0555b44706"} Oct 08 18:24:48 crc kubenswrapper[4988]: I1008 18:24:48.693782 4988 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="metallb-system/frr-k8s-xfwvv" Oct 08 18:24:48 crc kubenswrapper[4988]: I1008 18:24:48.734644 4988 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="metallb-system/frr-k8s-xfwvv" Oct 08 18:24:52 crc kubenswrapper[4988]: I1008 18:24:52.253818 4988 generic.go:334] "Generic (PLEG): container finished" podID="86be9f4e-06b6-4024-a098-820172b202db" containerID="c368283a2895ae97866bb7a6d6056f9ae50806e02ccdc151fa3242e70354bbba" exitCode=0 Oct 08 18:24:52 crc kubenswrapper[4988]: I1008 18:24:52.253880 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69rf72h" event={"ID":"86be9f4e-06b6-4024-a098-820172b202db","Type":"ContainerDied","Data":"c368283a2895ae97866bb7a6d6056f9ae50806e02ccdc151fa3242e70354bbba"} Oct 08 18:24:53 crc kubenswrapper[4988]: I1008 18:24:53.103840 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-webhook-server-64bf5d555-66f55" Oct 08 18:24:53 crc kubenswrapper[4988]: I1008 18:24:53.263038 4988 generic.go:334] "Generic (PLEG): container finished" podID="86be9f4e-06b6-4024-a098-820172b202db" containerID="87d3e581007766ff2e44267a783d2699c71b2e861764b72f549be425d0119bc5" exitCode=0 Oct 08 18:24:53 crc kubenswrapper[4988]: I1008 18:24:53.263098 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69rf72h" event={"ID":"86be9f4e-06b6-4024-a098-820172b202db","Type":"ContainerDied","Data":"87d3e581007766ff2e44267a783d2699c71b2e861764b72f549be425d0119bc5"} Oct 08 18:24:53 crc kubenswrapper[4988]: I1008 18:24:53.697348 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-xfwvv" Oct 08 18:24:54 crc kubenswrapper[4988]: I1008 18:24:54.530936 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69rf72h" Oct 08 18:24:54 crc kubenswrapper[4988]: I1008 18:24:54.560929 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/86be9f4e-06b6-4024-a098-820172b202db-util\") pod \"86be9f4e-06b6-4024-a098-820172b202db\" (UID: \"86be9f4e-06b6-4024-a098-820172b202db\") " Oct 08 18:24:54 crc kubenswrapper[4988]: I1008 18:24:54.560992 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/86be9f4e-06b6-4024-a098-820172b202db-bundle\") pod \"86be9f4e-06b6-4024-a098-820172b202db\" (UID: \"86be9f4e-06b6-4024-a098-820172b202db\") " Oct 08 18:24:54 crc kubenswrapper[4988]: I1008 18:24:54.561089 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-f2krj\" (UniqueName: \"kubernetes.io/projected/86be9f4e-06b6-4024-a098-820172b202db-kube-api-access-f2krj\") pod \"86be9f4e-06b6-4024-a098-820172b202db\" (UID: \"86be9f4e-06b6-4024-a098-820172b202db\") " Oct 08 18:24:54 crc kubenswrapper[4988]: I1008 18:24:54.562501 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/86be9f4e-06b6-4024-a098-820172b202db-bundle" (OuterVolumeSpecName: "bundle") pod "86be9f4e-06b6-4024-a098-820172b202db" (UID: "86be9f4e-06b6-4024-a098-820172b202db"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 18:24:54 crc kubenswrapper[4988]: I1008 18:24:54.567300 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/86be9f4e-06b6-4024-a098-820172b202db-kube-api-access-f2krj" (OuterVolumeSpecName: "kube-api-access-f2krj") pod "86be9f4e-06b6-4024-a098-820172b202db" (UID: "86be9f4e-06b6-4024-a098-820172b202db"). InnerVolumeSpecName "kube-api-access-f2krj". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:24:54 crc kubenswrapper[4988]: I1008 18:24:54.572859 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/86be9f4e-06b6-4024-a098-820172b202db-util" (OuterVolumeSpecName: "util") pod "86be9f4e-06b6-4024-a098-820172b202db" (UID: "86be9f4e-06b6-4024-a098-820172b202db"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 18:24:54 crc kubenswrapper[4988]: I1008 18:24:54.662901 4988 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/86be9f4e-06b6-4024-a098-820172b202db-util\") on node \"crc\" DevicePath \"\"" Oct 08 18:24:54 crc kubenswrapper[4988]: I1008 18:24:54.662935 4988 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/86be9f4e-06b6-4024-a098-820172b202db-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 18:24:54 crc kubenswrapper[4988]: I1008 18:24:54.662944 4988 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-f2krj\" (UniqueName: \"kubernetes.io/projected/86be9f4e-06b6-4024-a098-820172b202db-kube-api-access-f2krj\") on node \"crc\" DevicePath \"\"" Oct 08 18:24:55 crc kubenswrapper[4988]: I1008 18:24:55.281675 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69rf72h" event={"ID":"86be9f4e-06b6-4024-a098-820172b202db","Type":"ContainerDied","Data":"4223165cbc84b7f436c65efe062a6e02e032d4d6b451c94e796c84f85a706bec"} Oct 08 18:24:55 crc kubenswrapper[4988]: I1008 18:24:55.281970 4988 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4223165cbc84b7f436c65efe062a6e02e032d4d6b451c94e796c84f85a706bec" Oct 08 18:24:55 crc kubenswrapper[4988]: I1008 18:24:55.281712 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69rf72h" Oct 08 18:24:58 crc kubenswrapper[4988]: I1008 18:24:58.998944 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager-operator/cert-manager-operator-controller-manager-57cd46d6d-swdv5"] Oct 08 18:24:59 crc kubenswrapper[4988]: E1008 18:24:59.000917 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="86be9f4e-06b6-4024-a098-820172b202db" containerName="extract" Oct 08 18:24:59 crc kubenswrapper[4988]: I1008 18:24:59.001047 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="86be9f4e-06b6-4024-a098-820172b202db" containerName="extract" Oct 08 18:24:59 crc kubenswrapper[4988]: E1008 18:24:59.001147 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="86be9f4e-06b6-4024-a098-820172b202db" containerName="pull" Oct 08 18:24:59 crc kubenswrapper[4988]: I1008 18:24:59.001220 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="86be9f4e-06b6-4024-a098-820172b202db" containerName="pull" Oct 08 18:24:59 crc kubenswrapper[4988]: E1008 18:24:59.001308 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="86be9f4e-06b6-4024-a098-820172b202db" containerName="util" Oct 08 18:24:59 crc kubenswrapper[4988]: I1008 18:24:59.001382 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="86be9f4e-06b6-4024-a098-820172b202db" containerName="util" Oct 08 18:24:59 crc kubenswrapper[4988]: I1008 18:24:59.001651 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="86be9f4e-06b6-4024-a098-820172b202db" containerName="extract" Oct 08 18:24:59 crc kubenswrapper[4988]: I1008 18:24:59.002306 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager-operator/cert-manager-operator-controller-manager-57cd46d6d-swdv5" Oct 08 18:24:59 crc kubenswrapper[4988]: I1008 18:24:59.004013 4988 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager-operator"/"kube-root-ca.crt" Oct 08 18:24:59 crc kubenswrapper[4988]: I1008 18:24:59.004636 4988 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager-operator"/"cert-manager-operator-controller-manager-dockercfg-7v695" Oct 08 18:24:59 crc kubenswrapper[4988]: I1008 18:24:59.004650 4988 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager-operator"/"openshift-service-ca.crt" Oct 08 18:24:59 crc kubenswrapper[4988]: I1008 18:24:59.016845 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager-operator/cert-manager-operator-controller-manager-57cd46d6d-swdv5"] Oct 08 18:24:59 crc kubenswrapper[4988]: I1008 18:24:59.019176 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7xvxb\" (UniqueName: \"kubernetes.io/projected/2cb15138-60b1-41d3-8d46-45293094c160-kube-api-access-7xvxb\") pod \"cert-manager-operator-controller-manager-57cd46d6d-swdv5\" (UID: \"2cb15138-60b1-41d3-8d46-45293094c160\") " pod="cert-manager-operator/cert-manager-operator-controller-manager-57cd46d6d-swdv5" Oct 08 18:24:59 crc kubenswrapper[4988]: I1008 18:24:59.120850 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7xvxb\" (UniqueName: \"kubernetes.io/projected/2cb15138-60b1-41d3-8d46-45293094c160-kube-api-access-7xvxb\") pod \"cert-manager-operator-controller-manager-57cd46d6d-swdv5\" (UID: \"2cb15138-60b1-41d3-8d46-45293094c160\") " pod="cert-manager-operator/cert-manager-operator-controller-manager-57cd46d6d-swdv5" Oct 08 18:24:59 crc kubenswrapper[4988]: I1008 18:24:59.145945 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7xvxb\" (UniqueName: \"kubernetes.io/projected/2cb15138-60b1-41d3-8d46-45293094c160-kube-api-access-7xvxb\") pod \"cert-manager-operator-controller-manager-57cd46d6d-swdv5\" (UID: \"2cb15138-60b1-41d3-8d46-45293094c160\") " pod="cert-manager-operator/cert-manager-operator-controller-manager-57cd46d6d-swdv5" Oct 08 18:24:59 crc kubenswrapper[4988]: I1008 18:24:59.320499 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager-operator/cert-manager-operator-controller-manager-57cd46d6d-swdv5" Oct 08 18:24:59 crc kubenswrapper[4988]: I1008 18:24:59.859652 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager-operator/cert-manager-operator-controller-manager-57cd46d6d-swdv5"] Oct 08 18:24:59 crc kubenswrapper[4988]: W1008 18:24:59.870874 4988 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2cb15138_60b1_41d3_8d46_45293094c160.slice/crio-89ff0f0d7f8fb0491a37f535536ef09cabfe59e1f4efa7fb0151ef2bb77334b7 WatchSource:0}: Error finding container 89ff0f0d7f8fb0491a37f535536ef09cabfe59e1f4efa7fb0151ef2bb77334b7: Status 404 returned error can't find the container with id 89ff0f0d7f8fb0491a37f535536ef09cabfe59e1f4efa7fb0151ef2bb77334b7 Oct 08 18:25:00 crc kubenswrapper[4988]: I1008 18:25:00.309745 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager-operator/cert-manager-operator-controller-manager-57cd46d6d-swdv5" event={"ID":"2cb15138-60b1-41d3-8d46-45293094c160","Type":"ContainerStarted","Data":"89ff0f0d7f8fb0491a37f535536ef09cabfe59e1f4efa7fb0151ef2bb77334b7"} Oct 08 18:25:07 crc kubenswrapper[4988]: I1008 18:25:07.388187 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager-operator/cert-manager-operator-controller-manager-57cd46d6d-swdv5" event={"ID":"2cb15138-60b1-41d3-8d46-45293094c160","Type":"ContainerStarted","Data":"22ed19227008607356e3f34d0a7508960b8290cdbc9d70334e728feaaddd5723"} Oct 08 18:25:10 crc kubenswrapper[4988]: I1008 18:25:10.465930 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager-operator/cert-manager-operator-controller-manager-57cd46d6d-swdv5" podStartSLOduration=5.380487084 podStartE2EDuration="12.465915821s" podCreationTimestamp="2025-10-08 18:24:58 +0000 UTC" firstStartedPulling="2025-10-08 18:24:59.872678292 +0000 UTC m=+845.322521062" lastFinishedPulling="2025-10-08 18:25:06.958107029 +0000 UTC m=+852.407949799" observedRunningTime="2025-10-08 18:25:07.417373396 +0000 UTC m=+852.867216186" watchObservedRunningTime="2025-10-08 18:25:10.465915821 +0000 UTC m=+855.915758591" Oct 08 18:25:10 crc kubenswrapper[4988]: I1008 18:25:10.469250 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-webhook-d969966f-d8xbq"] Oct 08 18:25:10 crc kubenswrapper[4988]: I1008 18:25:10.470022 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-d969966f-d8xbq" Oct 08 18:25:10 crc kubenswrapper[4988]: I1008 18:25:10.472761 4988 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"openshift-service-ca.crt" Oct 08 18:25:10 crc kubenswrapper[4988]: I1008 18:25:10.472963 4988 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"kube-root-ca.crt" Oct 08 18:25:10 crc kubenswrapper[4988]: I1008 18:25:10.473216 4988 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-webhook-dockercfg-5ghnw" Oct 08 18:25:10 crc kubenswrapper[4988]: I1008 18:25:10.493337 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-d969966f-d8xbq"] Oct 08 18:25:10 crc kubenswrapper[4988]: I1008 18:25:10.636698 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/c718a13a-62b5-43c9-a151-bd3aab9e209d-bound-sa-token\") pod \"cert-manager-webhook-d969966f-d8xbq\" (UID: \"c718a13a-62b5-43c9-a151-bd3aab9e209d\") " pod="cert-manager/cert-manager-webhook-d969966f-d8xbq" Oct 08 18:25:10 crc kubenswrapper[4988]: I1008 18:25:10.637032 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nxdx6\" (UniqueName: \"kubernetes.io/projected/c718a13a-62b5-43c9-a151-bd3aab9e209d-kube-api-access-nxdx6\") pod \"cert-manager-webhook-d969966f-d8xbq\" (UID: \"c718a13a-62b5-43c9-a151-bd3aab9e209d\") " pod="cert-manager/cert-manager-webhook-d969966f-d8xbq" Oct 08 18:25:10 crc kubenswrapper[4988]: I1008 18:25:10.738800 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/c718a13a-62b5-43c9-a151-bd3aab9e209d-bound-sa-token\") pod \"cert-manager-webhook-d969966f-d8xbq\" (UID: \"c718a13a-62b5-43c9-a151-bd3aab9e209d\") " pod="cert-manager/cert-manager-webhook-d969966f-d8xbq" Oct 08 18:25:10 crc kubenswrapper[4988]: I1008 18:25:10.739210 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nxdx6\" (UniqueName: \"kubernetes.io/projected/c718a13a-62b5-43c9-a151-bd3aab9e209d-kube-api-access-nxdx6\") pod \"cert-manager-webhook-d969966f-d8xbq\" (UID: \"c718a13a-62b5-43c9-a151-bd3aab9e209d\") " pod="cert-manager/cert-manager-webhook-d969966f-d8xbq" Oct 08 18:25:10 crc kubenswrapper[4988]: I1008 18:25:10.773116 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/c718a13a-62b5-43c9-a151-bd3aab9e209d-bound-sa-token\") pod \"cert-manager-webhook-d969966f-d8xbq\" (UID: \"c718a13a-62b5-43c9-a151-bd3aab9e209d\") " pod="cert-manager/cert-manager-webhook-d969966f-d8xbq" Oct 08 18:25:10 crc kubenswrapper[4988]: I1008 18:25:10.773436 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nxdx6\" (UniqueName: \"kubernetes.io/projected/c718a13a-62b5-43c9-a151-bd3aab9e209d-kube-api-access-nxdx6\") pod \"cert-manager-webhook-d969966f-d8xbq\" (UID: \"c718a13a-62b5-43c9-a151-bd3aab9e209d\") " pod="cert-manager/cert-manager-webhook-d969966f-d8xbq" Oct 08 18:25:10 crc kubenswrapper[4988]: I1008 18:25:10.790584 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-d969966f-d8xbq" Oct 08 18:25:11 crc kubenswrapper[4988]: I1008 18:25:11.258940 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-d969966f-d8xbq"] Oct 08 18:25:11 crc kubenswrapper[4988]: I1008 18:25:11.419128 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-d969966f-d8xbq" event={"ID":"c718a13a-62b5-43c9-a151-bd3aab9e209d","Type":"ContainerStarted","Data":"00b9aec19f910984eaeac8331252b9f201d122b2950f1a8eda39f98ff62d6c35"} Oct 08 18:25:12 crc kubenswrapper[4988]: I1008 18:25:12.294409 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-cainjector-7d9f95dbf-qrjr7"] Oct 08 18:25:12 crc kubenswrapper[4988]: I1008 18:25:12.296627 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-7d9f95dbf-qrjr7" Oct 08 18:25:12 crc kubenswrapper[4988]: I1008 18:25:12.299755 4988 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-cainjector-dockercfg-c5gb9" Oct 08 18:25:12 crc kubenswrapper[4988]: I1008 18:25:12.317185 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-7d9f95dbf-qrjr7"] Oct 08 18:25:12 crc kubenswrapper[4988]: I1008 18:25:12.460673 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-grnhj\" (UniqueName: \"kubernetes.io/projected/75c13b54-1bce-4854-ae04-96c03b7b0333-kube-api-access-grnhj\") pod \"cert-manager-cainjector-7d9f95dbf-qrjr7\" (UID: \"75c13b54-1bce-4854-ae04-96c03b7b0333\") " pod="cert-manager/cert-manager-cainjector-7d9f95dbf-qrjr7" Oct 08 18:25:12 crc kubenswrapper[4988]: I1008 18:25:12.461058 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/75c13b54-1bce-4854-ae04-96c03b7b0333-bound-sa-token\") pod \"cert-manager-cainjector-7d9f95dbf-qrjr7\" (UID: \"75c13b54-1bce-4854-ae04-96c03b7b0333\") " pod="cert-manager/cert-manager-cainjector-7d9f95dbf-qrjr7" Oct 08 18:25:12 crc kubenswrapper[4988]: I1008 18:25:12.562636 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-grnhj\" (UniqueName: \"kubernetes.io/projected/75c13b54-1bce-4854-ae04-96c03b7b0333-kube-api-access-grnhj\") pod \"cert-manager-cainjector-7d9f95dbf-qrjr7\" (UID: \"75c13b54-1bce-4854-ae04-96c03b7b0333\") " pod="cert-manager/cert-manager-cainjector-7d9f95dbf-qrjr7" Oct 08 18:25:12 crc kubenswrapper[4988]: I1008 18:25:12.562695 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/75c13b54-1bce-4854-ae04-96c03b7b0333-bound-sa-token\") pod \"cert-manager-cainjector-7d9f95dbf-qrjr7\" (UID: \"75c13b54-1bce-4854-ae04-96c03b7b0333\") " pod="cert-manager/cert-manager-cainjector-7d9f95dbf-qrjr7" Oct 08 18:25:12 crc kubenswrapper[4988]: I1008 18:25:12.596736 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-grnhj\" (UniqueName: \"kubernetes.io/projected/75c13b54-1bce-4854-ae04-96c03b7b0333-kube-api-access-grnhj\") pod \"cert-manager-cainjector-7d9f95dbf-qrjr7\" (UID: \"75c13b54-1bce-4854-ae04-96c03b7b0333\") " pod="cert-manager/cert-manager-cainjector-7d9f95dbf-qrjr7" Oct 08 18:25:12 crc kubenswrapper[4988]: I1008 18:25:12.599905 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/75c13b54-1bce-4854-ae04-96c03b7b0333-bound-sa-token\") pod \"cert-manager-cainjector-7d9f95dbf-qrjr7\" (UID: \"75c13b54-1bce-4854-ae04-96c03b7b0333\") " pod="cert-manager/cert-manager-cainjector-7d9f95dbf-qrjr7" Oct 08 18:25:12 crc kubenswrapper[4988]: I1008 18:25:12.653447 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-7d9f95dbf-qrjr7" Oct 08 18:25:13 crc kubenswrapper[4988]: I1008 18:25:13.090152 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-7d9f95dbf-qrjr7"] Oct 08 18:25:13 crc kubenswrapper[4988]: I1008 18:25:13.440470 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-7d9f95dbf-qrjr7" event={"ID":"75c13b54-1bce-4854-ae04-96c03b7b0333","Type":"ContainerStarted","Data":"4bf7c9f9e12a5a526898f214f03f310b9a7dc9a69d7f1981857cb8e0ac8c36af"} Oct 08 18:25:16 crc kubenswrapper[4988]: I1008 18:25:16.458278 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-7d9f95dbf-qrjr7" event={"ID":"75c13b54-1bce-4854-ae04-96c03b7b0333","Type":"ContainerStarted","Data":"bfa11448e3981af785c06a1efafc28801a4cf942cd9d72138866ab84c34138ed"} Oct 08 18:25:16 crc kubenswrapper[4988]: I1008 18:25:16.460214 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-d969966f-d8xbq" event={"ID":"c718a13a-62b5-43c9-a151-bd3aab9e209d","Type":"ContainerStarted","Data":"9a7af8778f29471a4c81a68d90443e5837c37d8e4da8c2caa8db9aeca5fd235d"} Oct 08 18:25:16 crc kubenswrapper[4988]: I1008 18:25:16.460365 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="cert-manager/cert-manager-webhook-d969966f-d8xbq" Oct 08 18:25:16 crc kubenswrapper[4988]: I1008 18:25:16.479055 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-cainjector-7d9f95dbf-qrjr7" podStartSLOduration=2.07632911 podStartE2EDuration="4.479034866s" podCreationTimestamp="2025-10-08 18:25:12 +0000 UTC" firstStartedPulling="2025-10-08 18:25:13.099760572 +0000 UTC m=+858.549603342" lastFinishedPulling="2025-10-08 18:25:15.502466328 +0000 UTC m=+860.952309098" observedRunningTime="2025-10-08 18:25:16.474950595 +0000 UTC m=+861.924793365" watchObservedRunningTime="2025-10-08 18:25:16.479034866 +0000 UTC m=+861.928877636" Oct 08 18:25:16 crc kubenswrapper[4988]: I1008 18:25:16.503807 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-webhook-d969966f-d8xbq" podStartSLOduration=2.272787176 podStartE2EDuration="6.503787072s" podCreationTimestamp="2025-10-08 18:25:10 +0000 UTC" firstStartedPulling="2025-10-08 18:25:11.270958856 +0000 UTC m=+856.720801626" lastFinishedPulling="2025-10-08 18:25:15.501958752 +0000 UTC m=+860.951801522" observedRunningTime="2025-10-08 18:25:16.50002603 +0000 UTC m=+861.949868820" watchObservedRunningTime="2025-10-08 18:25:16.503787072 +0000 UTC m=+861.953629842" Oct 08 18:25:20 crc kubenswrapper[4988]: I1008 18:25:20.793364 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="cert-manager/cert-manager-webhook-d969966f-d8xbq" Oct 08 18:25:29 crc kubenswrapper[4988]: I1008 18:25:29.208981 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-7d4cc89fcb-4b68b"] Oct 08 18:25:29 crc kubenswrapper[4988]: I1008 18:25:29.211476 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-7d4cc89fcb-4b68b" Oct 08 18:25:29 crc kubenswrapper[4988]: I1008 18:25:29.214189 4988 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-dockercfg-695dw" Oct 08 18:25:29 crc kubenswrapper[4988]: I1008 18:25:29.231601 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-7d4cc89fcb-4b68b"] Oct 08 18:25:29 crc kubenswrapper[4988]: I1008 18:25:29.330550 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jtz9g\" (UniqueName: \"kubernetes.io/projected/d829b439-107e-4276-81dc-12052e1baf08-kube-api-access-jtz9g\") pod \"cert-manager-7d4cc89fcb-4b68b\" (UID: \"d829b439-107e-4276-81dc-12052e1baf08\") " pod="cert-manager/cert-manager-7d4cc89fcb-4b68b" Oct 08 18:25:29 crc kubenswrapper[4988]: I1008 18:25:29.330757 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/d829b439-107e-4276-81dc-12052e1baf08-bound-sa-token\") pod \"cert-manager-7d4cc89fcb-4b68b\" (UID: \"d829b439-107e-4276-81dc-12052e1baf08\") " pod="cert-manager/cert-manager-7d4cc89fcb-4b68b" Oct 08 18:25:29 crc kubenswrapper[4988]: I1008 18:25:29.431837 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jtz9g\" (UniqueName: \"kubernetes.io/projected/d829b439-107e-4276-81dc-12052e1baf08-kube-api-access-jtz9g\") pod \"cert-manager-7d4cc89fcb-4b68b\" (UID: \"d829b439-107e-4276-81dc-12052e1baf08\") " pod="cert-manager/cert-manager-7d4cc89fcb-4b68b" Oct 08 18:25:29 crc kubenswrapper[4988]: I1008 18:25:29.431919 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/d829b439-107e-4276-81dc-12052e1baf08-bound-sa-token\") pod \"cert-manager-7d4cc89fcb-4b68b\" (UID: \"d829b439-107e-4276-81dc-12052e1baf08\") " pod="cert-manager/cert-manager-7d4cc89fcb-4b68b" Oct 08 18:25:29 crc kubenswrapper[4988]: I1008 18:25:29.456208 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/d829b439-107e-4276-81dc-12052e1baf08-bound-sa-token\") pod \"cert-manager-7d4cc89fcb-4b68b\" (UID: \"d829b439-107e-4276-81dc-12052e1baf08\") " pod="cert-manager/cert-manager-7d4cc89fcb-4b68b" Oct 08 18:25:29 crc kubenswrapper[4988]: I1008 18:25:29.456206 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jtz9g\" (UniqueName: \"kubernetes.io/projected/d829b439-107e-4276-81dc-12052e1baf08-kube-api-access-jtz9g\") pod \"cert-manager-7d4cc89fcb-4b68b\" (UID: \"d829b439-107e-4276-81dc-12052e1baf08\") " pod="cert-manager/cert-manager-7d4cc89fcb-4b68b" Oct 08 18:25:29 crc kubenswrapper[4988]: I1008 18:25:29.532217 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-7d4cc89fcb-4b68b" Oct 08 18:25:29 crc kubenswrapper[4988]: I1008 18:25:29.974617 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-7d4cc89fcb-4b68b"] Oct 08 18:25:30 crc kubenswrapper[4988]: I1008 18:25:30.591980 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-7d4cc89fcb-4b68b" event={"ID":"d829b439-107e-4276-81dc-12052e1baf08","Type":"ContainerStarted","Data":"ffcddeb4e0030f589d9fbfc72967fd9b3a2f7a55f26b56d90f677cce3fa02e42"} Oct 08 18:25:30 crc kubenswrapper[4988]: I1008 18:25:30.592043 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-7d4cc89fcb-4b68b" event={"ID":"d829b439-107e-4276-81dc-12052e1baf08","Type":"ContainerStarted","Data":"f31734323c908b36c456880558cfbf38462265d1d86aab6058e66f28d585877c"} Oct 08 18:25:30 crc kubenswrapper[4988]: I1008 18:25:30.614456 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-7d4cc89fcb-4b68b" podStartSLOduration=1.614437997 podStartE2EDuration="1.614437997s" podCreationTimestamp="2025-10-08 18:25:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 18:25:30.614163609 +0000 UTC m=+876.064006419" watchObservedRunningTime="2025-10-08 18:25:30.614437997 +0000 UTC m=+876.064280787" Oct 08 18:25:35 crc kubenswrapper[4988]: I1008 18:25:35.088091 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-index-g8vgl"] Oct 08 18:25:35 crc kubenswrapper[4988]: I1008 18:25:35.089508 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-g8vgl" Oct 08 18:25:35 crc kubenswrapper[4988]: I1008 18:25:35.092156 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-index-dockercfg-bhkrz" Oct 08 18:25:35 crc kubenswrapper[4988]: I1008 18:25:35.092453 4988 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"openshift-service-ca.crt" Oct 08 18:25:35 crc kubenswrapper[4988]: I1008 18:25:35.092591 4988 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"kube-root-ca.crt" Oct 08 18:25:35 crc kubenswrapper[4988]: I1008 18:25:35.126079 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-g8vgl"] Oct 08 18:25:35 crc kubenswrapper[4988]: I1008 18:25:35.222861 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-54p6r\" (UniqueName: \"kubernetes.io/projected/834a8ce5-f140-4b45-80e4-2679b5730adf-kube-api-access-54p6r\") pod \"openstack-operator-index-g8vgl\" (UID: \"834a8ce5-f140-4b45-80e4-2679b5730adf\") " pod="openstack-operators/openstack-operator-index-g8vgl" Oct 08 18:25:35 crc kubenswrapper[4988]: I1008 18:25:35.324343 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-54p6r\" (UniqueName: \"kubernetes.io/projected/834a8ce5-f140-4b45-80e4-2679b5730adf-kube-api-access-54p6r\") pod \"openstack-operator-index-g8vgl\" (UID: \"834a8ce5-f140-4b45-80e4-2679b5730adf\") " pod="openstack-operators/openstack-operator-index-g8vgl" Oct 08 18:25:35 crc kubenswrapper[4988]: I1008 18:25:35.349429 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-54p6r\" (UniqueName: \"kubernetes.io/projected/834a8ce5-f140-4b45-80e4-2679b5730adf-kube-api-access-54p6r\") pod \"openstack-operator-index-g8vgl\" (UID: \"834a8ce5-f140-4b45-80e4-2679b5730adf\") " pod="openstack-operators/openstack-operator-index-g8vgl" Oct 08 18:25:35 crc kubenswrapper[4988]: I1008 18:25:35.413204 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-g8vgl" Oct 08 18:25:35 crc kubenswrapper[4988]: I1008 18:25:35.804161 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-g8vgl"] Oct 08 18:25:36 crc kubenswrapper[4988]: I1008 18:25:36.637122 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-g8vgl" event={"ID":"834a8ce5-f140-4b45-80e4-2679b5730adf","Type":"ContainerStarted","Data":"4aebee5c685941485ea5eb4564a30c2de5478d36245f54aa6c3b58bde5502265"} Oct 08 18:25:37 crc kubenswrapper[4988]: I1008 18:25:37.649186 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-g8vgl" event={"ID":"834a8ce5-f140-4b45-80e4-2679b5730adf","Type":"ContainerStarted","Data":"82acbb10cca6c557c6f2c6d16de3040f9a29cfb4271f34a86dd7c65ea085abd3"} Oct 08 18:25:37 crc kubenswrapper[4988]: I1008 18:25:37.668073 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-index-g8vgl" podStartSLOduration=1.719327549 podStartE2EDuration="2.668055771s" podCreationTimestamp="2025-10-08 18:25:35 +0000 UTC" firstStartedPulling="2025-10-08 18:25:35.815170581 +0000 UTC m=+881.265013351" lastFinishedPulling="2025-10-08 18:25:36.763898813 +0000 UTC m=+882.213741573" observedRunningTime="2025-10-08 18:25:37.665376682 +0000 UTC m=+883.115219462" watchObservedRunningTime="2025-10-08 18:25:37.668055771 +0000 UTC m=+883.117898541" Oct 08 18:25:38 crc kubenswrapper[4988]: I1008 18:25:38.259455 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/openstack-operator-index-g8vgl"] Oct 08 18:25:38 crc kubenswrapper[4988]: I1008 18:25:38.871312 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-index-5ccsc"] Oct 08 18:25:38 crc kubenswrapper[4988]: I1008 18:25:38.872328 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-5ccsc" Oct 08 18:25:38 crc kubenswrapper[4988]: I1008 18:25:38.883847 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-5ccsc"] Oct 08 18:25:38 crc kubenswrapper[4988]: I1008 18:25:38.972930 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wq8t5\" (UniqueName: \"kubernetes.io/projected/c1a9b179-3660-477f-8fdf-1aeb9cbb66df-kube-api-access-wq8t5\") pod \"openstack-operator-index-5ccsc\" (UID: \"c1a9b179-3660-477f-8fdf-1aeb9cbb66df\") " pod="openstack-operators/openstack-operator-index-5ccsc" Oct 08 18:25:39 crc kubenswrapper[4988]: I1008 18:25:39.074791 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wq8t5\" (UniqueName: \"kubernetes.io/projected/c1a9b179-3660-477f-8fdf-1aeb9cbb66df-kube-api-access-wq8t5\") pod \"openstack-operator-index-5ccsc\" (UID: \"c1a9b179-3660-477f-8fdf-1aeb9cbb66df\") " pod="openstack-operators/openstack-operator-index-5ccsc" Oct 08 18:25:39 crc kubenswrapper[4988]: I1008 18:25:39.097830 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wq8t5\" (UniqueName: \"kubernetes.io/projected/c1a9b179-3660-477f-8fdf-1aeb9cbb66df-kube-api-access-wq8t5\") pod \"openstack-operator-index-5ccsc\" (UID: \"c1a9b179-3660-477f-8fdf-1aeb9cbb66df\") " pod="openstack-operators/openstack-operator-index-5ccsc" Oct 08 18:25:39 crc kubenswrapper[4988]: I1008 18:25:39.225709 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-5ccsc" Oct 08 18:25:39 crc kubenswrapper[4988]: I1008 18:25:39.470007 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-5ccsc"] Oct 08 18:25:39 crc kubenswrapper[4988]: I1008 18:25:39.661425 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-5ccsc" event={"ID":"c1a9b179-3660-477f-8fdf-1aeb9cbb66df","Type":"ContainerStarted","Data":"c1921479874b61f5df68377dac07460aa929bf8a567205662878da74cca3fc22"} Oct 08 18:25:39 crc kubenswrapper[4988]: I1008 18:25:39.661554 4988 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack-operators/openstack-operator-index-g8vgl" podUID="834a8ce5-f140-4b45-80e4-2679b5730adf" containerName="registry-server" containerID="cri-o://82acbb10cca6c557c6f2c6d16de3040f9a29cfb4271f34a86dd7c65ea085abd3" gracePeriod=2 Oct 08 18:25:40 crc kubenswrapper[4988]: I1008 18:25:40.083191 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-g8vgl" Oct 08 18:25:40 crc kubenswrapper[4988]: I1008 18:25:40.189817 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-54p6r\" (UniqueName: \"kubernetes.io/projected/834a8ce5-f140-4b45-80e4-2679b5730adf-kube-api-access-54p6r\") pod \"834a8ce5-f140-4b45-80e4-2679b5730adf\" (UID: \"834a8ce5-f140-4b45-80e4-2679b5730adf\") " Oct 08 18:25:40 crc kubenswrapper[4988]: I1008 18:25:40.197056 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/834a8ce5-f140-4b45-80e4-2679b5730adf-kube-api-access-54p6r" (OuterVolumeSpecName: "kube-api-access-54p6r") pod "834a8ce5-f140-4b45-80e4-2679b5730adf" (UID: "834a8ce5-f140-4b45-80e4-2679b5730adf"). InnerVolumeSpecName "kube-api-access-54p6r". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:25:40 crc kubenswrapper[4988]: I1008 18:25:40.291620 4988 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-54p6r\" (UniqueName: \"kubernetes.io/projected/834a8ce5-f140-4b45-80e4-2679b5730adf-kube-api-access-54p6r\") on node \"crc\" DevicePath \"\"" Oct 08 18:25:40 crc kubenswrapper[4988]: I1008 18:25:40.670832 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-5ccsc" event={"ID":"c1a9b179-3660-477f-8fdf-1aeb9cbb66df","Type":"ContainerStarted","Data":"de45de6ba039ef1f98fbe3580c88b91a25465041fb9790c493c591a42fd3c352"} Oct 08 18:25:40 crc kubenswrapper[4988]: I1008 18:25:40.672823 4988 generic.go:334] "Generic (PLEG): container finished" podID="834a8ce5-f140-4b45-80e4-2679b5730adf" containerID="82acbb10cca6c557c6f2c6d16de3040f9a29cfb4271f34a86dd7c65ea085abd3" exitCode=0 Oct 08 18:25:40 crc kubenswrapper[4988]: I1008 18:25:40.672872 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-g8vgl" event={"ID":"834a8ce5-f140-4b45-80e4-2679b5730adf","Type":"ContainerDied","Data":"82acbb10cca6c557c6f2c6d16de3040f9a29cfb4271f34a86dd7c65ea085abd3"} Oct 08 18:25:40 crc kubenswrapper[4988]: I1008 18:25:40.672900 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-g8vgl" event={"ID":"834a8ce5-f140-4b45-80e4-2679b5730adf","Type":"ContainerDied","Data":"4aebee5c685941485ea5eb4564a30c2de5478d36245f54aa6c3b58bde5502265"} Oct 08 18:25:40 crc kubenswrapper[4988]: I1008 18:25:40.672918 4988 scope.go:117] "RemoveContainer" containerID="82acbb10cca6c557c6f2c6d16de3040f9a29cfb4271f34a86dd7c65ea085abd3" Oct 08 18:25:40 crc kubenswrapper[4988]: I1008 18:25:40.673233 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-g8vgl" Oct 08 18:25:40 crc kubenswrapper[4988]: I1008 18:25:40.693114 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-index-5ccsc" podStartSLOduration=2.076399716 podStartE2EDuration="2.693067837s" podCreationTimestamp="2025-10-08 18:25:38 +0000 UTC" firstStartedPulling="2025-10-08 18:25:39.493288621 +0000 UTC m=+884.943131401" lastFinishedPulling="2025-10-08 18:25:40.109956752 +0000 UTC m=+885.559799522" observedRunningTime="2025-10-08 18:25:40.692967574 +0000 UTC m=+886.142810394" watchObservedRunningTime="2025-10-08 18:25:40.693067837 +0000 UTC m=+886.142910617" Oct 08 18:25:40 crc kubenswrapper[4988]: I1008 18:25:40.698143 4988 scope.go:117] "RemoveContainer" containerID="82acbb10cca6c557c6f2c6d16de3040f9a29cfb4271f34a86dd7c65ea085abd3" Oct 08 18:25:40 crc kubenswrapper[4988]: E1008 18:25:40.698558 4988 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"82acbb10cca6c557c6f2c6d16de3040f9a29cfb4271f34a86dd7c65ea085abd3\": container with ID starting with 82acbb10cca6c557c6f2c6d16de3040f9a29cfb4271f34a86dd7c65ea085abd3 not found: ID does not exist" containerID="82acbb10cca6c557c6f2c6d16de3040f9a29cfb4271f34a86dd7c65ea085abd3" Oct 08 18:25:40 crc kubenswrapper[4988]: I1008 18:25:40.698602 4988 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"82acbb10cca6c557c6f2c6d16de3040f9a29cfb4271f34a86dd7c65ea085abd3"} err="failed to get container status \"82acbb10cca6c557c6f2c6d16de3040f9a29cfb4271f34a86dd7c65ea085abd3\": rpc error: code = NotFound desc = could not find container \"82acbb10cca6c557c6f2c6d16de3040f9a29cfb4271f34a86dd7c65ea085abd3\": container with ID starting with 82acbb10cca6c557c6f2c6d16de3040f9a29cfb4271f34a86dd7c65ea085abd3 not found: ID does not exist" Oct 08 18:25:40 crc kubenswrapper[4988]: I1008 18:25:40.720142 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/openstack-operator-index-g8vgl"] Oct 08 18:25:40 crc kubenswrapper[4988]: I1008 18:25:40.724850 4988 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/openstack-operator-index-g8vgl"] Oct 08 18:25:41 crc kubenswrapper[4988]: I1008 18:25:41.250745 4988 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="834a8ce5-f140-4b45-80e4-2679b5730adf" path="/var/lib/kubelet/pods/834a8ce5-f140-4b45-80e4-2679b5730adf/volumes" Oct 08 18:25:49 crc kubenswrapper[4988]: I1008 18:25:49.226717 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-index-5ccsc" Oct 08 18:25:49 crc kubenswrapper[4988]: I1008 18:25:49.227465 4988 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack-operators/openstack-operator-index-5ccsc" Oct 08 18:25:49 crc kubenswrapper[4988]: I1008 18:25:49.262521 4988 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack-operators/openstack-operator-index-5ccsc" Oct 08 18:25:49 crc kubenswrapper[4988]: I1008 18:25:49.773480 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-index-5ccsc" Oct 08 18:25:50 crc kubenswrapper[4988]: I1008 18:25:50.501949 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/1bfea55cf6f540853d0c0d919aed51b5fd6e2d105cc44a68cb64399f8bdkg77"] Oct 08 18:25:50 crc kubenswrapper[4988]: E1008 18:25:50.502186 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="834a8ce5-f140-4b45-80e4-2679b5730adf" containerName="registry-server" Oct 08 18:25:50 crc kubenswrapper[4988]: I1008 18:25:50.502201 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="834a8ce5-f140-4b45-80e4-2679b5730adf" containerName="registry-server" Oct 08 18:25:50 crc kubenswrapper[4988]: I1008 18:25:50.502315 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="834a8ce5-f140-4b45-80e4-2679b5730adf" containerName="registry-server" Oct 08 18:25:50 crc kubenswrapper[4988]: I1008 18:25:50.503165 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/1bfea55cf6f540853d0c0d919aed51b5fd6e2d105cc44a68cb64399f8bdkg77" Oct 08 18:25:50 crc kubenswrapper[4988]: I1008 18:25:50.505125 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"default-dockercfg-6hm8k" Oct 08 18:25:50 crc kubenswrapper[4988]: I1008 18:25:50.516523 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/1bfea55cf6f540853d0c0d919aed51b5fd6e2d105cc44a68cb64399f8bdkg77"] Oct 08 18:25:50 crc kubenswrapper[4988]: I1008 18:25:50.539135 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/75e43998-fcba-4311-8021-d102def09b3c-util\") pod \"1bfea55cf6f540853d0c0d919aed51b5fd6e2d105cc44a68cb64399f8bdkg77\" (UID: \"75e43998-fcba-4311-8021-d102def09b3c\") " pod="openstack-operators/1bfea55cf6f540853d0c0d919aed51b5fd6e2d105cc44a68cb64399f8bdkg77" Oct 08 18:25:50 crc kubenswrapper[4988]: I1008 18:25:50.539224 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rz4jg\" (UniqueName: \"kubernetes.io/projected/75e43998-fcba-4311-8021-d102def09b3c-kube-api-access-rz4jg\") pod \"1bfea55cf6f540853d0c0d919aed51b5fd6e2d105cc44a68cb64399f8bdkg77\" (UID: \"75e43998-fcba-4311-8021-d102def09b3c\") " pod="openstack-operators/1bfea55cf6f540853d0c0d919aed51b5fd6e2d105cc44a68cb64399f8bdkg77" Oct 08 18:25:50 crc kubenswrapper[4988]: I1008 18:25:50.539285 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/75e43998-fcba-4311-8021-d102def09b3c-bundle\") pod \"1bfea55cf6f540853d0c0d919aed51b5fd6e2d105cc44a68cb64399f8bdkg77\" (UID: \"75e43998-fcba-4311-8021-d102def09b3c\") " pod="openstack-operators/1bfea55cf6f540853d0c0d919aed51b5fd6e2d105cc44a68cb64399f8bdkg77" Oct 08 18:25:50 crc kubenswrapper[4988]: I1008 18:25:50.640157 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/75e43998-fcba-4311-8021-d102def09b3c-util\") pod \"1bfea55cf6f540853d0c0d919aed51b5fd6e2d105cc44a68cb64399f8bdkg77\" (UID: \"75e43998-fcba-4311-8021-d102def09b3c\") " pod="openstack-operators/1bfea55cf6f540853d0c0d919aed51b5fd6e2d105cc44a68cb64399f8bdkg77" Oct 08 18:25:50 crc kubenswrapper[4988]: I1008 18:25:50.640210 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rz4jg\" (UniqueName: \"kubernetes.io/projected/75e43998-fcba-4311-8021-d102def09b3c-kube-api-access-rz4jg\") pod \"1bfea55cf6f540853d0c0d919aed51b5fd6e2d105cc44a68cb64399f8bdkg77\" (UID: \"75e43998-fcba-4311-8021-d102def09b3c\") " pod="openstack-operators/1bfea55cf6f540853d0c0d919aed51b5fd6e2d105cc44a68cb64399f8bdkg77" Oct 08 18:25:50 crc kubenswrapper[4988]: I1008 18:25:50.640258 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/75e43998-fcba-4311-8021-d102def09b3c-bundle\") pod \"1bfea55cf6f540853d0c0d919aed51b5fd6e2d105cc44a68cb64399f8bdkg77\" (UID: \"75e43998-fcba-4311-8021-d102def09b3c\") " pod="openstack-operators/1bfea55cf6f540853d0c0d919aed51b5fd6e2d105cc44a68cb64399f8bdkg77" Oct 08 18:25:50 crc kubenswrapper[4988]: I1008 18:25:50.640818 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/75e43998-fcba-4311-8021-d102def09b3c-bundle\") pod \"1bfea55cf6f540853d0c0d919aed51b5fd6e2d105cc44a68cb64399f8bdkg77\" (UID: \"75e43998-fcba-4311-8021-d102def09b3c\") " pod="openstack-operators/1bfea55cf6f540853d0c0d919aed51b5fd6e2d105cc44a68cb64399f8bdkg77" Oct 08 18:25:50 crc kubenswrapper[4988]: I1008 18:25:50.641318 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/75e43998-fcba-4311-8021-d102def09b3c-util\") pod \"1bfea55cf6f540853d0c0d919aed51b5fd6e2d105cc44a68cb64399f8bdkg77\" (UID: \"75e43998-fcba-4311-8021-d102def09b3c\") " pod="openstack-operators/1bfea55cf6f540853d0c0d919aed51b5fd6e2d105cc44a68cb64399f8bdkg77" Oct 08 18:25:50 crc kubenswrapper[4988]: I1008 18:25:50.669285 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rz4jg\" (UniqueName: \"kubernetes.io/projected/75e43998-fcba-4311-8021-d102def09b3c-kube-api-access-rz4jg\") pod \"1bfea55cf6f540853d0c0d919aed51b5fd6e2d105cc44a68cb64399f8bdkg77\" (UID: \"75e43998-fcba-4311-8021-d102def09b3c\") " pod="openstack-operators/1bfea55cf6f540853d0c0d919aed51b5fd6e2d105cc44a68cb64399f8bdkg77" Oct 08 18:25:50 crc kubenswrapper[4988]: I1008 18:25:50.823010 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/1bfea55cf6f540853d0c0d919aed51b5fd6e2d105cc44a68cb64399f8bdkg77" Oct 08 18:25:51 crc kubenswrapper[4988]: I1008 18:25:51.249038 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/1bfea55cf6f540853d0c0d919aed51b5fd6e2d105cc44a68cb64399f8bdkg77"] Oct 08 18:25:51 crc kubenswrapper[4988]: W1008 18:25:51.264780 4988 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod75e43998_fcba_4311_8021_d102def09b3c.slice/crio-86a6fce16db1326a8dba4b1f31ea1664c7d42d95629605b2424d07bc4c868ae1 WatchSource:0}: Error finding container 86a6fce16db1326a8dba4b1f31ea1664c7d42d95629605b2424d07bc4c868ae1: Status 404 returned error can't find the container with id 86a6fce16db1326a8dba4b1f31ea1664c7d42d95629605b2424d07bc4c868ae1 Oct 08 18:25:51 crc kubenswrapper[4988]: I1008 18:25:51.748215 4988 generic.go:334] "Generic (PLEG): container finished" podID="75e43998-fcba-4311-8021-d102def09b3c" containerID="2cd1f70a3a75fa747a415aaaa3543bc6a978d824b344bd02addd356dc7d73523" exitCode=0 Oct 08 18:25:51 crc kubenswrapper[4988]: I1008 18:25:51.748262 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/1bfea55cf6f540853d0c0d919aed51b5fd6e2d105cc44a68cb64399f8bdkg77" event={"ID":"75e43998-fcba-4311-8021-d102def09b3c","Type":"ContainerDied","Data":"2cd1f70a3a75fa747a415aaaa3543bc6a978d824b344bd02addd356dc7d73523"} Oct 08 18:25:51 crc kubenswrapper[4988]: I1008 18:25:51.748292 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/1bfea55cf6f540853d0c0d919aed51b5fd6e2d105cc44a68cb64399f8bdkg77" event={"ID":"75e43998-fcba-4311-8021-d102def09b3c","Type":"ContainerStarted","Data":"86a6fce16db1326a8dba4b1f31ea1664c7d42d95629605b2424d07bc4c868ae1"} Oct 08 18:25:53 crc kubenswrapper[4988]: I1008 18:25:53.765206 4988 generic.go:334] "Generic (PLEG): container finished" podID="75e43998-fcba-4311-8021-d102def09b3c" containerID="20584727f7c35a182d70bc3275e270112320060fc6317c14ca72739a94eca9c2" exitCode=0 Oct 08 18:25:53 crc kubenswrapper[4988]: I1008 18:25:53.765291 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/1bfea55cf6f540853d0c0d919aed51b5fd6e2d105cc44a68cb64399f8bdkg77" event={"ID":"75e43998-fcba-4311-8021-d102def09b3c","Type":"ContainerDied","Data":"20584727f7c35a182d70bc3275e270112320060fc6317c14ca72739a94eca9c2"} Oct 08 18:25:54 crc kubenswrapper[4988]: I1008 18:25:54.780461 4988 generic.go:334] "Generic (PLEG): container finished" podID="75e43998-fcba-4311-8021-d102def09b3c" containerID="326c26871b82d629f7e0c849768a6d8788ee83cccff51c37dd86857b58afc755" exitCode=0 Oct 08 18:25:54 crc kubenswrapper[4988]: I1008 18:25:54.780591 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/1bfea55cf6f540853d0c0d919aed51b5fd6e2d105cc44a68cb64399f8bdkg77" event={"ID":"75e43998-fcba-4311-8021-d102def09b3c","Type":"ContainerDied","Data":"326c26871b82d629f7e0c849768a6d8788ee83cccff51c37dd86857b58afc755"} Oct 08 18:25:56 crc kubenswrapper[4988]: I1008 18:25:56.044426 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/1bfea55cf6f540853d0c0d919aed51b5fd6e2d105cc44a68cb64399f8bdkg77" Oct 08 18:25:56 crc kubenswrapper[4988]: I1008 18:25:56.115940 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rz4jg\" (UniqueName: \"kubernetes.io/projected/75e43998-fcba-4311-8021-d102def09b3c-kube-api-access-rz4jg\") pod \"75e43998-fcba-4311-8021-d102def09b3c\" (UID: \"75e43998-fcba-4311-8021-d102def09b3c\") " Oct 08 18:25:56 crc kubenswrapper[4988]: I1008 18:25:56.121444 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/75e43998-fcba-4311-8021-d102def09b3c-kube-api-access-rz4jg" (OuterVolumeSpecName: "kube-api-access-rz4jg") pod "75e43998-fcba-4311-8021-d102def09b3c" (UID: "75e43998-fcba-4311-8021-d102def09b3c"). InnerVolumeSpecName "kube-api-access-rz4jg". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:25:56 crc kubenswrapper[4988]: I1008 18:25:56.217539 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/75e43998-fcba-4311-8021-d102def09b3c-bundle\") pod \"75e43998-fcba-4311-8021-d102def09b3c\" (UID: \"75e43998-fcba-4311-8021-d102def09b3c\") " Oct 08 18:25:56 crc kubenswrapper[4988]: I1008 18:25:56.217880 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/75e43998-fcba-4311-8021-d102def09b3c-util\") pod \"75e43998-fcba-4311-8021-d102def09b3c\" (UID: \"75e43998-fcba-4311-8021-d102def09b3c\") " Oct 08 18:25:56 crc kubenswrapper[4988]: I1008 18:25:56.218719 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/75e43998-fcba-4311-8021-d102def09b3c-bundle" (OuterVolumeSpecName: "bundle") pod "75e43998-fcba-4311-8021-d102def09b3c" (UID: "75e43998-fcba-4311-8021-d102def09b3c"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 18:25:56 crc kubenswrapper[4988]: I1008 18:25:56.218904 4988 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rz4jg\" (UniqueName: \"kubernetes.io/projected/75e43998-fcba-4311-8021-d102def09b3c-kube-api-access-rz4jg\") on node \"crc\" DevicePath \"\"" Oct 08 18:25:56 crc kubenswrapper[4988]: I1008 18:25:56.246192 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/75e43998-fcba-4311-8021-d102def09b3c-util" (OuterVolumeSpecName: "util") pod "75e43998-fcba-4311-8021-d102def09b3c" (UID: "75e43998-fcba-4311-8021-d102def09b3c"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 18:25:56 crc kubenswrapper[4988]: I1008 18:25:56.320031 4988 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/75e43998-fcba-4311-8021-d102def09b3c-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 18:25:56 crc kubenswrapper[4988]: I1008 18:25:56.320089 4988 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/75e43998-fcba-4311-8021-d102def09b3c-util\") on node \"crc\" DevicePath \"\"" Oct 08 18:25:56 crc kubenswrapper[4988]: I1008 18:25:56.799147 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/1bfea55cf6f540853d0c0d919aed51b5fd6e2d105cc44a68cb64399f8bdkg77" event={"ID":"75e43998-fcba-4311-8021-d102def09b3c","Type":"ContainerDied","Data":"86a6fce16db1326a8dba4b1f31ea1664c7d42d95629605b2424d07bc4c868ae1"} Oct 08 18:25:56 crc kubenswrapper[4988]: I1008 18:25:56.799204 4988 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="86a6fce16db1326a8dba4b1f31ea1664c7d42d95629605b2424d07bc4c868ae1" Oct 08 18:25:56 crc kubenswrapper[4988]: I1008 18:25:56.799293 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/1bfea55cf6f540853d0c0d919aed51b5fd6e2d105cc44a68cb64399f8bdkg77" Oct 08 18:26:03 crc kubenswrapper[4988]: I1008 18:26:03.276584 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-controller-operator-bd6bc67fb-q9rwk"] Oct 08 18:26:03 crc kubenswrapper[4988]: E1008 18:26:03.277279 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="75e43998-fcba-4311-8021-d102def09b3c" containerName="util" Oct 08 18:26:03 crc kubenswrapper[4988]: I1008 18:26:03.277290 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="75e43998-fcba-4311-8021-d102def09b3c" containerName="util" Oct 08 18:26:03 crc kubenswrapper[4988]: E1008 18:26:03.277307 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="75e43998-fcba-4311-8021-d102def09b3c" containerName="pull" Oct 08 18:26:03 crc kubenswrapper[4988]: I1008 18:26:03.277313 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="75e43998-fcba-4311-8021-d102def09b3c" containerName="pull" Oct 08 18:26:03 crc kubenswrapper[4988]: E1008 18:26:03.277332 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="75e43998-fcba-4311-8021-d102def09b3c" containerName="extract" Oct 08 18:26:03 crc kubenswrapper[4988]: I1008 18:26:03.277338 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="75e43998-fcba-4311-8021-d102def09b3c" containerName="extract" Oct 08 18:26:03 crc kubenswrapper[4988]: I1008 18:26:03.277458 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="75e43998-fcba-4311-8021-d102def09b3c" containerName="extract" Oct 08 18:26:03 crc kubenswrapper[4988]: I1008 18:26:03.278057 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-operator-bd6bc67fb-q9rwk" Oct 08 18:26:03 crc kubenswrapper[4988]: I1008 18:26:03.279881 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-operator-dockercfg-qhkv6" Oct 08 18:26:03 crc kubenswrapper[4988]: I1008 18:26:03.309263 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-operator-bd6bc67fb-q9rwk"] Oct 08 18:26:03 crc kubenswrapper[4988]: I1008 18:26:03.412501 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gjtfz\" (UniqueName: \"kubernetes.io/projected/cbb26b26-1ffa-42b5-acce-5bca3a217c69-kube-api-access-gjtfz\") pod \"openstack-operator-controller-operator-bd6bc67fb-q9rwk\" (UID: \"cbb26b26-1ffa-42b5-acce-5bca3a217c69\") " pod="openstack-operators/openstack-operator-controller-operator-bd6bc67fb-q9rwk" Oct 08 18:26:03 crc kubenswrapper[4988]: I1008 18:26:03.513890 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gjtfz\" (UniqueName: \"kubernetes.io/projected/cbb26b26-1ffa-42b5-acce-5bca3a217c69-kube-api-access-gjtfz\") pod \"openstack-operator-controller-operator-bd6bc67fb-q9rwk\" (UID: \"cbb26b26-1ffa-42b5-acce-5bca3a217c69\") " pod="openstack-operators/openstack-operator-controller-operator-bd6bc67fb-q9rwk" Oct 08 18:26:03 crc kubenswrapper[4988]: I1008 18:26:03.532492 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gjtfz\" (UniqueName: \"kubernetes.io/projected/cbb26b26-1ffa-42b5-acce-5bca3a217c69-kube-api-access-gjtfz\") pod \"openstack-operator-controller-operator-bd6bc67fb-q9rwk\" (UID: \"cbb26b26-1ffa-42b5-acce-5bca3a217c69\") " pod="openstack-operators/openstack-operator-controller-operator-bd6bc67fb-q9rwk" Oct 08 18:26:03 crc kubenswrapper[4988]: I1008 18:26:03.596178 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-operator-bd6bc67fb-q9rwk" Oct 08 18:26:03 crc kubenswrapper[4988]: I1008 18:26:03.843138 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-operator-bd6bc67fb-q9rwk"] Oct 08 18:26:04 crc kubenswrapper[4988]: I1008 18:26:04.854628 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-bd6bc67fb-q9rwk" event={"ID":"cbb26b26-1ffa-42b5-acce-5bca3a217c69","Type":"ContainerStarted","Data":"f5c2ff613b86deecb7426f3b18312e32a54ed03210932c9007cd2d78b0be75ca"} Oct 08 18:26:08 crc kubenswrapper[4988]: I1008 18:26:08.877249 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-bd6bc67fb-q9rwk" event={"ID":"cbb26b26-1ffa-42b5-acce-5bca3a217c69","Type":"ContainerStarted","Data":"853b82d34078db8be8359488e4f59f6122d52d2630044237ab8a0aadd85f6470"} Oct 08 18:26:10 crc kubenswrapper[4988]: I1008 18:26:10.912270 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-bd6bc67fb-q9rwk" event={"ID":"cbb26b26-1ffa-42b5-acce-5bca3a217c69","Type":"ContainerStarted","Data":"c65a6e5ada5eff9621262d3fbd663e6634554005a41a025615eef5b2bb3091cd"} Oct 08 18:26:10 crc kubenswrapper[4988]: I1008 18:26:10.912637 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-operator-bd6bc67fb-q9rwk" Oct 08 18:26:10 crc kubenswrapper[4988]: I1008 18:26:10.955709 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-controller-operator-bd6bc67fb-q9rwk" podStartSLOduration=1.561282264 podStartE2EDuration="7.955689674s" podCreationTimestamp="2025-10-08 18:26:03 +0000 UTC" firstStartedPulling="2025-10-08 18:26:03.855494228 +0000 UTC m=+909.305337008" lastFinishedPulling="2025-10-08 18:26:10.249901648 +0000 UTC m=+915.699744418" observedRunningTime="2025-10-08 18:26:10.945227698 +0000 UTC m=+916.395070488" watchObservedRunningTime="2025-10-08 18:26:10.955689674 +0000 UTC m=+916.405532444" Oct 08 18:26:13 crc kubenswrapper[4988]: I1008 18:26:13.599769 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-operator-bd6bc67fb-q9rwk" Oct 08 18:26:23 crc kubenswrapper[4988]: I1008 18:26:23.337859 4988 patch_prober.go:28] interesting pod/machine-config-daemon-nm54f container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 18:26:23 crc kubenswrapper[4988]: I1008 18:26:23.338482 4988 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 18:26:30 crc kubenswrapper[4988]: I1008 18:26:30.365117 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/barbican-operator-controller-manager-658bdf4b74-cmnkr"] Oct 08 18:26:30 crc kubenswrapper[4988]: I1008 18:26:30.366718 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-controller-manager-658bdf4b74-cmnkr" Oct 08 18:26:30 crc kubenswrapper[4988]: I1008 18:26:30.370046 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"barbican-operator-controller-manager-dockercfg-xvttt" Oct 08 18:26:30 crc kubenswrapper[4988]: I1008 18:26:30.374247 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/cinder-operator-controller-manager-7b7fb68549-7xcz4"] Oct 08 18:26:30 crc kubenswrapper[4988]: I1008 18:26:30.375670 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-7b7fb68549-7xcz4" Oct 08 18:26:30 crc kubenswrapper[4988]: I1008 18:26:30.379704 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"cinder-operator-controller-manager-dockercfg-dzqn2" Oct 08 18:26:30 crc kubenswrapper[4988]: I1008 18:26:30.384791 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/barbican-operator-controller-manager-658bdf4b74-cmnkr"] Oct 08 18:26:30 crc kubenswrapper[4988]: I1008 18:26:30.400403 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-7b7fb68549-7xcz4"] Oct 08 18:26:30 crc kubenswrapper[4988]: I1008 18:26:30.432992 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/designate-operator-controller-manager-85d5d9dd78-94rjz"] Oct 08 18:26:30 crc kubenswrapper[4988]: I1008 18:26:30.433869 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/designate-operator-controller-manager-85d5d9dd78-94rjz" Oct 08 18:26:30 crc kubenswrapper[4988]: I1008 18:26:30.435632 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"designate-operator-controller-manager-dockercfg-lth5j" Oct 08 18:26:30 crc kubenswrapper[4988]: I1008 18:26:30.449743 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/designate-operator-controller-manager-85d5d9dd78-94rjz"] Oct 08 18:26:30 crc kubenswrapper[4988]: I1008 18:26:30.460192 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/glance-operator-controller-manager-84b9b84486-gpl58"] Oct 08 18:26:30 crc kubenswrapper[4988]: I1008 18:26:30.461430 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-84b9b84486-gpl58" Oct 08 18:26:30 crc kubenswrapper[4988]: I1008 18:26:30.463781 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"glance-operator-controller-manager-dockercfg-546s5" Oct 08 18:26:30 crc kubenswrapper[4988]: I1008 18:26:30.473748 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/heat-operator-controller-manager-858f76bbdd-rjzd5"] Oct 08 18:26:30 crc kubenswrapper[4988]: I1008 18:26:30.474917 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/heat-operator-controller-manager-858f76bbdd-rjzd5" Oct 08 18:26:30 crc kubenswrapper[4988]: I1008 18:26:30.482783 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"heat-operator-controller-manager-dockercfg-9lpmr" Oct 08 18:26:30 crc kubenswrapper[4988]: I1008 18:26:30.489283 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jj7h2\" (UniqueName: \"kubernetes.io/projected/c569c6ea-0929-4fc6-8d67-47578fa225c1-kube-api-access-jj7h2\") pod \"barbican-operator-controller-manager-658bdf4b74-cmnkr\" (UID: \"c569c6ea-0929-4fc6-8d67-47578fa225c1\") " pod="openstack-operators/barbican-operator-controller-manager-658bdf4b74-cmnkr" Oct 08 18:26:30 crc kubenswrapper[4988]: I1008 18:26:30.489372 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kqh2k\" (UniqueName: \"kubernetes.io/projected/420edeb5-79c9-4ece-9e95-d497d4be3b76-kube-api-access-kqh2k\") pod \"cinder-operator-controller-manager-7b7fb68549-7xcz4\" (UID: \"420edeb5-79c9-4ece-9e95-d497d4be3b76\") " pod="openstack-operators/cinder-operator-controller-manager-7b7fb68549-7xcz4" Oct 08 18:26:30 crc kubenswrapper[4988]: I1008 18:26:30.498486 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/horizon-operator-controller-manager-7ffbcb7588-jwr4r"] Oct 08 18:26:30 crc kubenswrapper[4988]: I1008 18:26:30.499916 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-7ffbcb7588-jwr4r" Oct 08 18:26:30 crc kubenswrapper[4988]: I1008 18:26:30.505742 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-84b9b84486-gpl58"] Oct 08 18:26:30 crc kubenswrapper[4988]: I1008 18:26:30.505929 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"horizon-operator-controller-manager-dockercfg-kng6j" Oct 08 18:26:30 crc kubenswrapper[4988]: I1008 18:26:30.512266 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/heat-operator-controller-manager-858f76bbdd-rjzd5"] Oct 08 18:26:30 crc kubenswrapper[4988]: I1008 18:26:30.542492 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-7ffbcb7588-jwr4r"] Oct 08 18:26:30 crc kubenswrapper[4988]: I1008 18:26:30.573897 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ironic-operator-controller-manager-9c5c78d49-fv7xm"] Oct 08 18:26:30 crc kubenswrapper[4988]: I1008 18:26:30.592327 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ironic-operator-controller-manager-9c5c78d49-fv7xm" Oct 08 18:26:30 crc kubenswrapper[4988]: I1008 18:26:30.596311 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"ironic-operator-controller-manager-dockercfg-gwkvw" Oct 08 18:26:30 crc kubenswrapper[4988]: I1008 18:26:30.600485 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kqh2k\" (UniqueName: \"kubernetes.io/projected/420edeb5-79c9-4ece-9e95-d497d4be3b76-kube-api-access-kqh2k\") pod \"cinder-operator-controller-manager-7b7fb68549-7xcz4\" (UID: \"420edeb5-79c9-4ece-9e95-d497d4be3b76\") " pod="openstack-operators/cinder-operator-controller-manager-7b7fb68549-7xcz4" Oct 08 18:26:30 crc kubenswrapper[4988]: I1008 18:26:30.600673 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jj7h2\" (UniqueName: \"kubernetes.io/projected/c569c6ea-0929-4fc6-8d67-47578fa225c1-kube-api-access-jj7h2\") pod \"barbican-operator-controller-manager-658bdf4b74-cmnkr\" (UID: \"c569c6ea-0929-4fc6-8d67-47578fa225c1\") " pod="openstack-operators/barbican-operator-controller-manager-658bdf4b74-cmnkr" Oct 08 18:26:30 crc kubenswrapper[4988]: I1008 18:26:30.600707 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j67qg\" (UniqueName: \"kubernetes.io/projected/2afedc0a-a797-4f9e-921f-5160fc437f79-kube-api-access-j67qg\") pod \"heat-operator-controller-manager-858f76bbdd-rjzd5\" (UID: \"2afedc0a-a797-4f9e-921f-5160fc437f79\") " pod="openstack-operators/heat-operator-controller-manager-858f76bbdd-rjzd5" Oct 08 18:26:30 crc kubenswrapper[4988]: I1008 18:26:30.601879 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lqlff\" (UniqueName: \"kubernetes.io/projected/81f789aa-012b-4ee6-a3c4-70bba5d64291-kube-api-access-lqlff\") pod \"designate-operator-controller-manager-85d5d9dd78-94rjz\" (UID: \"81f789aa-012b-4ee6-a3c4-70bba5d64291\") " pod="openstack-operators/designate-operator-controller-manager-85d5d9dd78-94rjz" Oct 08 18:26:30 crc kubenswrapper[4988]: I1008 18:26:30.603586 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6bj4h\" (UniqueName: \"kubernetes.io/projected/1af0d400-4968-4d6e-af43-0624f7bda6f0-kube-api-access-6bj4h\") pod \"glance-operator-controller-manager-84b9b84486-gpl58\" (UID: \"1af0d400-4968-4d6e-af43-0624f7bda6f0\") " pod="openstack-operators/glance-operator-controller-manager-84b9b84486-gpl58" Oct 08 18:26:30 crc kubenswrapper[4988]: I1008 18:26:30.628887 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/infra-operator-controller-manager-656bcbd775-2xfhc"] Oct 08 18:26:30 crc kubenswrapper[4988]: I1008 18:26:30.630175 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ironic-operator-controller-manager-9c5c78d49-fv7xm"] Oct 08 18:26:30 crc kubenswrapper[4988]: I1008 18:26:30.630292 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-656bcbd775-2xfhc" Oct 08 18:26:30 crc kubenswrapper[4988]: I1008 18:26:30.633309 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-controller-manager-dockercfg-gdtwb" Oct 08 18:26:30 crc kubenswrapper[4988]: I1008 18:26:30.647429 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kqh2k\" (UniqueName: \"kubernetes.io/projected/420edeb5-79c9-4ece-9e95-d497d4be3b76-kube-api-access-kqh2k\") pod \"cinder-operator-controller-manager-7b7fb68549-7xcz4\" (UID: \"420edeb5-79c9-4ece-9e95-d497d4be3b76\") " pod="openstack-operators/cinder-operator-controller-manager-7b7fb68549-7xcz4" Oct 08 18:26:30 crc kubenswrapper[4988]: I1008 18:26:30.650099 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jj7h2\" (UniqueName: \"kubernetes.io/projected/c569c6ea-0929-4fc6-8d67-47578fa225c1-kube-api-access-jj7h2\") pod \"barbican-operator-controller-manager-658bdf4b74-cmnkr\" (UID: \"c569c6ea-0929-4fc6-8d67-47578fa225c1\") " pod="openstack-operators/barbican-operator-controller-manager-658bdf4b74-cmnkr" Oct 08 18:26:30 crc kubenswrapper[4988]: I1008 18:26:30.653732 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-webhook-server-cert" Oct 08 18:26:30 crc kubenswrapper[4988]: I1008 18:26:30.681894 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-656bcbd775-2xfhc"] Oct 08 18:26:30 crc kubenswrapper[4988]: I1008 18:26:30.691241 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-controller-manager-658bdf4b74-cmnkr" Oct 08 18:26:30 crc kubenswrapper[4988]: I1008 18:26:30.706565 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/keystone-operator-controller-manager-55b6b7c7b8-4ts62"] Oct 08 18:26:30 crc kubenswrapper[4988]: I1008 18:26:30.707428 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-7b7fb68549-7xcz4" Oct 08 18:26:30 crc kubenswrapper[4988]: I1008 18:26:30.714209 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-55b6b7c7b8-4ts62"] Oct 08 18:26:30 crc kubenswrapper[4988]: I1008 18:26:30.714329 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-55b6b7c7b8-4ts62" Oct 08 18:26:30 crc kubenswrapper[4988]: I1008 18:26:30.720909 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"keystone-operator-controller-manager-dockercfg-9jj8f" Oct 08 18:26:30 crc kubenswrapper[4988]: I1008 18:26:30.721670 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/2ebb5c10-6685-4a61-8b57-bd39d36b3ce5-cert\") pod \"infra-operator-controller-manager-656bcbd775-2xfhc\" (UID: \"2ebb5c10-6685-4a61-8b57-bd39d36b3ce5\") " pod="openstack-operators/infra-operator-controller-manager-656bcbd775-2xfhc" Oct 08 18:26:30 crc kubenswrapper[4988]: I1008 18:26:30.721697 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w2lkk\" (UniqueName: \"kubernetes.io/projected/70495ce4-a8be-4853-be10-a5fe9e1020a7-kube-api-access-w2lkk\") pod \"ironic-operator-controller-manager-9c5c78d49-fv7xm\" (UID: \"70495ce4-a8be-4853-be10-a5fe9e1020a7\") " pod="openstack-operators/ironic-operator-controller-manager-9c5c78d49-fv7xm" Oct 08 18:26:30 crc kubenswrapper[4988]: I1008 18:26:30.721747 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j67qg\" (UniqueName: \"kubernetes.io/projected/2afedc0a-a797-4f9e-921f-5160fc437f79-kube-api-access-j67qg\") pod \"heat-operator-controller-manager-858f76bbdd-rjzd5\" (UID: \"2afedc0a-a797-4f9e-921f-5160fc437f79\") " pod="openstack-operators/heat-operator-controller-manager-858f76bbdd-rjzd5" Oct 08 18:26:30 crc kubenswrapper[4988]: I1008 18:26:30.721772 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lqlff\" (UniqueName: \"kubernetes.io/projected/81f789aa-012b-4ee6-a3c4-70bba5d64291-kube-api-access-lqlff\") pod \"designate-operator-controller-manager-85d5d9dd78-94rjz\" (UID: \"81f789aa-012b-4ee6-a3c4-70bba5d64291\") " pod="openstack-operators/designate-operator-controller-manager-85d5d9dd78-94rjz" Oct 08 18:26:30 crc kubenswrapper[4988]: I1008 18:26:30.721794 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gzbmj\" (UniqueName: \"kubernetes.io/projected/8600e1a1-b22a-4cdb-9e91-04e0a65b3756-kube-api-access-gzbmj\") pod \"horizon-operator-controller-manager-7ffbcb7588-jwr4r\" (UID: \"8600e1a1-b22a-4cdb-9e91-04e0a65b3756\") " pod="openstack-operators/horizon-operator-controller-manager-7ffbcb7588-jwr4r" Oct 08 18:26:30 crc kubenswrapper[4988]: I1008 18:26:30.721818 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6bj4h\" (UniqueName: \"kubernetes.io/projected/1af0d400-4968-4d6e-af43-0624f7bda6f0-kube-api-access-6bj4h\") pod \"glance-operator-controller-manager-84b9b84486-gpl58\" (UID: \"1af0d400-4968-4d6e-af43-0624f7bda6f0\") " pod="openstack-operators/glance-operator-controller-manager-84b9b84486-gpl58" Oct 08 18:26:30 crc kubenswrapper[4988]: I1008 18:26:30.721840 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4jdrl\" (UniqueName: \"kubernetes.io/projected/2ebb5c10-6685-4a61-8b57-bd39d36b3ce5-kube-api-access-4jdrl\") pod \"infra-operator-controller-manager-656bcbd775-2xfhc\" (UID: \"2ebb5c10-6685-4a61-8b57-bd39d36b3ce5\") " pod="openstack-operators/infra-operator-controller-manager-656bcbd775-2xfhc" Oct 08 18:26:30 crc kubenswrapper[4988]: I1008 18:26:30.740488 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-f9fb45f8f-mdxnm"] Oct 08 18:26:30 crc kubenswrapper[4988]: I1008 18:26:30.742032 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-f9fb45f8f-mdxnm" Oct 08 18:26:30 crc kubenswrapper[4988]: I1008 18:26:30.745156 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6bj4h\" (UniqueName: \"kubernetes.io/projected/1af0d400-4968-4d6e-af43-0624f7bda6f0-kube-api-access-6bj4h\") pod \"glance-operator-controller-manager-84b9b84486-gpl58\" (UID: \"1af0d400-4968-4d6e-af43-0624f7bda6f0\") " pod="openstack-operators/glance-operator-controller-manager-84b9b84486-gpl58" Oct 08 18:26:30 crc kubenswrapper[4988]: I1008 18:26:30.747871 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"mariadb-operator-controller-manager-dockercfg-cxrkz" Oct 08 18:26:30 crc kubenswrapper[4988]: I1008 18:26:30.750691 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lqlff\" (UniqueName: \"kubernetes.io/projected/81f789aa-012b-4ee6-a3c4-70bba5d64291-kube-api-access-lqlff\") pod \"designate-operator-controller-manager-85d5d9dd78-94rjz\" (UID: \"81f789aa-012b-4ee6-a3c4-70bba5d64291\") " pod="openstack-operators/designate-operator-controller-manager-85d5d9dd78-94rjz" Oct 08 18:26:30 crc kubenswrapper[4988]: I1008 18:26:30.750766 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/manila-operator-controller-manager-5f67fbc655-m7d7l"] Oct 08 18:26:30 crc kubenswrapper[4988]: I1008 18:26:30.756647 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-controller-manager-5f67fbc655-m7d7l" Oct 08 18:26:30 crc kubenswrapper[4988]: I1008 18:26:30.757331 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/designate-operator-controller-manager-85d5d9dd78-94rjz" Oct 08 18:26:30 crc kubenswrapper[4988]: I1008 18:26:30.759074 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"manila-operator-controller-manager-dockercfg-pk86d" Oct 08 18:26:30 crc kubenswrapper[4988]: I1008 18:26:30.765745 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-f9fb45f8f-mdxnm"] Oct 08 18:26:30 crc kubenswrapper[4988]: I1008 18:26:30.769600 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j67qg\" (UniqueName: \"kubernetes.io/projected/2afedc0a-a797-4f9e-921f-5160fc437f79-kube-api-access-j67qg\") pod \"heat-operator-controller-manager-858f76bbdd-rjzd5\" (UID: \"2afedc0a-a797-4f9e-921f-5160fc437f79\") " pod="openstack-operators/heat-operator-controller-manager-858f76bbdd-rjzd5" Oct 08 18:26:30 crc kubenswrapper[4988]: I1008 18:26:30.782530 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/manila-operator-controller-manager-5f67fbc655-m7d7l"] Oct 08 18:26:30 crc kubenswrapper[4988]: I1008 18:26:30.783786 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-84b9b84486-gpl58" Oct 08 18:26:30 crc kubenswrapper[4988]: I1008 18:26:30.787630 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/neutron-operator-controller-manager-79d585cb66-9cfhq"] Oct 08 18:26:30 crc kubenswrapper[4988]: I1008 18:26:30.790427 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/neutron-operator-controller-manager-79d585cb66-9cfhq" Oct 08 18:26:30 crc kubenswrapper[4988]: I1008 18:26:30.796012 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"neutron-operator-controller-manager-dockercfg-qm89w" Oct 08 18:26:30 crc kubenswrapper[4988]: I1008 18:26:30.809188 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/heat-operator-controller-manager-858f76bbdd-rjzd5" Oct 08 18:26:30 crc kubenswrapper[4988]: I1008 18:26:30.822738 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gzbmj\" (UniqueName: \"kubernetes.io/projected/8600e1a1-b22a-4cdb-9e91-04e0a65b3756-kube-api-access-gzbmj\") pod \"horizon-operator-controller-manager-7ffbcb7588-jwr4r\" (UID: \"8600e1a1-b22a-4cdb-9e91-04e0a65b3756\") " pod="openstack-operators/horizon-operator-controller-manager-7ffbcb7588-jwr4r" Oct 08 18:26:30 crc kubenswrapper[4988]: I1008 18:26:30.822785 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4jdrl\" (UniqueName: \"kubernetes.io/projected/2ebb5c10-6685-4a61-8b57-bd39d36b3ce5-kube-api-access-4jdrl\") pod \"infra-operator-controller-manager-656bcbd775-2xfhc\" (UID: \"2ebb5c10-6685-4a61-8b57-bd39d36b3ce5\") " pod="openstack-operators/infra-operator-controller-manager-656bcbd775-2xfhc" Oct 08 18:26:30 crc kubenswrapper[4988]: I1008 18:26:30.822819 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gbl9s\" (UniqueName: \"kubernetes.io/projected/70347d19-f528-4db3-81e7-6fbca659710e-kube-api-access-gbl9s\") pod \"keystone-operator-controller-manager-55b6b7c7b8-4ts62\" (UID: \"70347d19-f528-4db3-81e7-6fbca659710e\") " pod="openstack-operators/keystone-operator-controller-manager-55b6b7c7b8-4ts62" Oct 08 18:26:30 crc kubenswrapper[4988]: I1008 18:26:30.822843 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d62t7\" (UniqueName: \"kubernetes.io/projected/4c6dc76c-9b71-45aa-b1e9-0fd9deecbd39-kube-api-access-d62t7\") pod \"neutron-operator-controller-manager-79d585cb66-9cfhq\" (UID: \"4c6dc76c-9b71-45aa-b1e9-0fd9deecbd39\") " pod="openstack-operators/neutron-operator-controller-manager-79d585cb66-9cfhq" Oct 08 18:26:30 crc kubenswrapper[4988]: I1008 18:26:30.822878 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/2ebb5c10-6685-4a61-8b57-bd39d36b3ce5-cert\") pod \"infra-operator-controller-manager-656bcbd775-2xfhc\" (UID: \"2ebb5c10-6685-4a61-8b57-bd39d36b3ce5\") " pod="openstack-operators/infra-operator-controller-manager-656bcbd775-2xfhc" Oct 08 18:26:30 crc kubenswrapper[4988]: I1008 18:26:30.822896 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w2lkk\" (UniqueName: \"kubernetes.io/projected/70495ce4-a8be-4853-be10-a5fe9e1020a7-kube-api-access-w2lkk\") pod \"ironic-operator-controller-manager-9c5c78d49-fv7xm\" (UID: \"70495ce4-a8be-4853-be10-a5fe9e1020a7\") " pod="openstack-operators/ironic-operator-controller-manager-9c5c78d49-fv7xm" Oct 08 18:26:30 crc kubenswrapper[4988]: I1008 18:26:30.822919 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jdtdh\" (UniqueName: \"kubernetes.io/projected/cc3a04bc-795b-49e9-bc53-06b8534244ac-kube-api-access-jdtdh\") pod \"mariadb-operator-controller-manager-f9fb45f8f-mdxnm\" (UID: \"cc3a04bc-795b-49e9-bc53-06b8534244ac\") " pod="openstack-operators/mariadb-operator-controller-manager-f9fb45f8f-mdxnm" Oct 08 18:26:30 crc kubenswrapper[4988]: I1008 18:26:30.822955 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qb6z9\" (UniqueName: \"kubernetes.io/projected/128cb2db-23b2-423f-9ec1-a7c4ed246731-kube-api-access-qb6z9\") pod \"manila-operator-controller-manager-5f67fbc655-m7d7l\" (UID: \"128cb2db-23b2-423f-9ec1-a7c4ed246731\") " pod="openstack-operators/manila-operator-controller-manager-5f67fbc655-m7d7l" Oct 08 18:26:30 crc kubenswrapper[4988]: E1008 18:26:30.823427 4988 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Oct 08 18:26:30 crc kubenswrapper[4988]: E1008 18:26:30.823464 4988 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/2ebb5c10-6685-4a61-8b57-bd39d36b3ce5-cert podName:2ebb5c10-6685-4a61-8b57-bd39d36b3ce5 nodeName:}" failed. No retries permitted until 2025-10-08 18:26:31.323449837 +0000 UTC m=+936.773292597 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/2ebb5c10-6685-4a61-8b57-bd39d36b3ce5-cert") pod "infra-operator-controller-manager-656bcbd775-2xfhc" (UID: "2ebb5c10-6685-4a61-8b57-bd39d36b3ce5") : secret "infra-operator-webhook-server-cert" not found Oct 08 18:26:30 crc kubenswrapper[4988]: I1008 18:26:30.827537 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/neutron-operator-controller-manager-79d585cb66-9cfhq"] Oct 08 18:26:30 crc kubenswrapper[4988]: I1008 18:26:30.840713 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/nova-operator-controller-manager-5df598886f-8589q"] Oct 08 18:26:30 crc kubenswrapper[4988]: I1008 18:26:30.842102 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/nova-operator-controller-manager-5df598886f-8589q" Oct 08 18:26:30 crc kubenswrapper[4988]: I1008 18:26:30.845639 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"nova-operator-controller-manager-dockercfg-9zjm9" Oct 08 18:26:30 crc kubenswrapper[4988]: I1008 18:26:30.852965 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w2lkk\" (UniqueName: \"kubernetes.io/projected/70495ce4-a8be-4853-be10-a5fe9e1020a7-kube-api-access-w2lkk\") pod \"ironic-operator-controller-manager-9c5c78d49-fv7xm\" (UID: \"70495ce4-a8be-4853-be10-a5fe9e1020a7\") " pod="openstack-operators/ironic-operator-controller-manager-9c5c78d49-fv7xm" Oct 08 18:26:30 crc kubenswrapper[4988]: I1008 18:26:30.864327 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4jdrl\" (UniqueName: \"kubernetes.io/projected/2ebb5c10-6685-4a61-8b57-bd39d36b3ce5-kube-api-access-4jdrl\") pod \"infra-operator-controller-manager-656bcbd775-2xfhc\" (UID: \"2ebb5c10-6685-4a61-8b57-bd39d36b3ce5\") " pod="openstack-operators/infra-operator-controller-manager-656bcbd775-2xfhc" Oct 08 18:26:30 crc kubenswrapper[4988]: I1008 18:26:30.882110 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gzbmj\" (UniqueName: \"kubernetes.io/projected/8600e1a1-b22a-4cdb-9e91-04e0a65b3756-kube-api-access-gzbmj\") pod \"horizon-operator-controller-manager-7ffbcb7588-jwr4r\" (UID: \"8600e1a1-b22a-4cdb-9e91-04e0a65b3756\") " pod="openstack-operators/horizon-operator-controller-manager-7ffbcb7588-jwr4r" Oct 08 18:26:30 crc kubenswrapper[4988]: I1008 18:26:30.899014 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/nova-operator-controller-manager-5df598886f-8589q"] Oct 08 18:26:30 crc kubenswrapper[4988]: I1008 18:26:30.910456 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/octavia-operator-controller-manager-69fdcfc5f5-zd6wx"] Oct 08 18:26:30 crc kubenswrapper[4988]: I1008 18:26:30.912761 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/octavia-operator-controller-manager-69fdcfc5f5-zd6wx" Oct 08 18:26:30 crc kubenswrapper[4988]: I1008 18:26:30.915018 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"octavia-operator-controller-manager-dockercfg-2d8nr" Oct 08 18:26:30 crc kubenswrapper[4988]: I1008 18:26:30.924510 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jr44n\" (UniqueName: \"kubernetes.io/projected/db43233d-abb6-46e3-9b9f-946024ab477a-kube-api-access-jr44n\") pod \"octavia-operator-controller-manager-69fdcfc5f5-zd6wx\" (UID: \"db43233d-abb6-46e3-9b9f-946024ab477a\") " pod="openstack-operators/octavia-operator-controller-manager-69fdcfc5f5-zd6wx" Oct 08 18:26:30 crc kubenswrapper[4988]: I1008 18:26:30.924561 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qb6z9\" (UniqueName: \"kubernetes.io/projected/128cb2db-23b2-423f-9ec1-a7c4ed246731-kube-api-access-qb6z9\") pod \"manila-operator-controller-manager-5f67fbc655-m7d7l\" (UID: \"128cb2db-23b2-423f-9ec1-a7c4ed246731\") " pod="openstack-operators/manila-operator-controller-manager-5f67fbc655-m7d7l" Oct 08 18:26:30 crc kubenswrapper[4988]: I1008 18:26:30.924654 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-255cg\" (UniqueName: \"kubernetes.io/projected/f9ff7c98-5f49-4c45-a367-5ed4471083a8-kube-api-access-255cg\") pod \"nova-operator-controller-manager-5df598886f-8589q\" (UID: \"f9ff7c98-5f49-4c45-a367-5ed4471083a8\") " pod="openstack-operators/nova-operator-controller-manager-5df598886f-8589q" Oct 08 18:26:30 crc kubenswrapper[4988]: I1008 18:26:30.924873 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gbl9s\" (UniqueName: \"kubernetes.io/projected/70347d19-f528-4db3-81e7-6fbca659710e-kube-api-access-gbl9s\") pod \"keystone-operator-controller-manager-55b6b7c7b8-4ts62\" (UID: \"70347d19-f528-4db3-81e7-6fbca659710e\") " pod="openstack-operators/keystone-operator-controller-manager-55b6b7c7b8-4ts62" Oct 08 18:26:30 crc kubenswrapper[4988]: I1008 18:26:30.924921 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d62t7\" (UniqueName: \"kubernetes.io/projected/4c6dc76c-9b71-45aa-b1e9-0fd9deecbd39-kube-api-access-d62t7\") pod \"neutron-operator-controller-manager-79d585cb66-9cfhq\" (UID: \"4c6dc76c-9b71-45aa-b1e9-0fd9deecbd39\") " pod="openstack-operators/neutron-operator-controller-manager-79d585cb66-9cfhq" Oct 08 18:26:30 crc kubenswrapper[4988]: I1008 18:26:30.925085 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jdtdh\" (UniqueName: \"kubernetes.io/projected/cc3a04bc-795b-49e9-bc53-06b8534244ac-kube-api-access-jdtdh\") pod \"mariadb-operator-controller-manager-f9fb45f8f-mdxnm\" (UID: \"cc3a04bc-795b-49e9-bc53-06b8534244ac\") " pod="openstack-operators/mariadb-operator-controller-manager-f9fb45f8f-mdxnm" Oct 08 18:26:30 crc kubenswrapper[4988]: I1008 18:26:30.930133 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ironic-operator-controller-manager-9c5c78d49-fv7xm" Oct 08 18:26:30 crc kubenswrapper[4988]: I1008 18:26:30.932471 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/octavia-operator-controller-manager-69fdcfc5f5-zd6wx"] Oct 08 18:26:30 crc kubenswrapper[4988]: I1008 18:26:30.942453 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jdtdh\" (UniqueName: \"kubernetes.io/projected/cc3a04bc-795b-49e9-bc53-06b8534244ac-kube-api-access-jdtdh\") pod \"mariadb-operator-controller-manager-f9fb45f8f-mdxnm\" (UID: \"cc3a04bc-795b-49e9-bc53-06b8534244ac\") " pod="openstack-operators/mariadb-operator-controller-manager-f9fb45f8f-mdxnm" Oct 08 18:26:30 crc kubenswrapper[4988]: I1008 18:26:30.944895 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gbl9s\" (UniqueName: \"kubernetes.io/projected/70347d19-f528-4db3-81e7-6fbca659710e-kube-api-access-gbl9s\") pod \"keystone-operator-controller-manager-55b6b7c7b8-4ts62\" (UID: \"70347d19-f528-4db3-81e7-6fbca659710e\") " pod="openstack-operators/keystone-operator-controller-manager-55b6b7c7b8-4ts62" Oct 08 18:26:30 crc kubenswrapper[4988]: I1008 18:26:30.955156 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qb6z9\" (UniqueName: \"kubernetes.io/projected/128cb2db-23b2-423f-9ec1-a7c4ed246731-kube-api-access-qb6z9\") pod \"manila-operator-controller-manager-5f67fbc655-m7d7l\" (UID: \"128cb2db-23b2-423f-9ec1-a7c4ed246731\") " pod="openstack-operators/manila-operator-controller-manager-5f67fbc655-m7d7l" Oct 08 18:26:30 crc kubenswrapper[4988]: I1008 18:26:30.956322 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-677c5f5bff2dshw"] Oct 08 18:26:30 crc kubenswrapper[4988]: I1008 18:26:30.957762 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-baremetal-operator-controller-manager-677c5f5bff2dshw" Oct 08 18:26:30 crc kubenswrapper[4988]: I1008 18:26:30.965802 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-webhook-server-cert" Oct 08 18:26:30 crc kubenswrapper[4988]: I1008 18:26:30.966149 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ovn-operator-controller-manager-79db49b9fb-75xwr"] Oct 08 18:26:30 crc kubenswrapper[4988]: I1008 18:26:30.966221 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-controller-manager-dockercfg-fm84h" Oct 08 18:26:30 crc kubenswrapper[4988]: I1008 18:26:30.967369 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d62t7\" (UniqueName: \"kubernetes.io/projected/4c6dc76c-9b71-45aa-b1e9-0fd9deecbd39-kube-api-access-d62t7\") pod \"neutron-operator-controller-manager-79d585cb66-9cfhq\" (UID: \"4c6dc76c-9b71-45aa-b1e9-0fd9deecbd39\") " pod="openstack-operators/neutron-operator-controller-manager-79d585cb66-9cfhq" Oct 08 18:26:30 crc kubenswrapper[4988]: I1008 18:26:30.967921 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ovn-operator-controller-manager-79db49b9fb-75xwr" Oct 08 18:26:30 crc kubenswrapper[4988]: I1008 18:26:30.973488 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/placement-operator-controller-manager-68b6c87b68-rf945"] Oct 08 18:26:30 crc kubenswrapper[4988]: I1008 18:26:30.973959 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"ovn-operator-controller-manager-dockercfg-t788s" Oct 08 18:26:30 crc kubenswrapper[4988]: I1008 18:26:30.976060 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/placement-operator-controller-manager-68b6c87b68-rf945" Oct 08 18:26:30 crc kubenswrapper[4988]: I1008 18:26:30.977484 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/swift-operator-controller-manager-db6d7f97b-mxg6q"] Oct 08 18:26:30 crc kubenswrapper[4988]: I1008 18:26:30.978683 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-db6d7f97b-mxg6q" Oct 08 18:26:30 crc kubenswrapper[4988]: I1008 18:26:30.979609 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"placement-operator-controller-manager-dockercfg-66mvm" Oct 08 18:26:30 crc kubenswrapper[4988]: I1008 18:26:30.985001 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"swift-operator-controller-manager-dockercfg-gtfbv" Oct 08 18:26:30 crc kubenswrapper[4988]: I1008 18:26:30.996427 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/placement-operator-controller-manager-68b6c87b68-rf945"] Oct 08 18:26:31 crc kubenswrapper[4988]: I1008 18:26:31.008704 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-677c5f5bff2dshw"] Oct 08 18:26:31 crc kubenswrapper[4988]: I1008 18:26:31.028648 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/361b9b35-62d8-494a-9b8b-7d471adb246f-cert\") pod \"openstack-baremetal-operator-controller-manager-677c5f5bff2dshw\" (UID: \"361b9b35-62d8-494a-9b8b-7d471adb246f\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-677c5f5bff2dshw" Oct 08 18:26:31 crc kubenswrapper[4988]: I1008 18:26:31.028698 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tmhwm\" (UniqueName: \"kubernetes.io/projected/1154bd89-f7b0-4cce-b02c-9157b7866bf8-kube-api-access-tmhwm\") pod \"swift-operator-controller-manager-db6d7f97b-mxg6q\" (UID: \"1154bd89-f7b0-4cce-b02c-9157b7866bf8\") " pod="openstack-operators/swift-operator-controller-manager-db6d7f97b-mxg6q" Oct 08 18:26:31 crc kubenswrapper[4988]: I1008 18:26:31.028740 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jr44n\" (UniqueName: \"kubernetes.io/projected/db43233d-abb6-46e3-9b9f-946024ab477a-kube-api-access-jr44n\") pod \"octavia-operator-controller-manager-69fdcfc5f5-zd6wx\" (UID: \"db43233d-abb6-46e3-9b9f-946024ab477a\") " pod="openstack-operators/octavia-operator-controller-manager-69fdcfc5f5-zd6wx" Oct 08 18:26:31 crc kubenswrapper[4988]: I1008 18:26:31.028768 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8dfgv\" (UniqueName: \"kubernetes.io/projected/3433daff-1b78-4aa4-8eec-4bebfcbe1320-kube-api-access-8dfgv\") pod \"placement-operator-controller-manager-68b6c87b68-rf945\" (UID: \"3433daff-1b78-4aa4-8eec-4bebfcbe1320\") " pod="openstack-operators/placement-operator-controller-manager-68b6c87b68-rf945" Oct 08 18:26:31 crc kubenswrapper[4988]: I1008 18:26:31.028794 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-255cg\" (UniqueName: \"kubernetes.io/projected/f9ff7c98-5f49-4c45-a367-5ed4471083a8-kube-api-access-255cg\") pod \"nova-operator-controller-manager-5df598886f-8589q\" (UID: \"f9ff7c98-5f49-4c45-a367-5ed4471083a8\") " pod="openstack-operators/nova-operator-controller-manager-5df598886f-8589q" Oct 08 18:26:31 crc kubenswrapper[4988]: I1008 18:26:31.028823 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sldmw\" (UniqueName: \"kubernetes.io/projected/fc79827a-abba-4525-a705-129531293589-kube-api-access-sldmw\") pod \"ovn-operator-controller-manager-79db49b9fb-75xwr\" (UID: \"fc79827a-abba-4525-a705-129531293589\") " pod="openstack-operators/ovn-operator-controller-manager-79db49b9fb-75xwr" Oct 08 18:26:31 crc kubenswrapper[4988]: I1008 18:26:31.028845 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fnlmg\" (UniqueName: \"kubernetes.io/projected/361b9b35-62d8-494a-9b8b-7d471adb246f-kube-api-access-fnlmg\") pod \"openstack-baremetal-operator-controller-manager-677c5f5bff2dshw\" (UID: \"361b9b35-62d8-494a-9b8b-7d471adb246f\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-677c5f5bff2dshw" Oct 08 18:26:31 crc kubenswrapper[4988]: I1008 18:26:31.056523 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ovn-operator-controller-manager-79db49b9fb-75xwr"] Oct 08 18:26:31 crc kubenswrapper[4988]: I1008 18:26:31.068238 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-db6d7f97b-mxg6q"] Oct 08 18:26:31 crc kubenswrapper[4988]: I1008 18:26:31.104224 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-55b6b7c7b8-4ts62" Oct 08 18:26:31 crc kubenswrapper[4988]: I1008 18:26:31.115156 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-76796d4c6b-2szmt"] Oct 08 18:26:31 crc kubenswrapper[4988]: I1008 18:26:31.115848 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-255cg\" (UniqueName: \"kubernetes.io/projected/f9ff7c98-5f49-4c45-a367-5ed4471083a8-kube-api-access-255cg\") pod \"nova-operator-controller-manager-5df598886f-8589q\" (UID: \"f9ff7c98-5f49-4c45-a367-5ed4471083a8\") " pod="openstack-operators/nova-operator-controller-manager-5df598886f-8589q" Oct 08 18:26:31 crc kubenswrapper[4988]: I1008 18:26:31.116664 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jr44n\" (UniqueName: \"kubernetes.io/projected/db43233d-abb6-46e3-9b9f-946024ab477a-kube-api-access-jr44n\") pod \"octavia-operator-controller-manager-69fdcfc5f5-zd6wx\" (UID: \"db43233d-abb6-46e3-9b9f-946024ab477a\") " pod="openstack-operators/octavia-operator-controller-manager-69fdcfc5f5-zd6wx" Oct 08 18:26:31 crc kubenswrapper[4988]: I1008 18:26:31.117934 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-f9fb45f8f-mdxnm" Oct 08 18:26:31 crc kubenswrapper[4988]: I1008 18:26:31.123198 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/telemetry-operator-controller-manager-76796d4c6b-2szmt" Oct 08 18:26:31 crc kubenswrapper[4988]: I1008 18:26:31.132046 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-76796d4c6b-2szmt"] Oct 08 18:26:31 crc kubenswrapper[4988]: I1008 18:26:31.136364 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/361b9b35-62d8-494a-9b8b-7d471adb246f-cert\") pod \"openstack-baremetal-operator-controller-manager-677c5f5bff2dshw\" (UID: \"361b9b35-62d8-494a-9b8b-7d471adb246f\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-677c5f5bff2dshw" Oct 08 18:26:31 crc kubenswrapper[4988]: I1008 18:26:31.136443 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tmhwm\" (UniqueName: \"kubernetes.io/projected/1154bd89-f7b0-4cce-b02c-9157b7866bf8-kube-api-access-tmhwm\") pod \"swift-operator-controller-manager-db6d7f97b-mxg6q\" (UID: \"1154bd89-f7b0-4cce-b02c-9157b7866bf8\") " pod="openstack-operators/swift-operator-controller-manager-db6d7f97b-mxg6q" Oct 08 18:26:31 crc kubenswrapper[4988]: I1008 18:26:31.136486 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8dfgv\" (UniqueName: \"kubernetes.io/projected/3433daff-1b78-4aa4-8eec-4bebfcbe1320-kube-api-access-8dfgv\") pod \"placement-operator-controller-manager-68b6c87b68-rf945\" (UID: \"3433daff-1b78-4aa4-8eec-4bebfcbe1320\") " pod="openstack-operators/placement-operator-controller-manager-68b6c87b68-rf945" Oct 08 18:26:31 crc kubenswrapper[4988]: I1008 18:26:31.136518 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8wtq7\" (UniqueName: \"kubernetes.io/projected/c97bc0d9-b498-44ce-9fb5-f4cf7ca7561e-kube-api-access-8wtq7\") pod \"telemetry-operator-controller-manager-76796d4c6b-2szmt\" (UID: \"c97bc0d9-b498-44ce-9fb5-f4cf7ca7561e\") " pod="openstack-operators/telemetry-operator-controller-manager-76796d4c6b-2szmt" Oct 08 18:26:31 crc kubenswrapper[4988]: I1008 18:26:31.136563 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sldmw\" (UniqueName: \"kubernetes.io/projected/fc79827a-abba-4525-a705-129531293589-kube-api-access-sldmw\") pod \"ovn-operator-controller-manager-79db49b9fb-75xwr\" (UID: \"fc79827a-abba-4525-a705-129531293589\") " pod="openstack-operators/ovn-operator-controller-manager-79db49b9fb-75xwr" Oct 08 18:26:31 crc kubenswrapper[4988]: I1008 18:26:31.136588 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fnlmg\" (UniqueName: \"kubernetes.io/projected/361b9b35-62d8-494a-9b8b-7d471adb246f-kube-api-access-fnlmg\") pod \"openstack-baremetal-operator-controller-manager-677c5f5bff2dshw\" (UID: \"361b9b35-62d8-494a-9b8b-7d471adb246f\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-677c5f5bff2dshw" Oct 08 18:26:31 crc kubenswrapper[4988]: E1008 18:26:31.137008 4988 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Oct 08 18:26:31 crc kubenswrapper[4988]: E1008 18:26:31.137063 4988 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/361b9b35-62d8-494a-9b8b-7d471adb246f-cert podName:361b9b35-62d8-494a-9b8b-7d471adb246f nodeName:}" failed. No retries permitted until 2025-10-08 18:26:31.63704459 +0000 UTC m=+937.086887360 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/361b9b35-62d8-494a-9b8b-7d471adb246f-cert") pod "openstack-baremetal-operator-controller-manager-677c5f5bff2dshw" (UID: "361b9b35-62d8-494a-9b8b-7d471adb246f") : secret "openstack-baremetal-operator-webhook-server-cert" not found Oct 08 18:26:31 crc kubenswrapper[4988]: I1008 18:26:31.140428 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"telemetry-operator-controller-manager-dockercfg-rfnz8" Oct 08 18:26:31 crc kubenswrapper[4988]: I1008 18:26:31.145753 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-7ffbcb7588-jwr4r" Oct 08 18:26:31 crc kubenswrapper[4988]: I1008 18:26:31.156461 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-controller-manager-5f67fbc655-m7d7l" Oct 08 18:26:31 crc kubenswrapper[4988]: I1008 18:26:31.159947 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/test-operator-controller-manager-56c698c775-2p984"] Oct 08 18:26:31 crc kubenswrapper[4988]: I1008 18:26:31.160938 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-56c698c775-2p984" Oct 08 18:26:31 crc kubenswrapper[4988]: I1008 18:26:31.212139 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"test-operator-controller-manager-dockercfg-shcpb" Oct 08 18:26:31 crc kubenswrapper[4988]: I1008 18:26:31.213469 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/neutron-operator-controller-manager-79d585cb66-9cfhq" Oct 08 18:26:31 crc kubenswrapper[4988]: I1008 18:26:31.214301 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/nova-operator-controller-manager-5df598886f-8589q" Oct 08 18:26:31 crc kubenswrapper[4988]: I1008 18:26:31.228504 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-56c698c775-2p984"] Oct 08 18:26:31 crc kubenswrapper[4988]: I1008 18:26:31.266778 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8wtq7\" (UniqueName: \"kubernetes.io/projected/c97bc0d9-b498-44ce-9fb5-f4cf7ca7561e-kube-api-access-8wtq7\") pod \"telemetry-operator-controller-manager-76796d4c6b-2szmt\" (UID: \"c97bc0d9-b498-44ce-9fb5-f4cf7ca7561e\") " pod="openstack-operators/telemetry-operator-controller-manager-76796d4c6b-2szmt" Oct 08 18:26:31 crc kubenswrapper[4988]: I1008 18:26:31.266842 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tqhc7\" (UniqueName: \"kubernetes.io/projected/d3afe0af-aa6b-4321-99f3-39b2f64c2451-kube-api-access-tqhc7\") pod \"test-operator-controller-manager-56c698c775-2p984\" (UID: \"d3afe0af-aa6b-4321-99f3-39b2f64c2451\") " pod="openstack-operators/test-operator-controller-manager-56c698c775-2p984" Oct 08 18:26:31 crc kubenswrapper[4988]: I1008 18:26:31.267900 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/octavia-operator-controller-manager-69fdcfc5f5-zd6wx" Oct 08 18:26:31 crc kubenswrapper[4988]: I1008 18:26:31.268193 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8dfgv\" (UniqueName: \"kubernetes.io/projected/3433daff-1b78-4aa4-8eec-4bebfcbe1320-kube-api-access-8dfgv\") pod \"placement-operator-controller-manager-68b6c87b68-rf945\" (UID: \"3433daff-1b78-4aa4-8eec-4bebfcbe1320\") " pod="openstack-operators/placement-operator-controller-manager-68b6c87b68-rf945" Oct 08 18:26:31 crc kubenswrapper[4988]: I1008 18:26:31.260424 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tmhwm\" (UniqueName: \"kubernetes.io/projected/1154bd89-f7b0-4cce-b02c-9157b7866bf8-kube-api-access-tmhwm\") pod \"swift-operator-controller-manager-db6d7f97b-mxg6q\" (UID: \"1154bd89-f7b0-4cce-b02c-9157b7866bf8\") " pod="openstack-operators/swift-operator-controller-manager-db6d7f97b-mxg6q" Oct 08 18:26:31 crc kubenswrapper[4988]: I1008 18:26:31.278038 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sldmw\" (UniqueName: \"kubernetes.io/projected/fc79827a-abba-4525-a705-129531293589-kube-api-access-sldmw\") pod \"ovn-operator-controller-manager-79db49b9fb-75xwr\" (UID: \"fc79827a-abba-4525-a705-129531293589\") " pod="openstack-operators/ovn-operator-controller-manager-79db49b9fb-75xwr" Oct 08 18:26:31 crc kubenswrapper[4988]: I1008 18:26:31.279784 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fnlmg\" (UniqueName: \"kubernetes.io/projected/361b9b35-62d8-494a-9b8b-7d471adb246f-kube-api-access-fnlmg\") pod \"openstack-baremetal-operator-controller-manager-677c5f5bff2dshw\" (UID: \"361b9b35-62d8-494a-9b8b-7d471adb246f\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-677c5f5bff2dshw" Oct 08 18:26:31 crc kubenswrapper[4988]: I1008 18:26:31.337034 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8wtq7\" (UniqueName: \"kubernetes.io/projected/c97bc0d9-b498-44ce-9fb5-f4cf7ca7561e-kube-api-access-8wtq7\") pod \"telemetry-operator-controller-manager-76796d4c6b-2szmt\" (UID: \"c97bc0d9-b498-44ce-9fb5-f4cf7ca7561e\") " pod="openstack-operators/telemetry-operator-controller-manager-76796d4c6b-2szmt" Oct 08 18:26:31 crc kubenswrapper[4988]: I1008 18:26:31.370898 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tqhc7\" (UniqueName: \"kubernetes.io/projected/d3afe0af-aa6b-4321-99f3-39b2f64c2451-kube-api-access-tqhc7\") pod \"test-operator-controller-manager-56c698c775-2p984\" (UID: \"d3afe0af-aa6b-4321-99f3-39b2f64c2451\") " pod="openstack-operators/test-operator-controller-manager-56c698c775-2p984" Oct 08 18:26:31 crc kubenswrapper[4988]: I1008 18:26:31.371061 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/2ebb5c10-6685-4a61-8b57-bd39d36b3ce5-cert\") pod \"infra-operator-controller-manager-656bcbd775-2xfhc\" (UID: \"2ebb5c10-6685-4a61-8b57-bd39d36b3ce5\") " pod="openstack-operators/infra-operator-controller-manager-656bcbd775-2xfhc" Oct 08 18:26:31 crc kubenswrapper[4988]: I1008 18:26:31.377836 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/2ebb5c10-6685-4a61-8b57-bd39d36b3ce5-cert\") pod \"infra-operator-controller-manager-656bcbd775-2xfhc\" (UID: \"2ebb5c10-6685-4a61-8b57-bd39d36b3ce5\") " pod="openstack-operators/infra-operator-controller-manager-656bcbd775-2xfhc" Oct 08 18:26:31 crc kubenswrapper[4988]: I1008 18:26:31.416054 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tqhc7\" (UniqueName: \"kubernetes.io/projected/d3afe0af-aa6b-4321-99f3-39b2f64c2451-kube-api-access-tqhc7\") pod \"test-operator-controller-manager-56c698c775-2p984\" (UID: \"d3afe0af-aa6b-4321-99f3-39b2f64c2451\") " pod="openstack-operators/test-operator-controller-manager-56c698c775-2p984" Oct 08 18:26:31 crc kubenswrapper[4988]: I1008 18:26:31.451941 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/watcher-operator-controller-manager-7794bc6bd-wfq6r"] Oct 08 18:26:31 crc kubenswrapper[4988]: I1008 18:26:31.452976 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-7794bc6bd-wfq6r"] Oct 08 18:26:31 crc kubenswrapper[4988]: I1008 18:26:31.453034 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-controller-manager-598c4c5b5-mxr9w"] Oct 08 18:26:31 crc kubenswrapper[4988]: I1008 18:26:31.453881 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-manager-598c4c5b5-mxr9w"] Oct 08 18:26:31 crc kubenswrapper[4988]: I1008 18:26:31.453999 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-7794bc6bd-wfq6r" Oct 08 18:26:31 crc kubenswrapper[4988]: I1008 18:26:31.454023 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-manager-598c4c5b5-mxr9w" Oct 08 18:26:31 crc kubenswrapper[4988]: I1008 18:26:31.456139 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-25bz8"] Oct 08 18:26:31 crc kubenswrapper[4988]: I1008 18:26:31.457643 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-25bz8" Oct 08 18:26:31 crc kubenswrapper[4988]: I1008 18:26:31.465532 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-25bz8"] Oct 08 18:26:31 crc kubenswrapper[4988]: I1008 18:26:31.466058 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"rabbitmq-cluster-operator-controller-manager-dockercfg-qzqw5" Oct 08 18:26:31 crc kubenswrapper[4988]: I1008 18:26:31.466278 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"watcher-operator-controller-manager-dockercfg-ggz6s" Oct 08 18:26:31 crc kubenswrapper[4988]: I1008 18:26:31.466424 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"webhook-server-cert" Oct 08 18:26:31 crc kubenswrapper[4988]: I1008 18:26:31.466771 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-manager-dockercfg-cs5r8" Oct 08 18:26:31 crc kubenswrapper[4988]: I1008 18:26:31.482238 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ovn-operator-controller-manager-79db49b9fb-75xwr" Oct 08 18:26:31 crc kubenswrapper[4988]: I1008 18:26:31.491923 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/placement-operator-controller-manager-68b6c87b68-rf945" Oct 08 18:26:31 crc kubenswrapper[4988]: I1008 18:26:31.542755 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-db6d7f97b-mxg6q" Oct 08 18:26:31 crc kubenswrapper[4988]: I1008 18:26:31.576026 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sq6jx\" (UniqueName: \"kubernetes.io/projected/e4ef3040-d167-453d-aaa4-b807020a6876-kube-api-access-sq6jx\") pod \"rabbitmq-cluster-operator-manager-5f97d8c699-25bz8\" (UID: \"e4ef3040-d167-453d-aaa4-b807020a6876\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-25bz8" Oct 08 18:26:31 crc kubenswrapper[4988]: I1008 18:26:31.576111 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/3ed83b64-1d35-4ace-853f-2d246e27fc72-cert\") pod \"openstack-operator-controller-manager-598c4c5b5-mxr9w\" (UID: \"3ed83b64-1d35-4ace-853f-2d246e27fc72\") " pod="openstack-operators/openstack-operator-controller-manager-598c4c5b5-mxr9w" Oct 08 18:26:31 crc kubenswrapper[4988]: I1008 18:26:31.576179 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xtbd9\" (UniqueName: \"kubernetes.io/projected/3ed83b64-1d35-4ace-853f-2d246e27fc72-kube-api-access-xtbd9\") pod \"openstack-operator-controller-manager-598c4c5b5-mxr9w\" (UID: \"3ed83b64-1d35-4ace-853f-2d246e27fc72\") " pod="openstack-operators/openstack-operator-controller-manager-598c4c5b5-mxr9w" Oct 08 18:26:31 crc kubenswrapper[4988]: I1008 18:26:31.576203 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lxfvs\" (UniqueName: \"kubernetes.io/projected/47379f88-653e-4bda-a37b-a81857b7af36-kube-api-access-lxfvs\") pod \"watcher-operator-controller-manager-7794bc6bd-wfq6r\" (UID: \"47379f88-653e-4bda-a37b-a81857b7af36\") " pod="openstack-operators/watcher-operator-controller-manager-7794bc6bd-wfq6r" Oct 08 18:26:31 crc kubenswrapper[4988]: I1008 18:26:31.587811 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/telemetry-operator-controller-manager-76796d4c6b-2szmt" Oct 08 18:26:31 crc kubenswrapper[4988]: I1008 18:26:31.600194 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-656bcbd775-2xfhc" Oct 08 18:26:31 crc kubenswrapper[4988]: I1008 18:26:31.636169 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-56c698c775-2p984" Oct 08 18:26:31 crc kubenswrapper[4988]: I1008 18:26:31.678570 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xtbd9\" (UniqueName: \"kubernetes.io/projected/3ed83b64-1d35-4ace-853f-2d246e27fc72-kube-api-access-xtbd9\") pod \"openstack-operator-controller-manager-598c4c5b5-mxr9w\" (UID: \"3ed83b64-1d35-4ace-853f-2d246e27fc72\") " pod="openstack-operators/openstack-operator-controller-manager-598c4c5b5-mxr9w" Oct 08 18:26:31 crc kubenswrapper[4988]: I1008 18:26:31.678634 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lxfvs\" (UniqueName: \"kubernetes.io/projected/47379f88-653e-4bda-a37b-a81857b7af36-kube-api-access-lxfvs\") pod \"watcher-operator-controller-manager-7794bc6bd-wfq6r\" (UID: \"47379f88-653e-4bda-a37b-a81857b7af36\") " pod="openstack-operators/watcher-operator-controller-manager-7794bc6bd-wfq6r" Oct 08 18:26:31 crc kubenswrapper[4988]: I1008 18:26:31.678753 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sq6jx\" (UniqueName: \"kubernetes.io/projected/e4ef3040-d167-453d-aaa4-b807020a6876-kube-api-access-sq6jx\") pod \"rabbitmq-cluster-operator-manager-5f97d8c699-25bz8\" (UID: \"e4ef3040-d167-453d-aaa4-b807020a6876\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-25bz8" Oct 08 18:26:31 crc kubenswrapper[4988]: I1008 18:26:31.678817 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/3ed83b64-1d35-4ace-853f-2d246e27fc72-cert\") pod \"openstack-operator-controller-manager-598c4c5b5-mxr9w\" (UID: \"3ed83b64-1d35-4ace-853f-2d246e27fc72\") " pod="openstack-operators/openstack-operator-controller-manager-598c4c5b5-mxr9w" Oct 08 18:26:31 crc kubenswrapper[4988]: I1008 18:26:31.678857 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/361b9b35-62d8-494a-9b8b-7d471adb246f-cert\") pod \"openstack-baremetal-operator-controller-manager-677c5f5bff2dshw\" (UID: \"361b9b35-62d8-494a-9b8b-7d471adb246f\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-677c5f5bff2dshw" Oct 08 18:26:31 crc kubenswrapper[4988]: E1008 18:26:31.680074 4988 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Oct 08 18:26:31 crc kubenswrapper[4988]: E1008 18:26:31.680127 4988 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/3ed83b64-1d35-4ace-853f-2d246e27fc72-cert podName:3ed83b64-1d35-4ace-853f-2d246e27fc72 nodeName:}" failed. No retries permitted until 2025-10-08 18:26:32.180107705 +0000 UTC m=+937.629950475 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/3ed83b64-1d35-4ace-853f-2d246e27fc72-cert") pod "openstack-operator-controller-manager-598c4c5b5-mxr9w" (UID: "3ed83b64-1d35-4ace-853f-2d246e27fc72") : secret "webhook-server-cert" not found Oct 08 18:26:31 crc kubenswrapper[4988]: I1008 18:26:31.702223 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/361b9b35-62d8-494a-9b8b-7d471adb246f-cert\") pod \"openstack-baremetal-operator-controller-manager-677c5f5bff2dshw\" (UID: \"361b9b35-62d8-494a-9b8b-7d471adb246f\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-677c5f5bff2dshw" Oct 08 18:26:31 crc kubenswrapper[4988]: I1008 18:26:31.711983 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sq6jx\" (UniqueName: \"kubernetes.io/projected/e4ef3040-d167-453d-aaa4-b807020a6876-kube-api-access-sq6jx\") pod \"rabbitmq-cluster-operator-manager-5f97d8c699-25bz8\" (UID: \"e4ef3040-d167-453d-aaa4-b807020a6876\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-25bz8" Oct 08 18:26:31 crc kubenswrapper[4988]: I1008 18:26:31.721686 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lxfvs\" (UniqueName: \"kubernetes.io/projected/47379f88-653e-4bda-a37b-a81857b7af36-kube-api-access-lxfvs\") pod \"watcher-operator-controller-manager-7794bc6bd-wfq6r\" (UID: \"47379f88-653e-4bda-a37b-a81857b7af36\") " pod="openstack-operators/watcher-operator-controller-manager-7794bc6bd-wfq6r" Oct 08 18:26:31 crc kubenswrapper[4988]: I1008 18:26:31.737220 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xtbd9\" (UniqueName: \"kubernetes.io/projected/3ed83b64-1d35-4ace-853f-2d246e27fc72-kube-api-access-xtbd9\") pod \"openstack-operator-controller-manager-598c4c5b5-mxr9w\" (UID: \"3ed83b64-1d35-4ace-853f-2d246e27fc72\") " pod="openstack-operators/openstack-operator-controller-manager-598c4c5b5-mxr9w" Oct 08 18:26:31 crc kubenswrapper[4988]: I1008 18:26:31.887408 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-7794bc6bd-wfq6r" Oct 08 18:26:31 crc kubenswrapper[4988]: I1008 18:26:31.942308 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-baremetal-operator-controller-manager-677c5f5bff2dshw" Oct 08 18:26:31 crc kubenswrapper[4988]: I1008 18:26:31.963016 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-7b7fb68549-7xcz4"] Oct 08 18:26:32 crc kubenswrapper[4988]: I1008 18:26:32.014033 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-25bz8" Oct 08 18:26:32 crc kubenswrapper[4988]: I1008 18:26:32.104129 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-7b7fb68549-7xcz4" event={"ID":"420edeb5-79c9-4ece-9e95-d497d4be3b76","Type":"ContainerStarted","Data":"897f17d6afbe02fa75b602a72e64390c08d028a3a212dade406e0d3269961004"} Oct 08 18:26:32 crc kubenswrapper[4988]: I1008 18:26:32.108498 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/barbican-operator-controller-manager-658bdf4b74-cmnkr"] Oct 08 18:26:32 crc kubenswrapper[4988]: I1008 18:26:32.185337 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/3ed83b64-1d35-4ace-853f-2d246e27fc72-cert\") pod \"openstack-operator-controller-manager-598c4c5b5-mxr9w\" (UID: \"3ed83b64-1d35-4ace-853f-2d246e27fc72\") " pod="openstack-operators/openstack-operator-controller-manager-598c4c5b5-mxr9w" Oct 08 18:26:32 crc kubenswrapper[4988]: I1008 18:26:32.193953 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/3ed83b64-1d35-4ace-853f-2d246e27fc72-cert\") pod \"openstack-operator-controller-manager-598c4c5b5-mxr9w\" (UID: \"3ed83b64-1d35-4ace-853f-2d246e27fc72\") " pod="openstack-operators/openstack-operator-controller-manager-598c4c5b5-mxr9w" Oct 08 18:26:32 crc kubenswrapper[4988]: I1008 18:26:32.206937 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-84b9b84486-gpl58"] Oct 08 18:26:32 crc kubenswrapper[4988]: I1008 18:26:32.223983 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/heat-operator-controller-manager-858f76bbdd-rjzd5"] Oct 08 18:26:32 crc kubenswrapper[4988]: I1008 18:26:32.252616 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/designate-operator-controller-manager-85d5d9dd78-94rjz"] Oct 08 18:26:32 crc kubenswrapper[4988]: W1008 18:26:32.257845 4988 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod81f789aa_012b_4ee6_a3c4_70bba5d64291.slice/crio-e21650c53a51f961da0e00ab0e1689812aa34c34529641df30183848f1986e68 WatchSource:0}: Error finding container e21650c53a51f961da0e00ab0e1689812aa34c34529641df30183848f1986e68: Status 404 returned error can't find the container with id e21650c53a51f961da0e00ab0e1689812aa34c34529641df30183848f1986e68 Oct 08 18:26:32 crc kubenswrapper[4988]: I1008 18:26:32.344234 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-manager-598c4c5b5-mxr9w" Oct 08 18:26:32 crc kubenswrapper[4988]: I1008 18:26:32.391142 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/placement-operator-controller-manager-68b6c87b68-rf945"] Oct 08 18:26:32 crc kubenswrapper[4988]: I1008 18:26:32.401123 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/manila-operator-controller-manager-5f67fbc655-m7d7l"] Oct 08 18:26:32 crc kubenswrapper[4988]: I1008 18:26:32.406667 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-7ffbcb7588-jwr4r"] Oct 08 18:26:32 crc kubenswrapper[4988]: W1008 18:26:32.406859 4988 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8600e1a1_b22a_4cdb_9e91_04e0a65b3756.slice/crio-9c3616ae10e5a38d27af0b3e951ecaf60846162e977a4e149e1b81ee7b26f295 WatchSource:0}: Error finding container 9c3616ae10e5a38d27af0b3e951ecaf60846162e977a4e149e1b81ee7b26f295: Status 404 returned error can't find the container with id 9c3616ae10e5a38d27af0b3e951ecaf60846162e977a4e149e1b81ee7b26f295 Oct 08 18:26:32 crc kubenswrapper[4988]: I1008 18:26:32.413302 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/octavia-operator-controller-manager-69fdcfc5f5-zd6wx"] Oct 08 18:26:32 crc kubenswrapper[4988]: W1008 18:26:32.411699 4988 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod128cb2db_23b2_423f_9ec1_a7c4ed246731.slice/crio-84e2f1eab2f0a6dc1a9f284cc075b2ab96fadaa52264fa5f6db03f41139a7186 WatchSource:0}: Error finding container 84e2f1eab2f0a6dc1a9f284cc075b2ab96fadaa52264fa5f6db03f41139a7186: Status 404 returned error can't find the container with id 84e2f1eab2f0a6dc1a9f284cc075b2ab96fadaa52264fa5f6db03f41139a7186 Oct 08 18:26:32 crc kubenswrapper[4988]: W1008 18:26:32.421089 4988 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3433daff_1b78_4aa4_8eec_4bebfcbe1320.slice/crio-e76586e9a88d2829b2cd575a4a6dfddffbefa033e0d91c0a676f8e9c951372bf WatchSource:0}: Error finding container e76586e9a88d2829b2cd575a4a6dfddffbefa033e0d91c0a676f8e9c951372bf: Status 404 returned error can't find the container with id e76586e9a88d2829b2cd575a4a6dfddffbefa033e0d91c0a676f8e9c951372bf Oct 08 18:26:32 crc kubenswrapper[4988]: I1008 18:26:32.421262 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-f9fb45f8f-mdxnm"] Oct 08 18:26:32 crc kubenswrapper[4988]: W1008 18:26:32.421706 4988 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poddb43233d_abb6_46e3_9b9f_946024ab477a.slice/crio-a93eaca4e9a998eb0127e0950f1483753ea547e54571248e508b36d227e5ada7 WatchSource:0}: Error finding container a93eaca4e9a998eb0127e0950f1483753ea547e54571248e508b36d227e5ada7: Status 404 returned error can't find the container with id a93eaca4e9a998eb0127e0950f1483753ea547e54571248e508b36d227e5ada7 Oct 08 18:26:32 crc kubenswrapper[4988]: W1008 18:26:32.426103 4988 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podcc3a04bc_795b_49e9_bc53_06b8534244ac.slice/crio-e2fbe0aeb45d0a6717a02b78cca9d93766ceffa2a8fb8fe7cb444bb16029e304 WatchSource:0}: Error finding container e2fbe0aeb45d0a6717a02b78cca9d93766ceffa2a8fb8fe7cb444bb16029e304: Status 404 returned error can't find the container with id e2fbe0aeb45d0a6717a02b78cca9d93766ceffa2a8fb8fe7cb444bb16029e304 Oct 08 18:26:32 crc kubenswrapper[4988]: I1008 18:26:32.594020 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ironic-operator-controller-manager-9c5c78d49-fv7xm"] Oct 08 18:26:32 crc kubenswrapper[4988]: W1008 18:26:32.610580 4988 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod70495ce4_a8be_4853_be10_a5fe9e1020a7.slice/crio-5ef47f39d4fe67de4190c410a06ed8a3e54602d4316efe5906fb249e6a843068 WatchSource:0}: Error finding container 5ef47f39d4fe67de4190c410a06ed8a3e54602d4316efe5906fb249e6a843068: Status 404 returned error can't find the container with id 5ef47f39d4fe67de4190c410a06ed8a3e54602d4316efe5906fb249e6a843068 Oct 08 18:26:32 crc kubenswrapper[4988]: I1008 18:26:32.632908 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-db6d7f97b-mxg6q"] Oct 08 18:26:32 crc kubenswrapper[4988]: I1008 18:26:32.649018 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ovn-operator-controller-manager-79db49b9fb-75xwr"] Oct 08 18:26:32 crc kubenswrapper[4988]: E1008 18:26:32.699401 4988 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/test-operator@sha256:efa8fb78cffb573d299ffcc7bab1099affd2dbbab222152092b313074306e0a9,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-tqhc7,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod test-operator-controller-manager-56c698c775-2p984_openstack-operators(d3afe0af-aa6b-4321-99f3-39b2f64c2451): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Oct 08 18:26:32 crc kubenswrapper[4988]: E1008 18:26:32.699875 4988 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/ovn-operator@sha256:551b59e107c9812f7ad7aa06577376b0dcb58ff9498a41d5d5273e60e20ba7e4,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-sldmw,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ovn-operator-controller-manager-79db49b9fb-75xwr_openstack-operators(fc79827a-abba-4525-a705-129531293589): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Oct 08 18:26:32 crc kubenswrapper[4988]: I1008 18:26:32.700024 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-56c698c775-2p984"] Oct 08 18:26:32 crc kubenswrapper[4988]: I1008 18:26:32.719781 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/nova-operator-controller-manager-5df598886f-8589q"] Oct 08 18:26:32 crc kubenswrapper[4988]: I1008 18:26:32.869025 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-55b6b7c7b8-4ts62"] Oct 08 18:26:32 crc kubenswrapper[4988]: I1008 18:26:32.893045 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-25bz8"] Oct 08 18:26:32 crc kubenswrapper[4988]: I1008 18:26:32.895534 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-7794bc6bd-wfq6r"] Oct 08 18:26:32 crc kubenswrapper[4988]: I1008 18:26:32.898953 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-656bcbd775-2xfhc"] Oct 08 18:26:32 crc kubenswrapper[4988]: W1008 18:26:32.916627 4988 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod361b9b35_62d8_494a_9b8b_7d471adb246f.slice/crio-5430f4e6869c244c7cac8486aa213760d98862b9dd58dd8ba335dd1b5d91e8f0 WatchSource:0}: Error finding container 5430f4e6869c244c7cac8486aa213760d98862b9dd58dd8ba335dd1b5d91e8f0: Status 404 returned error can't find the container with id 5430f4e6869c244c7cac8486aa213760d98862b9dd58dd8ba335dd1b5d91e8f0 Oct 08 18:26:32 crc kubenswrapper[4988]: E1008 18:26:32.917145 4988 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:operator,Image:quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2,Command:[/manager],Args:[],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:metrics,HostPort:0,ContainerPort:9782,Protocol:TCP,HostIP:,},},Env:[]EnvVar{EnvVar{Name:OPERATOR_NAMESPACE,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{200 -3} {} 200m DecimalSI},memory: {{524288000 0} {} 500Mi BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-sq6jx,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod rabbitmq-cluster-operator-manager-5f97d8c699-25bz8_openstack-operators(e4ef3040-d167-453d-aaa4-b807020a6876): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Oct 08 18:26:32 crc kubenswrapper[4988]: E1008 18:26:32.918506 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-25bz8" podUID="e4ef3040-d167-453d-aaa4-b807020a6876" Oct 08 18:26:32 crc kubenswrapper[4988]: I1008 18:26:32.920801 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-76796d4c6b-2szmt"] Oct 08 18:26:32 crc kubenswrapper[4988]: I1008 18:26:32.927510 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-677c5f5bff2dshw"] Oct 08 18:26:32 crc kubenswrapper[4988]: W1008 18:26:32.927740 4988 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc97bc0d9_b498_44ce_9fb5_f4cf7ca7561e.slice/crio-a3b2234c16674808eec24b36c42d02a787574c571da000cde7c63b7351cc24a5 WatchSource:0}: Error finding container a3b2234c16674808eec24b36c42d02a787574c571da000cde7c63b7351cc24a5: Status 404 returned error can't find the container with id a3b2234c16674808eec24b36c42d02a787574c571da000cde7c63b7351cc24a5 Oct 08 18:26:32 crc kubenswrapper[4988]: E1008 18:26:32.928372 4988 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/openstack-baremetal-operator@sha256:a17fc270857869fd1efe5020b2a1cb8c2abbd838f08de88f3a6a59e8754ec351,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:true,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/openstack-k8s-operators/openstack-baremetal-operator-agent@sha256:03b4f3db4b373515f7e4095984b97197c05a14f87b2a0a525eb5d7be1d7bda66,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_ANSIBLEEE_IMAGE_URL_DEFAULT,Value:quay.io/openstack-k8s-operators/openstack-ansibleee-runner@sha256:6722a752fb7cbffbae811f6ad6567120fbd4ebbe8c38a83ec2df02850a3276bd,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_AODH_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-aodh-api@sha256:2115452234aedb505ed4efc6cd9b9a4ce3b9809aa7d0128d8fbeeee84dad1a69,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_AODH_EVALUATOR_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-aodh-evaluator@sha256:50597a8eaa6c4383f357574dcab8358b698729797b4156d932985a08ab86b7cd,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_AODH_LISTENER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-aodh-listener@sha256:cb4997d62c7b2534233a676cb92e19cf85dda07e2fb9fa642c28aab30489f69a,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_AODH_NOTIFIER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-aodh-notifier@sha256:1ccbf3f6cf24c9ee91bed71467491e22b8cb4b95bce90250f4174fae936b0fa1,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_APACHE_IMAGE_URL_DEFAULT,Value:registry.redhat.io/ubi9/httpd-24@sha256:e91d58021b54c46883595ff66be65882de54abdb3be2ca53c4162b20d18b5f48,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_BARBICAN_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-barbican-api@sha256:cbe345acb37e57986ecf6685d28c72d0e639bdb493a18e9d3ba947d6c3a16384,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_BARBICAN_KEYSTONE_LISTENER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-barbican-keystone-listener@sha256:e7dcc3bf23d5e0393ac173e3c43d4ae85f4613a4fd16b3c147dc32ae491d49bf,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_BARBICAN_WORKER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-barbican-worker@sha256:2a1a8b582c6e4cc31081bd8b0887acf45e31c1d14596c4e361d27d08fef0debf,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CEILOMETER_CENTRAL_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ceilometer-central@sha256:86daeb9c834bfcedb533086dff59a6b5b6e832b94ce2a9116337f8736bb80032,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CEILOMETER_COMPUTE_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ceilometer-compute@sha256:5d4fdf424fad33a3650163e9e7423f92e97de3305508c2b7c6435822e0313189,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CEILOMETER_IPMI_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ceilometer-ipmi@sha256:6d28de018f6e1672e775a75735e3bc16b63da41acd8fb5196ee0b06856c07133,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CEILOMETER_MYSQLD_EXPORTER_IMAGE_URL_DEFAULT,Value:quay.io/prometheus/mysqld-exporter@sha256:7211a617ec657701ca819aa0ba28e1d5750f5bf2c1391b755cc4a48cc360b0fa,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CEILOMETER_NOTIFICATION_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ceilometer-notification@sha256:c5fc9b72fc593bcf3b569c7ed24a256448eb1afab1504e668a3822e978be1306,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CEILOMETER_SGCORE_IMAGE_URL_DEFAULT,Value:quay.io/openstack-k8s-operators/sg-core@sha256:09b5017c95d7697e66b9c64846bc48ef5826a009cba89b956ec54561e5f4a2d1,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CINDER_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-cinder-api@sha256:85c75d60e1bd2f8a9ea0a2bb21a8df64c0a6f7b504cc1a05a355981d4b90e92f,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CINDER_BACKUP_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-cinder-backup@sha256:88b99249f15470f359fb554f7f3a56974b743f4655e3f0c982c0260f75a67697,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CINDER_SCHEDULER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-cinder-scheduler@sha256:e861d66785047d39eb68d9bac23e3f57ac84d9bd95593502d9b3b913b99fd1a4,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CINDER_VOLUME_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-cinder-volume@sha256:b95f09bf3d259f9eacf3b63931977483f5c3c332f49b95ee8a69d8e3fb71d082,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_DESIGNATE_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-designate-api@sha256:6fc7801c0d18d41b9f11484b1cdb342de9cebd93072ec2205dbe40945715184f,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_DESIGNATE_BACKENDBIND9_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-designate-backend-bind9@sha256:d4d824b80cbed683543d9e8c7045ac97e080774f45a5067ccbca26404e067821,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_DESIGNATE_CENTRAL_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-designate-central@sha256:182ec75938d8d3fb7d8f916373368add24062fec90489aa57776a81d0b36ea20,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_DESIGNATE_MDNS_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-designate-mdns@sha256:9507ba5ab74cbae902e2dc07f89c7b3b5b76d8079e444365fe0eee6000fd7aaa,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_DESIGNATE_PRODUCER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-designate-producer@sha256:17db080dcc4099f8a20aa0f238b6bca5c104672ae46743adeab9d1637725ecaa,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_DESIGNATE_UNBOUND_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-unbound@sha256:fd55cf3d73bfdc518419c9ba0b0cbef275140ae2d3bd0342a7310f81d57c2d78,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_DESIGNATE_WORKER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-designate-worker@sha256:d164a9bd383f50df69fc22e7422f4650cd5076c90ed19278fc0f04e54345a63d,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_FRR_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-frr@sha256:6beffe7d0bd75f9d1f495aeb7ab2334a2414af2c581d4833363df8441ed01018,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_ISCSID_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-iscsid@sha256:261e76f60c6bc6b172dc3608504552c63e83358a4fa3c0952a671544d83aa83f,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_KEPLER_IMAGE_URL_DEFAULT,Value:quay.io/sustainable_computing_io/kepler@sha256:581b65b646301e0fcb07582150ba63438f1353a85bf9acf1eb2acb4ce71c58bd,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_LOGROTATE_CROND_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-cron@sha256:2308c7b6c3d0aabbadfc9a06d84d67d2243f27fe8eed740ee96b1ce910203f62,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_MULTIPATHD_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-multipathd@sha256:02d33f59749441cd5751c319e9d7cff97ab1004844c0e992650d340c6e8fbf43,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_NEUTRON_DHCP_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent@sha256:9cf0ca292340f1f978603955ef682effbf24316d6e2376b1c89906d84c3f06d0,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_NEUTRON_METADATA_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn@sha256:c3e651f35b930bcf1a3084be8910c2f3f34d22a976c5379cf518a68d9994bfa7,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_NEUTRON_OVN_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-neutron-ovn-agent@sha256:58f678016d7f6c8fe579abe886fd138ef853642faa6766ca60639feac12d82ac,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_NEUTRON_SRIOV_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-neutron-sriov-agent@sha256:46f92909153aaf03a585374b77d103c536509747e3270558d9a533295c46a7c5,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_NODE_EXPORTER_IMAGE_URL_DEFAULT,Value:quay.io/prometheus/node-exporter@sha256:39c642b2b337e38c18e80266fb14383754178202f40103646337722a594d984c,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_OVN_BGP_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ovn-bgp-agent@sha256:7fe367f51638c5c302fd3f8e66a31b09cb3b11519a7f72ef142b6c6fe8b91694,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_PODMAN_EXPORTER_IMAGE_URL_DEFAULT,Value:quay.io/navidys/prometheus-podman-exporter@sha256:d339ba049bbd1adccb795962bf163f5b22fd84dea865d88b9eb525e46247d6bd,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_GLANCE_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-glance-api@sha256:9ebf424d4107275a2e3f21f7a18ef257ff2f97c1298109ac7c802a5a4f4794f2,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_HEAT_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-heat-api@sha256:4fcbe0d9a3c845708ecc32102ad4abbcbd947d87e5cf91f186de75b5d84ec681,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_HEAT_CFNAPI_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-heat-api-cfn@sha256:58a4e9a4dea86635c93ce37a2bb3c60ece62b3d656f6ee6a8845347cbb3e90fd,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_HEAT_ENGINE_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-heat-engine@sha256:6f2b843bc9f4ceb1ee873972d69e6bae6e1dbd378b486995bc3697d8bcff6339,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_HORIZON_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-horizon@sha256:03b4bb79b71d5ca7792d19c4c0ee08a5e5a407ad844c087305c42dd909ee7490,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_INFRA_MEMCACHED_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-memcached@sha256:773daada6402d9cad089cdc809d6c0335456d057ac1a25441ab5d82add2f70f4,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_INFRA_REDIS_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-redis@sha256:7323406a63fb3fdbb3eea4da0f7e8ed89c94c9bd0ad5ecd6c18fa4a4c2c550c4,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_IRONIC_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ironic-api@sha256:7ae82068011e2d2e5ddc88c943fd32ff4a11902793e7a1df729811b2e27122a0,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_IRONIC_CONDUCTOR_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ironic-conductor@sha256:0c762c15d9d98d39cc9dc3d1f9a70f9188fef58d4e2f3b0c69c896cab8da5e48,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_IRONIC_INSPECTOR_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ironic-inspector@sha256:febf65561eeef5b36b70d0d65ee83f6451e43ec97bfab4d826e14215da6ff19b,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_IRONIC_NEUTRON_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ironic-neutron-agent@sha256:b8aadfc3d547c5ef1e27fcb573d4760cf8c2f2271eefe1793c35a0d46b640837,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_IRONIC_PXE_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ironic-pxe@sha256:ecc91fd5079ee6d0c6ae1b11e97da790e33864d0e1930e574f959da2bddfa59a,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_IRONIC_PYTHON_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/ironic-python-agent@sha256:2e981e93f99c929a3f04e5e41c8f645d44d390a9aeee3c5193cce7ec2edcbf3a,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_KEYSTONE_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-keystone@sha256:1e5714637b6e1a24c2858fe6d9bbb3f00bc61d69ad74a657b1c23682bf4cb2b7,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_KSM_IMAGE_URL_DEFAULT,Value:registry.k8s.io/kube-state-metrics/kube-state-metrics@sha256:db384bf43222b066c378e77027a675d4cd9911107adba46c2922b3a55e10d6fb,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_MANILA_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-manila-api@sha256:35b8dcf27dc3b67f3840fa0e693ff312f74f7e22c634dff206a5c4d0133c716c,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_MANILA_SCHEDULER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-manila-scheduler@sha256:e109e4863e05e803dbfe04917756fd52231c560c65353170a2000be6cc2bb53d,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_MANILA_SHARE_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-manila-share@sha256:6df0bebd9318ce11624413249e7e9781311638f276f8877668d3b382fe90e62f,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_MARIADB_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-mariadb@sha256:56b75d97f4a48c8cf58b3a7c18c43618efb308bf0188124f6301142e61299b0c,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_NET_UTILS_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-netutils@sha256:a51ed62767206067aa501142dbf01f20b3d65325d30faf1b4d6424d5b17dfba5,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_NEUTRON_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-neutron-server@sha256:c4e71b2158fd939dad8b8e705273493051d3023273d23b279f2699dce6db33df,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_NOVA_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-nova-api@sha256:592e3cd32d3cc97a69093ad905b449aa374ffbb1b2644b738bb6c1434476d1f6,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_NOVA_COMPUTE_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-nova-compute@sha256:5f179b847f2dc32d9110b8f2be9fe65f1aeada1e18105dffdaf052981215d844,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_NOVA_CONDUCTOR_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-nova-conductor@sha256:9596452e283febbe08204d0ef0fd1992af3395d0969f7ac76663ed7c8be5b4d4,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_NOVA_NOVNC_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-nova-novncproxy@sha256:d61005a10bef1b37762a8a41e6755c1169241e36cc5f92886bca6f4f6b9c381a,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_NOVA_SCHEDULER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-nova-scheduler@sha256:e6a4335bcbeed3cd3e73ac879f754e314761e4a417a67539ca88e96a79346328,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OCTAVIA_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-octavia-api@sha256:97d88fc53421b699fc91983313d7beec4a0f177089e95bdf5ba15c3f521db9a9,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OCTAVIA_HEALTHMANAGER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-octavia-health-manager@sha256:5365e5c9c3ad2ede1b6945255b2cc6b009d642c39babdf25e0655282cfa646fe,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OCTAVIA_HOUSEKEEPING_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-octavia-housekeeping@sha256:5b55795d774e0ea160ff8a7fd491ed41cf2d93c7d821694abb3a879eaffcefeb,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OCTAVIA_RSYSLOG_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-rsyslog@sha256:26e955c46a6063eafcfeb79430bf3d9268dbe95687c00e63a624b3ec5a846f5a,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OCTAVIA_WORKER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-octavia-worker@sha256:58939baa18ab09e2b24996c5f3665ae52274b781f661ea06a67c991e9a832d5a,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OPENSTACK_CLIENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-openstackclient@sha256:b8bff6857fec93c3c1521f1a8c23de21bcb86fc0f960972e81f6c3f95d4185be,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OPENSTACK_MUST_GATHER_DEFAULT,Value:quay.io/openstack-k8s-operators/openstack-must-gather@sha256:943eee724277e252795909137538a553ef5284c8103ad01b9be7b0138c66d14d,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OPENSTACK_NETWORK_EXPORTER_IMAGE_URL_DEFAULT,Value:quay.io/openstack-k8s-operators/openstack-network-exporter@sha256:ecd56e6733c475f2d441344fd98f288c3eac0261ba113695fec7520a954ccbc7,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OS_CONTAINER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/edpm-hardened-uefi@sha256:d97b08fd421065c8c33a523973822ac468500cbe853069aa9214393fbda7a908,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OVN_CONTROLLER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ovn-controller@sha256:d76f7d6620930cc2e9ac070492bbeb525f83ce5ff4947463e3784bf1ce04a857,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OVN_CONTROLLER_OVS_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ovn-base@sha256:289dea3beea1cd4405895fc42e44372b35e4a941e31c59e102c333471a3ca9b7,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OVN_NB_DBCLUSTER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ovn-nb-db-server@sha256:9b19894fa67a81bf8ba4159b55b49f38877c670aeb97e2021c341cef2a9294e4,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OVN_NORTHD_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ovn-northd@sha256:ea164961ad30453ad0301c6b73364e1f1024f689634c88dd98265f9c7048e31d,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OVN_SB_DBCLUSTER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ovn-sb-db-server@sha256:6f9f2ea45f0271f6da8eb05a5f74cf5ce6769479346f5c2f407ee6f31a9c7ff3,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_PLACEMENT_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-placement-api@sha256:59448516174fc3bab679b9a8dd62cb9a9d16b5734aadbeb98e960e3b7c79bd22,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_RABBITMQ_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-rabbitmq@sha256:adcdeb8ecd601fb03c3b0901d5b5111af2ca48f7dd443e22224db6daaf08f5d0,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_SWIFT_ACCOUNT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-swift-account@sha256:2bf32d9b95899d7637dfe19d07cf1ecc9a06593984faff57a3c0dce060012edb,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_SWIFT_CONTAINER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-swift-container@sha256:7a452cd18b64d522e8a1e25bdcea543e9fe5f5b76e1c5e044c2b5334e06a326b,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_SWIFT_OBJECT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-swift-object@sha256:6a46aa13aa359b8e782a22d67db42db02bbf2bb7e35df4b684ac1daeda38cde3,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_SWIFT_PROXY_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-swift-proxy-server@sha256:f6824854bea6b2acbb00c34639799b4744818d4adbdd40e37dc5088f9ae18d58,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_TEST_TEMPEST_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-tempest-all@sha256:a66d2fdc21f25c690f02e643d2666dbe7df43a64cd55086ec33d6755e6d809b9,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_WATCHER_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-master-centos9/openstack-watcher-api@sha256:e0e84e3fb8f9f12e9280d3d28b415639abfad5b5e46787a61b42beb97f32c76e,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_WATCHER_APPLIER_IMAGE_URL_DEFAULT,Value:quay.io/podified-master-centos9/openstack-watcher-applier@sha256:bd5376b807eca8a409ea663906d9a990cd95a109b8f1f2c56c06f9201be1f6ca,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_WATCHER_DECISION_ENGINE_IMAGE_URL_DEFAULT,Value:quay.io/podified-master-centos9/openstack-watcher-decision-engine@sha256:366de9fadc79d3ea77de93ef1c81e4b8525764f0704e3f30d1a8afa2745c8166,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:cert,ReadOnly:true,MountPath:/tmp/k8s-webhook-server/serving-certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-fnlmg,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod openstack-baremetal-operator-controller-manager-677c5f5bff2dshw_openstack-operators(361b9b35-62d8-494a-9b8b-7d471adb246f): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Oct 08 18:26:32 crc kubenswrapper[4988]: W1008 18:26:32.929851 4988 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3ed83b64_1d35_4ace_853f_2d246e27fc72.slice/crio-7ed1ef8e993034ef4b096eeeb64402f6bdb52518e9edf22e1fded583c916bbb4 WatchSource:0}: Error finding container 7ed1ef8e993034ef4b096eeeb64402f6bdb52518e9edf22e1fded583c916bbb4: Status 404 returned error can't find the container with id 7ed1ef8e993034ef4b096eeeb64402f6bdb52518e9edf22e1fded583c916bbb4 Oct 08 18:26:32 crc kubenswrapper[4988]: E1008 18:26:32.930883 4988 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/telemetry-operator@sha256:9d26476523320d70d6d457b91663e8c233ed320d77032a7c57a89ce1aedd3931,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-8wtq7,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod telemetry-operator-controller-manager-76796d4c6b-2szmt_openstack-operators(c97bc0d9-b498-44ce-9fb5-f4cf7ca7561e): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Oct 08 18:26:32 crc kubenswrapper[4988]: I1008 18:26:32.931486 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/neutron-operator-controller-manager-79d585cb66-9cfhq"] Oct 08 18:26:32 crc kubenswrapper[4988]: I1008 18:26:32.936194 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-manager-598c4c5b5-mxr9w"] Oct 08 18:26:32 crc kubenswrapper[4988]: W1008 18:26:32.938416 4988 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2ebb5c10_6685_4a61_8b57_bd39d36b3ce5.slice/crio-40c1ac7a4a74a4fc321458a72d225d0765d381c9ce6976f50b86bca07d5663d3 WatchSource:0}: Error finding container 40c1ac7a4a74a4fc321458a72d225d0765d381c9ce6976f50b86bca07d5663d3: Status 404 returned error can't find the container with id 40c1ac7a4a74a4fc321458a72d225d0765d381c9ce6976f50b86bca07d5663d3 Oct 08 18:26:32 crc kubenswrapper[4988]: E1008 18:26:32.954868 4988 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/neutron-operator@sha256:33652e75a03a058769019fe8d8c51585a6eeefef5e1ecb96f9965434117954f2,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-d62t7,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod neutron-operator-controller-manager-79d585cb66-9cfhq_openstack-operators(4c6dc76c-9b71-45aa-b1e9-0fd9deecbd39): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Oct 08 18:26:32 crc kubenswrapper[4988]: E1008 18:26:32.955302 4988 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/infra-operator@sha256:5cfb2ae1092445950b39dd59caa9a8c9367f42fb8353a8c3848d3bc729f24492,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:true,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{600 -3} {} 600m DecimalSI},memory: {{2147483648 0} {} 2Gi BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{536870912 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:cert,ReadOnly:true,MountPath:/tmp/k8s-webhook-server/serving-certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-4jdrl,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod infra-operator-controller-manager-656bcbd775-2xfhc_openstack-operators(2ebb5c10-6685-4a61-8b57-bd39d36b3ce5): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Oct 08 18:26:32 crc kubenswrapper[4988]: E1008 18:26:32.956938 4988 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/watcher-operator@sha256:e4ae07e859166fc5e2cb4f8e0e2c3358b9d2e2d6721a3864d2e0c651d36698ca,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-lxfvs,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod watcher-operator-controller-manager-7794bc6bd-wfq6r_openstack-operators(47379f88-653e-4bda-a37b-a81857b7af36): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Oct 08 18:26:32 crc kubenswrapper[4988]: E1008 18:26:32.973913 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/test-operator-controller-manager-56c698c775-2p984" podUID="d3afe0af-aa6b-4321-99f3-39b2f64c2451" Oct 08 18:26:32 crc kubenswrapper[4988]: E1008 18:26:32.981371 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/ovn-operator-controller-manager-79db49b9fb-75xwr" podUID="fc79827a-abba-4525-a705-129531293589" Oct 08 18:26:33 crc kubenswrapper[4988]: I1008 18:26:33.127740 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-f9fb45f8f-mdxnm" event={"ID":"cc3a04bc-795b-49e9-bc53-06b8534244ac","Type":"ContainerStarted","Data":"e2fbe0aeb45d0a6717a02b78cca9d93766ceffa2a8fb8fe7cb444bb16029e304"} Oct 08 18:26:33 crc kubenswrapper[4988]: I1008 18:26:33.129226 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-656bcbd775-2xfhc" event={"ID":"2ebb5c10-6685-4a61-8b57-bd39d36b3ce5","Type":"ContainerStarted","Data":"40c1ac7a4a74a4fc321458a72d225d0765d381c9ce6976f50b86bca07d5663d3"} Oct 08 18:26:33 crc kubenswrapper[4988]: I1008 18:26:33.130080 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-55b6b7c7b8-4ts62" event={"ID":"70347d19-f528-4db3-81e7-6fbca659710e","Type":"ContainerStarted","Data":"68fbedd5817580b5dfae3b52b3c34a9dc58465c7edd9b71a87486537cc8fd85c"} Oct 08 18:26:33 crc kubenswrapper[4988]: I1008 18:26:33.131448 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-84b9b84486-gpl58" event={"ID":"1af0d400-4968-4d6e-af43-0624f7bda6f0","Type":"ContainerStarted","Data":"27e1c3356a5b6500aa560ffc9ab44979b6a6100e5089fb4f7553bc5860725e4b"} Oct 08 18:26:33 crc kubenswrapper[4988]: I1008 18:26:33.132948 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-79d585cb66-9cfhq" event={"ID":"4c6dc76c-9b71-45aa-b1e9-0fd9deecbd39","Type":"ContainerStarted","Data":"047a3cce5138efd6f51b8441d4fdaa0e19ebcaa6018ab40dea939f9f4ec5de32"} Oct 08 18:26:33 crc kubenswrapper[4988]: I1008 18:26:33.134606 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-85d5d9dd78-94rjz" event={"ID":"81f789aa-012b-4ee6-a3c4-70bba5d64291","Type":"ContainerStarted","Data":"e21650c53a51f961da0e00ab0e1689812aa34c34529641df30183848f1986e68"} Oct 08 18:26:33 crc kubenswrapper[4988]: I1008 18:26:33.150047 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-7ffbcb7588-jwr4r" event={"ID":"8600e1a1-b22a-4cdb-9e91-04e0a65b3756","Type":"ContainerStarted","Data":"9c3616ae10e5a38d27af0b3e951ecaf60846162e977a4e149e1b81ee7b26f295"} Oct 08 18:26:33 crc kubenswrapper[4988]: I1008 18:26:33.152200 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-68b6c87b68-rf945" event={"ID":"3433daff-1b78-4aa4-8eec-4bebfcbe1320","Type":"ContainerStarted","Data":"e76586e9a88d2829b2cd575a4a6dfddffbefa033e0d91c0a676f8e9c951372bf"} Oct 08 18:26:33 crc kubenswrapper[4988]: I1008 18:26:33.153200 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-db6d7f97b-mxg6q" event={"ID":"1154bd89-f7b0-4cce-b02c-9157b7866bf8","Type":"ContainerStarted","Data":"bd61dfc036aa9eef676ef1f5fe69c59455de2c59c66c51656fd420047250e090"} Oct 08 18:26:33 crc kubenswrapper[4988]: I1008 18:26:33.159171 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-76796d4c6b-2szmt" event={"ID":"c97bc0d9-b498-44ce-9fb5-f4cf7ca7561e","Type":"ContainerStarted","Data":"a3b2234c16674808eec24b36c42d02a787574c571da000cde7c63b7351cc24a5"} Oct 08 18:26:33 crc kubenswrapper[4988]: I1008 18:26:33.162777 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-69fdcfc5f5-zd6wx" event={"ID":"db43233d-abb6-46e3-9b9f-946024ab477a","Type":"ContainerStarted","Data":"a93eaca4e9a998eb0127e0950f1483753ea547e54571248e508b36d227e5ada7"} Oct 08 18:26:33 crc kubenswrapper[4988]: I1008 18:26:33.166693 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-56c698c775-2p984" event={"ID":"d3afe0af-aa6b-4321-99f3-39b2f64c2451","Type":"ContainerStarted","Data":"f94f8938d46ae02604af69b7888c8392a4bd2b713862e9d71cf648b0701192f1"} Oct 08 18:26:33 crc kubenswrapper[4988]: I1008 18:26:33.166755 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-56c698c775-2p984" event={"ID":"d3afe0af-aa6b-4321-99f3-39b2f64c2451","Type":"ContainerStarted","Data":"5a6620bdea3c93988a624d49221e37baf4e7ab4fad9253886db3d6fd526aa8fe"} Oct 08 18:26:33 crc kubenswrapper[4988]: E1008 18:26:33.168434 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/test-operator@sha256:efa8fb78cffb573d299ffcc7bab1099affd2dbbab222152092b313074306e0a9\\\"\"" pod="openstack-operators/test-operator-controller-manager-56c698c775-2p984" podUID="d3afe0af-aa6b-4321-99f3-39b2f64c2451" Oct 08 18:26:33 crc kubenswrapper[4988]: I1008 18:26:33.168791 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-79db49b9fb-75xwr" event={"ID":"fc79827a-abba-4525-a705-129531293589","Type":"ContainerStarted","Data":"75a907c8abeca23cf7f65037271da7c3aba4f26b5083d6351bb9b4c2cfe77855"} Oct 08 18:26:33 crc kubenswrapper[4988]: I1008 18:26:33.168825 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-79db49b9fb-75xwr" event={"ID":"fc79827a-abba-4525-a705-129531293589","Type":"ContainerStarted","Data":"6c50a1f434e2306e503cb986ebe631bda5ceb80f4446c05023a32f81851bebc7"} Oct 08 18:26:33 crc kubenswrapper[4988]: I1008 18:26:33.172953 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-658bdf4b74-cmnkr" event={"ID":"c569c6ea-0929-4fc6-8d67-47578fa225c1","Type":"ContainerStarted","Data":"a8b681f98d35516fb6e1e85fa5bb5e9fb1bb865892a0bfad25a11a05d5bc5a9f"} Oct 08 18:26:33 crc kubenswrapper[4988]: I1008 18:26:33.173938 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-5df598886f-8589q" event={"ID":"f9ff7c98-5f49-4c45-a367-5ed4471083a8","Type":"ContainerStarted","Data":"b8efc700bf6bd7062d31cbeb5e016b8ccb391b79df050e72728cb124cdd542b6"} Oct 08 18:26:33 crc kubenswrapper[4988]: E1008 18:26:33.174160 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/ovn-operator@sha256:551b59e107c9812f7ad7aa06577376b0dcb58ff9498a41d5d5273e60e20ba7e4\\\"\"" pod="openstack-operators/ovn-operator-controller-manager-79db49b9fb-75xwr" podUID="fc79827a-abba-4525-a705-129531293589" Oct 08 18:26:33 crc kubenswrapper[4988]: I1008 18:26:33.175123 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-598c4c5b5-mxr9w" event={"ID":"3ed83b64-1d35-4ace-853f-2d246e27fc72","Type":"ContainerStarted","Data":"7ed1ef8e993034ef4b096eeeb64402f6bdb52518e9edf22e1fded583c916bbb4"} Oct 08 18:26:33 crc kubenswrapper[4988]: I1008 18:26:33.176156 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-858f76bbdd-rjzd5" event={"ID":"2afedc0a-a797-4f9e-921f-5160fc437f79","Type":"ContainerStarted","Data":"a2c177fbb6214931b00944210662f0dbae908c7df7174068a38b1bff07f42f40"} Oct 08 18:26:33 crc kubenswrapper[4988]: I1008 18:26:33.177707 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-677c5f5bff2dshw" event={"ID":"361b9b35-62d8-494a-9b8b-7d471adb246f","Type":"ContainerStarted","Data":"5430f4e6869c244c7cac8486aa213760d98862b9dd58dd8ba335dd1b5d91e8f0"} Oct 08 18:26:33 crc kubenswrapper[4988]: I1008 18:26:33.195856 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-9c5c78d49-fv7xm" event={"ID":"70495ce4-a8be-4853-be10-a5fe9e1020a7","Type":"ContainerStarted","Data":"5ef47f39d4fe67de4190c410a06ed8a3e54602d4316efe5906fb249e6a843068"} Oct 08 18:26:33 crc kubenswrapper[4988]: I1008 18:26:33.209631 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-25bz8" event={"ID":"e4ef3040-d167-453d-aaa4-b807020a6876","Type":"ContainerStarted","Data":"56746cbfeb0bb92d0ded5b07ddb688cb3254851798a0121442880168ad70b563"} Oct 08 18:26:33 crc kubenswrapper[4988]: E1008 18:26:33.213701 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2\\\"\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-25bz8" podUID="e4ef3040-d167-453d-aaa4-b807020a6876" Oct 08 18:26:33 crc kubenswrapper[4988]: I1008 18:26:33.227798 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-7794bc6bd-wfq6r" event={"ID":"47379f88-653e-4bda-a37b-a81857b7af36","Type":"ContainerStarted","Data":"559aeb29c31323b87178e73a473bb1f9e672e4a7f936d9a7fd100548ffa3688e"} Oct 08 18:26:33 crc kubenswrapper[4988]: I1008 18:26:33.292695 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-5f67fbc655-m7d7l" event={"ID":"128cb2db-23b2-423f-9ec1-a7c4ed246731","Type":"ContainerStarted","Data":"84e2f1eab2f0a6dc1a9f284cc075b2ab96fadaa52264fa5f6db03f41139a7186"} Oct 08 18:26:33 crc kubenswrapper[4988]: E1008 18:26:33.405990 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/telemetry-operator-controller-manager-76796d4c6b-2szmt" podUID="c97bc0d9-b498-44ce-9fb5-f4cf7ca7561e" Oct 08 18:26:33 crc kubenswrapper[4988]: E1008 18:26:33.455462 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/openstack-baremetal-operator-controller-manager-677c5f5bff2dshw" podUID="361b9b35-62d8-494a-9b8b-7d471adb246f" Oct 08 18:26:33 crc kubenswrapper[4988]: E1008 18:26:33.510649 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/infra-operator-controller-manager-656bcbd775-2xfhc" podUID="2ebb5c10-6685-4a61-8b57-bd39d36b3ce5" Oct 08 18:26:33 crc kubenswrapper[4988]: E1008 18:26:33.612955 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/neutron-operator-controller-manager-79d585cb66-9cfhq" podUID="4c6dc76c-9b71-45aa-b1e9-0fd9deecbd39" Oct 08 18:26:33 crc kubenswrapper[4988]: E1008 18:26:33.637100 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/watcher-operator-controller-manager-7794bc6bd-wfq6r" podUID="47379f88-653e-4bda-a37b-a81857b7af36" Oct 08 18:26:34 crc kubenswrapper[4988]: I1008 18:26:34.338330 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-677c5f5bff2dshw" event={"ID":"361b9b35-62d8-494a-9b8b-7d471adb246f","Type":"ContainerStarted","Data":"1ceb466c41b4bd3705711dc467258e5719d2df8e3a73e957bfa71864017711d3"} Oct 08 18:26:34 crc kubenswrapper[4988]: I1008 18:26:34.342339 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-7794bc6bd-wfq6r" event={"ID":"47379f88-653e-4bda-a37b-a81857b7af36","Type":"ContainerStarted","Data":"7de5db252c1d54e2bb3f9dc2af231e4945b07159051af75184b87ec0eb21a312"} Oct 08 18:26:34 crc kubenswrapper[4988]: I1008 18:26:34.344601 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-656bcbd775-2xfhc" event={"ID":"2ebb5c10-6685-4a61-8b57-bd39d36b3ce5","Type":"ContainerStarted","Data":"13559939b667a8fa3b476613b09f278e1847a85c77d442bffac73d1e268dcdb4"} Oct 08 18:26:34 crc kubenswrapper[4988]: E1008 18:26:34.346323 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/openstack-baremetal-operator@sha256:a17fc270857869fd1efe5020b2a1cb8c2abbd838f08de88f3a6a59e8754ec351\\\"\"" pod="openstack-operators/openstack-baremetal-operator-controller-manager-677c5f5bff2dshw" podUID="361b9b35-62d8-494a-9b8b-7d471adb246f" Oct 08 18:26:34 crc kubenswrapper[4988]: I1008 18:26:34.346985 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-79d585cb66-9cfhq" event={"ID":"4c6dc76c-9b71-45aa-b1e9-0fd9deecbd39","Type":"ContainerStarted","Data":"a6002e4ea0537788c66b2b252dfd01c7ecb0ac82175cf50a09ab96b12fe1dbfa"} Oct 08 18:26:34 crc kubenswrapper[4988]: E1008 18:26:34.347207 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/watcher-operator@sha256:e4ae07e859166fc5e2cb4f8e0e2c3358b9d2e2d6721a3864d2e0c651d36698ca\\\"\"" pod="openstack-operators/watcher-operator-controller-manager-7794bc6bd-wfq6r" podUID="47379f88-653e-4bda-a37b-a81857b7af36" Oct 08 18:26:34 crc kubenswrapper[4988]: E1008 18:26:34.348067 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/infra-operator@sha256:5cfb2ae1092445950b39dd59caa9a8c9367f42fb8353a8c3848d3bc729f24492\\\"\"" pod="openstack-operators/infra-operator-controller-manager-656bcbd775-2xfhc" podUID="2ebb5c10-6685-4a61-8b57-bd39d36b3ce5" Oct 08 18:26:34 crc kubenswrapper[4988]: I1008 18:26:34.352115 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-76796d4c6b-2szmt" event={"ID":"c97bc0d9-b498-44ce-9fb5-f4cf7ca7561e","Type":"ContainerStarted","Data":"cedcf8b64a867c6f1c8285013d6f23e1909495bea231a7c58ee2f3125ce855ac"} Oct 08 18:26:34 crc kubenswrapper[4988]: E1008 18:26:34.369997 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/neutron-operator@sha256:33652e75a03a058769019fe8d8c51585a6eeefef5e1ecb96f9965434117954f2\\\"\"" pod="openstack-operators/neutron-operator-controller-manager-79d585cb66-9cfhq" podUID="4c6dc76c-9b71-45aa-b1e9-0fd9deecbd39" Oct 08 18:26:34 crc kubenswrapper[4988]: E1008 18:26:34.370358 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/telemetry-operator@sha256:9d26476523320d70d6d457b91663e8c233ed320d77032a7c57a89ce1aedd3931\\\"\"" pod="openstack-operators/telemetry-operator-controller-manager-76796d4c6b-2szmt" podUID="c97bc0d9-b498-44ce-9fb5-f4cf7ca7561e" Oct 08 18:26:34 crc kubenswrapper[4988]: I1008 18:26:34.394729 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-598c4c5b5-mxr9w" event={"ID":"3ed83b64-1d35-4ace-853f-2d246e27fc72","Type":"ContainerStarted","Data":"685bf1914b6922593e8625f4d8a9ccc2b5647389299efc9de015eb0299c7eb3f"} Oct 08 18:26:34 crc kubenswrapper[4988]: I1008 18:26:34.394777 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-598c4c5b5-mxr9w" event={"ID":"3ed83b64-1d35-4ace-853f-2d246e27fc72","Type":"ContainerStarted","Data":"1ec94da1d3d43f6e9a3ea31b6e7049675c37c97d627a8b3bed58026d1fa77d48"} Oct 08 18:26:34 crc kubenswrapper[4988]: I1008 18:26:34.395072 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-manager-598c4c5b5-mxr9w" Oct 08 18:26:34 crc kubenswrapper[4988]: E1008 18:26:34.422177 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2\\\"\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-25bz8" podUID="e4ef3040-d167-453d-aaa4-b807020a6876" Oct 08 18:26:34 crc kubenswrapper[4988]: E1008 18:26:34.422564 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/test-operator@sha256:efa8fb78cffb573d299ffcc7bab1099affd2dbbab222152092b313074306e0a9\\\"\"" pod="openstack-operators/test-operator-controller-manager-56c698c775-2p984" podUID="d3afe0af-aa6b-4321-99f3-39b2f64c2451" Oct 08 18:26:34 crc kubenswrapper[4988]: E1008 18:26:34.422618 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/ovn-operator@sha256:551b59e107c9812f7ad7aa06577376b0dcb58ff9498a41d5d5273e60e20ba7e4\\\"\"" pod="openstack-operators/ovn-operator-controller-manager-79db49b9fb-75xwr" podUID="fc79827a-abba-4525-a705-129531293589" Oct 08 18:26:34 crc kubenswrapper[4988]: I1008 18:26:34.862839 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-controller-manager-598c4c5b5-mxr9w" podStartSLOduration=3.862815366 podStartE2EDuration="3.862815366s" podCreationTimestamp="2025-10-08 18:26:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 18:26:34.843417384 +0000 UTC m=+940.293260184" watchObservedRunningTime="2025-10-08 18:26:34.862815366 +0000 UTC m=+940.312658136" Oct 08 18:26:35 crc kubenswrapper[4988]: E1008 18:26:35.417972 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/telemetry-operator@sha256:9d26476523320d70d6d457b91663e8c233ed320d77032a7c57a89ce1aedd3931\\\"\"" pod="openstack-operators/telemetry-operator-controller-manager-76796d4c6b-2szmt" podUID="c97bc0d9-b498-44ce-9fb5-f4cf7ca7561e" Oct 08 18:26:35 crc kubenswrapper[4988]: E1008 18:26:35.418473 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/openstack-baremetal-operator@sha256:a17fc270857869fd1efe5020b2a1cb8c2abbd838f08de88f3a6a59e8754ec351\\\"\"" pod="openstack-operators/openstack-baremetal-operator-controller-manager-677c5f5bff2dshw" podUID="361b9b35-62d8-494a-9b8b-7d471adb246f" Oct 08 18:26:35 crc kubenswrapper[4988]: E1008 18:26:35.419361 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/neutron-operator@sha256:33652e75a03a058769019fe8d8c51585a6eeefef5e1ecb96f9965434117954f2\\\"\"" pod="openstack-operators/neutron-operator-controller-manager-79d585cb66-9cfhq" podUID="4c6dc76c-9b71-45aa-b1e9-0fd9deecbd39" Oct 08 18:26:35 crc kubenswrapper[4988]: E1008 18:26:35.422124 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/infra-operator@sha256:5cfb2ae1092445950b39dd59caa9a8c9367f42fb8353a8c3848d3bc729f24492\\\"\"" pod="openstack-operators/infra-operator-controller-manager-656bcbd775-2xfhc" podUID="2ebb5c10-6685-4a61-8b57-bd39d36b3ce5" Oct 08 18:26:35 crc kubenswrapper[4988]: E1008 18:26:35.424516 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/watcher-operator@sha256:e4ae07e859166fc5e2cb4f8e0e2c3358b9d2e2d6721a3864d2e0c651d36698ca\\\"\"" pod="openstack-operators/watcher-operator-controller-manager-7794bc6bd-wfq6r" podUID="47379f88-653e-4bda-a37b-a81857b7af36" Oct 08 18:26:42 crc kubenswrapper[4988]: I1008 18:26:42.349427 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-manager-598c4c5b5-mxr9w" Oct 08 18:26:44 crc kubenswrapper[4988]: I1008 18:26:44.520241 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-7b7fb68549-7xcz4" event={"ID":"420edeb5-79c9-4ece-9e95-d497d4be3b76","Type":"ContainerStarted","Data":"fc077a7adbae9490c30af5db07e82a863dee995c738254922c75d2c6018371b0"} Oct 08 18:26:44 crc kubenswrapper[4988]: I1008 18:26:44.533536 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-9c5c78d49-fv7xm" event={"ID":"70495ce4-a8be-4853-be10-a5fe9e1020a7","Type":"ContainerStarted","Data":"230002bef3a479a9c74da09b2fc1ba3ed3e60507ce9ef6552750e19acf547118"} Oct 08 18:26:44 crc kubenswrapper[4988]: I1008 18:26:44.534835 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ironic-operator-controller-manager-9c5c78d49-fv7xm" Oct 08 18:26:44 crc kubenswrapper[4988]: I1008 18:26:44.559938 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-f9fb45f8f-mdxnm" event={"ID":"cc3a04bc-795b-49e9-bc53-06b8534244ac","Type":"ContainerStarted","Data":"8278e6deb4b7a758c41b220d479a19483686b7a1117b3fbef7e91451430763d1"} Oct 08 18:26:44 crc kubenswrapper[4988]: I1008 18:26:44.565905 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/ironic-operator-controller-manager-9c5c78d49-fv7xm" podStartSLOduration=3.407435104 podStartE2EDuration="14.565865413s" podCreationTimestamp="2025-10-08 18:26:30 +0000 UTC" firstStartedPulling="2025-10-08 18:26:32.617155427 +0000 UTC m=+938.066998197" lastFinishedPulling="2025-10-08 18:26:43.775585736 +0000 UTC m=+949.225428506" observedRunningTime="2025-10-08 18:26:44.562638703 +0000 UTC m=+950.012481493" watchObservedRunningTime="2025-10-08 18:26:44.565865413 +0000 UTC m=+950.015708183" Oct 08 18:26:44 crc kubenswrapper[4988]: I1008 18:26:44.579134 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-db6d7f97b-mxg6q" event={"ID":"1154bd89-f7b0-4cce-b02c-9157b7866bf8","Type":"ContainerStarted","Data":"6652a96dfd0f4377f12fa4f0affe4f581315cdd6719f4081a4a1b69a450b579d"} Oct 08 18:26:44 crc kubenswrapper[4988]: I1008 18:26:44.596235 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-84b9b84486-gpl58" event={"ID":"1af0d400-4968-4d6e-af43-0624f7bda6f0","Type":"ContainerStarted","Data":"ef3fe6cf9f1c76391e88b627cd73c2c73ca961543de2200239da1902525e4bfc"} Oct 08 18:26:44 crc kubenswrapper[4988]: I1008 18:26:44.611551 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-5df598886f-8589q" event={"ID":"f9ff7c98-5f49-4c45-a367-5ed4471083a8","Type":"ContainerStarted","Data":"cb4119ef092cae75249c41a750c0f2a0d5ccc4a681104c4d1f3a522c2b93c505"} Oct 08 18:26:44 crc kubenswrapper[4988]: I1008 18:26:44.626108 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-858f76bbdd-rjzd5" event={"ID":"2afedc0a-a797-4f9e-921f-5160fc437f79","Type":"ContainerStarted","Data":"023988fbc72cf0ee33404006fbe97d44597b4596993984cac6c61bac0fbbd201"} Oct 08 18:26:44 crc kubenswrapper[4988]: I1008 18:26:44.635084 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-85d5d9dd78-94rjz" event={"ID":"81f789aa-012b-4ee6-a3c4-70bba5d64291","Type":"ContainerStarted","Data":"d94bc4d1c6ddb9a1971c5d419bbeff5e5d8cd13ca9a30d45580af14a208137b8"} Oct 08 18:26:44 crc kubenswrapper[4988]: I1008 18:26:44.642659 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-5f67fbc655-m7d7l" event={"ID":"128cb2db-23b2-423f-9ec1-a7c4ed246731","Type":"ContainerStarted","Data":"a7ada31e27351240b528877f8889bfa876ea3acb8ee8cb0b306b73097eeab68d"} Oct 08 18:26:44 crc kubenswrapper[4988]: I1008 18:26:44.656760 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-68b6c87b68-rf945" event={"ID":"3433daff-1b78-4aa4-8eec-4bebfcbe1320","Type":"ContainerStarted","Data":"5bcdf76b5d2ffb060984a05b53e617d3edc5f0d4e1470f3b9f12ab71fbfc0cbc"} Oct 08 18:26:44 crc kubenswrapper[4988]: I1008 18:26:44.680678 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-658bdf4b74-cmnkr" event={"ID":"c569c6ea-0929-4fc6-8d67-47578fa225c1","Type":"ContainerStarted","Data":"7be742c382cc1ca31cc132994695a1d1ec94d724bfb21993d6f4dc8686a10fe3"} Oct 08 18:26:45 crc kubenswrapper[4988]: I1008 18:26:45.691445 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-7b7fb68549-7xcz4" event={"ID":"420edeb5-79c9-4ece-9e95-d497d4be3b76","Type":"ContainerStarted","Data":"a5f4359f21ab80056d3cecd1813aade0e8d32d70db3d994dd41ffa012e343a54"} Oct 08 18:26:45 crc kubenswrapper[4988]: I1008 18:26:45.692359 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/cinder-operator-controller-manager-7b7fb68549-7xcz4" Oct 08 18:26:45 crc kubenswrapper[4988]: I1008 18:26:45.694750 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-68b6c87b68-rf945" event={"ID":"3433daff-1b78-4aa4-8eec-4bebfcbe1320","Type":"ContainerStarted","Data":"0ec23249f347da2e7caa342f5b80db179849aafb587756e289c9ad805ca809c9"} Oct 08 18:26:45 crc kubenswrapper[4988]: I1008 18:26:45.695132 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/placement-operator-controller-manager-68b6c87b68-rf945" Oct 08 18:26:45 crc kubenswrapper[4988]: I1008 18:26:45.697930 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-7ffbcb7588-jwr4r" event={"ID":"8600e1a1-b22a-4cdb-9e91-04e0a65b3756","Type":"ContainerStarted","Data":"0284f840c1127a6aa9ea4a1723ffce667ea9b41a7f7d2505ab15135f7c0f9842"} Oct 08 18:26:45 crc kubenswrapper[4988]: I1008 18:26:45.697958 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-7ffbcb7588-jwr4r" event={"ID":"8600e1a1-b22a-4cdb-9e91-04e0a65b3756","Type":"ContainerStarted","Data":"cdf4f01f8b057091ad5d43c714a431083a041dc1ae4935a22d902ba69e803707"} Oct 08 18:26:45 crc kubenswrapper[4988]: I1008 18:26:45.698066 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/horizon-operator-controller-manager-7ffbcb7588-jwr4r" Oct 08 18:26:45 crc kubenswrapper[4988]: I1008 18:26:45.700323 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-55b6b7c7b8-4ts62" event={"ID":"70347d19-f528-4db3-81e7-6fbca659710e","Type":"ContainerStarted","Data":"1781e167c9ce52f9efeecef94163dbd2735177e51c43768a3c7fb3467b83f459"} Oct 08 18:26:45 crc kubenswrapper[4988]: I1008 18:26:45.700349 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-55b6b7c7b8-4ts62" event={"ID":"70347d19-f528-4db3-81e7-6fbca659710e","Type":"ContainerStarted","Data":"f88795e722e0bb8c7645886be470b6572cc0ce5c75efca652a45ac04f3bef42d"} Oct 08 18:26:45 crc kubenswrapper[4988]: I1008 18:26:45.700487 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/keystone-operator-controller-manager-55b6b7c7b8-4ts62" Oct 08 18:26:45 crc kubenswrapper[4988]: I1008 18:26:45.703092 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-84b9b84486-gpl58" event={"ID":"1af0d400-4968-4d6e-af43-0624f7bda6f0","Type":"ContainerStarted","Data":"9af14de8faa3f88c5113ddf17236fd3cfc42542b9ca882b2d6ba55953423f94f"} Oct 08 18:26:45 crc kubenswrapper[4988]: I1008 18:26:45.703172 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/glance-operator-controller-manager-84b9b84486-gpl58" Oct 08 18:26:45 crc kubenswrapper[4988]: I1008 18:26:45.705533 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-85d5d9dd78-94rjz" event={"ID":"81f789aa-012b-4ee6-a3c4-70bba5d64291","Type":"ContainerStarted","Data":"5d87b35be8003873304214c4ec46e174ac79f76a61ed78a266105a28f64fd436"} Oct 08 18:26:45 crc kubenswrapper[4988]: I1008 18:26:45.705668 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/designate-operator-controller-manager-85d5d9dd78-94rjz" Oct 08 18:26:45 crc kubenswrapper[4988]: I1008 18:26:45.708657 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-5f67fbc655-m7d7l" event={"ID":"128cb2db-23b2-423f-9ec1-a7c4ed246731","Type":"ContainerStarted","Data":"a2cdd1499ba72e769fb373a84c14661dcb9eecd82dda68a5a18af4103560707f"} Oct 08 18:26:45 crc kubenswrapper[4988]: I1008 18:26:45.708807 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/manila-operator-controller-manager-5f67fbc655-m7d7l" Oct 08 18:26:45 crc kubenswrapper[4988]: I1008 18:26:45.710753 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-9c5c78d49-fv7xm" event={"ID":"70495ce4-a8be-4853-be10-a5fe9e1020a7","Type":"ContainerStarted","Data":"29eb0b85fc92b5db479d9bef649e921f8d244ce0337ea392f3a88c960217eeb6"} Oct 08 18:26:45 crc kubenswrapper[4988]: I1008 18:26:45.713687 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-f9fb45f8f-mdxnm" event={"ID":"cc3a04bc-795b-49e9-bc53-06b8534244ac","Type":"ContainerStarted","Data":"5f87e2a91063bdcf8d35ab99dfa7be4716f27494a73a2380726abd131f74c7cc"} Oct 08 18:26:45 crc kubenswrapper[4988]: I1008 18:26:45.713923 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/mariadb-operator-controller-manager-f9fb45f8f-mdxnm" Oct 08 18:26:45 crc kubenswrapper[4988]: I1008 18:26:45.716405 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-658bdf4b74-cmnkr" event={"ID":"c569c6ea-0929-4fc6-8d67-47578fa225c1","Type":"ContainerStarted","Data":"58da89663a9a2d057bd31dfc81009505efb385e5905381b1e2017909fe1d205f"} Oct 08 18:26:45 crc kubenswrapper[4988]: I1008 18:26:45.716803 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/barbican-operator-controller-manager-658bdf4b74-cmnkr" Oct 08 18:26:45 crc kubenswrapper[4988]: I1008 18:26:45.719137 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-5df598886f-8589q" event={"ID":"f9ff7c98-5f49-4c45-a367-5ed4471083a8","Type":"ContainerStarted","Data":"c14715eecdddd69ffe36bbbd6f466ea2cdb8d238c2664525fd5fc791c5157289"} Oct 08 18:26:45 crc kubenswrapper[4988]: I1008 18:26:45.719499 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/nova-operator-controller-manager-5df598886f-8589q" Oct 08 18:26:45 crc kubenswrapper[4988]: I1008 18:26:45.721112 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-db6d7f97b-mxg6q" event={"ID":"1154bd89-f7b0-4cce-b02c-9157b7866bf8","Type":"ContainerStarted","Data":"c4e595186c63e73a989928567f20f7a58148c840a4c4ae59b38a8c797564dd06"} Oct 08 18:26:45 crc kubenswrapper[4988]: I1008 18:26:45.721525 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/swift-operator-controller-manager-db6d7f97b-mxg6q" Oct 08 18:26:45 crc kubenswrapper[4988]: I1008 18:26:45.723581 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-858f76bbdd-rjzd5" event={"ID":"2afedc0a-a797-4f9e-921f-5160fc437f79","Type":"ContainerStarted","Data":"31508748c4a514670c46c3af5fc7f56622bb8b018094d0f9683b498f700a7393"} Oct 08 18:26:45 crc kubenswrapper[4988]: I1008 18:26:45.723751 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/heat-operator-controller-manager-858f76bbdd-rjzd5" Oct 08 18:26:45 crc kubenswrapper[4988]: I1008 18:26:45.725996 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-69fdcfc5f5-zd6wx" event={"ID":"db43233d-abb6-46e3-9b9f-946024ab477a","Type":"ContainerStarted","Data":"8df704ff319020a26e8cf9de8d43690682dc481c0ba32c913d1f2505575d56d7"} Oct 08 18:26:45 crc kubenswrapper[4988]: I1008 18:26:45.726031 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-69fdcfc5f5-zd6wx" event={"ID":"db43233d-abb6-46e3-9b9f-946024ab477a","Type":"ContainerStarted","Data":"de3f4924d2890e7c01c2fa12e30d4d01defb0b51a58d1aaa272967fba28947b4"} Oct 08 18:26:45 crc kubenswrapper[4988]: I1008 18:26:45.726346 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/octavia-operator-controller-manager-69fdcfc5f5-zd6wx" Oct 08 18:26:45 crc kubenswrapper[4988]: I1008 18:26:45.737603 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/cinder-operator-controller-manager-7b7fb68549-7xcz4" podStartSLOduration=3.976703363 podStartE2EDuration="15.7375826s" podCreationTimestamp="2025-10-08 18:26:30 +0000 UTC" firstStartedPulling="2025-10-08 18:26:32.038717085 +0000 UTC m=+937.488559855" lastFinishedPulling="2025-10-08 18:26:43.799596322 +0000 UTC m=+949.249439092" observedRunningTime="2025-10-08 18:26:45.73051298 +0000 UTC m=+951.180355770" watchObservedRunningTime="2025-10-08 18:26:45.7375826 +0000 UTC m=+951.187425370" Oct 08 18:26:45 crc kubenswrapper[4988]: I1008 18:26:45.752703 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/barbican-operator-controller-manager-658bdf4b74-cmnkr" podStartSLOduration=4.093688744 podStartE2EDuration="15.752683618s" podCreationTimestamp="2025-10-08 18:26:30 +0000 UTC" firstStartedPulling="2025-10-08 18:26:32.116411236 +0000 UTC m=+937.566254006" lastFinishedPulling="2025-10-08 18:26:43.77540611 +0000 UTC m=+949.225248880" observedRunningTime="2025-10-08 18:26:45.749100366 +0000 UTC m=+951.198943156" watchObservedRunningTime="2025-10-08 18:26:45.752683618 +0000 UTC m=+951.202526388" Oct 08 18:26:45 crc kubenswrapper[4988]: I1008 18:26:45.772736 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/keystone-operator-controller-manager-55b6b7c7b8-4ts62" podStartSLOduration=4.910168354 podStartE2EDuration="15.77271532s" podCreationTimestamp="2025-10-08 18:26:30 +0000 UTC" firstStartedPulling="2025-10-08 18:26:32.912582437 +0000 UTC m=+938.362425207" lastFinishedPulling="2025-10-08 18:26:43.775129403 +0000 UTC m=+949.224972173" observedRunningTime="2025-10-08 18:26:45.767670713 +0000 UTC m=+951.217513483" watchObservedRunningTime="2025-10-08 18:26:45.77271532 +0000 UTC m=+951.222558090" Oct 08 18:26:45 crc kubenswrapper[4988]: I1008 18:26:45.835169 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/manila-operator-controller-manager-5f67fbc655-m7d7l" podStartSLOduration=4.467732534 podStartE2EDuration="15.835145178s" podCreationTimestamp="2025-10-08 18:26:30 +0000 UTC" firstStartedPulling="2025-10-08 18:26:32.414261621 +0000 UTC m=+937.864104391" lastFinishedPulling="2025-10-08 18:26:43.781674265 +0000 UTC m=+949.231517035" observedRunningTime="2025-10-08 18:26:45.804871147 +0000 UTC m=+951.254713937" watchObservedRunningTime="2025-10-08 18:26:45.835145178 +0000 UTC m=+951.284987958" Oct 08 18:26:45 crc kubenswrapper[4988]: I1008 18:26:45.836134 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/glance-operator-controller-manager-84b9b84486-gpl58" podStartSLOduration=4.248524908 podStartE2EDuration="15.836124767s" podCreationTimestamp="2025-10-08 18:26:30 +0000 UTC" firstStartedPulling="2025-10-08 18:26:32.231169298 +0000 UTC m=+937.681012068" lastFinishedPulling="2025-10-08 18:26:43.818769157 +0000 UTC m=+949.268611927" observedRunningTime="2025-10-08 18:26:45.830698909 +0000 UTC m=+951.280541689" watchObservedRunningTime="2025-10-08 18:26:45.836124767 +0000 UTC m=+951.285967537" Oct 08 18:26:45 crc kubenswrapper[4988]: I1008 18:26:45.860038 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/horizon-operator-controller-manager-7ffbcb7588-jwr4r" podStartSLOduration=4.44508894 podStartE2EDuration="15.860015319s" podCreationTimestamp="2025-10-08 18:26:30 +0000 UTC" firstStartedPulling="2025-10-08 18:26:32.411481214 +0000 UTC m=+937.861323984" lastFinishedPulling="2025-10-08 18:26:43.826407603 +0000 UTC m=+949.276250363" observedRunningTime="2025-10-08 18:26:45.858257634 +0000 UTC m=+951.308100414" watchObservedRunningTime="2025-10-08 18:26:45.860015319 +0000 UTC m=+951.309858099" Oct 08 18:26:45 crc kubenswrapper[4988]: I1008 18:26:45.879941 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/swift-operator-controller-manager-db6d7f97b-mxg6q" podStartSLOduration=4.704081068 podStartE2EDuration="15.879919967s" podCreationTimestamp="2025-10-08 18:26:30 +0000 UTC" firstStartedPulling="2025-10-08 18:26:32.652783383 +0000 UTC m=+938.102626153" lastFinishedPulling="2025-10-08 18:26:43.828622282 +0000 UTC m=+949.278465052" observedRunningTime="2025-10-08 18:26:45.876944375 +0000 UTC m=+951.326787165" watchObservedRunningTime="2025-10-08 18:26:45.879919967 +0000 UTC m=+951.329762737" Oct 08 18:26:45 crc kubenswrapper[4988]: I1008 18:26:45.908376 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/placement-operator-controller-manager-68b6c87b68-rf945" podStartSLOduration=4.540042787 podStartE2EDuration="15.90835664s" podCreationTimestamp="2025-10-08 18:26:30 +0000 UTC" firstStartedPulling="2025-10-08 18:26:32.437937595 +0000 UTC m=+937.887780355" lastFinishedPulling="2025-10-08 18:26:43.806251438 +0000 UTC m=+949.256094208" observedRunningTime="2025-10-08 18:26:45.905401258 +0000 UTC m=+951.355244038" watchObservedRunningTime="2025-10-08 18:26:45.90835664 +0000 UTC m=+951.358199410" Oct 08 18:26:45 crc kubenswrapper[4988]: I1008 18:26:45.932969 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/mariadb-operator-controller-manager-f9fb45f8f-mdxnm" podStartSLOduration=4.548245641 podStartE2EDuration="15.932952593s" podCreationTimestamp="2025-10-08 18:26:30 +0000 UTC" firstStartedPulling="2025-10-08 18:26:32.437194372 +0000 UTC m=+937.887037142" lastFinishedPulling="2025-10-08 18:26:43.821901334 +0000 UTC m=+949.271744094" observedRunningTime="2025-10-08 18:26:45.932901252 +0000 UTC m=+951.382744042" watchObservedRunningTime="2025-10-08 18:26:45.932952593 +0000 UTC m=+951.382795363" Oct 08 18:26:45 crc kubenswrapper[4988]: I1008 18:26:45.957881 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/heat-operator-controller-manager-858f76bbdd-rjzd5" podStartSLOduration=4.369158363 podStartE2EDuration="15.957862886s" podCreationTimestamp="2025-10-08 18:26:30 +0000 UTC" firstStartedPulling="2025-10-08 18:26:32.237298118 +0000 UTC m=+937.687140888" lastFinishedPulling="2025-10-08 18:26:43.826002641 +0000 UTC m=+949.275845411" observedRunningTime="2025-10-08 18:26:45.955749291 +0000 UTC m=+951.405592071" watchObservedRunningTime="2025-10-08 18:26:45.957862886 +0000 UTC m=+951.407705656" Oct 08 18:26:45 crc kubenswrapper[4988]: I1008 18:26:45.979231 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/nova-operator-controller-manager-5df598886f-8589q" podStartSLOduration=4.836678122 podStartE2EDuration="15.979209778s" podCreationTimestamp="2025-10-08 18:26:30 +0000 UTC" firstStartedPulling="2025-10-08 18:26:32.679618596 +0000 UTC m=+938.129461366" lastFinishedPulling="2025-10-08 18:26:43.822150252 +0000 UTC m=+949.271993022" observedRunningTime="2025-10-08 18:26:45.973402098 +0000 UTC m=+951.423244888" watchObservedRunningTime="2025-10-08 18:26:45.979209778 +0000 UTC m=+951.429052548" Oct 08 18:26:46 crc kubenswrapper[4988]: I1008 18:26:46.005329 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/octavia-operator-controller-manager-69fdcfc5f5-zd6wx" podStartSLOduration=4.606353034 podStartE2EDuration="16.005312489s" podCreationTimestamp="2025-10-08 18:26:30 +0000 UTC" firstStartedPulling="2025-10-08 18:26:32.42651602 +0000 UTC m=+937.876358790" lastFinishedPulling="2025-10-08 18:26:43.825475475 +0000 UTC m=+949.275318245" observedRunningTime="2025-10-08 18:26:46.001513211 +0000 UTC m=+951.451355991" watchObservedRunningTime="2025-10-08 18:26:46.005312489 +0000 UTC m=+951.455155259" Oct 08 18:26:46 crc kubenswrapper[4988]: I1008 18:26:46.028212 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/designate-operator-controller-manager-85d5d9dd78-94rjz" podStartSLOduration=4.496215866 podStartE2EDuration="16.028194049s" podCreationTimestamp="2025-10-08 18:26:30 +0000 UTC" firstStartedPulling="2025-10-08 18:26:32.259786056 +0000 UTC m=+937.709628826" lastFinishedPulling="2025-10-08 18:26:43.791764239 +0000 UTC m=+949.241607009" observedRunningTime="2025-10-08 18:26:46.02693096 +0000 UTC m=+951.476773730" watchObservedRunningTime="2025-10-08 18:26:46.028194049 +0000 UTC m=+951.478036819" Oct 08 18:26:48 crc kubenswrapper[4988]: I1008 18:26:48.751848 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-656bcbd775-2xfhc" event={"ID":"2ebb5c10-6685-4a61-8b57-bd39d36b3ce5","Type":"ContainerStarted","Data":"cb0a8dc4ed939f39dc8d9836ce2856a5d48875900f6a5d01bc1afd37256c4964"} Oct 08 18:26:48 crc kubenswrapper[4988]: I1008 18:26:48.752435 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/infra-operator-controller-manager-656bcbd775-2xfhc" Oct 08 18:26:48 crc kubenswrapper[4988]: I1008 18:26:48.755197 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-25bz8" event={"ID":"e4ef3040-d167-453d-aaa4-b807020a6876","Type":"ContainerStarted","Data":"956bc97073e56f19940e54b5e016e2a9c719b891172316092044b9a8efe5b270"} Oct 08 18:26:48 crc kubenswrapper[4988]: I1008 18:26:48.776927 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/infra-operator-controller-manager-656bcbd775-2xfhc" podStartSLOduration=3.354752139 podStartE2EDuration="18.776903079s" podCreationTimestamp="2025-10-08 18:26:30 +0000 UTC" firstStartedPulling="2025-10-08 18:26:32.955197719 +0000 UTC m=+938.405040489" lastFinishedPulling="2025-10-08 18:26:48.377348659 +0000 UTC m=+953.827191429" observedRunningTime="2025-10-08 18:26:48.77143476 +0000 UTC m=+954.221277550" watchObservedRunningTime="2025-10-08 18:26:48.776903079 +0000 UTC m=+954.226745849" Oct 08 18:26:50 crc kubenswrapper[4988]: I1008 18:26:50.694416 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/barbican-operator-controller-manager-658bdf4b74-cmnkr" Oct 08 18:26:50 crc kubenswrapper[4988]: I1008 18:26:50.714278 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-25bz8" podStartSLOduration=4.257438712 podStartE2EDuration="19.714254538s" podCreationTimestamp="2025-10-08 18:26:31 +0000 UTC" firstStartedPulling="2025-10-08 18:26:32.916710044 +0000 UTC m=+938.366552814" lastFinishedPulling="2025-10-08 18:26:48.37352587 +0000 UTC m=+953.823368640" observedRunningTime="2025-10-08 18:26:48.799764379 +0000 UTC m=+954.249607169" watchObservedRunningTime="2025-10-08 18:26:50.714254538 +0000 UTC m=+956.164097408" Oct 08 18:26:50 crc kubenswrapper[4988]: I1008 18:26:50.716494 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/cinder-operator-controller-manager-7b7fb68549-7xcz4" Oct 08 18:26:50 crc kubenswrapper[4988]: I1008 18:26:50.760044 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/designate-operator-controller-manager-85d5d9dd78-94rjz" Oct 08 18:26:50 crc kubenswrapper[4988]: I1008 18:26:50.787510 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/glance-operator-controller-manager-84b9b84486-gpl58" Oct 08 18:26:50 crc kubenswrapper[4988]: I1008 18:26:50.826655 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/heat-operator-controller-manager-858f76bbdd-rjzd5" Oct 08 18:26:50 crc kubenswrapper[4988]: I1008 18:26:50.939102 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ironic-operator-controller-manager-9c5c78d49-fv7xm" Oct 08 18:26:51 crc kubenswrapper[4988]: I1008 18:26:51.106768 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/keystone-operator-controller-manager-55b6b7c7b8-4ts62" Oct 08 18:26:51 crc kubenswrapper[4988]: I1008 18:26:51.120647 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/mariadb-operator-controller-manager-f9fb45f8f-mdxnm" Oct 08 18:26:51 crc kubenswrapper[4988]: I1008 18:26:51.148870 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/horizon-operator-controller-manager-7ffbcb7588-jwr4r" Oct 08 18:26:51 crc kubenswrapper[4988]: I1008 18:26:51.160257 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/manila-operator-controller-manager-5f67fbc655-m7d7l" Oct 08 18:26:51 crc kubenswrapper[4988]: I1008 18:26:51.218433 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/nova-operator-controller-manager-5df598886f-8589q" Oct 08 18:26:51 crc kubenswrapper[4988]: I1008 18:26:51.270913 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/octavia-operator-controller-manager-69fdcfc5f5-zd6wx" Oct 08 18:26:51 crc kubenswrapper[4988]: I1008 18:26:51.494672 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/placement-operator-controller-manager-68b6c87b68-rf945" Oct 08 18:26:51 crc kubenswrapper[4988]: I1008 18:26:51.548486 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/swift-operator-controller-manager-db6d7f97b-mxg6q" Oct 08 18:26:52 crc kubenswrapper[4988]: I1008 18:26:52.794958 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-79db49b9fb-75xwr" event={"ID":"fc79827a-abba-4525-a705-129531293589","Type":"ContainerStarted","Data":"41447c49befda21a7c3f130aed4202bb3a5536f835f9c558c2f5c6732195c83d"} Oct 08 18:26:52 crc kubenswrapper[4988]: I1008 18:26:52.795525 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ovn-operator-controller-manager-79db49b9fb-75xwr" Oct 08 18:26:52 crc kubenswrapper[4988]: I1008 18:26:52.798035 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-79d585cb66-9cfhq" event={"ID":"4c6dc76c-9b71-45aa-b1e9-0fd9deecbd39","Type":"ContainerStarted","Data":"77ce6a3b61c4d7bca86dbf384e42e59e98709ad69a590a6ff59ab1c2b285234f"} Oct 08 18:26:52 crc kubenswrapper[4988]: I1008 18:26:52.798921 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/neutron-operator-controller-manager-79d585cb66-9cfhq" Oct 08 18:26:52 crc kubenswrapper[4988]: I1008 18:26:52.800737 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-76796d4c6b-2szmt" event={"ID":"c97bc0d9-b498-44ce-9fb5-f4cf7ca7561e","Type":"ContainerStarted","Data":"7230ca3048e24dee34196b8f02ec0b54b1f5d839eeff8efcfd1beec010537492"} Oct 08 18:26:52 crc kubenswrapper[4988]: I1008 18:26:52.800929 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/telemetry-operator-controller-manager-76796d4c6b-2szmt" Oct 08 18:26:52 crc kubenswrapper[4988]: I1008 18:26:52.803148 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-7794bc6bd-wfq6r" event={"ID":"47379f88-653e-4bda-a37b-a81857b7af36","Type":"ContainerStarted","Data":"de314e02e31b59a8b17e8bc7fd2e46226d575a866d1e9d5d681b30aa2044a3b1"} Oct 08 18:26:52 crc kubenswrapper[4988]: I1008 18:26:52.803412 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/watcher-operator-controller-manager-7794bc6bd-wfq6r" Oct 08 18:26:52 crc kubenswrapper[4988]: I1008 18:26:52.819212 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/ovn-operator-controller-manager-79db49b9fb-75xwr" podStartSLOduration=3.907599648 podStartE2EDuration="22.819192027s" podCreationTimestamp="2025-10-08 18:26:30 +0000 UTC" firstStartedPulling="2025-10-08 18:26:32.699754831 +0000 UTC m=+938.149597601" lastFinishedPulling="2025-10-08 18:26:51.6113472 +0000 UTC m=+957.061189980" observedRunningTime="2025-10-08 18:26:52.815281436 +0000 UTC m=+958.265124226" watchObservedRunningTime="2025-10-08 18:26:52.819192027 +0000 UTC m=+958.269034807" Oct 08 18:26:52 crc kubenswrapper[4988]: I1008 18:26:52.841726 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/neutron-operator-controller-manager-79d585cb66-9cfhq" podStartSLOduration=4.199456476 podStartE2EDuration="22.841704186s" podCreationTimestamp="2025-10-08 18:26:30 +0000 UTC" firstStartedPulling="2025-10-08 18:26:32.954649262 +0000 UTC m=+938.404492032" lastFinishedPulling="2025-10-08 18:26:51.596896972 +0000 UTC m=+957.046739742" observedRunningTime="2025-10-08 18:26:52.836688591 +0000 UTC m=+958.286531361" watchObservedRunningTime="2025-10-08 18:26:52.841704186 +0000 UTC m=+958.291546966" Oct 08 18:26:52 crc kubenswrapper[4988]: I1008 18:26:52.864828 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/telemetry-operator-controller-manager-76796d4c6b-2szmt" podStartSLOduration=4.184841013 podStartE2EDuration="22.864809703s" podCreationTimestamp="2025-10-08 18:26:30 +0000 UTC" firstStartedPulling="2025-10-08 18:26:32.930762701 +0000 UTC m=+938.380605471" lastFinishedPulling="2025-10-08 18:26:51.610731391 +0000 UTC m=+957.060574161" observedRunningTime="2025-10-08 18:26:52.861990056 +0000 UTC m=+958.311832826" watchObservedRunningTime="2025-10-08 18:26:52.864809703 +0000 UTC m=+958.314652493" Oct 08 18:26:52 crc kubenswrapper[4988]: I1008 18:26:52.879672 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/watcher-operator-controller-manager-7794bc6bd-wfq6r" podStartSLOduration=4.236079132 podStartE2EDuration="22.879652943s" podCreationTimestamp="2025-10-08 18:26:30 +0000 UTC" firstStartedPulling="2025-10-08 18:26:32.95684917 +0000 UTC m=+938.406691940" lastFinishedPulling="2025-10-08 18:26:51.600422981 +0000 UTC m=+957.050265751" observedRunningTime="2025-10-08 18:26:52.87823798 +0000 UTC m=+958.328080750" watchObservedRunningTime="2025-10-08 18:26:52.879652943 +0000 UTC m=+958.329495713" Oct 08 18:26:53 crc kubenswrapper[4988]: I1008 18:26:53.338548 4988 patch_prober.go:28] interesting pod/machine-config-daemon-nm54f container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 18:26:53 crc kubenswrapper[4988]: I1008 18:26:53.338602 4988 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 18:26:53 crc kubenswrapper[4988]: I1008 18:26:53.813486 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-677c5f5bff2dshw" event={"ID":"361b9b35-62d8-494a-9b8b-7d471adb246f","Type":"ContainerStarted","Data":"bd1cc94c1893533216a87e695c799d115d7827c504598e7716b4954f4659211c"} Oct 08 18:26:53 crc kubenswrapper[4988]: I1008 18:26:53.813716 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-baremetal-operator-controller-manager-677c5f5bff2dshw" Oct 08 18:26:53 crc kubenswrapper[4988]: I1008 18:26:53.815555 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-56c698c775-2p984" event={"ID":"d3afe0af-aa6b-4321-99f3-39b2f64c2451","Type":"ContainerStarted","Data":"847ef0c4a98c43e250d8c2e6d58a5a7012b873ead7982f61a4e83a16c068edaf"} Oct 08 18:26:53 crc kubenswrapper[4988]: I1008 18:26:53.858936 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-baremetal-operator-controller-manager-677c5f5bff2dshw" podStartSLOduration=3.526045837 podStartE2EDuration="23.858915287s" podCreationTimestamp="2025-10-08 18:26:30 +0000 UTC" firstStartedPulling="2025-10-08 18:26:32.927771198 +0000 UTC m=+938.377613968" lastFinishedPulling="2025-10-08 18:26:53.260640648 +0000 UTC m=+958.710483418" observedRunningTime="2025-10-08 18:26:53.855249793 +0000 UTC m=+959.305092573" watchObservedRunningTime="2025-10-08 18:26:53.858915287 +0000 UTC m=+959.308758057" Oct 08 18:26:53 crc kubenswrapper[4988]: I1008 18:26:53.871318 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/test-operator-controller-manager-56c698c775-2p984" podStartSLOduration=3.314626074 podStartE2EDuration="23.871301611s" podCreationTimestamp="2025-10-08 18:26:30 +0000 UTC" firstStartedPulling="2025-10-08 18:26:32.69909159 +0000 UTC m=+938.148934360" lastFinishedPulling="2025-10-08 18:26:53.255767127 +0000 UTC m=+958.705609897" observedRunningTime="2025-10-08 18:26:53.869842846 +0000 UTC m=+959.319685616" watchObservedRunningTime="2025-10-08 18:26:53.871301611 +0000 UTC m=+959.321144381" Oct 08 18:27:01 crc kubenswrapper[4988]: I1008 18:27:01.217031 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/neutron-operator-controller-manager-79d585cb66-9cfhq" Oct 08 18:27:01 crc kubenswrapper[4988]: I1008 18:27:01.486632 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ovn-operator-controller-manager-79db49b9fb-75xwr" Oct 08 18:27:01 crc kubenswrapper[4988]: I1008 18:27:01.590606 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/telemetry-operator-controller-manager-76796d4c6b-2szmt" Oct 08 18:27:01 crc kubenswrapper[4988]: I1008 18:27:01.621377 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/infra-operator-controller-manager-656bcbd775-2xfhc" Oct 08 18:27:01 crc kubenswrapper[4988]: I1008 18:27:01.636964 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/test-operator-controller-manager-56c698c775-2p984" Oct 08 18:27:01 crc kubenswrapper[4988]: I1008 18:27:01.638683 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/test-operator-controller-manager-56c698c775-2p984" Oct 08 18:27:01 crc kubenswrapper[4988]: I1008 18:27:01.891208 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/watcher-operator-controller-manager-7794bc6bd-wfq6r" Oct 08 18:27:01 crc kubenswrapper[4988]: I1008 18:27:01.953077 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-baremetal-operator-controller-manager-677c5f5bff2dshw" Oct 08 18:27:22 crc kubenswrapper[4988]: I1008 18:27:22.587162 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-7bfcb9d745-v56k2"] Oct 08 18:27:22 crc kubenswrapper[4988]: I1008 18:27:22.588923 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7bfcb9d745-v56k2" Oct 08 18:27:22 crc kubenswrapper[4988]: I1008 18:27:22.597781 4988 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns" Oct 08 18:27:22 crc kubenswrapper[4988]: I1008 18:27:22.597868 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dnsmasq-dns-dockercfg-5hzsq" Oct 08 18:27:22 crc kubenswrapper[4988]: I1008 18:27:22.598077 4988 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"kube-root-ca.crt" Oct 08 18:27:22 crc kubenswrapper[4988]: I1008 18:27:22.598194 4988 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openshift-service-ca.crt" Oct 08 18:27:22 crc kubenswrapper[4988]: I1008 18:27:22.609117 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7bfcb9d745-v56k2"] Oct 08 18:27:22 crc kubenswrapper[4988]: I1008 18:27:22.708370 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-758b79db4c-qv72t"] Oct 08 18:27:22 crc kubenswrapper[4988]: I1008 18:27:22.709663 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-758b79db4c-qv72t" Oct 08 18:27:22 crc kubenswrapper[4988]: I1008 18:27:22.711217 4988 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns-svc" Oct 08 18:27:22 crc kubenswrapper[4988]: I1008 18:27:22.727471 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-758b79db4c-qv72t"] Oct 08 18:27:22 crc kubenswrapper[4988]: I1008 18:27:22.774840 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6gf4v\" (UniqueName: \"kubernetes.io/projected/f0c8efc3-858a-4d1d-b031-86d9e74a3d9c-kube-api-access-6gf4v\") pod \"dnsmasq-dns-7bfcb9d745-v56k2\" (UID: \"f0c8efc3-858a-4d1d-b031-86d9e74a3d9c\") " pod="openstack/dnsmasq-dns-7bfcb9d745-v56k2" Oct 08 18:27:22 crc kubenswrapper[4988]: I1008 18:27:22.774969 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f0c8efc3-858a-4d1d-b031-86d9e74a3d9c-config\") pod \"dnsmasq-dns-7bfcb9d745-v56k2\" (UID: \"f0c8efc3-858a-4d1d-b031-86d9e74a3d9c\") " pod="openstack/dnsmasq-dns-7bfcb9d745-v56k2" Oct 08 18:27:22 crc kubenswrapper[4988]: I1008 18:27:22.876183 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a61127a0-ab84-41a9-934a-68b130206578-config\") pod \"dnsmasq-dns-758b79db4c-qv72t\" (UID: \"a61127a0-ab84-41a9-934a-68b130206578\") " pod="openstack/dnsmasq-dns-758b79db4c-qv72t" Oct 08 18:27:22 crc kubenswrapper[4988]: I1008 18:27:22.876316 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a61127a0-ab84-41a9-934a-68b130206578-dns-svc\") pod \"dnsmasq-dns-758b79db4c-qv72t\" (UID: \"a61127a0-ab84-41a9-934a-68b130206578\") " pod="openstack/dnsmasq-dns-758b79db4c-qv72t" Oct 08 18:27:22 crc kubenswrapper[4988]: I1008 18:27:22.876511 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f0c8efc3-858a-4d1d-b031-86d9e74a3d9c-config\") pod \"dnsmasq-dns-7bfcb9d745-v56k2\" (UID: \"f0c8efc3-858a-4d1d-b031-86d9e74a3d9c\") " pod="openstack/dnsmasq-dns-7bfcb9d745-v56k2" Oct 08 18:27:22 crc kubenswrapper[4988]: I1008 18:27:22.876553 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9l88k\" (UniqueName: \"kubernetes.io/projected/a61127a0-ab84-41a9-934a-68b130206578-kube-api-access-9l88k\") pod \"dnsmasq-dns-758b79db4c-qv72t\" (UID: \"a61127a0-ab84-41a9-934a-68b130206578\") " pod="openstack/dnsmasq-dns-758b79db4c-qv72t" Oct 08 18:27:22 crc kubenswrapper[4988]: I1008 18:27:22.876686 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6gf4v\" (UniqueName: \"kubernetes.io/projected/f0c8efc3-858a-4d1d-b031-86d9e74a3d9c-kube-api-access-6gf4v\") pod \"dnsmasq-dns-7bfcb9d745-v56k2\" (UID: \"f0c8efc3-858a-4d1d-b031-86d9e74a3d9c\") " pod="openstack/dnsmasq-dns-7bfcb9d745-v56k2" Oct 08 18:27:22 crc kubenswrapper[4988]: I1008 18:27:22.877867 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f0c8efc3-858a-4d1d-b031-86d9e74a3d9c-config\") pod \"dnsmasq-dns-7bfcb9d745-v56k2\" (UID: \"f0c8efc3-858a-4d1d-b031-86d9e74a3d9c\") " pod="openstack/dnsmasq-dns-7bfcb9d745-v56k2" Oct 08 18:27:22 crc kubenswrapper[4988]: I1008 18:27:22.901357 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6gf4v\" (UniqueName: \"kubernetes.io/projected/f0c8efc3-858a-4d1d-b031-86d9e74a3d9c-kube-api-access-6gf4v\") pod \"dnsmasq-dns-7bfcb9d745-v56k2\" (UID: \"f0c8efc3-858a-4d1d-b031-86d9e74a3d9c\") " pod="openstack/dnsmasq-dns-7bfcb9d745-v56k2" Oct 08 18:27:22 crc kubenswrapper[4988]: I1008 18:27:22.918054 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7bfcb9d745-v56k2" Oct 08 18:27:22 crc kubenswrapper[4988]: I1008 18:27:22.978046 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a61127a0-ab84-41a9-934a-68b130206578-config\") pod \"dnsmasq-dns-758b79db4c-qv72t\" (UID: \"a61127a0-ab84-41a9-934a-68b130206578\") " pod="openstack/dnsmasq-dns-758b79db4c-qv72t" Oct 08 18:27:22 crc kubenswrapper[4988]: I1008 18:27:22.978148 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a61127a0-ab84-41a9-934a-68b130206578-dns-svc\") pod \"dnsmasq-dns-758b79db4c-qv72t\" (UID: \"a61127a0-ab84-41a9-934a-68b130206578\") " pod="openstack/dnsmasq-dns-758b79db4c-qv72t" Oct 08 18:27:22 crc kubenswrapper[4988]: I1008 18:27:22.978220 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9l88k\" (UniqueName: \"kubernetes.io/projected/a61127a0-ab84-41a9-934a-68b130206578-kube-api-access-9l88k\") pod \"dnsmasq-dns-758b79db4c-qv72t\" (UID: \"a61127a0-ab84-41a9-934a-68b130206578\") " pod="openstack/dnsmasq-dns-758b79db4c-qv72t" Oct 08 18:27:22 crc kubenswrapper[4988]: I1008 18:27:22.981136 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a61127a0-ab84-41a9-934a-68b130206578-dns-svc\") pod \"dnsmasq-dns-758b79db4c-qv72t\" (UID: \"a61127a0-ab84-41a9-934a-68b130206578\") " pod="openstack/dnsmasq-dns-758b79db4c-qv72t" Oct 08 18:27:22 crc kubenswrapper[4988]: I1008 18:27:22.982178 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a61127a0-ab84-41a9-934a-68b130206578-config\") pod \"dnsmasq-dns-758b79db4c-qv72t\" (UID: \"a61127a0-ab84-41a9-934a-68b130206578\") " pod="openstack/dnsmasq-dns-758b79db4c-qv72t" Oct 08 18:27:23 crc kubenswrapper[4988]: I1008 18:27:23.000226 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9l88k\" (UniqueName: \"kubernetes.io/projected/a61127a0-ab84-41a9-934a-68b130206578-kube-api-access-9l88k\") pod \"dnsmasq-dns-758b79db4c-qv72t\" (UID: \"a61127a0-ab84-41a9-934a-68b130206578\") " pod="openstack/dnsmasq-dns-758b79db4c-qv72t" Oct 08 18:27:23 crc kubenswrapper[4988]: I1008 18:27:23.025610 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-758b79db4c-qv72t" Oct 08 18:27:23 crc kubenswrapper[4988]: I1008 18:27:23.215229 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-758b79db4c-qv72t"] Oct 08 18:27:23 crc kubenswrapper[4988]: I1008 18:27:23.261524 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-644597f84c-nkqvt"] Oct 08 18:27:23 crc kubenswrapper[4988]: I1008 18:27:23.262724 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-644597f84c-nkqvt" Oct 08 18:27:23 crc kubenswrapper[4988]: I1008 18:27:23.270288 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-644597f84c-nkqvt"] Oct 08 18:27:23 crc kubenswrapper[4988]: I1008 18:27:23.337882 4988 patch_prober.go:28] interesting pod/machine-config-daemon-nm54f container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 18:27:23 crc kubenswrapper[4988]: I1008 18:27:23.337954 4988 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 18:27:23 crc kubenswrapper[4988]: I1008 18:27:23.338020 4988 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" Oct 08 18:27:23 crc kubenswrapper[4988]: I1008 18:27:23.339097 4988 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"a06add767e00bcfef86e27b752f06b6fb27241526ad4e59aaac22482ba9f2df4"} pod="openshift-machine-config-operator/machine-config-daemon-nm54f" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 08 18:27:23 crc kubenswrapper[4988]: I1008 18:27:23.339176 4988 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" containerName="machine-config-daemon" containerID="cri-o://a06add767e00bcfef86e27b752f06b6fb27241526ad4e59aaac22482ba9f2df4" gracePeriod=600 Oct 08 18:27:23 crc kubenswrapper[4988]: I1008 18:27:23.388155 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cp9pt\" (UniqueName: \"kubernetes.io/projected/e04834b5-6114-4fba-b3ac-9b6405b6712f-kube-api-access-cp9pt\") pod \"dnsmasq-dns-644597f84c-nkqvt\" (UID: \"e04834b5-6114-4fba-b3ac-9b6405b6712f\") " pod="openstack/dnsmasq-dns-644597f84c-nkqvt" Oct 08 18:27:23 crc kubenswrapper[4988]: I1008 18:27:23.388203 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e04834b5-6114-4fba-b3ac-9b6405b6712f-config\") pod \"dnsmasq-dns-644597f84c-nkqvt\" (UID: \"e04834b5-6114-4fba-b3ac-9b6405b6712f\") " pod="openstack/dnsmasq-dns-644597f84c-nkqvt" Oct 08 18:27:23 crc kubenswrapper[4988]: I1008 18:27:23.388284 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e04834b5-6114-4fba-b3ac-9b6405b6712f-dns-svc\") pod \"dnsmasq-dns-644597f84c-nkqvt\" (UID: \"e04834b5-6114-4fba-b3ac-9b6405b6712f\") " pod="openstack/dnsmasq-dns-644597f84c-nkqvt" Oct 08 18:27:23 crc kubenswrapper[4988]: I1008 18:27:23.467016 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7bfcb9d745-v56k2"] Oct 08 18:27:23 crc kubenswrapper[4988]: I1008 18:27:23.491301 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e04834b5-6114-4fba-b3ac-9b6405b6712f-dns-svc\") pod \"dnsmasq-dns-644597f84c-nkqvt\" (UID: \"e04834b5-6114-4fba-b3ac-9b6405b6712f\") " pod="openstack/dnsmasq-dns-644597f84c-nkqvt" Oct 08 18:27:23 crc kubenswrapper[4988]: I1008 18:27:23.491438 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cp9pt\" (UniqueName: \"kubernetes.io/projected/e04834b5-6114-4fba-b3ac-9b6405b6712f-kube-api-access-cp9pt\") pod \"dnsmasq-dns-644597f84c-nkqvt\" (UID: \"e04834b5-6114-4fba-b3ac-9b6405b6712f\") " pod="openstack/dnsmasq-dns-644597f84c-nkqvt" Oct 08 18:27:23 crc kubenswrapper[4988]: I1008 18:27:23.491463 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e04834b5-6114-4fba-b3ac-9b6405b6712f-config\") pod \"dnsmasq-dns-644597f84c-nkqvt\" (UID: \"e04834b5-6114-4fba-b3ac-9b6405b6712f\") " pod="openstack/dnsmasq-dns-644597f84c-nkqvt" Oct 08 18:27:23 crc kubenswrapper[4988]: I1008 18:27:23.492278 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e04834b5-6114-4fba-b3ac-9b6405b6712f-config\") pod \"dnsmasq-dns-644597f84c-nkqvt\" (UID: \"e04834b5-6114-4fba-b3ac-9b6405b6712f\") " pod="openstack/dnsmasq-dns-644597f84c-nkqvt" Oct 08 18:27:23 crc kubenswrapper[4988]: I1008 18:27:23.493188 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e04834b5-6114-4fba-b3ac-9b6405b6712f-dns-svc\") pod \"dnsmasq-dns-644597f84c-nkqvt\" (UID: \"e04834b5-6114-4fba-b3ac-9b6405b6712f\") " pod="openstack/dnsmasq-dns-644597f84c-nkqvt" Oct 08 18:27:23 crc kubenswrapper[4988]: I1008 18:27:23.523220 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cp9pt\" (UniqueName: \"kubernetes.io/projected/e04834b5-6114-4fba-b3ac-9b6405b6712f-kube-api-access-cp9pt\") pod \"dnsmasq-dns-644597f84c-nkqvt\" (UID: \"e04834b5-6114-4fba-b3ac-9b6405b6712f\") " pod="openstack/dnsmasq-dns-644597f84c-nkqvt" Oct 08 18:27:23 crc kubenswrapper[4988]: I1008 18:27:23.600920 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-644597f84c-nkqvt" Oct 08 18:27:23 crc kubenswrapper[4988]: I1008 18:27:23.616241 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-758b79db4c-qv72t"] Oct 08 18:27:23 crc kubenswrapper[4988]: I1008 18:27:23.745217 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7bfcb9d745-v56k2"] Oct 08 18:27:23 crc kubenswrapper[4988]: I1008 18:27:23.765040 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-77597f887-54xgv"] Oct 08 18:27:23 crc kubenswrapper[4988]: I1008 18:27:23.766176 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-77597f887-54xgv" Oct 08 18:27:23 crc kubenswrapper[4988]: I1008 18:27:23.789955 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-77597f887-54xgv"] Oct 08 18:27:23 crc kubenswrapper[4988]: I1008 18:27:23.799423 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5d556b85-3337-4855-9f43-a38946207378-dns-svc\") pod \"dnsmasq-dns-77597f887-54xgv\" (UID: \"5d556b85-3337-4855-9f43-a38946207378\") " pod="openstack/dnsmasq-dns-77597f887-54xgv" Oct 08 18:27:23 crc kubenswrapper[4988]: I1008 18:27:23.799503 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5d556b85-3337-4855-9f43-a38946207378-config\") pod \"dnsmasq-dns-77597f887-54xgv\" (UID: \"5d556b85-3337-4855-9f43-a38946207378\") " pod="openstack/dnsmasq-dns-77597f887-54xgv" Oct 08 18:27:23 crc kubenswrapper[4988]: I1008 18:27:23.799578 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tbtxg\" (UniqueName: \"kubernetes.io/projected/5d556b85-3337-4855-9f43-a38946207378-kube-api-access-tbtxg\") pod \"dnsmasq-dns-77597f887-54xgv\" (UID: \"5d556b85-3337-4855-9f43-a38946207378\") " pod="openstack/dnsmasq-dns-77597f887-54xgv" Oct 08 18:27:23 crc kubenswrapper[4988]: I1008 18:27:23.910663 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tbtxg\" (UniqueName: \"kubernetes.io/projected/5d556b85-3337-4855-9f43-a38946207378-kube-api-access-tbtxg\") pod \"dnsmasq-dns-77597f887-54xgv\" (UID: \"5d556b85-3337-4855-9f43-a38946207378\") " pod="openstack/dnsmasq-dns-77597f887-54xgv" Oct 08 18:27:23 crc kubenswrapper[4988]: I1008 18:27:23.910845 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5d556b85-3337-4855-9f43-a38946207378-dns-svc\") pod \"dnsmasq-dns-77597f887-54xgv\" (UID: \"5d556b85-3337-4855-9f43-a38946207378\") " pod="openstack/dnsmasq-dns-77597f887-54xgv" Oct 08 18:27:23 crc kubenswrapper[4988]: I1008 18:27:23.910937 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5d556b85-3337-4855-9f43-a38946207378-config\") pod \"dnsmasq-dns-77597f887-54xgv\" (UID: \"5d556b85-3337-4855-9f43-a38946207378\") " pod="openstack/dnsmasq-dns-77597f887-54xgv" Oct 08 18:27:23 crc kubenswrapper[4988]: I1008 18:27:23.911893 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5d556b85-3337-4855-9f43-a38946207378-config\") pod \"dnsmasq-dns-77597f887-54xgv\" (UID: \"5d556b85-3337-4855-9f43-a38946207378\") " pod="openstack/dnsmasq-dns-77597f887-54xgv" Oct 08 18:27:23 crc kubenswrapper[4988]: I1008 18:27:23.912378 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5d556b85-3337-4855-9f43-a38946207378-dns-svc\") pod \"dnsmasq-dns-77597f887-54xgv\" (UID: \"5d556b85-3337-4855-9f43-a38946207378\") " pod="openstack/dnsmasq-dns-77597f887-54xgv" Oct 08 18:27:23 crc kubenswrapper[4988]: I1008 18:27:23.929669 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tbtxg\" (UniqueName: \"kubernetes.io/projected/5d556b85-3337-4855-9f43-a38946207378-kube-api-access-tbtxg\") pod \"dnsmasq-dns-77597f887-54xgv\" (UID: \"5d556b85-3337-4855-9f43-a38946207378\") " pod="openstack/dnsmasq-dns-77597f887-54xgv" Oct 08 18:27:24 crc kubenswrapper[4988]: I1008 18:27:24.056210 4988 generic.go:334] "Generic (PLEG): container finished" podID="f328a20b-5283-407e-8a93-2ab2da599f18" containerID="a06add767e00bcfef86e27b752f06b6fb27241526ad4e59aaac22482ba9f2df4" exitCode=0 Oct 08 18:27:24 crc kubenswrapper[4988]: I1008 18:27:24.056281 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" event={"ID":"f328a20b-5283-407e-8a93-2ab2da599f18","Type":"ContainerDied","Data":"a06add767e00bcfef86e27b752f06b6fb27241526ad4e59aaac22482ba9f2df4"} Oct 08 18:27:24 crc kubenswrapper[4988]: I1008 18:27:24.056336 4988 scope.go:117] "RemoveContainer" containerID="0b7eb1da8fbdc295b99305bf18316c88ee5b435b3bee717d217e3efa0977c5a9" Oct 08 18:27:24 crc kubenswrapper[4988]: I1008 18:27:24.057641 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7bfcb9d745-v56k2" event={"ID":"f0c8efc3-858a-4d1d-b031-86d9e74a3d9c","Type":"ContainerStarted","Data":"bee727503b080650e80cd50d20b61054f08be6c7aa34fae491b3f54e202f56c4"} Oct 08 18:27:24 crc kubenswrapper[4988]: I1008 18:27:24.060122 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-758b79db4c-qv72t" event={"ID":"a61127a0-ab84-41a9-934a-68b130206578","Type":"ContainerStarted","Data":"4083d3f29623ad81e469ab7e4ac32353d8400e6a5e9903c66b25644d4dc0fbbc"} Oct 08 18:27:24 crc kubenswrapper[4988]: I1008 18:27:24.091403 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-77597f887-54xgv" Oct 08 18:27:24 crc kubenswrapper[4988]: I1008 18:27:24.136256 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-644597f84c-nkqvt"] Oct 08 18:27:24 crc kubenswrapper[4988]: W1008 18:27:24.145531 4988 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode04834b5_6114_4fba_b3ac_9b6405b6712f.slice/crio-1c0fe4bbc8a31a3adb33bbca81bb3d0d0b8154f1b6b2b87402836d7b6b9dfcf7 WatchSource:0}: Error finding container 1c0fe4bbc8a31a3adb33bbca81bb3d0d0b8154f1b6b2b87402836d7b6b9dfcf7: Status 404 returned error can't find the container with id 1c0fe4bbc8a31a3adb33bbca81bb3d0d0b8154f1b6b2b87402836d7b6b9dfcf7 Oct 08 18:27:24 crc kubenswrapper[4988]: I1008 18:27:24.378637 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-77597f887-54xgv"] Oct 08 18:27:24 crc kubenswrapper[4988]: I1008 18:27:24.411542 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-0"] Oct 08 18:27:24 crc kubenswrapper[4988]: I1008 18:27:24.412884 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Oct 08 18:27:24 crc kubenswrapper[4988]: I1008 18:27:24.417414 4988 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-server-conf" Oct 08 18:27:24 crc kubenswrapper[4988]: I1008 18:27:24.417638 4988 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-plugins-conf" Oct 08 18:27:24 crc kubenswrapper[4988]: I1008 18:27:24.417921 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-svc" Oct 08 18:27:24 crc kubenswrapper[4988]: I1008 18:27:24.418136 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-erlang-cookie" Oct 08 18:27:24 crc kubenswrapper[4988]: I1008 18:27:24.418328 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-server-dockercfg-7c5gq" Oct 08 18:27:24 crc kubenswrapper[4988]: I1008 18:27:24.418554 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-default-user" Oct 08 18:27:24 crc kubenswrapper[4988]: I1008 18:27:24.418749 4988 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-config-data" Oct 08 18:27:24 crc kubenswrapper[4988]: I1008 18:27:24.424534 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 08 18:27:24 crc kubenswrapper[4988]: I1008 18:27:24.525016 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/830d6ec5-0015-4a88-b8e8-0fd47f64fd46-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"830d6ec5-0015-4a88-b8e8-0fd47f64fd46\") " pod="openstack/rabbitmq-server-0" Oct 08 18:27:24 crc kubenswrapper[4988]: I1008 18:27:24.525099 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/830d6ec5-0015-4a88-b8e8-0fd47f64fd46-pod-info\") pod \"rabbitmq-server-0\" (UID: \"830d6ec5-0015-4a88-b8e8-0fd47f64fd46\") " pod="openstack/rabbitmq-server-0" Oct 08 18:27:24 crc kubenswrapper[4988]: I1008 18:27:24.525226 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/830d6ec5-0015-4a88-b8e8-0fd47f64fd46-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"830d6ec5-0015-4a88-b8e8-0fd47f64fd46\") " pod="openstack/rabbitmq-server-0" Oct 08 18:27:24 crc kubenswrapper[4988]: I1008 18:27:24.525274 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/830d6ec5-0015-4a88-b8e8-0fd47f64fd46-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"830d6ec5-0015-4a88-b8e8-0fd47f64fd46\") " pod="openstack/rabbitmq-server-0" Oct 08 18:27:24 crc kubenswrapper[4988]: I1008 18:27:24.525353 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/830d6ec5-0015-4a88-b8e8-0fd47f64fd46-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"830d6ec5-0015-4a88-b8e8-0fd47f64fd46\") " pod="openstack/rabbitmq-server-0" Oct 08 18:27:24 crc kubenswrapper[4988]: I1008 18:27:24.525455 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/830d6ec5-0015-4a88-b8e8-0fd47f64fd46-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"830d6ec5-0015-4a88-b8e8-0fd47f64fd46\") " pod="openstack/rabbitmq-server-0" Oct 08 18:27:24 crc kubenswrapper[4988]: I1008 18:27:24.525560 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/830d6ec5-0015-4a88-b8e8-0fd47f64fd46-server-conf\") pod \"rabbitmq-server-0\" (UID: \"830d6ec5-0015-4a88-b8e8-0fd47f64fd46\") " pod="openstack/rabbitmq-server-0" Oct 08 18:27:24 crc kubenswrapper[4988]: I1008 18:27:24.525597 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"rabbitmq-server-0\" (UID: \"830d6ec5-0015-4a88-b8e8-0fd47f64fd46\") " pod="openstack/rabbitmq-server-0" Oct 08 18:27:24 crc kubenswrapper[4988]: I1008 18:27:24.525623 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/830d6ec5-0015-4a88-b8e8-0fd47f64fd46-config-data\") pod \"rabbitmq-server-0\" (UID: \"830d6ec5-0015-4a88-b8e8-0fd47f64fd46\") " pod="openstack/rabbitmq-server-0" Oct 08 18:27:24 crc kubenswrapper[4988]: I1008 18:27:24.525735 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tmhl7\" (UniqueName: \"kubernetes.io/projected/830d6ec5-0015-4a88-b8e8-0fd47f64fd46-kube-api-access-tmhl7\") pod \"rabbitmq-server-0\" (UID: \"830d6ec5-0015-4a88-b8e8-0fd47f64fd46\") " pod="openstack/rabbitmq-server-0" Oct 08 18:27:24 crc kubenswrapper[4988]: I1008 18:27:24.525828 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/830d6ec5-0015-4a88-b8e8-0fd47f64fd46-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"830d6ec5-0015-4a88-b8e8-0fd47f64fd46\") " pod="openstack/rabbitmq-server-0" Oct 08 18:27:24 crc kubenswrapper[4988]: I1008 18:27:24.627799 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tmhl7\" (UniqueName: \"kubernetes.io/projected/830d6ec5-0015-4a88-b8e8-0fd47f64fd46-kube-api-access-tmhl7\") pod \"rabbitmq-server-0\" (UID: \"830d6ec5-0015-4a88-b8e8-0fd47f64fd46\") " pod="openstack/rabbitmq-server-0" Oct 08 18:27:24 crc kubenswrapper[4988]: I1008 18:27:24.627879 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/830d6ec5-0015-4a88-b8e8-0fd47f64fd46-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"830d6ec5-0015-4a88-b8e8-0fd47f64fd46\") " pod="openstack/rabbitmq-server-0" Oct 08 18:27:24 crc kubenswrapper[4988]: I1008 18:27:24.627914 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/830d6ec5-0015-4a88-b8e8-0fd47f64fd46-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"830d6ec5-0015-4a88-b8e8-0fd47f64fd46\") " pod="openstack/rabbitmq-server-0" Oct 08 18:27:24 crc kubenswrapper[4988]: I1008 18:27:24.627938 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/830d6ec5-0015-4a88-b8e8-0fd47f64fd46-pod-info\") pod \"rabbitmq-server-0\" (UID: \"830d6ec5-0015-4a88-b8e8-0fd47f64fd46\") " pod="openstack/rabbitmq-server-0" Oct 08 18:27:24 crc kubenswrapper[4988]: I1008 18:27:24.627969 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/830d6ec5-0015-4a88-b8e8-0fd47f64fd46-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"830d6ec5-0015-4a88-b8e8-0fd47f64fd46\") " pod="openstack/rabbitmq-server-0" Oct 08 18:27:24 crc kubenswrapper[4988]: I1008 18:27:24.627991 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/830d6ec5-0015-4a88-b8e8-0fd47f64fd46-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"830d6ec5-0015-4a88-b8e8-0fd47f64fd46\") " pod="openstack/rabbitmq-server-0" Oct 08 18:27:24 crc kubenswrapper[4988]: I1008 18:27:24.628032 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/830d6ec5-0015-4a88-b8e8-0fd47f64fd46-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"830d6ec5-0015-4a88-b8e8-0fd47f64fd46\") " pod="openstack/rabbitmq-server-0" Oct 08 18:27:24 crc kubenswrapper[4988]: I1008 18:27:24.628056 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/830d6ec5-0015-4a88-b8e8-0fd47f64fd46-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"830d6ec5-0015-4a88-b8e8-0fd47f64fd46\") " pod="openstack/rabbitmq-server-0" Oct 08 18:27:24 crc kubenswrapper[4988]: I1008 18:27:24.628092 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/830d6ec5-0015-4a88-b8e8-0fd47f64fd46-server-conf\") pod \"rabbitmq-server-0\" (UID: \"830d6ec5-0015-4a88-b8e8-0fd47f64fd46\") " pod="openstack/rabbitmq-server-0" Oct 08 18:27:24 crc kubenswrapper[4988]: I1008 18:27:24.628117 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"rabbitmq-server-0\" (UID: \"830d6ec5-0015-4a88-b8e8-0fd47f64fd46\") " pod="openstack/rabbitmq-server-0" Oct 08 18:27:24 crc kubenswrapper[4988]: I1008 18:27:24.628141 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/830d6ec5-0015-4a88-b8e8-0fd47f64fd46-config-data\") pod \"rabbitmq-server-0\" (UID: \"830d6ec5-0015-4a88-b8e8-0fd47f64fd46\") " pod="openstack/rabbitmq-server-0" Oct 08 18:27:24 crc kubenswrapper[4988]: I1008 18:27:24.629256 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/830d6ec5-0015-4a88-b8e8-0fd47f64fd46-config-data\") pod \"rabbitmq-server-0\" (UID: \"830d6ec5-0015-4a88-b8e8-0fd47f64fd46\") " pod="openstack/rabbitmq-server-0" Oct 08 18:27:24 crc kubenswrapper[4988]: I1008 18:27:24.629338 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/830d6ec5-0015-4a88-b8e8-0fd47f64fd46-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"830d6ec5-0015-4a88-b8e8-0fd47f64fd46\") " pod="openstack/rabbitmq-server-0" Oct 08 18:27:24 crc kubenswrapper[4988]: I1008 18:27:24.630752 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/830d6ec5-0015-4a88-b8e8-0fd47f64fd46-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"830d6ec5-0015-4a88-b8e8-0fd47f64fd46\") " pod="openstack/rabbitmq-server-0" Oct 08 18:27:24 crc kubenswrapper[4988]: I1008 18:27:24.631128 4988 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"rabbitmq-server-0\" (UID: \"830d6ec5-0015-4a88-b8e8-0fd47f64fd46\") device mount path \"/mnt/openstack/pv04\"" pod="openstack/rabbitmq-server-0" Oct 08 18:27:24 crc kubenswrapper[4988]: I1008 18:27:24.631249 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/830d6ec5-0015-4a88-b8e8-0fd47f64fd46-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"830d6ec5-0015-4a88-b8e8-0fd47f64fd46\") " pod="openstack/rabbitmq-server-0" Oct 08 18:27:24 crc kubenswrapper[4988]: I1008 18:27:24.632545 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/830d6ec5-0015-4a88-b8e8-0fd47f64fd46-server-conf\") pod \"rabbitmq-server-0\" (UID: \"830d6ec5-0015-4a88-b8e8-0fd47f64fd46\") " pod="openstack/rabbitmq-server-0" Oct 08 18:27:24 crc kubenswrapper[4988]: I1008 18:27:24.635619 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/830d6ec5-0015-4a88-b8e8-0fd47f64fd46-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"830d6ec5-0015-4a88-b8e8-0fd47f64fd46\") " pod="openstack/rabbitmq-server-0" Oct 08 18:27:24 crc kubenswrapper[4988]: I1008 18:27:24.640743 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/830d6ec5-0015-4a88-b8e8-0fd47f64fd46-pod-info\") pod \"rabbitmq-server-0\" (UID: \"830d6ec5-0015-4a88-b8e8-0fd47f64fd46\") " pod="openstack/rabbitmq-server-0" Oct 08 18:27:24 crc kubenswrapper[4988]: I1008 18:27:24.646802 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/830d6ec5-0015-4a88-b8e8-0fd47f64fd46-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"830d6ec5-0015-4a88-b8e8-0fd47f64fd46\") " pod="openstack/rabbitmq-server-0" Oct 08 18:27:24 crc kubenswrapper[4988]: I1008 18:27:24.648972 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/830d6ec5-0015-4a88-b8e8-0fd47f64fd46-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"830d6ec5-0015-4a88-b8e8-0fd47f64fd46\") " pod="openstack/rabbitmq-server-0" Oct 08 18:27:24 crc kubenswrapper[4988]: I1008 18:27:24.656946 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tmhl7\" (UniqueName: \"kubernetes.io/projected/830d6ec5-0015-4a88-b8e8-0fd47f64fd46-kube-api-access-tmhl7\") pod \"rabbitmq-server-0\" (UID: \"830d6ec5-0015-4a88-b8e8-0fd47f64fd46\") " pod="openstack/rabbitmq-server-0" Oct 08 18:27:24 crc kubenswrapper[4988]: I1008 18:27:24.662332 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"rabbitmq-server-0\" (UID: \"830d6ec5-0015-4a88-b8e8-0fd47f64fd46\") " pod="openstack/rabbitmq-server-0" Oct 08 18:27:24 crc kubenswrapper[4988]: I1008 18:27:24.765148 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Oct 08 18:27:24 crc kubenswrapper[4988]: I1008 18:27:24.923105 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 08 18:27:24 crc kubenswrapper[4988]: I1008 18:27:24.924597 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Oct 08 18:27:24 crc kubenswrapper[4988]: I1008 18:27:24.927133 4988 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-plugins-conf" Oct 08 18:27:24 crc kubenswrapper[4988]: I1008 18:27:24.927308 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-erlang-cookie" Oct 08 18:27:24 crc kubenswrapper[4988]: I1008 18:27:24.931213 4988 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-config-data" Oct 08 18:27:24 crc kubenswrapper[4988]: I1008 18:27:24.931363 4988 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-server-conf" Oct 08 18:27:24 crc kubenswrapper[4988]: I1008 18:27:24.931711 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-default-user" Oct 08 18:27:24 crc kubenswrapper[4988]: I1008 18:27:24.931901 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-server-dockercfg-mgktn" Oct 08 18:27:24 crc kubenswrapper[4988]: I1008 18:27:24.932037 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-cell1-svc" Oct 08 18:27:24 crc kubenswrapper[4988]: I1008 18:27:24.946812 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 08 18:27:25 crc kubenswrapper[4988]: I1008 18:27:25.033123 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/e5a8f819-c91d-429e-b848-e9c444739623-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"e5a8f819-c91d-429e-b848-e9c444739623\") " pod="openstack/rabbitmq-cell1-server-0" Oct 08 18:27:25 crc kubenswrapper[4988]: I1008 18:27:25.033236 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/e5a8f819-c91d-429e-b848-e9c444739623-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"e5a8f819-c91d-429e-b848-e9c444739623\") " pod="openstack/rabbitmq-cell1-server-0" Oct 08 18:27:25 crc kubenswrapper[4988]: I1008 18:27:25.033263 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/e5a8f819-c91d-429e-b848-e9c444739623-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"e5a8f819-c91d-429e-b848-e9c444739623\") " pod="openstack/rabbitmq-cell1-server-0" Oct 08 18:27:25 crc kubenswrapper[4988]: I1008 18:27:25.033280 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zv5wc\" (UniqueName: \"kubernetes.io/projected/e5a8f819-c91d-429e-b848-e9c444739623-kube-api-access-zv5wc\") pod \"rabbitmq-cell1-server-0\" (UID: \"e5a8f819-c91d-429e-b848-e9c444739623\") " pod="openstack/rabbitmq-cell1-server-0" Oct 08 18:27:25 crc kubenswrapper[4988]: I1008 18:27:25.033311 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/e5a8f819-c91d-429e-b848-e9c444739623-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"e5a8f819-c91d-429e-b848-e9c444739623\") " pod="openstack/rabbitmq-cell1-server-0" Oct 08 18:27:25 crc kubenswrapper[4988]: I1008 18:27:25.033560 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/e5a8f819-c91d-429e-b848-e9c444739623-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"e5a8f819-c91d-429e-b848-e9c444739623\") " pod="openstack/rabbitmq-cell1-server-0" Oct 08 18:27:25 crc kubenswrapper[4988]: I1008 18:27:25.033606 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/e5a8f819-c91d-429e-b848-e9c444739623-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"e5a8f819-c91d-429e-b848-e9c444739623\") " pod="openstack/rabbitmq-cell1-server-0" Oct 08 18:27:25 crc kubenswrapper[4988]: I1008 18:27:25.033701 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/e5a8f819-c91d-429e-b848-e9c444739623-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"e5a8f819-c91d-429e-b848-e9c444739623\") " pod="openstack/rabbitmq-cell1-server-0" Oct 08 18:27:25 crc kubenswrapper[4988]: I1008 18:27:25.033854 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/e5a8f819-c91d-429e-b848-e9c444739623-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"e5a8f819-c91d-429e-b848-e9c444739623\") " pod="openstack/rabbitmq-cell1-server-0" Oct 08 18:27:25 crc kubenswrapper[4988]: I1008 18:27:25.033896 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"e5a8f819-c91d-429e-b848-e9c444739623\") " pod="openstack/rabbitmq-cell1-server-0" Oct 08 18:27:25 crc kubenswrapper[4988]: I1008 18:27:25.033916 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/e5a8f819-c91d-429e-b848-e9c444739623-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"e5a8f819-c91d-429e-b848-e9c444739623\") " pod="openstack/rabbitmq-cell1-server-0" Oct 08 18:27:25 crc kubenswrapper[4988]: I1008 18:27:25.109707 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" event={"ID":"f328a20b-5283-407e-8a93-2ab2da599f18","Type":"ContainerStarted","Data":"33dcd4d530d69056c83e9e7f7862766d54a3df5b03043465d12788d2876eac58"} Oct 08 18:27:25 crc kubenswrapper[4988]: I1008 18:27:25.126712 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-77597f887-54xgv" event={"ID":"5d556b85-3337-4855-9f43-a38946207378","Type":"ContainerStarted","Data":"18c621fd710f1a97e662af27846b796420779e3bb2993ed758cfedb273d824d1"} Oct 08 18:27:25 crc kubenswrapper[4988]: I1008 18:27:25.140765 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/e5a8f819-c91d-429e-b848-e9c444739623-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"e5a8f819-c91d-429e-b848-e9c444739623\") " pod="openstack/rabbitmq-cell1-server-0" Oct 08 18:27:25 crc kubenswrapper[4988]: I1008 18:27:25.140859 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zv5wc\" (UniqueName: \"kubernetes.io/projected/e5a8f819-c91d-429e-b848-e9c444739623-kube-api-access-zv5wc\") pod \"rabbitmq-cell1-server-0\" (UID: \"e5a8f819-c91d-429e-b848-e9c444739623\") " pod="openstack/rabbitmq-cell1-server-0" Oct 08 18:27:25 crc kubenswrapper[4988]: I1008 18:27:25.141106 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-644597f84c-nkqvt" event={"ID":"e04834b5-6114-4fba-b3ac-9b6405b6712f","Type":"ContainerStarted","Data":"1c0fe4bbc8a31a3adb33bbca81bb3d0d0b8154f1b6b2b87402836d7b6b9dfcf7"} Oct 08 18:27:25 crc kubenswrapper[4988]: I1008 18:27:25.141137 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/e5a8f819-c91d-429e-b848-e9c444739623-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"e5a8f819-c91d-429e-b848-e9c444739623\") " pod="openstack/rabbitmq-cell1-server-0" Oct 08 18:27:25 crc kubenswrapper[4988]: I1008 18:27:25.141201 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/e5a8f819-c91d-429e-b848-e9c444739623-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"e5a8f819-c91d-429e-b848-e9c444739623\") " pod="openstack/rabbitmq-cell1-server-0" Oct 08 18:27:25 crc kubenswrapper[4988]: I1008 18:27:25.141222 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/e5a8f819-c91d-429e-b848-e9c444739623-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"e5a8f819-c91d-429e-b848-e9c444739623\") " pod="openstack/rabbitmq-cell1-server-0" Oct 08 18:27:25 crc kubenswrapper[4988]: I1008 18:27:25.141255 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/e5a8f819-c91d-429e-b848-e9c444739623-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"e5a8f819-c91d-429e-b848-e9c444739623\") " pod="openstack/rabbitmq-cell1-server-0" Oct 08 18:27:25 crc kubenswrapper[4988]: I1008 18:27:25.141360 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/e5a8f819-c91d-429e-b848-e9c444739623-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"e5a8f819-c91d-429e-b848-e9c444739623\") " pod="openstack/rabbitmq-cell1-server-0" Oct 08 18:27:25 crc kubenswrapper[4988]: I1008 18:27:25.141451 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"e5a8f819-c91d-429e-b848-e9c444739623\") " pod="openstack/rabbitmq-cell1-server-0" Oct 08 18:27:25 crc kubenswrapper[4988]: I1008 18:27:25.141468 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/e5a8f819-c91d-429e-b848-e9c444739623-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"e5a8f819-c91d-429e-b848-e9c444739623\") " pod="openstack/rabbitmq-cell1-server-0" Oct 08 18:27:25 crc kubenswrapper[4988]: I1008 18:27:25.141491 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/e5a8f819-c91d-429e-b848-e9c444739623-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"e5a8f819-c91d-429e-b848-e9c444739623\") " pod="openstack/rabbitmq-cell1-server-0" Oct 08 18:27:25 crc kubenswrapper[4988]: I1008 18:27:25.141538 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/e5a8f819-c91d-429e-b848-e9c444739623-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"e5a8f819-c91d-429e-b848-e9c444739623\") " pod="openstack/rabbitmq-cell1-server-0" Oct 08 18:27:25 crc kubenswrapper[4988]: I1008 18:27:25.143427 4988 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"e5a8f819-c91d-429e-b848-e9c444739623\") device mount path \"/mnt/openstack/pv07\"" pod="openstack/rabbitmq-cell1-server-0" Oct 08 18:27:25 crc kubenswrapper[4988]: I1008 18:27:25.145054 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/e5a8f819-c91d-429e-b848-e9c444739623-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"e5a8f819-c91d-429e-b848-e9c444739623\") " pod="openstack/rabbitmq-cell1-server-0" Oct 08 18:27:25 crc kubenswrapper[4988]: I1008 18:27:25.145247 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/e5a8f819-c91d-429e-b848-e9c444739623-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"e5a8f819-c91d-429e-b848-e9c444739623\") " pod="openstack/rabbitmq-cell1-server-0" Oct 08 18:27:25 crc kubenswrapper[4988]: I1008 18:27:25.145441 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/e5a8f819-c91d-429e-b848-e9c444739623-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"e5a8f819-c91d-429e-b848-e9c444739623\") " pod="openstack/rabbitmq-cell1-server-0" Oct 08 18:27:25 crc kubenswrapper[4988]: I1008 18:27:25.146205 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/e5a8f819-c91d-429e-b848-e9c444739623-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"e5a8f819-c91d-429e-b848-e9c444739623\") " pod="openstack/rabbitmq-cell1-server-0" Oct 08 18:27:25 crc kubenswrapper[4988]: I1008 18:27:25.148895 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/e5a8f819-c91d-429e-b848-e9c444739623-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"e5a8f819-c91d-429e-b848-e9c444739623\") " pod="openstack/rabbitmq-cell1-server-0" Oct 08 18:27:25 crc kubenswrapper[4988]: I1008 18:27:25.158450 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/e5a8f819-c91d-429e-b848-e9c444739623-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"e5a8f819-c91d-429e-b848-e9c444739623\") " pod="openstack/rabbitmq-cell1-server-0" Oct 08 18:27:25 crc kubenswrapper[4988]: I1008 18:27:25.158957 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/e5a8f819-c91d-429e-b848-e9c444739623-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"e5a8f819-c91d-429e-b848-e9c444739623\") " pod="openstack/rabbitmq-cell1-server-0" Oct 08 18:27:25 crc kubenswrapper[4988]: I1008 18:27:25.159306 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/e5a8f819-c91d-429e-b848-e9c444739623-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"e5a8f819-c91d-429e-b848-e9c444739623\") " pod="openstack/rabbitmq-cell1-server-0" Oct 08 18:27:25 crc kubenswrapper[4988]: I1008 18:27:25.164357 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/e5a8f819-c91d-429e-b848-e9c444739623-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"e5a8f819-c91d-429e-b848-e9c444739623\") " pod="openstack/rabbitmq-cell1-server-0" Oct 08 18:27:25 crc kubenswrapper[4988]: I1008 18:27:25.171833 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zv5wc\" (UniqueName: \"kubernetes.io/projected/e5a8f819-c91d-429e-b848-e9c444739623-kube-api-access-zv5wc\") pod \"rabbitmq-cell1-server-0\" (UID: \"e5a8f819-c91d-429e-b848-e9c444739623\") " pod="openstack/rabbitmq-cell1-server-0" Oct 08 18:27:25 crc kubenswrapper[4988]: I1008 18:27:25.174080 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"e5a8f819-c91d-429e-b848-e9c444739623\") " pod="openstack/rabbitmq-cell1-server-0" Oct 08 18:27:25 crc kubenswrapper[4988]: I1008 18:27:25.318631 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Oct 08 18:27:25 crc kubenswrapper[4988]: W1008 18:27:25.343686 4988 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod830d6ec5_0015_4a88_b8e8_0fd47f64fd46.slice/crio-a0bb4c09203e8a478b42ecb12f92dc6c845688e189426130aedb932153d1239c WatchSource:0}: Error finding container a0bb4c09203e8a478b42ecb12f92dc6c845688e189426130aedb932153d1239c: Status 404 returned error can't find the container with id a0bb4c09203e8a478b42ecb12f92dc6c845688e189426130aedb932153d1239c Oct 08 18:27:25 crc kubenswrapper[4988]: I1008 18:27:25.366032 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 08 18:27:25 crc kubenswrapper[4988]: I1008 18:27:25.831711 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 08 18:27:25 crc kubenswrapper[4988]: W1008 18:27:25.895758 4988 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode5a8f819_c91d_429e_b848_e9c444739623.slice/crio-2a6845add4d60a97c03ae806a22c38e8be0924372a13cb9a5e2ec5177a1ec2dc WatchSource:0}: Error finding container 2a6845add4d60a97c03ae806a22c38e8be0924372a13cb9a5e2ec5177a1ec2dc: Status 404 returned error can't find the container with id 2a6845add4d60a97c03ae806a22c38e8be0924372a13cb9a5e2ec5177a1ec2dc Oct 08 18:27:25 crc kubenswrapper[4988]: I1008 18:27:25.927728 4988 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 08 18:27:26 crc kubenswrapper[4988]: I1008 18:27:26.158545 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"e5a8f819-c91d-429e-b848-e9c444739623","Type":"ContainerStarted","Data":"2a6845add4d60a97c03ae806a22c38e8be0924372a13cb9a5e2ec5177a1ec2dc"} Oct 08 18:27:26 crc kubenswrapper[4988]: I1008 18:27:26.164483 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"830d6ec5-0015-4a88-b8e8-0fd47f64fd46","Type":"ContainerStarted","Data":"a0bb4c09203e8a478b42ecb12f92dc6c845688e189426130aedb932153d1239c"} Oct 08 18:27:27 crc kubenswrapper[4988]: I1008 18:27:27.185415 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-galera-0"] Oct 08 18:27:27 crc kubenswrapper[4988]: I1008 18:27:27.187050 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Oct 08 18:27:27 crc kubenswrapper[4988]: I1008 18:27:27.194083 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-svc" Oct 08 18:27:27 crc kubenswrapper[4988]: I1008 18:27:27.194133 4988 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config-data" Oct 08 18:27:27 crc kubenswrapper[4988]: I1008 18:27:27.194100 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Oct 08 18:27:27 crc kubenswrapper[4988]: I1008 18:27:27.194254 4988 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-scripts" Oct 08 18:27:27 crc kubenswrapper[4988]: I1008 18:27:27.194335 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-dockercfg-f6chc" Oct 08 18:27:27 crc kubenswrapper[4988]: I1008 18:27:27.207681 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"combined-ca-bundle" Oct 08 18:27:27 crc kubenswrapper[4988]: I1008 18:27:27.222132 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Oct 08 18:27:27 crc kubenswrapper[4988]: I1008 18:27:27.282727 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/72c64965-0b57-43e1-8d18-2bce6cabbfa1-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"72c64965-0b57-43e1-8d18-2bce6cabbfa1\") " pod="openstack/openstack-galera-0" Oct 08 18:27:27 crc kubenswrapper[4988]: I1008 18:27:27.282780 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/72c64965-0b57-43e1-8d18-2bce6cabbfa1-secrets\") pod \"openstack-galera-0\" (UID: \"72c64965-0b57-43e1-8d18-2bce6cabbfa1\") " pod="openstack/openstack-galera-0" Oct 08 18:27:27 crc kubenswrapper[4988]: I1008 18:27:27.282867 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"openstack-galera-0\" (UID: \"72c64965-0b57-43e1-8d18-2bce6cabbfa1\") " pod="openstack/openstack-galera-0" Oct 08 18:27:27 crc kubenswrapper[4988]: I1008 18:27:27.282935 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/72c64965-0b57-43e1-8d18-2bce6cabbfa1-config-data-generated\") pod \"openstack-galera-0\" (UID: \"72c64965-0b57-43e1-8d18-2bce6cabbfa1\") " pod="openstack/openstack-galera-0" Oct 08 18:27:27 crc kubenswrapper[4988]: I1008 18:27:27.282973 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qszql\" (UniqueName: \"kubernetes.io/projected/72c64965-0b57-43e1-8d18-2bce6cabbfa1-kube-api-access-qszql\") pod \"openstack-galera-0\" (UID: \"72c64965-0b57-43e1-8d18-2bce6cabbfa1\") " pod="openstack/openstack-galera-0" Oct 08 18:27:27 crc kubenswrapper[4988]: I1008 18:27:27.283002 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/72c64965-0b57-43e1-8d18-2bce6cabbfa1-kolla-config\") pod \"openstack-galera-0\" (UID: \"72c64965-0b57-43e1-8d18-2bce6cabbfa1\") " pod="openstack/openstack-galera-0" Oct 08 18:27:27 crc kubenswrapper[4988]: I1008 18:27:27.283115 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/72c64965-0b57-43e1-8d18-2bce6cabbfa1-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"72c64965-0b57-43e1-8d18-2bce6cabbfa1\") " pod="openstack/openstack-galera-0" Oct 08 18:27:27 crc kubenswrapper[4988]: I1008 18:27:27.283190 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/72c64965-0b57-43e1-8d18-2bce6cabbfa1-config-data-default\") pod \"openstack-galera-0\" (UID: \"72c64965-0b57-43e1-8d18-2bce6cabbfa1\") " pod="openstack/openstack-galera-0" Oct 08 18:27:27 crc kubenswrapper[4988]: I1008 18:27:27.283233 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/72c64965-0b57-43e1-8d18-2bce6cabbfa1-operator-scripts\") pod \"openstack-galera-0\" (UID: \"72c64965-0b57-43e1-8d18-2bce6cabbfa1\") " pod="openstack/openstack-galera-0" Oct 08 18:27:27 crc kubenswrapper[4988]: I1008 18:27:27.384444 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/72c64965-0b57-43e1-8d18-2bce6cabbfa1-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"72c64965-0b57-43e1-8d18-2bce6cabbfa1\") " pod="openstack/openstack-galera-0" Oct 08 18:27:27 crc kubenswrapper[4988]: I1008 18:27:27.384487 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/72c64965-0b57-43e1-8d18-2bce6cabbfa1-config-data-default\") pod \"openstack-galera-0\" (UID: \"72c64965-0b57-43e1-8d18-2bce6cabbfa1\") " pod="openstack/openstack-galera-0" Oct 08 18:27:27 crc kubenswrapper[4988]: I1008 18:27:27.384509 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/72c64965-0b57-43e1-8d18-2bce6cabbfa1-operator-scripts\") pod \"openstack-galera-0\" (UID: \"72c64965-0b57-43e1-8d18-2bce6cabbfa1\") " pod="openstack/openstack-galera-0" Oct 08 18:27:27 crc kubenswrapper[4988]: I1008 18:27:27.384550 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/72c64965-0b57-43e1-8d18-2bce6cabbfa1-secrets\") pod \"openstack-galera-0\" (UID: \"72c64965-0b57-43e1-8d18-2bce6cabbfa1\") " pod="openstack/openstack-galera-0" Oct 08 18:27:27 crc kubenswrapper[4988]: I1008 18:27:27.384565 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/72c64965-0b57-43e1-8d18-2bce6cabbfa1-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"72c64965-0b57-43e1-8d18-2bce6cabbfa1\") " pod="openstack/openstack-galera-0" Oct 08 18:27:27 crc kubenswrapper[4988]: I1008 18:27:27.384623 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"openstack-galera-0\" (UID: \"72c64965-0b57-43e1-8d18-2bce6cabbfa1\") " pod="openstack/openstack-galera-0" Oct 08 18:27:27 crc kubenswrapper[4988]: I1008 18:27:27.384664 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/72c64965-0b57-43e1-8d18-2bce6cabbfa1-config-data-generated\") pod \"openstack-galera-0\" (UID: \"72c64965-0b57-43e1-8d18-2bce6cabbfa1\") " pod="openstack/openstack-galera-0" Oct 08 18:27:27 crc kubenswrapper[4988]: I1008 18:27:27.384687 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qszql\" (UniqueName: \"kubernetes.io/projected/72c64965-0b57-43e1-8d18-2bce6cabbfa1-kube-api-access-qszql\") pod \"openstack-galera-0\" (UID: \"72c64965-0b57-43e1-8d18-2bce6cabbfa1\") " pod="openstack/openstack-galera-0" Oct 08 18:27:27 crc kubenswrapper[4988]: I1008 18:27:27.384709 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/72c64965-0b57-43e1-8d18-2bce6cabbfa1-kolla-config\") pod \"openstack-galera-0\" (UID: \"72c64965-0b57-43e1-8d18-2bce6cabbfa1\") " pod="openstack/openstack-galera-0" Oct 08 18:27:27 crc kubenswrapper[4988]: I1008 18:27:27.385401 4988 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"openstack-galera-0\" (UID: \"72c64965-0b57-43e1-8d18-2bce6cabbfa1\") device mount path \"/mnt/openstack/pv02\"" pod="openstack/openstack-galera-0" Oct 08 18:27:27 crc kubenswrapper[4988]: I1008 18:27:27.385450 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/72c64965-0b57-43e1-8d18-2bce6cabbfa1-kolla-config\") pod \"openstack-galera-0\" (UID: \"72c64965-0b57-43e1-8d18-2bce6cabbfa1\") " pod="openstack/openstack-galera-0" Oct 08 18:27:27 crc kubenswrapper[4988]: I1008 18:27:27.385977 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/72c64965-0b57-43e1-8d18-2bce6cabbfa1-config-data-generated\") pod \"openstack-galera-0\" (UID: \"72c64965-0b57-43e1-8d18-2bce6cabbfa1\") " pod="openstack/openstack-galera-0" Oct 08 18:27:27 crc kubenswrapper[4988]: I1008 18:27:27.386920 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/72c64965-0b57-43e1-8d18-2bce6cabbfa1-config-data-default\") pod \"openstack-galera-0\" (UID: \"72c64965-0b57-43e1-8d18-2bce6cabbfa1\") " pod="openstack/openstack-galera-0" Oct 08 18:27:27 crc kubenswrapper[4988]: I1008 18:27:27.387783 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/72c64965-0b57-43e1-8d18-2bce6cabbfa1-operator-scripts\") pod \"openstack-galera-0\" (UID: \"72c64965-0b57-43e1-8d18-2bce6cabbfa1\") " pod="openstack/openstack-galera-0" Oct 08 18:27:27 crc kubenswrapper[4988]: I1008 18:27:27.391115 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/72c64965-0b57-43e1-8d18-2bce6cabbfa1-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"72c64965-0b57-43e1-8d18-2bce6cabbfa1\") " pod="openstack/openstack-galera-0" Oct 08 18:27:27 crc kubenswrapper[4988]: I1008 18:27:27.391926 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/72c64965-0b57-43e1-8d18-2bce6cabbfa1-secrets\") pod \"openstack-galera-0\" (UID: \"72c64965-0b57-43e1-8d18-2bce6cabbfa1\") " pod="openstack/openstack-galera-0" Oct 08 18:27:27 crc kubenswrapper[4988]: I1008 18:27:27.401205 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qszql\" (UniqueName: \"kubernetes.io/projected/72c64965-0b57-43e1-8d18-2bce6cabbfa1-kube-api-access-qszql\") pod \"openstack-galera-0\" (UID: \"72c64965-0b57-43e1-8d18-2bce6cabbfa1\") " pod="openstack/openstack-galera-0" Oct 08 18:27:27 crc kubenswrapper[4988]: I1008 18:27:27.401351 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/72c64965-0b57-43e1-8d18-2bce6cabbfa1-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"72c64965-0b57-43e1-8d18-2bce6cabbfa1\") " pod="openstack/openstack-galera-0" Oct 08 18:27:27 crc kubenswrapper[4988]: I1008 18:27:27.411346 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"openstack-galera-0\" (UID: \"72c64965-0b57-43e1-8d18-2bce6cabbfa1\") " pod="openstack/openstack-galera-0" Oct 08 18:27:27 crc kubenswrapper[4988]: I1008 18:27:27.514664 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Oct 08 18:27:27 crc kubenswrapper[4988]: I1008 18:27:27.639671 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-cell1-galera-0"] Oct 08 18:27:27 crc kubenswrapper[4988]: I1008 18:27:27.641121 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Oct 08 18:27:27 crc kubenswrapper[4988]: I1008 18:27:27.647800 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-cell1-svc" Oct 08 18:27:27 crc kubenswrapper[4988]: I1008 18:27:27.647903 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-cell1-dockercfg-6wsbj" Oct 08 18:27:27 crc kubenswrapper[4988]: I1008 18:27:27.648209 4988 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-scripts" Oct 08 18:27:27 crc kubenswrapper[4988]: I1008 18:27:27.648899 4988 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-config-data" Oct 08 18:27:27 crc kubenswrapper[4988]: I1008 18:27:27.654322 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Oct 08 18:27:27 crc kubenswrapper[4988]: I1008 18:27:27.692111 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gz796\" (UniqueName: \"kubernetes.io/projected/3503942b-2825-4006-80e8-8c4610b89997-kube-api-access-gz796\") pod \"openstack-cell1-galera-0\" (UID: \"3503942b-2825-4006-80e8-8c4610b89997\") " pod="openstack/openstack-cell1-galera-0" Oct 08 18:27:27 crc kubenswrapper[4988]: I1008 18:27:27.692157 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3503942b-2825-4006-80e8-8c4610b89997-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"3503942b-2825-4006-80e8-8c4610b89997\") " pod="openstack/openstack-cell1-galera-0" Oct 08 18:27:27 crc kubenswrapper[4988]: I1008 18:27:27.692178 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3503942b-2825-4006-80e8-8c4610b89997-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"3503942b-2825-4006-80e8-8c4610b89997\") " pod="openstack/openstack-cell1-galera-0" Oct 08 18:27:27 crc kubenswrapper[4988]: I1008 18:27:27.692199 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/3503942b-2825-4006-80e8-8c4610b89997-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"3503942b-2825-4006-80e8-8c4610b89997\") " pod="openstack/openstack-cell1-galera-0" Oct 08 18:27:27 crc kubenswrapper[4988]: I1008 18:27:27.692219 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/3503942b-2825-4006-80e8-8c4610b89997-secrets\") pod \"openstack-cell1-galera-0\" (UID: \"3503942b-2825-4006-80e8-8c4610b89997\") " pod="openstack/openstack-cell1-galera-0" Oct 08 18:27:27 crc kubenswrapper[4988]: I1008 18:27:27.692610 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/3503942b-2825-4006-80e8-8c4610b89997-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"3503942b-2825-4006-80e8-8c4610b89997\") " pod="openstack/openstack-cell1-galera-0" Oct 08 18:27:27 crc kubenswrapper[4988]: I1008 18:27:27.692801 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"openstack-cell1-galera-0\" (UID: \"3503942b-2825-4006-80e8-8c4610b89997\") " pod="openstack/openstack-cell1-galera-0" Oct 08 18:27:27 crc kubenswrapper[4988]: I1008 18:27:27.692826 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/3503942b-2825-4006-80e8-8c4610b89997-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"3503942b-2825-4006-80e8-8c4610b89997\") " pod="openstack/openstack-cell1-galera-0" Oct 08 18:27:27 crc kubenswrapper[4988]: I1008 18:27:27.692926 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/3503942b-2825-4006-80e8-8c4610b89997-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"3503942b-2825-4006-80e8-8c4610b89997\") " pod="openstack/openstack-cell1-galera-0" Oct 08 18:27:27 crc kubenswrapper[4988]: I1008 18:27:27.796038 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"openstack-cell1-galera-0\" (UID: \"3503942b-2825-4006-80e8-8c4610b89997\") " pod="openstack/openstack-cell1-galera-0" Oct 08 18:27:27 crc kubenswrapper[4988]: I1008 18:27:27.796079 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/3503942b-2825-4006-80e8-8c4610b89997-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"3503942b-2825-4006-80e8-8c4610b89997\") " pod="openstack/openstack-cell1-galera-0" Oct 08 18:27:27 crc kubenswrapper[4988]: I1008 18:27:27.796117 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/3503942b-2825-4006-80e8-8c4610b89997-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"3503942b-2825-4006-80e8-8c4610b89997\") " pod="openstack/openstack-cell1-galera-0" Oct 08 18:27:27 crc kubenswrapper[4988]: I1008 18:27:27.796141 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gz796\" (UniqueName: \"kubernetes.io/projected/3503942b-2825-4006-80e8-8c4610b89997-kube-api-access-gz796\") pod \"openstack-cell1-galera-0\" (UID: \"3503942b-2825-4006-80e8-8c4610b89997\") " pod="openstack/openstack-cell1-galera-0" Oct 08 18:27:27 crc kubenswrapper[4988]: I1008 18:27:27.796163 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3503942b-2825-4006-80e8-8c4610b89997-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"3503942b-2825-4006-80e8-8c4610b89997\") " pod="openstack/openstack-cell1-galera-0" Oct 08 18:27:27 crc kubenswrapper[4988]: I1008 18:27:27.796179 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3503942b-2825-4006-80e8-8c4610b89997-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"3503942b-2825-4006-80e8-8c4610b89997\") " pod="openstack/openstack-cell1-galera-0" Oct 08 18:27:27 crc kubenswrapper[4988]: I1008 18:27:27.796196 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/3503942b-2825-4006-80e8-8c4610b89997-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"3503942b-2825-4006-80e8-8c4610b89997\") " pod="openstack/openstack-cell1-galera-0" Oct 08 18:27:27 crc kubenswrapper[4988]: I1008 18:27:27.796211 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/3503942b-2825-4006-80e8-8c4610b89997-secrets\") pod \"openstack-cell1-galera-0\" (UID: \"3503942b-2825-4006-80e8-8c4610b89997\") " pod="openstack/openstack-cell1-galera-0" Oct 08 18:27:27 crc kubenswrapper[4988]: I1008 18:27:27.796260 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/3503942b-2825-4006-80e8-8c4610b89997-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"3503942b-2825-4006-80e8-8c4610b89997\") " pod="openstack/openstack-cell1-galera-0" Oct 08 18:27:27 crc kubenswrapper[4988]: I1008 18:27:27.797017 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/3503942b-2825-4006-80e8-8c4610b89997-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"3503942b-2825-4006-80e8-8c4610b89997\") " pod="openstack/openstack-cell1-galera-0" Oct 08 18:27:27 crc kubenswrapper[4988]: I1008 18:27:27.797237 4988 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"openstack-cell1-galera-0\" (UID: \"3503942b-2825-4006-80e8-8c4610b89997\") device mount path \"/mnt/openstack/pv01\"" pod="openstack/openstack-cell1-galera-0" Oct 08 18:27:27 crc kubenswrapper[4988]: I1008 18:27:27.799142 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3503942b-2825-4006-80e8-8c4610b89997-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"3503942b-2825-4006-80e8-8c4610b89997\") " pod="openstack/openstack-cell1-galera-0" Oct 08 18:27:27 crc kubenswrapper[4988]: I1008 18:27:27.800109 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/3503942b-2825-4006-80e8-8c4610b89997-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"3503942b-2825-4006-80e8-8c4610b89997\") " pod="openstack/openstack-cell1-galera-0" Oct 08 18:27:27 crc kubenswrapper[4988]: I1008 18:27:27.800857 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/3503942b-2825-4006-80e8-8c4610b89997-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"3503942b-2825-4006-80e8-8c4610b89997\") " pod="openstack/openstack-cell1-galera-0" Oct 08 18:27:27 crc kubenswrapper[4988]: I1008 18:27:27.804145 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3503942b-2825-4006-80e8-8c4610b89997-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"3503942b-2825-4006-80e8-8c4610b89997\") " pod="openstack/openstack-cell1-galera-0" Oct 08 18:27:27 crc kubenswrapper[4988]: I1008 18:27:27.804204 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/3503942b-2825-4006-80e8-8c4610b89997-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"3503942b-2825-4006-80e8-8c4610b89997\") " pod="openstack/openstack-cell1-galera-0" Oct 08 18:27:27 crc kubenswrapper[4988]: I1008 18:27:27.805444 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/3503942b-2825-4006-80e8-8c4610b89997-secrets\") pod \"openstack-cell1-galera-0\" (UID: \"3503942b-2825-4006-80e8-8c4610b89997\") " pod="openstack/openstack-cell1-galera-0" Oct 08 18:27:27 crc kubenswrapper[4988]: I1008 18:27:27.824666 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"openstack-cell1-galera-0\" (UID: \"3503942b-2825-4006-80e8-8c4610b89997\") " pod="openstack/openstack-cell1-galera-0" Oct 08 18:27:27 crc kubenswrapper[4988]: I1008 18:27:27.826920 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gz796\" (UniqueName: \"kubernetes.io/projected/3503942b-2825-4006-80e8-8c4610b89997-kube-api-access-gz796\") pod \"openstack-cell1-galera-0\" (UID: \"3503942b-2825-4006-80e8-8c4610b89997\") " pod="openstack/openstack-cell1-galera-0" Oct 08 18:27:27 crc kubenswrapper[4988]: I1008 18:27:27.977149 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Oct 08 18:27:28 crc kubenswrapper[4988]: I1008 18:27:28.106285 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Oct 08 18:27:28 crc kubenswrapper[4988]: W1008 18:27:28.150788 4988 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod72c64965_0b57_43e1_8d18_2bce6cabbfa1.slice/crio-a0f4596eebd8fb2c500b21efcd1ffce02fd0b60bfebe08354d1aa52b412471be WatchSource:0}: Error finding container a0f4596eebd8fb2c500b21efcd1ffce02fd0b60bfebe08354d1aa52b412471be: Status 404 returned error can't find the container with id a0f4596eebd8fb2c500b21efcd1ffce02fd0b60bfebe08354d1aa52b412471be Oct 08 18:27:28 crc kubenswrapper[4988]: I1008 18:27:28.220926 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"72c64965-0b57-43e1-8d18-2bce6cabbfa1","Type":"ContainerStarted","Data":"a0f4596eebd8fb2c500b21efcd1ffce02fd0b60bfebe08354d1aa52b412471be"} Oct 08 18:27:28 crc kubenswrapper[4988]: I1008 18:27:28.251720 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/memcached-0"] Oct 08 18:27:28 crc kubenswrapper[4988]: I1008 18:27:28.253235 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Oct 08 18:27:28 crc kubenswrapper[4988]: I1008 18:27:28.257122 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Oct 08 18:27:28 crc kubenswrapper[4988]: I1008 18:27:28.260159 4988 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"memcached-config-data" Oct 08 18:27:28 crc kubenswrapper[4988]: I1008 18:27:28.260298 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"memcached-memcached-dockercfg-jw678" Oct 08 18:27:28 crc kubenswrapper[4988]: I1008 18:27:28.260448 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-memcached-svc" Oct 08 18:27:28 crc kubenswrapper[4988]: I1008 18:27:28.312365 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/32bb7401-6292-4ca6-98f4-15522125c3ee-combined-ca-bundle\") pod \"memcached-0\" (UID: \"32bb7401-6292-4ca6-98f4-15522125c3ee\") " pod="openstack/memcached-0" Oct 08 18:27:28 crc kubenswrapper[4988]: I1008 18:27:28.312527 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n9xkg\" (UniqueName: \"kubernetes.io/projected/32bb7401-6292-4ca6-98f4-15522125c3ee-kube-api-access-n9xkg\") pod \"memcached-0\" (UID: \"32bb7401-6292-4ca6-98f4-15522125c3ee\") " pod="openstack/memcached-0" Oct 08 18:27:28 crc kubenswrapper[4988]: I1008 18:27:28.312569 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/32bb7401-6292-4ca6-98f4-15522125c3ee-kolla-config\") pod \"memcached-0\" (UID: \"32bb7401-6292-4ca6-98f4-15522125c3ee\") " pod="openstack/memcached-0" Oct 08 18:27:28 crc kubenswrapper[4988]: I1008 18:27:28.312608 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/32bb7401-6292-4ca6-98f4-15522125c3ee-memcached-tls-certs\") pod \"memcached-0\" (UID: \"32bb7401-6292-4ca6-98f4-15522125c3ee\") " pod="openstack/memcached-0" Oct 08 18:27:28 crc kubenswrapper[4988]: I1008 18:27:28.312643 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/32bb7401-6292-4ca6-98f4-15522125c3ee-config-data\") pod \"memcached-0\" (UID: \"32bb7401-6292-4ca6-98f4-15522125c3ee\") " pod="openstack/memcached-0" Oct 08 18:27:28 crc kubenswrapper[4988]: I1008 18:27:28.414299 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/32bb7401-6292-4ca6-98f4-15522125c3ee-memcached-tls-certs\") pod \"memcached-0\" (UID: \"32bb7401-6292-4ca6-98f4-15522125c3ee\") " pod="openstack/memcached-0" Oct 08 18:27:28 crc kubenswrapper[4988]: I1008 18:27:28.414399 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/32bb7401-6292-4ca6-98f4-15522125c3ee-config-data\") pod \"memcached-0\" (UID: \"32bb7401-6292-4ca6-98f4-15522125c3ee\") " pod="openstack/memcached-0" Oct 08 18:27:28 crc kubenswrapper[4988]: I1008 18:27:28.414456 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/32bb7401-6292-4ca6-98f4-15522125c3ee-combined-ca-bundle\") pod \"memcached-0\" (UID: \"32bb7401-6292-4ca6-98f4-15522125c3ee\") " pod="openstack/memcached-0" Oct 08 18:27:28 crc kubenswrapper[4988]: I1008 18:27:28.414564 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n9xkg\" (UniqueName: \"kubernetes.io/projected/32bb7401-6292-4ca6-98f4-15522125c3ee-kube-api-access-n9xkg\") pod \"memcached-0\" (UID: \"32bb7401-6292-4ca6-98f4-15522125c3ee\") " pod="openstack/memcached-0" Oct 08 18:27:28 crc kubenswrapper[4988]: I1008 18:27:28.414598 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/32bb7401-6292-4ca6-98f4-15522125c3ee-kolla-config\") pod \"memcached-0\" (UID: \"32bb7401-6292-4ca6-98f4-15522125c3ee\") " pod="openstack/memcached-0" Oct 08 18:27:28 crc kubenswrapper[4988]: I1008 18:27:28.416262 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/32bb7401-6292-4ca6-98f4-15522125c3ee-kolla-config\") pod \"memcached-0\" (UID: \"32bb7401-6292-4ca6-98f4-15522125c3ee\") " pod="openstack/memcached-0" Oct 08 18:27:28 crc kubenswrapper[4988]: I1008 18:27:28.416723 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/32bb7401-6292-4ca6-98f4-15522125c3ee-config-data\") pod \"memcached-0\" (UID: \"32bb7401-6292-4ca6-98f4-15522125c3ee\") " pod="openstack/memcached-0" Oct 08 18:27:28 crc kubenswrapper[4988]: I1008 18:27:28.423008 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/32bb7401-6292-4ca6-98f4-15522125c3ee-memcached-tls-certs\") pod \"memcached-0\" (UID: \"32bb7401-6292-4ca6-98f4-15522125c3ee\") " pod="openstack/memcached-0" Oct 08 18:27:28 crc kubenswrapper[4988]: I1008 18:27:28.424830 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/32bb7401-6292-4ca6-98f4-15522125c3ee-combined-ca-bundle\") pod \"memcached-0\" (UID: \"32bb7401-6292-4ca6-98f4-15522125c3ee\") " pod="openstack/memcached-0" Oct 08 18:27:28 crc kubenswrapper[4988]: I1008 18:27:28.451550 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n9xkg\" (UniqueName: \"kubernetes.io/projected/32bb7401-6292-4ca6-98f4-15522125c3ee-kube-api-access-n9xkg\") pod \"memcached-0\" (UID: \"32bb7401-6292-4ca6-98f4-15522125c3ee\") " pod="openstack/memcached-0" Oct 08 18:27:28 crc kubenswrapper[4988]: I1008 18:27:28.576681 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Oct 08 18:27:28 crc kubenswrapper[4988]: I1008 18:27:28.683937 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Oct 08 18:27:29 crc kubenswrapper[4988]: I1008 18:27:29.147561 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Oct 08 18:27:29 crc kubenswrapper[4988]: W1008 18:27:29.159624 4988 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod32bb7401_6292_4ca6_98f4_15522125c3ee.slice/crio-f5dbe430bd2d56992aa14033ff6363ff154185ebd2038a740e4725f0755dc90e WatchSource:0}: Error finding container f5dbe430bd2d56992aa14033ff6363ff154185ebd2038a740e4725f0755dc90e: Status 404 returned error can't find the container with id f5dbe430bd2d56992aa14033ff6363ff154185ebd2038a740e4725f0755dc90e Oct 08 18:27:29 crc kubenswrapper[4988]: I1008 18:27:29.253547 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"32bb7401-6292-4ca6-98f4-15522125c3ee","Type":"ContainerStarted","Data":"f5dbe430bd2d56992aa14033ff6363ff154185ebd2038a740e4725f0755dc90e"} Oct 08 18:27:29 crc kubenswrapper[4988]: I1008 18:27:29.253594 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"3503942b-2825-4006-80e8-8c4610b89997","Type":"ContainerStarted","Data":"f883ee7dc1d600955ad76f9143dec7c4f14a01603d4965aa6075b83d56ed74f4"} Oct 08 18:27:29 crc kubenswrapper[4988]: I1008 18:27:29.959231 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Oct 08 18:27:29 crc kubenswrapper[4988]: I1008 18:27:29.977725 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 08 18:27:29 crc kubenswrapper[4988]: I1008 18:27:29.978070 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Oct 08 18:27:29 crc kubenswrapper[4988]: I1008 18:27:29.981364 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"telemetry-ceilometer-dockercfg-zrrkq" Oct 08 18:27:30 crc kubenswrapper[4988]: I1008 18:27:30.040092 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j2n7z\" (UniqueName: \"kubernetes.io/projected/731acc53-25a1-43a4-b9a8-16de25b1e7b1-kube-api-access-j2n7z\") pod \"kube-state-metrics-0\" (UID: \"731acc53-25a1-43a4-b9a8-16de25b1e7b1\") " pod="openstack/kube-state-metrics-0" Oct 08 18:27:30 crc kubenswrapper[4988]: I1008 18:27:30.141752 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j2n7z\" (UniqueName: \"kubernetes.io/projected/731acc53-25a1-43a4-b9a8-16de25b1e7b1-kube-api-access-j2n7z\") pod \"kube-state-metrics-0\" (UID: \"731acc53-25a1-43a4-b9a8-16de25b1e7b1\") " pod="openstack/kube-state-metrics-0" Oct 08 18:27:30 crc kubenswrapper[4988]: I1008 18:27:30.161665 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j2n7z\" (UniqueName: \"kubernetes.io/projected/731acc53-25a1-43a4-b9a8-16de25b1e7b1-kube-api-access-j2n7z\") pod \"kube-state-metrics-0\" (UID: \"731acc53-25a1-43a4-b9a8-16de25b1e7b1\") " pod="openstack/kube-state-metrics-0" Oct 08 18:27:30 crc kubenswrapper[4988]: I1008 18:27:30.339366 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Oct 08 18:27:33 crc kubenswrapper[4988]: I1008 18:27:33.864208 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-gp8f4"] Oct 08 18:27:33 crc kubenswrapper[4988]: I1008 18:27:33.865699 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-gp8f4" Oct 08 18:27:33 crc kubenswrapper[4988]: I1008 18:27:33.868369 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncontroller-ovncontroller-dockercfg-8dnr8" Oct 08 18:27:33 crc kubenswrapper[4988]: I1008 18:27:33.868647 4988 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-scripts" Oct 08 18:27:33 crc kubenswrapper[4988]: I1008 18:27:33.868800 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovncontroller-ovndbs" Oct 08 18:27:33 crc kubenswrapper[4988]: I1008 18:27:33.880051 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-ovs-kc7fw"] Oct 08 18:27:33 crc kubenswrapper[4988]: I1008 18:27:33.881800 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-kc7fw" Oct 08 18:27:33 crc kubenswrapper[4988]: I1008 18:27:33.885705 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-gp8f4"] Oct 08 18:27:33 crc kubenswrapper[4988]: I1008 18:27:33.897960 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-kc7fw"] Oct 08 18:27:34 crc kubenswrapper[4988]: I1008 18:27:34.003540 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/f1628013-e0b9-4c31-80c0-91aabaaef0f6-var-lib\") pod \"ovn-controller-ovs-kc7fw\" (UID: \"f1628013-e0b9-4c31-80c0-91aabaaef0f6\") " pod="openstack/ovn-controller-ovs-kc7fw" Oct 08 18:27:34 crc kubenswrapper[4988]: I1008 18:27:34.003584 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/c285e09d-c2f5-4642-9e6b-59b0ed3ed651-var-run-ovn\") pod \"ovn-controller-gp8f4\" (UID: \"c285e09d-c2f5-4642-9e6b-59b0ed3ed651\") " pod="openstack/ovn-controller-gp8f4" Oct 08 18:27:34 crc kubenswrapper[4988]: I1008 18:27:34.003607 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f1628013-e0b9-4c31-80c0-91aabaaef0f6-var-log\") pod \"ovn-controller-ovs-kc7fw\" (UID: \"f1628013-e0b9-4c31-80c0-91aabaaef0f6\") " pod="openstack/ovn-controller-ovs-kc7fw" Oct 08 18:27:34 crc kubenswrapper[4988]: I1008 18:27:34.003636 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c285e09d-c2f5-4642-9e6b-59b0ed3ed651-scripts\") pod \"ovn-controller-gp8f4\" (UID: \"c285e09d-c2f5-4642-9e6b-59b0ed3ed651\") " pod="openstack/ovn-controller-gp8f4" Oct 08 18:27:34 crc kubenswrapper[4988]: I1008 18:27:34.003699 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/c285e09d-c2f5-4642-9e6b-59b0ed3ed651-ovn-controller-tls-certs\") pod \"ovn-controller-gp8f4\" (UID: \"c285e09d-c2f5-4642-9e6b-59b0ed3ed651\") " pod="openstack/ovn-controller-gp8f4" Oct 08 18:27:34 crc kubenswrapper[4988]: I1008 18:27:34.003733 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v2tq6\" (UniqueName: \"kubernetes.io/projected/c285e09d-c2f5-4642-9e6b-59b0ed3ed651-kube-api-access-v2tq6\") pod \"ovn-controller-gp8f4\" (UID: \"c285e09d-c2f5-4642-9e6b-59b0ed3ed651\") " pod="openstack/ovn-controller-gp8f4" Oct 08 18:27:34 crc kubenswrapper[4988]: I1008 18:27:34.003749 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f1628013-e0b9-4c31-80c0-91aabaaef0f6-scripts\") pod \"ovn-controller-ovs-kc7fw\" (UID: \"f1628013-e0b9-4c31-80c0-91aabaaef0f6\") " pod="openstack/ovn-controller-ovs-kc7fw" Oct 08 18:27:34 crc kubenswrapper[4988]: I1008 18:27:34.003775 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mhxmj\" (UniqueName: \"kubernetes.io/projected/f1628013-e0b9-4c31-80c0-91aabaaef0f6-kube-api-access-mhxmj\") pod \"ovn-controller-ovs-kc7fw\" (UID: \"f1628013-e0b9-4c31-80c0-91aabaaef0f6\") " pod="openstack/ovn-controller-ovs-kc7fw" Oct 08 18:27:34 crc kubenswrapper[4988]: I1008 18:27:34.003797 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/f1628013-e0b9-4c31-80c0-91aabaaef0f6-etc-ovs\") pod \"ovn-controller-ovs-kc7fw\" (UID: \"f1628013-e0b9-4c31-80c0-91aabaaef0f6\") " pod="openstack/ovn-controller-ovs-kc7fw" Oct 08 18:27:34 crc kubenswrapper[4988]: I1008 18:27:34.003813 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/f1628013-e0b9-4c31-80c0-91aabaaef0f6-var-run\") pod \"ovn-controller-ovs-kc7fw\" (UID: \"f1628013-e0b9-4c31-80c0-91aabaaef0f6\") " pod="openstack/ovn-controller-ovs-kc7fw" Oct 08 18:27:34 crc kubenswrapper[4988]: I1008 18:27:34.003829 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/c285e09d-c2f5-4642-9e6b-59b0ed3ed651-var-log-ovn\") pod \"ovn-controller-gp8f4\" (UID: \"c285e09d-c2f5-4642-9e6b-59b0ed3ed651\") " pod="openstack/ovn-controller-gp8f4" Oct 08 18:27:34 crc kubenswrapper[4988]: I1008 18:27:34.003854 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c285e09d-c2f5-4642-9e6b-59b0ed3ed651-combined-ca-bundle\") pod \"ovn-controller-gp8f4\" (UID: \"c285e09d-c2f5-4642-9e6b-59b0ed3ed651\") " pod="openstack/ovn-controller-gp8f4" Oct 08 18:27:34 crc kubenswrapper[4988]: I1008 18:27:34.003870 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/c285e09d-c2f5-4642-9e6b-59b0ed3ed651-var-run\") pod \"ovn-controller-gp8f4\" (UID: \"c285e09d-c2f5-4642-9e6b-59b0ed3ed651\") " pod="openstack/ovn-controller-gp8f4" Oct 08 18:27:34 crc kubenswrapper[4988]: I1008 18:27:34.104827 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/f1628013-e0b9-4c31-80c0-91aabaaef0f6-var-lib\") pod \"ovn-controller-ovs-kc7fw\" (UID: \"f1628013-e0b9-4c31-80c0-91aabaaef0f6\") " pod="openstack/ovn-controller-ovs-kc7fw" Oct 08 18:27:34 crc kubenswrapper[4988]: I1008 18:27:34.104879 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/c285e09d-c2f5-4642-9e6b-59b0ed3ed651-var-run-ovn\") pod \"ovn-controller-gp8f4\" (UID: \"c285e09d-c2f5-4642-9e6b-59b0ed3ed651\") " pod="openstack/ovn-controller-gp8f4" Oct 08 18:27:34 crc kubenswrapper[4988]: I1008 18:27:34.104900 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f1628013-e0b9-4c31-80c0-91aabaaef0f6-var-log\") pod \"ovn-controller-ovs-kc7fw\" (UID: \"f1628013-e0b9-4c31-80c0-91aabaaef0f6\") " pod="openstack/ovn-controller-ovs-kc7fw" Oct 08 18:27:34 crc kubenswrapper[4988]: I1008 18:27:34.104929 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c285e09d-c2f5-4642-9e6b-59b0ed3ed651-scripts\") pod \"ovn-controller-gp8f4\" (UID: \"c285e09d-c2f5-4642-9e6b-59b0ed3ed651\") " pod="openstack/ovn-controller-gp8f4" Oct 08 18:27:34 crc kubenswrapper[4988]: I1008 18:27:34.104951 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/c285e09d-c2f5-4642-9e6b-59b0ed3ed651-ovn-controller-tls-certs\") pod \"ovn-controller-gp8f4\" (UID: \"c285e09d-c2f5-4642-9e6b-59b0ed3ed651\") " pod="openstack/ovn-controller-gp8f4" Oct 08 18:27:34 crc kubenswrapper[4988]: I1008 18:27:34.104976 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v2tq6\" (UniqueName: \"kubernetes.io/projected/c285e09d-c2f5-4642-9e6b-59b0ed3ed651-kube-api-access-v2tq6\") pod \"ovn-controller-gp8f4\" (UID: \"c285e09d-c2f5-4642-9e6b-59b0ed3ed651\") " pod="openstack/ovn-controller-gp8f4" Oct 08 18:27:34 crc kubenswrapper[4988]: I1008 18:27:34.104995 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f1628013-e0b9-4c31-80c0-91aabaaef0f6-scripts\") pod \"ovn-controller-ovs-kc7fw\" (UID: \"f1628013-e0b9-4c31-80c0-91aabaaef0f6\") " pod="openstack/ovn-controller-ovs-kc7fw" Oct 08 18:27:34 crc kubenswrapper[4988]: I1008 18:27:34.105030 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mhxmj\" (UniqueName: \"kubernetes.io/projected/f1628013-e0b9-4c31-80c0-91aabaaef0f6-kube-api-access-mhxmj\") pod \"ovn-controller-ovs-kc7fw\" (UID: \"f1628013-e0b9-4c31-80c0-91aabaaef0f6\") " pod="openstack/ovn-controller-ovs-kc7fw" Oct 08 18:27:34 crc kubenswrapper[4988]: I1008 18:27:34.105051 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/f1628013-e0b9-4c31-80c0-91aabaaef0f6-etc-ovs\") pod \"ovn-controller-ovs-kc7fw\" (UID: \"f1628013-e0b9-4c31-80c0-91aabaaef0f6\") " pod="openstack/ovn-controller-ovs-kc7fw" Oct 08 18:27:34 crc kubenswrapper[4988]: I1008 18:27:34.105069 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/f1628013-e0b9-4c31-80c0-91aabaaef0f6-var-run\") pod \"ovn-controller-ovs-kc7fw\" (UID: \"f1628013-e0b9-4c31-80c0-91aabaaef0f6\") " pod="openstack/ovn-controller-ovs-kc7fw" Oct 08 18:27:34 crc kubenswrapper[4988]: I1008 18:27:34.105091 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/c285e09d-c2f5-4642-9e6b-59b0ed3ed651-var-log-ovn\") pod \"ovn-controller-gp8f4\" (UID: \"c285e09d-c2f5-4642-9e6b-59b0ed3ed651\") " pod="openstack/ovn-controller-gp8f4" Oct 08 18:27:34 crc kubenswrapper[4988]: I1008 18:27:34.105124 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c285e09d-c2f5-4642-9e6b-59b0ed3ed651-combined-ca-bundle\") pod \"ovn-controller-gp8f4\" (UID: \"c285e09d-c2f5-4642-9e6b-59b0ed3ed651\") " pod="openstack/ovn-controller-gp8f4" Oct 08 18:27:34 crc kubenswrapper[4988]: I1008 18:27:34.105145 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/c285e09d-c2f5-4642-9e6b-59b0ed3ed651-var-run\") pod \"ovn-controller-gp8f4\" (UID: \"c285e09d-c2f5-4642-9e6b-59b0ed3ed651\") " pod="openstack/ovn-controller-gp8f4" Oct 08 18:27:34 crc kubenswrapper[4988]: I1008 18:27:34.105374 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/f1628013-e0b9-4c31-80c0-91aabaaef0f6-var-lib\") pod \"ovn-controller-ovs-kc7fw\" (UID: \"f1628013-e0b9-4c31-80c0-91aabaaef0f6\") " pod="openstack/ovn-controller-ovs-kc7fw" Oct 08 18:27:34 crc kubenswrapper[4988]: I1008 18:27:34.105491 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/c285e09d-c2f5-4642-9e6b-59b0ed3ed651-var-run-ovn\") pod \"ovn-controller-gp8f4\" (UID: \"c285e09d-c2f5-4642-9e6b-59b0ed3ed651\") " pod="openstack/ovn-controller-gp8f4" Oct 08 18:27:34 crc kubenswrapper[4988]: I1008 18:27:34.105539 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/c285e09d-c2f5-4642-9e6b-59b0ed3ed651-var-log-ovn\") pod \"ovn-controller-gp8f4\" (UID: \"c285e09d-c2f5-4642-9e6b-59b0ed3ed651\") " pod="openstack/ovn-controller-gp8f4" Oct 08 18:27:34 crc kubenswrapper[4988]: I1008 18:27:34.105645 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/f1628013-e0b9-4c31-80c0-91aabaaef0f6-etc-ovs\") pod \"ovn-controller-ovs-kc7fw\" (UID: \"f1628013-e0b9-4c31-80c0-91aabaaef0f6\") " pod="openstack/ovn-controller-ovs-kc7fw" Oct 08 18:27:34 crc kubenswrapper[4988]: I1008 18:27:34.105692 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f1628013-e0b9-4c31-80c0-91aabaaef0f6-var-log\") pod \"ovn-controller-ovs-kc7fw\" (UID: \"f1628013-e0b9-4c31-80c0-91aabaaef0f6\") " pod="openstack/ovn-controller-ovs-kc7fw" Oct 08 18:27:34 crc kubenswrapper[4988]: I1008 18:27:34.105725 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/f1628013-e0b9-4c31-80c0-91aabaaef0f6-var-run\") pod \"ovn-controller-ovs-kc7fw\" (UID: \"f1628013-e0b9-4c31-80c0-91aabaaef0f6\") " pod="openstack/ovn-controller-ovs-kc7fw" Oct 08 18:27:34 crc kubenswrapper[4988]: I1008 18:27:34.105985 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/c285e09d-c2f5-4642-9e6b-59b0ed3ed651-var-run\") pod \"ovn-controller-gp8f4\" (UID: \"c285e09d-c2f5-4642-9e6b-59b0ed3ed651\") " pod="openstack/ovn-controller-gp8f4" Oct 08 18:27:34 crc kubenswrapper[4988]: I1008 18:27:34.107667 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f1628013-e0b9-4c31-80c0-91aabaaef0f6-scripts\") pod \"ovn-controller-ovs-kc7fw\" (UID: \"f1628013-e0b9-4c31-80c0-91aabaaef0f6\") " pod="openstack/ovn-controller-ovs-kc7fw" Oct 08 18:27:34 crc kubenswrapper[4988]: I1008 18:27:34.108084 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c285e09d-c2f5-4642-9e6b-59b0ed3ed651-scripts\") pod \"ovn-controller-gp8f4\" (UID: \"c285e09d-c2f5-4642-9e6b-59b0ed3ed651\") " pod="openstack/ovn-controller-gp8f4" Oct 08 18:27:34 crc kubenswrapper[4988]: I1008 18:27:34.113768 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c285e09d-c2f5-4642-9e6b-59b0ed3ed651-combined-ca-bundle\") pod \"ovn-controller-gp8f4\" (UID: \"c285e09d-c2f5-4642-9e6b-59b0ed3ed651\") " pod="openstack/ovn-controller-gp8f4" Oct 08 18:27:34 crc kubenswrapper[4988]: I1008 18:27:34.117775 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/c285e09d-c2f5-4642-9e6b-59b0ed3ed651-ovn-controller-tls-certs\") pod \"ovn-controller-gp8f4\" (UID: \"c285e09d-c2f5-4642-9e6b-59b0ed3ed651\") " pod="openstack/ovn-controller-gp8f4" Oct 08 18:27:34 crc kubenswrapper[4988]: I1008 18:27:34.122134 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mhxmj\" (UniqueName: \"kubernetes.io/projected/f1628013-e0b9-4c31-80c0-91aabaaef0f6-kube-api-access-mhxmj\") pod \"ovn-controller-ovs-kc7fw\" (UID: \"f1628013-e0b9-4c31-80c0-91aabaaef0f6\") " pod="openstack/ovn-controller-ovs-kc7fw" Oct 08 18:27:34 crc kubenswrapper[4988]: I1008 18:27:34.125063 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v2tq6\" (UniqueName: \"kubernetes.io/projected/c285e09d-c2f5-4642-9e6b-59b0ed3ed651-kube-api-access-v2tq6\") pod \"ovn-controller-gp8f4\" (UID: \"c285e09d-c2f5-4642-9e6b-59b0ed3ed651\") " pod="openstack/ovn-controller-gp8f4" Oct 08 18:27:34 crc kubenswrapper[4988]: I1008 18:27:34.183675 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-gp8f4" Oct 08 18:27:34 crc kubenswrapper[4988]: I1008 18:27:34.201625 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-kc7fw" Oct 08 18:27:36 crc kubenswrapper[4988]: I1008 18:27:36.813894 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-nb-0"] Oct 08 18:27:36 crc kubenswrapper[4988]: I1008 18:27:36.815542 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Oct 08 18:27:36 crc kubenswrapper[4988]: I1008 18:27:36.815638 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Oct 08 18:27:36 crc kubenswrapper[4988]: I1008 18:27:36.821621 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovn-metrics" Oct 08 18:27:36 crc kubenswrapper[4988]: I1008 18:27:36.822137 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-nb-dockercfg-czkfp" Oct 08 18:27:36 crc kubenswrapper[4988]: I1008 18:27:36.822343 4988 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-scripts" Oct 08 18:27:36 crc kubenswrapper[4988]: I1008 18:27:36.824287 4988 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-config" Oct 08 18:27:36 crc kubenswrapper[4988]: I1008 18:27:36.824370 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovndbcluster-nb-ovndbs" Oct 08 18:27:36 crc kubenswrapper[4988]: I1008 18:27:36.948059 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b5dc1ee3-4897-45a2-ad00-0b91fbb2ae76-config\") pod \"ovsdbserver-nb-0\" (UID: \"b5dc1ee3-4897-45a2-ad00-0b91fbb2ae76\") " pod="openstack/ovsdbserver-nb-0" Oct 08 18:27:36 crc kubenswrapper[4988]: I1008 18:27:36.948150 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tvmgc\" (UniqueName: \"kubernetes.io/projected/b5dc1ee3-4897-45a2-ad00-0b91fbb2ae76-kube-api-access-tvmgc\") pod \"ovsdbserver-nb-0\" (UID: \"b5dc1ee3-4897-45a2-ad00-0b91fbb2ae76\") " pod="openstack/ovsdbserver-nb-0" Oct 08 18:27:36 crc kubenswrapper[4988]: I1008 18:27:36.948459 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/b5dc1ee3-4897-45a2-ad00-0b91fbb2ae76-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"b5dc1ee3-4897-45a2-ad00-0b91fbb2ae76\") " pod="openstack/ovsdbserver-nb-0" Oct 08 18:27:36 crc kubenswrapper[4988]: I1008 18:27:36.948667 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"ovsdbserver-nb-0\" (UID: \"b5dc1ee3-4897-45a2-ad00-0b91fbb2ae76\") " pod="openstack/ovsdbserver-nb-0" Oct 08 18:27:36 crc kubenswrapper[4988]: I1008 18:27:36.948739 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/b5dc1ee3-4897-45a2-ad00-0b91fbb2ae76-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"b5dc1ee3-4897-45a2-ad00-0b91fbb2ae76\") " pod="openstack/ovsdbserver-nb-0" Oct 08 18:27:36 crc kubenswrapper[4988]: I1008 18:27:36.948805 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b5dc1ee3-4897-45a2-ad00-0b91fbb2ae76-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"b5dc1ee3-4897-45a2-ad00-0b91fbb2ae76\") " pod="openstack/ovsdbserver-nb-0" Oct 08 18:27:36 crc kubenswrapper[4988]: I1008 18:27:36.948834 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/b5dc1ee3-4897-45a2-ad00-0b91fbb2ae76-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"b5dc1ee3-4897-45a2-ad00-0b91fbb2ae76\") " pod="openstack/ovsdbserver-nb-0" Oct 08 18:27:36 crc kubenswrapper[4988]: I1008 18:27:36.948886 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/b5dc1ee3-4897-45a2-ad00-0b91fbb2ae76-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"b5dc1ee3-4897-45a2-ad00-0b91fbb2ae76\") " pod="openstack/ovsdbserver-nb-0" Oct 08 18:27:37 crc kubenswrapper[4988]: I1008 18:27:37.008863 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-sb-0"] Oct 08 18:27:37 crc kubenswrapper[4988]: I1008 18:27:37.010419 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Oct 08 18:27:37 crc kubenswrapper[4988]: I1008 18:27:37.016426 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-sb-dockercfg-tznxm" Oct 08 18:27:37 crc kubenswrapper[4988]: I1008 18:27:37.016517 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovndbcluster-sb-ovndbs" Oct 08 18:27:37 crc kubenswrapper[4988]: I1008 18:27:37.016904 4988 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-scripts" Oct 08 18:27:37 crc kubenswrapper[4988]: I1008 18:27:37.017607 4988 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-config" Oct 08 18:27:37 crc kubenswrapper[4988]: I1008 18:27:37.038155 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Oct 08 18:27:37 crc kubenswrapper[4988]: I1008 18:27:37.049755 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b5dc1ee3-4897-45a2-ad00-0b91fbb2ae76-config\") pod \"ovsdbserver-nb-0\" (UID: \"b5dc1ee3-4897-45a2-ad00-0b91fbb2ae76\") " pod="openstack/ovsdbserver-nb-0" Oct 08 18:27:37 crc kubenswrapper[4988]: I1008 18:27:37.049794 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tvmgc\" (UniqueName: \"kubernetes.io/projected/b5dc1ee3-4897-45a2-ad00-0b91fbb2ae76-kube-api-access-tvmgc\") pod \"ovsdbserver-nb-0\" (UID: \"b5dc1ee3-4897-45a2-ad00-0b91fbb2ae76\") " pod="openstack/ovsdbserver-nb-0" Oct 08 18:27:37 crc kubenswrapper[4988]: I1008 18:27:37.049835 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/b5dc1ee3-4897-45a2-ad00-0b91fbb2ae76-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"b5dc1ee3-4897-45a2-ad00-0b91fbb2ae76\") " pod="openstack/ovsdbserver-nb-0" Oct 08 18:27:37 crc kubenswrapper[4988]: I1008 18:27:37.049882 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"ovsdbserver-nb-0\" (UID: \"b5dc1ee3-4897-45a2-ad00-0b91fbb2ae76\") " pod="openstack/ovsdbserver-nb-0" Oct 08 18:27:37 crc kubenswrapper[4988]: I1008 18:27:37.049909 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/b5dc1ee3-4897-45a2-ad00-0b91fbb2ae76-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"b5dc1ee3-4897-45a2-ad00-0b91fbb2ae76\") " pod="openstack/ovsdbserver-nb-0" Oct 08 18:27:37 crc kubenswrapper[4988]: I1008 18:27:37.049951 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b5dc1ee3-4897-45a2-ad00-0b91fbb2ae76-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"b5dc1ee3-4897-45a2-ad00-0b91fbb2ae76\") " pod="openstack/ovsdbserver-nb-0" Oct 08 18:27:37 crc kubenswrapper[4988]: I1008 18:27:37.049974 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/b5dc1ee3-4897-45a2-ad00-0b91fbb2ae76-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"b5dc1ee3-4897-45a2-ad00-0b91fbb2ae76\") " pod="openstack/ovsdbserver-nb-0" Oct 08 18:27:37 crc kubenswrapper[4988]: I1008 18:27:37.049995 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/b5dc1ee3-4897-45a2-ad00-0b91fbb2ae76-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"b5dc1ee3-4897-45a2-ad00-0b91fbb2ae76\") " pod="openstack/ovsdbserver-nb-0" Oct 08 18:27:37 crc kubenswrapper[4988]: I1008 18:27:37.050420 4988 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"ovsdbserver-nb-0\" (UID: \"b5dc1ee3-4897-45a2-ad00-0b91fbb2ae76\") device mount path \"/mnt/openstack/pv09\"" pod="openstack/ovsdbserver-nb-0" Oct 08 18:27:37 crc kubenswrapper[4988]: I1008 18:27:37.050486 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/b5dc1ee3-4897-45a2-ad00-0b91fbb2ae76-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"b5dc1ee3-4897-45a2-ad00-0b91fbb2ae76\") " pod="openstack/ovsdbserver-nb-0" Oct 08 18:27:37 crc kubenswrapper[4988]: I1008 18:27:37.050933 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b5dc1ee3-4897-45a2-ad00-0b91fbb2ae76-config\") pod \"ovsdbserver-nb-0\" (UID: \"b5dc1ee3-4897-45a2-ad00-0b91fbb2ae76\") " pod="openstack/ovsdbserver-nb-0" Oct 08 18:27:37 crc kubenswrapper[4988]: I1008 18:27:37.051122 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/b5dc1ee3-4897-45a2-ad00-0b91fbb2ae76-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"b5dc1ee3-4897-45a2-ad00-0b91fbb2ae76\") " pod="openstack/ovsdbserver-nb-0" Oct 08 18:27:37 crc kubenswrapper[4988]: I1008 18:27:37.069379 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tvmgc\" (UniqueName: \"kubernetes.io/projected/b5dc1ee3-4897-45a2-ad00-0b91fbb2ae76-kube-api-access-tvmgc\") pod \"ovsdbserver-nb-0\" (UID: \"b5dc1ee3-4897-45a2-ad00-0b91fbb2ae76\") " pod="openstack/ovsdbserver-nb-0" Oct 08 18:27:37 crc kubenswrapper[4988]: I1008 18:27:37.070379 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/b5dc1ee3-4897-45a2-ad00-0b91fbb2ae76-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"b5dc1ee3-4897-45a2-ad00-0b91fbb2ae76\") " pod="openstack/ovsdbserver-nb-0" Oct 08 18:27:37 crc kubenswrapper[4988]: I1008 18:27:37.071221 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"ovsdbserver-nb-0\" (UID: \"b5dc1ee3-4897-45a2-ad00-0b91fbb2ae76\") " pod="openstack/ovsdbserver-nb-0" Oct 08 18:27:37 crc kubenswrapper[4988]: I1008 18:27:37.073329 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b5dc1ee3-4897-45a2-ad00-0b91fbb2ae76-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"b5dc1ee3-4897-45a2-ad00-0b91fbb2ae76\") " pod="openstack/ovsdbserver-nb-0" Oct 08 18:27:37 crc kubenswrapper[4988]: I1008 18:27:37.075852 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/b5dc1ee3-4897-45a2-ad00-0b91fbb2ae76-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"b5dc1ee3-4897-45a2-ad00-0b91fbb2ae76\") " pod="openstack/ovsdbserver-nb-0" Oct 08 18:27:37 crc kubenswrapper[4988]: I1008 18:27:37.142851 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Oct 08 18:27:37 crc kubenswrapper[4988]: I1008 18:27:37.151487 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/1f36756f-acb3-439d-b10e-13573cbd252a-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"1f36756f-acb3-439d-b10e-13573cbd252a\") " pod="openstack/ovsdbserver-sb-0" Oct 08 18:27:37 crc kubenswrapper[4988]: I1008 18:27:37.151554 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1f36756f-acb3-439d-b10e-13573cbd252a-config\") pod \"ovsdbserver-sb-0\" (UID: \"1f36756f-acb3-439d-b10e-13573cbd252a\") " pod="openstack/ovsdbserver-sb-0" Oct 08 18:27:37 crc kubenswrapper[4988]: I1008 18:27:37.151579 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1f36756f-acb3-439d-b10e-13573cbd252a-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"1f36756f-acb3-439d-b10e-13573cbd252a\") " pod="openstack/ovsdbserver-sb-0" Oct 08 18:27:37 crc kubenswrapper[4988]: I1008 18:27:37.151620 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"ovsdbserver-sb-0\" (UID: \"1f36756f-acb3-439d-b10e-13573cbd252a\") " pod="openstack/ovsdbserver-sb-0" Oct 08 18:27:37 crc kubenswrapper[4988]: I1008 18:27:37.151637 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/1f36756f-acb3-439d-b10e-13573cbd252a-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"1f36756f-acb3-439d-b10e-13573cbd252a\") " pod="openstack/ovsdbserver-sb-0" Oct 08 18:27:37 crc kubenswrapper[4988]: I1008 18:27:37.151663 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/1f36756f-acb3-439d-b10e-13573cbd252a-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"1f36756f-acb3-439d-b10e-13573cbd252a\") " pod="openstack/ovsdbserver-sb-0" Oct 08 18:27:37 crc kubenswrapper[4988]: I1008 18:27:37.151688 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/1f36756f-acb3-439d-b10e-13573cbd252a-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"1f36756f-acb3-439d-b10e-13573cbd252a\") " pod="openstack/ovsdbserver-sb-0" Oct 08 18:27:37 crc kubenswrapper[4988]: I1008 18:27:37.151708 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w9pgj\" (UniqueName: \"kubernetes.io/projected/1f36756f-acb3-439d-b10e-13573cbd252a-kube-api-access-w9pgj\") pod \"ovsdbserver-sb-0\" (UID: \"1f36756f-acb3-439d-b10e-13573cbd252a\") " pod="openstack/ovsdbserver-sb-0" Oct 08 18:27:37 crc kubenswrapper[4988]: I1008 18:27:37.252840 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"ovsdbserver-sb-0\" (UID: \"1f36756f-acb3-439d-b10e-13573cbd252a\") " pod="openstack/ovsdbserver-sb-0" Oct 08 18:27:37 crc kubenswrapper[4988]: I1008 18:27:37.252890 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/1f36756f-acb3-439d-b10e-13573cbd252a-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"1f36756f-acb3-439d-b10e-13573cbd252a\") " pod="openstack/ovsdbserver-sb-0" Oct 08 18:27:37 crc kubenswrapper[4988]: I1008 18:27:37.252931 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/1f36756f-acb3-439d-b10e-13573cbd252a-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"1f36756f-acb3-439d-b10e-13573cbd252a\") " pod="openstack/ovsdbserver-sb-0" Oct 08 18:27:37 crc kubenswrapper[4988]: I1008 18:27:37.252959 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/1f36756f-acb3-439d-b10e-13573cbd252a-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"1f36756f-acb3-439d-b10e-13573cbd252a\") " pod="openstack/ovsdbserver-sb-0" Oct 08 18:27:37 crc kubenswrapper[4988]: I1008 18:27:37.252985 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w9pgj\" (UniqueName: \"kubernetes.io/projected/1f36756f-acb3-439d-b10e-13573cbd252a-kube-api-access-w9pgj\") pod \"ovsdbserver-sb-0\" (UID: \"1f36756f-acb3-439d-b10e-13573cbd252a\") " pod="openstack/ovsdbserver-sb-0" Oct 08 18:27:37 crc kubenswrapper[4988]: I1008 18:27:37.253043 4988 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"ovsdbserver-sb-0\" (UID: \"1f36756f-acb3-439d-b10e-13573cbd252a\") device mount path \"/mnt/openstack/pv03\"" pod="openstack/ovsdbserver-sb-0" Oct 08 18:27:37 crc kubenswrapper[4988]: I1008 18:27:37.253058 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/1f36756f-acb3-439d-b10e-13573cbd252a-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"1f36756f-acb3-439d-b10e-13573cbd252a\") " pod="openstack/ovsdbserver-sb-0" Oct 08 18:27:37 crc kubenswrapper[4988]: I1008 18:27:37.253121 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1f36756f-acb3-439d-b10e-13573cbd252a-config\") pod \"ovsdbserver-sb-0\" (UID: \"1f36756f-acb3-439d-b10e-13573cbd252a\") " pod="openstack/ovsdbserver-sb-0" Oct 08 18:27:37 crc kubenswrapper[4988]: I1008 18:27:37.253148 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1f36756f-acb3-439d-b10e-13573cbd252a-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"1f36756f-acb3-439d-b10e-13573cbd252a\") " pod="openstack/ovsdbserver-sb-0" Oct 08 18:27:37 crc kubenswrapper[4988]: I1008 18:27:37.262066 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1f36756f-acb3-439d-b10e-13573cbd252a-config\") pod \"ovsdbserver-sb-0\" (UID: \"1f36756f-acb3-439d-b10e-13573cbd252a\") " pod="openstack/ovsdbserver-sb-0" Oct 08 18:27:37 crc kubenswrapper[4988]: I1008 18:27:37.263348 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/1f36756f-acb3-439d-b10e-13573cbd252a-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"1f36756f-acb3-439d-b10e-13573cbd252a\") " pod="openstack/ovsdbserver-sb-0" Oct 08 18:27:37 crc kubenswrapper[4988]: I1008 18:27:37.263796 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/1f36756f-acb3-439d-b10e-13573cbd252a-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"1f36756f-acb3-439d-b10e-13573cbd252a\") " pod="openstack/ovsdbserver-sb-0" Oct 08 18:27:37 crc kubenswrapper[4988]: I1008 18:27:37.268286 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1f36756f-acb3-439d-b10e-13573cbd252a-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"1f36756f-acb3-439d-b10e-13573cbd252a\") " pod="openstack/ovsdbserver-sb-0" Oct 08 18:27:37 crc kubenswrapper[4988]: I1008 18:27:37.274419 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/1f36756f-acb3-439d-b10e-13573cbd252a-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"1f36756f-acb3-439d-b10e-13573cbd252a\") " pod="openstack/ovsdbserver-sb-0" Oct 08 18:27:37 crc kubenswrapper[4988]: I1008 18:27:37.275110 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/1f36756f-acb3-439d-b10e-13573cbd252a-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"1f36756f-acb3-439d-b10e-13573cbd252a\") " pod="openstack/ovsdbserver-sb-0" Oct 08 18:27:37 crc kubenswrapper[4988]: I1008 18:27:37.295172 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w9pgj\" (UniqueName: \"kubernetes.io/projected/1f36756f-acb3-439d-b10e-13573cbd252a-kube-api-access-w9pgj\") pod \"ovsdbserver-sb-0\" (UID: \"1f36756f-acb3-439d-b10e-13573cbd252a\") " pod="openstack/ovsdbserver-sb-0" Oct 08 18:27:37 crc kubenswrapper[4988]: I1008 18:27:37.326624 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"ovsdbserver-sb-0\" (UID: \"1f36756f-acb3-439d-b10e-13573cbd252a\") " pod="openstack/ovsdbserver-sb-0" Oct 08 18:27:37 crc kubenswrapper[4988]: I1008 18:27:37.336203 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Oct 08 18:27:45 crc kubenswrapper[4988]: E1008 18:27:45.915492 4988 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-mariadb@sha256:56b75d97f4a48c8cf58b3a7c18c43618efb308bf0188124f6301142e61299b0c" Oct 08 18:27:45 crc kubenswrapper[4988]: E1008 18:27:45.916165 4988 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:mysql-bootstrap,Image:quay.io/podified-antelope-centos9/openstack-mariadb@sha256:56b75d97f4a48c8cf58b3a7c18c43618efb308bf0188124f6301142e61299b0c,Command:[bash /var/lib/operator-scripts/mysql_bootstrap.sh],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:True,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},EnvVar{Name:DB_ROOT_PASSWORD,Value:,ValueFrom:&EnvVarSource{FieldRef:nil,ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:&SecretKeySelector{LocalObjectReference:LocalObjectReference{Name:osp-secret,},Key:DbRootPassword,Optional:nil,},},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:mysql-db,ReadOnly:false,MountPath:/var/lib/mysql,SubPath:mysql,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data-default,ReadOnly:true,MountPath:/var/lib/config-data/default,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data-generated,ReadOnly:false,MountPath:/var/lib/config-data/generated,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:secrets,ReadOnly:true,MountPath:/var/lib/secrets,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:operator-scripts,ReadOnly:true,MountPath:/var/lib/operator-scripts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kolla-config,ReadOnly:true,MountPath:/var/lib/kolla/config_files,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-gz796,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod openstack-cell1-galera-0_openstack(3503942b-2825-4006-80e8-8c4610b89997): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 08 18:27:45 crc kubenswrapper[4988]: E1008 18:27:45.917284 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"mysql-bootstrap\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/openstack-cell1-galera-0" podUID="3503942b-2825-4006-80e8-8c4610b89997" Oct 08 18:27:45 crc kubenswrapper[4988]: E1008 18:27:45.937797 4988 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-rabbitmq@sha256:adcdeb8ecd601fb03c3b0901d5b5111af2ca48f7dd443e22224db6daaf08f5d0" Oct 08 18:27:45 crc kubenswrapper[4988]: E1008 18:27:45.937967 4988 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:setup-container,Image:quay.io/podified-antelope-centos9/openstack-rabbitmq@sha256:adcdeb8ecd601fb03c3b0901d5b5111af2ca48f7dd443e22224db6daaf08f5d0,Command:[sh -c cp /tmp/erlang-cookie-secret/.erlang.cookie /var/lib/rabbitmq/.erlang.cookie && chmod 600 /var/lib/rabbitmq/.erlang.cookie ; cp /tmp/rabbitmq-plugins/enabled_plugins /operator/enabled_plugins ; echo '[default]' > /var/lib/rabbitmq/.rabbitmqadmin.conf && sed -e 's/default_user/username/' -e 's/default_pass/password/' /tmp/default_user.conf >> /var/lib/rabbitmq/.rabbitmqadmin.conf && chmod 600 /var/lib/rabbitmq/.rabbitmqadmin.conf ; sleep 30],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{20 -3} {} 20m DecimalSI},memory: {{67108864 0} {} BinarySI},},Requests:ResourceList{cpu: {{20 -3} {} 20m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:plugins-conf,ReadOnly:false,MountPath:/tmp/rabbitmq-plugins/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:rabbitmq-erlang-cookie,ReadOnly:false,MountPath:/var/lib/rabbitmq/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:erlang-cookie-secret,ReadOnly:false,MountPath:/tmp/erlang-cookie-secret/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:rabbitmq-plugins,ReadOnly:false,MountPath:/operator,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:persistence,ReadOnly:false,MountPath:/var/lib/rabbitmq/mnesia/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:rabbitmq-confd,ReadOnly:false,MountPath:/tmp/default_user.conf,SubPath:default_user.conf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-zv5wc,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod rabbitmq-cell1-server-0_openstack(e5a8f819-c91d-429e-b848-e9c444739623): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 08 18:27:45 crc kubenswrapper[4988]: E1008 18:27:45.939288 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"setup-container\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/rabbitmq-cell1-server-0" podUID="e5a8f819-c91d-429e-b848-e9c444739623" Oct 08 18:27:45 crc kubenswrapper[4988]: E1008 18:27:45.977113 4988 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-mariadb@sha256:56b75d97f4a48c8cf58b3a7c18c43618efb308bf0188124f6301142e61299b0c" Oct 08 18:27:45 crc kubenswrapper[4988]: E1008 18:27:45.977293 4988 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:mysql-bootstrap,Image:quay.io/podified-antelope-centos9/openstack-mariadb@sha256:56b75d97f4a48c8cf58b3a7c18c43618efb308bf0188124f6301142e61299b0c,Command:[bash /var/lib/operator-scripts/mysql_bootstrap.sh],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:True,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},EnvVar{Name:DB_ROOT_PASSWORD,Value:,ValueFrom:&EnvVarSource{FieldRef:nil,ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:&SecretKeySelector{LocalObjectReference:LocalObjectReference{Name:osp-secret,},Key:DbRootPassword,Optional:nil,},},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:mysql-db,ReadOnly:false,MountPath:/var/lib/mysql,SubPath:mysql,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data-default,ReadOnly:true,MountPath:/var/lib/config-data/default,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data-generated,ReadOnly:false,MountPath:/var/lib/config-data/generated,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:secrets,ReadOnly:true,MountPath:/var/lib/secrets,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:operator-scripts,ReadOnly:true,MountPath:/var/lib/operator-scripts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kolla-config,ReadOnly:true,MountPath:/var/lib/kolla/config_files,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-qszql,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod openstack-galera-0_openstack(72c64965-0b57-43e1-8d18-2bce6cabbfa1): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 08 18:27:45 crc kubenswrapper[4988]: E1008 18:27:45.978612 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"mysql-bootstrap\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/openstack-galera-0" podUID="72c64965-0b57-43e1-8d18-2bce6cabbfa1" Oct 08 18:27:46 crc kubenswrapper[4988]: E1008 18:27:46.417662 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"setup-container\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-rabbitmq@sha256:adcdeb8ecd601fb03c3b0901d5b5111af2ca48f7dd443e22224db6daaf08f5d0\\\"\"" pod="openstack/rabbitmq-cell1-server-0" podUID="e5a8f819-c91d-429e-b848-e9c444739623" Oct 08 18:27:46 crc kubenswrapper[4988]: E1008 18:27:46.417754 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"mysql-bootstrap\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-mariadb@sha256:56b75d97f4a48c8cf58b3a7c18c43618efb308bf0188124f6301142e61299b0c\\\"\"" pod="openstack/openstack-cell1-galera-0" podUID="3503942b-2825-4006-80e8-8c4610b89997" Oct 08 18:27:46 crc kubenswrapper[4988]: E1008 18:27:46.417809 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"mysql-bootstrap\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-mariadb@sha256:56b75d97f4a48c8cf58b3a7c18c43618efb308bf0188124f6301142e61299b0c\\\"\"" pod="openstack/openstack-galera-0" podUID="72c64965-0b57-43e1-8d18-2bce6cabbfa1" Oct 08 18:27:48 crc kubenswrapper[4988]: E1008 18:27:48.516416 4988 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-memcached@sha256:773daada6402d9cad089cdc809d6c0335456d057ac1a25441ab5d82add2f70f4" Oct 08 18:27:48 crc kubenswrapper[4988]: E1008 18:27:48.516922 4988 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:memcached,Image:quay.io/podified-antelope-centos9/openstack-memcached@sha256:773daada6402d9cad089cdc809d6c0335456d057ac1a25441ab5d82add2f70f4,Command:[/usr/bin/dumb-init -- /usr/local/bin/kolla_start],Args:[],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:memcached,HostPort:0,ContainerPort:11211,Protocol:TCP,HostIP:,},ContainerPort{Name:memcached-tls,HostPort:0,ContainerPort:11212,Protocol:TCP,HostIP:,},},Env:[]EnvVar{EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},EnvVar{Name:POD_IPS,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIPs,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},EnvVar{Name:CONFIG_HASH,Value:n5fh74hd5h57ch98h665h5d5h675h5c8hc4h7dh55fh586hdfh595hf5h555h66bhc7h5b8h65ch65dh5c4h559h556h645h58h97h548h687h667h566q,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/src,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kolla-config,ReadOnly:true,MountPath:/var/lib/kolla/config_files,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:memcached-tls-certs,ReadOnly:true,MountPath:/var/lib/config-data/tls/certs/memcached.crt,SubPath:tls.crt,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:memcached-tls-certs,ReadOnly:true,MountPath:/var/lib/config-data/tls/private/memcached.key,SubPath:tls.key,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-n9xkg,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:nil,TCPSocket:&TCPSocketAction{Port:{0 11211 },Host:,},GRPC:nil,},InitialDelaySeconds:3,TimeoutSeconds:5,PeriodSeconds:3,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:nil,TCPSocket:&TCPSocketAction{Port:{0 11211 },Host:,},GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:5,PeriodSeconds:5,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42457,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:*42457,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod memcached-0_openstack(32bb7401-6292-4ca6-98f4-15522125c3ee): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 08 18:27:48 crc kubenswrapper[4988]: E1008 18:27:48.518172 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"memcached\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/memcached-0" podUID="32bb7401-6292-4ca6-98f4-15522125c3ee" Oct 08 18:27:49 crc kubenswrapper[4988]: E1008 18:27:49.439402 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"memcached\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-memcached@sha256:773daada6402d9cad089cdc809d6c0335456d057ac1a25441ab5d82add2f70f4\\\"\"" pod="openstack/memcached-0" podUID="32bb7401-6292-4ca6-98f4-15522125c3ee" Oct 08 18:27:52 crc kubenswrapper[4988]: I1008 18:27:52.528132 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 08 18:27:52 crc kubenswrapper[4988]: E1008 18:27:52.946337 4988 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server@sha256:c4e71b2158fd939dad8b8e705273493051d3023273d23b279f2699dce6db33df" Oct 08 18:27:52 crc kubenswrapper[4988]: E1008 18:27:52.946542 4988 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server@sha256:c4e71b2158fd939dad8b8e705273493051d3023273d23b279f2699dce6db33df,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n659h4h664hbh658h587h67ch89h587h8fh679hc6hf9h55fh644h5d5h698h68dh5cdh5ffh669h54ch9h689hb8hd4h5bfhd8h5d7h5fh665h574q,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-tbtxg,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-77597f887-54xgv_openstack(5d556b85-3337-4855-9f43-a38946207378): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 08 18:27:52 crc kubenswrapper[4988]: E1008 18:27:52.947797 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-77597f887-54xgv" podUID="5d556b85-3337-4855-9f43-a38946207378" Oct 08 18:27:52 crc kubenswrapper[4988]: E1008 18:27:52.963145 4988 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server@sha256:c4e71b2158fd939dad8b8e705273493051d3023273d23b279f2699dce6db33df" Oct 08 18:27:52 crc kubenswrapper[4988]: E1008 18:27:52.963337 4988 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server@sha256:c4e71b2158fd939dad8b8e705273493051d3023273d23b279f2699dce6db33df,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n68chd6h679hbfh55fhc6h5ffh5d8h94h56ch589hb4hc5h57bh677hcdh655h8dh667h675h654h66ch567h8fh659h5b4h675h566h55bh54h67dh6dq,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-cp9pt,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-644597f84c-nkqvt_openstack(e04834b5-6114-4fba-b3ac-9b6405b6712f): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 08 18:27:52 crc kubenswrapper[4988]: E1008 18:27:52.964070 4988 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server@sha256:c4e71b2158fd939dad8b8e705273493051d3023273d23b279f2699dce6db33df" Oct 08 18:27:52 crc kubenswrapper[4988]: E1008 18:27:52.964321 4988 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server@sha256:c4e71b2158fd939dad8b8e705273493051d3023273d23b279f2699dce6db33df,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:nffh5bdhf4h5f8h79h55h77h58fh56dh7bh6fh578hbch55dh68h56bhd9h65dh57ch658hc9h566h666h688h58h65dh684h5d7h6ch575h5d6h88q,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-6gf4v,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-7bfcb9d745-v56k2_openstack(f0c8efc3-858a-4d1d-b031-86d9e74a3d9c): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 08 18:27:52 crc kubenswrapper[4988]: E1008 18:27:52.965252 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-644597f84c-nkqvt" podUID="e04834b5-6114-4fba-b3ac-9b6405b6712f" Oct 08 18:27:52 crc kubenswrapper[4988]: E1008 18:27:52.965721 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-7bfcb9d745-v56k2" podUID="f0c8efc3-858a-4d1d-b031-86d9e74a3d9c" Oct 08 18:27:52 crc kubenswrapper[4988]: E1008 18:27:52.974995 4988 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server@sha256:c4e71b2158fd939dad8b8e705273493051d3023273d23b279f2699dce6db33df" Oct 08 18:27:52 crc kubenswrapper[4988]: E1008 18:27:52.975205 4988 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server@sha256:c4e71b2158fd939dad8b8e705273493051d3023273d23b279f2699dce6db33df,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:ndfhb5h667h568h584h5f9h58dh565h664h587h597h577h64bh5c4h66fh647hbdh68ch5c5h68dh686h5f7h64hd7hc6h55fh57bh98h57fh87h5fh57fq,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-9l88k,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-758b79db4c-qv72t_openstack(a61127a0-ab84-41a9-934a-68b130206578): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 08 18:27:52 crc kubenswrapper[4988]: E1008 18:27:52.976684 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-758b79db4c-qv72t" podUID="a61127a0-ab84-41a9-934a-68b130206578" Oct 08 18:27:53 crc kubenswrapper[4988]: I1008 18:27:53.450072 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-gp8f4"] Oct 08 18:27:53 crc kubenswrapper[4988]: I1008 18:27:53.475261 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-kc7fw"] Oct 08 18:27:53 crc kubenswrapper[4988]: I1008 18:27:53.479607 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"731acc53-25a1-43a4-b9a8-16de25b1e7b1","Type":"ContainerStarted","Data":"b1e9368f2c64f0a3360049bc186a23f993a494e9080ad57b8a79456a4d59b220"} Oct 08 18:27:53 crc kubenswrapper[4988]: E1008 18:27:53.481521 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-neutron-server@sha256:c4e71b2158fd939dad8b8e705273493051d3023273d23b279f2699dce6db33df\\\"\"" pod="openstack/dnsmasq-dns-77597f887-54xgv" podUID="5d556b85-3337-4855-9f43-a38946207378" Oct 08 18:27:53 crc kubenswrapper[4988]: E1008 18:27:53.481664 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-neutron-server@sha256:c4e71b2158fd939dad8b8e705273493051d3023273d23b279f2699dce6db33df\\\"\"" pod="openstack/dnsmasq-dns-644597f84c-nkqvt" podUID="e04834b5-6114-4fba-b3ac-9b6405b6712f" Oct 08 18:27:53 crc kubenswrapper[4988]: I1008 18:27:53.601833 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Oct 08 18:27:54 crc kubenswrapper[4988]: I1008 18:27:54.188856 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Oct 08 18:27:54 crc kubenswrapper[4988]: W1008 18:27:54.264884 4988 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb5dc1ee3_4897_45a2_ad00_0b91fbb2ae76.slice/crio-74fae05a8629c65a280b6c23f9ffb0aee300b29c4f7758a9298dae712985fc5d WatchSource:0}: Error finding container 74fae05a8629c65a280b6c23f9ffb0aee300b29c4f7758a9298dae712985fc5d: Status 404 returned error can't find the container with id 74fae05a8629c65a280b6c23f9ffb0aee300b29c4f7758a9298dae712985fc5d Oct 08 18:27:54 crc kubenswrapper[4988]: W1008 18:27:54.268623 4988 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1f36756f_acb3_439d_b10e_13573cbd252a.slice/crio-b1a4cd2f1dc8c650e6916cf81b398b6e7853f8d12cb351a87c1894def087333c WatchSource:0}: Error finding container b1a4cd2f1dc8c650e6916cf81b398b6e7853f8d12cb351a87c1894def087333c: Status 404 returned error can't find the container with id b1a4cd2f1dc8c650e6916cf81b398b6e7853f8d12cb351a87c1894def087333c Oct 08 18:27:54 crc kubenswrapper[4988]: I1008 18:27:54.328265 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-758b79db4c-qv72t" Oct 08 18:27:54 crc kubenswrapper[4988]: I1008 18:27:54.334570 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7bfcb9d745-v56k2" Oct 08 18:27:54 crc kubenswrapper[4988]: I1008 18:27:54.439280 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6gf4v\" (UniqueName: \"kubernetes.io/projected/f0c8efc3-858a-4d1d-b031-86d9e74a3d9c-kube-api-access-6gf4v\") pod \"f0c8efc3-858a-4d1d-b031-86d9e74a3d9c\" (UID: \"f0c8efc3-858a-4d1d-b031-86d9e74a3d9c\") " Oct 08 18:27:54 crc kubenswrapper[4988]: I1008 18:27:54.439415 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a61127a0-ab84-41a9-934a-68b130206578-dns-svc\") pod \"a61127a0-ab84-41a9-934a-68b130206578\" (UID: \"a61127a0-ab84-41a9-934a-68b130206578\") " Oct 08 18:27:54 crc kubenswrapper[4988]: I1008 18:27:54.439490 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9l88k\" (UniqueName: \"kubernetes.io/projected/a61127a0-ab84-41a9-934a-68b130206578-kube-api-access-9l88k\") pod \"a61127a0-ab84-41a9-934a-68b130206578\" (UID: \"a61127a0-ab84-41a9-934a-68b130206578\") " Oct 08 18:27:54 crc kubenswrapper[4988]: I1008 18:27:54.439604 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a61127a0-ab84-41a9-934a-68b130206578-config\") pod \"a61127a0-ab84-41a9-934a-68b130206578\" (UID: \"a61127a0-ab84-41a9-934a-68b130206578\") " Oct 08 18:27:54 crc kubenswrapper[4988]: I1008 18:27:54.439628 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f0c8efc3-858a-4d1d-b031-86d9e74a3d9c-config\") pod \"f0c8efc3-858a-4d1d-b031-86d9e74a3d9c\" (UID: \"f0c8efc3-858a-4d1d-b031-86d9e74a3d9c\") " Oct 08 18:27:54 crc kubenswrapper[4988]: I1008 18:27:54.440210 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a61127a0-ab84-41a9-934a-68b130206578-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "a61127a0-ab84-41a9-934a-68b130206578" (UID: "a61127a0-ab84-41a9-934a-68b130206578"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 18:27:54 crc kubenswrapper[4988]: I1008 18:27:54.440631 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f0c8efc3-858a-4d1d-b031-86d9e74a3d9c-config" (OuterVolumeSpecName: "config") pod "f0c8efc3-858a-4d1d-b031-86d9e74a3d9c" (UID: "f0c8efc3-858a-4d1d-b031-86d9e74a3d9c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 18:27:54 crc kubenswrapper[4988]: I1008 18:27:54.440684 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a61127a0-ab84-41a9-934a-68b130206578-config" (OuterVolumeSpecName: "config") pod "a61127a0-ab84-41a9-934a-68b130206578" (UID: "a61127a0-ab84-41a9-934a-68b130206578"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 18:27:54 crc kubenswrapper[4988]: I1008 18:27:54.444878 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a61127a0-ab84-41a9-934a-68b130206578-kube-api-access-9l88k" (OuterVolumeSpecName: "kube-api-access-9l88k") pod "a61127a0-ab84-41a9-934a-68b130206578" (UID: "a61127a0-ab84-41a9-934a-68b130206578"). InnerVolumeSpecName "kube-api-access-9l88k". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:27:54 crc kubenswrapper[4988]: I1008 18:27:54.444985 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f0c8efc3-858a-4d1d-b031-86d9e74a3d9c-kube-api-access-6gf4v" (OuterVolumeSpecName: "kube-api-access-6gf4v") pod "f0c8efc3-858a-4d1d-b031-86d9e74a3d9c" (UID: "f0c8efc3-858a-4d1d-b031-86d9e74a3d9c"). InnerVolumeSpecName "kube-api-access-6gf4v". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:27:54 crc kubenswrapper[4988]: I1008 18:27:54.488128 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-gp8f4" event={"ID":"c285e09d-c2f5-4642-9e6b-59b0ed3ed651","Type":"ContainerStarted","Data":"3623831a4118466e308ca5fe5cc4eff6cb062697a66654a0488b0c44e1b18285"} Oct 08 18:27:54 crc kubenswrapper[4988]: I1008 18:27:54.489285 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"b5dc1ee3-4897-45a2-ad00-0b91fbb2ae76","Type":"ContainerStarted","Data":"74fae05a8629c65a280b6c23f9ffb0aee300b29c4f7758a9298dae712985fc5d"} Oct 08 18:27:54 crc kubenswrapper[4988]: I1008 18:27:54.490823 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"1f36756f-acb3-439d-b10e-13573cbd252a","Type":"ContainerStarted","Data":"b1a4cd2f1dc8c650e6916cf81b398b6e7853f8d12cb351a87c1894def087333c"} Oct 08 18:27:54 crc kubenswrapper[4988]: I1008 18:27:54.492358 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7bfcb9d745-v56k2" event={"ID":"f0c8efc3-858a-4d1d-b031-86d9e74a3d9c","Type":"ContainerDied","Data":"bee727503b080650e80cd50d20b61054f08be6c7aa34fae491b3f54e202f56c4"} Oct 08 18:27:54 crc kubenswrapper[4988]: I1008 18:27:54.492423 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7bfcb9d745-v56k2" Oct 08 18:27:54 crc kubenswrapper[4988]: I1008 18:27:54.493658 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-kc7fw" event={"ID":"f1628013-e0b9-4c31-80c0-91aabaaef0f6","Type":"ContainerStarted","Data":"8464f2b7c6120e931f4c293370cfc0d2b039813f7d469e2e29ad85a2d9185c98"} Oct 08 18:27:54 crc kubenswrapper[4988]: I1008 18:27:54.496615 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"830d6ec5-0015-4a88-b8e8-0fd47f64fd46","Type":"ContainerStarted","Data":"1002767e17c64f7c9060cbd0059630fd4f3dc91d7c2f7558f00143736e0751e0"} Oct 08 18:27:54 crc kubenswrapper[4988]: I1008 18:27:54.497559 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-758b79db4c-qv72t" event={"ID":"a61127a0-ab84-41a9-934a-68b130206578","Type":"ContainerDied","Data":"4083d3f29623ad81e469ab7e4ac32353d8400e6a5e9903c66b25644d4dc0fbbc"} Oct 08 18:27:54 crc kubenswrapper[4988]: I1008 18:27:54.497608 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-758b79db4c-qv72t" Oct 08 18:27:54 crc kubenswrapper[4988]: I1008 18:27:54.543502 4988 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a61127a0-ab84-41a9-934a-68b130206578-config\") on node \"crc\" DevicePath \"\"" Oct 08 18:27:54 crc kubenswrapper[4988]: I1008 18:27:54.543537 4988 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f0c8efc3-858a-4d1d-b031-86d9e74a3d9c-config\") on node \"crc\" DevicePath \"\"" Oct 08 18:27:54 crc kubenswrapper[4988]: I1008 18:27:54.543550 4988 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6gf4v\" (UniqueName: \"kubernetes.io/projected/f0c8efc3-858a-4d1d-b031-86d9e74a3d9c-kube-api-access-6gf4v\") on node \"crc\" DevicePath \"\"" Oct 08 18:27:54 crc kubenswrapper[4988]: I1008 18:27:54.543565 4988 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a61127a0-ab84-41a9-934a-68b130206578-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 08 18:27:54 crc kubenswrapper[4988]: I1008 18:27:54.543576 4988 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9l88k\" (UniqueName: \"kubernetes.io/projected/a61127a0-ab84-41a9-934a-68b130206578-kube-api-access-9l88k\") on node \"crc\" DevicePath \"\"" Oct 08 18:27:54 crc kubenswrapper[4988]: I1008 18:27:54.568118 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-758b79db4c-qv72t"] Oct 08 18:27:54 crc kubenswrapper[4988]: I1008 18:27:54.574163 4988 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-758b79db4c-qv72t"] Oct 08 18:27:54 crc kubenswrapper[4988]: I1008 18:27:54.593116 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7bfcb9d745-v56k2"] Oct 08 18:27:54 crc kubenswrapper[4988]: I1008 18:27:54.602553 4988 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-7bfcb9d745-v56k2"] Oct 08 18:27:55 crc kubenswrapper[4988]: I1008 18:27:55.270699 4988 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a61127a0-ab84-41a9-934a-68b130206578" path="/var/lib/kubelet/pods/a61127a0-ab84-41a9-934a-68b130206578/volumes" Oct 08 18:27:55 crc kubenswrapper[4988]: I1008 18:27:55.271325 4988 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f0c8efc3-858a-4d1d-b031-86d9e74a3d9c" path="/var/lib/kubelet/pods/f0c8efc3-858a-4d1d-b031-86d9e74a3d9c/volumes" Oct 08 18:27:57 crc kubenswrapper[4988]: I1008 18:27:57.534726 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-gp8f4" event={"ID":"c285e09d-c2f5-4642-9e6b-59b0ed3ed651","Type":"ContainerStarted","Data":"f3d6a7298803e1d31ec12101ceb0dd0c2936d1871327a38880d53ef51a7deceb"} Oct 08 18:27:57 crc kubenswrapper[4988]: I1008 18:27:57.535419 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-gp8f4" Oct 08 18:27:57 crc kubenswrapper[4988]: I1008 18:27:57.536614 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"731acc53-25a1-43a4-b9a8-16de25b1e7b1","Type":"ContainerStarted","Data":"bcd9a23419ff288fe7e9ecd15147e5affee5f1bc37f5d8bf5a0021d7900242f6"} Oct 08 18:27:57 crc kubenswrapper[4988]: I1008 18:27:57.536831 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Oct 08 18:27:57 crc kubenswrapper[4988]: I1008 18:27:57.538270 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"b5dc1ee3-4897-45a2-ad00-0b91fbb2ae76","Type":"ContainerStarted","Data":"74829553075c7054526031a117b72832271514093fbe3617d9278cd92e26275b"} Oct 08 18:27:57 crc kubenswrapper[4988]: I1008 18:27:57.540270 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"1f36756f-acb3-439d-b10e-13573cbd252a","Type":"ContainerStarted","Data":"a41e320aa9f3ac4edb764f4ac4f3c561a463787b781e74385ca8530f54e7b28e"} Oct 08 18:27:57 crc kubenswrapper[4988]: I1008 18:27:57.541732 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-kc7fw" event={"ID":"f1628013-e0b9-4c31-80c0-91aabaaef0f6","Type":"ContainerStarted","Data":"8d3bc1f9ff6541935bebe925efdef3e65cd202a230aed96556f734e8974eaf13"} Oct 08 18:27:57 crc kubenswrapper[4988]: I1008 18:27:57.557676 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-gp8f4" podStartSLOduration=21.08332984 podStartE2EDuration="24.557657214s" podCreationTimestamp="2025-10-08 18:27:33 +0000 UTC" firstStartedPulling="2025-10-08 18:27:53.57247221 +0000 UTC m=+1019.022314980" lastFinishedPulling="2025-10-08 18:27:57.046799584 +0000 UTC m=+1022.496642354" observedRunningTime="2025-10-08 18:27:57.551090107 +0000 UTC m=+1023.000932877" watchObservedRunningTime="2025-10-08 18:27:57.557657214 +0000 UTC m=+1023.007499984" Oct 08 18:27:57 crc kubenswrapper[4988]: I1008 18:27:57.591087 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=24.443200758 podStartE2EDuration="28.591057466s" podCreationTimestamp="2025-10-08 18:27:29 +0000 UTC" firstStartedPulling="2025-10-08 18:27:52.897107649 +0000 UTC m=+1018.346950419" lastFinishedPulling="2025-10-08 18:27:57.044964357 +0000 UTC m=+1022.494807127" observedRunningTime="2025-10-08 18:27:57.583981863 +0000 UTC m=+1023.033824653" watchObservedRunningTime="2025-10-08 18:27:57.591057466 +0000 UTC m=+1023.040900236" Oct 08 18:27:58 crc kubenswrapper[4988]: I1008 18:27:58.553284 4988 generic.go:334] "Generic (PLEG): container finished" podID="f1628013-e0b9-4c31-80c0-91aabaaef0f6" containerID="8d3bc1f9ff6541935bebe925efdef3e65cd202a230aed96556f734e8974eaf13" exitCode=0 Oct 08 18:27:58 crc kubenswrapper[4988]: I1008 18:27:58.553349 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-kc7fw" event={"ID":"f1628013-e0b9-4c31-80c0-91aabaaef0f6","Type":"ContainerDied","Data":"8d3bc1f9ff6541935bebe925efdef3e65cd202a230aed96556f734e8974eaf13"} Oct 08 18:27:59 crc kubenswrapper[4988]: I1008 18:27:59.564137 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-kc7fw" event={"ID":"f1628013-e0b9-4c31-80c0-91aabaaef0f6","Type":"ContainerStarted","Data":"863aa8ddeda792a78ba5af321605974ac7082d24fc8fb6f0400148fcf3f06126"} Oct 08 18:28:00 crc kubenswrapper[4988]: I1008 18:28:00.573322 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"e5a8f819-c91d-429e-b848-e9c444739623","Type":"ContainerStarted","Data":"586d76b71bf326ca8c68b69f484afc42dbe51e835567ee76b7228cb18c8b1384"} Oct 08 18:28:00 crc kubenswrapper[4988]: I1008 18:28:00.576276 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-kc7fw" event={"ID":"f1628013-e0b9-4c31-80c0-91aabaaef0f6","Type":"ContainerStarted","Data":"8a5332d3d9aa50bb56f12e04b1b32520399363308f9e75d5510ddd01f41593c3"} Oct 08 18:28:00 crc kubenswrapper[4988]: I1008 18:28:00.576426 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-kc7fw" Oct 08 18:28:00 crc kubenswrapper[4988]: I1008 18:28:00.623327 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-ovs-kc7fw" podStartSLOduration=24.133361572 podStartE2EDuration="27.623305857s" podCreationTimestamp="2025-10-08 18:27:33 +0000 UTC" firstStartedPulling="2025-10-08 18:27:53.557228231 +0000 UTC m=+1019.007071001" lastFinishedPulling="2025-10-08 18:27:57.047172516 +0000 UTC m=+1022.497015286" observedRunningTime="2025-10-08 18:28:00.618553816 +0000 UTC m=+1026.068396606" watchObservedRunningTime="2025-10-08 18:28:00.623305857 +0000 UTC m=+1026.073148627" Oct 08 18:28:01 crc kubenswrapper[4988]: I1008 18:28:01.591268 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"72c64965-0b57-43e1-8d18-2bce6cabbfa1","Type":"ContainerStarted","Data":"4a95ce795323840905212e6edf0cc0a3e009ff2a132cb0c21d30fe459fa29c5d"} Oct 08 18:28:01 crc kubenswrapper[4988]: I1008 18:28:01.594235 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"b5dc1ee3-4897-45a2-ad00-0b91fbb2ae76","Type":"ContainerStarted","Data":"b5799148279b22815a3dbd06ed629a27e96e308b0f484f69614b62b235afbb9f"} Oct 08 18:28:01 crc kubenswrapper[4988]: I1008 18:28:01.596284 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"1f36756f-acb3-439d-b10e-13573cbd252a","Type":"ContainerStarted","Data":"e8f79299e7fcca754183f5f8a133b9bb867b1fa107190b02e938976ad478eb0b"} Oct 08 18:28:01 crc kubenswrapper[4988]: I1008 18:28:01.598231 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"32bb7401-6292-4ca6-98f4-15522125c3ee","Type":"ContainerStarted","Data":"dcd1682ce23dc072a1b6bf3c5aeff64ca69361e18fa5cdcb9712e65170ddc37d"} Oct 08 18:28:01 crc kubenswrapper[4988]: I1008 18:28:01.598546 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/memcached-0" Oct 08 18:28:01 crc kubenswrapper[4988]: I1008 18:28:01.600370 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"3503942b-2825-4006-80e8-8c4610b89997","Type":"ContainerStarted","Data":"cefcf980cfc56a4cdb19408f13a9a6a84705b300fdac96063e98a4090209d49e"} Oct 08 18:28:01 crc kubenswrapper[4988]: I1008 18:28:01.600711 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-kc7fw" Oct 08 18:28:01 crc kubenswrapper[4988]: I1008 18:28:01.648890 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-nb-0" podStartSLOduration=20.523211429 podStartE2EDuration="26.648853606s" podCreationTimestamp="2025-10-08 18:27:35 +0000 UTC" firstStartedPulling="2025-10-08 18:27:54.267156679 +0000 UTC m=+1019.716999449" lastFinishedPulling="2025-10-08 18:28:00.392798856 +0000 UTC m=+1025.842641626" observedRunningTime="2025-10-08 18:28:01.645972085 +0000 UTC m=+1027.095814875" watchObservedRunningTime="2025-10-08 18:28:01.648853606 +0000 UTC m=+1027.098696376" Oct 08 18:28:01 crc kubenswrapper[4988]: I1008 18:28:01.712779 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-sb-0" podStartSLOduration=20.622112634 podStartE2EDuration="26.712759388s" podCreationTimestamp="2025-10-08 18:27:35 +0000 UTC" firstStartedPulling="2025-10-08 18:27:54.273010874 +0000 UTC m=+1019.722853664" lastFinishedPulling="2025-10-08 18:28:00.363657648 +0000 UTC m=+1025.813500418" observedRunningTime="2025-10-08 18:28:01.705527571 +0000 UTC m=+1027.155370371" watchObservedRunningTime="2025-10-08 18:28:01.712759388 +0000 UTC m=+1027.162602168" Oct 08 18:28:01 crc kubenswrapper[4988]: I1008 18:28:01.728317 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/memcached-0" podStartSLOduration=2.170510964 podStartE2EDuration="33.728297288s" podCreationTimestamp="2025-10-08 18:27:28 +0000 UTC" firstStartedPulling="2025-10-08 18:27:29.173494563 +0000 UTC m=+994.623337333" lastFinishedPulling="2025-10-08 18:28:00.731280887 +0000 UTC m=+1026.181123657" observedRunningTime="2025-10-08 18:28:01.726896983 +0000 UTC m=+1027.176739773" watchObservedRunningTime="2025-10-08 18:28:01.728297288 +0000 UTC m=+1027.178140068" Oct 08 18:28:02 crc kubenswrapper[4988]: I1008 18:28:02.143713 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-nb-0" Oct 08 18:28:02 crc kubenswrapper[4988]: I1008 18:28:02.336814 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-sb-0" Oct 08 18:28:04 crc kubenswrapper[4988]: I1008 18:28:04.144056 4988 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-nb-0" Oct 08 18:28:04 crc kubenswrapper[4988]: I1008 18:28:04.190976 4988 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-nb-0" Oct 08 18:28:04 crc kubenswrapper[4988]: I1008 18:28:04.337263 4988 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-sb-0" Oct 08 18:28:04 crc kubenswrapper[4988]: I1008 18:28:04.378213 4988 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-sb-0" Oct 08 18:28:04 crc kubenswrapper[4988]: I1008 18:28:04.666492 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-nb-0" Oct 08 18:28:04 crc kubenswrapper[4988]: I1008 18:28:04.666914 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-sb-0" Oct 08 18:28:04 crc kubenswrapper[4988]: I1008 18:28:04.850375 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-644597f84c-nkqvt"] Oct 08 18:28:04 crc kubenswrapper[4988]: I1008 18:28:04.908695 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-6d8fd57975-7jpt4"] Oct 08 18:28:04 crc kubenswrapper[4988]: I1008 18:28:04.910319 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6d8fd57975-7jpt4" Oct 08 18:28:04 crc kubenswrapper[4988]: I1008 18:28:04.914987 4988 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-nb" Oct 08 18:28:04 crc kubenswrapper[4988]: I1008 18:28:04.923263 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-metrics-w2zhj"] Oct 08 18:28:04 crc kubenswrapper[4988]: I1008 18:28:04.924539 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-w2zhj" Oct 08 18:28:04 crc kubenswrapper[4988]: I1008 18:28:04.926181 4988 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-metrics-config" Oct 08 18:28:04 crc kubenswrapper[4988]: I1008 18:28:04.929617 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6d8fd57975-7jpt4"] Oct 08 18:28:04 crc kubenswrapper[4988]: I1008 18:28:04.951426 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-w2zhj"] Oct 08 18:28:05 crc kubenswrapper[4988]: I1008 18:28:05.034708 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b92d978b-87c6-4210-9a69-99596b05cc18-combined-ca-bundle\") pod \"ovn-controller-metrics-w2zhj\" (UID: \"b92d978b-87c6-4210-9a69-99596b05cc18\") " pod="openstack/ovn-controller-metrics-w2zhj" Oct 08 18:28:05 crc kubenswrapper[4988]: I1008 18:28:05.034755 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b9thk\" (UniqueName: \"kubernetes.io/projected/b92d978b-87c6-4210-9a69-99596b05cc18-kube-api-access-b9thk\") pod \"ovn-controller-metrics-w2zhj\" (UID: \"b92d978b-87c6-4210-9a69-99596b05cc18\") " pod="openstack/ovn-controller-metrics-w2zhj" Oct 08 18:28:05 crc kubenswrapper[4988]: I1008 18:28:05.034791 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/94afe69f-30c7-4023-a485-f16998ed7fdd-ovsdbserver-nb\") pod \"dnsmasq-dns-6d8fd57975-7jpt4\" (UID: \"94afe69f-30c7-4023-a485-f16998ed7fdd\") " pod="openstack/dnsmasq-dns-6d8fd57975-7jpt4" Oct 08 18:28:05 crc kubenswrapper[4988]: I1008 18:28:05.034819 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/b92d978b-87c6-4210-9a69-99596b05cc18-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-w2zhj\" (UID: \"b92d978b-87c6-4210-9a69-99596b05cc18\") " pod="openstack/ovn-controller-metrics-w2zhj" Oct 08 18:28:05 crc kubenswrapper[4988]: I1008 18:28:05.034839 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/94afe69f-30c7-4023-a485-f16998ed7fdd-dns-svc\") pod \"dnsmasq-dns-6d8fd57975-7jpt4\" (UID: \"94afe69f-30c7-4023-a485-f16998ed7fdd\") " pod="openstack/dnsmasq-dns-6d8fd57975-7jpt4" Oct 08 18:28:05 crc kubenswrapper[4988]: I1008 18:28:05.034867 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/94afe69f-30c7-4023-a485-f16998ed7fdd-config\") pod \"dnsmasq-dns-6d8fd57975-7jpt4\" (UID: \"94afe69f-30c7-4023-a485-f16998ed7fdd\") " pod="openstack/dnsmasq-dns-6d8fd57975-7jpt4" Oct 08 18:28:05 crc kubenswrapper[4988]: I1008 18:28:05.034885 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b92d978b-87c6-4210-9a69-99596b05cc18-config\") pod \"ovn-controller-metrics-w2zhj\" (UID: \"b92d978b-87c6-4210-9a69-99596b05cc18\") " pod="openstack/ovn-controller-metrics-w2zhj" Oct 08 18:28:05 crc kubenswrapper[4988]: I1008 18:28:05.034788 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-77597f887-54xgv"] Oct 08 18:28:05 crc kubenswrapper[4988]: I1008 18:28:05.034911 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/b92d978b-87c6-4210-9a69-99596b05cc18-ovs-rundir\") pod \"ovn-controller-metrics-w2zhj\" (UID: \"b92d978b-87c6-4210-9a69-99596b05cc18\") " pod="openstack/ovn-controller-metrics-w2zhj" Oct 08 18:28:05 crc kubenswrapper[4988]: I1008 18:28:05.034953 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zstlc\" (UniqueName: \"kubernetes.io/projected/94afe69f-30c7-4023-a485-f16998ed7fdd-kube-api-access-zstlc\") pod \"dnsmasq-dns-6d8fd57975-7jpt4\" (UID: \"94afe69f-30c7-4023-a485-f16998ed7fdd\") " pod="openstack/dnsmasq-dns-6d8fd57975-7jpt4" Oct 08 18:28:05 crc kubenswrapper[4988]: I1008 18:28:05.034979 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/b92d978b-87c6-4210-9a69-99596b05cc18-ovn-rundir\") pod \"ovn-controller-metrics-w2zhj\" (UID: \"b92d978b-87c6-4210-9a69-99596b05cc18\") " pod="openstack/ovn-controller-metrics-w2zhj" Oct 08 18:28:05 crc kubenswrapper[4988]: I1008 18:28:05.075605 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-northd-0"] Oct 08 18:28:05 crc kubenswrapper[4988]: I1008 18:28:05.077083 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Oct 08 18:28:05 crc kubenswrapper[4988]: I1008 18:28:05.079488 4988 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-scripts" Oct 08 18:28:05 crc kubenswrapper[4988]: I1008 18:28:05.079740 4988 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-config" Oct 08 18:28:05 crc kubenswrapper[4988]: I1008 18:28:05.079858 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovnnorthd-ovndbs" Oct 08 18:28:05 crc kubenswrapper[4988]: I1008 18:28:05.080005 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovnnorthd-ovnnorthd-dockercfg-srpk9" Oct 08 18:28:05 crc kubenswrapper[4988]: I1008 18:28:05.081360 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-bc45f6dcf-m2rkl"] Oct 08 18:28:05 crc kubenswrapper[4988]: I1008 18:28:05.097604 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-bc45f6dcf-m2rkl" Oct 08 18:28:05 crc kubenswrapper[4988]: I1008 18:28:05.105622 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Oct 08 18:28:05 crc kubenswrapper[4988]: I1008 18:28:05.116733 4988 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-sb" Oct 08 18:28:05 crc kubenswrapper[4988]: I1008 18:28:05.118436 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-bc45f6dcf-m2rkl"] Oct 08 18:28:05 crc kubenswrapper[4988]: I1008 18:28:05.142281 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/b92d978b-87c6-4210-9a69-99596b05cc18-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-w2zhj\" (UID: \"b92d978b-87c6-4210-9a69-99596b05cc18\") " pod="openstack/ovn-controller-metrics-w2zhj" Oct 08 18:28:05 crc kubenswrapper[4988]: I1008 18:28:05.142317 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f026ec8f-ffb4-4538-a6fc-3116ce0e7805-config\") pod \"ovn-northd-0\" (UID: \"f026ec8f-ffb4-4538-a6fc-3116ce0e7805\") " pod="openstack/ovn-northd-0" Oct 08 18:28:05 crc kubenswrapper[4988]: I1008 18:28:05.142336 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/94afe69f-30c7-4023-a485-f16998ed7fdd-dns-svc\") pod \"dnsmasq-dns-6d8fd57975-7jpt4\" (UID: \"94afe69f-30c7-4023-a485-f16998ed7fdd\") " pod="openstack/dnsmasq-dns-6d8fd57975-7jpt4" Oct 08 18:28:05 crc kubenswrapper[4988]: I1008 18:28:05.142353 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/f026ec8f-ffb4-4538-a6fc-3116ce0e7805-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"f026ec8f-ffb4-4538-a6fc-3116ce0e7805\") " pod="openstack/ovn-northd-0" Oct 08 18:28:05 crc kubenswrapper[4988]: I1008 18:28:05.142373 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f026ec8f-ffb4-4538-a6fc-3116ce0e7805-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"f026ec8f-ffb4-4538-a6fc-3116ce0e7805\") " pod="openstack/ovn-northd-0" Oct 08 18:28:05 crc kubenswrapper[4988]: I1008 18:28:05.142414 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f026ec8f-ffb4-4538-a6fc-3116ce0e7805-scripts\") pod \"ovn-northd-0\" (UID: \"f026ec8f-ffb4-4538-a6fc-3116ce0e7805\") " pod="openstack/ovn-northd-0" Oct 08 18:28:05 crc kubenswrapper[4988]: I1008 18:28:05.142433 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/94afe69f-30c7-4023-a485-f16998ed7fdd-config\") pod \"dnsmasq-dns-6d8fd57975-7jpt4\" (UID: \"94afe69f-30c7-4023-a485-f16998ed7fdd\") " pod="openstack/dnsmasq-dns-6d8fd57975-7jpt4" Oct 08 18:28:05 crc kubenswrapper[4988]: I1008 18:28:05.142455 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b92d978b-87c6-4210-9a69-99596b05cc18-config\") pod \"ovn-controller-metrics-w2zhj\" (UID: \"b92d978b-87c6-4210-9a69-99596b05cc18\") " pod="openstack/ovn-controller-metrics-w2zhj" Oct 08 18:28:05 crc kubenswrapper[4988]: I1008 18:28:05.142476 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/b92d978b-87c6-4210-9a69-99596b05cc18-ovs-rundir\") pod \"ovn-controller-metrics-w2zhj\" (UID: \"b92d978b-87c6-4210-9a69-99596b05cc18\") " pod="openstack/ovn-controller-metrics-w2zhj" Oct 08 18:28:05 crc kubenswrapper[4988]: I1008 18:28:05.142499 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/f026ec8f-ffb4-4538-a6fc-3116ce0e7805-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"f026ec8f-ffb4-4538-a6fc-3116ce0e7805\") " pod="openstack/ovn-northd-0" Oct 08 18:28:05 crc kubenswrapper[4988]: I1008 18:28:05.142531 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zstlc\" (UniqueName: \"kubernetes.io/projected/94afe69f-30c7-4023-a485-f16998ed7fdd-kube-api-access-zstlc\") pod \"dnsmasq-dns-6d8fd57975-7jpt4\" (UID: \"94afe69f-30c7-4023-a485-f16998ed7fdd\") " pod="openstack/dnsmasq-dns-6d8fd57975-7jpt4" Oct 08 18:28:05 crc kubenswrapper[4988]: I1008 18:28:05.142548 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-765fn\" (UniqueName: \"kubernetes.io/projected/f026ec8f-ffb4-4538-a6fc-3116ce0e7805-kube-api-access-765fn\") pod \"ovn-northd-0\" (UID: \"f026ec8f-ffb4-4538-a6fc-3116ce0e7805\") " pod="openstack/ovn-northd-0" Oct 08 18:28:05 crc kubenswrapper[4988]: I1008 18:28:05.142571 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/b92d978b-87c6-4210-9a69-99596b05cc18-ovn-rundir\") pod \"ovn-controller-metrics-w2zhj\" (UID: \"b92d978b-87c6-4210-9a69-99596b05cc18\") " pod="openstack/ovn-controller-metrics-w2zhj" Oct 08 18:28:05 crc kubenswrapper[4988]: I1008 18:28:05.142590 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/f026ec8f-ffb4-4538-a6fc-3116ce0e7805-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"f026ec8f-ffb4-4538-a6fc-3116ce0e7805\") " pod="openstack/ovn-northd-0" Oct 08 18:28:05 crc kubenswrapper[4988]: I1008 18:28:05.142619 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b92d978b-87c6-4210-9a69-99596b05cc18-combined-ca-bundle\") pod \"ovn-controller-metrics-w2zhj\" (UID: \"b92d978b-87c6-4210-9a69-99596b05cc18\") " pod="openstack/ovn-controller-metrics-w2zhj" Oct 08 18:28:05 crc kubenswrapper[4988]: I1008 18:28:05.142638 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b9thk\" (UniqueName: \"kubernetes.io/projected/b92d978b-87c6-4210-9a69-99596b05cc18-kube-api-access-b9thk\") pod \"ovn-controller-metrics-w2zhj\" (UID: \"b92d978b-87c6-4210-9a69-99596b05cc18\") " pod="openstack/ovn-controller-metrics-w2zhj" Oct 08 18:28:05 crc kubenswrapper[4988]: I1008 18:28:05.142665 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/94afe69f-30c7-4023-a485-f16998ed7fdd-ovsdbserver-nb\") pod \"dnsmasq-dns-6d8fd57975-7jpt4\" (UID: \"94afe69f-30c7-4023-a485-f16998ed7fdd\") " pod="openstack/dnsmasq-dns-6d8fd57975-7jpt4" Oct 08 18:28:05 crc kubenswrapper[4988]: I1008 18:28:05.143939 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/94afe69f-30c7-4023-a485-f16998ed7fdd-ovsdbserver-nb\") pod \"dnsmasq-dns-6d8fd57975-7jpt4\" (UID: \"94afe69f-30c7-4023-a485-f16998ed7fdd\") " pod="openstack/dnsmasq-dns-6d8fd57975-7jpt4" Oct 08 18:28:05 crc kubenswrapper[4988]: I1008 18:28:05.144193 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/b92d978b-87c6-4210-9a69-99596b05cc18-ovs-rundir\") pod \"ovn-controller-metrics-w2zhj\" (UID: \"b92d978b-87c6-4210-9a69-99596b05cc18\") " pod="openstack/ovn-controller-metrics-w2zhj" Oct 08 18:28:05 crc kubenswrapper[4988]: I1008 18:28:05.144372 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b92d978b-87c6-4210-9a69-99596b05cc18-config\") pod \"ovn-controller-metrics-w2zhj\" (UID: \"b92d978b-87c6-4210-9a69-99596b05cc18\") " pod="openstack/ovn-controller-metrics-w2zhj" Oct 08 18:28:05 crc kubenswrapper[4988]: I1008 18:28:05.145852 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/b92d978b-87c6-4210-9a69-99596b05cc18-ovn-rundir\") pod \"ovn-controller-metrics-w2zhj\" (UID: \"b92d978b-87c6-4210-9a69-99596b05cc18\") " pod="openstack/ovn-controller-metrics-w2zhj" Oct 08 18:28:05 crc kubenswrapper[4988]: I1008 18:28:05.149659 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/94afe69f-30c7-4023-a485-f16998ed7fdd-dns-svc\") pod \"dnsmasq-dns-6d8fd57975-7jpt4\" (UID: \"94afe69f-30c7-4023-a485-f16998ed7fdd\") " pod="openstack/dnsmasq-dns-6d8fd57975-7jpt4" Oct 08 18:28:05 crc kubenswrapper[4988]: I1008 18:28:05.152784 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/94afe69f-30c7-4023-a485-f16998ed7fdd-config\") pod \"dnsmasq-dns-6d8fd57975-7jpt4\" (UID: \"94afe69f-30c7-4023-a485-f16998ed7fdd\") " pod="openstack/dnsmasq-dns-6d8fd57975-7jpt4" Oct 08 18:28:05 crc kubenswrapper[4988]: I1008 18:28:05.152970 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/b92d978b-87c6-4210-9a69-99596b05cc18-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-w2zhj\" (UID: \"b92d978b-87c6-4210-9a69-99596b05cc18\") " pod="openstack/ovn-controller-metrics-w2zhj" Oct 08 18:28:05 crc kubenswrapper[4988]: I1008 18:28:05.159297 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b92d978b-87c6-4210-9a69-99596b05cc18-combined-ca-bundle\") pod \"ovn-controller-metrics-w2zhj\" (UID: \"b92d978b-87c6-4210-9a69-99596b05cc18\") " pod="openstack/ovn-controller-metrics-w2zhj" Oct 08 18:28:05 crc kubenswrapper[4988]: I1008 18:28:05.172101 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zstlc\" (UniqueName: \"kubernetes.io/projected/94afe69f-30c7-4023-a485-f16998ed7fdd-kube-api-access-zstlc\") pod \"dnsmasq-dns-6d8fd57975-7jpt4\" (UID: \"94afe69f-30c7-4023-a485-f16998ed7fdd\") " pod="openstack/dnsmasq-dns-6d8fd57975-7jpt4" Oct 08 18:28:05 crc kubenswrapper[4988]: I1008 18:28:05.182006 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b9thk\" (UniqueName: \"kubernetes.io/projected/b92d978b-87c6-4210-9a69-99596b05cc18-kube-api-access-b9thk\") pod \"ovn-controller-metrics-w2zhj\" (UID: \"b92d978b-87c6-4210-9a69-99596b05cc18\") " pod="openstack/ovn-controller-metrics-w2zhj" Oct 08 18:28:05 crc kubenswrapper[4988]: I1008 18:28:05.244385 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-765fn\" (UniqueName: \"kubernetes.io/projected/f026ec8f-ffb4-4538-a6fc-3116ce0e7805-kube-api-access-765fn\") pod \"ovn-northd-0\" (UID: \"f026ec8f-ffb4-4538-a6fc-3116ce0e7805\") " pod="openstack/ovn-northd-0" Oct 08 18:28:05 crc kubenswrapper[4988]: I1008 18:28:05.244443 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/f026ec8f-ffb4-4538-a6fc-3116ce0e7805-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"f026ec8f-ffb4-4538-a6fc-3116ce0e7805\") " pod="openstack/ovn-northd-0" Oct 08 18:28:05 crc kubenswrapper[4988]: I1008 18:28:05.244465 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9d3361fd-fc49-4087-88b3-835f5bd31e44-dns-svc\") pod \"dnsmasq-dns-bc45f6dcf-m2rkl\" (UID: \"9d3361fd-fc49-4087-88b3-835f5bd31e44\") " pod="openstack/dnsmasq-dns-bc45f6dcf-m2rkl" Oct 08 18:28:05 crc kubenswrapper[4988]: I1008 18:28:05.244488 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-88c4j\" (UniqueName: \"kubernetes.io/projected/9d3361fd-fc49-4087-88b3-835f5bd31e44-kube-api-access-88c4j\") pod \"dnsmasq-dns-bc45f6dcf-m2rkl\" (UID: \"9d3361fd-fc49-4087-88b3-835f5bd31e44\") " pod="openstack/dnsmasq-dns-bc45f6dcf-m2rkl" Oct 08 18:28:05 crc kubenswrapper[4988]: I1008 18:28:05.244562 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9d3361fd-fc49-4087-88b3-835f5bd31e44-ovsdbserver-sb\") pod \"dnsmasq-dns-bc45f6dcf-m2rkl\" (UID: \"9d3361fd-fc49-4087-88b3-835f5bd31e44\") " pod="openstack/dnsmasq-dns-bc45f6dcf-m2rkl" Oct 08 18:28:05 crc kubenswrapper[4988]: I1008 18:28:05.244625 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d3361fd-fc49-4087-88b3-835f5bd31e44-config\") pod \"dnsmasq-dns-bc45f6dcf-m2rkl\" (UID: \"9d3361fd-fc49-4087-88b3-835f5bd31e44\") " pod="openstack/dnsmasq-dns-bc45f6dcf-m2rkl" Oct 08 18:28:05 crc kubenswrapper[4988]: I1008 18:28:05.244671 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9d3361fd-fc49-4087-88b3-835f5bd31e44-ovsdbserver-nb\") pod \"dnsmasq-dns-bc45f6dcf-m2rkl\" (UID: \"9d3361fd-fc49-4087-88b3-835f5bd31e44\") " pod="openstack/dnsmasq-dns-bc45f6dcf-m2rkl" Oct 08 18:28:05 crc kubenswrapper[4988]: I1008 18:28:05.244694 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f026ec8f-ffb4-4538-a6fc-3116ce0e7805-config\") pod \"ovn-northd-0\" (UID: \"f026ec8f-ffb4-4538-a6fc-3116ce0e7805\") " pod="openstack/ovn-northd-0" Oct 08 18:28:05 crc kubenswrapper[4988]: I1008 18:28:05.244723 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/f026ec8f-ffb4-4538-a6fc-3116ce0e7805-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"f026ec8f-ffb4-4538-a6fc-3116ce0e7805\") " pod="openstack/ovn-northd-0" Oct 08 18:28:05 crc kubenswrapper[4988]: I1008 18:28:05.244775 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f026ec8f-ffb4-4538-a6fc-3116ce0e7805-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"f026ec8f-ffb4-4538-a6fc-3116ce0e7805\") " pod="openstack/ovn-northd-0" Oct 08 18:28:05 crc kubenswrapper[4988]: I1008 18:28:05.244791 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f026ec8f-ffb4-4538-a6fc-3116ce0e7805-scripts\") pod \"ovn-northd-0\" (UID: \"f026ec8f-ffb4-4538-a6fc-3116ce0e7805\") " pod="openstack/ovn-northd-0" Oct 08 18:28:05 crc kubenswrapper[4988]: I1008 18:28:05.244828 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/f026ec8f-ffb4-4538-a6fc-3116ce0e7805-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"f026ec8f-ffb4-4538-a6fc-3116ce0e7805\") " pod="openstack/ovn-northd-0" Oct 08 18:28:05 crc kubenswrapper[4988]: I1008 18:28:05.245302 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/f026ec8f-ffb4-4538-a6fc-3116ce0e7805-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"f026ec8f-ffb4-4538-a6fc-3116ce0e7805\") " pod="openstack/ovn-northd-0" Oct 08 18:28:05 crc kubenswrapper[4988]: I1008 18:28:05.245966 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f026ec8f-ffb4-4538-a6fc-3116ce0e7805-config\") pod \"ovn-northd-0\" (UID: \"f026ec8f-ffb4-4538-a6fc-3116ce0e7805\") " pod="openstack/ovn-northd-0" Oct 08 18:28:05 crc kubenswrapper[4988]: I1008 18:28:05.246490 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f026ec8f-ffb4-4538-a6fc-3116ce0e7805-scripts\") pod \"ovn-northd-0\" (UID: \"f026ec8f-ffb4-4538-a6fc-3116ce0e7805\") " pod="openstack/ovn-northd-0" Oct 08 18:28:05 crc kubenswrapper[4988]: I1008 18:28:05.248211 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f026ec8f-ffb4-4538-a6fc-3116ce0e7805-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"f026ec8f-ffb4-4538-a6fc-3116ce0e7805\") " pod="openstack/ovn-northd-0" Oct 08 18:28:05 crc kubenswrapper[4988]: I1008 18:28:05.248444 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/f026ec8f-ffb4-4538-a6fc-3116ce0e7805-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"f026ec8f-ffb4-4538-a6fc-3116ce0e7805\") " pod="openstack/ovn-northd-0" Oct 08 18:28:05 crc kubenswrapper[4988]: I1008 18:28:05.250163 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/f026ec8f-ffb4-4538-a6fc-3116ce0e7805-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"f026ec8f-ffb4-4538-a6fc-3116ce0e7805\") " pod="openstack/ovn-northd-0" Oct 08 18:28:05 crc kubenswrapper[4988]: I1008 18:28:05.252053 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6d8fd57975-7jpt4" Oct 08 18:28:05 crc kubenswrapper[4988]: I1008 18:28:05.264039 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-w2zhj" Oct 08 18:28:05 crc kubenswrapper[4988]: I1008 18:28:05.268979 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-644597f84c-nkqvt" Oct 08 18:28:05 crc kubenswrapper[4988]: I1008 18:28:05.269584 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-765fn\" (UniqueName: \"kubernetes.io/projected/f026ec8f-ffb4-4538-a6fc-3116ce0e7805-kube-api-access-765fn\") pod \"ovn-northd-0\" (UID: \"f026ec8f-ffb4-4538-a6fc-3116ce0e7805\") " pod="openstack/ovn-northd-0" Oct 08 18:28:05 crc kubenswrapper[4988]: I1008 18:28:05.346033 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e04834b5-6114-4fba-b3ac-9b6405b6712f-dns-svc\") pod \"e04834b5-6114-4fba-b3ac-9b6405b6712f\" (UID: \"e04834b5-6114-4fba-b3ac-9b6405b6712f\") " Oct 08 18:28:05 crc kubenswrapper[4988]: I1008 18:28:05.346114 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cp9pt\" (UniqueName: \"kubernetes.io/projected/e04834b5-6114-4fba-b3ac-9b6405b6712f-kube-api-access-cp9pt\") pod \"e04834b5-6114-4fba-b3ac-9b6405b6712f\" (UID: \"e04834b5-6114-4fba-b3ac-9b6405b6712f\") " Oct 08 18:28:05 crc kubenswrapper[4988]: I1008 18:28:05.346284 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e04834b5-6114-4fba-b3ac-9b6405b6712f-config\") pod \"e04834b5-6114-4fba-b3ac-9b6405b6712f\" (UID: \"e04834b5-6114-4fba-b3ac-9b6405b6712f\") " Oct 08 18:28:05 crc kubenswrapper[4988]: I1008 18:28:05.346484 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9d3361fd-fc49-4087-88b3-835f5bd31e44-ovsdbserver-sb\") pod \"dnsmasq-dns-bc45f6dcf-m2rkl\" (UID: \"9d3361fd-fc49-4087-88b3-835f5bd31e44\") " pod="openstack/dnsmasq-dns-bc45f6dcf-m2rkl" Oct 08 18:28:05 crc kubenswrapper[4988]: I1008 18:28:05.346519 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d3361fd-fc49-4087-88b3-835f5bd31e44-config\") pod \"dnsmasq-dns-bc45f6dcf-m2rkl\" (UID: \"9d3361fd-fc49-4087-88b3-835f5bd31e44\") " pod="openstack/dnsmasq-dns-bc45f6dcf-m2rkl" Oct 08 18:28:05 crc kubenswrapper[4988]: I1008 18:28:05.346542 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9d3361fd-fc49-4087-88b3-835f5bd31e44-ovsdbserver-nb\") pod \"dnsmasq-dns-bc45f6dcf-m2rkl\" (UID: \"9d3361fd-fc49-4087-88b3-835f5bd31e44\") " pod="openstack/dnsmasq-dns-bc45f6dcf-m2rkl" Oct 08 18:28:05 crc kubenswrapper[4988]: I1008 18:28:05.346638 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e04834b5-6114-4fba-b3ac-9b6405b6712f-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "e04834b5-6114-4fba-b3ac-9b6405b6712f" (UID: "e04834b5-6114-4fba-b3ac-9b6405b6712f"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 18:28:05 crc kubenswrapper[4988]: I1008 18:28:05.346755 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9d3361fd-fc49-4087-88b3-835f5bd31e44-dns-svc\") pod \"dnsmasq-dns-bc45f6dcf-m2rkl\" (UID: \"9d3361fd-fc49-4087-88b3-835f5bd31e44\") " pod="openstack/dnsmasq-dns-bc45f6dcf-m2rkl" Oct 08 18:28:05 crc kubenswrapper[4988]: I1008 18:28:05.346789 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-88c4j\" (UniqueName: \"kubernetes.io/projected/9d3361fd-fc49-4087-88b3-835f5bd31e44-kube-api-access-88c4j\") pod \"dnsmasq-dns-bc45f6dcf-m2rkl\" (UID: \"9d3361fd-fc49-4087-88b3-835f5bd31e44\") " pod="openstack/dnsmasq-dns-bc45f6dcf-m2rkl" Oct 08 18:28:05 crc kubenswrapper[4988]: I1008 18:28:05.346842 4988 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e04834b5-6114-4fba-b3ac-9b6405b6712f-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 08 18:28:05 crc kubenswrapper[4988]: I1008 18:28:05.348607 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9d3361fd-fc49-4087-88b3-835f5bd31e44-dns-svc\") pod \"dnsmasq-dns-bc45f6dcf-m2rkl\" (UID: \"9d3361fd-fc49-4087-88b3-835f5bd31e44\") " pod="openstack/dnsmasq-dns-bc45f6dcf-m2rkl" Oct 08 18:28:05 crc kubenswrapper[4988]: I1008 18:28:05.349152 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e04834b5-6114-4fba-b3ac-9b6405b6712f-config" (OuterVolumeSpecName: "config") pod "e04834b5-6114-4fba-b3ac-9b6405b6712f" (UID: "e04834b5-6114-4fba-b3ac-9b6405b6712f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 18:28:05 crc kubenswrapper[4988]: I1008 18:28:05.349381 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9d3361fd-fc49-4087-88b3-835f5bd31e44-ovsdbserver-nb\") pod \"dnsmasq-dns-bc45f6dcf-m2rkl\" (UID: \"9d3361fd-fc49-4087-88b3-835f5bd31e44\") " pod="openstack/dnsmasq-dns-bc45f6dcf-m2rkl" Oct 08 18:28:05 crc kubenswrapper[4988]: I1008 18:28:05.349600 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d3361fd-fc49-4087-88b3-835f5bd31e44-config\") pod \"dnsmasq-dns-bc45f6dcf-m2rkl\" (UID: \"9d3361fd-fc49-4087-88b3-835f5bd31e44\") " pod="openstack/dnsmasq-dns-bc45f6dcf-m2rkl" Oct 08 18:28:05 crc kubenswrapper[4988]: I1008 18:28:05.349764 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9d3361fd-fc49-4087-88b3-835f5bd31e44-ovsdbserver-sb\") pod \"dnsmasq-dns-bc45f6dcf-m2rkl\" (UID: \"9d3361fd-fc49-4087-88b3-835f5bd31e44\") " pod="openstack/dnsmasq-dns-bc45f6dcf-m2rkl" Oct 08 18:28:05 crc kubenswrapper[4988]: I1008 18:28:05.352959 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e04834b5-6114-4fba-b3ac-9b6405b6712f-kube-api-access-cp9pt" (OuterVolumeSpecName: "kube-api-access-cp9pt") pod "e04834b5-6114-4fba-b3ac-9b6405b6712f" (UID: "e04834b5-6114-4fba-b3ac-9b6405b6712f"). InnerVolumeSpecName "kube-api-access-cp9pt". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:28:05 crc kubenswrapper[4988]: I1008 18:28:05.363154 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-88c4j\" (UniqueName: \"kubernetes.io/projected/9d3361fd-fc49-4087-88b3-835f5bd31e44-kube-api-access-88c4j\") pod \"dnsmasq-dns-bc45f6dcf-m2rkl\" (UID: \"9d3361fd-fc49-4087-88b3-835f5bd31e44\") " pod="openstack/dnsmasq-dns-bc45f6dcf-m2rkl" Oct 08 18:28:05 crc kubenswrapper[4988]: I1008 18:28:05.448058 4988 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e04834b5-6114-4fba-b3ac-9b6405b6712f-config\") on node \"crc\" DevicePath \"\"" Oct 08 18:28:05 crc kubenswrapper[4988]: I1008 18:28:05.448090 4988 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cp9pt\" (UniqueName: \"kubernetes.io/projected/e04834b5-6114-4fba-b3ac-9b6405b6712f-kube-api-access-cp9pt\") on node \"crc\" DevicePath \"\"" Oct 08 18:28:05 crc kubenswrapper[4988]: I1008 18:28:05.549416 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Oct 08 18:28:05 crc kubenswrapper[4988]: I1008 18:28:05.562327 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-bc45f6dcf-m2rkl" Oct 08 18:28:05 crc kubenswrapper[4988]: I1008 18:28:05.633579 4988 generic.go:334] "Generic (PLEG): container finished" podID="72c64965-0b57-43e1-8d18-2bce6cabbfa1" containerID="4a95ce795323840905212e6edf0cc0a3e009ff2a132cb0c21d30fe459fa29c5d" exitCode=0 Oct 08 18:28:05 crc kubenswrapper[4988]: I1008 18:28:05.633641 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"72c64965-0b57-43e1-8d18-2bce6cabbfa1","Type":"ContainerDied","Data":"4a95ce795323840905212e6edf0cc0a3e009ff2a132cb0c21d30fe459fa29c5d"} Oct 08 18:28:05 crc kubenswrapper[4988]: I1008 18:28:05.639262 4988 generic.go:334] "Generic (PLEG): container finished" podID="3503942b-2825-4006-80e8-8c4610b89997" containerID="cefcf980cfc56a4cdb19408f13a9a6a84705b300fdac96063e98a4090209d49e" exitCode=0 Oct 08 18:28:05 crc kubenswrapper[4988]: I1008 18:28:05.639348 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"3503942b-2825-4006-80e8-8c4610b89997","Type":"ContainerDied","Data":"cefcf980cfc56a4cdb19408f13a9a6a84705b300fdac96063e98a4090209d49e"} Oct 08 18:28:05 crc kubenswrapper[4988]: I1008 18:28:05.649304 4988 generic.go:334] "Generic (PLEG): container finished" podID="5d556b85-3337-4855-9f43-a38946207378" containerID="cd93b284a7c36928c08db5b9f623c3bce0b9770b5089f5b5f8fb251b93fa77b7" exitCode=0 Oct 08 18:28:05 crc kubenswrapper[4988]: I1008 18:28:05.649599 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-77597f887-54xgv" event={"ID":"5d556b85-3337-4855-9f43-a38946207378","Type":"ContainerDied","Data":"cd93b284a7c36928c08db5b9f623c3bce0b9770b5089f5b5f8fb251b93fa77b7"} Oct 08 18:28:05 crc kubenswrapper[4988]: I1008 18:28:05.660826 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-644597f84c-nkqvt" Oct 08 18:28:05 crc kubenswrapper[4988]: I1008 18:28:05.660901 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-644597f84c-nkqvt" event={"ID":"e04834b5-6114-4fba-b3ac-9b6405b6712f","Type":"ContainerDied","Data":"1c0fe4bbc8a31a3adb33bbca81bb3d0d0b8154f1b6b2b87402836d7b6b9dfcf7"} Oct 08 18:28:05 crc kubenswrapper[4988]: I1008 18:28:05.718429 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6d8fd57975-7jpt4"] Oct 08 18:28:05 crc kubenswrapper[4988]: W1008 18:28:05.721533 4988 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod94afe69f_30c7_4023_a485_f16998ed7fdd.slice/crio-424a85ff85d3385b98f63e9b5b64ca76eeb3192d5aac5b17f12fff4a6933de4a WatchSource:0}: Error finding container 424a85ff85d3385b98f63e9b5b64ca76eeb3192d5aac5b17f12fff4a6933de4a: Status 404 returned error can't find the container with id 424a85ff85d3385b98f63e9b5b64ca76eeb3192d5aac5b17f12fff4a6933de4a Oct 08 18:28:05 crc kubenswrapper[4988]: I1008 18:28:05.760704 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-644597f84c-nkqvt"] Oct 08 18:28:05 crc kubenswrapper[4988]: I1008 18:28:05.796344 4988 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-644597f84c-nkqvt"] Oct 08 18:28:05 crc kubenswrapper[4988]: I1008 18:28:05.823076 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-w2zhj"] Oct 08 18:28:05 crc kubenswrapper[4988]: I1008 18:28:05.956041 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Oct 08 18:28:06 crc kubenswrapper[4988]: I1008 18:28:06.054565 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-bc45f6dcf-m2rkl"] Oct 08 18:28:06 crc kubenswrapper[4988]: W1008 18:28:06.066333 4988 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9d3361fd_fc49_4087_88b3_835f5bd31e44.slice/crio-ea0ade9c9c48e4ce89492e90365cd54e0095195de61a78fd2e2865f31e4418fc WatchSource:0}: Error finding container ea0ade9c9c48e4ce89492e90365cd54e0095195de61a78fd2e2865f31e4418fc: Status 404 returned error can't find the container with id ea0ade9c9c48e4ce89492e90365cd54e0095195de61a78fd2e2865f31e4418fc Oct 08 18:28:06 crc kubenswrapper[4988]: I1008 18:28:06.283172 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-77597f887-54xgv" Oct 08 18:28:06 crc kubenswrapper[4988]: I1008 18:28:06.378205 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tbtxg\" (UniqueName: \"kubernetes.io/projected/5d556b85-3337-4855-9f43-a38946207378-kube-api-access-tbtxg\") pod \"5d556b85-3337-4855-9f43-a38946207378\" (UID: \"5d556b85-3337-4855-9f43-a38946207378\") " Oct 08 18:28:06 crc kubenswrapper[4988]: I1008 18:28:06.378332 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5d556b85-3337-4855-9f43-a38946207378-dns-svc\") pod \"5d556b85-3337-4855-9f43-a38946207378\" (UID: \"5d556b85-3337-4855-9f43-a38946207378\") " Oct 08 18:28:06 crc kubenswrapper[4988]: I1008 18:28:06.378444 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5d556b85-3337-4855-9f43-a38946207378-config\") pod \"5d556b85-3337-4855-9f43-a38946207378\" (UID: \"5d556b85-3337-4855-9f43-a38946207378\") " Oct 08 18:28:06 crc kubenswrapper[4988]: I1008 18:28:06.382751 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5d556b85-3337-4855-9f43-a38946207378-kube-api-access-tbtxg" (OuterVolumeSpecName: "kube-api-access-tbtxg") pod "5d556b85-3337-4855-9f43-a38946207378" (UID: "5d556b85-3337-4855-9f43-a38946207378"). InnerVolumeSpecName "kube-api-access-tbtxg". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:28:06 crc kubenswrapper[4988]: I1008 18:28:06.395970 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5d556b85-3337-4855-9f43-a38946207378-config" (OuterVolumeSpecName: "config") pod "5d556b85-3337-4855-9f43-a38946207378" (UID: "5d556b85-3337-4855-9f43-a38946207378"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 18:28:06 crc kubenswrapper[4988]: I1008 18:28:06.396013 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5d556b85-3337-4855-9f43-a38946207378-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "5d556b85-3337-4855-9f43-a38946207378" (UID: "5d556b85-3337-4855-9f43-a38946207378"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 18:28:06 crc kubenswrapper[4988]: I1008 18:28:06.480118 4988 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5d556b85-3337-4855-9f43-a38946207378-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 08 18:28:06 crc kubenswrapper[4988]: I1008 18:28:06.480148 4988 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5d556b85-3337-4855-9f43-a38946207378-config\") on node \"crc\" DevicePath \"\"" Oct 08 18:28:06 crc kubenswrapper[4988]: I1008 18:28:06.480159 4988 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tbtxg\" (UniqueName: \"kubernetes.io/projected/5d556b85-3337-4855-9f43-a38946207378-kube-api-access-tbtxg\") on node \"crc\" DevicePath \"\"" Oct 08 18:28:06 crc kubenswrapper[4988]: I1008 18:28:06.703934 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-w2zhj" event={"ID":"b92d978b-87c6-4210-9a69-99596b05cc18","Type":"ContainerStarted","Data":"80822f2e64d8fd8a2c3f08eea237f88c03ba055c04d5bb24498177ab69c61a2e"} Oct 08 18:28:06 crc kubenswrapper[4988]: I1008 18:28:06.703994 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-w2zhj" event={"ID":"b92d978b-87c6-4210-9a69-99596b05cc18","Type":"ContainerStarted","Data":"b1e0fd5f4062cfe174f7c631ad5c1f707bc5ae49d66c7afef5bcf92f995195fd"} Oct 08 18:28:06 crc kubenswrapper[4988]: I1008 18:28:06.722994 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"f026ec8f-ffb4-4538-a6fc-3116ce0e7805","Type":"ContainerStarted","Data":"2c4f9dbb0baa15e9b34f5b39299f44218d04d27e29dfc58039842941ca081b39"} Oct 08 18:28:06 crc kubenswrapper[4988]: I1008 18:28:06.736181 4988 generic.go:334] "Generic (PLEG): container finished" podID="94afe69f-30c7-4023-a485-f16998ed7fdd" containerID="e623903a25ba6b731e6957da4de44239c9b78815190afaeeebac2b2991024c3f" exitCode=0 Oct 08 18:28:06 crc kubenswrapper[4988]: I1008 18:28:06.736243 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6d8fd57975-7jpt4" event={"ID":"94afe69f-30c7-4023-a485-f16998ed7fdd","Type":"ContainerDied","Data":"e623903a25ba6b731e6957da4de44239c9b78815190afaeeebac2b2991024c3f"} Oct 08 18:28:06 crc kubenswrapper[4988]: I1008 18:28:06.736273 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6d8fd57975-7jpt4" event={"ID":"94afe69f-30c7-4023-a485-f16998ed7fdd","Type":"ContainerStarted","Data":"424a85ff85d3385b98f63e9b5b64ca76eeb3192d5aac5b17f12fff4a6933de4a"} Oct 08 18:28:06 crc kubenswrapper[4988]: I1008 18:28:06.739018 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-metrics-w2zhj" podStartSLOduration=2.73899665 podStartE2EDuration="2.73899665s" podCreationTimestamp="2025-10-08 18:28:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 18:28:06.733186897 +0000 UTC m=+1032.183029687" watchObservedRunningTime="2025-10-08 18:28:06.73899665 +0000 UTC m=+1032.188839420" Oct 08 18:28:06 crc kubenswrapper[4988]: I1008 18:28:06.754988 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"3503942b-2825-4006-80e8-8c4610b89997","Type":"ContainerStarted","Data":"e1a685c92e7a7b5fad1185fbe08be64ed9f96221d3c3a8c8c5d3364e9a593229"} Oct 08 18:28:06 crc kubenswrapper[4988]: I1008 18:28:06.803563 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-77597f887-54xgv" event={"ID":"5d556b85-3337-4855-9f43-a38946207378","Type":"ContainerDied","Data":"18c621fd710f1a97e662af27846b796420779e3bb2993ed758cfedb273d824d1"} Oct 08 18:28:06 crc kubenswrapper[4988]: I1008 18:28:06.803632 4988 scope.go:117] "RemoveContainer" containerID="cd93b284a7c36928c08db5b9f623c3bce0b9770b5089f5b5f8fb251b93fa77b7" Oct 08 18:28:06 crc kubenswrapper[4988]: I1008 18:28:06.803814 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-77597f887-54xgv" Oct 08 18:28:06 crc kubenswrapper[4988]: I1008 18:28:06.811139 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"72c64965-0b57-43e1-8d18-2bce6cabbfa1","Type":"ContainerStarted","Data":"239fe689949862e30d5f5b0021f848adb4380b18bdefdeb68e24c7cd3fb41c24"} Oct 08 18:28:06 crc kubenswrapper[4988]: I1008 18:28:06.814167 4988 generic.go:334] "Generic (PLEG): container finished" podID="9d3361fd-fc49-4087-88b3-835f5bd31e44" containerID="a622f60fa34c7e94c933213e1e70d6b951e184acb6ecf9cf248cfbbf3a54c7f4" exitCode=0 Oct 08 18:28:06 crc kubenswrapper[4988]: I1008 18:28:06.815102 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-bc45f6dcf-m2rkl" event={"ID":"9d3361fd-fc49-4087-88b3-835f5bd31e44","Type":"ContainerDied","Data":"a622f60fa34c7e94c933213e1e70d6b951e184acb6ecf9cf248cfbbf3a54c7f4"} Oct 08 18:28:06 crc kubenswrapper[4988]: I1008 18:28:06.815132 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-bc45f6dcf-m2rkl" event={"ID":"9d3361fd-fc49-4087-88b3-835f5bd31e44","Type":"ContainerStarted","Data":"ea0ade9c9c48e4ce89492e90365cd54e0095195de61a78fd2e2865f31e4418fc"} Oct 08 18:28:06 crc kubenswrapper[4988]: I1008 18:28:06.830198 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-cell1-galera-0" podStartSLOduration=9.155789865 podStartE2EDuration="40.830171221s" podCreationTimestamp="2025-10-08 18:27:26 +0000 UTC" firstStartedPulling="2025-10-08 18:27:28.721315852 +0000 UTC m=+994.171158622" lastFinishedPulling="2025-10-08 18:28:00.395697208 +0000 UTC m=+1025.845539978" observedRunningTime="2025-10-08 18:28:06.826689802 +0000 UTC m=+1032.276532572" watchObservedRunningTime="2025-10-08 18:28:06.830171221 +0000 UTC m=+1032.280013991" Oct 08 18:28:06 crc kubenswrapper[4988]: I1008 18:28:06.924277 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-77597f887-54xgv"] Oct 08 18:28:06 crc kubenswrapper[4988]: I1008 18:28:06.935113 4988 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-77597f887-54xgv"] Oct 08 18:28:07 crc kubenswrapper[4988]: I1008 18:28:07.249333 4988 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5d556b85-3337-4855-9f43-a38946207378" path="/var/lib/kubelet/pods/5d556b85-3337-4855-9f43-a38946207378/volumes" Oct 08 18:28:07 crc kubenswrapper[4988]: I1008 18:28:07.250743 4988 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e04834b5-6114-4fba-b3ac-9b6405b6712f" path="/var/lib/kubelet/pods/e04834b5-6114-4fba-b3ac-9b6405b6712f/volumes" Oct 08 18:28:07 crc kubenswrapper[4988]: I1008 18:28:07.515656 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-galera-0" Oct 08 18:28:07 crc kubenswrapper[4988]: I1008 18:28:07.515823 4988 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-galera-0" Oct 08 18:28:07 crc kubenswrapper[4988]: I1008 18:28:07.821295 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-bc45f6dcf-m2rkl" event={"ID":"9d3361fd-fc49-4087-88b3-835f5bd31e44","Type":"ContainerStarted","Data":"e1160555585e55ae8048a765c33e98e2dcecb0deea2736accaac322df3b8408e"} Oct 08 18:28:07 crc kubenswrapper[4988]: I1008 18:28:07.821714 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-bc45f6dcf-m2rkl" Oct 08 18:28:07 crc kubenswrapper[4988]: I1008 18:28:07.823974 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"f026ec8f-ffb4-4538-a6fc-3116ce0e7805","Type":"ContainerStarted","Data":"ada91b938d451d0675a819f062dafde62f4a75b785dc86a5448ce25e246cecbb"} Oct 08 18:28:07 crc kubenswrapper[4988]: I1008 18:28:07.824008 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"f026ec8f-ffb4-4538-a6fc-3116ce0e7805","Type":"ContainerStarted","Data":"e9efac4802d1c9a2377dc285da1c498c1363ff5f70695e54b4a257848bdea958"} Oct 08 18:28:07 crc kubenswrapper[4988]: I1008 18:28:07.824212 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-northd-0" Oct 08 18:28:07 crc kubenswrapper[4988]: I1008 18:28:07.826868 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6d8fd57975-7jpt4" event={"ID":"94afe69f-30c7-4023-a485-f16998ed7fdd","Type":"ContainerStarted","Data":"65ea95e1c8ef8edcae7c30eec6c6dbf2fcd831415768d6afb24644972378e9cb"} Oct 08 18:28:07 crc kubenswrapper[4988]: I1008 18:28:07.826990 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-6d8fd57975-7jpt4" Oct 08 18:28:07 crc kubenswrapper[4988]: I1008 18:28:07.837601 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-bc45f6dcf-m2rkl" podStartSLOduration=2.83757708 podStartE2EDuration="2.83757708s" podCreationTimestamp="2025-10-08 18:28:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 18:28:07.836675261 +0000 UTC m=+1033.286518061" watchObservedRunningTime="2025-10-08 18:28:07.83757708 +0000 UTC m=+1033.287419860" Oct 08 18:28:07 crc kubenswrapper[4988]: I1008 18:28:07.840979 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-galera-0" podStartSLOduration=9.633268542 podStartE2EDuration="41.840967776s" podCreationTimestamp="2025-10-08 18:27:26 +0000 UTC" firstStartedPulling="2025-10-08 18:27:28.190914186 +0000 UTC m=+993.640756956" lastFinishedPulling="2025-10-08 18:28:00.39861342 +0000 UTC m=+1025.848456190" observedRunningTime="2025-10-08 18:28:06.944182123 +0000 UTC m=+1032.394024893" watchObservedRunningTime="2025-10-08 18:28:07.840967776 +0000 UTC m=+1033.290810556" Oct 08 18:28:07 crc kubenswrapper[4988]: I1008 18:28:07.860811 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-6d8fd57975-7jpt4" podStartSLOduration=3.8607864899999997 podStartE2EDuration="3.86078649s" podCreationTimestamp="2025-10-08 18:28:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 18:28:07.853119729 +0000 UTC m=+1033.302962559" watchObservedRunningTime="2025-10-08 18:28:07.86078649 +0000 UTC m=+1033.310629260" Oct 08 18:28:07 crc kubenswrapper[4988]: I1008 18:28:07.880411 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-northd-0" podStartSLOduration=1.504767253 podStartE2EDuration="2.880378967s" podCreationTimestamp="2025-10-08 18:28:05 +0000 UTC" firstStartedPulling="2025-10-08 18:28:05.9928233 +0000 UTC m=+1031.442666070" lastFinishedPulling="2025-10-08 18:28:07.368435024 +0000 UTC m=+1032.818277784" observedRunningTime="2025-10-08 18:28:07.876294939 +0000 UTC m=+1033.326137729" watchObservedRunningTime="2025-10-08 18:28:07.880378967 +0000 UTC m=+1033.330221737" Oct 08 18:28:07 crc kubenswrapper[4988]: I1008 18:28:07.977954 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-cell1-galera-0" Oct 08 18:28:07 crc kubenswrapper[4988]: I1008 18:28:07.978220 4988 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-cell1-galera-0" Oct 08 18:28:08 crc kubenswrapper[4988]: I1008 18:28:08.577630 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/memcached-0" Oct 08 18:28:10 crc kubenswrapper[4988]: I1008 18:28:10.349481 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Oct 08 18:28:10 crc kubenswrapper[4988]: I1008 18:28:10.436620 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6d8fd57975-7jpt4"] Oct 08 18:28:10 crc kubenswrapper[4988]: I1008 18:28:10.436864 4988 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-6d8fd57975-7jpt4" podUID="94afe69f-30c7-4023-a485-f16998ed7fdd" containerName="dnsmasq-dns" containerID="cri-o://65ea95e1c8ef8edcae7c30eec6c6dbf2fcd831415768d6afb24644972378e9cb" gracePeriod=10 Oct 08 18:28:10 crc kubenswrapper[4988]: I1008 18:28:10.476227 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-57f58c7cff-42svl"] Oct 08 18:28:10 crc kubenswrapper[4988]: E1008 18:28:10.476532 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5d556b85-3337-4855-9f43-a38946207378" containerName="init" Oct 08 18:28:10 crc kubenswrapper[4988]: I1008 18:28:10.476547 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="5d556b85-3337-4855-9f43-a38946207378" containerName="init" Oct 08 18:28:10 crc kubenswrapper[4988]: I1008 18:28:10.476703 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="5d556b85-3337-4855-9f43-a38946207378" containerName="init" Oct 08 18:28:10 crc kubenswrapper[4988]: I1008 18:28:10.477602 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57f58c7cff-42svl" Oct 08 18:28:10 crc kubenswrapper[4988]: I1008 18:28:10.492184 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-57f58c7cff-42svl"] Oct 08 18:28:10 crc kubenswrapper[4988]: I1008 18:28:10.554104 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xbmst\" (UniqueName: \"kubernetes.io/projected/1487d461-4c56-4591-b498-b6ce9f314763-kube-api-access-xbmst\") pod \"dnsmasq-dns-57f58c7cff-42svl\" (UID: \"1487d461-4c56-4591-b498-b6ce9f314763\") " pod="openstack/dnsmasq-dns-57f58c7cff-42svl" Oct 08 18:28:10 crc kubenswrapper[4988]: I1008 18:28:10.554154 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/1487d461-4c56-4591-b498-b6ce9f314763-ovsdbserver-sb\") pod \"dnsmasq-dns-57f58c7cff-42svl\" (UID: \"1487d461-4c56-4591-b498-b6ce9f314763\") " pod="openstack/dnsmasq-dns-57f58c7cff-42svl" Oct 08 18:28:10 crc kubenswrapper[4988]: I1008 18:28:10.554204 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1487d461-4c56-4591-b498-b6ce9f314763-dns-svc\") pod \"dnsmasq-dns-57f58c7cff-42svl\" (UID: \"1487d461-4c56-4591-b498-b6ce9f314763\") " pod="openstack/dnsmasq-dns-57f58c7cff-42svl" Oct 08 18:28:10 crc kubenswrapper[4988]: I1008 18:28:10.554231 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/1487d461-4c56-4591-b498-b6ce9f314763-ovsdbserver-nb\") pod \"dnsmasq-dns-57f58c7cff-42svl\" (UID: \"1487d461-4c56-4591-b498-b6ce9f314763\") " pod="openstack/dnsmasq-dns-57f58c7cff-42svl" Oct 08 18:28:10 crc kubenswrapper[4988]: I1008 18:28:10.554269 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1487d461-4c56-4591-b498-b6ce9f314763-config\") pod \"dnsmasq-dns-57f58c7cff-42svl\" (UID: \"1487d461-4c56-4591-b498-b6ce9f314763\") " pod="openstack/dnsmasq-dns-57f58c7cff-42svl" Oct 08 18:28:10 crc kubenswrapper[4988]: I1008 18:28:10.656004 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xbmst\" (UniqueName: \"kubernetes.io/projected/1487d461-4c56-4591-b498-b6ce9f314763-kube-api-access-xbmst\") pod \"dnsmasq-dns-57f58c7cff-42svl\" (UID: \"1487d461-4c56-4591-b498-b6ce9f314763\") " pod="openstack/dnsmasq-dns-57f58c7cff-42svl" Oct 08 18:28:10 crc kubenswrapper[4988]: I1008 18:28:10.656083 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/1487d461-4c56-4591-b498-b6ce9f314763-ovsdbserver-sb\") pod \"dnsmasq-dns-57f58c7cff-42svl\" (UID: \"1487d461-4c56-4591-b498-b6ce9f314763\") " pod="openstack/dnsmasq-dns-57f58c7cff-42svl" Oct 08 18:28:10 crc kubenswrapper[4988]: I1008 18:28:10.656180 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1487d461-4c56-4591-b498-b6ce9f314763-dns-svc\") pod \"dnsmasq-dns-57f58c7cff-42svl\" (UID: \"1487d461-4c56-4591-b498-b6ce9f314763\") " pod="openstack/dnsmasq-dns-57f58c7cff-42svl" Oct 08 18:28:10 crc kubenswrapper[4988]: I1008 18:28:10.656236 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/1487d461-4c56-4591-b498-b6ce9f314763-ovsdbserver-nb\") pod \"dnsmasq-dns-57f58c7cff-42svl\" (UID: \"1487d461-4c56-4591-b498-b6ce9f314763\") " pod="openstack/dnsmasq-dns-57f58c7cff-42svl" Oct 08 18:28:10 crc kubenswrapper[4988]: I1008 18:28:10.656310 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1487d461-4c56-4591-b498-b6ce9f314763-config\") pod \"dnsmasq-dns-57f58c7cff-42svl\" (UID: \"1487d461-4c56-4591-b498-b6ce9f314763\") " pod="openstack/dnsmasq-dns-57f58c7cff-42svl" Oct 08 18:28:10 crc kubenswrapper[4988]: I1008 18:28:10.661895 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/1487d461-4c56-4591-b498-b6ce9f314763-ovsdbserver-sb\") pod \"dnsmasq-dns-57f58c7cff-42svl\" (UID: \"1487d461-4c56-4591-b498-b6ce9f314763\") " pod="openstack/dnsmasq-dns-57f58c7cff-42svl" Oct 08 18:28:10 crc kubenswrapper[4988]: I1008 18:28:10.661921 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1487d461-4c56-4591-b498-b6ce9f314763-config\") pod \"dnsmasq-dns-57f58c7cff-42svl\" (UID: \"1487d461-4c56-4591-b498-b6ce9f314763\") " pod="openstack/dnsmasq-dns-57f58c7cff-42svl" Oct 08 18:28:10 crc kubenswrapper[4988]: I1008 18:28:10.674514 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1487d461-4c56-4591-b498-b6ce9f314763-dns-svc\") pod \"dnsmasq-dns-57f58c7cff-42svl\" (UID: \"1487d461-4c56-4591-b498-b6ce9f314763\") " pod="openstack/dnsmasq-dns-57f58c7cff-42svl" Oct 08 18:28:10 crc kubenswrapper[4988]: I1008 18:28:10.675502 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/1487d461-4c56-4591-b498-b6ce9f314763-ovsdbserver-nb\") pod \"dnsmasq-dns-57f58c7cff-42svl\" (UID: \"1487d461-4c56-4591-b498-b6ce9f314763\") " pod="openstack/dnsmasq-dns-57f58c7cff-42svl" Oct 08 18:28:10 crc kubenswrapper[4988]: I1008 18:28:10.685342 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xbmst\" (UniqueName: \"kubernetes.io/projected/1487d461-4c56-4591-b498-b6ce9f314763-kube-api-access-xbmst\") pod \"dnsmasq-dns-57f58c7cff-42svl\" (UID: \"1487d461-4c56-4591-b498-b6ce9f314763\") " pod="openstack/dnsmasq-dns-57f58c7cff-42svl" Oct 08 18:28:10 crc kubenswrapper[4988]: I1008 18:28:10.792304 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57f58c7cff-42svl" Oct 08 18:28:10 crc kubenswrapper[4988]: I1008 18:28:10.857541 4988 generic.go:334] "Generic (PLEG): container finished" podID="94afe69f-30c7-4023-a485-f16998ed7fdd" containerID="65ea95e1c8ef8edcae7c30eec6c6dbf2fcd831415768d6afb24644972378e9cb" exitCode=0 Oct 08 18:28:10 crc kubenswrapper[4988]: I1008 18:28:10.857611 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6d8fd57975-7jpt4" event={"ID":"94afe69f-30c7-4023-a485-f16998ed7fdd","Type":"ContainerDied","Data":"65ea95e1c8ef8edcae7c30eec6c6dbf2fcd831415768d6afb24644972378e9cb"} Oct 08 18:28:11 crc kubenswrapper[4988]: I1008 18:28:11.288007 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-57f58c7cff-42svl"] Oct 08 18:28:11 crc kubenswrapper[4988]: I1008 18:28:11.519571 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-storage-0"] Oct 08 18:28:11 crc kubenswrapper[4988]: I1008 18:28:11.525026 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Oct 08 18:28:11 crc kubenswrapper[4988]: I1008 18:28:11.527022 4988 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-files" Oct 08 18:28:11 crc kubenswrapper[4988]: I1008 18:28:11.527086 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-conf" Oct 08 18:28:11 crc kubenswrapper[4988]: I1008 18:28:11.528072 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-swift-dockercfg-8kxb7" Oct 08 18:28:11 crc kubenswrapper[4988]: I1008 18:28:11.530418 4988 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-storage-config-data" Oct 08 18:28:11 crc kubenswrapper[4988]: I1008 18:28:11.538361 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-storage-0"] Oct 08 18:28:11 crc kubenswrapper[4988]: I1008 18:28:11.606666 4988 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-galera-0" Oct 08 18:28:11 crc kubenswrapper[4988]: I1008 18:28:11.662677 4988 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/openstack-galera-0" podUID="72c64965-0b57-43e1-8d18-2bce6cabbfa1" containerName="galera" probeResult="failure" output=< Oct 08 18:28:11 crc kubenswrapper[4988]: wsrep_local_state_comment (Joined) differs from Synced Oct 08 18:28:11 crc kubenswrapper[4988]: > Oct 08 18:28:11 crc kubenswrapper[4988]: I1008 18:28:11.677796 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/d04259e5-9595-447b-b804-feba577498d8-etc-swift\") pod \"swift-storage-0\" (UID: \"d04259e5-9595-447b-b804-feba577498d8\") " pod="openstack/swift-storage-0" Oct 08 18:28:11 crc kubenswrapper[4988]: I1008 18:28:11.678139 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/d04259e5-9595-447b-b804-feba577498d8-cache\") pod \"swift-storage-0\" (UID: \"d04259e5-9595-447b-b804-feba577498d8\") " pod="openstack/swift-storage-0" Oct 08 18:28:11 crc kubenswrapper[4988]: I1008 18:28:11.678193 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vfz4q\" (UniqueName: \"kubernetes.io/projected/d04259e5-9595-447b-b804-feba577498d8-kube-api-access-vfz4q\") pod \"swift-storage-0\" (UID: \"d04259e5-9595-447b-b804-feba577498d8\") " pod="openstack/swift-storage-0" Oct 08 18:28:11 crc kubenswrapper[4988]: I1008 18:28:11.678240 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/d04259e5-9595-447b-b804-feba577498d8-lock\") pod \"swift-storage-0\" (UID: \"d04259e5-9595-447b-b804-feba577498d8\") " pod="openstack/swift-storage-0" Oct 08 18:28:11 crc kubenswrapper[4988]: I1008 18:28:11.678269 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"swift-storage-0\" (UID: \"d04259e5-9595-447b-b804-feba577498d8\") " pod="openstack/swift-storage-0" Oct 08 18:28:11 crc kubenswrapper[4988]: I1008 18:28:11.779806 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/d04259e5-9595-447b-b804-feba577498d8-cache\") pod \"swift-storage-0\" (UID: \"d04259e5-9595-447b-b804-feba577498d8\") " pod="openstack/swift-storage-0" Oct 08 18:28:11 crc kubenswrapper[4988]: I1008 18:28:11.779895 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vfz4q\" (UniqueName: \"kubernetes.io/projected/d04259e5-9595-447b-b804-feba577498d8-kube-api-access-vfz4q\") pod \"swift-storage-0\" (UID: \"d04259e5-9595-447b-b804-feba577498d8\") " pod="openstack/swift-storage-0" Oct 08 18:28:11 crc kubenswrapper[4988]: I1008 18:28:11.779950 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/d04259e5-9595-447b-b804-feba577498d8-lock\") pod \"swift-storage-0\" (UID: \"d04259e5-9595-447b-b804-feba577498d8\") " pod="openstack/swift-storage-0" Oct 08 18:28:11 crc kubenswrapper[4988]: I1008 18:28:11.779987 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"swift-storage-0\" (UID: \"d04259e5-9595-447b-b804-feba577498d8\") " pod="openstack/swift-storage-0" Oct 08 18:28:11 crc kubenswrapper[4988]: I1008 18:28:11.780113 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/d04259e5-9595-447b-b804-feba577498d8-etc-swift\") pod \"swift-storage-0\" (UID: \"d04259e5-9595-447b-b804-feba577498d8\") " pod="openstack/swift-storage-0" Oct 08 18:28:11 crc kubenswrapper[4988]: E1008 18:28:11.780336 4988 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Oct 08 18:28:11 crc kubenswrapper[4988]: E1008 18:28:11.780366 4988 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Oct 08 18:28:11 crc kubenswrapper[4988]: E1008 18:28:11.780453 4988 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/d04259e5-9595-447b-b804-feba577498d8-etc-swift podName:d04259e5-9595-447b-b804-feba577498d8 nodeName:}" failed. No retries permitted until 2025-10-08 18:28:12.280427521 +0000 UTC m=+1037.730270311 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/d04259e5-9595-447b-b804-feba577498d8-etc-swift") pod "swift-storage-0" (UID: "d04259e5-9595-447b-b804-feba577498d8") : configmap "swift-ring-files" not found Oct 08 18:28:11 crc kubenswrapper[4988]: I1008 18:28:11.781143 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/d04259e5-9595-447b-b804-feba577498d8-cache\") pod \"swift-storage-0\" (UID: \"d04259e5-9595-447b-b804-feba577498d8\") " pod="openstack/swift-storage-0" Oct 08 18:28:11 crc kubenswrapper[4988]: I1008 18:28:11.781694 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/d04259e5-9595-447b-b804-feba577498d8-lock\") pod \"swift-storage-0\" (UID: \"d04259e5-9595-447b-b804-feba577498d8\") " pod="openstack/swift-storage-0" Oct 08 18:28:11 crc kubenswrapper[4988]: I1008 18:28:11.782030 4988 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"swift-storage-0\" (UID: \"d04259e5-9595-447b-b804-feba577498d8\") device mount path \"/mnt/openstack/pv06\"" pod="openstack/swift-storage-0" Oct 08 18:28:11 crc kubenswrapper[4988]: I1008 18:28:11.815315 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"swift-storage-0\" (UID: \"d04259e5-9595-447b-b804-feba577498d8\") " pod="openstack/swift-storage-0" Oct 08 18:28:11 crc kubenswrapper[4988]: I1008 18:28:11.817594 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vfz4q\" (UniqueName: \"kubernetes.io/projected/d04259e5-9595-447b-b804-feba577498d8-kube-api-access-vfz4q\") pod \"swift-storage-0\" (UID: \"d04259e5-9595-447b-b804-feba577498d8\") " pod="openstack/swift-storage-0" Oct 08 18:28:11 crc kubenswrapper[4988]: I1008 18:28:11.867675 4988 generic.go:334] "Generic (PLEG): container finished" podID="1487d461-4c56-4591-b498-b6ce9f314763" containerID="0ab1e605fc110bb68f6ddc8059bafeb690570e0b3d4bbdc86f04efaca042ac71" exitCode=0 Oct 08 18:28:11 crc kubenswrapper[4988]: I1008 18:28:11.867719 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57f58c7cff-42svl" event={"ID":"1487d461-4c56-4591-b498-b6ce9f314763","Type":"ContainerDied","Data":"0ab1e605fc110bb68f6ddc8059bafeb690570e0b3d4bbdc86f04efaca042ac71"} Oct 08 18:28:11 crc kubenswrapper[4988]: I1008 18:28:11.867748 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57f58c7cff-42svl" event={"ID":"1487d461-4c56-4591-b498-b6ce9f314763","Type":"ContainerStarted","Data":"70d8af15af815443cdd9a37912e3cb7997d6d9fee4c892217c40ea75e46d3691"} Oct 08 18:28:12 crc kubenswrapper[4988]: I1008 18:28:12.024601 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-ring-rebalance-9hjmc"] Oct 08 18:28:12 crc kubenswrapper[4988]: I1008 18:28:12.025927 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-9hjmc" Oct 08 18:28:12 crc kubenswrapper[4988]: I1008 18:28:12.029814 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-proxy-config-data" Oct 08 18:28:12 crc kubenswrapper[4988]: I1008 18:28:12.030135 4988 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-config-data" Oct 08 18:28:12 crc kubenswrapper[4988]: I1008 18:28:12.032934 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-9hjmc"] Oct 08 18:28:12 crc kubenswrapper[4988]: I1008 18:28:12.033722 4988 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-scripts" Oct 08 18:28:12 crc kubenswrapper[4988]: I1008 18:28:12.085860 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/567a67dc-01a7-4720-8081-d60470856578-combined-ca-bundle\") pod \"swift-ring-rebalance-9hjmc\" (UID: \"567a67dc-01a7-4720-8081-d60470856578\") " pod="openstack/swift-ring-rebalance-9hjmc" Oct 08 18:28:12 crc kubenswrapper[4988]: I1008 18:28:12.085990 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hqwhf\" (UniqueName: \"kubernetes.io/projected/567a67dc-01a7-4720-8081-d60470856578-kube-api-access-hqwhf\") pod \"swift-ring-rebalance-9hjmc\" (UID: \"567a67dc-01a7-4720-8081-d60470856578\") " pod="openstack/swift-ring-rebalance-9hjmc" Oct 08 18:28:12 crc kubenswrapper[4988]: I1008 18:28:12.086105 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/567a67dc-01a7-4720-8081-d60470856578-etc-swift\") pod \"swift-ring-rebalance-9hjmc\" (UID: \"567a67dc-01a7-4720-8081-d60470856578\") " pod="openstack/swift-ring-rebalance-9hjmc" Oct 08 18:28:12 crc kubenswrapper[4988]: I1008 18:28:12.086286 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/567a67dc-01a7-4720-8081-d60470856578-scripts\") pod \"swift-ring-rebalance-9hjmc\" (UID: \"567a67dc-01a7-4720-8081-d60470856578\") " pod="openstack/swift-ring-rebalance-9hjmc" Oct 08 18:28:12 crc kubenswrapper[4988]: I1008 18:28:12.086372 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/567a67dc-01a7-4720-8081-d60470856578-swiftconf\") pod \"swift-ring-rebalance-9hjmc\" (UID: \"567a67dc-01a7-4720-8081-d60470856578\") " pod="openstack/swift-ring-rebalance-9hjmc" Oct 08 18:28:12 crc kubenswrapper[4988]: I1008 18:28:12.086432 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/567a67dc-01a7-4720-8081-d60470856578-dispersionconf\") pod \"swift-ring-rebalance-9hjmc\" (UID: \"567a67dc-01a7-4720-8081-d60470856578\") " pod="openstack/swift-ring-rebalance-9hjmc" Oct 08 18:28:12 crc kubenswrapper[4988]: I1008 18:28:12.086628 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/567a67dc-01a7-4720-8081-d60470856578-ring-data-devices\") pod \"swift-ring-rebalance-9hjmc\" (UID: \"567a67dc-01a7-4720-8081-d60470856578\") " pod="openstack/swift-ring-rebalance-9hjmc" Oct 08 18:28:12 crc kubenswrapper[4988]: I1008 18:28:12.096305 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6d8fd57975-7jpt4" Oct 08 18:28:12 crc kubenswrapper[4988]: I1008 18:28:12.187320 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/94afe69f-30c7-4023-a485-f16998ed7fdd-dns-svc\") pod \"94afe69f-30c7-4023-a485-f16998ed7fdd\" (UID: \"94afe69f-30c7-4023-a485-f16998ed7fdd\") " Oct 08 18:28:12 crc kubenswrapper[4988]: I1008 18:28:12.189100 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/94afe69f-30c7-4023-a485-f16998ed7fdd-ovsdbserver-nb\") pod \"94afe69f-30c7-4023-a485-f16998ed7fdd\" (UID: \"94afe69f-30c7-4023-a485-f16998ed7fdd\") " Oct 08 18:28:12 crc kubenswrapper[4988]: I1008 18:28:12.189229 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zstlc\" (UniqueName: \"kubernetes.io/projected/94afe69f-30c7-4023-a485-f16998ed7fdd-kube-api-access-zstlc\") pod \"94afe69f-30c7-4023-a485-f16998ed7fdd\" (UID: \"94afe69f-30c7-4023-a485-f16998ed7fdd\") " Oct 08 18:28:12 crc kubenswrapper[4988]: I1008 18:28:12.189290 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/94afe69f-30c7-4023-a485-f16998ed7fdd-config\") pod \"94afe69f-30c7-4023-a485-f16998ed7fdd\" (UID: \"94afe69f-30c7-4023-a485-f16998ed7fdd\") " Oct 08 18:28:12 crc kubenswrapper[4988]: I1008 18:28:12.189707 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/567a67dc-01a7-4720-8081-d60470856578-scripts\") pod \"swift-ring-rebalance-9hjmc\" (UID: \"567a67dc-01a7-4720-8081-d60470856578\") " pod="openstack/swift-ring-rebalance-9hjmc" Oct 08 18:28:12 crc kubenswrapper[4988]: I1008 18:28:12.189795 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/567a67dc-01a7-4720-8081-d60470856578-swiftconf\") pod \"swift-ring-rebalance-9hjmc\" (UID: \"567a67dc-01a7-4720-8081-d60470856578\") " pod="openstack/swift-ring-rebalance-9hjmc" Oct 08 18:28:12 crc kubenswrapper[4988]: I1008 18:28:12.189824 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/567a67dc-01a7-4720-8081-d60470856578-dispersionconf\") pod \"swift-ring-rebalance-9hjmc\" (UID: \"567a67dc-01a7-4720-8081-d60470856578\") " pod="openstack/swift-ring-rebalance-9hjmc" Oct 08 18:28:12 crc kubenswrapper[4988]: I1008 18:28:12.189883 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/567a67dc-01a7-4720-8081-d60470856578-ring-data-devices\") pod \"swift-ring-rebalance-9hjmc\" (UID: \"567a67dc-01a7-4720-8081-d60470856578\") " pod="openstack/swift-ring-rebalance-9hjmc" Oct 08 18:28:12 crc kubenswrapper[4988]: I1008 18:28:12.190225 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/567a67dc-01a7-4720-8081-d60470856578-combined-ca-bundle\") pod \"swift-ring-rebalance-9hjmc\" (UID: \"567a67dc-01a7-4720-8081-d60470856578\") " pod="openstack/swift-ring-rebalance-9hjmc" Oct 08 18:28:12 crc kubenswrapper[4988]: I1008 18:28:12.190738 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/567a67dc-01a7-4720-8081-d60470856578-scripts\") pod \"swift-ring-rebalance-9hjmc\" (UID: \"567a67dc-01a7-4720-8081-d60470856578\") " pod="openstack/swift-ring-rebalance-9hjmc" Oct 08 18:28:12 crc kubenswrapper[4988]: I1008 18:28:12.190753 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hqwhf\" (UniqueName: \"kubernetes.io/projected/567a67dc-01a7-4720-8081-d60470856578-kube-api-access-hqwhf\") pod \"swift-ring-rebalance-9hjmc\" (UID: \"567a67dc-01a7-4720-8081-d60470856578\") " pod="openstack/swift-ring-rebalance-9hjmc" Oct 08 18:28:12 crc kubenswrapper[4988]: I1008 18:28:12.190856 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/567a67dc-01a7-4720-8081-d60470856578-etc-swift\") pod \"swift-ring-rebalance-9hjmc\" (UID: \"567a67dc-01a7-4720-8081-d60470856578\") " pod="openstack/swift-ring-rebalance-9hjmc" Oct 08 18:28:12 crc kubenswrapper[4988]: I1008 18:28:12.190877 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/567a67dc-01a7-4720-8081-d60470856578-ring-data-devices\") pod \"swift-ring-rebalance-9hjmc\" (UID: \"567a67dc-01a7-4720-8081-d60470856578\") " pod="openstack/swift-ring-rebalance-9hjmc" Oct 08 18:28:12 crc kubenswrapper[4988]: I1008 18:28:12.191159 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/567a67dc-01a7-4720-8081-d60470856578-etc-swift\") pod \"swift-ring-rebalance-9hjmc\" (UID: \"567a67dc-01a7-4720-8081-d60470856578\") " pod="openstack/swift-ring-rebalance-9hjmc" Oct 08 18:28:12 crc kubenswrapper[4988]: I1008 18:28:12.193897 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/567a67dc-01a7-4720-8081-d60470856578-dispersionconf\") pod \"swift-ring-rebalance-9hjmc\" (UID: \"567a67dc-01a7-4720-8081-d60470856578\") " pod="openstack/swift-ring-rebalance-9hjmc" Oct 08 18:28:12 crc kubenswrapper[4988]: I1008 18:28:12.194920 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/94afe69f-30c7-4023-a485-f16998ed7fdd-kube-api-access-zstlc" (OuterVolumeSpecName: "kube-api-access-zstlc") pod "94afe69f-30c7-4023-a485-f16998ed7fdd" (UID: "94afe69f-30c7-4023-a485-f16998ed7fdd"). InnerVolumeSpecName "kube-api-access-zstlc". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:28:12 crc kubenswrapper[4988]: I1008 18:28:12.195365 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/567a67dc-01a7-4720-8081-d60470856578-combined-ca-bundle\") pod \"swift-ring-rebalance-9hjmc\" (UID: \"567a67dc-01a7-4720-8081-d60470856578\") " pod="openstack/swift-ring-rebalance-9hjmc" Oct 08 18:28:12 crc kubenswrapper[4988]: I1008 18:28:12.196568 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/567a67dc-01a7-4720-8081-d60470856578-swiftconf\") pod \"swift-ring-rebalance-9hjmc\" (UID: \"567a67dc-01a7-4720-8081-d60470856578\") " pod="openstack/swift-ring-rebalance-9hjmc" Oct 08 18:28:12 crc kubenswrapper[4988]: I1008 18:28:12.206438 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hqwhf\" (UniqueName: \"kubernetes.io/projected/567a67dc-01a7-4720-8081-d60470856578-kube-api-access-hqwhf\") pod \"swift-ring-rebalance-9hjmc\" (UID: \"567a67dc-01a7-4720-8081-d60470856578\") " pod="openstack/swift-ring-rebalance-9hjmc" Oct 08 18:28:12 crc kubenswrapper[4988]: I1008 18:28:12.229790 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/94afe69f-30c7-4023-a485-f16998ed7fdd-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "94afe69f-30c7-4023-a485-f16998ed7fdd" (UID: "94afe69f-30c7-4023-a485-f16998ed7fdd"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 18:28:12 crc kubenswrapper[4988]: I1008 18:28:12.240954 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/94afe69f-30c7-4023-a485-f16998ed7fdd-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "94afe69f-30c7-4023-a485-f16998ed7fdd" (UID: "94afe69f-30c7-4023-a485-f16998ed7fdd"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 18:28:12 crc kubenswrapper[4988]: I1008 18:28:12.241333 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/94afe69f-30c7-4023-a485-f16998ed7fdd-config" (OuterVolumeSpecName: "config") pod "94afe69f-30c7-4023-a485-f16998ed7fdd" (UID: "94afe69f-30c7-4023-a485-f16998ed7fdd"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 18:28:12 crc kubenswrapper[4988]: I1008 18:28:12.292485 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/d04259e5-9595-447b-b804-feba577498d8-etc-swift\") pod \"swift-storage-0\" (UID: \"d04259e5-9595-447b-b804-feba577498d8\") " pod="openstack/swift-storage-0" Oct 08 18:28:12 crc kubenswrapper[4988]: I1008 18:28:12.292644 4988 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zstlc\" (UniqueName: \"kubernetes.io/projected/94afe69f-30c7-4023-a485-f16998ed7fdd-kube-api-access-zstlc\") on node \"crc\" DevicePath \"\"" Oct 08 18:28:12 crc kubenswrapper[4988]: I1008 18:28:12.292669 4988 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/94afe69f-30c7-4023-a485-f16998ed7fdd-config\") on node \"crc\" DevicePath \"\"" Oct 08 18:28:12 crc kubenswrapper[4988]: I1008 18:28:12.292677 4988 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/94afe69f-30c7-4023-a485-f16998ed7fdd-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 08 18:28:12 crc kubenswrapper[4988]: I1008 18:28:12.292687 4988 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/94afe69f-30c7-4023-a485-f16998ed7fdd-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 08 18:28:12 crc kubenswrapper[4988]: E1008 18:28:12.293022 4988 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Oct 08 18:28:12 crc kubenswrapper[4988]: E1008 18:28:12.293104 4988 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Oct 08 18:28:12 crc kubenswrapper[4988]: E1008 18:28:12.293213 4988 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/d04259e5-9595-447b-b804-feba577498d8-etc-swift podName:d04259e5-9595-447b-b804-feba577498d8 nodeName:}" failed. No retries permitted until 2025-10-08 18:28:13.293190661 +0000 UTC m=+1038.743033431 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/d04259e5-9595-447b-b804-feba577498d8-etc-swift") pod "swift-storage-0" (UID: "d04259e5-9595-447b-b804-feba577498d8") : configmap "swift-ring-files" not found Oct 08 18:28:12 crc kubenswrapper[4988]: I1008 18:28:12.390147 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-9hjmc" Oct 08 18:28:12 crc kubenswrapper[4988]: E1008 18:28:12.494569 4988 upgradeaware.go:427] Error proxying data from client to backend: readfrom tcp 38.102.83.102:58486->38.102.83.102:38495: write tcp 38.102.83.102:58486->38.102.83.102:38495: write: broken pipe Oct 08 18:28:12 crc kubenswrapper[4988]: I1008 18:28:12.877321 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57f58c7cff-42svl" event={"ID":"1487d461-4c56-4591-b498-b6ce9f314763","Type":"ContainerStarted","Data":"0bdd344781c086aea94f3fd316e2eefc84f2f8da29e9372eab7e14dd7a7a0061"} Oct 08 18:28:12 crc kubenswrapper[4988]: I1008 18:28:12.877586 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-57f58c7cff-42svl" Oct 08 18:28:12 crc kubenswrapper[4988]: I1008 18:28:12.879363 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6d8fd57975-7jpt4" event={"ID":"94afe69f-30c7-4023-a485-f16998ed7fdd","Type":"ContainerDied","Data":"424a85ff85d3385b98f63e9b5b64ca76eeb3192d5aac5b17f12fff4a6933de4a"} Oct 08 18:28:12 crc kubenswrapper[4988]: I1008 18:28:12.879408 4988 scope.go:117] "RemoveContainer" containerID="65ea95e1c8ef8edcae7c30eec6c6dbf2fcd831415768d6afb24644972378e9cb" Oct 08 18:28:12 crc kubenswrapper[4988]: I1008 18:28:12.879464 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6d8fd57975-7jpt4" Oct 08 18:28:12 crc kubenswrapper[4988]: I1008 18:28:12.900410 4988 scope.go:117] "RemoveContainer" containerID="e623903a25ba6b731e6957da4de44239c9b78815190afaeeebac2b2991024c3f" Oct 08 18:28:12 crc kubenswrapper[4988]: I1008 18:28:12.909236 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-57f58c7cff-42svl" podStartSLOduration=2.9092172720000002 podStartE2EDuration="2.909217272s" podCreationTimestamp="2025-10-08 18:28:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 18:28:12.898740253 +0000 UTC m=+1038.348583023" watchObservedRunningTime="2025-10-08 18:28:12.909217272 +0000 UTC m=+1038.359060042" Oct 08 18:28:12 crc kubenswrapper[4988]: I1008 18:28:12.958662 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6d8fd57975-7jpt4"] Oct 08 18:28:12 crc kubenswrapper[4988]: I1008 18:28:12.966473 4988 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-6d8fd57975-7jpt4"] Oct 08 18:28:12 crc kubenswrapper[4988]: I1008 18:28:12.972093 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-9hjmc"] Oct 08 18:28:13 crc kubenswrapper[4988]: I1008 18:28:13.248899 4988 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="94afe69f-30c7-4023-a485-f16998ed7fdd" path="/var/lib/kubelet/pods/94afe69f-30c7-4023-a485-f16998ed7fdd/volumes" Oct 08 18:28:13 crc kubenswrapper[4988]: I1008 18:28:13.312417 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/d04259e5-9595-447b-b804-feba577498d8-etc-swift\") pod \"swift-storage-0\" (UID: \"d04259e5-9595-447b-b804-feba577498d8\") " pod="openstack/swift-storage-0" Oct 08 18:28:13 crc kubenswrapper[4988]: E1008 18:28:13.312618 4988 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Oct 08 18:28:13 crc kubenswrapper[4988]: E1008 18:28:13.312824 4988 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Oct 08 18:28:13 crc kubenswrapper[4988]: E1008 18:28:13.312935 4988 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/d04259e5-9595-447b-b804-feba577498d8-etc-swift podName:d04259e5-9595-447b-b804-feba577498d8 nodeName:}" failed. No retries permitted until 2025-10-08 18:28:15.312902691 +0000 UTC m=+1040.762745511 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/d04259e5-9595-447b-b804-feba577498d8-etc-swift") pod "swift-storage-0" (UID: "d04259e5-9595-447b-b804-feba577498d8") : configmap "swift-ring-files" not found Oct 08 18:28:13 crc kubenswrapper[4988]: I1008 18:28:13.888637 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-9hjmc" event={"ID":"567a67dc-01a7-4720-8081-d60470856578","Type":"ContainerStarted","Data":"2aa0a658f88581d7181e004be979d6f5b5d85dbe40d43dc66d7c1747b511c24c"} Oct 08 18:28:15 crc kubenswrapper[4988]: I1008 18:28:15.345452 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/d04259e5-9595-447b-b804-feba577498d8-etc-swift\") pod \"swift-storage-0\" (UID: \"d04259e5-9595-447b-b804-feba577498d8\") " pod="openstack/swift-storage-0" Oct 08 18:28:15 crc kubenswrapper[4988]: E1008 18:28:15.345711 4988 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Oct 08 18:28:15 crc kubenswrapper[4988]: E1008 18:28:15.345756 4988 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Oct 08 18:28:15 crc kubenswrapper[4988]: E1008 18:28:15.345834 4988 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/d04259e5-9595-447b-b804-feba577498d8-etc-swift podName:d04259e5-9595-447b-b804-feba577498d8 nodeName:}" failed. No retries permitted until 2025-10-08 18:28:19.345804425 +0000 UTC m=+1044.795647235 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/d04259e5-9595-447b-b804-feba577498d8-etc-swift") pod "swift-storage-0" (UID: "d04259e5-9595-447b-b804-feba577498d8") : configmap "swift-ring-files" not found Oct 08 18:28:15 crc kubenswrapper[4988]: I1008 18:28:15.563711 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-bc45f6dcf-m2rkl" Oct 08 18:28:16 crc kubenswrapper[4988]: I1008 18:28:16.067408 4988 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-cell1-galera-0" Oct 08 18:28:16 crc kubenswrapper[4988]: I1008 18:28:16.118196 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-cell1-galera-0" Oct 08 18:28:17 crc kubenswrapper[4988]: I1008 18:28:17.563137 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-galera-0" Oct 08 18:28:17 crc kubenswrapper[4988]: I1008 18:28:17.921198 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-9hjmc" event={"ID":"567a67dc-01a7-4720-8081-d60470856578","Type":"ContainerStarted","Data":"adcb01b7b9d0f727f2f5ff67b48f5ae3862d402862564f8b83cd5f630ad8e178"} Oct 08 18:28:17 crc kubenswrapper[4988]: I1008 18:28:17.946998 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-ring-rebalance-9hjmc" podStartSLOduration=3.188452364 podStartE2EDuration="6.946975757s" podCreationTimestamp="2025-10-08 18:28:11 +0000 UTC" firstStartedPulling="2025-10-08 18:28:12.960850989 +0000 UTC m=+1038.410693759" lastFinishedPulling="2025-10-08 18:28:16.719374382 +0000 UTC m=+1042.169217152" observedRunningTime="2025-10-08 18:28:17.943421115 +0000 UTC m=+1043.393263935" watchObservedRunningTime="2025-10-08 18:28:17.946975757 +0000 UTC m=+1043.396818527" Oct 08 18:28:18 crc kubenswrapper[4988]: I1008 18:28:18.220845 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-create-fxlld"] Oct 08 18:28:18 crc kubenswrapper[4988]: E1008 18:28:18.221243 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="94afe69f-30c7-4023-a485-f16998ed7fdd" containerName="init" Oct 08 18:28:18 crc kubenswrapper[4988]: I1008 18:28:18.221264 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="94afe69f-30c7-4023-a485-f16998ed7fdd" containerName="init" Oct 08 18:28:18 crc kubenswrapper[4988]: E1008 18:28:18.221287 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="94afe69f-30c7-4023-a485-f16998ed7fdd" containerName="dnsmasq-dns" Oct 08 18:28:18 crc kubenswrapper[4988]: I1008 18:28:18.221295 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="94afe69f-30c7-4023-a485-f16998ed7fdd" containerName="dnsmasq-dns" Oct 08 18:28:18 crc kubenswrapper[4988]: I1008 18:28:18.221501 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="94afe69f-30c7-4023-a485-f16998ed7fdd" containerName="dnsmasq-dns" Oct 08 18:28:18 crc kubenswrapper[4988]: I1008 18:28:18.222203 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-fxlld" Oct 08 18:28:18 crc kubenswrapper[4988]: I1008 18:28:18.234921 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-fxlld"] Oct 08 18:28:18 crc kubenswrapper[4988]: I1008 18:28:18.300798 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dzpdk\" (UniqueName: \"kubernetes.io/projected/ad40a4d0-13f5-4e86-ba37-db0add482b46-kube-api-access-dzpdk\") pod \"keystone-db-create-fxlld\" (UID: \"ad40a4d0-13f5-4e86-ba37-db0add482b46\") " pod="openstack/keystone-db-create-fxlld" Oct 08 18:28:18 crc kubenswrapper[4988]: I1008 18:28:18.402781 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dzpdk\" (UniqueName: \"kubernetes.io/projected/ad40a4d0-13f5-4e86-ba37-db0add482b46-kube-api-access-dzpdk\") pod \"keystone-db-create-fxlld\" (UID: \"ad40a4d0-13f5-4e86-ba37-db0add482b46\") " pod="openstack/keystone-db-create-fxlld" Oct 08 18:28:18 crc kubenswrapper[4988]: I1008 18:28:18.441952 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dzpdk\" (UniqueName: \"kubernetes.io/projected/ad40a4d0-13f5-4e86-ba37-db0add482b46-kube-api-access-dzpdk\") pod \"keystone-db-create-fxlld\" (UID: \"ad40a4d0-13f5-4e86-ba37-db0add482b46\") " pod="openstack/keystone-db-create-fxlld" Oct 08 18:28:18 crc kubenswrapper[4988]: I1008 18:28:18.465143 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-create-xnv8d"] Oct 08 18:28:18 crc kubenswrapper[4988]: I1008 18:28:18.466166 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-xnv8d" Oct 08 18:28:18 crc kubenswrapper[4988]: I1008 18:28:18.481766 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-xnv8d"] Oct 08 18:28:18 crc kubenswrapper[4988]: I1008 18:28:18.504559 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mbncs\" (UniqueName: \"kubernetes.io/projected/fa0ac55f-9e87-488b-8626-36024ca51102-kube-api-access-mbncs\") pod \"placement-db-create-xnv8d\" (UID: \"fa0ac55f-9e87-488b-8626-36024ca51102\") " pod="openstack/placement-db-create-xnv8d" Oct 08 18:28:18 crc kubenswrapper[4988]: I1008 18:28:18.544469 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-fxlld" Oct 08 18:28:18 crc kubenswrapper[4988]: I1008 18:28:18.606464 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mbncs\" (UniqueName: \"kubernetes.io/projected/fa0ac55f-9e87-488b-8626-36024ca51102-kube-api-access-mbncs\") pod \"placement-db-create-xnv8d\" (UID: \"fa0ac55f-9e87-488b-8626-36024ca51102\") " pod="openstack/placement-db-create-xnv8d" Oct 08 18:28:18 crc kubenswrapper[4988]: I1008 18:28:18.627271 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mbncs\" (UniqueName: \"kubernetes.io/projected/fa0ac55f-9e87-488b-8626-36024ca51102-kube-api-access-mbncs\") pod \"placement-db-create-xnv8d\" (UID: \"fa0ac55f-9e87-488b-8626-36024ca51102\") " pod="openstack/placement-db-create-xnv8d" Oct 08 18:28:18 crc kubenswrapper[4988]: I1008 18:28:18.775478 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-create-pspkt"] Oct 08 18:28:18 crc kubenswrapper[4988]: I1008 18:28:18.776662 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-pspkt" Oct 08 18:28:18 crc kubenswrapper[4988]: I1008 18:28:18.781921 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-pspkt"] Oct 08 18:28:18 crc kubenswrapper[4988]: I1008 18:28:18.795357 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-xnv8d" Oct 08 18:28:18 crc kubenswrapper[4988]: I1008 18:28:18.910641 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x625s\" (UniqueName: \"kubernetes.io/projected/881d3e0f-adfd-4b37-a942-6698a6897963-kube-api-access-x625s\") pod \"glance-db-create-pspkt\" (UID: \"881d3e0f-adfd-4b37-a942-6698a6897963\") " pod="openstack/glance-db-create-pspkt" Oct 08 18:28:18 crc kubenswrapper[4988]: I1008 18:28:18.966087 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-fxlld"] Oct 08 18:28:18 crc kubenswrapper[4988]: W1008 18:28:18.973592 4988 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podad40a4d0_13f5_4e86_ba37_db0add482b46.slice/crio-8e3a7fc6dc0bc399afc6ef901a4c74e8b2dbbe66644790748691e9e4a75b75c8 WatchSource:0}: Error finding container 8e3a7fc6dc0bc399afc6ef901a4c74e8b2dbbe66644790748691e9e4a75b75c8: Status 404 returned error can't find the container with id 8e3a7fc6dc0bc399afc6ef901a4c74e8b2dbbe66644790748691e9e4a75b75c8 Oct 08 18:28:19 crc kubenswrapper[4988]: I1008 18:28:19.012769 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x625s\" (UniqueName: \"kubernetes.io/projected/881d3e0f-adfd-4b37-a942-6698a6897963-kube-api-access-x625s\") pod \"glance-db-create-pspkt\" (UID: \"881d3e0f-adfd-4b37-a942-6698a6897963\") " pod="openstack/glance-db-create-pspkt" Oct 08 18:28:19 crc kubenswrapper[4988]: I1008 18:28:19.032061 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x625s\" (UniqueName: \"kubernetes.io/projected/881d3e0f-adfd-4b37-a942-6698a6897963-kube-api-access-x625s\") pod \"glance-db-create-pspkt\" (UID: \"881d3e0f-adfd-4b37-a942-6698a6897963\") " pod="openstack/glance-db-create-pspkt" Oct 08 18:28:19 crc kubenswrapper[4988]: I1008 18:28:19.105068 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-pspkt" Oct 08 18:28:19 crc kubenswrapper[4988]: I1008 18:28:19.288332 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-xnv8d"] Oct 08 18:28:19 crc kubenswrapper[4988]: I1008 18:28:19.419963 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/d04259e5-9595-447b-b804-feba577498d8-etc-swift\") pod \"swift-storage-0\" (UID: \"d04259e5-9595-447b-b804-feba577498d8\") " pod="openstack/swift-storage-0" Oct 08 18:28:19 crc kubenswrapper[4988]: E1008 18:28:19.420228 4988 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Oct 08 18:28:19 crc kubenswrapper[4988]: E1008 18:28:19.420264 4988 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Oct 08 18:28:19 crc kubenswrapper[4988]: E1008 18:28:19.420324 4988 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/d04259e5-9595-447b-b804-feba577498d8-etc-swift podName:d04259e5-9595-447b-b804-feba577498d8 nodeName:}" failed. No retries permitted until 2025-10-08 18:28:27.420303687 +0000 UTC m=+1052.870146447 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/d04259e5-9595-447b-b804-feba577498d8-etc-swift") pod "swift-storage-0" (UID: "d04259e5-9595-447b-b804-feba577498d8") : configmap "swift-ring-files" not found Oct 08 18:28:19 crc kubenswrapper[4988]: I1008 18:28:19.544192 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-pspkt"] Oct 08 18:28:19 crc kubenswrapper[4988]: W1008 18:28:19.548638 4988 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod881d3e0f_adfd_4b37_a942_6698a6897963.slice/crio-389134aa50656a6d5345534afb9a3416e01fabb3d207811d294e2e0ef85253df WatchSource:0}: Error finding container 389134aa50656a6d5345534afb9a3416e01fabb3d207811d294e2e0ef85253df: Status 404 returned error can't find the container with id 389134aa50656a6d5345534afb9a3416e01fabb3d207811d294e2e0ef85253df Oct 08 18:28:19 crc kubenswrapper[4988]: I1008 18:28:19.944958 4988 generic.go:334] "Generic (PLEG): container finished" podID="881d3e0f-adfd-4b37-a942-6698a6897963" containerID="8623084f7aa4582501018d0249b0c75a282e2e977b5c5edd610997795d930f58" exitCode=0 Oct 08 18:28:19 crc kubenswrapper[4988]: I1008 18:28:19.945073 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-pspkt" event={"ID":"881d3e0f-adfd-4b37-a942-6698a6897963","Type":"ContainerDied","Data":"8623084f7aa4582501018d0249b0c75a282e2e977b5c5edd610997795d930f58"} Oct 08 18:28:19 crc kubenswrapper[4988]: I1008 18:28:19.945564 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-pspkt" event={"ID":"881d3e0f-adfd-4b37-a942-6698a6897963","Type":"ContainerStarted","Data":"389134aa50656a6d5345534afb9a3416e01fabb3d207811d294e2e0ef85253df"} Oct 08 18:28:19 crc kubenswrapper[4988]: I1008 18:28:19.949808 4988 generic.go:334] "Generic (PLEG): container finished" podID="fa0ac55f-9e87-488b-8626-36024ca51102" containerID="768f2a566b566e4f387ca37429cbe36110b0ea648df5d258af5492e387e65734" exitCode=0 Oct 08 18:28:19 crc kubenswrapper[4988]: I1008 18:28:19.949878 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-xnv8d" event={"ID":"fa0ac55f-9e87-488b-8626-36024ca51102","Type":"ContainerDied","Data":"768f2a566b566e4f387ca37429cbe36110b0ea648df5d258af5492e387e65734"} Oct 08 18:28:19 crc kubenswrapper[4988]: I1008 18:28:19.949904 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-xnv8d" event={"ID":"fa0ac55f-9e87-488b-8626-36024ca51102","Type":"ContainerStarted","Data":"d11fc3d54f875fb0b53df8b98c77fd6c98668965ae034844ece6a9b6f3d6b9d4"} Oct 08 18:28:19 crc kubenswrapper[4988]: I1008 18:28:19.952685 4988 generic.go:334] "Generic (PLEG): container finished" podID="ad40a4d0-13f5-4e86-ba37-db0add482b46" containerID="0bcd9f7810eb0aa145aba595b3f54d8a03770e0a5e29ab50c4eff8c3b0166474" exitCode=0 Oct 08 18:28:19 crc kubenswrapper[4988]: I1008 18:28:19.952717 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-fxlld" event={"ID":"ad40a4d0-13f5-4e86-ba37-db0add482b46","Type":"ContainerDied","Data":"0bcd9f7810eb0aa145aba595b3f54d8a03770e0a5e29ab50c4eff8c3b0166474"} Oct 08 18:28:19 crc kubenswrapper[4988]: I1008 18:28:19.952861 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-fxlld" event={"ID":"ad40a4d0-13f5-4e86-ba37-db0add482b46","Type":"ContainerStarted","Data":"8e3a7fc6dc0bc399afc6ef901a4c74e8b2dbbe66644790748691e9e4a75b75c8"} Oct 08 18:28:20 crc kubenswrapper[4988]: I1008 18:28:20.624460 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-northd-0" Oct 08 18:28:20 crc kubenswrapper[4988]: I1008 18:28:20.794466 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-57f58c7cff-42svl" Oct 08 18:28:20 crc kubenswrapper[4988]: I1008 18:28:20.844524 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-bc45f6dcf-m2rkl"] Oct 08 18:28:20 crc kubenswrapper[4988]: I1008 18:28:20.844991 4988 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-bc45f6dcf-m2rkl" podUID="9d3361fd-fc49-4087-88b3-835f5bd31e44" containerName="dnsmasq-dns" containerID="cri-o://e1160555585e55ae8048a765c33e98e2dcecb0deea2736accaac322df3b8408e" gracePeriod=10 Oct 08 18:28:21 crc kubenswrapper[4988]: I1008 18:28:21.370818 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-fxlld" Oct 08 18:28:21 crc kubenswrapper[4988]: I1008 18:28:21.456109 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dzpdk\" (UniqueName: \"kubernetes.io/projected/ad40a4d0-13f5-4e86-ba37-db0add482b46-kube-api-access-dzpdk\") pod \"ad40a4d0-13f5-4e86-ba37-db0add482b46\" (UID: \"ad40a4d0-13f5-4e86-ba37-db0add482b46\") " Oct 08 18:28:21 crc kubenswrapper[4988]: I1008 18:28:21.462510 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ad40a4d0-13f5-4e86-ba37-db0add482b46-kube-api-access-dzpdk" (OuterVolumeSpecName: "kube-api-access-dzpdk") pod "ad40a4d0-13f5-4e86-ba37-db0add482b46" (UID: "ad40a4d0-13f5-4e86-ba37-db0add482b46"). InnerVolumeSpecName "kube-api-access-dzpdk". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:28:21 crc kubenswrapper[4988]: I1008 18:28:21.488638 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-xnv8d" Oct 08 18:28:21 crc kubenswrapper[4988]: I1008 18:28:21.494554 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-pspkt" Oct 08 18:28:21 crc kubenswrapper[4988]: I1008 18:28:21.561115 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x625s\" (UniqueName: \"kubernetes.io/projected/881d3e0f-adfd-4b37-a942-6698a6897963-kube-api-access-x625s\") pod \"881d3e0f-adfd-4b37-a942-6698a6897963\" (UID: \"881d3e0f-adfd-4b37-a942-6698a6897963\") " Oct 08 18:28:21 crc kubenswrapper[4988]: I1008 18:28:21.561243 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mbncs\" (UniqueName: \"kubernetes.io/projected/fa0ac55f-9e87-488b-8626-36024ca51102-kube-api-access-mbncs\") pod \"fa0ac55f-9e87-488b-8626-36024ca51102\" (UID: \"fa0ac55f-9e87-488b-8626-36024ca51102\") " Oct 08 18:28:21 crc kubenswrapper[4988]: I1008 18:28:21.561607 4988 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dzpdk\" (UniqueName: \"kubernetes.io/projected/ad40a4d0-13f5-4e86-ba37-db0add482b46-kube-api-access-dzpdk\") on node \"crc\" DevicePath \"\"" Oct 08 18:28:21 crc kubenswrapper[4988]: I1008 18:28:21.564870 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fa0ac55f-9e87-488b-8626-36024ca51102-kube-api-access-mbncs" (OuterVolumeSpecName: "kube-api-access-mbncs") pod "fa0ac55f-9e87-488b-8626-36024ca51102" (UID: "fa0ac55f-9e87-488b-8626-36024ca51102"). InnerVolumeSpecName "kube-api-access-mbncs". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:28:21 crc kubenswrapper[4988]: I1008 18:28:21.565325 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/881d3e0f-adfd-4b37-a942-6698a6897963-kube-api-access-x625s" (OuterVolumeSpecName: "kube-api-access-x625s") pod "881d3e0f-adfd-4b37-a942-6698a6897963" (UID: "881d3e0f-adfd-4b37-a942-6698a6897963"). InnerVolumeSpecName "kube-api-access-x625s". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:28:21 crc kubenswrapper[4988]: I1008 18:28:21.662861 4988 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x625s\" (UniqueName: \"kubernetes.io/projected/881d3e0f-adfd-4b37-a942-6698a6897963-kube-api-access-x625s\") on node \"crc\" DevicePath \"\"" Oct 08 18:28:21 crc kubenswrapper[4988]: I1008 18:28:21.662897 4988 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mbncs\" (UniqueName: \"kubernetes.io/projected/fa0ac55f-9e87-488b-8626-36024ca51102-kube-api-access-mbncs\") on node \"crc\" DevicePath \"\"" Oct 08 18:28:21 crc kubenswrapper[4988]: I1008 18:28:21.843599 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-bc45f6dcf-m2rkl" Oct 08 18:28:21 crc kubenswrapper[4988]: I1008 18:28:21.967852 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-pspkt" event={"ID":"881d3e0f-adfd-4b37-a942-6698a6897963","Type":"ContainerDied","Data":"389134aa50656a6d5345534afb9a3416e01fabb3d207811d294e2e0ef85253df"} Oct 08 18:28:21 crc kubenswrapper[4988]: I1008 18:28:21.967904 4988 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="389134aa50656a6d5345534afb9a3416e01fabb3d207811d294e2e0ef85253df" Oct 08 18:28:21 crc kubenswrapper[4988]: I1008 18:28:21.967967 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-pspkt" Oct 08 18:28:21 crc kubenswrapper[4988]: I1008 18:28:21.968986 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-88c4j\" (UniqueName: \"kubernetes.io/projected/9d3361fd-fc49-4087-88b3-835f5bd31e44-kube-api-access-88c4j\") pod \"9d3361fd-fc49-4087-88b3-835f5bd31e44\" (UID: \"9d3361fd-fc49-4087-88b3-835f5bd31e44\") " Oct 08 18:28:21 crc kubenswrapper[4988]: I1008 18:28:21.969026 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d3361fd-fc49-4087-88b3-835f5bd31e44-config\") pod \"9d3361fd-fc49-4087-88b3-835f5bd31e44\" (UID: \"9d3361fd-fc49-4087-88b3-835f5bd31e44\") " Oct 08 18:28:21 crc kubenswrapper[4988]: I1008 18:28:21.969111 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9d3361fd-fc49-4087-88b3-835f5bd31e44-ovsdbserver-nb\") pod \"9d3361fd-fc49-4087-88b3-835f5bd31e44\" (UID: \"9d3361fd-fc49-4087-88b3-835f5bd31e44\") " Oct 08 18:28:21 crc kubenswrapper[4988]: I1008 18:28:21.969131 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9d3361fd-fc49-4087-88b3-835f5bd31e44-ovsdbserver-sb\") pod \"9d3361fd-fc49-4087-88b3-835f5bd31e44\" (UID: \"9d3361fd-fc49-4087-88b3-835f5bd31e44\") " Oct 08 18:28:21 crc kubenswrapper[4988]: I1008 18:28:21.969190 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9d3361fd-fc49-4087-88b3-835f5bd31e44-dns-svc\") pod \"9d3361fd-fc49-4087-88b3-835f5bd31e44\" (UID: \"9d3361fd-fc49-4087-88b3-835f5bd31e44\") " Oct 08 18:28:21 crc kubenswrapper[4988]: I1008 18:28:21.977518 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-xnv8d" event={"ID":"fa0ac55f-9e87-488b-8626-36024ca51102","Type":"ContainerDied","Data":"d11fc3d54f875fb0b53df8b98c77fd6c98668965ae034844ece6a9b6f3d6b9d4"} Oct 08 18:28:21 crc kubenswrapper[4988]: I1008 18:28:21.977562 4988 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d11fc3d54f875fb0b53df8b98c77fd6c98668965ae034844ece6a9b6f3d6b9d4" Oct 08 18:28:21 crc kubenswrapper[4988]: I1008 18:28:21.977618 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-xnv8d" Oct 08 18:28:21 crc kubenswrapper[4988]: I1008 18:28:21.977882 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9d3361fd-fc49-4087-88b3-835f5bd31e44-kube-api-access-88c4j" (OuterVolumeSpecName: "kube-api-access-88c4j") pod "9d3361fd-fc49-4087-88b3-835f5bd31e44" (UID: "9d3361fd-fc49-4087-88b3-835f5bd31e44"). InnerVolumeSpecName "kube-api-access-88c4j". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:28:21 crc kubenswrapper[4988]: I1008 18:28:21.982002 4988 generic.go:334] "Generic (PLEG): container finished" podID="9d3361fd-fc49-4087-88b3-835f5bd31e44" containerID="e1160555585e55ae8048a765c33e98e2dcecb0deea2736accaac322df3b8408e" exitCode=0 Oct 08 18:28:21 crc kubenswrapper[4988]: I1008 18:28:21.982108 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-bc45f6dcf-m2rkl" event={"ID":"9d3361fd-fc49-4087-88b3-835f5bd31e44","Type":"ContainerDied","Data":"e1160555585e55ae8048a765c33e98e2dcecb0deea2736accaac322df3b8408e"} Oct 08 18:28:21 crc kubenswrapper[4988]: I1008 18:28:21.982116 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-bc45f6dcf-m2rkl" Oct 08 18:28:21 crc kubenswrapper[4988]: I1008 18:28:21.982147 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-bc45f6dcf-m2rkl" event={"ID":"9d3361fd-fc49-4087-88b3-835f5bd31e44","Type":"ContainerDied","Data":"ea0ade9c9c48e4ce89492e90365cd54e0095195de61a78fd2e2865f31e4418fc"} Oct 08 18:28:21 crc kubenswrapper[4988]: I1008 18:28:21.982170 4988 scope.go:117] "RemoveContainer" containerID="e1160555585e55ae8048a765c33e98e2dcecb0deea2736accaac322df3b8408e" Oct 08 18:28:21 crc kubenswrapper[4988]: I1008 18:28:21.984008 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-fxlld" event={"ID":"ad40a4d0-13f5-4e86-ba37-db0add482b46","Type":"ContainerDied","Data":"8e3a7fc6dc0bc399afc6ef901a4c74e8b2dbbe66644790748691e9e4a75b75c8"} Oct 08 18:28:21 crc kubenswrapper[4988]: I1008 18:28:21.984042 4988 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8e3a7fc6dc0bc399afc6ef901a4c74e8b2dbbe66644790748691e9e4a75b75c8" Oct 08 18:28:21 crc kubenswrapper[4988]: I1008 18:28:21.984943 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-fxlld" Oct 08 18:28:22 crc kubenswrapper[4988]: I1008 18:28:22.013892 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d3361fd-fc49-4087-88b3-835f5bd31e44-config" (OuterVolumeSpecName: "config") pod "9d3361fd-fc49-4087-88b3-835f5bd31e44" (UID: "9d3361fd-fc49-4087-88b3-835f5bd31e44"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 18:28:22 crc kubenswrapper[4988]: I1008 18:28:22.030118 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d3361fd-fc49-4087-88b3-835f5bd31e44-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "9d3361fd-fc49-4087-88b3-835f5bd31e44" (UID: "9d3361fd-fc49-4087-88b3-835f5bd31e44"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 18:28:22 crc kubenswrapper[4988]: I1008 18:28:22.055521 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d3361fd-fc49-4087-88b3-835f5bd31e44-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "9d3361fd-fc49-4087-88b3-835f5bd31e44" (UID: "9d3361fd-fc49-4087-88b3-835f5bd31e44"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 18:28:22 crc kubenswrapper[4988]: I1008 18:28:22.065479 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d3361fd-fc49-4087-88b3-835f5bd31e44-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "9d3361fd-fc49-4087-88b3-835f5bd31e44" (UID: "9d3361fd-fc49-4087-88b3-835f5bd31e44"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 18:28:22 crc kubenswrapper[4988]: I1008 18:28:22.071545 4988 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9d3361fd-fc49-4087-88b3-835f5bd31e44-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 08 18:28:22 crc kubenswrapper[4988]: I1008 18:28:22.071575 4988 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9d3361fd-fc49-4087-88b3-835f5bd31e44-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 08 18:28:22 crc kubenswrapper[4988]: I1008 18:28:22.071584 4988 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9d3361fd-fc49-4087-88b3-835f5bd31e44-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 08 18:28:22 crc kubenswrapper[4988]: I1008 18:28:22.071595 4988 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-88c4j\" (UniqueName: \"kubernetes.io/projected/9d3361fd-fc49-4087-88b3-835f5bd31e44-kube-api-access-88c4j\") on node \"crc\" DevicePath \"\"" Oct 08 18:28:22 crc kubenswrapper[4988]: I1008 18:28:22.071604 4988 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d3361fd-fc49-4087-88b3-835f5bd31e44-config\") on node \"crc\" DevicePath \"\"" Oct 08 18:28:22 crc kubenswrapper[4988]: I1008 18:28:22.131672 4988 scope.go:117] "RemoveContainer" containerID="a622f60fa34c7e94c933213e1e70d6b951e184acb6ecf9cf248cfbbf3a54c7f4" Oct 08 18:28:22 crc kubenswrapper[4988]: I1008 18:28:22.152327 4988 scope.go:117] "RemoveContainer" containerID="e1160555585e55ae8048a765c33e98e2dcecb0deea2736accaac322df3b8408e" Oct 08 18:28:22 crc kubenswrapper[4988]: E1008 18:28:22.153611 4988 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e1160555585e55ae8048a765c33e98e2dcecb0deea2736accaac322df3b8408e\": container with ID starting with e1160555585e55ae8048a765c33e98e2dcecb0deea2736accaac322df3b8408e not found: ID does not exist" containerID="e1160555585e55ae8048a765c33e98e2dcecb0deea2736accaac322df3b8408e" Oct 08 18:28:22 crc kubenswrapper[4988]: I1008 18:28:22.153658 4988 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e1160555585e55ae8048a765c33e98e2dcecb0deea2736accaac322df3b8408e"} err="failed to get container status \"e1160555585e55ae8048a765c33e98e2dcecb0deea2736accaac322df3b8408e\": rpc error: code = NotFound desc = could not find container \"e1160555585e55ae8048a765c33e98e2dcecb0deea2736accaac322df3b8408e\": container with ID starting with e1160555585e55ae8048a765c33e98e2dcecb0deea2736accaac322df3b8408e not found: ID does not exist" Oct 08 18:28:22 crc kubenswrapper[4988]: I1008 18:28:22.153695 4988 scope.go:117] "RemoveContainer" containerID="a622f60fa34c7e94c933213e1e70d6b951e184acb6ecf9cf248cfbbf3a54c7f4" Oct 08 18:28:22 crc kubenswrapper[4988]: E1008 18:28:22.154044 4988 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a622f60fa34c7e94c933213e1e70d6b951e184acb6ecf9cf248cfbbf3a54c7f4\": container with ID starting with a622f60fa34c7e94c933213e1e70d6b951e184acb6ecf9cf248cfbbf3a54c7f4 not found: ID does not exist" containerID="a622f60fa34c7e94c933213e1e70d6b951e184acb6ecf9cf248cfbbf3a54c7f4" Oct 08 18:28:22 crc kubenswrapper[4988]: I1008 18:28:22.154066 4988 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a622f60fa34c7e94c933213e1e70d6b951e184acb6ecf9cf248cfbbf3a54c7f4"} err="failed to get container status \"a622f60fa34c7e94c933213e1e70d6b951e184acb6ecf9cf248cfbbf3a54c7f4\": rpc error: code = NotFound desc = could not find container \"a622f60fa34c7e94c933213e1e70d6b951e184acb6ecf9cf248cfbbf3a54c7f4\": container with ID starting with a622f60fa34c7e94c933213e1e70d6b951e184acb6ecf9cf248cfbbf3a54c7f4 not found: ID does not exist" Oct 08 18:28:22 crc kubenswrapper[4988]: I1008 18:28:22.311948 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-bc45f6dcf-m2rkl"] Oct 08 18:28:22 crc kubenswrapper[4988]: I1008 18:28:22.317229 4988 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-bc45f6dcf-m2rkl"] Oct 08 18:28:23 crc kubenswrapper[4988]: I1008 18:28:23.247809 4988 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9d3361fd-fc49-4087-88b3-835f5bd31e44" path="/var/lib/kubelet/pods/9d3361fd-fc49-4087-88b3-835f5bd31e44/volumes" Oct 08 18:28:27 crc kubenswrapper[4988]: I1008 18:28:27.031381 4988 generic.go:334] "Generic (PLEG): container finished" podID="567a67dc-01a7-4720-8081-d60470856578" containerID="adcb01b7b9d0f727f2f5ff67b48f5ae3862d402862564f8b83cd5f630ad8e178" exitCode=0 Oct 08 18:28:27 crc kubenswrapper[4988]: I1008 18:28:27.031486 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-9hjmc" event={"ID":"567a67dc-01a7-4720-8081-d60470856578","Type":"ContainerDied","Data":"adcb01b7b9d0f727f2f5ff67b48f5ae3862d402862564f8b83cd5f630ad8e178"} Oct 08 18:28:27 crc kubenswrapper[4988]: I1008 18:28:27.035476 4988 generic.go:334] "Generic (PLEG): container finished" podID="830d6ec5-0015-4a88-b8e8-0fd47f64fd46" containerID="1002767e17c64f7c9060cbd0059630fd4f3dc91d7c2f7558f00143736e0751e0" exitCode=0 Oct 08 18:28:27 crc kubenswrapper[4988]: I1008 18:28:27.035510 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"830d6ec5-0015-4a88-b8e8-0fd47f64fd46","Type":"ContainerDied","Data":"1002767e17c64f7c9060cbd0059630fd4f3dc91d7c2f7558f00143736e0751e0"} Oct 08 18:28:27 crc kubenswrapper[4988]: I1008 18:28:27.467063 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/d04259e5-9595-447b-b804-feba577498d8-etc-swift\") pod \"swift-storage-0\" (UID: \"d04259e5-9595-447b-b804-feba577498d8\") " pod="openstack/swift-storage-0" Oct 08 18:28:27 crc kubenswrapper[4988]: I1008 18:28:27.474544 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/d04259e5-9595-447b-b804-feba577498d8-etc-swift\") pod \"swift-storage-0\" (UID: \"d04259e5-9595-447b-b804-feba577498d8\") " pod="openstack/swift-storage-0" Oct 08 18:28:27 crc kubenswrapper[4988]: I1008 18:28:27.744839 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Oct 08 18:28:28 crc kubenswrapper[4988]: I1008 18:28:28.060152 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"830d6ec5-0015-4a88-b8e8-0fd47f64fd46","Type":"ContainerStarted","Data":"7aeabf72c789f76cad92e2b157ae1d52318b0fdd5da4ef59e8daac0afedb3f6a"} Oct 08 18:28:28 crc kubenswrapper[4988]: I1008 18:28:28.060862 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Oct 08 18:28:28 crc kubenswrapper[4988]: I1008 18:28:28.086691 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-0" podStartSLOduration=38.361634887 podStartE2EDuration="1m5.086674532s" podCreationTimestamp="2025-10-08 18:27:23 +0000 UTC" firstStartedPulling="2025-10-08 18:27:25.348606189 +0000 UTC m=+990.798448959" lastFinishedPulling="2025-10-08 18:27:52.073645804 +0000 UTC m=+1017.523488604" observedRunningTime="2025-10-08 18:28:28.083849732 +0000 UTC m=+1053.533692512" watchObservedRunningTime="2025-10-08 18:28:28.086674532 +0000 UTC m=+1053.536517292" Oct 08 18:28:28 crc kubenswrapper[4988]: I1008 18:28:28.260445 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-3e59-account-create-rv4kt"] Oct 08 18:28:28 crc kubenswrapper[4988]: E1008 18:28:28.260847 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9d3361fd-fc49-4087-88b3-835f5bd31e44" containerName="init" Oct 08 18:28:28 crc kubenswrapper[4988]: I1008 18:28:28.260866 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="9d3361fd-fc49-4087-88b3-835f5bd31e44" containerName="init" Oct 08 18:28:28 crc kubenswrapper[4988]: E1008 18:28:28.260880 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="881d3e0f-adfd-4b37-a942-6698a6897963" containerName="mariadb-database-create" Oct 08 18:28:28 crc kubenswrapper[4988]: I1008 18:28:28.260886 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="881d3e0f-adfd-4b37-a942-6698a6897963" containerName="mariadb-database-create" Oct 08 18:28:28 crc kubenswrapper[4988]: E1008 18:28:28.260902 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fa0ac55f-9e87-488b-8626-36024ca51102" containerName="mariadb-database-create" Oct 08 18:28:28 crc kubenswrapper[4988]: I1008 18:28:28.260909 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="fa0ac55f-9e87-488b-8626-36024ca51102" containerName="mariadb-database-create" Oct 08 18:28:28 crc kubenswrapper[4988]: E1008 18:28:28.260924 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9d3361fd-fc49-4087-88b3-835f5bd31e44" containerName="dnsmasq-dns" Oct 08 18:28:28 crc kubenswrapper[4988]: I1008 18:28:28.260930 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="9d3361fd-fc49-4087-88b3-835f5bd31e44" containerName="dnsmasq-dns" Oct 08 18:28:28 crc kubenswrapper[4988]: E1008 18:28:28.260939 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ad40a4d0-13f5-4e86-ba37-db0add482b46" containerName="mariadb-database-create" Oct 08 18:28:28 crc kubenswrapper[4988]: I1008 18:28:28.260945 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="ad40a4d0-13f5-4e86-ba37-db0add482b46" containerName="mariadb-database-create" Oct 08 18:28:28 crc kubenswrapper[4988]: I1008 18:28:28.261101 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="881d3e0f-adfd-4b37-a942-6698a6897963" containerName="mariadb-database-create" Oct 08 18:28:28 crc kubenswrapper[4988]: I1008 18:28:28.261114 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="9d3361fd-fc49-4087-88b3-835f5bd31e44" containerName="dnsmasq-dns" Oct 08 18:28:28 crc kubenswrapper[4988]: I1008 18:28:28.261124 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="ad40a4d0-13f5-4e86-ba37-db0add482b46" containerName="mariadb-database-create" Oct 08 18:28:28 crc kubenswrapper[4988]: I1008 18:28:28.261131 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="fa0ac55f-9e87-488b-8626-36024ca51102" containerName="mariadb-database-create" Oct 08 18:28:28 crc kubenswrapper[4988]: I1008 18:28:28.261653 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-3e59-account-create-rv4kt" Oct 08 18:28:28 crc kubenswrapper[4988]: I1008 18:28:28.263818 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-db-secret" Oct 08 18:28:28 crc kubenswrapper[4988]: I1008 18:28:28.281570 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-3e59-account-create-rv4kt"] Oct 08 18:28:28 crc kubenswrapper[4988]: I1008 18:28:28.327266 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-storage-0"] Oct 08 18:28:28 crc kubenswrapper[4988]: I1008 18:28:28.381418 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bhvqn\" (UniqueName: \"kubernetes.io/projected/b1336bd1-4eab-44f9-9b5b-9ae2aa3b379c-kube-api-access-bhvqn\") pod \"keystone-3e59-account-create-rv4kt\" (UID: \"b1336bd1-4eab-44f9-9b5b-9ae2aa3b379c\") " pod="openstack/keystone-3e59-account-create-rv4kt" Oct 08 18:28:28 crc kubenswrapper[4988]: I1008 18:28:28.395730 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-9hjmc" Oct 08 18:28:28 crc kubenswrapper[4988]: I1008 18:28:28.483068 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/567a67dc-01a7-4720-8081-d60470856578-swiftconf\") pod \"567a67dc-01a7-4720-8081-d60470856578\" (UID: \"567a67dc-01a7-4720-8081-d60470856578\") " Oct 08 18:28:28 crc kubenswrapper[4988]: I1008 18:28:28.483440 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/567a67dc-01a7-4720-8081-d60470856578-scripts\") pod \"567a67dc-01a7-4720-8081-d60470856578\" (UID: \"567a67dc-01a7-4720-8081-d60470856578\") " Oct 08 18:28:28 crc kubenswrapper[4988]: I1008 18:28:28.483478 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/567a67dc-01a7-4720-8081-d60470856578-etc-swift\") pod \"567a67dc-01a7-4720-8081-d60470856578\" (UID: \"567a67dc-01a7-4720-8081-d60470856578\") " Oct 08 18:28:28 crc kubenswrapper[4988]: I1008 18:28:28.483500 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hqwhf\" (UniqueName: \"kubernetes.io/projected/567a67dc-01a7-4720-8081-d60470856578-kube-api-access-hqwhf\") pod \"567a67dc-01a7-4720-8081-d60470856578\" (UID: \"567a67dc-01a7-4720-8081-d60470856578\") " Oct 08 18:28:28 crc kubenswrapper[4988]: I1008 18:28:28.483537 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/567a67dc-01a7-4720-8081-d60470856578-ring-data-devices\") pod \"567a67dc-01a7-4720-8081-d60470856578\" (UID: \"567a67dc-01a7-4720-8081-d60470856578\") " Oct 08 18:28:28 crc kubenswrapper[4988]: I1008 18:28:28.483594 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/567a67dc-01a7-4720-8081-d60470856578-combined-ca-bundle\") pod \"567a67dc-01a7-4720-8081-d60470856578\" (UID: \"567a67dc-01a7-4720-8081-d60470856578\") " Oct 08 18:28:28 crc kubenswrapper[4988]: I1008 18:28:28.483609 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/567a67dc-01a7-4720-8081-d60470856578-dispersionconf\") pod \"567a67dc-01a7-4720-8081-d60470856578\" (UID: \"567a67dc-01a7-4720-8081-d60470856578\") " Oct 08 18:28:28 crc kubenswrapper[4988]: I1008 18:28:28.483853 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bhvqn\" (UniqueName: \"kubernetes.io/projected/b1336bd1-4eab-44f9-9b5b-9ae2aa3b379c-kube-api-access-bhvqn\") pod \"keystone-3e59-account-create-rv4kt\" (UID: \"b1336bd1-4eab-44f9-9b5b-9ae2aa3b379c\") " pod="openstack/keystone-3e59-account-create-rv4kt" Oct 08 18:28:28 crc kubenswrapper[4988]: I1008 18:28:28.486103 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/567a67dc-01a7-4720-8081-d60470856578-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "567a67dc-01a7-4720-8081-d60470856578" (UID: "567a67dc-01a7-4720-8081-d60470856578"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 18:28:28 crc kubenswrapper[4988]: I1008 18:28:28.486927 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/567a67dc-01a7-4720-8081-d60470856578-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "567a67dc-01a7-4720-8081-d60470856578" (UID: "567a67dc-01a7-4720-8081-d60470856578"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 18:28:28 crc kubenswrapper[4988]: I1008 18:28:28.489659 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/567a67dc-01a7-4720-8081-d60470856578-kube-api-access-hqwhf" (OuterVolumeSpecName: "kube-api-access-hqwhf") pod "567a67dc-01a7-4720-8081-d60470856578" (UID: "567a67dc-01a7-4720-8081-d60470856578"). InnerVolumeSpecName "kube-api-access-hqwhf". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:28:28 crc kubenswrapper[4988]: I1008 18:28:28.492110 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/567a67dc-01a7-4720-8081-d60470856578-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "567a67dc-01a7-4720-8081-d60470856578" (UID: "567a67dc-01a7-4720-8081-d60470856578"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:28:28 crc kubenswrapper[4988]: I1008 18:28:28.503703 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bhvqn\" (UniqueName: \"kubernetes.io/projected/b1336bd1-4eab-44f9-9b5b-9ae2aa3b379c-kube-api-access-bhvqn\") pod \"keystone-3e59-account-create-rv4kt\" (UID: \"b1336bd1-4eab-44f9-9b5b-9ae2aa3b379c\") " pod="openstack/keystone-3e59-account-create-rv4kt" Oct 08 18:28:28 crc kubenswrapper[4988]: I1008 18:28:28.506589 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/567a67dc-01a7-4720-8081-d60470856578-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "567a67dc-01a7-4720-8081-d60470856578" (UID: "567a67dc-01a7-4720-8081-d60470856578"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:28:28 crc kubenswrapper[4988]: I1008 18:28:28.507870 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/567a67dc-01a7-4720-8081-d60470856578-scripts" (OuterVolumeSpecName: "scripts") pod "567a67dc-01a7-4720-8081-d60470856578" (UID: "567a67dc-01a7-4720-8081-d60470856578"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 18:28:28 crc kubenswrapper[4988]: I1008 18:28:28.509571 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/567a67dc-01a7-4720-8081-d60470856578-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "567a67dc-01a7-4720-8081-d60470856578" (UID: "567a67dc-01a7-4720-8081-d60470856578"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:28:28 crc kubenswrapper[4988]: I1008 18:28:28.585929 4988 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/567a67dc-01a7-4720-8081-d60470856578-scripts\") on node \"crc\" DevicePath \"\"" Oct 08 18:28:28 crc kubenswrapper[4988]: I1008 18:28:28.585972 4988 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/567a67dc-01a7-4720-8081-d60470856578-etc-swift\") on node \"crc\" DevicePath \"\"" Oct 08 18:28:28 crc kubenswrapper[4988]: I1008 18:28:28.585987 4988 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hqwhf\" (UniqueName: \"kubernetes.io/projected/567a67dc-01a7-4720-8081-d60470856578-kube-api-access-hqwhf\") on node \"crc\" DevicePath \"\"" Oct 08 18:28:28 crc kubenswrapper[4988]: I1008 18:28:28.586005 4988 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/567a67dc-01a7-4720-8081-d60470856578-ring-data-devices\") on node \"crc\" DevicePath \"\"" Oct 08 18:28:28 crc kubenswrapper[4988]: I1008 18:28:28.586016 4988 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/567a67dc-01a7-4720-8081-d60470856578-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 18:28:28 crc kubenswrapper[4988]: I1008 18:28:28.586024 4988 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/567a67dc-01a7-4720-8081-d60470856578-dispersionconf\") on node \"crc\" DevicePath \"\"" Oct 08 18:28:28 crc kubenswrapper[4988]: I1008 18:28:28.586032 4988 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/567a67dc-01a7-4720-8081-d60470856578-swiftconf\") on node \"crc\" DevicePath \"\"" Oct 08 18:28:28 crc kubenswrapper[4988]: I1008 18:28:28.586438 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-3e59-account-create-rv4kt" Oct 08 18:28:28 crc kubenswrapper[4988]: I1008 18:28:28.660994 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-79af-account-create-ddfvr"] Oct 08 18:28:28 crc kubenswrapper[4988]: E1008 18:28:28.661381 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="567a67dc-01a7-4720-8081-d60470856578" containerName="swift-ring-rebalance" Oct 08 18:28:28 crc kubenswrapper[4988]: I1008 18:28:28.661418 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="567a67dc-01a7-4720-8081-d60470856578" containerName="swift-ring-rebalance" Oct 08 18:28:28 crc kubenswrapper[4988]: I1008 18:28:28.661699 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="567a67dc-01a7-4720-8081-d60470856578" containerName="swift-ring-rebalance" Oct 08 18:28:28 crc kubenswrapper[4988]: I1008 18:28:28.662319 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-79af-account-create-ddfvr" Oct 08 18:28:28 crc kubenswrapper[4988]: I1008 18:28:28.663970 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-db-secret" Oct 08 18:28:28 crc kubenswrapper[4988]: I1008 18:28:28.685814 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-79af-account-create-ddfvr"] Oct 08 18:28:28 crc kubenswrapper[4988]: I1008 18:28:28.791985 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qwfr5\" (UniqueName: \"kubernetes.io/projected/f61ca6e6-56e3-4933-a8aa-5faa53c2b338-kube-api-access-qwfr5\") pod \"placement-79af-account-create-ddfvr\" (UID: \"f61ca6e6-56e3-4933-a8aa-5faa53c2b338\") " pod="openstack/placement-79af-account-create-ddfvr" Oct 08 18:28:28 crc kubenswrapper[4988]: I1008 18:28:28.855980 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-c9f7-account-create-p5k9l"] Oct 08 18:28:28 crc kubenswrapper[4988]: I1008 18:28:28.857648 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-c9f7-account-create-p5k9l" Oct 08 18:28:28 crc kubenswrapper[4988]: I1008 18:28:28.860942 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-db-secret" Oct 08 18:28:28 crc kubenswrapper[4988]: I1008 18:28:28.865843 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-c9f7-account-create-p5k9l"] Oct 08 18:28:28 crc kubenswrapper[4988]: I1008 18:28:28.895171 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xqbmj\" (UniqueName: \"kubernetes.io/projected/689970ec-f287-451e-bf7d-fde8af627bf1-kube-api-access-xqbmj\") pod \"glance-c9f7-account-create-p5k9l\" (UID: \"689970ec-f287-451e-bf7d-fde8af627bf1\") " pod="openstack/glance-c9f7-account-create-p5k9l" Oct 08 18:28:28 crc kubenswrapper[4988]: I1008 18:28:28.895239 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qwfr5\" (UniqueName: \"kubernetes.io/projected/f61ca6e6-56e3-4933-a8aa-5faa53c2b338-kube-api-access-qwfr5\") pod \"placement-79af-account-create-ddfvr\" (UID: \"f61ca6e6-56e3-4933-a8aa-5faa53c2b338\") " pod="openstack/placement-79af-account-create-ddfvr" Oct 08 18:28:28 crc kubenswrapper[4988]: I1008 18:28:28.910199 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qwfr5\" (UniqueName: \"kubernetes.io/projected/f61ca6e6-56e3-4933-a8aa-5faa53c2b338-kube-api-access-qwfr5\") pod \"placement-79af-account-create-ddfvr\" (UID: \"f61ca6e6-56e3-4933-a8aa-5faa53c2b338\") " pod="openstack/placement-79af-account-create-ddfvr" Oct 08 18:28:28 crc kubenswrapper[4988]: I1008 18:28:28.977287 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-79af-account-create-ddfvr" Oct 08 18:28:28 crc kubenswrapper[4988]: I1008 18:28:28.996145 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xqbmj\" (UniqueName: \"kubernetes.io/projected/689970ec-f287-451e-bf7d-fde8af627bf1-kube-api-access-xqbmj\") pod \"glance-c9f7-account-create-p5k9l\" (UID: \"689970ec-f287-451e-bf7d-fde8af627bf1\") " pod="openstack/glance-c9f7-account-create-p5k9l" Oct 08 18:28:29 crc kubenswrapper[4988]: I1008 18:28:29.015701 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xqbmj\" (UniqueName: \"kubernetes.io/projected/689970ec-f287-451e-bf7d-fde8af627bf1-kube-api-access-xqbmj\") pod \"glance-c9f7-account-create-p5k9l\" (UID: \"689970ec-f287-451e-bf7d-fde8af627bf1\") " pod="openstack/glance-c9f7-account-create-p5k9l" Oct 08 18:28:29 crc kubenswrapper[4988]: I1008 18:28:29.044878 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-3e59-account-create-rv4kt"] Oct 08 18:28:29 crc kubenswrapper[4988]: W1008 18:28:29.052934 4988 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb1336bd1_4eab_44f9_9b5b_9ae2aa3b379c.slice/crio-c48836d6537d6d046f00794d027a2e87bc5dc1bc62fc86bf0dd84a35ffcc0a55 WatchSource:0}: Error finding container c48836d6537d6d046f00794d027a2e87bc5dc1bc62fc86bf0dd84a35ffcc0a55: Status 404 returned error can't find the container with id c48836d6537d6d046f00794d027a2e87bc5dc1bc62fc86bf0dd84a35ffcc0a55 Oct 08 18:28:29 crc kubenswrapper[4988]: I1008 18:28:29.090801 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-3e59-account-create-rv4kt" event={"ID":"b1336bd1-4eab-44f9-9b5b-9ae2aa3b379c","Type":"ContainerStarted","Data":"c48836d6537d6d046f00794d027a2e87bc5dc1bc62fc86bf0dd84a35ffcc0a55"} Oct 08 18:28:29 crc kubenswrapper[4988]: I1008 18:28:29.098030 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-9hjmc" event={"ID":"567a67dc-01a7-4720-8081-d60470856578","Type":"ContainerDied","Data":"2aa0a658f88581d7181e004be979d6f5b5d85dbe40d43dc66d7c1747b511c24c"} Oct 08 18:28:29 crc kubenswrapper[4988]: I1008 18:28:29.098081 4988 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2aa0a658f88581d7181e004be979d6f5b5d85dbe40d43dc66d7c1747b511c24c" Oct 08 18:28:29 crc kubenswrapper[4988]: I1008 18:28:29.099157 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-9hjmc" Oct 08 18:28:29 crc kubenswrapper[4988]: I1008 18:28:29.100011 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"d04259e5-9595-447b-b804-feba577498d8","Type":"ContainerStarted","Data":"635706ae6f49645d3fc33dca27b64639e03636df81aa81c90a13bafd1aca31df"} Oct 08 18:28:29 crc kubenswrapper[4988]: I1008 18:28:29.181860 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-c9f7-account-create-p5k9l" Oct 08 18:28:29 crc kubenswrapper[4988]: I1008 18:28:29.338785 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-kc7fw" Oct 08 18:28:29 crc kubenswrapper[4988]: I1008 18:28:29.352922 4988 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-gp8f4" podUID="c285e09d-c2f5-4642-9e6b-59b0ed3ed651" containerName="ovn-controller" probeResult="failure" output=< Oct 08 18:28:29 crc kubenswrapper[4988]: ERROR - ovn-controller connection status is 'not connected', expecting 'connected' status Oct 08 18:28:29 crc kubenswrapper[4988]: > Oct 08 18:28:29 crc kubenswrapper[4988]: I1008 18:28:29.504453 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-79af-account-create-ddfvr"] Oct 08 18:28:29 crc kubenswrapper[4988]: W1008 18:28:29.507649 4988 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf61ca6e6_56e3_4933_a8aa_5faa53c2b338.slice/crio-93e0bc67d692469ebd6b61d83d51a7e844dc145c0a06069efc482b33d2f99ed8 WatchSource:0}: Error finding container 93e0bc67d692469ebd6b61d83d51a7e844dc145c0a06069efc482b33d2f99ed8: Status 404 returned error can't find the container with id 93e0bc67d692469ebd6b61d83d51a7e844dc145c0a06069efc482b33d2f99ed8 Oct 08 18:28:29 crc kubenswrapper[4988]: I1008 18:28:29.648115 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-c9f7-account-create-p5k9l"] Oct 08 18:28:29 crc kubenswrapper[4988]: W1008 18:28:29.654612 4988 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod689970ec_f287_451e_bf7d_fde8af627bf1.slice/crio-fab72c67ee67f3e5f66f52758f41961cbccd4fff08c574c445b8c46dd7bab8e5 WatchSource:0}: Error finding container fab72c67ee67f3e5f66f52758f41961cbccd4fff08c574c445b8c46dd7bab8e5: Status 404 returned error can't find the container with id fab72c67ee67f3e5f66f52758f41961cbccd4fff08c574c445b8c46dd7bab8e5 Oct 08 18:28:30 crc kubenswrapper[4988]: E1008 18:28:30.107569 4988 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod689970ec_f287_451e_bf7d_fde8af627bf1.slice/crio-conmon-a57d3af529ac7b94dd865e100104b122dc8c82e5f1730d8857445b107d9a020d.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf61ca6e6_56e3_4933_a8aa_5faa53c2b338.slice/crio-conmon-b870a785f013fafbe59c8f5824890a66f0b0d43937a44d2f40d45448ece61d72.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod689970ec_f287_451e_bf7d_fde8af627bf1.slice/crio-a57d3af529ac7b94dd865e100104b122dc8c82e5f1730d8857445b107d9a020d.scope\": RecentStats: unable to find data in memory cache]" Oct 08 18:28:30 crc kubenswrapper[4988]: I1008 18:28:30.108178 4988 generic.go:334] "Generic (PLEG): container finished" podID="f61ca6e6-56e3-4933-a8aa-5faa53c2b338" containerID="b870a785f013fafbe59c8f5824890a66f0b0d43937a44d2f40d45448ece61d72" exitCode=0 Oct 08 18:28:30 crc kubenswrapper[4988]: I1008 18:28:30.108203 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-79af-account-create-ddfvr" event={"ID":"f61ca6e6-56e3-4933-a8aa-5faa53c2b338","Type":"ContainerDied","Data":"b870a785f013fafbe59c8f5824890a66f0b0d43937a44d2f40d45448ece61d72"} Oct 08 18:28:30 crc kubenswrapper[4988]: I1008 18:28:30.108563 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-79af-account-create-ddfvr" event={"ID":"f61ca6e6-56e3-4933-a8aa-5faa53c2b338","Type":"ContainerStarted","Data":"93e0bc67d692469ebd6b61d83d51a7e844dc145c0a06069efc482b33d2f99ed8"} Oct 08 18:28:30 crc kubenswrapper[4988]: I1008 18:28:30.110574 4988 generic.go:334] "Generic (PLEG): container finished" podID="689970ec-f287-451e-bf7d-fde8af627bf1" containerID="a57d3af529ac7b94dd865e100104b122dc8c82e5f1730d8857445b107d9a020d" exitCode=0 Oct 08 18:28:30 crc kubenswrapper[4988]: I1008 18:28:30.110668 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-c9f7-account-create-p5k9l" event={"ID":"689970ec-f287-451e-bf7d-fde8af627bf1","Type":"ContainerDied","Data":"a57d3af529ac7b94dd865e100104b122dc8c82e5f1730d8857445b107d9a020d"} Oct 08 18:28:30 crc kubenswrapper[4988]: I1008 18:28:30.110683 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-c9f7-account-create-p5k9l" event={"ID":"689970ec-f287-451e-bf7d-fde8af627bf1","Type":"ContainerStarted","Data":"fab72c67ee67f3e5f66f52758f41961cbccd4fff08c574c445b8c46dd7bab8e5"} Oct 08 18:28:30 crc kubenswrapper[4988]: I1008 18:28:30.112713 4988 generic.go:334] "Generic (PLEG): container finished" podID="b1336bd1-4eab-44f9-9b5b-9ae2aa3b379c" containerID="632678ba07992badaec60f5c80d6cf410b41634603190807315e2c76b85f8b5c" exitCode=0 Oct 08 18:28:30 crc kubenswrapper[4988]: I1008 18:28:30.112743 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-3e59-account-create-rv4kt" event={"ID":"b1336bd1-4eab-44f9-9b5b-9ae2aa3b379c","Type":"ContainerDied","Data":"632678ba07992badaec60f5c80d6cf410b41634603190807315e2c76b85f8b5c"} Oct 08 18:28:31 crc kubenswrapper[4988]: I1008 18:28:31.605507 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-c9f7-account-create-p5k9l" Oct 08 18:28:31 crc kubenswrapper[4988]: I1008 18:28:31.622892 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-79af-account-create-ddfvr" Oct 08 18:28:31 crc kubenswrapper[4988]: I1008 18:28:31.629456 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-3e59-account-create-rv4kt" Oct 08 18:28:31 crc kubenswrapper[4988]: I1008 18:28:31.647981 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xqbmj\" (UniqueName: \"kubernetes.io/projected/689970ec-f287-451e-bf7d-fde8af627bf1-kube-api-access-xqbmj\") pod \"689970ec-f287-451e-bf7d-fde8af627bf1\" (UID: \"689970ec-f287-451e-bf7d-fde8af627bf1\") " Oct 08 18:28:31 crc kubenswrapper[4988]: I1008 18:28:31.648031 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qwfr5\" (UniqueName: \"kubernetes.io/projected/f61ca6e6-56e3-4933-a8aa-5faa53c2b338-kube-api-access-qwfr5\") pod \"f61ca6e6-56e3-4933-a8aa-5faa53c2b338\" (UID: \"f61ca6e6-56e3-4933-a8aa-5faa53c2b338\") " Oct 08 18:28:31 crc kubenswrapper[4988]: I1008 18:28:31.669320 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/689970ec-f287-451e-bf7d-fde8af627bf1-kube-api-access-xqbmj" (OuterVolumeSpecName: "kube-api-access-xqbmj") pod "689970ec-f287-451e-bf7d-fde8af627bf1" (UID: "689970ec-f287-451e-bf7d-fde8af627bf1"). InnerVolumeSpecName "kube-api-access-xqbmj". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:28:31 crc kubenswrapper[4988]: I1008 18:28:31.669467 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f61ca6e6-56e3-4933-a8aa-5faa53c2b338-kube-api-access-qwfr5" (OuterVolumeSpecName: "kube-api-access-qwfr5") pod "f61ca6e6-56e3-4933-a8aa-5faa53c2b338" (UID: "f61ca6e6-56e3-4933-a8aa-5faa53c2b338"). InnerVolumeSpecName "kube-api-access-qwfr5". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:28:31 crc kubenswrapper[4988]: I1008 18:28:31.749240 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bhvqn\" (UniqueName: \"kubernetes.io/projected/b1336bd1-4eab-44f9-9b5b-9ae2aa3b379c-kube-api-access-bhvqn\") pod \"b1336bd1-4eab-44f9-9b5b-9ae2aa3b379c\" (UID: \"b1336bd1-4eab-44f9-9b5b-9ae2aa3b379c\") " Oct 08 18:28:31 crc kubenswrapper[4988]: I1008 18:28:31.749745 4988 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xqbmj\" (UniqueName: \"kubernetes.io/projected/689970ec-f287-451e-bf7d-fde8af627bf1-kube-api-access-xqbmj\") on node \"crc\" DevicePath \"\"" Oct 08 18:28:31 crc kubenswrapper[4988]: I1008 18:28:31.749761 4988 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qwfr5\" (UniqueName: \"kubernetes.io/projected/f61ca6e6-56e3-4933-a8aa-5faa53c2b338-kube-api-access-qwfr5\") on node \"crc\" DevicePath \"\"" Oct 08 18:28:31 crc kubenswrapper[4988]: I1008 18:28:31.752255 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b1336bd1-4eab-44f9-9b5b-9ae2aa3b379c-kube-api-access-bhvqn" (OuterVolumeSpecName: "kube-api-access-bhvqn") pod "b1336bd1-4eab-44f9-9b5b-9ae2aa3b379c" (UID: "b1336bd1-4eab-44f9-9b5b-9ae2aa3b379c"). InnerVolumeSpecName "kube-api-access-bhvqn". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:28:31 crc kubenswrapper[4988]: I1008 18:28:31.851671 4988 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bhvqn\" (UniqueName: \"kubernetes.io/projected/b1336bd1-4eab-44f9-9b5b-9ae2aa3b379c-kube-api-access-bhvqn\") on node \"crc\" DevicePath \"\"" Oct 08 18:28:32 crc kubenswrapper[4988]: I1008 18:28:32.133574 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-79af-account-create-ddfvr" event={"ID":"f61ca6e6-56e3-4933-a8aa-5faa53c2b338","Type":"ContainerDied","Data":"93e0bc67d692469ebd6b61d83d51a7e844dc145c0a06069efc482b33d2f99ed8"} Oct 08 18:28:32 crc kubenswrapper[4988]: I1008 18:28:32.133626 4988 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="93e0bc67d692469ebd6b61d83d51a7e844dc145c0a06069efc482b33d2f99ed8" Oct 08 18:28:32 crc kubenswrapper[4988]: I1008 18:28:32.133695 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-79af-account-create-ddfvr" Oct 08 18:28:32 crc kubenswrapper[4988]: I1008 18:28:32.153959 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"d04259e5-9595-447b-b804-feba577498d8","Type":"ContainerStarted","Data":"e8f70c43dc552f28dc00baeb404e73ce329584c5d7563ccfd1bb7e213c973bc7"} Oct 08 18:28:32 crc kubenswrapper[4988]: I1008 18:28:32.154012 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"d04259e5-9595-447b-b804-feba577498d8","Type":"ContainerStarted","Data":"588710f5625320484d4ef400733b3ae63b0e2f936f619df2ca3b5435aa10a8c9"} Oct 08 18:28:32 crc kubenswrapper[4988]: I1008 18:28:32.154025 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"d04259e5-9595-447b-b804-feba577498d8","Type":"ContainerStarted","Data":"23ee960ca35b0ceb291c74103d14d93f4ff48fb0b2b08acdacc1fd66a0fb6327"} Oct 08 18:28:32 crc kubenswrapper[4988]: I1008 18:28:32.154036 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"d04259e5-9595-447b-b804-feba577498d8","Type":"ContainerStarted","Data":"d0e7528b976d6c731e5ab211ed178a97c7d155611aadc8e3e7dde446d2824719"} Oct 08 18:28:32 crc kubenswrapper[4988]: I1008 18:28:32.158458 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-c9f7-account-create-p5k9l" event={"ID":"689970ec-f287-451e-bf7d-fde8af627bf1","Type":"ContainerDied","Data":"fab72c67ee67f3e5f66f52758f41961cbccd4fff08c574c445b8c46dd7bab8e5"} Oct 08 18:28:32 crc kubenswrapper[4988]: I1008 18:28:32.158489 4988 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fab72c67ee67f3e5f66f52758f41961cbccd4fff08c574c445b8c46dd7bab8e5" Oct 08 18:28:32 crc kubenswrapper[4988]: I1008 18:28:32.158535 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-c9f7-account-create-p5k9l" Oct 08 18:28:32 crc kubenswrapper[4988]: I1008 18:28:32.166971 4988 generic.go:334] "Generic (PLEG): container finished" podID="e5a8f819-c91d-429e-b848-e9c444739623" containerID="586d76b71bf326ca8c68b69f484afc42dbe51e835567ee76b7228cb18c8b1384" exitCode=0 Oct 08 18:28:32 crc kubenswrapper[4988]: I1008 18:28:32.167032 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"e5a8f819-c91d-429e-b848-e9c444739623","Type":"ContainerDied","Data":"586d76b71bf326ca8c68b69f484afc42dbe51e835567ee76b7228cb18c8b1384"} Oct 08 18:28:32 crc kubenswrapper[4988]: I1008 18:28:32.169548 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-3e59-account-create-rv4kt" event={"ID":"b1336bd1-4eab-44f9-9b5b-9ae2aa3b379c","Type":"ContainerDied","Data":"c48836d6537d6d046f00794d027a2e87bc5dc1bc62fc86bf0dd84a35ffcc0a55"} Oct 08 18:28:32 crc kubenswrapper[4988]: I1008 18:28:32.169588 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-3e59-account-create-rv4kt" Oct 08 18:28:32 crc kubenswrapper[4988]: I1008 18:28:32.169595 4988 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c48836d6537d6d046f00794d027a2e87bc5dc1bc62fc86bf0dd84a35ffcc0a55" Oct 08 18:28:33 crc kubenswrapper[4988]: I1008 18:28:33.179654 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"e5a8f819-c91d-429e-b848-e9c444739623","Type":"ContainerStarted","Data":"bb5d4834c48dac50c4e9cebb831a5dc39438bd589b127972d2afef8419d96b2f"} Oct 08 18:28:33 crc kubenswrapper[4988]: I1008 18:28:33.180439 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Oct 08 18:28:34 crc kubenswrapper[4988]: I1008 18:28:34.082660 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-cell1-server-0" podStartSLOduration=-9223371965.772142 podStartE2EDuration="1m11.082632485s" podCreationTimestamp="2025-10-08 18:27:23 +0000 UTC" firstStartedPulling="2025-10-08 18:27:25.927202671 +0000 UTC m=+991.377045441" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 18:28:33.204544556 +0000 UTC m=+1058.654387336" watchObservedRunningTime="2025-10-08 18:28:34.082632485 +0000 UTC m=+1059.532475255" Oct 08 18:28:34 crc kubenswrapper[4988]: I1008 18:28:34.083732 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-sync-5v5z9"] Oct 08 18:28:34 crc kubenswrapper[4988]: E1008 18:28:34.084132 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b1336bd1-4eab-44f9-9b5b-9ae2aa3b379c" containerName="mariadb-account-create" Oct 08 18:28:34 crc kubenswrapper[4988]: I1008 18:28:34.084151 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="b1336bd1-4eab-44f9-9b5b-9ae2aa3b379c" containerName="mariadb-account-create" Oct 08 18:28:34 crc kubenswrapper[4988]: E1008 18:28:34.084180 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f61ca6e6-56e3-4933-a8aa-5faa53c2b338" containerName="mariadb-account-create" Oct 08 18:28:34 crc kubenswrapper[4988]: I1008 18:28:34.084189 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="f61ca6e6-56e3-4933-a8aa-5faa53c2b338" containerName="mariadb-account-create" Oct 08 18:28:34 crc kubenswrapper[4988]: E1008 18:28:34.084213 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="689970ec-f287-451e-bf7d-fde8af627bf1" containerName="mariadb-account-create" Oct 08 18:28:34 crc kubenswrapper[4988]: I1008 18:28:34.084221 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="689970ec-f287-451e-bf7d-fde8af627bf1" containerName="mariadb-account-create" Oct 08 18:28:34 crc kubenswrapper[4988]: I1008 18:28:34.084411 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="b1336bd1-4eab-44f9-9b5b-9ae2aa3b379c" containerName="mariadb-account-create" Oct 08 18:28:34 crc kubenswrapper[4988]: I1008 18:28:34.084443 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="f61ca6e6-56e3-4933-a8aa-5faa53c2b338" containerName="mariadb-account-create" Oct 08 18:28:34 crc kubenswrapper[4988]: I1008 18:28:34.084475 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="689970ec-f287-451e-bf7d-fde8af627bf1" containerName="mariadb-account-create" Oct 08 18:28:34 crc kubenswrapper[4988]: I1008 18:28:34.085071 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-5v5z9" Oct 08 18:28:34 crc kubenswrapper[4988]: I1008 18:28:34.093864 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-config-data" Oct 08 18:28:34 crc kubenswrapper[4988]: I1008 18:28:34.094030 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-5rwjf" Oct 08 18:28:34 crc kubenswrapper[4988]: I1008 18:28:34.095960 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-5v5z9"] Oct 08 18:28:34 crc kubenswrapper[4988]: I1008 18:28:34.185822 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c1f2511c-e7bc-4433-910b-eb3b94fcf09b-config-data\") pod \"glance-db-sync-5v5z9\" (UID: \"c1f2511c-e7bc-4433-910b-eb3b94fcf09b\") " pod="openstack/glance-db-sync-5v5z9" Oct 08 18:28:34 crc kubenswrapper[4988]: I1008 18:28:34.185867 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c1f2511c-e7bc-4433-910b-eb3b94fcf09b-combined-ca-bundle\") pod \"glance-db-sync-5v5z9\" (UID: \"c1f2511c-e7bc-4433-910b-eb3b94fcf09b\") " pod="openstack/glance-db-sync-5v5z9" Oct 08 18:28:34 crc kubenswrapper[4988]: I1008 18:28:34.185905 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/c1f2511c-e7bc-4433-910b-eb3b94fcf09b-db-sync-config-data\") pod \"glance-db-sync-5v5z9\" (UID: \"c1f2511c-e7bc-4433-910b-eb3b94fcf09b\") " pod="openstack/glance-db-sync-5v5z9" Oct 08 18:28:34 crc kubenswrapper[4988]: I1008 18:28:34.186133 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-654gh\" (UniqueName: \"kubernetes.io/projected/c1f2511c-e7bc-4433-910b-eb3b94fcf09b-kube-api-access-654gh\") pod \"glance-db-sync-5v5z9\" (UID: \"c1f2511c-e7bc-4433-910b-eb3b94fcf09b\") " pod="openstack/glance-db-sync-5v5z9" Oct 08 18:28:34 crc kubenswrapper[4988]: I1008 18:28:34.193561 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"d04259e5-9595-447b-b804-feba577498d8","Type":"ContainerStarted","Data":"c1f4256e3ed5eca7a8d3e7c8f01715b4bf6cc7778eac4b717acbddef65864b05"} Oct 08 18:28:34 crc kubenswrapper[4988]: I1008 18:28:34.193626 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"d04259e5-9595-447b-b804-feba577498d8","Type":"ContainerStarted","Data":"590b8a61b67584c2acbd2bea77795253fad7c7d2a79eb0fb7fde120c857233af"} Oct 08 18:28:34 crc kubenswrapper[4988]: I1008 18:28:34.236607 4988 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-gp8f4" podUID="c285e09d-c2f5-4642-9e6b-59b0ed3ed651" containerName="ovn-controller" probeResult="failure" output=< Oct 08 18:28:34 crc kubenswrapper[4988]: ERROR - ovn-controller connection status is 'not connected', expecting 'connected' status Oct 08 18:28:34 crc kubenswrapper[4988]: > Oct 08 18:28:34 crc kubenswrapper[4988]: I1008 18:28:34.285767 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-kc7fw" Oct 08 18:28:34 crc kubenswrapper[4988]: I1008 18:28:34.287199 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c1f2511c-e7bc-4433-910b-eb3b94fcf09b-combined-ca-bundle\") pod \"glance-db-sync-5v5z9\" (UID: \"c1f2511c-e7bc-4433-910b-eb3b94fcf09b\") " pod="openstack/glance-db-sync-5v5z9" Oct 08 18:28:34 crc kubenswrapper[4988]: I1008 18:28:34.287276 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c1f2511c-e7bc-4433-910b-eb3b94fcf09b-config-data\") pod \"glance-db-sync-5v5z9\" (UID: \"c1f2511c-e7bc-4433-910b-eb3b94fcf09b\") " pod="openstack/glance-db-sync-5v5z9" Oct 08 18:28:34 crc kubenswrapper[4988]: I1008 18:28:34.288376 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/c1f2511c-e7bc-4433-910b-eb3b94fcf09b-db-sync-config-data\") pod \"glance-db-sync-5v5z9\" (UID: \"c1f2511c-e7bc-4433-910b-eb3b94fcf09b\") " pod="openstack/glance-db-sync-5v5z9" Oct 08 18:28:34 crc kubenswrapper[4988]: I1008 18:28:34.288714 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-654gh\" (UniqueName: \"kubernetes.io/projected/c1f2511c-e7bc-4433-910b-eb3b94fcf09b-kube-api-access-654gh\") pod \"glance-db-sync-5v5z9\" (UID: \"c1f2511c-e7bc-4433-910b-eb3b94fcf09b\") " pod="openstack/glance-db-sync-5v5z9" Oct 08 18:28:34 crc kubenswrapper[4988]: I1008 18:28:34.303992 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c1f2511c-e7bc-4433-910b-eb3b94fcf09b-config-data\") pod \"glance-db-sync-5v5z9\" (UID: \"c1f2511c-e7bc-4433-910b-eb3b94fcf09b\") " pod="openstack/glance-db-sync-5v5z9" Oct 08 18:28:34 crc kubenswrapper[4988]: I1008 18:28:34.307697 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-654gh\" (UniqueName: \"kubernetes.io/projected/c1f2511c-e7bc-4433-910b-eb3b94fcf09b-kube-api-access-654gh\") pod \"glance-db-sync-5v5z9\" (UID: \"c1f2511c-e7bc-4433-910b-eb3b94fcf09b\") " pod="openstack/glance-db-sync-5v5z9" Oct 08 18:28:34 crc kubenswrapper[4988]: I1008 18:28:34.308121 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c1f2511c-e7bc-4433-910b-eb3b94fcf09b-combined-ca-bundle\") pod \"glance-db-sync-5v5z9\" (UID: \"c1f2511c-e7bc-4433-910b-eb3b94fcf09b\") " pod="openstack/glance-db-sync-5v5z9" Oct 08 18:28:34 crc kubenswrapper[4988]: I1008 18:28:34.316599 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/c1f2511c-e7bc-4433-910b-eb3b94fcf09b-db-sync-config-data\") pod \"glance-db-sync-5v5z9\" (UID: \"c1f2511c-e7bc-4433-910b-eb3b94fcf09b\") " pod="openstack/glance-db-sync-5v5z9" Oct 08 18:28:34 crc kubenswrapper[4988]: I1008 18:28:34.412979 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-5v5z9" Oct 08 18:28:34 crc kubenswrapper[4988]: I1008 18:28:34.505653 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-gp8f4-config-ql98h"] Oct 08 18:28:34 crc kubenswrapper[4988]: I1008 18:28:34.522963 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-gp8f4-config-ql98h" Oct 08 18:28:34 crc kubenswrapper[4988]: I1008 18:28:34.526682 4988 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-extra-scripts" Oct 08 18:28:34 crc kubenswrapper[4988]: I1008 18:28:34.556194 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-gp8f4-config-ql98h"] Oct 08 18:28:34 crc kubenswrapper[4988]: I1008 18:28:34.594460 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/39cc2e49-8d11-4254-a20e-ab36ffc23668-var-run-ovn\") pod \"ovn-controller-gp8f4-config-ql98h\" (UID: \"39cc2e49-8d11-4254-a20e-ab36ffc23668\") " pod="openstack/ovn-controller-gp8f4-config-ql98h" Oct 08 18:28:34 crc kubenswrapper[4988]: I1008 18:28:34.594512 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q5gpn\" (UniqueName: \"kubernetes.io/projected/39cc2e49-8d11-4254-a20e-ab36ffc23668-kube-api-access-q5gpn\") pod \"ovn-controller-gp8f4-config-ql98h\" (UID: \"39cc2e49-8d11-4254-a20e-ab36ffc23668\") " pod="openstack/ovn-controller-gp8f4-config-ql98h" Oct 08 18:28:34 crc kubenswrapper[4988]: I1008 18:28:34.594535 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/39cc2e49-8d11-4254-a20e-ab36ffc23668-additional-scripts\") pod \"ovn-controller-gp8f4-config-ql98h\" (UID: \"39cc2e49-8d11-4254-a20e-ab36ffc23668\") " pod="openstack/ovn-controller-gp8f4-config-ql98h" Oct 08 18:28:34 crc kubenswrapper[4988]: I1008 18:28:34.594574 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/39cc2e49-8d11-4254-a20e-ab36ffc23668-var-log-ovn\") pod \"ovn-controller-gp8f4-config-ql98h\" (UID: \"39cc2e49-8d11-4254-a20e-ab36ffc23668\") " pod="openstack/ovn-controller-gp8f4-config-ql98h" Oct 08 18:28:34 crc kubenswrapper[4988]: I1008 18:28:34.594603 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/39cc2e49-8d11-4254-a20e-ab36ffc23668-var-run\") pod \"ovn-controller-gp8f4-config-ql98h\" (UID: \"39cc2e49-8d11-4254-a20e-ab36ffc23668\") " pod="openstack/ovn-controller-gp8f4-config-ql98h" Oct 08 18:28:34 crc kubenswrapper[4988]: I1008 18:28:34.594657 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/39cc2e49-8d11-4254-a20e-ab36ffc23668-scripts\") pod \"ovn-controller-gp8f4-config-ql98h\" (UID: \"39cc2e49-8d11-4254-a20e-ab36ffc23668\") " pod="openstack/ovn-controller-gp8f4-config-ql98h" Oct 08 18:28:34 crc kubenswrapper[4988]: I1008 18:28:34.696967 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/39cc2e49-8d11-4254-a20e-ab36ffc23668-var-log-ovn\") pod \"ovn-controller-gp8f4-config-ql98h\" (UID: \"39cc2e49-8d11-4254-a20e-ab36ffc23668\") " pod="openstack/ovn-controller-gp8f4-config-ql98h" Oct 08 18:28:34 crc kubenswrapper[4988]: I1008 18:28:34.697040 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/39cc2e49-8d11-4254-a20e-ab36ffc23668-var-run\") pod \"ovn-controller-gp8f4-config-ql98h\" (UID: \"39cc2e49-8d11-4254-a20e-ab36ffc23668\") " pod="openstack/ovn-controller-gp8f4-config-ql98h" Oct 08 18:28:34 crc kubenswrapper[4988]: I1008 18:28:34.697098 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/39cc2e49-8d11-4254-a20e-ab36ffc23668-scripts\") pod \"ovn-controller-gp8f4-config-ql98h\" (UID: \"39cc2e49-8d11-4254-a20e-ab36ffc23668\") " pod="openstack/ovn-controller-gp8f4-config-ql98h" Oct 08 18:28:34 crc kubenswrapper[4988]: I1008 18:28:34.697142 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/39cc2e49-8d11-4254-a20e-ab36ffc23668-var-run-ovn\") pod \"ovn-controller-gp8f4-config-ql98h\" (UID: \"39cc2e49-8d11-4254-a20e-ab36ffc23668\") " pod="openstack/ovn-controller-gp8f4-config-ql98h" Oct 08 18:28:34 crc kubenswrapper[4988]: I1008 18:28:34.697174 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q5gpn\" (UniqueName: \"kubernetes.io/projected/39cc2e49-8d11-4254-a20e-ab36ffc23668-kube-api-access-q5gpn\") pod \"ovn-controller-gp8f4-config-ql98h\" (UID: \"39cc2e49-8d11-4254-a20e-ab36ffc23668\") " pod="openstack/ovn-controller-gp8f4-config-ql98h" Oct 08 18:28:34 crc kubenswrapper[4988]: I1008 18:28:34.697200 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/39cc2e49-8d11-4254-a20e-ab36ffc23668-additional-scripts\") pod \"ovn-controller-gp8f4-config-ql98h\" (UID: \"39cc2e49-8d11-4254-a20e-ab36ffc23668\") " pod="openstack/ovn-controller-gp8f4-config-ql98h" Oct 08 18:28:34 crc kubenswrapper[4988]: I1008 18:28:34.697421 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/39cc2e49-8d11-4254-a20e-ab36ffc23668-var-log-ovn\") pod \"ovn-controller-gp8f4-config-ql98h\" (UID: \"39cc2e49-8d11-4254-a20e-ab36ffc23668\") " pod="openstack/ovn-controller-gp8f4-config-ql98h" Oct 08 18:28:34 crc kubenswrapper[4988]: I1008 18:28:34.697457 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/39cc2e49-8d11-4254-a20e-ab36ffc23668-var-run-ovn\") pod \"ovn-controller-gp8f4-config-ql98h\" (UID: \"39cc2e49-8d11-4254-a20e-ab36ffc23668\") " pod="openstack/ovn-controller-gp8f4-config-ql98h" Oct 08 18:28:34 crc kubenswrapper[4988]: I1008 18:28:34.697468 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/39cc2e49-8d11-4254-a20e-ab36ffc23668-var-run\") pod \"ovn-controller-gp8f4-config-ql98h\" (UID: \"39cc2e49-8d11-4254-a20e-ab36ffc23668\") " pod="openstack/ovn-controller-gp8f4-config-ql98h" Oct 08 18:28:34 crc kubenswrapper[4988]: I1008 18:28:34.697907 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/39cc2e49-8d11-4254-a20e-ab36ffc23668-additional-scripts\") pod \"ovn-controller-gp8f4-config-ql98h\" (UID: \"39cc2e49-8d11-4254-a20e-ab36ffc23668\") " pod="openstack/ovn-controller-gp8f4-config-ql98h" Oct 08 18:28:34 crc kubenswrapper[4988]: I1008 18:28:34.701301 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/39cc2e49-8d11-4254-a20e-ab36ffc23668-scripts\") pod \"ovn-controller-gp8f4-config-ql98h\" (UID: \"39cc2e49-8d11-4254-a20e-ab36ffc23668\") " pod="openstack/ovn-controller-gp8f4-config-ql98h" Oct 08 18:28:34 crc kubenswrapper[4988]: I1008 18:28:34.721577 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q5gpn\" (UniqueName: \"kubernetes.io/projected/39cc2e49-8d11-4254-a20e-ab36ffc23668-kube-api-access-q5gpn\") pod \"ovn-controller-gp8f4-config-ql98h\" (UID: \"39cc2e49-8d11-4254-a20e-ab36ffc23668\") " pod="openstack/ovn-controller-gp8f4-config-ql98h" Oct 08 18:28:34 crc kubenswrapper[4988]: I1008 18:28:34.853709 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-gp8f4-config-ql98h" Oct 08 18:28:35 crc kubenswrapper[4988]: I1008 18:28:35.056441 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-5v5z9"] Oct 08 18:28:35 crc kubenswrapper[4988]: I1008 18:28:35.206455 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"d04259e5-9595-447b-b804-feba577498d8","Type":"ContainerStarted","Data":"31e32c3833aea9f83bda1aca6904be1c190d6d671f11ad9f856ce5abce9053fb"} Oct 08 18:28:35 crc kubenswrapper[4988]: I1008 18:28:35.206503 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"d04259e5-9595-447b-b804-feba577498d8","Type":"ContainerStarted","Data":"4f7780b6a0c60b8247000b0296848f37576f4cd48b317bfb6bd9807dc977f068"} Oct 08 18:28:35 crc kubenswrapper[4988]: I1008 18:28:35.207331 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-5v5z9" event={"ID":"c1f2511c-e7bc-4433-910b-eb3b94fcf09b","Type":"ContainerStarted","Data":"248d15ff74547a86da7c2dd855b4a5f67def74c5c30b8d05b574681289ed18f8"} Oct 08 18:28:35 crc kubenswrapper[4988]: W1008 18:28:35.302199 4988 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod39cc2e49_8d11_4254_a20e_ab36ffc23668.slice/crio-5a4cf695348d17a3d13bd37480eda43e323720b9a4eebb00758ad0bc82456115 WatchSource:0}: Error finding container 5a4cf695348d17a3d13bd37480eda43e323720b9a4eebb00758ad0bc82456115: Status 404 returned error can't find the container with id 5a4cf695348d17a3d13bd37480eda43e323720b9a4eebb00758ad0bc82456115 Oct 08 18:28:35 crc kubenswrapper[4988]: I1008 18:28:35.304078 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-gp8f4-config-ql98h"] Oct 08 18:28:36 crc kubenswrapper[4988]: I1008 18:28:36.227568 4988 generic.go:334] "Generic (PLEG): container finished" podID="39cc2e49-8d11-4254-a20e-ab36ffc23668" containerID="0f84577cfc3de7bfdb97674747df46d7d0dd3b852e078530650e529f2fc757bb" exitCode=0 Oct 08 18:28:36 crc kubenswrapper[4988]: I1008 18:28:36.227875 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-gp8f4-config-ql98h" event={"ID":"39cc2e49-8d11-4254-a20e-ab36ffc23668","Type":"ContainerDied","Data":"0f84577cfc3de7bfdb97674747df46d7d0dd3b852e078530650e529f2fc757bb"} Oct 08 18:28:36 crc kubenswrapper[4988]: I1008 18:28:36.227936 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-gp8f4-config-ql98h" event={"ID":"39cc2e49-8d11-4254-a20e-ab36ffc23668","Type":"ContainerStarted","Data":"5a4cf695348d17a3d13bd37480eda43e323720b9a4eebb00758ad0bc82456115"} Oct 08 18:28:37 crc kubenswrapper[4988]: I1008 18:28:37.246440 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"d04259e5-9595-447b-b804-feba577498d8","Type":"ContainerStarted","Data":"0f92f3d1af5000c50ceb5a04475cb2043188883ce2d85323918a574561017295"} Oct 08 18:28:37 crc kubenswrapper[4988]: I1008 18:28:37.246766 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"d04259e5-9595-447b-b804-feba577498d8","Type":"ContainerStarted","Data":"6c62379d0b1b37e153eff91e53dd273d7d1e85bb5f09dd3abdfc7a5b862d3459"} Oct 08 18:28:37 crc kubenswrapper[4988]: I1008 18:28:37.246776 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"d04259e5-9595-447b-b804-feba577498d8","Type":"ContainerStarted","Data":"94be3d977f16a3f88c65d1405dd7fd0737c52d3e6c73b44f0651409fc2428746"} Oct 08 18:28:37 crc kubenswrapper[4988]: I1008 18:28:37.497875 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-gp8f4-config-ql98h" Oct 08 18:28:37 crc kubenswrapper[4988]: I1008 18:28:37.553371 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q5gpn\" (UniqueName: \"kubernetes.io/projected/39cc2e49-8d11-4254-a20e-ab36ffc23668-kube-api-access-q5gpn\") pod \"39cc2e49-8d11-4254-a20e-ab36ffc23668\" (UID: \"39cc2e49-8d11-4254-a20e-ab36ffc23668\") " Oct 08 18:28:37 crc kubenswrapper[4988]: I1008 18:28:37.553455 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/39cc2e49-8d11-4254-a20e-ab36ffc23668-var-run-ovn\") pod \"39cc2e49-8d11-4254-a20e-ab36ffc23668\" (UID: \"39cc2e49-8d11-4254-a20e-ab36ffc23668\") " Oct 08 18:28:37 crc kubenswrapper[4988]: I1008 18:28:37.553487 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/39cc2e49-8d11-4254-a20e-ab36ffc23668-scripts\") pod \"39cc2e49-8d11-4254-a20e-ab36ffc23668\" (UID: \"39cc2e49-8d11-4254-a20e-ab36ffc23668\") " Oct 08 18:28:37 crc kubenswrapper[4988]: I1008 18:28:37.553509 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/39cc2e49-8d11-4254-a20e-ab36ffc23668-var-log-ovn\") pod \"39cc2e49-8d11-4254-a20e-ab36ffc23668\" (UID: \"39cc2e49-8d11-4254-a20e-ab36ffc23668\") " Oct 08 18:28:37 crc kubenswrapper[4988]: I1008 18:28:37.553533 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/39cc2e49-8d11-4254-a20e-ab36ffc23668-additional-scripts\") pod \"39cc2e49-8d11-4254-a20e-ab36ffc23668\" (UID: \"39cc2e49-8d11-4254-a20e-ab36ffc23668\") " Oct 08 18:28:37 crc kubenswrapper[4988]: I1008 18:28:37.553676 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/39cc2e49-8d11-4254-a20e-ab36ffc23668-var-run\") pod \"39cc2e49-8d11-4254-a20e-ab36ffc23668\" (UID: \"39cc2e49-8d11-4254-a20e-ab36ffc23668\") " Oct 08 18:28:37 crc kubenswrapper[4988]: I1008 18:28:37.553965 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/39cc2e49-8d11-4254-a20e-ab36ffc23668-var-run" (OuterVolumeSpecName: "var-run") pod "39cc2e49-8d11-4254-a20e-ab36ffc23668" (UID: "39cc2e49-8d11-4254-a20e-ab36ffc23668"). InnerVolumeSpecName "var-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 08 18:28:37 crc kubenswrapper[4988]: I1008 18:28:37.553992 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/39cc2e49-8d11-4254-a20e-ab36ffc23668-var-log-ovn" (OuterVolumeSpecName: "var-log-ovn") pod "39cc2e49-8d11-4254-a20e-ab36ffc23668" (UID: "39cc2e49-8d11-4254-a20e-ab36ffc23668"). InnerVolumeSpecName "var-log-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 08 18:28:37 crc kubenswrapper[4988]: I1008 18:28:37.554470 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/39cc2e49-8d11-4254-a20e-ab36ffc23668-additional-scripts" (OuterVolumeSpecName: "additional-scripts") pod "39cc2e49-8d11-4254-a20e-ab36ffc23668" (UID: "39cc2e49-8d11-4254-a20e-ab36ffc23668"). InnerVolumeSpecName "additional-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 18:28:37 crc kubenswrapper[4988]: I1008 18:28:37.554494 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/39cc2e49-8d11-4254-a20e-ab36ffc23668-var-run-ovn" (OuterVolumeSpecName: "var-run-ovn") pod "39cc2e49-8d11-4254-a20e-ab36ffc23668" (UID: "39cc2e49-8d11-4254-a20e-ab36ffc23668"). InnerVolumeSpecName "var-run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 08 18:28:37 crc kubenswrapper[4988]: I1008 18:28:37.554688 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/39cc2e49-8d11-4254-a20e-ab36ffc23668-scripts" (OuterVolumeSpecName: "scripts") pod "39cc2e49-8d11-4254-a20e-ab36ffc23668" (UID: "39cc2e49-8d11-4254-a20e-ab36ffc23668"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 18:28:37 crc kubenswrapper[4988]: I1008 18:28:37.559437 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/39cc2e49-8d11-4254-a20e-ab36ffc23668-kube-api-access-q5gpn" (OuterVolumeSpecName: "kube-api-access-q5gpn") pod "39cc2e49-8d11-4254-a20e-ab36ffc23668" (UID: "39cc2e49-8d11-4254-a20e-ab36ffc23668"). InnerVolumeSpecName "kube-api-access-q5gpn". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:28:37 crc kubenswrapper[4988]: I1008 18:28:37.655661 4988 reconciler_common.go:293] "Volume detached for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/39cc2e49-8d11-4254-a20e-ab36ffc23668-var-run\") on node \"crc\" DevicePath \"\"" Oct 08 18:28:37 crc kubenswrapper[4988]: I1008 18:28:37.655700 4988 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q5gpn\" (UniqueName: \"kubernetes.io/projected/39cc2e49-8d11-4254-a20e-ab36ffc23668-kube-api-access-q5gpn\") on node \"crc\" DevicePath \"\"" Oct 08 18:28:37 crc kubenswrapper[4988]: I1008 18:28:37.655712 4988 reconciler_common.go:293] "Volume detached for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/39cc2e49-8d11-4254-a20e-ab36ffc23668-var-run-ovn\") on node \"crc\" DevicePath \"\"" Oct 08 18:28:37 crc kubenswrapper[4988]: I1008 18:28:37.655724 4988 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/39cc2e49-8d11-4254-a20e-ab36ffc23668-scripts\") on node \"crc\" DevicePath \"\"" Oct 08 18:28:37 crc kubenswrapper[4988]: I1008 18:28:37.655737 4988 reconciler_common.go:293] "Volume detached for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/39cc2e49-8d11-4254-a20e-ab36ffc23668-var-log-ovn\") on node \"crc\" DevicePath \"\"" Oct 08 18:28:37 crc kubenswrapper[4988]: I1008 18:28:37.655750 4988 reconciler_common.go:293] "Volume detached for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/39cc2e49-8d11-4254-a20e-ab36ffc23668-additional-scripts\") on node \"crc\" DevicePath \"\"" Oct 08 18:28:38 crc kubenswrapper[4988]: I1008 18:28:38.257407 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"d04259e5-9595-447b-b804-feba577498d8","Type":"ContainerStarted","Data":"b723050435b720195f0b9b9fbdfc7bf95b8dd19f190a2d21120788a8cde0603f"} Oct 08 18:28:38 crc kubenswrapper[4988]: I1008 18:28:38.257451 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"d04259e5-9595-447b-b804-feba577498d8","Type":"ContainerStarted","Data":"3e8f7969a53f0a0b492bfa6e755ebabd4ef007e6b42af4f280d07a7401c0699d"} Oct 08 18:28:38 crc kubenswrapper[4988]: I1008 18:28:38.257460 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"d04259e5-9595-447b-b804-feba577498d8","Type":"ContainerStarted","Data":"f6d8e476ef92ad2a5ae0468a14ebdbcb37b6c35a714f5c83ec9ff493a5dd15a3"} Oct 08 18:28:38 crc kubenswrapper[4988]: I1008 18:28:38.257468 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"d04259e5-9595-447b-b804-feba577498d8","Type":"ContainerStarted","Data":"9935bea63badb3e4cb30261f8985a1224e374fbe56ec5af0a913e1c5fe57db04"} Oct 08 18:28:38 crc kubenswrapper[4988]: I1008 18:28:38.259452 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-gp8f4-config-ql98h" event={"ID":"39cc2e49-8d11-4254-a20e-ab36ffc23668","Type":"ContainerDied","Data":"5a4cf695348d17a3d13bd37480eda43e323720b9a4eebb00758ad0bc82456115"} Oct 08 18:28:38 crc kubenswrapper[4988]: I1008 18:28:38.259476 4988 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5a4cf695348d17a3d13bd37480eda43e323720b9a4eebb00758ad0bc82456115" Oct 08 18:28:38 crc kubenswrapper[4988]: I1008 18:28:38.259487 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-gp8f4-config-ql98h" Oct 08 18:28:38 crc kubenswrapper[4988]: I1008 18:28:38.297947 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-storage-0" podStartSLOduration=20.154797668 podStartE2EDuration="28.297920836s" podCreationTimestamp="2025-10-08 18:28:10 +0000 UTC" firstStartedPulling="2025-10-08 18:28:28.344540221 +0000 UTC m=+1053.794382991" lastFinishedPulling="2025-10-08 18:28:36.487663389 +0000 UTC m=+1061.937506159" observedRunningTime="2025-10-08 18:28:38.289864792 +0000 UTC m=+1063.739707552" watchObservedRunningTime="2025-10-08 18:28:38.297920836 +0000 UTC m=+1063.747763606" Oct 08 18:28:38 crc kubenswrapper[4988]: I1008 18:28:38.568632 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-59f45f6cf7-9snqd"] Oct 08 18:28:38 crc kubenswrapper[4988]: E1008 18:28:38.569311 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="39cc2e49-8d11-4254-a20e-ab36ffc23668" containerName="ovn-config" Oct 08 18:28:38 crc kubenswrapper[4988]: I1008 18:28:38.569427 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="39cc2e49-8d11-4254-a20e-ab36ffc23668" containerName="ovn-config" Oct 08 18:28:38 crc kubenswrapper[4988]: I1008 18:28:38.577727 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="39cc2e49-8d11-4254-a20e-ab36ffc23668" containerName="ovn-config" Oct 08 18:28:38 crc kubenswrapper[4988]: I1008 18:28:38.578773 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-59f45f6cf7-9snqd" Oct 08 18:28:38 crc kubenswrapper[4988]: I1008 18:28:38.593802 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-59f45f6cf7-9snqd"] Oct 08 18:28:38 crc kubenswrapper[4988]: I1008 18:28:38.594629 4988 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns-swift-storage-0" Oct 08 18:28:38 crc kubenswrapper[4988]: I1008 18:28:38.669913 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/095497e5-73fc-42a2-9db3-cdec5010b477-config\") pod \"dnsmasq-dns-59f45f6cf7-9snqd\" (UID: \"095497e5-73fc-42a2-9db3-cdec5010b477\") " pod="openstack/dnsmasq-dns-59f45f6cf7-9snqd" Oct 08 18:28:38 crc kubenswrapper[4988]: I1008 18:28:38.669996 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/095497e5-73fc-42a2-9db3-cdec5010b477-dns-svc\") pod \"dnsmasq-dns-59f45f6cf7-9snqd\" (UID: \"095497e5-73fc-42a2-9db3-cdec5010b477\") " pod="openstack/dnsmasq-dns-59f45f6cf7-9snqd" Oct 08 18:28:38 crc kubenswrapper[4988]: I1008 18:28:38.670042 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/095497e5-73fc-42a2-9db3-cdec5010b477-ovsdbserver-nb\") pod \"dnsmasq-dns-59f45f6cf7-9snqd\" (UID: \"095497e5-73fc-42a2-9db3-cdec5010b477\") " pod="openstack/dnsmasq-dns-59f45f6cf7-9snqd" Oct 08 18:28:38 crc kubenswrapper[4988]: I1008 18:28:38.670067 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tkmh8\" (UniqueName: \"kubernetes.io/projected/095497e5-73fc-42a2-9db3-cdec5010b477-kube-api-access-tkmh8\") pod \"dnsmasq-dns-59f45f6cf7-9snqd\" (UID: \"095497e5-73fc-42a2-9db3-cdec5010b477\") " pod="openstack/dnsmasq-dns-59f45f6cf7-9snqd" Oct 08 18:28:38 crc kubenswrapper[4988]: I1008 18:28:38.670147 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/095497e5-73fc-42a2-9db3-cdec5010b477-dns-swift-storage-0\") pod \"dnsmasq-dns-59f45f6cf7-9snqd\" (UID: \"095497e5-73fc-42a2-9db3-cdec5010b477\") " pod="openstack/dnsmasq-dns-59f45f6cf7-9snqd" Oct 08 18:28:38 crc kubenswrapper[4988]: I1008 18:28:38.670191 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/095497e5-73fc-42a2-9db3-cdec5010b477-ovsdbserver-sb\") pod \"dnsmasq-dns-59f45f6cf7-9snqd\" (UID: \"095497e5-73fc-42a2-9db3-cdec5010b477\") " pod="openstack/dnsmasq-dns-59f45f6cf7-9snqd" Oct 08 18:28:38 crc kubenswrapper[4988]: I1008 18:28:38.678920 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-gp8f4-config-ql98h"] Oct 08 18:28:38 crc kubenswrapper[4988]: I1008 18:28:38.684881 4988 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-gp8f4-config-ql98h"] Oct 08 18:28:38 crc kubenswrapper[4988]: I1008 18:28:38.771208 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/095497e5-73fc-42a2-9db3-cdec5010b477-ovsdbserver-sb\") pod \"dnsmasq-dns-59f45f6cf7-9snqd\" (UID: \"095497e5-73fc-42a2-9db3-cdec5010b477\") " pod="openstack/dnsmasq-dns-59f45f6cf7-9snqd" Oct 08 18:28:38 crc kubenswrapper[4988]: I1008 18:28:38.771634 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/095497e5-73fc-42a2-9db3-cdec5010b477-config\") pod \"dnsmasq-dns-59f45f6cf7-9snqd\" (UID: \"095497e5-73fc-42a2-9db3-cdec5010b477\") " pod="openstack/dnsmasq-dns-59f45f6cf7-9snqd" Oct 08 18:28:38 crc kubenswrapper[4988]: I1008 18:28:38.771695 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/095497e5-73fc-42a2-9db3-cdec5010b477-dns-svc\") pod \"dnsmasq-dns-59f45f6cf7-9snqd\" (UID: \"095497e5-73fc-42a2-9db3-cdec5010b477\") " pod="openstack/dnsmasq-dns-59f45f6cf7-9snqd" Oct 08 18:28:38 crc kubenswrapper[4988]: I1008 18:28:38.771749 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/095497e5-73fc-42a2-9db3-cdec5010b477-ovsdbserver-nb\") pod \"dnsmasq-dns-59f45f6cf7-9snqd\" (UID: \"095497e5-73fc-42a2-9db3-cdec5010b477\") " pod="openstack/dnsmasq-dns-59f45f6cf7-9snqd" Oct 08 18:28:38 crc kubenswrapper[4988]: I1008 18:28:38.771784 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tkmh8\" (UniqueName: \"kubernetes.io/projected/095497e5-73fc-42a2-9db3-cdec5010b477-kube-api-access-tkmh8\") pod \"dnsmasq-dns-59f45f6cf7-9snqd\" (UID: \"095497e5-73fc-42a2-9db3-cdec5010b477\") " pod="openstack/dnsmasq-dns-59f45f6cf7-9snqd" Oct 08 18:28:38 crc kubenswrapper[4988]: I1008 18:28:38.771825 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/095497e5-73fc-42a2-9db3-cdec5010b477-dns-swift-storage-0\") pod \"dnsmasq-dns-59f45f6cf7-9snqd\" (UID: \"095497e5-73fc-42a2-9db3-cdec5010b477\") " pod="openstack/dnsmasq-dns-59f45f6cf7-9snqd" Oct 08 18:28:38 crc kubenswrapper[4988]: I1008 18:28:38.772319 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/095497e5-73fc-42a2-9db3-cdec5010b477-ovsdbserver-sb\") pod \"dnsmasq-dns-59f45f6cf7-9snqd\" (UID: \"095497e5-73fc-42a2-9db3-cdec5010b477\") " pod="openstack/dnsmasq-dns-59f45f6cf7-9snqd" Oct 08 18:28:38 crc kubenswrapper[4988]: I1008 18:28:38.772356 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/095497e5-73fc-42a2-9db3-cdec5010b477-config\") pod \"dnsmasq-dns-59f45f6cf7-9snqd\" (UID: \"095497e5-73fc-42a2-9db3-cdec5010b477\") " pod="openstack/dnsmasq-dns-59f45f6cf7-9snqd" Oct 08 18:28:38 crc kubenswrapper[4988]: I1008 18:28:38.772547 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/095497e5-73fc-42a2-9db3-cdec5010b477-dns-svc\") pod \"dnsmasq-dns-59f45f6cf7-9snqd\" (UID: \"095497e5-73fc-42a2-9db3-cdec5010b477\") " pod="openstack/dnsmasq-dns-59f45f6cf7-9snqd" Oct 08 18:28:38 crc kubenswrapper[4988]: I1008 18:28:38.772667 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/095497e5-73fc-42a2-9db3-cdec5010b477-dns-swift-storage-0\") pod \"dnsmasq-dns-59f45f6cf7-9snqd\" (UID: \"095497e5-73fc-42a2-9db3-cdec5010b477\") " pod="openstack/dnsmasq-dns-59f45f6cf7-9snqd" Oct 08 18:28:38 crc kubenswrapper[4988]: I1008 18:28:38.772999 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/095497e5-73fc-42a2-9db3-cdec5010b477-ovsdbserver-nb\") pod \"dnsmasq-dns-59f45f6cf7-9snqd\" (UID: \"095497e5-73fc-42a2-9db3-cdec5010b477\") " pod="openstack/dnsmasq-dns-59f45f6cf7-9snqd" Oct 08 18:28:38 crc kubenswrapper[4988]: I1008 18:28:38.796887 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tkmh8\" (UniqueName: \"kubernetes.io/projected/095497e5-73fc-42a2-9db3-cdec5010b477-kube-api-access-tkmh8\") pod \"dnsmasq-dns-59f45f6cf7-9snqd\" (UID: \"095497e5-73fc-42a2-9db3-cdec5010b477\") " pod="openstack/dnsmasq-dns-59f45f6cf7-9snqd" Oct 08 18:28:38 crc kubenswrapper[4988]: I1008 18:28:38.895782 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-59f45f6cf7-9snqd" Oct 08 18:28:39 crc kubenswrapper[4988]: I1008 18:28:39.229074 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-gp8f4" Oct 08 18:28:39 crc kubenswrapper[4988]: I1008 18:28:39.252035 4988 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="39cc2e49-8d11-4254-a20e-ab36ffc23668" path="/var/lib/kubelet/pods/39cc2e49-8d11-4254-a20e-ab36ffc23668/volumes" Oct 08 18:28:39 crc kubenswrapper[4988]: I1008 18:28:39.389935 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-59f45f6cf7-9snqd"] Oct 08 18:28:39 crc kubenswrapper[4988]: W1008 18:28:39.401568 4988 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod095497e5_73fc_42a2_9db3_cdec5010b477.slice/crio-25762dd24b39ed92b4a68132faab7bd2ba589caada1a13db4ca52b0a2c29037b WatchSource:0}: Error finding container 25762dd24b39ed92b4a68132faab7bd2ba589caada1a13db4ca52b0a2c29037b: Status 404 returned error can't find the container with id 25762dd24b39ed92b4a68132faab7bd2ba589caada1a13db4ca52b0a2c29037b Oct 08 18:28:40 crc kubenswrapper[4988]: I1008 18:28:40.299608 4988 generic.go:334] "Generic (PLEG): container finished" podID="095497e5-73fc-42a2-9db3-cdec5010b477" containerID="84b53df54fd395ba6e0f9bca2e0ab63e769d5ef4812879f7d5885766e2fd8a3c" exitCode=0 Oct 08 18:28:40 crc kubenswrapper[4988]: I1008 18:28:40.299737 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-59f45f6cf7-9snqd" event={"ID":"095497e5-73fc-42a2-9db3-cdec5010b477","Type":"ContainerDied","Data":"84b53df54fd395ba6e0f9bca2e0ab63e769d5ef4812879f7d5885766e2fd8a3c"} Oct 08 18:28:40 crc kubenswrapper[4988]: I1008 18:28:40.299999 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-59f45f6cf7-9snqd" event={"ID":"095497e5-73fc-42a2-9db3-cdec5010b477","Type":"ContainerStarted","Data":"25762dd24b39ed92b4a68132faab7bd2ba589caada1a13db4ca52b0a2c29037b"} Oct 08 18:28:41 crc kubenswrapper[4988]: I1008 18:28:41.324423 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-59f45f6cf7-9snqd" event={"ID":"095497e5-73fc-42a2-9db3-cdec5010b477","Type":"ContainerStarted","Data":"d5fab394a095c11a56d6b6a9c80c90f7c9685834c5518456a3c18aac02301933"} Oct 08 18:28:41 crc kubenswrapper[4988]: I1008 18:28:41.324801 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-59f45f6cf7-9snqd" Oct 08 18:28:41 crc kubenswrapper[4988]: I1008 18:28:41.344147 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-59f45f6cf7-9snqd" podStartSLOduration=3.344130843 podStartE2EDuration="3.344130843s" podCreationTimestamp="2025-10-08 18:28:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 18:28:41.343020318 +0000 UTC m=+1066.792863088" watchObservedRunningTime="2025-10-08 18:28:41.344130843 +0000 UTC m=+1066.793973623" Oct 08 18:28:44 crc kubenswrapper[4988]: I1008 18:28:44.768648 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-0" Oct 08 18:28:45 crc kubenswrapper[4988]: I1008 18:28:45.022554 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-create-bt5zq"] Oct 08 18:28:45 crc kubenswrapper[4988]: I1008 18:28:45.023580 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-bt5zq" Oct 08 18:28:45 crc kubenswrapper[4988]: I1008 18:28:45.032950 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-bt5zq"] Oct 08 18:28:45 crc kubenswrapper[4988]: I1008 18:28:45.217292 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8vm47\" (UniqueName: \"kubernetes.io/projected/ac19d73a-13bf-4714-aa01-36b72ee937d9-kube-api-access-8vm47\") pod \"barbican-db-create-bt5zq\" (UID: \"ac19d73a-13bf-4714-aa01-36b72ee937d9\") " pod="openstack/barbican-db-create-bt5zq" Oct 08 18:28:45 crc kubenswrapper[4988]: I1008 18:28:45.228468 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-create-wt7bm"] Oct 08 18:28:45 crc kubenswrapper[4988]: I1008 18:28:45.229749 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-wt7bm" Oct 08 18:28:45 crc kubenswrapper[4988]: I1008 18:28:45.251222 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-wt7bm"] Oct 08 18:28:45 crc kubenswrapper[4988]: I1008 18:28:45.318784 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8vm47\" (UniqueName: \"kubernetes.io/projected/ac19d73a-13bf-4714-aa01-36b72ee937d9-kube-api-access-8vm47\") pod \"barbican-db-create-bt5zq\" (UID: \"ac19d73a-13bf-4714-aa01-36b72ee937d9\") " pod="openstack/barbican-db-create-bt5zq" Oct 08 18:28:45 crc kubenswrapper[4988]: I1008 18:28:45.321577 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-cell1-server-0" Oct 08 18:28:45 crc kubenswrapper[4988]: I1008 18:28:45.351212 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-create-8dclb"] Oct 08 18:28:45 crc kubenswrapper[4988]: I1008 18:28:45.353914 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-8dclb" Oct 08 18:28:45 crc kubenswrapper[4988]: I1008 18:28:45.381309 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8vm47\" (UniqueName: \"kubernetes.io/projected/ac19d73a-13bf-4714-aa01-36b72ee937d9-kube-api-access-8vm47\") pod \"barbican-db-create-bt5zq\" (UID: \"ac19d73a-13bf-4714-aa01-36b72ee937d9\") " pod="openstack/barbican-db-create-bt5zq" Oct 08 18:28:45 crc kubenswrapper[4988]: I1008 18:28:45.421036 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v7dpx\" (UniqueName: \"kubernetes.io/projected/ee22c4f7-5dc2-4f3c-99d7-5c1c578c9afe-kube-api-access-v7dpx\") pod \"cinder-db-create-wt7bm\" (UID: \"ee22c4f7-5dc2-4f3c-99d7-5c1c578c9afe\") " pod="openstack/cinder-db-create-wt7bm" Oct 08 18:28:45 crc kubenswrapper[4988]: I1008 18:28:45.426416 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-8dclb"] Oct 08 18:28:45 crc kubenswrapper[4988]: I1008 18:28:45.444296 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-sync-wzc5c"] Oct 08 18:28:45 crc kubenswrapper[4988]: I1008 18:28:45.446076 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-wzc5c" Oct 08 18:28:45 crc kubenswrapper[4988]: I1008 18:28:45.460054 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Oct 08 18:28:45 crc kubenswrapper[4988]: I1008 18:28:45.460207 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Oct 08 18:28:45 crc kubenswrapper[4988]: I1008 18:28:45.460263 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Oct 08 18:28:45 crc kubenswrapper[4988]: I1008 18:28:45.460565 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-srkzn" Oct 08 18:28:45 crc kubenswrapper[4988]: I1008 18:28:45.463642 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-wzc5c"] Oct 08 18:28:45 crc kubenswrapper[4988]: I1008 18:28:45.525491 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kwbkz\" (UniqueName: \"kubernetes.io/projected/99149e0d-f06b-4136-88da-313ab1705081-kube-api-access-kwbkz\") pod \"neutron-db-create-8dclb\" (UID: \"99149e0d-f06b-4136-88da-313ab1705081\") " pod="openstack/neutron-db-create-8dclb" Oct 08 18:28:45 crc kubenswrapper[4988]: I1008 18:28:45.525790 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v7dpx\" (UniqueName: \"kubernetes.io/projected/ee22c4f7-5dc2-4f3c-99d7-5c1c578c9afe-kube-api-access-v7dpx\") pod \"cinder-db-create-wt7bm\" (UID: \"ee22c4f7-5dc2-4f3c-99d7-5c1c578c9afe\") " pod="openstack/cinder-db-create-wt7bm" Oct 08 18:28:45 crc kubenswrapper[4988]: I1008 18:28:45.525899 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/434ac32e-3136-4d51-b68e-919adae50ffe-combined-ca-bundle\") pod \"keystone-db-sync-wzc5c\" (UID: \"434ac32e-3136-4d51-b68e-919adae50ffe\") " pod="openstack/keystone-db-sync-wzc5c" Oct 08 18:28:45 crc kubenswrapper[4988]: I1008 18:28:45.526016 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f9jd5\" (UniqueName: \"kubernetes.io/projected/434ac32e-3136-4d51-b68e-919adae50ffe-kube-api-access-f9jd5\") pod \"keystone-db-sync-wzc5c\" (UID: \"434ac32e-3136-4d51-b68e-919adae50ffe\") " pod="openstack/keystone-db-sync-wzc5c" Oct 08 18:28:45 crc kubenswrapper[4988]: I1008 18:28:45.526138 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/434ac32e-3136-4d51-b68e-919adae50ffe-config-data\") pod \"keystone-db-sync-wzc5c\" (UID: \"434ac32e-3136-4d51-b68e-919adae50ffe\") " pod="openstack/keystone-db-sync-wzc5c" Oct 08 18:28:45 crc kubenswrapper[4988]: I1008 18:28:45.547097 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v7dpx\" (UniqueName: \"kubernetes.io/projected/ee22c4f7-5dc2-4f3c-99d7-5c1c578c9afe-kube-api-access-v7dpx\") pod \"cinder-db-create-wt7bm\" (UID: \"ee22c4f7-5dc2-4f3c-99d7-5c1c578c9afe\") " pod="openstack/cinder-db-create-wt7bm" Oct 08 18:28:45 crc kubenswrapper[4988]: I1008 18:28:45.550822 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-wt7bm" Oct 08 18:28:45 crc kubenswrapper[4988]: I1008 18:28:45.629636 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f9jd5\" (UniqueName: \"kubernetes.io/projected/434ac32e-3136-4d51-b68e-919adae50ffe-kube-api-access-f9jd5\") pod \"keystone-db-sync-wzc5c\" (UID: \"434ac32e-3136-4d51-b68e-919adae50ffe\") " pod="openstack/keystone-db-sync-wzc5c" Oct 08 18:28:45 crc kubenswrapper[4988]: I1008 18:28:45.629733 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/434ac32e-3136-4d51-b68e-919adae50ffe-config-data\") pod \"keystone-db-sync-wzc5c\" (UID: \"434ac32e-3136-4d51-b68e-919adae50ffe\") " pod="openstack/keystone-db-sync-wzc5c" Oct 08 18:28:45 crc kubenswrapper[4988]: I1008 18:28:45.629822 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kwbkz\" (UniqueName: \"kubernetes.io/projected/99149e0d-f06b-4136-88da-313ab1705081-kube-api-access-kwbkz\") pod \"neutron-db-create-8dclb\" (UID: \"99149e0d-f06b-4136-88da-313ab1705081\") " pod="openstack/neutron-db-create-8dclb" Oct 08 18:28:45 crc kubenswrapper[4988]: I1008 18:28:45.629860 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/434ac32e-3136-4d51-b68e-919adae50ffe-combined-ca-bundle\") pod \"keystone-db-sync-wzc5c\" (UID: \"434ac32e-3136-4d51-b68e-919adae50ffe\") " pod="openstack/keystone-db-sync-wzc5c" Oct 08 18:28:45 crc kubenswrapper[4988]: I1008 18:28:45.636490 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/434ac32e-3136-4d51-b68e-919adae50ffe-config-data\") pod \"keystone-db-sync-wzc5c\" (UID: \"434ac32e-3136-4d51-b68e-919adae50ffe\") " pod="openstack/keystone-db-sync-wzc5c" Oct 08 18:28:45 crc kubenswrapper[4988]: I1008 18:28:45.644074 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-bt5zq" Oct 08 18:28:45 crc kubenswrapper[4988]: I1008 18:28:45.647457 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/434ac32e-3136-4d51-b68e-919adae50ffe-combined-ca-bundle\") pod \"keystone-db-sync-wzc5c\" (UID: \"434ac32e-3136-4d51-b68e-919adae50ffe\") " pod="openstack/keystone-db-sync-wzc5c" Oct 08 18:28:45 crc kubenswrapper[4988]: I1008 18:28:45.653679 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f9jd5\" (UniqueName: \"kubernetes.io/projected/434ac32e-3136-4d51-b68e-919adae50ffe-kube-api-access-f9jd5\") pod \"keystone-db-sync-wzc5c\" (UID: \"434ac32e-3136-4d51-b68e-919adae50ffe\") " pod="openstack/keystone-db-sync-wzc5c" Oct 08 18:28:45 crc kubenswrapper[4988]: I1008 18:28:45.659668 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kwbkz\" (UniqueName: \"kubernetes.io/projected/99149e0d-f06b-4136-88da-313ab1705081-kube-api-access-kwbkz\") pod \"neutron-db-create-8dclb\" (UID: \"99149e0d-f06b-4136-88da-313ab1705081\") " pod="openstack/neutron-db-create-8dclb" Oct 08 18:28:45 crc kubenswrapper[4988]: I1008 18:28:45.787072 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-8dclb" Oct 08 18:28:45 crc kubenswrapper[4988]: I1008 18:28:45.796645 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-wzc5c" Oct 08 18:28:48 crc kubenswrapper[4988]: I1008 18:28:48.897311 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-59f45f6cf7-9snqd" Oct 08 18:28:48 crc kubenswrapper[4988]: I1008 18:28:48.950370 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-57f58c7cff-42svl"] Oct 08 18:28:48 crc kubenswrapper[4988]: I1008 18:28:48.950669 4988 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-57f58c7cff-42svl" podUID="1487d461-4c56-4591-b498-b6ce9f314763" containerName="dnsmasq-dns" containerID="cri-o://0bdd344781c086aea94f3fd316e2eefc84f2f8da29e9372eab7e14dd7a7a0061" gracePeriod=10 Oct 08 18:28:50 crc kubenswrapper[4988]: I1008 18:28:50.793442 4988 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-57f58c7cff-42svl" podUID="1487d461-4c56-4591-b498-b6ce9f314763" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.117:5353: connect: connection refused" Oct 08 18:28:52 crc kubenswrapper[4988]: I1008 18:28:52.459275 4988 generic.go:334] "Generic (PLEG): container finished" podID="1487d461-4c56-4591-b498-b6ce9f314763" containerID="0bdd344781c086aea94f3fd316e2eefc84f2f8da29e9372eab7e14dd7a7a0061" exitCode=0 Oct 08 18:28:52 crc kubenswrapper[4988]: I1008 18:28:52.459321 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57f58c7cff-42svl" event={"ID":"1487d461-4c56-4591-b498-b6ce9f314763","Type":"ContainerDied","Data":"0bdd344781c086aea94f3fd316e2eefc84f2f8da29e9372eab7e14dd7a7a0061"} Oct 08 18:28:55 crc kubenswrapper[4988]: I1008 18:28:55.792851 4988 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-57f58c7cff-42svl" podUID="1487d461-4c56-4591-b498-b6ce9f314763" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.117:5353: connect: connection refused" Oct 08 18:28:55 crc kubenswrapper[4988]: E1008 18:28:55.934708 4988 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-glance-api@sha256:9ebf424d4107275a2e3f21f7a18ef257ff2f97c1298109ac7c802a5a4f4794f2" Oct 08 18:28:55 crc kubenswrapper[4988]: E1008 18:28:55.934880 4988 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:glance-db-sync,Image:quay.io/podified-antelope-centos9/openstack-glance-api@sha256:9ebf424d4107275a2e3f21f7a18ef257ff2f97c1298109ac7c802a5a4f4794f2,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:true,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:db-sync-config-data,ReadOnly:true,MountPath:/etc/glance/glance.conf.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/my.cnf,SubPath:my.cnf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:db-sync-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-654gh,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42415,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:*42415,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod glance-db-sync-5v5z9_openstack(c1f2511c-e7bc-4433-910b-eb3b94fcf09b): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 08 18:28:55 crc kubenswrapper[4988]: E1008 18:28:55.936262 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"glance-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/glance-db-sync-5v5z9" podUID="c1f2511c-e7bc-4433-910b-eb3b94fcf09b" Oct 08 18:28:56 crc kubenswrapper[4988]: I1008 18:28:56.206849 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-bt5zq"] Oct 08 18:28:56 crc kubenswrapper[4988]: I1008 18:28:56.213654 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57f58c7cff-42svl" Oct 08 18:28:56 crc kubenswrapper[4988]: I1008 18:28:56.300365 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1487d461-4c56-4591-b498-b6ce9f314763-config\") pod \"1487d461-4c56-4591-b498-b6ce9f314763\" (UID: \"1487d461-4c56-4591-b498-b6ce9f314763\") " Oct 08 18:28:56 crc kubenswrapper[4988]: I1008 18:28:56.300473 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1487d461-4c56-4591-b498-b6ce9f314763-dns-svc\") pod \"1487d461-4c56-4591-b498-b6ce9f314763\" (UID: \"1487d461-4c56-4591-b498-b6ce9f314763\") " Oct 08 18:28:56 crc kubenswrapper[4988]: I1008 18:28:56.300566 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/1487d461-4c56-4591-b498-b6ce9f314763-ovsdbserver-nb\") pod \"1487d461-4c56-4591-b498-b6ce9f314763\" (UID: \"1487d461-4c56-4591-b498-b6ce9f314763\") " Oct 08 18:28:56 crc kubenswrapper[4988]: I1008 18:28:56.300612 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xbmst\" (UniqueName: \"kubernetes.io/projected/1487d461-4c56-4591-b498-b6ce9f314763-kube-api-access-xbmst\") pod \"1487d461-4c56-4591-b498-b6ce9f314763\" (UID: \"1487d461-4c56-4591-b498-b6ce9f314763\") " Oct 08 18:28:56 crc kubenswrapper[4988]: I1008 18:28:56.300645 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/1487d461-4c56-4591-b498-b6ce9f314763-ovsdbserver-sb\") pod \"1487d461-4c56-4591-b498-b6ce9f314763\" (UID: \"1487d461-4c56-4591-b498-b6ce9f314763\") " Oct 08 18:28:56 crc kubenswrapper[4988]: I1008 18:28:56.308980 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1487d461-4c56-4591-b498-b6ce9f314763-kube-api-access-xbmst" (OuterVolumeSpecName: "kube-api-access-xbmst") pod "1487d461-4c56-4591-b498-b6ce9f314763" (UID: "1487d461-4c56-4591-b498-b6ce9f314763"). InnerVolumeSpecName "kube-api-access-xbmst". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:28:56 crc kubenswrapper[4988]: I1008 18:28:56.333553 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-wzc5c"] Oct 08 18:28:56 crc kubenswrapper[4988]: I1008 18:28:56.349139 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-wt7bm"] Oct 08 18:28:56 crc kubenswrapper[4988]: I1008 18:28:56.349774 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1487d461-4c56-4591-b498-b6ce9f314763-config" (OuterVolumeSpecName: "config") pod "1487d461-4c56-4591-b498-b6ce9f314763" (UID: "1487d461-4c56-4591-b498-b6ce9f314763"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 18:28:56 crc kubenswrapper[4988]: I1008 18:28:56.356878 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-8dclb"] Oct 08 18:28:56 crc kubenswrapper[4988]: W1008 18:28:56.361043 4988 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podee22c4f7_5dc2_4f3c_99d7_5c1c578c9afe.slice/crio-d063201094a200bf1eabd0a40b7a83bca69364da91bc49f3153ef09621f1d206 WatchSource:0}: Error finding container d063201094a200bf1eabd0a40b7a83bca69364da91bc49f3153ef09621f1d206: Status 404 returned error can't find the container with id d063201094a200bf1eabd0a40b7a83bca69364da91bc49f3153ef09621f1d206 Oct 08 18:28:56 crc kubenswrapper[4988]: I1008 18:28:56.378566 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1487d461-4c56-4591-b498-b6ce9f314763-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "1487d461-4c56-4591-b498-b6ce9f314763" (UID: "1487d461-4c56-4591-b498-b6ce9f314763"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 18:28:56 crc kubenswrapper[4988]: I1008 18:28:56.384065 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1487d461-4c56-4591-b498-b6ce9f314763-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "1487d461-4c56-4591-b498-b6ce9f314763" (UID: "1487d461-4c56-4591-b498-b6ce9f314763"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 18:28:56 crc kubenswrapper[4988]: I1008 18:28:56.392872 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1487d461-4c56-4591-b498-b6ce9f314763-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "1487d461-4c56-4591-b498-b6ce9f314763" (UID: "1487d461-4c56-4591-b498-b6ce9f314763"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 18:28:56 crc kubenswrapper[4988]: I1008 18:28:56.403162 4988 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xbmst\" (UniqueName: \"kubernetes.io/projected/1487d461-4c56-4591-b498-b6ce9f314763-kube-api-access-xbmst\") on node \"crc\" DevicePath \"\"" Oct 08 18:28:56 crc kubenswrapper[4988]: I1008 18:28:56.403195 4988 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/1487d461-4c56-4591-b498-b6ce9f314763-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 08 18:28:56 crc kubenswrapper[4988]: I1008 18:28:56.403207 4988 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1487d461-4c56-4591-b498-b6ce9f314763-config\") on node \"crc\" DevicePath \"\"" Oct 08 18:28:56 crc kubenswrapper[4988]: I1008 18:28:56.403218 4988 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1487d461-4c56-4591-b498-b6ce9f314763-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 08 18:28:56 crc kubenswrapper[4988]: I1008 18:28:56.403227 4988 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/1487d461-4c56-4591-b498-b6ce9f314763-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 08 18:28:56 crc kubenswrapper[4988]: I1008 18:28:56.497849 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57f58c7cff-42svl" event={"ID":"1487d461-4c56-4591-b498-b6ce9f314763","Type":"ContainerDied","Data":"70d8af15af815443cdd9a37912e3cb7997d6d9fee4c892217c40ea75e46d3691"} Oct 08 18:28:56 crc kubenswrapper[4988]: I1008 18:28:56.497905 4988 scope.go:117] "RemoveContainer" containerID="0bdd344781c086aea94f3fd316e2eefc84f2f8da29e9372eab7e14dd7a7a0061" Oct 08 18:28:56 crc kubenswrapper[4988]: I1008 18:28:56.497867 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57f58c7cff-42svl" Oct 08 18:28:56 crc kubenswrapper[4988]: I1008 18:28:56.518800 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-bt5zq" event={"ID":"ac19d73a-13bf-4714-aa01-36b72ee937d9","Type":"ContainerStarted","Data":"26ce2b8666b3f9444f2693312e4cfc9b72728425b1ed4dff0ae989f452ccc809"} Oct 08 18:28:56 crc kubenswrapper[4988]: I1008 18:28:56.518929 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-bt5zq" event={"ID":"ac19d73a-13bf-4714-aa01-36b72ee937d9","Type":"ContainerStarted","Data":"389c113c6c58bae6e673603cd28a23a63ddec1b35dc51285493bd6fe8eb0872f"} Oct 08 18:28:56 crc kubenswrapper[4988]: I1008 18:28:56.527650 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-wt7bm" event={"ID":"ee22c4f7-5dc2-4f3c-99d7-5c1c578c9afe","Type":"ContainerStarted","Data":"d063201094a200bf1eabd0a40b7a83bca69364da91bc49f3153ef09621f1d206"} Oct 08 18:28:56 crc kubenswrapper[4988]: I1008 18:28:56.529820 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-wzc5c" event={"ID":"434ac32e-3136-4d51-b68e-919adae50ffe","Type":"ContainerStarted","Data":"e2b1898b307d0726f0f6ff4d1d2b69f3e595e9b0e561204edf8d55f8a526cddb"} Oct 08 18:28:56 crc kubenswrapper[4988]: I1008 18:28:56.531413 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-8dclb" event={"ID":"99149e0d-f06b-4136-88da-313ab1705081","Type":"ContainerStarted","Data":"d5f878a0300b19da28ad213f53d2544c5f09e8484e75b1e23f6e9017bf96800e"} Oct 08 18:28:56 crc kubenswrapper[4988]: I1008 18:28:56.551131 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-db-create-bt5zq" podStartSLOduration=11.551110742 podStartE2EDuration="11.551110742s" podCreationTimestamp="2025-10-08 18:28:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 18:28:56.550091991 +0000 UTC m=+1081.999934781" watchObservedRunningTime="2025-10-08 18:28:56.551110742 +0000 UTC m=+1082.000953512" Oct 08 18:28:56 crc kubenswrapper[4988]: E1008 18:28:56.575901 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"glance-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-glance-api@sha256:9ebf424d4107275a2e3f21f7a18ef257ff2f97c1298109ac7c802a5a4f4794f2\\\"\"" pod="openstack/glance-db-sync-5v5z9" podUID="c1f2511c-e7bc-4433-910b-eb3b94fcf09b" Oct 08 18:28:56 crc kubenswrapper[4988]: I1008 18:28:56.576257 4988 scope.go:117] "RemoveContainer" containerID="0ab1e605fc110bb68f6ddc8059bafeb690570e0b3d4bbdc86f04efaca042ac71" Oct 08 18:28:56 crc kubenswrapper[4988]: I1008 18:28:56.662467 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-57f58c7cff-42svl"] Oct 08 18:28:56 crc kubenswrapper[4988]: I1008 18:28:56.673687 4988 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-57f58c7cff-42svl"] Oct 08 18:28:57 crc kubenswrapper[4988]: I1008 18:28:57.250236 4988 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1487d461-4c56-4591-b498-b6ce9f314763" path="/var/lib/kubelet/pods/1487d461-4c56-4591-b498-b6ce9f314763/volumes" Oct 08 18:28:57 crc kubenswrapper[4988]: I1008 18:28:57.543643 4988 generic.go:334] "Generic (PLEG): container finished" podID="ac19d73a-13bf-4714-aa01-36b72ee937d9" containerID="26ce2b8666b3f9444f2693312e4cfc9b72728425b1ed4dff0ae989f452ccc809" exitCode=0 Oct 08 18:28:57 crc kubenswrapper[4988]: I1008 18:28:57.543686 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-bt5zq" event={"ID":"ac19d73a-13bf-4714-aa01-36b72ee937d9","Type":"ContainerDied","Data":"26ce2b8666b3f9444f2693312e4cfc9b72728425b1ed4dff0ae989f452ccc809"} Oct 08 18:28:57 crc kubenswrapper[4988]: I1008 18:28:57.552403 4988 generic.go:334] "Generic (PLEG): container finished" podID="ee22c4f7-5dc2-4f3c-99d7-5c1c578c9afe" containerID="d8856aa4d121b65fbe8cb4f1653c4695099ac2b3a12881a31fcab24a31fa9319" exitCode=0 Oct 08 18:28:57 crc kubenswrapper[4988]: I1008 18:28:57.552599 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-wt7bm" event={"ID":"ee22c4f7-5dc2-4f3c-99d7-5c1c578c9afe","Type":"ContainerDied","Data":"d8856aa4d121b65fbe8cb4f1653c4695099ac2b3a12881a31fcab24a31fa9319"} Oct 08 18:28:57 crc kubenswrapper[4988]: I1008 18:28:57.554812 4988 generic.go:334] "Generic (PLEG): container finished" podID="99149e0d-f06b-4136-88da-313ab1705081" containerID="69d3e16a97ce13679cc0f4e17a4f51c18734213323f046dd107262854a6f6bb8" exitCode=0 Oct 08 18:28:57 crc kubenswrapper[4988]: I1008 18:28:57.554870 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-8dclb" event={"ID":"99149e0d-f06b-4136-88da-313ab1705081","Type":"ContainerDied","Data":"69d3e16a97ce13679cc0f4e17a4f51c18734213323f046dd107262854a6f6bb8"} Oct 08 18:29:00 crc kubenswrapper[4988]: I1008 18:29:00.737849 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-bt5zq" Oct 08 18:29:00 crc kubenswrapper[4988]: I1008 18:29:00.750635 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-wt7bm" Oct 08 18:29:00 crc kubenswrapper[4988]: I1008 18:29:00.759959 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-8dclb" Oct 08 18:29:00 crc kubenswrapper[4988]: I1008 18:29:00.809205 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8vm47\" (UniqueName: \"kubernetes.io/projected/ac19d73a-13bf-4714-aa01-36b72ee937d9-kube-api-access-8vm47\") pod \"ac19d73a-13bf-4714-aa01-36b72ee937d9\" (UID: \"ac19d73a-13bf-4714-aa01-36b72ee937d9\") " Oct 08 18:29:00 crc kubenswrapper[4988]: I1008 18:29:00.838131 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ac19d73a-13bf-4714-aa01-36b72ee937d9-kube-api-access-8vm47" (OuterVolumeSpecName: "kube-api-access-8vm47") pod "ac19d73a-13bf-4714-aa01-36b72ee937d9" (UID: "ac19d73a-13bf-4714-aa01-36b72ee937d9"). InnerVolumeSpecName "kube-api-access-8vm47". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:29:00 crc kubenswrapper[4988]: I1008 18:29:00.911268 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kwbkz\" (UniqueName: \"kubernetes.io/projected/99149e0d-f06b-4136-88da-313ab1705081-kube-api-access-kwbkz\") pod \"99149e0d-f06b-4136-88da-313ab1705081\" (UID: \"99149e0d-f06b-4136-88da-313ab1705081\") " Oct 08 18:29:00 crc kubenswrapper[4988]: I1008 18:29:00.911996 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v7dpx\" (UniqueName: \"kubernetes.io/projected/ee22c4f7-5dc2-4f3c-99d7-5c1c578c9afe-kube-api-access-v7dpx\") pod \"ee22c4f7-5dc2-4f3c-99d7-5c1c578c9afe\" (UID: \"ee22c4f7-5dc2-4f3c-99d7-5c1c578c9afe\") " Oct 08 18:29:00 crc kubenswrapper[4988]: I1008 18:29:00.912453 4988 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8vm47\" (UniqueName: \"kubernetes.io/projected/ac19d73a-13bf-4714-aa01-36b72ee937d9-kube-api-access-8vm47\") on node \"crc\" DevicePath \"\"" Oct 08 18:29:00 crc kubenswrapper[4988]: I1008 18:29:00.914368 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/99149e0d-f06b-4136-88da-313ab1705081-kube-api-access-kwbkz" (OuterVolumeSpecName: "kube-api-access-kwbkz") pod "99149e0d-f06b-4136-88da-313ab1705081" (UID: "99149e0d-f06b-4136-88da-313ab1705081"). InnerVolumeSpecName "kube-api-access-kwbkz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:29:00 crc kubenswrapper[4988]: I1008 18:29:00.914810 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ee22c4f7-5dc2-4f3c-99d7-5c1c578c9afe-kube-api-access-v7dpx" (OuterVolumeSpecName: "kube-api-access-v7dpx") pod "ee22c4f7-5dc2-4f3c-99d7-5c1c578c9afe" (UID: "ee22c4f7-5dc2-4f3c-99d7-5c1c578c9afe"). InnerVolumeSpecName "kube-api-access-v7dpx". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:29:01 crc kubenswrapper[4988]: I1008 18:29:01.014546 4988 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v7dpx\" (UniqueName: \"kubernetes.io/projected/ee22c4f7-5dc2-4f3c-99d7-5c1c578c9afe-kube-api-access-v7dpx\") on node \"crc\" DevicePath \"\"" Oct 08 18:29:01 crc kubenswrapper[4988]: I1008 18:29:01.014586 4988 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kwbkz\" (UniqueName: \"kubernetes.io/projected/99149e0d-f06b-4136-88da-313ab1705081-kube-api-access-kwbkz\") on node \"crc\" DevicePath \"\"" Oct 08 18:29:01 crc kubenswrapper[4988]: I1008 18:29:01.593918 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-wzc5c" event={"ID":"434ac32e-3136-4d51-b68e-919adae50ffe","Type":"ContainerStarted","Data":"57dcd881b0a41e0c7d3247b6644d30ca7dd361b249e7b89b3a8cd1fe5df4b6c9"} Oct 08 18:29:01 crc kubenswrapper[4988]: I1008 18:29:01.596230 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-8dclb" Oct 08 18:29:01 crc kubenswrapper[4988]: I1008 18:29:01.596238 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-8dclb" event={"ID":"99149e0d-f06b-4136-88da-313ab1705081","Type":"ContainerDied","Data":"d5f878a0300b19da28ad213f53d2544c5f09e8484e75b1e23f6e9017bf96800e"} Oct 08 18:29:01 crc kubenswrapper[4988]: I1008 18:29:01.596397 4988 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d5f878a0300b19da28ad213f53d2544c5f09e8484e75b1e23f6e9017bf96800e" Oct 08 18:29:01 crc kubenswrapper[4988]: I1008 18:29:01.598511 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-bt5zq" Oct 08 18:29:01 crc kubenswrapper[4988]: I1008 18:29:01.598972 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-bt5zq" event={"ID":"ac19d73a-13bf-4714-aa01-36b72ee937d9","Type":"ContainerDied","Data":"389c113c6c58bae6e673603cd28a23a63ddec1b35dc51285493bd6fe8eb0872f"} Oct 08 18:29:01 crc kubenswrapper[4988]: I1008 18:29:01.599001 4988 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="389c113c6c58bae6e673603cd28a23a63ddec1b35dc51285493bd6fe8eb0872f" Oct 08 18:29:01 crc kubenswrapper[4988]: I1008 18:29:01.601188 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-wt7bm" event={"ID":"ee22c4f7-5dc2-4f3c-99d7-5c1c578c9afe","Type":"ContainerDied","Data":"d063201094a200bf1eabd0a40b7a83bca69364da91bc49f3153ef09621f1d206"} Oct 08 18:29:01 crc kubenswrapper[4988]: I1008 18:29:01.601230 4988 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d063201094a200bf1eabd0a40b7a83bca69364da91bc49f3153ef09621f1d206" Oct 08 18:29:01 crc kubenswrapper[4988]: I1008 18:29:01.601293 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-wt7bm" Oct 08 18:29:01 crc kubenswrapper[4988]: I1008 18:29:01.625995 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-db-sync-wzc5c" podStartSLOduration=12.134900372 podStartE2EDuration="16.625970949s" podCreationTimestamp="2025-10-08 18:28:45 +0000 UTC" firstStartedPulling="2025-10-08 18:28:56.343937625 +0000 UTC m=+1081.793780395" lastFinishedPulling="2025-10-08 18:29:00.835008202 +0000 UTC m=+1086.284850972" observedRunningTime="2025-10-08 18:29:01.623039176 +0000 UTC m=+1087.072881996" watchObservedRunningTime="2025-10-08 18:29:01.625970949 +0000 UTC m=+1087.075813729" Oct 08 18:29:04 crc kubenswrapper[4988]: I1008 18:29:04.628501 4988 generic.go:334] "Generic (PLEG): container finished" podID="434ac32e-3136-4d51-b68e-919adae50ffe" containerID="57dcd881b0a41e0c7d3247b6644d30ca7dd361b249e7b89b3a8cd1fe5df4b6c9" exitCode=0 Oct 08 18:29:04 crc kubenswrapper[4988]: I1008 18:29:04.628610 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-wzc5c" event={"ID":"434ac32e-3136-4d51-b68e-919adae50ffe","Type":"ContainerDied","Data":"57dcd881b0a41e0c7d3247b6644d30ca7dd361b249e7b89b3a8cd1fe5df4b6c9"} Oct 08 18:29:05 crc kubenswrapper[4988]: I1008 18:29:05.925707 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-wzc5c" Oct 08 18:29:05 crc kubenswrapper[4988]: I1008 18:29:05.997541 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-f9jd5\" (UniqueName: \"kubernetes.io/projected/434ac32e-3136-4d51-b68e-919adae50ffe-kube-api-access-f9jd5\") pod \"434ac32e-3136-4d51-b68e-919adae50ffe\" (UID: \"434ac32e-3136-4d51-b68e-919adae50ffe\") " Oct 08 18:29:05 crc kubenswrapper[4988]: I1008 18:29:05.997604 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/434ac32e-3136-4d51-b68e-919adae50ffe-config-data\") pod \"434ac32e-3136-4d51-b68e-919adae50ffe\" (UID: \"434ac32e-3136-4d51-b68e-919adae50ffe\") " Oct 08 18:29:05 crc kubenswrapper[4988]: I1008 18:29:05.997676 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/434ac32e-3136-4d51-b68e-919adae50ffe-combined-ca-bundle\") pod \"434ac32e-3136-4d51-b68e-919adae50ffe\" (UID: \"434ac32e-3136-4d51-b68e-919adae50ffe\") " Oct 08 18:29:06 crc kubenswrapper[4988]: I1008 18:29:06.003922 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/434ac32e-3136-4d51-b68e-919adae50ffe-kube-api-access-f9jd5" (OuterVolumeSpecName: "kube-api-access-f9jd5") pod "434ac32e-3136-4d51-b68e-919adae50ffe" (UID: "434ac32e-3136-4d51-b68e-919adae50ffe"). InnerVolumeSpecName "kube-api-access-f9jd5". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:29:06 crc kubenswrapper[4988]: I1008 18:29:06.042207 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/434ac32e-3136-4d51-b68e-919adae50ffe-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "434ac32e-3136-4d51-b68e-919adae50ffe" (UID: "434ac32e-3136-4d51-b68e-919adae50ffe"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:29:06 crc kubenswrapper[4988]: I1008 18:29:06.042981 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/434ac32e-3136-4d51-b68e-919adae50ffe-config-data" (OuterVolumeSpecName: "config-data") pod "434ac32e-3136-4d51-b68e-919adae50ffe" (UID: "434ac32e-3136-4d51-b68e-919adae50ffe"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:29:06 crc kubenswrapper[4988]: I1008 18:29:06.099847 4988 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/434ac32e-3136-4d51-b68e-919adae50ffe-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 18:29:06 crc kubenswrapper[4988]: I1008 18:29:06.099904 4988 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-f9jd5\" (UniqueName: \"kubernetes.io/projected/434ac32e-3136-4d51-b68e-919adae50ffe-kube-api-access-f9jd5\") on node \"crc\" DevicePath \"\"" Oct 08 18:29:06 crc kubenswrapper[4988]: I1008 18:29:06.099935 4988 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/434ac32e-3136-4d51-b68e-919adae50ffe-config-data\") on node \"crc\" DevicePath \"\"" Oct 08 18:29:06 crc kubenswrapper[4988]: I1008 18:29:06.649688 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-wzc5c" event={"ID":"434ac32e-3136-4d51-b68e-919adae50ffe","Type":"ContainerDied","Data":"e2b1898b307d0726f0f6ff4d1d2b69f3e595e9b0e561204edf8d55f8a526cddb"} Oct 08 18:29:06 crc kubenswrapper[4988]: I1008 18:29:06.650378 4988 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e2b1898b307d0726f0f6ff4d1d2b69f3e595e9b0e561204edf8d55f8a526cddb" Oct 08 18:29:06 crc kubenswrapper[4988]: I1008 18:29:06.649765 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-wzc5c" Oct 08 18:29:06 crc kubenswrapper[4988]: I1008 18:29:06.911791 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-758b46c4c7-wgkvt"] Oct 08 18:29:06 crc kubenswrapper[4988]: E1008 18:29:06.912126 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1487d461-4c56-4591-b498-b6ce9f314763" containerName="init" Oct 08 18:29:06 crc kubenswrapper[4988]: I1008 18:29:06.912145 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="1487d461-4c56-4591-b498-b6ce9f314763" containerName="init" Oct 08 18:29:06 crc kubenswrapper[4988]: E1008 18:29:06.912155 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ac19d73a-13bf-4714-aa01-36b72ee937d9" containerName="mariadb-database-create" Oct 08 18:29:06 crc kubenswrapper[4988]: I1008 18:29:06.912162 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="ac19d73a-13bf-4714-aa01-36b72ee937d9" containerName="mariadb-database-create" Oct 08 18:29:06 crc kubenswrapper[4988]: E1008 18:29:06.912173 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1487d461-4c56-4591-b498-b6ce9f314763" containerName="dnsmasq-dns" Oct 08 18:29:06 crc kubenswrapper[4988]: I1008 18:29:06.912180 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="1487d461-4c56-4591-b498-b6ce9f314763" containerName="dnsmasq-dns" Oct 08 18:29:06 crc kubenswrapper[4988]: E1008 18:29:06.912208 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ee22c4f7-5dc2-4f3c-99d7-5c1c578c9afe" containerName="mariadb-database-create" Oct 08 18:29:06 crc kubenswrapper[4988]: I1008 18:29:06.912216 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="ee22c4f7-5dc2-4f3c-99d7-5c1c578c9afe" containerName="mariadb-database-create" Oct 08 18:29:06 crc kubenswrapper[4988]: E1008 18:29:06.912229 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="99149e0d-f06b-4136-88da-313ab1705081" containerName="mariadb-database-create" Oct 08 18:29:06 crc kubenswrapper[4988]: I1008 18:29:06.912237 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="99149e0d-f06b-4136-88da-313ab1705081" containerName="mariadb-database-create" Oct 08 18:29:06 crc kubenswrapper[4988]: E1008 18:29:06.912253 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="434ac32e-3136-4d51-b68e-919adae50ffe" containerName="keystone-db-sync" Oct 08 18:29:06 crc kubenswrapper[4988]: I1008 18:29:06.912262 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="434ac32e-3136-4d51-b68e-919adae50ffe" containerName="keystone-db-sync" Oct 08 18:29:06 crc kubenswrapper[4988]: I1008 18:29:06.912492 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="1487d461-4c56-4591-b498-b6ce9f314763" containerName="dnsmasq-dns" Oct 08 18:29:06 crc kubenswrapper[4988]: I1008 18:29:06.912513 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="ac19d73a-13bf-4714-aa01-36b72ee937d9" containerName="mariadb-database-create" Oct 08 18:29:06 crc kubenswrapper[4988]: I1008 18:29:06.912526 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="99149e0d-f06b-4136-88da-313ab1705081" containerName="mariadb-database-create" Oct 08 18:29:06 crc kubenswrapper[4988]: I1008 18:29:06.912545 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="ee22c4f7-5dc2-4f3c-99d7-5c1c578c9afe" containerName="mariadb-database-create" Oct 08 18:29:06 crc kubenswrapper[4988]: I1008 18:29:06.912554 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="434ac32e-3136-4d51-b68e-919adae50ffe" containerName="keystone-db-sync" Oct 08 18:29:06 crc kubenswrapper[4988]: I1008 18:29:06.913496 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-758b46c4c7-wgkvt" Oct 08 18:29:06 crc kubenswrapper[4988]: I1008 18:29:06.929958 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-758b46c4c7-wgkvt"] Oct 08 18:29:06 crc kubenswrapper[4988]: I1008 18:29:06.950343 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-jppx4"] Oct 08 18:29:06 crc kubenswrapper[4988]: I1008 18:29:06.951947 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-jppx4" Oct 08 18:29:06 crc kubenswrapper[4988]: I1008 18:29:06.956894 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Oct 08 18:29:06 crc kubenswrapper[4988]: I1008 18:29:06.957309 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Oct 08 18:29:06 crc kubenswrapper[4988]: I1008 18:29:06.962926 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-jppx4"] Oct 08 18:29:06 crc kubenswrapper[4988]: I1008 18:29:06.964802 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Oct 08 18:29:06 crc kubenswrapper[4988]: I1008 18:29:06.965038 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-srkzn" Oct 08 18:29:07 crc kubenswrapper[4988]: I1008 18:29:07.012769 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/53faaa07-2104-4c78-98ba-3f2f2c1926d7-fernet-keys\") pod \"keystone-bootstrap-jppx4\" (UID: \"53faaa07-2104-4c78-98ba-3f2f2c1926d7\") " pod="openstack/keystone-bootstrap-jppx4" Oct 08 18:29:07 crc kubenswrapper[4988]: I1008 18:29:07.012825 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ec575519-d902-40eb-81e9-5b8a6b64f70a-dns-svc\") pod \"dnsmasq-dns-758b46c4c7-wgkvt\" (UID: \"ec575519-d902-40eb-81e9-5b8a6b64f70a\") " pod="openstack/dnsmasq-dns-758b46c4c7-wgkvt" Oct 08 18:29:07 crc kubenswrapper[4988]: I1008 18:29:07.012859 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/53faaa07-2104-4c78-98ba-3f2f2c1926d7-combined-ca-bundle\") pod \"keystone-bootstrap-jppx4\" (UID: \"53faaa07-2104-4c78-98ba-3f2f2c1926d7\") " pod="openstack/keystone-bootstrap-jppx4" Oct 08 18:29:07 crc kubenswrapper[4988]: I1008 18:29:07.012916 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g9lg6\" (UniqueName: \"kubernetes.io/projected/ec575519-d902-40eb-81e9-5b8a6b64f70a-kube-api-access-g9lg6\") pod \"dnsmasq-dns-758b46c4c7-wgkvt\" (UID: \"ec575519-d902-40eb-81e9-5b8a6b64f70a\") " pod="openstack/dnsmasq-dns-758b46c4c7-wgkvt" Oct 08 18:29:07 crc kubenswrapper[4988]: I1008 18:29:07.012942 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/53faaa07-2104-4c78-98ba-3f2f2c1926d7-scripts\") pod \"keystone-bootstrap-jppx4\" (UID: \"53faaa07-2104-4c78-98ba-3f2f2c1926d7\") " pod="openstack/keystone-bootstrap-jppx4" Oct 08 18:29:07 crc kubenswrapper[4988]: I1008 18:29:07.012984 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/ec575519-d902-40eb-81e9-5b8a6b64f70a-dns-swift-storage-0\") pod \"dnsmasq-dns-758b46c4c7-wgkvt\" (UID: \"ec575519-d902-40eb-81e9-5b8a6b64f70a\") " pod="openstack/dnsmasq-dns-758b46c4c7-wgkvt" Oct 08 18:29:07 crc kubenswrapper[4988]: I1008 18:29:07.013008 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/53faaa07-2104-4c78-98ba-3f2f2c1926d7-credential-keys\") pod \"keystone-bootstrap-jppx4\" (UID: \"53faaa07-2104-4c78-98ba-3f2f2c1926d7\") " pod="openstack/keystone-bootstrap-jppx4" Oct 08 18:29:07 crc kubenswrapper[4988]: I1008 18:29:07.013102 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ec575519-d902-40eb-81e9-5b8a6b64f70a-ovsdbserver-nb\") pod \"dnsmasq-dns-758b46c4c7-wgkvt\" (UID: \"ec575519-d902-40eb-81e9-5b8a6b64f70a\") " pod="openstack/dnsmasq-dns-758b46c4c7-wgkvt" Oct 08 18:29:07 crc kubenswrapper[4988]: I1008 18:29:07.013195 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ec575519-d902-40eb-81e9-5b8a6b64f70a-ovsdbserver-sb\") pod \"dnsmasq-dns-758b46c4c7-wgkvt\" (UID: \"ec575519-d902-40eb-81e9-5b8a6b64f70a\") " pod="openstack/dnsmasq-dns-758b46c4c7-wgkvt" Oct 08 18:29:07 crc kubenswrapper[4988]: I1008 18:29:07.013356 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/53faaa07-2104-4c78-98ba-3f2f2c1926d7-config-data\") pod \"keystone-bootstrap-jppx4\" (UID: \"53faaa07-2104-4c78-98ba-3f2f2c1926d7\") " pod="openstack/keystone-bootstrap-jppx4" Oct 08 18:29:07 crc kubenswrapper[4988]: I1008 18:29:07.013418 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s4cp8\" (UniqueName: \"kubernetes.io/projected/53faaa07-2104-4c78-98ba-3f2f2c1926d7-kube-api-access-s4cp8\") pod \"keystone-bootstrap-jppx4\" (UID: \"53faaa07-2104-4c78-98ba-3f2f2c1926d7\") " pod="openstack/keystone-bootstrap-jppx4" Oct 08 18:29:07 crc kubenswrapper[4988]: I1008 18:29:07.013446 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ec575519-d902-40eb-81e9-5b8a6b64f70a-config\") pod \"dnsmasq-dns-758b46c4c7-wgkvt\" (UID: \"ec575519-d902-40eb-81e9-5b8a6b64f70a\") " pod="openstack/dnsmasq-dns-758b46c4c7-wgkvt" Oct 08 18:29:07 crc kubenswrapper[4988]: I1008 18:29:07.083662 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 08 18:29:07 crc kubenswrapper[4988]: I1008 18:29:07.085616 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 08 18:29:07 crc kubenswrapper[4988]: I1008 18:29:07.087970 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 08 18:29:07 crc kubenswrapper[4988]: I1008 18:29:07.088071 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 08 18:29:07 crc kubenswrapper[4988]: I1008 18:29:07.104261 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 08 18:29:07 crc kubenswrapper[4988]: I1008 18:29:07.117308 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/53faaa07-2104-4c78-98ba-3f2f2c1926d7-config-data\") pod \"keystone-bootstrap-jppx4\" (UID: \"53faaa07-2104-4c78-98ba-3f2f2c1926d7\") " pod="openstack/keystone-bootstrap-jppx4" Oct 08 18:29:07 crc kubenswrapper[4988]: I1008 18:29:07.117368 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s4cp8\" (UniqueName: \"kubernetes.io/projected/53faaa07-2104-4c78-98ba-3f2f2c1926d7-kube-api-access-s4cp8\") pod \"keystone-bootstrap-jppx4\" (UID: \"53faaa07-2104-4c78-98ba-3f2f2c1926d7\") " pod="openstack/keystone-bootstrap-jppx4" Oct 08 18:29:07 crc kubenswrapper[4988]: I1008 18:29:07.117416 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ec575519-d902-40eb-81e9-5b8a6b64f70a-config\") pod \"dnsmasq-dns-758b46c4c7-wgkvt\" (UID: \"ec575519-d902-40eb-81e9-5b8a6b64f70a\") " pod="openstack/dnsmasq-dns-758b46c4c7-wgkvt" Oct 08 18:29:07 crc kubenswrapper[4988]: I1008 18:29:07.117464 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/53faaa07-2104-4c78-98ba-3f2f2c1926d7-fernet-keys\") pod \"keystone-bootstrap-jppx4\" (UID: \"53faaa07-2104-4c78-98ba-3f2f2c1926d7\") " pod="openstack/keystone-bootstrap-jppx4" Oct 08 18:29:07 crc kubenswrapper[4988]: I1008 18:29:07.117488 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ec575519-d902-40eb-81e9-5b8a6b64f70a-dns-svc\") pod \"dnsmasq-dns-758b46c4c7-wgkvt\" (UID: \"ec575519-d902-40eb-81e9-5b8a6b64f70a\") " pod="openstack/dnsmasq-dns-758b46c4c7-wgkvt" Oct 08 18:29:07 crc kubenswrapper[4988]: I1008 18:29:07.117516 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/53faaa07-2104-4c78-98ba-3f2f2c1926d7-combined-ca-bundle\") pod \"keystone-bootstrap-jppx4\" (UID: \"53faaa07-2104-4c78-98ba-3f2f2c1926d7\") " pod="openstack/keystone-bootstrap-jppx4" Oct 08 18:29:07 crc kubenswrapper[4988]: I1008 18:29:07.117577 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g9lg6\" (UniqueName: \"kubernetes.io/projected/ec575519-d902-40eb-81e9-5b8a6b64f70a-kube-api-access-g9lg6\") pod \"dnsmasq-dns-758b46c4c7-wgkvt\" (UID: \"ec575519-d902-40eb-81e9-5b8a6b64f70a\") " pod="openstack/dnsmasq-dns-758b46c4c7-wgkvt" Oct 08 18:29:07 crc kubenswrapper[4988]: I1008 18:29:07.117605 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/53faaa07-2104-4c78-98ba-3f2f2c1926d7-scripts\") pod \"keystone-bootstrap-jppx4\" (UID: \"53faaa07-2104-4c78-98ba-3f2f2c1926d7\") " pod="openstack/keystone-bootstrap-jppx4" Oct 08 18:29:07 crc kubenswrapper[4988]: I1008 18:29:07.117651 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/ec575519-d902-40eb-81e9-5b8a6b64f70a-dns-swift-storage-0\") pod \"dnsmasq-dns-758b46c4c7-wgkvt\" (UID: \"ec575519-d902-40eb-81e9-5b8a6b64f70a\") " pod="openstack/dnsmasq-dns-758b46c4c7-wgkvt" Oct 08 18:29:07 crc kubenswrapper[4988]: I1008 18:29:07.117676 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/53faaa07-2104-4c78-98ba-3f2f2c1926d7-credential-keys\") pod \"keystone-bootstrap-jppx4\" (UID: \"53faaa07-2104-4c78-98ba-3f2f2c1926d7\") " pod="openstack/keystone-bootstrap-jppx4" Oct 08 18:29:07 crc kubenswrapper[4988]: I1008 18:29:07.117700 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ec575519-d902-40eb-81e9-5b8a6b64f70a-ovsdbserver-nb\") pod \"dnsmasq-dns-758b46c4c7-wgkvt\" (UID: \"ec575519-d902-40eb-81e9-5b8a6b64f70a\") " pod="openstack/dnsmasq-dns-758b46c4c7-wgkvt" Oct 08 18:29:07 crc kubenswrapper[4988]: I1008 18:29:07.117730 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ec575519-d902-40eb-81e9-5b8a6b64f70a-ovsdbserver-sb\") pod \"dnsmasq-dns-758b46c4c7-wgkvt\" (UID: \"ec575519-d902-40eb-81e9-5b8a6b64f70a\") " pod="openstack/dnsmasq-dns-758b46c4c7-wgkvt" Oct 08 18:29:07 crc kubenswrapper[4988]: I1008 18:29:07.119229 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ec575519-d902-40eb-81e9-5b8a6b64f70a-dns-svc\") pod \"dnsmasq-dns-758b46c4c7-wgkvt\" (UID: \"ec575519-d902-40eb-81e9-5b8a6b64f70a\") " pod="openstack/dnsmasq-dns-758b46c4c7-wgkvt" Oct 08 18:29:07 crc kubenswrapper[4988]: I1008 18:29:07.119260 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/ec575519-d902-40eb-81e9-5b8a6b64f70a-dns-swift-storage-0\") pod \"dnsmasq-dns-758b46c4c7-wgkvt\" (UID: \"ec575519-d902-40eb-81e9-5b8a6b64f70a\") " pod="openstack/dnsmasq-dns-758b46c4c7-wgkvt" Oct 08 18:29:07 crc kubenswrapper[4988]: I1008 18:29:07.119228 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ec575519-d902-40eb-81e9-5b8a6b64f70a-ovsdbserver-nb\") pod \"dnsmasq-dns-758b46c4c7-wgkvt\" (UID: \"ec575519-d902-40eb-81e9-5b8a6b64f70a\") " pod="openstack/dnsmasq-dns-758b46c4c7-wgkvt" Oct 08 18:29:07 crc kubenswrapper[4988]: I1008 18:29:07.119755 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ec575519-d902-40eb-81e9-5b8a6b64f70a-ovsdbserver-sb\") pod \"dnsmasq-dns-758b46c4c7-wgkvt\" (UID: \"ec575519-d902-40eb-81e9-5b8a6b64f70a\") " pod="openstack/dnsmasq-dns-758b46c4c7-wgkvt" Oct 08 18:29:07 crc kubenswrapper[4988]: I1008 18:29:07.122948 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ec575519-d902-40eb-81e9-5b8a6b64f70a-config\") pod \"dnsmasq-dns-758b46c4c7-wgkvt\" (UID: \"ec575519-d902-40eb-81e9-5b8a6b64f70a\") " pod="openstack/dnsmasq-dns-758b46c4c7-wgkvt" Oct 08 18:29:07 crc kubenswrapper[4988]: I1008 18:29:07.128725 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/53faaa07-2104-4c78-98ba-3f2f2c1926d7-combined-ca-bundle\") pod \"keystone-bootstrap-jppx4\" (UID: \"53faaa07-2104-4c78-98ba-3f2f2c1926d7\") " pod="openstack/keystone-bootstrap-jppx4" Oct 08 18:29:07 crc kubenswrapper[4988]: I1008 18:29:07.131930 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/53faaa07-2104-4c78-98ba-3f2f2c1926d7-credential-keys\") pod \"keystone-bootstrap-jppx4\" (UID: \"53faaa07-2104-4c78-98ba-3f2f2c1926d7\") " pod="openstack/keystone-bootstrap-jppx4" Oct 08 18:29:07 crc kubenswrapper[4988]: I1008 18:29:07.135441 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/53faaa07-2104-4c78-98ba-3f2f2c1926d7-config-data\") pod \"keystone-bootstrap-jppx4\" (UID: \"53faaa07-2104-4c78-98ba-3f2f2c1926d7\") " pod="openstack/keystone-bootstrap-jppx4" Oct 08 18:29:07 crc kubenswrapper[4988]: I1008 18:29:07.139872 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/53faaa07-2104-4c78-98ba-3f2f2c1926d7-scripts\") pod \"keystone-bootstrap-jppx4\" (UID: \"53faaa07-2104-4c78-98ba-3f2f2c1926d7\") " pod="openstack/keystone-bootstrap-jppx4" Oct 08 18:29:07 crc kubenswrapper[4988]: I1008 18:29:07.152163 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s4cp8\" (UniqueName: \"kubernetes.io/projected/53faaa07-2104-4c78-98ba-3f2f2c1926d7-kube-api-access-s4cp8\") pod \"keystone-bootstrap-jppx4\" (UID: \"53faaa07-2104-4c78-98ba-3f2f2c1926d7\") " pod="openstack/keystone-bootstrap-jppx4" Oct 08 18:29:07 crc kubenswrapper[4988]: I1008 18:29:07.152234 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/53faaa07-2104-4c78-98ba-3f2f2c1926d7-fernet-keys\") pod \"keystone-bootstrap-jppx4\" (UID: \"53faaa07-2104-4c78-98ba-3f2f2c1926d7\") " pod="openstack/keystone-bootstrap-jppx4" Oct 08 18:29:07 crc kubenswrapper[4988]: I1008 18:29:07.179349 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g9lg6\" (UniqueName: \"kubernetes.io/projected/ec575519-d902-40eb-81e9-5b8a6b64f70a-kube-api-access-g9lg6\") pod \"dnsmasq-dns-758b46c4c7-wgkvt\" (UID: \"ec575519-d902-40eb-81e9-5b8a6b64f70a\") " pod="openstack/dnsmasq-dns-758b46c4c7-wgkvt" Oct 08 18:29:07 crc kubenswrapper[4988]: I1008 18:29:07.222037 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8fed08cf-45cb-4fd2-92af-9f30513c0b72-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"8fed08cf-45cb-4fd2-92af-9f30513c0b72\") " pod="openstack/ceilometer-0" Oct 08 18:29:07 crc kubenswrapper[4988]: I1008 18:29:07.222135 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8fed08cf-45cb-4fd2-92af-9f30513c0b72-log-httpd\") pod \"ceilometer-0\" (UID: \"8fed08cf-45cb-4fd2-92af-9f30513c0b72\") " pod="openstack/ceilometer-0" Oct 08 18:29:07 crc kubenswrapper[4988]: I1008 18:29:07.222189 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8fed08cf-45cb-4fd2-92af-9f30513c0b72-config-data\") pod \"ceilometer-0\" (UID: \"8fed08cf-45cb-4fd2-92af-9f30513c0b72\") " pod="openstack/ceilometer-0" Oct 08 18:29:07 crc kubenswrapper[4988]: I1008 18:29:07.222220 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8fed08cf-45cb-4fd2-92af-9f30513c0b72-scripts\") pod \"ceilometer-0\" (UID: \"8fed08cf-45cb-4fd2-92af-9f30513c0b72\") " pod="openstack/ceilometer-0" Oct 08 18:29:07 crc kubenswrapper[4988]: I1008 18:29:07.222258 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4xpgf\" (UniqueName: \"kubernetes.io/projected/8fed08cf-45cb-4fd2-92af-9f30513c0b72-kube-api-access-4xpgf\") pod \"ceilometer-0\" (UID: \"8fed08cf-45cb-4fd2-92af-9f30513c0b72\") " pod="openstack/ceilometer-0" Oct 08 18:29:07 crc kubenswrapper[4988]: I1008 18:29:07.222299 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/8fed08cf-45cb-4fd2-92af-9f30513c0b72-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"8fed08cf-45cb-4fd2-92af-9f30513c0b72\") " pod="openstack/ceilometer-0" Oct 08 18:29:07 crc kubenswrapper[4988]: I1008 18:29:07.222329 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8fed08cf-45cb-4fd2-92af-9f30513c0b72-run-httpd\") pod \"ceilometer-0\" (UID: \"8fed08cf-45cb-4fd2-92af-9f30513c0b72\") " pod="openstack/ceilometer-0" Oct 08 18:29:07 crc kubenswrapper[4988]: I1008 18:29:07.238677 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-758b46c4c7-wgkvt" Oct 08 18:29:07 crc kubenswrapper[4988]: I1008 18:29:07.267088 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-758b46c4c7-wgkvt"] Oct 08 18:29:07 crc kubenswrapper[4988]: I1008 18:29:07.269252 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-jppx4" Oct 08 18:29:07 crc kubenswrapper[4988]: I1008 18:29:07.307015 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-69894dfcd9-948wp"] Oct 08 18:29:07 crc kubenswrapper[4988]: I1008 18:29:07.308821 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-69894dfcd9-948wp" Oct 08 18:29:07 crc kubenswrapper[4988]: I1008 18:29:07.341019 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8fed08cf-45cb-4fd2-92af-9f30513c0b72-log-httpd\") pod \"ceilometer-0\" (UID: \"8fed08cf-45cb-4fd2-92af-9f30513c0b72\") " pod="openstack/ceilometer-0" Oct 08 18:29:07 crc kubenswrapper[4988]: I1008 18:29:07.341458 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8fed08cf-45cb-4fd2-92af-9f30513c0b72-config-data\") pod \"ceilometer-0\" (UID: \"8fed08cf-45cb-4fd2-92af-9f30513c0b72\") " pod="openstack/ceilometer-0" Oct 08 18:29:07 crc kubenswrapper[4988]: I1008 18:29:07.341615 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8fed08cf-45cb-4fd2-92af-9f30513c0b72-scripts\") pod \"ceilometer-0\" (UID: \"8fed08cf-45cb-4fd2-92af-9f30513c0b72\") " pod="openstack/ceilometer-0" Oct 08 18:29:07 crc kubenswrapper[4988]: I1008 18:29:07.341736 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4xpgf\" (UniqueName: \"kubernetes.io/projected/8fed08cf-45cb-4fd2-92af-9f30513c0b72-kube-api-access-4xpgf\") pod \"ceilometer-0\" (UID: \"8fed08cf-45cb-4fd2-92af-9f30513c0b72\") " pod="openstack/ceilometer-0" Oct 08 18:29:07 crc kubenswrapper[4988]: I1008 18:29:07.341887 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/8fed08cf-45cb-4fd2-92af-9f30513c0b72-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"8fed08cf-45cb-4fd2-92af-9f30513c0b72\") " pod="openstack/ceilometer-0" Oct 08 18:29:07 crc kubenswrapper[4988]: I1008 18:29:07.342025 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8fed08cf-45cb-4fd2-92af-9f30513c0b72-run-httpd\") pod \"ceilometer-0\" (UID: \"8fed08cf-45cb-4fd2-92af-9f30513c0b72\") " pod="openstack/ceilometer-0" Oct 08 18:29:07 crc kubenswrapper[4988]: I1008 18:29:07.342267 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8fed08cf-45cb-4fd2-92af-9f30513c0b72-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"8fed08cf-45cb-4fd2-92af-9f30513c0b72\") " pod="openstack/ceilometer-0" Oct 08 18:29:07 crc kubenswrapper[4988]: I1008 18:29:07.343742 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8fed08cf-45cb-4fd2-92af-9f30513c0b72-log-httpd\") pod \"ceilometer-0\" (UID: \"8fed08cf-45cb-4fd2-92af-9f30513c0b72\") " pod="openstack/ceilometer-0" Oct 08 18:29:07 crc kubenswrapper[4988]: I1008 18:29:07.355808 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8fed08cf-45cb-4fd2-92af-9f30513c0b72-run-httpd\") pod \"ceilometer-0\" (UID: \"8fed08cf-45cb-4fd2-92af-9f30513c0b72\") " pod="openstack/ceilometer-0" Oct 08 18:29:07 crc kubenswrapper[4988]: I1008 18:29:07.365040 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8fed08cf-45cb-4fd2-92af-9f30513c0b72-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"8fed08cf-45cb-4fd2-92af-9f30513c0b72\") " pod="openstack/ceilometer-0" Oct 08 18:29:07 crc kubenswrapper[4988]: I1008 18:29:07.373544 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-69894dfcd9-948wp"] Oct 08 18:29:07 crc kubenswrapper[4988]: I1008 18:29:07.373729 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/8fed08cf-45cb-4fd2-92af-9f30513c0b72-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"8fed08cf-45cb-4fd2-92af-9f30513c0b72\") " pod="openstack/ceilometer-0" Oct 08 18:29:07 crc kubenswrapper[4988]: I1008 18:29:07.385483 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4xpgf\" (UniqueName: \"kubernetes.io/projected/8fed08cf-45cb-4fd2-92af-9f30513c0b72-kube-api-access-4xpgf\") pod \"ceilometer-0\" (UID: \"8fed08cf-45cb-4fd2-92af-9f30513c0b72\") " pod="openstack/ceilometer-0" Oct 08 18:29:07 crc kubenswrapper[4988]: I1008 18:29:07.385562 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-sync-gwkbl"] Oct 08 18:29:07 crc kubenswrapper[4988]: I1008 18:29:07.387350 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-gwkbl" Oct 08 18:29:07 crc kubenswrapper[4988]: I1008 18:29:07.388951 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Oct 08 18:29:07 crc kubenswrapper[4988]: I1008 18:29:07.390953 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8fed08cf-45cb-4fd2-92af-9f30513c0b72-scripts\") pod \"ceilometer-0\" (UID: \"8fed08cf-45cb-4fd2-92af-9f30513c0b72\") " pod="openstack/ceilometer-0" Oct 08 18:29:07 crc kubenswrapper[4988]: I1008 18:29:07.391235 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-mxq4l" Oct 08 18:29:07 crc kubenswrapper[4988]: I1008 18:29:07.391427 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Oct 08 18:29:07 crc kubenswrapper[4988]: I1008 18:29:07.395514 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8fed08cf-45cb-4fd2-92af-9f30513c0b72-config-data\") pod \"ceilometer-0\" (UID: \"8fed08cf-45cb-4fd2-92af-9f30513c0b72\") " pod="openstack/ceilometer-0" Oct 08 18:29:07 crc kubenswrapper[4988]: I1008 18:29:07.397675 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-gwkbl"] Oct 08 18:29:07 crc kubenswrapper[4988]: I1008 18:29:07.411994 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 08 18:29:07 crc kubenswrapper[4988]: I1008 18:29:07.445799 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/62b5df6a-6198-4608-8cd7-05bfa27bfd75-combined-ca-bundle\") pod \"placement-db-sync-gwkbl\" (UID: \"62b5df6a-6198-4608-8cd7-05bfa27bfd75\") " pod="openstack/placement-db-sync-gwkbl" Oct 08 18:29:07 crc kubenswrapper[4988]: I1008 18:29:07.445864 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/62b5df6a-6198-4608-8cd7-05bfa27bfd75-logs\") pod \"placement-db-sync-gwkbl\" (UID: \"62b5df6a-6198-4608-8cd7-05bfa27bfd75\") " pod="openstack/placement-db-sync-gwkbl" Oct 08 18:29:07 crc kubenswrapper[4988]: I1008 18:29:07.445905 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7a831457-b429-4339-b166-afed001c770d-ovsdbserver-nb\") pod \"dnsmasq-dns-69894dfcd9-948wp\" (UID: \"7a831457-b429-4339-b166-afed001c770d\") " pod="openstack/dnsmasq-dns-69894dfcd9-948wp" Oct 08 18:29:07 crc kubenswrapper[4988]: I1008 18:29:07.445933 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7a831457-b429-4339-b166-afed001c770d-dns-svc\") pod \"dnsmasq-dns-69894dfcd9-948wp\" (UID: \"7a831457-b429-4339-b166-afed001c770d\") " pod="openstack/dnsmasq-dns-69894dfcd9-948wp" Oct 08 18:29:07 crc kubenswrapper[4988]: I1008 18:29:07.445990 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7a831457-b429-4339-b166-afed001c770d-config\") pod \"dnsmasq-dns-69894dfcd9-948wp\" (UID: \"7a831457-b429-4339-b166-afed001c770d\") " pod="openstack/dnsmasq-dns-69894dfcd9-948wp" Oct 08 18:29:07 crc kubenswrapper[4988]: I1008 18:29:07.446024 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9vbc7\" (UniqueName: \"kubernetes.io/projected/7a831457-b429-4339-b166-afed001c770d-kube-api-access-9vbc7\") pod \"dnsmasq-dns-69894dfcd9-948wp\" (UID: \"7a831457-b429-4339-b166-afed001c770d\") " pod="openstack/dnsmasq-dns-69894dfcd9-948wp" Oct 08 18:29:07 crc kubenswrapper[4988]: I1008 18:29:07.446060 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/62b5df6a-6198-4608-8cd7-05bfa27bfd75-scripts\") pod \"placement-db-sync-gwkbl\" (UID: \"62b5df6a-6198-4608-8cd7-05bfa27bfd75\") " pod="openstack/placement-db-sync-gwkbl" Oct 08 18:29:07 crc kubenswrapper[4988]: I1008 18:29:07.446106 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/7a831457-b429-4339-b166-afed001c770d-dns-swift-storage-0\") pod \"dnsmasq-dns-69894dfcd9-948wp\" (UID: \"7a831457-b429-4339-b166-afed001c770d\") " pod="openstack/dnsmasq-dns-69894dfcd9-948wp" Oct 08 18:29:07 crc kubenswrapper[4988]: I1008 18:29:07.446131 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7a831457-b429-4339-b166-afed001c770d-ovsdbserver-sb\") pod \"dnsmasq-dns-69894dfcd9-948wp\" (UID: \"7a831457-b429-4339-b166-afed001c770d\") " pod="openstack/dnsmasq-dns-69894dfcd9-948wp" Oct 08 18:29:07 crc kubenswrapper[4988]: I1008 18:29:07.446165 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/62b5df6a-6198-4608-8cd7-05bfa27bfd75-config-data\") pod \"placement-db-sync-gwkbl\" (UID: \"62b5df6a-6198-4608-8cd7-05bfa27bfd75\") " pod="openstack/placement-db-sync-gwkbl" Oct 08 18:29:07 crc kubenswrapper[4988]: I1008 18:29:07.446191 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zxtbt\" (UniqueName: \"kubernetes.io/projected/62b5df6a-6198-4608-8cd7-05bfa27bfd75-kube-api-access-zxtbt\") pod \"placement-db-sync-gwkbl\" (UID: \"62b5df6a-6198-4608-8cd7-05bfa27bfd75\") " pod="openstack/placement-db-sync-gwkbl" Oct 08 18:29:07 crc kubenswrapper[4988]: I1008 18:29:07.547901 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9vbc7\" (UniqueName: \"kubernetes.io/projected/7a831457-b429-4339-b166-afed001c770d-kube-api-access-9vbc7\") pod \"dnsmasq-dns-69894dfcd9-948wp\" (UID: \"7a831457-b429-4339-b166-afed001c770d\") " pod="openstack/dnsmasq-dns-69894dfcd9-948wp" Oct 08 18:29:07 crc kubenswrapper[4988]: I1008 18:29:07.547983 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/62b5df6a-6198-4608-8cd7-05bfa27bfd75-scripts\") pod \"placement-db-sync-gwkbl\" (UID: \"62b5df6a-6198-4608-8cd7-05bfa27bfd75\") " pod="openstack/placement-db-sync-gwkbl" Oct 08 18:29:07 crc kubenswrapper[4988]: I1008 18:29:07.548037 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/7a831457-b429-4339-b166-afed001c770d-dns-swift-storage-0\") pod \"dnsmasq-dns-69894dfcd9-948wp\" (UID: \"7a831457-b429-4339-b166-afed001c770d\") " pod="openstack/dnsmasq-dns-69894dfcd9-948wp" Oct 08 18:29:07 crc kubenswrapper[4988]: I1008 18:29:07.548064 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7a831457-b429-4339-b166-afed001c770d-ovsdbserver-sb\") pod \"dnsmasq-dns-69894dfcd9-948wp\" (UID: \"7a831457-b429-4339-b166-afed001c770d\") " pod="openstack/dnsmasq-dns-69894dfcd9-948wp" Oct 08 18:29:07 crc kubenswrapper[4988]: I1008 18:29:07.548099 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/62b5df6a-6198-4608-8cd7-05bfa27bfd75-config-data\") pod \"placement-db-sync-gwkbl\" (UID: \"62b5df6a-6198-4608-8cd7-05bfa27bfd75\") " pod="openstack/placement-db-sync-gwkbl" Oct 08 18:29:07 crc kubenswrapper[4988]: I1008 18:29:07.552627 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zxtbt\" (UniqueName: \"kubernetes.io/projected/62b5df6a-6198-4608-8cd7-05bfa27bfd75-kube-api-access-zxtbt\") pod \"placement-db-sync-gwkbl\" (UID: \"62b5df6a-6198-4608-8cd7-05bfa27bfd75\") " pod="openstack/placement-db-sync-gwkbl" Oct 08 18:29:07 crc kubenswrapper[4988]: I1008 18:29:07.552752 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/62b5df6a-6198-4608-8cd7-05bfa27bfd75-combined-ca-bundle\") pod \"placement-db-sync-gwkbl\" (UID: \"62b5df6a-6198-4608-8cd7-05bfa27bfd75\") " pod="openstack/placement-db-sync-gwkbl" Oct 08 18:29:07 crc kubenswrapper[4988]: I1008 18:29:07.552816 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/62b5df6a-6198-4608-8cd7-05bfa27bfd75-logs\") pod \"placement-db-sync-gwkbl\" (UID: \"62b5df6a-6198-4608-8cd7-05bfa27bfd75\") " pod="openstack/placement-db-sync-gwkbl" Oct 08 18:29:07 crc kubenswrapper[4988]: I1008 18:29:07.552912 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7a831457-b429-4339-b166-afed001c770d-ovsdbserver-nb\") pod \"dnsmasq-dns-69894dfcd9-948wp\" (UID: \"7a831457-b429-4339-b166-afed001c770d\") " pod="openstack/dnsmasq-dns-69894dfcd9-948wp" Oct 08 18:29:07 crc kubenswrapper[4988]: I1008 18:29:07.552947 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7a831457-b429-4339-b166-afed001c770d-dns-svc\") pod \"dnsmasq-dns-69894dfcd9-948wp\" (UID: \"7a831457-b429-4339-b166-afed001c770d\") " pod="openstack/dnsmasq-dns-69894dfcd9-948wp" Oct 08 18:29:07 crc kubenswrapper[4988]: I1008 18:29:07.553066 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7a831457-b429-4339-b166-afed001c770d-config\") pod \"dnsmasq-dns-69894dfcd9-948wp\" (UID: \"7a831457-b429-4339-b166-afed001c770d\") " pod="openstack/dnsmasq-dns-69894dfcd9-948wp" Oct 08 18:29:07 crc kubenswrapper[4988]: I1008 18:29:07.554701 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7a831457-b429-4339-b166-afed001c770d-config\") pod \"dnsmasq-dns-69894dfcd9-948wp\" (UID: \"7a831457-b429-4339-b166-afed001c770d\") " pod="openstack/dnsmasq-dns-69894dfcd9-948wp" Oct 08 18:29:07 crc kubenswrapper[4988]: I1008 18:29:07.554974 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/7a831457-b429-4339-b166-afed001c770d-dns-swift-storage-0\") pod \"dnsmasq-dns-69894dfcd9-948wp\" (UID: \"7a831457-b429-4339-b166-afed001c770d\") " pod="openstack/dnsmasq-dns-69894dfcd9-948wp" Oct 08 18:29:07 crc kubenswrapper[4988]: I1008 18:29:07.556308 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/62b5df6a-6198-4608-8cd7-05bfa27bfd75-logs\") pod \"placement-db-sync-gwkbl\" (UID: \"62b5df6a-6198-4608-8cd7-05bfa27bfd75\") " pod="openstack/placement-db-sync-gwkbl" Oct 08 18:29:07 crc kubenswrapper[4988]: I1008 18:29:07.556593 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/62b5df6a-6198-4608-8cd7-05bfa27bfd75-scripts\") pod \"placement-db-sync-gwkbl\" (UID: \"62b5df6a-6198-4608-8cd7-05bfa27bfd75\") " pod="openstack/placement-db-sync-gwkbl" Oct 08 18:29:07 crc kubenswrapper[4988]: I1008 18:29:07.558230 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/62b5df6a-6198-4608-8cd7-05bfa27bfd75-config-data\") pod \"placement-db-sync-gwkbl\" (UID: \"62b5df6a-6198-4608-8cd7-05bfa27bfd75\") " pod="openstack/placement-db-sync-gwkbl" Oct 08 18:29:07 crc kubenswrapper[4988]: I1008 18:29:07.558687 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7a831457-b429-4339-b166-afed001c770d-dns-svc\") pod \"dnsmasq-dns-69894dfcd9-948wp\" (UID: \"7a831457-b429-4339-b166-afed001c770d\") " pod="openstack/dnsmasq-dns-69894dfcd9-948wp" Oct 08 18:29:07 crc kubenswrapper[4988]: I1008 18:29:07.558887 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7a831457-b429-4339-b166-afed001c770d-ovsdbserver-sb\") pod \"dnsmasq-dns-69894dfcd9-948wp\" (UID: \"7a831457-b429-4339-b166-afed001c770d\") " pod="openstack/dnsmasq-dns-69894dfcd9-948wp" Oct 08 18:29:07 crc kubenswrapper[4988]: I1008 18:29:07.560726 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7a831457-b429-4339-b166-afed001c770d-ovsdbserver-nb\") pod \"dnsmasq-dns-69894dfcd9-948wp\" (UID: \"7a831457-b429-4339-b166-afed001c770d\") " pod="openstack/dnsmasq-dns-69894dfcd9-948wp" Oct 08 18:29:07 crc kubenswrapper[4988]: I1008 18:29:07.566895 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/62b5df6a-6198-4608-8cd7-05bfa27bfd75-combined-ca-bundle\") pod \"placement-db-sync-gwkbl\" (UID: \"62b5df6a-6198-4608-8cd7-05bfa27bfd75\") " pod="openstack/placement-db-sync-gwkbl" Oct 08 18:29:07 crc kubenswrapper[4988]: I1008 18:29:07.579220 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9vbc7\" (UniqueName: \"kubernetes.io/projected/7a831457-b429-4339-b166-afed001c770d-kube-api-access-9vbc7\") pod \"dnsmasq-dns-69894dfcd9-948wp\" (UID: \"7a831457-b429-4339-b166-afed001c770d\") " pod="openstack/dnsmasq-dns-69894dfcd9-948wp" Oct 08 18:29:07 crc kubenswrapper[4988]: I1008 18:29:07.579973 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zxtbt\" (UniqueName: \"kubernetes.io/projected/62b5df6a-6198-4608-8cd7-05bfa27bfd75-kube-api-access-zxtbt\") pod \"placement-db-sync-gwkbl\" (UID: \"62b5df6a-6198-4608-8cd7-05bfa27bfd75\") " pod="openstack/placement-db-sync-gwkbl" Oct 08 18:29:07 crc kubenswrapper[4988]: I1008 18:29:07.621830 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-758b46c4c7-wgkvt"] Oct 08 18:29:07 crc kubenswrapper[4988]: I1008 18:29:07.657783 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-758b46c4c7-wgkvt" event={"ID":"ec575519-d902-40eb-81e9-5b8a6b64f70a","Type":"ContainerStarted","Data":"761f4d3bfde0172eb85a2cbb06479cbba368bb82727503d11adfe807d3da434b"} Oct 08 18:29:07 crc kubenswrapper[4988]: I1008 18:29:07.712805 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-69894dfcd9-948wp" Oct 08 18:29:07 crc kubenswrapper[4988]: I1008 18:29:07.732356 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-gwkbl" Oct 08 18:29:07 crc kubenswrapper[4988]: I1008 18:29:07.865834 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 08 18:29:07 crc kubenswrapper[4988]: I1008 18:29:07.926088 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-jppx4"] Oct 08 18:29:08 crc kubenswrapper[4988]: I1008 18:29:08.278552 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-gwkbl"] Oct 08 18:29:08 crc kubenswrapper[4988]: I1008 18:29:08.286708 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-69894dfcd9-948wp"] Oct 08 18:29:08 crc kubenswrapper[4988]: I1008 18:29:08.672519 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-jppx4" event={"ID":"53faaa07-2104-4c78-98ba-3f2f2c1926d7","Type":"ContainerStarted","Data":"8418ddc980b5b490b168e615b996b3b3baa5d93df3d598cef9398a441f9533d0"} Oct 08 18:29:08 crc kubenswrapper[4988]: I1008 18:29:08.672799 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-jppx4" event={"ID":"53faaa07-2104-4c78-98ba-3f2f2c1926d7","Type":"ContainerStarted","Data":"b424aa85db76d4be2f705b3d817e74158ee600961c861b45fb7a9ad847113399"} Oct 08 18:29:08 crc kubenswrapper[4988]: I1008 18:29:08.678984 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8fed08cf-45cb-4fd2-92af-9f30513c0b72","Type":"ContainerStarted","Data":"1f49914f5ed3f7f2d69b1d22ee1fdd1b474e9f2eed793d6fdef303b6808ac310"} Oct 08 18:29:08 crc kubenswrapper[4988]: I1008 18:29:08.680992 4988 generic.go:334] "Generic (PLEG): container finished" podID="7a831457-b429-4339-b166-afed001c770d" containerID="190609c635be9b56a48a82610585cd3e788b25a29cb94dabca71afefcecfec2f" exitCode=0 Oct 08 18:29:08 crc kubenswrapper[4988]: I1008 18:29:08.681076 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-69894dfcd9-948wp" event={"ID":"7a831457-b429-4339-b166-afed001c770d","Type":"ContainerDied","Data":"190609c635be9b56a48a82610585cd3e788b25a29cb94dabca71afefcecfec2f"} Oct 08 18:29:08 crc kubenswrapper[4988]: I1008 18:29:08.681111 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-69894dfcd9-948wp" event={"ID":"7a831457-b429-4339-b166-afed001c770d","Type":"ContainerStarted","Data":"5f3365d57e8b47bf123e83a83e5da3290a4622c50a10edc5ad829153757d864f"} Oct 08 18:29:08 crc kubenswrapper[4988]: I1008 18:29:08.686199 4988 generic.go:334] "Generic (PLEG): container finished" podID="ec575519-d902-40eb-81e9-5b8a6b64f70a" containerID="fb8da288036e4014cbfe9544f15585bed48a77a66584a38e5517d88e38f76aa4" exitCode=0 Oct 08 18:29:08 crc kubenswrapper[4988]: I1008 18:29:08.686271 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-758b46c4c7-wgkvt" event={"ID":"ec575519-d902-40eb-81e9-5b8a6b64f70a","Type":"ContainerDied","Data":"fb8da288036e4014cbfe9544f15585bed48a77a66584a38e5517d88e38f76aa4"} Oct 08 18:29:08 crc kubenswrapper[4988]: I1008 18:29:08.702208 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-jppx4" podStartSLOduration=2.70217957 podStartE2EDuration="2.70217957s" podCreationTimestamp="2025-10-08 18:29:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 18:29:08.696712038 +0000 UTC m=+1094.146554828" watchObservedRunningTime="2025-10-08 18:29:08.70217957 +0000 UTC m=+1094.152022360" Oct 08 18:29:08 crc kubenswrapper[4988]: I1008 18:29:08.707541 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-gwkbl" event={"ID":"62b5df6a-6198-4608-8cd7-05bfa27bfd75","Type":"ContainerStarted","Data":"26787b378d94e6d2c658f69be8e812ee1da4291764bd1d265e245e5e6021c91b"} Oct 08 18:29:08 crc kubenswrapper[4988]: I1008 18:29:08.757662 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 08 18:29:09 crc kubenswrapper[4988]: I1008 18:29:09.027784 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-758b46c4c7-wgkvt" Oct 08 18:29:09 crc kubenswrapper[4988]: I1008 18:29:09.078976 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ec575519-d902-40eb-81e9-5b8a6b64f70a-ovsdbserver-nb\") pod \"ec575519-d902-40eb-81e9-5b8a6b64f70a\" (UID: \"ec575519-d902-40eb-81e9-5b8a6b64f70a\") " Oct 08 18:29:09 crc kubenswrapper[4988]: I1008 18:29:09.079028 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-g9lg6\" (UniqueName: \"kubernetes.io/projected/ec575519-d902-40eb-81e9-5b8a6b64f70a-kube-api-access-g9lg6\") pod \"ec575519-d902-40eb-81e9-5b8a6b64f70a\" (UID: \"ec575519-d902-40eb-81e9-5b8a6b64f70a\") " Oct 08 18:29:09 crc kubenswrapper[4988]: I1008 18:29:09.079129 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/ec575519-d902-40eb-81e9-5b8a6b64f70a-dns-swift-storage-0\") pod \"ec575519-d902-40eb-81e9-5b8a6b64f70a\" (UID: \"ec575519-d902-40eb-81e9-5b8a6b64f70a\") " Oct 08 18:29:09 crc kubenswrapper[4988]: I1008 18:29:09.079286 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ec575519-d902-40eb-81e9-5b8a6b64f70a-dns-svc\") pod \"ec575519-d902-40eb-81e9-5b8a6b64f70a\" (UID: \"ec575519-d902-40eb-81e9-5b8a6b64f70a\") " Oct 08 18:29:09 crc kubenswrapper[4988]: I1008 18:29:09.079366 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ec575519-d902-40eb-81e9-5b8a6b64f70a-ovsdbserver-sb\") pod \"ec575519-d902-40eb-81e9-5b8a6b64f70a\" (UID: \"ec575519-d902-40eb-81e9-5b8a6b64f70a\") " Oct 08 18:29:09 crc kubenswrapper[4988]: I1008 18:29:09.079477 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ec575519-d902-40eb-81e9-5b8a6b64f70a-config\") pod \"ec575519-d902-40eb-81e9-5b8a6b64f70a\" (UID: \"ec575519-d902-40eb-81e9-5b8a6b64f70a\") " Oct 08 18:29:09 crc kubenswrapper[4988]: I1008 18:29:09.084263 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ec575519-d902-40eb-81e9-5b8a6b64f70a-kube-api-access-g9lg6" (OuterVolumeSpecName: "kube-api-access-g9lg6") pod "ec575519-d902-40eb-81e9-5b8a6b64f70a" (UID: "ec575519-d902-40eb-81e9-5b8a6b64f70a"). InnerVolumeSpecName "kube-api-access-g9lg6". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:29:09 crc kubenswrapper[4988]: I1008 18:29:09.101921 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ec575519-d902-40eb-81e9-5b8a6b64f70a-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "ec575519-d902-40eb-81e9-5b8a6b64f70a" (UID: "ec575519-d902-40eb-81e9-5b8a6b64f70a"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 18:29:09 crc kubenswrapper[4988]: I1008 18:29:09.101956 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ec575519-d902-40eb-81e9-5b8a6b64f70a-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "ec575519-d902-40eb-81e9-5b8a6b64f70a" (UID: "ec575519-d902-40eb-81e9-5b8a6b64f70a"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 18:29:09 crc kubenswrapper[4988]: I1008 18:29:09.105689 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ec575519-d902-40eb-81e9-5b8a6b64f70a-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "ec575519-d902-40eb-81e9-5b8a6b64f70a" (UID: "ec575519-d902-40eb-81e9-5b8a6b64f70a"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 18:29:09 crc kubenswrapper[4988]: I1008 18:29:09.109217 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ec575519-d902-40eb-81e9-5b8a6b64f70a-config" (OuterVolumeSpecName: "config") pod "ec575519-d902-40eb-81e9-5b8a6b64f70a" (UID: "ec575519-d902-40eb-81e9-5b8a6b64f70a"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 18:29:09 crc kubenswrapper[4988]: I1008 18:29:09.115883 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ec575519-d902-40eb-81e9-5b8a6b64f70a-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "ec575519-d902-40eb-81e9-5b8a6b64f70a" (UID: "ec575519-d902-40eb-81e9-5b8a6b64f70a"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 18:29:09 crc kubenswrapper[4988]: I1008 18:29:09.181124 4988 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ec575519-d902-40eb-81e9-5b8a6b64f70a-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 08 18:29:09 crc kubenswrapper[4988]: I1008 18:29:09.181156 4988 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-g9lg6\" (UniqueName: \"kubernetes.io/projected/ec575519-d902-40eb-81e9-5b8a6b64f70a-kube-api-access-g9lg6\") on node \"crc\" DevicePath \"\"" Oct 08 18:29:09 crc kubenswrapper[4988]: I1008 18:29:09.181166 4988 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/ec575519-d902-40eb-81e9-5b8a6b64f70a-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Oct 08 18:29:09 crc kubenswrapper[4988]: I1008 18:29:09.181176 4988 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ec575519-d902-40eb-81e9-5b8a6b64f70a-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 08 18:29:09 crc kubenswrapper[4988]: I1008 18:29:09.181184 4988 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ec575519-d902-40eb-81e9-5b8a6b64f70a-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 08 18:29:09 crc kubenswrapper[4988]: I1008 18:29:09.181192 4988 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ec575519-d902-40eb-81e9-5b8a6b64f70a-config\") on node \"crc\" DevicePath \"\"" Oct 08 18:29:09 crc kubenswrapper[4988]: I1008 18:29:09.717928 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-758b46c4c7-wgkvt" event={"ID":"ec575519-d902-40eb-81e9-5b8a6b64f70a","Type":"ContainerDied","Data":"761f4d3bfde0172eb85a2cbb06479cbba368bb82727503d11adfe807d3da434b"} Oct 08 18:29:09 crc kubenswrapper[4988]: I1008 18:29:09.718242 4988 scope.go:117] "RemoveContainer" containerID="fb8da288036e4014cbfe9544f15585bed48a77a66584a38e5517d88e38f76aa4" Oct 08 18:29:09 crc kubenswrapper[4988]: I1008 18:29:09.718348 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-758b46c4c7-wgkvt" Oct 08 18:29:09 crc kubenswrapper[4988]: I1008 18:29:09.727169 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-5v5z9" event={"ID":"c1f2511c-e7bc-4433-910b-eb3b94fcf09b","Type":"ContainerStarted","Data":"e0c58c7d3f2cbaa16eb7f1b22d366459ec6d6b096a0336a9554a9d97a33841f1"} Oct 08 18:29:09 crc kubenswrapper[4988]: I1008 18:29:09.732084 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-69894dfcd9-948wp" event={"ID":"7a831457-b429-4339-b166-afed001c770d","Type":"ContainerStarted","Data":"ac3f09e0a805492ad71b338aa6dd023325b8ab08859f1ce9590b691ca0ede0e9"} Oct 08 18:29:09 crc kubenswrapper[4988]: I1008 18:29:09.732136 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-69894dfcd9-948wp" Oct 08 18:29:09 crc kubenswrapper[4988]: I1008 18:29:09.811845 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-758b46c4c7-wgkvt"] Oct 08 18:29:09 crc kubenswrapper[4988]: I1008 18:29:09.845511 4988 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-758b46c4c7-wgkvt"] Oct 08 18:29:09 crc kubenswrapper[4988]: I1008 18:29:09.849928 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-69894dfcd9-948wp" podStartSLOduration=2.849906961 podStartE2EDuration="2.849906961s" podCreationTimestamp="2025-10-08 18:29:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 18:29:09.787334503 +0000 UTC m=+1095.237177283" watchObservedRunningTime="2025-10-08 18:29:09.849906961 +0000 UTC m=+1095.299749731" Oct 08 18:29:09 crc kubenswrapper[4988]: I1008 18:29:09.858744 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-db-sync-5v5z9" podStartSLOduration=2.085643606 podStartE2EDuration="35.85873161s" podCreationTimestamp="2025-10-08 18:28:34 +0000 UTC" firstStartedPulling="2025-10-08 18:28:35.038142392 +0000 UTC m=+1060.487985162" lastFinishedPulling="2025-10-08 18:29:08.811230396 +0000 UTC m=+1094.261073166" observedRunningTime="2025-10-08 18:29:09.812618782 +0000 UTC m=+1095.262461552" watchObservedRunningTime="2025-10-08 18:29:09.85873161 +0000 UTC m=+1095.308574380" Oct 08 18:29:11 crc kubenswrapper[4988]: I1008 18:29:11.246688 4988 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ec575519-d902-40eb-81e9-5b8a6b64f70a" path="/var/lib/kubelet/pods/ec575519-d902-40eb-81e9-5b8a6b64f70a/volumes" Oct 08 18:29:13 crc kubenswrapper[4988]: I1008 18:29:13.773922 4988 generic.go:334] "Generic (PLEG): container finished" podID="53faaa07-2104-4c78-98ba-3f2f2c1926d7" containerID="8418ddc980b5b490b168e615b996b3b3baa5d93df3d598cef9398a441f9533d0" exitCode=0 Oct 08 18:29:13 crc kubenswrapper[4988]: I1008 18:29:13.774109 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-jppx4" event={"ID":"53faaa07-2104-4c78-98ba-3f2f2c1926d7","Type":"ContainerDied","Data":"8418ddc980b5b490b168e615b996b3b3baa5d93df3d598cef9398a441f9533d0"} Oct 08 18:29:15 crc kubenswrapper[4988]: I1008 18:29:15.138629 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-jppx4" Oct 08 18:29:15 crc kubenswrapper[4988]: I1008 18:29:15.174578 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-32b1-account-create-tcnwc"] Oct 08 18:29:15 crc kubenswrapper[4988]: E1008 18:29:15.175021 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ec575519-d902-40eb-81e9-5b8a6b64f70a" containerName="init" Oct 08 18:29:15 crc kubenswrapper[4988]: I1008 18:29:15.175042 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="ec575519-d902-40eb-81e9-5b8a6b64f70a" containerName="init" Oct 08 18:29:15 crc kubenswrapper[4988]: E1008 18:29:15.175067 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="53faaa07-2104-4c78-98ba-3f2f2c1926d7" containerName="keystone-bootstrap" Oct 08 18:29:15 crc kubenswrapper[4988]: I1008 18:29:15.175074 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="53faaa07-2104-4c78-98ba-3f2f2c1926d7" containerName="keystone-bootstrap" Oct 08 18:29:15 crc kubenswrapper[4988]: I1008 18:29:15.175253 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="53faaa07-2104-4c78-98ba-3f2f2c1926d7" containerName="keystone-bootstrap" Oct 08 18:29:15 crc kubenswrapper[4988]: I1008 18:29:15.175282 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="ec575519-d902-40eb-81e9-5b8a6b64f70a" containerName="init" Oct 08 18:29:15 crc kubenswrapper[4988]: I1008 18:29:15.175849 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-32b1-account-create-tcnwc" Oct 08 18:29:15 crc kubenswrapper[4988]: I1008 18:29:15.176758 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s4cp8\" (UniqueName: \"kubernetes.io/projected/53faaa07-2104-4c78-98ba-3f2f2c1926d7-kube-api-access-s4cp8\") pod \"53faaa07-2104-4c78-98ba-3f2f2c1926d7\" (UID: \"53faaa07-2104-4c78-98ba-3f2f2c1926d7\") " Oct 08 18:29:15 crc kubenswrapper[4988]: I1008 18:29:15.176829 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/53faaa07-2104-4c78-98ba-3f2f2c1926d7-credential-keys\") pod \"53faaa07-2104-4c78-98ba-3f2f2c1926d7\" (UID: \"53faaa07-2104-4c78-98ba-3f2f2c1926d7\") " Oct 08 18:29:15 crc kubenswrapper[4988]: I1008 18:29:15.176970 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/53faaa07-2104-4c78-98ba-3f2f2c1926d7-scripts\") pod \"53faaa07-2104-4c78-98ba-3f2f2c1926d7\" (UID: \"53faaa07-2104-4c78-98ba-3f2f2c1926d7\") " Oct 08 18:29:15 crc kubenswrapper[4988]: I1008 18:29:15.177040 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/53faaa07-2104-4c78-98ba-3f2f2c1926d7-fernet-keys\") pod \"53faaa07-2104-4c78-98ba-3f2f2c1926d7\" (UID: \"53faaa07-2104-4c78-98ba-3f2f2c1926d7\") " Oct 08 18:29:15 crc kubenswrapper[4988]: I1008 18:29:15.177075 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/53faaa07-2104-4c78-98ba-3f2f2c1926d7-config-data\") pod \"53faaa07-2104-4c78-98ba-3f2f2c1926d7\" (UID: \"53faaa07-2104-4c78-98ba-3f2f2c1926d7\") " Oct 08 18:29:15 crc kubenswrapper[4988]: I1008 18:29:15.177113 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/53faaa07-2104-4c78-98ba-3f2f2c1926d7-combined-ca-bundle\") pod \"53faaa07-2104-4c78-98ba-3f2f2c1926d7\" (UID: \"53faaa07-2104-4c78-98ba-3f2f2c1926d7\") " Oct 08 18:29:15 crc kubenswrapper[4988]: I1008 18:29:15.177778 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-db-secret" Oct 08 18:29:15 crc kubenswrapper[4988]: I1008 18:29:15.182371 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-32b1-account-create-tcnwc"] Oct 08 18:29:15 crc kubenswrapper[4988]: I1008 18:29:15.182754 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/53faaa07-2104-4c78-98ba-3f2f2c1926d7-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "53faaa07-2104-4c78-98ba-3f2f2c1926d7" (UID: "53faaa07-2104-4c78-98ba-3f2f2c1926d7"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:29:15 crc kubenswrapper[4988]: I1008 18:29:15.182787 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/53faaa07-2104-4c78-98ba-3f2f2c1926d7-scripts" (OuterVolumeSpecName: "scripts") pod "53faaa07-2104-4c78-98ba-3f2f2c1926d7" (UID: "53faaa07-2104-4c78-98ba-3f2f2c1926d7"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:29:15 crc kubenswrapper[4988]: I1008 18:29:15.182865 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/53faaa07-2104-4c78-98ba-3f2f2c1926d7-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "53faaa07-2104-4c78-98ba-3f2f2c1926d7" (UID: "53faaa07-2104-4c78-98ba-3f2f2c1926d7"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:29:15 crc kubenswrapper[4988]: I1008 18:29:15.183437 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/53faaa07-2104-4c78-98ba-3f2f2c1926d7-kube-api-access-s4cp8" (OuterVolumeSpecName: "kube-api-access-s4cp8") pod "53faaa07-2104-4c78-98ba-3f2f2c1926d7" (UID: "53faaa07-2104-4c78-98ba-3f2f2c1926d7"). InnerVolumeSpecName "kube-api-access-s4cp8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:29:15 crc kubenswrapper[4988]: I1008 18:29:15.205651 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/53faaa07-2104-4c78-98ba-3f2f2c1926d7-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "53faaa07-2104-4c78-98ba-3f2f2c1926d7" (UID: "53faaa07-2104-4c78-98ba-3f2f2c1926d7"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:29:15 crc kubenswrapper[4988]: I1008 18:29:15.219691 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/53faaa07-2104-4c78-98ba-3f2f2c1926d7-config-data" (OuterVolumeSpecName: "config-data") pod "53faaa07-2104-4c78-98ba-3f2f2c1926d7" (UID: "53faaa07-2104-4c78-98ba-3f2f2c1926d7"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:29:15 crc kubenswrapper[4988]: I1008 18:29:15.280972 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cmbl9\" (UniqueName: \"kubernetes.io/projected/9b6f6c59-608b-4f5b-b4ef-3cb0e4c85345-kube-api-access-cmbl9\") pod \"barbican-32b1-account-create-tcnwc\" (UID: \"9b6f6c59-608b-4f5b-b4ef-3cb0e4c85345\") " pod="openstack/barbican-32b1-account-create-tcnwc" Oct 08 18:29:15 crc kubenswrapper[4988]: I1008 18:29:15.281418 4988 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/53faaa07-2104-4c78-98ba-3f2f2c1926d7-scripts\") on node \"crc\" DevicePath \"\"" Oct 08 18:29:15 crc kubenswrapper[4988]: I1008 18:29:15.281450 4988 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/53faaa07-2104-4c78-98ba-3f2f2c1926d7-fernet-keys\") on node \"crc\" DevicePath \"\"" Oct 08 18:29:15 crc kubenswrapper[4988]: I1008 18:29:15.281461 4988 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/53faaa07-2104-4c78-98ba-3f2f2c1926d7-config-data\") on node \"crc\" DevicePath \"\"" Oct 08 18:29:15 crc kubenswrapper[4988]: I1008 18:29:15.281469 4988 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/53faaa07-2104-4c78-98ba-3f2f2c1926d7-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 18:29:15 crc kubenswrapper[4988]: I1008 18:29:15.281482 4988 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s4cp8\" (UniqueName: \"kubernetes.io/projected/53faaa07-2104-4c78-98ba-3f2f2c1926d7-kube-api-access-s4cp8\") on node \"crc\" DevicePath \"\"" Oct 08 18:29:15 crc kubenswrapper[4988]: I1008 18:29:15.281491 4988 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/53faaa07-2104-4c78-98ba-3f2f2c1926d7-credential-keys\") on node \"crc\" DevicePath \"\"" Oct 08 18:29:15 crc kubenswrapper[4988]: I1008 18:29:15.366498 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-d8da-account-create-fxpc9"] Oct 08 18:29:15 crc kubenswrapper[4988]: I1008 18:29:15.367546 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-d8da-account-create-fxpc9" Oct 08 18:29:15 crc kubenswrapper[4988]: I1008 18:29:15.369704 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-db-secret" Oct 08 18:29:15 crc kubenswrapper[4988]: I1008 18:29:15.372458 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-d8da-account-create-fxpc9"] Oct 08 18:29:15 crc kubenswrapper[4988]: I1008 18:29:15.384766 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cmbl9\" (UniqueName: \"kubernetes.io/projected/9b6f6c59-608b-4f5b-b4ef-3cb0e4c85345-kube-api-access-cmbl9\") pod \"barbican-32b1-account-create-tcnwc\" (UID: \"9b6f6c59-608b-4f5b-b4ef-3cb0e4c85345\") " pod="openstack/barbican-32b1-account-create-tcnwc" Oct 08 18:29:15 crc kubenswrapper[4988]: I1008 18:29:15.404170 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cmbl9\" (UniqueName: \"kubernetes.io/projected/9b6f6c59-608b-4f5b-b4ef-3cb0e4c85345-kube-api-access-cmbl9\") pod \"barbican-32b1-account-create-tcnwc\" (UID: \"9b6f6c59-608b-4f5b-b4ef-3cb0e4c85345\") " pod="openstack/barbican-32b1-account-create-tcnwc" Oct 08 18:29:15 crc kubenswrapper[4988]: I1008 18:29:15.486087 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7xjqj\" (UniqueName: \"kubernetes.io/projected/feb5b694-6e5c-4778-8adf-6e3d89c0c875-kube-api-access-7xjqj\") pod \"cinder-d8da-account-create-fxpc9\" (UID: \"feb5b694-6e5c-4778-8adf-6e3d89c0c875\") " pod="openstack/cinder-d8da-account-create-fxpc9" Oct 08 18:29:15 crc kubenswrapper[4988]: I1008 18:29:15.494480 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-32b1-account-create-tcnwc" Oct 08 18:29:15 crc kubenswrapper[4988]: I1008 18:29:15.573295 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-3849-account-create-87mb6"] Oct 08 18:29:15 crc kubenswrapper[4988]: I1008 18:29:15.575846 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-3849-account-create-87mb6" Oct 08 18:29:15 crc kubenswrapper[4988]: I1008 18:29:15.581602 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-db-secret" Oct 08 18:29:15 crc kubenswrapper[4988]: I1008 18:29:15.595484 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7xjqj\" (UniqueName: \"kubernetes.io/projected/feb5b694-6e5c-4778-8adf-6e3d89c0c875-kube-api-access-7xjqj\") pod \"cinder-d8da-account-create-fxpc9\" (UID: \"feb5b694-6e5c-4778-8adf-6e3d89c0c875\") " pod="openstack/cinder-d8da-account-create-fxpc9" Oct 08 18:29:15 crc kubenswrapper[4988]: I1008 18:29:15.609825 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-3849-account-create-87mb6"] Oct 08 18:29:15 crc kubenswrapper[4988]: I1008 18:29:15.616684 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7xjqj\" (UniqueName: \"kubernetes.io/projected/feb5b694-6e5c-4778-8adf-6e3d89c0c875-kube-api-access-7xjqj\") pod \"cinder-d8da-account-create-fxpc9\" (UID: \"feb5b694-6e5c-4778-8adf-6e3d89c0c875\") " pod="openstack/cinder-d8da-account-create-fxpc9" Oct 08 18:29:15 crc kubenswrapper[4988]: I1008 18:29:15.685033 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-d8da-account-create-fxpc9" Oct 08 18:29:15 crc kubenswrapper[4988]: I1008 18:29:15.697508 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hm5gl\" (UniqueName: \"kubernetes.io/projected/8f236c17-7f75-40a3-9937-111f670a1aeb-kube-api-access-hm5gl\") pod \"neutron-3849-account-create-87mb6\" (UID: \"8f236c17-7f75-40a3-9937-111f670a1aeb\") " pod="openstack/neutron-3849-account-create-87mb6" Oct 08 18:29:15 crc kubenswrapper[4988]: I1008 18:29:15.800516 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hm5gl\" (UniqueName: \"kubernetes.io/projected/8f236c17-7f75-40a3-9937-111f670a1aeb-kube-api-access-hm5gl\") pod \"neutron-3849-account-create-87mb6\" (UID: \"8f236c17-7f75-40a3-9937-111f670a1aeb\") " pod="openstack/neutron-3849-account-create-87mb6" Oct 08 18:29:15 crc kubenswrapper[4988]: I1008 18:29:15.843021 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hm5gl\" (UniqueName: \"kubernetes.io/projected/8f236c17-7f75-40a3-9937-111f670a1aeb-kube-api-access-hm5gl\") pod \"neutron-3849-account-create-87mb6\" (UID: \"8f236c17-7f75-40a3-9937-111f670a1aeb\") " pod="openstack/neutron-3849-account-create-87mb6" Oct 08 18:29:15 crc kubenswrapper[4988]: I1008 18:29:15.851509 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-gwkbl" event={"ID":"62b5df6a-6198-4608-8cd7-05bfa27bfd75","Type":"ContainerStarted","Data":"46ab4b0ce428b8f21cfe1456c1a0cac36aef5111febc6ab77c67ce8016a55c70"} Oct 08 18:29:15 crc kubenswrapper[4988]: I1008 18:29:15.855337 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-jppx4" event={"ID":"53faaa07-2104-4c78-98ba-3f2f2c1926d7","Type":"ContainerDied","Data":"b424aa85db76d4be2f705b3d817e74158ee600961c861b45fb7a9ad847113399"} Oct 08 18:29:15 crc kubenswrapper[4988]: I1008 18:29:15.855526 4988 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b424aa85db76d4be2f705b3d817e74158ee600961c861b45fb7a9ad847113399" Oct 08 18:29:15 crc kubenswrapper[4988]: I1008 18:29:15.855754 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-jppx4" Oct 08 18:29:15 crc kubenswrapper[4988]: I1008 18:29:15.875046 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8fed08cf-45cb-4fd2-92af-9f30513c0b72","Type":"ContainerStarted","Data":"530180f4bca5eaf8c14c742c890bbe065b3ab6a329ac14ec714a341c4b5e100e"} Oct 08 18:29:15 crc kubenswrapper[4988]: I1008 18:29:15.885990 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-jppx4"] Oct 08 18:29:15 crc kubenswrapper[4988]: I1008 18:29:15.892136 4988 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-jppx4"] Oct 08 18:29:15 crc kubenswrapper[4988]: I1008 18:29:15.895008 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-db-sync-gwkbl" podStartSLOduration=2.428975729 podStartE2EDuration="8.894990808s" podCreationTimestamp="2025-10-08 18:29:07 +0000 UTC" firstStartedPulling="2025-10-08 18:29:08.288836978 +0000 UTC m=+1093.738679748" lastFinishedPulling="2025-10-08 18:29:14.754852047 +0000 UTC m=+1100.204694827" observedRunningTime="2025-10-08 18:29:15.881078978 +0000 UTC m=+1101.330921748" watchObservedRunningTime="2025-10-08 18:29:15.894990808 +0000 UTC m=+1101.344833578" Oct 08 18:29:15 crc kubenswrapper[4988]: I1008 18:29:15.906673 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-3849-account-create-87mb6" Oct 08 18:29:15 crc kubenswrapper[4988]: I1008 18:29:15.957256 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-748qj"] Oct 08 18:29:15 crc kubenswrapper[4988]: I1008 18:29:15.958482 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-748qj" Oct 08 18:29:15 crc kubenswrapper[4988]: I1008 18:29:15.963224 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Oct 08 18:29:15 crc kubenswrapper[4988]: I1008 18:29:15.963588 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-srkzn" Oct 08 18:29:15 crc kubenswrapper[4988]: I1008 18:29:15.963811 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Oct 08 18:29:15 crc kubenswrapper[4988]: I1008 18:29:15.963849 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Oct 08 18:29:15 crc kubenswrapper[4988]: I1008 18:29:15.976257 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-32b1-account-create-tcnwc"] Oct 08 18:29:15 crc kubenswrapper[4988]: I1008 18:29:15.986052 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-748qj"] Oct 08 18:29:16 crc kubenswrapper[4988]: I1008 18:29:16.008789 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b2e4e6bf-c872-4a45-be09-08decca63284-scripts\") pod \"keystone-bootstrap-748qj\" (UID: \"b2e4e6bf-c872-4a45-be09-08decca63284\") " pod="openstack/keystone-bootstrap-748qj" Oct 08 18:29:16 crc kubenswrapper[4988]: I1008 18:29:16.008990 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mlqtz\" (UniqueName: \"kubernetes.io/projected/b2e4e6bf-c872-4a45-be09-08decca63284-kube-api-access-mlqtz\") pod \"keystone-bootstrap-748qj\" (UID: \"b2e4e6bf-c872-4a45-be09-08decca63284\") " pod="openstack/keystone-bootstrap-748qj" Oct 08 18:29:16 crc kubenswrapper[4988]: I1008 18:29:16.009106 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/b2e4e6bf-c872-4a45-be09-08decca63284-credential-keys\") pod \"keystone-bootstrap-748qj\" (UID: \"b2e4e6bf-c872-4a45-be09-08decca63284\") " pod="openstack/keystone-bootstrap-748qj" Oct 08 18:29:16 crc kubenswrapper[4988]: I1008 18:29:16.009135 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/b2e4e6bf-c872-4a45-be09-08decca63284-fernet-keys\") pod \"keystone-bootstrap-748qj\" (UID: \"b2e4e6bf-c872-4a45-be09-08decca63284\") " pod="openstack/keystone-bootstrap-748qj" Oct 08 18:29:16 crc kubenswrapper[4988]: I1008 18:29:16.009207 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b2e4e6bf-c872-4a45-be09-08decca63284-config-data\") pod \"keystone-bootstrap-748qj\" (UID: \"b2e4e6bf-c872-4a45-be09-08decca63284\") " pod="openstack/keystone-bootstrap-748qj" Oct 08 18:29:16 crc kubenswrapper[4988]: I1008 18:29:16.009513 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b2e4e6bf-c872-4a45-be09-08decca63284-combined-ca-bundle\") pod \"keystone-bootstrap-748qj\" (UID: \"b2e4e6bf-c872-4a45-be09-08decca63284\") " pod="openstack/keystone-bootstrap-748qj" Oct 08 18:29:16 crc kubenswrapper[4988]: I1008 18:29:16.111018 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b2e4e6bf-c872-4a45-be09-08decca63284-config-data\") pod \"keystone-bootstrap-748qj\" (UID: \"b2e4e6bf-c872-4a45-be09-08decca63284\") " pod="openstack/keystone-bootstrap-748qj" Oct 08 18:29:16 crc kubenswrapper[4988]: I1008 18:29:16.111359 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b2e4e6bf-c872-4a45-be09-08decca63284-combined-ca-bundle\") pod \"keystone-bootstrap-748qj\" (UID: \"b2e4e6bf-c872-4a45-be09-08decca63284\") " pod="openstack/keystone-bootstrap-748qj" Oct 08 18:29:16 crc kubenswrapper[4988]: I1008 18:29:16.111416 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b2e4e6bf-c872-4a45-be09-08decca63284-scripts\") pod \"keystone-bootstrap-748qj\" (UID: \"b2e4e6bf-c872-4a45-be09-08decca63284\") " pod="openstack/keystone-bootstrap-748qj" Oct 08 18:29:16 crc kubenswrapper[4988]: I1008 18:29:16.111468 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mlqtz\" (UniqueName: \"kubernetes.io/projected/b2e4e6bf-c872-4a45-be09-08decca63284-kube-api-access-mlqtz\") pod \"keystone-bootstrap-748qj\" (UID: \"b2e4e6bf-c872-4a45-be09-08decca63284\") " pod="openstack/keystone-bootstrap-748qj" Oct 08 18:29:16 crc kubenswrapper[4988]: I1008 18:29:16.111499 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/b2e4e6bf-c872-4a45-be09-08decca63284-credential-keys\") pod \"keystone-bootstrap-748qj\" (UID: \"b2e4e6bf-c872-4a45-be09-08decca63284\") " pod="openstack/keystone-bootstrap-748qj" Oct 08 18:29:16 crc kubenswrapper[4988]: I1008 18:29:16.111518 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/b2e4e6bf-c872-4a45-be09-08decca63284-fernet-keys\") pod \"keystone-bootstrap-748qj\" (UID: \"b2e4e6bf-c872-4a45-be09-08decca63284\") " pod="openstack/keystone-bootstrap-748qj" Oct 08 18:29:16 crc kubenswrapper[4988]: I1008 18:29:16.120814 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/b2e4e6bf-c872-4a45-be09-08decca63284-credential-keys\") pod \"keystone-bootstrap-748qj\" (UID: \"b2e4e6bf-c872-4a45-be09-08decca63284\") " pod="openstack/keystone-bootstrap-748qj" Oct 08 18:29:16 crc kubenswrapper[4988]: I1008 18:29:16.121637 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b2e4e6bf-c872-4a45-be09-08decca63284-config-data\") pod \"keystone-bootstrap-748qj\" (UID: \"b2e4e6bf-c872-4a45-be09-08decca63284\") " pod="openstack/keystone-bootstrap-748qj" Oct 08 18:29:16 crc kubenswrapper[4988]: I1008 18:29:16.122333 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/b2e4e6bf-c872-4a45-be09-08decca63284-fernet-keys\") pod \"keystone-bootstrap-748qj\" (UID: \"b2e4e6bf-c872-4a45-be09-08decca63284\") " pod="openstack/keystone-bootstrap-748qj" Oct 08 18:29:16 crc kubenswrapper[4988]: I1008 18:29:16.123174 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b2e4e6bf-c872-4a45-be09-08decca63284-combined-ca-bundle\") pod \"keystone-bootstrap-748qj\" (UID: \"b2e4e6bf-c872-4a45-be09-08decca63284\") " pod="openstack/keystone-bootstrap-748qj" Oct 08 18:29:16 crc kubenswrapper[4988]: I1008 18:29:16.125822 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b2e4e6bf-c872-4a45-be09-08decca63284-scripts\") pod \"keystone-bootstrap-748qj\" (UID: \"b2e4e6bf-c872-4a45-be09-08decca63284\") " pod="openstack/keystone-bootstrap-748qj" Oct 08 18:29:16 crc kubenswrapper[4988]: I1008 18:29:16.129590 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-d8da-account-create-fxpc9"] Oct 08 18:29:16 crc kubenswrapper[4988]: I1008 18:29:16.134022 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mlqtz\" (UniqueName: \"kubernetes.io/projected/b2e4e6bf-c872-4a45-be09-08decca63284-kube-api-access-mlqtz\") pod \"keystone-bootstrap-748qj\" (UID: \"b2e4e6bf-c872-4a45-be09-08decca63284\") " pod="openstack/keystone-bootstrap-748qj" Oct 08 18:29:16 crc kubenswrapper[4988]: I1008 18:29:16.373109 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-3849-account-create-87mb6"] Oct 08 18:29:16 crc kubenswrapper[4988]: I1008 18:29:16.391149 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-748qj" Oct 08 18:29:16 crc kubenswrapper[4988]: I1008 18:29:16.886290 4988 generic.go:334] "Generic (PLEG): container finished" podID="9b6f6c59-608b-4f5b-b4ef-3cb0e4c85345" containerID="f3dcb22d4b89fb2ad16ba29f989be77152f28df772bd76ecfc8b14528c658bd2" exitCode=0 Oct 08 18:29:16 crc kubenswrapper[4988]: I1008 18:29:16.886361 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-32b1-account-create-tcnwc" event={"ID":"9b6f6c59-608b-4f5b-b4ef-3cb0e4c85345","Type":"ContainerDied","Data":"f3dcb22d4b89fb2ad16ba29f989be77152f28df772bd76ecfc8b14528c658bd2"} Oct 08 18:29:16 crc kubenswrapper[4988]: I1008 18:29:16.886589 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-32b1-account-create-tcnwc" event={"ID":"9b6f6c59-608b-4f5b-b4ef-3cb0e4c85345","Type":"ContainerStarted","Data":"4b543d463d7167f9afe9e4a440323bfef16f104082805c25032f5230e102e04e"} Oct 08 18:29:16 crc kubenswrapper[4988]: I1008 18:29:16.888638 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-d8da-account-create-fxpc9" event={"ID":"feb5b694-6e5c-4778-8adf-6e3d89c0c875","Type":"ContainerStarted","Data":"ee38f29c71febdb6cdc5a350ab90d5ad4cfc0fbe576a87a90cc2de98e24cac3d"} Oct 08 18:29:16 crc kubenswrapper[4988]: I1008 18:29:16.888663 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-d8da-account-create-fxpc9" event={"ID":"feb5b694-6e5c-4778-8adf-6e3d89c0c875","Type":"ContainerStarted","Data":"9ca280b030633cb8ae7fcfbfd072834eed11887ce6f909288392ee9258f1ef0c"} Oct 08 18:29:16 crc kubenswrapper[4988]: I1008 18:29:16.930159 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-d8da-account-create-fxpc9" podStartSLOduration=1.930141861 podStartE2EDuration="1.930141861s" podCreationTimestamp="2025-10-08 18:29:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 18:29:16.928008203 +0000 UTC m=+1102.377850993" watchObservedRunningTime="2025-10-08 18:29:16.930141861 +0000 UTC m=+1102.379984631" Oct 08 18:29:17 crc kubenswrapper[4988]: I1008 18:29:17.254843 4988 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="53faaa07-2104-4c78-98ba-3f2f2c1926d7" path="/var/lib/kubelet/pods/53faaa07-2104-4c78-98ba-3f2f2c1926d7/volumes" Oct 08 18:29:17 crc kubenswrapper[4988]: I1008 18:29:17.715333 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-69894dfcd9-948wp" Oct 08 18:29:17 crc kubenswrapper[4988]: I1008 18:29:17.793855 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-59f45f6cf7-9snqd"] Oct 08 18:29:17 crc kubenswrapper[4988]: I1008 18:29:17.794370 4988 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-59f45f6cf7-9snqd" podUID="095497e5-73fc-42a2-9db3-cdec5010b477" containerName="dnsmasq-dns" containerID="cri-o://d5fab394a095c11a56d6b6a9c80c90f7c9685834c5518456a3c18aac02301933" gracePeriod=10 Oct 08 18:29:17 crc kubenswrapper[4988]: I1008 18:29:17.837918 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-748qj"] Oct 08 18:29:17 crc kubenswrapper[4988]: I1008 18:29:17.902029 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-748qj" event={"ID":"b2e4e6bf-c872-4a45-be09-08decca63284","Type":"ContainerStarted","Data":"f22a00bcf752505f07c237044ba6fc9fb6deaeef7af5229a18cbd699754119e3"} Oct 08 18:29:17 crc kubenswrapper[4988]: I1008 18:29:17.907865 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8fed08cf-45cb-4fd2-92af-9f30513c0b72","Type":"ContainerStarted","Data":"2b5c761d77af17eace93d9cd355762d47e5b9791b61461447ac5dfb3542405b0"} Oct 08 18:29:17 crc kubenswrapper[4988]: I1008 18:29:17.911187 4988 generic.go:334] "Generic (PLEG): container finished" podID="feb5b694-6e5c-4778-8adf-6e3d89c0c875" containerID="ee38f29c71febdb6cdc5a350ab90d5ad4cfc0fbe576a87a90cc2de98e24cac3d" exitCode=0 Oct 08 18:29:17 crc kubenswrapper[4988]: I1008 18:29:17.911256 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-d8da-account-create-fxpc9" event={"ID":"feb5b694-6e5c-4778-8adf-6e3d89c0c875","Type":"ContainerDied","Data":"ee38f29c71febdb6cdc5a350ab90d5ad4cfc0fbe576a87a90cc2de98e24cac3d"} Oct 08 18:29:17 crc kubenswrapper[4988]: I1008 18:29:17.913765 4988 generic.go:334] "Generic (PLEG): container finished" podID="8f236c17-7f75-40a3-9937-111f670a1aeb" containerID="e634b0ebf42763a7900bb4a166f36326fba120e119d166d4e1c0843990c5daf2" exitCode=0 Oct 08 18:29:17 crc kubenswrapper[4988]: I1008 18:29:17.913996 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-3849-account-create-87mb6" event={"ID":"8f236c17-7f75-40a3-9937-111f670a1aeb","Type":"ContainerDied","Data":"e634b0ebf42763a7900bb4a166f36326fba120e119d166d4e1c0843990c5daf2"} Oct 08 18:29:17 crc kubenswrapper[4988]: I1008 18:29:17.914024 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-3849-account-create-87mb6" event={"ID":"8f236c17-7f75-40a3-9937-111f670a1aeb","Type":"ContainerStarted","Data":"31ea2f941a4e44b875c68438048b78a57eb5bda424745f1a726490e641f6729f"} Oct 08 18:29:18 crc kubenswrapper[4988]: I1008 18:29:18.303426 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-59f45f6cf7-9snqd" Oct 08 18:29:18 crc kubenswrapper[4988]: I1008 18:29:18.335041 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-32b1-account-create-tcnwc" Oct 08 18:29:18 crc kubenswrapper[4988]: I1008 18:29:18.362281 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tkmh8\" (UniqueName: \"kubernetes.io/projected/095497e5-73fc-42a2-9db3-cdec5010b477-kube-api-access-tkmh8\") pod \"095497e5-73fc-42a2-9db3-cdec5010b477\" (UID: \"095497e5-73fc-42a2-9db3-cdec5010b477\") " Oct 08 18:29:18 crc kubenswrapper[4988]: I1008 18:29:18.362359 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/095497e5-73fc-42a2-9db3-cdec5010b477-dns-swift-storage-0\") pod \"095497e5-73fc-42a2-9db3-cdec5010b477\" (UID: \"095497e5-73fc-42a2-9db3-cdec5010b477\") " Oct 08 18:29:18 crc kubenswrapper[4988]: I1008 18:29:18.362422 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/095497e5-73fc-42a2-9db3-cdec5010b477-config\") pod \"095497e5-73fc-42a2-9db3-cdec5010b477\" (UID: \"095497e5-73fc-42a2-9db3-cdec5010b477\") " Oct 08 18:29:18 crc kubenswrapper[4988]: I1008 18:29:18.362481 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/095497e5-73fc-42a2-9db3-cdec5010b477-ovsdbserver-sb\") pod \"095497e5-73fc-42a2-9db3-cdec5010b477\" (UID: \"095497e5-73fc-42a2-9db3-cdec5010b477\") " Oct 08 18:29:18 crc kubenswrapper[4988]: I1008 18:29:18.362500 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/095497e5-73fc-42a2-9db3-cdec5010b477-dns-svc\") pod \"095497e5-73fc-42a2-9db3-cdec5010b477\" (UID: \"095497e5-73fc-42a2-9db3-cdec5010b477\") " Oct 08 18:29:18 crc kubenswrapper[4988]: I1008 18:29:18.362601 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/095497e5-73fc-42a2-9db3-cdec5010b477-ovsdbserver-nb\") pod \"095497e5-73fc-42a2-9db3-cdec5010b477\" (UID: \"095497e5-73fc-42a2-9db3-cdec5010b477\") " Oct 08 18:29:18 crc kubenswrapper[4988]: I1008 18:29:18.371422 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/095497e5-73fc-42a2-9db3-cdec5010b477-kube-api-access-tkmh8" (OuterVolumeSpecName: "kube-api-access-tkmh8") pod "095497e5-73fc-42a2-9db3-cdec5010b477" (UID: "095497e5-73fc-42a2-9db3-cdec5010b477"). InnerVolumeSpecName "kube-api-access-tkmh8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:29:18 crc kubenswrapper[4988]: I1008 18:29:18.412770 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/095497e5-73fc-42a2-9db3-cdec5010b477-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "095497e5-73fc-42a2-9db3-cdec5010b477" (UID: "095497e5-73fc-42a2-9db3-cdec5010b477"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 18:29:18 crc kubenswrapper[4988]: I1008 18:29:18.422878 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/095497e5-73fc-42a2-9db3-cdec5010b477-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "095497e5-73fc-42a2-9db3-cdec5010b477" (UID: "095497e5-73fc-42a2-9db3-cdec5010b477"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 18:29:18 crc kubenswrapper[4988]: I1008 18:29:18.438954 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/095497e5-73fc-42a2-9db3-cdec5010b477-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "095497e5-73fc-42a2-9db3-cdec5010b477" (UID: "095497e5-73fc-42a2-9db3-cdec5010b477"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 18:29:18 crc kubenswrapper[4988]: I1008 18:29:18.446012 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/095497e5-73fc-42a2-9db3-cdec5010b477-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "095497e5-73fc-42a2-9db3-cdec5010b477" (UID: "095497e5-73fc-42a2-9db3-cdec5010b477"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 18:29:18 crc kubenswrapper[4988]: I1008 18:29:18.446067 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/095497e5-73fc-42a2-9db3-cdec5010b477-config" (OuterVolumeSpecName: "config") pod "095497e5-73fc-42a2-9db3-cdec5010b477" (UID: "095497e5-73fc-42a2-9db3-cdec5010b477"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 18:29:18 crc kubenswrapper[4988]: I1008 18:29:18.463571 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cmbl9\" (UniqueName: \"kubernetes.io/projected/9b6f6c59-608b-4f5b-b4ef-3cb0e4c85345-kube-api-access-cmbl9\") pod \"9b6f6c59-608b-4f5b-b4ef-3cb0e4c85345\" (UID: \"9b6f6c59-608b-4f5b-b4ef-3cb0e4c85345\") " Oct 08 18:29:18 crc kubenswrapper[4988]: I1008 18:29:18.464018 4988 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/095497e5-73fc-42a2-9db3-cdec5010b477-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Oct 08 18:29:18 crc kubenswrapper[4988]: I1008 18:29:18.464036 4988 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/095497e5-73fc-42a2-9db3-cdec5010b477-config\") on node \"crc\" DevicePath \"\"" Oct 08 18:29:18 crc kubenswrapper[4988]: I1008 18:29:18.464045 4988 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/095497e5-73fc-42a2-9db3-cdec5010b477-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 08 18:29:18 crc kubenswrapper[4988]: I1008 18:29:18.464054 4988 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/095497e5-73fc-42a2-9db3-cdec5010b477-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 08 18:29:18 crc kubenswrapper[4988]: I1008 18:29:18.464063 4988 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/095497e5-73fc-42a2-9db3-cdec5010b477-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 08 18:29:18 crc kubenswrapper[4988]: I1008 18:29:18.464071 4988 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tkmh8\" (UniqueName: \"kubernetes.io/projected/095497e5-73fc-42a2-9db3-cdec5010b477-kube-api-access-tkmh8\") on node \"crc\" DevicePath \"\"" Oct 08 18:29:18 crc kubenswrapper[4988]: I1008 18:29:18.467541 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9b6f6c59-608b-4f5b-b4ef-3cb0e4c85345-kube-api-access-cmbl9" (OuterVolumeSpecName: "kube-api-access-cmbl9") pod "9b6f6c59-608b-4f5b-b4ef-3cb0e4c85345" (UID: "9b6f6c59-608b-4f5b-b4ef-3cb0e4c85345"). InnerVolumeSpecName "kube-api-access-cmbl9". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:29:18 crc kubenswrapper[4988]: I1008 18:29:18.566595 4988 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cmbl9\" (UniqueName: \"kubernetes.io/projected/9b6f6c59-608b-4f5b-b4ef-3cb0e4c85345-kube-api-access-cmbl9\") on node \"crc\" DevicePath \"\"" Oct 08 18:29:18 crc kubenswrapper[4988]: I1008 18:29:18.926031 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-32b1-account-create-tcnwc" event={"ID":"9b6f6c59-608b-4f5b-b4ef-3cb0e4c85345","Type":"ContainerDied","Data":"4b543d463d7167f9afe9e4a440323bfef16f104082805c25032f5230e102e04e"} Oct 08 18:29:18 crc kubenswrapper[4988]: I1008 18:29:18.926353 4988 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4b543d463d7167f9afe9e4a440323bfef16f104082805c25032f5230e102e04e" Oct 08 18:29:18 crc kubenswrapper[4988]: I1008 18:29:18.926042 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-32b1-account-create-tcnwc" Oct 08 18:29:18 crc kubenswrapper[4988]: I1008 18:29:18.928115 4988 generic.go:334] "Generic (PLEG): container finished" podID="095497e5-73fc-42a2-9db3-cdec5010b477" containerID="d5fab394a095c11a56d6b6a9c80c90f7c9685834c5518456a3c18aac02301933" exitCode=0 Oct 08 18:29:18 crc kubenswrapper[4988]: I1008 18:29:18.928170 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-59f45f6cf7-9snqd" Oct 08 18:29:18 crc kubenswrapper[4988]: I1008 18:29:18.928186 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-59f45f6cf7-9snqd" event={"ID":"095497e5-73fc-42a2-9db3-cdec5010b477","Type":"ContainerDied","Data":"d5fab394a095c11a56d6b6a9c80c90f7c9685834c5518456a3c18aac02301933"} Oct 08 18:29:18 crc kubenswrapper[4988]: I1008 18:29:18.928264 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-59f45f6cf7-9snqd" event={"ID":"095497e5-73fc-42a2-9db3-cdec5010b477","Type":"ContainerDied","Data":"25762dd24b39ed92b4a68132faab7bd2ba589caada1a13db4ca52b0a2c29037b"} Oct 08 18:29:18 crc kubenswrapper[4988]: I1008 18:29:18.928290 4988 scope.go:117] "RemoveContainer" containerID="d5fab394a095c11a56d6b6a9c80c90f7c9685834c5518456a3c18aac02301933" Oct 08 18:29:18 crc kubenswrapper[4988]: I1008 18:29:18.931988 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-748qj" event={"ID":"b2e4e6bf-c872-4a45-be09-08decca63284","Type":"ContainerStarted","Data":"c6a4733dbf41fe0410a80b6de0ff01b8123be53bae21bb88ca041b937dd29455"} Oct 08 18:29:18 crc kubenswrapper[4988]: I1008 18:29:18.935222 4988 generic.go:334] "Generic (PLEG): container finished" podID="62b5df6a-6198-4608-8cd7-05bfa27bfd75" containerID="46ab4b0ce428b8f21cfe1456c1a0cac36aef5111febc6ab77c67ce8016a55c70" exitCode=0 Oct 08 18:29:18 crc kubenswrapper[4988]: I1008 18:29:18.935280 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-gwkbl" event={"ID":"62b5df6a-6198-4608-8cd7-05bfa27bfd75","Type":"ContainerDied","Data":"46ab4b0ce428b8f21cfe1456c1a0cac36aef5111febc6ab77c67ce8016a55c70"} Oct 08 18:29:18 crc kubenswrapper[4988]: I1008 18:29:18.955575 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-748qj" podStartSLOduration=3.955553896 podStartE2EDuration="3.955553896s" podCreationTimestamp="2025-10-08 18:29:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 18:29:18.950973312 +0000 UTC m=+1104.400816082" watchObservedRunningTime="2025-10-08 18:29:18.955553896 +0000 UTC m=+1104.405396666" Oct 08 18:29:18 crc kubenswrapper[4988]: I1008 18:29:18.987029 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-59f45f6cf7-9snqd"] Oct 08 18:29:18 crc kubenswrapper[4988]: I1008 18:29:18.995523 4988 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-59f45f6cf7-9snqd"] Oct 08 18:29:19 crc kubenswrapper[4988]: I1008 18:29:19.099829 4988 scope.go:117] "RemoveContainer" containerID="84b53df54fd395ba6e0f9bca2e0ab63e769d5ef4812879f7d5885766e2fd8a3c" Oct 08 18:29:19 crc kubenswrapper[4988]: I1008 18:29:19.138400 4988 scope.go:117] "RemoveContainer" containerID="d5fab394a095c11a56d6b6a9c80c90f7c9685834c5518456a3c18aac02301933" Oct 08 18:29:19 crc kubenswrapper[4988]: E1008 18:29:19.138860 4988 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d5fab394a095c11a56d6b6a9c80c90f7c9685834c5518456a3c18aac02301933\": container with ID starting with d5fab394a095c11a56d6b6a9c80c90f7c9685834c5518456a3c18aac02301933 not found: ID does not exist" containerID="d5fab394a095c11a56d6b6a9c80c90f7c9685834c5518456a3c18aac02301933" Oct 08 18:29:19 crc kubenswrapper[4988]: I1008 18:29:19.138913 4988 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d5fab394a095c11a56d6b6a9c80c90f7c9685834c5518456a3c18aac02301933"} err="failed to get container status \"d5fab394a095c11a56d6b6a9c80c90f7c9685834c5518456a3c18aac02301933\": rpc error: code = NotFound desc = could not find container \"d5fab394a095c11a56d6b6a9c80c90f7c9685834c5518456a3c18aac02301933\": container with ID starting with d5fab394a095c11a56d6b6a9c80c90f7c9685834c5518456a3c18aac02301933 not found: ID does not exist" Oct 08 18:29:19 crc kubenswrapper[4988]: I1008 18:29:19.139305 4988 scope.go:117] "RemoveContainer" containerID="84b53df54fd395ba6e0f9bca2e0ab63e769d5ef4812879f7d5885766e2fd8a3c" Oct 08 18:29:19 crc kubenswrapper[4988]: E1008 18:29:19.139593 4988 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"84b53df54fd395ba6e0f9bca2e0ab63e769d5ef4812879f7d5885766e2fd8a3c\": container with ID starting with 84b53df54fd395ba6e0f9bca2e0ab63e769d5ef4812879f7d5885766e2fd8a3c not found: ID does not exist" containerID="84b53df54fd395ba6e0f9bca2e0ab63e769d5ef4812879f7d5885766e2fd8a3c" Oct 08 18:29:19 crc kubenswrapper[4988]: I1008 18:29:19.139619 4988 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"84b53df54fd395ba6e0f9bca2e0ab63e769d5ef4812879f7d5885766e2fd8a3c"} err="failed to get container status \"84b53df54fd395ba6e0f9bca2e0ab63e769d5ef4812879f7d5885766e2fd8a3c\": rpc error: code = NotFound desc = could not find container \"84b53df54fd395ba6e0f9bca2e0ab63e769d5ef4812879f7d5885766e2fd8a3c\": container with ID starting with 84b53df54fd395ba6e0f9bca2e0ab63e769d5ef4812879f7d5885766e2fd8a3c not found: ID does not exist" Oct 08 18:29:19 crc kubenswrapper[4988]: I1008 18:29:19.282411 4988 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="095497e5-73fc-42a2-9db3-cdec5010b477" path="/var/lib/kubelet/pods/095497e5-73fc-42a2-9db3-cdec5010b477/volumes" Oct 08 18:29:19 crc kubenswrapper[4988]: I1008 18:29:19.411642 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-d8da-account-create-fxpc9" Oct 08 18:29:19 crc kubenswrapper[4988]: I1008 18:29:19.456061 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-3849-account-create-87mb6" Oct 08 18:29:19 crc kubenswrapper[4988]: I1008 18:29:19.484972 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7xjqj\" (UniqueName: \"kubernetes.io/projected/feb5b694-6e5c-4778-8adf-6e3d89c0c875-kube-api-access-7xjqj\") pod \"feb5b694-6e5c-4778-8adf-6e3d89c0c875\" (UID: \"feb5b694-6e5c-4778-8adf-6e3d89c0c875\") " Oct 08 18:29:19 crc kubenswrapper[4988]: I1008 18:29:19.490653 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/feb5b694-6e5c-4778-8adf-6e3d89c0c875-kube-api-access-7xjqj" (OuterVolumeSpecName: "kube-api-access-7xjqj") pod "feb5b694-6e5c-4778-8adf-6e3d89c0c875" (UID: "feb5b694-6e5c-4778-8adf-6e3d89c0c875"). InnerVolumeSpecName "kube-api-access-7xjqj". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:29:19 crc kubenswrapper[4988]: I1008 18:29:19.587147 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hm5gl\" (UniqueName: \"kubernetes.io/projected/8f236c17-7f75-40a3-9937-111f670a1aeb-kube-api-access-hm5gl\") pod \"8f236c17-7f75-40a3-9937-111f670a1aeb\" (UID: \"8f236c17-7f75-40a3-9937-111f670a1aeb\") " Oct 08 18:29:19 crc kubenswrapper[4988]: I1008 18:29:19.588510 4988 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7xjqj\" (UniqueName: \"kubernetes.io/projected/feb5b694-6e5c-4778-8adf-6e3d89c0c875-kube-api-access-7xjqj\") on node \"crc\" DevicePath \"\"" Oct 08 18:29:19 crc kubenswrapper[4988]: I1008 18:29:19.591248 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f236c17-7f75-40a3-9937-111f670a1aeb-kube-api-access-hm5gl" (OuterVolumeSpecName: "kube-api-access-hm5gl") pod "8f236c17-7f75-40a3-9937-111f670a1aeb" (UID: "8f236c17-7f75-40a3-9937-111f670a1aeb"). InnerVolumeSpecName "kube-api-access-hm5gl". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:29:19 crc kubenswrapper[4988]: I1008 18:29:19.690812 4988 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hm5gl\" (UniqueName: \"kubernetes.io/projected/8f236c17-7f75-40a3-9937-111f670a1aeb-kube-api-access-hm5gl\") on node \"crc\" DevicePath \"\"" Oct 08 18:29:19 crc kubenswrapper[4988]: I1008 18:29:19.950757 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-3849-account-create-87mb6" event={"ID":"8f236c17-7f75-40a3-9937-111f670a1aeb","Type":"ContainerDied","Data":"31ea2f941a4e44b875c68438048b78a57eb5bda424745f1a726490e641f6729f"} Oct 08 18:29:19 crc kubenswrapper[4988]: I1008 18:29:19.959028 4988 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="31ea2f941a4e44b875c68438048b78a57eb5bda424745f1a726490e641f6729f" Oct 08 18:29:19 crc kubenswrapper[4988]: I1008 18:29:19.959063 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-d8da-account-create-fxpc9" event={"ID":"feb5b694-6e5c-4778-8adf-6e3d89c0c875","Type":"ContainerDied","Data":"9ca280b030633cb8ae7fcfbfd072834eed11887ce6f909288392ee9258f1ef0c"} Oct 08 18:29:19 crc kubenswrapper[4988]: I1008 18:29:19.950927 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-3849-account-create-87mb6" Oct 08 18:29:19 crc kubenswrapper[4988]: I1008 18:29:19.959088 4988 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9ca280b030633cb8ae7fcfbfd072834eed11887ce6f909288392ee9258f1ef0c" Oct 08 18:29:19 crc kubenswrapper[4988]: I1008 18:29:19.956680 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-d8da-account-create-fxpc9" Oct 08 18:29:20 crc kubenswrapper[4988]: I1008 18:29:20.318269 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-gwkbl" Oct 08 18:29:20 crc kubenswrapper[4988]: I1008 18:29:20.406851 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/62b5df6a-6198-4608-8cd7-05bfa27bfd75-scripts\") pod \"62b5df6a-6198-4608-8cd7-05bfa27bfd75\" (UID: \"62b5df6a-6198-4608-8cd7-05bfa27bfd75\") " Oct 08 18:29:20 crc kubenswrapper[4988]: I1008 18:29:20.406887 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/62b5df6a-6198-4608-8cd7-05bfa27bfd75-logs\") pod \"62b5df6a-6198-4608-8cd7-05bfa27bfd75\" (UID: \"62b5df6a-6198-4608-8cd7-05bfa27bfd75\") " Oct 08 18:29:20 crc kubenswrapper[4988]: I1008 18:29:20.406940 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/62b5df6a-6198-4608-8cd7-05bfa27bfd75-config-data\") pod \"62b5df6a-6198-4608-8cd7-05bfa27bfd75\" (UID: \"62b5df6a-6198-4608-8cd7-05bfa27bfd75\") " Oct 08 18:29:20 crc kubenswrapper[4988]: I1008 18:29:20.406993 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/62b5df6a-6198-4608-8cd7-05bfa27bfd75-combined-ca-bundle\") pod \"62b5df6a-6198-4608-8cd7-05bfa27bfd75\" (UID: \"62b5df6a-6198-4608-8cd7-05bfa27bfd75\") " Oct 08 18:29:20 crc kubenswrapper[4988]: I1008 18:29:20.407073 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zxtbt\" (UniqueName: \"kubernetes.io/projected/62b5df6a-6198-4608-8cd7-05bfa27bfd75-kube-api-access-zxtbt\") pod \"62b5df6a-6198-4608-8cd7-05bfa27bfd75\" (UID: \"62b5df6a-6198-4608-8cd7-05bfa27bfd75\") " Oct 08 18:29:20 crc kubenswrapper[4988]: I1008 18:29:20.408147 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/62b5df6a-6198-4608-8cd7-05bfa27bfd75-logs" (OuterVolumeSpecName: "logs") pod "62b5df6a-6198-4608-8cd7-05bfa27bfd75" (UID: "62b5df6a-6198-4608-8cd7-05bfa27bfd75"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 18:29:20 crc kubenswrapper[4988]: I1008 18:29:20.411564 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/62b5df6a-6198-4608-8cd7-05bfa27bfd75-scripts" (OuterVolumeSpecName: "scripts") pod "62b5df6a-6198-4608-8cd7-05bfa27bfd75" (UID: "62b5df6a-6198-4608-8cd7-05bfa27bfd75"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:29:20 crc kubenswrapper[4988]: I1008 18:29:20.411778 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/62b5df6a-6198-4608-8cd7-05bfa27bfd75-kube-api-access-zxtbt" (OuterVolumeSpecName: "kube-api-access-zxtbt") pod "62b5df6a-6198-4608-8cd7-05bfa27bfd75" (UID: "62b5df6a-6198-4608-8cd7-05bfa27bfd75"). InnerVolumeSpecName "kube-api-access-zxtbt". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:29:20 crc kubenswrapper[4988]: I1008 18:29:20.456855 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/62b5df6a-6198-4608-8cd7-05bfa27bfd75-config-data" (OuterVolumeSpecName: "config-data") pod "62b5df6a-6198-4608-8cd7-05bfa27bfd75" (UID: "62b5df6a-6198-4608-8cd7-05bfa27bfd75"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:29:20 crc kubenswrapper[4988]: I1008 18:29:20.457746 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/62b5df6a-6198-4608-8cd7-05bfa27bfd75-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "62b5df6a-6198-4608-8cd7-05bfa27bfd75" (UID: "62b5df6a-6198-4608-8cd7-05bfa27bfd75"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:29:20 crc kubenswrapper[4988]: I1008 18:29:20.493542 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-sync-dm56p"] Oct 08 18:29:20 crc kubenswrapper[4988]: E1008 18:29:20.493945 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="62b5df6a-6198-4608-8cd7-05bfa27bfd75" containerName="placement-db-sync" Oct 08 18:29:20 crc kubenswrapper[4988]: I1008 18:29:20.493958 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="62b5df6a-6198-4608-8cd7-05bfa27bfd75" containerName="placement-db-sync" Oct 08 18:29:20 crc kubenswrapper[4988]: E1008 18:29:20.493973 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="feb5b694-6e5c-4778-8adf-6e3d89c0c875" containerName="mariadb-account-create" Oct 08 18:29:20 crc kubenswrapper[4988]: I1008 18:29:20.493979 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="feb5b694-6e5c-4778-8adf-6e3d89c0c875" containerName="mariadb-account-create" Oct 08 18:29:20 crc kubenswrapper[4988]: E1008 18:29:20.493988 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8f236c17-7f75-40a3-9937-111f670a1aeb" containerName="mariadb-account-create" Oct 08 18:29:20 crc kubenswrapper[4988]: I1008 18:29:20.493994 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="8f236c17-7f75-40a3-9937-111f670a1aeb" containerName="mariadb-account-create" Oct 08 18:29:20 crc kubenswrapper[4988]: E1008 18:29:20.494027 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9b6f6c59-608b-4f5b-b4ef-3cb0e4c85345" containerName="mariadb-account-create" Oct 08 18:29:20 crc kubenswrapper[4988]: I1008 18:29:20.494033 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="9b6f6c59-608b-4f5b-b4ef-3cb0e4c85345" containerName="mariadb-account-create" Oct 08 18:29:20 crc kubenswrapper[4988]: E1008 18:29:20.494046 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="095497e5-73fc-42a2-9db3-cdec5010b477" containerName="init" Oct 08 18:29:20 crc kubenswrapper[4988]: I1008 18:29:20.494051 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="095497e5-73fc-42a2-9db3-cdec5010b477" containerName="init" Oct 08 18:29:20 crc kubenswrapper[4988]: E1008 18:29:20.494060 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="095497e5-73fc-42a2-9db3-cdec5010b477" containerName="dnsmasq-dns" Oct 08 18:29:20 crc kubenswrapper[4988]: I1008 18:29:20.494068 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="095497e5-73fc-42a2-9db3-cdec5010b477" containerName="dnsmasq-dns" Oct 08 18:29:20 crc kubenswrapper[4988]: I1008 18:29:20.494221 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="8f236c17-7f75-40a3-9937-111f670a1aeb" containerName="mariadb-account-create" Oct 08 18:29:20 crc kubenswrapper[4988]: I1008 18:29:20.494234 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="9b6f6c59-608b-4f5b-b4ef-3cb0e4c85345" containerName="mariadb-account-create" Oct 08 18:29:20 crc kubenswrapper[4988]: I1008 18:29:20.494249 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="095497e5-73fc-42a2-9db3-cdec5010b477" containerName="dnsmasq-dns" Oct 08 18:29:20 crc kubenswrapper[4988]: I1008 18:29:20.494258 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="feb5b694-6e5c-4778-8adf-6e3d89c0c875" containerName="mariadb-account-create" Oct 08 18:29:20 crc kubenswrapper[4988]: I1008 18:29:20.494271 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="62b5df6a-6198-4608-8cd7-05bfa27bfd75" containerName="placement-db-sync" Oct 08 18:29:20 crc kubenswrapper[4988]: I1008 18:29:20.494809 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-dm56p" Oct 08 18:29:20 crc kubenswrapper[4988]: I1008 18:29:20.499088 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-vxtzc" Oct 08 18:29:20 crc kubenswrapper[4988]: I1008 18:29:20.499939 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Oct 08 18:29:20 crc kubenswrapper[4988]: I1008 18:29:20.508883 4988 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/62b5df6a-6198-4608-8cd7-05bfa27bfd75-config-data\") on node \"crc\" DevicePath \"\"" Oct 08 18:29:20 crc kubenswrapper[4988]: I1008 18:29:20.508921 4988 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/62b5df6a-6198-4608-8cd7-05bfa27bfd75-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 18:29:20 crc kubenswrapper[4988]: I1008 18:29:20.508932 4988 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zxtbt\" (UniqueName: \"kubernetes.io/projected/62b5df6a-6198-4608-8cd7-05bfa27bfd75-kube-api-access-zxtbt\") on node \"crc\" DevicePath \"\"" Oct 08 18:29:20 crc kubenswrapper[4988]: I1008 18:29:20.508941 4988 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/62b5df6a-6198-4608-8cd7-05bfa27bfd75-scripts\") on node \"crc\" DevicePath \"\"" Oct 08 18:29:20 crc kubenswrapper[4988]: I1008 18:29:20.508951 4988 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/62b5df6a-6198-4608-8cd7-05bfa27bfd75-logs\") on node \"crc\" DevicePath \"\"" Oct 08 18:29:20 crc kubenswrapper[4988]: I1008 18:29:20.511079 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-dm56p"] Oct 08 18:29:20 crc kubenswrapper[4988]: I1008 18:29:20.610984 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/d7f4f7aa-ccf2-45e1-8d32-db9684853231-db-sync-config-data\") pod \"barbican-db-sync-dm56p\" (UID: \"d7f4f7aa-ccf2-45e1-8d32-db9684853231\") " pod="openstack/barbican-db-sync-dm56p" Oct 08 18:29:20 crc kubenswrapper[4988]: I1008 18:29:20.611055 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8q9bd\" (UniqueName: \"kubernetes.io/projected/d7f4f7aa-ccf2-45e1-8d32-db9684853231-kube-api-access-8q9bd\") pod \"barbican-db-sync-dm56p\" (UID: \"d7f4f7aa-ccf2-45e1-8d32-db9684853231\") " pod="openstack/barbican-db-sync-dm56p" Oct 08 18:29:20 crc kubenswrapper[4988]: I1008 18:29:20.611231 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d7f4f7aa-ccf2-45e1-8d32-db9684853231-combined-ca-bundle\") pod \"barbican-db-sync-dm56p\" (UID: \"d7f4f7aa-ccf2-45e1-8d32-db9684853231\") " pod="openstack/barbican-db-sync-dm56p" Oct 08 18:29:20 crc kubenswrapper[4988]: I1008 18:29:20.712854 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/d7f4f7aa-ccf2-45e1-8d32-db9684853231-db-sync-config-data\") pod \"barbican-db-sync-dm56p\" (UID: \"d7f4f7aa-ccf2-45e1-8d32-db9684853231\") " pod="openstack/barbican-db-sync-dm56p" Oct 08 18:29:20 crc kubenswrapper[4988]: I1008 18:29:20.712930 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8q9bd\" (UniqueName: \"kubernetes.io/projected/d7f4f7aa-ccf2-45e1-8d32-db9684853231-kube-api-access-8q9bd\") pod \"barbican-db-sync-dm56p\" (UID: \"d7f4f7aa-ccf2-45e1-8d32-db9684853231\") " pod="openstack/barbican-db-sync-dm56p" Oct 08 18:29:20 crc kubenswrapper[4988]: I1008 18:29:20.712980 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d7f4f7aa-ccf2-45e1-8d32-db9684853231-combined-ca-bundle\") pod \"barbican-db-sync-dm56p\" (UID: \"d7f4f7aa-ccf2-45e1-8d32-db9684853231\") " pod="openstack/barbican-db-sync-dm56p" Oct 08 18:29:20 crc kubenswrapper[4988]: I1008 18:29:20.717345 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d7f4f7aa-ccf2-45e1-8d32-db9684853231-combined-ca-bundle\") pod \"barbican-db-sync-dm56p\" (UID: \"d7f4f7aa-ccf2-45e1-8d32-db9684853231\") " pod="openstack/barbican-db-sync-dm56p" Oct 08 18:29:20 crc kubenswrapper[4988]: I1008 18:29:20.717345 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/d7f4f7aa-ccf2-45e1-8d32-db9684853231-db-sync-config-data\") pod \"barbican-db-sync-dm56p\" (UID: \"d7f4f7aa-ccf2-45e1-8d32-db9684853231\") " pod="openstack/barbican-db-sync-dm56p" Oct 08 18:29:20 crc kubenswrapper[4988]: I1008 18:29:20.729572 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8q9bd\" (UniqueName: \"kubernetes.io/projected/d7f4f7aa-ccf2-45e1-8d32-db9684853231-kube-api-access-8q9bd\") pod \"barbican-db-sync-dm56p\" (UID: \"d7f4f7aa-ccf2-45e1-8d32-db9684853231\") " pod="openstack/barbican-db-sync-dm56p" Oct 08 18:29:20 crc kubenswrapper[4988]: I1008 18:29:20.802724 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-sync-wvjsl"] Oct 08 18:29:20 crc kubenswrapper[4988]: I1008 18:29:20.804502 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-wvjsl" Oct 08 18:29:20 crc kubenswrapper[4988]: I1008 18:29:20.807048 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Oct 08 18:29:20 crc kubenswrapper[4988]: I1008 18:29:20.807214 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-n2kmm" Oct 08 18:29:20 crc kubenswrapper[4988]: I1008 18:29:20.807597 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Oct 08 18:29:20 crc kubenswrapper[4988]: I1008 18:29:20.826855 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-dm56p" Oct 08 18:29:20 crc kubenswrapper[4988]: I1008 18:29:20.827352 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-wvjsl"] Oct 08 18:29:20 crc kubenswrapper[4988]: I1008 18:29:20.917803 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7vbb4\" (UniqueName: \"kubernetes.io/projected/de747e73-c280-424f-af48-3135bec91b55-kube-api-access-7vbb4\") pod \"neutron-db-sync-wvjsl\" (UID: \"de747e73-c280-424f-af48-3135bec91b55\") " pod="openstack/neutron-db-sync-wvjsl" Oct 08 18:29:20 crc kubenswrapper[4988]: I1008 18:29:20.917882 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/de747e73-c280-424f-af48-3135bec91b55-combined-ca-bundle\") pod \"neutron-db-sync-wvjsl\" (UID: \"de747e73-c280-424f-af48-3135bec91b55\") " pod="openstack/neutron-db-sync-wvjsl" Oct 08 18:29:20 crc kubenswrapper[4988]: I1008 18:29:20.917915 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/de747e73-c280-424f-af48-3135bec91b55-config\") pod \"neutron-db-sync-wvjsl\" (UID: \"de747e73-c280-424f-af48-3135bec91b55\") " pod="openstack/neutron-db-sync-wvjsl" Oct 08 18:29:20 crc kubenswrapper[4988]: I1008 18:29:20.968444 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-gwkbl" event={"ID":"62b5df6a-6198-4608-8cd7-05bfa27bfd75","Type":"ContainerDied","Data":"26787b378d94e6d2c658f69be8e812ee1da4291764bd1d265e245e5e6021c91b"} Oct 08 18:29:20 crc kubenswrapper[4988]: I1008 18:29:20.968479 4988 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="26787b378d94e6d2c658f69be8e812ee1da4291764bd1d265e245e5e6021c91b" Oct 08 18:29:20 crc kubenswrapper[4988]: I1008 18:29:20.968542 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-gwkbl" Oct 08 18:29:21 crc kubenswrapper[4988]: I1008 18:29:21.019562 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7vbb4\" (UniqueName: \"kubernetes.io/projected/de747e73-c280-424f-af48-3135bec91b55-kube-api-access-7vbb4\") pod \"neutron-db-sync-wvjsl\" (UID: \"de747e73-c280-424f-af48-3135bec91b55\") " pod="openstack/neutron-db-sync-wvjsl" Oct 08 18:29:21 crc kubenswrapper[4988]: I1008 18:29:21.019632 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/de747e73-c280-424f-af48-3135bec91b55-combined-ca-bundle\") pod \"neutron-db-sync-wvjsl\" (UID: \"de747e73-c280-424f-af48-3135bec91b55\") " pod="openstack/neutron-db-sync-wvjsl" Oct 08 18:29:21 crc kubenswrapper[4988]: I1008 18:29:21.019664 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/de747e73-c280-424f-af48-3135bec91b55-config\") pod \"neutron-db-sync-wvjsl\" (UID: \"de747e73-c280-424f-af48-3135bec91b55\") " pod="openstack/neutron-db-sync-wvjsl" Oct 08 18:29:21 crc kubenswrapper[4988]: I1008 18:29:21.023797 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/de747e73-c280-424f-af48-3135bec91b55-combined-ca-bundle\") pod \"neutron-db-sync-wvjsl\" (UID: \"de747e73-c280-424f-af48-3135bec91b55\") " pod="openstack/neutron-db-sync-wvjsl" Oct 08 18:29:21 crc kubenswrapper[4988]: I1008 18:29:21.024037 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/de747e73-c280-424f-af48-3135bec91b55-config\") pod \"neutron-db-sync-wvjsl\" (UID: \"de747e73-c280-424f-af48-3135bec91b55\") " pod="openstack/neutron-db-sync-wvjsl" Oct 08 18:29:21 crc kubenswrapper[4988]: I1008 18:29:21.036720 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7vbb4\" (UniqueName: \"kubernetes.io/projected/de747e73-c280-424f-af48-3135bec91b55-kube-api-access-7vbb4\") pod \"neutron-db-sync-wvjsl\" (UID: \"de747e73-c280-424f-af48-3135bec91b55\") " pod="openstack/neutron-db-sync-wvjsl" Oct 08 18:29:21 crc kubenswrapper[4988]: I1008 18:29:21.135838 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-wvjsl" Oct 08 18:29:21 crc kubenswrapper[4988]: I1008 18:29:21.138725 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-c44756996-dfw4k"] Oct 08 18:29:21 crc kubenswrapper[4988]: I1008 18:29:21.140664 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-c44756996-dfw4k" Oct 08 18:29:21 crc kubenswrapper[4988]: I1008 18:29:21.144077 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-placement-internal-svc" Oct 08 18:29:21 crc kubenswrapper[4988]: I1008 18:29:21.144124 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Oct 08 18:29:21 crc kubenswrapper[4988]: I1008 18:29:21.144167 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Oct 08 18:29:21 crc kubenswrapper[4988]: I1008 18:29:21.144343 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-placement-public-svc" Oct 08 18:29:21 crc kubenswrapper[4988]: I1008 18:29:21.144590 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-mxq4l" Oct 08 18:29:21 crc kubenswrapper[4988]: I1008 18:29:21.168814 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-c44756996-dfw4k"] Oct 08 18:29:21 crc kubenswrapper[4988]: I1008 18:29:21.225815 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f4e47608-89b4-45d1-88a6-fe75805d864d-logs\") pod \"placement-c44756996-dfw4k\" (UID: \"f4e47608-89b4-45d1-88a6-fe75805d864d\") " pod="openstack/placement-c44756996-dfw4k" Oct 08 18:29:21 crc kubenswrapper[4988]: I1008 18:29:21.225901 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/f4e47608-89b4-45d1-88a6-fe75805d864d-public-tls-certs\") pod \"placement-c44756996-dfw4k\" (UID: \"f4e47608-89b4-45d1-88a6-fe75805d864d\") " pod="openstack/placement-c44756996-dfw4k" Oct 08 18:29:21 crc kubenswrapper[4988]: I1008 18:29:21.225986 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f4e47608-89b4-45d1-88a6-fe75805d864d-combined-ca-bundle\") pod \"placement-c44756996-dfw4k\" (UID: \"f4e47608-89b4-45d1-88a6-fe75805d864d\") " pod="openstack/placement-c44756996-dfw4k" Oct 08 18:29:21 crc kubenswrapper[4988]: I1008 18:29:21.226021 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f4e47608-89b4-45d1-88a6-fe75805d864d-scripts\") pod \"placement-c44756996-dfw4k\" (UID: \"f4e47608-89b4-45d1-88a6-fe75805d864d\") " pod="openstack/placement-c44756996-dfw4k" Oct 08 18:29:21 crc kubenswrapper[4988]: I1008 18:29:21.226050 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/f4e47608-89b4-45d1-88a6-fe75805d864d-internal-tls-certs\") pod \"placement-c44756996-dfw4k\" (UID: \"f4e47608-89b4-45d1-88a6-fe75805d864d\") " pod="openstack/placement-c44756996-dfw4k" Oct 08 18:29:21 crc kubenswrapper[4988]: I1008 18:29:21.226072 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f4e47608-89b4-45d1-88a6-fe75805d864d-config-data\") pod \"placement-c44756996-dfw4k\" (UID: \"f4e47608-89b4-45d1-88a6-fe75805d864d\") " pod="openstack/placement-c44756996-dfw4k" Oct 08 18:29:21 crc kubenswrapper[4988]: I1008 18:29:21.226109 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tt47m\" (UniqueName: \"kubernetes.io/projected/f4e47608-89b4-45d1-88a6-fe75805d864d-kube-api-access-tt47m\") pod \"placement-c44756996-dfw4k\" (UID: \"f4e47608-89b4-45d1-88a6-fe75805d864d\") " pod="openstack/placement-c44756996-dfw4k" Oct 08 18:29:21 crc kubenswrapper[4988]: I1008 18:29:21.327216 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f4e47608-89b4-45d1-88a6-fe75805d864d-combined-ca-bundle\") pod \"placement-c44756996-dfw4k\" (UID: \"f4e47608-89b4-45d1-88a6-fe75805d864d\") " pod="openstack/placement-c44756996-dfw4k" Oct 08 18:29:21 crc kubenswrapper[4988]: I1008 18:29:21.327292 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f4e47608-89b4-45d1-88a6-fe75805d864d-scripts\") pod \"placement-c44756996-dfw4k\" (UID: \"f4e47608-89b4-45d1-88a6-fe75805d864d\") " pod="openstack/placement-c44756996-dfw4k" Oct 08 18:29:21 crc kubenswrapper[4988]: I1008 18:29:21.327320 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/f4e47608-89b4-45d1-88a6-fe75805d864d-internal-tls-certs\") pod \"placement-c44756996-dfw4k\" (UID: \"f4e47608-89b4-45d1-88a6-fe75805d864d\") " pod="openstack/placement-c44756996-dfw4k" Oct 08 18:29:21 crc kubenswrapper[4988]: I1008 18:29:21.327348 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f4e47608-89b4-45d1-88a6-fe75805d864d-config-data\") pod \"placement-c44756996-dfw4k\" (UID: \"f4e47608-89b4-45d1-88a6-fe75805d864d\") " pod="openstack/placement-c44756996-dfw4k" Oct 08 18:29:21 crc kubenswrapper[4988]: I1008 18:29:21.327375 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tt47m\" (UniqueName: \"kubernetes.io/projected/f4e47608-89b4-45d1-88a6-fe75805d864d-kube-api-access-tt47m\") pod \"placement-c44756996-dfw4k\" (UID: \"f4e47608-89b4-45d1-88a6-fe75805d864d\") " pod="openstack/placement-c44756996-dfw4k" Oct 08 18:29:21 crc kubenswrapper[4988]: I1008 18:29:21.327435 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f4e47608-89b4-45d1-88a6-fe75805d864d-logs\") pod \"placement-c44756996-dfw4k\" (UID: \"f4e47608-89b4-45d1-88a6-fe75805d864d\") " pod="openstack/placement-c44756996-dfw4k" Oct 08 18:29:21 crc kubenswrapper[4988]: I1008 18:29:21.327462 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/f4e47608-89b4-45d1-88a6-fe75805d864d-public-tls-certs\") pod \"placement-c44756996-dfw4k\" (UID: \"f4e47608-89b4-45d1-88a6-fe75805d864d\") " pod="openstack/placement-c44756996-dfw4k" Oct 08 18:29:21 crc kubenswrapper[4988]: I1008 18:29:21.328303 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f4e47608-89b4-45d1-88a6-fe75805d864d-logs\") pod \"placement-c44756996-dfw4k\" (UID: \"f4e47608-89b4-45d1-88a6-fe75805d864d\") " pod="openstack/placement-c44756996-dfw4k" Oct 08 18:29:21 crc kubenswrapper[4988]: I1008 18:29:21.333820 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f4e47608-89b4-45d1-88a6-fe75805d864d-scripts\") pod \"placement-c44756996-dfw4k\" (UID: \"f4e47608-89b4-45d1-88a6-fe75805d864d\") " pod="openstack/placement-c44756996-dfw4k" Oct 08 18:29:21 crc kubenswrapper[4988]: I1008 18:29:21.351298 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f4e47608-89b4-45d1-88a6-fe75805d864d-combined-ca-bundle\") pod \"placement-c44756996-dfw4k\" (UID: \"f4e47608-89b4-45d1-88a6-fe75805d864d\") " pod="openstack/placement-c44756996-dfw4k" Oct 08 18:29:21 crc kubenswrapper[4988]: I1008 18:29:21.353370 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/f4e47608-89b4-45d1-88a6-fe75805d864d-public-tls-certs\") pod \"placement-c44756996-dfw4k\" (UID: \"f4e47608-89b4-45d1-88a6-fe75805d864d\") " pod="openstack/placement-c44756996-dfw4k" Oct 08 18:29:21 crc kubenswrapper[4988]: I1008 18:29:21.353438 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f4e47608-89b4-45d1-88a6-fe75805d864d-config-data\") pod \"placement-c44756996-dfw4k\" (UID: \"f4e47608-89b4-45d1-88a6-fe75805d864d\") " pod="openstack/placement-c44756996-dfw4k" Oct 08 18:29:21 crc kubenswrapper[4988]: I1008 18:29:21.354573 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/f4e47608-89b4-45d1-88a6-fe75805d864d-internal-tls-certs\") pod \"placement-c44756996-dfw4k\" (UID: \"f4e47608-89b4-45d1-88a6-fe75805d864d\") " pod="openstack/placement-c44756996-dfw4k" Oct 08 18:29:21 crc kubenswrapper[4988]: I1008 18:29:21.358122 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tt47m\" (UniqueName: \"kubernetes.io/projected/f4e47608-89b4-45d1-88a6-fe75805d864d-kube-api-access-tt47m\") pod \"placement-c44756996-dfw4k\" (UID: \"f4e47608-89b4-45d1-88a6-fe75805d864d\") " pod="openstack/placement-c44756996-dfw4k" Oct 08 18:29:21 crc kubenswrapper[4988]: I1008 18:29:21.472978 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-c44756996-dfw4k" Oct 08 18:29:22 crc kubenswrapper[4988]: I1008 18:29:22.986433 4988 generic.go:334] "Generic (PLEG): container finished" podID="c1f2511c-e7bc-4433-910b-eb3b94fcf09b" containerID="e0c58c7d3f2cbaa16eb7f1b22d366459ec6d6b096a0336a9554a9d97a33841f1" exitCode=0 Oct 08 18:29:22 crc kubenswrapper[4988]: I1008 18:29:22.986529 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-5v5z9" event={"ID":"c1f2511c-e7bc-4433-910b-eb3b94fcf09b","Type":"ContainerDied","Data":"e0c58c7d3f2cbaa16eb7f1b22d366459ec6d6b096a0336a9554a9d97a33841f1"} Oct 08 18:29:22 crc kubenswrapper[4988]: I1008 18:29:22.990018 4988 generic.go:334] "Generic (PLEG): container finished" podID="b2e4e6bf-c872-4a45-be09-08decca63284" containerID="c6a4733dbf41fe0410a80b6de0ff01b8123be53bae21bb88ca041b937dd29455" exitCode=0 Oct 08 18:29:22 crc kubenswrapper[4988]: I1008 18:29:22.990055 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-748qj" event={"ID":"b2e4e6bf-c872-4a45-be09-08decca63284","Type":"ContainerDied","Data":"c6a4733dbf41fe0410a80b6de0ff01b8123be53bae21bb88ca041b937dd29455"} Oct 08 18:29:24 crc kubenswrapper[4988]: I1008 18:29:24.417647 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-c44756996-dfw4k"] Oct 08 18:29:24 crc kubenswrapper[4988]: W1008 18:29:24.423013 4988 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf4e47608_89b4_45d1_88a6_fe75805d864d.slice/crio-8cb906ede5c077bda5980c9dc0e8b3c6324dd51624529cfd6fe9c32c2ad8c301 WatchSource:0}: Error finding container 8cb906ede5c077bda5980c9dc0e8b3c6324dd51624529cfd6fe9c32c2ad8c301: Status 404 returned error can't find the container with id 8cb906ede5c077bda5980c9dc0e8b3c6324dd51624529cfd6fe9c32c2ad8c301 Oct 08 18:29:24 crc kubenswrapper[4988]: I1008 18:29:24.462733 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-748qj" Oct 08 18:29:24 crc kubenswrapper[4988]: I1008 18:29:24.464260 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-wvjsl"] Oct 08 18:29:24 crc kubenswrapper[4988]: W1008 18:29:24.470893 4988 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podde747e73_c280_424f_af48_3135bec91b55.slice/crio-d755e3c1d79ef62e9e57c675a8dc7c0d1d929f375f427db8ba59d7801f19ed4b WatchSource:0}: Error finding container d755e3c1d79ef62e9e57c675a8dc7c0d1d929f375f427db8ba59d7801f19ed4b: Status 404 returned error can't find the container with id d755e3c1d79ef62e9e57c675a8dc7c0d1d929f375f427db8ba59d7801f19ed4b Oct 08 18:29:24 crc kubenswrapper[4988]: I1008 18:29:24.552358 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-dm56p"] Oct 08 18:29:24 crc kubenswrapper[4988]: W1008 18:29:24.557649 4988 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd7f4f7aa_ccf2_45e1_8d32_db9684853231.slice/crio-932a4ebfb7ff061842fd0d0d71f8410a08a92a4f1aca818d7b9f074929ab0631 WatchSource:0}: Error finding container 932a4ebfb7ff061842fd0d0d71f8410a08a92a4f1aca818d7b9f074929ab0631: Status 404 returned error can't find the container with id 932a4ebfb7ff061842fd0d0d71f8410a08a92a4f1aca818d7b9f074929ab0631 Oct 08 18:29:24 crc kubenswrapper[4988]: I1008 18:29:24.584059 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b2e4e6bf-c872-4a45-be09-08decca63284-config-data\") pod \"b2e4e6bf-c872-4a45-be09-08decca63284\" (UID: \"b2e4e6bf-c872-4a45-be09-08decca63284\") " Oct 08 18:29:24 crc kubenswrapper[4988]: I1008 18:29:24.584110 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/b2e4e6bf-c872-4a45-be09-08decca63284-credential-keys\") pod \"b2e4e6bf-c872-4a45-be09-08decca63284\" (UID: \"b2e4e6bf-c872-4a45-be09-08decca63284\") " Oct 08 18:29:24 crc kubenswrapper[4988]: I1008 18:29:24.584182 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b2e4e6bf-c872-4a45-be09-08decca63284-scripts\") pod \"b2e4e6bf-c872-4a45-be09-08decca63284\" (UID: \"b2e4e6bf-c872-4a45-be09-08decca63284\") " Oct 08 18:29:24 crc kubenswrapper[4988]: I1008 18:29:24.584251 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b2e4e6bf-c872-4a45-be09-08decca63284-combined-ca-bundle\") pod \"b2e4e6bf-c872-4a45-be09-08decca63284\" (UID: \"b2e4e6bf-c872-4a45-be09-08decca63284\") " Oct 08 18:29:24 crc kubenswrapper[4988]: I1008 18:29:24.584349 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/b2e4e6bf-c872-4a45-be09-08decca63284-fernet-keys\") pod \"b2e4e6bf-c872-4a45-be09-08decca63284\" (UID: \"b2e4e6bf-c872-4a45-be09-08decca63284\") " Oct 08 18:29:24 crc kubenswrapper[4988]: I1008 18:29:24.584423 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mlqtz\" (UniqueName: \"kubernetes.io/projected/b2e4e6bf-c872-4a45-be09-08decca63284-kube-api-access-mlqtz\") pod \"b2e4e6bf-c872-4a45-be09-08decca63284\" (UID: \"b2e4e6bf-c872-4a45-be09-08decca63284\") " Oct 08 18:29:24 crc kubenswrapper[4988]: I1008 18:29:24.591437 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b2e4e6bf-c872-4a45-be09-08decca63284-scripts" (OuterVolumeSpecName: "scripts") pod "b2e4e6bf-c872-4a45-be09-08decca63284" (UID: "b2e4e6bf-c872-4a45-be09-08decca63284"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:29:24 crc kubenswrapper[4988]: I1008 18:29:24.591699 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b2e4e6bf-c872-4a45-be09-08decca63284-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "b2e4e6bf-c872-4a45-be09-08decca63284" (UID: "b2e4e6bf-c872-4a45-be09-08decca63284"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:29:24 crc kubenswrapper[4988]: I1008 18:29:24.592068 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b2e4e6bf-c872-4a45-be09-08decca63284-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "b2e4e6bf-c872-4a45-be09-08decca63284" (UID: "b2e4e6bf-c872-4a45-be09-08decca63284"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:29:24 crc kubenswrapper[4988]: I1008 18:29:24.609162 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b2e4e6bf-c872-4a45-be09-08decca63284-kube-api-access-mlqtz" (OuterVolumeSpecName: "kube-api-access-mlqtz") pod "b2e4e6bf-c872-4a45-be09-08decca63284" (UID: "b2e4e6bf-c872-4a45-be09-08decca63284"). InnerVolumeSpecName "kube-api-access-mlqtz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:29:24 crc kubenswrapper[4988]: I1008 18:29:24.631311 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b2e4e6bf-c872-4a45-be09-08decca63284-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "b2e4e6bf-c872-4a45-be09-08decca63284" (UID: "b2e4e6bf-c872-4a45-be09-08decca63284"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:29:24 crc kubenswrapper[4988]: I1008 18:29:24.635235 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b2e4e6bf-c872-4a45-be09-08decca63284-config-data" (OuterVolumeSpecName: "config-data") pod "b2e4e6bf-c872-4a45-be09-08decca63284" (UID: "b2e4e6bf-c872-4a45-be09-08decca63284"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:29:24 crc kubenswrapper[4988]: I1008 18:29:24.686402 4988 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b2e4e6bf-c872-4a45-be09-08decca63284-config-data\") on node \"crc\" DevicePath \"\"" Oct 08 18:29:24 crc kubenswrapper[4988]: I1008 18:29:24.686426 4988 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/b2e4e6bf-c872-4a45-be09-08decca63284-credential-keys\") on node \"crc\" DevicePath \"\"" Oct 08 18:29:24 crc kubenswrapper[4988]: I1008 18:29:24.686436 4988 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b2e4e6bf-c872-4a45-be09-08decca63284-scripts\") on node \"crc\" DevicePath \"\"" Oct 08 18:29:24 crc kubenswrapper[4988]: I1008 18:29:24.686445 4988 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b2e4e6bf-c872-4a45-be09-08decca63284-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 18:29:24 crc kubenswrapper[4988]: I1008 18:29:24.686454 4988 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/b2e4e6bf-c872-4a45-be09-08decca63284-fernet-keys\") on node \"crc\" DevicePath \"\"" Oct 08 18:29:24 crc kubenswrapper[4988]: I1008 18:29:24.686462 4988 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mlqtz\" (UniqueName: \"kubernetes.io/projected/b2e4e6bf-c872-4a45-be09-08decca63284-kube-api-access-mlqtz\") on node \"crc\" DevicePath \"\"" Oct 08 18:29:25 crc kubenswrapper[4988]: I1008 18:29:25.011079 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-dm56p" event={"ID":"d7f4f7aa-ccf2-45e1-8d32-db9684853231","Type":"ContainerStarted","Data":"932a4ebfb7ff061842fd0d0d71f8410a08a92a4f1aca818d7b9f074929ab0631"} Oct 08 18:29:25 crc kubenswrapper[4988]: I1008 18:29:25.012717 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-748qj" event={"ID":"b2e4e6bf-c872-4a45-be09-08decca63284","Type":"ContainerDied","Data":"f22a00bcf752505f07c237044ba6fc9fb6deaeef7af5229a18cbd699754119e3"} Oct 08 18:29:25 crc kubenswrapper[4988]: I1008 18:29:25.012826 4988 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f22a00bcf752505f07c237044ba6fc9fb6deaeef7af5229a18cbd699754119e3" Oct 08 18:29:25 crc kubenswrapper[4988]: I1008 18:29:25.012773 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-748qj" Oct 08 18:29:25 crc kubenswrapper[4988]: I1008 18:29:25.013921 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-wvjsl" event={"ID":"de747e73-c280-424f-af48-3135bec91b55","Type":"ContainerStarted","Data":"d755e3c1d79ef62e9e57c675a8dc7c0d1d929f375f427db8ba59d7801f19ed4b"} Oct 08 18:29:25 crc kubenswrapper[4988]: I1008 18:29:25.017063 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-c44756996-dfw4k" event={"ID":"f4e47608-89b4-45d1-88a6-fe75805d864d","Type":"ContainerStarted","Data":"8cb906ede5c077bda5980c9dc0e8b3c6324dd51624529cfd6fe9c32c2ad8c301"} Oct 08 18:29:25 crc kubenswrapper[4988]: I1008 18:29:25.122763 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-6dbc6dc9bd-hk8qw"] Oct 08 18:29:25 crc kubenswrapper[4988]: E1008 18:29:25.125278 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b2e4e6bf-c872-4a45-be09-08decca63284" containerName="keystone-bootstrap" Oct 08 18:29:25 crc kubenswrapper[4988]: I1008 18:29:25.125307 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="b2e4e6bf-c872-4a45-be09-08decca63284" containerName="keystone-bootstrap" Oct 08 18:29:25 crc kubenswrapper[4988]: I1008 18:29:25.125541 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="b2e4e6bf-c872-4a45-be09-08decca63284" containerName="keystone-bootstrap" Oct 08 18:29:25 crc kubenswrapper[4988]: I1008 18:29:25.131251 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-6dbc6dc9bd-hk8qw" Oct 08 18:29:25 crc kubenswrapper[4988]: I1008 18:29:25.138939 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Oct 08 18:29:25 crc kubenswrapper[4988]: I1008 18:29:25.139306 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-keystone-public-svc" Oct 08 18:29:25 crc kubenswrapper[4988]: I1008 18:29:25.139424 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-srkzn" Oct 08 18:29:25 crc kubenswrapper[4988]: I1008 18:29:25.139554 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Oct 08 18:29:25 crc kubenswrapper[4988]: I1008 18:29:25.140273 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-keystone-internal-svc" Oct 08 18:29:25 crc kubenswrapper[4988]: I1008 18:29:25.140471 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Oct 08 18:29:25 crc kubenswrapper[4988]: I1008 18:29:25.159499 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-6dbc6dc9bd-hk8qw"] Oct 08 18:29:25 crc kubenswrapper[4988]: I1008 18:29:25.194839 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/26a3b0a9-6aca-490d-a6fb-08addef86fbb-public-tls-certs\") pod \"keystone-6dbc6dc9bd-hk8qw\" (UID: \"26a3b0a9-6aca-490d-a6fb-08addef86fbb\") " pod="openstack/keystone-6dbc6dc9bd-hk8qw" Oct 08 18:29:25 crc kubenswrapper[4988]: I1008 18:29:25.194887 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/26a3b0a9-6aca-490d-a6fb-08addef86fbb-fernet-keys\") pod \"keystone-6dbc6dc9bd-hk8qw\" (UID: \"26a3b0a9-6aca-490d-a6fb-08addef86fbb\") " pod="openstack/keystone-6dbc6dc9bd-hk8qw" Oct 08 18:29:25 crc kubenswrapper[4988]: I1008 18:29:25.194925 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/26a3b0a9-6aca-490d-a6fb-08addef86fbb-combined-ca-bundle\") pod \"keystone-6dbc6dc9bd-hk8qw\" (UID: \"26a3b0a9-6aca-490d-a6fb-08addef86fbb\") " pod="openstack/keystone-6dbc6dc9bd-hk8qw" Oct 08 18:29:25 crc kubenswrapper[4988]: I1008 18:29:25.194960 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/26a3b0a9-6aca-490d-a6fb-08addef86fbb-credential-keys\") pod \"keystone-6dbc6dc9bd-hk8qw\" (UID: \"26a3b0a9-6aca-490d-a6fb-08addef86fbb\") " pod="openstack/keystone-6dbc6dc9bd-hk8qw" Oct 08 18:29:25 crc kubenswrapper[4988]: I1008 18:29:25.195001 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/26a3b0a9-6aca-490d-a6fb-08addef86fbb-config-data\") pod \"keystone-6dbc6dc9bd-hk8qw\" (UID: \"26a3b0a9-6aca-490d-a6fb-08addef86fbb\") " pod="openstack/keystone-6dbc6dc9bd-hk8qw" Oct 08 18:29:25 crc kubenswrapper[4988]: I1008 18:29:25.195023 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/26a3b0a9-6aca-490d-a6fb-08addef86fbb-internal-tls-certs\") pod \"keystone-6dbc6dc9bd-hk8qw\" (UID: \"26a3b0a9-6aca-490d-a6fb-08addef86fbb\") " pod="openstack/keystone-6dbc6dc9bd-hk8qw" Oct 08 18:29:25 crc kubenswrapper[4988]: I1008 18:29:25.195047 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/26a3b0a9-6aca-490d-a6fb-08addef86fbb-scripts\") pod \"keystone-6dbc6dc9bd-hk8qw\" (UID: \"26a3b0a9-6aca-490d-a6fb-08addef86fbb\") " pod="openstack/keystone-6dbc6dc9bd-hk8qw" Oct 08 18:29:25 crc kubenswrapper[4988]: I1008 18:29:25.195099 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hczwx\" (UniqueName: \"kubernetes.io/projected/26a3b0a9-6aca-490d-a6fb-08addef86fbb-kube-api-access-hczwx\") pod \"keystone-6dbc6dc9bd-hk8qw\" (UID: \"26a3b0a9-6aca-490d-a6fb-08addef86fbb\") " pod="openstack/keystone-6dbc6dc9bd-hk8qw" Oct 08 18:29:25 crc kubenswrapper[4988]: I1008 18:29:25.296711 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hczwx\" (UniqueName: \"kubernetes.io/projected/26a3b0a9-6aca-490d-a6fb-08addef86fbb-kube-api-access-hczwx\") pod \"keystone-6dbc6dc9bd-hk8qw\" (UID: \"26a3b0a9-6aca-490d-a6fb-08addef86fbb\") " pod="openstack/keystone-6dbc6dc9bd-hk8qw" Oct 08 18:29:25 crc kubenswrapper[4988]: I1008 18:29:25.296782 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/26a3b0a9-6aca-490d-a6fb-08addef86fbb-public-tls-certs\") pod \"keystone-6dbc6dc9bd-hk8qw\" (UID: \"26a3b0a9-6aca-490d-a6fb-08addef86fbb\") " pod="openstack/keystone-6dbc6dc9bd-hk8qw" Oct 08 18:29:25 crc kubenswrapper[4988]: I1008 18:29:25.296812 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/26a3b0a9-6aca-490d-a6fb-08addef86fbb-fernet-keys\") pod \"keystone-6dbc6dc9bd-hk8qw\" (UID: \"26a3b0a9-6aca-490d-a6fb-08addef86fbb\") " pod="openstack/keystone-6dbc6dc9bd-hk8qw" Oct 08 18:29:25 crc kubenswrapper[4988]: I1008 18:29:25.296867 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/26a3b0a9-6aca-490d-a6fb-08addef86fbb-combined-ca-bundle\") pod \"keystone-6dbc6dc9bd-hk8qw\" (UID: \"26a3b0a9-6aca-490d-a6fb-08addef86fbb\") " pod="openstack/keystone-6dbc6dc9bd-hk8qw" Oct 08 18:29:25 crc kubenswrapper[4988]: I1008 18:29:25.296935 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/26a3b0a9-6aca-490d-a6fb-08addef86fbb-credential-keys\") pod \"keystone-6dbc6dc9bd-hk8qw\" (UID: \"26a3b0a9-6aca-490d-a6fb-08addef86fbb\") " pod="openstack/keystone-6dbc6dc9bd-hk8qw" Oct 08 18:29:25 crc kubenswrapper[4988]: I1008 18:29:25.297005 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/26a3b0a9-6aca-490d-a6fb-08addef86fbb-config-data\") pod \"keystone-6dbc6dc9bd-hk8qw\" (UID: \"26a3b0a9-6aca-490d-a6fb-08addef86fbb\") " pod="openstack/keystone-6dbc6dc9bd-hk8qw" Oct 08 18:29:25 crc kubenswrapper[4988]: I1008 18:29:25.297032 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/26a3b0a9-6aca-490d-a6fb-08addef86fbb-internal-tls-certs\") pod \"keystone-6dbc6dc9bd-hk8qw\" (UID: \"26a3b0a9-6aca-490d-a6fb-08addef86fbb\") " pod="openstack/keystone-6dbc6dc9bd-hk8qw" Oct 08 18:29:25 crc kubenswrapper[4988]: I1008 18:29:25.297057 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/26a3b0a9-6aca-490d-a6fb-08addef86fbb-scripts\") pod \"keystone-6dbc6dc9bd-hk8qw\" (UID: \"26a3b0a9-6aca-490d-a6fb-08addef86fbb\") " pod="openstack/keystone-6dbc6dc9bd-hk8qw" Oct 08 18:29:25 crc kubenswrapper[4988]: I1008 18:29:25.301794 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/26a3b0a9-6aca-490d-a6fb-08addef86fbb-combined-ca-bundle\") pod \"keystone-6dbc6dc9bd-hk8qw\" (UID: \"26a3b0a9-6aca-490d-a6fb-08addef86fbb\") " pod="openstack/keystone-6dbc6dc9bd-hk8qw" Oct 08 18:29:25 crc kubenswrapper[4988]: I1008 18:29:25.302243 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/26a3b0a9-6aca-490d-a6fb-08addef86fbb-scripts\") pod \"keystone-6dbc6dc9bd-hk8qw\" (UID: \"26a3b0a9-6aca-490d-a6fb-08addef86fbb\") " pod="openstack/keystone-6dbc6dc9bd-hk8qw" Oct 08 18:29:25 crc kubenswrapper[4988]: I1008 18:29:25.303306 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/26a3b0a9-6aca-490d-a6fb-08addef86fbb-fernet-keys\") pod \"keystone-6dbc6dc9bd-hk8qw\" (UID: \"26a3b0a9-6aca-490d-a6fb-08addef86fbb\") " pod="openstack/keystone-6dbc6dc9bd-hk8qw" Oct 08 18:29:25 crc kubenswrapper[4988]: I1008 18:29:25.303449 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/26a3b0a9-6aca-490d-a6fb-08addef86fbb-internal-tls-certs\") pod \"keystone-6dbc6dc9bd-hk8qw\" (UID: \"26a3b0a9-6aca-490d-a6fb-08addef86fbb\") " pod="openstack/keystone-6dbc6dc9bd-hk8qw" Oct 08 18:29:25 crc kubenswrapper[4988]: I1008 18:29:25.306078 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/26a3b0a9-6aca-490d-a6fb-08addef86fbb-credential-keys\") pod \"keystone-6dbc6dc9bd-hk8qw\" (UID: \"26a3b0a9-6aca-490d-a6fb-08addef86fbb\") " pod="openstack/keystone-6dbc6dc9bd-hk8qw" Oct 08 18:29:25 crc kubenswrapper[4988]: I1008 18:29:25.306917 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/26a3b0a9-6aca-490d-a6fb-08addef86fbb-public-tls-certs\") pod \"keystone-6dbc6dc9bd-hk8qw\" (UID: \"26a3b0a9-6aca-490d-a6fb-08addef86fbb\") " pod="openstack/keystone-6dbc6dc9bd-hk8qw" Oct 08 18:29:25 crc kubenswrapper[4988]: I1008 18:29:25.307232 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/26a3b0a9-6aca-490d-a6fb-08addef86fbb-config-data\") pod \"keystone-6dbc6dc9bd-hk8qw\" (UID: \"26a3b0a9-6aca-490d-a6fb-08addef86fbb\") " pod="openstack/keystone-6dbc6dc9bd-hk8qw" Oct 08 18:29:25 crc kubenswrapper[4988]: I1008 18:29:25.324924 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hczwx\" (UniqueName: \"kubernetes.io/projected/26a3b0a9-6aca-490d-a6fb-08addef86fbb-kube-api-access-hczwx\") pod \"keystone-6dbc6dc9bd-hk8qw\" (UID: \"26a3b0a9-6aca-490d-a6fb-08addef86fbb\") " pod="openstack/keystone-6dbc6dc9bd-hk8qw" Oct 08 18:29:25 crc kubenswrapper[4988]: I1008 18:29:25.467894 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-6dbc6dc9bd-hk8qw" Oct 08 18:29:25 crc kubenswrapper[4988]: I1008 18:29:25.607468 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-sync-gm2vf"] Oct 08 18:29:25 crc kubenswrapper[4988]: I1008 18:29:25.608832 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-gm2vf" Oct 08 18:29:25 crc kubenswrapper[4988]: I1008 18:29:25.611938 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-pvckn" Oct 08 18:29:25 crc kubenswrapper[4988]: I1008 18:29:25.616533 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Oct 08 18:29:25 crc kubenswrapper[4988]: I1008 18:29:25.617319 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Oct 08 18:29:25 crc kubenswrapper[4988]: I1008 18:29:25.646763 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-gm2vf"] Oct 08 18:29:25 crc kubenswrapper[4988]: I1008 18:29:25.704040 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/5b6f72e9-8dc6-4290-b032-d8a570b91767-db-sync-config-data\") pod \"cinder-db-sync-gm2vf\" (UID: \"5b6f72e9-8dc6-4290-b032-d8a570b91767\") " pod="openstack/cinder-db-sync-gm2vf" Oct 08 18:29:25 crc kubenswrapper[4988]: I1008 18:29:25.704157 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lshc6\" (UniqueName: \"kubernetes.io/projected/5b6f72e9-8dc6-4290-b032-d8a570b91767-kube-api-access-lshc6\") pod \"cinder-db-sync-gm2vf\" (UID: \"5b6f72e9-8dc6-4290-b032-d8a570b91767\") " pod="openstack/cinder-db-sync-gm2vf" Oct 08 18:29:25 crc kubenswrapper[4988]: I1008 18:29:25.704234 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5b6f72e9-8dc6-4290-b032-d8a570b91767-scripts\") pod \"cinder-db-sync-gm2vf\" (UID: \"5b6f72e9-8dc6-4290-b032-d8a570b91767\") " pod="openstack/cinder-db-sync-gm2vf" Oct 08 18:29:25 crc kubenswrapper[4988]: I1008 18:29:25.704273 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/5b6f72e9-8dc6-4290-b032-d8a570b91767-etc-machine-id\") pod \"cinder-db-sync-gm2vf\" (UID: \"5b6f72e9-8dc6-4290-b032-d8a570b91767\") " pod="openstack/cinder-db-sync-gm2vf" Oct 08 18:29:25 crc kubenswrapper[4988]: I1008 18:29:25.704359 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5b6f72e9-8dc6-4290-b032-d8a570b91767-config-data\") pod \"cinder-db-sync-gm2vf\" (UID: \"5b6f72e9-8dc6-4290-b032-d8a570b91767\") " pod="openstack/cinder-db-sync-gm2vf" Oct 08 18:29:25 crc kubenswrapper[4988]: I1008 18:29:25.704437 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5b6f72e9-8dc6-4290-b032-d8a570b91767-combined-ca-bundle\") pod \"cinder-db-sync-gm2vf\" (UID: \"5b6f72e9-8dc6-4290-b032-d8a570b91767\") " pod="openstack/cinder-db-sync-gm2vf" Oct 08 18:29:25 crc kubenswrapper[4988]: I1008 18:29:25.805646 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5b6f72e9-8dc6-4290-b032-d8a570b91767-scripts\") pod \"cinder-db-sync-gm2vf\" (UID: \"5b6f72e9-8dc6-4290-b032-d8a570b91767\") " pod="openstack/cinder-db-sync-gm2vf" Oct 08 18:29:25 crc kubenswrapper[4988]: I1008 18:29:25.805720 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/5b6f72e9-8dc6-4290-b032-d8a570b91767-etc-machine-id\") pod \"cinder-db-sync-gm2vf\" (UID: \"5b6f72e9-8dc6-4290-b032-d8a570b91767\") " pod="openstack/cinder-db-sync-gm2vf" Oct 08 18:29:25 crc kubenswrapper[4988]: I1008 18:29:25.805844 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5b6f72e9-8dc6-4290-b032-d8a570b91767-config-data\") pod \"cinder-db-sync-gm2vf\" (UID: \"5b6f72e9-8dc6-4290-b032-d8a570b91767\") " pod="openstack/cinder-db-sync-gm2vf" Oct 08 18:29:25 crc kubenswrapper[4988]: I1008 18:29:25.805913 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5b6f72e9-8dc6-4290-b032-d8a570b91767-combined-ca-bundle\") pod \"cinder-db-sync-gm2vf\" (UID: \"5b6f72e9-8dc6-4290-b032-d8a570b91767\") " pod="openstack/cinder-db-sync-gm2vf" Oct 08 18:29:25 crc kubenswrapper[4988]: I1008 18:29:25.805952 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/5b6f72e9-8dc6-4290-b032-d8a570b91767-db-sync-config-data\") pod \"cinder-db-sync-gm2vf\" (UID: \"5b6f72e9-8dc6-4290-b032-d8a570b91767\") " pod="openstack/cinder-db-sync-gm2vf" Oct 08 18:29:25 crc kubenswrapper[4988]: I1008 18:29:25.806028 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lshc6\" (UniqueName: \"kubernetes.io/projected/5b6f72e9-8dc6-4290-b032-d8a570b91767-kube-api-access-lshc6\") pod \"cinder-db-sync-gm2vf\" (UID: \"5b6f72e9-8dc6-4290-b032-d8a570b91767\") " pod="openstack/cinder-db-sync-gm2vf" Oct 08 18:29:25 crc kubenswrapper[4988]: I1008 18:29:25.806980 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/5b6f72e9-8dc6-4290-b032-d8a570b91767-etc-machine-id\") pod \"cinder-db-sync-gm2vf\" (UID: \"5b6f72e9-8dc6-4290-b032-d8a570b91767\") " pod="openstack/cinder-db-sync-gm2vf" Oct 08 18:29:25 crc kubenswrapper[4988]: I1008 18:29:25.815376 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5b6f72e9-8dc6-4290-b032-d8a570b91767-combined-ca-bundle\") pod \"cinder-db-sync-gm2vf\" (UID: \"5b6f72e9-8dc6-4290-b032-d8a570b91767\") " pod="openstack/cinder-db-sync-gm2vf" Oct 08 18:29:25 crc kubenswrapper[4988]: I1008 18:29:25.815709 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5b6f72e9-8dc6-4290-b032-d8a570b91767-config-data\") pod \"cinder-db-sync-gm2vf\" (UID: \"5b6f72e9-8dc6-4290-b032-d8a570b91767\") " pod="openstack/cinder-db-sync-gm2vf" Oct 08 18:29:25 crc kubenswrapper[4988]: I1008 18:29:25.815749 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/5b6f72e9-8dc6-4290-b032-d8a570b91767-db-sync-config-data\") pod \"cinder-db-sync-gm2vf\" (UID: \"5b6f72e9-8dc6-4290-b032-d8a570b91767\") " pod="openstack/cinder-db-sync-gm2vf" Oct 08 18:29:25 crc kubenswrapper[4988]: I1008 18:29:25.816033 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5b6f72e9-8dc6-4290-b032-d8a570b91767-scripts\") pod \"cinder-db-sync-gm2vf\" (UID: \"5b6f72e9-8dc6-4290-b032-d8a570b91767\") " pod="openstack/cinder-db-sync-gm2vf" Oct 08 18:29:25 crc kubenswrapper[4988]: I1008 18:29:25.834504 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lshc6\" (UniqueName: \"kubernetes.io/projected/5b6f72e9-8dc6-4290-b032-d8a570b91767-kube-api-access-lshc6\") pod \"cinder-db-sync-gm2vf\" (UID: \"5b6f72e9-8dc6-4290-b032-d8a570b91767\") " pod="openstack/cinder-db-sync-gm2vf" Oct 08 18:29:25 crc kubenswrapper[4988]: I1008 18:29:25.947749 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-gm2vf" Oct 08 18:29:26 crc kubenswrapper[4988]: I1008 18:29:26.027615 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-wvjsl" event={"ID":"de747e73-c280-424f-af48-3135bec91b55","Type":"ContainerStarted","Data":"f82c2c46fdece62aac8a106f15e5fddfb69696f72a1f951b75f376ae4f50e07b"} Oct 08 18:29:26 crc kubenswrapper[4988]: I1008 18:29:26.034721 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-c44756996-dfw4k" event={"ID":"f4e47608-89b4-45d1-88a6-fe75805d864d","Type":"ContainerStarted","Data":"13f8a712b04c6d00db8bc2f724015fda6642a1723ac5c2e59ecbd2400fa35ece"} Oct 08 18:29:26 crc kubenswrapper[4988]: I1008 18:29:26.043644 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-db-sync-wvjsl" podStartSLOduration=6.043621914 podStartE2EDuration="6.043621914s" podCreationTimestamp="2025-10-08 18:29:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 18:29:26.039425271 +0000 UTC m=+1111.489268041" watchObservedRunningTime="2025-10-08 18:29:26.043621914 +0000 UTC m=+1111.493464694" Oct 08 18:29:26 crc kubenswrapper[4988]: I1008 18:29:26.438336 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-5v5z9" Oct 08 18:29:26 crc kubenswrapper[4988]: I1008 18:29:26.521927 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c1f2511c-e7bc-4433-910b-eb3b94fcf09b-combined-ca-bundle\") pod \"c1f2511c-e7bc-4433-910b-eb3b94fcf09b\" (UID: \"c1f2511c-e7bc-4433-910b-eb3b94fcf09b\") " Oct 08 18:29:26 crc kubenswrapper[4988]: I1008 18:29:26.522034 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c1f2511c-e7bc-4433-910b-eb3b94fcf09b-config-data\") pod \"c1f2511c-e7bc-4433-910b-eb3b94fcf09b\" (UID: \"c1f2511c-e7bc-4433-910b-eb3b94fcf09b\") " Oct 08 18:29:26 crc kubenswrapper[4988]: I1008 18:29:26.522076 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/c1f2511c-e7bc-4433-910b-eb3b94fcf09b-db-sync-config-data\") pod \"c1f2511c-e7bc-4433-910b-eb3b94fcf09b\" (UID: \"c1f2511c-e7bc-4433-910b-eb3b94fcf09b\") " Oct 08 18:29:26 crc kubenswrapper[4988]: I1008 18:29:26.522282 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-654gh\" (UniqueName: \"kubernetes.io/projected/c1f2511c-e7bc-4433-910b-eb3b94fcf09b-kube-api-access-654gh\") pod \"c1f2511c-e7bc-4433-910b-eb3b94fcf09b\" (UID: \"c1f2511c-e7bc-4433-910b-eb3b94fcf09b\") " Oct 08 18:29:26 crc kubenswrapper[4988]: I1008 18:29:26.533193 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c1f2511c-e7bc-4433-910b-eb3b94fcf09b-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "c1f2511c-e7bc-4433-910b-eb3b94fcf09b" (UID: "c1f2511c-e7bc-4433-910b-eb3b94fcf09b"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:29:26 crc kubenswrapper[4988]: I1008 18:29:26.533256 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c1f2511c-e7bc-4433-910b-eb3b94fcf09b-kube-api-access-654gh" (OuterVolumeSpecName: "kube-api-access-654gh") pod "c1f2511c-e7bc-4433-910b-eb3b94fcf09b" (UID: "c1f2511c-e7bc-4433-910b-eb3b94fcf09b"). InnerVolumeSpecName "kube-api-access-654gh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:29:26 crc kubenswrapper[4988]: I1008 18:29:26.556555 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-6dbc6dc9bd-hk8qw"] Oct 08 18:29:26 crc kubenswrapper[4988]: I1008 18:29:26.566260 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c1f2511c-e7bc-4433-910b-eb3b94fcf09b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "c1f2511c-e7bc-4433-910b-eb3b94fcf09b" (UID: "c1f2511c-e7bc-4433-910b-eb3b94fcf09b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:29:26 crc kubenswrapper[4988]: I1008 18:29:26.578446 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c1f2511c-e7bc-4433-910b-eb3b94fcf09b-config-data" (OuterVolumeSpecName: "config-data") pod "c1f2511c-e7bc-4433-910b-eb3b94fcf09b" (UID: "c1f2511c-e7bc-4433-910b-eb3b94fcf09b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:29:26 crc kubenswrapper[4988]: I1008 18:29:26.625026 4988 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c1f2511c-e7bc-4433-910b-eb3b94fcf09b-config-data\") on node \"crc\" DevicePath \"\"" Oct 08 18:29:26 crc kubenswrapper[4988]: I1008 18:29:26.625055 4988 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/c1f2511c-e7bc-4433-910b-eb3b94fcf09b-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Oct 08 18:29:26 crc kubenswrapper[4988]: I1008 18:29:26.625067 4988 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-654gh\" (UniqueName: \"kubernetes.io/projected/c1f2511c-e7bc-4433-910b-eb3b94fcf09b-kube-api-access-654gh\") on node \"crc\" DevicePath \"\"" Oct 08 18:29:26 crc kubenswrapper[4988]: I1008 18:29:26.625076 4988 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c1f2511c-e7bc-4433-910b-eb3b94fcf09b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 18:29:26 crc kubenswrapper[4988]: I1008 18:29:26.640041 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-gm2vf"] Oct 08 18:29:26 crc kubenswrapper[4988]: W1008 18:29:26.645545 4988 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5b6f72e9_8dc6_4290_b032_d8a570b91767.slice/crio-0445e0aa866fd32b4478f427736e174d9c619b3c4df681e2c48994219d561e07 WatchSource:0}: Error finding container 0445e0aa866fd32b4478f427736e174d9c619b3c4df681e2c48994219d561e07: Status 404 returned error can't find the container with id 0445e0aa866fd32b4478f427736e174d9c619b3c4df681e2c48994219d561e07 Oct 08 18:29:27 crc kubenswrapper[4988]: I1008 18:29:27.055088 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-gm2vf" event={"ID":"5b6f72e9-8dc6-4290-b032-d8a570b91767","Type":"ContainerStarted","Data":"0445e0aa866fd32b4478f427736e174d9c619b3c4df681e2c48994219d561e07"} Oct 08 18:29:27 crc kubenswrapper[4988]: I1008 18:29:27.057208 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-5v5z9" event={"ID":"c1f2511c-e7bc-4433-910b-eb3b94fcf09b","Type":"ContainerDied","Data":"248d15ff74547a86da7c2dd855b4a5f67def74c5c30b8d05b574681289ed18f8"} Oct 08 18:29:27 crc kubenswrapper[4988]: I1008 18:29:27.057238 4988 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="248d15ff74547a86da7c2dd855b4a5f67def74c5c30b8d05b574681289ed18f8" Oct 08 18:29:27 crc kubenswrapper[4988]: I1008 18:29:27.057261 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-5v5z9" Oct 08 18:29:27 crc kubenswrapper[4988]: I1008 18:29:27.067285 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-c44756996-dfw4k" event={"ID":"f4e47608-89b4-45d1-88a6-fe75805d864d","Type":"ContainerStarted","Data":"b226456d2f92fef2d798a8462e29b0c8f7c8a915dd5330879cb7257e0cc73d05"} Oct 08 18:29:27 crc kubenswrapper[4988]: I1008 18:29:27.067450 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-c44756996-dfw4k" Oct 08 18:29:27 crc kubenswrapper[4988]: I1008 18:29:27.073485 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-6dbc6dc9bd-hk8qw" event={"ID":"26a3b0a9-6aca-490d-a6fb-08addef86fbb","Type":"ContainerStarted","Data":"62d82328b02c86e078d53d3e4ba279c27a7c17092bf7900027389a74f0ad59d0"} Oct 08 18:29:27 crc kubenswrapper[4988]: I1008 18:29:27.073524 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-6dbc6dc9bd-hk8qw" event={"ID":"26a3b0a9-6aca-490d-a6fb-08addef86fbb","Type":"ContainerStarted","Data":"7ee10840af798462ce83a151c6143096bcd55ba704c55b5188685444a407217d"} Oct 08 18:29:27 crc kubenswrapper[4988]: I1008 18:29:27.073673 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/keystone-6dbc6dc9bd-hk8qw" Oct 08 18:29:27 crc kubenswrapper[4988]: I1008 18:29:27.077066 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8fed08cf-45cb-4fd2-92af-9f30513c0b72","Type":"ContainerStarted","Data":"f586d72bfd02084b33e3cbd2ad495a6f4732968d71f628cba78b3edd940ec1f1"} Oct 08 18:29:27 crc kubenswrapper[4988]: I1008 18:29:27.098729 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-c44756996-dfw4k" podStartSLOduration=6.09870587 podStartE2EDuration="6.09870587s" podCreationTimestamp="2025-10-08 18:29:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 18:29:27.08540626 +0000 UTC m=+1112.535249040" watchObservedRunningTime="2025-10-08 18:29:27.09870587 +0000 UTC m=+1112.548548650" Oct 08 18:29:27 crc kubenswrapper[4988]: I1008 18:29:27.112708 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-6dbc6dc9bd-hk8qw" podStartSLOduration=2.112689061 podStartE2EDuration="2.112689061s" podCreationTimestamp="2025-10-08 18:29:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 18:29:27.110090659 +0000 UTC m=+1112.559933429" watchObservedRunningTime="2025-10-08 18:29:27.112689061 +0000 UTC m=+1112.562531831" Oct 08 18:29:27 crc kubenswrapper[4988]: I1008 18:29:27.874053 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-77dd5cf987-5qhqw"] Oct 08 18:29:27 crc kubenswrapper[4988]: E1008 18:29:27.874601 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c1f2511c-e7bc-4433-910b-eb3b94fcf09b" containerName="glance-db-sync" Oct 08 18:29:27 crc kubenswrapper[4988]: I1008 18:29:27.874613 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="c1f2511c-e7bc-4433-910b-eb3b94fcf09b" containerName="glance-db-sync" Oct 08 18:29:27 crc kubenswrapper[4988]: I1008 18:29:27.874797 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="c1f2511c-e7bc-4433-910b-eb3b94fcf09b" containerName="glance-db-sync" Oct 08 18:29:27 crc kubenswrapper[4988]: I1008 18:29:27.875670 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-77dd5cf987-5qhqw" Oct 08 18:29:27 crc kubenswrapper[4988]: I1008 18:29:27.892181 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-77dd5cf987-5qhqw"] Oct 08 18:29:27 crc kubenswrapper[4988]: I1008 18:29:27.951626 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-566fx\" (UniqueName: \"kubernetes.io/projected/33fb9c50-b318-4279-af5c-605bb2f0d3c1-kube-api-access-566fx\") pod \"dnsmasq-dns-77dd5cf987-5qhqw\" (UID: \"33fb9c50-b318-4279-af5c-605bb2f0d3c1\") " pod="openstack/dnsmasq-dns-77dd5cf987-5qhqw" Oct 08 18:29:27 crc kubenswrapper[4988]: I1008 18:29:27.951720 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/33fb9c50-b318-4279-af5c-605bb2f0d3c1-ovsdbserver-nb\") pod \"dnsmasq-dns-77dd5cf987-5qhqw\" (UID: \"33fb9c50-b318-4279-af5c-605bb2f0d3c1\") " pod="openstack/dnsmasq-dns-77dd5cf987-5qhqw" Oct 08 18:29:27 crc kubenswrapper[4988]: I1008 18:29:27.951776 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/33fb9c50-b318-4279-af5c-605bb2f0d3c1-ovsdbserver-sb\") pod \"dnsmasq-dns-77dd5cf987-5qhqw\" (UID: \"33fb9c50-b318-4279-af5c-605bb2f0d3c1\") " pod="openstack/dnsmasq-dns-77dd5cf987-5qhqw" Oct 08 18:29:27 crc kubenswrapper[4988]: I1008 18:29:27.951812 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/33fb9c50-b318-4279-af5c-605bb2f0d3c1-dns-svc\") pod \"dnsmasq-dns-77dd5cf987-5qhqw\" (UID: \"33fb9c50-b318-4279-af5c-605bb2f0d3c1\") " pod="openstack/dnsmasq-dns-77dd5cf987-5qhqw" Oct 08 18:29:27 crc kubenswrapper[4988]: I1008 18:29:27.951834 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/33fb9c50-b318-4279-af5c-605bb2f0d3c1-dns-swift-storage-0\") pod \"dnsmasq-dns-77dd5cf987-5qhqw\" (UID: \"33fb9c50-b318-4279-af5c-605bb2f0d3c1\") " pod="openstack/dnsmasq-dns-77dd5cf987-5qhqw" Oct 08 18:29:27 crc kubenswrapper[4988]: I1008 18:29:27.951911 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/33fb9c50-b318-4279-af5c-605bb2f0d3c1-config\") pod \"dnsmasq-dns-77dd5cf987-5qhqw\" (UID: \"33fb9c50-b318-4279-af5c-605bb2f0d3c1\") " pod="openstack/dnsmasq-dns-77dd5cf987-5qhqw" Oct 08 18:29:28 crc kubenswrapper[4988]: I1008 18:29:28.055544 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/33fb9c50-b318-4279-af5c-605bb2f0d3c1-config\") pod \"dnsmasq-dns-77dd5cf987-5qhqw\" (UID: \"33fb9c50-b318-4279-af5c-605bb2f0d3c1\") " pod="openstack/dnsmasq-dns-77dd5cf987-5qhqw" Oct 08 18:29:28 crc kubenswrapper[4988]: I1008 18:29:28.055650 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-566fx\" (UniqueName: \"kubernetes.io/projected/33fb9c50-b318-4279-af5c-605bb2f0d3c1-kube-api-access-566fx\") pod \"dnsmasq-dns-77dd5cf987-5qhqw\" (UID: \"33fb9c50-b318-4279-af5c-605bb2f0d3c1\") " pod="openstack/dnsmasq-dns-77dd5cf987-5qhqw" Oct 08 18:29:28 crc kubenswrapper[4988]: I1008 18:29:28.055763 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/33fb9c50-b318-4279-af5c-605bb2f0d3c1-ovsdbserver-nb\") pod \"dnsmasq-dns-77dd5cf987-5qhqw\" (UID: \"33fb9c50-b318-4279-af5c-605bb2f0d3c1\") " pod="openstack/dnsmasq-dns-77dd5cf987-5qhqw" Oct 08 18:29:28 crc kubenswrapper[4988]: I1008 18:29:28.055874 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/33fb9c50-b318-4279-af5c-605bb2f0d3c1-ovsdbserver-sb\") pod \"dnsmasq-dns-77dd5cf987-5qhqw\" (UID: \"33fb9c50-b318-4279-af5c-605bb2f0d3c1\") " pod="openstack/dnsmasq-dns-77dd5cf987-5qhqw" Oct 08 18:29:28 crc kubenswrapper[4988]: I1008 18:29:28.055958 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/33fb9c50-b318-4279-af5c-605bb2f0d3c1-dns-svc\") pod \"dnsmasq-dns-77dd5cf987-5qhqw\" (UID: \"33fb9c50-b318-4279-af5c-605bb2f0d3c1\") " pod="openstack/dnsmasq-dns-77dd5cf987-5qhqw" Oct 08 18:29:28 crc kubenswrapper[4988]: I1008 18:29:28.055996 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/33fb9c50-b318-4279-af5c-605bb2f0d3c1-dns-swift-storage-0\") pod \"dnsmasq-dns-77dd5cf987-5qhqw\" (UID: \"33fb9c50-b318-4279-af5c-605bb2f0d3c1\") " pod="openstack/dnsmasq-dns-77dd5cf987-5qhqw" Oct 08 18:29:28 crc kubenswrapper[4988]: I1008 18:29:28.056693 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/33fb9c50-b318-4279-af5c-605bb2f0d3c1-config\") pod \"dnsmasq-dns-77dd5cf987-5qhqw\" (UID: \"33fb9c50-b318-4279-af5c-605bb2f0d3c1\") " pod="openstack/dnsmasq-dns-77dd5cf987-5qhqw" Oct 08 18:29:28 crc kubenswrapper[4988]: I1008 18:29:28.056912 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/33fb9c50-b318-4279-af5c-605bb2f0d3c1-ovsdbserver-nb\") pod \"dnsmasq-dns-77dd5cf987-5qhqw\" (UID: \"33fb9c50-b318-4279-af5c-605bb2f0d3c1\") " pod="openstack/dnsmasq-dns-77dd5cf987-5qhqw" Oct 08 18:29:28 crc kubenswrapper[4988]: I1008 18:29:28.057062 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/33fb9c50-b318-4279-af5c-605bb2f0d3c1-dns-swift-storage-0\") pod \"dnsmasq-dns-77dd5cf987-5qhqw\" (UID: \"33fb9c50-b318-4279-af5c-605bb2f0d3c1\") " pod="openstack/dnsmasq-dns-77dd5cf987-5qhqw" Oct 08 18:29:28 crc kubenswrapper[4988]: I1008 18:29:28.057184 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/33fb9c50-b318-4279-af5c-605bb2f0d3c1-ovsdbserver-sb\") pod \"dnsmasq-dns-77dd5cf987-5qhqw\" (UID: \"33fb9c50-b318-4279-af5c-605bb2f0d3c1\") " pod="openstack/dnsmasq-dns-77dd5cf987-5qhqw" Oct 08 18:29:28 crc kubenswrapper[4988]: I1008 18:29:28.057962 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/33fb9c50-b318-4279-af5c-605bb2f0d3c1-dns-svc\") pod \"dnsmasq-dns-77dd5cf987-5qhqw\" (UID: \"33fb9c50-b318-4279-af5c-605bb2f0d3c1\") " pod="openstack/dnsmasq-dns-77dd5cf987-5qhqw" Oct 08 18:29:28 crc kubenswrapper[4988]: I1008 18:29:28.077930 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-566fx\" (UniqueName: \"kubernetes.io/projected/33fb9c50-b318-4279-af5c-605bb2f0d3c1-kube-api-access-566fx\") pod \"dnsmasq-dns-77dd5cf987-5qhqw\" (UID: \"33fb9c50-b318-4279-af5c-605bb2f0d3c1\") " pod="openstack/dnsmasq-dns-77dd5cf987-5qhqw" Oct 08 18:29:28 crc kubenswrapper[4988]: I1008 18:29:28.085033 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-c44756996-dfw4k" Oct 08 18:29:28 crc kubenswrapper[4988]: I1008 18:29:28.197493 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-77dd5cf987-5qhqw" Oct 08 18:29:28 crc kubenswrapper[4988]: I1008 18:29:28.777044 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Oct 08 18:29:28 crc kubenswrapper[4988]: I1008 18:29:28.779250 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 08 18:29:28 crc kubenswrapper[4988]: I1008 18:29:28.781827 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-scripts" Oct 08 18:29:28 crc kubenswrapper[4988]: I1008 18:29:28.781869 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-5rwjf" Oct 08 18:29:28 crc kubenswrapper[4988]: I1008 18:29:28.782095 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Oct 08 18:29:28 crc kubenswrapper[4988]: I1008 18:29:28.799940 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 08 18:29:28 crc kubenswrapper[4988]: I1008 18:29:28.869849 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1b98c68e-2be9-41fd-a311-8bb8bb74939f-config-data\") pod \"glance-default-external-api-0\" (UID: \"1b98c68e-2be9-41fd-a311-8bb8bb74939f\") " pod="openstack/glance-default-external-api-0" Oct 08 18:29:28 crc kubenswrapper[4988]: I1008 18:29:28.869920 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1b98c68e-2be9-41fd-a311-8bb8bb74939f-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"1b98c68e-2be9-41fd-a311-8bb8bb74939f\") " pod="openstack/glance-default-external-api-0" Oct 08 18:29:28 crc kubenswrapper[4988]: I1008 18:29:28.870000 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1b98c68e-2be9-41fd-a311-8bb8bb74939f-scripts\") pod \"glance-default-external-api-0\" (UID: \"1b98c68e-2be9-41fd-a311-8bb8bb74939f\") " pod="openstack/glance-default-external-api-0" Oct 08 18:29:28 crc kubenswrapper[4988]: I1008 18:29:28.870073 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1b98c68e-2be9-41fd-a311-8bb8bb74939f-logs\") pod \"glance-default-external-api-0\" (UID: \"1b98c68e-2be9-41fd-a311-8bb8bb74939f\") " pod="openstack/glance-default-external-api-0" Oct 08 18:29:28 crc kubenswrapper[4988]: I1008 18:29:28.870238 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xtjp4\" (UniqueName: \"kubernetes.io/projected/1b98c68e-2be9-41fd-a311-8bb8bb74939f-kube-api-access-xtjp4\") pod \"glance-default-external-api-0\" (UID: \"1b98c68e-2be9-41fd-a311-8bb8bb74939f\") " pod="openstack/glance-default-external-api-0" Oct 08 18:29:28 crc kubenswrapper[4988]: I1008 18:29:28.870350 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"glance-default-external-api-0\" (UID: \"1b98c68e-2be9-41fd-a311-8bb8bb74939f\") " pod="openstack/glance-default-external-api-0" Oct 08 18:29:28 crc kubenswrapper[4988]: I1008 18:29:28.870514 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/1b98c68e-2be9-41fd-a311-8bb8bb74939f-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"1b98c68e-2be9-41fd-a311-8bb8bb74939f\") " pod="openstack/glance-default-external-api-0" Oct 08 18:29:28 crc kubenswrapper[4988]: I1008 18:29:28.972949 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"glance-default-external-api-0\" (UID: \"1b98c68e-2be9-41fd-a311-8bb8bb74939f\") " pod="openstack/glance-default-external-api-0" Oct 08 18:29:28 crc kubenswrapper[4988]: I1008 18:29:28.973129 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/1b98c68e-2be9-41fd-a311-8bb8bb74939f-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"1b98c68e-2be9-41fd-a311-8bb8bb74939f\") " pod="openstack/glance-default-external-api-0" Oct 08 18:29:28 crc kubenswrapper[4988]: I1008 18:29:28.973336 4988 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"glance-default-external-api-0\" (UID: \"1b98c68e-2be9-41fd-a311-8bb8bb74939f\") device mount path \"/mnt/openstack/pv11\"" pod="openstack/glance-default-external-api-0" Oct 08 18:29:28 crc kubenswrapper[4988]: I1008 18:29:28.974130 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/1b98c68e-2be9-41fd-a311-8bb8bb74939f-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"1b98c68e-2be9-41fd-a311-8bb8bb74939f\") " pod="openstack/glance-default-external-api-0" Oct 08 18:29:28 crc kubenswrapper[4988]: I1008 18:29:28.975136 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1b98c68e-2be9-41fd-a311-8bb8bb74939f-config-data\") pod \"glance-default-external-api-0\" (UID: \"1b98c68e-2be9-41fd-a311-8bb8bb74939f\") " pod="openstack/glance-default-external-api-0" Oct 08 18:29:28 crc kubenswrapper[4988]: I1008 18:29:28.975250 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1b98c68e-2be9-41fd-a311-8bb8bb74939f-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"1b98c68e-2be9-41fd-a311-8bb8bb74939f\") " pod="openstack/glance-default-external-api-0" Oct 08 18:29:28 crc kubenswrapper[4988]: I1008 18:29:28.975507 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1b98c68e-2be9-41fd-a311-8bb8bb74939f-scripts\") pod \"glance-default-external-api-0\" (UID: \"1b98c68e-2be9-41fd-a311-8bb8bb74939f\") " pod="openstack/glance-default-external-api-0" Oct 08 18:29:28 crc kubenswrapper[4988]: I1008 18:29:28.975795 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1b98c68e-2be9-41fd-a311-8bb8bb74939f-logs\") pod \"glance-default-external-api-0\" (UID: \"1b98c68e-2be9-41fd-a311-8bb8bb74939f\") " pod="openstack/glance-default-external-api-0" Oct 08 18:29:28 crc kubenswrapper[4988]: I1008 18:29:28.975901 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xtjp4\" (UniqueName: \"kubernetes.io/projected/1b98c68e-2be9-41fd-a311-8bb8bb74939f-kube-api-access-xtjp4\") pod \"glance-default-external-api-0\" (UID: \"1b98c68e-2be9-41fd-a311-8bb8bb74939f\") " pod="openstack/glance-default-external-api-0" Oct 08 18:29:28 crc kubenswrapper[4988]: I1008 18:29:28.978541 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1b98c68e-2be9-41fd-a311-8bb8bb74939f-logs\") pod \"glance-default-external-api-0\" (UID: \"1b98c68e-2be9-41fd-a311-8bb8bb74939f\") " pod="openstack/glance-default-external-api-0" Oct 08 18:29:28 crc kubenswrapper[4988]: I1008 18:29:28.986360 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1b98c68e-2be9-41fd-a311-8bb8bb74939f-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"1b98c68e-2be9-41fd-a311-8bb8bb74939f\") " pod="openstack/glance-default-external-api-0" Oct 08 18:29:28 crc kubenswrapper[4988]: I1008 18:29:28.987709 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1b98c68e-2be9-41fd-a311-8bb8bb74939f-scripts\") pod \"glance-default-external-api-0\" (UID: \"1b98c68e-2be9-41fd-a311-8bb8bb74939f\") " pod="openstack/glance-default-external-api-0" Oct 08 18:29:28 crc kubenswrapper[4988]: I1008 18:29:28.987844 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1b98c68e-2be9-41fd-a311-8bb8bb74939f-config-data\") pod \"glance-default-external-api-0\" (UID: \"1b98c68e-2be9-41fd-a311-8bb8bb74939f\") " pod="openstack/glance-default-external-api-0" Oct 08 18:29:28 crc kubenswrapper[4988]: I1008 18:29:28.994872 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xtjp4\" (UniqueName: \"kubernetes.io/projected/1b98c68e-2be9-41fd-a311-8bb8bb74939f-kube-api-access-xtjp4\") pod \"glance-default-external-api-0\" (UID: \"1b98c68e-2be9-41fd-a311-8bb8bb74939f\") " pod="openstack/glance-default-external-api-0" Oct 08 18:29:29 crc kubenswrapper[4988]: I1008 18:29:29.015693 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"glance-default-external-api-0\" (UID: \"1b98c68e-2be9-41fd-a311-8bb8bb74939f\") " pod="openstack/glance-default-external-api-0" Oct 08 18:29:29 crc kubenswrapper[4988]: I1008 18:29:29.046002 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 08 18:29:29 crc kubenswrapper[4988]: I1008 18:29:29.055664 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 08 18:29:29 crc kubenswrapper[4988]: I1008 18:29:29.055916 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 08 18:29:29 crc kubenswrapper[4988]: I1008 18:29:29.058634 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Oct 08 18:29:29 crc kubenswrapper[4988]: I1008 18:29:29.114635 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 08 18:29:29 crc kubenswrapper[4988]: I1008 18:29:29.178945 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-94t5w\" (UniqueName: \"kubernetes.io/projected/d77174cf-9f42-4fc9-83d6-355b2b728289-kube-api-access-94t5w\") pod \"glance-default-internal-api-0\" (UID: \"d77174cf-9f42-4fc9-83d6-355b2b728289\") " pod="openstack/glance-default-internal-api-0" Oct 08 18:29:29 crc kubenswrapper[4988]: I1008 18:29:29.179056 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d77174cf-9f42-4fc9-83d6-355b2b728289-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"d77174cf-9f42-4fc9-83d6-355b2b728289\") " pod="openstack/glance-default-internal-api-0" Oct 08 18:29:29 crc kubenswrapper[4988]: I1008 18:29:29.179114 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d77174cf-9f42-4fc9-83d6-355b2b728289-scripts\") pod \"glance-default-internal-api-0\" (UID: \"d77174cf-9f42-4fc9-83d6-355b2b728289\") " pod="openstack/glance-default-internal-api-0" Oct 08 18:29:29 crc kubenswrapper[4988]: I1008 18:29:29.179162 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/d77174cf-9f42-4fc9-83d6-355b2b728289-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"d77174cf-9f42-4fc9-83d6-355b2b728289\") " pod="openstack/glance-default-internal-api-0" Oct 08 18:29:29 crc kubenswrapper[4988]: I1008 18:29:29.179241 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d77174cf-9f42-4fc9-83d6-355b2b728289-logs\") pod \"glance-default-internal-api-0\" (UID: \"d77174cf-9f42-4fc9-83d6-355b2b728289\") " pod="openstack/glance-default-internal-api-0" Oct 08 18:29:29 crc kubenswrapper[4988]: I1008 18:29:29.179279 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-internal-api-0\" (UID: \"d77174cf-9f42-4fc9-83d6-355b2b728289\") " pod="openstack/glance-default-internal-api-0" Oct 08 18:29:29 crc kubenswrapper[4988]: I1008 18:29:29.179300 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d77174cf-9f42-4fc9-83d6-355b2b728289-config-data\") pod \"glance-default-internal-api-0\" (UID: \"d77174cf-9f42-4fc9-83d6-355b2b728289\") " pod="openstack/glance-default-internal-api-0" Oct 08 18:29:29 crc kubenswrapper[4988]: I1008 18:29:29.281060 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d77174cf-9f42-4fc9-83d6-355b2b728289-logs\") pod \"glance-default-internal-api-0\" (UID: \"d77174cf-9f42-4fc9-83d6-355b2b728289\") " pod="openstack/glance-default-internal-api-0" Oct 08 18:29:29 crc kubenswrapper[4988]: I1008 18:29:29.281373 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-internal-api-0\" (UID: \"d77174cf-9f42-4fc9-83d6-355b2b728289\") " pod="openstack/glance-default-internal-api-0" Oct 08 18:29:29 crc kubenswrapper[4988]: I1008 18:29:29.281421 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d77174cf-9f42-4fc9-83d6-355b2b728289-config-data\") pod \"glance-default-internal-api-0\" (UID: \"d77174cf-9f42-4fc9-83d6-355b2b728289\") " pod="openstack/glance-default-internal-api-0" Oct 08 18:29:29 crc kubenswrapper[4988]: I1008 18:29:29.281489 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-94t5w\" (UniqueName: \"kubernetes.io/projected/d77174cf-9f42-4fc9-83d6-355b2b728289-kube-api-access-94t5w\") pod \"glance-default-internal-api-0\" (UID: \"d77174cf-9f42-4fc9-83d6-355b2b728289\") " pod="openstack/glance-default-internal-api-0" Oct 08 18:29:29 crc kubenswrapper[4988]: I1008 18:29:29.281531 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d77174cf-9f42-4fc9-83d6-355b2b728289-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"d77174cf-9f42-4fc9-83d6-355b2b728289\") " pod="openstack/glance-default-internal-api-0" Oct 08 18:29:29 crc kubenswrapper[4988]: I1008 18:29:29.281575 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d77174cf-9f42-4fc9-83d6-355b2b728289-scripts\") pod \"glance-default-internal-api-0\" (UID: \"d77174cf-9f42-4fc9-83d6-355b2b728289\") " pod="openstack/glance-default-internal-api-0" Oct 08 18:29:29 crc kubenswrapper[4988]: I1008 18:29:29.281619 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/d77174cf-9f42-4fc9-83d6-355b2b728289-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"d77174cf-9f42-4fc9-83d6-355b2b728289\") " pod="openstack/glance-default-internal-api-0" Oct 08 18:29:29 crc kubenswrapper[4988]: I1008 18:29:29.281814 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d77174cf-9f42-4fc9-83d6-355b2b728289-logs\") pod \"glance-default-internal-api-0\" (UID: \"d77174cf-9f42-4fc9-83d6-355b2b728289\") " pod="openstack/glance-default-internal-api-0" Oct 08 18:29:29 crc kubenswrapper[4988]: I1008 18:29:29.282582 4988 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-internal-api-0\" (UID: \"d77174cf-9f42-4fc9-83d6-355b2b728289\") device mount path \"/mnt/openstack/pv05\"" pod="openstack/glance-default-internal-api-0" Oct 08 18:29:29 crc kubenswrapper[4988]: I1008 18:29:29.282916 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/d77174cf-9f42-4fc9-83d6-355b2b728289-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"d77174cf-9f42-4fc9-83d6-355b2b728289\") " pod="openstack/glance-default-internal-api-0" Oct 08 18:29:29 crc kubenswrapper[4988]: I1008 18:29:29.285577 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d77174cf-9f42-4fc9-83d6-355b2b728289-scripts\") pod \"glance-default-internal-api-0\" (UID: \"d77174cf-9f42-4fc9-83d6-355b2b728289\") " pod="openstack/glance-default-internal-api-0" Oct 08 18:29:29 crc kubenswrapper[4988]: I1008 18:29:29.286450 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d77174cf-9f42-4fc9-83d6-355b2b728289-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"d77174cf-9f42-4fc9-83d6-355b2b728289\") " pod="openstack/glance-default-internal-api-0" Oct 08 18:29:29 crc kubenswrapper[4988]: I1008 18:29:29.290377 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d77174cf-9f42-4fc9-83d6-355b2b728289-config-data\") pod \"glance-default-internal-api-0\" (UID: \"d77174cf-9f42-4fc9-83d6-355b2b728289\") " pod="openstack/glance-default-internal-api-0" Oct 08 18:29:29 crc kubenswrapper[4988]: I1008 18:29:29.307183 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-94t5w\" (UniqueName: \"kubernetes.io/projected/d77174cf-9f42-4fc9-83d6-355b2b728289-kube-api-access-94t5w\") pod \"glance-default-internal-api-0\" (UID: \"d77174cf-9f42-4fc9-83d6-355b2b728289\") " pod="openstack/glance-default-internal-api-0" Oct 08 18:29:29 crc kubenswrapper[4988]: I1008 18:29:29.308638 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-internal-api-0\" (UID: \"d77174cf-9f42-4fc9-83d6-355b2b728289\") " pod="openstack/glance-default-internal-api-0" Oct 08 18:29:29 crc kubenswrapper[4988]: I1008 18:29:29.397818 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 08 18:29:29 crc kubenswrapper[4988]: I1008 18:29:29.889901 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 08 18:29:29 crc kubenswrapper[4988]: I1008 18:29:29.945266 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 08 18:29:31 crc kubenswrapper[4988]: I1008 18:29:31.983102 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-77dd5cf987-5qhqw"] Oct 08 18:29:32 crc kubenswrapper[4988]: I1008 18:29:32.060832 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 08 18:29:32 crc kubenswrapper[4988]: W1008 18:29:32.066348 4988 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1b98c68e_2be9_41fd_a311_8bb8bb74939f.slice/crio-dfbbcaf67a6f048a0c393ec47b97825384a8dc0076b8d56cc1d58925eb275a41 WatchSource:0}: Error finding container dfbbcaf67a6f048a0c393ec47b97825384a8dc0076b8d56cc1d58925eb275a41: Status 404 returned error can't find the container with id dfbbcaf67a6f048a0c393ec47b97825384a8dc0076b8d56cc1d58925eb275a41 Oct 08 18:29:32 crc kubenswrapper[4988]: I1008 18:29:32.131765 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"1b98c68e-2be9-41fd-a311-8bb8bb74939f","Type":"ContainerStarted","Data":"dfbbcaf67a6f048a0c393ec47b97825384a8dc0076b8d56cc1d58925eb275a41"} Oct 08 18:29:32 crc kubenswrapper[4988]: I1008 18:29:32.142522 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-dm56p" event={"ID":"d7f4f7aa-ccf2-45e1-8d32-db9684853231","Type":"ContainerStarted","Data":"75ad7b164d49b05e8fa5d093445ac8c6b637d6459bbbacb342ee57317efae57a"} Oct 08 18:29:32 crc kubenswrapper[4988]: I1008 18:29:32.147724 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-77dd5cf987-5qhqw" event={"ID":"33fb9c50-b318-4279-af5c-605bb2f0d3c1","Type":"ContainerStarted","Data":"2f487ca795d8c82b64f5a7ae4058c27ee5391b6dcd345e70c10bf248546580d9"} Oct 08 18:29:32 crc kubenswrapper[4988]: I1008 18:29:32.167623 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-db-sync-dm56p" podStartSLOduration=5.28406862 podStartE2EDuration="12.167602602s" podCreationTimestamp="2025-10-08 18:29:20 +0000 UTC" firstStartedPulling="2025-10-08 18:29:24.560882172 +0000 UTC m=+1110.010724942" lastFinishedPulling="2025-10-08 18:29:31.444416154 +0000 UTC m=+1116.894258924" observedRunningTime="2025-10-08 18:29:32.158577867 +0000 UTC m=+1117.608420637" watchObservedRunningTime="2025-10-08 18:29:32.167602602 +0000 UTC m=+1117.617445362" Oct 08 18:29:32 crc kubenswrapper[4988]: I1008 18:29:32.756134 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 08 18:29:32 crc kubenswrapper[4988]: W1008 18:29:32.759189 4988 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd77174cf_9f42_4fc9_83d6_355b2b728289.slice/crio-104765904ff68e5a4e1baba7586655b91a823e990a9fd4fc19decf1edd61aa50 WatchSource:0}: Error finding container 104765904ff68e5a4e1baba7586655b91a823e990a9fd4fc19decf1edd61aa50: Status 404 returned error can't find the container with id 104765904ff68e5a4e1baba7586655b91a823e990a9fd4fc19decf1edd61aa50 Oct 08 18:29:33 crc kubenswrapper[4988]: I1008 18:29:33.163705 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"1b98c68e-2be9-41fd-a311-8bb8bb74939f","Type":"ContainerStarted","Data":"c2c1cc04e03bf2c844c4b2b292dd8b58cae0aeb74871b25ff6675b936a938eed"} Oct 08 18:29:33 crc kubenswrapper[4988]: I1008 18:29:33.163760 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"1b98c68e-2be9-41fd-a311-8bb8bb74939f","Type":"ContainerStarted","Data":"ae8d3969526856e7c45e181fb9448caddd6ee057b902ae77af235edd0b24b74a"} Oct 08 18:29:33 crc kubenswrapper[4988]: I1008 18:29:33.163844 4988 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="1b98c68e-2be9-41fd-a311-8bb8bb74939f" containerName="glance-log" containerID="cri-o://ae8d3969526856e7c45e181fb9448caddd6ee057b902ae77af235edd0b24b74a" gracePeriod=30 Oct 08 18:29:33 crc kubenswrapper[4988]: I1008 18:29:33.164122 4988 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="1b98c68e-2be9-41fd-a311-8bb8bb74939f" containerName="glance-httpd" containerID="cri-o://c2c1cc04e03bf2c844c4b2b292dd8b58cae0aeb74871b25ff6675b936a938eed" gracePeriod=30 Oct 08 18:29:33 crc kubenswrapper[4988]: I1008 18:29:33.177160 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"d77174cf-9f42-4fc9-83d6-355b2b728289","Type":"ContainerStarted","Data":"104765904ff68e5a4e1baba7586655b91a823e990a9fd4fc19decf1edd61aa50"} Oct 08 18:29:33 crc kubenswrapper[4988]: I1008 18:29:33.180006 4988 generic.go:334] "Generic (PLEG): container finished" podID="33fb9c50-b318-4279-af5c-605bb2f0d3c1" containerID="32e6bd21bb3120a82182cffbe11ccd289e40b61aa9ea29f4916e76d61813e4d4" exitCode=0 Oct 08 18:29:33 crc kubenswrapper[4988]: I1008 18:29:33.180099 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-77dd5cf987-5qhqw" event={"ID":"33fb9c50-b318-4279-af5c-605bb2f0d3c1","Type":"ContainerDied","Data":"32e6bd21bb3120a82182cffbe11ccd289e40b61aa9ea29f4916e76d61813e4d4"} Oct 08 18:29:33 crc kubenswrapper[4988]: I1008 18:29:33.198220 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=6.198204207 podStartE2EDuration="6.198204207s" podCreationTimestamp="2025-10-08 18:29:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 18:29:33.193696725 +0000 UTC m=+1118.643539495" watchObservedRunningTime="2025-10-08 18:29:33.198204207 +0000 UTC m=+1118.648046977" Oct 08 18:29:33 crc kubenswrapper[4988]: I1008 18:29:33.779054 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 08 18:29:33 crc kubenswrapper[4988]: I1008 18:29:33.874955 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1b98c68e-2be9-41fd-a311-8bb8bb74939f-config-data\") pod \"1b98c68e-2be9-41fd-a311-8bb8bb74939f\" (UID: \"1b98c68e-2be9-41fd-a311-8bb8bb74939f\") " Oct 08 18:29:33 crc kubenswrapper[4988]: I1008 18:29:33.875727 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xtjp4\" (UniqueName: \"kubernetes.io/projected/1b98c68e-2be9-41fd-a311-8bb8bb74939f-kube-api-access-xtjp4\") pod \"1b98c68e-2be9-41fd-a311-8bb8bb74939f\" (UID: \"1b98c68e-2be9-41fd-a311-8bb8bb74939f\") " Oct 08 18:29:33 crc kubenswrapper[4988]: I1008 18:29:33.875765 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1b98c68e-2be9-41fd-a311-8bb8bb74939f-combined-ca-bundle\") pod \"1b98c68e-2be9-41fd-a311-8bb8bb74939f\" (UID: \"1b98c68e-2be9-41fd-a311-8bb8bb74939f\") " Oct 08 18:29:33 crc kubenswrapper[4988]: I1008 18:29:33.875814 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1b98c68e-2be9-41fd-a311-8bb8bb74939f-logs\") pod \"1b98c68e-2be9-41fd-a311-8bb8bb74939f\" (UID: \"1b98c68e-2be9-41fd-a311-8bb8bb74939f\") " Oct 08 18:29:33 crc kubenswrapper[4988]: I1008 18:29:33.875913 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/1b98c68e-2be9-41fd-a311-8bb8bb74939f-httpd-run\") pod \"1b98c68e-2be9-41fd-a311-8bb8bb74939f\" (UID: \"1b98c68e-2be9-41fd-a311-8bb8bb74939f\") " Oct 08 18:29:33 crc kubenswrapper[4988]: I1008 18:29:33.876354 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1b98c68e-2be9-41fd-a311-8bb8bb74939f-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "1b98c68e-2be9-41fd-a311-8bb8bb74939f" (UID: "1b98c68e-2be9-41fd-a311-8bb8bb74939f"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 18:29:33 crc kubenswrapper[4988]: I1008 18:29:33.876418 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1b98c68e-2be9-41fd-a311-8bb8bb74939f-scripts\") pod \"1b98c68e-2be9-41fd-a311-8bb8bb74939f\" (UID: \"1b98c68e-2be9-41fd-a311-8bb8bb74939f\") " Oct 08 18:29:33 crc kubenswrapper[4988]: I1008 18:29:33.876285 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1b98c68e-2be9-41fd-a311-8bb8bb74939f-logs" (OuterVolumeSpecName: "logs") pod "1b98c68e-2be9-41fd-a311-8bb8bb74939f" (UID: "1b98c68e-2be9-41fd-a311-8bb8bb74939f"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 18:29:33 crc kubenswrapper[4988]: I1008 18:29:33.876470 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"1b98c68e-2be9-41fd-a311-8bb8bb74939f\" (UID: \"1b98c68e-2be9-41fd-a311-8bb8bb74939f\") " Oct 08 18:29:33 crc kubenswrapper[4988]: I1008 18:29:33.877283 4988 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/1b98c68e-2be9-41fd-a311-8bb8bb74939f-httpd-run\") on node \"crc\" DevicePath \"\"" Oct 08 18:29:33 crc kubenswrapper[4988]: I1008 18:29:33.877299 4988 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1b98c68e-2be9-41fd-a311-8bb8bb74939f-logs\") on node \"crc\" DevicePath \"\"" Oct 08 18:29:33 crc kubenswrapper[4988]: I1008 18:29:33.882503 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage11-crc" (OuterVolumeSpecName: "glance") pod "1b98c68e-2be9-41fd-a311-8bb8bb74939f" (UID: "1b98c68e-2be9-41fd-a311-8bb8bb74939f"). InnerVolumeSpecName "local-storage11-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 08 18:29:33 crc kubenswrapper[4988]: I1008 18:29:33.882638 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1b98c68e-2be9-41fd-a311-8bb8bb74939f-scripts" (OuterVolumeSpecName: "scripts") pod "1b98c68e-2be9-41fd-a311-8bb8bb74939f" (UID: "1b98c68e-2be9-41fd-a311-8bb8bb74939f"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:29:33 crc kubenswrapper[4988]: I1008 18:29:33.885467 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1b98c68e-2be9-41fd-a311-8bb8bb74939f-kube-api-access-xtjp4" (OuterVolumeSpecName: "kube-api-access-xtjp4") pod "1b98c68e-2be9-41fd-a311-8bb8bb74939f" (UID: "1b98c68e-2be9-41fd-a311-8bb8bb74939f"). InnerVolumeSpecName "kube-api-access-xtjp4". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:29:33 crc kubenswrapper[4988]: I1008 18:29:33.905456 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1b98c68e-2be9-41fd-a311-8bb8bb74939f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "1b98c68e-2be9-41fd-a311-8bb8bb74939f" (UID: "1b98c68e-2be9-41fd-a311-8bb8bb74939f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:29:33 crc kubenswrapper[4988]: I1008 18:29:33.930641 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1b98c68e-2be9-41fd-a311-8bb8bb74939f-config-data" (OuterVolumeSpecName: "config-data") pod "1b98c68e-2be9-41fd-a311-8bb8bb74939f" (UID: "1b98c68e-2be9-41fd-a311-8bb8bb74939f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:29:33 crc kubenswrapper[4988]: I1008 18:29:33.978588 4988 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xtjp4\" (UniqueName: \"kubernetes.io/projected/1b98c68e-2be9-41fd-a311-8bb8bb74939f-kube-api-access-xtjp4\") on node \"crc\" DevicePath \"\"" Oct 08 18:29:33 crc kubenswrapper[4988]: I1008 18:29:33.978624 4988 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1b98c68e-2be9-41fd-a311-8bb8bb74939f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 18:29:33 crc kubenswrapper[4988]: I1008 18:29:33.978632 4988 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1b98c68e-2be9-41fd-a311-8bb8bb74939f-scripts\") on node \"crc\" DevicePath \"\"" Oct 08 18:29:33 crc kubenswrapper[4988]: I1008 18:29:33.978665 4988 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") on node \"crc\" " Oct 08 18:29:33 crc kubenswrapper[4988]: I1008 18:29:33.978676 4988 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1b98c68e-2be9-41fd-a311-8bb8bb74939f-config-data\") on node \"crc\" DevicePath \"\"" Oct 08 18:29:33 crc kubenswrapper[4988]: I1008 18:29:33.998656 4988 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage11-crc" (UniqueName: "kubernetes.io/local-volume/local-storage11-crc") on node "crc" Oct 08 18:29:34 crc kubenswrapper[4988]: I1008 18:29:34.080508 4988 reconciler_common.go:293] "Volume detached for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") on node \"crc\" DevicePath \"\"" Oct 08 18:29:34 crc kubenswrapper[4988]: I1008 18:29:34.192656 4988 generic.go:334] "Generic (PLEG): container finished" podID="1b98c68e-2be9-41fd-a311-8bb8bb74939f" containerID="c2c1cc04e03bf2c844c4b2b292dd8b58cae0aeb74871b25ff6675b936a938eed" exitCode=143 Oct 08 18:29:34 crc kubenswrapper[4988]: I1008 18:29:34.192698 4988 generic.go:334] "Generic (PLEG): container finished" podID="1b98c68e-2be9-41fd-a311-8bb8bb74939f" containerID="ae8d3969526856e7c45e181fb9448caddd6ee057b902ae77af235edd0b24b74a" exitCode=143 Oct 08 18:29:34 crc kubenswrapper[4988]: I1008 18:29:34.192831 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 08 18:29:34 crc kubenswrapper[4988]: I1008 18:29:34.193408 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"1b98c68e-2be9-41fd-a311-8bb8bb74939f","Type":"ContainerDied","Data":"c2c1cc04e03bf2c844c4b2b292dd8b58cae0aeb74871b25ff6675b936a938eed"} Oct 08 18:29:34 crc kubenswrapper[4988]: I1008 18:29:34.193457 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"1b98c68e-2be9-41fd-a311-8bb8bb74939f","Type":"ContainerDied","Data":"ae8d3969526856e7c45e181fb9448caddd6ee057b902ae77af235edd0b24b74a"} Oct 08 18:29:34 crc kubenswrapper[4988]: I1008 18:29:34.193478 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"1b98c68e-2be9-41fd-a311-8bb8bb74939f","Type":"ContainerDied","Data":"dfbbcaf67a6f048a0c393ec47b97825384a8dc0076b8d56cc1d58925eb275a41"} Oct 08 18:29:34 crc kubenswrapper[4988]: I1008 18:29:34.193498 4988 scope.go:117] "RemoveContainer" containerID="c2c1cc04e03bf2c844c4b2b292dd8b58cae0aeb74871b25ff6675b936a938eed" Oct 08 18:29:34 crc kubenswrapper[4988]: I1008 18:29:34.197822 4988 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="d77174cf-9f42-4fc9-83d6-355b2b728289" containerName="glance-log" containerID="cri-o://a0ac407571d8352f322c503b5844b1316dfe3d08f823d141074c056a397d2684" gracePeriod=30 Oct 08 18:29:34 crc kubenswrapper[4988]: I1008 18:29:34.197853 4988 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="d77174cf-9f42-4fc9-83d6-355b2b728289" containerName="glance-httpd" containerID="cri-o://5a3e3b94cc7cf1dee8c248bb8567bd97c7e17296edb15d81cfce00dd723be9b4" gracePeriod=30 Oct 08 18:29:34 crc kubenswrapper[4988]: I1008 18:29:34.197750 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"d77174cf-9f42-4fc9-83d6-355b2b728289","Type":"ContainerStarted","Data":"5a3e3b94cc7cf1dee8c248bb8567bd97c7e17296edb15d81cfce00dd723be9b4"} Oct 08 18:29:34 crc kubenswrapper[4988]: I1008 18:29:34.197964 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"d77174cf-9f42-4fc9-83d6-355b2b728289","Type":"ContainerStarted","Data":"a0ac407571d8352f322c503b5844b1316dfe3d08f823d141074c056a397d2684"} Oct 08 18:29:34 crc kubenswrapper[4988]: I1008 18:29:34.201564 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-77dd5cf987-5qhqw" event={"ID":"33fb9c50-b318-4279-af5c-605bb2f0d3c1","Type":"ContainerStarted","Data":"8ac057a4f4bc0915ef6b7c8284d0fb581b4eadb20d3d7d433254afafbe52a8a2"} Oct 08 18:29:34 crc kubenswrapper[4988]: I1008 18:29:34.202324 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-77dd5cf987-5qhqw" Oct 08 18:29:34 crc kubenswrapper[4988]: I1008 18:29:34.223454 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=7.22336948 podStartE2EDuration="7.22336948s" podCreationTimestamp="2025-10-08 18:29:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 18:29:34.217747843 +0000 UTC m=+1119.667590633" watchObservedRunningTime="2025-10-08 18:29:34.22336948 +0000 UTC m=+1119.673212250" Oct 08 18:29:34 crc kubenswrapper[4988]: I1008 18:29:34.277441 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 08 18:29:34 crc kubenswrapper[4988]: I1008 18:29:34.304467 4988 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 08 18:29:34 crc kubenswrapper[4988]: I1008 18:29:34.307839 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-77dd5cf987-5qhqw" podStartSLOduration=7.307824812 podStartE2EDuration="7.307824812s" podCreationTimestamp="2025-10-08 18:29:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 18:29:34.270501565 +0000 UTC m=+1119.720344345" watchObservedRunningTime="2025-10-08 18:29:34.307824812 +0000 UTC m=+1119.757667582" Oct 08 18:29:34 crc kubenswrapper[4988]: I1008 18:29:34.321322 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Oct 08 18:29:34 crc kubenswrapper[4988]: E1008 18:29:34.338823 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1b98c68e-2be9-41fd-a311-8bb8bb74939f" containerName="glance-httpd" Oct 08 18:29:34 crc kubenswrapper[4988]: I1008 18:29:34.338856 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="1b98c68e-2be9-41fd-a311-8bb8bb74939f" containerName="glance-httpd" Oct 08 18:29:34 crc kubenswrapper[4988]: E1008 18:29:34.338907 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1b98c68e-2be9-41fd-a311-8bb8bb74939f" containerName="glance-log" Oct 08 18:29:34 crc kubenswrapper[4988]: I1008 18:29:34.338915 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="1b98c68e-2be9-41fd-a311-8bb8bb74939f" containerName="glance-log" Oct 08 18:29:34 crc kubenswrapper[4988]: I1008 18:29:34.339210 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="1b98c68e-2be9-41fd-a311-8bb8bb74939f" containerName="glance-log" Oct 08 18:29:34 crc kubenswrapper[4988]: I1008 18:29:34.339225 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="1b98c68e-2be9-41fd-a311-8bb8bb74939f" containerName="glance-httpd" Oct 08 18:29:34 crc kubenswrapper[4988]: I1008 18:29:34.341664 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 08 18:29:34 crc kubenswrapper[4988]: I1008 18:29:34.344466 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Oct 08 18:29:34 crc kubenswrapper[4988]: I1008 18:29:34.344733 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 08 18:29:34 crc kubenswrapper[4988]: I1008 18:29:34.345070 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Oct 08 18:29:34 crc kubenswrapper[4988]: I1008 18:29:34.492892 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b98d358d-c785-4494-932c-1d6662944b74-config-data\") pod \"glance-default-external-api-0\" (UID: \"b98d358d-c785-4494-932c-1d6662944b74\") " pod="openstack/glance-default-external-api-0" Oct 08 18:29:34 crc kubenswrapper[4988]: I1008 18:29:34.492966 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b98d358d-c785-4494-932c-1d6662944b74-scripts\") pod \"glance-default-external-api-0\" (UID: \"b98d358d-c785-4494-932c-1d6662944b74\") " pod="openstack/glance-default-external-api-0" Oct 08 18:29:34 crc kubenswrapper[4988]: I1008 18:29:34.492992 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"glance-default-external-api-0\" (UID: \"b98d358d-c785-4494-932c-1d6662944b74\") " pod="openstack/glance-default-external-api-0" Oct 08 18:29:34 crc kubenswrapper[4988]: I1008 18:29:34.493019 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b98d358d-c785-4494-932c-1d6662944b74-logs\") pod \"glance-default-external-api-0\" (UID: \"b98d358d-c785-4494-932c-1d6662944b74\") " pod="openstack/glance-default-external-api-0" Oct 08 18:29:34 crc kubenswrapper[4988]: I1008 18:29:34.493051 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/b98d358d-c785-4494-932c-1d6662944b74-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"b98d358d-c785-4494-932c-1d6662944b74\") " pod="openstack/glance-default-external-api-0" Oct 08 18:29:34 crc kubenswrapper[4988]: I1008 18:29:34.493084 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-df76b\" (UniqueName: \"kubernetes.io/projected/b98d358d-c785-4494-932c-1d6662944b74-kube-api-access-df76b\") pod \"glance-default-external-api-0\" (UID: \"b98d358d-c785-4494-932c-1d6662944b74\") " pod="openstack/glance-default-external-api-0" Oct 08 18:29:34 crc kubenswrapper[4988]: I1008 18:29:34.493105 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/b98d358d-c785-4494-932c-1d6662944b74-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"b98d358d-c785-4494-932c-1d6662944b74\") " pod="openstack/glance-default-external-api-0" Oct 08 18:29:34 crc kubenswrapper[4988]: I1008 18:29:34.493149 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b98d358d-c785-4494-932c-1d6662944b74-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"b98d358d-c785-4494-932c-1d6662944b74\") " pod="openstack/glance-default-external-api-0" Oct 08 18:29:34 crc kubenswrapper[4988]: I1008 18:29:34.595019 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b98d358d-c785-4494-932c-1d6662944b74-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"b98d358d-c785-4494-932c-1d6662944b74\") " pod="openstack/glance-default-external-api-0" Oct 08 18:29:34 crc kubenswrapper[4988]: I1008 18:29:34.595122 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b98d358d-c785-4494-932c-1d6662944b74-config-data\") pod \"glance-default-external-api-0\" (UID: \"b98d358d-c785-4494-932c-1d6662944b74\") " pod="openstack/glance-default-external-api-0" Oct 08 18:29:34 crc kubenswrapper[4988]: I1008 18:29:34.595217 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b98d358d-c785-4494-932c-1d6662944b74-scripts\") pod \"glance-default-external-api-0\" (UID: \"b98d358d-c785-4494-932c-1d6662944b74\") " pod="openstack/glance-default-external-api-0" Oct 08 18:29:34 crc kubenswrapper[4988]: I1008 18:29:34.595276 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"glance-default-external-api-0\" (UID: \"b98d358d-c785-4494-932c-1d6662944b74\") " pod="openstack/glance-default-external-api-0" Oct 08 18:29:34 crc kubenswrapper[4988]: I1008 18:29:34.595335 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b98d358d-c785-4494-932c-1d6662944b74-logs\") pod \"glance-default-external-api-0\" (UID: \"b98d358d-c785-4494-932c-1d6662944b74\") " pod="openstack/glance-default-external-api-0" Oct 08 18:29:34 crc kubenswrapper[4988]: I1008 18:29:34.595395 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/b98d358d-c785-4494-932c-1d6662944b74-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"b98d358d-c785-4494-932c-1d6662944b74\") " pod="openstack/glance-default-external-api-0" Oct 08 18:29:34 crc kubenswrapper[4988]: I1008 18:29:34.595433 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-df76b\" (UniqueName: \"kubernetes.io/projected/b98d358d-c785-4494-932c-1d6662944b74-kube-api-access-df76b\") pod \"glance-default-external-api-0\" (UID: \"b98d358d-c785-4494-932c-1d6662944b74\") " pod="openstack/glance-default-external-api-0" Oct 08 18:29:34 crc kubenswrapper[4988]: I1008 18:29:34.595454 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/b98d358d-c785-4494-932c-1d6662944b74-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"b98d358d-c785-4494-932c-1d6662944b74\") " pod="openstack/glance-default-external-api-0" Oct 08 18:29:34 crc kubenswrapper[4988]: I1008 18:29:34.595857 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/b98d358d-c785-4494-932c-1d6662944b74-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"b98d358d-c785-4494-932c-1d6662944b74\") " pod="openstack/glance-default-external-api-0" Oct 08 18:29:34 crc kubenswrapper[4988]: I1008 18:29:34.595963 4988 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"glance-default-external-api-0\" (UID: \"b98d358d-c785-4494-932c-1d6662944b74\") device mount path \"/mnt/openstack/pv11\"" pod="openstack/glance-default-external-api-0" Oct 08 18:29:34 crc kubenswrapper[4988]: I1008 18:29:34.596195 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b98d358d-c785-4494-932c-1d6662944b74-logs\") pod \"glance-default-external-api-0\" (UID: \"b98d358d-c785-4494-932c-1d6662944b74\") " pod="openstack/glance-default-external-api-0" Oct 08 18:29:34 crc kubenswrapper[4988]: I1008 18:29:34.600234 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b98d358d-c785-4494-932c-1d6662944b74-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"b98d358d-c785-4494-932c-1d6662944b74\") " pod="openstack/glance-default-external-api-0" Oct 08 18:29:34 crc kubenswrapper[4988]: I1008 18:29:34.600594 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b98d358d-c785-4494-932c-1d6662944b74-config-data\") pod \"glance-default-external-api-0\" (UID: \"b98d358d-c785-4494-932c-1d6662944b74\") " pod="openstack/glance-default-external-api-0" Oct 08 18:29:34 crc kubenswrapper[4988]: I1008 18:29:34.605773 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/b98d358d-c785-4494-932c-1d6662944b74-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"b98d358d-c785-4494-932c-1d6662944b74\") " pod="openstack/glance-default-external-api-0" Oct 08 18:29:34 crc kubenswrapper[4988]: I1008 18:29:34.605853 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b98d358d-c785-4494-932c-1d6662944b74-scripts\") pod \"glance-default-external-api-0\" (UID: \"b98d358d-c785-4494-932c-1d6662944b74\") " pod="openstack/glance-default-external-api-0" Oct 08 18:29:34 crc kubenswrapper[4988]: I1008 18:29:34.613651 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-df76b\" (UniqueName: \"kubernetes.io/projected/b98d358d-c785-4494-932c-1d6662944b74-kube-api-access-df76b\") pod \"glance-default-external-api-0\" (UID: \"b98d358d-c785-4494-932c-1d6662944b74\") " pod="openstack/glance-default-external-api-0" Oct 08 18:29:34 crc kubenswrapper[4988]: I1008 18:29:34.634893 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"glance-default-external-api-0\" (UID: \"b98d358d-c785-4494-932c-1d6662944b74\") " pod="openstack/glance-default-external-api-0" Oct 08 18:29:34 crc kubenswrapper[4988]: I1008 18:29:34.717928 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 08 18:29:35 crc kubenswrapper[4988]: I1008 18:29:35.212785 4988 generic.go:334] "Generic (PLEG): container finished" podID="d77174cf-9f42-4fc9-83d6-355b2b728289" containerID="5a3e3b94cc7cf1dee8c248bb8567bd97c7e17296edb15d81cfce00dd723be9b4" exitCode=0 Oct 08 18:29:35 crc kubenswrapper[4988]: I1008 18:29:35.213135 4988 generic.go:334] "Generic (PLEG): container finished" podID="d77174cf-9f42-4fc9-83d6-355b2b728289" containerID="a0ac407571d8352f322c503b5844b1316dfe3d08f823d141074c056a397d2684" exitCode=143 Oct 08 18:29:35 crc kubenswrapper[4988]: I1008 18:29:35.212891 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"d77174cf-9f42-4fc9-83d6-355b2b728289","Type":"ContainerDied","Data":"5a3e3b94cc7cf1dee8c248bb8567bd97c7e17296edb15d81cfce00dd723be9b4"} Oct 08 18:29:35 crc kubenswrapper[4988]: I1008 18:29:35.213195 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"d77174cf-9f42-4fc9-83d6-355b2b728289","Type":"ContainerDied","Data":"a0ac407571d8352f322c503b5844b1316dfe3d08f823d141074c056a397d2684"} Oct 08 18:29:35 crc kubenswrapper[4988]: I1008 18:29:35.254555 4988 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1b98c68e-2be9-41fd-a311-8bb8bb74939f" path="/var/lib/kubelet/pods/1b98c68e-2be9-41fd-a311-8bb8bb74939f/volumes" Oct 08 18:29:37 crc kubenswrapper[4988]: I1008 18:29:37.784124 4988 scope.go:117] "RemoveContainer" containerID="ae8d3969526856e7c45e181fb9448caddd6ee057b902ae77af235edd0b24b74a" Oct 08 18:29:38 crc kubenswrapper[4988]: I1008 18:29:38.198611 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-77dd5cf987-5qhqw" Oct 08 18:29:38 crc kubenswrapper[4988]: I1008 18:29:38.283847 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-69894dfcd9-948wp"] Oct 08 18:29:38 crc kubenswrapper[4988]: I1008 18:29:38.284463 4988 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-69894dfcd9-948wp" podUID="7a831457-b429-4339-b166-afed001c770d" containerName="dnsmasq-dns" containerID="cri-o://ac3f09e0a805492ad71b338aa6dd023325b8ab08859f1ce9590b691ca0ede0e9" gracePeriod=10 Oct 08 18:29:39 crc kubenswrapper[4988]: I1008 18:29:39.267908 4988 generic.go:334] "Generic (PLEG): container finished" podID="7a831457-b429-4339-b166-afed001c770d" containerID="ac3f09e0a805492ad71b338aa6dd023325b8ab08859f1ce9590b691ca0ede0e9" exitCode=0 Oct 08 18:29:39 crc kubenswrapper[4988]: I1008 18:29:39.267987 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-69894dfcd9-948wp" event={"ID":"7a831457-b429-4339-b166-afed001c770d","Type":"ContainerDied","Data":"ac3f09e0a805492ad71b338aa6dd023325b8ab08859f1ce9590b691ca0ede0e9"} Oct 08 18:29:39 crc kubenswrapper[4988]: I1008 18:29:39.270312 4988 generic.go:334] "Generic (PLEG): container finished" podID="d7f4f7aa-ccf2-45e1-8d32-db9684853231" containerID="75ad7b164d49b05e8fa5d093445ac8c6b637d6459bbbacb342ee57317efae57a" exitCode=0 Oct 08 18:29:39 crc kubenswrapper[4988]: I1008 18:29:39.270339 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-dm56p" event={"ID":"d7f4f7aa-ccf2-45e1-8d32-db9684853231","Type":"ContainerDied","Data":"75ad7b164d49b05e8fa5d093445ac8c6b637d6459bbbacb342ee57317efae57a"} Oct 08 18:29:42 crc kubenswrapper[4988]: I1008 18:29:42.713913 4988 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-69894dfcd9-948wp" podUID="7a831457-b429-4339-b166-afed001c770d" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.136:5353: connect: connection refused" Oct 08 18:29:47 crc kubenswrapper[4988]: E1008 18:29:47.259090 4988 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/ubi9/httpd-24@sha256:e91d58021b54c46883595ff66be65882de54abdb3be2ca53c4162b20d18b5f48" Oct 08 18:29:47 crc kubenswrapper[4988]: E1008 18:29:47.260087 4988 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:proxy-httpd,Image:registry.redhat.io/ubi9/httpd-24@sha256:e91d58021b54c46883595ff66be65882de54abdb3be2ca53c4162b20d18b5f48,Command:[/usr/sbin/httpd],Args:[-DFOREGROUND],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:proxy-httpd,HostPort:0,ContainerPort:3000,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/httpd/conf/httpd.conf,SubPath:httpd.conf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/httpd/conf.d/ssl.conf,SubPath:ssl.conf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:run-httpd,ReadOnly:false,MountPath:/run/httpd,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:log-httpd,ReadOnly:false,MountPath:/var/log/httpd,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-4xpgf,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/,Port:{0 3000 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:300,TimeoutSeconds:30,PeriodSeconds:30,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/,Port:{0 3000 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:10,TimeoutSeconds:30,PeriodSeconds:30,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*0,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ceilometer-0_openstack(8fed08cf-45cb-4fd2-92af-9f30513c0b72): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Oct 08 18:29:47 crc kubenswrapper[4988]: E1008 18:29:47.261500 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"proxy-httpd\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openstack/ceilometer-0" podUID="8fed08cf-45cb-4fd2-92af-9f30513c0b72" Oct 08 18:29:47 crc kubenswrapper[4988]: I1008 18:29:47.268993 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-dm56p" Oct 08 18:29:47 crc kubenswrapper[4988]: I1008 18:29:47.278144 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 08 18:29:47 crc kubenswrapper[4988]: I1008 18:29:47.365715 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 08 18:29:47 crc kubenswrapper[4988]: I1008 18:29:47.365711 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"d77174cf-9f42-4fc9-83d6-355b2b728289","Type":"ContainerDied","Data":"104765904ff68e5a4e1baba7586655b91a823e990a9fd4fc19decf1edd61aa50"} Oct 08 18:29:47 crc kubenswrapper[4988]: I1008 18:29:47.373180 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-dm56p" event={"ID":"d7f4f7aa-ccf2-45e1-8d32-db9684853231","Type":"ContainerDied","Data":"932a4ebfb7ff061842fd0d0d71f8410a08a92a4f1aca818d7b9f074929ab0631"} Oct 08 18:29:47 crc kubenswrapper[4988]: I1008 18:29:47.373227 4988 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="932a4ebfb7ff061842fd0d0d71f8410a08a92a4f1aca818d7b9f074929ab0631" Oct 08 18:29:47 crc kubenswrapper[4988]: I1008 18:29:47.373242 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-dm56p" Oct 08 18:29:47 crc kubenswrapper[4988]: I1008 18:29:47.373287 4988 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="8fed08cf-45cb-4fd2-92af-9f30513c0b72" containerName="ceilometer-central-agent" containerID="cri-o://530180f4bca5eaf8c14c742c890bbe065b3ab6a329ac14ec714a341c4b5e100e" gracePeriod=30 Oct 08 18:29:47 crc kubenswrapper[4988]: I1008 18:29:47.373351 4988 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="8fed08cf-45cb-4fd2-92af-9f30513c0b72" containerName="sg-core" containerID="cri-o://f586d72bfd02084b33e3cbd2ad495a6f4732968d71f628cba78b3edd940ec1f1" gracePeriod=30 Oct 08 18:29:47 crc kubenswrapper[4988]: I1008 18:29:47.373428 4988 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="8fed08cf-45cb-4fd2-92af-9f30513c0b72" containerName="ceilometer-notification-agent" containerID="cri-o://2b5c761d77af17eace93d9cd355762d47e5b9791b61461447ac5dfb3542405b0" gracePeriod=30 Oct 08 18:29:47 crc kubenswrapper[4988]: I1008 18:29:47.416072 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/d77174cf-9f42-4fc9-83d6-355b2b728289-httpd-run\") pod \"d77174cf-9f42-4fc9-83d6-355b2b728289\" (UID: \"d77174cf-9f42-4fc9-83d6-355b2b728289\") " Oct 08 18:29:47 crc kubenswrapper[4988]: I1008 18:29:47.416451 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d77174cf-9f42-4fc9-83d6-355b2b728289-scripts\") pod \"d77174cf-9f42-4fc9-83d6-355b2b728289\" (UID: \"d77174cf-9f42-4fc9-83d6-355b2b728289\") " Oct 08 18:29:47 crc kubenswrapper[4988]: I1008 18:29:47.416480 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d77174cf-9f42-4fc9-83d6-355b2b728289-combined-ca-bundle\") pod \"d77174cf-9f42-4fc9-83d6-355b2b728289\" (UID: \"d77174cf-9f42-4fc9-83d6-355b2b728289\") " Oct 08 18:29:47 crc kubenswrapper[4988]: I1008 18:29:47.416501 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-94t5w\" (UniqueName: \"kubernetes.io/projected/d77174cf-9f42-4fc9-83d6-355b2b728289-kube-api-access-94t5w\") pod \"d77174cf-9f42-4fc9-83d6-355b2b728289\" (UID: \"d77174cf-9f42-4fc9-83d6-355b2b728289\") " Oct 08 18:29:47 crc kubenswrapper[4988]: I1008 18:29:47.416593 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8q9bd\" (UniqueName: \"kubernetes.io/projected/d7f4f7aa-ccf2-45e1-8d32-db9684853231-kube-api-access-8q9bd\") pod \"d7f4f7aa-ccf2-45e1-8d32-db9684853231\" (UID: \"d7f4f7aa-ccf2-45e1-8d32-db9684853231\") " Oct 08 18:29:47 crc kubenswrapper[4988]: I1008 18:29:47.416617 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d77174cf-9f42-4fc9-83d6-355b2b728289-config-data\") pod \"d77174cf-9f42-4fc9-83d6-355b2b728289\" (UID: \"d77174cf-9f42-4fc9-83d6-355b2b728289\") " Oct 08 18:29:47 crc kubenswrapper[4988]: I1008 18:29:47.416638 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"d77174cf-9f42-4fc9-83d6-355b2b728289\" (UID: \"d77174cf-9f42-4fc9-83d6-355b2b728289\") " Oct 08 18:29:47 crc kubenswrapper[4988]: I1008 18:29:47.416668 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/d7f4f7aa-ccf2-45e1-8d32-db9684853231-db-sync-config-data\") pod \"d7f4f7aa-ccf2-45e1-8d32-db9684853231\" (UID: \"d7f4f7aa-ccf2-45e1-8d32-db9684853231\") " Oct 08 18:29:47 crc kubenswrapper[4988]: I1008 18:29:47.416724 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d77174cf-9f42-4fc9-83d6-355b2b728289-logs\") pod \"d77174cf-9f42-4fc9-83d6-355b2b728289\" (UID: \"d77174cf-9f42-4fc9-83d6-355b2b728289\") " Oct 08 18:29:47 crc kubenswrapper[4988]: I1008 18:29:47.416782 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d7f4f7aa-ccf2-45e1-8d32-db9684853231-combined-ca-bundle\") pod \"d7f4f7aa-ccf2-45e1-8d32-db9684853231\" (UID: \"d7f4f7aa-ccf2-45e1-8d32-db9684853231\") " Oct 08 18:29:47 crc kubenswrapper[4988]: I1008 18:29:47.416818 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d77174cf-9f42-4fc9-83d6-355b2b728289-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "d77174cf-9f42-4fc9-83d6-355b2b728289" (UID: "d77174cf-9f42-4fc9-83d6-355b2b728289"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 18:29:47 crc kubenswrapper[4988]: I1008 18:29:47.418692 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d77174cf-9f42-4fc9-83d6-355b2b728289-logs" (OuterVolumeSpecName: "logs") pod "d77174cf-9f42-4fc9-83d6-355b2b728289" (UID: "d77174cf-9f42-4fc9-83d6-355b2b728289"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 18:29:47 crc kubenswrapper[4988]: I1008 18:29:47.431313 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d7f4f7aa-ccf2-45e1-8d32-db9684853231-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "d7f4f7aa-ccf2-45e1-8d32-db9684853231" (UID: "d7f4f7aa-ccf2-45e1-8d32-db9684853231"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:29:47 crc kubenswrapper[4988]: I1008 18:29:47.434142 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d7f4f7aa-ccf2-45e1-8d32-db9684853231-kube-api-access-8q9bd" (OuterVolumeSpecName: "kube-api-access-8q9bd") pod "d7f4f7aa-ccf2-45e1-8d32-db9684853231" (UID: "d7f4f7aa-ccf2-45e1-8d32-db9684853231"). InnerVolumeSpecName "kube-api-access-8q9bd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:29:47 crc kubenswrapper[4988]: I1008 18:29:47.434835 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d77174cf-9f42-4fc9-83d6-355b2b728289-kube-api-access-94t5w" (OuterVolumeSpecName: "kube-api-access-94t5w") pod "d77174cf-9f42-4fc9-83d6-355b2b728289" (UID: "d77174cf-9f42-4fc9-83d6-355b2b728289"). InnerVolumeSpecName "kube-api-access-94t5w". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:29:47 crc kubenswrapper[4988]: I1008 18:29:47.436098 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d77174cf-9f42-4fc9-83d6-355b2b728289-scripts" (OuterVolumeSpecName: "scripts") pod "d77174cf-9f42-4fc9-83d6-355b2b728289" (UID: "d77174cf-9f42-4fc9-83d6-355b2b728289"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:29:47 crc kubenswrapper[4988]: I1008 18:29:47.441138 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage05-crc" (OuterVolumeSpecName: "glance") pod "d77174cf-9f42-4fc9-83d6-355b2b728289" (UID: "d77174cf-9f42-4fc9-83d6-355b2b728289"). InnerVolumeSpecName "local-storage05-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 08 18:29:47 crc kubenswrapper[4988]: I1008 18:29:47.455687 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d77174cf-9f42-4fc9-83d6-355b2b728289-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d77174cf-9f42-4fc9-83d6-355b2b728289" (UID: "d77174cf-9f42-4fc9-83d6-355b2b728289"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:29:47 crc kubenswrapper[4988]: I1008 18:29:47.455845 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d7f4f7aa-ccf2-45e1-8d32-db9684853231-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d7f4f7aa-ccf2-45e1-8d32-db9684853231" (UID: "d7f4f7aa-ccf2-45e1-8d32-db9684853231"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:29:47 crc kubenswrapper[4988]: I1008 18:29:47.470709 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d77174cf-9f42-4fc9-83d6-355b2b728289-config-data" (OuterVolumeSpecName: "config-data") pod "d77174cf-9f42-4fc9-83d6-355b2b728289" (UID: "d77174cf-9f42-4fc9-83d6-355b2b728289"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:29:47 crc kubenswrapper[4988]: I1008 18:29:47.518978 4988 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d77174cf-9f42-4fc9-83d6-355b2b728289-config-data\") on node \"crc\" DevicePath \"\"" Oct 08 18:29:47 crc kubenswrapper[4988]: I1008 18:29:47.519030 4988 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") on node \"crc\" " Oct 08 18:29:47 crc kubenswrapper[4988]: I1008 18:29:47.519044 4988 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/d7f4f7aa-ccf2-45e1-8d32-db9684853231-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Oct 08 18:29:47 crc kubenswrapper[4988]: I1008 18:29:47.519059 4988 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d77174cf-9f42-4fc9-83d6-355b2b728289-logs\") on node \"crc\" DevicePath \"\"" Oct 08 18:29:47 crc kubenswrapper[4988]: I1008 18:29:47.519071 4988 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d7f4f7aa-ccf2-45e1-8d32-db9684853231-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 18:29:47 crc kubenswrapper[4988]: I1008 18:29:47.519084 4988 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/d77174cf-9f42-4fc9-83d6-355b2b728289-httpd-run\") on node \"crc\" DevicePath \"\"" Oct 08 18:29:47 crc kubenswrapper[4988]: I1008 18:29:47.519096 4988 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d77174cf-9f42-4fc9-83d6-355b2b728289-scripts\") on node \"crc\" DevicePath \"\"" Oct 08 18:29:47 crc kubenswrapper[4988]: I1008 18:29:47.519108 4988 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d77174cf-9f42-4fc9-83d6-355b2b728289-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 18:29:47 crc kubenswrapper[4988]: I1008 18:29:47.519120 4988 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-94t5w\" (UniqueName: \"kubernetes.io/projected/d77174cf-9f42-4fc9-83d6-355b2b728289-kube-api-access-94t5w\") on node \"crc\" DevicePath \"\"" Oct 08 18:29:47 crc kubenswrapper[4988]: I1008 18:29:47.519134 4988 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8q9bd\" (UniqueName: \"kubernetes.io/projected/d7f4f7aa-ccf2-45e1-8d32-db9684853231-kube-api-access-8q9bd\") on node \"crc\" DevicePath \"\"" Oct 08 18:29:47 crc kubenswrapper[4988]: I1008 18:29:47.544788 4988 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage05-crc" (UniqueName: "kubernetes.io/local-volume/local-storage05-crc") on node "crc" Oct 08 18:29:47 crc kubenswrapper[4988]: I1008 18:29:47.620802 4988 reconciler_common.go:293] "Volume detached for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") on node \"crc\" DevicePath \"\"" Oct 08 18:29:47 crc kubenswrapper[4988]: I1008 18:29:47.701255 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 08 18:29:47 crc kubenswrapper[4988]: I1008 18:29:47.708885 4988 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 08 18:29:47 crc kubenswrapper[4988]: I1008 18:29:47.730875 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 08 18:29:47 crc kubenswrapper[4988]: E1008 18:29:47.731635 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d7f4f7aa-ccf2-45e1-8d32-db9684853231" containerName="barbican-db-sync" Oct 08 18:29:47 crc kubenswrapper[4988]: I1008 18:29:47.731666 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="d7f4f7aa-ccf2-45e1-8d32-db9684853231" containerName="barbican-db-sync" Oct 08 18:29:47 crc kubenswrapper[4988]: E1008 18:29:47.731696 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d77174cf-9f42-4fc9-83d6-355b2b728289" containerName="glance-log" Oct 08 18:29:47 crc kubenswrapper[4988]: I1008 18:29:47.731704 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="d77174cf-9f42-4fc9-83d6-355b2b728289" containerName="glance-log" Oct 08 18:29:47 crc kubenswrapper[4988]: E1008 18:29:47.731726 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d77174cf-9f42-4fc9-83d6-355b2b728289" containerName="glance-httpd" Oct 08 18:29:47 crc kubenswrapper[4988]: I1008 18:29:47.731735 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="d77174cf-9f42-4fc9-83d6-355b2b728289" containerName="glance-httpd" Oct 08 18:29:47 crc kubenswrapper[4988]: I1008 18:29:47.731968 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="d77174cf-9f42-4fc9-83d6-355b2b728289" containerName="glance-log" Oct 08 18:29:47 crc kubenswrapper[4988]: I1008 18:29:47.732003 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="d7f4f7aa-ccf2-45e1-8d32-db9684853231" containerName="barbican-db-sync" Oct 08 18:29:47 crc kubenswrapper[4988]: I1008 18:29:47.732019 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="d77174cf-9f42-4fc9-83d6-355b2b728289" containerName="glance-httpd" Oct 08 18:29:47 crc kubenswrapper[4988]: I1008 18:29:47.733414 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 08 18:29:47 crc kubenswrapper[4988]: I1008 18:29:47.738154 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Oct 08 18:29:47 crc kubenswrapper[4988]: I1008 18:29:47.738234 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Oct 08 18:29:47 crc kubenswrapper[4988]: I1008 18:29:47.754744 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 08 18:29:47 crc kubenswrapper[4988]: I1008 18:29:47.824513 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-internal-api-0\" (UID: \"97a836ba-13f2-4fba-8768-b8323af05c76\") " pod="openstack/glance-default-internal-api-0" Oct 08 18:29:47 crc kubenswrapper[4988]: I1008 18:29:47.824614 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/97a836ba-13f2-4fba-8768-b8323af05c76-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"97a836ba-13f2-4fba-8768-b8323af05c76\") " pod="openstack/glance-default-internal-api-0" Oct 08 18:29:47 crc kubenswrapper[4988]: I1008 18:29:47.824635 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/97a836ba-13f2-4fba-8768-b8323af05c76-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"97a836ba-13f2-4fba-8768-b8323af05c76\") " pod="openstack/glance-default-internal-api-0" Oct 08 18:29:47 crc kubenswrapper[4988]: I1008 18:29:47.824659 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/97a836ba-13f2-4fba-8768-b8323af05c76-logs\") pod \"glance-default-internal-api-0\" (UID: \"97a836ba-13f2-4fba-8768-b8323af05c76\") " pod="openstack/glance-default-internal-api-0" Oct 08 18:29:47 crc kubenswrapper[4988]: I1008 18:29:47.824684 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/97a836ba-13f2-4fba-8768-b8323af05c76-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"97a836ba-13f2-4fba-8768-b8323af05c76\") " pod="openstack/glance-default-internal-api-0" Oct 08 18:29:47 crc kubenswrapper[4988]: I1008 18:29:47.824703 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/97a836ba-13f2-4fba-8768-b8323af05c76-config-data\") pod \"glance-default-internal-api-0\" (UID: \"97a836ba-13f2-4fba-8768-b8323af05c76\") " pod="openstack/glance-default-internal-api-0" Oct 08 18:29:47 crc kubenswrapper[4988]: I1008 18:29:47.824724 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8mmxb\" (UniqueName: \"kubernetes.io/projected/97a836ba-13f2-4fba-8768-b8323af05c76-kube-api-access-8mmxb\") pod \"glance-default-internal-api-0\" (UID: \"97a836ba-13f2-4fba-8768-b8323af05c76\") " pod="openstack/glance-default-internal-api-0" Oct 08 18:29:47 crc kubenswrapper[4988]: I1008 18:29:47.824760 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/97a836ba-13f2-4fba-8768-b8323af05c76-scripts\") pod \"glance-default-internal-api-0\" (UID: \"97a836ba-13f2-4fba-8768-b8323af05c76\") " pod="openstack/glance-default-internal-api-0" Oct 08 18:29:47 crc kubenswrapper[4988]: I1008 18:29:47.927189 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/97a836ba-13f2-4fba-8768-b8323af05c76-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"97a836ba-13f2-4fba-8768-b8323af05c76\") " pod="openstack/glance-default-internal-api-0" Oct 08 18:29:47 crc kubenswrapper[4988]: I1008 18:29:47.927235 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/97a836ba-13f2-4fba-8768-b8323af05c76-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"97a836ba-13f2-4fba-8768-b8323af05c76\") " pod="openstack/glance-default-internal-api-0" Oct 08 18:29:47 crc kubenswrapper[4988]: I1008 18:29:47.927263 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/97a836ba-13f2-4fba-8768-b8323af05c76-logs\") pod \"glance-default-internal-api-0\" (UID: \"97a836ba-13f2-4fba-8768-b8323af05c76\") " pod="openstack/glance-default-internal-api-0" Oct 08 18:29:47 crc kubenswrapper[4988]: I1008 18:29:47.927294 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/97a836ba-13f2-4fba-8768-b8323af05c76-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"97a836ba-13f2-4fba-8768-b8323af05c76\") " pod="openstack/glance-default-internal-api-0" Oct 08 18:29:47 crc kubenswrapper[4988]: I1008 18:29:47.927312 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/97a836ba-13f2-4fba-8768-b8323af05c76-config-data\") pod \"glance-default-internal-api-0\" (UID: \"97a836ba-13f2-4fba-8768-b8323af05c76\") " pod="openstack/glance-default-internal-api-0" Oct 08 18:29:47 crc kubenswrapper[4988]: I1008 18:29:47.927335 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8mmxb\" (UniqueName: \"kubernetes.io/projected/97a836ba-13f2-4fba-8768-b8323af05c76-kube-api-access-8mmxb\") pod \"glance-default-internal-api-0\" (UID: \"97a836ba-13f2-4fba-8768-b8323af05c76\") " pod="openstack/glance-default-internal-api-0" Oct 08 18:29:47 crc kubenswrapper[4988]: I1008 18:29:47.927375 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/97a836ba-13f2-4fba-8768-b8323af05c76-scripts\") pod \"glance-default-internal-api-0\" (UID: \"97a836ba-13f2-4fba-8768-b8323af05c76\") " pod="openstack/glance-default-internal-api-0" Oct 08 18:29:47 crc kubenswrapper[4988]: I1008 18:29:47.927420 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-internal-api-0\" (UID: \"97a836ba-13f2-4fba-8768-b8323af05c76\") " pod="openstack/glance-default-internal-api-0" Oct 08 18:29:47 crc kubenswrapper[4988]: I1008 18:29:47.927589 4988 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-internal-api-0\" (UID: \"97a836ba-13f2-4fba-8768-b8323af05c76\") device mount path \"/mnt/openstack/pv05\"" pod="openstack/glance-default-internal-api-0" Oct 08 18:29:47 crc kubenswrapper[4988]: I1008 18:29:47.928151 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/97a836ba-13f2-4fba-8768-b8323af05c76-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"97a836ba-13f2-4fba-8768-b8323af05c76\") " pod="openstack/glance-default-internal-api-0" Oct 08 18:29:47 crc kubenswrapper[4988]: I1008 18:29:47.928951 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/97a836ba-13f2-4fba-8768-b8323af05c76-logs\") pod \"glance-default-internal-api-0\" (UID: \"97a836ba-13f2-4fba-8768-b8323af05c76\") " pod="openstack/glance-default-internal-api-0" Oct 08 18:29:47 crc kubenswrapper[4988]: I1008 18:29:47.932860 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/97a836ba-13f2-4fba-8768-b8323af05c76-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"97a836ba-13f2-4fba-8768-b8323af05c76\") " pod="openstack/glance-default-internal-api-0" Oct 08 18:29:47 crc kubenswrapper[4988]: I1008 18:29:47.933634 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/97a836ba-13f2-4fba-8768-b8323af05c76-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"97a836ba-13f2-4fba-8768-b8323af05c76\") " pod="openstack/glance-default-internal-api-0" Oct 08 18:29:47 crc kubenswrapper[4988]: I1008 18:29:47.934328 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/97a836ba-13f2-4fba-8768-b8323af05c76-config-data\") pod \"glance-default-internal-api-0\" (UID: \"97a836ba-13f2-4fba-8768-b8323af05c76\") " pod="openstack/glance-default-internal-api-0" Oct 08 18:29:47 crc kubenswrapper[4988]: I1008 18:29:47.936701 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/97a836ba-13f2-4fba-8768-b8323af05c76-scripts\") pod \"glance-default-internal-api-0\" (UID: \"97a836ba-13f2-4fba-8768-b8323af05c76\") " pod="openstack/glance-default-internal-api-0" Oct 08 18:29:47 crc kubenswrapper[4988]: I1008 18:29:47.945468 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8mmxb\" (UniqueName: \"kubernetes.io/projected/97a836ba-13f2-4fba-8768-b8323af05c76-kube-api-access-8mmxb\") pod \"glance-default-internal-api-0\" (UID: \"97a836ba-13f2-4fba-8768-b8323af05c76\") " pod="openstack/glance-default-internal-api-0" Oct 08 18:29:47 crc kubenswrapper[4988]: I1008 18:29:47.964591 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-internal-api-0\" (UID: \"97a836ba-13f2-4fba-8768-b8323af05c76\") " pod="openstack/glance-default-internal-api-0" Oct 08 18:29:48 crc kubenswrapper[4988]: I1008 18:29:48.065050 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 08 18:29:48 crc kubenswrapper[4988]: I1008 18:29:48.383469 4988 generic.go:334] "Generic (PLEG): container finished" podID="8fed08cf-45cb-4fd2-92af-9f30513c0b72" containerID="f586d72bfd02084b33e3cbd2ad495a6f4732968d71f628cba78b3edd940ec1f1" exitCode=2 Oct 08 18:29:48 crc kubenswrapper[4988]: I1008 18:29:48.383813 4988 generic.go:334] "Generic (PLEG): container finished" podID="8fed08cf-45cb-4fd2-92af-9f30513c0b72" containerID="530180f4bca5eaf8c14c742c890bbe065b3ab6a329ac14ec714a341c4b5e100e" exitCode=0 Oct 08 18:29:48 crc kubenswrapper[4988]: I1008 18:29:48.383528 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8fed08cf-45cb-4fd2-92af-9f30513c0b72","Type":"ContainerDied","Data":"f586d72bfd02084b33e3cbd2ad495a6f4732968d71f628cba78b3edd940ec1f1"} Oct 08 18:29:48 crc kubenswrapper[4988]: I1008 18:29:48.383859 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8fed08cf-45cb-4fd2-92af-9f30513c0b72","Type":"ContainerDied","Data":"530180f4bca5eaf8c14c742c890bbe065b3ab6a329ac14ec714a341c4b5e100e"} Oct 08 18:29:48 crc kubenswrapper[4988]: I1008 18:29:48.603447 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-worker-7bf697f8df-h54ft"] Oct 08 18:29:48 crc kubenswrapper[4988]: I1008 18:29:48.605626 4988 scope.go:117] "RemoveContainer" containerID="c2c1cc04e03bf2c844c4b2b292dd8b58cae0aeb74871b25ff6675b936a938eed" Oct 08 18:29:48 crc kubenswrapper[4988]: I1008 18:29:48.607033 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-7bf697f8df-h54ft" Oct 08 18:29:48 crc kubenswrapper[4988]: E1008 18:29:48.608849 4988 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c2c1cc04e03bf2c844c4b2b292dd8b58cae0aeb74871b25ff6675b936a938eed\": container with ID starting with c2c1cc04e03bf2c844c4b2b292dd8b58cae0aeb74871b25ff6675b936a938eed not found: ID does not exist" containerID="c2c1cc04e03bf2c844c4b2b292dd8b58cae0aeb74871b25ff6675b936a938eed" Oct 08 18:29:48 crc kubenswrapper[4988]: I1008 18:29:48.608897 4988 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c2c1cc04e03bf2c844c4b2b292dd8b58cae0aeb74871b25ff6675b936a938eed"} err="failed to get container status \"c2c1cc04e03bf2c844c4b2b292dd8b58cae0aeb74871b25ff6675b936a938eed\": rpc error: code = NotFound desc = could not find container \"c2c1cc04e03bf2c844c4b2b292dd8b58cae0aeb74871b25ff6675b936a938eed\": container with ID starting with c2c1cc04e03bf2c844c4b2b292dd8b58cae0aeb74871b25ff6675b936a938eed not found: ID does not exist" Oct 08 18:29:48 crc kubenswrapper[4988]: I1008 18:29:48.608924 4988 scope.go:117] "RemoveContainer" containerID="ae8d3969526856e7c45e181fb9448caddd6ee057b902ae77af235edd0b24b74a" Oct 08 18:29:48 crc kubenswrapper[4988]: E1008 18:29:48.609455 4988 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ae8d3969526856e7c45e181fb9448caddd6ee057b902ae77af235edd0b24b74a\": container with ID starting with ae8d3969526856e7c45e181fb9448caddd6ee057b902ae77af235edd0b24b74a not found: ID does not exist" containerID="ae8d3969526856e7c45e181fb9448caddd6ee057b902ae77af235edd0b24b74a" Oct 08 18:29:48 crc kubenswrapper[4988]: I1008 18:29:48.609480 4988 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ae8d3969526856e7c45e181fb9448caddd6ee057b902ae77af235edd0b24b74a"} err="failed to get container status \"ae8d3969526856e7c45e181fb9448caddd6ee057b902ae77af235edd0b24b74a\": rpc error: code = NotFound desc = could not find container \"ae8d3969526856e7c45e181fb9448caddd6ee057b902ae77af235edd0b24b74a\": container with ID starting with ae8d3969526856e7c45e181fb9448caddd6ee057b902ae77af235edd0b24b74a not found: ID does not exist" Oct 08 18:29:48 crc kubenswrapper[4988]: I1008 18:29:48.609496 4988 scope.go:117] "RemoveContainer" containerID="c2c1cc04e03bf2c844c4b2b292dd8b58cae0aeb74871b25ff6675b936a938eed" Oct 08 18:29:48 crc kubenswrapper[4988]: I1008 18:29:48.609874 4988 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c2c1cc04e03bf2c844c4b2b292dd8b58cae0aeb74871b25ff6675b936a938eed"} err="failed to get container status \"c2c1cc04e03bf2c844c4b2b292dd8b58cae0aeb74871b25ff6675b936a938eed\": rpc error: code = NotFound desc = could not find container \"c2c1cc04e03bf2c844c4b2b292dd8b58cae0aeb74871b25ff6675b936a938eed\": container with ID starting with c2c1cc04e03bf2c844c4b2b292dd8b58cae0aeb74871b25ff6675b936a938eed not found: ID does not exist" Oct 08 18:29:48 crc kubenswrapper[4988]: I1008 18:29:48.609939 4988 scope.go:117] "RemoveContainer" containerID="ae8d3969526856e7c45e181fb9448caddd6ee057b902ae77af235edd0b24b74a" Oct 08 18:29:48 crc kubenswrapper[4988]: I1008 18:29:48.610206 4988 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ae8d3969526856e7c45e181fb9448caddd6ee057b902ae77af235edd0b24b74a"} err="failed to get container status \"ae8d3969526856e7c45e181fb9448caddd6ee057b902ae77af235edd0b24b74a\": rpc error: code = NotFound desc = could not find container \"ae8d3969526856e7c45e181fb9448caddd6ee057b902ae77af235edd0b24b74a\": container with ID starting with ae8d3969526856e7c45e181fb9448caddd6ee057b902ae77af235edd0b24b74a not found: ID does not exist" Oct 08 18:29:48 crc kubenswrapper[4988]: I1008 18:29:48.610230 4988 scope.go:117] "RemoveContainer" containerID="5a3e3b94cc7cf1dee8c248bb8567bd97c7e17296edb15d81cfce00dd723be9b4" Oct 08 18:29:48 crc kubenswrapper[4988]: I1008 18:29:48.612839 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-worker-config-data" Oct 08 18:29:48 crc kubenswrapper[4988]: I1008 18:29:48.613116 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-vxtzc" Oct 08 18:29:48 crc kubenswrapper[4988]: I1008 18:29:48.613358 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Oct 08 18:29:48 crc kubenswrapper[4988]: I1008 18:29:48.634049 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-keystone-listener-8694c889b8-lhm4g"] Oct 08 18:29:48 crc kubenswrapper[4988]: I1008 18:29:48.635607 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-8694c889b8-lhm4g" Oct 08 18:29:48 crc kubenswrapper[4988]: I1008 18:29:48.643548 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-keystone-listener-config-data" Oct 08 18:29:48 crc kubenswrapper[4988]: I1008 18:29:48.682165 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-7bf697f8df-h54ft"] Oct 08 18:29:48 crc kubenswrapper[4988]: I1008 18:29:48.718436 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-8694c889b8-lhm4g"] Oct 08 18:29:48 crc kubenswrapper[4988]: E1008 18:29:48.727299 4988 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-cinder-api@sha256:85c75d60e1bd2f8a9ea0a2bb21a8df64c0a6f7b504cc1a05a355981d4b90e92f" Oct 08 18:29:48 crc kubenswrapper[4988]: E1008 18:29:48.727445 4988 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:cinder-db-sync,Image:quay.io/podified-antelope-centos9/openstack-cinder-api@sha256:85c75d60e1bd2f8a9ea0a2bb21a8df64c0a6f7b504cc1a05a355981d4b90e92f,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_set_configs && /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:TRUE,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:etc-machine-id,ReadOnly:true,MountPath:/etc/machine-id,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:scripts,ReadOnly:true,MountPath:/usr/local/bin/container-scripts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/config-data/merged,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/my.cnf,SubPath:my.cnf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:db-sync-config-data,ReadOnly:true,MountPath:/etc/cinder/cinder.conf.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:db-sync-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-lshc6,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:*0,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod cinder-db-sync-gm2vf_openstack(5b6f72e9-8dc6-4290-b032-d8a570b91767): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 08 18:29:48 crc kubenswrapper[4988]: E1008 18:29:48.732976 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/cinder-db-sync-gm2vf" podUID="5b6f72e9-8dc6-4290-b032-d8a570b91767" Oct 08 18:29:48 crc kubenswrapper[4988]: I1008 18:29:48.733358 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-54df4b685c-fr6bz"] Oct 08 18:29:48 crc kubenswrapper[4988]: I1008 18:29:48.734812 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-54df4b685c-fr6bz" Oct 08 18:29:48 crc kubenswrapper[4988]: I1008 18:29:48.739696 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-54df4b685c-fr6bz"] Oct 08 18:29:48 crc kubenswrapper[4988]: I1008 18:29:48.744111 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c9f172a3-9188-4f5f-bf3a-fbc3b0dc5206-config-data\") pod \"barbican-keystone-listener-8694c889b8-lhm4g\" (UID: \"c9f172a3-9188-4f5f-bf3a-fbc3b0dc5206\") " pod="openstack/barbican-keystone-listener-8694c889b8-lhm4g" Oct 08 18:29:48 crc kubenswrapper[4988]: I1008 18:29:48.744151 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-54jth\" (UniqueName: \"kubernetes.io/projected/9159c915-07a7-4b41-8672-8b9d6a9572ca-kube-api-access-54jth\") pod \"barbican-worker-7bf697f8df-h54ft\" (UID: \"9159c915-07a7-4b41-8672-8b9d6a9572ca\") " pod="openstack/barbican-worker-7bf697f8df-h54ft" Oct 08 18:29:48 crc kubenswrapper[4988]: I1008 18:29:48.744209 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c9f172a3-9188-4f5f-bf3a-fbc3b0dc5206-logs\") pod \"barbican-keystone-listener-8694c889b8-lhm4g\" (UID: \"c9f172a3-9188-4f5f-bf3a-fbc3b0dc5206\") " pod="openstack/barbican-keystone-listener-8694c889b8-lhm4g" Oct 08 18:29:48 crc kubenswrapper[4988]: I1008 18:29:48.744227 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9159c915-07a7-4b41-8672-8b9d6a9572ca-logs\") pod \"barbican-worker-7bf697f8df-h54ft\" (UID: \"9159c915-07a7-4b41-8672-8b9d6a9572ca\") " pod="openstack/barbican-worker-7bf697f8df-h54ft" Oct 08 18:29:48 crc kubenswrapper[4988]: I1008 18:29:48.744481 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d2ttv\" (UniqueName: \"kubernetes.io/projected/c9f172a3-9188-4f5f-bf3a-fbc3b0dc5206-kube-api-access-d2ttv\") pod \"barbican-keystone-listener-8694c889b8-lhm4g\" (UID: \"c9f172a3-9188-4f5f-bf3a-fbc3b0dc5206\") " pod="openstack/barbican-keystone-listener-8694c889b8-lhm4g" Oct 08 18:29:48 crc kubenswrapper[4988]: I1008 18:29:48.744616 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c9f172a3-9188-4f5f-bf3a-fbc3b0dc5206-combined-ca-bundle\") pod \"barbican-keystone-listener-8694c889b8-lhm4g\" (UID: \"c9f172a3-9188-4f5f-bf3a-fbc3b0dc5206\") " pod="openstack/barbican-keystone-listener-8694c889b8-lhm4g" Oct 08 18:29:48 crc kubenswrapper[4988]: I1008 18:29:48.744693 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9159c915-07a7-4b41-8672-8b9d6a9572ca-combined-ca-bundle\") pod \"barbican-worker-7bf697f8df-h54ft\" (UID: \"9159c915-07a7-4b41-8672-8b9d6a9572ca\") " pod="openstack/barbican-worker-7bf697f8df-h54ft" Oct 08 18:29:48 crc kubenswrapper[4988]: I1008 18:29:48.744748 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/c9f172a3-9188-4f5f-bf3a-fbc3b0dc5206-config-data-custom\") pod \"barbican-keystone-listener-8694c889b8-lhm4g\" (UID: \"c9f172a3-9188-4f5f-bf3a-fbc3b0dc5206\") " pod="openstack/barbican-keystone-listener-8694c889b8-lhm4g" Oct 08 18:29:48 crc kubenswrapper[4988]: I1008 18:29:48.744779 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9159c915-07a7-4b41-8672-8b9d6a9572ca-config-data\") pod \"barbican-worker-7bf697f8df-h54ft\" (UID: \"9159c915-07a7-4b41-8672-8b9d6a9572ca\") " pod="openstack/barbican-worker-7bf697f8df-h54ft" Oct 08 18:29:48 crc kubenswrapper[4988]: I1008 18:29:48.744832 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/9159c915-07a7-4b41-8672-8b9d6a9572ca-config-data-custom\") pod \"barbican-worker-7bf697f8df-h54ft\" (UID: \"9159c915-07a7-4b41-8672-8b9d6a9572ca\") " pod="openstack/barbican-worker-7bf697f8df-h54ft" Oct 08 18:29:48 crc kubenswrapper[4988]: I1008 18:29:48.757721 4988 scope.go:117] "RemoveContainer" containerID="a0ac407571d8352f322c503b5844b1316dfe3d08f823d141074c056a397d2684" Oct 08 18:29:48 crc kubenswrapper[4988]: I1008 18:29:48.773907 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-69894dfcd9-948wp" Oct 08 18:29:48 crc kubenswrapper[4988]: I1008 18:29:48.856994 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7a831457-b429-4339-b166-afed001c770d-dns-svc\") pod \"7a831457-b429-4339-b166-afed001c770d\" (UID: \"7a831457-b429-4339-b166-afed001c770d\") " Oct 08 18:29:48 crc kubenswrapper[4988]: I1008 18:29:48.857402 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7a831457-b429-4339-b166-afed001c770d-ovsdbserver-nb\") pod \"7a831457-b429-4339-b166-afed001c770d\" (UID: \"7a831457-b429-4339-b166-afed001c770d\") " Oct 08 18:29:48 crc kubenswrapper[4988]: I1008 18:29:48.857505 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/7a831457-b429-4339-b166-afed001c770d-dns-swift-storage-0\") pod \"7a831457-b429-4339-b166-afed001c770d\" (UID: \"7a831457-b429-4339-b166-afed001c770d\") " Oct 08 18:29:48 crc kubenswrapper[4988]: I1008 18:29:48.857548 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7a831457-b429-4339-b166-afed001c770d-config\") pod \"7a831457-b429-4339-b166-afed001c770d\" (UID: \"7a831457-b429-4339-b166-afed001c770d\") " Oct 08 18:29:48 crc kubenswrapper[4988]: I1008 18:29:48.857577 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7a831457-b429-4339-b166-afed001c770d-ovsdbserver-sb\") pod \"7a831457-b429-4339-b166-afed001c770d\" (UID: \"7a831457-b429-4339-b166-afed001c770d\") " Oct 08 18:29:48 crc kubenswrapper[4988]: I1008 18:29:48.857602 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9vbc7\" (UniqueName: \"kubernetes.io/projected/7a831457-b429-4339-b166-afed001c770d-kube-api-access-9vbc7\") pod \"7a831457-b429-4339-b166-afed001c770d\" (UID: \"7a831457-b429-4339-b166-afed001c770d\") " Oct 08 18:29:48 crc kubenswrapper[4988]: I1008 18:29:48.857842 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c9f172a3-9188-4f5f-bf3a-fbc3b0dc5206-logs\") pod \"barbican-keystone-listener-8694c889b8-lhm4g\" (UID: \"c9f172a3-9188-4f5f-bf3a-fbc3b0dc5206\") " pod="openstack/barbican-keystone-listener-8694c889b8-lhm4g" Oct 08 18:29:48 crc kubenswrapper[4988]: I1008 18:29:48.857868 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9159c915-07a7-4b41-8672-8b9d6a9572ca-logs\") pod \"barbican-worker-7bf697f8df-h54ft\" (UID: \"9159c915-07a7-4b41-8672-8b9d6a9572ca\") " pod="openstack/barbican-worker-7bf697f8df-h54ft" Oct 08 18:29:48 crc kubenswrapper[4988]: I1008 18:29:48.857933 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d9f52cd3-500f-45b9-9f83-8e8244e511d6-dns-svc\") pod \"dnsmasq-dns-54df4b685c-fr6bz\" (UID: \"d9f52cd3-500f-45b9-9f83-8e8244e511d6\") " pod="openstack/dnsmasq-dns-54df4b685c-fr6bz" Oct 08 18:29:48 crc kubenswrapper[4988]: I1008 18:29:48.857955 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d2ttv\" (UniqueName: \"kubernetes.io/projected/c9f172a3-9188-4f5f-bf3a-fbc3b0dc5206-kube-api-access-d2ttv\") pod \"barbican-keystone-listener-8694c889b8-lhm4g\" (UID: \"c9f172a3-9188-4f5f-bf3a-fbc3b0dc5206\") " pod="openstack/barbican-keystone-listener-8694c889b8-lhm4g" Oct 08 18:29:48 crc kubenswrapper[4988]: I1008 18:29:48.857993 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c9f172a3-9188-4f5f-bf3a-fbc3b0dc5206-combined-ca-bundle\") pod \"barbican-keystone-listener-8694c889b8-lhm4g\" (UID: \"c9f172a3-9188-4f5f-bf3a-fbc3b0dc5206\") " pod="openstack/barbican-keystone-listener-8694c889b8-lhm4g" Oct 08 18:29:48 crc kubenswrapper[4988]: I1008 18:29:48.858018 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wx89j\" (UniqueName: \"kubernetes.io/projected/d9f52cd3-500f-45b9-9f83-8e8244e511d6-kube-api-access-wx89j\") pod \"dnsmasq-dns-54df4b685c-fr6bz\" (UID: \"d9f52cd3-500f-45b9-9f83-8e8244e511d6\") " pod="openstack/dnsmasq-dns-54df4b685c-fr6bz" Oct 08 18:29:48 crc kubenswrapper[4988]: I1008 18:29:48.858045 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9159c915-07a7-4b41-8672-8b9d6a9572ca-combined-ca-bundle\") pod \"barbican-worker-7bf697f8df-h54ft\" (UID: \"9159c915-07a7-4b41-8672-8b9d6a9572ca\") " pod="openstack/barbican-worker-7bf697f8df-h54ft" Oct 08 18:29:48 crc kubenswrapper[4988]: I1008 18:29:48.858072 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/d9f52cd3-500f-45b9-9f83-8e8244e511d6-dns-swift-storage-0\") pod \"dnsmasq-dns-54df4b685c-fr6bz\" (UID: \"d9f52cd3-500f-45b9-9f83-8e8244e511d6\") " pod="openstack/dnsmasq-dns-54df4b685c-fr6bz" Oct 08 18:29:48 crc kubenswrapper[4988]: I1008 18:29:48.858088 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d9f52cd3-500f-45b9-9f83-8e8244e511d6-ovsdbserver-sb\") pod \"dnsmasq-dns-54df4b685c-fr6bz\" (UID: \"d9f52cd3-500f-45b9-9f83-8e8244e511d6\") " pod="openstack/dnsmasq-dns-54df4b685c-fr6bz" Oct 08 18:29:48 crc kubenswrapper[4988]: I1008 18:29:48.858106 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/c9f172a3-9188-4f5f-bf3a-fbc3b0dc5206-config-data-custom\") pod \"barbican-keystone-listener-8694c889b8-lhm4g\" (UID: \"c9f172a3-9188-4f5f-bf3a-fbc3b0dc5206\") " pod="openstack/barbican-keystone-listener-8694c889b8-lhm4g" Oct 08 18:29:48 crc kubenswrapper[4988]: I1008 18:29:48.858124 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9159c915-07a7-4b41-8672-8b9d6a9572ca-config-data\") pod \"barbican-worker-7bf697f8df-h54ft\" (UID: \"9159c915-07a7-4b41-8672-8b9d6a9572ca\") " pod="openstack/barbican-worker-7bf697f8df-h54ft" Oct 08 18:29:48 crc kubenswrapper[4988]: I1008 18:29:48.858147 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/9159c915-07a7-4b41-8672-8b9d6a9572ca-config-data-custom\") pod \"barbican-worker-7bf697f8df-h54ft\" (UID: \"9159c915-07a7-4b41-8672-8b9d6a9572ca\") " pod="openstack/barbican-worker-7bf697f8df-h54ft" Oct 08 18:29:48 crc kubenswrapper[4988]: I1008 18:29:48.858178 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c9f172a3-9188-4f5f-bf3a-fbc3b0dc5206-config-data\") pod \"barbican-keystone-listener-8694c889b8-lhm4g\" (UID: \"c9f172a3-9188-4f5f-bf3a-fbc3b0dc5206\") " pod="openstack/barbican-keystone-listener-8694c889b8-lhm4g" Oct 08 18:29:48 crc kubenswrapper[4988]: I1008 18:29:48.858199 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-54jth\" (UniqueName: \"kubernetes.io/projected/9159c915-07a7-4b41-8672-8b9d6a9572ca-kube-api-access-54jth\") pod \"barbican-worker-7bf697f8df-h54ft\" (UID: \"9159c915-07a7-4b41-8672-8b9d6a9572ca\") " pod="openstack/barbican-worker-7bf697f8df-h54ft" Oct 08 18:29:48 crc kubenswrapper[4988]: I1008 18:29:48.858224 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d9f52cd3-500f-45b9-9f83-8e8244e511d6-ovsdbserver-nb\") pod \"dnsmasq-dns-54df4b685c-fr6bz\" (UID: \"d9f52cd3-500f-45b9-9f83-8e8244e511d6\") " pod="openstack/dnsmasq-dns-54df4b685c-fr6bz" Oct 08 18:29:48 crc kubenswrapper[4988]: I1008 18:29:48.858256 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d9f52cd3-500f-45b9-9f83-8e8244e511d6-config\") pod \"dnsmasq-dns-54df4b685c-fr6bz\" (UID: \"d9f52cd3-500f-45b9-9f83-8e8244e511d6\") " pod="openstack/dnsmasq-dns-54df4b685c-fr6bz" Oct 08 18:29:48 crc kubenswrapper[4988]: I1008 18:29:48.883270 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c9f172a3-9188-4f5f-bf3a-fbc3b0dc5206-combined-ca-bundle\") pod \"barbican-keystone-listener-8694c889b8-lhm4g\" (UID: \"c9f172a3-9188-4f5f-bf3a-fbc3b0dc5206\") " pod="openstack/barbican-keystone-listener-8694c889b8-lhm4g" Oct 08 18:29:48 crc kubenswrapper[4988]: I1008 18:29:48.883694 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/c9f172a3-9188-4f5f-bf3a-fbc3b0dc5206-config-data-custom\") pod \"barbican-keystone-listener-8694c889b8-lhm4g\" (UID: \"c9f172a3-9188-4f5f-bf3a-fbc3b0dc5206\") " pod="openstack/barbican-keystone-listener-8694c889b8-lhm4g" Oct 08 18:29:48 crc kubenswrapper[4988]: I1008 18:29:48.883911 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c9f172a3-9188-4f5f-bf3a-fbc3b0dc5206-logs\") pod \"barbican-keystone-listener-8694c889b8-lhm4g\" (UID: \"c9f172a3-9188-4f5f-bf3a-fbc3b0dc5206\") " pod="openstack/barbican-keystone-listener-8694c889b8-lhm4g" Oct 08 18:29:48 crc kubenswrapper[4988]: I1008 18:29:48.884898 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9159c915-07a7-4b41-8672-8b9d6a9572ca-logs\") pod \"barbican-worker-7bf697f8df-h54ft\" (UID: \"9159c915-07a7-4b41-8672-8b9d6a9572ca\") " pod="openstack/barbican-worker-7bf697f8df-h54ft" Oct 08 18:29:48 crc kubenswrapper[4988]: I1008 18:29:48.907626 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c9f172a3-9188-4f5f-bf3a-fbc3b0dc5206-config-data\") pod \"barbican-keystone-listener-8694c889b8-lhm4g\" (UID: \"c9f172a3-9188-4f5f-bf3a-fbc3b0dc5206\") " pod="openstack/barbican-keystone-listener-8694c889b8-lhm4g" Oct 08 18:29:48 crc kubenswrapper[4988]: I1008 18:29:48.908008 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-54jth\" (UniqueName: \"kubernetes.io/projected/9159c915-07a7-4b41-8672-8b9d6a9572ca-kube-api-access-54jth\") pod \"barbican-worker-7bf697f8df-h54ft\" (UID: \"9159c915-07a7-4b41-8672-8b9d6a9572ca\") " pod="openstack/barbican-worker-7bf697f8df-h54ft" Oct 08 18:29:48 crc kubenswrapper[4988]: I1008 18:29:48.912825 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-76f7bb8cb8-x44rw"] Oct 08 18:29:48 crc kubenswrapper[4988]: E1008 18:29:48.913220 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7a831457-b429-4339-b166-afed001c770d" containerName="dnsmasq-dns" Oct 08 18:29:48 crc kubenswrapper[4988]: I1008 18:29:48.913231 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="7a831457-b429-4339-b166-afed001c770d" containerName="dnsmasq-dns" Oct 08 18:29:48 crc kubenswrapper[4988]: E1008 18:29:48.913252 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7a831457-b429-4339-b166-afed001c770d" containerName="init" Oct 08 18:29:48 crc kubenswrapper[4988]: I1008 18:29:48.913258 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="7a831457-b429-4339-b166-afed001c770d" containerName="init" Oct 08 18:29:48 crc kubenswrapper[4988]: I1008 18:29:48.913442 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="7a831457-b429-4339-b166-afed001c770d" containerName="dnsmasq-dns" Oct 08 18:29:48 crc kubenswrapper[4988]: I1008 18:29:48.914018 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/9159c915-07a7-4b41-8672-8b9d6a9572ca-config-data-custom\") pod \"barbican-worker-7bf697f8df-h54ft\" (UID: \"9159c915-07a7-4b41-8672-8b9d6a9572ca\") " pod="openstack/barbican-worker-7bf697f8df-h54ft" Oct 08 18:29:48 crc kubenswrapper[4988]: I1008 18:29:48.914341 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-76f7bb8cb8-x44rw" Oct 08 18:29:48 crc kubenswrapper[4988]: I1008 18:29:48.915752 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9159c915-07a7-4b41-8672-8b9d6a9572ca-combined-ca-bundle\") pod \"barbican-worker-7bf697f8df-h54ft\" (UID: \"9159c915-07a7-4b41-8672-8b9d6a9572ca\") " pod="openstack/barbican-worker-7bf697f8df-h54ft" Oct 08 18:29:48 crc kubenswrapper[4988]: I1008 18:29:48.923189 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9159c915-07a7-4b41-8672-8b9d6a9572ca-config-data\") pod \"barbican-worker-7bf697f8df-h54ft\" (UID: \"9159c915-07a7-4b41-8672-8b9d6a9572ca\") " pod="openstack/barbican-worker-7bf697f8df-h54ft" Oct 08 18:29:48 crc kubenswrapper[4988]: I1008 18:29:48.931278 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-api-config-data" Oct 08 18:29:48 crc kubenswrapper[4988]: I1008 18:29:48.951270 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d2ttv\" (UniqueName: \"kubernetes.io/projected/c9f172a3-9188-4f5f-bf3a-fbc3b0dc5206-kube-api-access-d2ttv\") pod \"barbican-keystone-listener-8694c889b8-lhm4g\" (UID: \"c9f172a3-9188-4f5f-bf3a-fbc3b0dc5206\") " pod="openstack/barbican-keystone-listener-8694c889b8-lhm4g" Oct 08 18:29:48 crc kubenswrapper[4988]: I1008 18:29:48.962299 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d9f52cd3-500f-45b9-9f83-8e8244e511d6-config\") pod \"dnsmasq-dns-54df4b685c-fr6bz\" (UID: \"d9f52cd3-500f-45b9-9f83-8e8244e511d6\") " pod="openstack/dnsmasq-dns-54df4b685c-fr6bz" Oct 08 18:29:48 crc kubenswrapper[4988]: I1008 18:29:48.962400 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d9f52cd3-500f-45b9-9f83-8e8244e511d6-dns-svc\") pod \"dnsmasq-dns-54df4b685c-fr6bz\" (UID: \"d9f52cd3-500f-45b9-9f83-8e8244e511d6\") " pod="openstack/dnsmasq-dns-54df4b685c-fr6bz" Oct 08 18:29:48 crc kubenswrapper[4988]: I1008 18:29:48.962444 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wx89j\" (UniqueName: \"kubernetes.io/projected/d9f52cd3-500f-45b9-9f83-8e8244e511d6-kube-api-access-wx89j\") pod \"dnsmasq-dns-54df4b685c-fr6bz\" (UID: \"d9f52cd3-500f-45b9-9f83-8e8244e511d6\") " pod="openstack/dnsmasq-dns-54df4b685c-fr6bz" Oct 08 18:29:48 crc kubenswrapper[4988]: I1008 18:29:48.962477 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/d9f52cd3-500f-45b9-9f83-8e8244e511d6-dns-swift-storage-0\") pod \"dnsmasq-dns-54df4b685c-fr6bz\" (UID: \"d9f52cd3-500f-45b9-9f83-8e8244e511d6\") " pod="openstack/dnsmasq-dns-54df4b685c-fr6bz" Oct 08 18:29:48 crc kubenswrapper[4988]: I1008 18:29:48.962496 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d9f52cd3-500f-45b9-9f83-8e8244e511d6-ovsdbserver-sb\") pod \"dnsmasq-dns-54df4b685c-fr6bz\" (UID: \"d9f52cd3-500f-45b9-9f83-8e8244e511d6\") " pod="openstack/dnsmasq-dns-54df4b685c-fr6bz" Oct 08 18:29:48 crc kubenswrapper[4988]: I1008 18:29:48.962545 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d9f52cd3-500f-45b9-9f83-8e8244e511d6-ovsdbserver-nb\") pod \"dnsmasq-dns-54df4b685c-fr6bz\" (UID: \"d9f52cd3-500f-45b9-9f83-8e8244e511d6\") " pod="openstack/dnsmasq-dns-54df4b685c-fr6bz" Oct 08 18:29:48 crc kubenswrapper[4988]: I1008 18:29:48.963289 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d9f52cd3-500f-45b9-9f83-8e8244e511d6-dns-svc\") pod \"dnsmasq-dns-54df4b685c-fr6bz\" (UID: \"d9f52cd3-500f-45b9-9f83-8e8244e511d6\") " pod="openstack/dnsmasq-dns-54df4b685c-fr6bz" Oct 08 18:29:48 crc kubenswrapper[4988]: I1008 18:29:48.963806 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d9f52cd3-500f-45b9-9f83-8e8244e511d6-config\") pod \"dnsmasq-dns-54df4b685c-fr6bz\" (UID: \"d9f52cd3-500f-45b9-9f83-8e8244e511d6\") " pod="openstack/dnsmasq-dns-54df4b685c-fr6bz" Oct 08 18:29:48 crc kubenswrapper[4988]: I1008 18:29:48.964297 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/d9f52cd3-500f-45b9-9f83-8e8244e511d6-dns-swift-storage-0\") pod \"dnsmasq-dns-54df4b685c-fr6bz\" (UID: \"d9f52cd3-500f-45b9-9f83-8e8244e511d6\") " pod="openstack/dnsmasq-dns-54df4b685c-fr6bz" Oct 08 18:29:48 crc kubenswrapper[4988]: I1008 18:29:48.965062 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d9f52cd3-500f-45b9-9f83-8e8244e511d6-ovsdbserver-sb\") pod \"dnsmasq-dns-54df4b685c-fr6bz\" (UID: \"d9f52cd3-500f-45b9-9f83-8e8244e511d6\") " pod="openstack/dnsmasq-dns-54df4b685c-fr6bz" Oct 08 18:29:48 crc kubenswrapper[4988]: I1008 18:29:48.983904 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d9f52cd3-500f-45b9-9f83-8e8244e511d6-ovsdbserver-nb\") pod \"dnsmasq-dns-54df4b685c-fr6bz\" (UID: \"d9f52cd3-500f-45b9-9f83-8e8244e511d6\") " pod="openstack/dnsmasq-dns-54df4b685c-fr6bz" Oct 08 18:29:48 crc kubenswrapper[4988]: I1008 18:29:48.986973 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-76f7bb8cb8-x44rw"] Oct 08 18:29:49 crc kubenswrapper[4988]: I1008 18:29:49.011086 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wx89j\" (UniqueName: \"kubernetes.io/projected/d9f52cd3-500f-45b9-9f83-8e8244e511d6-kube-api-access-wx89j\") pod \"dnsmasq-dns-54df4b685c-fr6bz\" (UID: \"d9f52cd3-500f-45b9-9f83-8e8244e511d6\") " pod="openstack/dnsmasq-dns-54df4b685c-fr6bz" Oct 08 18:29:49 crc kubenswrapper[4988]: I1008 18:29:49.011511 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7a831457-b429-4339-b166-afed001c770d-kube-api-access-9vbc7" (OuterVolumeSpecName: "kube-api-access-9vbc7") pod "7a831457-b429-4339-b166-afed001c770d" (UID: "7a831457-b429-4339-b166-afed001c770d"). InnerVolumeSpecName "kube-api-access-9vbc7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:29:49 crc kubenswrapper[4988]: I1008 18:29:49.034677 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7a831457-b429-4339-b166-afed001c770d-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "7a831457-b429-4339-b166-afed001c770d" (UID: "7a831457-b429-4339-b166-afed001c770d"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 18:29:49 crc kubenswrapper[4988]: I1008 18:29:49.053537 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-7bf697f8df-h54ft" Oct 08 18:29:49 crc kubenswrapper[4988]: I1008 18:29:49.064716 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fe49d0c8-90f0-4673-b527-ff112cfd3037-combined-ca-bundle\") pod \"barbican-api-76f7bb8cb8-x44rw\" (UID: \"fe49d0c8-90f0-4673-b527-ff112cfd3037\") " pod="openstack/barbican-api-76f7bb8cb8-x44rw" Oct 08 18:29:49 crc kubenswrapper[4988]: I1008 18:29:49.064759 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fe49d0c8-90f0-4673-b527-ff112cfd3037-logs\") pod \"barbican-api-76f7bb8cb8-x44rw\" (UID: \"fe49d0c8-90f0-4673-b527-ff112cfd3037\") " pod="openstack/barbican-api-76f7bb8cb8-x44rw" Oct 08 18:29:49 crc kubenswrapper[4988]: I1008 18:29:49.064799 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cjvcj\" (UniqueName: \"kubernetes.io/projected/fe49d0c8-90f0-4673-b527-ff112cfd3037-kube-api-access-cjvcj\") pod \"barbican-api-76f7bb8cb8-x44rw\" (UID: \"fe49d0c8-90f0-4673-b527-ff112cfd3037\") " pod="openstack/barbican-api-76f7bb8cb8-x44rw" Oct 08 18:29:49 crc kubenswrapper[4988]: I1008 18:29:49.064929 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fe49d0c8-90f0-4673-b527-ff112cfd3037-config-data\") pod \"barbican-api-76f7bb8cb8-x44rw\" (UID: \"fe49d0c8-90f0-4673-b527-ff112cfd3037\") " pod="openstack/barbican-api-76f7bb8cb8-x44rw" Oct 08 18:29:49 crc kubenswrapper[4988]: I1008 18:29:49.065121 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/fe49d0c8-90f0-4673-b527-ff112cfd3037-config-data-custom\") pod \"barbican-api-76f7bb8cb8-x44rw\" (UID: \"fe49d0c8-90f0-4673-b527-ff112cfd3037\") " pod="openstack/barbican-api-76f7bb8cb8-x44rw" Oct 08 18:29:49 crc kubenswrapper[4988]: I1008 18:29:49.065288 4988 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/7a831457-b429-4339-b166-afed001c770d-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Oct 08 18:29:49 crc kubenswrapper[4988]: I1008 18:29:49.065301 4988 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9vbc7\" (UniqueName: \"kubernetes.io/projected/7a831457-b429-4339-b166-afed001c770d-kube-api-access-9vbc7\") on node \"crc\" DevicePath \"\"" Oct 08 18:29:49 crc kubenswrapper[4988]: I1008 18:29:49.072129 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7a831457-b429-4339-b166-afed001c770d-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "7a831457-b429-4339-b166-afed001c770d" (UID: "7a831457-b429-4339-b166-afed001c770d"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 18:29:49 crc kubenswrapper[4988]: I1008 18:29:49.075891 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7a831457-b429-4339-b166-afed001c770d-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "7a831457-b429-4339-b166-afed001c770d" (UID: "7a831457-b429-4339-b166-afed001c770d"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 18:29:49 crc kubenswrapper[4988]: I1008 18:29:49.078627 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-8694c889b8-lhm4g" Oct 08 18:29:49 crc kubenswrapper[4988]: I1008 18:29:49.091684 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7a831457-b429-4339-b166-afed001c770d-config" (OuterVolumeSpecName: "config") pod "7a831457-b429-4339-b166-afed001c770d" (UID: "7a831457-b429-4339-b166-afed001c770d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 18:29:49 crc kubenswrapper[4988]: I1008 18:29:49.106945 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7a831457-b429-4339-b166-afed001c770d-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "7a831457-b429-4339-b166-afed001c770d" (UID: "7a831457-b429-4339-b166-afed001c770d"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 18:29:49 crc kubenswrapper[4988]: I1008 18:29:49.118075 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-54df4b685c-fr6bz" Oct 08 18:29:49 crc kubenswrapper[4988]: I1008 18:29:49.182574 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fe49d0c8-90f0-4673-b527-ff112cfd3037-config-data\") pod \"barbican-api-76f7bb8cb8-x44rw\" (UID: \"fe49d0c8-90f0-4673-b527-ff112cfd3037\") " pod="openstack/barbican-api-76f7bb8cb8-x44rw" Oct 08 18:29:49 crc kubenswrapper[4988]: I1008 18:29:49.182672 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/fe49d0c8-90f0-4673-b527-ff112cfd3037-config-data-custom\") pod \"barbican-api-76f7bb8cb8-x44rw\" (UID: \"fe49d0c8-90f0-4673-b527-ff112cfd3037\") " pod="openstack/barbican-api-76f7bb8cb8-x44rw" Oct 08 18:29:49 crc kubenswrapper[4988]: I1008 18:29:49.182827 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fe49d0c8-90f0-4673-b527-ff112cfd3037-combined-ca-bundle\") pod \"barbican-api-76f7bb8cb8-x44rw\" (UID: \"fe49d0c8-90f0-4673-b527-ff112cfd3037\") " pod="openstack/barbican-api-76f7bb8cb8-x44rw" Oct 08 18:29:49 crc kubenswrapper[4988]: I1008 18:29:49.182856 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fe49d0c8-90f0-4673-b527-ff112cfd3037-logs\") pod \"barbican-api-76f7bb8cb8-x44rw\" (UID: \"fe49d0c8-90f0-4673-b527-ff112cfd3037\") " pod="openstack/barbican-api-76f7bb8cb8-x44rw" Oct 08 18:29:49 crc kubenswrapper[4988]: I1008 18:29:49.182916 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cjvcj\" (UniqueName: \"kubernetes.io/projected/fe49d0c8-90f0-4673-b527-ff112cfd3037-kube-api-access-cjvcj\") pod \"barbican-api-76f7bb8cb8-x44rw\" (UID: \"fe49d0c8-90f0-4673-b527-ff112cfd3037\") " pod="openstack/barbican-api-76f7bb8cb8-x44rw" Oct 08 18:29:49 crc kubenswrapper[4988]: I1008 18:29:49.183066 4988 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7a831457-b429-4339-b166-afed001c770d-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 08 18:29:49 crc kubenswrapper[4988]: I1008 18:29:49.183077 4988 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7a831457-b429-4339-b166-afed001c770d-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 08 18:29:49 crc kubenswrapper[4988]: I1008 18:29:49.183090 4988 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7a831457-b429-4339-b166-afed001c770d-config\") on node \"crc\" DevicePath \"\"" Oct 08 18:29:49 crc kubenswrapper[4988]: I1008 18:29:49.183099 4988 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7a831457-b429-4339-b166-afed001c770d-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 08 18:29:49 crc kubenswrapper[4988]: I1008 18:29:49.184046 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fe49d0c8-90f0-4673-b527-ff112cfd3037-logs\") pod \"barbican-api-76f7bb8cb8-x44rw\" (UID: \"fe49d0c8-90f0-4673-b527-ff112cfd3037\") " pod="openstack/barbican-api-76f7bb8cb8-x44rw" Oct 08 18:29:49 crc kubenswrapper[4988]: I1008 18:29:49.194738 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/fe49d0c8-90f0-4673-b527-ff112cfd3037-config-data-custom\") pod \"barbican-api-76f7bb8cb8-x44rw\" (UID: \"fe49d0c8-90f0-4673-b527-ff112cfd3037\") " pod="openstack/barbican-api-76f7bb8cb8-x44rw" Oct 08 18:29:49 crc kubenswrapper[4988]: I1008 18:29:49.197639 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fe49d0c8-90f0-4673-b527-ff112cfd3037-config-data\") pod \"barbican-api-76f7bb8cb8-x44rw\" (UID: \"fe49d0c8-90f0-4673-b527-ff112cfd3037\") " pod="openstack/barbican-api-76f7bb8cb8-x44rw" Oct 08 18:29:49 crc kubenswrapper[4988]: I1008 18:29:49.198309 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fe49d0c8-90f0-4673-b527-ff112cfd3037-combined-ca-bundle\") pod \"barbican-api-76f7bb8cb8-x44rw\" (UID: \"fe49d0c8-90f0-4673-b527-ff112cfd3037\") " pod="openstack/barbican-api-76f7bb8cb8-x44rw" Oct 08 18:29:49 crc kubenswrapper[4988]: I1008 18:29:49.212885 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cjvcj\" (UniqueName: \"kubernetes.io/projected/fe49d0c8-90f0-4673-b527-ff112cfd3037-kube-api-access-cjvcj\") pod \"barbican-api-76f7bb8cb8-x44rw\" (UID: \"fe49d0c8-90f0-4673-b527-ff112cfd3037\") " pod="openstack/barbican-api-76f7bb8cb8-x44rw" Oct 08 18:29:49 crc kubenswrapper[4988]: I1008 18:29:49.299037 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-76f7bb8cb8-x44rw" Oct 08 18:29:49 crc kubenswrapper[4988]: I1008 18:29:49.345878 4988 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d77174cf-9f42-4fc9-83d6-355b2b728289" path="/var/lib/kubelet/pods/d77174cf-9f42-4fc9-83d6-355b2b728289/volumes" Oct 08 18:29:49 crc kubenswrapper[4988]: I1008 18:29:49.437435 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-69894dfcd9-948wp" event={"ID":"7a831457-b429-4339-b166-afed001c770d","Type":"ContainerDied","Data":"5f3365d57e8b47bf123e83a83e5da3290a4622c50a10edc5ad829153757d864f"} Oct 08 18:29:49 crc kubenswrapper[4988]: I1008 18:29:49.437681 4988 scope.go:117] "RemoveContainer" containerID="ac3f09e0a805492ad71b338aa6dd023325b8ab08859f1ce9590b691ca0ede0e9" Oct 08 18:29:49 crc kubenswrapper[4988]: I1008 18:29:49.437791 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-69894dfcd9-948wp" Oct 08 18:29:49 crc kubenswrapper[4988]: E1008 18:29:49.442094 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-cinder-api@sha256:85c75d60e1bd2f8a9ea0a2bb21a8df64c0a6f7b504cc1a05a355981d4b90e92f\\\"\"" pod="openstack/cinder-db-sync-gm2vf" podUID="5b6f72e9-8dc6-4290-b032-d8a570b91767" Oct 08 18:29:49 crc kubenswrapper[4988]: I1008 18:29:49.466224 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-69894dfcd9-948wp"] Oct 08 18:29:49 crc kubenswrapper[4988]: I1008 18:29:49.471612 4988 scope.go:117] "RemoveContainer" containerID="190609c635be9b56a48a82610585cd3e788b25a29cb94dabca71afefcecfec2f" Oct 08 18:29:49 crc kubenswrapper[4988]: I1008 18:29:49.477714 4988 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-69894dfcd9-948wp"] Oct 08 18:29:49 crc kubenswrapper[4988]: I1008 18:29:49.507728 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 08 18:29:49 crc kubenswrapper[4988]: I1008 18:29:49.590210 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 08 18:29:49 crc kubenswrapper[4988]: I1008 18:29:49.726002 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-7bf697f8df-h54ft"] Oct 08 18:29:49 crc kubenswrapper[4988]: W1008 18:29:49.726966 4988 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9159c915_07a7_4b41_8672_8b9d6a9572ca.slice/crio-fb877c896c174d8672045022e21f1ab69fe359bfbc9c5f531d8c89eae810dc71 WatchSource:0}: Error finding container fb877c896c174d8672045022e21f1ab69fe359bfbc9c5f531d8c89eae810dc71: Status 404 returned error can't find the container with id fb877c896c174d8672045022e21f1ab69fe359bfbc9c5f531d8c89eae810dc71 Oct 08 18:29:49 crc kubenswrapper[4988]: W1008 18:29:49.731454 4988 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc9f172a3_9188_4f5f_bf3a_fbc3b0dc5206.slice/crio-77cc9e1e0ea4e222419c071fbb94d7b4a3b6fcc421e688d2501dd61fe2fa13ce WatchSource:0}: Error finding container 77cc9e1e0ea4e222419c071fbb94d7b4a3b6fcc421e688d2501dd61fe2fa13ce: Status 404 returned error can't find the container with id 77cc9e1e0ea4e222419c071fbb94d7b4a3b6fcc421e688d2501dd61fe2fa13ce Oct 08 18:29:49 crc kubenswrapper[4988]: I1008 18:29:49.733901 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-8694c889b8-lhm4g"] Oct 08 18:29:49 crc kubenswrapper[4988]: I1008 18:29:49.865966 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-54df4b685c-fr6bz"] Oct 08 18:29:49 crc kubenswrapper[4988]: W1008 18:29:49.871177 4988 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd9f52cd3_500f_45b9_9f83_8e8244e511d6.slice/crio-a1cdc1b3db5fda1d178b0965e413f8b36b8c31907bf682098c127f991ed50430 WatchSource:0}: Error finding container a1cdc1b3db5fda1d178b0965e413f8b36b8c31907bf682098c127f991ed50430: Status 404 returned error can't find the container with id a1cdc1b3db5fda1d178b0965e413f8b36b8c31907bf682098c127f991ed50430 Oct 08 18:29:49 crc kubenswrapper[4988]: I1008 18:29:49.939478 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-76f7bb8cb8-x44rw"] Oct 08 18:29:49 crc kubenswrapper[4988]: W1008 18:29:49.945712 4988 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podfe49d0c8_90f0_4673_b527_ff112cfd3037.slice/crio-2e61b2af6a360d8f3ce49333eb603bdf2a9800e8b88f055d0906ef61d8be49b4 WatchSource:0}: Error finding container 2e61b2af6a360d8f3ce49333eb603bdf2a9800e8b88f055d0906ef61d8be49b4: Status 404 returned error can't find the container with id 2e61b2af6a360d8f3ce49333eb603bdf2a9800e8b88f055d0906ef61d8be49b4 Oct 08 18:29:50 crc kubenswrapper[4988]: I1008 18:29:50.453122 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"b98d358d-c785-4494-932c-1d6662944b74","Type":"ContainerStarted","Data":"efa73d2a94e9d47c508d221c8632c29f721f7e3a2ae514318ae8a6b5fc355138"} Oct 08 18:29:50 crc kubenswrapper[4988]: I1008 18:29:50.453556 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"b98d358d-c785-4494-932c-1d6662944b74","Type":"ContainerStarted","Data":"60b27f51694809ff52bff2c604cbfa056ed4778073da33d8b9ba28366e4d86f8"} Oct 08 18:29:50 crc kubenswrapper[4988]: I1008 18:29:50.456929 4988 generic.go:334] "Generic (PLEG): container finished" podID="8fed08cf-45cb-4fd2-92af-9f30513c0b72" containerID="2b5c761d77af17eace93d9cd355762d47e5b9791b61461447ac5dfb3542405b0" exitCode=0 Oct 08 18:29:50 crc kubenswrapper[4988]: I1008 18:29:50.457017 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8fed08cf-45cb-4fd2-92af-9f30513c0b72","Type":"ContainerDied","Data":"2b5c761d77af17eace93d9cd355762d47e5b9791b61461447ac5dfb3542405b0"} Oct 08 18:29:50 crc kubenswrapper[4988]: I1008 18:29:50.461559 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-8694c889b8-lhm4g" event={"ID":"c9f172a3-9188-4f5f-bf3a-fbc3b0dc5206","Type":"ContainerStarted","Data":"77cc9e1e0ea4e222419c071fbb94d7b4a3b6fcc421e688d2501dd61fe2fa13ce"} Oct 08 18:29:50 crc kubenswrapper[4988]: I1008 18:29:50.463542 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"97a836ba-13f2-4fba-8768-b8323af05c76","Type":"ContainerStarted","Data":"169065f869ca95a8a25e30ecb0c2b7316835c33372ad4a1a673366dc1694dc25"} Oct 08 18:29:50 crc kubenswrapper[4988]: I1008 18:29:50.463567 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"97a836ba-13f2-4fba-8768-b8323af05c76","Type":"ContainerStarted","Data":"3e3a1a6247bdb3ec156fbb7d8dc0c0f995772c9aa883ffb9d5cc63e7923786bb"} Oct 08 18:29:50 crc kubenswrapper[4988]: I1008 18:29:50.468706 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-76f7bb8cb8-x44rw" event={"ID":"fe49d0c8-90f0-4673-b527-ff112cfd3037","Type":"ContainerStarted","Data":"0e25441dbd7ef8b78df421475b3d19064e1a51130eb299a3433a4940e3737266"} Oct 08 18:29:50 crc kubenswrapper[4988]: I1008 18:29:50.468747 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-76f7bb8cb8-x44rw" event={"ID":"fe49d0c8-90f0-4673-b527-ff112cfd3037","Type":"ContainerStarted","Data":"2e61b2af6a360d8f3ce49333eb603bdf2a9800e8b88f055d0906ef61d8be49b4"} Oct 08 18:29:50 crc kubenswrapper[4988]: I1008 18:29:50.468789 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-76f7bb8cb8-x44rw" Oct 08 18:29:50 crc kubenswrapper[4988]: I1008 18:29:50.468809 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-76f7bb8cb8-x44rw" Oct 08 18:29:50 crc kubenswrapper[4988]: I1008 18:29:50.472623 4988 generic.go:334] "Generic (PLEG): container finished" podID="d9f52cd3-500f-45b9-9f83-8e8244e511d6" containerID="629b7f05d0b6b84e81db7afa778e7f08381c6ce616a2416e8b5bb6bd08dd4c6a" exitCode=0 Oct 08 18:29:50 crc kubenswrapper[4988]: I1008 18:29:50.472698 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-54df4b685c-fr6bz" event={"ID":"d9f52cd3-500f-45b9-9f83-8e8244e511d6","Type":"ContainerDied","Data":"629b7f05d0b6b84e81db7afa778e7f08381c6ce616a2416e8b5bb6bd08dd4c6a"} Oct 08 18:29:50 crc kubenswrapper[4988]: I1008 18:29:50.472741 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-54df4b685c-fr6bz" event={"ID":"d9f52cd3-500f-45b9-9f83-8e8244e511d6","Type":"ContainerStarted","Data":"a1cdc1b3db5fda1d178b0965e413f8b36b8c31907bf682098c127f991ed50430"} Oct 08 18:29:50 crc kubenswrapper[4988]: I1008 18:29:50.475023 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-7bf697f8df-h54ft" event={"ID":"9159c915-07a7-4b41-8672-8b9d6a9572ca","Type":"ContainerStarted","Data":"fb877c896c174d8672045022e21f1ab69fe359bfbc9c5f531d8c89eae810dc71"} Oct 08 18:29:50 crc kubenswrapper[4988]: I1008 18:29:50.500215 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-76f7bb8cb8-x44rw" podStartSLOduration=2.500198539 podStartE2EDuration="2.500198539s" podCreationTimestamp="2025-10-08 18:29:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 18:29:50.485658251 +0000 UTC m=+1135.935501031" watchObservedRunningTime="2025-10-08 18:29:50.500198539 +0000 UTC m=+1135.950041309" Oct 08 18:29:50 crc kubenswrapper[4988]: I1008 18:29:50.543571 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 08 18:29:50 crc kubenswrapper[4988]: I1008 18:29:50.734021 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8fed08cf-45cb-4fd2-92af-9f30513c0b72-run-httpd\") pod \"8fed08cf-45cb-4fd2-92af-9f30513c0b72\" (UID: \"8fed08cf-45cb-4fd2-92af-9f30513c0b72\") " Oct 08 18:29:50 crc kubenswrapper[4988]: I1008 18:29:50.734064 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8fed08cf-45cb-4fd2-92af-9f30513c0b72-scripts\") pod \"8fed08cf-45cb-4fd2-92af-9f30513c0b72\" (UID: \"8fed08cf-45cb-4fd2-92af-9f30513c0b72\") " Oct 08 18:29:50 crc kubenswrapper[4988]: I1008 18:29:50.734113 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8fed08cf-45cb-4fd2-92af-9f30513c0b72-combined-ca-bundle\") pod \"8fed08cf-45cb-4fd2-92af-9f30513c0b72\" (UID: \"8fed08cf-45cb-4fd2-92af-9f30513c0b72\") " Oct 08 18:29:50 crc kubenswrapper[4988]: I1008 18:29:50.734143 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/8fed08cf-45cb-4fd2-92af-9f30513c0b72-sg-core-conf-yaml\") pod \"8fed08cf-45cb-4fd2-92af-9f30513c0b72\" (UID: \"8fed08cf-45cb-4fd2-92af-9f30513c0b72\") " Oct 08 18:29:50 crc kubenswrapper[4988]: I1008 18:29:50.734264 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4xpgf\" (UniqueName: \"kubernetes.io/projected/8fed08cf-45cb-4fd2-92af-9f30513c0b72-kube-api-access-4xpgf\") pod \"8fed08cf-45cb-4fd2-92af-9f30513c0b72\" (UID: \"8fed08cf-45cb-4fd2-92af-9f30513c0b72\") " Oct 08 18:29:50 crc kubenswrapper[4988]: I1008 18:29:50.734324 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8fed08cf-45cb-4fd2-92af-9f30513c0b72-log-httpd\") pod \"8fed08cf-45cb-4fd2-92af-9f30513c0b72\" (UID: \"8fed08cf-45cb-4fd2-92af-9f30513c0b72\") " Oct 08 18:29:50 crc kubenswrapper[4988]: I1008 18:29:50.734343 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8fed08cf-45cb-4fd2-92af-9f30513c0b72-config-data\") pod \"8fed08cf-45cb-4fd2-92af-9f30513c0b72\" (UID: \"8fed08cf-45cb-4fd2-92af-9f30513c0b72\") " Oct 08 18:29:50 crc kubenswrapper[4988]: I1008 18:29:50.736805 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8fed08cf-45cb-4fd2-92af-9f30513c0b72-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "8fed08cf-45cb-4fd2-92af-9f30513c0b72" (UID: "8fed08cf-45cb-4fd2-92af-9f30513c0b72"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 18:29:50 crc kubenswrapper[4988]: I1008 18:29:50.737153 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8fed08cf-45cb-4fd2-92af-9f30513c0b72-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "8fed08cf-45cb-4fd2-92af-9f30513c0b72" (UID: "8fed08cf-45cb-4fd2-92af-9f30513c0b72"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 18:29:50 crc kubenswrapper[4988]: I1008 18:29:50.741499 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8fed08cf-45cb-4fd2-92af-9f30513c0b72-scripts" (OuterVolumeSpecName: "scripts") pod "8fed08cf-45cb-4fd2-92af-9f30513c0b72" (UID: "8fed08cf-45cb-4fd2-92af-9f30513c0b72"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:29:50 crc kubenswrapper[4988]: I1008 18:29:50.741672 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8fed08cf-45cb-4fd2-92af-9f30513c0b72-kube-api-access-4xpgf" (OuterVolumeSpecName: "kube-api-access-4xpgf") pod "8fed08cf-45cb-4fd2-92af-9f30513c0b72" (UID: "8fed08cf-45cb-4fd2-92af-9f30513c0b72"). InnerVolumeSpecName "kube-api-access-4xpgf". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:29:50 crc kubenswrapper[4988]: I1008 18:29:50.784038 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8fed08cf-45cb-4fd2-92af-9f30513c0b72-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "8fed08cf-45cb-4fd2-92af-9f30513c0b72" (UID: "8fed08cf-45cb-4fd2-92af-9f30513c0b72"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:29:50 crc kubenswrapper[4988]: I1008 18:29:50.807353 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8fed08cf-45cb-4fd2-92af-9f30513c0b72-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "8fed08cf-45cb-4fd2-92af-9f30513c0b72" (UID: "8fed08cf-45cb-4fd2-92af-9f30513c0b72"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:29:50 crc kubenswrapper[4988]: I1008 18:29:50.835952 4988 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8fed08cf-45cb-4fd2-92af-9f30513c0b72-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 08 18:29:50 crc kubenswrapper[4988]: I1008 18:29:50.836196 4988 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8fed08cf-45cb-4fd2-92af-9f30513c0b72-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 08 18:29:50 crc kubenswrapper[4988]: I1008 18:29:50.836205 4988 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8fed08cf-45cb-4fd2-92af-9f30513c0b72-scripts\") on node \"crc\" DevicePath \"\"" Oct 08 18:29:50 crc kubenswrapper[4988]: I1008 18:29:50.836213 4988 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8fed08cf-45cb-4fd2-92af-9f30513c0b72-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 18:29:50 crc kubenswrapper[4988]: I1008 18:29:50.836222 4988 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/8fed08cf-45cb-4fd2-92af-9f30513c0b72-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Oct 08 18:29:50 crc kubenswrapper[4988]: I1008 18:29:50.836230 4988 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4xpgf\" (UniqueName: \"kubernetes.io/projected/8fed08cf-45cb-4fd2-92af-9f30513c0b72-kube-api-access-4xpgf\") on node \"crc\" DevicePath \"\"" Oct 08 18:29:50 crc kubenswrapper[4988]: I1008 18:29:50.841928 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8fed08cf-45cb-4fd2-92af-9f30513c0b72-config-data" (OuterVolumeSpecName: "config-data") pod "8fed08cf-45cb-4fd2-92af-9f30513c0b72" (UID: "8fed08cf-45cb-4fd2-92af-9f30513c0b72"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:29:50 crc kubenswrapper[4988]: I1008 18:29:50.938129 4988 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8fed08cf-45cb-4fd2-92af-9f30513c0b72-config-data\") on node \"crc\" DevicePath \"\"" Oct 08 18:29:51 crc kubenswrapper[4988]: I1008 18:29:51.246219 4988 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7a831457-b429-4339-b166-afed001c770d" path="/var/lib/kubelet/pods/7a831457-b429-4339-b166-afed001c770d/volumes" Oct 08 18:29:51 crc kubenswrapper[4988]: I1008 18:29:51.491631 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"97a836ba-13f2-4fba-8768-b8323af05c76","Type":"ContainerStarted","Data":"826b4afcb111f1466543dcc847edf3bfe65bed172e3c9408bb0e660047b6800f"} Oct 08 18:29:51 crc kubenswrapper[4988]: I1008 18:29:51.497812 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-76f7bb8cb8-x44rw" event={"ID":"fe49d0c8-90f0-4673-b527-ff112cfd3037","Type":"ContainerStarted","Data":"dfe3a72d9e18e94651f136aa2b98ffa4d52065039c4bc92724da4663ef94fddf"} Oct 08 18:29:51 crc kubenswrapper[4988]: I1008 18:29:51.500198 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-54df4b685c-fr6bz" event={"ID":"d9f52cd3-500f-45b9-9f83-8e8244e511d6","Type":"ContainerStarted","Data":"9a9ae238ebc775efdca8a60be60a294b733e168bb9d325901a7175d4ac2574bb"} Oct 08 18:29:51 crc kubenswrapper[4988]: I1008 18:29:51.500557 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-54df4b685c-fr6bz" Oct 08 18:29:51 crc kubenswrapper[4988]: I1008 18:29:51.506532 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"b98d358d-c785-4494-932c-1d6662944b74","Type":"ContainerStarted","Data":"97c41cb131e40fdfd704edbc22cffe0cdc5412ff8904ccc8c9505df6fb46f714"} Oct 08 18:29:51 crc kubenswrapper[4988]: I1008 18:29:51.509883 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8fed08cf-45cb-4fd2-92af-9f30513c0b72","Type":"ContainerDied","Data":"1f49914f5ed3f7f2d69b1d22ee1fdd1b474e9f2eed793d6fdef303b6808ac310"} Oct 08 18:29:51 crc kubenswrapper[4988]: I1008 18:29:51.509935 4988 scope.go:117] "RemoveContainer" containerID="f586d72bfd02084b33e3cbd2ad495a6f4732968d71f628cba78b3edd940ec1f1" Oct 08 18:29:51 crc kubenswrapper[4988]: I1008 18:29:51.510115 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 08 18:29:51 crc kubenswrapper[4988]: I1008 18:29:51.517277 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=4.517263166 podStartE2EDuration="4.517263166s" podCreationTimestamp="2025-10-08 18:29:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 18:29:51.513782128 +0000 UTC m=+1136.963624898" watchObservedRunningTime="2025-10-08 18:29:51.517263166 +0000 UTC m=+1136.967105936" Oct 08 18:29:51 crc kubenswrapper[4988]: I1008 18:29:51.541870 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-6c486d5744-j86zn"] Oct 08 18:29:51 crc kubenswrapper[4988]: E1008 18:29:51.542326 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8fed08cf-45cb-4fd2-92af-9f30513c0b72" containerName="sg-core" Oct 08 18:29:51 crc kubenswrapper[4988]: I1008 18:29:51.542348 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="8fed08cf-45cb-4fd2-92af-9f30513c0b72" containerName="sg-core" Oct 08 18:29:51 crc kubenswrapper[4988]: E1008 18:29:51.542400 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8fed08cf-45cb-4fd2-92af-9f30513c0b72" containerName="ceilometer-central-agent" Oct 08 18:29:51 crc kubenswrapper[4988]: I1008 18:29:51.542412 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="8fed08cf-45cb-4fd2-92af-9f30513c0b72" containerName="ceilometer-central-agent" Oct 08 18:29:51 crc kubenswrapper[4988]: E1008 18:29:51.542434 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8fed08cf-45cb-4fd2-92af-9f30513c0b72" containerName="ceilometer-notification-agent" Oct 08 18:29:51 crc kubenswrapper[4988]: I1008 18:29:51.542443 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="8fed08cf-45cb-4fd2-92af-9f30513c0b72" containerName="ceilometer-notification-agent" Oct 08 18:29:51 crc kubenswrapper[4988]: I1008 18:29:51.542654 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="8fed08cf-45cb-4fd2-92af-9f30513c0b72" containerName="ceilometer-central-agent" Oct 08 18:29:51 crc kubenswrapper[4988]: I1008 18:29:51.542675 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="8fed08cf-45cb-4fd2-92af-9f30513c0b72" containerName="ceilometer-notification-agent" Oct 08 18:29:51 crc kubenswrapper[4988]: I1008 18:29:51.542697 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="8fed08cf-45cb-4fd2-92af-9f30513c0b72" containerName="sg-core" Oct 08 18:29:51 crc kubenswrapper[4988]: I1008 18:29:51.544287 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-6c486d5744-j86zn" Oct 08 18:29:51 crc kubenswrapper[4988]: I1008 18:29:51.544491 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=17.544474705 podStartE2EDuration="17.544474705s" podCreationTimestamp="2025-10-08 18:29:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 18:29:51.539785706 +0000 UTC m=+1136.989628486" watchObservedRunningTime="2025-10-08 18:29:51.544474705 +0000 UTC m=+1136.994317475" Oct 08 18:29:51 crc kubenswrapper[4988]: I1008 18:29:51.546117 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-barbican-internal-svc" Oct 08 18:29:51 crc kubenswrapper[4988]: I1008 18:29:51.546481 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-barbican-public-svc" Oct 08 18:29:51 crc kubenswrapper[4988]: I1008 18:29:51.566538 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-6c486d5744-j86zn"] Oct 08 18:29:51 crc kubenswrapper[4988]: I1008 18:29:51.577022 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-54df4b685c-fr6bz" podStartSLOduration=3.577007369 podStartE2EDuration="3.577007369s" podCreationTimestamp="2025-10-08 18:29:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 18:29:51.575225294 +0000 UTC m=+1137.025068074" watchObservedRunningTime="2025-10-08 18:29:51.577007369 +0000 UTC m=+1137.026850139" Oct 08 18:29:51 crc kubenswrapper[4988]: I1008 18:29:51.619042 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 08 18:29:51 crc kubenswrapper[4988]: I1008 18:29:51.626815 4988 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Oct 08 18:29:51 crc kubenswrapper[4988]: I1008 18:29:51.650874 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/3a57537b-6115-4249-8d16-3ab0c4b6d21e-internal-tls-certs\") pod \"barbican-api-6c486d5744-j86zn\" (UID: \"3a57537b-6115-4249-8d16-3ab0c4b6d21e\") " pod="openstack/barbican-api-6c486d5744-j86zn" Oct 08 18:29:51 crc kubenswrapper[4988]: I1008 18:29:51.651043 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3a57537b-6115-4249-8d16-3ab0c4b6d21e-config-data\") pod \"barbican-api-6c486d5744-j86zn\" (UID: \"3a57537b-6115-4249-8d16-3ab0c4b6d21e\") " pod="openstack/barbican-api-6c486d5744-j86zn" Oct 08 18:29:51 crc kubenswrapper[4988]: I1008 18:29:51.651066 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gfb6j\" (UniqueName: \"kubernetes.io/projected/3a57537b-6115-4249-8d16-3ab0c4b6d21e-kube-api-access-gfb6j\") pod \"barbican-api-6c486d5744-j86zn\" (UID: \"3a57537b-6115-4249-8d16-3ab0c4b6d21e\") " pod="openstack/barbican-api-6c486d5744-j86zn" Oct 08 18:29:51 crc kubenswrapper[4988]: I1008 18:29:51.651185 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/3a57537b-6115-4249-8d16-3ab0c4b6d21e-public-tls-certs\") pod \"barbican-api-6c486d5744-j86zn\" (UID: \"3a57537b-6115-4249-8d16-3ab0c4b6d21e\") " pod="openstack/barbican-api-6c486d5744-j86zn" Oct 08 18:29:51 crc kubenswrapper[4988]: I1008 18:29:51.651216 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3a57537b-6115-4249-8d16-3ab0c4b6d21e-logs\") pod \"barbican-api-6c486d5744-j86zn\" (UID: \"3a57537b-6115-4249-8d16-3ab0c4b6d21e\") " pod="openstack/barbican-api-6c486d5744-j86zn" Oct 08 18:29:51 crc kubenswrapper[4988]: I1008 18:29:51.651231 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/3a57537b-6115-4249-8d16-3ab0c4b6d21e-config-data-custom\") pod \"barbican-api-6c486d5744-j86zn\" (UID: \"3a57537b-6115-4249-8d16-3ab0c4b6d21e\") " pod="openstack/barbican-api-6c486d5744-j86zn" Oct 08 18:29:51 crc kubenswrapper[4988]: I1008 18:29:51.651267 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3a57537b-6115-4249-8d16-3ab0c4b6d21e-combined-ca-bundle\") pod \"barbican-api-6c486d5744-j86zn\" (UID: \"3a57537b-6115-4249-8d16-3ab0c4b6d21e\") " pod="openstack/barbican-api-6c486d5744-j86zn" Oct 08 18:29:51 crc kubenswrapper[4988]: I1008 18:29:51.655754 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 08 18:29:51 crc kubenswrapper[4988]: I1008 18:29:51.658371 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 08 18:29:51 crc kubenswrapper[4988]: I1008 18:29:51.661063 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 08 18:29:51 crc kubenswrapper[4988]: I1008 18:29:51.668291 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 08 18:29:51 crc kubenswrapper[4988]: I1008 18:29:51.673732 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 08 18:29:51 crc kubenswrapper[4988]: I1008 18:29:51.759427 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3a57537b-6115-4249-8d16-3ab0c4b6d21e-combined-ca-bundle\") pod \"barbican-api-6c486d5744-j86zn\" (UID: \"3a57537b-6115-4249-8d16-3ab0c4b6d21e\") " pod="openstack/barbican-api-6c486d5744-j86zn" Oct 08 18:29:51 crc kubenswrapper[4988]: I1008 18:29:51.759711 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/9e8e39d2-57c8-4ea6-9c83-dcb8d75274cd-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"9e8e39d2-57c8-4ea6-9c83-dcb8d75274cd\") " pod="openstack/ceilometer-0" Oct 08 18:29:51 crc kubenswrapper[4988]: I1008 18:29:51.759848 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/3a57537b-6115-4249-8d16-3ab0c4b6d21e-internal-tls-certs\") pod \"barbican-api-6c486d5744-j86zn\" (UID: \"3a57537b-6115-4249-8d16-3ab0c4b6d21e\") " pod="openstack/barbican-api-6c486d5744-j86zn" Oct 08 18:29:51 crc kubenswrapper[4988]: I1008 18:29:51.760013 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9e8e39d2-57c8-4ea6-9c83-dcb8d75274cd-scripts\") pod \"ceilometer-0\" (UID: \"9e8e39d2-57c8-4ea6-9c83-dcb8d75274cd\") " pod="openstack/ceilometer-0" Oct 08 18:29:51 crc kubenswrapper[4988]: I1008 18:29:51.760115 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9e8e39d2-57c8-4ea6-9c83-dcb8d75274cd-log-httpd\") pod \"ceilometer-0\" (UID: \"9e8e39d2-57c8-4ea6-9c83-dcb8d75274cd\") " pod="openstack/ceilometer-0" Oct 08 18:29:51 crc kubenswrapper[4988]: I1008 18:29:51.760238 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9e8e39d2-57c8-4ea6-9c83-dcb8d75274cd-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"9e8e39d2-57c8-4ea6-9c83-dcb8d75274cd\") " pod="openstack/ceilometer-0" Oct 08 18:29:51 crc kubenswrapper[4988]: I1008 18:29:51.760353 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3a57537b-6115-4249-8d16-3ab0c4b6d21e-config-data\") pod \"barbican-api-6c486d5744-j86zn\" (UID: \"3a57537b-6115-4249-8d16-3ab0c4b6d21e\") " pod="openstack/barbican-api-6c486d5744-j86zn" Oct 08 18:29:51 crc kubenswrapper[4988]: I1008 18:29:51.760481 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9e8e39d2-57c8-4ea6-9c83-dcb8d75274cd-run-httpd\") pod \"ceilometer-0\" (UID: \"9e8e39d2-57c8-4ea6-9c83-dcb8d75274cd\") " pod="openstack/ceilometer-0" Oct 08 18:29:51 crc kubenswrapper[4988]: I1008 18:29:51.760575 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gfb6j\" (UniqueName: \"kubernetes.io/projected/3a57537b-6115-4249-8d16-3ab0c4b6d21e-kube-api-access-gfb6j\") pod \"barbican-api-6c486d5744-j86zn\" (UID: \"3a57537b-6115-4249-8d16-3ab0c4b6d21e\") " pod="openstack/barbican-api-6c486d5744-j86zn" Oct 08 18:29:51 crc kubenswrapper[4988]: I1008 18:29:51.760667 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9e8e39d2-57c8-4ea6-9c83-dcb8d75274cd-config-data\") pod \"ceilometer-0\" (UID: \"9e8e39d2-57c8-4ea6-9c83-dcb8d75274cd\") " pod="openstack/ceilometer-0" Oct 08 18:29:51 crc kubenswrapper[4988]: I1008 18:29:51.760791 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4txp6\" (UniqueName: \"kubernetes.io/projected/9e8e39d2-57c8-4ea6-9c83-dcb8d75274cd-kube-api-access-4txp6\") pod \"ceilometer-0\" (UID: \"9e8e39d2-57c8-4ea6-9c83-dcb8d75274cd\") " pod="openstack/ceilometer-0" Oct 08 18:29:51 crc kubenswrapper[4988]: I1008 18:29:51.760894 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/3a57537b-6115-4249-8d16-3ab0c4b6d21e-public-tls-certs\") pod \"barbican-api-6c486d5744-j86zn\" (UID: \"3a57537b-6115-4249-8d16-3ab0c4b6d21e\") " pod="openstack/barbican-api-6c486d5744-j86zn" Oct 08 18:29:51 crc kubenswrapper[4988]: I1008 18:29:51.760979 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/3a57537b-6115-4249-8d16-3ab0c4b6d21e-config-data-custom\") pod \"barbican-api-6c486d5744-j86zn\" (UID: \"3a57537b-6115-4249-8d16-3ab0c4b6d21e\") " pod="openstack/barbican-api-6c486d5744-j86zn" Oct 08 18:29:51 crc kubenswrapper[4988]: I1008 18:29:51.761045 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3a57537b-6115-4249-8d16-3ab0c4b6d21e-logs\") pod \"barbican-api-6c486d5744-j86zn\" (UID: \"3a57537b-6115-4249-8d16-3ab0c4b6d21e\") " pod="openstack/barbican-api-6c486d5744-j86zn" Oct 08 18:29:51 crc kubenswrapper[4988]: I1008 18:29:51.761399 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3a57537b-6115-4249-8d16-3ab0c4b6d21e-logs\") pod \"barbican-api-6c486d5744-j86zn\" (UID: \"3a57537b-6115-4249-8d16-3ab0c4b6d21e\") " pod="openstack/barbican-api-6c486d5744-j86zn" Oct 08 18:29:51 crc kubenswrapper[4988]: I1008 18:29:51.766462 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/3a57537b-6115-4249-8d16-3ab0c4b6d21e-public-tls-certs\") pod \"barbican-api-6c486d5744-j86zn\" (UID: \"3a57537b-6115-4249-8d16-3ab0c4b6d21e\") " pod="openstack/barbican-api-6c486d5744-j86zn" Oct 08 18:29:51 crc kubenswrapper[4988]: I1008 18:29:51.767057 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/3a57537b-6115-4249-8d16-3ab0c4b6d21e-config-data-custom\") pod \"barbican-api-6c486d5744-j86zn\" (UID: \"3a57537b-6115-4249-8d16-3ab0c4b6d21e\") " pod="openstack/barbican-api-6c486d5744-j86zn" Oct 08 18:29:51 crc kubenswrapper[4988]: I1008 18:29:51.767352 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3a57537b-6115-4249-8d16-3ab0c4b6d21e-combined-ca-bundle\") pod \"barbican-api-6c486d5744-j86zn\" (UID: \"3a57537b-6115-4249-8d16-3ab0c4b6d21e\") " pod="openstack/barbican-api-6c486d5744-j86zn" Oct 08 18:29:51 crc kubenswrapper[4988]: I1008 18:29:51.772873 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/3a57537b-6115-4249-8d16-3ab0c4b6d21e-internal-tls-certs\") pod \"barbican-api-6c486d5744-j86zn\" (UID: \"3a57537b-6115-4249-8d16-3ab0c4b6d21e\") " pod="openstack/barbican-api-6c486d5744-j86zn" Oct 08 18:29:51 crc kubenswrapper[4988]: I1008 18:29:51.780092 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gfb6j\" (UniqueName: \"kubernetes.io/projected/3a57537b-6115-4249-8d16-3ab0c4b6d21e-kube-api-access-gfb6j\") pod \"barbican-api-6c486d5744-j86zn\" (UID: \"3a57537b-6115-4249-8d16-3ab0c4b6d21e\") " pod="openstack/barbican-api-6c486d5744-j86zn" Oct 08 18:29:51 crc kubenswrapper[4988]: I1008 18:29:51.780431 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3a57537b-6115-4249-8d16-3ab0c4b6d21e-config-data\") pod \"barbican-api-6c486d5744-j86zn\" (UID: \"3a57537b-6115-4249-8d16-3ab0c4b6d21e\") " pod="openstack/barbican-api-6c486d5744-j86zn" Oct 08 18:29:51 crc kubenswrapper[4988]: I1008 18:29:51.865047 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9e8e39d2-57c8-4ea6-9c83-dcb8d75274cd-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"9e8e39d2-57c8-4ea6-9c83-dcb8d75274cd\") " pod="openstack/ceilometer-0" Oct 08 18:29:51 crc kubenswrapper[4988]: I1008 18:29:51.865110 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9e8e39d2-57c8-4ea6-9c83-dcb8d75274cd-run-httpd\") pod \"ceilometer-0\" (UID: \"9e8e39d2-57c8-4ea6-9c83-dcb8d75274cd\") " pod="openstack/ceilometer-0" Oct 08 18:29:51 crc kubenswrapper[4988]: I1008 18:29:51.865132 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9e8e39d2-57c8-4ea6-9c83-dcb8d75274cd-config-data\") pod \"ceilometer-0\" (UID: \"9e8e39d2-57c8-4ea6-9c83-dcb8d75274cd\") " pod="openstack/ceilometer-0" Oct 08 18:29:51 crc kubenswrapper[4988]: I1008 18:29:51.865171 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4txp6\" (UniqueName: \"kubernetes.io/projected/9e8e39d2-57c8-4ea6-9c83-dcb8d75274cd-kube-api-access-4txp6\") pod \"ceilometer-0\" (UID: \"9e8e39d2-57c8-4ea6-9c83-dcb8d75274cd\") " pod="openstack/ceilometer-0" Oct 08 18:29:51 crc kubenswrapper[4988]: I1008 18:29:51.865234 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/9e8e39d2-57c8-4ea6-9c83-dcb8d75274cd-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"9e8e39d2-57c8-4ea6-9c83-dcb8d75274cd\") " pod="openstack/ceilometer-0" Oct 08 18:29:51 crc kubenswrapper[4988]: I1008 18:29:51.865290 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9e8e39d2-57c8-4ea6-9c83-dcb8d75274cd-scripts\") pod \"ceilometer-0\" (UID: \"9e8e39d2-57c8-4ea6-9c83-dcb8d75274cd\") " pod="openstack/ceilometer-0" Oct 08 18:29:51 crc kubenswrapper[4988]: I1008 18:29:51.865308 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9e8e39d2-57c8-4ea6-9c83-dcb8d75274cd-log-httpd\") pod \"ceilometer-0\" (UID: \"9e8e39d2-57c8-4ea6-9c83-dcb8d75274cd\") " pod="openstack/ceilometer-0" Oct 08 18:29:51 crc kubenswrapper[4988]: I1008 18:29:51.866199 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9e8e39d2-57c8-4ea6-9c83-dcb8d75274cd-run-httpd\") pod \"ceilometer-0\" (UID: \"9e8e39d2-57c8-4ea6-9c83-dcb8d75274cd\") " pod="openstack/ceilometer-0" Oct 08 18:29:51 crc kubenswrapper[4988]: I1008 18:29:51.866238 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-6c486d5744-j86zn" Oct 08 18:29:51 crc kubenswrapper[4988]: I1008 18:29:51.866277 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9e8e39d2-57c8-4ea6-9c83-dcb8d75274cd-log-httpd\") pod \"ceilometer-0\" (UID: \"9e8e39d2-57c8-4ea6-9c83-dcb8d75274cd\") " pod="openstack/ceilometer-0" Oct 08 18:29:51 crc kubenswrapper[4988]: I1008 18:29:51.869138 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9e8e39d2-57c8-4ea6-9c83-dcb8d75274cd-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"9e8e39d2-57c8-4ea6-9c83-dcb8d75274cd\") " pod="openstack/ceilometer-0" Oct 08 18:29:51 crc kubenswrapper[4988]: I1008 18:29:51.870800 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9e8e39d2-57c8-4ea6-9c83-dcb8d75274cd-scripts\") pod \"ceilometer-0\" (UID: \"9e8e39d2-57c8-4ea6-9c83-dcb8d75274cd\") " pod="openstack/ceilometer-0" Oct 08 18:29:51 crc kubenswrapper[4988]: I1008 18:29:51.883593 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9e8e39d2-57c8-4ea6-9c83-dcb8d75274cd-config-data\") pod \"ceilometer-0\" (UID: \"9e8e39d2-57c8-4ea6-9c83-dcb8d75274cd\") " pod="openstack/ceilometer-0" Oct 08 18:29:51 crc kubenswrapper[4988]: I1008 18:29:51.886444 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4txp6\" (UniqueName: \"kubernetes.io/projected/9e8e39d2-57c8-4ea6-9c83-dcb8d75274cd-kube-api-access-4txp6\") pod \"ceilometer-0\" (UID: \"9e8e39d2-57c8-4ea6-9c83-dcb8d75274cd\") " pod="openstack/ceilometer-0" Oct 08 18:29:51 crc kubenswrapper[4988]: I1008 18:29:51.887787 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/9e8e39d2-57c8-4ea6-9c83-dcb8d75274cd-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"9e8e39d2-57c8-4ea6-9c83-dcb8d75274cd\") " pod="openstack/ceilometer-0" Oct 08 18:29:51 crc kubenswrapper[4988]: I1008 18:29:51.890430 4988 scope.go:117] "RemoveContainer" containerID="2b5c761d77af17eace93d9cd355762d47e5b9791b61461447ac5dfb3542405b0" Oct 08 18:29:51 crc kubenswrapper[4988]: I1008 18:29:51.975002 4988 scope.go:117] "RemoveContainer" containerID="530180f4bca5eaf8c14c742c890bbe065b3ab6a329ac14ec714a341c4b5e100e" Oct 08 18:29:51 crc kubenswrapper[4988]: I1008 18:29:51.977279 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 08 18:29:52 crc kubenswrapper[4988]: I1008 18:29:52.509406 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-6c486d5744-j86zn"] Oct 08 18:29:52 crc kubenswrapper[4988]: W1008 18:29:52.511786 4988 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3a57537b_6115_4249_8d16_3ab0c4b6d21e.slice/crio-fe2b1b18022fcd542d4117b8811974fd4ee1907095a986d6338c7dd0db4ca3b9 WatchSource:0}: Error finding container fe2b1b18022fcd542d4117b8811974fd4ee1907095a986d6338c7dd0db4ca3b9: Status 404 returned error can't find the container with id fe2b1b18022fcd542d4117b8811974fd4ee1907095a986d6338c7dd0db4ca3b9 Oct 08 18:29:52 crc kubenswrapper[4988]: I1008 18:29:52.520613 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-8694c889b8-lhm4g" event={"ID":"c9f172a3-9188-4f5f-bf3a-fbc3b0dc5206","Type":"ContainerStarted","Data":"c7e441bdd6242790ab99582ae7f776a6bd322dc0c9ca3985c984d45c5cc43a5c"} Oct 08 18:29:52 crc kubenswrapper[4988]: I1008 18:29:52.521977 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-7bf697f8df-h54ft" event={"ID":"9159c915-07a7-4b41-8672-8b9d6a9572ca","Type":"ContainerStarted","Data":"2927fb800c44516806e01d37c729bc04efb6e3981dccab523101ed4405525b47"} Oct 08 18:29:52 crc kubenswrapper[4988]: I1008 18:29:52.604009 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 08 18:29:52 crc kubenswrapper[4988]: I1008 18:29:52.716424 4988 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-69894dfcd9-948wp" podUID="7a831457-b429-4339-b166-afed001c770d" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.136:5353: i/o timeout" Oct 08 18:29:52 crc kubenswrapper[4988]: I1008 18:29:52.955177 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-c44756996-dfw4k" Oct 08 18:29:52 crc kubenswrapper[4988]: I1008 18:29:52.956985 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-c44756996-dfw4k" Oct 08 18:29:53 crc kubenswrapper[4988]: I1008 18:29:53.248220 4988 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8fed08cf-45cb-4fd2-92af-9f30513c0b72" path="/var/lib/kubelet/pods/8fed08cf-45cb-4fd2-92af-9f30513c0b72/volumes" Oct 08 18:29:53 crc kubenswrapper[4988]: I1008 18:29:53.338145 4988 patch_prober.go:28] interesting pod/machine-config-daemon-nm54f container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 18:29:53 crc kubenswrapper[4988]: I1008 18:29:53.338436 4988 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 18:29:53 crc kubenswrapper[4988]: I1008 18:29:53.532657 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-6c486d5744-j86zn" event={"ID":"3a57537b-6115-4249-8d16-3ab0c4b6d21e","Type":"ContainerStarted","Data":"0f23a2df4066a02929dd3edb8dc8d068809211dbfad9bf36b14e3f21916284e7"} Oct 08 18:29:53 crc kubenswrapper[4988]: I1008 18:29:53.533617 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-6c486d5744-j86zn" event={"ID":"3a57537b-6115-4249-8d16-3ab0c4b6d21e","Type":"ContainerStarted","Data":"8ef71a96a488da8760db257dae569210b2903674028f06127874742bfeb624fd"} Oct 08 18:29:53 crc kubenswrapper[4988]: I1008 18:29:53.533691 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-6c486d5744-j86zn" Oct 08 18:29:53 crc kubenswrapper[4988]: I1008 18:29:53.533769 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-6c486d5744-j86zn" Oct 08 18:29:53 crc kubenswrapper[4988]: I1008 18:29:53.533832 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-6c486d5744-j86zn" event={"ID":"3a57537b-6115-4249-8d16-3ab0c4b6d21e","Type":"ContainerStarted","Data":"fe2b1b18022fcd542d4117b8811974fd4ee1907095a986d6338c7dd0db4ca3b9"} Oct 08 18:29:53 crc kubenswrapper[4988]: I1008 18:29:53.534521 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-7bf697f8df-h54ft" event={"ID":"9159c915-07a7-4b41-8672-8b9d6a9572ca","Type":"ContainerStarted","Data":"1f77c788702456fc178f26b1d4a4449a94e1ac2f3f33ad505531b3885e2f8f15"} Oct 08 18:29:53 crc kubenswrapper[4988]: I1008 18:29:53.536397 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9e8e39d2-57c8-4ea6-9c83-dcb8d75274cd","Type":"ContainerStarted","Data":"9640cc556082067881ad4130d32e24914d42ee959a828c2f3e071e7ca9d49656"} Oct 08 18:29:53 crc kubenswrapper[4988]: I1008 18:29:53.536449 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9e8e39d2-57c8-4ea6-9c83-dcb8d75274cd","Type":"ContainerStarted","Data":"c44e85e774111fa2a83bd1e90e873d37a116d12f0d295b4b11395053d875cf29"} Oct 08 18:29:53 crc kubenswrapper[4988]: I1008 18:29:53.538163 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-8694c889b8-lhm4g" event={"ID":"c9f172a3-9188-4f5f-bf3a-fbc3b0dc5206","Type":"ContainerStarted","Data":"3251e38d412d5c6a4e41496647e54dfc1bc7728f57aa1932956313fe00c0a86b"} Oct 08 18:29:53 crc kubenswrapper[4988]: I1008 18:29:53.560131 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-6c486d5744-j86zn" podStartSLOduration=2.560114788 podStartE2EDuration="2.560114788s" podCreationTimestamp="2025-10-08 18:29:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 18:29:53.552640812 +0000 UTC m=+1139.002483602" watchObservedRunningTime="2025-10-08 18:29:53.560114788 +0000 UTC m=+1139.009957558" Oct 08 18:29:53 crc kubenswrapper[4988]: I1008 18:29:53.581804 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-keystone-listener-8694c889b8-lhm4g" podStartSLOduration=3.412535647 podStartE2EDuration="5.581784861s" podCreationTimestamp="2025-10-08 18:29:48 +0000 UTC" firstStartedPulling="2025-10-08 18:29:49.733925143 +0000 UTC m=+1135.183767913" lastFinishedPulling="2025-10-08 18:29:51.903174357 +0000 UTC m=+1137.353017127" observedRunningTime="2025-10-08 18:29:53.576820044 +0000 UTC m=+1139.026662824" watchObservedRunningTime="2025-10-08 18:29:53.581784861 +0000 UTC m=+1139.031627631" Oct 08 18:29:53 crc kubenswrapper[4988]: I1008 18:29:53.600871 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-worker-7bf697f8df-h54ft" podStartSLOduration=3.419589368 podStartE2EDuration="5.600853681s" podCreationTimestamp="2025-10-08 18:29:48 +0000 UTC" firstStartedPulling="2025-10-08 18:29:49.733313364 +0000 UTC m=+1135.183156134" lastFinishedPulling="2025-10-08 18:29:51.914577677 +0000 UTC m=+1137.364420447" observedRunningTime="2025-10-08 18:29:53.598209028 +0000 UTC m=+1139.048051808" watchObservedRunningTime="2025-10-08 18:29:53.600853681 +0000 UTC m=+1139.050696451" Oct 08 18:29:54 crc kubenswrapper[4988]: I1008 18:29:54.719184 4988 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Oct 08 18:29:54 crc kubenswrapper[4988]: I1008 18:29:54.720794 4988 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Oct 08 18:29:54 crc kubenswrapper[4988]: I1008 18:29:54.745155 4988 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Oct 08 18:29:54 crc kubenswrapper[4988]: I1008 18:29:54.760184 4988 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Oct 08 18:29:55 crc kubenswrapper[4988]: I1008 18:29:55.561654 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9e8e39d2-57c8-4ea6-9c83-dcb8d75274cd","Type":"ContainerStarted","Data":"46c9d841d83e09ba232a18c37dc58ed0a217dca805674f3203c8f5d7cbddb5cb"} Oct 08 18:29:55 crc kubenswrapper[4988]: I1008 18:29:55.562053 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Oct 08 18:29:55 crc kubenswrapper[4988]: I1008 18:29:55.562080 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Oct 08 18:29:56 crc kubenswrapper[4988]: I1008 18:29:56.598291 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9e8e39d2-57c8-4ea6-9c83-dcb8d75274cd","Type":"ContainerStarted","Data":"3611c104fb9f40d7e4eb1645a5029783c73aa03a8d819c857e23f703653a58e3"} Oct 08 18:29:57 crc kubenswrapper[4988]: I1008 18:29:57.222592 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/keystone-6dbc6dc9bd-hk8qw" Oct 08 18:29:57 crc kubenswrapper[4988]: I1008 18:29:57.699634 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Oct 08 18:29:57 crc kubenswrapper[4988]: I1008 18:29:57.700030 4988 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 08 18:29:57 crc kubenswrapper[4988]: I1008 18:29:57.778118 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Oct 08 18:29:58 crc kubenswrapper[4988]: I1008 18:29:58.065702 4988 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Oct 08 18:29:58 crc kubenswrapper[4988]: I1008 18:29:58.065758 4988 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Oct 08 18:29:58 crc kubenswrapper[4988]: I1008 18:29:58.102334 4988 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Oct 08 18:29:58 crc kubenswrapper[4988]: I1008 18:29:58.110479 4988 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Oct 08 18:29:58 crc kubenswrapper[4988]: I1008 18:29:58.634673 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9e8e39d2-57c8-4ea6-9c83-dcb8d75274cd","Type":"ContainerStarted","Data":"1e28870498bf7017bb7aaf972c99e33b912ab2db76680063b2dd790406e59691"} Oct 08 18:29:58 crc kubenswrapper[4988]: I1008 18:29:58.636190 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Oct 08 18:29:58 crc kubenswrapper[4988]: I1008 18:29:58.636233 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Oct 08 18:29:58 crc kubenswrapper[4988]: I1008 18:29:58.636249 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 08 18:29:58 crc kubenswrapper[4988]: I1008 18:29:58.667290 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.848550786 podStartE2EDuration="7.667273786s" podCreationTimestamp="2025-10-08 18:29:51 +0000 UTC" firstStartedPulling="2025-10-08 18:29:52.615360918 +0000 UTC m=+1138.065203688" lastFinishedPulling="2025-10-08 18:29:57.434083918 +0000 UTC m=+1142.883926688" observedRunningTime="2025-10-08 18:29:58.66010343 +0000 UTC m=+1144.109946200" watchObservedRunningTime="2025-10-08 18:29:58.667273786 +0000 UTC m=+1144.117116556" Oct 08 18:29:59 crc kubenswrapper[4988]: I1008 18:29:59.120216 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-54df4b685c-fr6bz" Oct 08 18:29:59 crc kubenswrapper[4988]: I1008 18:29:59.220854 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-77dd5cf987-5qhqw"] Oct 08 18:29:59 crc kubenswrapper[4988]: I1008 18:29:59.221111 4988 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-77dd5cf987-5qhqw" podUID="33fb9c50-b318-4279-af5c-605bb2f0d3c1" containerName="dnsmasq-dns" containerID="cri-o://8ac057a4f4bc0915ef6b7c8284d0fb581b4eadb20d3d7d433254afafbe52a8a2" gracePeriod=10 Oct 08 18:29:59 crc kubenswrapper[4988]: I1008 18:29:59.647581 4988 generic.go:334] "Generic (PLEG): container finished" podID="33fb9c50-b318-4279-af5c-605bb2f0d3c1" containerID="8ac057a4f4bc0915ef6b7c8284d0fb581b4eadb20d3d7d433254afafbe52a8a2" exitCode=0 Oct 08 18:29:59 crc kubenswrapper[4988]: I1008 18:29:59.647687 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-77dd5cf987-5qhqw" event={"ID":"33fb9c50-b318-4279-af5c-605bb2f0d3c1","Type":"ContainerDied","Data":"8ac057a4f4bc0915ef6b7c8284d0fb581b4eadb20d3d7d433254afafbe52a8a2"} Oct 08 18:29:59 crc kubenswrapper[4988]: I1008 18:29:59.654571 4988 generic.go:334] "Generic (PLEG): container finished" podID="de747e73-c280-424f-af48-3135bec91b55" containerID="f82c2c46fdece62aac8a106f15e5fddfb69696f72a1f951b75f376ae4f50e07b" exitCode=0 Oct 08 18:29:59 crc kubenswrapper[4988]: I1008 18:29:59.655788 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-wvjsl" event={"ID":"de747e73-c280-424f-af48-3135bec91b55","Type":"ContainerDied","Data":"f82c2c46fdece62aac8a106f15e5fddfb69696f72a1f951b75f376ae4f50e07b"} Oct 08 18:29:59 crc kubenswrapper[4988]: I1008 18:29:59.810921 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstackclient"] Oct 08 18:29:59 crc kubenswrapper[4988]: I1008 18:29:59.817289 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Oct 08 18:29:59 crc kubenswrapper[4988]: I1008 18:29:59.822785 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstackclient-openstackclient-dockercfg-hnrkw" Oct 08 18:29:59 crc kubenswrapper[4988]: I1008 18:29:59.823049 4988 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config" Oct 08 18:29:59 crc kubenswrapper[4988]: I1008 18:29:59.823261 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-config-secret" Oct 08 18:29:59 crc kubenswrapper[4988]: I1008 18:29:59.833662 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-77dd5cf987-5qhqw" Oct 08 18:29:59 crc kubenswrapper[4988]: I1008 18:29:59.847659 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Oct 08 18:29:59 crc kubenswrapper[4988]: I1008 18:29:59.908112 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pm5r2\" (UniqueName: \"kubernetes.io/projected/c97aab35-bbee-41a6-88b2-801b0b097a6d-kube-api-access-pm5r2\") pod \"openstackclient\" (UID: \"c97aab35-bbee-41a6-88b2-801b0b097a6d\") " pod="openstack/openstackclient" Oct 08 18:29:59 crc kubenswrapper[4988]: I1008 18:29:59.908160 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/c97aab35-bbee-41a6-88b2-801b0b097a6d-openstack-config\") pod \"openstackclient\" (UID: \"c97aab35-bbee-41a6-88b2-801b0b097a6d\") " pod="openstack/openstackclient" Oct 08 18:29:59 crc kubenswrapper[4988]: I1008 18:29:59.908197 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c97aab35-bbee-41a6-88b2-801b0b097a6d-combined-ca-bundle\") pod \"openstackclient\" (UID: \"c97aab35-bbee-41a6-88b2-801b0b097a6d\") " pod="openstack/openstackclient" Oct 08 18:29:59 crc kubenswrapper[4988]: I1008 18:29:59.908462 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/c97aab35-bbee-41a6-88b2-801b0b097a6d-openstack-config-secret\") pod \"openstackclient\" (UID: \"c97aab35-bbee-41a6-88b2-801b0b097a6d\") " pod="openstack/openstackclient" Oct 08 18:30:00 crc kubenswrapper[4988]: I1008 18:30:00.009444 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/33fb9c50-b318-4279-af5c-605bb2f0d3c1-dns-svc\") pod \"33fb9c50-b318-4279-af5c-605bb2f0d3c1\" (UID: \"33fb9c50-b318-4279-af5c-605bb2f0d3c1\") " Oct 08 18:30:00 crc kubenswrapper[4988]: I1008 18:30:00.009511 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/33fb9c50-b318-4279-af5c-605bb2f0d3c1-ovsdbserver-sb\") pod \"33fb9c50-b318-4279-af5c-605bb2f0d3c1\" (UID: \"33fb9c50-b318-4279-af5c-605bb2f0d3c1\") " Oct 08 18:30:00 crc kubenswrapper[4988]: I1008 18:30:00.009611 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-566fx\" (UniqueName: \"kubernetes.io/projected/33fb9c50-b318-4279-af5c-605bb2f0d3c1-kube-api-access-566fx\") pod \"33fb9c50-b318-4279-af5c-605bb2f0d3c1\" (UID: \"33fb9c50-b318-4279-af5c-605bb2f0d3c1\") " Oct 08 18:30:00 crc kubenswrapper[4988]: I1008 18:30:00.009632 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/33fb9c50-b318-4279-af5c-605bb2f0d3c1-dns-swift-storage-0\") pod \"33fb9c50-b318-4279-af5c-605bb2f0d3c1\" (UID: \"33fb9c50-b318-4279-af5c-605bb2f0d3c1\") " Oct 08 18:30:00 crc kubenswrapper[4988]: I1008 18:30:00.009696 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/33fb9c50-b318-4279-af5c-605bb2f0d3c1-ovsdbserver-nb\") pod \"33fb9c50-b318-4279-af5c-605bb2f0d3c1\" (UID: \"33fb9c50-b318-4279-af5c-605bb2f0d3c1\") " Oct 08 18:30:00 crc kubenswrapper[4988]: I1008 18:30:00.009711 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/33fb9c50-b318-4279-af5c-605bb2f0d3c1-config\") pod \"33fb9c50-b318-4279-af5c-605bb2f0d3c1\" (UID: \"33fb9c50-b318-4279-af5c-605bb2f0d3c1\") " Oct 08 18:30:00 crc kubenswrapper[4988]: I1008 18:30:00.009919 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pm5r2\" (UniqueName: \"kubernetes.io/projected/c97aab35-bbee-41a6-88b2-801b0b097a6d-kube-api-access-pm5r2\") pod \"openstackclient\" (UID: \"c97aab35-bbee-41a6-88b2-801b0b097a6d\") " pod="openstack/openstackclient" Oct 08 18:30:00 crc kubenswrapper[4988]: I1008 18:30:00.009948 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/c97aab35-bbee-41a6-88b2-801b0b097a6d-openstack-config\") pod \"openstackclient\" (UID: \"c97aab35-bbee-41a6-88b2-801b0b097a6d\") " pod="openstack/openstackclient" Oct 08 18:30:00 crc kubenswrapper[4988]: I1008 18:30:00.009980 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c97aab35-bbee-41a6-88b2-801b0b097a6d-combined-ca-bundle\") pod \"openstackclient\" (UID: \"c97aab35-bbee-41a6-88b2-801b0b097a6d\") " pod="openstack/openstackclient" Oct 08 18:30:00 crc kubenswrapper[4988]: I1008 18:30:00.010044 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/c97aab35-bbee-41a6-88b2-801b0b097a6d-openstack-config-secret\") pod \"openstackclient\" (UID: \"c97aab35-bbee-41a6-88b2-801b0b097a6d\") " pod="openstack/openstackclient" Oct 08 18:30:00 crc kubenswrapper[4988]: I1008 18:30:00.011048 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/c97aab35-bbee-41a6-88b2-801b0b097a6d-openstack-config\") pod \"openstackclient\" (UID: \"c97aab35-bbee-41a6-88b2-801b0b097a6d\") " pod="openstack/openstackclient" Oct 08 18:30:00 crc kubenswrapper[4988]: I1008 18:30:00.022024 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/c97aab35-bbee-41a6-88b2-801b0b097a6d-openstack-config-secret\") pod \"openstackclient\" (UID: \"c97aab35-bbee-41a6-88b2-801b0b097a6d\") " pod="openstack/openstackclient" Oct 08 18:30:00 crc kubenswrapper[4988]: I1008 18:30:00.027552 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/33fb9c50-b318-4279-af5c-605bb2f0d3c1-kube-api-access-566fx" (OuterVolumeSpecName: "kube-api-access-566fx") pod "33fb9c50-b318-4279-af5c-605bb2f0d3c1" (UID: "33fb9c50-b318-4279-af5c-605bb2f0d3c1"). InnerVolumeSpecName "kube-api-access-566fx". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:30:00 crc kubenswrapper[4988]: I1008 18:30:00.029463 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c97aab35-bbee-41a6-88b2-801b0b097a6d-combined-ca-bundle\") pod \"openstackclient\" (UID: \"c97aab35-bbee-41a6-88b2-801b0b097a6d\") " pod="openstack/openstackclient" Oct 08 18:30:00 crc kubenswrapper[4988]: I1008 18:30:00.031365 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pm5r2\" (UniqueName: \"kubernetes.io/projected/c97aab35-bbee-41a6-88b2-801b0b097a6d-kube-api-access-pm5r2\") pod \"openstackclient\" (UID: \"c97aab35-bbee-41a6-88b2-801b0b097a6d\") " pod="openstack/openstackclient" Oct 08 18:30:00 crc kubenswrapper[4988]: I1008 18:30:00.095988 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/33fb9c50-b318-4279-af5c-605bb2f0d3c1-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "33fb9c50-b318-4279-af5c-605bb2f0d3c1" (UID: "33fb9c50-b318-4279-af5c-605bb2f0d3c1"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 18:30:00 crc kubenswrapper[4988]: I1008 18:30:00.104019 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/33fb9c50-b318-4279-af5c-605bb2f0d3c1-config" (OuterVolumeSpecName: "config") pod "33fb9c50-b318-4279-af5c-605bb2f0d3c1" (UID: "33fb9c50-b318-4279-af5c-605bb2f0d3c1"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 18:30:00 crc kubenswrapper[4988]: I1008 18:30:00.106588 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/33fb9c50-b318-4279-af5c-605bb2f0d3c1-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "33fb9c50-b318-4279-af5c-605bb2f0d3c1" (UID: "33fb9c50-b318-4279-af5c-605bb2f0d3c1"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 18:30:00 crc kubenswrapper[4988]: I1008 18:30:00.112706 4988 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/33fb9c50-b318-4279-af5c-605bb2f0d3c1-config\") on node \"crc\" DevicePath \"\"" Oct 08 18:30:00 crc kubenswrapper[4988]: I1008 18:30:00.112736 4988 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/33fb9c50-b318-4279-af5c-605bb2f0d3c1-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 08 18:30:00 crc kubenswrapper[4988]: I1008 18:30:00.112780 4988 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/33fb9c50-b318-4279-af5c-605bb2f0d3c1-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 08 18:30:00 crc kubenswrapper[4988]: I1008 18:30:00.112792 4988 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-566fx\" (UniqueName: \"kubernetes.io/projected/33fb9c50-b318-4279-af5c-605bb2f0d3c1-kube-api-access-566fx\") on node \"crc\" DevicePath \"\"" Oct 08 18:30:00 crc kubenswrapper[4988]: I1008 18:30:00.128862 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/33fb9c50-b318-4279-af5c-605bb2f0d3c1-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "33fb9c50-b318-4279-af5c-605bb2f0d3c1" (UID: "33fb9c50-b318-4279-af5c-605bb2f0d3c1"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 18:30:00 crc kubenswrapper[4988]: I1008 18:30:00.146711 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/33fb9c50-b318-4279-af5c-605bb2f0d3c1-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "33fb9c50-b318-4279-af5c-605bb2f0d3c1" (UID: "33fb9c50-b318-4279-af5c-605bb2f0d3c1"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 18:30:00 crc kubenswrapper[4988]: I1008 18:30:00.146978 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29332470-lbpdq"] Oct 08 18:30:00 crc kubenswrapper[4988]: E1008 18:30:00.147464 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="33fb9c50-b318-4279-af5c-605bb2f0d3c1" containerName="dnsmasq-dns" Oct 08 18:30:00 crc kubenswrapper[4988]: I1008 18:30:00.149513 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="33fb9c50-b318-4279-af5c-605bb2f0d3c1" containerName="dnsmasq-dns" Oct 08 18:30:00 crc kubenswrapper[4988]: E1008 18:30:00.149644 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="33fb9c50-b318-4279-af5c-605bb2f0d3c1" containerName="init" Oct 08 18:30:00 crc kubenswrapper[4988]: I1008 18:30:00.149668 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="33fb9c50-b318-4279-af5c-605bb2f0d3c1" containerName="init" Oct 08 18:30:00 crc kubenswrapper[4988]: I1008 18:30:00.150008 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="33fb9c50-b318-4279-af5c-605bb2f0d3c1" containerName="dnsmasq-dns" Oct 08 18:30:00 crc kubenswrapper[4988]: I1008 18:30:00.150901 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29332470-lbpdq" Oct 08 18:30:00 crc kubenswrapper[4988]: I1008 18:30:00.152338 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Oct 08 18:30:00 crc kubenswrapper[4988]: I1008 18:30:00.161659 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 08 18:30:00 crc kubenswrapper[4988]: I1008 18:30:00.162746 4988 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 08 18:30:00 crc kubenswrapper[4988]: I1008 18:30:00.168504 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29332470-lbpdq"] Oct 08 18:30:00 crc kubenswrapper[4988]: I1008 18:30:00.217158 4988 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/33fb9c50-b318-4279-af5c-605bb2f0d3c1-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 08 18:30:00 crc kubenswrapper[4988]: I1008 18:30:00.217198 4988 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/33fb9c50-b318-4279-af5c-605bb2f0d3c1-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Oct 08 18:30:00 crc kubenswrapper[4988]: I1008 18:30:00.320399 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9wrtn\" (UniqueName: \"kubernetes.io/projected/b039de59-9a58-4b20-b279-1447415e8843-kube-api-access-9wrtn\") pod \"collect-profiles-29332470-lbpdq\" (UID: \"b039de59-9a58-4b20-b279-1447415e8843\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29332470-lbpdq" Oct 08 18:30:00 crc kubenswrapper[4988]: I1008 18:30:00.320799 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/b039de59-9a58-4b20-b279-1447415e8843-secret-volume\") pod \"collect-profiles-29332470-lbpdq\" (UID: \"b039de59-9a58-4b20-b279-1447415e8843\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29332470-lbpdq" Oct 08 18:30:00 crc kubenswrapper[4988]: I1008 18:30:00.320823 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b039de59-9a58-4b20-b279-1447415e8843-config-volume\") pod \"collect-profiles-29332470-lbpdq\" (UID: \"b039de59-9a58-4b20-b279-1447415e8843\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29332470-lbpdq" Oct 08 18:30:00 crc kubenswrapper[4988]: I1008 18:30:00.424121 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/b039de59-9a58-4b20-b279-1447415e8843-secret-volume\") pod \"collect-profiles-29332470-lbpdq\" (UID: \"b039de59-9a58-4b20-b279-1447415e8843\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29332470-lbpdq" Oct 08 18:30:00 crc kubenswrapper[4988]: I1008 18:30:00.424161 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b039de59-9a58-4b20-b279-1447415e8843-config-volume\") pod \"collect-profiles-29332470-lbpdq\" (UID: \"b039de59-9a58-4b20-b279-1447415e8843\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29332470-lbpdq" Oct 08 18:30:00 crc kubenswrapper[4988]: I1008 18:30:00.424215 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9wrtn\" (UniqueName: \"kubernetes.io/projected/b039de59-9a58-4b20-b279-1447415e8843-kube-api-access-9wrtn\") pod \"collect-profiles-29332470-lbpdq\" (UID: \"b039de59-9a58-4b20-b279-1447415e8843\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29332470-lbpdq" Oct 08 18:30:00 crc kubenswrapper[4988]: I1008 18:30:00.426323 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b039de59-9a58-4b20-b279-1447415e8843-config-volume\") pod \"collect-profiles-29332470-lbpdq\" (UID: \"b039de59-9a58-4b20-b279-1447415e8843\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29332470-lbpdq" Oct 08 18:30:00 crc kubenswrapper[4988]: I1008 18:30:00.429722 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/b039de59-9a58-4b20-b279-1447415e8843-secret-volume\") pod \"collect-profiles-29332470-lbpdq\" (UID: \"b039de59-9a58-4b20-b279-1447415e8843\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29332470-lbpdq" Oct 08 18:30:00 crc kubenswrapper[4988]: I1008 18:30:00.442451 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9wrtn\" (UniqueName: \"kubernetes.io/projected/b039de59-9a58-4b20-b279-1447415e8843-kube-api-access-9wrtn\") pod \"collect-profiles-29332470-lbpdq\" (UID: \"b039de59-9a58-4b20-b279-1447415e8843\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29332470-lbpdq" Oct 08 18:30:00 crc kubenswrapper[4988]: I1008 18:30:00.485938 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Oct 08 18:30:00 crc kubenswrapper[4988]: W1008 18:30:00.500073 4988 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc97aab35_bbee_41a6_88b2_801b0b097a6d.slice/crio-3cbde5fa510c4fb3ae5b05af41867d2f1e1d39e486c406523d480312f4db2ff7 WatchSource:0}: Error finding container 3cbde5fa510c4fb3ae5b05af41867d2f1e1d39e486c406523d480312f4db2ff7: Status 404 returned error can't find the container with id 3cbde5fa510c4fb3ae5b05af41867d2f1e1d39e486c406523d480312f4db2ff7 Oct 08 18:30:00 crc kubenswrapper[4988]: I1008 18:30:00.505352 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29332470-lbpdq" Oct 08 18:30:00 crc kubenswrapper[4988]: I1008 18:30:00.669613 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-77dd5cf987-5qhqw" event={"ID":"33fb9c50-b318-4279-af5c-605bb2f0d3c1","Type":"ContainerDied","Data":"2f487ca795d8c82b64f5a7ae4058c27ee5391b6dcd345e70c10bf248546580d9"} Oct 08 18:30:00 crc kubenswrapper[4988]: I1008 18:30:00.669948 4988 scope.go:117] "RemoveContainer" containerID="8ac057a4f4bc0915ef6b7c8284d0fb581b4eadb20d3d7d433254afafbe52a8a2" Oct 08 18:30:00 crc kubenswrapper[4988]: I1008 18:30:00.669872 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-77dd5cf987-5qhqw" Oct 08 18:30:00 crc kubenswrapper[4988]: I1008 18:30:00.679026 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"c97aab35-bbee-41a6-88b2-801b0b097a6d","Type":"ContainerStarted","Data":"3cbde5fa510c4fb3ae5b05af41867d2f1e1d39e486c406523d480312f4db2ff7"} Oct 08 18:30:00 crc kubenswrapper[4988]: I1008 18:30:00.679113 4988 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 08 18:30:00 crc kubenswrapper[4988]: I1008 18:30:00.679125 4988 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 08 18:30:00 crc kubenswrapper[4988]: I1008 18:30:00.741695 4988 scope.go:117] "RemoveContainer" containerID="32e6bd21bb3120a82182cffbe11ccd289e40b61aa9ea29f4916e76d61813e4d4" Oct 08 18:30:00 crc kubenswrapper[4988]: I1008 18:30:00.745285 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-77dd5cf987-5qhqw"] Oct 08 18:30:00 crc kubenswrapper[4988]: I1008 18:30:00.759483 4988 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-77dd5cf987-5qhqw"] Oct 08 18:30:01 crc kubenswrapper[4988]: I1008 18:30:01.044689 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29332470-lbpdq"] Oct 08 18:30:01 crc kubenswrapper[4988]: I1008 18:30:01.088817 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-wvjsl" Oct 08 18:30:01 crc kubenswrapper[4988]: I1008 18:30:01.142925 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/de747e73-c280-424f-af48-3135bec91b55-combined-ca-bundle\") pod \"de747e73-c280-424f-af48-3135bec91b55\" (UID: \"de747e73-c280-424f-af48-3135bec91b55\") " Oct 08 18:30:01 crc kubenswrapper[4988]: I1008 18:30:01.142984 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7vbb4\" (UniqueName: \"kubernetes.io/projected/de747e73-c280-424f-af48-3135bec91b55-kube-api-access-7vbb4\") pod \"de747e73-c280-424f-af48-3135bec91b55\" (UID: \"de747e73-c280-424f-af48-3135bec91b55\") " Oct 08 18:30:01 crc kubenswrapper[4988]: I1008 18:30:01.143033 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/de747e73-c280-424f-af48-3135bec91b55-config\") pod \"de747e73-c280-424f-af48-3135bec91b55\" (UID: \"de747e73-c280-424f-af48-3135bec91b55\") " Oct 08 18:30:01 crc kubenswrapper[4988]: I1008 18:30:01.150799 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/de747e73-c280-424f-af48-3135bec91b55-kube-api-access-7vbb4" (OuterVolumeSpecName: "kube-api-access-7vbb4") pod "de747e73-c280-424f-af48-3135bec91b55" (UID: "de747e73-c280-424f-af48-3135bec91b55"). InnerVolumeSpecName "kube-api-access-7vbb4". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:30:01 crc kubenswrapper[4988]: I1008 18:30:01.187522 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/de747e73-c280-424f-af48-3135bec91b55-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "de747e73-c280-424f-af48-3135bec91b55" (UID: "de747e73-c280-424f-af48-3135bec91b55"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:30:01 crc kubenswrapper[4988]: I1008 18:30:01.195526 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/de747e73-c280-424f-af48-3135bec91b55-config" (OuterVolumeSpecName: "config") pod "de747e73-c280-424f-af48-3135bec91b55" (UID: "de747e73-c280-424f-af48-3135bec91b55"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:30:01 crc kubenswrapper[4988]: I1008 18:30:01.198654 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Oct 08 18:30:01 crc kubenswrapper[4988]: I1008 18:30:01.245846 4988 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/de747e73-c280-424f-af48-3135bec91b55-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 18:30:01 crc kubenswrapper[4988]: I1008 18:30:01.245875 4988 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7vbb4\" (UniqueName: \"kubernetes.io/projected/de747e73-c280-424f-af48-3135bec91b55-kube-api-access-7vbb4\") on node \"crc\" DevicePath \"\"" Oct 08 18:30:01 crc kubenswrapper[4988]: I1008 18:30:01.245888 4988 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/de747e73-c280-424f-af48-3135bec91b55-config\") on node \"crc\" DevicePath \"\"" Oct 08 18:30:01 crc kubenswrapper[4988]: I1008 18:30:01.278353 4988 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="33fb9c50-b318-4279-af5c-605bb2f0d3c1" path="/var/lib/kubelet/pods/33fb9c50-b318-4279-af5c-605bb2f0d3c1/volumes" Oct 08 18:30:01 crc kubenswrapper[4988]: I1008 18:30:01.382983 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Oct 08 18:30:01 crc kubenswrapper[4988]: I1008 18:30:01.500946 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-76f7bb8cb8-x44rw" Oct 08 18:30:01 crc kubenswrapper[4988]: I1008 18:30:01.711774 4988 generic.go:334] "Generic (PLEG): container finished" podID="b039de59-9a58-4b20-b279-1447415e8843" containerID="75398b89f53af48fd5e5ebbacf47ccffc023704db78076e20861c7bf6ee6b5ee" exitCode=0 Oct 08 18:30:01 crc kubenswrapper[4988]: I1008 18:30:01.711870 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29332470-lbpdq" event={"ID":"b039de59-9a58-4b20-b279-1447415e8843","Type":"ContainerDied","Data":"75398b89f53af48fd5e5ebbacf47ccffc023704db78076e20861c7bf6ee6b5ee"} Oct 08 18:30:01 crc kubenswrapper[4988]: I1008 18:30:01.711904 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29332470-lbpdq" event={"ID":"b039de59-9a58-4b20-b279-1447415e8843","Type":"ContainerStarted","Data":"94f2395041d84f3db3c5c7f88bd3a1bd1804017f20b0e8ab8ceca842f248e266"} Oct 08 18:30:01 crc kubenswrapper[4988]: I1008 18:30:01.738784 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-wvjsl" Oct 08 18:30:01 crc kubenswrapper[4988]: I1008 18:30:01.739301 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-wvjsl" event={"ID":"de747e73-c280-424f-af48-3135bec91b55","Type":"ContainerDied","Data":"d755e3c1d79ef62e9e57c675a8dc7c0d1d929f375f427db8ba59d7801f19ed4b"} Oct 08 18:30:01 crc kubenswrapper[4988]: I1008 18:30:01.739327 4988 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d755e3c1d79ef62e9e57c675a8dc7c0d1d929f375f427db8ba59d7801f19ed4b" Oct 08 18:30:01 crc kubenswrapper[4988]: I1008 18:30:01.929589 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-cddb74997-npcnc"] Oct 08 18:30:01 crc kubenswrapper[4988]: E1008 18:30:01.930049 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="de747e73-c280-424f-af48-3135bec91b55" containerName="neutron-db-sync" Oct 08 18:30:01 crc kubenswrapper[4988]: I1008 18:30:01.930082 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="de747e73-c280-424f-af48-3135bec91b55" containerName="neutron-db-sync" Oct 08 18:30:01 crc kubenswrapper[4988]: I1008 18:30:01.930281 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="de747e73-c280-424f-af48-3135bec91b55" containerName="neutron-db-sync" Oct 08 18:30:01 crc kubenswrapper[4988]: I1008 18:30:01.938622 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-cddb74997-npcnc" Oct 08 18:30:01 crc kubenswrapper[4988]: I1008 18:30:01.956105 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-cddb74997-npcnc"] Oct 08 18:30:01 crc kubenswrapper[4988]: I1008 18:30:01.982737 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-8554744ccd-jsq45"] Oct 08 18:30:01 crc kubenswrapper[4988]: I1008 18:30:01.995659 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-8554744ccd-jsq45" Oct 08 18:30:02 crc kubenswrapper[4988]: I1008 18:30:02.000565 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-n2kmm" Oct 08 18:30:02 crc kubenswrapper[4988]: I1008 18:30:02.001810 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Oct 08 18:30:02 crc kubenswrapper[4988]: I1008 18:30:02.002241 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Oct 08 18:30:02 crc kubenswrapper[4988]: I1008 18:30:02.002400 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-ovndbs" Oct 08 18:30:02 crc kubenswrapper[4988]: I1008 18:30:02.070798 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6cbd4338-3c8f-481a-9523-fbe1d5964292-dns-svc\") pod \"dnsmasq-dns-cddb74997-npcnc\" (UID: \"6cbd4338-3c8f-481a-9523-fbe1d5964292\") " pod="openstack/dnsmasq-dns-cddb74997-npcnc" Oct 08 18:30:02 crc kubenswrapper[4988]: I1008 18:30:02.070862 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/60b75acf-c263-4e5f-af6e-4195fc6be9e4-config\") pod \"neutron-8554744ccd-jsq45\" (UID: \"60b75acf-c263-4e5f-af6e-4195fc6be9e4\") " pod="openstack/neutron-8554744ccd-jsq45" Oct 08 18:30:02 crc kubenswrapper[4988]: I1008 18:30:02.070884 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/60b75acf-c263-4e5f-af6e-4195fc6be9e4-httpd-config\") pod \"neutron-8554744ccd-jsq45\" (UID: \"60b75acf-c263-4e5f-af6e-4195fc6be9e4\") " pod="openstack/neutron-8554744ccd-jsq45" Oct 08 18:30:02 crc kubenswrapper[4988]: I1008 18:30:02.070905 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/60b75acf-c263-4e5f-af6e-4195fc6be9e4-ovndb-tls-certs\") pod \"neutron-8554744ccd-jsq45\" (UID: \"60b75acf-c263-4e5f-af6e-4195fc6be9e4\") " pod="openstack/neutron-8554744ccd-jsq45" Oct 08 18:30:02 crc kubenswrapper[4988]: I1008 18:30:02.085085 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mf9vj\" (UniqueName: \"kubernetes.io/projected/6cbd4338-3c8f-481a-9523-fbe1d5964292-kube-api-access-mf9vj\") pod \"dnsmasq-dns-cddb74997-npcnc\" (UID: \"6cbd4338-3c8f-481a-9523-fbe1d5964292\") " pod="openstack/dnsmasq-dns-cddb74997-npcnc" Oct 08 18:30:02 crc kubenswrapper[4988]: I1008 18:30:02.085127 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6cbd4338-3c8f-481a-9523-fbe1d5964292-config\") pod \"dnsmasq-dns-cddb74997-npcnc\" (UID: \"6cbd4338-3c8f-481a-9523-fbe1d5964292\") " pod="openstack/dnsmasq-dns-cddb74997-npcnc" Oct 08 18:30:02 crc kubenswrapper[4988]: I1008 18:30:02.085205 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6cbd4338-3c8f-481a-9523-fbe1d5964292-ovsdbserver-sb\") pod \"dnsmasq-dns-cddb74997-npcnc\" (UID: \"6cbd4338-3c8f-481a-9523-fbe1d5964292\") " pod="openstack/dnsmasq-dns-cddb74997-npcnc" Oct 08 18:30:02 crc kubenswrapper[4988]: I1008 18:30:02.085241 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-phv2m\" (UniqueName: \"kubernetes.io/projected/60b75acf-c263-4e5f-af6e-4195fc6be9e4-kube-api-access-phv2m\") pod \"neutron-8554744ccd-jsq45\" (UID: \"60b75acf-c263-4e5f-af6e-4195fc6be9e4\") " pod="openstack/neutron-8554744ccd-jsq45" Oct 08 18:30:02 crc kubenswrapper[4988]: I1008 18:30:02.085273 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/6cbd4338-3c8f-481a-9523-fbe1d5964292-dns-swift-storage-0\") pod \"dnsmasq-dns-cddb74997-npcnc\" (UID: \"6cbd4338-3c8f-481a-9523-fbe1d5964292\") " pod="openstack/dnsmasq-dns-cddb74997-npcnc" Oct 08 18:30:02 crc kubenswrapper[4988]: I1008 18:30:02.085301 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6cbd4338-3c8f-481a-9523-fbe1d5964292-ovsdbserver-nb\") pod \"dnsmasq-dns-cddb74997-npcnc\" (UID: \"6cbd4338-3c8f-481a-9523-fbe1d5964292\") " pod="openstack/dnsmasq-dns-cddb74997-npcnc" Oct 08 18:30:02 crc kubenswrapper[4988]: I1008 18:30:02.085480 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/60b75acf-c263-4e5f-af6e-4195fc6be9e4-combined-ca-bundle\") pod \"neutron-8554744ccd-jsq45\" (UID: \"60b75acf-c263-4e5f-af6e-4195fc6be9e4\") " pod="openstack/neutron-8554744ccd-jsq45" Oct 08 18:30:02 crc kubenswrapper[4988]: I1008 18:30:02.134099 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-8554744ccd-jsq45"] Oct 08 18:30:02 crc kubenswrapper[4988]: I1008 18:30:02.189296 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/60b75acf-c263-4e5f-af6e-4195fc6be9e4-httpd-config\") pod \"neutron-8554744ccd-jsq45\" (UID: \"60b75acf-c263-4e5f-af6e-4195fc6be9e4\") " pod="openstack/neutron-8554744ccd-jsq45" Oct 08 18:30:02 crc kubenswrapper[4988]: I1008 18:30:02.189344 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/60b75acf-c263-4e5f-af6e-4195fc6be9e4-ovndb-tls-certs\") pod \"neutron-8554744ccd-jsq45\" (UID: \"60b75acf-c263-4e5f-af6e-4195fc6be9e4\") " pod="openstack/neutron-8554744ccd-jsq45" Oct 08 18:30:02 crc kubenswrapper[4988]: I1008 18:30:02.189427 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mf9vj\" (UniqueName: \"kubernetes.io/projected/6cbd4338-3c8f-481a-9523-fbe1d5964292-kube-api-access-mf9vj\") pod \"dnsmasq-dns-cddb74997-npcnc\" (UID: \"6cbd4338-3c8f-481a-9523-fbe1d5964292\") " pod="openstack/dnsmasq-dns-cddb74997-npcnc" Oct 08 18:30:02 crc kubenswrapper[4988]: I1008 18:30:02.189448 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6cbd4338-3c8f-481a-9523-fbe1d5964292-config\") pod \"dnsmasq-dns-cddb74997-npcnc\" (UID: \"6cbd4338-3c8f-481a-9523-fbe1d5964292\") " pod="openstack/dnsmasq-dns-cddb74997-npcnc" Oct 08 18:30:02 crc kubenswrapper[4988]: I1008 18:30:02.189476 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6cbd4338-3c8f-481a-9523-fbe1d5964292-ovsdbserver-sb\") pod \"dnsmasq-dns-cddb74997-npcnc\" (UID: \"6cbd4338-3c8f-481a-9523-fbe1d5964292\") " pod="openstack/dnsmasq-dns-cddb74997-npcnc" Oct 08 18:30:02 crc kubenswrapper[4988]: I1008 18:30:02.189495 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-phv2m\" (UniqueName: \"kubernetes.io/projected/60b75acf-c263-4e5f-af6e-4195fc6be9e4-kube-api-access-phv2m\") pod \"neutron-8554744ccd-jsq45\" (UID: \"60b75acf-c263-4e5f-af6e-4195fc6be9e4\") " pod="openstack/neutron-8554744ccd-jsq45" Oct 08 18:30:02 crc kubenswrapper[4988]: I1008 18:30:02.189519 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/6cbd4338-3c8f-481a-9523-fbe1d5964292-dns-swift-storage-0\") pod \"dnsmasq-dns-cddb74997-npcnc\" (UID: \"6cbd4338-3c8f-481a-9523-fbe1d5964292\") " pod="openstack/dnsmasq-dns-cddb74997-npcnc" Oct 08 18:30:02 crc kubenswrapper[4988]: I1008 18:30:02.189541 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6cbd4338-3c8f-481a-9523-fbe1d5964292-ovsdbserver-nb\") pod \"dnsmasq-dns-cddb74997-npcnc\" (UID: \"6cbd4338-3c8f-481a-9523-fbe1d5964292\") " pod="openstack/dnsmasq-dns-cddb74997-npcnc" Oct 08 18:30:02 crc kubenswrapper[4988]: I1008 18:30:02.189593 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/60b75acf-c263-4e5f-af6e-4195fc6be9e4-combined-ca-bundle\") pod \"neutron-8554744ccd-jsq45\" (UID: \"60b75acf-c263-4e5f-af6e-4195fc6be9e4\") " pod="openstack/neutron-8554744ccd-jsq45" Oct 08 18:30:02 crc kubenswrapper[4988]: I1008 18:30:02.189626 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6cbd4338-3c8f-481a-9523-fbe1d5964292-dns-svc\") pod \"dnsmasq-dns-cddb74997-npcnc\" (UID: \"6cbd4338-3c8f-481a-9523-fbe1d5964292\") " pod="openstack/dnsmasq-dns-cddb74997-npcnc" Oct 08 18:30:02 crc kubenswrapper[4988]: I1008 18:30:02.189646 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/60b75acf-c263-4e5f-af6e-4195fc6be9e4-config\") pod \"neutron-8554744ccd-jsq45\" (UID: \"60b75acf-c263-4e5f-af6e-4195fc6be9e4\") " pod="openstack/neutron-8554744ccd-jsq45" Oct 08 18:30:02 crc kubenswrapper[4988]: I1008 18:30:02.200420 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/60b75acf-c263-4e5f-af6e-4195fc6be9e4-config\") pod \"neutron-8554744ccd-jsq45\" (UID: \"60b75acf-c263-4e5f-af6e-4195fc6be9e4\") " pod="openstack/neutron-8554744ccd-jsq45" Oct 08 18:30:02 crc kubenswrapper[4988]: I1008 18:30:02.210909 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/60b75acf-c263-4e5f-af6e-4195fc6be9e4-combined-ca-bundle\") pod \"neutron-8554744ccd-jsq45\" (UID: \"60b75acf-c263-4e5f-af6e-4195fc6be9e4\") " pod="openstack/neutron-8554744ccd-jsq45" Oct 08 18:30:02 crc kubenswrapper[4988]: I1008 18:30:02.211256 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6cbd4338-3c8f-481a-9523-fbe1d5964292-ovsdbserver-sb\") pod \"dnsmasq-dns-cddb74997-npcnc\" (UID: \"6cbd4338-3c8f-481a-9523-fbe1d5964292\") " pod="openstack/dnsmasq-dns-cddb74997-npcnc" Oct 08 18:30:02 crc kubenswrapper[4988]: I1008 18:30:02.211930 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6cbd4338-3c8f-481a-9523-fbe1d5964292-ovsdbserver-nb\") pod \"dnsmasq-dns-cddb74997-npcnc\" (UID: \"6cbd4338-3c8f-481a-9523-fbe1d5964292\") " pod="openstack/dnsmasq-dns-cddb74997-npcnc" Oct 08 18:30:02 crc kubenswrapper[4988]: I1008 18:30:02.212051 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6cbd4338-3c8f-481a-9523-fbe1d5964292-config\") pod \"dnsmasq-dns-cddb74997-npcnc\" (UID: \"6cbd4338-3c8f-481a-9523-fbe1d5964292\") " pod="openstack/dnsmasq-dns-cddb74997-npcnc" Oct 08 18:30:02 crc kubenswrapper[4988]: I1008 18:30:02.215906 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6cbd4338-3c8f-481a-9523-fbe1d5964292-dns-svc\") pod \"dnsmasq-dns-cddb74997-npcnc\" (UID: \"6cbd4338-3c8f-481a-9523-fbe1d5964292\") " pod="openstack/dnsmasq-dns-cddb74997-npcnc" Oct 08 18:30:02 crc kubenswrapper[4988]: I1008 18:30:02.224167 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/60b75acf-c263-4e5f-af6e-4195fc6be9e4-ovndb-tls-certs\") pod \"neutron-8554744ccd-jsq45\" (UID: \"60b75acf-c263-4e5f-af6e-4195fc6be9e4\") " pod="openstack/neutron-8554744ccd-jsq45" Oct 08 18:30:02 crc kubenswrapper[4988]: I1008 18:30:02.224685 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/6cbd4338-3c8f-481a-9523-fbe1d5964292-dns-swift-storage-0\") pod \"dnsmasq-dns-cddb74997-npcnc\" (UID: \"6cbd4338-3c8f-481a-9523-fbe1d5964292\") " pod="openstack/dnsmasq-dns-cddb74997-npcnc" Oct 08 18:30:02 crc kubenswrapper[4988]: I1008 18:30:02.231075 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/60b75acf-c263-4e5f-af6e-4195fc6be9e4-httpd-config\") pod \"neutron-8554744ccd-jsq45\" (UID: \"60b75acf-c263-4e5f-af6e-4195fc6be9e4\") " pod="openstack/neutron-8554744ccd-jsq45" Oct 08 18:30:02 crc kubenswrapper[4988]: I1008 18:30:02.235970 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mf9vj\" (UniqueName: \"kubernetes.io/projected/6cbd4338-3c8f-481a-9523-fbe1d5964292-kube-api-access-mf9vj\") pod \"dnsmasq-dns-cddb74997-npcnc\" (UID: \"6cbd4338-3c8f-481a-9523-fbe1d5964292\") " pod="openstack/dnsmasq-dns-cddb74997-npcnc" Oct 08 18:30:02 crc kubenswrapper[4988]: I1008 18:30:02.237972 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-phv2m\" (UniqueName: \"kubernetes.io/projected/60b75acf-c263-4e5f-af6e-4195fc6be9e4-kube-api-access-phv2m\") pod \"neutron-8554744ccd-jsq45\" (UID: \"60b75acf-c263-4e5f-af6e-4195fc6be9e4\") " pod="openstack/neutron-8554744ccd-jsq45" Oct 08 18:30:02 crc kubenswrapper[4988]: I1008 18:30:02.270372 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-cddb74997-npcnc" Oct 08 18:30:02 crc kubenswrapper[4988]: I1008 18:30:02.355951 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-76f7bb8cb8-x44rw" Oct 08 18:30:02 crc kubenswrapper[4988]: I1008 18:30:02.378741 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-8554744ccd-jsq45" Oct 08 18:30:02 crc kubenswrapper[4988]: I1008 18:30:02.781613 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-gm2vf" event={"ID":"5b6f72e9-8dc6-4290-b032-d8a570b91767","Type":"ContainerStarted","Data":"ffe051d7124cf2f0d968449026dd9267afba1ef3e9305140464f3494f98986b8"} Oct 08 18:30:03 crc kubenswrapper[4988]: I1008 18:30:03.098553 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-db-sync-gm2vf" podStartSLOduration=3.884002625 podStartE2EDuration="38.098530796s" podCreationTimestamp="2025-10-08 18:29:25 +0000 UTC" firstStartedPulling="2025-10-08 18:29:26.648870775 +0000 UTC m=+1112.098713545" lastFinishedPulling="2025-10-08 18:30:00.863398946 +0000 UTC m=+1146.313241716" observedRunningTime="2025-10-08 18:30:02.811942306 +0000 UTC m=+1148.261785076" watchObservedRunningTime="2025-10-08 18:30:03.098530796 +0000 UTC m=+1148.548373567" Oct 08 18:30:03 crc kubenswrapper[4988]: I1008 18:30:03.123610 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-cddb74997-npcnc"] Oct 08 18:30:03 crc kubenswrapper[4988]: I1008 18:30:03.322062 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29332470-lbpdq" Oct 08 18:30:03 crc kubenswrapper[4988]: I1008 18:30:03.418005 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b039de59-9a58-4b20-b279-1447415e8843-config-volume\") pod \"b039de59-9a58-4b20-b279-1447415e8843\" (UID: \"b039de59-9a58-4b20-b279-1447415e8843\") " Oct 08 18:30:03 crc kubenswrapper[4988]: I1008 18:30:03.418284 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/b039de59-9a58-4b20-b279-1447415e8843-secret-volume\") pod \"b039de59-9a58-4b20-b279-1447415e8843\" (UID: \"b039de59-9a58-4b20-b279-1447415e8843\") " Oct 08 18:30:03 crc kubenswrapper[4988]: I1008 18:30:03.419084 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9wrtn\" (UniqueName: \"kubernetes.io/projected/b039de59-9a58-4b20-b279-1447415e8843-kube-api-access-9wrtn\") pod \"b039de59-9a58-4b20-b279-1447415e8843\" (UID: \"b039de59-9a58-4b20-b279-1447415e8843\") " Oct 08 18:30:03 crc kubenswrapper[4988]: I1008 18:30:03.420213 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b039de59-9a58-4b20-b279-1447415e8843-config-volume" (OuterVolumeSpecName: "config-volume") pod "b039de59-9a58-4b20-b279-1447415e8843" (UID: "b039de59-9a58-4b20-b279-1447415e8843"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 18:30:03 crc kubenswrapper[4988]: I1008 18:30:03.431617 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b039de59-9a58-4b20-b279-1447415e8843-kube-api-access-9wrtn" (OuterVolumeSpecName: "kube-api-access-9wrtn") pod "b039de59-9a58-4b20-b279-1447415e8843" (UID: "b039de59-9a58-4b20-b279-1447415e8843"). InnerVolumeSpecName "kube-api-access-9wrtn". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:30:03 crc kubenswrapper[4988]: I1008 18:30:03.441581 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b039de59-9a58-4b20-b279-1447415e8843-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "b039de59-9a58-4b20-b279-1447415e8843" (UID: "b039de59-9a58-4b20-b279-1447415e8843"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:30:03 crc kubenswrapper[4988]: I1008 18:30:03.441949 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-8554744ccd-jsq45"] Oct 08 18:30:03 crc kubenswrapper[4988]: I1008 18:30:03.521655 4988 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9wrtn\" (UniqueName: \"kubernetes.io/projected/b039de59-9a58-4b20-b279-1447415e8843-kube-api-access-9wrtn\") on node \"crc\" DevicePath \"\"" Oct 08 18:30:03 crc kubenswrapper[4988]: I1008 18:30:03.522000 4988 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b039de59-9a58-4b20-b279-1447415e8843-config-volume\") on node \"crc\" DevicePath \"\"" Oct 08 18:30:03 crc kubenswrapper[4988]: I1008 18:30:03.522014 4988 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/b039de59-9a58-4b20-b279-1447415e8843-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 08 18:30:03 crc kubenswrapper[4988]: I1008 18:30:03.798938 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29332470-lbpdq" event={"ID":"b039de59-9a58-4b20-b279-1447415e8843","Type":"ContainerDied","Data":"94f2395041d84f3db3c5c7f88bd3a1bd1804017f20b0e8ab8ceca842f248e266"} Oct 08 18:30:03 crc kubenswrapper[4988]: I1008 18:30:03.799260 4988 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="94f2395041d84f3db3c5c7f88bd3a1bd1804017f20b0e8ab8ceca842f248e266" Oct 08 18:30:03 crc kubenswrapper[4988]: I1008 18:30:03.799433 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29332470-lbpdq" Oct 08 18:30:03 crc kubenswrapper[4988]: I1008 18:30:03.820103 4988 generic.go:334] "Generic (PLEG): container finished" podID="6cbd4338-3c8f-481a-9523-fbe1d5964292" containerID="8b74a595124b875ce655fa386eb777f3fcb8a872fe2678fd2b94e4af489fb8e3" exitCode=0 Oct 08 18:30:03 crc kubenswrapper[4988]: I1008 18:30:03.820177 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-cddb74997-npcnc" event={"ID":"6cbd4338-3c8f-481a-9523-fbe1d5964292","Type":"ContainerDied","Data":"8b74a595124b875ce655fa386eb777f3fcb8a872fe2678fd2b94e4af489fb8e3"} Oct 08 18:30:03 crc kubenswrapper[4988]: I1008 18:30:03.820206 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-cddb74997-npcnc" event={"ID":"6cbd4338-3c8f-481a-9523-fbe1d5964292","Type":"ContainerStarted","Data":"cbee9fba14a3d1f6bb97b53358cbb2255b1f7ff83ccb8de7764c700a0d98afe9"} Oct 08 18:30:03 crc kubenswrapper[4988]: I1008 18:30:03.831677 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-8554744ccd-jsq45" event={"ID":"60b75acf-c263-4e5f-af6e-4195fc6be9e4","Type":"ContainerStarted","Data":"1af046392b464d870001852927bcb817b391578c44077d5241550bdd64b98e22"} Oct 08 18:30:03 crc kubenswrapper[4988]: I1008 18:30:03.831731 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-8554744ccd-jsq45" event={"ID":"60b75acf-c263-4e5f-af6e-4195fc6be9e4","Type":"ContainerStarted","Data":"c8b811412a9b67f92430740a4f1df53dfc66d18cdb37e09a505a6fbf0176739b"} Oct 08 18:30:04 crc kubenswrapper[4988]: I1008 18:30:04.798294 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-6c486d5744-j86zn" Oct 08 18:30:04 crc kubenswrapper[4988]: I1008 18:30:04.845428 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-cddb74997-npcnc" event={"ID":"6cbd4338-3c8f-481a-9523-fbe1d5964292","Type":"ContainerStarted","Data":"7b822b6e3018338b30995c41fa217c4e11a7bcbc7709c7746e9fcbe90f2a0a84"} Oct 08 18:30:04 crc kubenswrapper[4988]: I1008 18:30:04.845524 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-cddb74997-npcnc" Oct 08 18:30:04 crc kubenswrapper[4988]: I1008 18:30:04.851880 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-8554744ccd-jsq45" event={"ID":"60b75acf-c263-4e5f-af6e-4195fc6be9e4","Type":"ContainerStarted","Data":"ce7818666ce3c638983016985eb011d4d71fe1d55928608343a134abd73c80de"} Oct 08 18:30:04 crc kubenswrapper[4988]: I1008 18:30:04.852236 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-8554744ccd-jsq45" Oct 08 18:30:04 crc kubenswrapper[4988]: I1008 18:30:04.871747 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-cddb74997-npcnc" podStartSLOduration=3.87172809 podStartE2EDuration="3.87172809s" podCreationTimestamp="2025-10-08 18:30:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 18:30:04.866722422 +0000 UTC m=+1150.316565192" watchObservedRunningTime="2025-10-08 18:30:04.87172809 +0000 UTC m=+1150.321570860" Oct 08 18:30:05 crc kubenswrapper[4988]: I1008 18:30:05.039072 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-8554744ccd-jsq45" podStartSLOduration=4.039054133 podStartE2EDuration="4.039054133s" podCreationTimestamp="2025-10-08 18:30:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 18:30:04.893832856 +0000 UTC m=+1150.343675636" watchObservedRunningTime="2025-10-08 18:30:05.039054133 +0000 UTC m=+1150.488896893" Oct 08 18:30:05 crc kubenswrapper[4988]: I1008 18:30:05.045414 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-6bb998f7d5-h4g2x"] Oct 08 18:30:05 crc kubenswrapper[4988]: E1008 18:30:05.045774 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b039de59-9a58-4b20-b279-1447415e8843" containerName="collect-profiles" Oct 08 18:30:05 crc kubenswrapper[4988]: I1008 18:30:05.045793 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="b039de59-9a58-4b20-b279-1447415e8843" containerName="collect-profiles" Oct 08 18:30:05 crc kubenswrapper[4988]: I1008 18:30:05.046002 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="b039de59-9a58-4b20-b279-1447415e8843" containerName="collect-profiles" Oct 08 18:30:05 crc kubenswrapper[4988]: I1008 18:30:05.046868 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-6bb998f7d5-h4g2x" Oct 08 18:30:05 crc kubenswrapper[4988]: I1008 18:30:05.051435 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-public-svc" Oct 08 18:30:05 crc kubenswrapper[4988]: I1008 18:30:05.052015 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-internal-svc" Oct 08 18:30:05 crc kubenswrapper[4988]: I1008 18:30:05.058805 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-6bb998f7d5-h4g2x"] Oct 08 18:30:05 crc kubenswrapper[4988]: I1008 18:30:05.155537 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/6d2df307-23d3-48fd-9bb6-f09e500aaf9e-public-tls-certs\") pod \"neutron-6bb998f7d5-h4g2x\" (UID: \"6d2df307-23d3-48fd-9bb6-f09e500aaf9e\") " pod="openstack/neutron-6bb998f7d5-h4g2x" Oct 08 18:30:05 crc kubenswrapper[4988]: I1008 18:30:05.155585 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/6d2df307-23d3-48fd-9bb6-f09e500aaf9e-httpd-config\") pod \"neutron-6bb998f7d5-h4g2x\" (UID: \"6d2df307-23d3-48fd-9bb6-f09e500aaf9e\") " pod="openstack/neutron-6bb998f7d5-h4g2x" Oct 08 18:30:05 crc kubenswrapper[4988]: I1008 18:30:05.155609 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6d2df307-23d3-48fd-9bb6-f09e500aaf9e-combined-ca-bundle\") pod \"neutron-6bb998f7d5-h4g2x\" (UID: \"6d2df307-23d3-48fd-9bb6-f09e500aaf9e\") " pod="openstack/neutron-6bb998f7d5-h4g2x" Oct 08 18:30:05 crc kubenswrapper[4988]: I1008 18:30:05.155637 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/6d2df307-23d3-48fd-9bb6-f09e500aaf9e-ovndb-tls-certs\") pod \"neutron-6bb998f7d5-h4g2x\" (UID: \"6d2df307-23d3-48fd-9bb6-f09e500aaf9e\") " pod="openstack/neutron-6bb998f7d5-h4g2x" Oct 08 18:30:05 crc kubenswrapper[4988]: I1008 18:30:05.155666 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8gplz\" (UniqueName: \"kubernetes.io/projected/6d2df307-23d3-48fd-9bb6-f09e500aaf9e-kube-api-access-8gplz\") pod \"neutron-6bb998f7d5-h4g2x\" (UID: \"6d2df307-23d3-48fd-9bb6-f09e500aaf9e\") " pod="openstack/neutron-6bb998f7d5-h4g2x" Oct 08 18:30:05 crc kubenswrapper[4988]: I1008 18:30:05.155935 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/6d2df307-23d3-48fd-9bb6-f09e500aaf9e-internal-tls-certs\") pod \"neutron-6bb998f7d5-h4g2x\" (UID: \"6d2df307-23d3-48fd-9bb6-f09e500aaf9e\") " pod="openstack/neutron-6bb998f7d5-h4g2x" Oct 08 18:30:05 crc kubenswrapper[4988]: I1008 18:30:05.156043 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/6d2df307-23d3-48fd-9bb6-f09e500aaf9e-config\") pod \"neutron-6bb998f7d5-h4g2x\" (UID: \"6d2df307-23d3-48fd-9bb6-f09e500aaf9e\") " pod="openstack/neutron-6bb998f7d5-h4g2x" Oct 08 18:30:05 crc kubenswrapper[4988]: I1008 18:30:05.265817 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/6d2df307-23d3-48fd-9bb6-f09e500aaf9e-internal-tls-certs\") pod \"neutron-6bb998f7d5-h4g2x\" (UID: \"6d2df307-23d3-48fd-9bb6-f09e500aaf9e\") " pod="openstack/neutron-6bb998f7d5-h4g2x" Oct 08 18:30:05 crc kubenswrapper[4988]: I1008 18:30:05.265880 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/6d2df307-23d3-48fd-9bb6-f09e500aaf9e-config\") pod \"neutron-6bb998f7d5-h4g2x\" (UID: \"6d2df307-23d3-48fd-9bb6-f09e500aaf9e\") " pod="openstack/neutron-6bb998f7d5-h4g2x" Oct 08 18:30:05 crc kubenswrapper[4988]: I1008 18:30:05.265945 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/6d2df307-23d3-48fd-9bb6-f09e500aaf9e-public-tls-certs\") pod \"neutron-6bb998f7d5-h4g2x\" (UID: \"6d2df307-23d3-48fd-9bb6-f09e500aaf9e\") " pod="openstack/neutron-6bb998f7d5-h4g2x" Oct 08 18:30:05 crc kubenswrapper[4988]: I1008 18:30:05.265971 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/6d2df307-23d3-48fd-9bb6-f09e500aaf9e-httpd-config\") pod \"neutron-6bb998f7d5-h4g2x\" (UID: \"6d2df307-23d3-48fd-9bb6-f09e500aaf9e\") " pod="openstack/neutron-6bb998f7d5-h4g2x" Oct 08 18:30:05 crc kubenswrapper[4988]: I1008 18:30:05.265990 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6d2df307-23d3-48fd-9bb6-f09e500aaf9e-combined-ca-bundle\") pod \"neutron-6bb998f7d5-h4g2x\" (UID: \"6d2df307-23d3-48fd-9bb6-f09e500aaf9e\") " pod="openstack/neutron-6bb998f7d5-h4g2x" Oct 08 18:30:05 crc kubenswrapper[4988]: I1008 18:30:05.266014 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/6d2df307-23d3-48fd-9bb6-f09e500aaf9e-ovndb-tls-certs\") pod \"neutron-6bb998f7d5-h4g2x\" (UID: \"6d2df307-23d3-48fd-9bb6-f09e500aaf9e\") " pod="openstack/neutron-6bb998f7d5-h4g2x" Oct 08 18:30:05 crc kubenswrapper[4988]: I1008 18:30:05.266043 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8gplz\" (UniqueName: \"kubernetes.io/projected/6d2df307-23d3-48fd-9bb6-f09e500aaf9e-kube-api-access-8gplz\") pod \"neutron-6bb998f7d5-h4g2x\" (UID: \"6d2df307-23d3-48fd-9bb6-f09e500aaf9e\") " pod="openstack/neutron-6bb998f7d5-h4g2x" Oct 08 18:30:05 crc kubenswrapper[4988]: I1008 18:30:05.273941 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/6d2df307-23d3-48fd-9bb6-f09e500aaf9e-public-tls-certs\") pod \"neutron-6bb998f7d5-h4g2x\" (UID: \"6d2df307-23d3-48fd-9bb6-f09e500aaf9e\") " pod="openstack/neutron-6bb998f7d5-h4g2x" Oct 08 18:30:05 crc kubenswrapper[4988]: I1008 18:30:05.274206 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/6d2df307-23d3-48fd-9bb6-f09e500aaf9e-httpd-config\") pod \"neutron-6bb998f7d5-h4g2x\" (UID: \"6d2df307-23d3-48fd-9bb6-f09e500aaf9e\") " pod="openstack/neutron-6bb998f7d5-h4g2x" Oct 08 18:30:05 crc kubenswrapper[4988]: I1008 18:30:05.274229 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/6d2df307-23d3-48fd-9bb6-f09e500aaf9e-config\") pod \"neutron-6bb998f7d5-h4g2x\" (UID: \"6d2df307-23d3-48fd-9bb6-f09e500aaf9e\") " pod="openstack/neutron-6bb998f7d5-h4g2x" Oct 08 18:30:05 crc kubenswrapper[4988]: I1008 18:30:05.274318 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/6d2df307-23d3-48fd-9bb6-f09e500aaf9e-internal-tls-certs\") pod \"neutron-6bb998f7d5-h4g2x\" (UID: \"6d2df307-23d3-48fd-9bb6-f09e500aaf9e\") " pod="openstack/neutron-6bb998f7d5-h4g2x" Oct 08 18:30:05 crc kubenswrapper[4988]: I1008 18:30:05.274536 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/6d2df307-23d3-48fd-9bb6-f09e500aaf9e-ovndb-tls-certs\") pod \"neutron-6bb998f7d5-h4g2x\" (UID: \"6d2df307-23d3-48fd-9bb6-f09e500aaf9e\") " pod="openstack/neutron-6bb998f7d5-h4g2x" Oct 08 18:30:05 crc kubenswrapper[4988]: I1008 18:30:05.274920 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6d2df307-23d3-48fd-9bb6-f09e500aaf9e-combined-ca-bundle\") pod \"neutron-6bb998f7d5-h4g2x\" (UID: \"6d2df307-23d3-48fd-9bb6-f09e500aaf9e\") " pod="openstack/neutron-6bb998f7d5-h4g2x" Oct 08 18:30:05 crc kubenswrapper[4988]: I1008 18:30:05.288149 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8gplz\" (UniqueName: \"kubernetes.io/projected/6d2df307-23d3-48fd-9bb6-f09e500aaf9e-kube-api-access-8gplz\") pod \"neutron-6bb998f7d5-h4g2x\" (UID: \"6d2df307-23d3-48fd-9bb6-f09e500aaf9e\") " pod="openstack/neutron-6bb998f7d5-h4g2x" Oct 08 18:30:05 crc kubenswrapper[4988]: I1008 18:30:05.429957 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-6bb998f7d5-h4g2x" Oct 08 18:30:06 crc kubenswrapper[4988]: I1008 18:30:06.072294 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-6c486d5744-j86zn" Oct 08 18:30:06 crc kubenswrapper[4988]: I1008 18:30:06.150539 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-76f7bb8cb8-x44rw"] Oct 08 18:30:06 crc kubenswrapper[4988]: I1008 18:30:06.150763 4988 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-76f7bb8cb8-x44rw" podUID="fe49d0c8-90f0-4673-b527-ff112cfd3037" containerName="barbican-api-log" containerID="cri-o://0e25441dbd7ef8b78df421475b3d19064e1a51130eb299a3433a4940e3737266" gracePeriod=30 Oct 08 18:30:06 crc kubenswrapper[4988]: I1008 18:30:06.151141 4988 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-76f7bb8cb8-x44rw" podUID="fe49d0c8-90f0-4673-b527-ff112cfd3037" containerName="barbican-api" containerID="cri-o://dfe3a72d9e18e94651f136aa2b98ffa4d52065039c4bc92724da4663ef94fddf" gracePeriod=30 Oct 08 18:30:06 crc kubenswrapper[4988]: I1008 18:30:06.330369 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-6bb998f7d5-h4g2x"] Oct 08 18:30:06 crc kubenswrapper[4988]: W1008 18:30:06.350005 4988 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6d2df307_23d3_48fd_9bb6_f09e500aaf9e.slice/crio-8dd0d4be4a76c246ed2902066cd503da58a22dc94e2ffe053725437104391724 WatchSource:0}: Error finding container 8dd0d4be4a76c246ed2902066cd503da58a22dc94e2ffe053725437104391724: Status 404 returned error can't find the container with id 8dd0d4be4a76c246ed2902066cd503da58a22dc94e2ffe053725437104391724 Oct 08 18:30:06 crc kubenswrapper[4988]: I1008 18:30:06.894117 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-6bb998f7d5-h4g2x" event={"ID":"6d2df307-23d3-48fd-9bb6-f09e500aaf9e","Type":"ContainerStarted","Data":"ed0f1f9337e028007e93109bea68d88246a9ba24b639073b907414a6728b9835"} Oct 08 18:30:06 crc kubenswrapper[4988]: I1008 18:30:06.894481 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-6bb998f7d5-h4g2x" event={"ID":"6d2df307-23d3-48fd-9bb6-f09e500aaf9e","Type":"ContainerStarted","Data":"8dd0d4be4a76c246ed2902066cd503da58a22dc94e2ffe053725437104391724"} Oct 08 18:30:06 crc kubenswrapper[4988]: I1008 18:30:06.895514 4988 generic.go:334] "Generic (PLEG): container finished" podID="fe49d0c8-90f0-4673-b527-ff112cfd3037" containerID="0e25441dbd7ef8b78df421475b3d19064e1a51130eb299a3433a4940e3737266" exitCode=143 Oct 08 18:30:06 crc kubenswrapper[4988]: I1008 18:30:06.895558 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-76f7bb8cb8-x44rw" event={"ID":"fe49d0c8-90f0-4673-b527-ff112cfd3037","Type":"ContainerDied","Data":"0e25441dbd7ef8b78df421475b3d19064e1a51130eb299a3433a4940e3737266"} Oct 08 18:30:07 crc kubenswrapper[4988]: I1008 18:30:07.908142 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-6bb998f7d5-h4g2x" event={"ID":"6d2df307-23d3-48fd-9bb6-f09e500aaf9e","Type":"ContainerStarted","Data":"db863ec0ad9c9861590e77462403049f0e352d2f01da17f2db5e4682b149163e"} Oct 08 18:30:07 crc kubenswrapper[4988]: I1008 18:30:07.908755 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-6bb998f7d5-h4g2x" Oct 08 18:30:07 crc kubenswrapper[4988]: I1008 18:30:07.932265 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-6bb998f7d5-h4g2x" podStartSLOduration=2.932246048 podStartE2EDuration="2.932246048s" podCreationTimestamp="2025-10-08 18:30:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 18:30:07.929604205 +0000 UTC m=+1153.379446995" watchObservedRunningTime="2025-10-08 18:30:07.932246048 +0000 UTC m=+1153.382088818" Oct 08 18:30:08 crc kubenswrapper[4988]: I1008 18:30:08.446774 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-proxy-7b7cf6bf4f-s6vrs"] Oct 08 18:30:08 crc kubenswrapper[4988]: I1008 18:30:08.448912 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-7b7cf6bf4f-s6vrs" Oct 08 18:30:08 crc kubenswrapper[4988]: I1008 18:30:08.451943 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-swift-internal-svc" Oct 08 18:30:08 crc kubenswrapper[4988]: I1008 18:30:08.452243 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-proxy-config-data" Oct 08 18:30:08 crc kubenswrapper[4988]: I1008 18:30:08.452422 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-swift-public-svc" Oct 08 18:30:08 crc kubenswrapper[4988]: I1008 18:30:08.463351 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-proxy-7b7cf6bf4f-s6vrs"] Oct 08 18:30:08 crc kubenswrapper[4988]: I1008 18:30:08.539986 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/857686e3-0c6b-4d08-a8d6-f89e4c832139-run-httpd\") pod \"swift-proxy-7b7cf6bf4f-s6vrs\" (UID: \"857686e3-0c6b-4d08-a8d6-f89e4c832139\") " pod="openstack/swift-proxy-7b7cf6bf4f-s6vrs" Oct 08 18:30:08 crc kubenswrapper[4988]: I1008 18:30:08.540049 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k5jwg\" (UniqueName: \"kubernetes.io/projected/857686e3-0c6b-4d08-a8d6-f89e4c832139-kube-api-access-k5jwg\") pod \"swift-proxy-7b7cf6bf4f-s6vrs\" (UID: \"857686e3-0c6b-4d08-a8d6-f89e4c832139\") " pod="openstack/swift-proxy-7b7cf6bf4f-s6vrs" Oct 08 18:30:08 crc kubenswrapper[4988]: I1008 18:30:08.540281 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/857686e3-0c6b-4d08-a8d6-f89e4c832139-public-tls-certs\") pod \"swift-proxy-7b7cf6bf4f-s6vrs\" (UID: \"857686e3-0c6b-4d08-a8d6-f89e4c832139\") " pod="openstack/swift-proxy-7b7cf6bf4f-s6vrs" Oct 08 18:30:08 crc kubenswrapper[4988]: I1008 18:30:08.540358 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/857686e3-0c6b-4d08-a8d6-f89e4c832139-log-httpd\") pod \"swift-proxy-7b7cf6bf4f-s6vrs\" (UID: \"857686e3-0c6b-4d08-a8d6-f89e4c832139\") " pod="openstack/swift-proxy-7b7cf6bf4f-s6vrs" Oct 08 18:30:08 crc kubenswrapper[4988]: I1008 18:30:08.540452 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/857686e3-0c6b-4d08-a8d6-f89e4c832139-config-data\") pod \"swift-proxy-7b7cf6bf4f-s6vrs\" (UID: \"857686e3-0c6b-4d08-a8d6-f89e4c832139\") " pod="openstack/swift-proxy-7b7cf6bf4f-s6vrs" Oct 08 18:30:08 crc kubenswrapper[4988]: I1008 18:30:08.540510 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/857686e3-0c6b-4d08-a8d6-f89e4c832139-etc-swift\") pod \"swift-proxy-7b7cf6bf4f-s6vrs\" (UID: \"857686e3-0c6b-4d08-a8d6-f89e4c832139\") " pod="openstack/swift-proxy-7b7cf6bf4f-s6vrs" Oct 08 18:30:08 crc kubenswrapper[4988]: I1008 18:30:08.540629 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/857686e3-0c6b-4d08-a8d6-f89e4c832139-combined-ca-bundle\") pod \"swift-proxy-7b7cf6bf4f-s6vrs\" (UID: \"857686e3-0c6b-4d08-a8d6-f89e4c832139\") " pod="openstack/swift-proxy-7b7cf6bf4f-s6vrs" Oct 08 18:30:08 crc kubenswrapper[4988]: I1008 18:30:08.540736 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/857686e3-0c6b-4d08-a8d6-f89e4c832139-internal-tls-certs\") pod \"swift-proxy-7b7cf6bf4f-s6vrs\" (UID: \"857686e3-0c6b-4d08-a8d6-f89e4c832139\") " pod="openstack/swift-proxy-7b7cf6bf4f-s6vrs" Oct 08 18:30:08 crc kubenswrapper[4988]: I1008 18:30:08.642598 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/857686e3-0c6b-4d08-a8d6-f89e4c832139-public-tls-certs\") pod \"swift-proxy-7b7cf6bf4f-s6vrs\" (UID: \"857686e3-0c6b-4d08-a8d6-f89e4c832139\") " pod="openstack/swift-proxy-7b7cf6bf4f-s6vrs" Oct 08 18:30:08 crc kubenswrapper[4988]: I1008 18:30:08.642648 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/857686e3-0c6b-4d08-a8d6-f89e4c832139-log-httpd\") pod \"swift-proxy-7b7cf6bf4f-s6vrs\" (UID: \"857686e3-0c6b-4d08-a8d6-f89e4c832139\") " pod="openstack/swift-proxy-7b7cf6bf4f-s6vrs" Oct 08 18:30:08 crc kubenswrapper[4988]: I1008 18:30:08.642691 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/857686e3-0c6b-4d08-a8d6-f89e4c832139-config-data\") pod \"swift-proxy-7b7cf6bf4f-s6vrs\" (UID: \"857686e3-0c6b-4d08-a8d6-f89e4c832139\") " pod="openstack/swift-proxy-7b7cf6bf4f-s6vrs" Oct 08 18:30:08 crc kubenswrapper[4988]: I1008 18:30:08.642732 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/857686e3-0c6b-4d08-a8d6-f89e4c832139-etc-swift\") pod \"swift-proxy-7b7cf6bf4f-s6vrs\" (UID: \"857686e3-0c6b-4d08-a8d6-f89e4c832139\") " pod="openstack/swift-proxy-7b7cf6bf4f-s6vrs" Oct 08 18:30:08 crc kubenswrapper[4988]: I1008 18:30:08.642775 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/857686e3-0c6b-4d08-a8d6-f89e4c832139-combined-ca-bundle\") pod \"swift-proxy-7b7cf6bf4f-s6vrs\" (UID: \"857686e3-0c6b-4d08-a8d6-f89e4c832139\") " pod="openstack/swift-proxy-7b7cf6bf4f-s6vrs" Oct 08 18:30:08 crc kubenswrapper[4988]: I1008 18:30:08.642814 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/857686e3-0c6b-4d08-a8d6-f89e4c832139-internal-tls-certs\") pod \"swift-proxy-7b7cf6bf4f-s6vrs\" (UID: \"857686e3-0c6b-4d08-a8d6-f89e4c832139\") " pod="openstack/swift-proxy-7b7cf6bf4f-s6vrs" Oct 08 18:30:08 crc kubenswrapper[4988]: I1008 18:30:08.642853 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/857686e3-0c6b-4d08-a8d6-f89e4c832139-run-httpd\") pod \"swift-proxy-7b7cf6bf4f-s6vrs\" (UID: \"857686e3-0c6b-4d08-a8d6-f89e4c832139\") " pod="openstack/swift-proxy-7b7cf6bf4f-s6vrs" Oct 08 18:30:08 crc kubenswrapper[4988]: I1008 18:30:08.642877 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k5jwg\" (UniqueName: \"kubernetes.io/projected/857686e3-0c6b-4d08-a8d6-f89e4c832139-kube-api-access-k5jwg\") pod \"swift-proxy-7b7cf6bf4f-s6vrs\" (UID: \"857686e3-0c6b-4d08-a8d6-f89e4c832139\") " pod="openstack/swift-proxy-7b7cf6bf4f-s6vrs" Oct 08 18:30:08 crc kubenswrapper[4988]: I1008 18:30:08.643310 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/857686e3-0c6b-4d08-a8d6-f89e4c832139-run-httpd\") pod \"swift-proxy-7b7cf6bf4f-s6vrs\" (UID: \"857686e3-0c6b-4d08-a8d6-f89e4c832139\") " pod="openstack/swift-proxy-7b7cf6bf4f-s6vrs" Oct 08 18:30:08 crc kubenswrapper[4988]: I1008 18:30:08.643320 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/857686e3-0c6b-4d08-a8d6-f89e4c832139-log-httpd\") pod \"swift-proxy-7b7cf6bf4f-s6vrs\" (UID: \"857686e3-0c6b-4d08-a8d6-f89e4c832139\") " pod="openstack/swift-proxy-7b7cf6bf4f-s6vrs" Oct 08 18:30:08 crc kubenswrapper[4988]: I1008 18:30:08.652091 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/857686e3-0c6b-4d08-a8d6-f89e4c832139-internal-tls-certs\") pod \"swift-proxy-7b7cf6bf4f-s6vrs\" (UID: \"857686e3-0c6b-4d08-a8d6-f89e4c832139\") " pod="openstack/swift-proxy-7b7cf6bf4f-s6vrs" Oct 08 18:30:08 crc kubenswrapper[4988]: I1008 18:30:08.652477 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/857686e3-0c6b-4d08-a8d6-f89e4c832139-combined-ca-bundle\") pod \"swift-proxy-7b7cf6bf4f-s6vrs\" (UID: \"857686e3-0c6b-4d08-a8d6-f89e4c832139\") " pod="openstack/swift-proxy-7b7cf6bf4f-s6vrs" Oct 08 18:30:08 crc kubenswrapper[4988]: I1008 18:30:08.652645 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/857686e3-0c6b-4d08-a8d6-f89e4c832139-config-data\") pod \"swift-proxy-7b7cf6bf4f-s6vrs\" (UID: \"857686e3-0c6b-4d08-a8d6-f89e4c832139\") " pod="openstack/swift-proxy-7b7cf6bf4f-s6vrs" Oct 08 18:30:08 crc kubenswrapper[4988]: I1008 18:30:08.652840 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/857686e3-0c6b-4d08-a8d6-f89e4c832139-etc-swift\") pod \"swift-proxy-7b7cf6bf4f-s6vrs\" (UID: \"857686e3-0c6b-4d08-a8d6-f89e4c832139\") " pod="openstack/swift-proxy-7b7cf6bf4f-s6vrs" Oct 08 18:30:08 crc kubenswrapper[4988]: I1008 18:30:08.662955 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/857686e3-0c6b-4d08-a8d6-f89e4c832139-public-tls-certs\") pod \"swift-proxy-7b7cf6bf4f-s6vrs\" (UID: \"857686e3-0c6b-4d08-a8d6-f89e4c832139\") " pod="openstack/swift-proxy-7b7cf6bf4f-s6vrs" Oct 08 18:30:08 crc kubenswrapper[4988]: I1008 18:30:08.667274 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k5jwg\" (UniqueName: \"kubernetes.io/projected/857686e3-0c6b-4d08-a8d6-f89e4c832139-kube-api-access-k5jwg\") pod \"swift-proxy-7b7cf6bf4f-s6vrs\" (UID: \"857686e3-0c6b-4d08-a8d6-f89e4c832139\") " pod="openstack/swift-proxy-7b7cf6bf4f-s6vrs" Oct 08 18:30:08 crc kubenswrapper[4988]: I1008 18:30:08.765948 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-7b7cf6bf4f-s6vrs" Oct 08 18:30:09 crc kubenswrapper[4988]: I1008 18:30:09.120166 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 08 18:30:09 crc kubenswrapper[4988]: I1008 18:30:09.120533 4988 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="9e8e39d2-57c8-4ea6-9c83-dcb8d75274cd" containerName="proxy-httpd" containerID="cri-o://1e28870498bf7017bb7aaf972c99e33b912ab2db76680063b2dd790406e59691" gracePeriod=30 Oct 08 18:30:09 crc kubenswrapper[4988]: I1008 18:30:09.120601 4988 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="9e8e39d2-57c8-4ea6-9c83-dcb8d75274cd" containerName="sg-core" containerID="cri-o://3611c104fb9f40d7e4eb1645a5029783c73aa03a8d819c857e23f703653a58e3" gracePeriod=30 Oct 08 18:30:09 crc kubenswrapper[4988]: I1008 18:30:09.120476 4988 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="9e8e39d2-57c8-4ea6-9c83-dcb8d75274cd" containerName="ceilometer-central-agent" containerID="cri-o://9640cc556082067881ad4130d32e24914d42ee959a828c2f3e071e7ca9d49656" gracePeriod=30 Oct 08 18:30:09 crc kubenswrapper[4988]: I1008 18:30:09.120593 4988 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="9e8e39d2-57c8-4ea6-9c83-dcb8d75274cd" containerName="ceilometer-notification-agent" containerID="cri-o://46c9d841d83e09ba232a18c37dc58ed0a217dca805674f3203c8f5d7cbddb5cb" gracePeriod=30 Oct 08 18:30:09 crc kubenswrapper[4988]: I1008 18:30:09.134809 4988 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ceilometer-0" podUID="9e8e39d2-57c8-4ea6-9c83-dcb8d75274cd" containerName="proxy-httpd" probeResult="failure" output="Get \"http://10.217.0.157:3000/\": EOF" Oct 08 18:30:09 crc kubenswrapper[4988]: I1008 18:30:09.584065 4988 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-76f7bb8cb8-x44rw" podUID="fe49d0c8-90f0-4673-b527-ff112cfd3037" containerName="barbican-api-log" probeResult="failure" output="Get \"http://10.217.0.155:9311/healthcheck\": read tcp 10.217.0.2:60992->10.217.0.155:9311: read: connection reset by peer" Oct 08 18:30:09 crc kubenswrapper[4988]: I1008 18:30:09.584086 4988 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-76f7bb8cb8-x44rw" podUID="fe49d0c8-90f0-4673-b527-ff112cfd3037" containerName="barbican-api" probeResult="failure" output="Get \"http://10.217.0.155:9311/healthcheck\": read tcp 10.217.0.2:32774->10.217.0.155:9311: read: connection reset by peer" Oct 08 18:30:09 crc kubenswrapper[4988]: I1008 18:30:09.934943 4988 generic.go:334] "Generic (PLEG): container finished" podID="fe49d0c8-90f0-4673-b527-ff112cfd3037" containerID="dfe3a72d9e18e94651f136aa2b98ffa4d52065039c4bc92724da4663ef94fddf" exitCode=0 Oct 08 18:30:09 crc kubenswrapper[4988]: I1008 18:30:09.934976 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-76f7bb8cb8-x44rw" event={"ID":"fe49d0c8-90f0-4673-b527-ff112cfd3037","Type":"ContainerDied","Data":"dfe3a72d9e18e94651f136aa2b98ffa4d52065039c4bc92724da4663ef94fddf"} Oct 08 18:30:09 crc kubenswrapper[4988]: I1008 18:30:09.940583 4988 generic.go:334] "Generic (PLEG): container finished" podID="9e8e39d2-57c8-4ea6-9c83-dcb8d75274cd" containerID="1e28870498bf7017bb7aaf972c99e33b912ab2db76680063b2dd790406e59691" exitCode=0 Oct 08 18:30:09 crc kubenswrapper[4988]: I1008 18:30:09.940636 4988 generic.go:334] "Generic (PLEG): container finished" podID="9e8e39d2-57c8-4ea6-9c83-dcb8d75274cd" containerID="3611c104fb9f40d7e4eb1645a5029783c73aa03a8d819c857e23f703653a58e3" exitCode=2 Oct 08 18:30:09 crc kubenswrapper[4988]: I1008 18:30:09.940647 4988 generic.go:334] "Generic (PLEG): container finished" podID="9e8e39d2-57c8-4ea6-9c83-dcb8d75274cd" containerID="9640cc556082067881ad4130d32e24914d42ee959a828c2f3e071e7ca9d49656" exitCode=0 Oct 08 18:30:09 crc kubenswrapper[4988]: I1008 18:30:09.940639 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9e8e39d2-57c8-4ea6-9c83-dcb8d75274cd","Type":"ContainerDied","Data":"1e28870498bf7017bb7aaf972c99e33b912ab2db76680063b2dd790406e59691"} Oct 08 18:30:09 crc kubenswrapper[4988]: I1008 18:30:09.940687 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9e8e39d2-57c8-4ea6-9c83-dcb8d75274cd","Type":"ContainerDied","Data":"3611c104fb9f40d7e4eb1645a5029783c73aa03a8d819c857e23f703653a58e3"} Oct 08 18:30:09 crc kubenswrapper[4988]: I1008 18:30:09.940705 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9e8e39d2-57c8-4ea6-9c83-dcb8d75274cd","Type":"ContainerDied","Data":"9640cc556082067881ad4130d32e24914d42ee959a828c2f3e071e7ca9d49656"} Oct 08 18:30:10 crc kubenswrapper[4988]: I1008 18:30:10.952863 4988 generic.go:334] "Generic (PLEG): container finished" podID="5b6f72e9-8dc6-4290-b032-d8a570b91767" containerID="ffe051d7124cf2f0d968449026dd9267afba1ef3e9305140464f3494f98986b8" exitCode=0 Oct 08 18:30:10 crc kubenswrapper[4988]: I1008 18:30:10.952948 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-gm2vf" event={"ID":"5b6f72e9-8dc6-4290-b032-d8a570b91767","Type":"ContainerDied","Data":"ffe051d7124cf2f0d968449026dd9267afba1ef3e9305140464f3494f98986b8"} Oct 08 18:30:11 crc kubenswrapper[4988]: I1008 18:30:11.966327 4988 generic.go:334] "Generic (PLEG): container finished" podID="9e8e39d2-57c8-4ea6-9c83-dcb8d75274cd" containerID="46c9d841d83e09ba232a18c37dc58ed0a217dca805674f3203c8f5d7cbddb5cb" exitCode=0 Oct 08 18:30:11 crc kubenswrapper[4988]: I1008 18:30:11.966468 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9e8e39d2-57c8-4ea6-9c83-dcb8d75274cd","Type":"ContainerDied","Data":"46c9d841d83e09ba232a18c37dc58ed0a217dca805674f3203c8f5d7cbddb5cb"} Oct 08 18:30:12 crc kubenswrapper[4988]: I1008 18:30:12.272661 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-cddb74997-npcnc" Oct 08 18:30:12 crc kubenswrapper[4988]: I1008 18:30:12.338118 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-54df4b685c-fr6bz"] Oct 08 18:30:12 crc kubenswrapper[4988]: I1008 18:30:12.338492 4988 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-54df4b685c-fr6bz" podUID="d9f52cd3-500f-45b9-9f83-8e8244e511d6" containerName="dnsmasq-dns" containerID="cri-o://9a9ae238ebc775efdca8a60be60a294b733e168bb9d325901a7175d4ac2574bb" gracePeriod=10 Oct 08 18:30:12 crc kubenswrapper[4988]: I1008 18:30:12.983648 4988 generic.go:334] "Generic (PLEG): container finished" podID="d9f52cd3-500f-45b9-9f83-8e8244e511d6" containerID="9a9ae238ebc775efdca8a60be60a294b733e168bb9d325901a7175d4ac2574bb" exitCode=0 Oct 08 18:30:12 crc kubenswrapper[4988]: I1008 18:30:12.983688 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-54df4b685c-fr6bz" event={"ID":"d9f52cd3-500f-45b9-9f83-8e8244e511d6","Type":"ContainerDied","Data":"9a9ae238ebc775efdca8a60be60a294b733e168bb9d325901a7175d4ac2574bb"} Oct 08 18:30:14 crc kubenswrapper[4988]: I1008 18:30:14.119233 4988 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-54df4b685c-fr6bz" podUID="d9f52cd3-500f-45b9-9f83-8e8244e511d6" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.154:5353: connect: connection refused" Oct 08 18:30:14 crc kubenswrapper[4988]: I1008 18:30:14.304246 4988 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-76f7bb8cb8-x44rw" podUID="fe49d0c8-90f0-4673-b527-ff112cfd3037" containerName="barbican-api-log" probeResult="failure" output="Get \"http://10.217.0.155:9311/healthcheck\": dial tcp 10.217.0.155:9311: connect: connection refused" Oct 08 18:30:14 crc kubenswrapper[4988]: I1008 18:30:14.304575 4988 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-76f7bb8cb8-x44rw" podUID="fe49d0c8-90f0-4673-b527-ff112cfd3037" containerName="barbican-api" probeResult="failure" output="Get \"http://10.217.0.155:9311/healthcheck\": dial tcp 10.217.0.155:9311: connect: connection refused" Oct 08 18:30:14 crc kubenswrapper[4988]: I1008 18:30:14.545636 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-gm2vf" Oct 08 18:30:14 crc kubenswrapper[4988]: I1008 18:30:14.666855 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5b6f72e9-8dc6-4290-b032-d8a570b91767-scripts\") pod \"5b6f72e9-8dc6-4290-b032-d8a570b91767\" (UID: \"5b6f72e9-8dc6-4290-b032-d8a570b91767\") " Oct 08 18:30:14 crc kubenswrapper[4988]: I1008 18:30:14.667058 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5b6f72e9-8dc6-4290-b032-d8a570b91767-combined-ca-bundle\") pod \"5b6f72e9-8dc6-4290-b032-d8a570b91767\" (UID: \"5b6f72e9-8dc6-4290-b032-d8a570b91767\") " Oct 08 18:30:14 crc kubenswrapper[4988]: I1008 18:30:14.667086 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/5b6f72e9-8dc6-4290-b032-d8a570b91767-etc-machine-id\") pod \"5b6f72e9-8dc6-4290-b032-d8a570b91767\" (UID: \"5b6f72e9-8dc6-4290-b032-d8a570b91767\") " Oct 08 18:30:14 crc kubenswrapper[4988]: I1008 18:30:14.667128 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/5b6f72e9-8dc6-4290-b032-d8a570b91767-db-sync-config-data\") pod \"5b6f72e9-8dc6-4290-b032-d8a570b91767\" (UID: \"5b6f72e9-8dc6-4290-b032-d8a570b91767\") " Oct 08 18:30:14 crc kubenswrapper[4988]: I1008 18:30:14.667187 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5b6f72e9-8dc6-4290-b032-d8a570b91767-config-data\") pod \"5b6f72e9-8dc6-4290-b032-d8a570b91767\" (UID: \"5b6f72e9-8dc6-4290-b032-d8a570b91767\") " Oct 08 18:30:14 crc kubenswrapper[4988]: I1008 18:30:14.667242 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lshc6\" (UniqueName: \"kubernetes.io/projected/5b6f72e9-8dc6-4290-b032-d8a570b91767-kube-api-access-lshc6\") pod \"5b6f72e9-8dc6-4290-b032-d8a570b91767\" (UID: \"5b6f72e9-8dc6-4290-b032-d8a570b91767\") " Oct 08 18:30:14 crc kubenswrapper[4988]: I1008 18:30:14.667858 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/5b6f72e9-8dc6-4290-b032-d8a570b91767-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "5b6f72e9-8dc6-4290-b032-d8a570b91767" (UID: "5b6f72e9-8dc6-4290-b032-d8a570b91767"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 08 18:30:14 crc kubenswrapper[4988]: I1008 18:30:14.678301 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5b6f72e9-8dc6-4290-b032-d8a570b91767-kube-api-access-lshc6" (OuterVolumeSpecName: "kube-api-access-lshc6") pod "5b6f72e9-8dc6-4290-b032-d8a570b91767" (UID: "5b6f72e9-8dc6-4290-b032-d8a570b91767"). InnerVolumeSpecName "kube-api-access-lshc6". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:30:14 crc kubenswrapper[4988]: I1008 18:30:14.678305 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5b6f72e9-8dc6-4290-b032-d8a570b91767-scripts" (OuterVolumeSpecName: "scripts") pod "5b6f72e9-8dc6-4290-b032-d8a570b91767" (UID: "5b6f72e9-8dc6-4290-b032-d8a570b91767"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:30:14 crc kubenswrapper[4988]: I1008 18:30:14.684913 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5b6f72e9-8dc6-4290-b032-d8a570b91767-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "5b6f72e9-8dc6-4290-b032-d8a570b91767" (UID: "5b6f72e9-8dc6-4290-b032-d8a570b91767"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:30:14 crc kubenswrapper[4988]: I1008 18:30:14.728772 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5b6f72e9-8dc6-4290-b032-d8a570b91767-config-data" (OuterVolumeSpecName: "config-data") pod "5b6f72e9-8dc6-4290-b032-d8a570b91767" (UID: "5b6f72e9-8dc6-4290-b032-d8a570b91767"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:30:14 crc kubenswrapper[4988]: I1008 18:30:14.740525 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5b6f72e9-8dc6-4290-b032-d8a570b91767-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "5b6f72e9-8dc6-4290-b032-d8a570b91767" (UID: "5b6f72e9-8dc6-4290-b032-d8a570b91767"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:30:14 crc kubenswrapper[4988]: I1008 18:30:14.750298 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-76f7bb8cb8-x44rw" Oct 08 18:30:14 crc kubenswrapper[4988]: I1008 18:30:14.772723 4988 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5b6f72e9-8dc6-4290-b032-d8a570b91767-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 18:30:14 crc kubenswrapper[4988]: I1008 18:30:14.772769 4988 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/5b6f72e9-8dc6-4290-b032-d8a570b91767-etc-machine-id\") on node \"crc\" DevicePath \"\"" Oct 08 18:30:14 crc kubenswrapper[4988]: I1008 18:30:14.772785 4988 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/5b6f72e9-8dc6-4290-b032-d8a570b91767-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Oct 08 18:30:14 crc kubenswrapper[4988]: I1008 18:30:14.772797 4988 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5b6f72e9-8dc6-4290-b032-d8a570b91767-config-data\") on node \"crc\" DevicePath \"\"" Oct 08 18:30:14 crc kubenswrapper[4988]: I1008 18:30:14.772809 4988 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lshc6\" (UniqueName: \"kubernetes.io/projected/5b6f72e9-8dc6-4290-b032-d8a570b91767-kube-api-access-lshc6\") on node \"crc\" DevicePath \"\"" Oct 08 18:30:14 crc kubenswrapper[4988]: I1008 18:30:14.772822 4988 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5b6f72e9-8dc6-4290-b032-d8a570b91767-scripts\") on node \"crc\" DevicePath \"\"" Oct 08 18:30:14 crc kubenswrapper[4988]: I1008 18:30:14.847707 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-54df4b685c-fr6bz" Oct 08 18:30:14 crc kubenswrapper[4988]: I1008 18:30:14.878341 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/fe49d0c8-90f0-4673-b527-ff112cfd3037-config-data-custom\") pod \"fe49d0c8-90f0-4673-b527-ff112cfd3037\" (UID: \"fe49d0c8-90f0-4673-b527-ff112cfd3037\") " Oct 08 18:30:14 crc kubenswrapper[4988]: I1008 18:30:14.878513 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fe49d0c8-90f0-4673-b527-ff112cfd3037-logs\") pod \"fe49d0c8-90f0-4673-b527-ff112cfd3037\" (UID: \"fe49d0c8-90f0-4673-b527-ff112cfd3037\") " Oct 08 18:30:14 crc kubenswrapper[4988]: I1008 18:30:14.878548 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fe49d0c8-90f0-4673-b527-ff112cfd3037-combined-ca-bundle\") pod \"fe49d0c8-90f0-4673-b527-ff112cfd3037\" (UID: \"fe49d0c8-90f0-4673-b527-ff112cfd3037\") " Oct 08 18:30:14 crc kubenswrapper[4988]: I1008 18:30:14.878679 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cjvcj\" (UniqueName: \"kubernetes.io/projected/fe49d0c8-90f0-4673-b527-ff112cfd3037-kube-api-access-cjvcj\") pod \"fe49d0c8-90f0-4673-b527-ff112cfd3037\" (UID: \"fe49d0c8-90f0-4673-b527-ff112cfd3037\") " Oct 08 18:30:14 crc kubenswrapper[4988]: I1008 18:30:14.878725 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fe49d0c8-90f0-4673-b527-ff112cfd3037-config-data\") pod \"fe49d0c8-90f0-4673-b527-ff112cfd3037\" (UID: \"fe49d0c8-90f0-4673-b527-ff112cfd3037\") " Oct 08 18:30:14 crc kubenswrapper[4988]: I1008 18:30:14.883199 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fe49d0c8-90f0-4673-b527-ff112cfd3037-logs" (OuterVolumeSpecName: "logs") pod "fe49d0c8-90f0-4673-b527-ff112cfd3037" (UID: "fe49d0c8-90f0-4673-b527-ff112cfd3037"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 18:30:14 crc kubenswrapper[4988]: I1008 18:30:14.886093 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fe49d0c8-90f0-4673-b527-ff112cfd3037-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "fe49d0c8-90f0-4673-b527-ff112cfd3037" (UID: "fe49d0c8-90f0-4673-b527-ff112cfd3037"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:30:14 crc kubenswrapper[4988]: I1008 18:30:14.887197 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fe49d0c8-90f0-4673-b527-ff112cfd3037-kube-api-access-cjvcj" (OuterVolumeSpecName: "kube-api-access-cjvcj") pod "fe49d0c8-90f0-4673-b527-ff112cfd3037" (UID: "fe49d0c8-90f0-4673-b527-ff112cfd3037"). InnerVolumeSpecName "kube-api-access-cjvcj". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:30:14 crc kubenswrapper[4988]: I1008 18:30:14.889004 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 08 18:30:14 crc kubenswrapper[4988]: I1008 18:30:14.911881 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fe49d0c8-90f0-4673-b527-ff112cfd3037-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "fe49d0c8-90f0-4673-b527-ff112cfd3037" (UID: "fe49d0c8-90f0-4673-b527-ff112cfd3037"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:30:14 crc kubenswrapper[4988]: I1008 18:30:14.965008 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fe49d0c8-90f0-4673-b527-ff112cfd3037-config-data" (OuterVolumeSpecName: "config-data") pod "fe49d0c8-90f0-4673-b527-ff112cfd3037" (UID: "fe49d0c8-90f0-4673-b527-ff112cfd3037"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:30:14 crc kubenswrapper[4988]: I1008 18:30:14.979888 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/d9f52cd3-500f-45b9-9f83-8e8244e511d6-dns-swift-storage-0\") pod \"d9f52cd3-500f-45b9-9f83-8e8244e511d6\" (UID: \"d9f52cd3-500f-45b9-9f83-8e8244e511d6\") " Oct 08 18:30:14 crc kubenswrapper[4988]: I1008 18:30:14.979927 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9e8e39d2-57c8-4ea6-9c83-dcb8d75274cd-log-httpd\") pod \"9e8e39d2-57c8-4ea6-9c83-dcb8d75274cd\" (UID: \"9e8e39d2-57c8-4ea6-9c83-dcb8d75274cd\") " Oct 08 18:30:14 crc kubenswrapper[4988]: I1008 18:30:14.979954 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4txp6\" (UniqueName: \"kubernetes.io/projected/9e8e39d2-57c8-4ea6-9c83-dcb8d75274cd-kube-api-access-4txp6\") pod \"9e8e39d2-57c8-4ea6-9c83-dcb8d75274cd\" (UID: \"9e8e39d2-57c8-4ea6-9c83-dcb8d75274cd\") " Oct 08 18:30:14 crc kubenswrapper[4988]: I1008 18:30:14.980003 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d9f52cd3-500f-45b9-9f83-8e8244e511d6-ovsdbserver-sb\") pod \"d9f52cd3-500f-45b9-9f83-8e8244e511d6\" (UID: \"d9f52cd3-500f-45b9-9f83-8e8244e511d6\") " Oct 08 18:30:14 crc kubenswrapper[4988]: I1008 18:30:14.980031 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9e8e39d2-57c8-4ea6-9c83-dcb8d75274cd-config-data\") pod \"9e8e39d2-57c8-4ea6-9c83-dcb8d75274cd\" (UID: \"9e8e39d2-57c8-4ea6-9c83-dcb8d75274cd\") " Oct 08 18:30:14 crc kubenswrapper[4988]: I1008 18:30:14.980055 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wx89j\" (UniqueName: \"kubernetes.io/projected/d9f52cd3-500f-45b9-9f83-8e8244e511d6-kube-api-access-wx89j\") pod \"d9f52cd3-500f-45b9-9f83-8e8244e511d6\" (UID: \"d9f52cd3-500f-45b9-9f83-8e8244e511d6\") " Oct 08 18:30:14 crc kubenswrapper[4988]: I1008 18:30:14.980097 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d9f52cd3-500f-45b9-9f83-8e8244e511d6-ovsdbserver-nb\") pod \"d9f52cd3-500f-45b9-9f83-8e8244e511d6\" (UID: \"d9f52cd3-500f-45b9-9f83-8e8244e511d6\") " Oct 08 18:30:14 crc kubenswrapper[4988]: I1008 18:30:14.980145 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d9f52cd3-500f-45b9-9f83-8e8244e511d6-dns-svc\") pod \"d9f52cd3-500f-45b9-9f83-8e8244e511d6\" (UID: \"d9f52cd3-500f-45b9-9f83-8e8244e511d6\") " Oct 08 18:30:14 crc kubenswrapper[4988]: I1008 18:30:14.980164 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d9f52cd3-500f-45b9-9f83-8e8244e511d6-config\") pod \"d9f52cd3-500f-45b9-9f83-8e8244e511d6\" (UID: \"d9f52cd3-500f-45b9-9f83-8e8244e511d6\") " Oct 08 18:30:14 crc kubenswrapper[4988]: I1008 18:30:14.980188 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9e8e39d2-57c8-4ea6-9c83-dcb8d75274cd-run-httpd\") pod \"9e8e39d2-57c8-4ea6-9c83-dcb8d75274cd\" (UID: \"9e8e39d2-57c8-4ea6-9c83-dcb8d75274cd\") " Oct 08 18:30:14 crc kubenswrapper[4988]: I1008 18:30:14.980234 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/9e8e39d2-57c8-4ea6-9c83-dcb8d75274cd-sg-core-conf-yaml\") pod \"9e8e39d2-57c8-4ea6-9c83-dcb8d75274cd\" (UID: \"9e8e39d2-57c8-4ea6-9c83-dcb8d75274cd\") " Oct 08 18:30:14 crc kubenswrapper[4988]: I1008 18:30:14.980271 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9e8e39d2-57c8-4ea6-9c83-dcb8d75274cd-scripts\") pod \"9e8e39d2-57c8-4ea6-9c83-dcb8d75274cd\" (UID: \"9e8e39d2-57c8-4ea6-9c83-dcb8d75274cd\") " Oct 08 18:30:14 crc kubenswrapper[4988]: I1008 18:30:14.980305 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9e8e39d2-57c8-4ea6-9c83-dcb8d75274cd-combined-ca-bundle\") pod \"9e8e39d2-57c8-4ea6-9c83-dcb8d75274cd\" (UID: \"9e8e39d2-57c8-4ea6-9c83-dcb8d75274cd\") " Oct 08 18:30:14 crc kubenswrapper[4988]: I1008 18:30:14.980495 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9e8e39d2-57c8-4ea6-9c83-dcb8d75274cd-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "9e8e39d2-57c8-4ea6-9c83-dcb8d75274cd" (UID: "9e8e39d2-57c8-4ea6-9c83-dcb8d75274cd"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 18:30:14 crc kubenswrapper[4988]: I1008 18:30:14.981022 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9e8e39d2-57c8-4ea6-9c83-dcb8d75274cd-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "9e8e39d2-57c8-4ea6-9c83-dcb8d75274cd" (UID: "9e8e39d2-57c8-4ea6-9c83-dcb8d75274cd"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 18:30:14 crc kubenswrapper[4988]: I1008 18:30:14.981410 4988 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cjvcj\" (UniqueName: \"kubernetes.io/projected/fe49d0c8-90f0-4673-b527-ff112cfd3037-kube-api-access-cjvcj\") on node \"crc\" DevicePath \"\"" Oct 08 18:30:14 crc kubenswrapper[4988]: I1008 18:30:14.981434 4988 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fe49d0c8-90f0-4673-b527-ff112cfd3037-config-data\") on node \"crc\" DevicePath \"\"" Oct 08 18:30:14 crc kubenswrapper[4988]: I1008 18:30:14.981451 4988 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9e8e39d2-57c8-4ea6-9c83-dcb8d75274cd-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 08 18:30:14 crc kubenswrapper[4988]: I1008 18:30:14.981485 4988 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/fe49d0c8-90f0-4673-b527-ff112cfd3037-config-data-custom\") on node \"crc\" DevicePath \"\"" Oct 08 18:30:14 crc kubenswrapper[4988]: I1008 18:30:14.982089 4988 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fe49d0c8-90f0-4673-b527-ff112cfd3037-logs\") on node \"crc\" DevicePath \"\"" Oct 08 18:30:14 crc kubenswrapper[4988]: I1008 18:30:14.982112 4988 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fe49d0c8-90f0-4673-b527-ff112cfd3037-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 18:30:14 crc kubenswrapper[4988]: I1008 18:30:14.982126 4988 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9e8e39d2-57c8-4ea6-9c83-dcb8d75274cd-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 08 18:30:14 crc kubenswrapper[4988]: I1008 18:30:14.986673 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9e8e39d2-57c8-4ea6-9c83-dcb8d75274cd-scripts" (OuterVolumeSpecName: "scripts") pod "9e8e39d2-57c8-4ea6-9c83-dcb8d75274cd" (UID: "9e8e39d2-57c8-4ea6-9c83-dcb8d75274cd"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:30:14 crc kubenswrapper[4988]: I1008 18:30:14.996686 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9e8e39d2-57c8-4ea6-9c83-dcb8d75274cd-kube-api-access-4txp6" (OuterVolumeSpecName: "kube-api-access-4txp6") pod "9e8e39d2-57c8-4ea6-9c83-dcb8d75274cd" (UID: "9e8e39d2-57c8-4ea6-9c83-dcb8d75274cd"). InnerVolumeSpecName "kube-api-access-4txp6". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:30:15 crc kubenswrapper[4988]: I1008 18:30:15.009689 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d9f52cd3-500f-45b9-9f83-8e8244e511d6-kube-api-access-wx89j" (OuterVolumeSpecName: "kube-api-access-wx89j") pod "d9f52cd3-500f-45b9-9f83-8e8244e511d6" (UID: "d9f52cd3-500f-45b9-9f83-8e8244e511d6"). InnerVolumeSpecName "kube-api-access-wx89j". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:30:15 crc kubenswrapper[4988]: I1008 18:30:15.012997 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-gm2vf" event={"ID":"5b6f72e9-8dc6-4290-b032-d8a570b91767","Type":"ContainerDied","Data":"0445e0aa866fd32b4478f427736e174d9c619b3c4df681e2c48994219d561e07"} Oct 08 18:30:15 crc kubenswrapper[4988]: I1008 18:30:15.013044 4988 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0445e0aa866fd32b4478f427736e174d9c619b3c4df681e2c48994219d561e07" Oct 08 18:30:15 crc kubenswrapper[4988]: I1008 18:30:15.013279 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-gm2vf" Oct 08 18:30:15 crc kubenswrapper[4988]: I1008 18:30:15.016814 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"c97aab35-bbee-41a6-88b2-801b0b097a6d","Type":"ContainerStarted","Data":"47dce63c2014855045e7a231270ccde5910760705845d62023d29e100f929478"} Oct 08 18:30:15 crc kubenswrapper[4988]: I1008 18:30:15.034353 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-76f7bb8cb8-x44rw" event={"ID":"fe49d0c8-90f0-4673-b527-ff112cfd3037","Type":"ContainerDied","Data":"2e61b2af6a360d8f3ce49333eb603bdf2a9800e8b88f055d0906ef61d8be49b4"} Oct 08 18:30:15 crc kubenswrapper[4988]: I1008 18:30:15.034412 4988 scope.go:117] "RemoveContainer" containerID="dfe3a72d9e18e94651f136aa2b98ffa4d52065039c4bc92724da4663ef94fddf" Oct 08 18:30:15 crc kubenswrapper[4988]: I1008 18:30:15.034529 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-76f7bb8cb8-x44rw" Oct 08 18:30:15 crc kubenswrapper[4988]: I1008 18:30:15.053177 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-54df4b685c-fr6bz" event={"ID":"d9f52cd3-500f-45b9-9f83-8e8244e511d6","Type":"ContainerDied","Data":"a1cdc1b3db5fda1d178b0965e413f8b36b8c31907bf682098c127f991ed50430"} Oct 08 18:30:15 crc kubenswrapper[4988]: I1008 18:30:15.053232 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-54df4b685c-fr6bz" Oct 08 18:30:15 crc kubenswrapper[4988]: I1008 18:30:15.058418 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstackclient" podStartSLOduration=2.140015704 podStartE2EDuration="16.058402076s" podCreationTimestamp="2025-10-08 18:29:59 +0000 UTC" firstStartedPulling="2025-10-08 18:30:00.502948609 +0000 UTC m=+1145.952791379" lastFinishedPulling="2025-10-08 18:30:14.421334971 +0000 UTC m=+1159.871177751" observedRunningTime="2025-10-08 18:30:15.034056738 +0000 UTC m=+1160.483899508" watchObservedRunningTime="2025-10-08 18:30:15.058402076 +0000 UTC m=+1160.508244846" Oct 08 18:30:15 crc kubenswrapper[4988]: I1008 18:30:15.063623 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9e8e39d2-57c8-4ea6-9c83-dcb8d75274cd-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "9e8e39d2-57c8-4ea6-9c83-dcb8d75274cd" (UID: "9e8e39d2-57c8-4ea6-9c83-dcb8d75274cd"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:30:15 crc kubenswrapper[4988]: I1008 18:30:15.063723 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9e8e39d2-57c8-4ea6-9c83-dcb8d75274cd","Type":"ContainerDied","Data":"c44e85e774111fa2a83bd1e90e873d37a116d12f0d295b4b11395053d875cf29"} Oct 08 18:30:15 crc kubenswrapper[4988]: I1008 18:30:15.063772 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 08 18:30:15 crc kubenswrapper[4988]: I1008 18:30:15.083657 4988 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/9e8e39d2-57c8-4ea6-9c83-dcb8d75274cd-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Oct 08 18:30:15 crc kubenswrapper[4988]: I1008 18:30:15.083684 4988 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9e8e39d2-57c8-4ea6-9c83-dcb8d75274cd-scripts\") on node \"crc\" DevicePath \"\"" Oct 08 18:30:15 crc kubenswrapper[4988]: I1008 18:30:15.083694 4988 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4txp6\" (UniqueName: \"kubernetes.io/projected/9e8e39d2-57c8-4ea6-9c83-dcb8d75274cd-kube-api-access-4txp6\") on node \"crc\" DevicePath \"\"" Oct 08 18:30:15 crc kubenswrapper[4988]: I1008 18:30:15.083704 4988 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wx89j\" (UniqueName: \"kubernetes.io/projected/d9f52cd3-500f-45b9-9f83-8e8244e511d6-kube-api-access-wx89j\") on node \"crc\" DevicePath \"\"" Oct 08 18:30:15 crc kubenswrapper[4988]: I1008 18:30:15.087237 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d9f52cd3-500f-45b9-9f83-8e8244e511d6-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "d9f52cd3-500f-45b9-9f83-8e8244e511d6" (UID: "d9f52cd3-500f-45b9-9f83-8e8244e511d6"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 18:30:15 crc kubenswrapper[4988]: I1008 18:30:15.096213 4988 scope.go:117] "RemoveContainer" containerID="0e25441dbd7ef8b78df421475b3d19064e1a51130eb299a3433a4940e3737266" Oct 08 18:30:15 crc kubenswrapper[4988]: I1008 18:30:15.104707 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d9f52cd3-500f-45b9-9f83-8e8244e511d6-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "d9f52cd3-500f-45b9-9f83-8e8244e511d6" (UID: "d9f52cd3-500f-45b9-9f83-8e8244e511d6"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 18:30:15 crc kubenswrapper[4988]: I1008 18:30:15.107064 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-proxy-7b7cf6bf4f-s6vrs"] Oct 08 18:30:15 crc kubenswrapper[4988]: I1008 18:30:15.108410 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d9f52cd3-500f-45b9-9f83-8e8244e511d6-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "d9f52cd3-500f-45b9-9f83-8e8244e511d6" (UID: "d9f52cd3-500f-45b9-9f83-8e8244e511d6"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 18:30:15 crc kubenswrapper[4988]: I1008 18:30:15.114455 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-76f7bb8cb8-x44rw"] Oct 08 18:30:15 crc kubenswrapper[4988]: I1008 18:30:15.114580 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d9f52cd3-500f-45b9-9f83-8e8244e511d6-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "d9f52cd3-500f-45b9-9f83-8e8244e511d6" (UID: "d9f52cd3-500f-45b9-9f83-8e8244e511d6"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 18:30:15 crc kubenswrapper[4988]: I1008 18:30:15.122534 4988 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-api-76f7bb8cb8-x44rw"] Oct 08 18:30:15 crc kubenswrapper[4988]: I1008 18:30:15.128104 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d9f52cd3-500f-45b9-9f83-8e8244e511d6-config" (OuterVolumeSpecName: "config") pod "d9f52cd3-500f-45b9-9f83-8e8244e511d6" (UID: "d9f52cd3-500f-45b9-9f83-8e8244e511d6"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 18:30:15 crc kubenswrapper[4988]: I1008 18:30:15.131670 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9e8e39d2-57c8-4ea6-9c83-dcb8d75274cd-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "9e8e39d2-57c8-4ea6-9c83-dcb8d75274cd" (UID: "9e8e39d2-57c8-4ea6-9c83-dcb8d75274cd"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:30:15 crc kubenswrapper[4988]: W1008 18:30:15.132922 4988 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod857686e3_0c6b_4d08_a8d6_f89e4c832139.slice/crio-9bfa3aa2a36b16efd6317af958966d09bd036f9b524035fdf3e77884e6bb79c6 WatchSource:0}: Error finding container 9bfa3aa2a36b16efd6317af958966d09bd036f9b524035fdf3e77884e6bb79c6: Status 404 returned error can't find the container with id 9bfa3aa2a36b16efd6317af958966d09bd036f9b524035fdf3e77884e6bb79c6 Oct 08 18:30:15 crc kubenswrapper[4988]: I1008 18:30:15.142689 4988 scope.go:117] "RemoveContainer" containerID="9a9ae238ebc775efdca8a60be60a294b733e168bb9d325901a7175d4ac2574bb" Oct 08 18:30:15 crc kubenswrapper[4988]: I1008 18:30:15.163589 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9e8e39d2-57c8-4ea6-9c83-dcb8d75274cd-config-data" (OuterVolumeSpecName: "config-data") pod "9e8e39d2-57c8-4ea6-9c83-dcb8d75274cd" (UID: "9e8e39d2-57c8-4ea6-9c83-dcb8d75274cd"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:30:15 crc kubenswrapper[4988]: I1008 18:30:15.164590 4988 scope.go:117] "RemoveContainer" containerID="629b7f05d0b6b84e81db7afa778e7f08381c6ce616a2416e8b5bb6bd08dd4c6a" Oct 08 18:30:15 crc kubenswrapper[4988]: I1008 18:30:15.185339 4988 scope.go:117] "RemoveContainer" containerID="1e28870498bf7017bb7aaf972c99e33b912ab2db76680063b2dd790406e59691" Oct 08 18:30:15 crc kubenswrapper[4988]: I1008 18:30:15.185430 4988 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d9f52cd3-500f-45b9-9f83-8e8244e511d6-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 08 18:30:15 crc kubenswrapper[4988]: I1008 18:30:15.185457 4988 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d9f52cd3-500f-45b9-9f83-8e8244e511d6-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 08 18:30:15 crc kubenswrapper[4988]: I1008 18:30:15.185466 4988 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d9f52cd3-500f-45b9-9f83-8e8244e511d6-config\") on node \"crc\" DevicePath \"\"" Oct 08 18:30:15 crc kubenswrapper[4988]: I1008 18:30:15.185475 4988 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9e8e39d2-57c8-4ea6-9c83-dcb8d75274cd-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 18:30:15 crc kubenswrapper[4988]: I1008 18:30:15.185485 4988 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/d9f52cd3-500f-45b9-9f83-8e8244e511d6-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Oct 08 18:30:15 crc kubenswrapper[4988]: I1008 18:30:15.185495 4988 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d9f52cd3-500f-45b9-9f83-8e8244e511d6-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 08 18:30:15 crc kubenswrapper[4988]: I1008 18:30:15.185503 4988 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9e8e39d2-57c8-4ea6-9c83-dcb8d75274cd-config-data\") on node \"crc\" DevicePath \"\"" Oct 08 18:30:15 crc kubenswrapper[4988]: I1008 18:30:15.225157 4988 scope.go:117] "RemoveContainer" containerID="3611c104fb9f40d7e4eb1645a5029783c73aa03a8d819c857e23f703653a58e3" Oct 08 18:30:15 crc kubenswrapper[4988]: I1008 18:30:15.246101 4988 scope.go:117] "RemoveContainer" containerID="46c9d841d83e09ba232a18c37dc58ed0a217dca805674f3203c8f5d7cbddb5cb" Oct 08 18:30:15 crc kubenswrapper[4988]: I1008 18:30:15.249493 4988 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fe49d0c8-90f0-4673-b527-ff112cfd3037" path="/var/lib/kubelet/pods/fe49d0c8-90f0-4673-b527-ff112cfd3037/volumes" Oct 08 18:30:15 crc kubenswrapper[4988]: I1008 18:30:15.278982 4988 scope.go:117] "RemoveContainer" containerID="9640cc556082067881ad4130d32e24914d42ee959a828c2f3e071e7ca9d49656" Oct 08 18:30:15 crc kubenswrapper[4988]: I1008 18:30:15.497849 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 08 18:30:15 crc kubenswrapper[4988]: I1008 18:30:15.514534 4988 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Oct 08 18:30:15 crc kubenswrapper[4988]: I1008 18:30:15.524123 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-54df4b685c-fr6bz"] Oct 08 18:30:15 crc kubenswrapper[4988]: I1008 18:30:15.532889 4988 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-54df4b685c-fr6bz"] Oct 08 18:30:15 crc kubenswrapper[4988]: I1008 18:30:15.541979 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 08 18:30:15 crc kubenswrapper[4988]: E1008 18:30:15.542424 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5b6f72e9-8dc6-4290-b032-d8a570b91767" containerName="cinder-db-sync" Oct 08 18:30:15 crc kubenswrapper[4988]: I1008 18:30:15.542443 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="5b6f72e9-8dc6-4290-b032-d8a570b91767" containerName="cinder-db-sync" Oct 08 18:30:15 crc kubenswrapper[4988]: E1008 18:30:15.542454 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d9f52cd3-500f-45b9-9f83-8e8244e511d6" containerName="init" Oct 08 18:30:15 crc kubenswrapper[4988]: I1008 18:30:15.542460 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="d9f52cd3-500f-45b9-9f83-8e8244e511d6" containerName="init" Oct 08 18:30:15 crc kubenswrapper[4988]: E1008 18:30:15.542471 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9e8e39d2-57c8-4ea6-9c83-dcb8d75274cd" containerName="ceilometer-notification-agent" Oct 08 18:30:15 crc kubenswrapper[4988]: I1008 18:30:15.542478 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="9e8e39d2-57c8-4ea6-9c83-dcb8d75274cd" containerName="ceilometer-notification-agent" Oct 08 18:30:15 crc kubenswrapper[4988]: E1008 18:30:15.542496 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d9f52cd3-500f-45b9-9f83-8e8244e511d6" containerName="dnsmasq-dns" Oct 08 18:30:15 crc kubenswrapper[4988]: I1008 18:30:15.542501 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="d9f52cd3-500f-45b9-9f83-8e8244e511d6" containerName="dnsmasq-dns" Oct 08 18:30:15 crc kubenswrapper[4988]: E1008 18:30:15.542522 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9e8e39d2-57c8-4ea6-9c83-dcb8d75274cd" containerName="proxy-httpd" Oct 08 18:30:15 crc kubenswrapper[4988]: I1008 18:30:15.542527 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="9e8e39d2-57c8-4ea6-9c83-dcb8d75274cd" containerName="proxy-httpd" Oct 08 18:30:15 crc kubenswrapper[4988]: E1008 18:30:15.542535 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9e8e39d2-57c8-4ea6-9c83-dcb8d75274cd" containerName="ceilometer-central-agent" Oct 08 18:30:15 crc kubenswrapper[4988]: I1008 18:30:15.542541 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="9e8e39d2-57c8-4ea6-9c83-dcb8d75274cd" containerName="ceilometer-central-agent" Oct 08 18:30:15 crc kubenswrapper[4988]: E1008 18:30:15.542550 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9e8e39d2-57c8-4ea6-9c83-dcb8d75274cd" containerName="sg-core" Oct 08 18:30:15 crc kubenswrapper[4988]: I1008 18:30:15.542555 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="9e8e39d2-57c8-4ea6-9c83-dcb8d75274cd" containerName="sg-core" Oct 08 18:30:15 crc kubenswrapper[4988]: E1008 18:30:15.542566 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fe49d0c8-90f0-4673-b527-ff112cfd3037" containerName="barbican-api-log" Oct 08 18:30:15 crc kubenswrapper[4988]: I1008 18:30:15.542572 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="fe49d0c8-90f0-4673-b527-ff112cfd3037" containerName="barbican-api-log" Oct 08 18:30:15 crc kubenswrapper[4988]: E1008 18:30:15.542580 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fe49d0c8-90f0-4673-b527-ff112cfd3037" containerName="barbican-api" Oct 08 18:30:15 crc kubenswrapper[4988]: I1008 18:30:15.542586 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="fe49d0c8-90f0-4673-b527-ff112cfd3037" containerName="barbican-api" Oct 08 18:30:15 crc kubenswrapper[4988]: I1008 18:30:15.542741 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="9e8e39d2-57c8-4ea6-9c83-dcb8d75274cd" containerName="proxy-httpd" Oct 08 18:30:15 crc kubenswrapper[4988]: I1008 18:30:15.542756 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="5b6f72e9-8dc6-4290-b032-d8a570b91767" containerName="cinder-db-sync" Oct 08 18:30:15 crc kubenswrapper[4988]: I1008 18:30:15.542769 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="fe49d0c8-90f0-4673-b527-ff112cfd3037" containerName="barbican-api-log" Oct 08 18:30:15 crc kubenswrapper[4988]: I1008 18:30:15.542780 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="9e8e39d2-57c8-4ea6-9c83-dcb8d75274cd" containerName="ceilometer-central-agent" Oct 08 18:30:15 crc kubenswrapper[4988]: I1008 18:30:15.542788 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="d9f52cd3-500f-45b9-9f83-8e8244e511d6" containerName="dnsmasq-dns" Oct 08 18:30:15 crc kubenswrapper[4988]: I1008 18:30:15.542794 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="fe49d0c8-90f0-4673-b527-ff112cfd3037" containerName="barbican-api" Oct 08 18:30:15 crc kubenswrapper[4988]: I1008 18:30:15.542801 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="9e8e39d2-57c8-4ea6-9c83-dcb8d75274cd" containerName="sg-core" Oct 08 18:30:15 crc kubenswrapper[4988]: I1008 18:30:15.542808 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="9e8e39d2-57c8-4ea6-9c83-dcb8d75274cd" containerName="ceilometer-notification-agent" Oct 08 18:30:15 crc kubenswrapper[4988]: I1008 18:30:15.545315 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 08 18:30:15 crc kubenswrapper[4988]: I1008 18:30:15.549160 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 08 18:30:15 crc kubenswrapper[4988]: I1008 18:30:15.549515 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 08 18:30:15 crc kubenswrapper[4988]: I1008 18:30:15.577574 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 08 18:30:15 crc kubenswrapper[4988]: I1008 18:30:15.595651 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/dd22b6ca-9c02-46f8-b202-e9634e79261e-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"dd22b6ca-9c02-46f8-b202-e9634e79261e\") " pod="openstack/ceilometer-0" Oct 08 18:30:15 crc kubenswrapper[4988]: I1008 18:30:15.595799 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q4hm6\" (UniqueName: \"kubernetes.io/projected/dd22b6ca-9c02-46f8-b202-e9634e79261e-kube-api-access-q4hm6\") pod \"ceilometer-0\" (UID: \"dd22b6ca-9c02-46f8-b202-e9634e79261e\") " pod="openstack/ceilometer-0" Oct 08 18:30:15 crc kubenswrapper[4988]: I1008 18:30:15.595842 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dd22b6ca-9c02-46f8-b202-e9634e79261e-config-data\") pod \"ceilometer-0\" (UID: \"dd22b6ca-9c02-46f8-b202-e9634e79261e\") " pod="openstack/ceilometer-0" Oct 08 18:30:15 crc kubenswrapper[4988]: I1008 18:30:15.595876 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dd22b6ca-9c02-46f8-b202-e9634e79261e-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"dd22b6ca-9c02-46f8-b202-e9634e79261e\") " pod="openstack/ceilometer-0" Oct 08 18:30:15 crc kubenswrapper[4988]: I1008 18:30:15.595922 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/dd22b6ca-9c02-46f8-b202-e9634e79261e-scripts\") pod \"ceilometer-0\" (UID: \"dd22b6ca-9c02-46f8-b202-e9634e79261e\") " pod="openstack/ceilometer-0" Oct 08 18:30:15 crc kubenswrapper[4988]: I1008 18:30:15.595950 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/dd22b6ca-9c02-46f8-b202-e9634e79261e-run-httpd\") pod \"ceilometer-0\" (UID: \"dd22b6ca-9c02-46f8-b202-e9634e79261e\") " pod="openstack/ceilometer-0" Oct 08 18:30:15 crc kubenswrapper[4988]: I1008 18:30:15.595976 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/dd22b6ca-9c02-46f8-b202-e9634e79261e-log-httpd\") pod \"ceilometer-0\" (UID: \"dd22b6ca-9c02-46f8-b202-e9634e79261e\") " pod="openstack/ceilometer-0" Oct 08 18:30:15 crc kubenswrapper[4988]: I1008 18:30:15.697346 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q4hm6\" (UniqueName: \"kubernetes.io/projected/dd22b6ca-9c02-46f8-b202-e9634e79261e-kube-api-access-q4hm6\") pod \"ceilometer-0\" (UID: \"dd22b6ca-9c02-46f8-b202-e9634e79261e\") " pod="openstack/ceilometer-0" Oct 08 18:30:15 crc kubenswrapper[4988]: I1008 18:30:15.697441 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dd22b6ca-9c02-46f8-b202-e9634e79261e-config-data\") pod \"ceilometer-0\" (UID: \"dd22b6ca-9c02-46f8-b202-e9634e79261e\") " pod="openstack/ceilometer-0" Oct 08 18:30:15 crc kubenswrapper[4988]: I1008 18:30:15.697487 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dd22b6ca-9c02-46f8-b202-e9634e79261e-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"dd22b6ca-9c02-46f8-b202-e9634e79261e\") " pod="openstack/ceilometer-0" Oct 08 18:30:15 crc kubenswrapper[4988]: I1008 18:30:15.697548 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/dd22b6ca-9c02-46f8-b202-e9634e79261e-scripts\") pod \"ceilometer-0\" (UID: \"dd22b6ca-9c02-46f8-b202-e9634e79261e\") " pod="openstack/ceilometer-0" Oct 08 18:30:15 crc kubenswrapper[4988]: I1008 18:30:15.697586 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/dd22b6ca-9c02-46f8-b202-e9634e79261e-run-httpd\") pod \"ceilometer-0\" (UID: \"dd22b6ca-9c02-46f8-b202-e9634e79261e\") " pod="openstack/ceilometer-0" Oct 08 18:30:15 crc kubenswrapper[4988]: I1008 18:30:15.697615 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/dd22b6ca-9c02-46f8-b202-e9634e79261e-log-httpd\") pod \"ceilometer-0\" (UID: \"dd22b6ca-9c02-46f8-b202-e9634e79261e\") " pod="openstack/ceilometer-0" Oct 08 18:30:15 crc kubenswrapper[4988]: I1008 18:30:15.697660 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/dd22b6ca-9c02-46f8-b202-e9634e79261e-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"dd22b6ca-9c02-46f8-b202-e9634e79261e\") " pod="openstack/ceilometer-0" Oct 08 18:30:15 crc kubenswrapper[4988]: I1008 18:30:15.698315 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/dd22b6ca-9c02-46f8-b202-e9634e79261e-run-httpd\") pod \"ceilometer-0\" (UID: \"dd22b6ca-9c02-46f8-b202-e9634e79261e\") " pod="openstack/ceilometer-0" Oct 08 18:30:15 crc kubenswrapper[4988]: I1008 18:30:15.698365 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/dd22b6ca-9c02-46f8-b202-e9634e79261e-log-httpd\") pod \"ceilometer-0\" (UID: \"dd22b6ca-9c02-46f8-b202-e9634e79261e\") " pod="openstack/ceilometer-0" Oct 08 18:30:15 crc kubenswrapper[4988]: I1008 18:30:15.703785 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/dd22b6ca-9c02-46f8-b202-e9634e79261e-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"dd22b6ca-9c02-46f8-b202-e9634e79261e\") " pod="openstack/ceilometer-0" Oct 08 18:30:15 crc kubenswrapper[4988]: I1008 18:30:15.704270 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dd22b6ca-9c02-46f8-b202-e9634e79261e-config-data\") pod \"ceilometer-0\" (UID: \"dd22b6ca-9c02-46f8-b202-e9634e79261e\") " pod="openstack/ceilometer-0" Oct 08 18:30:15 crc kubenswrapper[4988]: I1008 18:30:15.704594 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/dd22b6ca-9c02-46f8-b202-e9634e79261e-scripts\") pod \"ceilometer-0\" (UID: \"dd22b6ca-9c02-46f8-b202-e9634e79261e\") " pod="openstack/ceilometer-0" Oct 08 18:30:15 crc kubenswrapper[4988]: I1008 18:30:15.714464 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dd22b6ca-9c02-46f8-b202-e9634e79261e-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"dd22b6ca-9c02-46f8-b202-e9634e79261e\") " pod="openstack/ceilometer-0" Oct 08 18:30:15 crc kubenswrapper[4988]: I1008 18:30:15.727986 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q4hm6\" (UniqueName: \"kubernetes.io/projected/dd22b6ca-9c02-46f8-b202-e9634e79261e-kube-api-access-q4hm6\") pod \"ceilometer-0\" (UID: \"dd22b6ca-9c02-46f8-b202-e9634e79261e\") " pod="openstack/ceilometer-0" Oct 08 18:30:15 crc kubenswrapper[4988]: I1008 18:30:15.826569 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Oct 08 18:30:15 crc kubenswrapper[4988]: I1008 18:30:15.827972 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Oct 08 18:30:15 crc kubenswrapper[4988]: I1008 18:30:15.835785 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Oct 08 18:30:15 crc kubenswrapper[4988]: I1008 18:30:15.836023 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Oct 08 18:30:15 crc kubenswrapper[4988]: I1008 18:30:15.842871 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-pvckn" Oct 08 18:30:15 crc kubenswrapper[4988]: I1008 18:30:15.851829 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Oct 08 18:30:15 crc kubenswrapper[4988]: I1008 18:30:15.870558 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 08 18:30:15 crc kubenswrapper[4988]: I1008 18:30:15.901377 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/48acdfd7-797c-4545-9dd5-a9087ec4a525-scripts\") pod \"cinder-scheduler-0\" (UID: \"48acdfd7-797c-4545-9dd5-a9087ec4a525\") " pod="openstack/cinder-scheduler-0" Oct 08 18:30:15 crc kubenswrapper[4988]: I1008 18:30:15.901778 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/48acdfd7-797c-4545-9dd5-a9087ec4a525-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"48acdfd7-797c-4545-9dd5-a9087ec4a525\") " pod="openstack/cinder-scheduler-0" Oct 08 18:30:15 crc kubenswrapper[4988]: I1008 18:30:15.901814 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/48acdfd7-797c-4545-9dd5-a9087ec4a525-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"48acdfd7-797c-4545-9dd5-a9087ec4a525\") " pod="openstack/cinder-scheduler-0" Oct 08 18:30:15 crc kubenswrapper[4988]: I1008 18:30:15.901889 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/48acdfd7-797c-4545-9dd5-a9087ec4a525-config-data\") pod \"cinder-scheduler-0\" (UID: \"48acdfd7-797c-4545-9dd5-a9087ec4a525\") " pod="openstack/cinder-scheduler-0" Oct 08 18:30:15 crc kubenswrapper[4988]: I1008 18:30:15.901912 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/48acdfd7-797c-4545-9dd5-a9087ec4a525-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"48acdfd7-797c-4545-9dd5-a9087ec4a525\") " pod="openstack/cinder-scheduler-0" Oct 08 18:30:15 crc kubenswrapper[4988]: I1008 18:30:15.902013 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6vf2x\" (UniqueName: \"kubernetes.io/projected/48acdfd7-797c-4545-9dd5-a9087ec4a525-kube-api-access-6vf2x\") pod \"cinder-scheduler-0\" (UID: \"48acdfd7-797c-4545-9dd5-a9087ec4a525\") " pod="openstack/cinder-scheduler-0" Oct 08 18:30:15 crc kubenswrapper[4988]: I1008 18:30:15.922470 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 08 18:30:15 crc kubenswrapper[4988]: I1008 18:30:15.986855 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-db-create-nxrck"] Oct 08 18:30:15 crc kubenswrapper[4988]: I1008 18:30:15.987981 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-nxrck" Oct 08 18:30:16 crc kubenswrapper[4988]: I1008 18:30:16.000140 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-nxrck"] Oct 08 18:30:16 crc kubenswrapper[4988]: I1008 18:30:16.004863 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/48acdfd7-797c-4545-9dd5-a9087ec4a525-scripts\") pod \"cinder-scheduler-0\" (UID: \"48acdfd7-797c-4545-9dd5-a9087ec4a525\") " pod="openstack/cinder-scheduler-0" Oct 08 18:30:16 crc kubenswrapper[4988]: I1008 18:30:16.004897 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/48acdfd7-797c-4545-9dd5-a9087ec4a525-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"48acdfd7-797c-4545-9dd5-a9087ec4a525\") " pod="openstack/cinder-scheduler-0" Oct 08 18:30:16 crc kubenswrapper[4988]: I1008 18:30:16.004922 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/48acdfd7-797c-4545-9dd5-a9087ec4a525-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"48acdfd7-797c-4545-9dd5-a9087ec4a525\") " pod="openstack/cinder-scheduler-0" Oct 08 18:30:16 crc kubenswrapper[4988]: I1008 18:30:16.004973 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/48acdfd7-797c-4545-9dd5-a9087ec4a525-config-data\") pod \"cinder-scheduler-0\" (UID: \"48acdfd7-797c-4545-9dd5-a9087ec4a525\") " pod="openstack/cinder-scheduler-0" Oct 08 18:30:16 crc kubenswrapper[4988]: I1008 18:30:16.004993 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/48acdfd7-797c-4545-9dd5-a9087ec4a525-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"48acdfd7-797c-4545-9dd5-a9087ec4a525\") " pod="openstack/cinder-scheduler-0" Oct 08 18:30:16 crc kubenswrapper[4988]: I1008 18:30:16.005062 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6vf2x\" (UniqueName: \"kubernetes.io/projected/48acdfd7-797c-4545-9dd5-a9087ec4a525-kube-api-access-6vf2x\") pod \"cinder-scheduler-0\" (UID: \"48acdfd7-797c-4545-9dd5-a9087ec4a525\") " pod="openstack/cinder-scheduler-0" Oct 08 18:30:16 crc kubenswrapper[4988]: I1008 18:30:16.008481 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/48acdfd7-797c-4545-9dd5-a9087ec4a525-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"48acdfd7-797c-4545-9dd5-a9087ec4a525\") " pod="openstack/cinder-scheduler-0" Oct 08 18:30:16 crc kubenswrapper[4988]: I1008 18:30:16.009403 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/48acdfd7-797c-4545-9dd5-a9087ec4a525-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"48acdfd7-797c-4545-9dd5-a9087ec4a525\") " pod="openstack/cinder-scheduler-0" Oct 08 18:30:16 crc kubenswrapper[4988]: I1008 18:30:16.014265 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/48acdfd7-797c-4545-9dd5-a9087ec4a525-scripts\") pod \"cinder-scheduler-0\" (UID: \"48acdfd7-797c-4545-9dd5-a9087ec4a525\") " pod="openstack/cinder-scheduler-0" Oct 08 18:30:16 crc kubenswrapper[4988]: I1008 18:30:16.015504 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/48acdfd7-797c-4545-9dd5-a9087ec4a525-config-data\") pod \"cinder-scheduler-0\" (UID: \"48acdfd7-797c-4545-9dd5-a9087ec4a525\") " pod="openstack/cinder-scheduler-0" Oct 08 18:30:16 crc kubenswrapper[4988]: I1008 18:30:16.017006 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/48acdfd7-797c-4545-9dd5-a9087ec4a525-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"48acdfd7-797c-4545-9dd5-a9087ec4a525\") " pod="openstack/cinder-scheduler-0" Oct 08 18:30:16 crc kubenswrapper[4988]: I1008 18:30:16.018883 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-59b9656b65-wtfvw"] Oct 08 18:30:16 crc kubenswrapper[4988]: I1008 18:30:16.022470 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-59b9656b65-wtfvw" Oct 08 18:30:16 crc kubenswrapper[4988]: I1008 18:30:16.039284 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-59b9656b65-wtfvw"] Oct 08 18:30:16 crc kubenswrapper[4988]: I1008 18:30:16.074986 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6vf2x\" (UniqueName: \"kubernetes.io/projected/48acdfd7-797c-4545-9dd5-a9087ec4a525-kube-api-access-6vf2x\") pod \"cinder-scheduler-0\" (UID: \"48acdfd7-797c-4545-9dd5-a9087ec4a525\") " pod="openstack/cinder-scheduler-0" Oct 08 18:30:16 crc kubenswrapper[4988]: I1008 18:30:16.118423 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f6134524-f05a-43be-b80d-a862f3e7b390-ovsdbserver-sb\") pod \"dnsmasq-dns-59b9656b65-wtfvw\" (UID: \"f6134524-f05a-43be-b80d-a862f3e7b390\") " pod="openstack/dnsmasq-dns-59b9656b65-wtfvw" Oct 08 18:30:16 crc kubenswrapper[4988]: I1008 18:30:16.118469 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h77st\" (UniqueName: \"kubernetes.io/projected/a86ec23c-c0e4-4e88-901f-ac5e66f103b7-kube-api-access-h77st\") pod \"nova-api-db-create-nxrck\" (UID: \"a86ec23c-c0e4-4e88-901f-ac5e66f103b7\") " pod="openstack/nova-api-db-create-nxrck" Oct 08 18:30:16 crc kubenswrapper[4988]: I1008 18:30:16.118522 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/f6134524-f05a-43be-b80d-a862f3e7b390-dns-swift-storage-0\") pod \"dnsmasq-dns-59b9656b65-wtfvw\" (UID: \"f6134524-f05a-43be-b80d-a862f3e7b390\") " pod="openstack/dnsmasq-dns-59b9656b65-wtfvw" Oct 08 18:30:16 crc kubenswrapper[4988]: I1008 18:30:16.118549 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jpv87\" (UniqueName: \"kubernetes.io/projected/f6134524-f05a-43be-b80d-a862f3e7b390-kube-api-access-jpv87\") pod \"dnsmasq-dns-59b9656b65-wtfvw\" (UID: \"f6134524-f05a-43be-b80d-a862f3e7b390\") " pod="openstack/dnsmasq-dns-59b9656b65-wtfvw" Oct 08 18:30:16 crc kubenswrapper[4988]: I1008 18:30:16.118607 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f6134524-f05a-43be-b80d-a862f3e7b390-config\") pod \"dnsmasq-dns-59b9656b65-wtfvw\" (UID: \"f6134524-f05a-43be-b80d-a862f3e7b390\") " pod="openstack/dnsmasq-dns-59b9656b65-wtfvw" Oct 08 18:30:16 crc kubenswrapper[4988]: I1008 18:30:16.118663 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f6134524-f05a-43be-b80d-a862f3e7b390-ovsdbserver-nb\") pod \"dnsmasq-dns-59b9656b65-wtfvw\" (UID: \"f6134524-f05a-43be-b80d-a862f3e7b390\") " pod="openstack/dnsmasq-dns-59b9656b65-wtfvw" Oct 08 18:30:16 crc kubenswrapper[4988]: I1008 18:30:16.118693 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f6134524-f05a-43be-b80d-a862f3e7b390-dns-svc\") pod \"dnsmasq-dns-59b9656b65-wtfvw\" (UID: \"f6134524-f05a-43be-b80d-a862f3e7b390\") " pod="openstack/dnsmasq-dns-59b9656b65-wtfvw" Oct 08 18:30:16 crc kubenswrapper[4988]: I1008 18:30:16.148328 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Oct 08 18:30:16 crc kubenswrapper[4988]: I1008 18:30:16.169102 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-7b7cf6bf4f-s6vrs" event={"ID":"857686e3-0c6b-4d08-a8d6-f89e4c832139","Type":"ContainerStarted","Data":"ee0c93d867a043a6ace66d0ccc23af52eb47335f2734cbb64f7f04eb41db6e3c"} Oct 08 18:30:16 crc kubenswrapper[4988]: I1008 18:30:16.169144 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-7b7cf6bf4f-s6vrs" event={"ID":"857686e3-0c6b-4d08-a8d6-f89e4c832139","Type":"ContainerStarted","Data":"42e44fdf6315337e912e8f08c104630a1408f1ece9b5a29dc6c2ddc4b1e15bf2"} Oct 08 18:30:16 crc kubenswrapper[4988]: I1008 18:30:16.169153 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-7b7cf6bf4f-s6vrs" event={"ID":"857686e3-0c6b-4d08-a8d6-f89e4c832139","Type":"ContainerStarted","Data":"9bfa3aa2a36b16efd6317af958966d09bd036f9b524035fdf3e77884e6bb79c6"} Oct 08 18:30:16 crc kubenswrapper[4988]: I1008 18:30:16.171623 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-db-create-vbbnt"] Oct 08 18:30:16 crc kubenswrapper[4988]: I1008 18:30:16.172788 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-vbbnt" Oct 08 18:30:16 crc kubenswrapper[4988]: I1008 18:30:16.195422 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-vbbnt"] Oct 08 18:30:16 crc kubenswrapper[4988]: I1008 18:30:16.221365 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f6134524-f05a-43be-b80d-a862f3e7b390-ovsdbserver-nb\") pod \"dnsmasq-dns-59b9656b65-wtfvw\" (UID: \"f6134524-f05a-43be-b80d-a862f3e7b390\") " pod="openstack/dnsmasq-dns-59b9656b65-wtfvw" Oct 08 18:30:16 crc kubenswrapper[4988]: I1008 18:30:16.224113 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f6134524-f05a-43be-b80d-a862f3e7b390-dns-svc\") pod \"dnsmasq-dns-59b9656b65-wtfvw\" (UID: \"f6134524-f05a-43be-b80d-a862f3e7b390\") " pod="openstack/dnsmasq-dns-59b9656b65-wtfvw" Oct 08 18:30:16 crc kubenswrapper[4988]: I1008 18:30:16.224165 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f6134524-f05a-43be-b80d-a862f3e7b390-ovsdbserver-sb\") pod \"dnsmasq-dns-59b9656b65-wtfvw\" (UID: \"f6134524-f05a-43be-b80d-a862f3e7b390\") " pod="openstack/dnsmasq-dns-59b9656b65-wtfvw" Oct 08 18:30:16 crc kubenswrapper[4988]: I1008 18:30:16.222810 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f6134524-f05a-43be-b80d-a862f3e7b390-ovsdbserver-nb\") pod \"dnsmasq-dns-59b9656b65-wtfvw\" (UID: \"f6134524-f05a-43be-b80d-a862f3e7b390\") " pod="openstack/dnsmasq-dns-59b9656b65-wtfvw" Oct 08 18:30:16 crc kubenswrapper[4988]: I1008 18:30:16.224209 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h77st\" (UniqueName: \"kubernetes.io/projected/a86ec23c-c0e4-4e88-901f-ac5e66f103b7-kube-api-access-h77st\") pod \"nova-api-db-create-nxrck\" (UID: \"a86ec23c-c0e4-4e88-901f-ac5e66f103b7\") " pod="openstack/nova-api-db-create-nxrck" Oct 08 18:30:16 crc kubenswrapper[4988]: I1008 18:30:16.224241 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/f6134524-f05a-43be-b80d-a862f3e7b390-dns-swift-storage-0\") pod \"dnsmasq-dns-59b9656b65-wtfvw\" (UID: \"f6134524-f05a-43be-b80d-a862f3e7b390\") " pod="openstack/dnsmasq-dns-59b9656b65-wtfvw" Oct 08 18:30:16 crc kubenswrapper[4988]: I1008 18:30:16.224284 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jpv87\" (UniqueName: \"kubernetes.io/projected/f6134524-f05a-43be-b80d-a862f3e7b390-kube-api-access-jpv87\") pod \"dnsmasq-dns-59b9656b65-wtfvw\" (UID: \"f6134524-f05a-43be-b80d-a862f3e7b390\") " pod="openstack/dnsmasq-dns-59b9656b65-wtfvw" Oct 08 18:30:16 crc kubenswrapper[4988]: I1008 18:30:16.224421 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f6134524-f05a-43be-b80d-a862f3e7b390-config\") pod \"dnsmasq-dns-59b9656b65-wtfvw\" (UID: \"f6134524-f05a-43be-b80d-a862f3e7b390\") " pod="openstack/dnsmasq-dns-59b9656b65-wtfvw" Oct 08 18:30:16 crc kubenswrapper[4988]: I1008 18:30:16.224723 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f6134524-f05a-43be-b80d-a862f3e7b390-dns-svc\") pod \"dnsmasq-dns-59b9656b65-wtfvw\" (UID: \"f6134524-f05a-43be-b80d-a862f3e7b390\") " pod="openstack/dnsmasq-dns-59b9656b65-wtfvw" Oct 08 18:30:16 crc kubenswrapper[4988]: I1008 18:30:16.224929 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-proxy-7b7cf6bf4f-s6vrs" podStartSLOduration=8.224914993 podStartE2EDuration="8.224914993s" podCreationTimestamp="2025-10-08 18:30:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 18:30:16.214563587 +0000 UTC m=+1161.664406347" watchObservedRunningTime="2025-10-08 18:30:16.224914993 +0000 UTC m=+1161.674757783" Oct 08 18:30:16 crc kubenswrapper[4988]: I1008 18:30:16.225304 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/f6134524-f05a-43be-b80d-a862f3e7b390-dns-swift-storage-0\") pod \"dnsmasq-dns-59b9656b65-wtfvw\" (UID: \"f6134524-f05a-43be-b80d-a862f3e7b390\") " pod="openstack/dnsmasq-dns-59b9656b65-wtfvw" Oct 08 18:30:16 crc kubenswrapper[4988]: I1008 18:30:16.225895 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f6134524-f05a-43be-b80d-a862f3e7b390-config\") pod \"dnsmasq-dns-59b9656b65-wtfvw\" (UID: \"f6134524-f05a-43be-b80d-a862f3e7b390\") " pod="openstack/dnsmasq-dns-59b9656b65-wtfvw" Oct 08 18:30:16 crc kubenswrapper[4988]: I1008 18:30:16.236540 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f6134524-f05a-43be-b80d-a862f3e7b390-ovsdbserver-sb\") pod \"dnsmasq-dns-59b9656b65-wtfvw\" (UID: \"f6134524-f05a-43be-b80d-a862f3e7b390\") " pod="openstack/dnsmasq-dns-59b9656b65-wtfvw" Oct 08 18:30:16 crc kubenswrapper[4988]: I1008 18:30:16.256348 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jpv87\" (UniqueName: \"kubernetes.io/projected/f6134524-f05a-43be-b80d-a862f3e7b390-kube-api-access-jpv87\") pod \"dnsmasq-dns-59b9656b65-wtfvw\" (UID: \"f6134524-f05a-43be-b80d-a862f3e7b390\") " pod="openstack/dnsmasq-dns-59b9656b65-wtfvw" Oct 08 18:30:16 crc kubenswrapper[4988]: I1008 18:30:16.270519 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Oct 08 18:30:16 crc kubenswrapper[4988]: I1008 18:30:16.272464 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 08 18:30:16 crc kubenswrapper[4988]: I1008 18:30:16.277968 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h77st\" (UniqueName: \"kubernetes.io/projected/a86ec23c-c0e4-4e88-901f-ac5e66f103b7-kube-api-access-h77st\") pod \"nova-api-db-create-nxrck\" (UID: \"a86ec23c-c0e4-4e88-901f-ac5e66f103b7\") " pod="openstack/nova-api-db-create-nxrck" Oct 08 18:30:16 crc kubenswrapper[4988]: I1008 18:30:16.279983 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Oct 08 18:30:16 crc kubenswrapper[4988]: I1008 18:30:16.323337 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Oct 08 18:30:16 crc kubenswrapper[4988]: I1008 18:30:16.326006 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nkhb9\" (UniqueName: \"kubernetes.io/projected/d4d4bedc-1012-484f-8c73-ea8e6436b685-kube-api-access-nkhb9\") pod \"cinder-api-0\" (UID: \"d4d4bedc-1012-484f-8c73-ea8e6436b685\") " pod="openstack/cinder-api-0" Oct 08 18:30:16 crc kubenswrapper[4988]: I1008 18:30:16.326068 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d4d4bedc-1012-484f-8c73-ea8e6436b685-config-data\") pod \"cinder-api-0\" (UID: \"d4d4bedc-1012-484f-8c73-ea8e6436b685\") " pod="openstack/cinder-api-0" Oct 08 18:30:16 crc kubenswrapper[4988]: I1008 18:30:16.326332 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d4d4bedc-1012-484f-8c73-ea8e6436b685-scripts\") pod \"cinder-api-0\" (UID: \"d4d4bedc-1012-484f-8c73-ea8e6436b685\") " pod="openstack/cinder-api-0" Oct 08 18:30:16 crc kubenswrapper[4988]: I1008 18:30:16.326392 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d4d4bedc-1012-484f-8c73-ea8e6436b685-logs\") pod \"cinder-api-0\" (UID: \"d4d4bedc-1012-484f-8c73-ea8e6436b685\") " pod="openstack/cinder-api-0" Oct 08 18:30:16 crc kubenswrapper[4988]: I1008 18:30:16.326418 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d4d4bedc-1012-484f-8c73-ea8e6436b685-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"d4d4bedc-1012-484f-8c73-ea8e6436b685\") " pod="openstack/cinder-api-0" Oct 08 18:30:16 crc kubenswrapper[4988]: I1008 18:30:16.326466 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sspds\" (UniqueName: \"kubernetes.io/projected/caebe15b-a16d-4e0d-9f75-9d78958e6f9a-kube-api-access-sspds\") pod \"nova-cell0-db-create-vbbnt\" (UID: \"caebe15b-a16d-4e0d-9f75-9d78958e6f9a\") " pod="openstack/nova-cell0-db-create-vbbnt" Oct 08 18:30:16 crc kubenswrapper[4988]: I1008 18:30:16.326559 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d4d4bedc-1012-484f-8c73-ea8e6436b685-config-data-custom\") pod \"cinder-api-0\" (UID: \"d4d4bedc-1012-484f-8c73-ea8e6436b685\") " pod="openstack/cinder-api-0" Oct 08 18:30:16 crc kubenswrapper[4988]: I1008 18:30:16.326641 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/d4d4bedc-1012-484f-8c73-ea8e6436b685-etc-machine-id\") pod \"cinder-api-0\" (UID: \"d4d4bedc-1012-484f-8c73-ea8e6436b685\") " pod="openstack/cinder-api-0" Oct 08 18:30:16 crc kubenswrapper[4988]: I1008 18:30:16.333595 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-db-create-bz99d"] Oct 08 18:30:16 crc kubenswrapper[4988]: I1008 18:30:16.335079 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-bz99d" Oct 08 18:30:16 crc kubenswrapper[4988]: I1008 18:30:16.337786 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-nxrck" Oct 08 18:30:16 crc kubenswrapper[4988]: I1008 18:30:16.352274 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-bz99d"] Oct 08 18:30:16 crc kubenswrapper[4988]: I1008 18:30:16.362735 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-59b9656b65-wtfvw" Oct 08 18:30:16 crc kubenswrapper[4988]: I1008 18:30:16.429298 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d4d4bedc-1012-484f-8c73-ea8e6436b685-scripts\") pod \"cinder-api-0\" (UID: \"d4d4bedc-1012-484f-8c73-ea8e6436b685\") " pod="openstack/cinder-api-0" Oct 08 18:30:16 crc kubenswrapper[4988]: I1008 18:30:16.429340 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d4d4bedc-1012-484f-8c73-ea8e6436b685-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"d4d4bedc-1012-484f-8c73-ea8e6436b685\") " pod="openstack/cinder-api-0" Oct 08 18:30:16 crc kubenswrapper[4988]: I1008 18:30:16.429361 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sspds\" (UniqueName: \"kubernetes.io/projected/caebe15b-a16d-4e0d-9f75-9d78958e6f9a-kube-api-access-sspds\") pod \"nova-cell0-db-create-vbbnt\" (UID: \"caebe15b-a16d-4e0d-9f75-9d78958e6f9a\") " pod="openstack/nova-cell0-db-create-vbbnt" Oct 08 18:30:16 crc kubenswrapper[4988]: I1008 18:30:16.429375 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d4d4bedc-1012-484f-8c73-ea8e6436b685-logs\") pod \"cinder-api-0\" (UID: \"d4d4bedc-1012-484f-8c73-ea8e6436b685\") " pod="openstack/cinder-api-0" Oct 08 18:30:16 crc kubenswrapper[4988]: I1008 18:30:16.429416 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d4d4bedc-1012-484f-8c73-ea8e6436b685-config-data-custom\") pod \"cinder-api-0\" (UID: \"d4d4bedc-1012-484f-8c73-ea8e6436b685\") " pod="openstack/cinder-api-0" Oct 08 18:30:16 crc kubenswrapper[4988]: I1008 18:30:16.429454 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/d4d4bedc-1012-484f-8c73-ea8e6436b685-etc-machine-id\") pod \"cinder-api-0\" (UID: \"d4d4bedc-1012-484f-8c73-ea8e6436b685\") " pod="openstack/cinder-api-0" Oct 08 18:30:16 crc kubenswrapper[4988]: I1008 18:30:16.429485 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7hx5p\" (UniqueName: \"kubernetes.io/projected/41f6a322-8491-4b4c-9535-4122214d14b5-kube-api-access-7hx5p\") pod \"nova-cell1-db-create-bz99d\" (UID: \"41f6a322-8491-4b4c-9535-4122214d14b5\") " pod="openstack/nova-cell1-db-create-bz99d" Oct 08 18:30:16 crc kubenswrapper[4988]: I1008 18:30:16.429523 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nkhb9\" (UniqueName: \"kubernetes.io/projected/d4d4bedc-1012-484f-8c73-ea8e6436b685-kube-api-access-nkhb9\") pod \"cinder-api-0\" (UID: \"d4d4bedc-1012-484f-8c73-ea8e6436b685\") " pod="openstack/cinder-api-0" Oct 08 18:30:16 crc kubenswrapper[4988]: I1008 18:30:16.429546 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d4d4bedc-1012-484f-8c73-ea8e6436b685-config-data\") pod \"cinder-api-0\" (UID: \"d4d4bedc-1012-484f-8c73-ea8e6436b685\") " pod="openstack/cinder-api-0" Oct 08 18:30:16 crc kubenswrapper[4988]: I1008 18:30:16.429936 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/d4d4bedc-1012-484f-8c73-ea8e6436b685-etc-machine-id\") pod \"cinder-api-0\" (UID: \"d4d4bedc-1012-484f-8c73-ea8e6436b685\") " pod="openstack/cinder-api-0" Oct 08 18:30:16 crc kubenswrapper[4988]: I1008 18:30:16.430621 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d4d4bedc-1012-484f-8c73-ea8e6436b685-logs\") pod \"cinder-api-0\" (UID: \"d4d4bedc-1012-484f-8c73-ea8e6436b685\") " pod="openstack/cinder-api-0" Oct 08 18:30:16 crc kubenswrapper[4988]: I1008 18:30:16.433215 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d4d4bedc-1012-484f-8c73-ea8e6436b685-config-data\") pod \"cinder-api-0\" (UID: \"d4d4bedc-1012-484f-8c73-ea8e6436b685\") " pod="openstack/cinder-api-0" Oct 08 18:30:16 crc kubenswrapper[4988]: I1008 18:30:16.435177 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d4d4bedc-1012-484f-8c73-ea8e6436b685-scripts\") pod \"cinder-api-0\" (UID: \"d4d4bedc-1012-484f-8c73-ea8e6436b685\") " pod="openstack/cinder-api-0" Oct 08 18:30:16 crc kubenswrapper[4988]: I1008 18:30:16.436556 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d4d4bedc-1012-484f-8c73-ea8e6436b685-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"d4d4bedc-1012-484f-8c73-ea8e6436b685\") " pod="openstack/cinder-api-0" Oct 08 18:30:16 crc kubenswrapper[4988]: I1008 18:30:16.447304 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d4d4bedc-1012-484f-8c73-ea8e6436b685-config-data-custom\") pod \"cinder-api-0\" (UID: \"d4d4bedc-1012-484f-8c73-ea8e6436b685\") " pod="openstack/cinder-api-0" Oct 08 18:30:16 crc kubenswrapper[4988]: I1008 18:30:16.453011 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nkhb9\" (UniqueName: \"kubernetes.io/projected/d4d4bedc-1012-484f-8c73-ea8e6436b685-kube-api-access-nkhb9\") pod \"cinder-api-0\" (UID: \"d4d4bedc-1012-484f-8c73-ea8e6436b685\") " pod="openstack/cinder-api-0" Oct 08 18:30:16 crc kubenswrapper[4988]: I1008 18:30:16.478404 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sspds\" (UniqueName: \"kubernetes.io/projected/caebe15b-a16d-4e0d-9f75-9d78958e6f9a-kube-api-access-sspds\") pod \"nova-cell0-db-create-vbbnt\" (UID: \"caebe15b-a16d-4e0d-9f75-9d78958e6f9a\") " pod="openstack/nova-cell0-db-create-vbbnt" Oct 08 18:30:16 crc kubenswrapper[4988]: I1008 18:30:16.508376 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-vbbnt" Oct 08 18:30:16 crc kubenswrapper[4988]: I1008 18:30:16.533040 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7hx5p\" (UniqueName: \"kubernetes.io/projected/41f6a322-8491-4b4c-9535-4122214d14b5-kube-api-access-7hx5p\") pod \"nova-cell1-db-create-bz99d\" (UID: \"41f6a322-8491-4b4c-9535-4122214d14b5\") " pod="openstack/nova-cell1-db-create-bz99d" Oct 08 18:30:16 crc kubenswrapper[4988]: I1008 18:30:16.589364 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7hx5p\" (UniqueName: \"kubernetes.io/projected/41f6a322-8491-4b4c-9535-4122214d14b5-kube-api-access-7hx5p\") pod \"nova-cell1-db-create-bz99d\" (UID: \"41f6a322-8491-4b4c-9535-4122214d14b5\") " pod="openstack/nova-cell1-db-create-bz99d" Oct 08 18:30:16 crc kubenswrapper[4988]: I1008 18:30:16.660236 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 08 18:30:16 crc kubenswrapper[4988]: I1008 18:30:16.665262 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 08 18:30:16 crc kubenswrapper[4988]: I1008 18:30:16.690335 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-bz99d" Oct 08 18:30:16 crc kubenswrapper[4988]: I1008 18:30:16.787623 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 08 18:30:16 crc kubenswrapper[4988]: W1008 18:30:16.807225 4988 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod48acdfd7_797c_4545_9dd5_a9087ec4a525.slice/crio-55fd61079ab74343b42def895b389ce2a94d829d4515de632dedb5b8eaea5519 WatchSource:0}: Error finding container 55fd61079ab74343b42def895b389ce2a94d829d4515de632dedb5b8eaea5519: Status 404 returned error can't find the container with id 55fd61079ab74343b42def895b389ce2a94d829d4515de632dedb5b8eaea5519 Oct 08 18:30:16 crc kubenswrapper[4988]: I1008 18:30:16.972268 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-59b9656b65-wtfvw"] Oct 08 18:30:16 crc kubenswrapper[4988]: I1008 18:30:16.984223 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-nxrck"] Oct 08 18:30:16 crc kubenswrapper[4988]: W1008 18:30:16.986765 4988 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda86ec23c_c0e4_4e88_901f_ac5e66f103b7.slice/crio-f51abca3558eed0ad64832a045a5d1c167b06bc212a918c976ef5cfb515482d7 WatchSource:0}: Error finding container f51abca3558eed0ad64832a045a5d1c167b06bc212a918c976ef5cfb515482d7: Status 404 returned error can't find the container with id f51abca3558eed0ad64832a045a5d1c167b06bc212a918c976ef5cfb515482d7 Oct 08 18:30:16 crc kubenswrapper[4988]: W1008 18:30:16.989656 4988 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf6134524_f05a_43be_b80d_a862f3e7b390.slice/crio-0711b6e67a4760649ffaf998e791bdad1d1c4995cfb81ece8dea0d05efb3d97e WatchSource:0}: Error finding container 0711b6e67a4760649ffaf998e791bdad1d1c4995cfb81ece8dea0d05efb3d97e: Status 404 returned error can't find the container with id 0711b6e67a4760649ffaf998e791bdad1d1c4995cfb81ece8dea0d05efb3d97e Oct 08 18:30:17 crc kubenswrapper[4988]: I1008 18:30:17.381493 4988 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9e8e39d2-57c8-4ea6-9c83-dcb8d75274cd" path="/var/lib/kubelet/pods/9e8e39d2-57c8-4ea6-9c83-dcb8d75274cd/volumes" Oct 08 18:30:17 crc kubenswrapper[4988]: I1008 18:30:17.398705 4988 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d9f52cd3-500f-45b9-9f83-8e8244e511d6" path="/var/lib/kubelet/pods/d9f52cd3-500f-45b9-9f83-8e8244e511d6/volumes" Oct 08 18:30:17 crc kubenswrapper[4988]: I1008 18:30:17.399334 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"48acdfd7-797c-4545-9dd5-a9087ec4a525","Type":"ContainerStarted","Data":"55fd61079ab74343b42def895b389ce2a94d829d4515de632dedb5b8eaea5519"} Oct 08 18:30:17 crc kubenswrapper[4988]: I1008 18:30:17.399357 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"dd22b6ca-9c02-46f8-b202-e9634e79261e","Type":"ContainerStarted","Data":"37996652b3268f27bc8cf8ff09856cfb3e863ab4f12744f45d0fc8f9c03ecc0f"} Oct 08 18:30:17 crc kubenswrapper[4988]: I1008 18:30:17.399370 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-vbbnt"] Oct 08 18:30:17 crc kubenswrapper[4988]: I1008 18:30:17.399403 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Oct 08 18:30:17 crc kubenswrapper[4988]: I1008 18:30:17.399413 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-bz99d"] Oct 08 18:30:17 crc kubenswrapper[4988]: I1008 18:30:17.400622 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-nxrck" event={"ID":"a86ec23c-c0e4-4e88-901f-ac5e66f103b7","Type":"ContainerStarted","Data":"d8e23c92099ddad91fb908ab60379b8040c04ad905cb12b8b115b0fe8da69e06"} Oct 08 18:30:17 crc kubenswrapper[4988]: I1008 18:30:17.400666 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-nxrck" event={"ID":"a86ec23c-c0e4-4e88-901f-ac5e66f103b7","Type":"ContainerStarted","Data":"f51abca3558eed0ad64832a045a5d1c167b06bc212a918c976ef5cfb515482d7"} Oct 08 18:30:17 crc kubenswrapper[4988]: I1008 18:30:17.422500 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-db-create-nxrck" podStartSLOduration=2.422478259 podStartE2EDuration="2.422478259s" podCreationTimestamp="2025-10-08 18:30:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 18:30:17.416211971 +0000 UTC m=+1162.866054761" watchObservedRunningTime="2025-10-08 18:30:17.422478259 +0000 UTC m=+1162.872321029" Oct 08 18:30:17 crc kubenswrapper[4988]: I1008 18:30:17.432614 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-59b9656b65-wtfvw" event={"ID":"f6134524-f05a-43be-b80d-a862f3e7b390","Type":"ContainerStarted","Data":"0711b6e67a4760649ffaf998e791bdad1d1c4995cfb81ece8dea0d05efb3d97e"} Oct 08 18:30:17 crc kubenswrapper[4988]: I1008 18:30:17.432686 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/swift-proxy-7b7cf6bf4f-s6vrs" Oct 08 18:30:17 crc kubenswrapper[4988]: I1008 18:30:17.432768 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/swift-proxy-7b7cf6bf4f-s6vrs" Oct 08 18:30:18 crc kubenswrapper[4988]: I1008 18:30:18.447878 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"dd22b6ca-9c02-46f8-b202-e9634e79261e","Type":"ContainerStarted","Data":"e7f72cbf44e45568e6e9b978548b5543f7f049dadbaa5ec5eeba3c0abaf598a5"} Oct 08 18:30:18 crc kubenswrapper[4988]: I1008 18:30:18.454109 4988 generic.go:334] "Generic (PLEG): container finished" podID="a86ec23c-c0e4-4e88-901f-ac5e66f103b7" containerID="d8e23c92099ddad91fb908ab60379b8040c04ad905cb12b8b115b0fe8da69e06" exitCode=0 Oct 08 18:30:18 crc kubenswrapper[4988]: I1008 18:30:18.454257 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-nxrck" event={"ID":"a86ec23c-c0e4-4e88-901f-ac5e66f103b7","Type":"ContainerDied","Data":"d8e23c92099ddad91fb908ab60379b8040c04ad905cb12b8b115b0fe8da69e06"} Oct 08 18:30:18 crc kubenswrapper[4988]: I1008 18:30:18.457796 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-vbbnt" event={"ID":"caebe15b-a16d-4e0d-9f75-9d78958e6f9a","Type":"ContainerStarted","Data":"22daa187f81f4dead5f47bc9340c19470b641fcc7b4541232f44df8458409f63"} Oct 08 18:30:18 crc kubenswrapper[4988]: I1008 18:30:18.457828 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-vbbnt" event={"ID":"caebe15b-a16d-4e0d-9f75-9d78958e6f9a","Type":"ContainerStarted","Data":"febaf0ebee58e23a1f08268b2babac507b832f8bbf7116fefb20bc38577b7e6e"} Oct 08 18:30:18 crc kubenswrapper[4988]: I1008 18:30:18.470145 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"d4d4bedc-1012-484f-8c73-ea8e6436b685","Type":"ContainerStarted","Data":"c5ccaca5a43181b9f64a6893e06d0aae10774e78192795cd53d7e1a83869f39f"} Oct 08 18:30:18 crc kubenswrapper[4988]: I1008 18:30:18.470189 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"d4d4bedc-1012-484f-8c73-ea8e6436b685","Type":"ContainerStarted","Data":"2195418014a8f15e7544775be552b1cc6eb81e475aa42a22dcde7498f5b3d6e9"} Oct 08 18:30:18 crc kubenswrapper[4988]: I1008 18:30:18.474344 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-bz99d" event={"ID":"41f6a322-8491-4b4c-9535-4122214d14b5","Type":"ContainerStarted","Data":"bd8657ab54dc617e4190cb1a864b28e10b3355e63b38eb13c2d1d5db69eb55ac"} Oct 08 18:30:18 crc kubenswrapper[4988]: I1008 18:30:18.474423 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-bz99d" event={"ID":"41f6a322-8491-4b4c-9535-4122214d14b5","Type":"ContainerStarted","Data":"77ff76c7c0bf5f90c9c5edc55ba1913daa2e689fbb6c47fb3c0855a642033ec4"} Oct 08 18:30:18 crc kubenswrapper[4988]: I1008 18:30:18.491525 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-db-create-vbbnt" podStartSLOduration=2.491505944 podStartE2EDuration="2.491505944s" podCreationTimestamp="2025-10-08 18:30:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 18:30:18.483848633 +0000 UTC m=+1163.933691403" watchObservedRunningTime="2025-10-08 18:30:18.491505944 +0000 UTC m=+1163.941348724" Oct 08 18:30:18 crc kubenswrapper[4988]: I1008 18:30:18.493599 4988 generic.go:334] "Generic (PLEG): container finished" podID="f6134524-f05a-43be-b80d-a862f3e7b390" containerID="ede0a83a2428dcb8b487d4ec98b6590b8bc67faf808232a9a9cecb4b352d76b6" exitCode=0 Oct 08 18:30:18 crc kubenswrapper[4988]: I1008 18:30:18.493781 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-59b9656b65-wtfvw" event={"ID":"f6134524-f05a-43be-b80d-a862f3e7b390","Type":"ContainerDied","Data":"ede0a83a2428dcb8b487d4ec98b6590b8bc67faf808232a9a9cecb4b352d76b6"} Oct 08 18:30:18 crc kubenswrapper[4988]: I1008 18:30:18.500367 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-db-create-bz99d" podStartSLOduration=2.500350713 podStartE2EDuration="2.500350713s" podCreationTimestamp="2025-10-08 18:30:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 18:30:18.498825865 +0000 UTC m=+1163.948668645" watchObservedRunningTime="2025-10-08 18:30:18.500350713 +0000 UTC m=+1163.950193493" Oct 08 18:30:18 crc kubenswrapper[4988]: I1008 18:30:18.848429 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Oct 08 18:30:19 crc kubenswrapper[4988]: I1008 18:30:19.512283 4988 generic.go:334] "Generic (PLEG): container finished" podID="41f6a322-8491-4b4c-9535-4122214d14b5" containerID="bd8657ab54dc617e4190cb1a864b28e10b3355e63b38eb13c2d1d5db69eb55ac" exitCode=0 Oct 08 18:30:19 crc kubenswrapper[4988]: I1008 18:30:19.512632 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-bz99d" event={"ID":"41f6a322-8491-4b4c-9535-4122214d14b5","Type":"ContainerDied","Data":"bd8657ab54dc617e4190cb1a864b28e10b3355e63b38eb13c2d1d5db69eb55ac"} Oct 08 18:30:19 crc kubenswrapper[4988]: I1008 18:30:19.515811 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-59b9656b65-wtfvw" event={"ID":"f6134524-f05a-43be-b80d-a862f3e7b390","Type":"ContainerStarted","Data":"d2cfa238fcf3427c4917a7f32e8de7b6a8c3479bf5cc18d79302827c17d4c2d4"} Oct 08 18:30:19 crc kubenswrapper[4988]: I1008 18:30:19.515932 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-59b9656b65-wtfvw" Oct 08 18:30:19 crc kubenswrapper[4988]: I1008 18:30:19.519736 4988 generic.go:334] "Generic (PLEG): container finished" podID="caebe15b-a16d-4e0d-9f75-9d78958e6f9a" containerID="22daa187f81f4dead5f47bc9340c19470b641fcc7b4541232f44df8458409f63" exitCode=0 Oct 08 18:30:19 crc kubenswrapper[4988]: I1008 18:30:19.519801 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-vbbnt" event={"ID":"caebe15b-a16d-4e0d-9f75-9d78958e6f9a","Type":"ContainerDied","Data":"22daa187f81f4dead5f47bc9340c19470b641fcc7b4541232f44df8458409f63"} Oct 08 18:30:19 crc kubenswrapper[4988]: I1008 18:30:19.521809 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"d4d4bedc-1012-484f-8c73-ea8e6436b685","Type":"ContainerStarted","Data":"c31c2d7d0c808b506aa0e58a9ae84ee5955871961f3f26b50483eb9778e51fd2"} Oct 08 18:30:19 crc kubenswrapper[4988]: I1008 18:30:19.545344 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-59b9656b65-wtfvw" podStartSLOduration=4.54532052 podStartE2EDuration="4.54532052s" podCreationTimestamp="2025-10-08 18:30:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 18:30:19.53959121 +0000 UTC m=+1164.989433990" watchObservedRunningTime="2025-10-08 18:30:19.54532052 +0000 UTC m=+1164.995163300" Oct 08 18:30:19 crc kubenswrapper[4988]: I1008 18:30:19.993284 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-nxrck" Oct 08 18:30:20 crc kubenswrapper[4988]: I1008 18:30:20.072795 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h77st\" (UniqueName: \"kubernetes.io/projected/a86ec23c-c0e4-4e88-901f-ac5e66f103b7-kube-api-access-h77st\") pod \"a86ec23c-c0e4-4e88-901f-ac5e66f103b7\" (UID: \"a86ec23c-c0e4-4e88-901f-ac5e66f103b7\") " Oct 08 18:30:20 crc kubenswrapper[4988]: I1008 18:30:20.077567 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a86ec23c-c0e4-4e88-901f-ac5e66f103b7-kube-api-access-h77st" (OuterVolumeSpecName: "kube-api-access-h77st") pod "a86ec23c-c0e4-4e88-901f-ac5e66f103b7" (UID: "a86ec23c-c0e4-4e88-901f-ac5e66f103b7"). InnerVolumeSpecName "kube-api-access-h77st". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:30:20 crc kubenswrapper[4988]: I1008 18:30:20.175232 4988 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h77st\" (UniqueName: \"kubernetes.io/projected/a86ec23c-c0e4-4e88-901f-ac5e66f103b7-kube-api-access-h77st\") on node \"crc\" DevicePath \"\"" Oct 08 18:30:20 crc kubenswrapper[4988]: I1008 18:30:20.534264 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-nxrck" event={"ID":"a86ec23c-c0e4-4e88-901f-ac5e66f103b7","Type":"ContainerDied","Data":"f51abca3558eed0ad64832a045a5d1c167b06bc212a918c976ef5cfb515482d7"} Oct 08 18:30:20 crc kubenswrapper[4988]: I1008 18:30:20.534582 4988 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f51abca3558eed0ad64832a045a5d1c167b06bc212a918c976ef5cfb515482d7" Oct 08 18:30:20 crc kubenswrapper[4988]: I1008 18:30:20.534720 4988 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="d4d4bedc-1012-484f-8c73-ea8e6436b685" containerName="cinder-api-log" containerID="cri-o://c5ccaca5a43181b9f64a6893e06d0aae10774e78192795cd53d7e1a83869f39f" gracePeriod=30 Oct 08 18:30:20 crc kubenswrapper[4988]: I1008 18:30:20.535128 4988 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="d4d4bedc-1012-484f-8c73-ea8e6436b685" containerName="cinder-api" containerID="cri-o://c31c2d7d0c808b506aa0e58a9ae84ee5955871961f3f26b50483eb9778e51fd2" gracePeriod=30 Oct 08 18:30:20 crc kubenswrapper[4988]: I1008 18:30:20.535320 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-nxrck" Oct 08 18:30:20 crc kubenswrapper[4988]: I1008 18:30:20.561118 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=4.561097138 podStartE2EDuration="4.561097138s" podCreationTimestamp="2025-10-08 18:30:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 18:30:20.555909124 +0000 UTC m=+1166.005751904" watchObservedRunningTime="2025-10-08 18:30:20.561097138 +0000 UTC m=+1166.010939908" Oct 08 18:30:21 crc kubenswrapper[4988]: I1008 18:30:21.069654 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-vbbnt" Oct 08 18:30:21 crc kubenswrapper[4988]: I1008 18:30:21.081092 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-bz99d" Oct 08 18:30:21 crc kubenswrapper[4988]: I1008 18:30:21.098667 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sspds\" (UniqueName: \"kubernetes.io/projected/caebe15b-a16d-4e0d-9f75-9d78958e6f9a-kube-api-access-sspds\") pod \"caebe15b-a16d-4e0d-9f75-9d78958e6f9a\" (UID: \"caebe15b-a16d-4e0d-9f75-9d78958e6f9a\") " Oct 08 18:30:21 crc kubenswrapper[4988]: I1008 18:30:21.099280 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7hx5p\" (UniqueName: \"kubernetes.io/projected/41f6a322-8491-4b4c-9535-4122214d14b5-kube-api-access-7hx5p\") pod \"41f6a322-8491-4b4c-9535-4122214d14b5\" (UID: \"41f6a322-8491-4b4c-9535-4122214d14b5\") " Oct 08 18:30:21 crc kubenswrapper[4988]: I1008 18:30:21.119888 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/caebe15b-a16d-4e0d-9f75-9d78958e6f9a-kube-api-access-sspds" (OuterVolumeSpecName: "kube-api-access-sspds") pod "caebe15b-a16d-4e0d-9f75-9d78958e6f9a" (UID: "caebe15b-a16d-4e0d-9f75-9d78958e6f9a"). InnerVolumeSpecName "kube-api-access-sspds". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:30:21 crc kubenswrapper[4988]: I1008 18:30:21.134602 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/41f6a322-8491-4b4c-9535-4122214d14b5-kube-api-access-7hx5p" (OuterVolumeSpecName: "kube-api-access-7hx5p") pod "41f6a322-8491-4b4c-9535-4122214d14b5" (UID: "41f6a322-8491-4b4c-9535-4122214d14b5"). InnerVolumeSpecName "kube-api-access-7hx5p". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:30:21 crc kubenswrapper[4988]: I1008 18:30:21.202292 4988 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sspds\" (UniqueName: \"kubernetes.io/projected/caebe15b-a16d-4e0d-9f75-9d78958e6f9a-kube-api-access-sspds\") on node \"crc\" DevicePath \"\"" Oct 08 18:30:21 crc kubenswrapper[4988]: I1008 18:30:21.202324 4988 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7hx5p\" (UniqueName: \"kubernetes.io/projected/41f6a322-8491-4b4c-9535-4122214d14b5-kube-api-access-7hx5p\") on node \"crc\" DevicePath \"\"" Oct 08 18:30:21 crc kubenswrapper[4988]: I1008 18:30:21.313650 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 08 18:30:21 crc kubenswrapper[4988]: I1008 18:30:21.404712 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d4d4bedc-1012-484f-8c73-ea8e6436b685-config-data-custom\") pod \"d4d4bedc-1012-484f-8c73-ea8e6436b685\" (UID: \"d4d4bedc-1012-484f-8c73-ea8e6436b685\") " Oct 08 18:30:21 crc kubenswrapper[4988]: I1008 18:30:21.405113 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nkhb9\" (UniqueName: \"kubernetes.io/projected/d4d4bedc-1012-484f-8c73-ea8e6436b685-kube-api-access-nkhb9\") pod \"d4d4bedc-1012-484f-8c73-ea8e6436b685\" (UID: \"d4d4bedc-1012-484f-8c73-ea8e6436b685\") " Oct 08 18:30:21 crc kubenswrapper[4988]: I1008 18:30:21.405178 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d4d4bedc-1012-484f-8c73-ea8e6436b685-combined-ca-bundle\") pod \"d4d4bedc-1012-484f-8c73-ea8e6436b685\" (UID: \"d4d4bedc-1012-484f-8c73-ea8e6436b685\") " Oct 08 18:30:21 crc kubenswrapper[4988]: I1008 18:30:21.405282 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d4d4bedc-1012-484f-8c73-ea8e6436b685-config-data\") pod \"d4d4bedc-1012-484f-8c73-ea8e6436b685\" (UID: \"d4d4bedc-1012-484f-8c73-ea8e6436b685\") " Oct 08 18:30:21 crc kubenswrapper[4988]: I1008 18:30:21.405367 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d4d4bedc-1012-484f-8c73-ea8e6436b685-logs\") pod \"d4d4bedc-1012-484f-8c73-ea8e6436b685\" (UID: \"d4d4bedc-1012-484f-8c73-ea8e6436b685\") " Oct 08 18:30:21 crc kubenswrapper[4988]: I1008 18:30:21.405448 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/d4d4bedc-1012-484f-8c73-ea8e6436b685-etc-machine-id\") pod \"d4d4bedc-1012-484f-8c73-ea8e6436b685\" (UID: \"d4d4bedc-1012-484f-8c73-ea8e6436b685\") " Oct 08 18:30:21 crc kubenswrapper[4988]: I1008 18:30:21.405483 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d4d4bedc-1012-484f-8c73-ea8e6436b685-scripts\") pod \"d4d4bedc-1012-484f-8c73-ea8e6436b685\" (UID: \"d4d4bedc-1012-484f-8c73-ea8e6436b685\") " Oct 08 18:30:21 crc kubenswrapper[4988]: I1008 18:30:21.406446 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d4d4bedc-1012-484f-8c73-ea8e6436b685-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "d4d4bedc-1012-484f-8c73-ea8e6436b685" (UID: "d4d4bedc-1012-484f-8c73-ea8e6436b685"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 08 18:30:21 crc kubenswrapper[4988]: I1008 18:30:21.406731 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d4d4bedc-1012-484f-8c73-ea8e6436b685-logs" (OuterVolumeSpecName: "logs") pod "d4d4bedc-1012-484f-8c73-ea8e6436b685" (UID: "d4d4bedc-1012-484f-8c73-ea8e6436b685"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 18:30:21 crc kubenswrapper[4988]: I1008 18:30:21.419613 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d4d4bedc-1012-484f-8c73-ea8e6436b685-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "d4d4bedc-1012-484f-8c73-ea8e6436b685" (UID: "d4d4bedc-1012-484f-8c73-ea8e6436b685"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:30:21 crc kubenswrapper[4988]: I1008 18:30:21.421519 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d4d4bedc-1012-484f-8c73-ea8e6436b685-scripts" (OuterVolumeSpecName: "scripts") pod "d4d4bedc-1012-484f-8c73-ea8e6436b685" (UID: "d4d4bedc-1012-484f-8c73-ea8e6436b685"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:30:21 crc kubenswrapper[4988]: I1008 18:30:21.421609 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d4d4bedc-1012-484f-8c73-ea8e6436b685-kube-api-access-nkhb9" (OuterVolumeSpecName: "kube-api-access-nkhb9") pod "d4d4bedc-1012-484f-8c73-ea8e6436b685" (UID: "d4d4bedc-1012-484f-8c73-ea8e6436b685"). InnerVolumeSpecName "kube-api-access-nkhb9". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:30:21 crc kubenswrapper[4988]: I1008 18:30:21.451466 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d4d4bedc-1012-484f-8c73-ea8e6436b685-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d4d4bedc-1012-484f-8c73-ea8e6436b685" (UID: "d4d4bedc-1012-484f-8c73-ea8e6436b685"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:30:21 crc kubenswrapper[4988]: I1008 18:30:21.470313 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d4d4bedc-1012-484f-8c73-ea8e6436b685-config-data" (OuterVolumeSpecName: "config-data") pod "d4d4bedc-1012-484f-8c73-ea8e6436b685" (UID: "d4d4bedc-1012-484f-8c73-ea8e6436b685"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:30:21 crc kubenswrapper[4988]: I1008 18:30:21.507945 4988 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/d4d4bedc-1012-484f-8c73-ea8e6436b685-etc-machine-id\") on node \"crc\" DevicePath \"\"" Oct 08 18:30:21 crc kubenswrapper[4988]: I1008 18:30:21.507993 4988 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d4d4bedc-1012-484f-8c73-ea8e6436b685-scripts\") on node \"crc\" DevicePath \"\"" Oct 08 18:30:21 crc kubenswrapper[4988]: I1008 18:30:21.508011 4988 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d4d4bedc-1012-484f-8c73-ea8e6436b685-config-data-custom\") on node \"crc\" DevicePath \"\"" Oct 08 18:30:21 crc kubenswrapper[4988]: I1008 18:30:21.508028 4988 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nkhb9\" (UniqueName: \"kubernetes.io/projected/d4d4bedc-1012-484f-8c73-ea8e6436b685-kube-api-access-nkhb9\") on node \"crc\" DevicePath \"\"" Oct 08 18:30:21 crc kubenswrapper[4988]: I1008 18:30:21.508042 4988 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d4d4bedc-1012-484f-8c73-ea8e6436b685-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 18:30:21 crc kubenswrapper[4988]: I1008 18:30:21.508053 4988 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d4d4bedc-1012-484f-8c73-ea8e6436b685-config-data\") on node \"crc\" DevicePath \"\"" Oct 08 18:30:21 crc kubenswrapper[4988]: I1008 18:30:21.508063 4988 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d4d4bedc-1012-484f-8c73-ea8e6436b685-logs\") on node \"crc\" DevicePath \"\"" Oct 08 18:30:21 crc kubenswrapper[4988]: I1008 18:30:21.560274 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"dd22b6ca-9c02-46f8-b202-e9634e79261e","Type":"ContainerStarted","Data":"f68fab85a12f4d29f55021002a6a561eb9662e3a85deffc9606602abbf47a9dc"} Oct 08 18:30:21 crc kubenswrapper[4988]: I1008 18:30:21.562270 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-vbbnt" event={"ID":"caebe15b-a16d-4e0d-9f75-9d78958e6f9a","Type":"ContainerDied","Data":"febaf0ebee58e23a1f08268b2babac507b832f8bbf7116fefb20bc38577b7e6e"} Oct 08 18:30:21 crc kubenswrapper[4988]: I1008 18:30:21.562321 4988 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="febaf0ebee58e23a1f08268b2babac507b832f8bbf7116fefb20bc38577b7e6e" Oct 08 18:30:21 crc kubenswrapper[4988]: I1008 18:30:21.562290 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-vbbnt" Oct 08 18:30:21 crc kubenswrapper[4988]: I1008 18:30:21.564318 4988 generic.go:334] "Generic (PLEG): container finished" podID="d4d4bedc-1012-484f-8c73-ea8e6436b685" containerID="c31c2d7d0c808b506aa0e58a9ae84ee5955871961f3f26b50483eb9778e51fd2" exitCode=0 Oct 08 18:30:21 crc kubenswrapper[4988]: I1008 18:30:21.564340 4988 generic.go:334] "Generic (PLEG): container finished" podID="d4d4bedc-1012-484f-8c73-ea8e6436b685" containerID="c5ccaca5a43181b9f64a6893e06d0aae10774e78192795cd53d7e1a83869f39f" exitCode=143 Oct 08 18:30:21 crc kubenswrapper[4988]: I1008 18:30:21.564374 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"d4d4bedc-1012-484f-8c73-ea8e6436b685","Type":"ContainerDied","Data":"c31c2d7d0c808b506aa0e58a9ae84ee5955871961f3f26b50483eb9778e51fd2"} Oct 08 18:30:21 crc kubenswrapper[4988]: I1008 18:30:21.564412 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"d4d4bedc-1012-484f-8c73-ea8e6436b685","Type":"ContainerDied","Data":"c5ccaca5a43181b9f64a6893e06d0aae10774e78192795cd53d7e1a83869f39f"} Oct 08 18:30:21 crc kubenswrapper[4988]: I1008 18:30:21.564421 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"d4d4bedc-1012-484f-8c73-ea8e6436b685","Type":"ContainerDied","Data":"2195418014a8f15e7544775be552b1cc6eb81e475aa42a22dcde7498f5b3d6e9"} Oct 08 18:30:21 crc kubenswrapper[4988]: I1008 18:30:21.564435 4988 scope.go:117] "RemoveContainer" containerID="c31c2d7d0c808b506aa0e58a9ae84ee5955871961f3f26b50483eb9778e51fd2" Oct 08 18:30:21 crc kubenswrapper[4988]: I1008 18:30:21.564496 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 08 18:30:21 crc kubenswrapper[4988]: I1008 18:30:21.567684 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-bz99d" event={"ID":"41f6a322-8491-4b4c-9535-4122214d14b5","Type":"ContainerDied","Data":"77ff76c7c0bf5f90c9c5edc55ba1913daa2e689fbb6c47fb3c0855a642033ec4"} Oct 08 18:30:21 crc kubenswrapper[4988]: I1008 18:30:21.567709 4988 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="77ff76c7c0bf5f90c9c5edc55ba1913daa2e689fbb6c47fb3c0855a642033ec4" Oct 08 18:30:21 crc kubenswrapper[4988]: I1008 18:30:21.567778 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-bz99d" Oct 08 18:30:21 crc kubenswrapper[4988]: I1008 18:30:21.570887 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"48acdfd7-797c-4545-9dd5-a9087ec4a525","Type":"ContainerStarted","Data":"9fbef399ac4284687057dc2c47897eb4be0635cba1dc233289a1f03bfaa51bf3"} Oct 08 18:30:21 crc kubenswrapper[4988]: I1008 18:30:21.570947 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"48acdfd7-797c-4545-9dd5-a9087ec4a525","Type":"ContainerStarted","Data":"9b7da6032d176b505a5dd95802130a58e053c8b0dbf445c13c049d3deee7e1cf"} Oct 08 18:30:21 crc kubenswrapper[4988]: I1008 18:30:21.586643 4988 scope.go:117] "RemoveContainer" containerID="c5ccaca5a43181b9f64a6893e06d0aae10774e78192795cd53d7e1a83869f39f" Oct 08 18:30:21 crc kubenswrapper[4988]: I1008 18:30:21.596287 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=3.738852058 podStartE2EDuration="6.596265886s" podCreationTimestamp="2025-10-08 18:30:15 +0000 UTC" firstStartedPulling="2025-10-08 18:30:16.830197136 +0000 UTC m=+1162.280039906" lastFinishedPulling="2025-10-08 18:30:19.687610964 +0000 UTC m=+1165.137453734" observedRunningTime="2025-10-08 18:30:21.593198869 +0000 UTC m=+1167.043041639" watchObservedRunningTime="2025-10-08 18:30:21.596265886 +0000 UTC m=+1167.046108646" Oct 08 18:30:21 crc kubenswrapper[4988]: I1008 18:30:21.612804 4988 scope.go:117] "RemoveContainer" containerID="c31c2d7d0c808b506aa0e58a9ae84ee5955871961f3f26b50483eb9778e51fd2" Oct 08 18:30:21 crc kubenswrapper[4988]: E1008 18:30:21.613251 4988 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c31c2d7d0c808b506aa0e58a9ae84ee5955871961f3f26b50483eb9778e51fd2\": container with ID starting with c31c2d7d0c808b506aa0e58a9ae84ee5955871961f3f26b50483eb9778e51fd2 not found: ID does not exist" containerID="c31c2d7d0c808b506aa0e58a9ae84ee5955871961f3f26b50483eb9778e51fd2" Oct 08 18:30:21 crc kubenswrapper[4988]: I1008 18:30:21.613291 4988 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c31c2d7d0c808b506aa0e58a9ae84ee5955871961f3f26b50483eb9778e51fd2"} err="failed to get container status \"c31c2d7d0c808b506aa0e58a9ae84ee5955871961f3f26b50483eb9778e51fd2\": rpc error: code = NotFound desc = could not find container \"c31c2d7d0c808b506aa0e58a9ae84ee5955871961f3f26b50483eb9778e51fd2\": container with ID starting with c31c2d7d0c808b506aa0e58a9ae84ee5955871961f3f26b50483eb9778e51fd2 not found: ID does not exist" Oct 08 18:30:21 crc kubenswrapper[4988]: I1008 18:30:21.613316 4988 scope.go:117] "RemoveContainer" containerID="c5ccaca5a43181b9f64a6893e06d0aae10774e78192795cd53d7e1a83869f39f" Oct 08 18:30:21 crc kubenswrapper[4988]: E1008 18:30:21.613710 4988 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c5ccaca5a43181b9f64a6893e06d0aae10774e78192795cd53d7e1a83869f39f\": container with ID starting with c5ccaca5a43181b9f64a6893e06d0aae10774e78192795cd53d7e1a83869f39f not found: ID does not exist" containerID="c5ccaca5a43181b9f64a6893e06d0aae10774e78192795cd53d7e1a83869f39f" Oct 08 18:30:21 crc kubenswrapper[4988]: I1008 18:30:21.613736 4988 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c5ccaca5a43181b9f64a6893e06d0aae10774e78192795cd53d7e1a83869f39f"} err="failed to get container status \"c5ccaca5a43181b9f64a6893e06d0aae10774e78192795cd53d7e1a83869f39f\": rpc error: code = NotFound desc = could not find container \"c5ccaca5a43181b9f64a6893e06d0aae10774e78192795cd53d7e1a83869f39f\": container with ID starting with c5ccaca5a43181b9f64a6893e06d0aae10774e78192795cd53d7e1a83869f39f not found: ID does not exist" Oct 08 18:30:21 crc kubenswrapper[4988]: I1008 18:30:21.613758 4988 scope.go:117] "RemoveContainer" containerID="c31c2d7d0c808b506aa0e58a9ae84ee5955871961f3f26b50483eb9778e51fd2" Oct 08 18:30:21 crc kubenswrapper[4988]: I1008 18:30:21.614383 4988 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c31c2d7d0c808b506aa0e58a9ae84ee5955871961f3f26b50483eb9778e51fd2"} err="failed to get container status \"c31c2d7d0c808b506aa0e58a9ae84ee5955871961f3f26b50483eb9778e51fd2\": rpc error: code = NotFound desc = could not find container \"c31c2d7d0c808b506aa0e58a9ae84ee5955871961f3f26b50483eb9778e51fd2\": container with ID starting with c31c2d7d0c808b506aa0e58a9ae84ee5955871961f3f26b50483eb9778e51fd2 not found: ID does not exist" Oct 08 18:30:21 crc kubenswrapper[4988]: I1008 18:30:21.614415 4988 scope.go:117] "RemoveContainer" containerID="c5ccaca5a43181b9f64a6893e06d0aae10774e78192795cd53d7e1a83869f39f" Oct 08 18:30:21 crc kubenswrapper[4988]: I1008 18:30:21.614623 4988 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c5ccaca5a43181b9f64a6893e06d0aae10774e78192795cd53d7e1a83869f39f"} err="failed to get container status \"c5ccaca5a43181b9f64a6893e06d0aae10774e78192795cd53d7e1a83869f39f\": rpc error: code = NotFound desc = could not find container \"c5ccaca5a43181b9f64a6893e06d0aae10774e78192795cd53d7e1a83869f39f\": container with ID starting with c5ccaca5a43181b9f64a6893e06d0aae10774e78192795cd53d7e1a83869f39f not found: ID does not exist" Oct 08 18:30:21 crc kubenswrapper[4988]: I1008 18:30:21.620458 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Oct 08 18:30:21 crc kubenswrapper[4988]: I1008 18:30:21.628230 4988 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-api-0"] Oct 08 18:30:21 crc kubenswrapper[4988]: I1008 18:30:21.648113 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Oct 08 18:30:21 crc kubenswrapper[4988]: E1008 18:30:21.648601 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d4d4bedc-1012-484f-8c73-ea8e6436b685" containerName="cinder-api-log" Oct 08 18:30:21 crc kubenswrapper[4988]: I1008 18:30:21.648616 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="d4d4bedc-1012-484f-8c73-ea8e6436b685" containerName="cinder-api-log" Oct 08 18:30:21 crc kubenswrapper[4988]: E1008 18:30:21.648629 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="caebe15b-a16d-4e0d-9f75-9d78958e6f9a" containerName="mariadb-database-create" Oct 08 18:30:21 crc kubenswrapper[4988]: I1008 18:30:21.648635 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="caebe15b-a16d-4e0d-9f75-9d78958e6f9a" containerName="mariadb-database-create" Oct 08 18:30:21 crc kubenswrapper[4988]: E1008 18:30:21.648660 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a86ec23c-c0e4-4e88-901f-ac5e66f103b7" containerName="mariadb-database-create" Oct 08 18:30:21 crc kubenswrapper[4988]: I1008 18:30:21.648666 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="a86ec23c-c0e4-4e88-901f-ac5e66f103b7" containerName="mariadb-database-create" Oct 08 18:30:21 crc kubenswrapper[4988]: E1008 18:30:21.648679 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="41f6a322-8491-4b4c-9535-4122214d14b5" containerName="mariadb-database-create" Oct 08 18:30:21 crc kubenswrapper[4988]: I1008 18:30:21.648685 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="41f6a322-8491-4b4c-9535-4122214d14b5" containerName="mariadb-database-create" Oct 08 18:30:21 crc kubenswrapper[4988]: E1008 18:30:21.648696 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d4d4bedc-1012-484f-8c73-ea8e6436b685" containerName="cinder-api" Oct 08 18:30:21 crc kubenswrapper[4988]: I1008 18:30:21.648702 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="d4d4bedc-1012-484f-8c73-ea8e6436b685" containerName="cinder-api" Oct 08 18:30:21 crc kubenswrapper[4988]: I1008 18:30:21.648886 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="d4d4bedc-1012-484f-8c73-ea8e6436b685" containerName="cinder-api-log" Oct 08 18:30:21 crc kubenswrapper[4988]: I1008 18:30:21.648901 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="a86ec23c-c0e4-4e88-901f-ac5e66f103b7" containerName="mariadb-database-create" Oct 08 18:30:21 crc kubenswrapper[4988]: I1008 18:30:21.648911 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="caebe15b-a16d-4e0d-9f75-9d78958e6f9a" containerName="mariadb-database-create" Oct 08 18:30:21 crc kubenswrapper[4988]: I1008 18:30:21.648920 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="d4d4bedc-1012-484f-8c73-ea8e6436b685" containerName="cinder-api" Oct 08 18:30:21 crc kubenswrapper[4988]: I1008 18:30:21.648933 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="41f6a322-8491-4b4c-9535-4122214d14b5" containerName="mariadb-database-create" Oct 08 18:30:21 crc kubenswrapper[4988]: I1008 18:30:21.649867 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 08 18:30:21 crc kubenswrapper[4988]: I1008 18:30:21.653289 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-public-svc" Oct 08 18:30:21 crc kubenswrapper[4988]: I1008 18:30:21.653488 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-internal-svc" Oct 08 18:30:21 crc kubenswrapper[4988]: I1008 18:30:21.656915 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Oct 08 18:30:21 crc kubenswrapper[4988]: I1008 18:30:21.662674 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Oct 08 18:30:21 crc kubenswrapper[4988]: I1008 18:30:21.713314 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b6krt\" (UniqueName: \"kubernetes.io/projected/99d77eed-40f0-4879-be96-abea3a807cd9-kube-api-access-b6krt\") pod \"cinder-api-0\" (UID: \"99d77eed-40f0-4879-be96-abea3a807cd9\") " pod="openstack/cinder-api-0" Oct 08 18:30:21 crc kubenswrapper[4988]: I1008 18:30:21.713364 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/99d77eed-40f0-4879-be96-abea3a807cd9-etc-machine-id\") pod \"cinder-api-0\" (UID: \"99d77eed-40f0-4879-be96-abea3a807cd9\") " pod="openstack/cinder-api-0" Oct 08 18:30:21 crc kubenswrapper[4988]: I1008 18:30:21.713445 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/99d77eed-40f0-4879-be96-abea3a807cd9-config-data\") pod \"cinder-api-0\" (UID: \"99d77eed-40f0-4879-be96-abea3a807cd9\") " pod="openstack/cinder-api-0" Oct 08 18:30:21 crc kubenswrapper[4988]: I1008 18:30:21.713479 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/99d77eed-40f0-4879-be96-abea3a807cd9-scripts\") pod \"cinder-api-0\" (UID: \"99d77eed-40f0-4879-be96-abea3a807cd9\") " pod="openstack/cinder-api-0" Oct 08 18:30:21 crc kubenswrapper[4988]: I1008 18:30:21.713667 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/99d77eed-40f0-4879-be96-abea3a807cd9-logs\") pod \"cinder-api-0\" (UID: \"99d77eed-40f0-4879-be96-abea3a807cd9\") " pod="openstack/cinder-api-0" Oct 08 18:30:21 crc kubenswrapper[4988]: I1008 18:30:21.713722 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/99d77eed-40f0-4879-be96-abea3a807cd9-public-tls-certs\") pod \"cinder-api-0\" (UID: \"99d77eed-40f0-4879-be96-abea3a807cd9\") " pod="openstack/cinder-api-0" Oct 08 18:30:21 crc kubenswrapper[4988]: I1008 18:30:21.713856 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/99d77eed-40f0-4879-be96-abea3a807cd9-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"99d77eed-40f0-4879-be96-abea3a807cd9\") " pod="openstack/cinder-api-0" Oct 08 18:30:21 crc kubenswrapper[4988]: I1008 18:30:21.713922 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/99d77eed-40f0-4879-be96-abea3a807cd9-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"99d77eed-40f0-4879-be96-abea3a807cd9\") " pod="openstack/cinder-api-0" Oct 08 18:30:21 crc kubenswrapper[4988]: I1008 18:30:21.713946 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/99d77eed-40f0-4879-be96-abea3a807cd9-config-data-custom\") pod \"cinder-api-0\" (UID: \"99d77eed-40f0-4879-be96-abea3a807cd9\") " pod="openstack/cinder-api-0" Oct 08 18:30:21 crc kubenswrapper[4988]: I1008 18:30:21.755405 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 08 18:30:21 crc kubenswrapper[4988]: I1008 18:30:21.755866 4988 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="b98d358d-c785-4494-932c-1d6662944b74" containerName="glance-log" containerID="cri-o://efa73d2a94e9d47c508d221c8632c29f721f7e3a2ae514318ae8a6b5fc355138" gracePeriod=30 Oct 08 18:30:21 crc kubenswrapper[4988]: I1008 18:30:21.755977 4988 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="b98d358d-c785-4494-932c-1d6662944b74" containerName="glance-httpd" containerID="cri-o://97c41cb131e40fdfd704edbc22cffe0cdc5412ff8904ccc8c9505df6fb46f714" gracePeriod=30 Oct 08 18:30:21 crc kubenswrapper[4988]: I1008 18:30:21.815777 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b6krt\" (UniqueName: \"kubernetes.io/projected/99d77eed-40f0-4879-be96-abea3a807cd9-kube-api-access-b6krt\") pod \"cinder-api-0\" (UID: \"99d77eed-40f0-4879-be96-abea3a807cd9\") " pod="openstack/cinder-api-0" Oct 08 18:30:21 crc kubenswrapper[4988]: I1008 18:30:21.815836 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/99d77eed-40f0-4879-be96-abea3a807cd9-etc-machine-id\") pod \"cinder-api-0\" (UID: \"99d77eed-40f0-4879-be96-abea3a807cd9\") " pod="openstack/cinder-api-0" Oct 08 18:30:21 crc kubenswrapper[4988]: I1008 18:30:21.815866 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/99d77eed-40f0-4879-be96-abea3a807cd9-config-data\") pod \"cinder-api-0\" (UID: \"99d77eed-40f0-4879-be96-abea3a807cd9\") " pod="openstack/cinder-api-0" Oct 08 18:30:21 crc kubenswrapper[4988]: I1008 18:30:21.815902 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/99d77eed-40f0-4879-be96-abea3a807cd9-scripts\") pod \"cinder-api-0\" (UID: \"99d77eed-40f0-4879-be96-abea3a807cd9\") " pod="openstack/cinder-api-0" Oct 08 18:30:21 crc kubenswrapper[4988]: I1008 18:30:21.815958 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/99d77eed-40f0-4879-be96-abea3a807cd9-logs\") pod \"cinder-api-0\" (UID: \"99d77eed-40f0-4879-be96-abea3a807cd9\") " pod="openstack/cinder-api-0" Oct 08 18:30:21 crc kubenswrapper[4988]: I1008 18:30:21.815976 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/99d77eed-40f0-4879-be96-abea3a807cd9-public-tls-certs\") pod \"cinder-api-0\" (UID: \"99d77eed-40f0-4879-be96-abea3a807cd9\") " pod="openstack/cinder-api-0" Oct 08 18:30:21 crc kubenswrapper[4988]: I1008 18:30:21.815996 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/99d77eed-40f0-4879-be96-abea3a807cd9-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"99d77eed-40f0-4879-be96-abea3a807cd9\") " pod="openstack/cinder-api-0" Oct 08 18:30:21 crc kubenswrapper[4988]: I1008 18:30:21.816001 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/99d77eed-40f0-4879-be96-abea3a807cd9-etc-machine-id\") pod \"cinder-api-0\" (UID: \"99d77eed-40f0-4879-be96-abea3a807cd9\") " pod="openstack/cinder-api-0" Oct 08 18:30:21 crc kubenswrapper[4988]: I1008 18:30:21.816037 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/99d77eed-40f0-4879-be96-abea3a807cd9-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"99d77eed-40f0-4879-be96-abea3a807cd9\") " pod="openstack/cinder-api-0" Oct 08 18:30:21 crc kubenswrapper[4988]: I1008 18:30:21.816054 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/99d77eed-40f0-4879-be96-abea3a807cd9-config-data-custom\") pod \"cinder-api-0\" (UID: \"99d77eed-40f0-4879-be96-abea3a807cd9\") " pod="openstack/cinder-api-0" Oct 08 18:30:21 crc kubenswrapper[4988]: I1008 18:30:21.817036 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/99d77eed-40f0-4879-be96-abea3a807cd9-logs\") pod \"cinder-api-0\" (UID: \"99d77eed-40f0-4879-be96-abea3a807cd9\") " pod="openstack/cinder-api-0" Oct 08 18:30:21 crc kubenswrapper[4988]: I1008 18:30:21.819370 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/99d77eed-40f0-4879-be96-abea3a807cd9-scripts\") pod \"cinder-api-0\" (UID: \"99d77eed-40f0-4879-be96-abea3a807cd9\") " pod="openstack/cinder-api-0" Oct 08 18:30:21 crc kubenswrapper[4988]: I1008 18:30:21.820227 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/99d77eed-40f0-4879-be96-abea3a807cd9-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"99d77eed-40f0-4879-be96-abea3a807cd9\") " pod="openstack/cinder-api-0" Oct 08 18:30:21 crc kubenswrapper[4988]: I1008 18:30:21.823865 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/99d77eed-40f0-4879-be96-abea3a807cd9-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"99d77eed-40f0-4879-be96-abea3a807cd9\") " pod="openstack/cinder-api-0" Oct 08 18:30:21 crc kubenswrapper[4988]: I1008 18:30:21.824290 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/99d77eed-40f0-4879-be96-abea3a807cd9-config-data-custom\") pod \"cinder-api-0\" (UID: \"99d77eed-40f0-4879-be96-abea3a807cd9\") " pod="openstack/cinder-api-0" Oct 08 18:30:21 crc kubenswrapper[4988]: I1008 18:30:21.824685 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/99d77eed-40f0-4879-be96-abea3a807cd9-public-tls-certs\") pod \"cinder-api-0\" (UID: \"99d77eed-40f0-4879-be96-abea3a807cd9\") " pod="openstack/cinder-api-0" Oct 08 18:30:21 crc kubenswrapper[4988]: I1008 18:30:21.835062 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/99d77eed-40f0-4879-be96-abea3a807cd9-config-data\") pod \"cinder-api-0\" (UID: \"99d77eed-40f0-4879-be96-abea3a807cd9\") " pod="openstack/cinder-api-0" Oct 08 18:30:21 crc kubenswrapper[4988]: I1008 18:30:21.846970 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b6krt\" (UniqueName: \"kubernetes.io/projected/99d77eed-40f0-4879-be96-abea3a807cd9-kube-api-access-b6krt\") pod \"cinder-api-0\" (UID: \"99d77eed-40f0-4879-be96-abea3a807cd9\") " pod="openstack/cinder-api-0" Oct 08 18:30:21 crc kubenswrapper[4988]: I1008 18:30:21.975448 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 08 18:30:22 crc kubenswrapper[4988]: W1008 18:30:22.449666 4988 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod99d77eed_40f0_4879_be96_abea3a807cd9.slice/crio-26ea215b02c8f513e1a65b6a89053fad73394e2112598a6fe6ab5ff048ac368f WatchSource:0}: Error finding container 26ea215b02c8f513e1a65b6a89053fad73394e2112598a6fe6ab5ff048ac368f: Status 404 returned error can't find the container with id 26ea215b02c8f513e1a65b6a89053fad73394e2112598a6fe6ab5ff048ac368f Oct 08 18:30:22 crc kubenswrapper[4988]: I1008 18:30:22.451638 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Oct 08 18:30:22 crc kubenswrapper[4988]: I1008 18:30:22.582477 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"99d77eed-40f0-4879-be96-abea3a807cd9","Type":"ContainerStarted","Data":"26ea215b02c8f513e1a65b6a89053fad73394e2112598a6fe6ab5ff048ac368f"} Oct 08 18:30:22 crc kubenswrapper[4988]: I1008 18:30:22.583988 4988 generic.go:334] "Generic (PLEG): container finished" podID="b98d358d-c785-4494-932c-1d6662944b74" containerID="efa73d2a94e9d47c508d221c8632c29f721f7e3a2ae514318ae8a6b5fc355138" exitCode=143 Oct 08 18:30:22 crc kubenswrapper[4988]: I1008 18:30:22.584031 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"b98d358d-c785-4494-932c-1d6662944b74","Type":"ContainerDied","Data":"efa73d2a94e9d47c508d221c8632c29f721f7e3a2ae514318ae8a6b5fc355138"} Oct 08 18:30:22 crc kubenswrapper[4988]: I1008 18:30:22.586535 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"dd22b6ca-9c02-46f8-b202-e9634e79261e","Type":"ContainerStarted","Data":"586b686a9fb503615c3706eb8f4495b64a82deed17b0d9af763fe2f7ba7b9c1a"} Oct 08 18:30:23 crc kubenswrapper[4988]: I1008 18:30:23.252330 4988 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d4d4bedc-1012-484f-8c73-ea8e6436b685" path="/var/lib/kubelet/pods/d4d4bedc-1012-484f-8c73-ea8e6436b685/volumes" Oct 08 18:30:23 crc kubenswrapper[4988]: I1008 18:30:23.337743 4988 patch_prober.go:28] interesting pod/machine-config-daemon-nm54f container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 18:30:23 crc kubenswrapper[4988]: I1008 18:30:23.337797 4988 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 18:30:23 crc kubenswrapper[4988]: I1008 18:30:23.631965 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"99d77eed-40f0-4879-be96-abea3a807cd9","Type":"ContainerStarted","Data":"1c61da6cc2966ef990d9bcdbb3737986e369db20fe8b42772368a61ef7dc1f03"} Oct 08 18:30:23 crc kubenswrapper[4988]: I1008 18:30:23.785507 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/swift-proxy-7b7cf6bf4f-s6vrs" Oct 08 18:30:23 crc kubenswrapper[4988]: I1008 18:30:23.786144 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/swift-proxy-7b7cf6bf4f-s6vrs" Oct 08 18:30:24 crc kubenswrapper[4988]: I1008 18:30:24.294058 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 08 18:30:24 crc kubenswrapper[4988]: I1008 18:30:24.294426 4988 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="97a836ba-13f2-4fba-8768-b8323af05c76" containerName="glance-httpd" containerID="cri-o://826b4afcb111f1466543dcc847edf3bfe65bed172e3c9408bb0e660047b6800f" gracePeriod=30 Oct 08 18:30:24 crc kubenswrapper[4988]: I1008 18:30:24.294326 4988 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="97a836ba-13f2-4fba-8768-b8323af05c76" containerName="glance-log" containerID="cri-o://169065f869ca95a8a25e30ecb0c2b7316835c33372ad4a1a673366dc1694dc25" gracePeriod=30 Oct 08 18:30:24 crc kubenswrapper[4988]: I1008 18:30:24.644373 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"dd22b6ca-9c02-46f8-b202-e9634e79261e","Type":"ContainerStarted","Data":"e5ca3b5706491f622025bb64d5490aa3dc7dfa0d1f5ee57a27c7d5ad68838da9"} Oct 08 18:30:24 crc kubenswrapper[4988]: I1008 18:30:24.647545 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 08 18:30:24 crc kubenswrapper[4988]: I1008 18:30:24.649543 4988 generic.go:334] "Generic (PLEG): container finished" podID="97a836ba-13f2-4fba-8768-b8323af05c76" containerID="169065f869ca95a8a25e30ecb0c2b7316835c33372ad4a1a673366dc1694dc25" exitCode=143 Oct 08 18:30:24 crc kubenswrapper[4988]: I1008 18:30:24.649622 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"97a836ba-13f2-4fba-8768-b8323af05c76","Type":"ContainerDied","Data":"169065f869ca95a8a25e30ecb0c2b7316835c33372ad4a1a673366dc1694dc25"} Oct 08 18:30:24 crc kubenswrapper[4988]: I1008 18:30:24.652520 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"99d77eed-40f0-4879-be96-abea3a807cd9","Type":"ContainerStarted","Data":"d2face2dfb0693cf7a2c3531e80c082e310d388b04c8005247e18e79ea8bbab3"} Oct 08 18:30:24 crc kubenswrapper[4988]: I1008 18:30:24.652658 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Oct 08 18:30:24 crc kubenswrapper[4988]: I1008 18:30:24.668689 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.4600582109999998 podStartE2EDuration="9.668667668s" podCreationTimestamp="2025-10-08 18:30:15 +0000 UTC" firstStartedPulling="2025-10-08 18:30:16.690328939 +0000 UTC m=+1162.140171709" lastFinishedPulling="2025-10-08 18:30:23.898938396 +0000 UTC m=+1169.348781166" observedRunningTime="2025-10-08 18:30:24.663779512 +0000 UTC m=+1170.113622282" watchObservedRunningTime="2025-10-08 18:30:24.668667668 +0000 UTC m=+1170.118510438" Oct 08 18:30:24 crc kubenswrapper[4988]: I1008 18:30:24.693628 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=3.693612675 podStartE2EDuration="3.693612675s" podCreationTimestamp="2025-10-08 18:30:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 18:30:24.687523441 +0000 UTC m=+1170.137366241" watchObservedRunningTime="2025-10-08 18:30:24.693612675 +0000 UTC m=+1170.143455446" Oct 08 18:30:24 crc kubenswrapper[4988]: I1008 18:30:24.840645 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 08 18:30:25 crc kubenswrapper[4988]: I1008 18:30:25.475876 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 08 18:30:25 crc kubenswrapper[4988]: I1008 18:30:25.499335 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b98d358d-c785-4494-932c-1d6662944b74-logs\") pod \"b98d358d-c785-4494-932c-1d6662944b74\" (UID: \"b98d358d-c785-4494-932c-1d6662944b74\") " Oct 08 18:30:25 crc kubenswrapper[4988]: I1008 18:30:25.499413 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/b98d358d-c785-4494-932c-1d6662944b74-httpd-run\") pod \"b98d358d-c785-4494-932c-1d6662944b74\" (UID: \"b98d358d-c785-4494-932c-1d6662944b74\") " Oct 08 18:30:25 crc kubenswrapper[4988]: I1008 18:30:25.499465 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"b98d358d-c785-4494-932c-1d6662944b74\" (UID: \"b98d358d-c785-4494-932c-1d6662944b74\") " Oct 08 18:30:25 crc kubenswrapper[4988]: I1008 18:30:25.499497 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-df76b\" (UniqueName: \"kubernetes.io/projected/b98d358d-c785-4494-932c-1d6662944b74-kube-api-access-df76b\") pod \"b98d358d-c785-4494-932c-1d6662944b74\" (UID: \"b98d358d-c785-4494-932c-1d6662944b74\") " Oct 08 18:30:25 crc kubenswrapper[4988]: I1008 18:30:25.499519 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b98d358d-c785-4494-932c-1d6662944b74-scripts\") pod \"b98d358d-c785-4494-932c-1d6662944b74\" (UID: \"b98d358d-c785-4494-932c-1d6662944b74\") " Oct 08 18:30:25 crc kubenswrapper[4988]: I1008 18:30:25.499549 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/b98d358d-c785-4494-932c-1d6662944b74-public-tls-certs\") pod \"b98d358d-c785-4494-932c-1d6662944b74\" (UID: \"b98d358d-c785-4494-932c-1d6662944b74\") " Oct 08 18:30:25 crc kubenswrapper[4988]: I1008 18:30:25.499582 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b98d358d-c785-4494-932c-1d6662944b74-combined-ca-bundle\") pod \"b98d358d-c785-4494-932c-1d6662944b74\" (UID: \"b98d358d-c785-4494-932c-1d6662944b74\") " Oct 08 18:30:25 crc kubenswrapper[4988]: I1008 18:30:25.499687 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b98d358d-c785-4494-932c-1d6662944b74-config-data\") pod \"b98d358d-c785-4494-932c-1d6662944b74\" (UID: \"b98d358d-c785-4494-932c-1d6662944b74\") " Oct 08 18:30:25 crc kubenswrapper[4988]: I1008 18:30:25.500233 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b98d358d-c785-4494-932c-1d6662944b74-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "b98d358d-c785-4494-932c-1d6662944b74" (UID: "b98d358d-c785-4494-932c-1d6662944b74"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 18:30:25 crc kubenswrapper[4988]: I1008 18:30:25.501931 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b98d358d-c785-4494-932c-1d6662944b74-logs" (OuterVolumeSpecName: "logs") pod "b98d358d-c785-4494-932c-1d6662944b74" (UID: "b98d358d-c785-4494-932c-1d6662944b74"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 18:30:25 crc kubenswrapper[4988]: I1008 18:30:25.518709 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b98d358d-c785-4494-932c-1d6662944b74-scripts" (OuterVolumeSpecName: "scripts") pod "b98d358d-c785-4494-932c-1d6662944b74" (UID: "b98d358d-c785-4494-932c-1d6662944b74"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:30:25 crc kubenswrapper[4988]: I1008 18:30:25.531038 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage11-crc" (OuterVolumeSpecName: "glance") pod "b98d358d-c785-4494-932c-1d6662944b74" (UID: "b98d358d-c785-4494-932c-1d6662944b74"). InnerVolumeSpecName "local-storage11-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 08 18:30:25 crc kubenswrapper[4988]: I1008 18:30:25.543240 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b98d358d-c785-4494-932c-1d6662944b74-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "b98d358d-c785-4494-932c-1d6662944b74" (UID: "b98d358d-c785-4494-932c-1d6662944b74"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:30:25 crc kubenswrapper[4988]: I1008 18:30:25.564585 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b98d358d-c785-4494-932c-1d6662944b74-kube-api-access-df76b" (OuterVolumeSpecName: "kube-api-access-df76b") pod "b98d358d-c785-4494-932c-1d6662944b74" (UID: "b98d358d-c785-4494-932c-1d6662944b74"). InnerVolumeSpecName "kube-api-access-df76b". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:30:25 crc kubenswrapper[4988]: I1008 18:30:25.589842 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b98d358d-c785-4494-932c-1d6662944b74-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "b98d358d-c785-4494-932c-1d6662944b74" (UID: "b98d358d-c785-4494-932c-1d6662944b74"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:30:25 crc kubenswrapper[4988]: I1008 18:30:25.605551 4988 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b98d358d-c785-4494-932c-1d6662944b74-logs\") on node \"crc\" DevicePath \"\"" Oct 08 18:30:25 crc kubenswrapper[4988]: I1008 18:30:25.605582 4988 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/b98d358d-c785-4494-932c-1d6662944b74-httpd-run\") on node \"crc\" DevicePath \"\"" Oct 08 18:30:25 crc kubenswrapper[4988]: I1008 18:30:25.605607 4988 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") on node \"crc\" " Oct 08 18:30:25 crc kubenswrapper[4988]: I1008 18:30:25.605618 4988 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-df76b\" (UniqueName: \"kubernetes.io/projected/b98d358d-c785-4494-932c-1d6662944b74-kube-api-access-df76b\") on node \"crc\" DevicePath \"\"" Oct 08 18:30:25 crc kubenswrapper[4988]: I1008 18:30:25.605629 4988 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b98d358d-c785-4494-932c-1d6662944b74-scripts\") on node \"crc\" DevicePath \"\"" Oct 08 18:30:25 crc kubenswrapper[4988]: I1008 18:30:25.605636 4988 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/b98d358d-c785-4494-932c-1d6662944b74-public-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 08 18:30:25 crc kubenswrapper[4988]: I1008 18:30:25.605645 4988 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b98d358d-c785-4494-932c-1d6662944b74-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 18:30:25 crc kubenswrapper[4988]: I1008 18:30:25.630975 4988 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage11-crc" (UniqueName: "kubernetes.io/local-volume/local-storage11-crc") on node "crc" Oct 08 18:30:25 crc kubenswrapper[4988]: I1008 18:30:25.648861 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b98d358d-c785-4494-932c-1d6662944b74-config-data" (OuterVolumeSpecName: "config-data") pod "b98d358d-c785-4494-932c-1d6662944b74" (UID: "b98d358d-c785-4494-932c-1d6662944b74"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:30:25 crc kubenswrapper[4988]: I1008 18:30:25.667198 4988 generic.go:334] "Generic (PLEG): container finished" podID="b98d358d-c785-4494-932c-1d6662944b74" containerID="97c41cb131e40fdfd704edbc22cffe0cdc5412ff8904ccc8c9505df6fb46f714" exitCode=0 Oct 08 18:30:25 crc kubenswrapper[4988]: I1008 18:30:25.667519 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 08 18:30:25 crc kubenswrapper[4988]: I1008 18:30:25.668348 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"b98d358d-c785-4494-932c-1d6662944b74","Type":"ContainerDied","Data":"97c41cb131e40fdfd704edbc22cffe0cdc5412ff8904ccc8c9505df6fb46f714"} Oct 08 18:30:25 crc kubenswrapper[4988]: I1008 18:30:25.668408 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"b98d358d-c785-4494-932c-1d6662944b74","Type":"ContainerDied","Data":"60b27f51694809ff52bff2c604cbfa056ed4778073da33d8b9ba28366e4d86f8"} Oct 08 18:30:25 crc kubenswrapper[4988]: I1008 18:30:25.668430 4988 scope.go:117] "RemoveContainer" containerID="97c41cb131e40fdfd704edbc22cffe0cdc5412ff8904ccc8c9505df6fb46f714" Oct 08 18:30:25 crc kubenswrapper[4988]: I1008 18:30:25.702506 4988 scope.go:117] "RemoveContainer" containerID="efa73d2a94e9d47c508d221c8632c29f721f7e3a2ae514318ae8a6b5fc355138" Oct 08 18:30:25 crc kubenswrapper[4988]: I1008 18:30:25.704719 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 08 18:30:25 crc kubenswrapper[4988]: I1008 18:30:25.706819 4988 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b98d358d-c785-4494-932c-1d6662944b74-config-data\") on node \"crc\" DevicePath \"\"" Oct 08 18:30:25 crc kubenswrapper[4988]: I1008 18:30:25.706843 4988 reconciler_common.go:293] "Volume detached for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") on node \"crc\" DevicePath \"\"" Oct 08 18:30:25 crc kubenswrapper[4988]: I1008 18:30:25.720081 4988 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 08 18:30:25 crc kubenswrapper[4988]: I1008 18:30:25.733236 4988 scope.go:117] "RemoveContainer" containerID="97c41cb131e40fdfd704edbc22cffe0cdc5412ff8904ccc8c9505df6fb46f714" Oct 08 18:30:25 crc kubenswrapper[4988]: E1008 18:30:25.733806 4988 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"97c41cb131e40fdfd704edbc22cffe0cdc5412ff8904ccc8c9505df6fb46f714\": container with ID starting with 97c41cb131e40fdfd704edbc22cffe0cdc5412ff8904ccc8c9505df6fb46f714 not found: ID does not exist" containerID="97c41cb131e40fdfd704edbc22cffe0cdc5412ff8904ccc8c9505df6fb46f714" Oct 08 18:30:25 crc kubenswrapper[4988]: I1008 18:30:25.733918 4988 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"97c41cb131e40fdfd704edbc22cffe0cdc5412ff8904ccc8c9505df6fb46f714"} err="failed to get container status \"97c41cb131e40fdfd704edbc22cffe0cdc5412ff8904ccc8c9505df6fb46f714\": rpc error: code = NotFound desc = could not find container \"97c41cb131e40fdfd704edbc22cffe0cdc5412ff8904ccc8c9505df6fb46f714\": container with ID starting with 97c41cb131e40fdfd704edbc22cffe0cdc5412ff8904ccc8c9505df6fb46f714 not found: ID does not exist" Oct 08 18:30:25 crc kubenswrapper[4988]: I1008 18:30:25.734009 4988 scope.go:117] "RemoveContainer" containerID="efa73d2a94e9d47c508d221c8632c29f721f7e3a2ae514318ae8a6b5fc355138" Oct 08 18:30:25 crc kubenswrapper[4988]: E1008 18:30:25.734554 4988 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"efa73d2a94e9d47c508d221c8632c29f721f7e3a2ae514318ae8a6b5fc355138\": container with ID starting with efa73d2a94e9d47c508d221c8632c29f721f7e3a2ae514318ae8a6b5fc355138 not found: ID does not exist" containerID="efa73d2a94e9d47c508d221c8632c29f721f7e3a2ae514318ae8a6b5fc355138" Oct 08 18:30:25 crc kubenswrapper[4988]: I1008 18:30:25.734677 4988 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"efa73d2a94e9d47c508d221c8632c29f721f7e3a2ae514318ae8a6b5fc355138"} err="failed to get container status \"efa73d2a94e9d47c508d221c8632c29f721f7e3a2ae514318ae8a6b5fc355138\": rpc error: code = NotFound desc = could not find container \"efa73d2a94e9d47c508d221c8632c29f721f7e3a2ae514318ae8a6b5fc355138\": container with ID starting with efa73d2a94e9d47c508d221c8632c29f721f7e3a2ae514318ae8a6b5fc355138 not found: ID does not exist" Oct 08 18:30:25 crc kubenswrapper[4988]: I1008 18:30:25.736555 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Oct 08 18:30:25 crc kubenswrapper[4988]: E1008 18:30:25.737036 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b98d358d-c785-4494-932c-1d6662944b74" containerName="glance-httpd" Oct 08 18:30:25 crc kubenswrapper[4988]: I1008 18:30:25.737059 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="b98d358d-c785-4494-932c-1d6662944b74" containerName="glance-httpd" Oct 08 18:30:25 crc kubenswrapper[4988]: E1008 18:30:25.737080 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b98d358d-c785-4494-932c-1d6662944b74" containerName="glance-log" Oct 08 18:30:25 crc kubenswrapper[4988]: I1008 18:30:25.737087 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="b98d358d-c785-4494-932c-1d6662944b74" containerName="glance-log" Oct 08 18:30:25 crc kubenswrapper[4988]: I1008 18:30:25.737285 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="b98d358d-c785-4494-932c-1d6662944b74" containerName="glance-httpd" Oct 08 18:30:25 crc kubenswrapper[4988]: I1008 18:30:25.737311 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="b98d358d-c785-4494-932c-1d6662944b74" containerName="glance-log" Oct 08 18:30:25 crc kubenswrapper[4988]: I1008 18:30:25.738408 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 08 18:30:25 crc kubenswrapper[4988]: I1008 18:30:25.743415 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Oct 08 18:30:25 crc kubenswrapper[4988]: I1008 18:30:25.743615 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Oct 08 18:30:25 crc kubenswrapper[4988]: I1008 18:30:25.765636 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 08 18:30:25 crc kubenswrapper[4988]: I1008 18:30:25.808249 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/56702a03-175d-4a54-b84f-e25cab037431-config-data\") pod \"glance-default-external-api-0\" (UID: \"56702a03-175d-4a54-b84f-e25cab037431\") " pod="openstack/glance-default-external-api-0" Oct 08 18:30:25 crc kubenswrapper[4988]: I1008 18:30:25.808296 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"glance-default-external-api-0\" (UID: \"56702a03-175d-4a54-b84f-e25cab037431\") " pod="openstack/glance-default-external-api-0" Oct 08 18:30:25 crc kubenswrapper[4988]: I1008 18:30:25.808336 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/56702a03-175d-4a54-b84f-e25cab037431-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"56702a03-175d-4a54-b84f-e25cab037431\") " pod="openstack/glance-default-external-api-0" Oct 08 18:30:25 crc kubenswrapper[4988]: I1008 18:30:25.808360 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/56702a03-175d-4a54-b84f-e25cab037431-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"56702a03-175d-4a54-b84f-e25cab037431\") " pod="openstack/glance-default-external-api-0" Oct 08 18:30:25 crc kubenswrapper[4988]: I1008 18:30:25.808398 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/56702a03-175d-4a54-b84f-e25cab037431-logs\") pod \"glance-default-external-api-0\" (UID: \"56702a03-175d-4a54-b84f-e25cab037431\") " pod="openstack/glance-default-external-api-0" Oct 08 18:30:25 crc kubenswrapper[4988]: I1008 18:30:25.808427 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g74gb\" (UniqueName: \"kubernetes.io/projected/56702a03-175d-4a54-b84f-e25cab037431-kube-api-access-g74gb\") pod \"glance-default-external-api-0\" (UID: \"56702a03-175d-4a54-b84f-e25cab037431\") " pod="openstack/glance-default-external-api-0" Oct 08 18:30:25 crc kubenswrapper[4988]: I1008 18:30:25.808455 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/56702a03-175d-4a54-b84f-e25cab037431-scripts\") pod \"glance-default-external-api-0\" (UID: \"56702a03-175d-4a54-b84f-e25cab037431\") " pod="openstack/glance-default-external-api-0" Oct 08 18:30:25 crc kubenswrapper[4988]: I1008 18:30:25.808478 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/56702a03-175d-4a54-b84f-e25cab037431-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"56702a03-175d-4a54-b84f-e25cab037431\") " pod="openstack/glance-default-external-api-0" Oct 08 18:30:25 crc kubenswrapper[4988]: I1008 18:30:25.910908 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/56702a03-175d-4a54-b84f-e25cab037431-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"56702a03-175d-4a54-b84f-e25cab037431\") " pod="openstack/glance-default-external-api-0" Oct 08 18:30:25 crc kubenswrapper[4988]: I1008 18:30:25.910959 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/56702a03-175d-4a54-b84f-e25cab037431-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"56702a03-175d-4a54-b84f-e25cab037431\") " pod="openstack/glance-default-external-api-0" Oct 08 18:30:25 crc kubenswrapper[4988]: I1008 18:30:25.910996 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/56702a03-175d-4a54-b84f-e25cab037431-logs\") pod \"glance-default-external-api-0\" (UID: \"56702a03-175d-4a54-b84f-e25cab037431\") " pod="openstack/glance-default-external-api-0" Oct 08 18:30:25 crc kubenswrapper[4988]: I1008 18:30:25.911058 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g74gb\" (UniqueName: \"kubernetes.io/projected/56702a03-175d-4a54-b84f-e25cab037431-kube-api-access-g74gb\") pod \"glance-default-external-api-0\" (UID: \"56702a03-175d-4a54-b84f-e25cab037431\") " pod="openstack/glance-default-external-api-0" Oct 08 18:30:25 crc kubenswrapper[4988]: I1008 18:30:25.911623 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/56702a03-175d-4a54-b84f-e25cab037431-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"56702a03-175d-4a54-b84f-e25cab037431\") " pod="openstack/glance-default-external-api-0" Oct 08 18:30:25 crc kubenswrapper[4988]: I1008 18:30:25.911655 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/56702a03-175d-4a54-b84f-e25cab037431-scripts\") pod \"glance-default-external-api-0\" (UID: \"56702a03-175d-4a54-b84f-e25cab037431\") " pod="openstack/glance-default-external-api-0" Oct 08 18:30:25 crc kubenswrapper[4988]: I1008 18:30:25.911639 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/56702a03-175d-4a54-b84f-e25cab037431-logs\") pod \"glance-default-external-api-0\" (UID: \"56702a03-175d-4a54-b84f-e25cab037431\") " pod="openstack/glance-default-external-api-0" Oct 08 18:30:25 crc kubenswrapper[4988]: I1008 18:30:25.911738 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/56702a03-175d-4a54-b84f-e25cab037431-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"56702a03-175d-4a54-b84f-e25cab037431\") " pod="openstack/glance-default-external-api-0" Oct 08 18:30:25 crc kubenswrapper[4988]: I1008 18:30:25.911936 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/56702a03-175d-4a54-b84f-e25cab037431-config-data\") pod \"glance-default-external-api-0\" (UID: \"56702a03-175d-4a54-b84f-e25cab037431\") " pod="openstack/glance-default-external-api-0" Oct 08 18:30:25 crc kubenswrapper[4988]: I1008 18:30:25.911997 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"glance-default-external-api-0\" (UID: \"56702a03-175d-4a54-b84f-e25cab037431\") " pod="openstack/glance-default-external-api-0" Oct 08 18:30:25 crc kubenswrapper[4988]: I1008 18:30:25.912256 4988 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"glance-default-external-api-0\" (UID: \"56702a03-175d-4a54-b84f-e25cab037431\") device mount path \"/mnt/openstack/pv11\"" pod="openstack/glance-default-external-api-0" Oct 08 18:30:25 crc kubenswrapper[4988]: I1008 18:30:25.916480 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/56702a03-175d-4a54-b84f-e25cab037431-config-data\") pod \"glance-default-external-api-0\" (UID: \"56702a03-175d-4a54-b84f-e25cab037431\") " pod="openstack/glance-default-external-api-0" Oct 08 18:30:25 crc kubenswrapper[4988]: I1008 18:30:25.916584 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/56702a03-175d-4a54-b84f-e25cab037431-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"56702a03-175d-4a54-b84f-e25cab037431\") " pod="openstack/glance-default-external-api-0" Oct 08 18:30:25 crc kubenswrapper[4988]: I1008 18:30:25.916658 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/56702a03-175d-4a54-b84f-e25cab037431-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"56702a03-175d-4a54-b84f-e25cab037431\") " pod="openstack/glance-default-external-api-0" Oct 08 18:30:25 crc kubenswrapper[4988]: I1008 18:30:25.922569 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/56702a03-175d-4a54-b84f-e25cab037431-scripts\") pod \"glance-default-external-api-0\" (UID: \"56702a03-175d-4a54-b84f-e25cab037431\") " pod="openstack/glance-default-external-api-0" Oct 08 18:30:25 crc kubenswrapper[4988]: I1008 18:30:25.926774 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g74gb\" (UniqueName: \"kubernetes.io/projected/56702a03-175d-4a54-b84f-e25cab037431-kube-api-access-g74gb\") pod \"glance-default-external-api-0\" (UID: \"56702a03-175d-4a54-b84f-e25cab037431\") " pod="openstack/glance-default-external-api-0" Oct 08 18:30:25 crc kubenswrapper[4988]: I1008 18:30:25.955318 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"glance-default-external-api-0\" (UID: \"56702a03-175d-4a54-b84f-e25cab037431\") " pod="openstack/glance-default-external-api-0" Oct 08 18:30:26 crc kubenswrapper[4988]: I1008 18:30:26.058346 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 08 18:30:26 crc kubenswrapper[4988]: I1008 18:30:26.149251 4988 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Oct 08 18:30:26 crc kubenswrapper[4988]: I1008 18:30:26.364676 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-59b9656b65-wtfvw" Oct 08 18:30:26 crc kubenswrapper[4988]: I1008 18:30:26.378496 4988 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Oct 08 18:30:26 crc kubenswrapper[4988]: I1008 18:30:26.426566 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-cddb74997-npcnc"] Oct 08 18:30:26 crc kubenswrapper[4988]: I1008 18:30:26.426814 4988 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-cddb74997-npcnc" podUID="6cbd4338-3c8f-481a-9523-fbe1d5964292" containerName="dnsmasq-dns" containerID="cri-o://7b822b6e3018338b30995c41fa217c4e11a7bcbc7709c7746e9fcbe90f2a0a84" gracePeriod=10 Oct 08 18:30:26 crc kubenswrapper[4988]: I1008 18:30:26.629822 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 08 18:30:26 crc kubenswrapper[4988]: W1008 18:30:26.643607 4988 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod56702a03_175d_4a54_b84f_e25cab037431.slice/crio-05fc98e1e038eeb4c35b2cae978bb1945dce3667d3c4eaf00c83a5cde3df55e3 WatchSource:0}: Error finding container 05fc98e1e038eeb4c35b2cae978bb1945dce3667d3c4eaf00c83a5cde3df55e3: Status 404 returned error can't find the container with id 05fc98e1e038eeb4c35b2cae978bb1945dce3667d3c4eaf00c83a5cde3df55e3 Oct 08 18:30:26 crc kubenswrapper[4988]: I1008 18:30:26.706128 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"56702a03-175d-4a54-b84f-e25cab037431","Type":"ContainerStarted","Data":"05fc98e1e038eeb4c35b2cae978bb1945dce3667d3c4eaf00c83a5cde3df55e3"} Oct 08 18:30:26 crc kubenswrapper[4988]: I1008 18:30:26.711476 4988 generic.go:334] "Generic (PLEG): container finished" podID="6cbd4338-3c8f-481a-9523-fbe1d5964292" containerID="7b822b6e3018338b30995c41fa217c4e11a7bcbc7709c7746e9fcbe90f2a0a84" exitCode=0 Oct 08 18:30:26 crc kubenswrapper[4988]: I1008 18:30:26.711743 4988 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="dd22b6ca-9c02-46f8-b202-e9634e79261e" containerName="ceilometer-central-agent" containerID="cri-o://e7f72cbf44e45568e6e9b978548b5543f7f049dadbaa5ec5eeba3c0abaf598a5" gracePeriod=30 Oct 08 18:30:26 crc kubenswrapper[4988]: I1008 18:30:26.712000 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-cddb74997-npcnc" event={"ID":"6cbd4338-3c8f-481a-9523-fbe1d5964292","Type":"ContainerDied","Data":"7b822b6e3018338b30995c41fa217c4e11a7bcbc7709c7746e9fcbe90f2a0a84"} Oct 08 18:30:26 crc kubenswrapper[4988]: I1008 18:30:26.713073 4988 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="dd22b6ca-9c02-46f8-b202-e9634e79261e" containerName="proxy-httpd" containerID="cri-o://e5ca3b5706491f622025bb64d5490aa3dc7dfa0d1f5ee57a27c7d5ad68838da9" gracePeriod=30 Oct 08 18:30:26 crc kubenswrapper[4988]: I1008 18:30:26.713152 4988 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="dd22b6ca-9c02-46f8-b202-e9634e79261e" containerName="sg-core" containerID="cri-o://586b686a9fb503615c3706eb8f4495b64a82deed17b0d9af763fe2f7ba7b9c1a" gracePeriod=30 Oct 08 18:30:26 crc kubenswrapper[4988]: I1008 18:30:26.713202 4988 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="dd22b6ca-9c02-46f8-b202-e9634e79261e" containerName="ceilometer-notification-agent" containerID="cri-o://f68fab85a12f4d29f55021002a6a561eb9662e3a85deffc9606602abbf47a9dc" gracePeriod=30 Oct 08 18:30:26 crc kubenswrapper[4988]: I1008 18:30:26.784498 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 08 18:30:27 crc kubenswrapper[4988]: I1008 18:30:27.247284 4988 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b98d358d-c785-4494-932c-1d6662944b74" path="/var/lib/kubelet/pods/b98d358d-c785-4494-932c-1d6662944b74/volumes" Oct 08 18:30:27 crc kubenswrapper[4988]: I1008 18:30:27.304907 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-cddb74997-npcnc" Oct 08 18:30:27 crc kubenswrapper[4988]: I1008 18:30:27.436733 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/6cbd4338-3c8f-481a-9523-fbe1d5964292-dns-swift-storage-0\") pod \"6cbd4338-3c8f-481a-9523-fbe1d5964292\" (UID: \"6cbd4338-3c8f-481a-9523-fbe1d5964292\") " Oct 08 18:30:27 crc kubenswrapper[4988]: I1008 18:30:27.436821 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6cbd4338-3c8f-481a-9523-fbe1d5964292-config\") pod \"6cbd4338-3c8f-481a-9523-fbe1d5964292\" (UID: \"6cbd4338-3c8f-481a-9523-fbe1d5964292\") " Oct 08 18:30:27 crc kubenswrapper[4988]: I1008 18:30:27.436846 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6cbd4338-3c8f-481a-9523-fbe1d5964292-ovsdbserver-sb\") pod \"6cbd4338-3c8f-481a-9523-fbe1d5964292\" (UID: \"6cbd4338-3c8f-481a-9523-fbe1d5964292\") " Oct 08 18:30:27 crc kubenswrapper[4988]: I1008 18:30:27.436909 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mf9vj\" (UniqueName: \"kubernetes.io/projected/6cbd4338-3c8f-481a-9523-fbe1d5964292-kube-api-access-mf9vj\") pod \"6cbd4338-3c8f-481a-9523-fbe1d5964292\" (UID: \"6cbd4338-3c8f-481a-9523-fbe1d5964292\") " Oct 08 18:30:27 crc kubenswrapper[4988]: I1008 18:30:27.436978 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6cbd4338-3c8f-481a-9523-fbe1d5964292-dns-svc\") pod \"6cbd4338-3c8f-481a-9523-fbe1d5964292\" (UID: \"6cbd4338-3c8f-481a-9523-fbe1d5964292\") " Oct 08 18:30:27 crc kubenswrapper[4988]: I1008 18:30:27.437019 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6cbd4338-3c8f-481a-9523-fbe1d5964292-ovsdbserver-nb\") pod \"6cbd4338-3c8f-481a-9523-fbe1d5964292\" (UID: \"6cbd4338-3c8f-481a-9523-fbe1d5964292\") " Oct 08 18:30:27 crc kubenswrapper[4988]: I1008 18:30:27.447440 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6cbd4338-3c8f-481a-9523-fbe1d5964292-kube-api-access-mf9vj" (OuterVolumeSpecName: "kube-api-access-mf9vj") pod "6cbd4338-3c8f-481a-9523-fbe1d5964292" (UID: "6cbd4338-3c8f-481a-9523-fbe1d5964292"). InnerVolumeSpecName "kube-api-access-mf9vj". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:30:27 crc kubenswrapper[4988]: I1008 18:30:27.495233 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6cbd4338-3c8f-481a-9523-fbe1d5964292-config" (OuterVolumeSpecName: "config") pod "6cbd4338-3c8f-481a-9523-fbe1d5964292" (UID: "6cbd4338-3c8f-481a-9523-fbe1d5964292"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 18:30:27 crc kubenswrapper[4988]: I1008 18:30:27.505177 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6cbd4338-3c8f-481a-9523-fbe1d5964292-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "6cbd4338-3c8f-481a-9523-fbe1d5964292" (UID: "6cbd4338-3c8f-481a-9523-fbe1d5964292"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 18:30:27 crc kubenswrapper[4988]: I1008 18:30:27.511886 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6cbd4338-3c8f-481a-9523-fbe1d5964292-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "6cbd4338-3c8f-481a-9523-fbe1d5964292" (UID: "6cbd4338-3c8f-481a-9523-fbe1d5964292"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 18:30:27 crc kubenswrapper[4988]: I1008 18:30:27.532335 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6cbd4338-3c8f-481a-9523-fbe1d5964292-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "6cbd4338-3c8f-481a-9523-fbe1d5964292" (UID: "6cbd4338-3c8f-481a-9523-fbe1d5964292"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 18:30:27 crc kubenswrapper[4988]: I1008 18:30:27.535116 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6cbd4338-3c8f-481a-9523-fbe1d5964292-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "6cbd4338-3c8f-481a-9523-fbe1d5964292" (UID: "6cbd4338-3c8f-481a-9523-fbe1d5964292"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 18:30:27 crc kubenswrapper[4988]: I1008 18:30:27.539079 4988 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mf9vj\" (UniqueName: \"kubernetes.io/projected/6cbd4338-3c8f-481a-9523-fbe1d5964292-kube-api-access-mf9vj\") on node \"crc\" DevicePath \"\"" Oct 08 18:30:27 crc kubenswrapper[4988]: I1008 18:30:27.539113 4988 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6cbd4338-3c8f-481a-9523-fbe1d5964292-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 08 18:30:27 crc kubenswrapper[4988]: I1008 18:30:27.539125 4988 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6cbd4338-3c8f-481a-9523-fbe1d5964292-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 08 18:30:27 crc kubenswrapper[4988]: I1008 18:30:27.539135 4988 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/6cbd4338-3c8f-481a-9523-fbe1d5964292-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Oct 08 18:30:27 crc kubenswrapper[4988]: I1008 18:30:27.539143 4988 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6cbd4338-3c8f-481a-9523-fbe1d5964292-config\") on node \"crc\" DevicePath \"\"" Oct 08 18:30:27 crc kubenswrapper[4988]: I1008 18:30:27.539150 4988 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6cbd4338-3c8f-481a-9523-fbe1d5964292-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 08 18:30:27 crc kubenswrapper[4988]: I1008 18:30:27.568548 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 08 18:30:27 crc kubenswrapper[4988]: I1008 18:30:27.727500 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"56702a03-175d-4a54-b84f-e25cab037431","Type":"ContainerStarted","Data":"9f5e2ee884e53117065985ec938aeba8be9ae37755acfba2392c435832c65f0e"} Oct 08 18:30:27 crc kubenswrapper[4988]: I1008 18:30:27.730838 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-cddb74997-npcnc" event={"ID":"6cbd4338-3c8f-481a-9523-fbe1d5964292","Type":"ContainerDied","Data":"cbee9fba14a3d1f6bb97b53358cbb2255b1f7ff83ccb8de7764c700a0d98afe9"} Oct 08 18:30:27 crc kubenswrapper[4988]: I1008 18:30:27.730869 4988 scope.go:117] "RemoveContainer" containerID="7b822b6e3018338b30995c41fa217c4e11a7bcbc7709c7746e9fcbe90f2a0a84" Oct 08 18:30:27 crc kubenswrapper[4988]: I1008 18:30:27.731005 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-cddb74997-npcnc" Oct 08 18:30:27 crc kubenswrapper[4988]: I1008 18:30:27.736412 4988 generic.go:334] "Generic (PLEG): container finished" podID="dd22b6ca-9c02-46f8-b202-e9634e79261e" containerID="e5ca3b5706491f622025bb64d5490aa3dc7dfa0d1f5ee57a27c7d5ad68838da9" exitCode=0 Oct 08 18:30:27 crc kubenswrapper[4988]: I1008 18:30:27.736433 4988 generic.go:334] "Generic (PLEG): container finished" podID="dd22b6ca-9c02-46f8-b202-e9634e79261e" containerID="586b686a9fb503615c3706eb8f4495b64a82deed17b0d9af763fe2f7ba7b9c1a" exitCode=2 Oct 08 18:30:27 crc kubenswrapper[4988]: I1008 18:30:27.736442 4988 generic.go:334] "Generic (PLEG): container finished" podID="dd22b6ca-9c02-46f8-b202-e9634e79261e" containerID="f68fab85a12f4d29f55021002a6a561eb9662e3a85deffc9606602abbf47a9dc" exitCode=0 Oct 08 18:30:27 crc kubenswrapper[4988]: I1008 18:30:27.736449 4988 generic.go:334] "Generic (PLEG): container finished" podID="dd22b6ca-9c02-46f8-b202-e9634e79261e" containerID="e7f72cbf44e45568e6e9b978548b5543f7f049dadbaa5ec5eeba3c0abaf598a5" exitCode=0 Oct 08 18:30:27 crc kubenswrapper[4988]: I1008 18:30:27.736488 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"dd22b6ca-9c02-46f8-b202-e9634e79261e","Type":"ContainerDied","Data":"e5ca3b5706491f622025bb64d5490aa3dc7dfa0d1f5ee57a27c7d5ad68838da9"} Oct 08 18:30:27 crc kubenswrapper[4988]: I1008 18:30:27.736514 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"dd22b6ca-9c02-46f8-b202-e9634e79261e","Type":"ContainerDied","Data":"586b686a9fb503615c3706eb8f4495b64a82deed17b0d9af763fe2f7ba7b9c1a"} Oct 08 18:30:27 crc kubenswrapper[4988]: I1008 18:30:27.736523 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"dd22b6ca-9c02-46f8-b202-e9634e79261e","Type":"ContainerDied","Data":"f68fab85a12f4d29f55021002a6a561eb9662e3a85deffc9606602abbf47a9dc"} Oct 08 18:30:27 crc kubenswrapper[4988]: I1008 18:30:27.736532 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"dd22b6ca-9c02-46f8-b202-e9634e79261e","Type":"ContainerDied","Data":"e7f72cbf44e45568e6e9b978548b5543f7f049dadbaa5ec5eeba3c0abaf598a5"} Oct 08 18:30:27 crc kubenswrapper[4988]: I1008 18:30:27.736541 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"dd22b6ca-9c02-46f8-b202-e9634e79261e","Type":"ContainerDied","Data":"37996652b3268f27bc8cf8ff09856cfb3e863ab4f12744f45d0fc8f9c03ecc0f"} Oct 08 18:30:27 crc kubenswrapper[4988]: I1008 18:30:27.736595 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 08 18:30:27 crc kubenswrapper[4988]: I1008 18:30:27.742246 4988 generic.go:334] "Generic (PLEG): container finished" podID="97a836ba-13f2-4fba-8768-b8323af05c76" containerID="826b4afcb111f1466543dcc847edf3bfe65bed172e3c9408bb0e660047b6800f" exitCode=0 Oct 08 18:30:27 crc kubenswrapper[4988]: I1008 18:30:27.742290 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"97a836ba-13f2-4fba-8768-b8323af05c76","Type":"ContainerDied","Data":"826b4afcb111f1466543dcc847edf3bfe65bed172e3c9408bb0e660047b6800f"} Oct 08 18:30:27 crc kubenswrapper[4988]: I1008 18:30:27.742466 4988 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="48acdfd7-797c-4545-9dd5-a9087ec4a525" containerName="cinder-scheduler" containerID="cri-o://9b7da6032d176b505a5dd95802130a58e053c8b0dbf445c13c049d3deee7e1cf" gracePeriod=30 Oct 08 18:30:27 crc kubenswrapper[4988]: I1008 18:30:27.742533 4988 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="48acdfd7-797c-4545-9dd5-a9087ec4a525" containerName="probe" containerID="cri-o://9fbef399ac4284687057dc2c47897eb4be0635cba1dc233289a1f03bfaa51bf3" gracePeriod=30 Oct 08 18:30:27 crc kubenswrapper[4988]: I1008 18:30:27.743152 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/dd22b6ca-9c02-46f8-b202-e9634e79261e-sg-core-conf-yaml\") pod \"dd22b6ca-9c02-46f8-b202-e9634e79261e\" (UID: \"dd22b6ca-9c02-46f8-b202-e9634e79261e\") " Oct 08 18:30:27 crc kubenswrapper[4988]: I1008 18:30:27.743186 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q4hm6\" (UniqueName: \"kubernetes.io/projected/dd22b6ca-9c02-46f8-b202-e9634e79261e-kube-api-access-q4hm6\") pod \"dd22b6ca-9c02-46f8-b202-e9634e79261e\" (UID: \"dd22b6ca-9c02-46f8-b202-e9634e79261e\") " Oct 08 18:30:27 crc kubenswrapper[4988]: I1008 18:30:27.743276 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dd22b6ca-9c02-46f8-b202-e9634e79261e-combined-ca-bundle\") pod \"dd22b6ca-9c02-46f8-b202-e9634e79261e\" (UID: \"dd22b6ca-9c02-46f8-b202-e9634e79261e\") " Oct 08 18:30:27 crc kubenswrapper[4988]: I1008 18:30:27.743352 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/dd22b6ca-9c02-46f8-b202-e9634e79261e-scripts\") pod \"dd22b6ca-9c02-46f8-b202-e9634e79261e\" (UID: \"dd22b6ca-9c02-46f8-b202-e9634e79261e\") " Oct 08 18:30:27 crc kubenswrapper[4988]: I1008 18:30:27.743690 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/dd22b6ca-9c02-46f8-b202-e9634e79261e-log-httpd\") pod \"dd22b6ca-9c02-46f8-b202-e9634e79261e\" (UID: \"dd22b6ca-9c02-46f8-b202-e9634e79261e\") " Oct 08 18:30:27 crc kubenswrapper[4988]: I1008 18:30:27.743711 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dd22b6ca-9c02-46f8-b202-e9634e79261e-config-data\") pod \"dd22b6ca-9c02-46f8-b202-e9634e79261e\" (UID: \"dd22b6ca-9c02-46f8-b202-e9634e79261e\") " Oct 08 18:30:27 crc kubenswrapper[4988]: I1008 18:30:27.743727 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/dd22b6ca-9c02-46f8-b202-e9634e79261e-run-httpd\") pod \"dd22b6ca-9c02-46f8-b202-e9634e79261e\" (UID: \"dd22b6ca-9c02-46f8-b202-e9634e79261e\") " Oct 08 18:30:27 crc kubenswrapper[4988]: I1008 18:30:27.744344 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/dd22b6ca-9c02-46f8-b202-e9634e79261e-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "dd22b6ca-9c02-46f8-b202-e9634e79261e" (UID: "dd22b6ca-9c02-46f8-b202-e9634e79261e"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 18:30:27 crc kubenswrapper[4988]: I1008 18:30:27.744797 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/dd22b6ca-9c02-46f8-b202-e9634e79261e-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "dd22b6ca-9c02-46f8-b202-e9634e79261e" (UID: "dd22b6ca-9c02-46f8-b202-e9634e79261e"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 18:30:27 crc kubenswrapper[4988]: I1008 18:30:27.750204 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dd22b6ca-9c02-46f8-b202-e9634e79261e-scripts" (OuterVolumeSpecName: "scripts") pod "dd22b6ca-9c02-46f8-b202-e9634e79261e" (UID: "dd22b6ca-9c02-46f8-b202-e9634e79261e"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:30:27 crc kubenswrapper[4988]: I1008 18:30:27.753840 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dd22b6ca-9c02-46f8-b202-e9634e79261e-kube-api-access-q4hm6" (OuterVolumeSpecName: "kube-api-access-q4hm6") pod "dd22b6ca-9c02-46f8-b202-e9634e79261e" (UID: "dd22b6ca-9c02-46f8-b202-e9634e79261e"). InnerVolumeSpecName "kube-api-access-q4hm6". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:30:27 crc kubenswrapper[4988]: I1008 18:30:27.762262 4988 scope.go:117] "RemoveContainer" containerID="8b74a595124b875ce655fa386eb777f3fcb8a872fe2678fd2b94e4af489fb8e3" Oct 08 18:30:27 crc kubenswrapper[4988]: I1008 18:30:27.783133 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-cddb74997-npcnc"] Oct 08 18:30:27 crc kubenswrapper[4988]: I1008 18:30:27.791917 4988 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-cddb74997-npcnc"] Oct 08 18:30:27 crc kubenswrapper[4988]: I1008 18:30:27.841858 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dd22b6ca-9c02-46f8-b202-e9634e79261e-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "dd22b6ca-9c02-46f8-b202-e9634e79261e" (UID: "dd22b6ca-9c02-46f8-b202-e9634e79261e"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:30:27 crc kubenswrapper[4988]: I1008 18:30:27.846354 4988 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q4hm6\" (UniqueName: \"kubernetes.io/projected/dd22b6ca-9c02-46f8-b202-e9634e79261e-kube-api-access-q4hm6\") on node \"crc\" DevicePath \"\"" Oct 08 18:30:27 crc kubenswrapper[4988]: I1008 18:30:27.846482 4988 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/dd22b6ca-9c02-46f8-b202-e9634e79261e-scripts\") on node \"crc\" DevicePath \"\"" Oct 08 18:30:27 crc kubenswrapper[4988]: I1008 18:30:27.846508 4988 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/dd22b6ca-9c02-46f8-b202-e9634e79261e-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 08 18:30:27 crc kubenswrapper[4988]: I1008 18:30:27.846521 4988 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/dd22b6ca-9c02-46f8-b202-e9634e79261e-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 08 18:30:27 crc kubenswrapper[4988]: I1008 18:30:27.846530 4988 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/dd22b6ca-9c02-46f8-b202-e9634e79261e-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Oct 08 18:30:27 crc kubenswrapper[4988]: I1008 18:30:27.871719 4988 scope.go:117] "RemoveContainer" containerID="e5ca3b5706491f622025bb64d5490aa3dc7dfa0d1f5ee57a27c7d5ad68838da9" Oct 08 18:30:27 crc kubenswrapper[4988]: I1008 18:30:27.889250 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dd22b6ca-9c02-46f8-b202-e9634e79261e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "dd22b6ca-9c02-46f8-b202-e9634e79261e" (UID: "dd22b6ca-9c02-46f8-b202-e9634e79261e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:30:27 crc kubenswrapper[4988]: I1008 18:30:27.911510 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dd22b6ca-9c02-46f8-b202-e9634e79261e-config-data" (OuterVolumeSpecName: "config-data") pod "dd22b6ca-9c02-46f8-b202-e9634e79261e" (UID: "dd22b6ca-9c02-46f8-b202-e9634e79261e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:30:27 crc kubenswrapper[4988]: I1008 18:30:27.925868 4988 scope.go:117] "RemoveContainer" containerID="586b686a9fb503615c3706eb8f4495b64a82deed17b0d9af763fe2f7ba7b9c1a" Oct 08 18:30:27 crc kubenswrapper[4988]: I1008 18:30:27.943729 4988 scope.go:117] "RemoveContainer" containerID="f68fab85a12f4d29f55021002a6a561eb9662e3a85deffc9606602abbf47a9dc" Oct 08 18:30:27 crc kubenswrapper[4988]: I1008 18:30:27.948425 4988 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dd22b6ca-9c02-46f8-b202-e9634e79261e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 18:30:27 crc kubenswrapper[4988]: I1008 18:30:27.948453 4988 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dd22b6ca-9c02-46f8-b202-e9634e79261e-config-data\") on node \"crc\" DevicePath \"\"" Oct 08 18:30:27 crc kubenswrapper[4988]: I1008 18:30:27.963911 4988 scope.go:117] "RemoveContainer" containerID="e7f72cbf44e45568e6e9b978548b5543f7f049dadbaa5ec5eeba3c0abaf598a5" Oct 08 18:30:27 crc kubenswrapper[4988]: I1008 18:30:27.996300 4988 scope.go:117] "RemoveContainer" containerID="e5ca3b5706491f622025bb64d5490aa3dc7dfa0d1f5ee57a27c7d5ad68838da9" Oct 08 18:30:27 crc kubenswrapper[4988]: E1008 18:30:27.996894 4988 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e5ca3b5706491f622025bb64d5490aa3dc7dfa0d1f5ee57a27c7d5ad68838da9\": container with ID starting with e5ca3b5706491f622025bb64d5490aa3dc7dfa0d1f5ee57a27c7d5ad68838da9 not found: ID does not exist" containerID="e5ca3b5706491f622025bb64d5490aa3dc7dfa0d1f5ee57a27c7d5ad68838da9" Oct 08 18:30:27 crc kubenswrapper[4988]: I1008 18:30:27.996939 4988 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e5ca3b5706491f622025bb64d5490aa3dc7dfa0d1f5ee57a27c7d5ad68838da9"} err="failed to get container status \"e5ca3b5706491f622025bb64d5490aa3dc7dfa0d1f5ee57a27c7d5ad68838da9\": rpc error: code = NotFound desc = could not find container \"e5ca3b5706491f622025bb64d5490aa3dc7dfa0d1f5ee57a27c7d5ad68838da9\": container with ID starting with e5ca3b5706491f622025bb64d5490aa3dc7dfa0d1f5ee57a27c7d5ad68838da9 not found: ID does not exist" Oct 08 18:30:27 crc kubenswrapper[4988]: I1008 18:30:27.996965 4988 scope.go:117] "RemoveContainer" containerID="586b686a9fb503615c3706eb8f4495b64a82deed17b0d9af763fe2f7ba7b9c1a" Oct 08 18:30:27 crc kubenswrapper[4988]: E1008 18:30:27.998729 4988 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"586b686a9fb503615c3706eb8f4495b64a82deed17b0d9af763fe2f7ba7b9c1a\": container with ID starting with 586b686a9fb503615c3706eb8f4495b64a82deed17b0d9af763fe2f7ba7b9c1a not found: ID does not exist" containerID="586b686a9fb503615c3706eb8f4495b64a82deed17b0d9af763fe2f7ba7b9c1a" Oct 08 18:30:27 crc kubenswrapper[4988]: I1008 18:30:27.998763 4988 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"586b686a9fb503615c3706eb8f4495b64a82deed17b0d9af763fe2f7ba7b9c1a"} err="failed to get container status \"586b686a9fb503615c3706eb8f4495b64a82deed17b0d9af763fe2f7ba7b9c1a\": rpc error: code = NotFound desc = could not find container \"586b686a9fb503615c3706eb8f4495b64a82deed17b0d9af763fe2f7ba7b9c1a\": container with ID starting with 586b686a9fb503615c3706eb8f4495b64a82deed17b0d9af763fe2f7ba7b9c1a not found: ID does not exist" Oct 08 18:30:27 crc kubenswrapper[4988]: I1008 18:30:27.998783 4988 scope.go:117] "RemoveContainer" containerID="f68fab85a12f4d29f55021002a6a561eb9662e3a85deffc9606602abbf47a9dc" Oct 08 18:30:27 crc kubenswrapper[4988]: E1008 18:30:27.999230 4988 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f68fab85a12f4d29f55021002a6a561eb9662e3a85deffc9606602abbf47a9dc\": container with ID starting with f68fab85a12f4d29f55021002a6a561eb9662e3a85deffc9606602abbf47a9dc not found: ID does not exist" containerID="f68fab85a12f4d29f55021002a6a561eb9662e3a85deffc9606602abbf47a9dc" Oct 08 18:30:27 crc kubenswrapper[4988]: I1008 18:30:27.999300 4988 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f68fab85a12f4d29f55021002a6a561eb9662e3a85deffc9606602abbf47a9dc"} err="failed to get container status \"f68fab85a12f4d29f55021002a6a561eb9662e3a85deffc9606602abbf47a9dc\": rpc error: code = NotFound desc = could not find container \"f68fab85a12f4d29f55021002a6a561eb9662e3a85deffc9606602abbf47a9dc\": container with ID starting with f68fab85a12f4d29f55021002a6a561eb9662e3a85deffc9606602abbf47a9dc not found: ID does not exist" Oct 08 18:30:27 crc kubenswrapper[4988]: I1008 18:30:27.999360 4988 scope.go:117] "RemoveContainer" containerID="e7f72cbf44e45568e6e9b978548b5543f7f049dadbaa5ec5eeba3c0abaf598a5" Oct 08 18:30:27 crc kubenswrapper[4988]: E1008 18:30:27.999874 4988 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e7f72cbf44e45568e6e9b978548b5543f7f049dadbaa5ec5eeba3c0abaf598a5\": container with ID starting with e7f72cbf44e45568e6e9b978548b5543f7f049dadbaa5ec5eeba3c0abaf598a5 not found: ID does not exist" containerID="e7f72cbf44e45568e6e9b978548b5543f7f049dadbaa5ec5eeba3c0abaf598a5" Oct 08 18:30:27 crc kubenswrapper[4988]: I1008 18:30:27.999932 4988 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e7f72cbf44e45568e6e9b978548b5543f7f049dadbaa5ec5eeba3c0abaf598a5"} err="failed to get container status \"e7f72cbf44e45568e6e9b978548b5543f7f049dadbaa5ec5eeba3c0abaf598a5\": rpc error: code = NotFound desc = could not find container \"e7f72cbf44e45568e6e9b978548b5543f7f049dadbaa5ec5eeba3c0abaf598a5\": container with ID starting with e7f72cbf44e45568e6e9b978548b5543f7f049dadbaa5ec5eeba3c0abaf598a5 not found: ID does not exist" Oct 08 18:30:28 crc kubenswrapper[4988]: I1008 18:30:27.999972 4988 scope.go:117] "RemoveContainer" containerID="e5ca3b5706491f622025bb64d5490aa3dc7dfa0d1f5ee57a27c7d5ad68838da9" Oct 08 18:30:28 crc kubenswrapper[4988]: I1008 18:30:28.000864 4988 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e5ca3b5706491f622025bb64d5490aa3dc7dfa0d1f5ee57a27c7d5ad68838da9"} err="failed to get container status \"e5ca3b5706491f622025bb64d5490aa3dc7dfa0d1f5ee57a27c7d5ad68838da9\": rpc error: code = NotFound desc = could not find container \"e5ca3b5706491f622025bb64d5490aa3dc7dfa0d1f5ee57a27c7d5ad68838da9\": container with ID starting with e5ca3b5706491f622025bb64d5490aa3dc7dfa0d1f5ee57a27c7d5ad68838da9 not found: ID does not exist" Oct 08 18:30:28 crc kubenswrapper[4988]: I1008 18:30:28.000887 4988 scope.go:117] "RemoveContainer" containerID="586b686a9fb503615c3706eb8f4495b64a82deed17b0d9af763fe2f7ba7b9c1a" Oct 08 18:30:28 crc kubenswrapper[4988]: I1008 18:30:28.002172 4988 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"586b686a9fb503615c3706eb8f4495b64a82deed17b0d9af763fe2f7ba7b9c1a"} err="failed to get container status \"586b686a9fb503615c3706eb8f4495b64a82deed17b0d9af763fe2f7ba7b9c1a\": rpc error: code = NotFound desc = could not find container \"586b686a9fb503615c3706eb8f4495b64a82deed17b0d9af763fe2f7ba7b9c1a\": container with ID starting with 586b686a9fb503615c3706eb8f4495b64a82deed17b0d9af763fe2f7ba7b9c1a not found: ID does not exist" Oct 08 18:30:28 crc kubenswrapper[4988]: I1008 18:30:28.002197 4988 scope.go:117] "RemoveContainer" containerID="f68fab85a12f4d29f55021002a6a561eb9662e3a85deffc9606602abbf47a9dc" Oct 08 18:30:28 crc kubenswrapper[4988]: I1008 18:30:28.002378 4988 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f68fab85a12f4d29f55021002a6a561eb9662e3a85deffc9606602abbf47a9dc"} err="failed to get container status \"f68fab85a12f4d29f55021002a6a561eb9662e3a85deffc9606602abbf47a9dc\": rpc error: code = NotFound desc = could not find container \"f68fab85a12f4d29f55021002a6a561eb9662e3a85deffc9606602abbf47a9dc\": container with ID starting with f68fab85a12f4d29f55021002a6a561eb9662e3a85deffc9606602abbf47a9dc not found: ID does not exist" Oct 08 18:30:28 crc kubenswrapper[4988]: I1008 18:30:28.002420 4988 scope.go:117] "RemoveContainer" containerID="e7f72cbf44e45568e6e9b978548b5543f7f049dadbaa5ec5eeba3c0abaf598a5" Oct 08 18:30:28 crc kubenswrapper[4988]: I1008 18:30:28.003259 4988 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e7f72cbf44e45568e6e9b978548b5543f7f049dadbaa5ec5eeba3c0abaf598a5"} err="failed to get container status \"e7f72cbf44e45568e6e9b978548b5543f7f049dadbaa5ec5eeba3c0abaf598a5\": rpc error: code = NotFound desc = could not find container \"e7f72cbf44e45568e6e9b978548b5543f7f049dadbaa5ec5eeba3c0abaf598a5\": container with ID starting with e7f72cbf44e45568e6e9b978548b5543f7f049dadbaa5ec5eeba3c0abaf598a5 not found: ID does not exist" Oct 08 18:30:28 crc kubenswrapper[4988]: I1008 18:30:28.003309 4988 scope.go:117] "RemoveContainer" containerID="e5ca3b5706491f622025bb64d5490aa3dc7dfa0d1f5ee57a27c7d5ad68838da9" Oct 08 18:30:28 crc kubenswrapper[4988]: I1008 18:30:28.005168 4988 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e5ca3b5706491f622025bb64d5490aa3dc7dfa0d1f5ee57a27c7d5ad68838da9"} err="failed to get container status \"e5ca3b5706491f622025bb64d5490aa3dc7dfa0d1f5ee57a27c7d5ad68838da9\": rpc error: code = NotFound desc = could not find container \"e5ca3b5706491f622025bb64d5490aa3dc7dfa0d1f5ee57a27c7d5ad68838da9\": container with ID starting with e5ca3b5706491f622025bb64d5490aa3dc7dfa0d1f5ee57a27c7d5ad68838da9 not found: ID does not exist" Oct 08 18:30:28 crc kubenswrapper[4988]: I1008 18:30:28.005200 4988 scope.go:117] "RemoveContainer" containerID="586b686a9fb503615c3706eb8f4495b64a82deed17b0d9af763fe2f7ba7b9c1a" Oct 08 18:30:28 crc kubenswrapper[4988]: I1008 18:30:28.005825 4988 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"586b686a9fb503615c3706eb8f4495b64a82deed17b0d9af763fe2f7ba7b9c1a"} err="failed to get container status \"586b686a9fb503615c3706eb8f4495b64a82deed17b0d9af763fe2f7ba7b9c1a\": rpc error: code = NotFound desc = could not find container \"586b686a9fb503615c3706eb8f4495b64a82deed17b0d9af763fe2f7ba7b9c1a\": container with ID starting with 586b686a9fb503615c3706eb8f4495b64a82deed17b0d9af763fe2f7ba7b9c1a not found: ID does not exist" Oct 08 18:30:28 crc kubenswrapper[4988]: I1008 18:30:28.005853 4988 scope.go:117] "RemoveContainer" containerID="f68fab85a12f4d29f55021002a6a561eb9662e3a85deffc9606602abbf47a9dc" Oct 08 18:30:28 crc kubenswrapper[4988]: I1008 18:30:28.006157 4988 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f68fab85a12f4d29f55021002a6a561eb9662e3a85deffc9606602abbf47a9dc"} err="failed to get container status \"f68fab85a12f4d29f55021002a6a561eb9662e3a85deffc9606602abbf47a9dc\": rpc error: code = NotFound desc = could not find container \"f68fab85a12f4d29f55021002a6a561eb9662e3a85deffc9606602abbf47a9dc\": container with ID starting with f68fab85a12f4d29f55021002a6a561eb9662e3a85deffc9606602abbf47a9dc not found: ID does not exist" Oct 08 18:30:28 crc kubenswrapper[4988]: I1008 18:30:28.006180 4988 scope.go:117] "RemoveContainer" containerID="e7f72cbf44e45568e6e9b978548b5543f7f049dadbaa5ec5eeba3c0abaf598a5" Oct 08 18:30:28 crc kubenswrapper[4988]: I1008 18:30:28.006459 4988 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e7f72cbf44e45568e6e9b978548b5543f7f049dadbaa5ec5eeba3c0abaf598a5"} err="failed to get container status \"e7f72cbf44e45568e6e9b978548b5543f7f049dadbaa5ec5eeba3c0abaf598a5\": rpc error: code = NotFound desc = could not find container \"e7f72cbf44e45568e6e9b978548b5543f7f049dadbaa5ec5eeba3c0abaf598a5\": container with ID starting with e7f72cbf44e45568e6e9b978548b5543f7f049dadbaa5ec5eeba3c0abaf598a5 not found: ID does not exist" Oct 08 18:30:28 crc kubenswrapper[4988]: I1008 18:30:28.006504 4988 scope.go:117] "RemoveContainer" containerID="e5ca3b5706491f622025bb64d5490aa3dc7dfa0d1f5ee57a27c7d5ad68838da9" Oct 08 18:30:28 crc kubenswrapper[4988]: I1008 18:30:28.015634 4988 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e5ca3b5706491f622025bb64d5490aa3dc7dfa0d1f5ee57a27c7d5ad68838da9"} err="failed to get container status \"e5ca3b5706491f622025bb64d5490aa3dc7dfa0d1f5ee57a27c7d5ad68838da9\": rpc error: code = NotFound desc = could not find container \"e5ca3b5706491f622025bb64d5490aa3dc7dfa0d1f5ee57a27c7d5ad68838da9\": container with ID starting with e5ca3b5706491f622025bb64d5490aa3dc7dfa0d1f5ee57a27c7d5ad68838da9 not found: ID does not exist" Oct 08 18:30:28 crc kubenswrapper[4988]: I1008 18:30:28.015702 4988 scope.go:117] "RemoveContainer" containerID="586b686a9fb503615c3706eb8f4495b64a82deed17b0d9af763fe2f7ba7b9c1a" Oct 08 18:30:28 crc kubenswrapper[4988]: I1008 18:30:28.016865 4988 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"586b686a9fb503615c3706eb8f4495b64a82deed17b0d9af763fe2f7ba7b9c1a"} err="failed to get container status \"586b686a9fb503615c3706eb8f4495b64a82deed17b0d9af763fe2f7ba7b9c1a\": rpc error: code = NotFound desc = could not find container \"586b686a9fb503615c3706eb8f4495b64a82deed17b0d9af763fe2f7ba7b9c1a\": container with ID starting with 586b686a9fb503615c3706eb8f4495b64a82deed17b0d9af763fe2f7ba7b9c1a not found: ID does not exist" Oct 08 18:30:28 crc kubenswrapper[4988]: I1008 18:30:28.016910 4988 scope.go:117] "RemoveContainer" containerID="f68fab85a12f4d29f55021002a6a561eb9662e3a85deffc9606602abbf47a9dc" Oct 08 18:30:28 crc kubenswrapper[4988]: I1008 18:30:28.017611 4988 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f68fab85a12f4d29f55021002a6a561eb9662e3a85deffc9606602abbf47a9dc"} err="failed to get container status \"f68fab85a12f4d29f55021002a6a561eb9662e3a85deffc9606602abbf47a9dc\": rpc error: code = NotFound desc = could not find container \"f68fab85a12f4d29f55021002a6a561eb9662e3a85deffc9606602abbf47a9dc\": container with ID starting with f68fab85a12f4d29f55021002a6a561eb9662e3a85deffc9606602abbf47a9dc not found: ID does not exist" Oct 08 18:30:28 crc kubenswrapper[4988]: I1008 18:30:28.017630 4988 scope.go:117] "RemoveContainer" containerID="e7f72cbf44e45568e6e9b978548b5543f7f049dadbaa5ec5eeba3c0abaf598a5" Oct 08 18:30:28 crc kubenswrapper[4988]: I1008 18:30:28.019179 4988 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e7f72cbf44e45568e6e9b978548b5543f7f049dadbaa5ec5eeba3c0abaf598a5"} err="failed to get container status \"e7f72cbf44e45568e6e9b978548b5543f7f049dadbaa5ec5eeba3c0abaf598a5\": rpc error: code = NotFound desc = could not find container \"e7f72cbf44e45568e6e9b978548b5543f7f049dadbaa5ec5eeba3c0abaf598a5\": container with ID starting with e7f72cbf44e45568e6e9b978548b5543f7f049dadbaa5ec5eeba3c0abaf598a5 not found: ID does not exist" Oct 08 18:30:28 crc kubenswrapper[4988]: I1008 18:30:28.137364 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 08 18:30:28 crc kubenswrapper[4988]: I1008 18:30:28.158506 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 08 18:30:28 crc kubenswrapper[4988]: I1008 18:30:28.192703 4988 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Oct 08 18:30:28 crc kubenswrapper[4988]: I1008 18:30:28.206737 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 08 18:30:28 crc kubenswrapper[4988]: E1008 18:30:28.207252 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dd22b6ca-9c02-46f8-b202-e9634e79261e" containerName="proxy-httpd" Oct 08 18:30:28 crc kubenswrapper[4988]: I1008 18:30:28.207279 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="dd22b6ca-9c02-46f8-b202-e9634e79261e" containerName="proxy-httpd" Oct 08 18:30:28 crc kubenswrapper[4988]: E1008 18:30:28.207305 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="97a836ba-13f2-4fba-8768-b8323af05c76" containerName="glance-log" Oct 08 18:30:28 crc kubenswrapper[4988]: I1008 18:30:28.207315 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="97a836ba-13f2-4fba-8768-b8323af05c76" containerName="glance-log" Oct 08 18:30:28 crc kubenswrapper[4988]: E1008 18:30:28.207335 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dd22b6ca-9c02-46f8-b202-e9634e79261e" containerName="ceilometer-central-agent" Oct 08 18:30:28 crc kubenswrapper[4988]: I1008 18:30:28.207344 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="dd22b6ca-9c02-46f8-b202-e9634e79261e" containerName="ceilometer-central-agent" Oct 08 18:30:28 crc kubenswrapper[4988]: E1008 18:30:28.207356 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6cbd4338-3c8f-481a-9523-fbe1d5964292" containerName="init" Oct 08 18:30:28 crc kubenswrapper[4988]: I1008 18:30:28.207364 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="6cbd4338-3c8f-481a-9523-fbe1d5964292" containerName="init" Oct 08 18:30:28 crc kubenswrapper[4988]: E1008 18:30:28.207487 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dd22b6ca-9c02-46f8-b202-e9634e79261e" containerName="ceilometer-notification-agent" Oct 08 18:30:28 crc kubenswrapper[4988]: I1008 18:30:28.207498 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="dd22b6ca-9c02-46f8-b202-e9634e79261e" containerName="ceilometer-notification-agent" Oct 08 18:30:28 crc kubenswrapper[4988]: E1008 18:30:28.207528 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dd22b6ca-9c02-46f8-b202-e9634e79261e" containerName="sg-core" Oct 08 18:30:28 crc kubenswrapper[4988]: I1008 18:30:28.207569 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="dd22b6ca-9c02-46f8-b202-e9634e79261e" containerName="sg-core" Oct 08 18:30:28 crc kubenswrapper[4988]: E1008 18:30:28.207584 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="97a836ba-13f2-4fba-8768-b8323af05c76" containerName="glance-httpd" Oct 08 18:30:28 crc kubenswrapper[4988]: I1008 18:30:28.207593 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="97a836ba-13f2-4fba-8768-b8323af05c76" containerName="glance-httpd" Oct 08 18:30:28 crc kubenswrapper[4988]: E1008 18:30:28.207609 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6cbd4338-3c8f-481a-9523-fbe1d5964292" containerName="dnsmasq-dns" Oct 08 18:30:28 crc kubenswrapper[4988]: I1008 18:30:28.207617 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="6cbd4338-3c8f-481a-9523-fbe1d5964292" containerName="dnsmasq-dns" Oct 08 18:30:28 crc kubenswrapper[4988]: I1008 18:30:28.207972 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="97a836ba-13f2-4fba-8768-b8323af05c76" containerName="glance-log" Oct 08 18:30:28 crc kubenswrapper[4988]: I1008 18:30:28.208003 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="dd22b6ca-9c02-46f8-b202-e9634e79261e" containerName="proxy-httpd" Oct 08 18:30:28 crc kubenswrapper[4988]: I1008 18:30:28.208021 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="dd22b6ca-9c02-46f8-b202-e9634e79261e" containerName="sg-core" Oct 08 18:30:28 crc kubenswrapper[4988]: I1008 18:30:28.208038 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="dd22b6ca-9c02-46f8-b202-e9634e79261e" containerName="ceilometer-central-agent" Oct 08 18:30:28 crc kubenswrapper[4988]: I1008 18:30:28.208054 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="97a836ba-13f2-4fba-8768-b8323af05c76" containerName="glance-httpd" Oct 08 18:30:28 crc kubenswrapper[4988]: I1008 18:30:28.208065 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="dd22b6ca-9c02-46f8-b202-e9634e79261e" containerName="ceilometer-notification-agent" Oct 08 18:30:28 crc kubenswrapper[4988]: I1008 18:30:28.208075 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="6cbd4338-3c8f-481a-9523-fbe1d5964292" containerName="dnsmasq-dns" Oct 08 18:30:28 crc kubenswrapper[4988]: I1008 18:30:28.210222 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 08 18:30:28 crc kubenswrapper[4988]: I1008 18:30:28.217293 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 08 18:30:28 crc kubenswrapper[4988]: I1008 18:30:28.217356 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 08 18:30:28 crc kubenswrapper[4988]: I1008 18:30:28.251351 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 08 18:30:28 crc kubenswrapper[4988]: I1008 18:30:28.254079 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/97a836ba-13f2-4fba-8768-b8323af05c76-scripts\") pod \"97a836ba-13f2-4fba-8768-b8323af05c76\" (UID: \"97a836ba-13f2-4fba-8768-b8323af05c76\") " Oct 08 18:30:28 crc kubenswrapper[4988]: I1008 18:30:28.254125 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/97a836ba-13f2-4fba-8768-b8323af05c76-config-data\") pod \"97a836ba-13f2-4fba-8768-b8323af05c76\" (UID: \"97a836ba-13f2-4fba-8768-b8323af05c76\") " Oct 08 18:30:28 crc kubenswrapper[4988]: I1008 18:30:28.254184 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/97a836ba-13f2-4fba-8768-b8323af05c76-logs\") pod \"97a836ba-13f2-4fba-8768-b8323af05c76\" (UID: \"97a836ba-13f2-4fba-8768-b8323af05c76\") " Oct 08 18:30:28 crc kubenswrapper[4988]: I1008 18:30:28.254281 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/97a836ba-13f2-4fba-8768-b8323af05c76-httpd-run\") pod \"97a836ba-13f2-4fba-8768-b8323af05c76\" (UID: \"97a836ba-13f2-4fba-8768-b8323af05c76\") " Oct 08 18:30:28 crc kubenswrapper[4988]: I1008 18:30:28.254321 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8mmxb\" (UniqueName: \"kubernetes.io/projected/97a836ba-13f2-4fba-8768-b8323af05c76-kube-api-access-8mmxb\") pod \"97a836ba-13f2-4fba-8768-b8323af05c76\" (UID: \"97a836ba-13f2-4fba-8768-b8323af05c76\") " Oct 08 18:30:28 crc kubenswrapper[4988]: I1008 18:30:28.254370 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/97a836ba-13f2-4fba-8768-b8323af05c76-internal-tls-certs\") pod \"97a836ba-13f2-4fba-8768-b8323af05c76\" (UID: \"97a836ba-13f2-4fba-8768-b8323af05c76\") " Oct 08 18:30:28 crc kubenswrapper[4988]: I1008 18:30:28.254407 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"97a836ba-13f2-4fba-8768-b8323af05c76\" (UID: \"97a836ba-13f2-4fba-8768-b8323af05c76\") " Oct 08 18:30:28 crc kubenswrapper[4988]: I1008 18:30:28.254461 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/97a836ba-13f2-4fba-8768-b8323af05c76-combined-ca-bundle\") pod \"97a836ba-13f2-4fba-8768-b8323af05c76\" (UID: \"97a836ba-13f2-4fba-8768-b8323af05c76\") " Oct 08 18:30:28 crc kubenswrapper[4988]: I1008 18:30:28.255666 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/97a836ba-13f2-4fba-8768-b8323af05c76-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "97a836ba-13f2-4fba-8768-b8323af05c76" (UID: "97a836ba-13f2-4fba-8768-b8323af05c76"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 18:30:28 crc kubenswrapper[4988]: I1008 18:30:28.256631 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/97a836ba-13f2-4fba-8768-b8323af05c76-logs" (OuterVolumeSpecName: "logs") pod "97a836ba-13f2-4fba-8768-b8323af05c76" (UID: "97a836ba-13f2-4fba-8768-b8323af05c76"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 18:30:28 crc kubenswrapper[4988]: I1008 18:30:28.264528 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage05-crc" (OuterVolumeSpecName: "glance") pod "97a836ba-13f2-4fba-8768-b8323af05c76" (UID: "97a836ba-13f2-4fba-8768-b8323af05c76"). InnerVolumeSpecName "local-storage05-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 08 18:30:28 crc kubenswrapper[4988]: I1008 18:30:28.272663 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/97a836ba-13f2-4fba-8768-b8323af05c76-kube-api-access-8mmxb" (OuterVolumeSpecName: "kube-api-access-8mmxb") pod "97a836ba-13f2-4fba-8768-b8323af05c76" (UID: "97a836ba-13f2-4fba-8768-b8323af05c76"). InnerVolumeSpecName "kube-api-access-8mmxb". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:30:28 crc kubenswrapper[4988]: I1008 18:30:28.275515 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/97a836ba-13f2-4fba-8768-b8323af05c76-scripts" (OuterVolumeSpecName: "scripts") pod "97a836ba-13f2-4fba-8768-b8323af05c76" (UID: "97a836ba-13f2-4fba-8768-b8323af05c76"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:30:28 crc kubenswrapper[4988]: I1008 18:30:28.313531 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/97a836ba-13f2-4fba-8768-b8323af05c76-config-data" (OuterVolumeSpecName: "config-data") pod "97a836ba-13f2-4fba-8768-b8323af05c76" (UID: "97a836ba-13f2-4fba-8768-b8323af05c76"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:30:28 crc kubenswrapper[4988]: I1008 18:30:28.324770 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/97a836ba-13f2-4fba-8768-b8323af05c76-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "97a836ba-13f2-4fba-8768-b8323af05c76" (UID: "97a836ba-13f2-4fba-8768-b8323af05c76"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:30:28 crc kubenswrapper[4988]: I1008 18:30:28.331468 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/97a836ba-13f2-4fba-8768-b8323af05c76-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "97a836ba-13f2-4fba-8768-b8323af05c76" (UID: "97a836ba-13f2-4fba-8768-b8323af05c76"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:30:28 crc kubenswrapper[4988]: I1008 18:30:28.356340 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-khgnt\" (UniqueName: \"kubernetes.io/projected/1920c047-744f-4365-937d-b2618118f20a-kube-api-access-khgnt\") pod \"ceilometer-0\" (UID: \"1920c047-744f-4365-937d-b2618118f20a\") " pod="openstack/ceilometer-0" Oct 08 18:30:28 crc kubenswrapper[4988]: I1008 18:30:28.356418 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1920c047-744f-4365-937d-b2618118f20a-log-httpd\") pod \"ceilometer-0\" (UID: \"1920c047-744f-4365-937d-b2618118f20a\") " pod="openstack/ceilometer-0" Oct 08 18:30:28 crc kubenswrapper[4988]: I1008 18:30:28.356443 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/1920c047-744f-4365-937d-b2618118f20a-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"1920c047-744f-4365-937d-b2618118f20a\") " pod="openstack/ceilometer-0" Oct 08 18:30:28 crc kubenswrapper[4988]: I1008 18:30:28.356479 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1920c047-744f-4365-937d-b2618118f20a-config-data\") pod \"ceilometer-0\" (UID: \"1920c047-744f-4365-937d-b2618118f20a\") " pod="openstack/ceilometer-0" Oct 08 18:30:28 crc kubenswrapper[4988]: I1008 18:30:28.356497 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1920c047-744f-4365-937d-b2618118f20a-run-httpd\") pod \"ceilometer-0\" (UID: \"1920c047-744f-4365-937d-b2618118f20a\") " pod="openstack/ceilometer-0" Oct 08 18:30:28 crc kubenswrapper[4988]: I1008 18:30:28.356539 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1920c047-744f-4365-937d-b2618118f20a-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"1920c047-744f-4365-937d-b2618118f20a\") " pod="openstack/ceilometer-0" Oct 08 18:30:28 crc kubenswrapper[4988]: I1008 18:30:28.356558 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1920c047-744f-4365-937d-b2618118f20a-scripts\") pod \"ceilometer-0\" (UID: \"1920c047-744f-4365-937d-b2618118f20a\") " pod="openstack/ceilometer-0" Oct 08 18:30:28 crc kubenswrapper[4988]: I1008 18:30:28.356682 4988 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8mmxb\" (UniqueName: \"kubernetes.io/projected/97a836ba-13f2-4fba-8768-b8323af05c76-kube-api-access-8mmxb\") on node \"crc\" DevicePath \"\"" Oct 08 18:30:28 crc kubenswrapper[4988]: I1008 18:30:28.356692 4988 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/97a836ba-13f2-4fba-8768-b8323af05c76-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 08 18:30:28 crc kubenswrapper[4988]: I1008 18:30:28.356711 4988 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") on node \"crc\" " Oct 08 18:30:28 crc kubenswrapper[4988]: I1008 18:30:28.356721 4988 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/97a836ba-13f2-4fba-8768-b8323af05c76-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 18:30:28 crc kubenswrapper[4988]: I1008 18:30:28.356729 4988 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/97a836ba-13f2-4fba-8768-b8323af05c76-scripts\") on node \"crc\" DevicePath \"\"" Oct 08 18:30:28 crc kubenswrapper[4988]: I1008 18:30:28.356737 4988 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/97a836ba-13f2-4fba-8768-b8323af05c76-config-data\") on node \"crc\" DevicePath \"\"" Oct 08 18:30:28 crc kubenswrapper[4988]: I1008 18:30:28.356744 4988 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/97a836ba-13f2-4fba-8768-b8323af05c76-logs\") on node \"crc\" DevicePath \"\"" Oct 08 18:30:28 crc kubenswrapper[4988]: I1008 18:30:28.356751 4988 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/97a836ba-13f2-4fba-8768-b8323af05c76-httpd-run\") on node \"crc\" DevicePath \"\"" Oct 08 18:30:28 crc kubenswrapper[4988]: I1008 18:30:28.376362 4988 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage05-crc" (UniqueName: "kubernetes.io/local-volume/local-storage05-crc") on node "crc" Oct 08 18:30:28 crc kubenswrapper[4988]: I1008 18:30:28.460532 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-khgnt\" (UniqueName: \"kubernetes.io/projected/1920c047-744f-4365-937d-b2618118f20a-kube-api-access-khgnt\") pod \"ceilometer-0\" (UID: \"1920c047-744f-4365-937d-b2618118f20a\") " pod="openstack/ceilometer-0" Oct 08 18:30:28 crc kubenswrapper[4988]: I1008 18:30:28.460594 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1920c047-744f-4365-937d-b2618118f20a-log-httpd\") pod \"ceilometer-0\" (UID: \"1920c047-744f-4365-937d-b2618118f20a\") " pod="openstack/ceilometer-0" Oct 08 18:30:28 crc kubenswrapper[4988]: I1008 18:30:28.460634 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/1920c047-744f-4365-937d-b2618118f20a-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"1920c047-744f-4365-937d-b2618118f20a\") " pod="openstack/ceilometer-0" Oct 08 18:30:28 crc kubenswrapper[4988]: I1008 18:30:28.460659 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1920c047-744f-4365-937d-b2618118f20a-config-data\") pod \"ceilometer-0\" (UID: \"1920c047-744f-4365-937d-b2618118f20a\") " pod="openstack/ceilometer-0" Oct 08 18:30:28 crc kubenswrapper[4988]: I1008 18:30:28.460679 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1920c047-744f-4365-937d-b2618118f20a-run-httpd\") pod \"ceilometer-0\" (UID: \"1920c047-744f-4365-937d-b2618118f20a\") " pod="openstack/ceilometer-0" Oct 08 18:30:28 crc kubenswrapper[4988]: I1008 18:30:28.460711 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1920c047-744f-4365-937d-b2618118f20a-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"1920c047-744f-4365-937d-b2618118f20a\") " pod="openstack/ceilometer-0" Oct 08 18:30:28 crc kubenswrapper[4988]: I1008 18:30:28.460731 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1920c047-744f-4365-937d-b2618118f20a-scripts\") pod \"ceilometer-0\" (UID: \"1920c047-744f-4365-937d-b2618118f20a\") " pod="openstack/ceilometer-0" Oct 08 18:30:28 crc kubenswrapper[4988]: I1008 18:30:28.460784 4988 reconciler_common.go:293] "Volume detached for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") on node \"crc\" DevicePath \"\"" Oct 08 18:30:28 crc kubenswrapper[4988]: I1008 18:30:28.462114 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1920c047-744f-4365-937d-b2618118f20a-run-httpd\") pod \"ceilometer-0\" (UID: \"1920c047-744f-4365-937d-b2618118f20a\") " pod="openstack/ceilometer-0" Oct 08 18:30:28 crc kubenswrapper[4988]: I1008 18:30:28.463713 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1920c047-744f-4365-937d-b2618118f20a-log-httpd\") pod \"ceilometer-0\" (UID: \"1920c047-744f-4365-937d-b2618118f20a\") " pod="openstack/ceilometer-0" Oct 08 18:30:28 crc kubenswrapper[4988]: I1008 18:30:28.468038 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1920c047-744f-4365-937d-b2618118f20a-scripts\") pod \"ceilometer-0\" (UID: \"1920c047-744f-4365-937d-b2618118f20a\") " pod="openstack/ceilometer-0" Oct 08 18:30:28 crc kubenswrapper[4988]: I1008 18:30:28.468600 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1920c047-744f-4365-937d-b2618118f20a-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"1920c047-744f-4365-937d-b2618118f20a\") " pod="openstack/ceilometer-0" Oct 08 18:30:28 crc kubenswrapper[4988]: I1008 18:30:28.469773 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/1920c047-744f-4365-937d-b2618118f20a-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"1920c047-744f-4365-937d-b2618118f20a\") " pod="openstack/ceilometer-0" Oct 08 18:30:28 crc kubenswrapper[4988]: I1008 18:30:28.469858 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1920c047-744f-4365-937d-b2618118f20a-config-data\") pod \"ceilometer-0\" (UID: \"1920c047-744f-4365-937d-b2618118f20a\") " pod="openstack/ceilometer-0" Oct 08 18:30:28 crc kubenswrapper[4988]: I1008 18:30:28.486516 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-khgnt\" (UniqueName: \"kubernetes.io/projected/1920c047-744f-4365-937d-b2618118f20a-kube-api-access-khgnt\") pod \"ceilometer-0\" (UID: \"1920c047-744f-4365-937d-b2618118f20a\") " pod="openstack/ceilometer-0" Oct 08 18:30:28 crc kubenswrapper[4988]: I1008 18:30:28.551117 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 08 18:30:28 crc kubenswrapper[4988]: I1008 18:30:28.756723 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 08 18:30:28 crc kubenswrapper[4988]: I1008 18:30:28.756763 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"97a836ba-13f2-4fba-8768-b8323af05c76","Type":"ContainerDied","Data":"3e3a1a6247bdb3ec156fbb7d8dc0c0f995772c9aa883ffb9d5cc63e7923786bb"} Oct 08 18:30:28 crc kubenswrapper[4988]: I1008 18:30:28.757105 4988 scope.go:117] "RemoveContainer" containerID="826b4afcb111f1466543dcc847edf3bfe65bed172e3c9408bb0e660047b6800f" Oct 08 18:30:28 crc kubenswrapper[4988]: I1008 18:30:28.764478 4988 generic.go:334] "Generic (PLEG): container finished" podID="48acdfd7-797c-4545-9dd5-a9087ec4a525" containerID="9fbef399ac4284687057dc2c47897eb4be0635cba1dc233289a1f03bfaa51bf3" exitCode=0 Oct 08 18:30:28 crc kubenswrapper[4988]: I1008 18:30:28.764583 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"48acdfd7-797c-4545-9dd5-a9087ec4a525","Type":"ContainerDied","Data":"9fbef399ac4284687057dc2c47897eb4be0635cba1dc233289a1f03bfaa51bf3"} Oct 08 18:30:28 crc kubenswrapper[4988]: I1008 18:30:28.769378 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"56702a03-175d-4a54-b84f-e25cab037431","Type":"ContainerStarted","Data":"23da0ff1f117dc601e37e2bb0a79de17b4662200fa002a471f0c5938e67d484b"} Oct 08 18:30:28 crc kubenswrapper[4988]: I1008 18:30:28.793896 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=3.793879509 podStartE2EDuration="3.793879509s" podCreationTimestamp="2025-10-08 18:30:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 18:30:28.791236685 +0000 UTC m=+1174.241079465" watchObservedRunningTime="2025-10-08 18:30:28.793879509 +0000 UTC m=+1174.243722279" Oct 08 18:30:28 crc kubenswrapper[4988]: I1008 18:30:28.818081 4988 scope.go:117] "RemoveContainer" containerID="169065f869ca95a8a25e30ecb0c2b7316835c33372ad4a1a673366dc1694dc25" Oct 08 18:30:28 crc kubenswrapper[4988]: I1008 18:30:28.831571 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 08 18:30:28 crc kubenswrapper[4988]: I1008 18:30:28.841879 4988 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 08 18:30:28 crc kubenswrapper[4988]: I1008 18:30:28.853516 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 08 18:30:28 crc kubenswrapper[4988]: I1008 18:30:28.855013 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 08 18:30:28 crc kubenswrapper[4988]: I1008 18:30:28.860001 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 08 18:30:28 crc kubenswrapper[4988]: I1008 18:30:28.860566 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Oct 08 18:30:28 crc kubenswrapper[4988]: I1008 18:30:28.860830 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Oct 08 18:30:28 crc kubenswrapper[4988]: I1008 18:30:28.973408 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/0fcc9499-a06b-4cd1-a5da-ea8108916050-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"0fcc9499-a06b-4cd1-a5da-ea8108916050\") " pod="openstack/glance-default-internal-api-0" Oct 08 18:30:28 crc kubenswrapper[4988]: I1008 18:30:28.973749 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g9rvl\" (UniqueName: \"kubernetes.io/projected/0fcc9499-a06b-4cd1-a5da-ea8108916050-kube-api-access-g9rvl\") pod \"glance-default-internal-api-0\" (UID: \"0fcc9499-a06b-4cd1-a5da-ea8108916050\") " pod="openstack/glance-default-internal-api-0" Oct 08 18:30:28 crc kubenswrapper[4988]: I1008 18:30:28.973783 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-internal-api-0\" (UID: \"0fcc9499-a06b-4cd1-a5da-ea8108916050\") " pod="openstack/glance-default-internal-api-0" Oct 08 18:30:28 crc kubenswrapper[4988]: I1008 18:30:28.973821 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/0fcc9499-a06b-4cd1-a5da-ea8108916050-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"0fcc9499-a06b-4cd1-a5da-ea8108916050\") " pod="openstack/glance-default-internal-api-0" Oct 08 18:30:28 crc kubenswrapper[4988]: I1008 18:30:28.973845 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0fcc9499-a06b-4cd1-a5da-ea8108916050-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"0fcc9499-a06b-4cd1-a5da-ea8108916050\") " pod="openstack/glance-default-internal-api-0" Oct 08 18:30:28 crc kubenswrapper[4988]: I1008 18:30:28.973860 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0fcc9499-a06b-4cd1-a5da-ea8108916050-logs\") pod \"glance-default-internal-api-0\" (UID: \"0fcc9499-a06b-4cd1-a5da-ea8108916050\") " pod="openstack/glance-default-internal-api-0" Oct 08 18:30:28 crc kubenswrapper[4988]: I1008 18:30:28.973915 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0fcc9499-a06b-4cd1-a5da-ea8108916050-scripts\") pod \"glance-default-internal-api-0\" (UID: \"0fcc9499-a06b-4cd1-a5da-ea8108916050\") " pod="openstack/glance-default-internal-api-0" Oct 08 18:30:28 crc kubenswrapper[4988]: I1008 18:30:28.973937 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0fcc9499-a06b-4cd1-a5da-ea8108916050-config-data\") pod \"glance-default-internal-api-0\" (UID: \"0fcc9499-a06b-4cd1-a5da-ea8108916050\") " pod="openstack/glance-default-internal-api-0" Oct 08 18:30:29 crc kubenswrapper[4988]: I1008 18:30:29.019780 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 08 18:30:29 crc kubenswrapper[4988]: W1008 18:30:29.024112 4988 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1920c047_744f_4365_937d_b2618118f20a.slice/crio-e5c62e8aa7d7b5023992aa9dc11e06db48fa32b76a39e6fe765d4a87060fddf3 WatchSource:0}: Error finding container e5c62e8aa7d7b5023992aa9dc11e06db48fa32b76a39e6fe765d4a87060fddf3: Status 404 returned error can't find the container with id e5c62e8aa7d7b5023992aa9dc11e06db48fa32b76a39e6fe765d4a87060fddf3 Oct 08 18:30:29 crc kubenswrapper[4988]: I1008 18:30:29.075752 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0fcc9499-a06b-4cd1-a5da-ea8108916050-scripts\") pod \"glance-default-internal-api-0\" (UID: \"0fcc9499-a06b-4cd1-a5da-ea8108916050\") " pod="openstack/glance-default-internal-api-0" Oct 08 18:30:29 crc kubenswrapper[4988]: I1008 18:30:29.076099 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0fcc9499-a06b-4cd1-a5da-ea8108916050-config-data\") pod \"glance-default-internal-api-0\" (UID: \"0fcc9499-a06b-4cd1-a5da-ea8108916050\") " pod="openstack/glance-default-internal-api-0" Oct 08 18:30:29 crc kubenswrapper[4988]: I1008 18:30:29.076315 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/0fcc9499-a06b-4cd1-a5da-ea8108916050-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"0fcc9499-a06b-4cd1-a5da-ea8108916050\") " pod="openstack/glance-default-internal-api-0" Oct 08 18:30:29 crc kubenswrapper[4988]: I1008 18:30:29.076440 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g9rvl\" (UniqueName: \"kubernetes.io/projected/0fcc9499-a06b-4cd1-a5da-ea8108916050-kube-api-access-g9rvl\") pod \"glance-default-internal-api-0\" (UID: \"0fcc9499-a06b-4cd1-a5da-ea8108916050\") " pod="openstack/glance-default-internal-api-0" Oct 08 18:30:29 crc kubenswrapper[4988]: I1008 18:30:29.076563 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-internal-api-0\" (UID: \"0fcc9499-a06b-4cd1-a5da-ea8108916050\") " pod="openstack/glance-default-internal-api-0" Oct 08 18:30:29 crc kubenswrapper[4988]: I1008 18:30:29.076719 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/0fcc9499-a06b-4cd1-a5da-ea8108916050-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"0fcc9499-a06b-4cd1-a5da-ea8108916050\") " pod="openstack/glance-default-internal-api-0" Oct 08 18:30:29 crc kubenswrapper[4988]: I1008 18:30:29.076839 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0fcc9499-a06b-4cd1-a5da-ea8108916050-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"0fcc9499-a06b-4cd1-a5da-ea8108916050\") " pod="openstack/glance-default-internal-api-0" Oct 08 18:30:29 crc kubenswrapper[4988]: I1008 18:30:29.076919 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/0fcc9499-a06b-4cd1-a5da-ea8108916050-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"0fcc9499-a06b-4cd1-a5da-ea8108916050\") " pod="openstack/glance-default-internal-api-0" Oct 08 18:30:29 crc kubenswrapper[4988]: I1008 18:30:29.077019 4988 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-internal-api-0\" (UID: \"0fcc9499-a06b-4cd1-a5da-ea8108916050\") device mount path \"/mnt/openstack/pv05\"" pod="openstack/glance-default-internal-api-0" Oct 08 18:30:29 crc kubenswrapper[4988]: I1008 18:30:29.077123 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0fcc9499-a06b-4cd1-a5da-ea8108916050-logs\") pod \"glance-default-internal-api-0\" (UID: \"0fcc9499-a06b-4cd1-a5da-ea8108916050\") " pod="openstack/glance-default-internal-api-0" Oct 08 18:30:29 crc kubenswrapper[4988]: I1008 18:30:29.077293 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0fcc9499-a06b-4cd1-a5da-ea8108916050-logs\") pod \"glance-default-internal-api-0\" (UID: \"0fcc9499-a06b-4cd1-a5da-ea8108916050\") " pod="openstack/glance-default-internal-api-0" Oct 08 18:30:29 crc kubenswrapper[4988]: I1008 18:30:29.083193 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/0fcc9499-a06b-4cd1-a5da-ea8108916050-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"0fcc9499-a06b-4cd1-a5da-ea8108916050\") " pod="openstack/glance-default-internal-api-0" Oct 08 18:30:29 crc kubenswrapper[4988]: I1008 18:30:29.083623 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0fcc9499-a06b-4cd1-a5da-ea8108916050-config-data\") pod \"glance-default-internal-api-0\" (UID: \"0fcc9499-a06b-4cd1-a5da-ea8108916050\") " pod="openstack/glance-default-internal-api-0" Oct 08 18:30:29 crc kubenswrapper[4988]: I1008 18:30:29.083620 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0fcc9499-a06b-4cd1-a5da-ea8108916050-scripts\") pod \"glance-default-internal-api-0\" (UID: \"0fcc9499-a06b-4cd1-a5da-ea8108916050\") " pod="openstack/glance-default-internal-api-0" Oct 08 18:30:29 crc kubenswrapper[4988]: I1008 18:30:29.085300 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0fcc9499-a06b-4cd1-a5da-ea8108916050-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"0fcc9499-a06b-4cd1-a5da-ea8108916050\") " pod="openstack/glance-default-internal-api-0" Oct 08 18:30:29 crc kubenswrapper[4988]: I1008 18:30:29.099184 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g9rvl\" (UniqueName: \"kubernetes.io/projected/0fcc9499-a06b-4cd1-a5da-ea8108916050-kube-api-access-g9rvl\") pod \"glance-default-internal-api-0\" (UID: \"0fcc9499-a06b-4cd1-a5da-ea8108916050\") " pod="openstack/glance-default-internal-api-0" Oct 08 18:30:29 crc kubenswrapper[4988]: I1008 18:30:29.109853 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-internal-api-0\" (UID: \"0fcc9499-a06b-4cd1-a5da-ea8108916050\") " pod="openstack/glance-default-internal-api-0" Oct 08 18:30:29 crc kubenswrapper[4988]: I1008 18:30:29.175443 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 08 18:30:29 crc kubenswrapper[4988]: I1008 18:30:29.258045 4988 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6cbd4338-3c8f-481a-9523-fbe1d5964292" path="/var/lib/kubelet/pods/6cbd4338-3c8f-481a-9523-fbe1d5964292/volumes" Oct 08 18:30:29 crc kubenswrapper[4988]: I1008 18:30:29.259147 4988 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="97a836ba-13f2-4fba-8768-b8323af05c76" path="/var/lib/kubelet/pods/97a836ba-13f2-4fba-8768-b8323af05c76/volumes" Oct 08 18:30:29 crc kubenswrapper[4988]: I1008 18:30:29.259900 4988 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="dd22b6ca-9c02-46f8-b202-e9634e79261e" path="/var/lib/kubelet/pods/dd22b6ca-9c02-46f8-b202-e9634e79261e/volumes" Oct 08 18:30:29 crc kubenswrapper[4988]: I1008 18:30:29.752110 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 08 18:30:29 crc kubenswrapper[4988]: I1008 18:30:29.786186 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"0fcc9499-a06b-4cd1-a5da-ea8108916050","Type":"ContainerStarted","Data":"13b42e11464da2a16b26f60e40ed4b477086a4e26e88e8184281465eec0c04c9"} Oct 08 18:30:29 crc kubenswrapper[4988]: I1008 18:30:29.787507 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"1920c047-744f-4365-937d-b2618118f20a","Type":"ContainerStarted","Data":"e5c62e8aa7d7b5023992aa9dc11e06db48fa32b76a39e6fe765d4a87060fddf3"} Oct 08 18:30:29 crc kubenswrapper[4988]: I1008 18:30:29.836482 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 08 18:30:30 crc kubenswrapper[4988]: I1008 18:30:30.851545 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"0fcc9499-a06b-4cd1-a5da-ea8108916050","Type":"ContainerStarted","Data":"7d3dbf100ba64891a97f34dd540d20ec061ba6822f245dec78c63c17c5d178fa"} Oct 08 18:30:30 crc kubenswrapper[4988]: I1008 18:30:30.868007 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"1920c047-744f-4365-937d-b2618118f20a","Type":"ContainerStarted","Data":"9cb9807398778f96d1bc49c73ecd4c900ad4814d7e983b782c92e5145fc9490c"} Oct 08 18:30:30 crc kubenswrapper[4988]: I1008 18:30:30.868049 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"1920c047-744f-4365-937d-b2618118f20a","Type":"ContainerStarted","Data":"50f43557736bb7d729755e3a24c5bffebf5e4b3f810a71a05545b4c5793d7ad5"} Oct 08 18:30:31 crc kubenswrapper[4988]: I1008 18:30:31.881231 4988 generic.go:334] "Generic (PLEG): container finished" podID="48acdfd7-797c-4545-9dd5-a9087ec4a525" containerID="9b7da6032d176b505a5dd95802130a58e053c8b0dbf445c13c049d3deee7e1cf" exitCode=0 Oct 08 18:30:31 crc kubenswrapper[4988]: I1008 18:30:31.881309 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"48acdfd7-797c-4545-9dd5-a9087ec4a525","Type":"ContainerDied","Data":"9b7da6032d176b505a5dd95802130a58e053c8b0dbf445c13c049d3deee7e1cf"} Oct 08 18:30:31 crc kubenswrapper[4988]: I1008 18:30:31.886453 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"1920c047-744f-4365-937d-b2618118f20a","Type":"ContainerStarted","Data":"f7bcf58fcffdcb57bf35d7ebba08dd99e688d5e331252bf0101612b163458960"} Oct 08 18:30:31 crc kubenswrapper[4988]: I1008 18:30:31.891988 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"0fcc9499-a06b-4cd1-a5da-ea8108916050","Type":"ContainerStarted","Data":"9315fcc87c2b03972c71e57728c4fdbc3b30200edc121e010b7ac2e7d7658e92"} Oct 08 18:30:31 crc kubenswrapper[4988]: I1008 18:30:31.915351 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=3.915334786 podStartE2EDuration="3.915334786s" podCreationTimestamp="2025-10-08 18:30:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 18:30:31.90950334 +0000 UTC m=+1177.359346120" watchObservedRunningTime="2025-10-08 18:30:31.915334786 +0000 UTC m=+1177.365177556" Oct 08 18:30:32 crc kubenswrapper[4988]: I1008 18:30:32.271701 4988 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-cddb74997-npcnc" podUID="6cbd4338-3c8f-481a-9523-fbe1d5964292" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.160:5353: i/o timeout" Oct 08 18:30:32 crc kubenswrapper[4988]: I1008 18:30:32.291436 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Oct 08 18:30:32 crc kubenswrapper[4988]: I1008 18:30:32.361676 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/48acdfd7-797c-4545-9dd5-a9087ec4a525-etc-machine-id\") pod \"48acdfd7-797c-4545-9dd5-a9087ec4a525\" (UID: \"48acdfd7-797c-4545-9dd5-a9087ec4a525\") " Oct 08 18:30:32 crc kubenswrapper[4988]: I1008 18:30:32.361877 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/48acdfd7-797c-4545-9dd5-a9087ec4a525-config-data\") pod \"48acdfd7-797c-4545-9dd5-a9087ec4a525\" (UID: \"48acdfd7-797c-4545-9dd5-a9087ec4a525\") " Oct 08 18:30:32 crc kubenswrapper[4988]: I1008 18:30:32.361944 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/48acdfd7-797c-4545-9dd5-a9087ec4a525-scripts\") pod \"48acdfd7-797c-4545-9dd5-a9087ec4a525\" (UID: \"48acdfd7-797c-4545-9dd5-a9087ec4a525\") " Oct 08 18:30:32 crc kubenswrapper[4988]: I1008 18:30:32.361985 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/48acdfd7-797c-4545-9dd5-a9087ec4a525-combined-ca-bundle\") pod \"48acdfd7-797c-4545-9dd5-a9087ec4a525\" (UID: \"48acdfd7-797c-4545-9dd5-a9087ec4a525\") " Oct 08 18:30:32 crc kubenswrapper[4988]: I1008 18:30:32.362048 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6vf2x\" (UniqueName: \"kubernetes.io/projected/48acdfd7-797c-4545-9dd5-a9087ec4a525-kube-api-access-6vf2x\") pod \"48acdfd7-797c-4545-9dd5-a9087ec4a525\" (UID: \"48acdfd7-797c-4545-9dd5-a9087ec4a525\") " Oct 08 18:30:32 crc kubenswrapper[4988]: I1008 18:30:32.362083 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/48acdfd7-797c-4545-9dd5-a9087ec4a525-config-data-custom\") pod \"48acdfd7-797c-4545-9dd5-a9087ec4a525\" (UID: \"48acdfd7-797c-4545-9dd5-a9087ec4a525\") " Oct 08 18:30:32 crc kubenswrapper[4988]: I1008 18:30:32.363188 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/48acdfd7-797c-4545-9dd5-a9087ec4a525-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "48acdfd7-797c-4545-9dd5-a9087ec4a525" (UID: "48acdfd7-797c-4545-9dd5-a9087ec4a525"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 08 18:30:32 crc kubenswrapper[4988]: I1008 18:30:32.375507 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/48acdfd7-797c-4545-9dd5-a9087ec4a525-kube-api-access-6vf2x" (OuterVolumeSpecName: "kube-api-access-6vf2x") pod "48acdfd7-797c-4545-9dd5-a9087ec4a525" (UID: "48acdfd7-797c-4545-9dd5-a9087ec4a525"). InnerVolumeSpecName "kube-api-access-6vf2x". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:30:32 crc kubenswrapper[4988]: I1008 18:30:32.386527 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/48acdfd7-797c-4545-9dd5-a9087ec4a525-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "48acdfd7-797c-4545-9dd5-a9087ec4a525" (UID: "48acdfd7-797c-4545-9dd5-a9087ec4a525"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:30:32 crc kubenswrapper[4988]: I1008 18:30:32.400077 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-8554744ccd-jsq45" Oct 08 18:30:32 crc kubenswrapper[4988]: I1008 18:30:32.400458 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/48acdfd7-797c-4545-9dd5-a9087ec4a525-scripts" (OuterVolumeSpecName: "scripts") pod "48acdfd7-797c-4545-9dd5-a9087ec4a525" (UID: "48acdfd7-797c-4545-9dd5-a9087ec4a525"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:30:32 crc kubenswrapper[4988]: I1008 18:30:32.464623 4988 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/48acdfd7-797c-4545-9dd5-a9087ec4a525-config-data-custom\") on node \"crc\" DevicePath \"\"" Oct 08 18:30:32 crc kubenswrapper[4988]: I1008 18:30:32.464834 4988 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/48acdfd7-797c-4545-9dd5-a9087ec4a525-etc-machine-id\") on node \"crc\" DevicePath \"\"" Oct 08 18:30:32 crc kubenswrapper[4988]: I1008 18:30:32.464913 4988 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/48acdfd7-797c-4545-9dd5-a9087ec4a525-scripts\") on node \"crc\" DevicePath \"\"" Oct 08 18:30:32 crc kubenswrapper[4988]: I1008 18:30:32.464986 4988 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6vf2x\" (UniqueName: \"kubernetes.io/projected/48acdfd7-797c-4545-9dd5-a9087ec4a525-kube-api-access-6vf2x\") on node \"crc\" DevicePath \"\"" Oct 08 18:30:32 crc kubenswrapper[4988]: I1008 18:30:32.482531 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/48acdfd7-797c-4545-9dd5-a9087ec4a525-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "48acdfd7-797c-4545-9dd5-a9087ec4a525" (UID: "48acdfd7-797c-4545-9dd5-a9087ec4a525"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:30:32 crc kubenswrapper[4988]: I1008 18:30:32.547294 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/48acdfd7-797c-4545-9dd5-a9087ec4a525-config-data" (OuterVolumeSpecName: "config-data") pod "48acdfd7-797c-4545-9dd5-a9087ec4a525" (UID: "48acdfd7-797c-4545-9dd5-a9087ec4a525"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:30:32 crc kubenswrapper[4988]: I1008 18:30:32.566686 4988 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/48acdfd7-797c-4545-9dd5-a9087ec4a525-config-data\") on node \"crc\" DevicePath \"\"" Oct 08 18:30:32 crc kubenswrapper[4988]: I1008 18:30:32.566721 4988 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/48acdfd7-797c-4545-9dd5-a9087ec4a525-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 18:30:32 crc kubenswrapper[4988]: I1008 18:30:32.930003 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"48acdfd7-797c-4545-9dd5-a9087ec4a525","Type":"ContainerDied","Data":"55fd61079ab74343b42def895b389ce2a94d829d4515de632dedb5b8eaea5519"} Oct 08 18:30:32 crc kubenswrapper[4988]: I1008 18:30:32.930444 4988 scope.go:117] "RemoveContainer" containerID="9fbef399ac4284687057dc2c47897eb4be0635cba1dc233289a1f03bfaa51bf3" Oct 08 18:30:32 crc kubenswrapper[4988]: I1008 18:30:32.930037 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Oct 08 18:30:32 crc kubenswrapper[4988]: I1008 18:30:32.956125 4988 scope.go:117] "RemoveContainer" containerID="9b7da6032d176b505a5dd95802130a58e053c8b0dbf445c13c049d3deee7e1cf" Oct 08 18:30:32 crc kubenswrapper[4988]: I1008 18:30:32.974169 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 08 18:30:33 crc kubenswrapper[4988]: I1008 18:30:33.002696 4988 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 08 18:30:33 crc kubenswrapper[4988]: I1008 18:30:33.022825 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Oct 08 18:30:33 crc kubenswrapper[4988]: E1008 18:30:33.029797 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="48acdfd7-797c-4545-9dd5-a9087ec4a525" containerName="cinder-scheduler" Oct 08 18:30:33 crc kubenswrapper[4988]: I1008 18:30:33.029840 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="48acdfd7-797c-4545-9dd5-a9087ec4a525" containerName="cinder-scheduler" Oct 08 18:30:33 crc kubenswrapper[4988]: E1008 18:30:33.029878 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="48acdfd7-797c-4545-9dd5-a9087ec4a525" containerName="probe" Oct 08 18:30:33 crc kubenswrapper[4988]: I1008 18:30:33.029890 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="48acdfd7-797c-4545-9dd5-a9087ec4a525" containerName="probe" Oct 08 18:30:33 crc kubenswrapper[4988]: I1008 18:30:33.030085 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="48acdfd7-797c-4545-9dd5-a9087ec4a525" containerName="probe" Oct 08 18:30:33 crc kubenswrapper[4988]: I1008 18:30:33.030109 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="48acdfd7-797c-4545-9dd5-a9087ec4a525" containerName="cinder-scheduler" Oct 08 18:30:33 crc kubenswrapper[4988]: I1008 18:30:33.031303 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Oct 08 18:30:33 crc kubenswrapper[4988]: I1008 18:30:33.034704 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Oct 08 18:30:33 crc kubenswrapper[4988]: I1008 18:30:33.035129 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 08 18:30:33 crc kubenswrapper[4988]: I1008 18:30:33.074828 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9glwn\" (UniqueName: \"kubernetes.io/projected/0ec345a3-1744-4e78-bf13-91f0325296e4-kube-api-access-9glwn\") pod \"cinder-scheduler-0\" (UID: \"0ec345a3-1744-4e78-bf13-91f0325296e4\") " pod="openstack/cinder-scheduler-0" Oct 08 18:30:33 crc kubenswrapper[4988]: I1008 18:30:33.074992 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/0ec345a3-1744-4e78-bf13-91f0325296e4-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"0ec345a3-1744-4e78-bf13-91f0325296e4\") " pod="openstack/cinder-scheduler-0" Oct 08 18:30:33 crc kubenswrapper[4988]: I1008 18:30:33.075048 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0ec345a3-1744-4e78-bf13-91f0325296e4-config-data\") pod \"cinder-scheduler-0\" (UID: \"0ec345a3-1744-4e78-bf13-91f0325296e4\") " pod="openstack/cinder-scheduler-0" Oct 08 18:30:33 crc kubenswrapper[4988]: I1008 18:30:33.075120 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0ec345a3-1744-4e78-bf13-91f0325296e4-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"0ec345a3-1744-4e78-bf13-91f0325296e4\") " pod="openstack/cinder-scheduler-0" Oct 08 18:30:33 crc kubenswrapper[4988]: I1008 18:30:33.075172 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/0ec345a3-1744-4e78-bf13-91f0325296e4-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"0ec345a3-1744-4e78-bf13-91f0325296e4\") " pod="openstack/cinder-scheduler-0" Oct 08 18:30:33 crc kubenswrapper[4988]: I1008 18:30:33.075269 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0ec345a3-1744-4e78-bf13-91f0325296e4-scripts\") pod \"cinder-scheduler-0\" (UID: \"0ec345a3-1744-4e78-bf13-91f0325296e4\") " pod="openstack/cinder-scheduler-0" Oct 08 18:30:33 crc kubenswrapper[4988]: I1008 18:30:33.176642 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0ec345a3-1744-4e78-bf13-91f0325296e4-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"0ec345a3-1744-4e78-bf13-91f0325296e4\") " pod="openstack/cinder-scheduler-0" Oct 08 18:30:33 crc kubenswrapper[4988]: I1008 18:30:33.176702 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/0ec345a3-1744-4e78-bf13-91f0325296e4-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"0ec345a3-1744-4e78-bf13-91f0325296e4\") " pod="openstack/cinder-scheduler-0" Oct 08 18:30:33 crc kubenswrapper[4988]: I1008 18:30:33.176745 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0ec345a3-1744-4e78-bf13-91f0325296e4-scripts\") pod \"cinder-scheduler-0\" (UID: \"0ec345a3-1744-4e78-bf13-91f0325296e4\") " pod="openstack/cinder-scheduler-0" Oct 08 18:30:33 crc kubenswrapper[4988]: I1008 18:30:33.176810 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9glwn\" (UniqueName: \"kubernetes.io/projected/0ec345a3-1744-4e78-bf13-91f0325296e4-kube-api-access-9glwn\") pod \"cinder-scheduler-0\" (UID: \"0ec345a3-1744-4e78-bf13-91f0325296e4\") " pod="openstack/cinder-scheduler-0" Oct 08 18:30:33 crc kubenswrapper[4988]: I1008 18:30:33.176864 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/0ec345a3-1744-4e78-bf13-91f0325296e4-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"0ec345a3-1744-4e78-bf13-91f0325296e4\") " pod="openstack/cinder-scheduler-0" Oct 08 18:30:33 crc kubenswrapper[4988]: I1008 18:30:33.176875 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/0ec345a3-1744-4e78-bf13-91f0325296e4-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"0ec345a3-1744-4e78-bf13-91f0325296e4\") " pod="openstack/cinder-scheduler-0" Oct 08 18:30:33 crc kubenswrapper[4988]: I1008 18:30:33.176974 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0ec345a3-1744-4e78-bf13-91f0325296e4-config-data\") pod \"cinder-scheduler-0\" (UID: \"0ec345a3-1744-4e78-bf13-91f0325296e4\") " pod="openstack/cinder-scheduler-0" Oct 08 18:30:33 crc kubenswrapper[4988]: I1008 18:30:33.182992 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0ec345a3-1744-4e78-bf13-91f0325296e4-config-data\") pod \"cinder-scheduler-0\" (UID: \"0ec345a3-1744-4e78-bf13-91f0325296e4\") " pod="openstack/cinder-scheduler-0" Oct 08 18:30:33 crc kubenswrapper[4988]: I1008 18:30:33.195858 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0ec345a3-1744-4e78-bf13-91f0325296e4-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"0ec345a3-1744-4e78-bf13-91f0325296e4\") " pod="openstack/cinder-scheduler-0" Oct 08 18:30:33 crc kubenswrapper[4988]: I1008 18:30:33.196507 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/0ec345a3-1744-4e78-bf13-91f0325296e4-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"0ec345a3-1744-4e78-bf13-91f0325296e4\") " pod="openstack/cinder-scheduler-0" Oct 08 18:30:33 crc kubenswrapper[4988]: I1008 18:30:33.211407 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0ec345a3-1744-4e78-bf13-91f0325296e4-scripts\") pod \"cinder-scheduler-0\" (UID: \"0ec345a3-1744-4e78-bf13-91f0325296e4\") " pod="openstack/cinder-scheduler-0" Oct 08 18:30:33 crc kubenswrapper[4988]: I1008 18:30:33.219041 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9glwn\" (UniqueName: \"kubernetes.io/projected/0ec345a3-1744-4e78-bf13-91f0325296e4-kube-api-access-9glwn\") pod \"cinder-scheduler-0\" (UID: \"0ec345a3-1744-4e78-bf13-91f0325296e4\") " pod="openstack/cinder-scheduler-0" Oct 08 18:30:33 crc kubenswrapper[4988]: I1008 18:30:33.263719 4988 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="48acdfd7-797c-4545-9dd5-a9087ec4a525" path="/var/lib/kubelet/pods/48acdfd7-797c-4545-9dd5-a9087ec4a525/volumes" Oct 08 18:30:33 crc kubenswrapper[4988]: I1008 18:30:33.345837 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Oct 08 18:30:33 crc kubenswrapper[4988]: I1008 18:30:33.815937 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 08 18:30:33 crc kubenswrapper[4988]: W1008 18:30:33.816054 4988 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0ec345a3_1744_4e78_bf13_91f0325296e4.slice/crio-d7041603bf5b31608ccc2efc18fa6bb62e518086af0360055ed352353750dabb WatchSource:0}: Error finding container d7041603bf5b31608ccc2efc18fa6bb62e518086af0360055ed352353750dabb: Status 404 returned error can't find the container with id d7041603bf5b31608ccc2efc18fa6bb62e518086af0360055ed352353750dabb Oct 08 18:30:33 crc kubenswrapper[4988]: I1008 18:30:33.942131 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"1920c047-744f-4365-937d-b2618118f20a","Type":"ContainerStarted","Data":"5d3a10489953cbd8a8a5d887d5c53277c3ed173e80ce163dedf775da865c2188"} Oct 08 18:30:33 crc kubenswrapper[4988]: I1008 18:30:33.942516 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 08 18:30:33 crc kubenswrapper[4988]: I1008 18:30:33.942765 4988 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="1920c047-744f-4365-937d-b2618118f20a" containerName="ceilometer-notification-agent" containerID="cri-o://9cb9807398778f96d1bc49c73ecd4c900ad4814d7e983b782c92e5145fc9490c" gracePeriod=30 Oct 08 18:30:33 crc kubenswrapper[4988]: I1008 18:30:33.942759 4988 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="1920c047-744f-4365-937d-b2618118f20a" containerName="sg-core" containerID="cri-o://f7bcf58fcffdcb57bf35d7ebba08dd99e688d5e331252bf0101612b163458960" gracePeriod=30 Oct 08 18:30:33 crc kubenswrapper[4988]: I1008 18:30:33.942854 4988 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="1920c047-744f-4365-937d-b2618118f20a" containerName="proxy-httpd" containerID="cri-o://5d3a10489953cbd8a8a5d887d5c53277c3ed173e80ce163dedf775da865c2188" gracePeriod=30 Oct 08 18:30:33 crc kubenswrapper[4988]: I1008 18:30:33.942742 4988 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="1920c047-744f-4365-937d-b2618118f20a" containerName="ceilometer-central-agent" containerID="cri-o://50f43557736bb7d729755e3a24c5bffebf5e4b3f810a71a05545b4c5793d7ad5" gracePeriod=30 Oct 08 18:30:33 crc kubenswrapper[4988]: I1008 18:30:33.944827 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"0ec345a3-1744-4e78-bf13-91f0325296e4","Type":"ContainerStarted","Data":"d7041603bf5b31608ccc2efc18fa6bb62e518086af0360055ed352353750dabb"} Oct 08 18:30:34 crc kubenswrapper[4988]: I1008 18:30:34.325203 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/cinder-api-0" Oct 08 18:30:34 crc kubenswrapper[4988]: I1008 18:30:34.347261 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.220893928 podStartE2EDuration="6.347242026s" podCreationTimestamp="2025-10-08 18:30:28 +0000 UTC" firstStartedPulling="2025-10-08 18:30:29.026135605 +0000 UTC m=+1174.475978375" lastFinishedPulling="2025-10-08 18:30:33.152483703 +0000 UTC m=+1178.602326473" observedRunningTime="2025-10-08 18:30:33.967776872 +0000 UTC m=+1179.417619652" watchObservedRunningTime="2025-10-08 18:30:34.347242026 +0000 UTC m=+1179.797084796" Oct 08 18:30:34 crc kubenswrapper[4988]: I1008 18:30:34.960134 4988 generic.go:334] "Generic (PLEG): container finished" podID="1920c047-744f-4365-937d-b2618118f20a" containerID="5d3a10489953cbd8a8a5d887d5c53277c3ed173e80ce163dedf775da865c2188" exitCode=0 Oct 08 18:30:34 crc kubenswrapper[4988]: I1008 18:30:34.960451 4988 generic.go:334] "Generic (PLEG): container finished" podID="1920c047-744f-4365-937d-b2618118f20a" containerID="f7bcf58fcffdcb57bf35d7ebba08dd99e688d5e331252bf0101612b163458960" exitCode=2 Oct 08 18:30:34 crc kubenswrapper[4988]: I1008 18:30:34.960463 4988 generic.go:334] "Generic (PLEG): container finished" podID="1920c047-744f-4365-937d-b2618118f20a" containerID="9cb9807398778f96d1bc49c73ecd4c900ad4814d7e983b782c92e5145fc9490c" exitCode=0 Oct 08 18:30:34 crc kubenswrapper[4988]: I1008 18:30:34.960282 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"1920c047-744f-4365-937d-b2618118f20a","Type":"ContainerDied","Data":"5d3a10489953cbd8a8a5d887d5c53277c3ed173e80ce163dedf775da865c2188"} Oct 08 18:30:34 crc kubenswrapper[4988]: I1008 18:30:34.960532 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"1920c047-744f-4365-937d-b2618118f20a","Type":"ContainerDied","Data":"f7bcf58fcffdcb57bf35d7ebba08dd99e688d5e331252bf0101612b163458960"} Oct 08 18:30:34 crc kubenswrapper[4988]: I1008 18:30:34.960551 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"1920c047-744f-4365-937d-b2618118f20a","Type":"ContainerDied","Data":"9cb9807398778f96d1bc49c73ecd4c900ad4814d7e983b782c92e5145fc9490c"} Oct 08 18:30:34 crc kubenswrapper[4988]: I1008 18:30:34.962941 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"0ec345a3-1744-4e78-bf13-91f0325296e4","Type":"ContainerStarted","Data":"590ae1792458f3fd70bbade362261ee71cafb4633235b6af45da637905058707"} Oct 08 18:30:35 crc kubenswrapper[4988]: I1008 18:30:35.452529 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-6bb998f7d5-h4g2x" Oct 08 18:30:35 crc kubenswrapper[4988]: I1008 18:30:35.535699 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-8554744ccd-jsq45"] Oct 08 18:30:35 crc kubenswrapper[4988]: I1008 18:30:35.536072 4988 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-8554744ccd-jsq45" podUID="60b75acf-c263-4e5f-af6e-4195fc6be9e4" containerName="neutron-api" containerID="cri-o://1af046392b464d870001852927bcb817b391578c44077d5241550bdd64b98e22" gracePeriod=30 Oct 08 18:30:35 crc kubenswrapper[4988]: I1008 18:30:35.536626 4988 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-8554744ccd-jsq45" podUID="60b75acf-c263-4e5f-af6e-4195fc6be9e4" containerName="neutron-httpd" containerID="cri-o://ce7818666ce3c638983016985eb011d4d71fe1d55928608343a134abd73c80de" gracePeriod=30 Oct 08 18:30:36 crc kubenswrapper[4988]: I1008 18:30:36.000105 4988 generic.go:334] "Generic (PLEG): container finished" podID="60b75acf-c263-4e5f-af6e-4195fc6be9e4" containerID="ce7818666ce3c638983016985eb011d4d71fe1d55928608343a134abd73c80de" exitCode=0 Oct 08 18:30:36 crc kubenswrapper[4988]: I1008 18:30:36.000687 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-8554744ccd-jsq45" event={"ID":"60b75acf-c263-4e5f-af6e-4195fc6be9e4","Type":"ContainerDied","Data":"ce7818666ce3c638983016985eb011d4d71fe1d55928608343a134abd73c80de"} Oct 08 18:30:36 crc kubenswrapper[4988]: I1008 18:30:36.007486 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-ebe4-account-create-c64w6"] Oct 08 18:30:36 crc kubenswrapper[4988]: I1008 18:30:36.008944 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-ebe4-account-create-c64w6" Oct 08 18:30:36 crc kubenswrapper[4988]: I1008 18:30:36.012120 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-db-secret" Oct 08 18:30:36 crc kubenswrapper[4988]: I1008 18:30:36.015415 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-ebe4-account-create-c64w6"] Oct 08 18:30:36 crc kubenswrapper[4988]: I1008 18:30:36.030837 4988 generic.go:334] "Generic (PLEG): container finished" podID="1920c047-744f-4365-937d-b2618118f20a" containerID="50f43557736bb7d729755e3a24c5bffebf5e4b3f810a71a05545b4c5793d7ad5" exitCode=0 Oct 08 18:30:36 crc kubenswrapper[4988]: I1008 18:30:36.030895 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"1920c047-744f-4365-937d-b2618118f20a","Type":"ContainerDied","Data":"50f43557736bb7d729755e3a24c5bffebf5e4b3f810a71a05545b4c5793d7ad5"} Oct 08 18:30:36 crc kubenswrapper[4988]: I1008 18:30:36.035143 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"0ec345a3-1744-4e78-bf13-91f0325296e4","Type":"ContainerStarted","Data":"19c72634c92910f59b5f356f56971cc3e2867c6c02b8cdd4a997a62cb452e3cb"} Oct 08 18:30:36 crc kubenswrapper[4988]: I1008 18:30:36.055893 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=4.055877638 podStartE2EDuration="4.055877638s" podCreationTimestamp="2025-10-08 18:30:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 18:30:36.052892232 +0000 UTC m=+1181.502735002" watchObservedRunningTime="2025-10-08 18:30:36.055877638 +0000 UTC m=+1181.505720408" Oct 08 18:30:36 crc kubenswrapper[4988]: I1008 18:30:36.059238 4988 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Oct 08 18:30:36 crc kubenswrapper[4988]: I1008 18:30:36.059274 4988 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Oct 08 18:30:36 crc kubenswrapper[4988]: I1008 18:30:36.108939 4988 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Oct 08 18:30:36 crc kubenswrapper[4988]: I1008 18:30:36.109373 4988 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Oct 08 18:30:36 crc kubenswrapper[4988]: I1008 18:30:36.137089 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zgjf8\" (UniqueName: \"kubernetes.io/projected/e3e8eb1b-f0b5-4e56-9ba7-d068435355c5-kube-api-access-zgjf8\") pod \"nova-api-ebe4-account-create-c64w6\" (UID: \"e3e8eb1b-f0b5-4e56-9ba7-d068435355c5\") " pod="openstack/nova-api-ebe4-account-create-c64w6" Oct 08 18:30:36 crc kubenswrapper[4988]: I1008 18:30:36.179346 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 08 18:30:36 crc kubenswrapper[4988]: I1008 18:30:36.192494 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-2f70-account-create-tdkwz"] Oct 08 18:30:36 crc kubenswrapper[4988]: E1008 18:30:36.192897 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1920c047-744f-4365-937d-b2618118f20a" containerName="proxy-httpd" Oct 08 18:30:36 crc kubenswrapper[4988]: I1008 18:30:36.192916 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="1920c047-744f-4365-937d-b2618118f20a" containerName="proxy-httpd" Oct 08 18:30:36 crc kubenswrapper[4988]: E1008 18:30:36.192934 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1920c047-744f-4365-937d-b2618118f20a" containerName="sg-core" Oct 08 18:30:36 crc kubenswrapper[4988]: I1008 18:30:36.192940 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="1920c047-744f-4365-937d-b2618118f20a" containerName="sg-core" Oct 08 18:30:36 crc kubenswrapper[4988]: E1008 18:30:36.192958 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1920c047-744f-4365-937d-b2618118f20a" containerName="ceilometer-central-agent" Oct 08 18:30:36 crc kubenswrapper[4988]: I1008 18:30:36.192965 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="1920c047-744f-4365-937d-b2618118f20a" containerName="ceilometer-central-agent" Oct 08 18:30:36 crc kubenswrapper[4988]: E1008 18:30:36.192983 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1920c047-744f-4365-937d-b2618118f20a" containerName="ceilometer-notification-agent" Oct 08 18:30:36 crc kubenswrapper[4988]: I1008 18:30:36.192989 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="1920c047-744f-4365-937d-b2618118f20a" containerName="ceilometer-notification-agent" Oct 08 18:30:36 crc kubenswrapper[4988]: I1008 18:30:36.193151 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="1920c047-744f-4365-937d-b2618118f20a" containerName="proxy-httpd" Oct 08 18:30:36 crc kubenswrapper[4988]: I1008 18:30:36.193170 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="1920c047-744f-4365-937d-b2618118f20a" containerName="sg-core" Oct 08 18:30:36 crc kubenswrapper[4988]: I1008 18:30:36.193183 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="1920c047-744f-4365-937d-b2618118f20a" containerName="ceilometer-central-agent" Oct 08 18:30:36 crc kubenswrapper[4988]: I1008 18:30:36.193192 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="1920c047-744f-4365-937d-b2618118f20a" containerName="ceilometer-notification-agent" Oct 08 18:30:36 crc kubenswrapper[4988]: I1008 18:30:36.193830 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-2f70-account-create-tdkwz" Oct 08 18:30:36 crc kubenswrapper[4988]: I1008 18:30:36.197572 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-db-secret" Oct 08 18:30:36 crc kubenswrapper[4988]: I1008 18:30:36.223275 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-2f70-account-create-tdkwz"] Oct 08 18:30:36 crc kubenswrapper[4988]: I1008 18:30:36.237973 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-khgnt\" (UniqueName: \"kubernetes.io/projected/1920c047-744f-4365-937d-b2618118f20a-kube-api-access-khgnt\") pod \"1920c047-744f-4365-937d-b2618118f20a\" (UID: \"1920c047-744f-4365-937d-b2618118f20a\") " Oct 08 18:30:36 crc kubenswrapper[4988]: I1008 18:30:36.238360 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1920c047-744f-4365-937d-b2618118f20a-combined-ca-bundle\") pod \"1920c047-744f-4365-937d-b2618118f20a\" (UID: \"1920c047-744f-4365-937d-b2618118f20a\") " Oct 08 18:30:36 crc kubenswrapper[4988]: I1008 18:30:36.238483 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1920c047-744f-4365-937d-b2618118f20a-config-data\") pod \"1920c047-744f-4365-937d-b2618118f20a\" (UID: \"1920c047-744f-4365-937d-b2618118f20a\") " Oct 08 18:30:36 crc kubenswrapper[4988]: I1008 18:30:36.238649 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1920c047-744f-4365-937d-b2618118f20a-scripts\") pod \"1920c047-744f-4365-937d-b2618118f20a\" (UID: \"1920c047-744f-4365-937d-b2618118f20a\") " Oct 08 18:30:36 crc kubenswrapper[4988]: I1008 18:30:36.239110 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1920c047-744f-4365-937d-b2618118f20a-run-httpd\") pod \"1920c047-744f-4365-937d-b2618118f20a\" (UID: \"1920c047-744f-4365-937d-b2618118f20a\") " Oct 08 18:30:36 crc kubenswrapper[4988]: I1008 18:30:36.239229 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1920c047-744f-4365-937d-b2618118f20a-log-httpd\") pod \"1920c047-744f-4365-937d-b2618118f20a\" (UID: \"1920c047-744f-4365-937d-b2618118f20a\") " Oct 08 18:30:36 crc kubenswrapper[4988]: I1008 18:30:36.239360 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/1920c047-744f-4365-937d-b2618118f20a-sg-core-conf-yaml\") pod \"1920c047-744f-4365-937d-b2618118f20a\" (UID: \"1920c047-744f-4365-937d-b2618118f20a\") " Oct 08 18:30:36 crc kubenswrapper[4988]: I1008 18:30:36.239436 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1920c047-744f-4365-937d-b2618118f20a-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "1920c047-744f-4365-937d-b2618118f20a" (UID: "1920c047-744f-4365-937d-b2618118f20a"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 18:30:36 crc kubenswrapper[4988]: I1008 18:30:36.239731 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1920c047-744f-4365-937d-b2618118f20a-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "1920c047-744f-4365-937d-b2618118f20a" (UID: "1920c047-744f-4365-937d-b2618118f20a"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 18:30:36 crc kubenswrapper[4988]: I1008 18:30:36.239898 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9tkpb\" (UniqueName: \"kubernetes.io/projected/5f273f27-4991-4aa3-a734-dec827cfb78b-kube-api-access-9tkpb\") pod \"nova-cell0-2f70-account-create-tdkwz\" (UID: \"5f273f27-4991-4aa3-a734-dec827cfb78b\") " pod="openstack/nova-cell0-2f70-account-create-tdkwz" Oct 08 18:30:36 crc kubenswrapper[4988]: I1008 18:30:36.240029 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zgjf8\" (UniqueName: \"kubernetes.io/projected/e3e8eb1b-f0b5-4e56-9ba7-d068435355c5-kube-api-access-zgjf8\") pod \"nova-api-ebe4-account-create-c64w6\" (UID: \"e3e8eb1b-f0b5-4e56-9ba7-d068435355c5\") " pod="openstack/nova-api-ebe4-account-create-c64w6" Oct 08 18:30:36 crc kubenswrapper[4988]: I1008 18:30:36.240254 4988 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1920c047-744f-4365-937d-b2618118f20a-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 08 18:30:36 crc kubenswrapper[4988]: I1008 18:30:36.240336 4988 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1920c047-744f-4365-937d-b2618118f20a-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 08 18:30:36 crc kubenswrapper[4988]: I1008 18:30:36.249515 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1920c047-744f-4365-937d-b2618118f20a-scripts" (OuterVolumeSpecName: "scripts") pod "1920c047-744f-4365-937d-b2618118f20a" (UID: "1920c047-744f-4365-937d-b2618118f20a"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:30:36 crc kubenswrapper[4988]: I1008 18:30:36.257532 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1920c047-744f-4365-937d-b2618118f20a-kube-api-access-khgnt" (OuterVolumeSpecName: "kube-api-access-khgnt") pod "1920c047-744f-4365-937d-b2618118f20a" (UID: "1920c047-744f-4365-937d-b2618118f20a"). InnerVolumeSpecName "kube-api-access-khgnt". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:30:36 crc kubenswrapper[4988]: I1008 18:30:36.263127 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zgjf8\" (UniqueName: \"kubernetes.io/projected/e3e8eb1b-f0b5-4e56-9ba7-d068435355c5-kube-api-access-zgjf8\") pod \"nova-api-ebe4-account-create-c64w6\" (UID: \"e3e8eb1b-f0b5-4e56-9ba7-d068435355c5\") " pod="openstack/nova-api-ebe4-account-create-c64w6" Oct 08 18:30:36 crc kubenswrapper[4988]: I1008 18:30:36.268576 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1920c047-744f-4365-937d-b2618118f20a-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "1920c047-744f-4365-937d-b2618118f20a" (UID: "1920c047-744f-4365-937d-b2618118f20a"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:30:36 crc kubenswrapper[4988]: I1008 18:30:36.318659 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1920c047-744f-4365-937d-b2618118f20a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "1920c047-744f-4365-937d-b2618118f20a" (UID: "1920c047-744f-4365-937d-b2618118f20a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:30:36 crc kubenswrapper[4988]: I1008 18:30:36.341570 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9tkpb\" (UniqueName: \"kubernetes.io/projected/5f273f27-4991-4aa3-a734-dec827cfb78b-kube-api-access-9tkpb\") pod \"nova-cell0-2f70-account-create-tdkwz\" (UID: \"5f273f27-4991-4aa3-a734-dec827cfb78b\") " pod="openstack/nova-cell0-2f70-account-create-tdkwz" Oct 08 18:30:36 crc kubenswrapper[4988]: I1008 18:30:36.341702 4988 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/1920c047-744f-4365-937d-b2618118f20a-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Oct 08 18:30:36 crc kubenswrapper[4988]: I1008 18:30:36.341715 4988 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-khgnt\" (UniqueName: \"kubernetes.io/projected/1920c047-744f-4365-937d-b2618118f20a-kube-api-access-khgnt\") on node \"crc\" DevicePath \"\"" Oct 08 18:30:36 crc kubenswrapper[4988]: I1008 18:30:36.341725 4988 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1920c047-744f-4365-937d-b2618118f20a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 18:30:36 crc kubenswrapper[4988]: I1008 18:30:36.341733 4988 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1920c047-744f-4365-937d-b2618118f20a-scripts\") on node \"crc\" DevicePath \"\"" Oct 08 18:30:36 crc kubenswrapper[4988]: I1008 18:30:36.344507 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-ebe4-account-create-c64w6" Oct 08 18:30:36 crc kubenswrapper[4988]: I1008 18:30:36.349139 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1920c047-744f-4365-937d-b2618118f20a-config-data" (OuterVolumeSpecName: "config-data") pod "1920c047-744f-4365-937d-b2618118f20a" (UID: "1920c047-744f-4365-937d-b2618118f20a"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:30:36 crc kubenswrapper[4988]: I1008 18:30:36.357907 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9tkpb\" (UniqueName: \"kubernetes.io/projected/5f273f27-4991-4aa3-a734-dec827cfb78b-kube-api-access-9tkpb\") pod \"nova-cell0-2f70-account-create-tdkwz\" (UID: \"5f273f27-4991-4aa3-a734-dec827cfb78b\") " pod="openstack/nova-cell0-2f70-account-create-tdkwz" Oct 08 18:30:36 crc kubenswrapper[4988]: I1008 18:30:36.397626 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-4316-account-create-5kjmj"] Oct 08 18:30:36 crc kubenswrapper[4988]: I1008 18:30:36.398930 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-4316-account-create-5kjmj" Oct 08 18:30:36 crc kubenswrapper[4988]: I1008 18:30:36.403424 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-db-secret" Oct 08 18:30:36 crc kubenswrapper[4988]: I1008 18:30:36.421715 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-4316-account-create-5kjmj"] Oct 08 18:30:36 crc kubenswrapper[4988]: I1008 18:30:36.445046 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dcst5\" (UniqueName: \"kubernetes.io/projected/0fd89f92-b9aa-4a7b-a3ed-130abc4ce161-kube-api-access-dcst5\") pod \"nova-cell1-4316-account-create-5kjmj\" (UID: \"0fd89f92-b9aa-4a7b-a3ed-130abc4ce161\") " pod="openstack/nova-cell1-4316-account-create-5kjmj" Oct 08 18:30:36 crc kubenswrapper[4988]: I1008 18:30:36.445187 4988 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1920c047-744f-4365-937d-b2618118f20a-config-data\") on node \"crc\" DevicePath \"\"" Oct 08 18:30:36 crc kubenswrapper[4988]: I1008 18:30:36.524073 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-2f70-account-create-tdkwz" Oct 08 18:30:36 crc kubenswrapper[4988]: I1008 18:30:36.547625 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dcst5\" (UniqueName: \"kubernetes.io/projected/0fd89f92-b9aa-4a7b-a3ed-130abc4ce161-kube-api-access-dcst5\") pod \"nova-cell1-4316-account-create-5kjmj\" (UID: \"0fd89f92-b9aa-4a7b-a3ed-130abc4ce161\") " pod="openstack/nova-cell1-4316-account-create-5kjmj" Oct 08 18:30:36 crc kubenswrapper[4988]: I1008 18:30:36.567004 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dcst5\" (UniqueName: \"kubernetes.io/projected/0fd89f92-b9aa-4a7b-a3ed-130abc4ce161-kube-api-access-dcst5\") pod \"nova-cell1-4316-account-create-5kjmj\" (UID: \"0fd89f92-b9aa-4a7b-a3ed-130abc4ce161\") " pod="openstack/nova-cell1-4316-account-create-5kjmj" Oct 08 18:30:36 crc kubenswrapper[4988]: I1008 18:30:36.727319 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-4316-account-create-5kjmj" Oct 08 18:30:36 crc kubenswrapper[4988]: I1008 18:30:36.839877 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-ebe4-account-create-c64w6"] Oct 08 18:30:36 crc kubenswrapper[4988]: I1008 18:30:36.985030 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-2f70-account-create-tdkwz"] Oct 08 18:30:37 crc kubenswrapper[4988]: I1008 18:30:37.045612 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-ebe4-account-create-c64w6" event={"ID":"e3e8eb1b-f0b5-4e56-9ba7-d068435355c5","Type":"ContainerStarted","Data":"efcd09d870061017ba284fc9969469d96b6835d1902fdf2b2303c413da0a2a27"} Oct 08 18:30:37 crc kubenswrapper[4988]: I1008 18:30:37.046667 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-2f70-account-create-tdkwz" event={"ID":"5f273f27-4991-4aa3-a734-dec827cfb78b","Type":"ContainerStarted","Data":"47a8a394273b79a84efe27e920d5b8ced645193d3bd45847a88c0713f026ea5e"} Oct 08 18:30:37 crc kubenswrapper[4988]: I1008 18:30:37.049649 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"1920c047-744f-4365-937d-b2618118f20a","Type":"ContainerDied","Data":"e5c62e8aa7d7b5023992aa9dc11e06db48fa32b76a39e6fe765d4a87060fddf3"} Oct 08 18:30:37 crc kubenswrapper[4988]: I1008 18:30:37.049727 4988 scope.go:117] "RemoveContainer" containerID="5d3a10489953cbd8a8a5d887d5c53277c3ed173e80ce163dedf775da865c2188" Oct 08 18:30:37 crc kubenswrapper[4988]: I1008 18:30:37.049741 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 08 18:30:37 crc kubenswrapper[4988]: I1008 18:30:37.050200 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Oct 08 18:30:37 crc kubenswrapper[4988]: I1008 18:30:37.050234 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Oct 08 18:30:37 crc kubenswrapper[4988]: I1008 18:30:37.077508 4988 scope.go:117] "RemoveContainer" containerID="f7bcf58fcffdcb57bf35d7ebba08dd99e688d5e331252bf0101612b163458960" Oct 08 18:30:37 crc kubenswrapper[4988]: I1008 18:30:37.090805 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 08 18:30:37 crc kubenswrapper[4988]: I1008 18:30:37.100486 4988 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Oct 08 18:30:37 crc kubenswrapper[4988]: I1008 18:30:37.124135 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 08 18:30:37 crc kubenswrapper[4988]: I1008 18:30:37.126891 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 08 18:30:37 crc kubenswrapper[4988]: I1008 18:30:37.129519 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 08 18:30:37 crc kubenswrapper[4988]: I1008 18:30:37.142145 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 08 18:30:37 crc kubenswrapper[4988]: I1008 18:30:37.146574 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 08 18:30:37 crc kubenswrapper[4988]: I1008 18:30:37.147253 4988 scope.go:117] "RemoveContainer" containerID="9cb9807398778f96d1bc49c73ecd4c900ad4814d7e983b782c92e5145fc9490c" Oct 08 18:30:37 crc kubenswrapper[4988]: I1008 18:30:37.159864 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/872d504e-cc28-4f84-a509-1e5e3dc809ec-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"872d504e-cc28-4f84-a509-1e5e3dc809ec\") " pod="openstack/ceilometer-0" Oct 08 18:30:37 crc kubenswrapper[4988]: I1008 18:30:37.160559 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2nrhr\" (UniqueName: \"kubernetes.io/projected/872d504e-cc28-4f84-a509-1e5e3dc809ec-kube-api-access-2nrhr\") pod \"ceilometer-0\" (UID: \"872d504e-cc28-4f84-a509-1e5e3dc809ec\") " pod="openstack/ceilometer-0" Oct 08 18:30:37 crc kubenswrapper[4988]: I1008 18:30:37.160657 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/872d504e-cc28-4f84-a509-1e5e3dc809ec-run-httpd\") pod \"ceilometer-0\" (UID: \"872d504e-cc28-4f84-a509-1e5e3dc809ec\") " pod="openstack/ceilometer-0" Oct 08 18:30:37 crc kubenswrapper[4988]: I1008 18:30:37.160789 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/872d504e-cc28-4f84-a509-1e5e3dc809ec-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"872d504e-cc28-4f84-a509-1e5e3dc809ec\") " pod="openstack/ceilometer-0" Oct 08 18:30:37 crc kubenswrapper[4988]: I1008 18:30:37.160907 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/872d504e-cc28-4f84-a509-1e5e3dc809ec-scripts\") pod \"ceilometer-0\" (UID: \"872d504e-cc28-4f84-a509-1e5e3dc809ec\") " pod="openstack/ceilometer-0" Oct 08 18:30:37 crc kubenswrapper[4988]: I1008 18:30:37.160978 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/872d504e-cc28-4f84-a509-1e5e3dc809ec-log-httpd\") pod \"ceilometer-0\" (UID: \"872d504e-cc28-4f84-a509-1e5e3dc809ec\") " pod="openstack/ceilometer-0" Oct 08 18:30:37 crc kubenswrapper[4988]: I1008 18:30:37.161170 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/872d504e-cc28-4f84-a509-1e5e3dc809ec-config-data\") pod \"ceilometer-0\" (UID: \"872d504e-cc28-4f84-a509-1e5e3dc809ec\") " pod="openstack/ceilometer-0" Oct 08 18:30:37 crc kubenswrapper[4988]: I1008 18:30:37.187545 4988 scope.go:117] "RemoveContainer" containerID="50f43557736bb7d729755e3a24c5bffebf5e4b3f810a71a05545b4c5793d7ad5" Oct 08 18:30:37 crc kubenswrapper[4988]: I1008 18:30:37.232790 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-4316-account-create-5kjmj"] Oct 08 18:30:37 crc kubenswrapper[4988]: I1008 18:30:37.256024 4988 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1920c047-744f-4365-937d-b2618118f20a" path="/var/lib/kubelet/pods/1920c047-744f-4365-937d-b2618118f20a/volumes" Oct 08 18:30:37 crc kubenswrapper[4988]: I1008 18:30:37.262757 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/872d504e-cc28-4f84-a509-1e5e3dc809ec-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"872d504e-cc28-4f84-a509-1e5e3dc809ec\") " pod="openstack/ceilometer-0" Oct 08 18:30:37 crc kubenswrapper[4988]: I1008 18:30:37.262824 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2nrhr\" (UniqueName: \"kubernetes.io/projected/872d504e-cc28-4f84-a509-1e5e3dc809ec-kube-api-access-2nrhr\") pod \"ceilometer-0\" (UID: \"872d504e-cc28-4f84-a509-1e5e3dc809ec\") " pod="openstack/ceilometer-0" Oct 08 18:30:37 crc kubenswrapper[4988]: I1008 18:30:37.262850 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/872d504e-cc28-4f84-a509-1e5e3dc809ec-run-httpd\") pod \"ceilometer-0\" (UID: \"872d504e-cc28-4f84-a509-1e5e3dc809ec\") " pod="openstack/ceilometer-0" Oct 08 18:30:37 crc kubenswrapper[4988]: I1008 18:30:37.262905 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/872d504e-cc28-4f84-a509-1e5e3dc809ec-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"872d504e-cc28-4f84-a509-1e5e3dc809ec\") " pod="openstack/ceilometer-0" Oct 08 18:30:37 crc kubenswrapper[4988]: I1008 18:30:37.262978 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/872d504e-cc28-4f84-a509-1e5e3dc809ec-scripts\") pod \"ceilometer-0\" (UID: \"872d504e-cc28-4f84-a509-1e5e3dc809ec\") " pod="openstack/ceilometer-0" Oct 08 18:30:37 crc kubenswrapper[4988]: I1008 18:30:37.263002 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/872d504e-cc28-4f84-a509-1e5e3dc809ec-log-httpd\") pod \"ceilometer-0\" (UID: \"872d504e-cc28-4f84-a509-1e5e3dc809ec\") " pod="openstack/ceilometer-0" Oct 08 18:30:37 crc kubenswrapper[4988]: I1008 18:30:37.263039 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/872d504e-cc28-4f84-a509-1e5e3dc809ec-config-data\") pod \"ceilometer-0\" (UID: \"872d504e-cc28-4f84-a509-1e5e3dc809ec\") " pod="openstack/ceilometer-0" Oct 08 18:30:37 crc kubenswrapper[4988]: I1008 18:30:37.264197 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/872d504e-cc28-4f84-a509-1e5e3dc809ec-run-httpd\") pod \"ceilometer-0\" (UID: \"872d504e-cc28-4f84-a509-1e5e3dc809ec\") " pod="openstack/ceilometer-0" Oct 08 18:30:37 crc kubenswrapper[4988]: I1008 18:30:37.264420 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/872d504e-cc28-4f84-a509-1e5e3dc809ec-log-httpd\") pod \"ceilometer-0\" (UID: \"872d504e-cc28-4f84-a509-1e5e3dc809ec\") " pod="openstack/ceilometer-0" Oct 08 18:30:37 crc kubenswrapper[4988]: I1008 18:30:37.269545 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/872d504e-cc28-4f84-a509-1e5e3dc809ec-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"872d504e-cc28-4f84-a509-1e5e3dc809ec\") " pod="openstack/ceilometer-0" Oct 08 18:30:37 crc kubenswrapper[4988]: I1008 18:30:37.270992 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/872d504e-cc28-4f84-a509-1e5e3dc809ec-config-data\") pod \"ceilometer-0\" (UID: \"872d504e-cc28-4f84-a509-1e5e3dc809ec\") " pod="openstack/ceilometer-0" Oct 08 18:30:37 crc kubenswrapper[4988]: I1008 18:30:37.273602 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/872d504e-cc28-4f84-a509-1e5e3dc809ec-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"872d504e-cc28-4f84-a509-1e5e3dc809ec\") " pod="openstack/ceilometer-0" Oct 08 18:30:37 crc kubenswrapper[4988]: I1008 18:30:37.283500 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2nrhr\" (UniqueName: \"kubernetes.io/projected/872d504e-cc28-4f84-a509-1e5e3dc809ec-kube-api-access-2nrhr\") pod \"ceilometer-0\" (UID: \"872d504e-cc28-4f84-a509-1e5e3dc809ec\") " pod="openstack/ceilometer-0" Oct 08 18:30:37 crc kubenswrapper[4988]: I1008 18:30:37.284953 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/872d504e-cc28-4f84-a509-1e5e3dc809ec-scripts\") pod \"ceilometer-0\" (UID: \"872d504e-cc28-4f84-a509-1e5e3dc809ec\") " pod="openstack/ceilometer-0" Oct 08 18:30:37 crc kubenswrapper[4988]: I1008 18:30:37.468463 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 08 18:30:37 crc kubenswrapper[4988]: I1008 18:30:37.933799 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 08 18:30:38 crc kubenswrapper[4988]: I1008 18:30:38.059077 4988 generic.go:334] "Generic (PLEG): container finished" podID="0fd89f92-b9aa-4a7b-a3ed-130abc4ce161" containerID="415de07802af7873cf01a1c532b1c1de922fc28c756bb42991542444c0b9ed64" exitCode=0 Oct 08 18:30:38 crc kubenswrapper[4988]: I1008 18:30:38.059470 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-4316-account-create-5kjmj" event={"ID":"0fd89f92-b9aa-4a7b-a3ed-130abc4ce161","Type":"ContainerDied","Data":"415de07802af7873cf01a1c532b1c1de922fc28c756bb42991542444c0b9ed64"} Oct 08 18:30:38 crc kubenswrapper[4988]: I1008 18:30:38.059504 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-4316-account-create-5kjmj" event={"ID":"0fd89f92-b9aa-4a7b-a3ed-130abc4ce161","Type":"ContainerStarted","Data":"c9fb7fe6f640ffac5cd31525c845960828be0df1fd543461b251b855af48aec6"} Oct 08 18:30:38 crc kubenswrapper[4988]: I1008 18:30:38.061506 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"872d504e-cc28-4f84-a509-1e5e3dc809ec","Type":"ContainerStarted","Data":"5ad380b52cd6782b558d4e85f1ef9348a5a24e4503f9480d14c2f9680479466c"} Oct 08 18:30:38 crc kubenswrapper[4988]: I1008 18:30:38.064400 4988 generic.go:334] "Generic (PLEG): container finished" podID="e3e8eb1b-f0b5-4e56-9ba7-d068435355c5" containerID="5c2684680583eebae32531e937601c8903ed775d6ac916679e0120825b5358f8" exitCode=0 Oct 08 18:30:38 crc kubenswrapper[4988]: I1008 18:30:38.064470 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-ebe4-account-create-c64w6" event={"ID":"e3e8eb1b-f0b5-4e56-9ba7-d068435355c5","Type":"ContainerDied","Data":"5c2684680583eebae32531e937601c8903ed775d6ac916679e0120825b5358f8"} Oct 08 18:30:38 crc kubenswrapper[4988]: I1008 18:30:38.066080 4988 generic.go:334] "Generic (PLEG): container finished" podID="5f273f27-4991-4aa3-a734-dec827cfb78b" containerID="b532a4185e833359abeaceb113ab2b9b1e3c1758139d29685926ea848c821a01" exitCode=0 Oct 08 18:30:38 crc kubenswrapper[4988]: I1008 18:30:38.066866 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-2f70-account-create-tdkwz" event={"ID":"5f273f27-4991-4aa3-a734-dec827cfb78b","Type":"ContainerDied","Data":"b532a4185e833359abeaceb113ab2b9b1e3c1758139d29685926ea848c821a01"} Oct 08 18:30:38 crc kubenswrapper[4988]: I1008 18:30:38.346885 4988 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Oct 08 18:30:39 crc kubenswrapper[4988]: I1008 18:30:39.078297 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"872d504e-cc28-4f84-a509-1e5e3dc809ec","Type":"ContainerStarted","Data":"d72be3d5571a4599f581e687b5b5ea86da6f84ff1b611e46cd3f46b526852946"} Oct 08 18:30:39 crc kubenswrapper[4988]: I1008 18:30:39.138204 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Oct 08 18:30:39 crc kubenswrapper[4988]: I1008 18:30:39.138296 4988 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 08 18:30:39 crc kubenswrapper[4988]: I1008 18:30:39.139678 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Oct 08 18:30:39 crc kubenswrapper[4988]: I1008 18:30:39.176557 4988 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Oct 08 18:30:39 crc kubenswrapper[4988]: I1008 18:30:39.177473 4988 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Oct 08 18:30:39 crc kubenswrapper[4988]: I1008 18:30:39.299233 4988 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Oct 08 18:30:39 crc kubenswrapper[4988]: I1008 18:30:39.302646 4988 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Oct 08 18:30:39 crc kubenswrapper[4988]: I1008 18:30:39.533899 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-ebe4-account-create-c64w6" Oct 08 18:30:39 crc kubenswrapper[4988]: I1008 18:30:39.633708 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zgjf8\" (UniqueName: \"kubernetes.io/projected/e3e8eb1b-f0b5-4e56-9ba7-d068435355c5-kube-api-access-zgjf8\") pod \"e3e8eb1b-f0b5-4e56-9ba7-d068435355c5\" (UID: \"e3e8eb1b-f0b5-4e56-9ba7-d068435355c5\") " Oct 08 18:30:39 crc kubenswrapper[4988]: I1008 18:30:39.640682 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e3e8eb1b-f0b5-4e56-9ba7-d068435355c5-kube-api-access-zgjf8" (OuterVolumeSpecName: "kube-api-access-zgjf8") pod "e3e8eb1b-f0b5-4e56-9ba7-d068435355c5" (UID: "e3e8eb1b-f0b5-4e56-9ba7-d068435355c5"). InnerVolumeSpecName "kube-api-access-zgjf8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:30:39 crc kubenswrapper[4988]: I1008 18:30:39.736218 4988 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zgjf8\" (UniqueName: \"kubernetes.io/projected/e3e8eb1b-f0b5-4e56-9ba7-d068435355c5-kube-api-access-zgjf8\") on node \"crc\" DevicePath \"\"" Oct 08 18:30:39 crc kubenswrapper[4988]: I1008 18:30:39.741618 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-4316-account-create-5kjmj" Oct 08 18:30:39 crc kubenswrapper[4988]: I1008 18:30:39.748321 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-2f70-account-create-tdkwz" Oct 08 18:30:39 crc kubenswrapper[4988]: I1008 18:30:39.845299 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9tkpb\" (UniqueName: \"kubernetes.io/projected/5f273f27-4991-4aa3-a734-dec827cfb78b-kube-api-access-9tkpb\") pod \"5f273f27-4991-4aa3-a734-dec827cfb78b\" (UID: \"5f273f27-4991-4aa3-a734-dec827cfb78b\") " Oct 08 18:30:39 crc kubenswrapper[4988]: I1008 18:30:39.845536 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dcst5\" (UniqueName: \"kubernetes.io/projected/0fd89f92-b9aa-4a7b-a3ed-130abc4ce161-kube-api-access-dcst5\") pod \"0fd89f92-b9aa-4a7b-a3ed-130abc4ce161\" (UID: \"0fd89f92-b9aa-4a7b-a3ed-130abc4ce161\") " Oct 08 18:30:39 crc kubenswrapper[4988]: I1008 18:30:39.850667 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0fd89f92-b9aa-4a7b-a3ed-130abc4ce161-kube-api-access-dcst5" (OuterVolumeSpecName: "kube-api-access-dcst5") pod "0fd89f92-b9aa-4a7b-a3ed-130abc4ce161" (UID: "0fd89f92-b9aa-4a7b-a3ed-130abc4ce161"). InnerVolumeSpecName "kube-api-access-dcst5". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:30:39 crc kubenswrapper[4988]: I1008 18:30:39.851892 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5f273f27-4991-4aa3-a734-dec827cfb78b-kube-api-access-9tkpb" (OuterVolumeSpecName: "kube-api-access-9tkpb") pod "5f273f27-4991-4aa3-a734-dec827cfb78b" (UID: "5f273f27-4991-4aa3-a734-dec827cfb78b"). InnerVolumeSpecName "kube-api-access-9tkpb". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:30:39 crc kubenswrapper[4988]: I1008 18:30:39.947893 4988 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dcst5\" (UniqueName: \"kubernetes.io/projected/0fd89f92-b9aa-4a7b-a3ed-130abc4ce161-kube-api-access-dcst5\") on node \"crc\" DevicePath \"\"" Oct 08 18:30:39 crc kubenswrapper[4988]: I1008 18:30:39.948130 4988 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9tkpb\" (UniqueName: \"kubernetes.io/projected/5f273f27-4991-4aa3-a734-dec827cfb78b-kube-api-access-9tkpb\") on node \"crc\" DevicePath \"\"" Oct 08 18:30:40 crc kubenswrapper[4988]: I1008 18:30:40.086856 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-ebe4-account-create-c64w6" event={"ID":"e3e8eb1b-f0b5-4e56-9ba7-d068435355c5","Type":"ContainerDied","Data":"efcd09d870061017ba284fc9969469d96b6835d1902fdf2b2303c413da0a2a27"} Oct 08 18:30:40 crc kubenswrapper[4988]: I1008 18:30:40.087305 4988 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="efcd09d870061017ba284fc9969469d96b6835d1902fdf2b2303c413da0a2a27" Oct 08 18:30:40 crc kubenswrapper[4988]: I1008 18:30:40.087374 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-ebe4-account-create-c64w6" Oct 08 18:30:40 crc kubenswrapper[4988]: I1008 18:30:40.092310 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-2f70-account-create-tdkwz" event={"ID":"5f273f27-4991-4aa3-a734-dec827cfb78b","Type":"ContainerDied","Data":"47a8a394273b79a84efe27e920d5b8ced645193d3bd45847a88c0713f026ea5e"} Oct 08 18:30:40 crc kubenswrapper[4988]: I1008 18:30:40.092361 4988 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="47a8a394273b79a84efe27e920d5b8ced645193d3bd45847a88c0713f026ea5e" Oct 08 18:30:40 crc kubenswrapper[4988]: I1008 18:30:40.092465 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-2f70-account-create-tdkwz" Oct 08 18:30:40 crc kubenswrapper[4988]: I1008 18:30:40.099209 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-4316-account-create-5kjmj" event={"ID":"0fd89f92-b9aa-4a7b-a3ed-130abc4ce161","Type":"ContainerDied","Data":"c9fb7fe6f640ffac5cd31525c845960828be0df1fd543461b251b855af48aec6"} Oct 08 18:30:40 crc kubenswrapper[4988]: I1008 18:30:40.099244 4988 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c9fb7fe6f640ffac5cd31525c845960828be0df1fd543461b251b855af48aec6" Oct 08 18:30:40 crc kubenswrapper[4988]: I1008 18:30:40.099266 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Oct 08 18:30:40 crc kubenswrapper[4988]: I1008 18:30:40.099313 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-4316-account-create-5kjmj" Oct 08 18:30:40 crc kubenswrapper[4988]: I1008 18:30:40.104862 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Oct 08 18:30:41 crc kubenswrapper[4988]: I1008 18:30:41.089466 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-8554744ccd-jsq45" Oct 08 18:30:41 crc kubenswrapper[4988]: I1008 18:30:41.141602 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"872d504e-cc28-4f84-a509-1e5e3dc809ec","Type":"ContainerStarted","Data":"fa177885dd7b88258aae0fa58b600daee7a6b174bb239d9302c4dfbddff36f3a"} Oct 08 18:30:41 crc kubenswrapper[4988]: I1008 18:30:41.147986 4988 generic.go:334] "Generic (PLEG): container finished" podID="60b75acf-c263-4e5f-af6e-4195fc6be9e4" containerID="1af046392b464d870001852927bcb817b391578c44077d5241550bdd64b98e22" exitCode=0 Oct 08 18:30:41 crc kubenswrapper[4988]: I1008 18:30:41.148123 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-8554744ccd-jsq45" event={"ID":"60b75acf-c263-4e5f-af6e-4195fc6be9e4","Type":"ContainerDied","Data":"1af046392b464d870001852927bcb817b391578c44077d5241550bdd64b98e22"} Oct 08 18:30:41 crc kubenswrapper[4988]: I1008 18:30:41.148166 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-8554744ccd-jsq45" event={"ID":"60b75acf-c263-4e5f-af6e-4195fc6be9e4","Type":"ContainerDied","Data":"c8b811412a9b67f92430740a4f1df53dfc66d18cdb37e09a505a6fbf0176739b"} Oct 08 18:30:41 crc kubenswrapper[4988]: I1008 18:30:41.148187 4988 scope.go:117] "RemoveContainer" containerID="ce7818666ce3c638983016985eb011d4d71fe1d55928608343a134abd73c80de" Oct 08 18:30:41 crc kubenswrapper[4988]: I1008 18:30:41.148197 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-8554744ccd-jsq45" Oct 08 18:30:41 crc kubenswrapper[4988]: I1008 18:30:41.169068 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/60b75acf-c263-4e5f-af6e-4195fc6be9e4-combined-ca-bundle\") pod \"60b75acf-c263-4e5f-af6e-4195fc6be9e4\" (UID: \"60b75acf-c263-4e5f-af6e-4195fc6be9e4\") " Oct 08 18:30:41 crc kubenswrapper[4988]: I1008 18:30:41.169176 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-phv2m\" (UniqueName: \"kubernetes.io/projected/60b75acf-c263-4e5f-af6e-4195fc6be9e4-kube-api-access-phv2m\") pod \"60b75acf-c263-4e5f-af6e-4195fc6be9e4\" (UID: \"60b75acf-c263-4e5f-af6e-4195fc6be9e4\") " Oct 08 18:30:41 crc kubenswrapper[4988]: I1008 18:30:41.169288 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/60b75acf-c263-4e5f-af6e-4195fc6be9e4-httpd-config\") pod \"60b75acf-c263-4e5f-af6e-4195fc6be9e4\" (UID: \"60b75acf-c263-4e5f-af6e-4195fc6be9e4\") " Oct 08 18:30:41 crc kubenswrapper[4988]: I1008 18:30:41.169417 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/60b75acf-c263-4e5f-af6e-4195fc6be9e4-config\") pod \"60b75acf-c263-4e5f-af6e-4195fc6be9e4\" (UID: \"60b75acf-c263-4e5f-af6e-4195fc6be9e4\") " Oct 08 18:30:41 crc kubenswrapper[4988]: I1008 18:30:41.169493 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/60b75acf-c263-4e5f-af6e-4195fc6be9e4-ovndb-tls-certs\") pod \"60b75acf-c263-4e5f-af6e-4195fc6be9e4\" (UID: \"60b75acf-c263-4e5f-af6e-4195fc6be9e4\") " Oct 08 18:30:41 crc kubenswrapper[4988]: I1008 18:30:41.173636 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/60b75acf-c263-4e5f-af6e-4195fc6be9e4-kube-api-access-phv2m" (OuterVolumeSpecName: "kube-api-access-phv2m") pod "60b75acf-c263-4e5f-af6e-4195fc6be9e4" (UID: "60b75acf-c263-4e5f-af6e-4195fc6be9e4"). InnerVolumeSpecName "kube-api-access-phv2m". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:30:41 crc kubenswrapper[4988]: I1008 18:30:41.173663 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/60b75acf-c263-4e5f-af6e-4195fc6be9e4-httpd-config" (OuterVolumeSpecName: "httpd-config") pod "60b75acf-c263-4e5f-af6e-4195fc6be9e4" (UID: "60b75acf-c263-4e5f-af6e-4195fc6be9e4"). InnerVolumeSpecName "httpd-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:30:41 crc kubenswrapper[4988]: I1008 18:30:41.192179 4988 scope.go:117] "RemoveContainer" containerID="1af046392b464d870001852927bcb817b391578c44077d5241550bdd64b98e22" Oct 08 18:30:41 crc kubenswrapper[4988]: I1008 18:30:41.221478 4988 scope.go:117] "RemoveContainer" containerID="ce7818666ce3c638983016985eb011d4d71fe1d55928608343a134abd73c80de" Oct 08 18:30:41 crc kubenswrapper[4988]: I1008 18:30:41.222724 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/60b75acf-c263-4e5f-af6e-4195fc6be9e4-config" (OuterVolumeSpecName: "config") pod "60b75acf-c263-4e5f-af6e-4195fc6be9e4" (UID: "60b75acf-c263-4e5f-af6e-4195fc6be9e4"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:30:41 crc kubenswrapper[4988]: E1008 18:30:41.222809 4988 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ce7818666ce3c638983016985eb011d4d71fe1d55928608343a134abd73c80de\": container with ID starting with ce7818666ce3c638983016985eb011d4d71fe1d55928608343a134abd73c80de not found: ID does not exist" containerID="ce7818666ce3c638983016985eb011d4d71fe1d55928608343a134abd73c80de" Oct 08 18:30:41 crc kubenswrapper[4988]: I1008 18:30:41.222868 4988 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ce7818666ce3c638983016985eb011d4d71fe1d55928608343a134abd73c80de"} err="failed to get container status \"ce7818666ce3c638983016985eb011d4d71fe1d55928608343a134abd73c80de\": rpc error: code = NotFound desc = could not find container \"ce7818666ce3c638983016985eb011d4d71fe1d55928608343a134abd73c80de\": container with ID starting with ce7818666ce3c638983016985eb011d4d71fe1d55928608343a134abd73c80de not found: ID does not exist" Oct 08 18:30:41 crc kubenswrapper[4988]: I1008 18:30:41.222908 4988 scope.go:117] "RemoveContainer" containerID="1af046392b464d870001852927bcb817b391578c44077d5241550bdd64b98e22" Oct 08 18:30:41 crc kubenswrapper[4988]: I1008 18:30:41.223253 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/60b75acf-c263-4e5f-af6e-4195fc6be9e4-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "60b75acf-c263-4e5f-af6e-4195fc6be9e4" (UID: "60b75acf-c263-4e5f-af6e-4195fc6be9e4"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:30:41 crc kubenswrapper[4988]: E1008 18:30:41.223281 4988 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1af046392b464d870001852927bcb817b391578c44077d5241550bdd64b98e22\": container with ID starting with 1af046392b464d870001852927bcb817b391578c44077d5241550bdd64b98e22 not found: ID does not exist" containerID="1af046392b464d870001852927bcb817b391578c44077d5241550bdd64b98e22" Oct 08 18:30:41 crc kubenswrapper[4988]: I1008 18:30:41.223310 4988 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1af046392b464d870001852927bcb817b391578c44077d5241550bdd64b98e22"} err="failed to get container status \"1af046392b464d870001852927bcb817b391578c44077d5241550bdd64b98e22\": rpc error: code = NotFound desc = could not find container \"1af046392b464d870001852927bcb817b391578c44077d5241550bdd64b98e22\": container with ID starting with 1af046392b464d870001852927bcb817b391578c44077d5241550bdd64b98e22 not found: ID does not exist" Oct 08 18:30:41 crc kubenswrapper[4988]: I1008 18:30:41.260520 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/60b75acf-c263-4e5f-af6e-4195fc6be9e4-ovndb-tls-certs" (OuterVolumeSpecName: "ovndb-tls-certs") pod "60b75acf-c263-4e5f-af6e-4195fc6be9e4" (UID: "60b75acf-c263-4e5f-af6e-4195fc6be9e4"). InnerVolumeSpecName "ovndb-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:30:41 crc kubenswrapper[4988]: I1008 18:30:41.272035 4988 reconciler_common.go:293] "Volume detached for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/60b75acf-c263-4e5f-af6e-4195fc6be9e4-httpd-config\") on node \"crc\" DevicePath \"\"" Oct 08 18:30:41 crc kubenswrapper[4988]: I1008 18:30:41.272086 4988 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/60b75acf-c263-4e5f-af6e-4195fc6be9e4-config\") on node \"crc\" DevicePath \"\"" Oct 08 18:30:41 crc kubenswrapper[4988]: I1008 18:30:41.272102 4988 reconciler_common.go:293] "Volume detached for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/60b75acf-c263-4e5f-af6e-4195fc6be9e4-ovndb-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 08 18:30:41 crc kubenswrapper[4988]: I1008 18:30:41.272114 4988 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/60b75acf-c263-4e5f-af6e-4195fc6be9e4-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 18:30:41 crc kubenswrapper[4988]: I1008 18:30:41.272126 4988 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-phv2m\" (UniqueName: \"kubernetes.io/projected/60b75acf-c263-4e5f-af6e-4195fc6be9e4-kube-api-access-phv2m\") on node \"crc\" DevicePath \"\"" Oct 08 18:30:41 crc kubenswrapper[4988]: I1008 18:30:41.489460 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-db-sync-hzsjj"] Oct 08 18:30:41 crc kubenswrapper[4988]: E1008 18:30:41.489847 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e3e8eb1b-f0b5-4e56-9ba7-d068435355c5" containerName="mariadb-account-create" Oct 08 18:30:41 crc kubenswrapper[4988]: I1008 18:30:41.489866 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="e3e8eb1b-f0b5-4e56-9ba7-d068435355c5" containerName="mariadb-account-create" Oct 08 18:30:41 crc kubenswrapper[4988]: E1008 18:30:41.489881 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="60b75acf-c263-4e5f-af6e-4195fc6be9e4" containerName="neutron-httpd" Oct 08 18:30:41 crc kubenswrapper[4988]: I1008 18:30:41.489890 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="60b75acf-c263-4e5f-af6e-4195fc6be9e4" containerName="neutron-httpd" Oct 08 18:30:41 crc kubenswrapper[4988]: E1008 18:30:41.489915 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5f273f27-4991-4aa3-a734-dec827cfb78b" containerName="mariadb-account-create" Oct 08 18:30:41 crc kubenswrapper[4988]: I1008 18:30:41.489924 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="5f273f27-4991-4aa3-a734-dec827cfb78b" containerName="mariadb-account-create" Oct 08 18:30:41 crc kubenswrapper[4988]: E1008 18:30:41.489936 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="60b75acf-c263-4e5f-af6e-4195fc6be9e4" containerName="neutron-api" Oct 08 18:30:41 crc kubenswrapper[4988]: I1008 18:30:41.489944 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="60b75acf-c263-4e5f-af6e-4195fc6be9e4" containerName="neutron-api" Oct 08 18:30:41 crc kubenswrapper[4988]: E1008 18:30:41.489971 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0fd89f92-b9aa-4a7b-a3ed-130abc4ce161" containerName="mariadb-account-create" Oct 08 18:30:41 crc kubenswrapper[4988]: I1008 18:30:41.489980 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="0fd89f92-b9aa-4a7b-a3ed-130abc4ce161" containerName="mariadb-account-create" Oct 08 18:30:41 crc kubenswrapper[4988]: I1008 18:30:41.490219 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="5f273f27-4991-4aa3-a734-dec827cfb78b" containerName="mariadb-account-create" Oct 08 18:30:41 crc kubenswrapper[4988]: I1008 18:30:41.490250 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="e3e8eb1b-f0b5-4e56-9ba7-d068435355c5" containerName="mariadb-account-create" Oct 08 18:30:41 crc kubenswrapper[4988]: I1008 18:30:41.490265 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="60b75acf-c263-4e5f-af6e-4195fc6be9e4" containerName="neutron-httpd" Oct 08 18:30:41 crc kubenswrapper[4988]: I1008 18:30:41.490279 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="0fd89f92-b9aa-4a7b-a3ed-130abc4ce161" containerName="mariadb-account-create" Oct 08 18:30:41 crc kubenswrapper[4988]: I1008 18:30:41.490288 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="60b75acf-c263-4e5f-af6e-4195fc6be9e4" containerName="neutron-api" Oct 08 18:30:41 crc kubenswrapper[4988]: I1008 18:30:41.491031 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-hzsjj" Oct 08 18:30:41 crc kubenswrapper[4988]: I1008 18:30:41.493797 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-scripts" Oct 08 18:30:41 crc kubenswrapper[4988]: I1008 18:30:41.493982 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-vmdxg" Oct 08 18:30:41 crc kubenswrapper[4988]: I1008 18:30:41.494453 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Oct 08 18:30:41 crc kubenswrapper[4988]: I1008 18:30:41.526449 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-hzsjj"] Oct 08 18:30:41 crc kubenswrapper[4988]: I1008 18:30:41.579363 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/32722c37-a19b-42e1-be71-432f3499af15-config-data\") pod \"nova-cell0-conductor-db-sync-hzsjj\" (UID: \"32722c37-a19b-42e1-be71-432f3499af15\") " pod="openstack/nova-cell0-conductor-db-sync-hzsjj" Oct 08 18:30:41 crc kubenswrapper[4988]: I1008 18:30:41.579634 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/32722c37-a19b-42e1-be71-432f3499af15-scripts\") pod \"nova-cell0-conductor-db-sync-hzsjj\" (UID: \"32722c37-a19b-42e1-be71-432f3499af15\") " pod="openstack/nova-cell0-conductor-db-sync-hzsjj" Oct 08 18:30:41 crc kubenswrapper[4988]: I1008 18:30:41.579720 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bdkvj\" (UniqueName: \"kubernetes.io/projected/32722c37-a19b-42e1-be71-432f3499af15-kube-api-access-bdkvj\") pod \"nova-cell0-conductor-db-sync-hzsjj\" (UID: \"32722c37-a19b-42e1-be71-432f3499af15\") " pod="openstack/nova-cell0-conductor-db-sync-hzsjj" Oct 08 18:30:41 crc kubenswrapper[4988]: I1008 18:30:41.579763 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/32722c37-a19b-42e1-be71-432f3499af15-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-hzsjj\" (UID: \"32722c37-a19b-42e1-be71-432f3499af15\") " pod="openstack/nova-cell0-conductor-db-sync-hzsjj" Oct 08 18:30:41 crc kubenswrapper[4988]: I1008 18:30:41.579838 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-8554744ccd-jsq45"] Oct 08 18:30:41 crc kubenswrapper[4988]: I1008 18:30:41.596321 4988 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-8554744ccd-jsq45"] Oct 08 18:30:41 crc kubenswrapper[4988]: I1008 18:30:41.681421 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bdkvj\" (UniqueName: \"kubernetes.io/projected/32722c37-a19b-42e1-be71-432f3499af15-kube-api-access-bdkvj\") pod \"nova-cell0-conductor-db-sync-hzsjj\" (UID: \"32722c37-a19b-42e1-be71-432f3499af15\") " pod="openstack/nova-cell0-conductor-db-sync-hzsjj" Oct 08 18:30:41 crc kubenswrapper[4988]: I1008 18:30:41.681494 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/32722c37-a19b-42e1-be71-432f3499af15-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-hzsjj\" (UID: \"32722c37-a19b-42e1-be71-432f3499af15\") " pod="openstack/nova-cell0-conductor-db-sync-hzsjj" Oct 08 18:30:41 crc kubenswrapper[4988]: I1008 18:30:41.681548 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/32722c37-a19b-42e1-be71-432f3499af15-config-data\") pod \"nova-cell0-conductor-db-sync-hzsjj\" (UID: \"32722c37-a19b-42e1-be71-432f3499af15\") " pod="openstack/nova-cell0-conductor-db-sync-hzsjj" Oct 08 18:30:41 crc kubenswrapper[4988]: I1008 18:30:41.681566 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/32722c37-a19b-42e1-be71-432f3499af15-scripts\") pod \"nova-cell0-conductor-db-sync-hzsjj\" (UID: \"32722c37-a19b-42e1-be71-432f3499af15\") " pod="openstack/nova-cell0-conductor-db-sync-hzsjj" Oct 08 18:30:41 crc kubenswrapper[4988]: I1008 18:30:41.688155 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/32722c37-a19b-42e1-be71-432f3499af15-scripts\") pod \"nova-cell0-conductor-db-sync-hzsjj\" (UID: \"32722c37-a19b-42e1-be71-432f3499af15\") " pod="openstack/nova-cell0-conductor-db-sync-hzsjj" Oct 08 18:30:41 crc kubenswrapper[4988]: I1008 18:30:41.691024 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/32722c37-a19b-42e1-be71-432f3499af15-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-hzsjj\" (UID: \"32722c37-a19b-42e1-be71-432f3499af15\") " pod="openstack/nova-cell0-conductor-db-sync-hzsjj" Oct 08 18:30:41 crc kubenswrapper[4988]: I1008 18:30:41.695085 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/32722c37-a19b-42e1-be71-432f3499af15-config-data\") pod \"nova-cell0-conductor-db-sync-hzsjj\" (UID: \"32722c37-a19b-42e1-be71-432f3499af15\") " pod="openstack/nova-cell0-conductor-db-sync-hzsjj" Oct 08 18:30:41 crc kubenswrapper[4988]: I1008 18:30:41.702887 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bdkvj\" (UniqueName: \"kubernetes.io/projected/32722c37-a19b-42e1-be71-432f3499af15-kube-api-access-bdkvj\") pod \"nova-cell0-conductor-db-sync-hzsjj\" (UID: \"32722c37-a19b-42e1-be71-432f3499af15\") " pod="openstack/nova-cell0-conductor-db-sync-hzsjj" Oct 08 18:30:41 crc kubenswrapper[4988]: I1008 18:30:41.823790 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-hzsjj" Oct 08 18:30:42 crc kubenswrapper[4988]: I1008 18:30:42.158557 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"872d504e-cc28-4f84-a509-1e5e3dc809ec","Type":"ContainerStarted","Data":"0e685f9876e7842310fd195d1e095bbeeddcf0845f46c775c8b0534054387d31"} Oct 08 18:30:42 crc kubenswrapper[4988]: I1008 18:30:42.159776 4988 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 08 18:30:42 crc kubenswrapper[4988]: I1008 18:30:42.159787 4988 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 08 18:30:42 crc kubenswrapper[4988]: I1008 18:30:42.309731 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Oct 08 18:30:42 crc kubenswrapper[4988]: I1008 18:30:42.311756 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-hzsjj"] Oct 08 18:30:42 crc kubenswrapper[4988]: I1008 18:30:42.313099 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Oct 08 18:30:42 crc kubenswrapper[4988]: W1008 18:30:42.316717 4988 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod32722c37_a19b_42e1_be71_432f3499af15.slice/crio-168dd7ae71b862261bfff84ac9282738bd6092f8ec8b1d76f4171a38fd86d4bd WatchSource:0}: Error finding container 168dd7ae71b862261bfff84ac9282738bd6092f8ec8b1d76f4171a38fd86d4bd: Status 404 returned error can't find the container with id 168dd7ae71b862261bfff84ac9282738bd6092f8ec8b1d76f4171a38fd86d4bd Oct 08 18:30:43 crc kubenswrapper[4988]: I1008 18:30:43.181703 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-hzsjj" event={"ID":"32722c37-a19b-42e1-be71-432f3499af15","Type":"ContainerStarted","Data":"168dd7ae71b862261bfff84ac9282738bd6092f8ec8b1d76f4171a38fd86d4bd"} Oct 08 18:30:43 crc kubenswrapper[4988]: I1008 18:30:43.290609 4988 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="60b75acf-c263-4e5f-af6e-4195fc6be9e4" path="/var/lib/kubelet/pods/60b75acf-c263-4e5f-af6e-4195fc6be9e4/volumes" Oct 08 18:30:43 crc kubenswrapper[4988]: I1008 18:30:43.584709 4988 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Oct 08 18:30:44 crc kubenswrapper[4988]: I1008 18:30:44.196652 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"872d504e-cc28-4f84-a509-1e5e3dc809ec","Type":"ContainerStarted","Data":"800f3af68885f1a55a2dc6c08861f2a662598d8fbefb296152e2105d10e1d4be"} Oct 08 18:30:44 crc kubenswrapper[4988]: I1008 18:30:44.225871 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.096259273 podStartE2EDuration="7.225825478s" podCreationTimestamp="2025-10-08 18:30:37 +0000 UTC" firstStartedPulling="2025-10-08 18:30:37.937183122 +0000 UTC m=+1183.387025892" lastFinishedPulling="2025-10-08 18:30:43.066749327 +0000 UTC m=+1188.516592097" observedRunningTime="2025-10-08 18:30:44.217352737 +0000 UTC m=+1189.667195517" watchObservedRunningTime="2025-10-08 18:30:44.225825478 +0000 UTC m=+1189.675668248" Oct 08 18:30:45 crc kubenswrapper[4988]: I1008 18:30:45.205483 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 08 18:30:47 crc kubenswrapper[4988]: I1008 18:30:47.817014 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 08 18:30:47 crc kubenswrapper[4988]: I1008 18:30:47.817599 4988 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="872d504e-cc28-4f84-a509-1e5e3dc809ec" containerName="sg-core" containerID="cri-o://0e685f9876e7842310fd195d1e095bbeeddcf0845f46c775c8b0534054387d31" gracePeriod=30 Oct 08 18:30:47 crc kubenswrapper[4988]: I1008 18:30:47.817623 4988 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="872d504e-cc28-4f84-a509-1e5e3dc809ec" containerName="ceilometer-notification-agent" containerID="cri-o://fa177885dd7b88258aae0fa58b600daee7a6b174bb239d9302c4dfbddff36f3a" gracePeriod=30 Oct 08 18:30:47 crc kubenswrapper[4988]: I1008 18:30:47.817675 4988 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="872d504e-cc28-4f84-a509-1e5e3dc809ec" containerName="proxy-httpd" containerID="cri-o://800f3af68885f1a55a2dc6c08861f2a662598d8fbefb296152e2105d10e1d4be" gracePeriod=30 Oct 08 18:30:47 crc kubenswrapper[4988]: I1008 18:30:47.817858 4988 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="872d504e-cc28-4f84-a509-1e5e3dc809ec" containerName="ceilometer-central-agent" containerID="cri-o://d72be3d5571a4599f581e687b5b5ea86da6f84ff1b611e46cd3f46b526852946" gracePeriod=30 Oct 08 18:30:48 crc kubenswrapper[4988]: I1008 18:30:48.242087 4988 generic.go:334] "Generic (PLEG): container finished" podID="872d504e-cc28-4f84-a509-1e5e3dc809ec" containerID="0e685f9876e7842310fd195d1e095bbeeddcf0845f46c775c8b0534054387d31" exitCode=2 Oct 08 18:30:48 crc kubenswrapper[4988]: I1008 18:30:48.242133 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"872d504e-cc28-4f84-a509-1e5e3dc809ec","Type":"ContainerDied","Data":"0e685f9876e7842310fd195d1e095bbeeddcf0845f46c775c8b0534054387d31"} Oct 08 18:30:49 crc kubenswrapper[4988]: I1008 18:30:49.255489 4988 generic.go:334] "Generic (PLEG): container finished" podID="872d504e-cc28-4f84-a509-1e5e3dc809ec" containerID="800f3af68885f1a55a2dc6c08861f2a662598d8fbefb296152e2105d10e1d4be" exitCode=0 Oct 08 18:30:49 crc kubenswrapper[4988]: I1008 18:30:49.255758 4988 generic.go:334] "Generic (PLEG): container finished" podID="872d504e-cc28-4f84-a509-1e5e3dc809ec" containerID="fa177885dd7b88258aae0fa58b600daee7a6b174bb239d9302c4dfbddff36f3a" exitCode=0 Oct 08 18:30:49 crc kubenswrapper[4988]: I1008 18:30:49.255769 4988 generic.go:334] "Generic (PLEG): container finished" podID="872d504e-cc28-4f84-a509-1e5e3dc809ec" containerID="d72be3d5571a4599f581e687b5b5ea86da6f84ff1b611e46cd3f46b526852946" exitCode=0 Oct 08 18:30:49 crc kubenswrapper[4988]: I1008 18:30:49.255540 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"872d504e-cc28-4f84-a509-1e5e3dc809ec","Type":"ContainerDied","Data":"800f3af68885f1a55a2dc6c08861f2a662598d8fbefb296152e2105d10e1d4be"} Oct 08 18:30:49 crc kubenswrapper[4988]: I1008 18:30:49.255807 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"872d504e-cc28-4f84-a509-1e5e3dc809ec","Type":"ContainerDied","Data":"fa177885dd7b88258aae0fa58b600daee7a6b174bb239d9302c4dfbddff36f3a"} Oct 08 18:30:49 crc kubenswrapper[4988]: I1008 18:30:49.255822 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"872d504e-cc28-4f84-a509-1e5e3dc809ec","Type":"ContainerDied","Data":"d72be3d5571a4599f581e687b5b5ea86da6f84ff1b611e46cd3f46b526852946"} Oct 08 18:30:49 crc kubenswrapper[4988]: I1008 18:30:49.985331 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 08 18:30:50 crc kubenswrapper[4988]: I1008 18:30:50.049175 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/872d504e-cc28-4f84-a509-1e5e3dc809ec-scripts\") pod \"872d504e-cc28-4f84-a509-1e5e3dc809ec\" (UID: \"872d504e-cc28-4f84-a509-1e5e3dc809ec\") " Oct 08 18:30:50 crc kubenswrapper[4988]: I1008 18:30:50.049290 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/872d504e-cc28-4f84-a509-1e5e3dc809ec-run-httpd\") pod \"872d504e-cc28-4f84-a509-1e5e3dc809ec\" (UID: \"872d504e-cc28-4f84-a509-1e5e3dc809ec\") " Oct 08 18:30:50 crc kubenswrapper[4988]: I1008 18:30:50.049422 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2nrhr\" (UniqueName: \"kubernetes.io/projected/872d504e-cc28-4f84-a509-1e5e3dc809ec-kube-api-access-2nrhr\") pod \"872d504e-cc28-4f84-a509-1e5e3dc809ec\" (UID: \"872d504e-cc28-4f84-a509-1e5e3dc809ec\") " Oct 08 18:30:50 crc kubenswrapper[4988]: I1008 18:30:50.049985 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/872d504e-cc28-4f84-a509-1e5e3dc809ec-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "872d504e-cc28-4f84-a509-1e5e3dc809ec" (UID: "872d504e-cc28-4f84-a509-1e5e3dc809ec"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 18:30:50 crc kubenswrapper[4988]: I1008 18:30:50.050330 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/872d504e-cc28-4f84-a509-1e5e3dc809ec-combined-ca-bundle\") pod \"872d504e-cc28-4f84-a509-1e5e3dc809ec\" (UID: \"872d504e-cc28-4f84-a509-1e5e3dc809ec\") " Oct 08 18:30:50 crc kubenswrapper[4988]: I1008 18:30:50.050358 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/872d504e-cc28-4f84-a509-1e5e3dc809ec-config-data\") pod \"872d504e-cc28-4f84-a509-1e5e3dc809ec\" (UID: \"872d504e-cc28-4f84-a509-1e5e3dc809ec\") " Oct 08 18:30:50 crc kubenswrapper[4988]: I1008 18:30:50.050500 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/872d504e-cc28-4f84-a509-1e5e3dc809ec-sg-core-conf-yaml\") pod \"872d504e-cc28-4f84-a509-1e5e3dc809ec\" (UID: \"872d504e-cc28-4f84-a509-1e5e3dc809ec\") " Oct 08 18:30:50 crc kubenswrapper[4988]: I1008 18:30:50.050550 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/872d504e-cc28-4f84-a509-1e5e3dc809ec-log-httpd\") pod \"872d504e-cc28-4f84-a509-1e5e3dc809ec\" (UID: \"872d504e-cc28-4f84-a509-1e5e3dc809ec\") " Oct 08 18:30:50 crc kubenswrapper[4988]: I1008 18:30:50.051151 4988 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/872d504e-cc28-4f84-a509-1e5e3dc809ec-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 08 18:30:50 crc kubenswrapper[4988]: I1008 18:30:50.051532 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/872d504e-cc28-4f84-a509-1e5e3dc809ec-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "872d504e-cc28-4f84-a509-1e5e3dc809ec" (UID: "872d504e-cc28-4f84-a509-1e5e3dc809ec"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 18:30:50 crc kubenswrapper[4988]: I1008 18:30:50.055666 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/872d504e-cc28-4f84-a509-1e5e3dc809ec-scripts" (OuterVolumeSpecName: "scripts") pod "872d504e-cc28-4f84-a509-1e5e3dc809ec" (UID: "872d504e-cc28-4f84-a509-1e5e3dc809ec"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:30:50 crc kubenswrapper[4988]: I1008 18:30:50.057568 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/872d504e-cc28-4f84-a509-1e5e3dc809ec-kube-api-access-2nrhr" (OuterVolumeSpecName: "kube-api-access-2nrhr") pod "872d504e-cc28-4f84-a509-1e5e3dc809ec" (UID: "872d504e-cc28-4f84-a509-1e5e3dc809ec"). InnerVolumeSpecName "kube-api-access-2nrhr". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:30:50 crc kubenswrapper[4988]: I1008 18:30:50.091469 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/872d504e-cc28-4f84-a509-1e5e3dc809ec-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "872d504e-cc28-4f84-a509-1e5e3dc809ec" (UID: "872d504e-cc28-4f84-a509-1e5e3dc809ec"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:30:50 crc kubenswrapper[4988]: I1008 18:30:50.131037 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/872d504e-cc28-4f84-a509-1e5e3dc809ec-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "872d504e-cc28-4f84-a509-1e5e3dc809ec" (UID: "872d504e-cc28-4f84-a509-1e5e3dc809ec"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:30:50 crc kubenswrapper[4988]: I1008 18:30:50.153350 4988 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/872d504e-cc28-4f84-a509-1e5e3dc809ec-scripts\") on node \"crc\" DevicePath \"\"" Oct 08 18:30:50 crc kubenswrapper[4988]: I1008 18:30:50.153406 4988 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2nrhr\" (UniqueName: \"kubernetes.io/projected/872d504e-cc28-4f84-a509-1e5e3dc809ec-kube-api-access-2nrhr\") on node \"crc\" DevicePath \"\"" Oct 08 18:30:50 crc kubenswrapper[4988]: I1008 18:30:50.153422 4988 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/872d504e-cc28-4f84-a509-1e5e3dc809ec-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 18:30:50 crc kubenswrapper[4988]: I1008 18:30:50.153433 4988 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/872d504e-cc28-4f84-a509-1e5e3dc809ec-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Oct 08 18:30:50 crc kubenswrapper[4988]: I1008 18:30:50.153444 4988 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/872d504e-cc28-4f84-a509-1e5e3dc809ec-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 08 18:30:50 crc kubenswrapper[4988]: I1008 18:30:50.173429 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/872d504e-cc28-4f84-a509-1e5e3dc809ec-config-data" (OuterVolumeSpecName: "config-data") pod "872d504e-cc28-4f84-a509-1e5e3dc809ec" (UID: "872d504e-cc28-4f84-a509-1e5e3dc809ec"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:30:50 crc kubenswrapper[4988]: I1008 18:30:50.255134 4988 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/872d504e-cc28-4f84-a509-1e5e3dc809ec-config-data\") on node \"crc\" DevicePath \"\"" Oct 08 18:30:50 crc kubenswrapper[4988]: I1008 18:30:50.265076 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-hzsjj" event={"ID":"32722c37-a19b-42e1-be71-432f3499af15","Type":"ContainerStarted","Data":"3bb3e6305405075980f0e6b13cc2d64214327a9883596c00b4d0ce337b183212"} Oct 08 18:30:50 crc kubenswrapper[4988]: I1008 18:30:50.267712 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"872d504e-cc28-4f84-a509-1e5e3dc809ec","Type":"ContainerDied","Data":"5ad380b52cd6782b558d4e85f1ef9348a5a24e4503f9480d14c2f9680479466c"} Oct 08 18:30:50 crc kubenswrapper[4988]: I1008 18:30:50.267750 4988 scope.go:117] "RemoveContainer" containerID="800f3af68885f1a55a2dc6c08861f2a662598d8fbefb296152e2105d10e1d4be" Oct 08 18:30:50 crc kubenswrapper[4988]: I1008 18:30:50.267759 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 08 18:30:50 crc kubenswrapper[4988]: I1008 18:30:50.300530 4988 scope.go:117] "RemoveContainer" containerID="0e685f9876e7842310fd195d1e095bbeeddcf0845f46c775c8b0534054387d31" Oct 08 18:30:50 crc kubenswrapper[4988]: I1008 18:30:50.357622 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-db-sync-hzsjj" podStartSLOduration=1.941052818 podStartE2EDuration="9.3576029s" podCreationTimestamp="2025-10-08 18:30:41 +0000 UTC" firstStartedPulling="2025-10-08 18:30:42.321131156 +0000 UTC m=+1187.770973926" lastFinishedPulling="2025-10-08 18:30:49.737681238 +0000 UTC m=+1195.187524008" observedRunningTime="2025-10-08 18:30:50.295091621 +0000 UTC m=+1195.744934391" watchObservedRunningTime="2025-10-08 18:30:50.3576029 +0000 UTC m=+1195.807445670" Oct 08 18:30:50 crc kubenswrapper[4988]: I1008 18:30:50.379271 4988 scope.go:117] "RemoveContainer" containerID="fa177885dd7b88258aae0fa58b600daee7a6b174bb239d9302c4dfbddff36f3a" Oct 08 18:30:50 crc kubenswrapper[4988]: I1008 18:30:50.380676 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 08 18:30:50 crc kubenswrapper[4988]: I1008 18:30:50.412531 4988 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Oct 08 18:30:50 crc kubenswrapper[4988]: I1008 18:30:50.422446 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 08 18:30:50 crc kubenswrapper[4988]: E1008 18:30:50.422841 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="872d504e-cc28-4f84-a509-1e5e3dc809ec" containerName="ceilometer-notification-agent" Oct 08 18:30:50 crc kubenswrapper[4988]: I1008 18:30:50.422853 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="872d504e-cc28-4f84-a509-1e5e3dc809ec" containerName="ceilometer-notification-agent" Oct 08 18:30:50 crc kubenswrapper[4988]: E1008 18:30:50.422874 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="872d504e-cc28-4f84-a509-1e5e3dc809ec" containerName="ceilometer-central-agent" Oct 08 18:30:50 crc kubenswrapper[4988]: I1008 18:30:50.422879 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="872d504e-cc28-4f84-a509-1e5e3dc809ec" containerName="ceilometer-central-agent" Oct 08 18:30:50 crc kubenswrapper[4988]: E1008 18:30:50.422893 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="872d504e-cc28-4f84-a509-1e5e3dc809ec" containerName="sg-core" Oct 08 18:30:50 crc kubenswrapper[4988]: I1008 18:30:50.422899 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="872d504e-cc28-4f84-a509-1e5e3dc809ec" containerName="sg-core" Oct 08 18:30:50 crc kubenswrapper[4988]: E1008 18:30:50.422920 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="872d504e-cc28-4f84-a509-1e5e3dc809ec" containerName="proxy-httpd" Oct 08 18:30:50 crc kubenswrapper[4988]: I1008 18:30:50.422926 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="872d504e-cc28-4f84-a509-1e5e3dc809ec" containerName="proxy-httpd" Oct 08 18:30:50 crc kubenswrapper[4988]: I1008 18:30:50.423109 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="872d504e-cc28-4f84-a509-1e5e3dc809ec" containerName="ceilometer-notification-agent" Oct 08 18:30:50 crc kubenswrapper[4988]: I1008 18:30:50.423125 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="872d504e-cc28-4f84-a509-1e5e3dc809ec" containerName="proxy-httpd" Oct 08 18:30:50 crc kubenswrapper[4988]: I1008 18:30:50.423135 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="872d504e-cc28-4f84-a509-1e5e3dc809ec" containerName="sg-core" Oct 08 18:30:50 crc kubenswrapper[4988]: I1008 18:30:50.423153 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="872d504e-cc28-4f84-a509-1e5e3dc809ec" containerName="ceilometer-central-agent" Oct 08 18:30:50 crc kubenswrapper[4988]: I1008 18:30:50.424790 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 08 18:30:50 crc kubenswrapper[4988]: I1008 18:30:50.429457 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 08 18:30:50 crc kubenswrapper[4988]: I1008 18:30:50.429526 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 08 18:30:50 crc kubenswrapper[4988]: I1008 18:30:50.432211 4988 scope.go:117] "RemoveContainer" containerID="d72be3d5571a4599f581e687b5b5ea86da6f84ff1b611e46cd3f46b526852946" Oct 08 18:30:50 crc kubenswrapper[4988]: I1008 18:30:50.435627 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 08 18:30:50 crc kubenswrapper[4988]: I1008 18:30:50.562592 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1c9bf85f-df23-42e0-b705-69f3938d8208-scripts\") pod \"ceilometer-0\" (UID: \"1c9bf85f-df23-42e0-b705-69f3938d8208\") " pod="openstack/ceilometer-0" Oct 08 18:30:50 crc kubenswrapper[4988]: I1008 18:30:50.562984 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/1c9bf85f-df23-42e0-b705-69f3938d8208-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"1c9bf85f-df23-42e0-b705-69f3938d8208\") " pod="openstack/ceilometer-0" Oct 08 18:30:50 crc kubenswrapper[4988]: I1008 18:30:50.563059 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1c9bf85f-df23-42e0-b705-69f3938d8208-config-data\") pod \"ceilometer-0\" (UID: \"1c9bf85f-df23-42e0-b705-69f3938d8208\") " pod="openstack/ceilometer-0" Oct 08 18:30:50 crc kubenswrapper[4988]: I1008 18:30:50.563110 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1c9bf85f-df23-42e0-b705-69f3938d8208-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"1c9bf85f-df23-42e0-b705-69f3938d8208\") " pod="openstack/ceilometer-0" Oct 08 18:30:50 crc kubenswrapper[4988]: I1008 18:30:50.563156 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1c9bf85f-df23-42e0-b705-69f3938d8208-run-httpd\") pod \"ceilometer-0\" (UID: \"1c9bf85f-df23-42e0-b705-69f3938d8208\") " pod="openstack/ceilometer-0" Oct 08 18:30:50 crc kubenswrapper[4988]: I1008 18:30:50.563188 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qhfz7\" (UniqueName: \"kubernetes.io/projected/1c9bf85f-df23-42e0-b705-69f3938d8208-kube-api-access-qhfz7\") pod \"ceilometer-0\" (UID: \"1c9bf85f-df23-42e0-b705-69f3938d8208\") " pod="openstack/ceilometer-0" Oct 08 18:30:50 crc kubenswrapper[4988]: I1008 18:30:50.563219 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1c9bf85f-df23-42e0-b705-69f3938d8208-log-httpd\") pod \"ceilometer-0\" (UID: \"1c9bf85f-df23-42e0-b705-69f3938d8208\") " pod="openstack/ceilometer-0" Oct 08 18:30:50 crc kubenswrapper[4988]: I1008 18:30:50.664312 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1c9bf85f-df23-42e0-b705-69f3938d8208-config-data\") pod \"ceilometer-0\" (UID: \"1c9bf85f-df23-42e0-b705-69f3938d8208\") " pod="openstack/ceilometer-0" Oct 08 18:30:50 crc kubenswrapper[4988]: I1008 18:30:50.664414 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1c9bf85f-df23-42e0-b705-69f3938d8208-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"1c9bf85f-df23-42e0-b705-69f3938d8208\") " pod="openstack/ceilometer-0" Oct 08 18:30:50 crc kubenswrapper[4988]: I1008 18:30:50.664457 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1c9bf85f-df23-42e0-b705-69f3938d8208-run-httpd\") pod \"ceilometer-0\" (UID: \"1c9bf85f-df23-42e0-b705-69f3938d8208\") " pod="openstack/ceilometer-0" Oct 08 18:30:50 crc kubenswrapper[4988]: I1008 18:30:50.664482 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qhfz7\" (UniqueName: \"kubernetes.io/projected/1c9bf85f-df23-42e0-b705-69f3938d8208-kube-api-access-qhfz7\") pod \"ceilometer-0\" (UID: \"1c9bf85f-df23-42e0-b705-69f3938d8208\") " pod="openstack/ceilometer-0" Oct 08 18:30:50 crc kubenswrapper[4988]: I1008 18:30:50.664509 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1c9bf85f-df23-42e0-b705-69f3938d8208-log-httpd\") pod \"ceilometer-0\" (UID: \"1c9bf85f-df23-42e0-b705-69f3938d8208\") " pod="openstack/ceilometer-0" Oct 08 18:30:50 crc kubenswrapper[4988]: I1008 18:30:50.664558 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1c9bf85f-df23-42e0-b705-69f3938d8208-scripts\") pod \"ceilometer-0\" (UID: \"1c9bf85f-df23-42e0-b705-69f3938d8208\") " pod="openstack/ceilometer-0" Oct 08 18:30:50 crc kubenswrapper[4988]: I1008 18:30:50.664580 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/1c9bf85f-df23-42e0-b705-69f3938d8208-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"1c9bf85f-df23-42e0-b705-69f3938d8208\") " pod="openstack/ceilometer-0" Oct 08 18:30:50 crc kubenswrapper[4988]: I1008 18:30:50.665150 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1c9bf85f-df23-42e0-b705-69f3938d8208-run-httpd\") pod \"ceilometer-0\" (UID: \"1c9bf85f-df23-42e0-b705-69f3938d8208\") " pod="openstack/ceilometer-0" Oct 08 18:30:50 crc kubenswrapper[4988]: I1008 18:30:50.665465 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1c9bf85f-df23-42e0-b705-69f3938d8208-log-httpd\") pod \"ceilometer-0\" (UID: \"1c9bf85f-df23-42e0-b705-69f3938d8208\") " pod="openstack/ceilometer-0" Oct 08 18:30:50 crc kubenswrapper[4988]: I1008 18:30:50.669145 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1c9bf85f-df23-42e0-b705-69f3938d8208-config-data\") pod \"ceilometer-0\" (UID: \"1c9bf85f-df23-42e0-b705-69f3938d8208\") " pod="openstack/ceilometer-0" Oct 08 18:30:50 crc kubenswrapper[4988]: I1008 18:30:50.669511 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/1c9bf85f-df23-42e0-b705-69f3938d8208-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"1c9bf85f-df23-42e0-b705-69f3938d8208\") " pod="openstack/ceilometer-0" Oct 08 18:30:50 crc kubenswrapper[4988]: I1008 18:30:50.669756 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1c9bf85f-df23-42e0-b705-69f3938d8208-scripts\") pod \"ceilometer-0\" (UID: \"1c9bf85f-df23-42e0-b705-69f3938d8208\") " pod="openstack/ceilometer-0" Oct 08 18:30:50 crc kubenswrapper[4988]: I1008 18:30:50.670741 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1c9bf85f-df23-42e0-b705-69f3938d8208-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"1c9bf85f-df23-42e0-b705-69f3938d8208\") " pod="openstack/ceilometer-0" Oct 08 18:30:50 crc kubenswrapper[4988]: I1008 18:30:50.692059 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qhfz7\" (UniqueName: \"kubernetes.io/projected/1c9bf85f-df23-42e0-b705-69f3938d8208-kube-api-access-qhfz7\") pod \"ceilometer-0\" (UID: \"1c9bf85f-df23-42e0-b705-69f3938d8208\") " pod="openstack/ceilometer-0" Oct 08 18:30:50 crc kubenswrapper[4988]: I1008 18:30:50.759648 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 08 18:30:51 crc kubenswrapper[4988]: I1008 18:30:51.232907 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 08 18:30:51 crc kubenswrapper[4988]: I1008 18:30:51.253194 4988 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="872d504e-cc28-4f84-a509-1e5e3dc809ec" path="/var/lib/kubelet/pods/872d504e-cc28-4f84-a509-1e5e3dc809ec/volumes" Oct 08 18:30:51 crc kubenswrapper[4988]: I1008 18:30:51.277877 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"1c9bf85f-df23-42e0-b705-69f3938d8208","Type":"ContainerStarted","Data":"180f4d5f4c235a912e54557a28353c8678d7b0bfe83565a80b8b0cb9e02655e9"} Oct 08 18:30:51 crc kubenswrapper[4988]: I1008 18:30:51.717796 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 08 18:30:52 crc kubenswrapper[4988]: I1008 18:30:52.286805 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"1c9bf85f-df23-42e0-b705-69f3938d8208","Type":"ContainerStarted","Data":"ef6c7ed3962eab0108765f77a0b9afaa599c1c96b4e1a6070f7466d2c9ac9095"} Oct 08 18:30:53 crc kubenswrapper[4988]: I1008 18:30:53.295899 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"1c9bf85f-df23-42e0-b705-69f3938d8208","Type":"ContainerStarted","Data":"ec98ddb44e93e83be1bc8710bfd30c06d306076c89effcf0e3e08a8e03278fdc"} Oct 08 18:30:53 crc kubenswrapper[4988]: I1008 18:30:53.337781 4988 patch_prober.go:28] interesting pod/machine-config-daemon-nm54f container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 18:30:53 crc kubenswrapper[4988]: I1008 18:30:53.337845 4988 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 18:30:53 crc kubenswrapper[4988]: I1008 18:30:53.337890 4988 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" Oct 08 18:30:53 crc kubenswrapper[4988]: I1008 18:30:53.338651 4988 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"33dcd4d530d69056c83e9e7f7862766d54a3df5b03043465d12788d2876eac58"} pod="openshift-machine-config-operator/machine-config-daemon-nm54f" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 08 18:30:53 crc kubenswrapper[4988]: I1008 18:30:53.338719 4988 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" containerName="machine-config-daemon" containerID="cri-o://33dcd4d530d69056c83e9e7f7862766d54a3df5b03043465d12788d2876eac58" gracePeriod=600 Oct 08 18:30:54 crc kubenswrapper[4988]: I1008 18:30:54.307296 4988 generic.go:334] "Generic (PLEG): container finished" podID="f328a20b-5283-407e-8a93-2ab2da599f18" containerID="33dcd4d530d69056c83e9e7f7862766d54a3df5b03043465d12788d2876eac58" exitCode=0 Oct 08 18:30:54 crc kubenswrapper[4988]: I1008 18:30:54.307368 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" event={"ID":"f328a20b-5283-407e-8a93-2ab2da599f18","Type":"ContainerDied","Data":"33dcd4d530d69056c83e9e7f7862766d54a3df5b03043465d12788d2876eac58"} Oct 08 18:30:54 crc kubenswrapper[4988]: I1008 18:30:54.308215 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" event={"ID":"f328a20b-5283-407e-8a93-2ab2da599f18","Type":"ContainerStarted","Data":"e0a9bc6112323fbe5f5506d9ba72b002a9b9faef1191582e4a39898dc43fcbbb"} Oct 08 18:30:54 crc kubenswrapper[4988]: I1008 18:30:54.308246 4988 scope.go:117] "RemoveContainer" containerID="a06add767e00bcfef86e27b752f06b6fb27241526ad4e59aaac22482ba9f2df4" Oct 08 18:30:54 crc kubenswrapper[4988]: I1008 18:30:54.315617 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"1c9bf85f-df23-42e0-b705-69f3938d8208","Type":"ContainerStarted","Data":"8e328a0e54231b368732806a23906cef674b0e9d96a85757598c9092ca88b264"} Oct 08 18:30:56 crc kubenswrapper[4988]: I1008 18:30:56.341337 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"1c9bf85f-df23-42e0-b705-69f3938d8208","Type":"ContainerStarted","Data":"1a2b842baa8e68898cbb9bb630636107329c9d552b7cb0fe66c7526141ce9629"} Oct 08 18:30:56 crc kubenswrapper[4988]: I1008 18:30:56.341799 4988 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="1c9bf85f-df23-42e0-b705-69f3938d8208" containerName="proxy-httpd" containerID="cri-o://1a2b842baa8e68898cbb9bb630636107329c9d552b7cb0fe66c7526141ce9629" gracePeriod=30 Oct 08 18:30:56 crc kubenswrapper[4988]: I1008 18:30:56.341818 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 08 18:30:56 crc kubenswrapper[4988]: I1008 18:30:56.341784 4988 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="1c9bf85f-df23-42e0-b705-69f3938d8208" containerName="ceilometer-notification-agent" containerID="cri-o://ec98ddb44e93e83be1bc8710bfd30c06d306076c89effcf0e3e08a8e03278fdc" gracePeriod=30 Oct 08 18:30:56 crc kubenswrapper[4988]: I1008 18:30:56.341545 4988 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="1c9bf85f-df23-42e0-b705-69f3938d8208" containerName="ceilometer-central-agent" containerID="cri-o://ef6c7ed3962eab0108765f77a0b9afaa599c1c96b4e1a6070f7466d2c9ac9095" gracePeriod=30 Oct 08 18:30:56 crc kubenswrapper[4988]: I1008 18:30:56.341767 4988 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="1c9bf85f-df23-42e0-b705-69f3938d8208" containerName="sg-core" containerID="cri-o://8e328a0e54231b368732806a23906cef674b0e9d96a85757598c9092ca88b264" gracePeriod=30 Oct 08 18:30:56 crc kubenswrapper[4988]: I1008 18:30:56.386303 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.485848338 podStartE2EDuration="6.386279233s" podCreationTimestamp="2025-10-08 18:30:50 +0000 UTC" firstStartedPulling="2025-10-08 18:30:51.244756786 +0000 UTC m=+1196.694599566" lastFinishedPulling="2025-10-08 18:30:55.145187691 +0000 UTC m=+1200.595030461" observedRunningTime="2025-10-08 18:30:56.3642869 +0000 UTC m=+1201.814129680" watchObservedRunningTime="2025-10-08 18:30:56.386279233 +0000 UTC m=+1201.836122003" Oct 08 18:30:57 crc kubenswrapper[4988]: I1008 18:30:57.356064 4988 generic.go:334] "Generic (PLEG): container finished" podID="1c9bf85f-df23-42e0-b705-69f3938d8208" containerID="1a2b842baa8e68898cbb9bb630636107329c9d552b7cb0fe66c7526141ce9629" exitCode=0 Oct 08 18:30:57 crc kubenswrapper[4988]: I1008 18:30:57.356420 4988 generic.go:334] "Generic (PLEG): container finished" podID="1c9bf85f-df23-42e0-b705-69f3938d8208" containerID="8e328a0e54231b368732806a23906cef674b0e9d96a85757598c9092ca88b264" exitCode=2 Oct 08 18:30:57 crc kubenswrapper[4988]: I1008 18:30:57.356434 4988 generic.go:334] "Generic (PLEG): container finished" podID="1c9bf85f-df23-42e0-b705-69f3938d8208" containerID="ec98ddb44e93e83be1bc8710bfd30c06d306076c89effcf0e3e08a8e03278fdc" exitCode=0 Oct 08 18:30:57 crc kubenswrapper[4988]: I1008 18:30:57.356142 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"1c9bf85f-df23-42e0-b705-69f3938d8208","Type":"ContainerDied","Data":"1a2b842baa8e68898cbb9bb630636107329c9d552b7cb0fe66c7526141ce9629"} Oct 08 18:30:57 crc kubenswrapper[4988]: I1008 18:30:57.356481 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"1c9bf85f-df23-42e0-b705-69f3938d8208","Type":"ContainerDied","Data":"8e328a0e54231b368732806a23906cef674b0e9d96a85757598c9092ca88b264"} Oct 08 18:30:57 crc kubenswrapper[4988]: I1008 18:30:57.356501 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"1c9bf85f-df23-42e0-b705-69f3938d8208","Type":"ContainerDied","Data":"ec98ddb44e93e83be1bc8710bfd30c06d306076c89effcf0e3e08a8e03278fdc"} Oct 08 18:31:00 crc kubenswrapper[4988]: I1008 18:31:00.394004 4988 generic.go:334] "Generic (PLEG): container finished" podID="1c9bf85f-df23-42e0-b705-69f3938d8208" containerID="ef6c7ed3962eab0108765f77a0b9afaa599c1c96b4e1a6070f7466d2c9ac9095" exitCode=0 Oct 08 18:31:00 crc kubenswrapper[4988]: I1008 18:31:00.394072 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"1c9bf85f-df23-42e0-b705-69f3938d8208","Type":"ContainerDied","Data":"ef6c7ed3962eab0108765f77a0b9afaa599c1c96b4e1a6070f7466d2c9ac9095"} Oct 08 18:31:00 crc kubenswrapper[4988]: I1008 18:31:00.696859 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 08 18:31:00 crc kubenswrapper[4988]: I1008 18:31:00.734287 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1c9bf85f-df23-42e0-b705-69f3938d8208-log-httpd\") pod \"1c9bf85f-df23-42e0-b705-69f3938d8208\" (UID: \"1c9bf85f-df23-42e0-b705-69f3938d8208\") " Oct 08 18:31:00 crc kubenswrapper[4988]: I1008 18:31:00.734358 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1c9bf85f-df23-42e0-b705-69f3938d8208-scripts\") pod \"1c9bf85f-df23-42e0-b705-69f3938d8208\" (UID: \"1c9bf85f-df23-42e0-b705-69f3938d8208\") " Oct 08 18:31:00 crc kubenswrapper[4988]: I1008 18:31:00.734469 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/1c9bf85f-df23-42e0-b705-69f3938d8208-sg-core-conf-yaml\") pod \"1c9bf85f-df23-42e0-b705-69f3938d8208\" (UID: \"1c9bf85f-df23-42e0-b705-69f3938d8208\") " Oct 08 18:31:00 crc kubenswrapper[4988]: I1008 18:31:00.734525 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1c9bf85f-df23-42e0-b705-69f3938d8208-combined-ca-bundle\") pod \"1c9bf85f-df23-42e0-b705-69f3938d8208\" (UID: \"1c9bf85f-df23-42e0-b705-69f3938d8208\") " Oct 08 18:31:00 crc kubenswrapper[4988]: I1008 18:31:00.734568 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1c9bf85f-df23-42e0-b705-69f3938d8208-run-httpd\") pod \"1c9bf85f-df23-42e0-b705-69f3938d8208\" (UID: \"1c9bf85f-df23-42e0-b705-69f3938d8208\") " Oct 08 18:31:00 crc kubenswrapper[4988]: I1008 18:31:00.734648 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qhfz7\" (UniqueName: \"kubernetes.io/projected/1c9bf85f-df23-42e0-b705-69f3938d8208-kube-api-access-qhfz7\") pod \"1c9bf85f-df23-42e0-b705-69f3938d8208\" (UID: \"1c9bf85f-df23-42e0-b705-69f3938d8208\") " Oct 08 18:31:00 crc kubenswrapper[4988]: I1008 18:31:00.734700 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1c9bf85f-df23-42e0-b705-69f3938d8208-config-data\") pod \"1c9bf85f-df23-42e0-b705-69f3938d8208\" (UID: \"1c9bf85f-df23-42e0-b705-69f3938d8208\") " Oct 08 18:31:00 crc kubenswrapper[4988]: I1008 18:31:00.734869 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1c9bf85f-df23-42e0-b705-69f3938d8208-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "1c9bf85f-df23-42e0-b705-69f3938d8208" (UID: "1c9bf85f-df23-42e0-b705-69f3938d8208"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 18:31:00 crc kubenswrapper[4988]: I1008 18:31:00.735238 4988 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1c9bf85f-df23-42e0-b705-69f3938d8208-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 08 18:31:00 crc kubenswrapper[4988]: I1008 18:31:00.735762 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1c9bf85f-df23-42e0-b705-69f3938d8208-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "1c9bf85f-df23-42e0-b705-69f3938d8208" (UID: "1c9bf85f-df23-42e0-b705-69f3938d8208"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 18:31:00 crc kubenswrapper[4988]: I1008 18:31:00.746545 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1c9bf85f-df23-42e0-b705-69f3938d8208-scripts" (OuterVolumeSpecName: "scripts") pod "1c9bf85f-df23-42e0-b705-69f3938d8208" (UID: "1c9bf85f-df23-42e0-b705-69f3938d8208"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:31:00 crc kubenswrapper[4988]: I1008 18:31:00.749764 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1c9bf85f-df23-42e0-b705-69f3938d8208-kube-api-access-qhfz7" (OuterVolumeSpecName: "kube-api-access-qhfz7") pod "1c9bf85f-df23-42e0-b705-69f3938d8208" (UID: "1c9bf85f-df23-42e0-b705-69f3938d8208"). InnerVolumeSpecName "kube-api-access-qhfz7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:31:00 crc kubenswrapper[4988]: I1008 18:31:00.768766 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1c9bf85f-df23-42e0-b705-69f3938d8208-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "1c9bf85f-df23-42e0-b705-69f3938d8208" (UID: "1c9bf85f-df23-42e0-b705-69f3938d8208"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:31:00 crc kubenswrapper[4988]: I1008 18:31:00.821193 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1c9bf85f-df23-42e0-b705-69f3938d8208-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "1c9bf85f-df23-42e0-b705-69f3938d8208" (UID: "1c9bf85f-df23-42e0-b705-69f3938d8208"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:31:00 crc kubenswrapper[4988]: I1008 18:31:00.836626 4988 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1c9bf85f-df23-42e0-b705-69f3938d8208-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 08 18:31:00 crc kubenswrapper[4988]: I1008 18:31:00.836654 4988 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qhfz7\" (UniqueName: \"kubernetes.io/projected/1c9bf85f-df23-42e0-b705-69f3938d8208-kube-api-access-qhfz7\") on node \"crc\" DevicePath \"\"" Oct 08 18:31:00 crc kubenswrapper[4988]: I1008 18:31:00.836666 4988 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1c9bf85f-df23-42e0-b705-69f3938d8208-scripts\") on node \"crc\" DevicePath \"\"" Oct 08 18:31:00 crc kubenswrapper[4988]: I1008 18:31:00.836674 4988 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/1c9bf85f-df23-42e0-b705-69f3938d8208-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Oct 08 18:31:00 crc kubenswrapper[4988]: I1008 18:31:00.836683 4988 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1c9bf85f-df23-42e0-b705-69f3938d8208-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 18:31:00 crc kubenswrapper[4988]: I1008 18:31:00.861372 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1c9bf85f-df23-42e0-b705-69f3938d8208-config-data" (OuterVolumeSpecName: "config-data") pod "1c9bf85f-df23-42e0-b705-69f3938d8208" (UID: "1c9bf85f-df23-42e0-b705-69f3938d8208"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:31:00 crc kubenswrapper[4988]: I1008 18:31:00.938421 4988 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1c9bf85f-df23-42e0-b705-69f3938d8208-config-data\") on node \"crc\" DevicePath \"\"" Oct 08 18:31:01 crc kubenswrapper[4988]: I1008 18:31:01.407507 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"1c9bf85f-df23-42e0-b705-69f3938d8208","Type":"ContainerDied","Data":"180f4d5f4c235a912e54557a28353c8678d7b0bfe83565a80b8b0cb9e02655e9"} Oct 08 18:31:01 crc kubenswrapper[4988]: I1008 18:31:01.407604 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 08 18:31:01 crc kubenswrapper[4988]: I1008 18:31:01.408497 4988 scope.go:117] "RemoveContainer" containerID="1a2b842baa8e68898cbb9bb630636107329c9d552b7cb0fe66c7526141ce9629" Oct 08 18:31:01 crc kubenswrapper[4988]: I1008 18:31:01.436590 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 08 18:31:01 crc kubenswrapper[4988]: I1008 18:31:01.443754 4988 scope.go:117] "RemoveContainer" containerID="8e328a0e54231b368732806a23906cef674b0e9d96a85757598c9092ca88b264" Oct 08 18:31:01 crc kubenswrapper[4988]: I1008 18:31:01.451301 4988 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Oct 08 18:31:01 crc kubenswrapper[4988]: I1008 18:31:01.469375 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 08 18:31:01 crc kubenswrapper[4988]: E1008 18:31:01.469958 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1c9bf85f-df23-42e0-b705-69f3938d8208" containerName="sg-core" Oct 08 18:31:01 crc kubenswrapper[4988]: I1008 18:31:01.470000 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="1c9bf85f-df23-42e0-b705-69f3938d8208" containerName="sg-core" Oct 08 18:31:01 crc kubenswrapper[4988]: E1008 18:31:01.470023 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1c9bf85f-df23-42e0-b705-69f3938d8208" containerName="ceilometer-central-agent" Oct 08 18:31:01 crc kubenswrapper[4988]: I1008 18:31:01.470032 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="1c9bf85f-df23-42e0-b705-69f3938d8208" containerName="ceilometer-central-agent" Oct 08 18:31:01 crc kubenswrapper[4988]: E1008 18:31:01.470068 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1c9bf85f-df23-42e0-b705-69f3938d8208" containerName="proxy-httpd" Oct 08 18:31:01 crc kubenswrapper[4988]: I1008 18:31:01.470114 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="1c9bf85f-df23-42e0-b705-69f3938d8208" containerName="proxy-httpd" Oct 08 18:31:01 crc kubenswrapper[4988]: E1008 18:31:01.470136 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1c9bf85f-df23-42e0-b705-69f3938d8208" containerName="ceilometer-notification-agent" Oct 08 18:31:01 crc kubenswrapper[4988]: I1008 18:31:01.470144 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="1c9bf85f-df23-42e0-b705-69f3938d8208" containerName="ceilometer-notification-agent" Oct 08 18:31:01 crc kubenswrapper[4988]: I1008 18:31:01.470356 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="1c9bf85f-df23-42e0-b705-69f3938d8208" containerName="sg-core" Oct 08 18:31:01 crc kubenswrapper[4988]: I1008 18:31:01.470417 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="1c9bf85f-df23-42e0-b705-69f3938d8208" containerName="ceilometer-central-agent" Oct 08 18:31:01 crc kubenswrapper[4988]: I1008 18:31:01.470430 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="1c9bf85f-df23-42e0-b705-69f3938d8208" containerName="proxy-httpd" Oct 08 18:31:01 crc kubenswrapper[4988]: I1008 18:31:01.470448 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="1c9bf85f-df23-42e0-b705-69f3938d8208" containerName="ceilometer-notification-agent" Oct 08 18:31:01 crc kubenswrapper[4988]: I1008 18:31:01.472278 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 08 18:31:01 crc kubenswrapper[4988]: I1008 18:31:01.473605 4988 scope.go:117] "RemoveContainer" containerID="ec98ddb44e93e83be1bc8710bfd30c06d306076c89effcf0e3e08a8e03278fdc" Oct 08 18:31:01 crc kubenswrapper[4988]: I1008 18:31:01.480086 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 08 18:31:01 crc kubenswrapper[4988]: I1008 18:31:01.481711 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 08 18:31:01 crc kubenswrapper[4988]: I1008 18:31:01.481901 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 08 18:31:01 crc kubenswrapper[4988]: I1008 18:31:01.517177 4988 scope.go:117] "RemoveContainer" containerID="ef6c7ed3962eab0108765f77a0b9afaa599c1c96b4e1a6070f7466d2c9ac9095" Oct 08 18:31:01 crc kubenswrapper[4988]: I1008 18:31:01.575729 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fa5061f0-e7f7-4059-9f69-48d6a1cb98d7-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"fa5061f0-e7f7-4059-9f69-48d6a1cb98d7\") " pod="openstack/ceilometer-0" Oct 08 18:31:01 crc kubenswrapper[4988]: I1008 18:31:01.575790 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fa5061f0-e7f7-4059-9f69-48d6a1cb98d7-scripts\") pod \"ceilometer-0\" (UID: \"fa5061f0-e7f7-4059-9f69-48d6a1cb98d7\") " pod="openstack/ceilometer-0" Oct 08 18:31:01 crc kubenswrapper[4988]: I1008 18:31:01.575826 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fa5061f0-e7f7-4059-9f69-48d6a1cb98d7-config-data\") pod \"ceilometer-0\" (UID: \"fa5061f0-e7f7-4059-9f69-48d6a1cb98d7\") " pod="openstack/ceilometer-0" Oct 08 18:31:01 crc kubenswrapper[4988]: I1008 18:31:01.575920 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/fa5061f0-e7f7-4059-9f69-48d6a1cb98d7-run-httpd\") pod \"ceilometer-0\" (UID: \"fa5061f0-e7f7-4059-9f69-48d6a1cb98d7\") " pod="openstack/ceilometer-0" Oct 08 18:31:01 crc kubenswrapper[4988]: I1008 18:31:01.575944 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jcxnj\" (UniqueName: \"kubernetes.io/projected/fa5061f0-e7f7-4059-9f69-48d6a1cb98d7-kube-api-access-jcxnj\") pod \"ceilometer-0\" (UID: \"fa5061f0-e7f7-4059-9f69-48d6a1cb98d7\") " pod="openstack/ceilometer-0" Oct 08 18:31:01 crc kubenswrapper[4988]: I1008 18:31:01.575983 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/fa5061f0-e7f7-4059-9f69-48d6a1cb98d7-log-httpd\") pod \"ceilometer-0\" (UID: \"fa5061f0-e7f7-4059-9f69-48d6a1cb98d7\") " pod="openstack/ceilometer-0" Oct 08 18:31:01 crc kubenswrapper[4988]: I1008 18:31:01.576022 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/fa5061f0-e7f7-4059-9f69-48d6a1cb98d7-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"fa5061f0-e7f7-4059-9f69-48d6a1cb98d7\") " pod="openstack/ceilometer-0" Oct 08 18:31:01 crc kubenswrapper[4988]: I1008 18:31:01.676927 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/fa5061f0-e7f7-4059-9f69-48d6a1cb98d7-run-httpd\") pod \"ceilometer-0\" (UID: \"fa5061f0-e7f7-4059-9f69-48d6a1cb98d7\") " pod="openstack/ceilometer-0" Oct 08 18:31:01 crc kubenswrapper[4988]: I1008 18:31:01.676975 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jcxnj\" (UniqueName: \"kubernetes.io/projected/fa5061f0-e7f7-4059-9f69-48d6a1cb98d7-kube-api-access-jcxnj\") pod \"ceilometer-0\" (UID: \"fa5061f0-e7f7-4059-9f69-48d6a1cb98d7\") " pod="openstack/ceilometer-0" Oct 08 18:31:01 crc kubenswrapper[4988]: I1008 18:31:01.677021 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/fa5061f0-e7f7-4059-9f69-48d6a1cb98d7-log-httpd\") pod \"ceilometer-0\" (UID: \"fa5061f0-e7f7-4059-9f69-48d6a1cb98d7\") " pod="openstack/ceilometer-0" Oct 08 18:31:01 crc kubenswrapper[4988]: I1008 18:31:01.677061 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/fa5061f0-e7f7-4059-9f69-48d6a1cb98d7-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"fa5061f0-e7f7-4059-9f69-48d6a1cb98d7\") " pod="openstack/ceilometer-0" Oct 08 18:31:01 crc kubenswrapper[4988]: I1008 18:31:01.677125 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fa5061f0-e7f7-4059-9f69-48d6a1cb98d7-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"fa5061f0-e7f7-4059-9f69-48d6a1cb98d7\") " pod="openstack/ceilometer-0" Oct 08 18:31:01 crc kubenswrapper[4988]: I1008 18:31:01.677154 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fa5061f0-e7f7-4059-9f69-48d6a1cb98d7-scripts\") pod \"ceilometer-0\" (UID: \"fa5061f0-e7f7-4059-9f69-48d6a1cb98d7\") " pod="openstack/ceilometer-0" Oct 08 18:31:01 crc kubenswrapper[4988]: I1008 18:31:01.677178 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fa5061f0-e7f7-4059-9f69-48d6a1cb98d7-config-data\") pod \"ceilometer-0\" (UID: \"fa5061f0-e7f7-4059-9f69-48d6a1cb98d7\") " pod="openstack/ceilometer-0" Oct 08 18:31:01 crc kubenswrapper[4988]: I1008 18:31:01.678498 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/fa5061f0-e7f7-4059-9f69-48d6a1cb98d7-log-httpd\") pod \"ceilometer-0\" (UID: \"fa5061f0-e7f7-4059-9f69-48d6a1cb98d7\") " pod="openstack/ceilometer-0" Oct 08 18:31:01 crc kubenswrapper[4988]: I1008 18:31:01.678868 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/fa5061f0-e7f7-4059-9f69-48d6a1cb98d7-run-httpd\") pod \"ceilometer-0\" (UID: \"fa5061f0-e7f7-4059-9f69-48d6a1cb98d7\") " pod="openstack/ceilometer-0" Oct 08 18:31:01 crc kubenswrapper[4988]: I1008 18:31:01.681493 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fa5061f0-e7f7-4059-9f69-48d6a1cb98d7-config-data\") pod \"ceilometer-0\" (UID: \"fa5061f0-e7f7-4059-9f69-48d6a1cb98d7\") " pod="openstack/ceilometer-0" Oct 08 18:31:01 crc kubenswrapper[4988]: I1008 18:31:01.682087 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fa5061f0-e7f7-4059-9f69-48d6a1cb98d7-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"fa5061f0-e7f7-4059-9f69-48d6a1cb98d7\") " pod="openstack/ceilometer-0" Oct 08 18:31:01 crc kubenswrapper[4988]: I1008 18:31:01.682445 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fa5061f0-e7f7-4059-9f69-48d6a1cb98d7-scripts\") pod \"ceilometer-0\" (UID: \"fa5061f0-e7f7-4059-9f69-48d6a1cb98d7\") " pod="openstack/ceilometer-0" Oct 08 18:31:01 crc kubenswrapper[4988]: I1008 18:31:01.690843 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/fa5061f0-e7f7-4059-9f69-48d6a1cb98d7-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"fa5061f0-e7f7-4059-9f69-48d6a1cb98d7\") " pod="openstack/ceilometer-0" Oct 08 18:31:01 crc kubenswrapper[4988]: I1008 18:31:01.693566 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jcxnj\" (UniqueName: \"kubernetes.io/projected/fa5061f0-e7f7-4059-9f69-48d6a1cb98d7-kube-api-access-jcxnj\") pod \"ceilometer-0\" (UID: \"fa5061f0-e7f7-4059-9f69-48d6a1cb98d7\") " pod="openstack/ceilometer-0" Oct 08 18:31:01 crc kubenswrapper[4988]: I1008 18:31:01.800742 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 08 18:31:02 crc kubenswrapper[4988]: I1008 18:31:02.247848 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 08 18:31:02 crc kubenswrapper[4988]: I1008 18:31:02.415952 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"fa5061f0-e7f7-4059-9f69-48d6a1cb98d7","Type":"ContainerStarted","Data":"cef716ddb91f513c196627b27d20f52e59e2523d565d59958302795fbca835a6"} Oct 08 18:31:02 crc kubenswrapper[4988]: I1008 18:31:02.418899 4988 generic.go:334] "Generic (PLEG): container finished" podID="32722c37-a19b-42e1-be71-432f3499af15" containerID="3bb3e6305405075980f0e6b13cc2d64214327a9883596c00b4d0ce337b183212" exitCode=0 Oct 08 18:31:02 crc kubenswrapper[4988]: I1008 18:31:02.418938 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-hzsjj" event={"ID":"32722c37-a19b-42e1-be71-432f3499af15","Type":"ContainerDied","Data":"3bb3e6305405075980f0e6b13cc2d64214327a9883596c00b4d0ce337b183212"} Oct 08 18:31:03 crc kubenswrapper[4988]: I1008 18:31:03.248695 4988 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1c9bf85f-df23-42e0-b705-69f3938d8208" path="/var/lib/kubelet/pods/1c9bf85f-df23-42e0-b705-69f3938d8208/volumes" Oct 08 18:31:03 crc kubenswrapper[4988]: I1008 18:31:03.430800 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"fa5061f0-e7f7-4059-9f69-48d6a1cb98d7","Type":"ContainerStarted","Data":"4abb52d2340f790b7dbc4c91a708a887af9a63619f61ed6defbeb30cce9e793b"} Oct 08 18:31:03 crc kubenswrapper[4988]: I1008 18:31:03.792907 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-hzsjj" Oct 08 18:31:03 crc kubenswrapper[4988]: I1008 18:31:03.930220 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/32722c37-a19b-42e1-be71-432f3499af15-combined-ca-bundle\") pod \"32722c37-a19b-42e1-be71-432f3499af15\" (UID: \"32722c37-a19b-42e1-be71-432f3499af15\") " Oct 08 18:31:03 crc kubenswrapper[4988]: I1008 18:31:03.930419 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/32722c37-a19b-42e1-be71-432f3499af15-scripts\") pod \"32722c37-a19b-42e1-be71-432f3499af15\" (UID: \"32722c37-a19b-42e1-be71-432f3499af15\") " Oct 08 18:31:03 crc kubenswrapper[4988]: I1008 18:31:03.930458 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/32722c37-a19b-42e1-be71-432f3499af15-config-data\") pod \"32722c37-a19b-42e1-be71-432f3499af15\" (UID: \"32722c37-a19b-42e1-be71-432f3499af15\") " Oct 08 18:31:03 crc kubenswrapper[4988]: I1008 18:31:03.930537 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bdkvj\" (UniqueName: \"kubernetes.io/projected/32722c37-a19b-42e1-be71-432f3499af15-kube-api-access-bdkvj\") pod \"32722c37-a19b-42e1-be71-432f3499af15\" (UID: \"32722c37-a19b-42e1-be71-432f3499af15\") " Oct 08 18:31:03 crc kubenswrapper[4988]: I1008 18:31:03.934549 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/32722c37-a19b-42e1-be71-432f3499af15-kube-api-access-bdkvj" (OuterVolumeSpecName: "kube-api-access-bdkvj") pod "32722c37-a19b-42e1-be71-432f3499af15" (UID: "32722c37-a19b-42e1-be71-432f3499af15"). InnerVolumeSpecName "kube-api-access-bdkvj". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:31:03 crc kubenswrapper[4988]: I1008 18:31:03.943639 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/32722c37-a19b-42e1-be71-432f3499af15-scripts" (OuterVolumeSpecName: "scripts") pod "32722c37-a19b-42e1-be71-432f3499af15" (UID: "32722c37-a19b-42e1-be71-432f3499af15"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:31:03 crc kubenswrapper[4988]: I1008 18:31:03.965813 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/32722c37-a19b-42e1-be71-432f3499af15-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "32722c37-a19b-42e1-be71-432f3499af15" (UID: "32722c37-a19b-42e1-be71-432f3499af15"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:31:03 crc kubenswrapper[4988]: I1008 18:31:03.966928 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/32722c37-a19b-42e1-be71-432f3499af15-config-data" (OuterVolumeSpecName: "config-data") pod "32722c37-a19b-42e1-be71-432f3499af15" (UID: "32722c37-a19b-42e1-be71-432f3499af15"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:31:04 crc kubenswrapper[4988]: I1008 18:31:04.033321 4988 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/32722c37-a19b-42e1-be71-432f3499af15-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 18:31:04 crc kubenswrapper[4988]: I1008 18:31:04.033368 4988 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/32722c37-a19b-42e1-be71-432f3499af15-scripts\") on node \"crc\" DevicePath \"\"" Oct 08 18:31:04 crc kubenswrapper[4988]: I1008 18:31:04.033398 4988 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/32722c37-a19b-42e1-be71-432f3499af15-config-data\") on node \"crc\" DevicePath \"\"" Oct 08 18:31:04 crc kubenswrapper[4988]: I1008 18:31:04.033411 4988 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bdkvj\" (UniqueName: \"kubernetes.io/projected/32722c37-a19b-42e1-be71-432f3499af15-kube-api-access-bdkvj\") on node \"crc\" DevicePath \"\"" Oct 08 18:31:04 crc kubenswrapper[4988]: I1008 18:31:04.440244 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"fa5061f0-e7f7-4059-9f69-48d6a1cb98d7","Type":"ContainerStarted","Data":"4edcb300385ece56d49e71ae3fa0161264857a94d80352e9f4dfc1128294a789"} Oct 08 18:31:04 crc kubenswrapper[4988]: I1008 18:31:04.442041 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-hzsjj" event={"ID":"32722c37-a19b-42e1-be71-432f3499af15","Type":"ContainerDied","Data":"168dd7ae71b862261bfff84ac9282738bd6092f8ec8b1d76f4171a38fd86d4bd"} Oct 08 18:31:04 crc kubenswrapper[4988]: I1008 18:31:04.442079 4988 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="168dd7ae71b862261bfff84ac9282738bd6092f8ec8b1d76f4171a38fd86d4bd" Oct 08 18:31:04 crc kubenswrapper[4988]: I1008 18:31:04.442148 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-hzsjj" Oct 08 18:31:04 crc kubenswrapper[4988]: I1008 18:31:04.527455 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-0"] Oct 08 18:31:04 crc kubenswrapper[4988]: E1008 18:31:04.527858 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="32722c37-a19b-42e1-be71-432f3499af15" containerName="nova-cell0-conductor-db-sync" Oct 08 18:31:04 crc kubenswrapper[4988]: I1008 18:31:04.527876 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="32722c37-a19b-42e1-be71-432f3499af15" containerName="nova-cell0-conductor-db-sync" Oct 08 18:31:04 crc kubenswrapper[4988]: I1008 18:31:04.528074 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="32722c37-a19b-42e1-be71-432f3499af15" containerName="nova-cell0-conductor-db-sync" Oct 08 18:31:04 crc kubenswrapper[4988]: I1008 18:31:04.528808 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Oct 08 18:31:04 crc kubenswrapper[4988]: I1008 18:31:04.531210 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Oct 08 18:31:04 crc kubenswrapper[4988]: I1008 18:31:04.531540 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-vmdxg" Oct 08 18:31:04 crc kubenswrapper[4988]: I1008 18:31:04.568561 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Oct 08 18:31:04 crc kubenswrapper[4988]: I1008 18:31:04.647499 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zb8kd\" (UniqueName: \"kubernetes.io/projected/bd9cfcdf-4770-4efa-a41e-7867ec85206f-kube-api-access-zb8kd\") pod \"nova-cell0-conductor-0\" (UID: \"bd9cfcdf-4770-4efa-a41e-7867ec85206f\") " pod="openstack/nova-cell0-conductor-0" Oct 08 18:31:04 crc kubenswrapper[4988]: I1008 18:31:04.647570 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bd9cfcdf-4770-4efa-a41e-7867ec85206f-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"bd9cfcdf-4770-4efa-a41e-7867ec85206f\") " pod="openstack/nova-cell0-conductor-0" Oct 08 18:31:04 crc kubenswrapper[4988]: I1008 18:31:04.647601 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bd9cfcdf-4770-4efa-a41e-7867ec85206f-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"bd9cfcdf-4770-4efa-a41e-7867ec85206f\") " pod="openstack/nova-cell0-conductor-0" Oct 08 18:31:04 crc kubenswrapper[4988]: I1008 18:31:04.748962 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bd9cfcdf-4770-4efa-a41e-7867ec85206f-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"bd9cfcdf-4770-4efa-a41e-7867ec85206f\") " pod="openstack/nova-cell0-conductor-0" Oct 08 18:31:04 crc kubenswrapper[4988]: I1008 18:31:04.749193 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zb8kd\" (UniqueName: \"kubernetes.io/projected/bd9cfcdf-4770-4efa-a41e-7867ec85206f-kube-api-access-zb8kd\") pod \"nova-cell0-conductor-0\" (UID: \"bd9cfcdf-4770-4efa-a41e-7867ec85206f\") " pod="openstack/nova-cell0-conductor-0" Oct 08 18:31:04 crc kubenswrapper[4988]: I1008 18:31:04.749264 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bd9cfcdf-4770-4efa-a41e-7867ec85206f-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"bd9cfcdf-4770-4efa-a41e-7867ec85206f\") " pod="openstack/nova-cell0-conductor-0" Oct 08 18:31:04 crc kubenswrapper[4988]: I1008 18:31:04.759359 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bd9cfcdf-4770-4efa-a41e-7867ec85206f-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"bd9cfcdf-4770-4efa-a41e-7867ec85206f\") " pod="openstack/nova-cell0-conductor-0" Oct 08 18:31:04 crc kubenswrapper[4988]: I1008 18:31:04.768829 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bd9cfcdf-4770-4efa-a41e-7867ec85206f-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"bd9cfcdf-4770-4efa-a41e-7867ec85206f\") " pod="openstack/nova-cell0-conductor-0" Oct 08 18:31:04 crc kubenswrapper[4988]: I1008 18:31:04.780581 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zb8kd\" (UniqueName: \"kubernetes.io/projected/bd9cfcdf-4770-4efa-a41e-7867ec85206f-kube-api-access-zb8kd\") pod \"nova-cell0-conductor-0\" (UID: \"bd9cfcdf-4770-4efa-a41e-7867ec85206f\") " pod="openstack/nova-cell0-conductor-0" Oct 08 18:31:04 crc kubenswrapper[4988]: I1008 18:31:04.857747 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Oct 08 18:31:05 crc kubenswrapper[4988]: I1008 18:31:05.275660 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Oct 08 18:31:05 crc kubenswrapper[4988]: W1008 18:31:05.278541 4988 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podbd9cfcdf_4770_4efa_a41e_7867ec85206f.slice/crio-fecc6278563eb7471936c98f9e1f21c2d2390670f6fae45a4a28048909f867b7 WatchSource:0}: Error finding container fecc6278563eb7471936c98f9e1f21c2d2390670f6fae45a4a28048909f867b7: Status 404 returned error can't find the container with id fecc6278563eb7471936c98f9e1f21c2d2390670f6fae45a4a28048909f867b7 Oct 08 18:31:05 crc kubenswrapper[4988]: I1008 18:31:05.452778 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"fa5061f0-e7f7-4059-9f69-48d6a1cb98d7","Type":"ContainerStarted","Data":"e19138f089a607cc5db443a8dfe9a62efdb451bc27329b434278497a10f97446"} Oct 08 18:31:05 crc kubenswrapper[4988]: I1008 18:31:05.454568 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"bd9cfcdf-4770-4efa-a41e-7867ec85206f","Type":"ContainerStarted","Data":"fecc6278563eb7471936c98f9e1f21c2d2390670f6fae45a4a28048909f867b7"} Oct 08 18:31:06 crc kubenswrapper[4988]: I1008 18:31:06.469794 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"fa5061f0-e7f7-4059-9f69-48d6a1cb98d7","Type":"ContainerStarted","Data":"7f1af50932c5ecb7086c67d8b967a7c5d6ed5ac407c3ca0a5e4d0a729102c747"} Oct 08 18:31:06 crc kubenswrapper[4988]: I1008 18:31:06.470722 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 08 18:31:06 crc kubenswrapper[4988]: I1008 18:31:06.472575 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"bd9cfcdf-4770-4efa-a41e-7867ec85206f","Type":"ContainerStarted","Data":"17f00a030a73692f1b5f14c014e24ee1e9bf7251c877957ee31dd0cb3579f43c"} Oct 08 18:31:06 crc kubenswrapper[4988]: I1008 18:31:06.472717 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell0-conductor-0" Oct 08 18:31:06 crc kubenswrapper[4988]: I1008 18:31:06.499225 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=1.664187657 podStartE2EDuration="5.499196609s" podCreationTimestamp="2025-10-08 18:31:01 +0000 UTC" firstStartedPulling="2025-10-08 18:31:02.24395522 +0000 UTC m=+1207.693798030" lastFinishedPulling="2025-10-08 18:31:06.078964212 +0000 UTC m=+1211.528806982" observedRunningTime="2025-10-08 18:31:06.492162155 +0000 UTC m=+1211.942004925" watchObservedRunningTime="2025-10-08 18:31:06.499196609 +0000 UTC m=+1211.949039379" Oct 08 18:31:06 crc kubenswrapper[4988]: I1008 18:31:06.518849 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-0" podStartSLOduration=2.518827617 podStartE2EDuration="2.518827617s" podCreationTimestamp="2025-10-08 18:31:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 18:31:06.512598948 +0000 UTC m=+1211.962441728" watchObservedRunningTime="2025-10-08 18:31:06.518827617 +0000 UTC m=+1211.968670387" Oct 08 18:31:14 crc kubenswrapper[4988]: I1008 18:31:14.903868 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell0-conductor-0" Oct 08 18:31:15 crc kubenswrapper[4988]: I1008 18:31:15.458172 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-cell-mapping-tm2wd"] Oct 08 18:31:15 crc kubenswrapper[4988]: I1008 18:31:15.459871 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-tm2wd" Oct 08 18:31:15 crc kubenswrapper[4988]: I1008 18:31:15.466964 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-config-data" Oct 08 18:31:15 crc kubenswrapper[4988]: I1008 18:31:15.467000 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-scripts" Oct 08 18:31:15 crc kubenswrapper[4988]: I1008 18:31:15.476652 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-tm2wd"] Oct 08 18:31:15 crc kubenswrapper[4988]: I1008 18:31:15.557129 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cs5rt\" (UniqueName: \"kubernetes.io/projected/2e2a1e47-ca67-445c-b358-ad7e5bce08ea-kube-api-access-cs5rt\") pod \"nova-cell0-cell-mapping-tm2wd\" (UID: \"2e2a1e47-ca67-445c-b358-ad7e5bce08ea\") " pod="openstack/nova-cell0-cell-mapping-tm2wd" Oct 08 18:31:15 crc kubenswrapper[4988]: I1008 18:31:15.558456 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2e2a1e47-ca67-445c-b358-ad7e5bce08ea-scripts\") pod \"nova-cell0-cell-mapping-tm2wd\" (UID: \"2e2a1e47-ca67-445c-b358-ad7e5bce08ea\") " pod="openstack/nova-cell0-cell-mapping-tm2wd" Oct 08 18:31:15 crc kubenswrapper[4988]: I1008 18:31:15.558487 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2e2a1e47-ca67-445c-b358-ad7e5bce08ea-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-tm2wd\" (UID: \"2e2a1e47-ca67-445c-b358-ad7e5bce08ea\") " pod="openstack/nova-cell0-cell-mapping-tm2wd" Oct 08 18:31:15 crc kubenswrapper[4988]: I1008 18:31:15.558509 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2e2a1e47-ca67-445c-b358-ad7e5bce08ea-config-data\") pod \"nova-cell0-cell-mapping-tm2wd\" (UID: \"2e2a1e47-ca67-445c-b358-ad7e5bce08ea\") " pod="openstack/nova-cell0-cell-mapping-tm2wd" Oct 08 18:31:15 crc kubenswrapper[4988]: I1008 18:31:15.618505 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Oct 08 18:31:15 crc kubenswrapper[4988]: I1008 18:31:15.620281 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 08 18:31:15 crc kubenswrapper[4988]: I1008 18:31:15.623293 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Oct 08 18:31:15 crc kubenswrapper[4988]: I1008 18:31:15.636963 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 08 18:31:15 crc kubenswrapper[4988]: I1008 18:31:15.661161 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a116a7eb-0c4f-4d6b-93ef-93ac26f22441-config-data\") pod \"nova-api-0\" (UID: \"a116a7eb-0c4f-4d6b-93ef-93ac26f22441\") " pod="openstack/nova-api-0" Oct 08 18:31:15 crc kubenswrapper[4988]: I1008 18:31:15.661519 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2e2a1e47-ca67-445c-b358-ad7e5bce08ea-scripts\") pod \"nova-cell0-cell-mapping-tm2wd\" (UID: \"2e2a1e47-ca67-445c-b358-ad7e5bce08ea\") " pod="openstack/nova-cell0-cell-mapping-tm2wd" Oct 08 18:31:15 crc kubenswrapper[4988]: I1008 18:31:15.661579 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2e2a1e47-ca67-445c-b358-ad7e5bce08ea-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-tm2wd\" (UID: \"2e2a1e47-ca67-445c-b358-ad7e5bce08ea\") " pod="openstack/nova-cell0-cell-mapping-tm2wd" Oct 08 18:31:15 crc kubenswrapper[4988]: I1008 18:31:15.661621 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a116a7eb-0c4f-4d6b-93ef-93ac26f22441-logs\") pod \"nova-api-0\" (UID: \"a116a7eb-0c4f-4d6b-93ef-93ac26f22441\") " pod="openstack/nova-api-0" Oct 08 18:31:15 crc kubenswrapper[4988]: I1008 18:31:15.661652 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2e2a1e47-ca67-445c-b358-ad7e5bce08ea-config-data\") pod \"nova-cell0-cell-mapping-tm2wd\" (UID: \"2e2a1e47-ca67-445c-b358-ad7e5bce08ea\") " pod="openstack/nova-cell0-cell-mapping-tm2wd" Oct 08 18:31:15 crc kubenswrapper[4988]: I1008 18:31:15.661791 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a116a7eb-0c4f-4d6b-93ef-93ac26f22441-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"a116a7eb-0c4f-4d6b-93ef-93ac26f22441\") " pod="openstack/nova-api-0" Oct 08 18:31:15 crc kubenswrapper[4988]: I1008 18:31:15.661862 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cs5rt\" (UniqueName: \"kubernetes.io/projected/2e2a1e47-ca67-445c-b358-ad7e5bce08ea-kube-api-access-cs5rt\") pod \"nova-cell0-cell-mapping-tm2wd\" (UID: \"2e2a1e47-ca67-445c-b358-ad7e5bce08ea\") " pod="openstack/nova-cell0-cell-mapping-tm2wd" Oct 08 18:31:15 crc kubenswrapper[4988]: I1008 18:31:15.661915 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cd84r\" (UniqueName: \"kubernetes.io/projected/a116a7eb-0c4f-4d6b-93ef-93ac26f22441-kube-api-access-cd84r\") pod \"nova-api-0\" (UID: \"a116a7eb-0c4f-4d6b-93ef-93ac26f22441\") " pod="openstack/nova-api-0" Oct 08 18:31:15 crc kubenswrapper[4988]: I1008 18:31:15.686290 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2e2a1e47-ca67-445c-b358-ad7e5bce08ea-scripts\") pod \"nova-cell0-cell-mapping-tm2wd\" (UID: \"2e2a1e47-ca67-445c-b358-ad7e5bce08ea\") " pod="openstack/nova-cell0-cell-mapping-tm2wd" Oct 08 18:31:15 crc kubenswrapper[4988]: I1008 18:31:15.688476 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2e2a1e47-ca67-445c-b358-ad7e5bce08ea-config-data\") pod \"nova-cell0-cell-mapping-tm2wd\" (UID: \"2e2a1e47-ca67-445c-b358-ad7e5bce08ea\") " pod="openstack/nova-cell0-cell-mapping-tm2wd" Oct 08 18:31:15 crc kubenswrapper[4988]: I1008 18:31:15.690152 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2e2a1e47-ca67-445c-b358-ad7e5bce08ea-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-tm2wd\" (UID: \"2e2a1e47-ca67-445c-b358-ad7e5bce08ea\") " pod="openstack/nova-cell0-cell-mapping-tm2wd" Oct 08 18:31:15 crc kubenswrapper[4988]: I1008 18:31:15.699131 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cs5rt\" (UniqueName: \"kubernetes.io/projected/2e2a1e47-ca67-445c-b358-ad7e5bce08ea-kube-api-access-cs5rt\") pod \"nova-cell0-cell-mapping-tm2wd\" (UID: \"2e2a1e47-ca67-445c-b358-ad7e5bce08ea\") " pod="openstack/nova-cell0-cell-mapping-tm2wd" Oct 08 18:31:15 crc kubenswrapper[4988]: I1008 18:31:15.726457 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Oct 08 18:31:15 crc kubenswrapper[4988]: I1008 18:31:15.734618 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 08 18:31:15 crc kubenswrapper[4988]: I1008 18:31:15.749789 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Oct 08 18:31:15 crc kubenswrapper[4988]: I1008 18:31:15.767424 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a116a7eb-0c4f-4d6b-93ef-93ac26f22441-config-data\") pod \"nova-api-0\" (UID: \"a116a7eb-0c4f-4d6b-93ef-93ac26f22441\") " pod="openstack/nova-api-0" Oct 08 18:31:15 crc kubenswrapper[4988]: I1008 18:31:15.767510 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a116a7eb-0c4f-4d6b-93ef-93ac26f22441-logs\") pod \"nova-api-0\" (UID: \"a116a7eb-0c4f-4d6b-93ef-93ac26f22441\") " pod="openstack/nova-api-0" Oct 08 18:31:15 crc kubenswrapper[4988]: I1008 18:31:15.767567 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a116a7eb-0c4f-4d6b-93ef-93ac26f22441-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"a116a7eb-0c4f-4d6b-93ef-93ac26f22441\") " pod="openstack/nova-api-0" Oct 08 18:31:15 crc kubenswrapper[4988]: I1008 18:31:15.767601 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cd84r\" (UniqueName: \"kubernetes.io/projected/a116a7eb-0c4f-4d6b-93ef-93ac26f22441-kube-api-access-cd84r\") pod \"nova-api-0\" (UID: \"a116a7eb-0c4f-4d6b-93ef-93ac26f22441\") " pod="openstack/nova-api-0" Oct 08 18:31:15 crc kubenswrapper[4988]: I1008 18:31:15.768890 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a116a7eb-0c4f-4d6b-93ef-93ac26f22441-logs\") pod \"nova-api-0\" (UID: \"a116a7eb-0c4f-4d6b-93ef-93ac26f22441\") " pod="openstack/nova-api-0" Oct 08 18:31:15 crc kubenswrapper[4988]: I1008 18:31:15.784078 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a116a7eb-0c4f-4d6b-93ef-93ac26f22441-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"a116a7eb-0c4f-4d6b-93ef-93ac26f22441\") " pod="openstack/nova-api-0" Oct 08 18:31:15 crc kubenswrapper[4988]: I1008 18:31:15.784901 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a116a7eb-0c4f-4d6b-93ef-93ac26f22441-config-data\") pod \"nova-api-0\" (UID: \"a116a7eb-0c4f-4d6b-93ef-93ac26f22441\") " pod="openstack/nova-api-0" Oct 08 18:31:15 crc kubenswrapper[4988]: I1008 18:31:15.792997 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 08 18:31:15 crc kubenswrapper[4988]: I1008 18:31:15.797000 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cd84r\" (UniqueName: \"kubernetes.io/projected/a116a7eb-0c4f-4d6b-93ef-93ac26f22441-kube-api-access-cd84r\") pod \"nova-api-0\" (UID: \"a116a7eb-0c4f-4d6b-93ef-93ac26f22441\") " pod="openstack/nova-api-0" Oct 08 18:31:15 crc kubenswrapper[4988]: I1008 18:31:15.797370 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-tm2wd" Oct 08 18:31:15 crc kubenswrapper[4988]: I1008 18:31:15.868729 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6ca97861-a9a4-47c4-ac53-050d5106c0ba-config-data\") pod \"nova-metadata-0\" (UID: \"6ca97861-a9a4-47c4-ac53-050d5106c0ba\") " pod="openstack/nova-metadata-0" Oct 08 18:31:15 crc kubenswrapper[4988]: I1008 18:31:15.868768 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6ca97861-a9a4-47c4-ac53-050d5106c0ba-logs\") pod \"nova-metadata-0\" (UID: \"6ca97861-a9a4-47c4-ac53-050d5106c0ba\") " pod="openstack/nova-metadata-0" Oct 08 18:31:15 crc kubenswrapper[4988]: I1008 18:31:15.868836 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6ca97861-a9a4-47c4-ac53-050d5106c0ba-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"6ca97861-a9a4-47c4-ac53-050d5106c0ba\") " pod="openstack/nova-metadata-0" Oct 08 18:31:15 crc kubenswrapper[4988]: I1008 18:31:15.868890 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-59kzp\" (UniqueName: \"kubernetes.io/projected/6ca97861-a9a4-47c4-ac53-050d5106c0ba-kube-api-access-59kzp\") pod \"nova-metadata-0\" (UID: \"6ca97861-a9a4-47c4-ac53-050d5106c0ba\") " pod="openstack/nova-metadata-0" Oct 08 18:31:15 crc kubenswrapper[4988]: I1008 18:31:15.945468 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 08 18:31:15 crc kubenswrapper[4988]: I1008 18:31:15.972357 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6ca97861-a9a4-47c4-ac53-050d5106c0ba-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"6ca97861-a9a4-47c4-ac53-050d5106c0ba\") " pod="openstack/nova-metadata-0" Oct 08 18:31:15 crc kubenswrapper[4988]: I1008 18:31:15.972456 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-59kzp\" (UniqueName: \"kubernetes.io/projected/6ca97861-a9a4-47c4-ac53-050d5106c0ba-kube-api-access-59kzp\") pod \"nova-metadata-0\" (UID: \"6ca97861-a9a4-47c4-ac53-050d5106c0ba\") " pod="openstack/nova-metadata-0" Oct 08 18:31:15 crc kubenswrapper[4988]: I1008 18:31:15.972520 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6ca97861-a9a4-47c4-ac53-050d5106c0ba-config-data\") pod \"nova-metadata-0\" (UID: \"6ca97861-a9a4-47c4-ac53-050d5106c0ba\") " pod="openstack/nova-metadata-0" Oct 08 18:31:15 crc kubenswrapper[4988]: I1008 18:31:15.972570 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6ca97861-a9a4-47c4-ac53-050d5106c0ba-logs\") pod \"nova-metadata-0\" (UID: \"6ca97861-a9a4-47c4-ac53-050d5106c0ba\") " pod="openstack/nova-metadata-0" Oct 08 18:31:15 crc kubenswrapper[4988]: I1008 18:31:15.973629 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6ca97861-a9a4-47c4-ac53-050d5106c0ba-logs\") pod \"nova-metadata-0\" (UID: \"6ca97861-a9a4-47c4-ac53-050d5106c0ba\") " pod="openstack/nova-metadata-0" Oct 08 18:31:15 crc kubenswrapper[4988]: I1008 18:31:15.980853 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6ca97861-a9a4-47c4-ac53-050d5106c0ba-config-data\") pod \"nova-metadata-0\" (UID: \"6ca97861-a9a4-47c4-ac53-050d5106c0ba\") " pod="openstack/nova-metadata-0" Oct 08 18:31:15 crc kubenswrapper[4988]: I1008 18:31:15.986098 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6ca97861-a9a4-47c4-ac53-050d5106c0ba-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"6ca97861-a9a4-47c4-ac53-050d5106c0ba\") " pod="openstack/nova-metadata-0" Oct 08 18:31:16 crc kubenswrapper[4988]: I1008 18:31:16.003222 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 08 18:31:16 crc kubenswrapper[4988]: I1008 18:31:16.004695 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Oct 08 18:31:16 crc kubenswrapper[4988]: I1008 18:31:16.008196 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Oct 08 18:31:16 crc kubenswrapper[4988]: I1008 18:31:16.014046 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-59kzp\" (UniqueName: \"kubernetes.io/projected/6ca97861-a9a4-47c4-ac53-050d5106c0ba-kube-api-access-59kzp\") pod \"nova-metadata-0\" (UID: \"6ca97861-a9a4-47c4-ac53-050d5106c0ba\") " pod="openstack/nova-metadata-0" Oct 08 18:31:16 crc kubenswrapper[4988]: I1008 18:31:16.029905 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-64d8d96789-8bhgs"] Oct 08 18:31:16 crc kubenswrapper[4988]: I1008 18:31:16.047465 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-64d8d96789-8bhgs" Oct 08 18:31:16 crc kubenswrapper[4988]: I1008 18:31:16.063240 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 08 18:31:16 crc kubenswrapper[4988]: I1008 18:31:16.072687 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Oct 08 18:31:16 crc kubenswrapper[4988]: I1008 18:31:16.073856 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 08 18:31:16 crc kubenswrapper[4988]: I1008 18:31:16.075197 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d8520f38-f3e7-4a05-ac90-ed48db1b2ba0-ovsdbserver-nb\") pod \"dnsmasq-dns-64d8d96789-8bhgs\" (UID: \"d8520f38-f3e7-4a05-ac90-ed48db1b2ba0\") " pod="openstack/dnsmasq-dns-64d8d96789-8bhgs" Oct 08 18:31:16 crc kubenswrapper[4988]: I1008 18:31:16.075336 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lv7tb\" (UniqueName: \"kubernetes.io/projected/110929bd-ec2a-4f95-8cb1-ace9284ec881-kube-api-access-lv7tb\") pod \"nova-cell1-novncproxy-0\" (UID: \"110929bd-ec2a-4f95-8cb1-ace9284ec881\") " pod="openstack/nova-cell1-novncproxy-0" Oct 08 18:31:16 crc kubenswrapper[4988]: I1008 18:31:16.075473 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d8520f38-f3e7-4a05-ac90-ed48db1b2ba0-dns-svc\") pod \"dnsmasq-dns-64d8d96789-8bhgs\" (UID: \"d8520f38-f3e7-4a05-ac90-ed48db1b2ba0\") " pod="openstack/dnsmasq-dns-64d8d96789-8bhgs" Oct 08 18:31:16 crc kubenswrapper[4988]: I1008 18:31:16.075502 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d8520f38-f3e7-4a05-ac90-ed48db1b2ba0-config\") pod \"dnsmasq-dns-64d8d96789-8bhgs\" (UID: \"d8520f38-f3e7-4a05-ac90-ed48db1b2ba0\") " pod="openstack/dnsmasq-dns-64d8d96789-8bhgs" Oct 08 18:31:16 crc kubenswrapper[4988]: I1008 18:31:16.075825 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/110929bd-ec2a-4f95-8cb1-ace9284ec881-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"110929bd-ec2a-4f95-8cb1-ace9284ec881\") " pod="openstack/nova-cell1-novncproxy-0" Oct 08 18:31:16 crc kubenswrapper[4988]: I1008 18:31:16.075853 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rrgsq\" (UniqueName: \"kubernetes.io/projected/d8520f38-f3e7-4a05-ac90-ed48db1b2ba0-kube-api-access-rrgsq\") pod \"dnsmasq-dns-64d8d96789-8bhgs\" (UID: \"d8520f38-f3e7-4a05-ac90-ed48db1b2ba0\") " pod="openstack/dnsmasq-dns-64d8d96789-8bhgs" Oct 08 18:31:16 crc kubenswrapper[4988]: I1008 18:31:16.075911 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/110929bd-ec2a-4f95-8cb1-ace9284ec881-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"110929bd-ec2a-4f95-8cb1-ace9284ec881\") " pod="openstack/nova-cell1-novncproxy-0" Oct 08 18:31:16 crc kubenswrapper[4988]: I1008 18:31:16.076079 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d8520f38-f3e7-4a05-ac90-ed48db1b2ba0-ovsdbserver-sb\") pod \"dnsmasq-dns-64d8d96789-8bhgs\" (UID: \"d8520f38-f3e7-4a05-ac90-ed48db1b2ba0\") " pod="openstack/dnsmasq-dns-64d8d96789-8bhgs" Oct 08 18:31:16 crc kubenswrapper[4988]: I1008 18:31:16.076136 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/d8520f38-f3e7-4a05-ac90-ed48db1b2ba0-dns-swift-storage-0\") pod \"dnsmasq-dns-64d8d96789-8bhgs\" (UID: \"d8520f38-f3e7-4a05-ac90-ed48db1b2ba0\") " pod="openstack/dnsmasq-dns-64d8d96789-8bhgs" Oct 08 18:31:16 crc kubenswrapper[4988]: I1008 18:31:16.077318 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Oct 08 18:31:16 crc kubenswrapper[4988]: I1008 18:31:16.166688 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-64d8d96789-8bhgs"] Oct 08 18:31:16 crc kubenswrapper[4988]: I1008 18:31:16.177308 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/d8520f38-f3e7-4a05-ac90-ed48db1b2ba0-dns-swift-storage-0\") pod \"dnsmasq-dns-64d8d96789-8bhgs\" (UID: \"d8520f38-f3e7-4a05-ac90-ed48db1b2ba0\") " pod="openstack/dnsmasq-dns-64d8d96789-8bhgs" Oct 08 18:31:16 crc kubenswrapper[4988]: I1008 18:31:16.177379 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d8520f38-f3e7-4a05-ac90-ed48db1b2ba0-ovsdbserver-nb\") pod \"dnsmasq-dns-64d8d96789-8bhgs\" (UID: \"d8520f38-f3e7-4a05-ac90-ed48db1b2ba0\") " pod="openstack/dnsmasq-dns-64d8d96789-8bhgs" Oct 08 18:31:16 crc kubenswrapper[4988]: I1008 18:31:16.177412 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lv7tb\" (UniqueName: \"kubernetes.io/projected/110929bd-ec2a-4f95-8cb1-ace9284ec881-kube-api-access-lv7tb\") pod \"nova-cell1-novncproxy-0\" (UID: \"110929bd-ec2a-4f95-8cb1-ace9284ec881\") " pod="openstack/nova-cell1-novncproxy-0" Oct 08 18:31:16 crc kubenswrapper[4988]: I1008 18:31:16.177449 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d8520f38-f3e7-4a05-ac90-ed48db1b2ba0-dns-svc\") pod \"dnsmasq-dns-64d8d96789-8bhgs\" (UID: \"d8520f38-f3e7-4a05-ac90-ed48db1b2ba0\") " pod="openstack/dnsmasq-dns-64d8d96789-8bhgs" Oct 08 18:31:16 crc kubenswrapper[4988]: I1008 18:31:16.177468 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d8520f38-f3e7-4a05-ac90-ed48db1b2ba0-config\") pod \"dnsmasq-dns-64d8d96789-8bhgs\" (UID: \"d8520f38-f3e7-4a05-ac90-ed48db1b2ba0\") " pod="openstack/dnsmasq-dns-64d8d96789-8bhgs" Oct 08 18:31:16 crc kubenswrapper[4988]: I1008 18:31:16.177504 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0e1348e9-ee35-49e7-a616-af30789c7ce0-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"0e1348e9-ee35-49e7-a616-af30789c7ce0\") " pod="openstack/nova-scheduler-0" Oct 08 18:31:16 crc kubenswrapper[4988]: I1008 18:31:16.177535 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/110929bd-ec2a-4f95-8cb1-ace9284ec881-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"110929bd-ec2a-4f95-8cb1-ace9284ec881\") " pod="openstack/nova-cell1-novncproxy-0" Oct 08 18:31:16 crc kubenswrapper[4988]: I1008 18:31:16.177559 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rrgsq\" (UniqueName: \"kubernetes.io/projected/d8520f38-f3e7-4a05-ac90-ed48db1b2ba0-kube-api-access-rrgsq\") pod \"dnsmasq-dns-64d8d96789-8bhgs\" (UID: \"d8520f38-f3e7-4a05-ac90-ed48db1b2ba0\") " pod="openstack/dnsmasq-dns-64d8d96789-8bhgs" Oct 08 18:31:16 crc kubenswrapper[4988]: I1008 18:31:16.177582 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sfj8s\" (UniqueName: \"kubernetes.io/projected/0e1348e9-ee35-49e7-a616-af30789c7ce0-kube-api-access-sfj8s\") pod \"nova-scheduler-0\" (UID: \"0e1348e9-ee35-49e7-a616-af30789c7ce0\") " pod="openstack/nova-scheduler-0" Oct 08 18:31:16 crc kubenswrapper[4988]: I1008 18:31:16.177600 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0e1348e9-ee35-49e7-a616-af30789c7ce0-config-data\") pod \"nova-scheduler-0\" (UID: \"0e1348e9-ee35-49e7-a616-af30789c7ce0\") " pod="openstack/nova-scheduler-0" Oct 08 18:31:16 crc kubenswrapper[4988]: I1008 18:31:16.177618 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/110929bd-ec2a-4f95-8cb1-ace9284ec881-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"110929bd-ec2a-4f95-8cb1-ace9284ec881\") " pod="openstack/nova-cell1-novncproxy-0" Oct 08 18:31:16 crc kubenswrapper[4988]: I1008 18:31:16.177652 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d8520f38-f3e7-4a05-ac90-ed48db1b2ba0-ovsdbserver-sb\") pod \"dnsmasq-dns-64d8d96789-8bhgs\" (UID: \"d8520f38-f3e7-4a05-ac90-ed48db1b2ba0\") " pod="openstack/dnsmasq-dns-64d8d96789-8bhgs" Oct 08 18:31:16 crc kubenswrapper[4988]: I1008 18:31:16.183280 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/d8520f38-f3e7-4a05-ac90-ed48db1b2ba0-dns-swift-storage-0\") pod \"dnsmasq-dns-64d8d96789-8bhgs\" (UID: \"d8520f38-f3e7-4a05-ac90-ed48db1b2ba0\") " pod="openstack/dnsmasq-dns-64d8d96789-8bhgs" Oct 08 18:31:16 crc kubenswrapper[4988]: I1008 18:31:16.183881 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d8520f38-f3e7-4a05-ac90-ed48db1b2ba0-ovsdbserver-nb\") pod \"dnsmasq-dns-64d8d96789-8bhgs\" (UID: \"d8520f38-f3e7-4a05-ac90-ed48db1b2ba0\") " pod="openstack/dnsmasq-dns-64d8d96789-8bhgs" Oct 08 18:31:16 crc kubenswrapper[4988]: I1008 18:31:16.184776 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d8520f38-f3e7-4a05-ac90-ed48db1b2ba0-dns-svc\") pod \"dnsmasq-dns-64d8d96789-8bhgs\" (UID: \"d8520f38-f3e7-4a05-ac90-ed48db1b2ba0\") " pod="openstack/dnsmasq-dns-64d8d96789-8bhgs" Oct 08 18:31:16 crc kubenswrapper[4988]: I1008 18:31:16.185302 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d8520f38-f3e7-4a05-ac90-ed48db1b2ba0-config\") pod \"dnsmasq-dns-64d8d96789-8bhgs\" (UID: \"d8520f38-f3e7-4a05-ac90-ed48db1b2ba0\") " pod="openstack/dnsmasq-dns-64d8d96789-8bhgs" Oct 08 18:31:16 crc kubenswrapper[4988]: I1008 18:31:16.195894 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d8520f38-f3e7-4a05-ac90-ed48db1b2ba0-ovsdbserver-sb\") pod \"dnsmasq-dns-64d8d96789-8bhgs\" (UID: \"d8520f38-f3e7-4a05-ac90-ed48db1b2ba0\") " pod="openstack/dnsmasq-dns-64d8d96789-8bhgs" Oct 08 18:31:16 crc kubenswrapper[4988]: I1008 18:31:16.202357 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/110929bd-ec2a-4f95-8cb1-ace9284ec881-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"110929bd-ec2a-4f95-8cb1-ace9284ec881\") " pod="openstack/nova-cell1-novncproxy-0" Oct 08 18:31:16 crc kubenswrapper[4988]: I1008 18:31:16.211089 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/110929bd-ec2a-4f95-8cb1-ace9284ec881-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"110929bd-ec2a-4f95-8cb1-ace9284ec881\") " pod="openstack/nova-cell1-novncproxy-0" Oct 08 18:31:16 crc kubenswrapper[4988]: I1008 18:31:16.211229 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lv7tb\" (UniqueName: \"kubernetes.io/projected/110929bd-ec2a-4f95-8cb1-ace9284ec881-kube-api-access-lv7tb\") pod \"nova-cell1-novncproxy-0\" (UID: \"110929bd-ec2a-4f95-8cb1-ace9284ec881\") " pod="openstack/nova-cell1-novncproxy-0" Oct 08 18:31:16 crc kubenswrapper[4988]: I1008 18:31:16.246469 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rrgsq\" (UniqueName: \"kubernetes.io/projected/d8520f38-f3e7-4a05-ac90-ed48db1b2ba0-kube-api-access-rrgsq\") pod \"dnsmasq-dns-64d8d96789-8bhgs\" (UID: \"d8520f38-f3e7-4a05-ac90-ed48db1b2ba0\") " pod="openstack/dnsmasq-dns-64d8d96789-8bhgs" Oct 08 18:31:16 crc kubenswrapper[4988]: I1008 18:31:16.259744 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Oct 08 18:31:16 crc kubenswrapper[4988]: I1008 18:31:16.277707 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 08 18:31:16 crc kubenswrapper[4988]: I1008 18:31:16.279288 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0e1348e9-ee35-49e7-a616-af30789c7ce0-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"0e1348e9-ee35-49e7-a616-af30789c7ce0\") " pod="openstack/nova-scheduler-0" Oct 08 18:31:16 crc kubenswrapper[4988]: I1008 18:31:16.279341 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sfj8s\" (UniqueName: \"kubernetes.io/projected/0e1348e9-ee35-49e7-a616-af30789c7ce0-kube-api-access-sfj8s\") pod \"nova-scheduler-0\" (UID: \"0e1348e9-ee35-49e7-a616-af30789c7ce0\") " pod="openstack/nova-scheduler-0" Oct 08 18:31:16 crc kubenswrapper[4988]: I1008 18:31:16.279361 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0e1348e9-ee35-49e7-a616-af30789c7ce0-config-data\") pod \"nova-scheduler-0\" (UID: \"0e1348e9-ee35-49e7-a616-af30789c7ce0\") " pod="openstack/nova-scheduler-0" Oct 08 18:31:16 crc kubenswrapper[4988]: I1008 18:31:16.304319 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0e1348e9-ee35-49e7-a616-af30789c7ce0-config-data\") pod \"nova-scheduler-0\" (UID: \"0e1348e9-ee35-49e7-a616-af30789c7ce0\") " pod="openstack/nova-scheduler-0" Oct 08 18:31:16 crc kubenswrapper[4988]: I1008 18:31:16.324366 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0e1348e9-ee35-49e7-a616-af30789c7ce0-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"0e1348e9-ee35-49e7-a616-af30789c7ce0\") " pod="openstack/nova-scheduler-0" Oct 08 18:31:16 crc kubenswrapper[4988]: I1008 18:31:16.338302 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sfj8s\" (UniqueName: \"kubernetes.io/projected/0e1348e9-ee35-49e7-a616-af30789c7ce0-kube-api-access-sfj8s\") pod \"nova-scheduler-0\" (UID: \"0e1348e9-ee35-49e7-a616-af30789c7ce0\") " pod="openstack/nova-scheduler-0" Oct 08 18:31:16 crc kubenswrapper[4988]: I1008 18:31:16.348825 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Oct 08 18:31:16 crc kubenswrapper[4988]: I1008 18:31:16.388261 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-64d8d96789-8bhgs" Oct 08 18:31:16 crc kubenswrapper[4988]: I1008 18:31:16.445890 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 08 18:31:16 crc kubenswrapper[4988]: I1008 18:31:16.647918 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-tm2wd"] Oct 08 18:31:16 crc kubenswrapper[4988]: I1008 18:31:16.839852 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 08 18:31:16 crc kubenswrapper[4988]: W1008 18:31:16.852525 4988 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda116a7eb_0c4f_4d6b_93ef_93ac26f22441.slice/crio-3f333c9fc45c29edb605c4f707b65a326fff530a56e6dded0a50bcb4776049ee WatchSource:0}: Error finding container 3f333c9fc45c29edb605c4f707b65a326fff530a56e6dded0a50bcb4776049ee: Status 404 returned error can't find the container with id 3f333c9fc45c29edb605c4f707b65a326fff530a56e6dded0a50bcb4776049ee Oct 08 18:31:17 crc kubenswrapper[4988]: I1008 18:31:16.997456 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-db-sync-rx7cw"] Oct 08 18:31:17 crc kubenswrapper[4988]: I1008 18:31:16.999165 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-rx7cw" Oct 08 18:31:17 crc kubenswrapper[4988]: I1008 18:31:17.009650 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-scripts" Oct 08 18:31:17 crc kubenswrapper[4988]: I1008 18:31:17.009919 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Oct 08 18:31:17 crc kubenswrapper[4988]: I1008 18:31:17.014673 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 08 18:31:17 crc kubenswrapper[4988]: I1008 18:31:17.028127 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-rx7cw"] Oct 08 18:31:17 crc kubenswrapper[4988]: I1008 18:31:17.103243 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f4q8v\" (UniqueName: \"kubernetes.io/projected/7aaaa954-1cc6-4304-8a21-86d9dfe4fd23-kube-api-access-f4q8v\") pod \"nova-cell1-conductor-db-sync-rx7cw\" (UID: \"7aaaa954-1cc6-4304-8a21-86d9dfe4fd23\") " pod="openstack/nova-cell1-conductor-db-sync-rx7cw" Oct 08 18:31:17 crc kubenswrapper[4988]: I1008 18:31:17.104317 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7aaaa954-1cc6-4304-8a21-86d9dfe4fd23-scripts\") pod \"nova-cell1-conductor-db-sync-rx7cw\" (UID: \"7aaaa954-1cc6-4304-8a21-86d9dfe4fd23\") " pod="openstack/nova-cell1-conductor-db-sync-rx7cw" Oct 08 18:31:17 crc kubenswrapper[4988]: I1008 18:31:17.104399 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7aaaa954-1cc6-4304-8a21-86d9dfe4fd23-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-rx7cw\" (UID: \"7aaaa954-1cc6-4304-8a21-86d9dfe4fd23\") " pod="openstack/nova-cell1-conductor-db-sync-rx7cw" Oct 08 18:31:17 crc kubenswrapper[4988]: I1008 18:31:17.104638 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7aaaa954-1cc6-4304-8a21-86d9dfe4fd23-config-data\") pod \"nova-cell1-conductor-db-sync-rx7cw\" (UID: \"7aaaa954-1cc6-4304-8a21-86d9dfe4fd23\") " pod="openstack/nova-cell1-conductor-db-sync-rx7cw" Oct 08 18:31:17 crc kubenswrapper[4988]: I1008 18:31:17.110022 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Oct 08 18:31:17 crc kubenswrapper[4988]: I1008 18:31:17.132873 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 08 18:31:17 crc kubenswrapper[4988]: I1008 18:31:17.139858 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-64d8d96789-8bhgs"] Oct 08 18:31:17 crc kubenswrapper[4988]: I1008 18:31:17.206599 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7aaaa954-1cc6-4304-8a21-86d9dfe4fd23-config-data\") pod \"nova-cell1-conductor-db-sync-rx7cw\" (UID: \"7aaaa954-1cc6-4304-8a21-86d9dfe4fd23\") " pod="openstack/nova-cell1-conductor-db-sync-rx7cw" Oct 08 18:31:17 crc kubenswrapper[4988]: I1008 18:31:17.206688 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f4q8v\" (UniqueName: \"kubernetes.io/projected/7aaaa954-1cc6-4304-8a21-86d9dfe4fd23-kube-api-access-f4q8v\") pod \"nova-cell1-conductor-db-sync-rx7cw\" (UID: \"7aaaa954-1cc6-4304-8a21-86d9dfe4fd23\") " pod="openstack/nova-cell1-conductor-db-sync-rx7cw" Oct 08 18:31:17 crc kubenswrapper[4988]: I1008 18:31:17.206767 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7aaaa954-1cc6-4304-8a21-86d9dfe4fd23-scripts\") pod \"nova-cell1-conductor-db-sync-rx7cw\" (UID: \"7aaaa954-1cc6-4304-8a21-86d9dfe4fd23\") " pod="openstack/nova-cell1-conductor-db-sync-rx7cw" Oct 08 18:31:17 crc kubenswrapper[4988]: I1008 18:31:17.206863 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7aaaa954-1cc6-4304-8a21-86d9dfe4fd23-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-rx7cw\" (UID: \"7aaaa954-1cc6-4304-8a21-86d9dfe4fd23\") " pod="openstack/nova-cell1-conductor-db-sync-rx7cw" Oct 08 18:31:17 crc kubenswrapper[4988]: I1008 18:31:17.212587 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7aaaa954-1cc6-4304-8a21-86d9dfe4fd23-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-rx7cw\" (UID: \"7aaaa954-1cc6-4304-8a21-86d9dfe4fd23\") " pod="openstack/nova-cell1-conductor-db-sync-rx7cw" Oct 08 18:31:17 crc kubenswrapper[4988]: I1008 18:31:17.217574 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7aaaa954-1cc6-4304-8a21-86d9dfe4fd23-config-data\") pod \"nova-cell1-conductor-db-sync-rx7cw\" (UID: \"7aaaa954-1cc6-4304-8a21-86d9dfe4fd23\") " pod="openstack/nova-cell1-conductor-db-sync-rx7cw" Oct 08 18:31:17 crc kubenswrapper[4988]: I1008 18:31:17.217905 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7aaaa954-1cc6-4304-8a21-86d9dfe4fd23-scripts\") pod \"nova-cell1-conductor-db-sync-rx7cw\" (UID: \"7aaaa954-1cc6-4304-8a21-86d9dfe4fd23\") " pod="openstack/nova-cell1-conductor-db-sync-rx7cw" Oct 08 18:31:17 crc kubenswrapper[4988]: I1008 18:31:17.230961 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f4q8v\" (UniqueName: \"kubernetes.io/projected/7aaaa954-1cc6-4304-8a21-86d9dfe4fd23-kube-api-access-f4q8v\") pod \"nova-cell1-conductor-db-sync-rx7cw\" (UID: \"7aaaa954-1cc6-4304-8a21-86d9dfe4fd23\") " pod="openstack/nova-cell1-conductor-db-sync-rx7cw" Oct 08 18:31:17 crc kubenswrapper[4988]: I1008 18:31:17.366533 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-rx7cw" Oct 08 18:31:17 crc kubenswrapper[4988]: I1008 18:31:17.596302 4988 generic.go:334] "Generic (PLEG): container finished" podID="d8520f38-f3e7-4a05-ac90-ed48db1b2ba0" containerID="290de12fdada84cf292143bedd05a23659d15e3a50f137b76d4f26b1c2aad8af" exitCode=0 Oct 08 18:31:17 crc kubenswrapper[4988]: I1008 18:31:17.596634 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-64d8d96789-8bhgs" event={"ID":"d8520f38-f3e7-4a05-ac90-ed48db1b2ba0","Type":"ContainerDied","Data":"290de12fdada84cf292143bedd05a23659d15e3a50f137b76d4f26b1c2aad8af"} Oct 08 18:31:17 crc kubenswrapper[4988]: I1008 18:31:17.596661 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-64d8d96789-8bhgs" event={"ID":"d8520f38-f3e7-4a05-ac90-ed48db1b2ba0","Type":"ContainerStarted","Data":"13baa2b9d20cd176d4848124d8f8a2a626e9072d1f3957f1137263212c3dfdd4"} Oct 08 18:31:17 crc kubenswrapper[4988]: I1008 18:31:17.600667 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"0e1348e9-ee35-49e7-a616-af30789c7ce0","Type":"ContainerStarted","Data":"385df180470d78bda7bfcc67ab85a705f50ca98fbc5b551aef915c3c148e720f"} Oct 08 18:31:17 crc kubenswrapper[4988]: I1008 18:31:17.613660 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"110929bd-ec2a-4f95-8cb1-ace9284ec881","Type":"ContainerStarted","Data":"d4d78f18f9b8f0e83ea0c0c7fb72ff91fbf5ac5068d047c2f09df2854eef18d9"} Oct 08 18:31:17 crc kubenswrapper[4988]: I1008 18:31:17.618491 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"6ca97861-a9a4-47c4-ac53-050d5106c0ba","Type":"ContainerStarted","Data":"7e0dcad3d50707ce3c55ba1ac671954489c015eafeb27294a7c91400c102e365"} Oct 08 18:31:17 crc kubenswrapper[4988]: I1008 18:31:17.643423 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"a116a7eb-0c4f-4d6b-93ef-93ac26f22441","Type":"ContainerStarted","Data":"3f333c9fc45c29edb605c4f707b65a326fff530a56e6dded0a50bcb4776049ee"} Oct 08 18:31:17 crc kubenswrapper[4988]: I1008 18:31:17.658001 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-tm2wd" event={"ID":"2e2a1e47-ca67-445c-b358-ad7e5bce08ea","Type":"ContainerStarted","Data":"8187466829ece57145a6f580c49fd29ced64f87be389192fb463c718a5acae13"} Oct 08 18:31:17 crc kubenswrapper[4988]: I1008 18:31:17.658048 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-tm2wd" event={"ID":"2e2a1e47-ca67-445c-b358-ad7e5bce08ea","Type":"ContainerStarted","Data":"a9b44797feb4f30d8638636a4c736665a21d3738f85f1f855295676ef65fc84a"} Oct 08 18:31:17 crc kubenswrapper[4988]: I1008 18:31:17.676928 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-cell-mapping-tm2wd" podStartSLOduration=2.6769104219999997 podStartE2EDuration="2.676910422s" podCreationTimestamp="2025-10-08 18:31:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 18:31:17.672558853 +0000 UTC m=+1223.122401643" watchObservedRunningTime="2025-10-08 18:31:17.676910422 +0000 UTC m=+1223.126753192" Oct 08 18:31:17 crc kubenswrapper[4988]: I1008 18:31:17.926350 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-rx7cw"] Oct 08 18:31:17 crc kubenswrapper[4988]: W1008 18:31:17.963261 4988 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7aaaa954_1cc6_4304_8a21_86d9dfe4fd23.slice/crio-61929a33e24fafc1b48794f1d99db9657b578e9d2af7db1a64e60d1547f47bfd WatchSource:0}: Error finding container 61929a33e24fafc1b48794f1d99db9657b578e9d2af7db1a64e60d1547f47bfd: Status 404 returned error can't find the container with id 61929a33e24fafc1b48794f1d99db9657b578e9d2af7db1a64e60d1547f47bfd Oct 08 18:31:18 crc kubenswrapper[4988]: I1008 18:31:18.668100 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-64d8d96789-8bhgs" event={"ID":"d8520f38-f3e7-4a05-ac90-ed48db1b2ba0","Type":"ContainerStarted","Data":"e73abd7244e715c31f953913bd7ab824186e18b2167e4cab3bfaa5f8522ce848"} Oct 08 18:31:18 crc kubenswrapper[4988]: I1008 18:31:18.668645 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-64d8d96789-8bhgs" Oct 08 18:31:18 crc kubenswrapper[4988]: I1008 18:31:18.670541 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-rx7cw" event={"ID":"7aaaa954-1cc6-4304-8a21-86d9dfe4fd23","Type":"ContainerStarted","Data":"170776f5aecd85fd95ec5552397ec9b090c5fec990524b633f8d6b1d51a1c74e"} Oct 08 18:31:18 crc kubenswrapper[4988]: I1008 18:31:18.670573 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-rx7cw" event={"ID":"7aaaa954-1cc6-4304-8a21-86d9dfe4fd23","Type":"ContainerStarted","Data":"61929a33e24fafc1b48794f1d99db9657b578e9d2af7db1a64e60d1547f47bfd"} Oct 08 18:31:18 crc kubenswrapper[4988]: I1008 18:31:18.701273 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-64d8d96789-8bhgs" podStartSLOduration=3.7012542550000003 podStartE2EDuration="3.701254255s" podCreationTimestamp="2025-10-08 18:31:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 18:31:18.687887127 +0000 UTC m=+1224.137729897" watchObservedRunningTime="2025-10-08 18:31:18.701254255 +0000 UTC m=+1224.151097025" Oct 08 18:31:18 crc kubenswrapper[4988]: I1008 18:31:18.709243 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-db-sync-rx7cw" podStartSLOduration=2.70922854 podStartE2EDuration="2.70922854s" podCreationTimestamp="2025-10-08 18:31:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 18:31:18.704951763 +0000 UTC m=+1224.154794533" watchObservedRunningTime="2025-10-08 18:31:18.70922854 +0000 UTC m=+1224.159071310" Oct 08 18:31:19 crc kubenswrapper[4988]: I1008 18:31:19.576209 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Oct 08 18:31:19 crc kubenswrapper[4988]: I1008 18:31:19.589763 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 08 18:31:21 crc kubenswrapper[4988]: I1008 18:31:21.695783 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"0e1348e9-ee35-49e7-a616-af30789c7ce0","Type":"ContainerStarted","Data":"2d3963fa42e1cd55fec9ceb64219ccf309ed4efb871f3ffc012626613b72c62c"} Oct 08 18:31:21 crc kubenswrapper[4988]: I1008 18:31:21.698020 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"110929bd-ec2a-4f95-8cb1-ace9284ec881","Type":"ContainerStarted","Data":"31ef7ae1b9b13e2a236ac5de546a23be6667e216cee78fec4f9e3fb4e96f6183"} Oct 08 18:31:21 crc kubenswrapper[4988]: I1008 18:31:21.698058 4988 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell1-novncproxy-0" podUID="110929bd-ec2a-4f95-8cb1-ace9284ec881" containerName="nova-cell1-novncproxy-novncproxy" containerID="cri-o://31ef7ae1b9b13e2a236ac5de546a23be6667e216cee78fec4f9e3fb4e96f6183" gracePeriod=30 Oct 08 18:31:21 crc kubenswrapper[4988]: I1008 18:31:21.700657 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"6ca97861-a9a4-47c4-ac53-050d5106c0ba","Type":"ContainerStarted","Data":"75d10d7be99bacab7f577b0eddbb16950c6741fadc52e5ce925d458a2e0b95dc"} Oct 08 18:31:21 crc kubenswrapper[4988]: I1008 18:31:21.700699 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"6ca97861-a9a4-47c4-ac53-050d5106c0ba","Type":"ContainerStarted","Data":"f657d957d5f97effb90019c1de5adf1d0a048476716b1ea587df56ca1b86b231"} Oct 08 18:31:21 crc kubenswrapper[4988]: I1008 18:31:21.700740 4988 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="6ca97861-a9a4-47c4-ac53-050d5106c0ba" containerName="nova-metadata-log" containerID="cri-o://f657d957d5f97effb90019c1de5adf1d0a048476716b1ea587df56ca1b86b231" gracePeriod=30 Oct 08 18:31:21 crc kubenswrapper[4988]: I1008 18:31:21.700756 4988 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="6ca97861-a9a4-47c4-ac53-050d5106c0ba" containerName="nova-metadata-metadata" containerID="cri-o://75d10d7be99bacab7f577b0eddbb16950c6741fadc52e5ce925d458a2e0b95dc" gracePeriod=30 Oct 08 18:31:21 crc kubenswrapper[4988]: I1008 18:31:21.707816 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"a116a7eb-0c4f-4d6b-93ef-93ac26f22441","Type":"ContainerStarted","Data":"f1e67556c910c6012c3493e80adf12b19a788da6c247b9ef602743fd99cb3b7a"} Oct 08 18:31:21 crc kubenswrapper[4988]: I1008 18:31:21.707867 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"a116a7eb-0c4f-4d6b-93ef-93ac26f22441","Type":"ContainerStarted","Data":"8b312e4e6552af848a89843736fde9707853545f97ff16dd7eb3c47b05d0c32d"} Oct 08 18:31:21 crc kubenswrapper[4988]: I1008 18:31:21.725578 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=3.422711269 podStartE2EDuration="6.725530955s" podCreationTimestamp="2025-10-08 18:31:15 +0000 UTC" firstStartedPulling="2025-10-08 18:31:17.112861897 +0000 UTC m=+1222.562704667" lastFinishedPulling="2025-10-08 18:31:20.415681573 +0000 UTC m=+1225.865524353" observedRunningTime="2025-10-08 18:31:21.721355722 +0000 UTC m=+1227.171198512" watchObservedRunningTime="2025-10-08 18:31:21.725530955 +0000 UTC m=+1227.175373725" Oct 08 18:31:21 crc kubenswrapper[4988]: I1008 18:31:21.739986 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=3.488313857 podStartE2EDuration="6.739967616s" podCreationTimestamp="2025-10-08 18:31:15 +0000 UTC" firstStartedPulling="2025-10-08 18:31:17.14764281 +0000 UTC m=+1222.597485580" lastFinishedPulling="2025-10-08 18:31:20.399296569 +0000 UTC m=+1225.849139339" observedRunningTime="2025-10-08 18:31:21.734351287 +0000 UTC m=+1227.184194057" watchObservedRunningTime="2025-10-08 18:31:21.739967616 +0000 UTC m=+1227.189810386" Oct 08 18:31:21 crc kubenswrapper[4988]: I1008 18:31:21.763029 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=3.330031556 podStartE2EDuration="6.763013763s" podCreationTimestamp="2025-10-08 18:31:15 +0000 UTC" firstStartedPulling="2025-10-08 18:31:17.012891491 +0000 UTC m=+1222.462734261" lastFinishedPulling="2025-10-08 18:31:20.445873698 +0000 UTC m=+1225.895716468" observedRunningTime="2025-10-08 18:31:21.752496107 +0000 UTC m=+1227.202338887" watchObservedRunningTime="2025-10-08 18:31:21.763013763 +0000 UTC m=+1227.212856533" Oct 08 18:31:21 crc kubenswrapper[4988]: I1008 18:31:21.784502 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=3.221462384 podStartE2EDuration="6.784480179s" podCreationTimestamp="2025-10-08 18:31:15 +0000 UTC" firstStartedPulling="2025-10-08 18:31:16.855089105 +0000 UTC m=+1222.304931875" lastFinishedPulling="2025-10-08 18:31:20.4181069 +0000 UTC m=+1225.867949670" observedRunningTime="2025-10-08 18:31:21.778181139 +0000 UTC m=+1227.228023909" watchObservedRunningTime="2025-10-08 18:31:21.784480179 +0000 UTC m=+1227.234322949" Oct 08 18:31:22 crc kubenswrapper[4988]: I1008 18:31:22.273454 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 08 18:31:22 crc kubenswrapper[4988]: I1008 18:31:22.343411 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6ca97861-a9a4-47c4-ac53-050d5106c0ba-logs\") pod \"6ca97861-a9a4-47c4-ac53-050d5106c0ba\" (UID: \"6ca97861-a9a4-47c4-ac53-050d5106c0ba\") " Oct 08 18:31:22 crc kubenswrapper[4988]: I1008 18:31:22.343873 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6ca97861-a9a4-47c4-ac53-050d5106c0ba-logs" (OuterVolumeSpecName: "logs") pod "6ca97861-a9a4-47c4-ac53-050d5106c0ba" (UID: "6ca97861-a9a4-47c4-ac53-050d5106c0ba"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 18:31:22 crc kubenswrapper[4988]: I1008 18:31:22.343929 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6ca97861-a9a4-47c4-ac53-050d5106c0ba-combined-ca-bundle\") pod \"6ca97861-a9a4-47c4-ac53-050d5106c0ba\" (UID: \"6ca97861-a9a4-47c4-ac53-050d5106c0ba\") " Oct 08 18:31:22 crc kubenswrapper[4988]: I1008 18:31:22.344000 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6ca97861-a9a4-47c4-ac53-050d5106c0ba-config-data\") pod \"6ca97861-a9a4-47c4-ac53-050d5106c0ba\" (UID: \"6ca97861-a9a4-47c4-ac53-050d5106c0ba\") " Oct 08 18:31:22 crc kubenswrapper[4988]: I1008 18:31:22.344090 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-59kzp\" (UniqueName: \"kubernetes.io/projected/6ca97861-a9a4-47c4-ac53-050d5106c0ba-kube-api-access-59kzp\") pod \"6ca97861-a9a4-47c4-ac53-050d5106c0ba\" (UID: \"6ca97861-a9a4-47c4-ac53-050d5106c0ba\") " Oct 08 18:31:22 crc kubenswrapper[4988]: I1008 18:31:22.344617 4988 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6ca97861-a9a4-47c4-ac53-050d5106c0ba-logs\") on node \"crc\" DevicePath \"\"" Oct 08 18:31:22 crc kubenswrapper[4988]: I1008 18:31:22.350803 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6ca97861-a9a4-47c4-ac53-050d5106c0ba-kube-api-access-59kzp" (OuterVolumeSpecName: "kube-api-access-59kzp") pod "6ca97861-a9a4-47c4-ac53-050d5106c0ba" (UID: "6ca97861-a9a4-47c4-ac53-050d5106c0ba"). InnerVolumeSpecName "kube-api-access-59kzp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:31:22 crc kubenswrapper[4988]: I1008 18:31:22.374629 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6ca97861-a9a4-47c4-ac53-050d5106c0ba-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "6ca97861-a9a4-47c4-ac53-050d5106c0ba" (UID: "6ca97861-a9a4-47c4-ac53-050d5106c0ba"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:31:22 crc kubenswrapper[4988]: I1008 18:31:22.384569 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6ca97861-a9a4-47c4-ac53-050d5106c0ba-config-data" (OuterVolumeSpecName: "config-data") pod "6ca97861-a9a4-47c4-ac53-050d5106c0ba" (UID: "6ca97861-a9a4-47c4-ac53-050d5106c0ba"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:31:22 crc kubenswrapper[4988]: I1008 18:31:22.446708 4988 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6ca97861-a9a4-47c4-ac53-050d5106c0ba-config-data\") on node \"crc\" DevicePath \"\"" Oct 08 18:31:22 crc kubenswrapper[4988]: I1008 18:31:22.446740 4988 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-59kzp\" (UniqueName: \"kubernetes.io/projected/6ca97861-a9a4-47c4-ac53-050d5106c0ba-kube-api-access-59kzp\") on node \"crc\" DevicePath \"\"" Oct 08 18:31:22 crc kubenswrapper[4988]: I1008 18:31:22.446752 4988 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6ca97861-a9a4-47c4-ac53-050d5106c0ba-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 18:31:22 crc kubenswrapper[4988]: I1008 18:31:22.718461 4988 generic.go:334] "Generic (PLEG): container finished" podID="6ca97861-a9a4-47c4-ac53-050d5106c0ba" containerID="75d10d7be99bacab7f577b0eddbb16950c6741fadc52e5ce925d458a2e0b95dc" exitCode=0 Oct 08 18:31:22 crc kubenswrapper[4988]: I1008 18:31:22.718492 4988 generic.go:334] "Generic (PLEG): container finished" podID="6ca97861-a9a4-47c4-ac53-050d5106c0ba" containerID="f657d957d5f97effb90019c1de5adf1d0a048476716b1ea587df56ca1b86b231" exitCode=143 Oct 08 18:31:22 crc kubenswrapper[4988]: I1008 18:31:22.718527 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"6ca97861-a9a4-47c4-ac53-050d5106c0ba","Type":"ContainerDied","Data":"75d10d7be99bacab7f577b0eddbb16950c6741fadc52e5ce925d458a2e0b95dc"} Oct 08 18:31:22 crc kubenswrapper[4988]: I1008 18:31:22.718568 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"6ca97861-a9a4-47c4-ac53-050d5106c0ba","Type":"ContainerDied","Data":"f657d957d5f97effb90019c1de5adf1d0a048476716b1ea587df56ca1b86b231"} Oct 08 18:31:22 crc kubenswrapper[4988]: I1008 18:31:22.718581 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"6ca97861-a9a4-47c4-ac53-050d5106c0ba","Type":"ContainerDied","Data":"7e0dcad3d50707ce3c55ba1ac671954489c015eafeb27294a7c91400c102e365"} Oct 08 18:31:22 crc kubenswrapper[4988]: I1008 18:31:22.718565 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 08 18:31:22 crc kubenswrapper[4988]: I1008 18:31:22.718587 4988 scope.go:117] "RemoveContainer" containerID="75d10d7be99bacab7f577b0eddbb16950c6741fadc52e5ce925d458a2e0b95dc" Oct 08 18:31:22 crc kubenswrapper[4988]: I1008 18:31:22.753711 4988 scope.go:117] "RemoveContainer" containerID="f657d957d5f97effb90019c1de5adf1d0a048476716b1ea587df56ca1b86b231" Oct 08 18:31:22 crc kubenswrapper[4988]: I1008 18:31:22.772037 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Oct 08 18:31:22 crc kubenswrapper[4988]: I1008 18:31:22.784124 4988 scope.go:117] "RemoveContainer" containerID="75d10d7be99bacab7f577b0eddbb16950c6741fadc52e5ce925d458a2e0b95dc" Oct 08 18:31:22 crc kubenswrapper[4988]: E1008 18:31:22.784628 4988 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"75d10d7be99bacab7f577b0eddbb16950c6741fadc52e5ce925d458a2e0b95dc\": container with ID starting with 75d10d7be99bacab7f577b0eddbb16950c6741fadc52e5ce925d458a2e0b95dc not found: ID does not exist" containerID="75d10d7be99bacab7f577b0eddbb16950c6741fadc52e5ce925d458a2e0b95dc" Oct 08 18:31:22 crc kubenswrapper[4988]: I1008 18:31:22.784666 4988 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Oct 08 18:31:22 crc kubenswrapper[4988]: I1008 18:31:22.784676 4988 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"75d10d7be99bacab7f577b0eddbb16950c6741fadc52e5ce925d458a2e0b95dc"} err="failed to get container status \"75d10d7be99bacab7f577b0eddbb16950c6741fadc52e5ce925d458a2e0b95dc\": rpc error: code = NotFound desc = could not find container \"75d10d7be99bacab7f577b0eddbb16950c6741fadc52e5ce925d458a2e0b95dc\": container with ID starting with 75d10d7be99bacab7f577b0eddbb16950c6741fadc52e5ce925d458a2e0b95dc not found: ID does not exist" Oct 08 18:31:22 crc kubenswrapper[4988]: I1008 18:31:22.784780 4988 scope.go:117] "RemoveContainer" containerID="f657d957d5f97effb90019c1de5adf1d0a048476716b1ea587df56ca1b86b231" Oct 08 18:31:22 crc kubenswrapper[4988]: E1008 18:31:22.785469 4988 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f657d957d5f97effb90019c1de5adf1d0a048476716b1ea587df56ca1b86b231\": container with ID starting with f657d957d5f97effb90019c1de5adf1d0a048476716b1ea587df56ca1b86b231 not found: ID does not exist" containerID="f657d957d5f97effb90019c1de5adf1d0a048476716b1ea587df56ca1b86b231" Oct 08 18:31:22 crc kubenswrapper[4988]: I1008 18:31:22.785508 4988 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f657d957d5f97effb90019c1de5adf1d0a048476716b1ea587df56ca1b86b231"} err="failed to get container status \"f657d957d5f97effb90019c1de5adf1d0a048476716b1ea587df56ca1b86b231\": rpc error: code = NotFound desc = could not find container \"f657d957d5f97effb90019c1de5adf1d0a048476716b1ea587df56ca1b86b231\": container with ID starting with f657d957d5f97effb90019c1de5adf1d0a048476716b1ea587df56ca1b86b231 not found: ID does not exist" Oct 08 18:31:22 crc kubenswrapper[4988]: I1008 18:31:22.785529 4988 scope.go:117] "RemoveContainer" containerID="75d10d7be99bacab7f577b0eddbb16950c6741fadc52e5ce925d458a2e0b95dc" Oct 08 18:31:22 crc kubenswrapper[4988]: I1008 18:31:22.789538 4988 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"75d10d7be99bacab7f577b0eddbb16950c6741fadc52e5ce925d458a2e0b95dc"} err="failed to get container status \"75d10d7be99bacab7f577b0eddbb16950c6741fadc52e5ce925d458a2e0b95dc\": rpc error: code = NotFound desc = could not find container \"75d10d7be99bacab7f577b0eddbb16950c6741fadc52e5ce925d458a2e0b95dc\": container with ID starting with 75d10d7be99bacab7f577b0eddbb16950c6741fadc52e5ce925d458a2e0b95dc not found: ID does not exist" Oct 08 18:31:22 crc kubenswrapper[4988]: I1008 18:31:22.789591 4988 scope.go:117] "RemoveContainer" containerID="f657d957d5f97effb90019c1de5adf1d0a048476716b1ea587df56ca1b86b231" Oct 08 18:31:22 crc kubenswrapper[4988]: I1008 18:31:22.789964 4988 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f657d957d5f97effb90019c1de5adf1d0a048476716b1ea587df56ca1b86b231"} err="failed to get container status \"f657d957d5f97effb90019c1de5adf1d0a048476716b1ea587df56ca1b86b231\": rpc error: code = NotFound desc = could not find container \"f657d957d5f97effb90019c1de5adf1d0a048476716b1ea587df56ca1b86b231\": container with ID starting with f657d957d5f97effb90019c1de5adf1d0a048476716b1ea587df56ca1b86b231 not found: ID does not exist" Oct 08 18:31:22 crc kubenswrapper[4988]: I1008 18:31:22.802358 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Oct 08 18:31:22 crc kubenswrapper[4988]: E1008 18:31:22.802846 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6ca97861-a9a4-47c4-ac53-050d5106c0ba" containerName="nova-metadata-log" Oct 08 18:31:22 crc kubenswrapper[4988]: I1008 18:31:22.802866 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="6ca97861-a9a4-47c4-ac53-050d5106c0ba" containerName="nova-metadata-log" Oct 08 18:31:22 crc kubenswrapper[4988]: E1008 18:31:22.802894 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6ca97861-a9a4-47c4-ac53-050d5106c0ba" containerName="nova-metadata-metadata" Oct 08 18:31:22 crc kubenswrapper[4988]: I1008 18:31:22.802905 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="6ca97861-a9a4-47c4-ac53-050d5106c0ba" containerName="nova-metadata-metadata" Oct 08 18:31:22 crc kubenswrapper[4988]: I1008 18:31:22.803133 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="6ca97861-a9a4-47c4-ac53-050d5106c0ba" containerName="nova-metadata-metadata" Oct 08 18:31:22 crc kubenswrapper[4988]: I1008 18:31:22.803165 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="6ca97861-a9a4-47c4-ac53-050d5106c0ba" containerName="nova-metadata-log" Oct 08 18:31:22 crc kubenswrapper[4988]: I1008 18:31:22.804341 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 08 18:31:22 crc kubenswrapper[4988]: I1008 18:31:22.807858 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Oct 08 18:31:22 crc kubenswrapper[4988]: I1008 18:31:22.808233 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Oct 08 18:31:22 crc kubenswrapper[4988]: I1008 18:31:22.824168 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 08 18:31:22 crc kubenswrapper[4988]: I1008 18:31:22.856530 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b5a36680-3152-4d1f-a97d-92a3f1a3bc6e-logs\") pod \"nova-metadata-0\" (UID: \"b5a36680-3152-4d1f-a97d-92a3f1a3bc6e\") " pod="openstack/nova-metadata-0" Oct 08 18:31:22 crc kubenswrapper[4988]: I1008 18:31:22.856586 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xfv8k\" (UniqueName: \"kubernetes.io/projected/b5a36680-3152-4d1f-a97d-92a3f1a3bc6e-kube-api-access-xfv8k\") pod \"nova-metadata-0\" (UID: \"b5a36680-3152-4d1f-a97d-92a3f1a3bc6e\") " pod="openstack/nova-metadata-0" Oct 08 18:31:22 crc kubenswrapper[4988]: I1008 18:31:22.856620 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/b5a36680-3152-4d1f-a97d-92a3f1a3bc6e-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"b5a36680-3152-4d1f-a97d-92a3f1a3bc6e\") " pod="openstack/nova-metadata-0" Oct 08 18:31:22 crc kubenswrapper[4988]: I1008 18:31:22.856690 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b5a36680-3152-4d1f-a97d-92a3f1a3bc6e-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"b5a36680-3152-4d1f-a97d-92a3f1a3bc6e\") " pod="openstack/nova-metadata-0" Oct 08 18:31:22 crc kubenswrapper[4988]: I1008 18:31:22.856762 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b5a36680-3152-4d1f-a97d-92a3f1a3bc6e-config-data\") pod \"nova-metadata-0\" (UID: \"b5a36680-3152-4d1f-a97d-92a3f1a3bc6e\") " pod="openstack/nova-metadata-0" Oct 08 18:31:22 crc kubenswrapper[4988]: I1008 18:31:22.958738 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b5a36680-3152-4d1f-a97d-92a3f1a3bc6e-logs\") pod \"nova-metadata-0\" (UID: \"b5a36680-3152-4d1f-a97d-92a3f1a3bc6e\") " pod="openstack/nova-metadata-0" Oct 08 18:31:22 crc kubenswrapper[4988]: I1008 18:31:22.958795 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xfv8k\" (UniqueName: \"kubernetes.io/projected/b5a36680-3152-4d1f-a97d-92a3f1a3bc6e-kube-api-access-xfv8k\") pod \"nova-metadata-0\" (UID: \"b5a36680-3152-4d1f-a97d-92a3f1a3bc6e\") " pod="openstack/nova-metadata-0" Oct 08 18:31:22 crc kubenswrapper[4988]: I1008 18:31:22.958833 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/b5a36680-3152-4d1f-a97d-92a3f1a3bc6e-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"b5a36680-3152-4d1f-a97d-92a3f1a3bc6e\") " pod="openstack/nova-metadata-0" Oct 08 18:31:22 crc kubenswrapper[4988]: I1008 18:31:22.958890 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b5a36680-3152-4d1f-a97d-92a3f1a3bc6e-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"b5a36680-3152-4d1f-a97d-92a3f1a3bc6e\") " pod="openstack/nova-metadata-0" Oct 08 18:31:22 crc kubenswrapper[4988]: I1008 18:31:22.958944 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b5a36680-3152-4d1f-a97d-92a3f1a3bc6e-config-data\") pod \"nova-metadata-0\" (UID: \"b5a36680-3152-4d1f-a97d-92a3f1a3bc6e\") " pod="openstack/nova-metadata-0" Oct 08 18:31:22 crc kubenswrapper[4988]: I1008 18:31:22.959323 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b5a36680-3152-4d1f-a97d-92a3f1a3bc6e-logs\") pod \"nova-metadata-0\" (UID: \"b5a36680-3152-4d1f-a97d-92a3f1a3bc6e\") " pod="openstack/nova-metadata-0" Oct 08 18:31:22 crc kubenswrapper[4988]: I1008 18:31:22.963039 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b5a36680-3152-4d1f-a97d-92a3f1a3bc6e-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"b5a36680-3152-4d1f-a97d-92a3f1a3bc6e\") " pod="openstack/nova-metadata-0" Oct 08 18:31:22 crc kubenswrapper[4988]: I1008 18:31:22.965005 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/b5a36680-3152-4d1f-a97d-92a3f1a3bc6e-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"b5a36680-3152-4d1f-a97d-92a3f1a3bc6e\") " pod="openstack/nova-metadata-0" Oct 08 18:31:22 crc kubenswrapper[4988]: I1008 18:31:22.978011 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b5a36680-3152-4d1f-a97d-92a3f1a3bc6e-config-data\") pod \"nova-metadata-0\" (UID: \"b5a36680-3152-4d1f-a97d-92a3f1a3bc6e\") " pod="openstack/nova-metadata-0" Oct 08 18:31:22 crc kubenswrapper[4988]: I1008 18:31:22.978074 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xfv8k\" (UniqueName: \"kubernetes.io/projected/b5a36680-3152-4d1f-a97d-92a3f1a3bc6e-kube-api-access-xfv8k\") pod \"nova-metadata-0\" (UID: \"b5a36680-3152-4d1f-a97d-92a3f1a3bc6e\") " pod="openstack/nova-metadata-0" Oct 08 18:31:23 crc kubenswrapper[4988]: I1008 18:31:23.127572 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 08 18:31:23 crc kubenswrapper[4988]: I1008 18:31:23.253794 4988 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6ca97861-a9a4-47c4-ac53-050d5106c0ba" path="/var/lib/kubelet/pods/6ca97861-a9a4-47c4-ac53-050d5106c0ba/volumes" Oct 08 18:31:23 crc kubenswrapper[4988]: I1008 18:31:23.651464 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 08 18:31:23 crc kubenswrapper[4988]: I1008 18:31:23.731751 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"b5a36680-3152-4d1f-a97d-92a3f1a3bc6e","Type":"ContainerStarted","Data":"36195b6ddcc1c5790944a7a2f25a56451d2550ab56480548b12800f80aa8d62a"} Oct 08 18:31:24 crc kubenswrapper[4988]: I1008 18:31:24.746008 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"b5a36680-3152-4d1f-a97d-92a3f1a3bc6e","Type":"ContainerStarted","Data":"b92e21969a50a5f6d64686041134b3dbd486d24db430c79d7bc370516ca76213"} Oct 08 18:31:24 crc kubenswrapper[4988]: I1008 18:31:24.746667 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"b5a36680-3152-4d1f-a97d-92a3f1a3bc6e","Type":"ContainerStarted","Data":"8cfd7246d5f871d4b56104e40d243e23c05324b876da5e779fccc912e4df7472"} Oct 08 18:31:24 crc kubenswrapper[4988]: I1008 18:31:24.747911 4988 generic.go:334] "Generic (PLEG): container finished" podID="2e2a1e47-ca67-445c-b358-ad7e5bce08ea" containerID="8187466829ece57145a6f580c49fd29ced64f87be389192fb463c718a5acae13" exitCode=0 Oct 08 18:31:24 crc kubenswrapper[4988]: I1008 18:31:24.747948 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-tm2wd" event={"ID":"2e2a1e47-ca67-445c-b358-ad7e5bce08ea","Type":"ContainerDied","Data":"8187466829ece57145a6f580c49fd29ced64f87be389192fb463c718a5acae13"} Oct 08 18:31:24 crc kubenswrapper[4988]: I1008 18:31:24.788610 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.788583574 podStartE2EDuration="2.788583574s" podCreationTimestamp="2025-10-08 18:31:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 18:31:24.782474629 +0000 UTC m=+1230.232317409" watchObservedRunningTime="2025-10-08 18:31:24.788583574 +0000 UTC m=+1230.238426354" Oct 08 18:31:25 crc kubenswrapper[4988]: I1008 18:31:25.756449 4988 generic.go:334] "Generic (PLEG): container finished" podID="7aaaa954-1cc6-4304-8a21-86d9dfe4fd23" containerID="170776f5aecd85fd95ec5552397ec9b090c5fec990524b633f8d6b1d51a1c74e" exitCode=0 Oct 08 18:31:25 crc kubenswrapper[4988]: I1008 18:31:25.756546 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-rx7cw" event={"ID":"7aaaa954-1cc6-4304-8a21-86d9dfe4fd23","Type":"ContainerDied","Data":"170776f5aecd85fd95ec5552397ec9b090c5fec990524b633f8d6b1d51a1c74e"} Oct 08 18:31:25 crc kubenswrapper[4988]: I1008 18:31:25.945726 4988 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 08 18:31:25 crc kubenswrapper[4988]: I1008 18:31:25.945827 4988 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 08 18:31:26 crc kubenswrapper[4988]: I1008 18:31:26.127895 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-tm2wd" Oct 08 18:31:26 crc kubenswrapper[4988]: I1008 18:31:26.222139 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cs5rt\" (UniqueName: \"kubernetes.io/projected/2e2a1e47-ca67-445c-b358-ad7e5bce08ea-kube-api-access-cs5rt\") pod \"2e2a1e47-ca67-445c-b358-ad7e5bce08ea\" (UID: \"2e2a1e47-ca67-445c-b358-ad7e5bce08ea\") " Oct 08 18:31:26 crc kubenswrapper[4988]: I1008 18:31:26.222212 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2e2a1e47-ca67-445c-b358-ad7e5bce08ea-combined-ca-bundle\") pod \"2e2a1e47-ca67-445c-b358-ad7e5bce08ea\" (UID: \"2e2a1e47-ca67-445c-b358-ad7e5bce08ea\") " Oct 08 18:31:26 crc kubenswrapper[4988]: I1008 18:31:26.222260 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2e2a1e47-ca67-445c-b358-ad7e5bce08ea-config-data\") pod \"2e2a1e47-ca67-445c-b358-ad7e5bce08ea\" (UID: \"2e2a1e47-ca67-445c-b358-ad7e5bce08ea\") " Oct 08 18:31:26 crc kubenswrapper[4988]: I1008 18:31:26.222360 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2e2a1e47-ca67-445c-b358-ad7e5bce08ea-scripts\") pod \"2e2a1e47-ca67-445c-b358-ad7e5bce08ea\" (UID: \"2e2a1e47-ca67-445c-b358-ad7e5bce08ea\") " Oct 08 18:31:26 crc kubenswrapper[4988]: I1008 18:31:26.228735 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2e2a1e47-ca67-445c-b358-ad7e5bce08ea-kube-api-access-cs5rt" (OuterVolumeSpecName: "kube-api-access-cs5rt") pod "2e2a1e47-ca67-445c-b358-ad7e5bce08ea" (UID: "2e2a1e47-ca67-445c-b358-ad7e5bce08ea"). InnerVolumeSpecName "kube-api-access-cs5rt". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:31:26 crc kubenswrapper[4988]: I1008 18:31:26.242399 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2e2a1e47-ca67-445c-b358-ad7e5bce08ea-scripts" (OuterVolumeSpecName: "scripts") pod "2e2a1e47-ca67-445c-b358-ad7e5bce08ea" (UID: "2e2a1e47-ca67-445c-b358-ad7e5bce08ea"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:31:26 crc kubenswrapper[4988]: I1008 18:31:26.249977 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2e2a1e47-ca67-445c-b358-ad7e5bce08ea-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "2e2a1e47-ca67-445c-b358-ad7e5bce08ea" (UID: "2e2a1e47-ca67-445c-b358-ad7e5bce08ea"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:31:26 crc kubenswrapper[4988]: I1008 18:31:26.260890 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2e2a1e47-ca67-445c-b358-ad7e5bce08ea-config-data" (OuterVolumeSpecName: "config-data") pod "2e2a1e47-ca67-445c-b358-ad7e5bce08ea" (UID: "2e2a1e47-ca67-445c-b358-ad7e5bce08ea"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:31:26 crc kubenswrapper[4988]: I1008 18:31:26.325350 4988 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cs5rt\" (UniqueName: \"kubernetes.io/projected/2e2a1e47-ca67-445c-b358-ad7e5bce08ea-kube-api-access-cs5rt\") on node \"crc\" DevicePath \"\"" Oct 08 18:31:26 crc kubenswrapper[4988]: I1008 18:31:26.325568 4988 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2e2a1e47-ca67-445c-b358-ad7e5bce08ea-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 18:31:26 crc kubenswrapper[4988]: I1008 18:31:26.325643 4988 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2e2a1e47-ca67-445c-b358-ad7e5bce08ea-config-data\") on node \"crc\" DevicePath \"\"" Oct 08 18:31:26 crc kubenswrapper[4988]: I1008 18:31:26.325707 4988 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2e2a1e47-ca67-445c-b358-ad7e5bce08ea-scripts\") on node \"crc\" DevicePath \"\"" Oct 08 18:31:26 crc kubenswrapper[4988]: I1008 18:31:26.349955 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Oct 08 18:31:26 crc kubenswrapper[4988]: I1008 18:31:26.389519 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-64d8d96789-8bhgs" Oct 08 18:31:26 crc kubenswrapper[4988]: I1008 18:31:26.446587 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Oct 08 18:31:26 crc kubenswrapper[4988]: I1008 18:31:26.446813 4988 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Oct 08 18:31:26 crc kubenswrapper[4988]: I1008 18:31:26.468732 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-59b9656b65-wtfvw"] Oct 08 18:31:26 crc kubenswrapper[4988]: I1008 18:31:26.468984 4988 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-59b9656b65-wtfvw" podUID="f6134524-f05a-43be-b80d-a862f3e7b390" containerName="dnsmasq-dns" containerID="cri-o://d2cfa238fcf3427c4917a7f32e8de7b6a8c3479bf5cc18d79302827c17d4c2d4" gracePeriod=10 Oct 08 18:31:26 crc kubenswrapper[4988]: I1008 18:31:26.480644 4988 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Oct 08 18:31:26 crc kubenswrapper[4988]: I1008 18:31:26.791980 4988 generic.go:334] "Generic (PLEG): container finished" podID="f6134524-f05a-43be-b80d-a862f3e7b390" containerID="d2cfa238fcf3427c4917a7f32e8de7b6a8c3479bf5cc18d79302827c17d4c2d4" exitCode=0 Oct 08 18:31:26 crc kubenswrapper[4988]: I1008 18:31:26.792044 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-59b9656b65-wtfvw" event={"ID":"f6134524-f05a-43be-b80d-a862f3e7b390","Type":"ContainerDied","Data":"d2cfa238fcf3427c4917a7f32e8de7b6a8c3479bf5cc18d79302827c17d4c2d4"} Oct 08 18:31:26 crc kubenswrapper[4988]: I1008 18:31:26.794272 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-tm2wd" Oct 08 18:31:26 crc kubenswrapper[4988]: I1008 18:31:26.795625 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-tm2wd" event={"ID":"2e2a1e47-ca67-445c-b358-ad7e5bce08ea","Type":"ContainerDied","Data":"a9b44797feb4f30d8638636a4c736665a21d3738f85f1f855295676ef65fc84a"} Oct 08 18:31:26 crc kubenswrapper[4988]: I1008 18:31:26.795670 4988 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a9b44797feb4f30d8638636a4c736665a21d3738f85f1f855295676ef65fc84a" Oct 08 18:31:26 crc kubenswrapper[4988]: I1008 18:31:26.841080 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Oct 08 18:31:26 crc kubenswrapper[4988]: I1008 18:31:26.982346 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-59b9656b65-wtfvw" Oct 08 18:31:27 crc kubenswrapper[4988]: I1008 18:31:27.042850 4988 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="a116a7eb-0c4f-4d6b-93ef-93ac26f22441" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.0.185:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 08 18:31:27 crc kubenswrapper[4988]: I1008 18:31:27.043209 4988 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="a116a7eb-0c4f-4d6b-93ef-93ac26f22441" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.0.185:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 08 18:31:27 crc kubenswrapper[4988]: I1008 18:31:27.044492 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f6134524-f05a-43be-b80d-a862f3e7b390-ovsdbserver-nb\") pod \"f6134524-f05a-43be-b80d-a862f3e7b390\" (UID: \"f6134524-f05a-43be-b80d-a862f3e7b390\") " Oct 08 18:31:27 crc kubenswrapper[4988]: I1008 18:31:27.070168 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f6134524-f05a-43be-b80d-a862f3e7b390-ovsdbserver-sb\") pod \"f6134524-f05a-43be-b80d-a862f3e7b390\" (UID: \"f6134524-f05a-43be-b80d-a862f3e7b390\") " Oct 08 18:31:27 crc kubenswrapper[4988]: I1008 18:31:27.070248 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/f6134524-f05a-43be-b80d-a862f3e7b390-dns-swift-storage-0\") pod \"f6134524-f05a-43be-b80d-a862f3e7b390\" (UID: \"f6134524-f05a-43be-b80d-a862f3e7b390\") " Oct 08 18:31:27 crc kubenswrapper[4988]: I1008 18:31:27.070328 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jpv87\" (UniqueName: \"kubernetes.io/projected/f6134524-f05a-43be-b80d-a862f3e7b390-kube-api-access-jpv87\") pod \"f6134524-f05a-43be-b80d-a862f3e7b390\" (UID: \"f6134524-f05a-43be-b80d-a862f3e7b390\") " Oct 08 18:31:27 crc kubenswrapper[4988]: I1008 18:31:27.070414 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f6134524-f05a-43be-b80d-a862f3e7b390-dns-svc\") pod \"f6134524-f05a-43be-b80d-a862f3e7b390\" (UID: \"f6134524-f05a-43be-b80d-a862f3e7b390\") " Oct 08 18:31:27 crc kubenswrapper[4988]: I1008 18:31:27.070540 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f6134524-f05a-43be-b80d-a862f3e7b390-config\") pod \"f6134524-f05a-43be-b80d-a862f3e7b390\" (UID: \"f6134524-f05a-43be-b80d-a862f3e7b390\") " Oct 08 18:31:27 crc kubenswrapper[4988]: I1008 18:31:27.074283 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f6134524-f05a-43be-b80d-a862f3e7b390-kube-api-access-jpv87" (OuterVolumeSpecName: "kube-api-access-jpv87") pod "f6134524-f05a-43be-b80d-a862f3e7b390" (UID: "f6134524-f05a-43be-b80d-a862f3e7b390"). InnerVolumeSpecName "kube-api-access-jpv87". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:31:27 crc kubenswrapper[4988]: I1008 18:31:27.136318 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 08 18:31:27 crc kubenswrapper[4988]: I1008 18:31:27.136620 4988 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="a116a7eb-0c4f-4d6b-93ef-93ac26f22441" containerName="nova-api-log" containerID="cri-o://8b312e4e6552af848a89843736fde9707853545f97ff16dd7eb3c47b05d0c32d" gracePeriod=30 Oct 08 18:31:27 crc kubenswrapper[4988]: I1008 18:31:27.137666 4988 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="a116a7eb-0c4f-4d6b-93ef-93ac26f22441" containerName="nova-api-api" containerID="cri-o://f1e67556c910c6012c3493e80adf12b19a788da6c247b9ef602743fd99cb3b7a" gracePeriod=30 Oct 08 18:31:27 crc kubenswrapper[4988]: I1008 18:31:27.163071 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f6134524-f05a-43be-b80d-a862f3e7b390-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "f6134524-f05a-43be-b80d-a862f3e7b390" (UID: "f6134524-f05a-43be-b80d-a862f3e7b390"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 18:31:27 crc kubenswrapper[4988]: I1008 18:31:27.165760 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Oct 08 18:31:27 crc kubenswrapper[4988]: I1008 18:31:27.165983 4988 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="b5a36680-3152-4d1f-a97d-92a3f1a3bc6e" containerName="nova-metadata-log" containerID="cri-o://8cfd7246d5f871d4b56104e40d243e23c05324b876da5e779fccc912e4df7472" gracePeriod=30 Oct 08 18:31:27 crc kubenswrapper[4988]: I1008 18:31:27.166367 4988 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="b5a36680-3152-4d1f-a97d-92a3f1a3bc6e" containerName="nova-metadata-metadata" containerID="cri-o://b92e21969a50a5f6d64686041134b3dbd486d24db430c79d7bc370516ca76213" gracePeriod=30 Oct 08 18:31:27 crc kubenswrapper[4988]: I1008 18:31:27.178589 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f6134524-f05a-43be-b80d-a862f3e7b390-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "f6134524-f05a-43be-b80d-a862f3e7b390" (UID: "f6134524-f05a-43be-b80d-a862f3e7b390"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 18:31:27 crc kubenswrapper[4988]: I1008 18:31:27.195210 4988 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f6134524-f05a-43be-b80d-a862f3e7b390-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 08 18:31:27 crc kubenswrapper[4988]: I1008 18:31:27.196467 4988 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/f6134524-f05a-43be-b80d-a862f3e7b390-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Oct 08 18:31:27 crc kubenswrapper[4988]: I1008 18:31:27.196548 4988 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jpv87\" (UniqueName: \"kubernetes.io/projected/f6134524-f05a-43be-b80d-a862f3e7b390-kube-api-access-jpv87\") on node \"crc\" DevicePath \"\"" Oct 08 18:31:27 crc kubenswrapper[4988]: I1008 18:31:27.200845 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f6134524-f05a-43be-b80d-a862f3e7b390-config" (OuterVolumeSpecName: "config") pod "f6134524-f05a-43be-b80d-a862f3e7b390" (UID: "f6134524-f05a-43be-b80d-a862f3e7b390"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 18:31:27 crc kubenswrapper[4988]: I1008 18:31:27.220006 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f6134524-f05a-43be-b80d-a862f3e7b390-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "f6134524-f05a-43be-b80d-a862f3e7b390" (UID: "f6134524-f05a-43be-b80d-a862f3e7b390"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 18:31:27 crc kubenswrapper[4988]: I1008 18:31:27.233371 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f6134524-f05a-43be-b80d-a862f3e7b390-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "f6134524-f05a-43be-b80d-a862f3e7b390" (UID: "f6134524-f05a-43be-b80d-a862f3e7b390"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 18:31:27 crc kubenswrapper[4988]: I1008 18:31:27.298786 4988 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f6134524-f05a-43be-b80d-a862f3e7b390-config\") on node \"crc\" DevicePath \"\"" Oct 08 18:31:27 crc kubenswrapper[4988]: I1008 18:31:27.298828 4988 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f6134524-f05a-43be-b80d-a862f3e7b390-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 08 18:31:27 crc kubenswrapper[4988]: I1008 18:31:27.298838 4988 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f6134524-f05a-43be-b80d-a862f3e7b390-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 08 18:31:27 crc kubenswrapper[4988]: I1008 18:31:27.314074 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-rx7cw" Oct 08 18:31:27 crc kubenswrapper[4988]: I1008 18:31:27.319157 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Oct 08 18:31:27 crc kubenswrapper[4988]: I1008 18:31:27.502614 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7aaaa954-1cc6-4304-8a21-86d9dfe4fd23-combined-ca-bundle\") pod \"7aaaa954-1cc6-4304-8a21-86d9dfe4fd23\" (UID: \"7aaaa954-1cc6-4304-8a21-86d9dfe4fd23\") " Oct 08 18:31:27 crc kubenswrapper[4988]: I1008 18:31:27.502687 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7aaaa954-1cc6-4304-8a21-86d9dfe4fd23-config-data\") pod \"7aaaa954-1cc6-4304-8a21-86d9dfe4fd23\" (UID: \"7aaaa954-1cc6-4304-8a21-86d9dfe4fd23\") " Oct 08 18:31:27 crc kubenswrapper[4988]: I1008 18:31:27.502761 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7aaaa954-1cc6-4304-8a21-86d9dfe4fd23-scripts\") pod \"7aaaa954-1cc6-4304-8a21-86d9dfe4fd23\" (UID: \"7aaaa954-1cc6-4304-8a21-86d9dfe4fd23\") " Oct 08 18:31:27 crc kubenswrapper[4988]: I1008 18:31:27.503005 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-f4q8v\" (UniqueName: \"kubernetes.io/projected/7aaaa954-1cc6-4304-8a21-86d9dfe4fd23-kube-api-access-f4q8v\") pod \"7aaaa954-1cc6-4304-8a21-86d9dfe4fd23\" (UID: \"7aaaa954-1cc6-4304-8a21-86d9dfe4fd23\") " Oct 08 18:31:27 crc kubenswrapper[4988]: I1008 18:31:27.507326 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7aaaa954-1cc6-4304-8a21-86d9dfe4fd23-kube-api-access-f4q8v" (OuterVolumeSpecName: "kube-api-access-f4q8v") pod "7aaaa954-1cc6-4304-8a21-86d9dfe4fd23" (UID: "7aaaa954-1cc6-4304-8a21-86d9dfe4fd23"). InnerVolumeSpecName "kube-api-access-f4q8v". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:31:27 crc kubenswrapper[4988]: I1008 18:31:27.507623 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7aaaa954-1cc6-4304-8a21-86d9dfe4fd23-scripts" (OuterVolumeSpecName: "scripts") pod "7aaaa954-1cc6-4304-8a21-86d9dfe4fd23" (UID: "7aaaa954-1cc6-4304-8a21-86d9dfe4fd23"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:31:27 crc kubenswrapper[4988]: I1008 18:31:27.538241 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7aaaa954-1cc6-4304-8a21-86d9dfe4fd23-config-data" (OuterVolumeSpecName: "config-data") pod "7aaaa954-1cc6-4304-8a21-86d9dfe4fd23" (UID: "7aaaa954-1cc6-4304-8a21-86d9dfe4fd23"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:31:27 crc kubenswrapper[4988]: I1008 18:31:27.544840 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7aaaa954-1cc6-4304-8a21-86d9dfe4fd23-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "7aaaa954-1cc6-4304-8a21-86d9dfe4fd23" (UID: "7aaaa954-1cc6-4304-8a21-86d9dfe4fd23"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:31:27 crc kubenswrapper[4988]: I1008 18:31:27.608376 4988 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-f4q8v\" (UniqueName: \"kubernetes.io/projected/7aaaa954-1cc6-4304-8a21-86d9dfe4fd23-kube-api-access-f4q8v\") on node \"crc\" DevicePath \"\"" Oct 08 18:31:27 crc kubenswrapper[4988]: I1008 18:31:27.608714 4988 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7aaaa954-1cc6-4304-8a21-86d9dfe4fd23-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 18:31:27 crc kubenswrapper[4988]: I1008 18:31:27.608729 4988 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7aaaa954-1cc6-4304-8a21-86d9dfe4fd23-config-data\") on node \"crc\" DevicePath \"\"" Oct 08 18:31:27 crc kubenswrapper[4988]: I1008 18:31:27.608741 4988 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7aaaa954-1cc6-4304-8a21-86d9dfe4fd23-scripts\") on node \"crc\" DevicePath \"\"" Oct 08 18:31:27 crc kubenswrapper[4988]: I1008 18:31:27.684426 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 08 18:31:27 crc kubenswrapper[4988]: I1008 18:31:27.709799 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xfv8k\" (UniqueName: \"kubernetes.io/projected/b5a36680-3152-4d1f-a97d-92a3f1a3bc6e-kube-api-access-xfv8k\") pod \"b5a36680-3152-4d1f-a97d-92a3f1a3bc6e\" (UID: \"b5a36680-3152-4d1f-a97d-92a3f1a3bc6e\") " Oct 08 18:31:27 crc kubenswrapper[4988]: I1008 18:31:27.709892 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b5a36680-3152-4d1f-a97d-92a3f1a3bc6e-config-data\") pod \"b5a36680-3152-4d1f-a97d-92a3f1a3bc6e\" (UID: \"b5a36680-3152-4d1f-a97d-92a3f1a3bc6e\") " Oct 08 18:31:27 crc kubenswrapper[4988]: I1008 18:31:27.710022 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/b5a36680-3152-4d1f-a97d-92a3f1a3bc6e-nova-metadata-tls-certs\") pod \"b5a36680-3152-4d1f-a97d-92a3f1a3bc6e\" (UID: \"b5a36680-3152-4d1f-a97d-92a3f1a3bc6e\") " Oct 08 18:31:27 crc kubenswrapper[4988]: I1008 18:31:27.710068 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b5a36680-3152-4d1f-a97d-92a3f1a3bc6e-combined-ca-bundle\") pod \"b5a36680-3152-4d1f-a97d-92a3f1a3bc6e\" (UID: \"b5a36680-3152-4d1f-a97d-92a3f1a3bc6e\") " Oct 08 18:31:27 crc kubenswrapper[4988]: I1008 18:31:27.710102 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b5a36680-3152-4d1f-a97d-92a3f1a3bc6e-logs\") pod \"b5a36680-3152-4d1f-a97d-92a3f1a3bc6e\" (UID: \"b5a36680-3152-4d1f-a97d-92a3f1a3bc6e\") " Oct 08 18:31:27 crc kubenswrapper[4988]: I1008 18:31:27.710622 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b5a36680-3152-4d1f-a97d-92a3f1a3bc6e-logs" (OuterVolumeSpecName: "logs") pod "b5a36680-3152-4d1f-a97d-92a3f1a3bc6e" (UID: "b5a36680-3152-4d1f-a97d-92a3f1a3bc6e"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 18:31:27 crc kubenswrapper[4988]: I1008 18:31:27.715837 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b5a36680-3152-4d1f-a97d-92a3f1a3bc6e-kube-api-access-xfv8k" (OuterVolumeSpecName: "kube-api-access-xfv8k") pod "b5a36680-3152-4d1f-a97d-92a3f1a3bc6e" (UID: "b5a36680-3152-4d1f-a97d-92a3f1a3bc6e"). InnerVolumeSpecName "kube-api-access-xfv8k". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:31:27 crc kubenswrapper[4988]: I1008 18:31:27.747058 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b5a36680-3152-4d1f-a97d-92a3f1a3bc6e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "b5a36680-3152-4d1f-a97d-92a3f1a3bc6e" (UID: "b5a36680-3152-4d1f-a97d-92a3f1a3bc6e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:31:27 crc kubenswrapper[4988]: I1008 18:31:27.747777 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b5a36680-3152-4d1f-a97d-92a3f1a3bc6e-config-data" (OuterVolumeSpecName: "config-data") pod "b5a36680-3152-4d1f-a97d-92a3f1a3bc6e" (UID: "b5a36680-3152-4d1f-a97d-92a3f1a3bc6e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:31:27 crc kubenswrapper[4988]: I1008 18:31:27.769667 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b5a36680-3152-4d1f-a97d-92a3f1a3bc6e-nova-metadata-tls-certs" (OuterVolumeSpecName: "nova-metadata-tls-certs") pod "b5a36680-3152-4d1f-a97d-92a3f1a3bc6e" (UID: "b5a36680-3152-4d1f-a97d-92a3f1a3bc6e"). InnerVolumeSpecName "nova-metadata-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:31:27 crc kubenswrapper[4988]: I1008 18:31:27.811673 4988 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b5a36680-3152-4d1f-a97d-92a3f1a3bc6e-logs\") on node \"crc\" DevicePath \"\"" Oct 08 18:31:27 crc kubenswrapper[4988]: I1008 18:31:27.811703 4988 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xfv8k\" (UniqueName: \"kubernetes.io/projected/b5a36680-3152-4d1f-a97d-92a3f1a3bc6e-kube-api-access-xfv8k\") on node \"crc\" DevicePath \"\"" Oct 08 18:31:27 crc kubenswrapper[4988]: I1008 18:31:27.811712 4988 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b5a36680-3152-4d1f-a97d-92a3f1a3bc6e-config-data\") on node \"crc\" DevicePath \"\"" Oct 08 18:31:27 crc kubenswrapper[4988]: I1008 18:31:27.811722 4988 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/b5a36680-3152-4d1f-a97d-92a3f1a3bc6e-nova-metadata-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 08 18:31:27 crc kubenswrapper[4988]: I1008 18:31:27.811730 4988 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b5a36680-3152-4d1f-a97d-92a3f1a3bc6e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 18:31:27 crc kubenswrapper[4988]: I1008 18:31:27.813034 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-rx7cw" Oct 08 18:31:27 crc kubenswrapper[4988]: I1008 18:31:27.813203 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-rx7cw" event={"ID":"7aaaa954-1cc6-4304-8a21-86d9dfe4fd23","Type":"ContainerDied","Data":"61929a33e24fafc1b48794f1d99db9657b578e9d2af7db1a64e60d1547f47bfd"} Oct 08 18:31:27 crc kubenswrapper[4988]: I1008 18:31:27.813229 4988 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="61929a33e24fafc1b48794f1d99db9657b578e9d2af7db1a64e60d1547f47bfd" Oct 08 18:31:27 crc kubenswrapper[4988]: I1008 18:31:27.822477 4988 generic.go:334] "Generic (PLEG): container finished" podID="b5a36680-3152-4d1f-a97d-92a3f1a3bc6e" containerID="b92e21969a50a5f6d64686041134b3dbd486d24db430c79d7bc370516ca76213" exitCode=0 Oct 08 18:31:27 crc kubenswrapper[4988]: I1008 18:31:27.822725 4988 generic.go:334] "Generic (PLEG): container finished" podID="b5a36680-3152-4d1f-a97d-92a3f1a3bc6e" containerID="8cfd7246d5f871d4b56104e40d243e23c05324b876da5e779fccc912e4df7472" exitCode=143 Oct 08 18:31:27 crc kubenswrapper[4988]: I1008 18:31:27.822525 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"b5a36680-3152-4d1f-a97d-92a3f1a3bc6e","Type":"ContainerDied","Data":"b92e21969a50a5f6d64686041134b3dbd486d24db430c79d7bc370516ca76213"} Oct 08 18:31:27 crc kubenswrapper[4988]: I1008 18:31:27.822870 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"b5a36680-3152-4d1f-a97d-92a3f1a3bc6e","Type":"ContainerDied","Data":"8cfd7246d5f871d4b56104e40d243e23c05324b876da5e779fccc912e4df7472"} Oct 08 18:31:27 crc kubenswrapper[4988]: I1008 18:31:27.822904 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"b5a36680-3152-4d1f-a97d-92a3f1a3bc6e","Type":"ContainerDied","Data":"36195b6ddcc1c5790944a7a2f25a56451d2550ab56480548b12800f80aa8d62a"} Oct 08 18:31:27 crc kubenswrapper[4988]: I1008 18:31:27.822925 4988 scope.go:117] "RemoveContainer" containerID="b92e21969a50a5f6d64686041134b3dbd486d24db430c79d7bc370516ca76213" Oct 08 18:31:27 crc kubenswrapper[4988]: I1008 18:31:27.822553 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 08 18:31:27 crc kubenswrapper[4988]: I1008 18:31:27.835488 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-59b9656b65-wtfvw" event={"ID":"f6134524-f05a-43be-b80d-a862f3e7b390","Type":"ContainerDied","Data":"0711b6e67a4760649ffaf998e791bdad1d1c4995cfb81ece8dea0d05efb3d97e"} Oct 08 18:31:27 crc kubenswrapper[4988]: I1008 18:31:27.835617 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-59b9656b65-wtfvw" Oct 08 18:31:27 crc kubenswrapper[4988]: I1008 18:31:27.843217 4988 generic.go:334] "Generic (PLEG): container finished" podID="a116a7eb-0c4f-4d6b-93ef-93ac26f22441" containerID="8b312e4e6552af848a89843736fde9707853545f97ff16dd7eb3c47b05d0c32d" exitCode=143 Oct 08 18:31:27 crc kubenswrapper[4988]: I1008 18:31:27.843327 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"a116a7eb-0c4f-4d6b-93ef-93ac26f22441","Type":"ContainerDied","Data":"8b312e4e6552af848a89843736fde9707853545f97ff16dd7eb3c47b05d0c32d"} Oct 08 18:31:27 crc kubenswrapper[4988]: I1008 18:31:27.864493 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-0"] Oct 08 18:31:27 crc kubenswrapper[4988]: E1008 18:31:27.865145 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2e2a1e47-ca67-445c-b358-ad7e5bce08ea" containerName="nova-manage" Oct 08 18:31:27 crc kubenswrapper[4988]: I1008 18:31:27.865227 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="2e2a1e47-ca67-445c-b358-ad7e5bce08ea" containerName="nova-manage" Oct 08 18:31:27 crc kubenswrapper[4988]: E1008 18:31:27.865293 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b5a36680-3152-4d1f-a97d-92a3f1a3bc6e" containerName="nova-metadata-log" Oct 08 18:31:27 crc kubenswrapper[4988]: I1008 18:31:27.865374 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="b5a36680-3152-4d1f-a97d-92a3f1a3bc6e" containerName="nova-metadata-log" Oct 08 18:31:27 crc kubenswrapper[4988]: E1008 18:31:27.865462 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f6134524-f05a-43be-b80d-a862f3e7b390" containerName="init" Oct 08 18:31:27 crc kubenswrapper[4988]: I1008 18:31:27.865531 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="f6134524-f05a-43be-b80d-a862f3e7b390" containerName="init" Oct 08 18:31:27 crc kubenswrapper[4988]: E1008 18:31:27.865598 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7aaaa954-1cc6-4304-8a21-86d9dfe4fd23" containerName="nova-cell1-conductor-db-sync" Oct 08 18:31:27 crc kubenswrapper[4988]: I1008 18:31:27.865652 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="7aaaa954-1cc6-4304-8a21-86d9dfe4fd23" containerName="nova-cell1-conductor-db-sync" Oct 08 18:31:27 crc kubenswrapper[4988]: E1008 18:31:27.865722 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b5a36680-3152-4d1f-a97d-92a3f1a3bc6e" containerName="nova-metadata-metadata" Oct 08 18:31:27 crc kubenswrapper[4988]: I1008 18:31:27.865772 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="b5a36680-3152-4d1f-a97d-92a3f1a3bc6e" containerName="nova-metadata-metadata" Oct 08 18:31:27 crc kubenswrapper[4988]: E1008 18:31:27.865826 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f6134524-f05a-43be-b80d-a862f3e7b390" containerName="dnsmasq-dns" Oct 08 18:31:27 crc kubenswrapper[4988]: I1008 18:31:27.865888 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="f6134524-f05a-43be-b80d-a862f3e7b390" containerName="dnsmasq-dns" Oct 08 18:31:27 crc kubenswrapper[4988]: I1008 18:31:27.866116 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="b5a36680-3152-4d1f-a97d-92a3f1a3bc6e" containerName="nova-metadata-log" Oct 08 18:31:27 crc kubenswrapper[4988]: I1008 18:31:27.866927 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="b5a36680-3152-4d1f-a97d-92a3f1a3bc6e" containerName="nova-metadata-metadata" Oct 08 18:31:27 crc kubenswrapper[4988]: I1008 18:31:27.867012 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="2e2a1e47-ca67-445c-b358-ad7e5bce08ea" containerName="nova-manage" Oct 08 18:31:27 crc kubenswrapper[4988]: I1008 18:31:27.867069 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="7aaaa954-1cc6-4304-8a21-86d9dfe4fd23" containerName="nova-cell1-conductor-db-sync" Oct 08 18:31:27 crc kubenswrapper[4988]: I1008 18:31:27.867132 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="f6134524-f05a-43be-b80d-a862f3e7b390" containerName="dnsmasq-dns" Oct 08 18:31:27 crc kubenswrapper[4988]: I1008 18:31:27.867824 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Oct 08 18:31:27 crc kubenswrapper[4988]: I1008 18:31:27.870378 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Oct 08 18:31:27 crc kubenswrapper[4988]: I1008 18:31:27.870436 4988 scope.go:117] "RemoveContainer" containerID="8cfd7246d5f871d4b56104e40d243e23c05324b876da5e779fccc912e4df7472" Oct 08 18:31:27 crc kubenswrapper[4988]: I1008 18:31:27.872285 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Oct 08 18:31:27 crc kubenswrapper[4988]: I1008 18:31:27.901466 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-59b9656b65-wtfvw"] Oct 08 18:31:27 crc kubenswrapper[4988]: I1008 18:31:27.925527 4988 scope.go:117] "RemoveContainer" containerID="b92e21969a50a5f6d64686041134b3dbd486d24db430c79d7bc370516ca76213" Oct 08 18:31:27 crc kubenswrapper[4988]: I1008 18:31:27.925664 4988 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-59b9656b65-wtfvw"] Oct 08 18:31:27 crc kubenswrapper[4988]: E1008 18:31:27.928988 4988 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b92e21969a50a5f6d64686041134b3dbd486d24db430c79d7bc370516ca76213\": container with ID starting with b92e21969a50a5f6d64686041134b3dbd486d24db430c79d7bc370516ca76213 not found: ID does not exist" containerID="b92e21969a50a5f6d64686041134b3dbd486d24db430c79d7bc370516ca76213" Oct 08 18:31:27 crc kubenswrapper[4988]: I1008 18:31:27.929025 4988 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b92e21969a50a5f6d64686041134b3dbd486d24db430c79d7bc370516ca76213"} err="failed to get container status \"b92e21969a50a5f6d64686041134b3dbd486d24db430c79d7bc370516ca76213\": rpc error: code = NotFound desc = could not find container \"b92e21969a50a5f6d64686041134b3dbd486d24db430c79d7bc370516ca76213\": container with ID starting with b92e21969a50a5f6d64686041134b3dbd486d24db430c79d7bc370516ca76213 not found: ID does not exist" Oct 08 18:31:27 crc kubenswrapper[4988]: I1008 18:31:27.929056 4988 scope.go:117] "RemoveContainer" containerID="8cfd7246d5f871d4b56104e40d243e23c05324b876da5e779fccc912e4df7472" Oct 08 18:31:27 crc kubenswrapper[4988]: E1008 18:31:27.932474 4988 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8cfd7246d5f871d4b56104e40d243e23c05324b876da5e779fccc912e4df7472\": container with ID starting with 8cfd7246d5f871d4b56104e40d243e23c05324b876da5e779fccc912e4df7472 not found: ID does not exist" containerID="8cfd7246d5f871d4b56104e40d243e23c05324b876da5e779fccc912e4df7472" Oct 08 18:31:27 crc kubenswrapper[4988]: I1008 18:31:27.932511 4988 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8cfd7246d5f871d4b56104e40d243e23c05324b876da5e779fccc912e4df7472"} err="failed to get container status \"8cfd7246d5f871d4b56104e40d243e23c05324b876da5e779fccc912e4df7472\": rpc error: code = NotFound desc = could not find container \"8cfd7246d5f871d4b56104e40d243e23c05324b876da5e779fccc912e4df7472\": container with ID starting with 8cfd7246d5f871d4b56104e40d243e23c05324b876da5e779fccc912e4df7472 not found: ID does not exist" Oct 08 18:31:27 crc kubenswrapper[4988]: I1008 18:31:27.932531 4988 scope.go:117] "RemoveContainer" containerID="b92e21969a50a5f6d64686041134b3dbd486d24db430c79d7bc370516ca76213" Oct 08 18:31:27 crc kubenswrapper[4988]: I1008 18:31:27.936469 4988 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b92e21969a50a5f6d64686041134b3dbd486d24db430c79d7bc370516ca76213"} err="failed to get container status \"b92e21969a50a5f6d64686041134b3dbd486d24db430c79d7bc370516ca76213\": rpc error: code = NotFound desc = could not find container \"b92e21969a50a5f6d64686041134b3dbd486d24db430c79d7bc370516ca76213\": container with ID starting with b92e21969a50a5f6d64686041134b3dbd486d24db430c79d7bc370516ca76213 not found: ID does not exist" Oct 08 18:31:27 crc kubenswrapper[4988]: I1008 18:31:27.936503 4988 scope.go:117] "RemoveContainer" containerID="8cfd7246d5f871d4b56104e40d243e23c05324b876da5e779fccc912e4df7472" Oct 08 18:31:27 crc kubenswrapper[4988]: I1008 18:31:27.940460 4988 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8cfd7246d5f871d4b56104e40d243e23c05324b876da5e779fccc912e4df7472"} err="failed to get container status \"8cfd7246d5f871d4b56104e40d243e23c05324b876da5e779fccc912e4df7472\": rpc error: code = NotFound desc = could not find container \"8cfd7246d5f871d4b56104e40d243e23c05324b876da5e779fccc912e4df7472\": container with ID starting with 8cfd7246d5f871d4b56104e40d243e23c05324b876da5e779fccc912e4df7472 not found: ID does not exist" Oct 08 18:31:27 crc kubenswrapper[4988]: I1008 18:31:27.940493 4988 scope.go:117] "RemoveContainer" containerID="d2cfa238fcf3427c4917a7f32e8de7b6a8c3479bf5cc18d79302827c17d4c2d4" Oct 08 18:31:27 crc kubenswrapper[4988]: I1008 18:31:27.955169 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Oct 08 18:31:27 crc kubenswrapper[4988]: I1008 18:31:27.960962 4988 scope.go:117] "RemoveContainer" containerID="ede0a83a2428dcb8b487d4ec98b6590b8bc67faf808232a9a9cecb4b352d76b6" Oct 08 18:31:27 crc kubenswrapper[4988]: I1008 18:31:27.968637 4988 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Oct 08 18:31:27 crc kubenswrapper[4988]: I1008 18:31:27.984226 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Oct 08 18:31:27 crc kubenswrapper[4988]: I1008 18:31:27.985998 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 08 18:31:27 crc kubenswrapper[4988]: I1008 18:31:27.989471 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Oct 08 18:31:27 crc kubenswrapper[4988]: I1008 18:31:27.989722 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Oct 08 18:31:27 crc kubenswrapper[4988]: I1008 18:31:27.991401 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 08 18:31:28 crc kubenswrapper[4988]: I1008 18:31:28.017085 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zrmcr\" (UniqueName: \"kubernetes.io/projected/85bc7f8f-fdb3-4abb-88d3-ed0c10982c4c-kube-api-access-zrmcr\") pod \"nova-cell1-conductor-0\" (UID: \"85bc7f8f-fdb3-4abb-88d3-ed0c10982c4c\") " pod="openstack/nova-cell1-conductor-0" Oct 08 18:31:28 crc kubenswrapper[4988]: I1008 18:31:28.017164 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/85bc7f8f-fdb3-4abb-88d3-ed0c10982c4c-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"85bc7f8f-fdb3-4abb-88d3-ed0c10982c4c\") " pod="openstack/nova-cell1-conductor-0" Oct 08 18:31:28 crc kubenswrapper[4988]: I1008 18:31:28.017192 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/85bc7f8f-fdb3-4abb-88d3-ed0c10982c4c-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"85bc7f8f-fdb3-4abb-88d3-ed0c10982c4c\") " pod="openstack/nova-cell1-conductor-0" Oct 08 18:31:28 crc kubenswrapper[4988]: I1008 18:31:28.119219 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zk494\" (UniqueName: \"kubernetes.io/projected/c51f2710-19e3-457c-ab1b-6c7e19877213-kube-api-access-zk494\") pod \"nova-metadata-0\" (UID: \"c51f2710-19e3-457c-ab1b-6c7e19877213\") " pod="openstack/nova-metadata-0" Oct 08 18:31:28 crc kubenswrapper[4988]: I1008 18:31:28.119267 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c51f2710-19e3-457c-ab1b-6c7e19877213-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"c51f2710-19e3-457c-ab1b-6c7e19877213\") " pod="openstack/nova-metadata-0" Oct 08 18:31:28 crc kubenswrapper[4988]: I1008 18:31:28.119341 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zrmcr\" (UniqueName: \"kubernetes.io/projected/85bc7f8f-fdb3-4abb-88d3-ed0c10982c4c-kube-api-access-zrmcr\") pod \"nova-cell1-conductor-0\" (UID: \"85bc7f8f-fdb3-4abb-88d3-ed0c10982c4c\") " pod="openstack/nova-cell1-conductor-0" Oct 08 18:31:28 crc kubenswrapper[4988]: I1008 18:31:28.119399 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c51f2710-19e3-457c-ab1b-6c7e19877213-logs\") pod \"nova-metadata-0\" (UID: \"c51f2710-19e3-457c-ab1b-6c7e19877213\") " pod="openstack/nova-metadata-0" Oct 08 18:31:28 crc kubenswrapper[4988]: I1008 18:31:28.119432 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/c51f2710-19e3-457c-ab1b-6c7e19877213-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"c51f2710-19e3-457c-ab1b-6c7e19877213\") " pod="openstack/nova-metadata-0" Oct 08 18:31:28 crc kubenswrapper[4988]: I1008 18:31:28.119461 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/85bc7f8f-fdb3-4abb-88d3-ed0c10982c4c-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"85bc7f8f-fdb3-4abb-88d3-ed0c10982c4c\") " pod="openstack/nova-cell1-conductor-0" Oct 08 18:31:28 crc kubenswrapper[4988]: I1008 18:31:28.119486 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/85bc7f8f-fdb3-4abb-88d3-ed0c10982c4c-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"85bc7f8f-fdb3-4abb-88d3-ed0c10982c4c\") " pod="openstack/nova-cell1-conductor-0" Oct 08 18:31:28 crc kubenswrapper[4988]: I1008 18:31:28.119522 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c51f2710-19e3-457c-ab1b-6c7e19877213-config-data\") pod \"nova-metadata-0\" (UID: \"c51f2710-19e3-457c-ab1b-6c7e19877213\") " pod="openstack/nova-metadata-0" Oct 08 18:31:28 crc kubenswrapper[4988]: I1008 18:31:28.123115 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/85bc7f8f-fdb3-4abb-88d3-ed0c10982c4c-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"85bc7f8f-fdb3-4abb-88d3-ed0c10982c4c\") " pod="openstack/nova-cell1-conductor-0" Oct 08 18:31:28 crc kubenswrapper[4988]: I1008 18:31:28.123714 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/85bc7f8f-fdb3-4abb-88d3-ed0c10982c4c-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"85bc7f8f-fdb3-4abb-88d3-ed0c10982c4c\") " pod="openstack/nova-cell1-conductor-0" Oct 08 18:31:28 crc kubenswrapper[4988]: I1008 18:31:28.136107 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zrmcr\" (UniqueName: \"kubernetes.io/projected/85bc7f8f-fdb3-4abb-88d3-ed0c10982c4c-kube-api-access-zrmcr\") pod \"nova-cell1-conductor-0\" (UID: \"85bc7f8f-fdb3-4abb-88d3-ed0c10982c4c\") " pod="openstack/nova-cell1-conductor-0" Oct 08 18:31:28 crc kubenswrapper[4988]: I1008 18:31:28.196090 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Oct 08 18:31:28 crc kubenswrapper[4988]: I1008 18:31:28.227511 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c51f2710-19e3-457c-ab1b-6c7e19877213-logs\") pod \"nova-metadata-0\" (UID: \"c51f2710-19e3-457c-ab1b-6c7e19877213\") " pod="openstack/nova-metadata-0" Oct 08 18:31:28 crc kubenswrapper[4988]: I1008 18:31:28.227570 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/c51f2710-19e3-457c-ab1b-6c7e19877213-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"c51f2710-19e3-457c-ab1b-6c7e19877213\") " pod="openstack/nova-metadata-0" Oct 08 18:31:28 crc kubenswrapper[4988]: I1008 18:31:28.227611 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c51f2710-19e3-457c-ab1b-6c7e19877213-config-data\") pod \"nova-metadata-0\" (UID: \"c51f2710-19e3-457c-ab1b-6c7e19877213\") " pod="openstack/nova-metadata-0" Oct 08 18:31:28 crc kubenswrapper[4988]: I1008 18:31:28.227671 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zk494\" (UniqueName: \"kubernetes.io/projected/c51f2710-19e3-457c-ab1b-6c7e19877213-kube-api-access-zk494\") pod \"nova-metadata-0\" (UID: \"c51f2710-19e3-457c-ab1b-6c7e19877213\") " pod="openstack/nova-metadata-0" Oct 08 18:31:28 crc kubenswrapper[4988]: I1008 18:31:28.227693 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c51f2710-19e3-457c-ab1b-6c7e19877213-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"c51f2710-19e3-457c-ab1b-6c7e19877213\") " pod="openstack/nova-metadata-0" Oct 08 18:31:28 crc kubenswrapper[4988]: I1008 18:31:28.231830 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c51f2710-19e3-457c-ab1b-6c7e19877213-logs\") pod \"nova-metadata-0\" (UID: \"c51f2710-19e3-457c-ab1b-6c7e19877213\") " pod="openstack/nova-metadata-0" Oct 08 18:31:28 crc kubenswrapper[4988]: I1008 18:31:28.235419 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/c51f2710-19e3-457c-ab1b-6c7e19877213-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"c51f2710-19e3-457c-ab1b-6c7e19877213\") " pod="openstack/nova-metadata-0" Oct 08 18:31:28 crc kubenswrapper[4988]: I1008 18:31:28.243954 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c51f2710-19e3-457c-ab1b-6c7e19877213-config-data\") pod \"nova-metadata-0\" (UID: \"c51f2710-19e3-457c-ab1b-6c7e19877213\") " pod="openstack/nova-metadata-0" Oct 08 18:31:28 crc kubenswrapper[4988]: I1008 18:31:28.245353 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c51f2710-19e3-457c-ab1b-6c7e19877213-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"c51f2710-19e3-457c-ab1b-6c7e19877213\") " pod="openstack/nova-metadata-0" Oct 08 18:31:28 crc kubenswrapper[4988]: I1008 18:31:28.255647 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zk494\" (UniqueName: \"kubernetes.io/projected/c51f2710-19e3-457c-ab1b-6c7e19877213-kube-api-access-zk494\") pod \"nova-metadata-0\" (UID: \"c51f2710-19e3-457c-ab1b-6c7e19877213\") " pod="openstack/nova-metadata-0" Oct 08 18:31:28 crc kubenswrapper[4988]: I1008 18:31:28.372947 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 08 18:31:28 crc kubenswrapper[4988]: I1008 18:31:28.669843 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Oct 08 18:31:28 crc kubenswrapper[4988]: W1008 18:31:28.676718 4988 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod85bc7f8f_fdb3_4abb_88d3_ed0c10982c4c.slice/crio-20bc7fbba36bae4b2acfad1b694961fa583a6bb472efe5ab62f9bf0a088c1414 WatchSource:0}: Error finding container 20bc7fbba36bae4b2acfad1b694961fa583a6bb472efe5ab62f9bf0a088c1414: Status 404 returned error can't find the container with id 20bc7fbba36bae4b2acfad1b694961fa583a6bb472efe5ab62f9bf0a088c1414 Oct 08 18:31:28 crc kubenswrapper[4988]: I1008 18:31:28.845146 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 08 18:31:28 crc kubenswrapper[4988]: I1008 18:31:28.852532 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"85bc7f8f-fdb3-4abb-88d3-ed0c10982c4c","Type":"ContainerStarted","Data":"20bc7fbba36bae4b2acfad1b694961fa583a6bb472efe5ab62f9bf0a088c1414"} Oct 08 18:31:28 crc kubenswrapper[4988]: I1008 18:31:28.854647 4988 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="0e1348e9-ee35-49e7-a616-af30789c7ce0" containerName="nova-scheduler-scheduler" containerID="cri-o://2d3963fa42e1cd55fec9ceb64219ccf309ed4efb871f3ffc012626613b72c62c" gracePeriod=30 Oct 08 18:31:28 crc kubenswrapper[4988]: W1008 18:31:28.859821 4988 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc51f2710_19e3_457c_ab1b_6c7e19877213.slice/crio-854b0b91e2dd7bd74cfd3ea918cb2a1a74151fe5a7d547e3def9a1900ce4f5d7 WatchSource:0}: Error finding container 854b0b91e2dd7bd74cfd3ea918cb2a1a74151fe5a7d547e3def9a1900ce4f5d7: Status 404 returned error can't find the container with id 854b0b91e2dd7bd74cfd3ea918cb2a1a74151fe5a7d547e3def9a1900ce4f5d7 Oct 08 18:31:29 crc kubenswrapper[4988]: I1008 18:31:29.259410 4988 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b5a36680-3152-4d1f-a97d-92a3f1a3bc6e" path="/var/lib/kubelet/pods/b5a36680-3152-4d1f-a97d-92a3f1a3bc6e/volumes" Oct 08 18:31:29 crc kubenswrapper[4988]: I1008 18:31:29.260553 4988 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f6134524-f05a-43be-b80d-a862f3e7b390" path="/var/lib/kubelet/pods/f6134524-f05a-43be-b80d-a862f3e7b390/volumes" Oct 08 18:31:29 crc kubenswrapper[4988]: I1008 18:31:29.873593 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"c51f2710-19e3-457c-ab1b-6c7e19877213","Type":"ContainerStarted","Data":"8673b565fae2659ae04aabd4852445f45c49c3733848e2571f9ff5b470e6a680"} Oct 08 18:31:29 crc kubenswrapper[4988]: I1008 18:31:29.874037 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"c51f2710-19e3-457c-ab1b-6c7e19877213","Type":"ContainerStarted","Data":"15a065b5102872a8fc42dc14c1114cd59430e0c88890fc13c4496e3e17a3c6bc"} Oct 08 18:31:29 crc kubenswrapper[4988]: I1008 18:31:29.874097 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"c51f2710-19e3-457c-ab1b-6c7e19877213","Type":"ContainerStarted","Data":"854b0b91e2dd7bd74cfd3ea918cb2a1a74151fe5a7d547e3def9a1900ce4f5d7"} Oct 08 18:31:29 crc kubenswrapper[4988]: I1008 18:31:29.879795 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"85bc7f8f-fdb3-4abb-88d3-ed0c10982c4c","Type":"ContainerStarted","Data":"e18a49c9de8366d161727792956633e70d61d0fc8c2799f046b7a6f94af795e5"} Oct 08 18:31:29 crc kubenswrapper[4988]: I1008 18:31:29.880001 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-conductor-0" Oct 08 18:31:29 crc kubenswrapper[4988]: I1008 18:31:29.912302 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.912276816 podStartE2EDuration="2.912276816s" podCreationTimestamp="2025-10-08 18:31:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 18:31:29.894264059 +0000 UTC m=+1235.344106879" watchObservedRunningTime="2025-10-08 18:31:29.912276816 +0000 UTC m=+1235.362119606" Oct 08 18:31:29 crc kubenswrapper[4988]: I1008 18:31:29.933970 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-0" podStartSLOduration=2.93395048 podStartE2EDuration="2.93395048s" podCreationTimestamp="2025-10-08 18:31:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 18:31:29.920846241 +0000 UTC m=+1235.370689031" watchObservedRunningTime="2025-10-08 18:31:29.93395048 +0000 UTC m=+1235.383793260" Oct 08 18:31:31 crc kubenswrapper[4988]: E1008 18:31:31.450766 4988 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="2d3963fa42e1cd55fec9ceb64219ccf309ed4efb871f3ffc012626613b72c62c" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Oct 08 18:31:31 crc kubenswrapper[4988]: E1008 18:31:31.453514 4988 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="2d3963fa42e1cd55fec9ceb64219ccf309ed4efb871f3ffc012626613b72c62c" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Oct 08 18:31:31 crc kubenswrapper[4988]: E1008 18:31:31.456613 4988 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="2d3963fa42e1cd55fec9ceb64219ccf309ed4efb871f3ffc012626613b72c62c" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Oct 08 18:31:31 crc kubenswrapper[4988]: E1008 18:31:31.456653 4988 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-scheduler-0" podUID="0e1348e9-ee35-49e7-a616-af30789c7ce0" containerName="nova-scheduler-scheduler" Oct 08 18:31:31 crc kubenswrapper[4988]: I1008 18:31:31.811241 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Oct 08 18:31:32 crc kubenswrapper[4988]: I1008 18:31:32.911205 4988 generic.go:334] "Generic (PLEG): container finished" podID="0e1348e9-ee35-49e7-a616-af30789c7ce0" containerID="2d3963fa42e1cd55fec9ceb64219ccf309ed4efb871f3ffc012626613b72c62c" exitCode=0 Oct 08 18:31:32 crc kubenswrapper[4988]: I1008 18:31:32.911268 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"0e1348e9-ee35-49e7-a616-af30789c7ce0","Type":"ContainerDied","Data":"2d3963fa42e1cd55fec9ceb64219ccf309ed4efb871f3ffc012626613b72c62c"} Oct 08 18:31:32 crc kubenswrapper[4988]: I1008 18:31:32.916309 4988 generic.go:334] "Generic (PLEG): container finished" podID="a116a7eb-0c4f-4d6b-93ef-93ac26f22441" containerID="f1e67556c910c6012c3493e80adf12b19a788da6c247b9ef602743fd99cb3b7a" exitCode=0 Oct 08 18:31:32 crc kubenswrapper[4988]: I1008 18:31:32.916465 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"a116a7eb-0c4f-4d6b-93ef-93ac26f22441","Type":"ContainerDied","Data":"f1e67556c910c6012c3493e80adf12b19a788da6c247b9ef602743fd99cb3b7a"} Oct 08 18:31:32 crc kubenswrapper[4988]: I1008 18:31:32.916668 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"a116a7eb-0c4f-4d6b-93ef-93ac26f22441","Type":"ContainerDied","Data":"3f333c9fc45c29edb605c4f707b65a326fff530a56e6dded0a50bcb4776049ee"} Oct 08 18:31:32 crc kubenswrapper[4988]: I1008 18:31:32.916683 4988 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3f333c9fc45c29edb605c4f707b65a326fff530a56e6dded0a50bcb4776049ee" Oct 08 18:31:32 crc kubenswrapper[4988]: I1008 18:31:32.942502 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 08 18:31:33 crc kubenswrapper[4988]: I1008 18:31:33.081636 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 08 18:31:33 crc kubenswrapper[4988]: I1008 18:31:33.126284 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a116a7eb-0c4f-4d6b-93ef-93ac26f22441-config-data\") pod \"a116a7eb-0c4f-4d6b-93ef-93ac26f22441\" (UID: \"a116a7eb-0c4f-4d6b-93ef-93ac26f22441\") " Oct 08 18:31:33 crc kubenswrapper[4988]: I1008 18:31:33.126474 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a116a7eb-0c4f-4d6b-93ef-93ac26f22441-combined-ca-bundle\") pod \"a116a7eb-0c4f-4d6b-93ef-93ac26f22441\" (UID: \"a116a7eb-0c4f-4d6b-93ef-93ac26f22441\") " Oct 08 18:31:33 crc kubenswrapper[4988]: I1008 18:31:33.126535 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a116a7eb-0c4f-4d6b-93ef-93ac26f22441-logs\") pod \"a116a7eb-0c4f-4d6b-93ef-93ac26f22441\" (UID: \"a116a7eb-0c4f-4d6b-93ef-93ac26f22441\") " Oct 08 18:31:33 crc kubenswrapper[4988]: I1008 18:31:33.126725 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cd84r\" (UniqueName: \"kubernetes.io/projected/a116a7eb-0c4f-4d6b-93ef-93ac26f22441-kube-api-access-cd84r\") pod \"a116a7eb-0c4f-4d6b-93ef-93ac26f22441\" (UID: \"a116a7eb-0c4f-4d6b-93ef-93ac26f22441\") " Oct 08 18:31:33 crc kubenswrapper[4988]: I1008 18:31:33.127313 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a116a7eb-0c4f-4d6b-93ef-93ac26f22441-logs" (OuterVolumeSpecName: "logs") pod "a116a7eb-0c4f-4d6b-93ef-93ac26f22441" (UID: "a116a7eb-0c4f-4d6b-93ef-93ac26f22441"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 18:31:33 crc kubenswrapper[4988]: I1008 18:31:33.127458 4988 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a116a7eb-0c4f-4d6b-93ef-93ac26f22441-logs\") on node \"crc\" DevicePath \"\"" Oct 08 18:31:33 crc kubenswrapper[4988]: I1008 18:31:33.131682 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a116a7eb-0c4f-4d6b-93ef-93ac26f22441-kube-api-access-cd84r" (OuterVolumeSpecName: "kube-api-access-cd84r") pod "a116a7eb-0c4f-4d6b-93ef-93ac26f22441" (UID: "a116a7eb-0c4f-4d6b-93ef-93ac26f22441"). InnerVolumeSpecName "kube-api-access-cd84r". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:31:33 crc kubenswrapper[4988]: I1008 18:31:33.151791 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a116a7eb-0c4f-4d6b-93ef-93ac26f22441-config-data" (OuterVolumeSpecName: "config-data") pod "a116a7eb-0c4f-4d6b-93ef-93ac26f22441" (UID: "a116a7eb-0c4f-4d6b-93ef-93ac26f22441"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:31:33 crc kubenswrapper[4988]: I1008 18:31:33.157269 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a116a7eb-0c4f-4d6b-93ef-93ac26f22441-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a116a7eb-0c4f-4d6b-93ef-93ac26f22441" (UID: "a116a7eb-0c4f-4d6b-93ef-93ac26f22441"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:31:33 crc kubenswrapper[4988]: I1008 18:31:33.228583 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0e1348e9-ee35-49e7-a616-af30789c7ce0-combined-ca-bundle\") pod \"0e1348e9-ee35-49e7-a616-af30789c7ce0\" (UID: \"0e1348e9-ee35-49e7-a616-af30789c7ce0\") " Oct 08 18:31:33 crc kubenswrapper[4988]: I1008 18:31:33.228811 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0e1348e9-ee35-49e7-a616-af30789c7ce0-config-data\") pod \"0e1348e9-ee35-49e7-a616-af30789c7ce0\" (UID: \"0e1348e9-ee35-49e7-a616-af30789c7ce0\") " Oct 08 18:31:33 crc kubenswrapper[4988]: I1008 18:31:33.228852 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sfj8s\" (UniqueName: \"kubernetes.io/projected/0e1348e9-ee35-49e7-a616-af30789c7ce0-kube-api-access-sfj8s\") pod \"0e1348e9-ee35-49e7-a616-af30789c7ce0\" (UID: \"0e1348e9-ee35-49e7-a616-af30789c7ce0\") " Oct 08 18:31:33 crc kubenswrapper[4988]: I1008 18:31:33.229738 4988 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cd84r\" (UniqueName: \"kubernetes.io/projected/a116a7eb-0c4f-4d6b-93ef-93ac26f22441-kube-api-access-cd84r\") on node \"crc\" DevicePath \"\"" Oct 08 18:31:33 crc kubenswrapper[4988]: I1008 18:31:33.229761 4988 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a116a7eb-0c4f-4d6b-93ef-93ac26f22441-config-data\") on node \"crc\" DevicePath \"\"" Oct 08 18:31:33 crc kubenswrapper[4988]: I1008 18:31:33.229771 4988 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a116a7eb-0c4f-4d6b-93ef-93ac26f22441-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 18:31:33 crc kubenswrapper[4988]: I1008 18:31:33.232376 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0e1348e9-ee35-49e7-a616-af30789c7ce0-kube-api-access-sfj8s" (OuterVolumeSpecName: "kube-api-access-sfj8s") pod "0e1348e9-ee35-49e7-a616-af30789c7ce0" (UID: "0e1348e9-ee35-49e7-a616-af30789c7ce0"). InnerVolumeSpecName "kube-api-access-sfj8s". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:31:33 crc kubenswrapper[4988]: I1008 18:31:33.255916 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0e1348e9-ee35-49e7-a616-af30789c7ce0-config-data" (OuterVolumeSpecName: "config-data") pod "0e1348e9-ee35-49e7-a616-af30789c7ce0" (UID: "0e1348e9-ee35-49e7-a616-af30789c7ce0"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:31:33 crc kubenswrapper[4988]: I1008 18:31:33.259625 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0e1348e9-ee35-49e7-a616-af30789c7ce0-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "0e1348e9-ee35-49e7-a616-af30789c7ce0" (UID: "0e1348e9-ee35-49e7-a616-af30789c7ce0"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:31:33 crc kubenswrapper[4988]: I1008 18:31:33.342734 4988 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0e1348e9-ee35-49e7-a616-af30789c7ce0-config-data\") on node \"crc\" DevicePath \"\"" Oct 08 18:31:33 crc kubenswrapper[4988]: I1008 18:31:33.343297 4988 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sfj8s\" (UniqueName: \"kubernetes.io/projected/0e1348e9-ee35-49e7-a616-af30789c7ce0-kube-api-access-sfj8s\") on node \"crc\" DevicePath \"\"" Oct 08 18:31:33 crc kubenswrapper[4988]: I1008 18:31:33.343420 4988 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0e1348e9-ee35-49e7-a616-af30789c7ce0-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 18:31:33 crc kubenswrapper[4988]: I1008 18:31:33.374207 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Oct 08 18:31:33 crc kubenswrapper[4988]: I1008 18:31:33.374265 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Oct 08 18:31:33 crc kubenswrapper[4988]: I1008 18:31:33.925524 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"0e1348e9-ee35-49e7-a616-af30789c7ce0","Type":"ContainerDied","Data":"385df180470d78bda7bfcc67ab85a705f50ca98fbc5b551aef915c3c148e720f"} Oct 08 18:31:33 crc kubenswrapper[4988]: I1008 18:31:33.925589 4988 scope.go:117] "RemoveContainer" containerID="2d3963fa42e1cd55fec9ceb64219ccf309ed4efb871f3ffc012626613b72c62c" Oct 08 18:31:33 crc kubenswrapper[4988]: I1008 18:31:33.925592 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 08 18:31:33 crc kubenswrapper[4988]: I1008 18:31:33.925543 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 08 18:31:33 crc kubenswrapper[4988]: I1008 18:31:33.957572 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 08 18:31:33 crc kubenswrapper[4988]: I1008 18:31:33.974803 4988 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Oct 08 18:31:33 crc kubenswrapper[4988]: I1008 18:31:33.990988 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Oct 08 18:31:34 crc kubenswrapper[4988]: I1008 18:31:34.000712 4988 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Oct 08 18:31:34 crc kubenswrapper[4988]: I1008 18:31:34.025147 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Oct 08 18:31:34 crc kubenswrapper[4988]: E1008 18:31:34.025763 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a116a7eb-0c4f-4d6b-93ef-93ac26f22441" containerName="nova-api-log" Oct 08 18:31:34 crc kubenswrapper[4988]: I1008 18:31:34.025834 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="a116a7eb-0c4f-4d6b-93ef-93ac26f22441" containerName="nova-api-log" Oct 08 18:31:34 crc kubenswrapper[4988]: E1008 18:31:34.025899 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0e1348e9-ee35-49e7-a616-af30789c7ce0" containerName="nova-scheduler-scheduler" Oct 08 18:31:34 crc kubenswrapper[4988]: I1008 18:31:34.025990 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="0e1348e9-ee35-49e7-a616-af30789c7ce0" containerName="nova-scheduler-scheduler" Oct 08 18:31:34 crc kubenswrapper[4988]: E1008 18:31:34.026069 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a116a7eb-0c4f-4d6b-93ef-93ac26f22441" containerName="nova-api-api" Oct 08 18:31:34 crc kubenswrapper[4988]: I1008 18:31:34.026138 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="a116a7eb-0c4f-4d6b-93ef-93ac26f22441" containerName="nova-api-api" Oct 08 18:31:34 crc kubenswrapper[4988]: I1008 18:31:34.026405 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="a116a7eb-0c4f-4d6b-93ef-93ac26f22441" containerName="nova-api-log" Oct 08 18:31:34 crc kubenswrapper[4988]: I1008 18:31:34.026483 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="a116a7eb-0c4f-4d6b-93ef-93ac26f22441" containerName="nova-api-api" Oct 08 18:31:34 crc kubenswrapper[4988]: I1008 18:31:34.026553 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="0e1348e9-ee35-49e7-a616-af30789c7ce0" containerName="nova-scheduler-scheduler" Oct 08 18:31:34 crc kubenswrapper[4988]: I1008 18:31:34.027567 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Oct 08 18:31:34 crc kubenswrapper[4988]: I1008 18:31:34.028461 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 08 18:31:34 crc kubenswrapper[4988]: I1008 18:31:34.028608 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 08 18:31:34 crc kubenswrapper[4988]: I1008 18:31:34.037001 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Oct 08 18:31:34 crc kubenswrapper[4988]: I1008 18:31:34.037216 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 08 18:31:34 crc kubenswrapper[4988]: I1008 18:31:34.052306 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Oct 08 18:31:34 crc kubenswrapper[4988]: I1008 18:31:34.052358 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Oct 08 18:31:34 crc kubenswrapper[4988]: I1008 18:31:34.055507 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xvrsx\" (UniqueName: \"kubernetes.io/projected/6c8adde4-41e0-4d33-961a-17a76088b3b8-kube-api-access-xvrsx\") pod \"nova-api-0\" (UID: \"6c8adde4-41e0-4d33-961a-17a76088b3b8\") " pod="openstack/nova-api-0" Oct 08 18:31:34 crc kubenswrapper[4988]: I1008 18:31:34.055591 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6c8adde4-41e0-4d33-961a-17a76088b3b8-logs\") pod \"nova-api-0\" (UID: \"6c8adde4-41e0-4d33-961a-17a76088b3b8\") " pod="openstack/nova-api-0" Oct 08 18:31:34 crc kubenswrapper[4988]: I1008 18:31:34.055614 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6c8adde4-41e0-4d33-961a-17a76088b3b8-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"6c8adde4-41e0-4d33-961a-17a76088b3b8\") " pod="openstack/nova-api-0" Oct 08 18:31:34 crc kubenswrapper[4988]: I1008 18:31:34.055633 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/be97711d-1138-4866-a82d-c0ddbcec58c8-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"be97711d-1138-4866-a82d-c0ddbcec58c8\") " pod="openstack/nova-scheduler-0" Oct 08 18:31:34 crc kubenswrapper[4988]: I1008 18:31:34.055685 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xgfbc\" (UniqueName: \"kubernetes.io/projected/be97711d-1138-4866-a82d-c0ddbcec58c8-kube-api-access-xgfbc\") pod \"nova-scheduler-0\" (UID: \"be97711d-1138-4866-a82d-c0ddbcec58c8\") " pod="openstack/nova-scheduler-0" Oct 08 18:31:34 crc kubenswrapper[4988]: I1008 18:31:34.055710 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6c8adde4-41e0-4d33-961a-17a76088b3b8-config-data\") pod \"nova-api-0\" (UID: \"6c8adde4-41e0-4d33-961a-17a76088b3b8\") " pod="openstack/nova-api-0" Oct 08 18:31:34 crc kubenswrapper[4988]: I1008 18:31:34.055736 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/be97711d-1138-4866-a82d-c0ddbcec58c8-config-data\") pod \"nova-scheduler-0\" (UID: \"be97711d-1138-4866-a82d-c0ddbcec58c8\") " pod="openstack/nova-scheduler-0" Oct 08 18:31:34 crc kubenswrapper[4988]: I1008 18:31:34.156520 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xvrsx\" (UniqueName: \"kubernetes.io/projected/6c8adde4-41e0-4d33-961a-17a76088b3b8-kube-api-access-xvrsx\") pod \"nova-api-0\" (UID: \"6c8adde4-41e0-4d33-961a-17a76088b3b8\") " pod="openstack/nova-api-0" Oct 08 18:31:34 crc kubenswrapper[4988]: I1008 18:31:34.156592 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6c8adde4-41e0-4d33-961a-17a76088b3b8-logs\") pod \"nova-api-0\" (UID: \"6c8adde4-41e0-4d33-961a-17a76088b3b8\") " pod="openstack/nova-api-0" Oct 08 18:31:34 crc kubenswrapper[4988]: I1008 18:31:34.156615 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6c8adde4-41e0-4d33-961a-17a76088b3b8-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"6c8adde4-41e0-4d33-961a-17a76088b3b8\") " pod="openstack/nova-api-0" Oct 08 18:31:34 crc kubenswrapper[4988]: I1008 18:31:34.156636 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/be97711d-1138-4866-a82d-c0ddbcec58c8-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"be97711d-1138-4866-a82d-c0ddbcec58c8\") " pod="openstack/nova-scheduler-0" Oct 08 18:31:34 crc kubenswrapper[4988]: I1008 18:31:34.156682 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xgfbc\" (UniqueName: \"kubernetes.io/projected/be97711d-1138-4866-a82d-c0ddbcec58c8-kube-api-access-xgfbc\") pod \"nova-scheduler-0\" (UID: \"be97711d-1138-4866-a82d-c0ddbcec58c8\") " pod="openstack/nova-scheduler-0" Oct 08 18:31:34 crc kubenswrapper[4988]: I1008 18:31:34.156710 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6c8adde4-41e0-4d33-961a-17a76088b3b8-config-data\") pod \"nova-api-0\" (UID: \"6c8adde4-41e0-4d33-961a-17a76088b3b8\") " pod="openstack/nova-api-0" Oct 08 18:31:34 crc kubenswrapper[4988]: I1008 18:31:34.156734 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/be97711d-1138-4866-a82d-c0ddbcec58c8-config-data\") pod \"nova-scheduler-0\" (UID: \"be97711d-1138-4866-a82d-c0ddbcec58c8\") " pod="openstack/nova-scheduler-0" Oct 08 18:31:34 crc kubenswrapper[4988]: I1008 18:31:34.157019 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6c8adde4-41e0-4d33-961a-17a76088b3b8-logs\") pod \"nova-api-0\" (UID: \"6c8adde4-41e0-4d33-961a-17a76088b3b8\") " pod="openstack/nova-api-0" Oct 08 18:31:34 crc kubenswrapper[4988]: I1008 18:31:34.161869 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/be97711d-1138-4866-a82d-c0ddbcec58c8-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"be97711d-1138-4866-a82d-c0ddbcec58c8\") " pod="openstack/nova-scheduler-0" Oct 08 18:31:34 crc kubenswrapper[4988]: I1008 18:31:34.162377 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6c8adde4-41e0-4d33-961a-17a76088b3b8-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"6c8adde4-41e0-4d33-961a-17a76088b3b8\") " pod="openstack/nova-api-0" Oct 08 18:31:34 crc kubenswrapper[4988]: I1008 18:31:34.162616 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6c8adde4-41e0-4d33-961a-17a76088b3b8-config-data\") pod \"nova-api-0\" (UID: \"6c8adde4-41e0-4d33-961a-17a76088b3b8\") " pod="openstack/nova-api-0" Oct 08 18:31:34 crc kubenswrapper[4988]: I1008 18:31:34.163898 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/be97711d-1138-4866-a82d-c0ddbcec58c8-config-data\") pod \"nova-scheduler-0\" (UID: \"be97711d-1138-4866-a82d-c0ddbcec58c8\") " pod="openstack/nova-scheduler-0" Oct 08 18:31:34 crc kubenswrapper[4988]: I1008 18:31:34.175950 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xgfbc\" (UniqueName: \"kubernetes.io/projected/be97711d-1138-4866-a82d-c0ddbcec58c8-kube-api-access-xgfbc\") pod \"nova-scheduler-0\" (UID: \"be97711d-1138-4866-a82d-c0ddbcec58c8\") " pod="openstack/nova-scheduler-0" Oct 08 18:31:34 crc kubenswrapper[4988]: I1008 18:31:34.188213 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xvrsx\" (UniqueName: \"kubernetes.io/projected/6c8adde4-41e0-4d33-961a-17a76088b3b8-kube-api-access-xvrsx\") pod \"nova-api-0\" (UID: \"6c8adde4-41e0-4d33-961a-17a76088b3b8\") " pod="openstack/nova-api-0" Oct 08 18:31:34 crc kubenswrapper[4988]: I1008 18:31:34.372636 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 08 18:31:34 crc kubenswrapper[4988]: I1008 18:31:34.382740 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 08 18:31:34 crc kubenswrapper[4988]: I1008 18:31:34.919690 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 08 18:31:34 crc kubenswrapper[4988]: I1008 18:31:34.934249 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Oct 08 18:31:34 crc kubenswrapper[4988]: I1008 18:31:34.945440 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"6c8adde4-41e0-4d33-961a-17a76088b3b8","Type":"ContainerStarted","Data":"3510a6108a346aa0be640613afe19a7f162113ead0d4f39c8a2d108f4df106d7"} Oct 08 18:31:35 crc kubenswrapper[4988]: I1008 18:31:35.279038 4988 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0e1348e9-ee35-49e7-a616-af30789c7ce0" path="/var/lib/kubelet/pods/0e1348e9-ee35-49e7-a616-af30789c7ce0/volumes" Oct 08 18:31:35 crc kubenswrapper[4988]: I1008 18:31:35.285946 4988 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a116a7eb-0c4f-4d6b-93ef-93ac26f22441" path="/var/lib/kubelet/pods/a116a7eb-0c4f-4d6b-93ef-93ac26f22441/volumes" Oct 08 18:31:35 crc kubenswrapper[4988]: I1008 18:31:35.964976 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"be97711d-1138-4866-a82d-c0ddbcec58c8","Type":"ContainerStarted","Data":"3f453b3226c2919bd0521ef15a3bf5e939375ac3f298ac9733160689d3fda89b"} Oct 08 18:31:35 crc kubenswrapper[4988]: I1008 18:31:35.965341 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"be97711d-1138-4866-a82d-c0ddbcec58c8","Type":"ContainerStarted","Data":"36549afe6c5d0a01f18a6ff16f92b8eb840f2473ae9b0af84b6e4afcaf57e47c"} Oct 08 18:31:35 crc kubenswrapper[4988]: I1008 18:31:35.967923 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"6c8adde4-41e0-4d33-961a-17a76088b3b8","Type":"ContainerStarted","Data":"56adcfbff76321bed6638d6eb70e3ed96f4f090daebaacfb04d44073983e39a8"} Oct 08 18:31:35 crc kubenswrapper[4988]: I1008 18:31:35.967948 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"6c8adde4-41e0-4d33-961a-17a76088b3b8","Type":"ContainerStarted","Data":"8a29e42569215a39afa39fa848957fbb8c88d7c870e4e76ffcd20fafb98282df"} Oct 08 18:31:35 crc kubenswrapper[4988]: I1008 18:31:35.988884 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.988868326 podStartE2EDuration="2.988868326s" podCreationTimestamp="2025-10-08 18:31:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 18:31:35.981972706 +0000 UTC m=+1241.431815476" watchObservedRunningTime="2025-10-08 18:31:35.988868326 +0000 UTC m=+1241.438711096" Oct 08 18:31:36 crc kubenswrapper[4988]: I1008 18:31:36.101792 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=3.101773483 podStartE2EDuration="3.101773483s" podCreationTimestamp="2025-10-08 18:31:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 18:31:36.003014339 +0000 UTC m=+1241.452857099" watchObservedRunningTime="2025-10-08 18:31:36.101773483 +0000 UTC m=+1241.551616253" Oct 08 18:31:36 crc kubenswrapper[4988]: I1008 18:31:36.113090 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 08 18:31:36 crc kubenswrapper[4988]: I1008 18:31:36.113294 4988 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/kube-state-metrics-0" podUID="731acc53-25a1-43a4-b9a8-16de25b1e7b1" containerName="kube-state-metrics" containerID="cri-o://bcd9a23419ff288fe7e9ecd15147e5affee5f1bc37f5d8bf5a0021d7900242f6" gracePeriod=30 Oct 08 18:31:36 crc kubenswrapper[4988]: I1008 18:31:36.630275 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Oct 08 18:31:36 crc kubenswrapper[4988]: I1008 18:31:36.811617 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-j2n7z\" (UniqueName: \"kubernetes.io/projected/731acc53-25a1-43a4-b9a8-16de25b1e7b1-kube-api-access-j2n7z\") pod \"731acc53-25a1-43a4-b9a8-16de25b1e7b1\" (UID: \"731acc53-25a1-43a4-b9a8-16de25b1e7b1\") " Oct 08 18:31:36 crc kubenswrapper[4988]: I1008 18:31:36.817583 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/731acc53-25a1-43a4-b9a8-16de25b1e7b1-kube-api-access-j2n7z" (OuterVolumeSpecName: "kube-api-access-j2n7z") pod "731acc53-25a1-43a4-b9a8-16de25b1e7b1" (UID: "731acc53-25a1-43a4-b9a8-16de25b1e7b1"). InnerVolumeSpecName "kube-api-access-j2n7z". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:31:36 crc kubenswrapper[4988]: I1008 18:31:36.913763 4988 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-j2n7z\" (UniqueName: \"kubernetes.io/projected/731acc53-25a1-43a4-b9a8-16de25b1e7b1-kube-api-access-j2n7z\") on node \"crc\" DevicePath \"\"" Oct 08 18:31:36 crc kubenswrapper[4988]: I1008 18:31:36.981918 4988 generic.go:334] "Generic (PLEG): container finished" podID="731acc53-25a1-43a4-b9a8-16de25b1e7b1" containerID="bcd9a23419ff288fe7e9ecd15147e5affee5f1bc37f5d8bf5a0021d7900242f6" exitCode=2 Oct 08 18:31:36 crc kubenswrapper[4988]: I1008 18:31:36.981995 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"731acc53-25a1-43a4-b9a8-16de25b1e7b1","Type":"ContainerDied","Data":"bcd9a23419ff288fe7e9ecd15147e5affee5f1bc37f5d8bf5a0021d7900242f6"} Oct 08 18:31:36 crc kubenswrapper[4988]: I1008 18:31:36.982083 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"731acc53-25a1-43a4-b9a8-16de25b1e7b1","Type":"ContainerDied","Data":"b1e9368f2c64f0a3360049bc186a23f993a494e9080ad57b8a79456a4d59b220"} Oct 08 18:31:36 crc kubenswrapper[4988]: I1008 18:31:36.982112 4988 scope.go:117] "RemoveContainer" containerID="bcd9a23419ff288fe7e9ecd15147e5affee5f1bc37f5d8bf5a0021d7900242f6" Oct 08 18:31:36 crc kubenswrapper[4988]: I1008 18:31:36.982290 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Oct 08 18:31:37 crc kubenswrapper[4988]: I1008 18:31:37.012806 4988 scope.go:117] "RemoveContainer" containerID="bcd9a23419ff288fe7e9ecd15147e5affee5f1bc37f5d8bf5a0021d7900242f6" Oct 08 18:31:37 crc kubenswrapper[4988]: E1008 18:31:37.013317 4988 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bcd9a23419ff288fe7e9ecd15147e5affee5f1bc37f5d8bf5a0021d7900242f6\": container with ID starting with bcd9a23419ff288fe7e9ecd15147e5affee5f1bc37f5d8bf5a0021d7900242f6 not found: ID does not exist" containerID="bcd9a23419ff288fe7e9ecd15147e5affee5f1bc37f5d8bf5a0021d7900242f6" Oct 08 18:31:37 crc kubenswrapper[4988]: I1008 18:31:37.013362 4988 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bcd9a23419ff288fe7e9ecd15147e5affee5f1bc37f5d8bf5a0021d7900242f6"} err="failed to get container status \"bcd9a23419ff288fe7e9ecd15147e5affee5f1bc37f5d8bf5a0021d7900242f6\": rpc error: code = NotFound desc = could not find container \"bcd9a23419ff288fe7e9ecd15147e5affee5f1bc37f5d8bf5a0021d7900242f6\": container with ID starting with bcd9a23419ff288fe7e9ecd15147e5affee5f1bc37f5d8bf5a0021d7900242f6 not found: ID does not exist" Oct 08 18:31:37 crc kubenswrapper[4988]: I1008 18:31:37.038121 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 08 18:31:37 crc kubenswrapper[4988]: I1008 18:31:37.051140 4988 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 08 18:31:37 crc kubenswrapper[4988]: I1008 18:31:37.062580 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Oct 08 18:31:37 crc kubenswrapper[4988]: E1008 18:31:37.063045 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="731acc53-25a1-43a4-b9a8-16de25b1e7b1" containerName="kube-state-metrics" Oct 08 18:31:37 crc kubenswrapper[4988]: I1008 18:31:37.063063 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="731acc53-25a1-43a4-b9a8-16de25b1e7b1" containerName="kube-state-metrics" Oct 08 18:31:37 crc kubenswrapper[4988]: I1008 18:31:37.063275 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="731acc53-25a1-43a4-b9a8-16de25b1e7b1" containerName="kube-state-metrics" Oct 08 18:31:37 crc kubenswrapper[4988]: I1008 18:31:37.063949 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Oct 08 18:31:37 crc kubenswrapper[4988]: I1008 18:31:37.066791 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-kube-state-metrics-svc" Oct 08 18:31:37 crc kubenswrapper[4988]: I1008 18:31:37.066906 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"kube-state-metrics-tls-config" Oct 08 18:31:37 crc kubenswrapper[4988]: I1008 18:31:37.071931 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 08 18:31:37 crc kubenswrapper[4988]: I1008 18:31:37.219039 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/b6d1becf-4229-40bf-b53a-d8a9eb4b580e-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"b6d1becf-4229-40bf-b53a-d8a9eb4b580e\") " pod="openstack/kube-state-metrics-0" Oct 08 18:31:37 crc kubenswrapper[4988]: I1008 18:31:37.219184 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2hv4g\" (UniqueName: \"kubernetes.io/projected/b6d1becf-4229-40bf-b53a-d8a9eb4b580e-kube-api-access-2hv4g\") pod \"kube-state-metrics-0\" (UID: \"b6d1becf-4229-40bf-b53a-d8a9eb4b580e\") " pod="openstack/kube-state-metrics-0" Oct 08 18:31:37 crc kubenswrapper[4988]: I1008 18:31:37.219259 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b6d1becf-4229-40bf-b53a-d8a9eb4b580e-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"b6d1becf-4229-40bf-b53a-d8a9eb4b580e\") " pod="openstack/kube-state-metrics-0" Oct 08 18:31:37 crc kubenswrapper[4988]: I1008 18:31:37.219298 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/b6d1becf-4229-40bf-b53a-d8a9eb4b580e-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"b6d1becf-4229-40bf-b53a-d8a9eb4b580e\") " pod="openstack/kube-state-metrics-0" Oct 08 18:31:37 crc kubenswrapper[4988]: I1008 18:31:37.250197 4988 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="731acc53-25a1-43a4-b9a8-16de25b1e7b1" path="/var/lib/kubelet/pods/731acc53-25a1-43a4-b9a8-16de25b1e7b1/volumes" Oct 08 18:31:37 crc kubenswrapper[4988]: I1008 18:31:37.320875 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2hv4g\" (UniqueName: \"kubernetes.io/projected/b6d1becf-4229-40bf-b53a-d8a9eb4b580e-kube-api-access-2hv4g\") pod \"kube-state-metrics-0\" (UID: \"b6d1becf-4229-40bf-b53a-d8a9eb4b580e\") " pod="openstack/kube-state-metrics-0" Oct 08 18:31:37 crc kubenswrapper[4988]: I1008 18:31:37.320975 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b6d1becf-4229-40bf-b53a-d8a9eb4b580e-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"b6d1becf-4229-40bf-b53a-d8a9eb4b580e\") " pod="openstack/kube-state-metrics-0" Oct 08 18:31:37 crc kubenswrapper[4988]: I1008 18:31:37.321014 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/b6d1becf-4229-40bf-b53a-d8a9eb4b580e-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"b6d1becf-4229-40bf-b53a-d8a9eb4b580e\") " pod="openstack/kube-state-metrics-0" Oct 08 18:31:37 crc kubenswrapper[4988]: I1008 18:31:37.321081 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/b6d1becf-4229-40bf-b53a-d8a9eb4b580e-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"b6d1becf-4229-40bf-b53a-d8a9eb4b580e\") " pod="openstack/kube-state-metrics-0" Oct 08 18:31:37 crc kubenswrapper[4988]: I1008 18:31:37.326796 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b6d1becf-4229-40bf-b53a-d8a9eb4b580e-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"b6d1becf-4229-40bf-b53a-d8a9eb4b580e\") " pod="openstack/kube-state-metrics-0" Oct 08 18:31:37 crc kubenswrapper[4988]: I1008 18:31:37.328456 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/b6d1becf-4229-40bf-b53a-d8a9eb4b580e-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"b6d1becf-4229-40bf-b53a-d8a9eb4b580e\") " pod="openstack/kube-state-metrics-0" Oct 08 18:31:37 crc kubenswrapper[4988]: I1008 18:31:37.338747 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/b6d1becf-4229-40bf-b53a-d8a9eb4b580e-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"b6d1becf-4229-40bf-b53a-d8a9eb4b580e\") " pod="openstack/kube-state-metrics-0" Oct 08 18:31:37 crc kubenswrapper[4988]: I1008 18:31:37.346486 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2hv4g\" (UniqueName: \"kubernetes.io/projected/b6d1becf-4229-40bf-b53a-d8a9eb4b580e-kube-api-access-2hv4g\") pod \"kube-state-metrics-0\" (UID: \"b6d1becf-4229-40bf-b53a-d8a9eb4b580e\") " pod="openstack/kube-state-metrics-0" Oct 08 18:31:37 crc kubenswrapper[4988]: I1008 18:31:37.388461 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Oct 08 18:31:37 crc kubenswrapper[4988]: I1008 18:31:37.942486 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 08 18:31:38 crc kubenswrapper[4988]: I1008 18:31:38.001175 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"b6d1becf-4229-40bf-b53a-d8a9eb4b580e","Type":"ContainerStarted","Data":"f064454d5dc1aaecb4c0413e7e223abb0b5fa88fbea49db04428e68ae3072ec7"} Oct 08 18:31:38 crc kubenswrapper[4988]: I1008 18:31:38.193348 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 08 18:31:38 crc kubenswrapper[4988]: I1008 18:31:38.195586 4988 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="fa5061f0-e7f7-4059-9f69-48d6a1cb98d7" containerName="ceilometer-central-agent" containerID="cri-o://4abb52d2340f790b7dbc4c91a708a887af9a63619f61ed6defbeb30cce9e793b" gracePeriod=30 Oct 08 18:31:38 crc kubenswrapper[4988]: I1008 18:31:38.195738 4988 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="fa5061f0-e7f7-4059-9f69-48d6a1cb98d7" containerName="ceilometer-notification-agent" containerID="cri-o://4edcb300385ece56d49e71ae3fa0161264857a94d80352e9f4dfc1128294a789" gracePeriod=30 Oct 08 18:31:38 crc kubenswrapper[4988]: I1008 18:31:38.195732 4988 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="fa5061f0-e7f7-4059-9f69-48d6a1cb98d7" containerName="sg-core" containerID="cri-o://e19138f089a607cc5db443a8dfe9a62efdb451bc27329b434278497a10f97446" gracePeriod=30 Oct 08 18:31:38 crc kubenswrapper[4988]: I1008 18:31:38.195791 4988 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="fa5061f0-e7f7-4059-9f69-48d6a1cb98d7" containerName="proxy-httpd" containerID="cri-o://7f1af50932c5ecb7086c67d8b967a7c5d6ed5ac407c3ca0a5e4d0a729102c747" gracePeriod=30 Oct 08 18:31:38 crc kubenswrapper[4988]: I1008 18:31:38.247990 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-conductor-0" Oct 08 18:31:38 crc kubenswrapper[4988]: I1008 18:31:38.375322 4988 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Oct 08 18:31:38 crc kubenswrapper[4988]: I1008 18:31:38.375948 4988 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Oct 08 18:31:39 crc kubenswrapper[4988]: I1008 18:31:39.011629 4988 generic.go:334] "Generic (PLEG): container finished" podID="fa5061f0-e7f7-4059-9f69-48d6a1cb98d7" containerID="7f1af50932c5ecb7086c67d8b967a7c5d6ed5ac407c3ca0a5e4d0a729102c747" exitCode=0 Oct 08 18:31:39 crc kubenswrapper[4988]: I1008 18:31:39.011953 4988 generic.go:334] "Generic (PLEG): container finished" podID="fa5061f0-e7f7-4059-9f69-48d6a1cb98d7" containerID="e19138f089a607cc5db443a8dfe9a62efdb451bc27329b434278497a10f97446" exitCode=2 Oct 08 18:31:39 crc kubenswrapper[4988]: I1008 18:31:39.011968 4988 generic.go:334] "Generic (PLEG): container finished" podID="fa5061f0-e7f7-4059-9f69-48d6a1cb98d7" containerID="4abb52d2340f790b7dbc4c91a708a887af9a63619f61ed6defbeb30cce9e793b" exitCode=0 Oct 08 18:31:39 crc kubenswrapper[4988]: I1008 18:31:39.011709 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"fa5061f0-e7f7-4059-9f69-48d6a1cb98d7","Type":"ContainerDied","Data":"7f1af50932c5ecb7086c67d8b967a7c5d6ed5ac407c3ca0a5e4d0a729102c747"} Oct 08 18:31:39 crc kubenswrapper[4988]: I1008 18:31:39.012042 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"fa5061f0-e7f7-4059-9f69-48d6a1cb98d7","Type":"ContainerDied","Data":"e19138f089a607cc5db443a8dfe9a62efdb451bc27329b434278497a10f97446"} Oct 08 18:31:39 crc kubenswrapper[4988]: I1008 18:31:39.012062 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"fa5061f0-e7f7-4059-9f69-48d6a1cb98d7","Type":"ContainerDied","Data":"4abb52d2340f790b7dbc4c91a708a887af9a63619f61ed6defbeb30cce9e793b"} Oct 08 18:31:39 crc kubenswrapper[4988]: I1008 18:31:39.013815 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"b6d1becf-4229-40bf-b53a-d8a9eb4b580e","Type":"ContainerStarted","Data":"0f31eca0db70bc881a635808747404e9ac717d3d38c16e693305f6196d241dac"} Oct 08 18:31:39 crc kubenswrapper[4988]: I1008 18:31:39.013979 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Oct 08 18:31:39 crc kubenswrapper[4988]: I1008 18:31:39.031276 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=1.6569785750000001 podStartE2EDuration="2.031255633s" podCreationTimestamp="2025-10-08 18:31:37 +0000 UTC" firstStartedPulling="2025-10-08 18:31:37.950483463 +0000 UTC m=+1243.400326233" lastFinishedPulling="2025-10-08 18:31:38.324760521 +0000 UTC m=+1243.774603291" observedRunningTime="2025-10-08 18:31:39.029040542 +0000 UTC m=+1244.478883322" watchObservedRunningTime="2025-10-08 18:31:39.031255633 +0000 UTC m=+1244.481098403" Oct 08 18:31:39 crc kubenswrapper[4988]: I1008 18:31:39.373531 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Oct 08 18:31:39 crc kubenswrapper[4988]: I1008 18:31:39.387564 4988 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="c51f2710-19e3-457c-ab1b-6c7e19877213" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.193:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 08 18:31:39 crc kubenswrapper[4988]: I1008 18:31:39.387638 4988 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="c51f2710-19e3-457c-ab1b-6c7e19877213" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.193:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 08 18:31:40 crc kubenswrapper[4988]: I1008 18:31:40.046887 4988 generic.go:334] "Generic (PLEG): container finished" podID="fa5061f0-e7f7-4059-9f69-48d6a1cb98d7" containerID="4edcb300385ece56d49e71ae3fa0161264857a94d80352e9f4dfc1128294a789" exitCode=0 Oct 08 18:31:40 crc kubenswrapper[4988]: I1008 18:31:40.048030 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"fa5061f0-e7f7-4059-9f69-48d6a1cb98d7","Type":"ContainerDied","Data":"4edcb300385ece56d49e71ae3fa0161264857a94d80352e9f4dfc1128294a789"} Oct 08 18:31:40 crc kubenswrapper[4988]: I1008 18:31:40.445947 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 08 18:31:40 crc kubenswrapper[4988]: I1008 18:31:40.579844 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fa5061f0-e7f7-4059-9f69-48d6a1cb98d7-combined-ca-bundle\") pod \"fa5061f0-e7f7-4059-9f69-48d6a1cb98d7\" (UID: \"fa5061f0-e7f7-4059-9f69-48d6a1cb98d7\") " Oct 08 18:31:40 crc kubenswrapper[4988]: I1008 18:31:40.579941 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/fa5061f0-e7f7-4059-9f69-48d6a1cb98d7-log-httpd\") pod \"fa5061f0-e7f7-4059-9f69-48d6a1cb98d7\" (UID: \"fa5061f0-e7f7-4059-9f69-48d6a1cb98d7\") " Oct 08 18:31:40 crc kubenswrapper[4988]: I1008 18:31:40.580026 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jcxnj\" (UniqueName: \"kubernetes.io/projected/fa5061f0-e7f7-4059-9f69-48d6a1cb98d7-kube-api-access-jcxnj\") pod \"fa5061f0-e7f7-4059-9f69-48d6a1cb98d7\" (UID: \"fa5061f0-e7f7-4059-9f69-48d6a1cb98d7\") " Oct 08 18:31:40 crc kubenswrapper[4988]: I1008 18:31:40.580070 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fa5061f0-e7f7-4059-9f69-48d6a1cb98d7-scripts\") pod \"fa5061f0-e7f7-4059-9f69-48d6a1cb98d7\" (UID: \"fa5061f0-e7f7-4059-9f69-48d6a1cb98d7\") " Oct 08 18:31:40 crc kubenswrapper[4988]: I1008 18:31:40.580135 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fa5061f0-e7f7-4059-9f69-48d6a1cb98d7-config-data\") pod \"fa5061f0-e7f7-4059-9f69-48d6a1cb98d7\" (UID: \"fa5061f0-e7f7-4059-9f69-48d6a1cb98d7\") " Oct 08 18:31:40 crc kubenswrapper[4988]: I1008 18:31:40.580181 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/fa5061f0-e7f7-4059-9f69-48d6a1cb98d7-sg-core-conf-yaml\") pod \"fa5061f0-e7f7-4059-9f69-48d6a1cb98d7\" (UID: \"fa5061f0-e7f7-4059-9f69-48d6a1cb98d7\") " Oct 08 18:31:40 crc kubenswrapper[4988]: I1008 18:31:40.580237 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/fa5061f0-e7f7-4059-9f69-48d6a1cb98d7-run-httpd\") pod \"fa5061f0-e7f7-4059-9f69-48d6a1cb98d7\" (UID: \"fa5061f0-e7f7-4059-9f69-48d6a1cb98d7\") " Oct 08 18:31:40 crc kubenswrapper[4988]: I1008 18:31:40.581106 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fa5061f0-e7f7-4059-9f69-48d6a1cb98d7-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "fa5061f0-e7f7-4059-9f69-48d6a1cb98d7" (UID: "fa5061f0-e7f7-4059-9f69-48d6a1cb98d7"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 18:31:40 crc kubenswrapper[4988]: I1008 18:31:40.583164 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fa5061f0-e7f7-4059-9f69-48d6a1cb98d7-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "fa5061f0-e7f7-4059-9f69-48d6a1cb98d7" (UID: "fa5061f0-e7f7-4059-9f69-48d6a1cb98d7"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 18:31:40 crc kubenswrapper[4988]: I1008 18:31:40.592486 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fa5061f0-e7f7-4059-9f69-48d6a1cb98d7-scripts" (OuterVolumeSpecName: "scripts") pod "fa5061f0-e7f7-4059-9f69-48d6a1cb98d7" (UID: "fa5061f0-e7f7-4059-9f69-48d6a1cb98d7"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:31:40 crc kubenswrapper[4988]: I1008 18:31:40.597974 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fa5061f0-e7f7-4059-9f69-48d6a1cb98d7-kube-api-access-jcxnj" (OuterVolumeSpecName: "kube-api-access-jcxnj") pod "fa5061f0-e7f7-4059-9f69-48d6a1cb98d7" (UID: "fa5061f0-e7f7-4059-9f69-48d6a1cb98d7"). InnerVolumeSpecName "kube-api-access-jcxnj". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:31:40 crc kubenswrapper[4988]: I1008 18:31:40.623285 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fa5061f0-e7f7-4059-9f69-48d6a1cb98d7-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "fa5061f0-e7f7-4059-9f69-48d6a1cb98d7" (UID: "fa5061f0-e7f7-4059-9f69-48d6a1cb98d7"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:31:40 crc kubenswrapper[4988]: I1008 18:31:40.668993 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fa5061f0-e7f7-4059-9f69-48d6a1cb98d7-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "fa5061f0-e7f7-4059-9f69-48d6a1cb98d7" (UID: "fa5061f0-e7f7-4059-9f69-48d6a1cb98d7"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:31:40 crc kubenswrapper[4988]: I1008 18:31:40.682755 4988 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/fa5061f0-e7f7-4059-9f69-48d6a1cb98d7-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Oct 08 18:31:40 crc kubenswrapper[4988]: I1008 18:31:40.682803 4988 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/fa5061f0-e7f7-4059-9f69-48d6a1cb98d7-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 08 18:31:40 crc kubenswrapper[4988]: I1008 18:31:40.682815 4988 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fa5061f0-e7f7-4059-9f69-48d6a1cb98d7-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 18:31:40 crc kubenswrapper[4988]: I1008 18:31:40.682831 4988 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/fa5061f0-e7f7-4059-9f69-48d6a1cb98d7-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 08 18:31:40 crc kubenswrapper[4988]: I1008 18:31:40.682845 4988 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jcxnj\" (UniqueName: \"kubernetes.io/projected/fa5061f0-e7f7-4059-9f69-48d6a1cb98d7-kube-api-access-jcxnj\") on node \"crc\" DevicePath \"\"" Oct 08 18:31:40 crc kubenswrapper[4988]: I1008 18:31:40.682859 4988 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fa5061f0-e7f7-4059-9f69-48d6a1cb98d7-scripts\") on node \"crc\" DevicePath \"\"" Oct 08 18:31:40 crc kubenswrapper[4988]: I1008 18:31:40.706030 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fa5061f0-e7f7-4059-9f69-48d6a1cb98d7-config-data" (OuterVolumeSpecName: "config-data") pod "fa5061f0-e7f7-4059-9f69-48d6a1cb98d7" (UID: "fa5061f0-e7f7-4059-9f69-48d6a1cb98d7"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:31:40 crc kubenswrapper[4988]: I1008 18:31:40.784789 4988 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fa5061f0-e7f7-4059-9f69-48d6a1cb98d7-config-data\") on node \"crc\" DevicePath \"\"" Oct 08 18:31:41 crc kubenswrapper[4988]: I1008 18:31:41.057731 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"fa5061f0-e7f7-4059-9f69-48d6a1cb98d7","Type":"ContainerDied","Data":"cef716ddb91f513c196627b27d20f52e59e2523d565d59958302795fbca835a6"} Oct 08 18:31:41 crc kubenswrapper[4988]: I1008 18:31:41.057804 4988 scope.go:117] "RemoveContainer" containerID="7f1af50932c5ecb7086c67d8b967a7c5d6ed5ac407c3ca0a5e4d0a729102c747" Oct 08 18:31:41 crc kubenswrapper[4988]: I1008 18:31:41.057982 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 08 18:31:41 crc kubenswrapper[4988]: I1008 18:31:41.081415 4988 scope.go:117] "RemoveContainer" containerID="e19138f089a607cc5db443a8dfe9a62efdb451bc27329b434278497a10f97446" Oct 08 18:31:41 crc kubenswrapper[4988]: I1008 18:31:41.100105 4988 scope.go:117] "RemoveContainer" containerID="4edcb300385ece56d49e71ae3fa0161264857a94d80352e9f4dfc1128294a789" Oct 08 18:31:41 crc kubenswrapper[4988]: I1008 18:31:41.132427 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 08 18:31:41 crc kubenswrapper[4988]: I1008 18:31:41.139461 4988 scope.go:117] "RemoveContainer" containerID="4abb52d2340f790b7dbc4c91a708a887af9a63619f61ed6defbeb30cce9e793b" Oct 08 18:31:41 crc kubenswrapper[4988]: I1008 18:31:41.158562 4988 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Oct 08 18:31:41 crc kubenswrapper[4988]: I1008 18:31:41.166436 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 08 18:31:41 crc kubenswrapper[4988]: E1008 18:31:41.166801 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fa5061f0-e7f7-4059-9f69-48d6a1cb98d7" containerName="sg-core" Oct 08 18:31:41 crc kubenswrapper[4988]: I1008 18:31:41.166817 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="fa5061f0-e7f7-4059-9f69-48d6a1cb98d7" containerName="sg-core" Oct 08 18:31:41 crc kubenswrapper[4988]: E1008 18:31:41.166831 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fa5061f0-e7f7-4059-9f69-48d6a1cb98d7" containerName="proxy-httpd" Oct 08 18:31:41 crc kubenswrapper[4988]: I1008 18:31:41.166839 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="fa5061f0-e7f7-4059-9f69-48d6a1cb98d7" containerName="proxy-httpd" Oct 08 18:31:41 crc kubenswrapper[4988]: E1008 18:31:41.166857 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fa5061f0-e7f7-4059-9f69-48d6a1cb98d7" containerName="ceilometer-notification-agent" Oct 08 18:31:41 crc kubenswrapper[4988]: I1008 18:31:41.166863 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="fa5061f0-e7f7-4059-9f69-48d6a1cb98d7" containerName="ceilometer-notification-agent" Oct 08 18:31:41 crc kubenswrapper[4988]: E1008 18:31:41.166878 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fa5061f0-e7f7-4059-9f69-48d6a1cb98d7" containerName="ceilometer-central-agent" Oct 08 18:31:41 crc kubenswrapper[4988]: I1008 18:31:41.166883 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="fa5061f0-e7f7-4059-9f69-48d6a1cb98d7" containerName="ceilometer-central-agent" Oct 08 18:31:41 crc kubenswrapper[4988]: I1008 18:31:41.167043 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="fa5061f0-e7f7-4059-9f69-48d6a1cb98d7" containerName="sg-core" Oct 08 18:31:41 crc kubenswrapper[4988]: I1008 18:31:41.167057 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="fa5061f0-e7f7-4059-9f69-48d6a1cb98d7" containerName="ceilometer-central-agent" Oct 08 18:31:41 crc kubenswrapper[4988]: I1008 18:31:41.167074 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="fa5061f0-e7f7-4059-9f69-48d6a1cb98d7" containerName="proxy-httpd" Oct 08 18:31:41 crc kubenswrapper[4988]: I1008 18:31:41.167085 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="fa5061f0-e7f7-4059-9f69-48d6a1cb98d7" containerName="ceilometer-notification-agent" Oct 08 18:31:41 crc kubenswrapper[4988]: I1008 18:31:41.168694 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 08 18:31:41 crc kubenswrapper[4988]: I1008 18:31:41.174088 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Oct 08 18:31:41 crc kubenswrapper[4988]: I1008 18:31:41.174301 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 08 18:31:41 crc kubenswrapper[4988]: I1008 18:31:41.177516 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 08 18:31:41 crc kubenswrapper[4988]: I1008 18:31:41.179429 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 08 18:31:41 crc kubenswrapper[4988]: I1008 18:31:41.246904 4988 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fa5061f0-e7f7-4059-9f69-48d6a1cb98d7" path="/var/lib/kubelet/pods/fa5061f0-e7f7-4059-9f69-48d6a1cb98d7/volumes" Oct 08 18:31:41 crc kubenswrapper[4988]: I1008 18:31:41.292872 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2df23cd3-cc55-48df-84f1-1b244b1bff09-log-httpd\") pod \"ceilometer-0\" (UID: \"2df23cd3-cc55-48df-84f1-1b244b1bff09\") " pod="openstack/ceilometer-0" Oct 08 18:31:41 crc kubenswrapper[4988]: I1008 18:31:41.292942 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2df23cd3-cc55-48df-84f1-1b244b1bff09-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"2df23cd3-cc55-48df-84f1-1b244b1bff09\") " pod="openstack/ceilometer-0" Oct 08 18:31:41 crc kubenswrapper[4988]: I1008 18:31:41.292992 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2df23cd3-cc55-48df-84f1-1b244b1bff09-scripts\") pod \"ceilometer-0\" (UID: \"2df23cd3-cc55-48df-84f1-1b244b1bff09\") " pod="openstack/ceilometer-0" Oct 08 18:31:41 crc kubenswrapper[4988]: I1008 18:31:41.293306 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/2df23cd3-cc55-48df-84f1-1b244b1bff09-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"2df23cd3-cc55-48df-84f1-1b244b1bff09\") " pod="openstack/ceilometer-0" Oct 08 18:31:41 crc kubenswrapper[4988]: I1008 18:31:41.293344 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2df23cd3-cc55-48df-84f1-1b244b1bff09-config-data\") pod \"ceilometer-0\" (UID: \"2df23cd3-cc55-48df-84f1-1b244b1bff09\") " pod="openstack/ceilometer-0" Oct 08 18:31:41 crc kubenswrapper[4988]: I1008 18:31:41.293461 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v8dbl\" (UniqueName: \"kubernetes.io/projected/2df23cd3-cc55-48df-84f1-1b244b1bff09-kube-api-access-v8dbl\") pod \"ceilometer-0\" (UID: \"2df23cd3-cc55-48df-84f1-1b244b1bff09\") " pod="openstack/ceilometer-0" Oct 08 18:31:41 crc kubenswrapper[4988]: I1008 18:31:41.293482 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2df23cd3-cc55-48df-84f1-1b244b1bff09-run-httpd\") pod \"ceilometer-0\" (UID: \"2df23cd3-cc55-48df-84f1-1b244b1bff09\") " pod="openstack/ceilometer-0" Oct 08 18:31:41 crc kubenswrapper[4988]: I1008 18:31:41.293497 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/2df23cd3-cc55-48df-84f1-1b244b1bff09-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"2df23cd3-cc55-48df-84f1-1b244b1bff09\") " pod="openstack/ceilometer-0" Oct 08 18:31:41 crc kubenswrapper[4988]: I1008 18:31:41.395087 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/2df23cd3-cc55-48df-84f1-1b244b1bff09-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"2df23cd3-cc55-48df-84f1-1b244b1bff09\") " pod="openstack/ceilometer-0" Oct 08 18:31:41 crc kubenswrapper[4988]: I1008 18:31:41.395177 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2df23cd3-cc55-48df-84f1-1b244b1bff09-config-data\") pod \"ceilometer-0\" (UID: \"2df23cd3-cc55-48df-84f1-1b244b1bff09\") " pod="openstack/ceilometer-0" Oct 08 18:31:41 crc kubenswrapper[4988]: I1008 18:31:41.395277 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v8dbl\" (UniqueName: \"kubernetes.io/projected/2df23cd3-cc55-48df-84f1-1b244b1bff09-kube-api-access-v8dbl\") pod \"ceilometer-0\" (UID: \"2df23cd3-cc55-48df-84f1-1b244b1bff09\") " pod="openstack/ceilometer-0" Oct 08 18:31:41 crc kubenswrapper[4988]: I1008 18:31:41.395295 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2df23cd3-cc55-48df-84f1-1b244b1bff09-run-httpd\") pod \"ceilometer-0\" (UID: \"2df23cd3-cc55-48df-84f1-1b244b1bff09\") " pod="openstack/ceilometer-0" Oct 08 18:31:41 crc kubenswrapper[4988]: I1008 18:31:41.395313 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/2df23cd3-cc55-48df-84f1-1b244b1bff09-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"2df23cd3-cc55-48df-84f1-1b244b1bff09\") " pod="openstack/ceilometer-0" Oct 08 18:31:41 crc kubenswrapper[4988]: I1008 18:31:41.395375 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2df23cd3-cc55-48df-84f1-1b244b1bff09-log-httpd\") pod \"ceilometer-0\" (UID: \"2df23cd3-cc55-48df-84f1-1b244b1bff09\") " pod="openstack/ceilometer-0" Oct 08 18:31:41 crc kubenswrapper[4988]: I1008 18:31:41.395418 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2df23cd3-cc55-48df-84f1-1b244b1bff09-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"2df23cd3-cc55-48df-84f1-1b244b1bff09\") " pod="openstack/ceilometer-0" Oct 08 18:31:41 crc kubenswrapper[4988]: I1008 18:31:41.395467 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2df23cd3-cc55-48df-84f1-1b244b1bff09-scripts\") pod \"ceilometer-0\" (UID: \"2df23cd3-cc55-48df-84f1-1b244b1bff09\") " pod="openstack/ceilometer-0" Oct 08 18:31:41 crc kubenswrapper[4988]: I1008 18:31:41.396096 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2df23cd3-cc55-48df-84f1-1b244b1bff09-run-httpd\") pod \"ceilometer-0\" (UID: \"2df23cd3-cc55-48df-84f1-1b244b1bff09\") " pod="openstack/ceilometer-0" Oct 08 18:31:41 crc kubenswrapper[4988]: I1008 18:31:41.396748 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2df23cd3-cc55-48df-84f1-1b244b1bff09-log-httpd\") pod \"ceilometer-0\" (UID: \"2df23cd3-cc55-48df-84f1-1b244b1bff09\") " pod="openstack/ceilometer-0" Oct 08 18:31:41 crc kubenswrapper[4988]: I1008 18:31:41.400025 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/2df23cd3-cc55-48df-84f1-1b244b1bff09-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"2df23cd3-cc55-48df-84f1-1b244b1bff09\") " pod="openstack/ceilometer-0" Oct 08 18:31:41 crc kubenswrapper[4988]: I1008 18:31:41.400351 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/2df23cd3-cc55-48df-84f1-1b244b1bff09-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"2df23cd3-cc55-48df-84f1-1b244b1bff09\") " pod="openstack/ceilometer-0" Oct 08 18:31:41 crc kubenswrapper[4988]: I1008 18:31:41.401019 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2df23cd3-cc55-48df-84f1-1b244b1bff09-config-data\") pod \"ceilometer-0\" (UID: \"2df23cd3-cc55-48df-84f1-1b244b1bff09\") " pod="openstack/ceilometer-0" Oct 08 18:31:41 crc kubenswrapper[4988]: I1008 18:31:41.405808 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2df23cd3-cc55-48df-84f1-1b244b1bff09-scripts\") pod \"ceilometer-0\" (UID: \"2df23cd3-cc55-48df-84f1-1b244b1bff09\") " pod="openstack/ceilometer-0" Oct 08 18:31:41 crc kubenswrapper[4988]: I1008 18:31:41.406588 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2df23cd3-cc55-48df-84f1-1b244b1bff09-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"2df23cd3-cc55-48df-84f1-1b244b1bff09\") " pod="openstack/ceilometer-0" Oct 08 18:31:41 crc kubenswrapper[4988]: I1008 18:31:41.425844 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v8dbl\" (UniqueName: \"kubernetes.io/projected/2df23cd3-cc55-48df-84f1-1b244b1bff09-kube-api-access-v8dbl\") pod \"ceilometer-0\" (UID: \"2df23cd3-cc55-48df-84f1-1b244b1bff09\") " pod="openstack/ceilometer-0" Oct 08 18:31:41 crc kubenswrapper[4988]: I1008 18:31:41.490121 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 08 18:31:41 crc kubenswrapper[4988]: I1008 18:31:41.908922 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 08 18:31:42 crc kubenswrapper[4988]: I1008 18:31:42.068033 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2df23cd3-cc55-48df-84f1-1b244b1bff09","Type":"ContainerStarted","Data":"54d3ca04da26b716bb45abe2dafc73b6daa59f7d5b9464e5ba249135d0085066"} Oct 08 18:31:43 crc kubenswrapper[4988]: I1008 18:31:43.081246 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2df23cd3-cc55-48df-84f1-1b244b1bff09","Type":"ContainerStarted","Data":"d397074a7b71070cc40ef8027e42e3ecd64f50a2047c9dea24791be7fb82d6ca"} Oct 08 18:31:44 crc kubenswrapper[4988]: I1008 18:31:44.090588 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2df23cd3-cc55-48df-84f1-1b244b1bff09","Type":"ContainerStarted","Data":"d82c51909c3813179ef9e4c9077f8b2443df11cf73f43f8967bbc410e11bb01b"} Oct 08 18:31:44 crc kubenswrapper[4988]: I1008 18:31:44.373273 4988 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Oct 08 18:31:44 crc kubenswrapper[4988]: I1008 18:31:44.385132 4988 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 08 18:31:44 crc kubenswrapper[4988]: I1008 18:31:44.385179 4988 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 08 18:31:44 crc kubenswrapper[4988]: I1008 18:31:44.425207 4988 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Oct 08 18:31:45 crc kubenswrapper[4988]: I1008 18:31:45.103484 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2df23cd3-cc55-48df-84f1-1b244b1bff09","Type":"ContainerStarted","Data":"e3574329636b8c9e97dcdb6530a126abbbb07009c757689a079cb8463fc1f56e"} Oct 08 18:31:45 crc kubenswrapper[4988]: I1008 18:31:45.137743 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Oct 08 18:31:45 crc kubenswrapper[4988]: I1008 18:31:45.473755 4988 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="6c8adde4-41e0-4d33-961a-17a76088b3b8" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.0.194:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 08 18:31:45 crc kubenswrapper[4988]: I1008 18:31:45.474451 4988 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="6c8adde4-41e0-4d33-961a-17a76088b3b8" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.0.194:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 08 18:31:47 crc kubenswrapper[4988]: I1008 18:31:47.124423 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2df23cd3-cc55-48df-84f1-1b244b1bff09","Type":"ContainerStarted","Data":"fd6ebabdffd1c5a641c7fe0ca4540b07bcf6bfc6a284e37c9305dfd3e60cffa1"} Oct 08 18:31:47 crc kubenswrapper[4988]: I1008 18:31:47.126539 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 08 18:31:47 crc kubenswrapper[4988]: I1008 18:31:47.164912 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=1.851360501 podStartE2EDuration="6.164894265s" podCreationTimestamp="2025-10-08 18:31:41 +0000 UTC" firstStartedPulling="2025-10-08 18:31:41.914225562 +0000 UTC m=+1247.364068332" lastFinishedPulling="2025-10-08 18:31:46.227759326 +0000 UTC m=+1251.677602096" observedRunningTime="2025-10-08 18:31:47.149269285 +0000 UTC m=+1252.599112075" watchObservedRunningTime="2025-10-08 18:31:47.164894265 +0000 UTC m=+1252.614737035" Oct 08 18:31:47 crc kubenswrapper[4988]: I1008 18:31:47.400886 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Oct 08 18:31:48 crc kubenswrapper[4988]: I1008 18:31:48.379001 4988 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Oct 08 18:31:48 crc kubenswrapper[4988]: I1008 18:31:48.385493 4988 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Oct 08 18:31:48 crc kubenswrapper[4988]: I1008 18:31:48.395367 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Oct 08 18:31:49 crc kubenswrapper[4988]: I1008 18:31:49.148314 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Oct 08 18:31:52 crc kubenswrapper[4988]: I1008 18:31:52.088908 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Oct 08 18:31:52 crc kubenswrapper[4988]: I1008 18:31:52.172523 4988 generic.go:334] "Generic (PLEG): container finished" podID="110929bd-ec2a-4f95-8cb1-ace9284ec881" containerID="31ef7ae1b9b13e2a236ac5de546a23be6667e216cee78fec4f9e3fb4e96f6183" exitCode=137 Oct 08 18:31:52 crc kubenswrapper[4988]: I1008 18:31:52.172604 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Oct 08 18:31:52 crc kubenswrapper[4988]: I1008 18:31:52.172629 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"110929bd-ec2a-4f95-8cb1-ace9284ec881","Type":"ContainerDied","Data":"31ef7ae1b9b13e2a236ac5de546a23be6667e216cee78fec4f9e3fb4e96f6183"} Oct 08 18:31:52 crc kubenswrapper[4988]: I1008 18:31:52.173798 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"110929bd-ec2a-4f95-8cb1-ace9284ec881","Type":"ContainerDied","Data":"d4d78f18f9b8f0e83ea0c0c7fb72ff91fbf5ac5068d047c2f09df2854eef18d9"} Oct 08 18:31:52 crc kubenswrapper[4988]: I1008 18:31:52.173916 4988 scope.go:117] "RemoveContainer" containerID="31ef7ae1b9b13e2a236ac5de546a23be6667e216cee78fec4f9e3fb4e96f6183" Oct 08 18:31:52 crc kubenswrapper[4988]: I1008 18:31:52.195075 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/110929bd-ec2a-4f95-8cb1-ace9284ec881-combined-ca-bundle\") pod \"110929bd-ec2a-4f95-8cb1-ace9284ec881\" (UID: \"110929bd-ec2a-4f95-8cb1-ace9284ec881\") " Oct 08 18:31:52 crc kubenswrapper[4988]: I1008 18:31:52.195335 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/110929bd-ec2a-4f95-8cb1-ace9284ec881-config-data\") pod \"110929bd-ec2a-4f95-8cb1-ace9284ec881\" (UID: \"110929bd-ec2a-4f95-8cb1-ace9284ec881\") " Oct 08 18:31:52 crc kubenswrapper[4988]: I1008 18:31:52.195399 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lv7tb\" (UniqueName: \"kubernetes.io/projected/110929bd-ec2a-4f95-8cb1-ace9284ec881-kube-api-access-lv7tb\") pod \"110929bd-ec2a-4f95-8cb1-ace9284ec881\" (UID: \"110929bd-ec2a-4f95-8cb1-ace9284ec881\") " Oct 08 18:31:52 crc kubenswrapper[4988]: I1008 18:31:52.216493 4988 scope.go:117] "RemoveContainer" containerID="31ef7ae1b9b13e2a236ac5de546a23be6667e216cee78fec4f9e3fb4e96f6183" Oct 08 18:31:52 crc kubenswrapper[4988]: E1008 18:31:52.218242 4988 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"31ef7ae1b9b13e2a236ac5de546a23be6667e216cee78fec4f9e3fb4e96f6183\": container with ID starting with 31ef7ae1b9b13e2a236ac5de546a23be6667e216cee78fec4f9e3fb4e96f6183 not found: ID does not exist" containerID="31ef7ae1b9b13e2a236ac5de546a23be6667e216cee78fec4f9e3fb4e96f6183" Oct 08 18:31:52 crc kubenswrapper[4988]: I1008 18:31:52.218272 4988 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"31ef7ae1b9b13e2a236ac5de546a23be6667e216cee78fec4f9e3fb4e96f6183"} err="failed to get container status \"31ef7ae1b9b13e2a236ac5de546a23be6667e216cee78fec4f9e3fb4e96f6183\": rpc error: code = NotFound desc = could not find container \"31ef7ae1b9b13e2a236ac5de546a23be6667e216cee78fec4f9e3fb4e96f6183\": container with ID starting with 31ef7ae1b9b13e2a236ac5de546a23be6667e216cee78fec4f9e3fb4e96f6183 not found: ID does not exist" Oct 08 18:31:52 crc kubenswrapper[4988]: I1008 18:31:52.218509 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/110929bd-ec2a-4f95-8cb1-ace9284ec881-kube-api-access-lv7tb" (OuterVolumeSpecName: "kube-api-access-lv7tb") pod "110929bd-ec2a-4f95-8cb1-ace9284ec881" (UID: "110929bd-ec2a-4f95-8cb1-ace9284ec881"). InnerVolumeSpecName "kube-api-access-lv7tb". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:31:52 crc kubenswrapper[4988]: I1008 18:31:52.230054 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/110929bd-ec2a-4f95-8cb1-ace9284ec881-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "110929bd-ec2a-4f95-8cb1-ace9284ec881" (UID: "110929bd-ec2a-4f95-8cb1-ace9284ec881"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:31:52 crc kubenswrapper[4988]: I1008 18:31:52.236692 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/110929bd-ec2a-4f95-8cb1-ace9284ec881-config-data" (OuterVolumeSpecName: "config-data") pod "110929bd-ec2a-4f95-8cb1-ace9284ec881" (UID: "110929bd-ec2a-4f95-8cb1-ace9284ec881"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:31:52 crc kubenswrapper[4988]: I1008 18:31:52.296909 4988 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/110929bd-ec2a-4f95-8cb1-ace9284ec881-config-data\") on node \"crc\" DevicePath \"\"" Oct 08 18:31:52 crc kubenswrapper[4988]: I1008 18:31:52.296966 4988 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lv7tb\" (UniqueName: \"kubernetes.io/projected/110929bd-ec2a-4f95-8cb1-ace9284ec881-kube-api-access-lv7tb\") on node \"crc\" DevicePath \"\"" Oct 08 18:31:52 crc kubenswrapper[4988]: I1008 18:31:52.296979 4988 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/110929bd-ec2a-4f95-8cb1-ace9284ec881-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 18:31:52 crc kubenswrapper[4988]: I1008 18:31:52.511556 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 08 18:31:52 crc kubenswrapper[4988]: I1008 18:31:52.521525 4988 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 08 18:31:52 crc kubenswrapper[4988]: I1008 18:31:52.532131 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 08 18:31:52 crc kubenswrapper[4988]: E1008 18:31:52.532585 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="110929bd-ec2a-4f95-8cb1-ace9284ec881" containerName="nova-cell1-novncproxy-novncproxy" Oct 08 18:31:52 crc kubenswrapper[4988]: I1008 18:31:52.532600 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="110929bd-ec2a-4f95-8cb1-ace9284ec881" containerName="nova-cell1-novncproxy-novncproxy" Oct 08 18:31:52 crc kubenswrapper[4988]: I1008 18:31:52.532827 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="110929bd-ec2a-4f95-8cb1-ace9284ec881" containerName="nova-cell1-novncproxy-novncproxy" Oct 08 18:31:52 crc kubenswrapper[4988]: I1008 18:31:52.533607 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Oct 08 18:31:52 crc kubenswrapper[4988]: I1008 18:31:52.536555 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-novncproxy-cell1-public-svc" Oct 08 18:31:52 crc kubenswrapper[4988]: I1008 18:31:52.536786 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-novncproxy-cell1-vencrypt" Oct 08 18:31:52 crc kubenswrapper[4988]: I1008 18:31:52.542121 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Oct 08 18:31:52 crc kubenswrapper[4988]: I1008 18:31:52.544962 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 08 18:31:52 crc kubenswrapper[4988]: I1008 18:31:52.607929 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/816b358d-4428-4fae-b91c-03c0171faa23-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"816b358d-4428-4fae-b91c-03c0171faa23\") " pod="openstack/nova-cell1-novncproxy-0" Oct 08 18:31:52 crc kubenswrapper[4988]: I1008 18:31:52.608218 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/816b358d-4428-4fae-b91c-03c0171faa23-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"816b358d-4428-4fae-b91c-03c0171faa23\") " pod="openstack/nova-cell1-novncproxy-0" Oct 08 18:31:52 crc kubenswrapper[4988]: I1008 18:31:52.608343 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/816b358d-4428-4fae-b91c-03c0171faa23-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"816b358d-4428-4fae-b91c-03c0171faa23\") " pod="openstack/nova-cell1-novncproxy-0" Oct 08 18:31:52 crc kubenswrapper[4988]: I1008 18:31:52.608502 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9d54x\" (UniqueName: \"kubernetes.io/projected/816b358d-4428-4fae-b91c-03c0171faa23-kube-api-access-9d54x\") pod \"nova-cell1-novncproxy-0\" (UID: \"816b358d-4428-4fae-b91c-03c0171faa23\") " pod="openstack/nova-cell1-novncproxy-0" Oct 08 18:31:52 crc kubenswrapper[4988]: I1008 18:31:52.608671 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/816b358d-4428-4fae-b91c-03c0171faa23-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"816b358d-4428-4fae-b91c-03c0171faa23\") " pod="openstack/nova-cell1-novncproxy-0" Oct 08 18:31:52 crc kubenswrapper[4988]: I1008 18:31:52.710417 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/816b358d-4428-4fae-b91c-03c0171faa23-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"816b358d-4428-4fae-b91c-03c0171faa23\") " pod="openstack/nova-cell1-novncproxy-0" Oct 08 18:31:52 crc kubenswrapper[4988]: I1008 18:31:52.710498 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/816b358d-4428-4fae-b91c-03c0171faa23-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"816b358d-4428-4fae-b91c-03c0171faa23\") " pod="openstack/nova-cell1-novncproxy-0" Oct 08 18:31:52 crc kubenswrapper[4988]: I1008 18:31:52.710531 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/816b358d-4428-4fae-b91c-03c0171faa23-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"816b358d-4428-4fae-b91c-03c0171faa23\") " pod="openstack/nova-cell1-novncproxy-0" Oct 08 18:31:52 crc kubenswrapper[4988]: I1008 18:31:52.710558 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/816b358d-4428-4fae-b91c-03c0171faa23-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"816b358d-4428-4fae-b91c-03c0171faa23\") " pod="openstack/nova-cell1-novncproxy-0" Oct 08 18:31:52 crc kubenswrapper[4988]: I1008 18:31:52.710592 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9d54x\" (UniqueName: \"kubernetes.io/projected/816b358d-4428-4fae-b91c-03c0171faa23-kube-api-access-9d54x\") pod \"nova-cell1-novncproxy-0\" (UID: \"816b358d-4428-4fae-b91c-03c0171faa23\") " pod="openstack/nova-cell1-novncproxy-0" Oct 08 18:31:52 crc kubenswrapper[4988]: I1008 18:31:52.714943 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/816b358d-4428-4fae-b91c-03c0171faa23-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"816b358d-4428-4fae-b91c-03c0171faa23\") " pod="openstack/nova-cell1-novncproxy-0" Oct 08 18:31:52 crc kubenswrapper[4988]: I1008 18:31:52.714954 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/816b358d-4428-4fae-b91c-03c0171faa23-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"816b358d-4428-4fae-b91c-03c0171faa23\") " pod="openstack/nova-cell1-novncproxy-0" Oct 08 18:31:52 crc kubenswrapper[4988]: I1008 18:31:52.716153 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/816b358d-4428-4fae-b91c-03c0171faa23-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"816b358d-4428-4fae-b91c-03c0171faa23\") " pod="openstack/nova-cell1-novncproxy-0" Oct 08 18:31:52 crc kubenswrapper[4988]: I1008 18:31:52.716427 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/816b358d-4428-4fae-b91c-03c0171faa23-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"816b358d-4428-4fae-b91c-03c0171faa23\") " pod="openstack/nova-cell1-novncproxy-0" Oct 08 18:31:52 crc kubenswrapper[4988]: I1008 18:31:52.728488 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9d54x\" (UniqueName: \"kubernetes.io/projected/816b358d-4428-4fae-b91c-03c0171faa23-kube-api-access-9d54x\") pod \"nova-cell1-novncproxy-0\" (UID: \"816b358d-4428-4fae-b91c-03c0171faa23\") " pod="openstack/nova-cell1-novncproxy-0" Oct 08 18:31:52 crc kubenswrapper[4988]: I1008 18:31:52.851832 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Oct 08 18:31:53 crc kubenswrapper[4988]: I1008 18:31:53.247308 4988 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="110929bd-ec2a-4f95-8cb1-ace9284ec881" path="/var/lib/kubelet/pods/110929bd-ec2a-4f95-8cb1-ace9284ec881/volumes" Oct 08 18:31:53 crc kubenswrapper[4988]: I1008 18:31:53.335925 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 08 18:31:54 crc kubenswrapper[4988]: I1008 18:31:54.196202 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"816b358d-4428-4fae-b91c-03c0171faa23","Type":"ContainerStarted","Data":"a947f771405cb73547e24443c664920da93801df8ff9dd6369c101deea715564"} Oct 08 18:31:54 crc kubenswrapper[4988]: I1008 18:31:54.196517 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"816b358d-4428-4fae-b91c-03c0171faa23","Type":"ContainerStarted","Data":"8a452108aab53fec7dc4838c8537c45f71851137a3ecea9fd46ecab51a3db81f"} Oct 08 18:31:54 crc kubenswrapper[4988]: I1008 18:31:54.223450 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=2.22342806 podStartE2EDuration="2.22342806s" podCreationTimestamp="2025-10-08 18:31:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 18:31:54.215189036 +0000 UTC m=+1259.665031816" watchObservedRunningTime="2025-10-08 18:31:54.22342806 +0000 UTC m=+1259.673270870" Oct 08 18:31:54 crc kubenswrapper[4988]: I1008 18:31:54.421239 4988 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Oct 08 18:31:54 crc kubenswrapper[4988]: I1008 18:31:54.422098 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Oct 08 18:31:54 crc kubenswrapper[4988]: I1008 18:31:54.426090 4988 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Oct 08 18:31:54 crc kubenswrapper[4988]: I1008 18:31:54.433843 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Oct 08 18:31:55 crc kubenswrapper[4988]: I1008 18:31:55.224058 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Oct 08 18:31:55 crc kubenswrapper[4988]: I1008 18:31:55.254087 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Oct 08 18:31:55 crc kubenswrapper[4988]: I1008 18:31:55.431614 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5967cc9597-r6jrv"] Oct 08 18:31:55 crc kubenswrapper[4988]: I1008 18:31:55.433027 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5967cc9597-r6jrv" Oct 08 18:31:55 crc kubenswrapper[4988]: I1008 18:31:55.460670 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5967cc9597-r6jrv"] Oct 08 18:31:55 crc kubenswrapper[4988]: I1008 18:31:55.568800 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/467e736b-562a-4f31-af64-2fce55e965c8-dns-svc\") pod \"dnsmasq-dns-5967cc9597-r6jrv\" (UID: \"467e736b-562a-4f31-af64-2fce55e965c8\") " pod="openstack/dnsmasq-dns-5967cc9597-r6jrv" Oct 08 18:31:55 crc kubenswrapper[4988]: I1008 18:31:55.568894 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/467e736b-562a-4f31-af64-2fce55e965c8-config\") pod \"dnsmasq-dns-5967cc9597-r6jrv\" (UID: \"467e736b-562a-4f31-af64-2fce55e965c8\") " pod="openstack/dnsmasq-dns-5967cc9597-r6jrv" Oct 08 18:31:55 crc kubenswrapper[4988]: I1008 18:31:55.568944 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/467e736b-562a-4f31-af64-2fce55e965c8-dns-swift-storage-0\") pod \"dnsmasq-dns-5967cc9597-r6jrv\" (UID: \"467e736b-562a-4f31-af64-2fce55e965c8\") " pod="openstack/dnsmasq-dns-5967cc9597-r6jrv" Oct 08 18:31:55 crc kubenswrapper[4988]: I1008 18:31:55.568966 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/467e736b-562a-4f31-af64-2fce55e965c8-ovsdbserver-sb\") pod \"dnsmasq-dns-5967cc9597-r6jrv\" (UID: \"467e736b-562a-4f31-af64-2fce55e965c8\") " pod="openstack/dnsmasq-dns-5967cc9597-r6jrv" Oct 08 18:31:55 crc kubenswrapper[4988]: I1008 18:31:55.569039 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m8nnt\" (UniqueName: \"kubernetes.io/projected/467e736b-562a-4f31-af64-2fce55e965c8-kube-api-access-m8nnt\") pod \"dnsmasq-dns-5967cc9597-r6jrv\" (UID: \"467e736b-562a-4f31-af64-2fce55e965c8\") " pod="openstack/dnsmasq-dns-5967cc9597-r6jrv" Oct 08 18:31:55 crc kubenswrapper[4988]: I1008 18:31:55.569074 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/467e736b-562a-4f31-af64-2fce55e965c8-ovsdbserver-nb\") pod \"dnsmasq-dns-5967cc9597-r6jrv\" (UID: \"467e736b-562a-4f31-af64-2fce55e965c8\") " pod="openstack/dnsmasq-dns-5967cc9597-r6jrv" Oct 08 18:31:55 crc kubenswrapper[4988]: I1008 18:31:55.676092 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/467e736b-562a-4f31-af64-2fce55e965c8-config\") pod \"dnsmasq-dns-5967cc9597-r6jrv\" (UID: \"467e736b-562a-4f31-af64-2fce55e965c8\") " pod="openstack/dnsmasq-dns-5967cc9597-r6jrv" Oct 08 18:31:55 crc kubenswrapper[4988]: I1008 18:31:55.676344 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/467e736b-562a-4f31-af64-2fce55e965c8-dns-swift-storage-0\") pod \"dnsmasq-dns-5967cc9597-r6jrv\" (UID: \"467e736b-562a-4f31-af64-2fce55e965c8\") " pod="openstack/dnsmasq-dns-5967cc9597-r6jrv" Oct 08 18:31:55 crc kubenswrapper[4988]: I1008 18:31:55.676378 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/467e736b-562a-4f31-af64-2fce55e965c8-ovsdbserver-sb\") pod \"dnsmasq-dns-5967cc9597-r6jrv\" (UID: \"467e736b-562a-4f31-af64-2fce55e965c8\") " pod="openstack/dnsmasq-dns-5967cc9597-r6jrv" Oct 08 18:31:55 crc kubenswrapper[4988]: I1008 18:31:55.676649 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m8nnt\" (UniqueName: \"kubernetes.io/projected/467e736b-562a-4f31-af64-2fce55e965c8-kube-api-access-m8nnt\") pod \"dnsmasq-dns-5967cc9597-r6jrv\" (UID: \"467e736b-562a-4f31-af64-2fce55e965c8\") " pod="openstack/dnsmasq-dns-5967cc9597-r6jrv" Oct 08 18:31:55 crc kubenswrapper[4988]: I1008 18:31:55.676783 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/467e736b-562a-4f31-af64-2fce55e965c8-ovsdbserver-nb\") pod \"dnsmasq-dns-5967cc9597-r6jrv\" (UID: \"467e736b-562a-4f31-af64-2fce55e965c8\") " pod="openstack/dnsmasq-dns-5967cc9597-r6jrv" Oct 08 18:31:55 crc kubenswrapper[4988]: I1008 18:31:55.676908 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/467e736b-562a-4f31-af64-2fce55e965c8-dns-svc\") pod \"dnsmasq-dns-5967cc9597-r6jrv\" (UID: \"467e736b-562a-4f31-af64-2fce55e965c8\") " pod="openstack/dnsmasq-dns-5967cc9597-r6jrv" Oct 08 18:31:55 crc kubenswrapper[4988]: I1008 18:31:55.677438 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/467e736b-562a-4f31-af64-2fce55e965c8-config\") pod \"dnsmasq-dns-5967cc9597-r6jrv\" (UID: \"467e736b-562a-4f31-af64-2fce55e965c8\") " pod="openstack/dnsmasq-dns-5967cc9597-r6jrv" Oct 08 18:31:55 crc kubenswrapper[4988]: I1008 18:31:55.677483 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/467e736b-562a-4f31-af64-2fce55e965c8-dns-swift-storage-0\") pod \"dnsmasq-dns-5967cc9597-r6jrv\" (UID: \"467e736b-562a-4f31-af64-2fce55e965c8\") " pod="openstack/dnsmasq-dns-5967cc9597-r6jrv" Oct 08 18:31:55 crc kubenswrapper[4988]: I1008 18:31:55.678775 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/467e736b-562a-4f31-af64-2fce55e965c8-ovsdbserver-sb\") pod \"dnsmasq-dns-5967cc9597-r6jrv\" (UID: \"467e736b-562a-4f31-af64-2fce55e965c8\") " pod="openstack/dnsmasq-dns-5967cc9597-r6jrv" Oct 08 18:31:55 crc kubenswrapper[4988]: I1008 18:31:55.681195 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/467e736b-562a-4f31-af64-2fce55e965c8-dns-svc\") pod \"dnsmasq-dns-5967cc9597-r6jrv\" (UID: \"467e736b-562a-4f31-af64-2fce55e965c8\") " pod="openstack/dnsmasq-dns-5967cc9597-r6jrv" Oct 08 18:31:55 crc kubenswrapper[4988]: I1008 18:31:55.681265 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/467e736b-562a-4f31-af64-2fce55e965c8-ovsdbserver-nb\") pod \"dnsmasq-dns-5967cc9597-r6jrv\" (UID: \"467e736b-562a-4f31-af64-2fce55e965c8\") " pod="openstack/dnsmasq-dns-5967cc9597-r6jrv" Oct 08 18:31:55 crc kubenswrapper[4988]: I1008 18:31:55.719645 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m8nnt\" (UniqueName: \"kubernetes.io/projected/467e736b-562a-4f31-af64-2fce55e965c8-kube-api-access-m8nnt\") pod \"dnsmasq-dns-5967cc9597-r6jrv\" (UID: \"467e736b-562a-4f31-af64-2fce55e965c8\") " pod="openstack/dnsmasq-dns-5967cc9597-r6jrv" Oct 08 18:31:55 crc kubenswrapper[4988]: I1008 18:31:55.756896 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5967cc9597-r6jrv" Oct 08 18:31:56 crc kubenswrapper[4988]: I1008 18:31:56.221594 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5967cc9597-r6jrv"] Oct 08 18:31:57 crc kubenswrapper[4988]: I1008 18:31:57.243942 4988 generic.go:334] "Generic (PLEG): container finished" podID="467e736b-562a-4f31-af64-2fce55e965c8" containerID="4cfd36bf140608a1fc5f7709dc1e3e6caea4bad2d51f7170f308befa5e5888fd" exitCode=0 Oct 08 18:31:57 crc kubenswrapper[4988]: I1008 18:31:57.247851 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5967cc9597-r6jrv" event={"ID":"467e736b-562a-4f31-af64-2fce55e965c8","Type":"ContainerDied","Data":"4cfd36bf140608a1fc5f7709dc1e3e6caea4bad2d51f7170f308befa5e5888fd"} Oct 08 18:31:57 crc kubenswrapper[4988]: I1008 18:31:57.247892 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5967cc9597-r6jrv" event={"ID":"467e736b-562a-4f31-af64-2fce55e965c8","Type":"ContainerStarted","Data":"d9b29ee2ea5ef3070f75c3d04f7fcb2db21a70a8b07adf0ecf69890c24328f79"} Oct 08 18:31:57 crc kubenswrapper[4988]: I1008 18:31:57.628398 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 08 18:31:57 crc kubenswrapper[4988]: I1008 18:31:57.629005 4988 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="2df23cd3-cc55-48df-84f1-1b244b1bff09" containerName="ceilometer-central-agent" containerID="cri-o://d397074a7b71070cc40ef8027e42e3ecd64f50a2047c9dea24791be7fb82d6ca" gracePeriod=30 Oct 08 18:31:57 crc kubenswrapper[4988]: I1008 18:31:57.629850 4988 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="2df23cd3-cc55-48df-84f1-1b244b1bff09" containerName="proxy-httpd" containerID="cri-o://fd6ebabdffd1c5a641c7fe0ca4540b07bcf6bfc6a284e37c9305dfd3e60cffa1" gracePeriod=30 Oct 08 18:31:57 crc kubenswrapper[4988]: I1008 18:31:57.629935 4988 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="2df23cd3-cc55-48df-84f1-1b244b1bff09" containerName="ceilometer-notification-agent" containerID="cri-o://d82c51909c3813179ef9e4c9077f8b2443df11cf73f43f8967bbc410e11bb01b" gracePeriod=30 Oct 08 18:31:57 crc kubenswrapper[4988]: I1008 18:31:57.630063 4988 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="2df23cd3-cc55-48df-84f1-1b244b1bff09" containerName="sg-core" containerID="cri-o://e3574329636b8c9e97dcdb6530a126abbbb07009c757689a079cb8463fc1f56e" gracePeriod=30 Oct 08 18:31:57 crc kubenswrapper[4988]: I1008 18:31:57.641581 4988 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ceilometer-0" podUID="2df23cd3-cc55-48df-84f1-1b244b1bff09" containerName="proxy-httpd" probeResult="failure" output="Get \"https://10.217.0.197:3000/\": EOF" Oct 08 18:31:57 crc kubenswrapper[4988]: I1008 18:31:57.852587 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Oct 08 18:31:58 crc kubenswrapper[4988]: I1008 18:31:58.081183 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 08 18:31:58 crc kubenswrapper[4988]: I1008 18:31:58.255439 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5967cc9597-r6jrv" event={"ID":"467e736b-562a-4f31-af64-2fce55e965c8","Type":"ContainerStarted","Data":"9933c83a6d7a8c9610edd4f70255f0f91bccecf1b14bb8fd883bcb6b9506666f"} Oct 08 18:31:58 crc kubenswrapper[4988]: I1008 18:31:58.255580 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-5967cc9597-r6jrv" Oct 08 18:31:58 crc kubenswrapper[4988]: I1008 18:31:58.262420 4988 generic.go:334] "Generic (PLEG): container finished" podID="2df23cd3-cc55-48df-84f1-1b244b1bff09" containerID="fd6ebabdffd1c5a641c7fe0ca4540b07bcf6bfc6a284e37c9305dfd3e60cffa1" exitCode=0 Oct 08 18:31:58 crc kubenswrapper[4988]: I1008 18:31:58.262448 4988 generic.go:334] "Generic (PLEG): container finished" podID="2df23cd3-cc55-48df-84f1-1b244b1bff09" containerID="e3574329636b8c9e97dcdb6530a126abbbb07009c757689a079cb8463fc1f56e" exitCode=2 Oct 08 18:31:58 crc kubenswrapper[4988]: I1008 18:31:58.262457 4988 generic.go:334] "Generic (PLEG): container finished" podID="2df23cd3-cc55-48df-84f1-1b244b1bff09" containerID="d397074a7b71070cc40ef8027e42e3ecd64f50a2047c9dea24791be7fb82d6ca" exitCode=0 Oct 08 18:31:58 crc kubenswrapper[4988]: I1008 18:31:58.262485 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2df23cd3-cc55-48df-84f1-1b244b1bff09","Type":"ContainerDied","Data":"fd6ebabdffd1c5a641c7fe0ca4540b07bcf6bfc6a284e37c9305dfd3e60cffa1"} Oct 08 18:31:58 crc kubenswrapper[4988]: I1008 18:31:58.262522 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2df23cd3-cc55-48df-84f1-1b244b1bff09","Type":"ContainerDied","Data":"e3574329636b8c9e97dcdb6530a126abbbb07009c757689a079cb8463fc1f56e"} Oct 08 18:31:58 crc kubenswrapper[4988]: I1008 18:31:58.262534 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2df23cd3-cc55-48df-84f1-1b244b1bff09","Type":"ContainerDied","Data":"d397074a7b71070cc40ef8027e42e3ecd64f50a2047c9dea24791be7fb82d6ca"} Oct 08 18:31:58 crc kubenswrapper[4988]: I1008 18:31:58.262613 4988 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="6c8adde4-41e0-4d33-961a-17a76088b3b8" containerName="nova-api-log" containerID="cri-o://8a29e42569215a39afa39fa848957fbb8c88d7c870e4e76ffcd20fafb98282df" gracePeriod=30 Oct 08 18:31:58 crc kubenswrapper[4988]: I1008 18:31:58.262704 4988 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="6c8adde4-41e0-4d33-961a-17a76088b3b8" containerName="nova-api-api" containerID="cri-o://56adcfbff76321bed6638d6eb70e3ed96f4f090daebaacfb04d44073983e39a8" gracePeriod=30 Oct 08 18:31:58 crc kubenswrapper[4988]: I1008 18:31:58.288486 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-5967cc9597-r6jrv" podStartSLOduration=3.288462454 podStartE2EDuration="3.288462454s" podCreationTimestamp="2025-10-08 18:31:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 18:31:58.275989715 +0000 UTC m=+1263.725832495" watchObservedRunningTime="2025-10-08 18:31:58.288462454 +0000 UTC m=+1263.738305224" Oct 08 18:31:59 crc kubenswrapper[4988]: I1008 18:31:59.276313 4988 generic.go:334] "Generic (PLEG): container finished" podID="6c8adde4-41e0-4d33-961a-17a76088b3b8" containerID="8a29e42569215a39afa39fa848957fbb8c88d7c870e4e76ffcd20fafb98282df" exitCode=143 Oct 08 18:31:59 crc kubenswrapper[4988]: I1008 18:31:59.277491 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"6c8adde4-41e0-4d33-961a-17a76088b3b8","Type":"ContainerDied","Data":"8a29e42569215a39afa39fa848957fbb8c88d7c870e4e76ffcd20fafb98282df"} Oct 08 18:32:00 crc kubenswrapper[4988]: I1008 18:32:00.194554 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 08 18:32:00 crc kubenswrapper[4988]: I1008 18:32:00.288519 4988 generic.go:334] "Generic (PLEG): container finished" podID="2df23cd3-cc55-48df-84f1-1b244b1bff09" containerID="d82c51909c3813179ef9e4c9077f8b2443df11cf73f43f8967bbc410e11bb01b" exitCode=0 Oct 08 18:32:00 crc kubenswrapper[4988]: I1008 18:32:00.288557 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2df23cd3-cc55-48df-84f1-1b244b1bff09","Type":"ContainerDied","Data":"d82c51909c3813179ef9e4c9077f8b2443df11cf73f43f8967bbc410e11bb01b"} Oct 08 18:32:00 crc kubenswrapper[4988]: I1008 18:32:00.288583 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2df23cd3-cc55-48df-84f1-1b244b1bff09","Type":"ContainerDied","Data":"54d3ca04da26b716bb45abe2dafc73b6daa59f7d5b9464e5ba249135d0085066"} Oct 08 18:32:00 crc kubenswrapper[4988]: I1008 18:32:00.288600 4988 scope.go:117] "RemoveContainer" containerID="fd6ebabdffd1c5a641c7fe0ca4540b07bcf6bfc6a284e37c9305dfd3e60cffa1" Oct 08 18:32:00 crc kubenswrapper[4988]: I1008 18:32:00.288612 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 08 18:32:00 crc kubenswrapper[4988]: I1008 18:32:00.320183 4988 scope.go:117] "RemoveContainer" containerID="e3574329636b8c9e97dcdb6530a126abbbb07009c757689a079cb8463fc1f56e" Oct 08 18:32:00 crc kubenswrapper[4988]: I1008 18:32:00.343434 4988 scope.go:117] "RemoveContainer" containerID="d82c51909c3813179ef9e4c9077f8b2443df11cf73f43f8967bbc410e11bb01b" Oct 08 18:32:00 crc kubenswrapper[4988]: I1008 18:32:00.367865 4988 scope.go:117] "RemoveContainer" containerID="d397074a7b71070cc40ef8027e42e3ecd64f50a2047c9dea24791be7fb82d6ca" Oct 08 18:32:00 crc kubenswrapper[4988]: I1008 18:32:00.396656 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v8dbl\" (UniqueName: \"kubernetes.io/projected/2df23cd3-cc55-48df-84f1-1b244b1bff09-kube-api-access-v8dbl\") pod \"2df23cd3-cc55-48df-84f1-1b244b1bff09\" (UID: \"2df23cd3-cc55-48df-84f1-1b244b1bff09\") " Oct 08 18:32:00 crc kubenswrapper[4988]: I1008 18:32:00.396797 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/2df23cd3-cc55-48df-84f1-1b244b1bff09-ceilometer-tls-certs\") pod \"2df23cd3-cc55-48df-84f1-1b244b1bff09\" (UID: \"2df23cd3-cc55-48df-84f1-1b244b1bff09\") " Oct 08 18:32:00 crc kubenswrapper[4988]: I1008 18:32:00.396874 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2df23cd3-cc55-48df-84f1-1b244b1bff09-combined-ca-bundle\") pod \"2df23cd3-cc55-48df-84f1-1b244b1bff09\" (UID: \"2df23cd3-cc55-48df-84f1-1b244b1bff09\") " Oct 08 18:32:00 crc kubenswrapper[4988]: I1008 18:32:00.396912 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/2df23cd3-cc55-48df-84f1-1b244b1bff09-sg-core-conf-yaml\") pod \"2df23cd3-cc55-48df-84f1-1b244b1bff09\" (UID: \"2df23cd3-cc55-48df-84f1-1b244b1bff09\") " Oct 08 18:32:00 crc kubenswrapper[4988]: I1008 18:32:00.396962 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2df23cd3-cc55-48df-84f1-1b244b1bff09-config-data\") pod \"2df23cd3-cc55-48df-84f1-1b244b1bff09\" (UID: \"2df23cd3-cc55-48df-84f1-1b244b1bff09\") " Oct 08 18:32:00 crc kubenswrapper[4988]: I1008 18:32:00.396997 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2df23cd3-cc55-48df-84f1-1b244b1bff09-log-httpd\") pod \"2df23cd3-cc55-48df-84f1-1b244b1bff09\" (UID: \"2df23cd3-cc55-48df-84f1-1b244b1bff09\") " Oct 08 18:32:00 crc kubenswrapper[4988]: I1008 18:32:00.397032 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2df23cd3-cc55-48df-84f1-1b244b1bff09-scripts\") pod \"2df23cd3-cc55-48df-84f1-1b244b1bff09\" (UID: \"2df23cd3-cc55-48df-84f1-1b244b1bff09\") " Oct 08 18:32:00 crc kubenswrapper[4988]: I1008 18:32:00.397063 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2df23cd3-cc55-48df-84f1-1b244b1bff09-run-httpd\") pod \"2df23cd3-cc55-48df-84f1-1b244b1bff09\" (UID: \"2df23cd3-cc55-48df-84f1-1b244b1bff09\") " Oct 08 18:32:00 crc kubenswrapper[4988]: I1008 18:32:00.397894 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2df23cd3-cc55-48df-84f1-1b244b1bff09-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "2df23cd3-cc55-48df-84f1-1b244b1bff09" (UID: "2df23cd3-cc55-48df-84f1-1b244b1bff09"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 18:32:00 crc kubenswrapper[4988]: I1008 18:32:00.411716 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2df23cd3-cc55-48df-84f1-1b244b1bff09-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "2df23cd3-cc55-48df-84f1-1b244b1bff09" (UID: "2df23cd3-cc55-48df-84f1-1b244b1bff09"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 18:32:00 crc kubenswrapper[4988]: I1008 18:32:00.413560 4988 scope.go:117] "RemoveContainer" containerID="fd6ebabdffd1c5a641c7fe0ca4540b07bcf6bfc6a284e37c9305dfd3e60cffa1" Oct 08 18:32:00 crc kubenswrapper[4988]: E1008 18:32:00.414054 4988 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fd6ebabdffd1c5a641c7fe0ca4540b07bcf6bfc6a284e37c9305dfd3e60cffa1\": container with ID starting with fd6ebabdffd1c5a641c7fe0ca4540b07bcf6bfc6a284e37c9305dfd3e60cffa1 not found: ID does not exist" containerID="fd6ebabdffd1c5a641c7fe0ca4540b07bcf6bfc6a284e37c9305dfd3e60cffa1" Oct 08 18:32:00 crc kubenswrapper[4988]: I1008 18:32:00.414084 4988 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fd6ebabdffd1c5a641c7fe0ca4540b07bcf6bfc6a284e37c9305dfd3e60cffa1"} err="failed to get container status \"fd6ebabdffd1c5a641c7fe0ca4540b07bcf6bfc6a284e37c9305dfd3e60cffa1\": rpc error: code = NotFound desc = could not find container \"fd6ebabdffd1c5a641c7fe0ca4540b07bcf6bfc6a284e37c9305dfd3e60cffa1\": container with ID starting with fd6ebabdffd1c5a641c7fe0ca4540b07bcf6bfc6a284e37c9305dfd3e60cffa1 not found: ID does not exist" Oct 08 18:32:00 crc kubenswrapper[4988]: I1008 18:32:00.414105 4988 scope.go:117] "RemoveContainer" containerID="e3574329636b8c9e97dcdb6530a126abbbb07009c757689a079cb8463fc1f56e" Oct 08 18:32:00 crc kubenswrapper[4988]: E1008 18:32:00.414373 4988 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e3574329636b8c9e97dcdb6530a126abbbb07009c757689a079cb8463fc1f56e\": container with ID starting with e3574329636b8c9e97dcdb6530a126abbbb07009c757689a079cb8463fc1f56e not found: ID does not exist" containerID="e3574329636b8c9e97dcdb6530a126abbbb07009c757689a079cb8463fc1f56e" Oct 08 18:32:00 crc kubenswrapper[4988]: I1008 18:32:00.414408 4988 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e3574329636b8c9e97dcdb6530a126abbbb07009c757689a079cb8463fc1f56e"} err="failed to get container status \"e3574329636b8c9e97dcdb6530a126abbbb07009c757689a079cb8463fc1f56e\": rpc error: code = NotFound desc = could not find container \"e3574329636b8c9e97dcdb6530a126abbbb07009c757689a079cb8463fc1f56e\": container with ID starting with e3574329636b8c9e97dcdb6530a126abbbb07009c757689a079cb8463fc1f56e not found: ID does not exist" Oct 08 18:32:00 crc kubenswrapper[4988]: I1008 18:32:00.414423 4988 scope.go:117] "RemoveContainer" containerID="d82c51909c3813179ef9e4c9077f8b2443df11cf73f43f8967bbc410e11bb01b" Oct 08 18:32:00 crc kubenswrapper[4988]: E1008 18:32:00.414635 4988 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d82c51909c3813179ef9e4c9077f8b2443df11cf73f43f8967bbc410e11bb01b\": container with ID starting with d82c51909c3813179ef9e4c9077f8b2443df11cf73f43f8967bbc410e11bb01b not found: ID does not exist" containerID="d82c51909c3813179ef9e4c9077f8b2443df11cf73f43f8967bbc410e11bb01b" Oct 08 18:32:00 crc kubenswrapper[4988]: I1008 18:32:00.414655 4988 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d82c51909c3813179ef9e4c9077f8b2443df11cf73f43f8967bbc410e11bb01b"} err="failed to get container status \"d82c51909c3813179ef9e4c9077f8b2443df11cf73f43f8967bbc410e11bb01b\": rpc error: code = NotFound desc = could not find container \"d82c51909c3813179ef9e4c9077f8b2443df11cf73f43f8967bbc410e11bb01b\": container with ID starting with d82c51909c3813179ef9e4c9077f8b2443df11cf73f43f8967bbc410e11bb01b not found: ID does not exist" Oct 08 18:32:00 crc kubenswrapper[4988]: I1008 18:32:00.414667 4988 scope.go:117] "RemoveContainer" containerID="d397074a7b71070cc40ef8027e42e3ecd64f50a2047c9dea24791be7fb82d6ca" Oct 08 18:32:00 crc kubenswrapper[4988]: E1008 18:32:00.414891 4988 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d397074a7b71070cc40ef8027e42e3ecd64f50a2047c9dea24791be7fb82d6ca\": container with ID starting with d397074a7b71070cc40ef8027e42e3ecd64f50a2047c9dea24791be7fb82d6ca not found: ID does not exist" containerID="d397074a7b71070cc40ef8027e42e3ecd64f50a2047c9dea24791be7fb82d6ca" Oct 08 18:32:00 crc kubenswrapper[4988]: I1008 18:32:00.414909 4988 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d397074a7b71070cc40ef8027e42e3ecd64f50a2047c9dea24791be7fb82d6ca"} err="failed to get container status \"d397074a7b71070cc40ef8027e42e3ecd64f50a2047c9dea24791be7fb82d6ca\": rpc error: code = NotFound desc = could not find container \"d397074a7b71070cc40ef8027e42e3ecd64f50a2047c9dea24791be7fb82d6ca\": container with ID starting with d397074a7b71070cc40ef8027e42e3ecd64f50a2047c9dea24791be7fb82d6ca not found: ID does not exist" Oct 08 18:32:00 crc kubenswrapper[4988]: I1008 18:32:00.426649 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2df23cd3-cc55-48df-84f1-1b244b1bff09-scripts" (OuterVolumeSpecName: "scripts") pod "2df23cd3-cc55-48df-84f1-1b244b1bff09" (UID: "2df23cd3-cc55-48df-84f1-1b244b1bff09"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:32:00 crc kubenswrapper[4988]: I1008 18:32:00.444553 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2df23cd3-cc55-48df-84f1-1b244b1bff09-kube-api-access-v8dbl" (OuterVolumeSpecName: "kube-api-access-v8dbl") pod "2df23cd3-cc55-48df-84f1-1b244b1bff09" (UID: "2df23cd3-cc55-48df-84f1-1b244b1bff09"). InnerVolumeSpecName "kube-api-access-v8dbl". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:32:00 crc kubenswrapper[4988]: I1008 18:32:00.461199 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2df23cd3-cc55-48df-84f1-1b244b1bff09-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "2df23cd3-cc55-48df-84f1-1b244b1bff09" (UID: "2df23cd3-cc55-48df-84f1-1b244b1bff09"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:32:00 crc kubenswrapper[4988]: I1008 18:32:00.500735 4988 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/2df23cd3-cc55-48df-84f1-1b244b1bff09-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Oct 08 18:32:00 crc kubenswrapper[4988]: I1008 18:32:00.500777 4988 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2df23cd3-cc55-48df-84f1-1b244b1bff09-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 08 18:32:00 crc kubenswrapper[4988]: I1008 18:32:00.500788 4988 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2df23cd3-cc55-48df-84f1-1b244b1bff09-scripts\") on node \"crc\" DevicePath \"\"" Oct 08 18:32:00 crc kubenswrapper[4988]: I1008 18:32:00.500798 4988 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2df23cd3-cc55-48df-84f1-1b244b1bff09-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 08 18:32:00 crc kubenswrapper[4988]: I1008 18:32:00.500809 4988 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v8dbl\" (UniqueName: \"kubernetes.io/projected/2df23cd3-cc55-48df-84f1-1b244b1bff09-kube-api-access-v8dbl\") on node \"crc\" DevicePath \"\"" Oct 08 18:32:00 crc kubenswrapper[4988]: I1008 18:32:00.567914 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2df23cd3-cc55-48df-84f1-1b244b1bff09-ceilometer-tls-certs" (OuterVolumeSpecName: "ceilometer-tls-certs") pod "2df23cd3-cc55-48df-84f1-1b244b1bff09" (UID: "2df23cd3-cc55-48df-84f1-1b244b1bff09"). InnerVolumeSpecName "ceilometer-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:32:00 crc kubenswrapper[4988]: I1008 18:32:00.585536 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2df23cd3-cc55-48df-84f1-1b244b1bff09-config-data" (OuterVolumeSpecName: "config-data") pod "2df23cd3-cc55-48df-84f1-1b244b1bff09" (UID: "2df23cd3-cc55-48df-84f1-1b244b1bff09"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:32:00 crc kubenswrapper[4988]: I1008 18:32:00.600314 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2df23cd3-cc55-48df-84f1-1b244b1bff09-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "2df23cd3-cc55-48df-84f1-1b244b1bff09" (UID: "2df23cd3-cc55-48df-84f1-1b244b1bff09"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:32:00 crc kubenswrapper[4988]: I1008 18:32:00.602528 4988 reconciler_common.go:293] "Volume detached for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/2df23cd3-cc55-48df-84f1-1b244b1bff09-ceilometer-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 08 18:32:00 crc kubenswrapper[4988]: I1008 18:32:00.602555 4988 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2df23cd3-cc55-48df-84f1-1b244b1bff09-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 18:32:00 crc kubenswrapper[4988]: I1008 18:32:00.602566 4988 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2df23cd3-cc55-48df-84f1-1b244b1bff09-config-data\") on node \"crc\" DevicePath \"\"" Oct 08 18:32:00 crc kubenswrapper[4988]: I1008 18:32:00.921427 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 08 18:32:00 crc kubenswrapper[4988]: I1008 18:32:00.931573 4988 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Oct 08 18:32:00 crc kubenswrapper[4988]: I1008 18:32:00.948371 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 08 18:32:00 crc kubenswrapper[4988]: E1008 18:32:00.948852 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2df23cd3-cc55-48df-84f1-1b244b1bff09" containerName="proxy-httpd" Oct 08 18:32:00 crc kubenswrapper[4988]: I1008 18:32:00.948870 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="2df23cd3-cc55-48df-84f1-1b244b1bff09" containerName="proxy-httpd" Oct 08 18:32:00 crc kubenswrapper[4988]: E1008 18:32:00.948901 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2df23cd3-cc55-48df-84f1-1b244b1bff09" containerName="sg-core" Oct 08 18:32:00 crc kubenswrapper[4988]: I1008 18:32:00.948911 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="2df23cd3-cc55-48df-84f1-1b244b1bff09" containerName="sg-core" Oct 08 18:32:00 crc kubenswrapper[4988]: E1008 18:32:00.948931 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2df23cd3-cc55-48df-84f1-1b244b1bff09" containerName="ceilometer-central-agent" Oct 08 18:32:00 crc kubenswrapper[4988]: I1008 18:32:00.948944 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="2df23cd3-cc55-48df-84f1-1b244b1bff09" containerName="ceilometer-central-agent" Oct 08 18:32:00 crc kubenswrapper[4988]: E1008 18:32:00.948978 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2df23cd3-cc55-48df-84f1-1b244b1bff09" containerName="ceilometer-notification-agent" Oct 08 18:32:00 crc kubenswrapper[4988]: I1008 18:32:00.948989 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="2df23cd3-cc55-48df-84f1-1b244b1bff09" containerName="ceilometer-notification-agent" Oct 08 18:32:00 crc kubenswrapper[4988]: I1008 18:32:00.949224 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="2df23cd3-cc55-48df-84f1-1b244b1bff09" containerName="ceilometer-notification-agent" Oct 08 18:32:00 crc kubenswrapper[4988]: I1008 18:32:00.949257 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="2df23cd3-cc55-48df-84f1-1b244b1bff09" containerName="proxy-httpd" Oct 08 18:32:00 crc kubenswrapper[4988]: I1008 18:32:00.949275 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="2df23cd3-cc55-48df-84f1-1b244b1bff09" containerName="sg-core" Oct 08 18:32:00 crc kubenswrapper[4988]: I1008 18:32:00.949291 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="2df23cd3-cc55-48df-84f1-1b244b1bff09" containerName="ceilometer-central-agent" Oct 08 18:32:00 crc kubenswrapper[4988]: I1008 18:32:00.951762 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 08 18:32:00 crc kubenswrapper[4988]: I1008 18:32:00.953809 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Oct 08 18:32:00 crc kubenswrapper[4988]: I1008 18:32:00.954104 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 08 18:32:00 crc kubenswrapper[4988]: I1008 18:32:00.955374 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 08 18:32:00 crc kubenswrapper[4988]: I1008 18:32:00.961290 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 08 18:32:01 crc kubenswrapper[4988]: I1008 18:32:01.110505 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ccd2f8fa-1c16-4c58-9f87-e60ba3a2f5fb-config-data\") pod \"ceilometer-0\" (UID: \"ccd2f8fa-1c16-4c58-9f87-e60ba3a2f5fb\") " pod="openstack/ceilometer-0" Oct 08 18:32:01 crc kubenswrapper[4988]: I1008 18:32:01.110575 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/ccd2f8fa-1c16-4c58-9f87-e60ba3a2f5fb-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"ccd2f8fa-1c16-4c58-9f87-e60ba3a2f5fb\") " pod="openstack/ceilometer-0" Oct 08 18:32:01 crc kubenswrapper[4988]: I1008 18:32:01.110624 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ccd2f8fa-1c16-4c58-9f87-e60ba3a2f5fb-scripts\") pod \"ceilometer-0\" (UID: \"ccd2f8fa-1c16-4c58-9f87-e60ba3a2f5fb\") " pod="openstack/ceilometer-0" Oct 08 18:32:01 crc kubenswrapper[4988]: I1008 18:32:01.110640 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/ccd2f8fa-1c16-4c58-9f87-e60ba3a2f5fb-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"ccd2f8fa-1c16-4c58-9f87-e60ba3a2f5fb\") " pod="openstack/ceilometer-0" Oct 08 18:32:01 crc kubenswrapper[4988]: I1008 18:32:01.110664 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ccd2f8fa-1c16-4c58-9f87-e60ba3a2f5fb-log-httpd\") pod \"ceilometer-0\" (UID: \"ccd2f8fa-1c16-4c58-9f87-e60ba3a2f5fb\") " pod="openstack/ceilometer-0" Oct 08 18:32:01 crc kubenswrapper[4988]: I1008 18:32:01.110688 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ccd2f8fa-1c16-4c58-9f87-e60ba3a2f5fb-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"ccd2f8fa-1c16-4c58-9f87-e60ba3a2f5fb\") " pod="openstack/ceilometer-0" Oct 08 18:32:01 crc kubenswrapper[4988]: I1008 18:32:01.110750 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ccd2f8fa-1c16-4c58-9f87-e60ba3a2f5fb-run-httpd\") pod \"ceilometer-0\" (UID: \"ccd2f8fa-1c16-4c58-9f87-e60ba3a2f5fb\") " pod="openstack/ceilometer-0" Oct 08 18:32:01 crc kubenswrapper[4988]: I1008 18:32:01.110814 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8p2rl\" (UniqueName: \"kubernetes.io/projected/ccd2f8fa-1c16-4c58-9f87-e60ba3a2f5fb-kube-api-access-8p2rl\") pod \"ceilometer-0\" (UID: \"ccd2f8fa-1c16-4c58-9f87-e60ba3a2f5fb\") " pod="openstack/ceilometer-0" Oct 08 18:32:01 crc kubenswrapper[4988]: I1008 18:32:01.212280 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ccd2f8fa-1c16-4c58-9f87-e60ba3a2f5fb-log-httpd\") pod \"ceilometer-0\" (UID: \"ccd2f8fa-1c16-4c58-9f87-e60ba3a2f5fb\") " pod="openstack/ceilometer-0" Oct 08 18:32:01 crc kubenswrapper[4988]: I1008 18:32:01.212582 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ccd2f8fa-1c16-4c58-9f87-e60ba3a2f5fb-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"ccd2f8fa-1c16-4c58-9f87-e60ba3a2f5fb\") " pod="openstack/ceilometer-0" Oct 08 18:32:01 crc kubenswrapper[4988]: I1008 18:32:01.212611 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ccd2f8fa-1c16-4c58-9f87-e60ba3a2f5fb-run-httpd\") pod \"ceilometer-0\" (UID: \"ccd2f8fa-1c16-4c58-9f87-e60ba3a2f5fb\") " pod="openstack/ceilometer-0" Oct 08 18:32:01 crc kubenswrapper[4988]: I1008 18:32:01.212665 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8p2rl\" (UniqueName: \"kubernetes.io/projected/ccd2f8fa-1c16-4c58-9f87-e60ba3a2f5fb-kube-api-access-8p2rl\") pod \"ceilometer-0\" (UID: \"ccd2f8fa-1c16-4c58-9f87-e60ba3a2f5fb\") " pod="openstack/ceilometer-0" Oct 08 18:32:01 crc kubenswrapper[4988]: I1008 18:32:01.212745 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ccd2f8fa-1c16-4c58-9f87-e60ba3a2f5fb-config-data\") pod \"ceilometer-0\" (UID: \"ccd2f8fa-1c16-4c58-9f87-e60ba3a2f5fb\") " pod="openstack/ceilometer-0" Oct 08 18:32:01 crc kubenswrapper[4988]: I1008 18:32:01.212796 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/ccd2f8fa-1c16-4c58-9f87-e60ba3a2f5fb-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"ccd2f8fa-1c16-4c58-9f87-e60ba3a2f5fb\") " pod="openstack/ceilometer-0" Oct 08 18:32:01 crc kubenswrapper[4988]: I1008 18:32:01.212809 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ccd2f8fa-1c16-4c58-9f87-e60ba3a2f5fb-log-httpd\") pod \"ceilometer-0\" (UID: \"ccd2f8fa-1c16-4c58-9f87-e60ba3a2f5fb\") " pod="openstack/ceilometer-0" Oct 08 18:32:01 crc kubenswrapper[4988]: I1008 18:32:01.212834 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ccd2f8fa-1c16-4c58-9f87-e60ba3a2f5fb-scripts\") pod \"ceilometer-0\" (UID: \"ccd2f8fa-1c16-4c58-9f87-e60ba3a2f5fb\") " pod="openstack/ceilometer-0" Oct 08 18:32:01 crc kubenswrapper[4988]: I1008 18:32:01.212851 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/ccd2f8fa-1c16-4c58-9f87-e60ba3a2f5fb-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"ccd2f8fa-1c16-4c58-9f87-e60ba3a2f5fb\") " pod="openstack/ceilometer-0" Oct 08 18:32:01 crc kubenswrapper[4988]: I1008 18:32:01.213115 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ccd2f8fa-1c16-4c58-9f87-e60ba3a2f5fb-run-httpd\") pod \"ceilometer-0\" (UID: \"ccd2f8fa-1c16-4c58-9f87-e60ba3a2f5fb\") " pod="openstack/ceilometer-0" Oct 08 18:32:01 crc kubenswrapper[4988]: I1008 18:32:01.218807 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ccd2f8fa-1c16-4c58-9f87-e60ba3a2f5fb-scripts\") pod \"ceilometer-0\" (UID: \"ccd2f8fa-1c16-4c58-9f87-e60ba3a2f5fb\") " pod="openstack/ceilometer-0" Oct 08 18:32:01 crc kubenswrapper[4988]: I1008 18:32:01.219187 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ccd2f8fa-1c16-4c58-9f87-e60ba3a2f5fb-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"ccd2f8fa-1c16-4c58-9f87-e60ba3a2f5fb\") " pod="openstack/ceilometer-0" Oct 08 18:32:01 crc kubenswrapper[4988]: I1008 18:32:01.219276 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/ccd2f8fa-1c16-4c58-9f87-e60ba3a2f5fb-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"ccd2f8fa-1c16-4c58-9f87-e60ba3a2f5fb\") " pod="openstack/ceilometer-0" Oct 08 18:32:01 crc kubenswrapper[4988]: I1008 18:32:01.219783 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ccd2f8fa-1c16-4c58-9f87-e60ba3a2f5fb-config-data\") pod \"ceilometer-0\" (UID: \"ccd2f8fa-1c16-4c58-9f87-e60ba3a2f5fb\") " pod="openstack/ceilometer-0" Oct 08 18:32:01 crc kubenswrapper[4988]: I1008 18:32:01.221876 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/ccd2f8fa-1c16-4c58-9f87-e60ba3a2f5fb-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"ccd2f8fa-1c16-4c58-9f87-e60ba3a2f5fb\") " pod="openstack/ceilometer-0" Oct 08 18:32:01 crc kubenswrapper[4988]: I1008 18:32:01.232578 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8p2rl\" (UniqueName: \"kubernetes.io/projected/ccd2f8fa-1c16-4c58-9f87-e60ba3a2f5fb-kube-api-access-8p2rl\") pod \"ceilometer-0\" (UID: \"ccd2f8fa-1c16-4c58-9f87-e60ba3a2f5fb\") " pod="openstack/ceilometer-0" Oct 08 18:32:01 crc kubenswrapper[4988]: I1008 18:32:01.248367 4988 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2df23cd3-cc55-48df-84f1-1b244b1bff09" path="/var/lib/kubelet/pods/2df23cd3-cc55-48df-84f1-1b244b1bff09/volumes" Oct 08 18:32:01 crc kubenswrapper[4988]: I1008 18:32:01.289158 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 08 18:32:01 crc kubenswrapper[4988]: I1008 18:32:01.762950 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 08 18:32:01 crc kubenswrapper[4988]: I1008 18:32:01.791820 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 08 18:32:01 crc kubenswrapper[4988]: I1008 18:32:01.934956 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6c8adde4-41e0-4d33-961a-17a76088b3b8-logs\") pod \"6c8adde4-41e0-4d33-961a-17a76088b3b8\" (UID: \"6c8adde4-41e0-4d33-961a-17a76088b3b8\") " Oct 08 18:32:01 crc kubenswrapper[4988]: I1008 18:32:01.935061 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xvrsx\" (UniqueName: \"kubernetes.io/projected/6c8adde4-41e0-4d33-961a-17a76088b3b8-kube-api-access-xvrsx\") pod \"6c8adde4-41e0-4d33-961a-17a76088b3b8\" (UID: \"6c8adde4-41e0-4d33-961a-17a76088b3b8\") " Oct 08 18:32:01 crc kubenswrapper[4988]: I1008 18:32:01.935100 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6c8adde4-41e0-4d33-961a-17a76088b3b8-combined-ca-bundle\") pod \"6c8adde4-41e0-4d33-961a-17a76088b3b8\" (UID: \"6c8adde4-41e0-4d33-961a-17a76088b3b8\") " Oct 08 18:32:01 crc kubenswrapper[4988]: I1008 18:32:01.935140 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6c8adde4-41e0-4d33-961a-17a76088b3b8-config-data\") pod \"6c8adde4-41e0-4d33-961a-17a76088b3b8\" (UID: \"6c8adde4-41e0-4d33-961a-17a76088b3b8\") " Oct 08 18:32:01 crc kubenswrapper[4988]: I1008 18:32:01.936651 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6c8adde4-41e0-4d33-961a-17a76088b3b8-logs" (OuterVolumeSpecName: "logs") pod "6c8adde4-41e0-4d33-961a-17a76088b3b8" (UID: "6c8adde4-41e0-4d33-961a-17a76088b3b8"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 18:32:01 crc kubenswrapper[4988]: I1008 18:32:01.948181 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6c8adde4-41e0-4d33-961a-17a76088b3b8-kube-api-access-xvrsx" (OuterVolumeSpecName: "kube-api-access-xvrsx") pod "6c8adde4-41e0-4d33-961a-17a76088b3b8" (UID: "6c8adde4-41e0-4d33-961a-17a76088b3b8"). InnerVolumeSpecName "kube-api-access-xvrsx". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:32:01 crc kubenswrapper[4988]: I1008 18:32:01.973143 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6c8adde4-41e0-4d33-961a-17a76088b3b8-config-data" (OuterVolumeSpecName: "config-data") pod "6c8adde4-41e0-4d33-961a-17a76088b3b8" (UID: "6c8adde4-41e0-4d33-961a-17a76088b3b8"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:32:01 crc kubenswrapper[4988]: I1008 18:32:01.981678 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6c8adde4-41e0-4d33-961a-17a76088b3b8-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "6c8adde4-41e0-4d33-961a-17a76088b3b8" (UID: "6c8adde4-41e0-4d33-961a-17a76088b3b8"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:32:02 crc kubenswrapper[4988]: I1008 18:32:02.036502 4988 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6c8adde4-41e0-4d33-961a-17a76088b3b8-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 18:32:02 crc kubenswrapper[4988]: I1008 18:32:02.036533 4988 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6c8adde4-41e0-4d33-961a-17a76088b3b8-config-data\") on node \"crc\" DevicePath \"\"" Oct 08 18:32:02 crc kubenswrapper[4988]: I1008 18:32:02.036542 4988 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6c8adde4-41e0-4d33-961a-17a76088b3b8-logs\") on node \"crc\" DevicePath \"\"" Oct 08 18:32:02 crc kubenswrapper[4988]: I1008 18:32:02.036551 4988 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xvrsx\" (UniqueName: \"kubernetes.io/projected/6c8adde4-41e0-4d33-961a-17a76088b3b8-kube-api-access-xvrsx\") on node \"crc\" DevicePath \"\"" Oct 08 18:32:02 crc kubenswrapper[4988]: I1008 18:32:02.311006 4988 generic.go:334] "Generic (PLEG): container finished" podID="6c8adde4-41e0-4d33-961a-17a76088b3b8" containerID="56adcfbff76321bed6638d6eb70e3ed96f4f090daebaacfb04d44073983e39a8" exitCode=0 Oct 08 18:32:02 crc kubenswrapper[4988]: I1008 18:32:02.311089 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"6c8adde4-41e0-4d33-961a-17a76088b3b8","Type":"ContainerDied","Data":"56adcfbff76321bed6638d6eb70e3ed96f4f090daebaacfb04d44073983e39a8"} Oct 08 18:32:02 crc kubenswrapper[4988]: I1008 18:32:02.311128 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"6c8adde4-41e0-4d33-961a-17a76088b3b8","Type":"ContainerDied","Data":"3510a6108a346aa0be640613afe19a7f162113ead0d4f39c8a2d108f4df106d7"} Oct 08 18:32:02 crc kubenswrapper[4988]: I1008 18:32:02.311094 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 08 18:32:02 crc kubenswrapper[4988]: I1008 18:32:02.311172 4988 scope.go:117] "RemoveContainer" containerID="56adcfbff76321bed6638d6eb70e3ed96f4f090daebaacfb04d44073983e39a8" Oct 08 18:32:02 crc kubenswrapper[4988]: I1008 18:32:02.314317 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ccd2f8fa-1c16-4c58-9f87-e60ba3a2f5fb","Type":"ContainerStarted","Data":"fd69699740aac8c03e9405c7eb6b657c410818eb3005cb0c61692c403508b677"} Oct 08 18:32:02 crc kubenswrapper[4988]: I1008 18:32:02.345167 4988 scope.go:117] "RemoveContainer" containerID="8a29e42569215a39afa39fa848957fbb8c88d7c870e4e76ffcd20fafb98282df" Oct 08 18:32:02 crc kubenswrapper[4988]: I1008 18:32:02.345950 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 08 18:32:02 crc kubenswrapper[4988]: I1008 18:32:02.365603 4988 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Oct 08 18:32:02 crc kubenswrapper[4988]: I1008 18:32:02.374884 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Oct 08 18:32:02 crc kubenswrapper[4988]: E1008 18:32:02.375244 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6c8adde4-41e0-4d33-961a-17a76088b3b8" containerName="nova-api-api" Oct 08 18:32:02 crc kubenswrapper[4988]: I1008 18:32:02.375262 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="6c8adde4-41e0-4d33-961a-17a76088b3b8" containerName="nova-api-api" Oct 08 18:32:02 crc kubenswrapper[4988]: E1008 18:32:02.375284 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6c8adde4-41e0-4d33-961a-17a76088b3b8" containerName="nova-api-log" Oct 08 18:32:02 crc kubenswrapper[4988]: I1008 18:32:02.375291 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="6c8adde4-41e0-4d33-961a-17a76088b3b8" containerName="nova-api-log" Oct 08 18:32:02 crc kubenswrapper[4988]: I1008 18:32:02.375486 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="6c8adde4-41e0-4d33-961a-17a76088b3b8" containerName="nova-api-api" Oct 08 18:32:02 crc kubenswrapper[4988]: I1008 18:32:02.375505 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="6c8adde4-41e0-4d33-961a-17a76088b3b8" containerName="nova-api-log" Oct 08 18:32:02 crc kubenswrapper[4988]: I1008 18:32:02.376559 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 08 18:32:02 crc kubenswrapper[4988]: I1008 18:32:02.383049 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-public-svc" Oct 08 18:32:02 crc kubenswrapper[4988]: I1008 18:32:02.383204 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-internal-svc" Oct 08 18:32:02 crc kubenswrapper[4988]: I1008 18:32:02.383448 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Oct 08 18:32:02 crc kubenswrapper[4988]: I1008 18:32:02.387321 4988 scope.go:117] "RemoveContainer" containerID="56adcfbff76321bed6638d6eb70e3ed96f4f090daebaacfb04d44073983e39a8" Oct 08 18:32:02 crc kubenswrapper[4988]: I1008 18:32:02.387635 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 08 18:32:02 crc kubenswrapper[4988]: E1008 18:32:02.387941 4988 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"56adcfbff76321bed6638d6eb70e3ed96f4f090daebaacfb04d44073983e39a8\": container with ID starting with 56adcfbff76321bed6638d6eb70e3ed96f4f090daebaacfb04d44073983e39a8 not found: ID does not exist" containerID="56adcfbff76321bed6638d6eb70e3ed96f4f090daebaacfb04d44073983e39a8" Oct 08 18:32:02 crc kubenswrapper[4988]: I1008 18:32:02.387987 4988 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"56adcfbff76321bed6638d6eb70e3ed96f4f090daebaacfb04d44073983e39a8"} err="failed to get container status \"56adcfbff76321bed6638d6eb70e3ed96f4f090daebaacfb04d44073983e39a8\": rpc error: code = NotFound desc = could not find container \"56adcfbff76321bed6638d6eb70e3ed96f4f090daebaacfb04d44073983e39a8\": container with ID starting with 56adcfbff76321bed6638d6eb70e3ed96f4f090daebaacfb04d44073983e39a8 not found: ID does not exist" Oct 08 18:32:02 crc kubenswrapper[4988]: I1008 18:32:02.388013 4988 scope.go:117] "RemoveContainer" containerID="8a29e42569215a39afa39fa848957fbb8c88d7c870e4e76ffcd20fafb98282df" Oct 08 18:32:02 crc kubenswrapper[4988]: E1008 18:32:02.388299 4988 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8a29e42569215a39afa39fa848957fbb8c88d7c870e4e76ffcd20fafb98282df\": container with ID starting with 8a29e42569215a39afa39fa848957fbb8c88d7c870e4e76ffcd20fafb98282df not found: ID does not exist" containerID="8a29e42569215a39afa39fa848957fbb8c88d7c870e4e76ffcd20fafb98282df" Oct 08 18:32:02 crc kubenswrapper[4988]: I1008 18:32:02.388333 4988 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8a29e42569215a39afa39fa848957fbb8c88d7c870e4e76ffcd20fafb98282df"} err="failed to get container status \"8a29e42569215a39afa39fa848957fbb8c88d7c870e4e76ffcd20fafb98282df\": rpc error: code = NotFound desc = could not find container \"8a29e42569215a39afa39fa848957fbb8c88d7c870e4e76ffcd20fafb98282df\": container with ID starting with 8a29e42569215a39afa39fa848957fbb8c88d7c870e4e76ffcd20fafb98282df not found: ID does not exist" Oct 08 18:32:02 crc kubenswrapper[4988]: I1008 18:32:02.543775 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/38babc76-590a-423e-b104-46e0e58fd3cc-config-data\") pod \"nova-api-0\" (UID: \"38babc76-590a-423e-b104-46e0e58fd3cc\") " pod="openstack/nova-api-0" Oct 08 18:32:02 crc kubenswrapper[4988]: I1008 18:32:02.543865 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-45xm7\" (UniqueName: \"kubernetes.io/projected/38babc76-590a-423e-b104-46e0e58fd3cc-kube-api-access-45xm7\") pod \"nova-api-0\" (UID: \"38babc76-590a-423e-b104-46e0e58fd3cc\") " pod="openstack/nova-api-0" Oct 08 18:32:02 crc kubenswrapper[4988]: I1008 18:32:02.543957 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/38babc76-590a-423e-b104-46e0e58fd3cc-public-tls-certs\") pod \"nova-api-0\" (UID: \"38babc76-590a-423e-b104-46e0e58fd3cc\") " pod="openstack/nova-api-0" Oct 08 18:32:02 crc kubenswrapper[4988]: I1008 18:32:02.543977 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/38babc76-590a-423e-b104-46e0e58fd3cc-internal-tls-certs\") pod \"nova-api-0\" (UID: \"38babc76-590a-423e-b104-46e0e58fd3cc\") " pod="openstack/nova-api-0" Oct 08 18:32:02 crc kubenswrapper[4988]: I1008 18:32:02.543999 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/38babc76-590a-423e-b104-46e0e58fd3cc-logs\") pod \"nova-api-0\" (UID: \"38babc76-590a-423e-b104-46e0e58fd3cc\") " pod="openstack/nova-api-0" Oct 08 18:32:02 crc kubenswrapper[4988]: I1008 18:32:02.544057 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/38babc76-590a-423e-b104-46e0e58fd3cc-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"38babc76-590a-423e-b104-46e0e58fd3cc\") " pod="openstack/nova-api-0" Oct 08 18:32:02 crc kubenswrapper[4988]: I1008 18:32:02.646036 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/38babc76-590a-423e-b104-46e0e58fd3cc-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"38babc76-590a-423e-b104-46e0e58fd3cc\") " pod="openstack/nova-api-0" Oct 08 18:32:02 crc kubenswrapper[4988]: I1008 18:32:02.646356 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/38babc76-590a-423e-b104-46e0e58fd3cc-config-data\") pod \"nova-api-0\" (UID: \"38babc76-590a-423e-b104-46e0e58fd3cc\") " pod="openstack/nova-api-0" Oct 08 18:32:02 crc kubenswrapper[4988]: I1008 18:32:02.646399 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-45xm7\" (UniqueName: \"kubernetes.io/projected/38babc76-590a-423e-b104-46e0e58fd3cc-kube-api-access-45xm7\") pod \"nova-api-0\" (UID: \"38babc76-590a-423e-b104-46e0e58fd3cc\") " pod="openstack/nova-api-0" Oct 08 18:32:02 crc kubenswrapper[4988]: I1008 18:32:02.646484 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/38babc76-590a-423e-b104-46e0e58fd3cc-public-tls-certs\") pod \"nova-api-0\" (UID: \"38babc76-590a-423e-b104-46e0e58fd3cc\") " pod="openstack/nova-api-0" Oct 08 18:32:02 crc kubenswrapper[4988]: I1008 18:32:02.646505 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/38babc76-590a-423e-b104-46e0e58fd3cc-internal-tls-certs\") pod \"nova-api-0\" (UID: \"38babc76-590a-423e-b104-46e0e58fd3cc\") " pod="openstack/nova-api-0" Oct 08 18:32:02 crc kubenswrapper[4988]: I1008 18:32:02.646525 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/38babc76-590a-423e-b104-46e0e58fd3cc-logs\") pod \"nova-api-0\" (UID: \"38babc76-590a-423e-b104-46e0e58fd3cc\") " pod="openstack/nova-api-0" Oct 08 18:32:02 crc kubenswrapper[4988]: I1008 18:32:02.646947 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/38babc76-590a-423e-b104-46e0e58fd3cc-logs\") pod \"nova-api-0\" (UID: \"38babc76-590a-423e-b104-46e0e58fd3cc\") " pod="openstack/nova-api-0" Oct 08 18:32:02 crc kubenswrapper[4988]: I1008 18:32:02.650024 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/38babc76-590a-423e-b104-46e0e58fd3cc-internal-tls-certs\") pod \"nova-api-0\" (UID: \"38babc76-590a-423e-b104-46e0e58fd3cc\") " pod="openstack/nova-api-0" Oct 08 18:32:02 crc kubenswrapper[4988]: I1008 18:32:02.650661 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/38babc76-590a-423e-b104-46e0e58fd3cc-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"38babc76-590a-423e-b104-46e0e58fd3cc\") " pod="openstack/nova-api-0" Oct 08 18:32:02 crc kubenswrapper[4988]: I1008 18:32:02.651731 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/38babc76-590a-423e-b104-46e0e58fd3cc-config-data\") pod \"nova-api-0\" (UID: \"38babc76-590a-423e-b104-46e0e58fd3cc\") " pod="openstack/nova-api-0" Oct 08 18:32:02 crc kubenswrapper[4988]: I1008 18:32:02.652615 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/38babc76-590a-423e-b104-46e0e58fd3cc-public-tls-certs\") pod \"nova-api-0\" (UID: \"38babc76-590a-423e-b104-46e0e58fd3cc\") " pod="openstack/nova-api-0" Oct 08 18:32:02 crc kubenswrapper[4988]: I1008 18:32:02.663437 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-45xm7\" (UniqueName: \"kubernetes.io/projected/38babc76-590a-423e-b104-46e0e58fd3cc-kube-api-access-45xm7\") pod \"nova-api-0\" (UID: \"38babc76-590a-423e-b104-46e0e58fd3cc\") " pod="openstack/nova-api-0" Oct 08 18:32:02 crc kubenswrapper[4988]: I1008 18:32:02.711635 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 08 18:32:02 crc kubenswrapper[4988]: I1008 18:32:02.852350 4988 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-cell1-novncproxy-0" Oct 08 18:32:02 crc kubenswrapper[4988]: I1008 18:32:02.884770 4988 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-cell1-novncproxy-0" Oct 08 18:32:03 crc kubenswrapper[4988]: I1008 18:32:03.228938 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 08 18:32:03 crc kubenswrapper[4988]: W1008 18:32:03.234781 4988 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod38babc76_590a_423e_b104_46e0e58fd3cc.slice/crio-958dbdf57bb6d3003dfe38be859959524bf9bccc068bf1735c1789240e600430 WatchSource:0}: Error finding container 958dbdf57bb6d3003dfe38be859959524bf9bccc068bf1735c1789240e600430: Status 404 returned error can't find the container with id 958dbdf57bb6d3003dfe38be859959524bf9bccc068bf1735c1789240e600430 Oct 08 18:32:03 crc kubenswrapper[4988]: I1008 18:32:03.250365 4988 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6c8adde4-41e0-4d33-961a-17a76088b3b8" path="/var/lib/kubelet/pods/6c8adde4-41e0-4d33-961a-17a76088b3b8/volumes" Oct 08 18:32:03 crc kubenswrapper[4988]: I1008 18:32:03.324826 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ccd2f8fa-1c16-4c58-9f87-e60ba3a2f5fb","Type":"ContainerStarted","Data":"c7ddcc51fc71ff2377802bb6d1c0c0545ab7211828f994c467b39e3b24eaac16"} Oct 08 18:32:03 crc kubenswrapper[4988]: I1008 18:32:03.327852 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"38babc76-590a-423e-b104-46e0e58fd3cc","Type":"ContainerStarted","Data":"958dbdf57bb6d3003dfe38be859959524bf9bccc068bf1735c1789240e600430"} Oct 08 18:32:03 crc kubenswrapper[4988]: I1008 18:32:03.347909 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-novncproxy-0" Oct 08 18:32:03 crc kubenswrapper[4988]: I1008 18:32:03.490361 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-cell-mapping-92jjg"] Oct 08 18:32:03 crc kubenswrapper[4988]: I1008 18:32:03.492238 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-92jjg" Oct 08 18:32:03 crc kubenswrapper[4988]: I1008 18:32:03.493906 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-config-data" Oct 08 18:32:03 crc kubenswrapper[4988]: I1008 18:32:03.496819 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-scripts" Oct 08 18:32:03 crc kubenswrapper[4988]: I1008 18:32:03.502595 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-92jjg"] Oct 08 18:32:03 crc kubenswrapper[4988]: I1008 18:32:03.671316 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c132f5da-fe6a-4616-9d2f-bc0c046244a1-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-92jjg\" (UID: \"c132f5da-fe6a-4616-9d2f-bc0c046244a1\") " pod="openstack/nova-cell1-cell-mapping-92jjg" Oct 08 18:32:03 crc kubenswrapper[4988]: I1008 18:32:03.671616 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c132f5da-fe6a-4616-9d2f-bc0c046244a1-config-data\") pod \"nova-cell1-cell-mapping-92jjg\" (UID: \"c132f5da-fe6a-4616-9d2f-bc0c046244a1\") " pod="openstack/nova-cell1-cell-mapping-92jjg" Oct 08 18:32:03 crc kubenswrapper[4988]: I1008 18:32:03.671652 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dggs9\" (UniqueName: \"kubernetes.io/projected/c132f5da-fe6a-4616-9d2f-bc0c046244a1-kube-api-access-dggs9\") pod \"nova-cell1-cell-mapping-92jjg\" (UID: \"c132f5da-fe6a-4616-9d2f-bc0c046244a1\") " pod="openstack/nova-cell1-cell-mapping-92jjg" Oct 08 18:32:03 crc kubenswrapper[4988]: I1008 18:32:03.671786 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c132f5da-fe6a-4616-9d2f-bc0c046244a1-scripts\") pod \"nova-cell1-cell-mapping-92jjg\" (UID: \"c132f5da-fe6a-4616-9d2f-bc0c046244a1\") " pod="openstack/nova-cell1-cell-mapping-92jjg" Oct 08 18:32:03 crc kubenswrapper[4988]: I1008 18:32:03.773667 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c132f5da-fe6a-4616-9d2f-bc0c046244a1-scripts\") pod \"nova-cell1-cell-mapping-92jjg\" (UID: \"c132f5da-fe6a-4616-9d2f-bc0c046244a1\") " pod="openstack/nova-cell1-cell-mapping-92jjg" Oct 08 18:32:03 crc kubenswrapper[4988]: I1008 18:32:03.773742 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c132f5da-fe6a-4616-9d2f-bc0c046244a1-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-92jjg\" (UID: \"c132f5da-fe6a-4616-9d2f-bc0c046244a1\") " pod="openstack/nova-cell1-cell-mapping-92jjg" Oct 08 18:32:03 crc kubenswrapper[4988]: I1008 18:32:03.773819 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c132f5da-fe6a-4616-9d2f-bc0c046244a1-config-data\") pod \"nova-cell1-cell-mapping-92jjg\" (UID: \"c132f5da-fe6a-4616-9d2f-bc0c046244a1\") " pod="openstack/nova-cell1-cell-mapping-92jjg" Oct 08 18:32:03 crc kubenswrapper[4988]: I1008 18:32:03.773842 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dggs9\" (UniqueName: \"kubernetes.io/projected/c132f5da-fe6a-4616-9d2f-bc0c046244a1-kube-api-access-dggs9\") pod \"nova-cell1-cell-mapping-92jjg\" (UID: \"c132f5da-fe6a-4616-9d2f-bc0c046244a1\") " pod="openstack/nova-cell1-cell-mapping-92jjg" Oct 08 18:32:03 crc kubenswrapper[4988]: I1008 18:32:03.778460 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c132f5da-fe6a-4616-9d2f-bc0c046244a1-config-data\") pod \"nova-cell1-cell-mapping-92jjg\" (UID: \"c132f5da-fe6a-4616-9d2f-bc0c046244a1\") " pod="openstack/nova-cell1-cell-mapping-92jjg" Oct 08 18:32:03 crc kubenswrapper[4988]: I1008 18:32:03.779870 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c132f5da-fe6a-4616-9d2f-bc0c046244a1-scripts\") pod \"nova-cell1-cell-mapping-92jjg\" (UID: \"c132f5da-fe6a-4616-9d2f-bc0c046244a1\") " pod="openstack/nova-cell1-cell-mapping-92jjg" Oct 08 18:32:03 crc kubenswrapper[4988]: I1008 18:32:03.780314 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c132f5da-fe6a-4616-9d2f-bc0c046244a1-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-92jjg\" (UID: \"c132f5da-fe6a-4616-9d2f-bc0c046244a1\") " pod="openstack/nova-cell1-cell-mapping-92jjg" Oct 08 18:32:03 crc kubenswrapper[4988]: I1008 18:32:03.796706 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dggs9\" (UniqueName: \"kubernetes.io/projected/c132f5da-fe6a-4616-9d2f-bc0c046244a1-kube-api-access-dggs9\") pod \"nova-cell1-cell-mapping-92jjg\" (UID: \"c132f5da-fe6a-4616-9d2f-bc0c046244a1\") " pod="openstack/nova-cell1-cell-mapping-92jjg" Oct 08 18:32:03 crc kubenswrapper[4988]: I1008 18:32:03.904560 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-92jjg" Oct 08 18:32:04 crc kubenswrapper[4988]: I1008 18:32:04.342501 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ccd2f8fa-1c16-4c58-9f87-e60ba3a2f5fb","Type":"ContainerStarted","Data":"be90eff6d6a94661069208a3760ad429da6e736c3f64018e09b59d748a843d33"} Oct 08 18:32:04 crc kubenswrapper[4988]: I1008 18:32:04.342966 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ccd2f8fa-1c16-4c58-9f87-e60ba3a2f5fb","Type":"ContainerStarted","Data":"06f5039b59f5eef23b916412ccb9a5658d672fcb48cf3d4a8e55e5a55e9519c0"} Oct 08 18:32:04 crc kubenswrapper[4988]: I1008 18:32:04.344965 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"38babc76-590a-423e-b104-46e0e58fd3cc","Type":"ContainerStarted","Data":"59b7126406dfff05a4bdc7b64861e9856b465d2f98514267dc274671417640dc"} Oct 08 18:32:04 crc kubenswrapper[4988]: I1008 18:32:04.344993 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"38babc76-590a-423e-b104-46e0e58fd3cc","Type":"ContainerStarted","Data":"2fdad3ce1cf67b244156aa498c775fba81c6157100710cf51356b7373addcde5"} Oct 08 18:32:04 crc kubenswrapper[4988]: I1008 18:32:04.368783 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.368759153 podStartE2EDuration="2.368759153s" podCreationTimestamp="2025-10-08 18:32:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 18:32:04.361759799 +0000 UTC m=+1269.811602609" watchObservedRunningTime="2025-10-08 18:32:04.368759153 +0000 UTC m=+1269.818601943" Oct 08 18:32:04 crc kubenswrapper[4988]: W1008 18:32:04.406555 4988 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc132f5da_fe6a_4616_9d2f_bc0c046244a1.slice/crio-9a0f75fe7471f48aa300108b7062fdc46cf5df92b5edb0ac5fc5937494c99621 WatchSource:0}: Error finding container 9a0f75fe7471f48aa300108b7062fdc46cf5df92b5edb0ac5fc5937494c99621: Status 404 returned error can't find the container with id 9a0f75fe7471f48aa300108b7062fdc46cf5df92b5edb0ac5fc5937494c99621 Oct 08 18:32:04 crc kubenswrapper[4988]: I1008 18:32:04.410255 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-92jjg"] Oct 08 18:32:05 crc kubenswrapper[4988]: I1008 18:32:05.355511 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-92jjg" event={"ID":"c132f5da-fe6a-4616-9d2f-bc0c046244a1","Type":"ContainerStarted","Data":"85a5febf4bb240fccaef0fc62af632ef9fb88c921ce282445ce2e5cfef1289b0"} Oct 08 18:32:05 crc kubenswrapper[4988]: I1008 18:32:05.355850 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-92jjg" event={"ID":"c132f5da-fe6a-4616-9d2f-bc0c046244a1","Type":"ContainerStarted","Data":"9a0f75fe7471f48aa300108b7062fdc46cf5df92b5edb0ac5fc5937494c99621"} Oct 08 18:32:05 crc kubenswrapper[4988]: I1008 18:32:05.373480 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-cell-mapping-92jjg" podStartSLOduration=2.373464226 podStartE2EDuration="2.373464226s" podCreationTimestamp="2025-10-08 18:32:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 18:32:05.368035633 +0000 UTC m=+1270.817878423" watchObservedRunningTime="2025-10-08 18:32:05.373464226 +0000 UTC m=+1270.823306996" Oct 08 18:32:05 crc kubenswrapper[4988]: I1008 18:32:05.758701 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-5967cc9597-r6jrv" Oct 08 18:32:05 crc kubenswrapper[4988]: I1008 18:32:05.825766 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-64d8d96789-8bhgs"] Oct 08 18:32:05 crc kubenswrapper[4988]: I1008 18:32:05.830081 4988 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-64d8d96789-8bhgs" podUID="d8520f38-f3e7-4a05-ac90-ed48db1b2ba0" containerName="dnsmasq-dns" containerID="cri-o://e73abd7244e715c31f953913bd7ab824186e18b2167e4cab3bfaa5f8522ce848" gracePeriod=10 Oct 08 18:32:06 crc kubenswrapper[4988]: E1008 18:32:06.320758 4988 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd8520f38_f3e7_4a05_ac90_ed48db1b2ba0.slice/crio-conmon-e73abd7244e715c31f953913bd7ab824186e18b2167e4cab3bfaa5f8522ce848.scope\": RecentStats: unable to find data in memory cache]" Oct 08 18:32:06 crc kubenswrapper[4988]: I1008 18:32:06.388485 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 08 18:32:06 crc kubenswrapper[4988]: I1008 18:32:06.388945 4988 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-64d8d96789-8bhgs" podUID="d8520f38-f3e7-4a05-ac90-ed48db1b2ba0" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.188:5353: connect: connection refused" Oct 08 18:32:06 crc kubenswrapper[4988]: I1008 18:32:06.391839 4988 generic.go:334] "Generic (PLEG): container finished" podID="d8520f38-f3e7-4a05-ac90-ed48db1b2ba0" containerID="e73abd7244e715c31f953913bd7ab824186e18b2167e4cab3bfaa5f8522ce848" exitCode=0 Oct 08 18:32:06 crc kubenswrapper[4988]: I1008 18:32:06.392473 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-64d8d96789-8bhgs" event={"ID":"d8520f38-f3e7-4a05-ac90-ed48db1b2ba0","Type":"ContainerDied","Data":"e73abd7244e715c31f953913bd7ab824186e18b2167e4cab3bfaa5f8522ce848"} Oct 08 18:32:06 crc kubenswrapper[4988]: I1008 18:32:06.412929 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.338366315 podStartE2EDuration="6.412911703s" podCreationTimestamp="2025-10-08 18:32:00 +0000 UTC" firstStartedPulling="2025-10-08 18:32:01.769089179 +0000 UTC m=+1267.218931949" lastFinishedPulling="2025-10-08 18:32:05.843634567 +0000 UTC m=+1271.293477337" observedRunningTime="2025-10-08 18:32:06.405202996 +0000 UTC m=+1271.855045766" watchObservedRunningTime="2025-10-08 18:32:06.412911703 +0000 UTC m=+1271.862754473" Oct 08 18:32:06 crc kubenswrapper[4988]: I1008 18:32:06.804841 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-64d8d96789-8bhgs" Oct 08 18:32:06 crc kubenswrapper[4988]: I1008 18:32:06.934901 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d8520f38-f3e7-4a05-ac90-ed48db1b2ba0-dns-svc\") pod \"d8520f38-f3e7-4a05-ac90-ed48db1b2ba0\" (UID: \"d8520f38-f3e7-4a05-ac90-ed48db1b2ba0\") " Oct 08 18:32:06 crc kubenswrapper[4988]: I1008 18:32:06.934971 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d8520f38-f3e7-4a05-ac90-ed48db1b2ba0-ovsdbserver-nb\") pod \"d8520f38-f3e7-4a05-ac90-ed48db1b2ba0\" (UID: \"d8520f38-f3e7-4a05-ac90-ed48db1b2ba0\") " Oct 08 18:32:06 crc kubenswrapper[4988]: I1008 18:32:06.935110 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d8520f38-f3e7-4a05-ac90-ed48db1b2ba0-config\") pod \"d8520f38-f3e7-4a05-ac90-ed48db1b2ba0\" (UID: \"d8520f38-f3e7-4a05-ac90-ed48db1b2ba0\") " Oct 08 18:32:06 crc kubenswrapper[4988]: I1008 18:32:06.935161 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rrgsq\" (UniqueName: \"kubernetes.io/projected/d8520f38-f3e7-4a05-ac90-ed48db1b2ba0-kube-api-access-rrgsq\") pod \"d8520f38-f3e7-4a05-ac90-ed48db1b2ba0\" (UID: \"d8520f38-f3e7-4a05-ac90-ed48db1b2ba0\") " Oct 08 18:32:06 crc kubenswrapper[4988]: I1008 18:32:06.935232 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d8520f38-f3e7-4a05-ac90-ed48db1b2ba0-ovsdbserver-sb\") pod \"d8520f38-f3e7-4a05-ac90-ed48db1b2ba0\" (UID: \"d8520f38-f3e7-4a05-ac90-ed48db1b2ba0\") " Oct 08 18:32:06 crc kubenswrapper[4988]: I1008 18:32:06.935257 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/d8520f38-f3e7-4a05-ac90-ed48db1b2ba0-dns-swift-storage-0\") pod \"d8520f38-f3e7-4a05-ac90-ed48db1b2ba0\" (UID: \"d8520f38-f3e7-4a05-ac90-ed48db1b2ba0\") " Oct 08 18:32:06 crc kubenswrapper[4988]: I1008 18:32:06.943566 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d8520f38-f3e7-4a05-ac90-ed48db1b2ba0-kube-api-access-rrgsq" (OuterVolumeSpecName: "kube-api-access-rrgsq") pod "d8520f38-f3e7-4a05-ac90-ed48db1b2ba0" (UID: "d8520f38-f3e7-4a05-ac90-ed48db1b2ba0"). InnerVolumeSpecName "kube-api-access-rrgsq". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:32:06 crc kubenswrapper[4988]: I1008 18:32:06.989929 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d8520f38-f3e7-4a05-ac90-ed48db1b2ba0-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "d8520f38-f3e7-4a05-ac90-ed48db1b2ba0" (UID: "d8520f38-f3e7-4a05-ac90-ed48db1b2ba0"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 18:32:06 crc kubenswrapper[4988]: I1008 18:32:06.994660 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d8520f38-f3e7-4a05-ac90-ed48db1b2ba0-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "d8520f38-f3e7-4a05-ac90-ed48db1b2ba0" (UID: "d8520f38-f3e7-4a05-ac90-ed48db1b2ba0"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 18:32:06 crc kubenswrapper[4988]: I1008 18:32:06.996459 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d8520f38-f3e7-4a05-ac90-ed48db1b2ba0-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "d8520f38-f3e7-4a05-ac90-ed48db1b2ba0" (UID: "d8520f38-f3e7-4a05-ac90-ed48db1b2ba0"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 18:32:06 crc kubenswrapper[4988]: I1008 18:32:06.999493 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d8520f38-f3e7-4a05-ac90-ed48db1b2ba0-config" (OuterVolumeSpecName: "config") pod "d8520f38-f3e7-4a05-ac90-ed48db1b2ba0" (UID: "d8520f38-f3e7-4a05-ac90-ed48db1b2ba0"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 18:32:07 crc kubenswrapper[4988]: I1008 18:32:07.001431 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d8520f38-f3e7-4a05-ac90-ed48db1b2ba0-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "d8520f38-f3e7-4a05-ac90-ed48db1b2ba0" (UID: "d8520f38-f3e7-4a05-ac90-ed48db1b2ba0"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 18:32:07 crc kubenswrapper[4988]: I1008 18:32:07.037634 4988 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rrgsq\" (UniqueName: \"kubernetes.io/projected/d8520f38-f3e7-4a05-ac90-ed48db1b2ba0-kube-api-access-rrgsq\") on node \"crc\" DevicePath \"\"" Oct 08 18:32:07 crc kubenswrapper[4988]: I1008 18:32:07.037668 4988 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d8520f38-f3e7-4a05-ac90-ed48db1b2ba0-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 08 18:32:07 crc kubenswrapper[4988]: I1008 18:32:07.037703 4988 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/d8520f38-f3e7-4a05-ac90-ed48db1b2ba0-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Oct 08 18:32:07 crc kubenswrapper[4988]: I1008 18:32:07.037717 4988 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d8520f38-f3e7-4a05-ac90-ed48db1b2ba0-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 08 18:32:07 crc kubenswrapper[4988]: I1008 18:32:07.037725 4988 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d8520f38-f3e7-4a05-ac90-ed48db1b2ba0-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 08 18:32:07 crc kubenswrapper[4988]: I1008 18:32:07.037733 4988 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d8520f38-f3e7-4a05-ac90-ed48db1b2ba0-config\") on node \"crc\" DevicePath \"\"" Oct 08 18:32:07 crc kubenswrapper[4988]: I1008 18:32:07.404376 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-64d8d96789-8bhgs" event={"ID":"d8520f38-f3e7-4a05-ac90-ed48db1b2ba0","Type":"ContainerDied","Data":"13baa2b9d20cd176d4848124d8f8a2a626e9072d1f3957f1137263212c3dfdd4"} Oct 08 18:32:07 crc kubenswrapper[4988]: I1008 18:32:07.404450 4988 scope.go:117] "RemoveContainer" containerID="e73abd7244e715c31f953913bd7ab824186e18b2167e4cab3bfaa5f8522ce848" Oct 08 18:32:07 crc kubenswrapper[4988]: I1008 18:32:07.404591 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-64d8d96789-8bhgs" Oct 08 18:32:07 crc kubenswrapper[4988]: I1008 18:32:07.411759 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ccd2f8fa-1c16-4c58-9f87-e60ba3a2f5fb","Type":"ContainerStarted","Data":"f64fdd9df8d4aff7bcd8cfab3af523fa24041d42ea9cfb0addd7bf856026fe76"} Oct 08 18:32:07 crc kubenswrapper[4988]: I1008 18:32:07.437289 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-64d8d96789-8bhgs"] Oct 08 18:32:07 crc kubenswrapper[4988]: I1008 18:32:07.444901 4988 scope.go:117] "RemoveContainer" containerID="290de12fdada84cf292143bedd05a23659d15e3a50f137b76d4f26b1c2aad8af" Oct 08 18:32:07 crc kubenswrapper[4988]: I1008 18:32:07.450698 4988 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-64d8d96789-8bhgs"] Oct 08 18:32:09 crc kubenswrapper[4988]: I1008 18:32:09.253627 4988 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d8520f38-f3e7-4a05-ac90-ed48db1b2ba0" path="/var/lib/kubelet/pods/d8520f38-f3e7-4a05-ac90-ed48db1b2ba0/volumes" Oct 08 18:32:09 crc kubenswrapper[4988]: I1008 18:32:09.438817 4988 generic.go:334] "Generic (PLEG): container finished" podID="c132f5da-fe6a-4616-9d2f-bc0c046244a1" containerID="85a5febf4bb240fccaef0fc62af632ef9fb88c921ce282445ce2e5cfef1289b0" exitCode=0 Oct 08 18:32:09 crc kubenswrapper[4988]: I1008 18:32:09.438907 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-92jjg" event={"ID":"c132f5da-fe6a-4616-9d2f-bc0c046244a1","Type":"ContainerDied","Data":"85a5febf4bb240fccaef0fc62af632ef9fb88c921ce282445ce2e5cfef1289b0"} Oct 08 18:32:10 crc kubenswrapper[4988]: I1008 18:32:10.824095 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-92jjg" Oct 08 18:32:11 crc kubenswrapper[4988]: I1008 18:32:11.010960 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dggs9\" (UniqueName: \"kubernetes.io/projected/c132f5da-fe6a-4616-9d2f-bc0c046244a1-kube-api-access-dggs9\") pod \"c132f5da-fe6a-4616-9d2f-bc0c046244a1\" (UID: \"c132f5da-fe6a-4616-9d2f-bc0c046244a1\") " Oct 08 18:32:11 crc kubenswrapper[4988]: I1008 18:32:11.011770 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c132f5da-fe6a-4616-9d2f-bc0c046244a1-combined-ca-bundle\") pod \"c132f5da-fe6a-4616-9d2f-bc0c046244a1\" (UID: \"c132f5da-fe6a-4616-9d2f-bc0c046244a1\") " Oct 08 18:32:11 crc kubenswrapper[4988]: I1008 18:32:11.011936 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c132f5da-fe6a-4616-9d2f-bc0c046244a1-scripts\") pod \"c132f5da-fe6a-4616-9d2f-bc0c046244a1\" (UID: \"c132f5da-fe6a-4616-9d2f-bc0c046244a1\") " Oct 08 18:32:11 crc kubenswrapper[4988]: I1008 18:32:11.012111 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c132f5da-fe6a-4616-9d2f-bc0c046244a1-config-data\") pod \"c132f5da-fe6a-4616-9d2f-bc0c046244a1\" (UID: \"c132f5da-fe6a-4616-9d2f-bc0c046244a1\") " Oct 08 18:32:11 crc kubenswrapper[4988]: I1008 18:32:11.017477 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c132f5da-fe6a-4616-9d2f-bc0c046244a1-scripts" (OuterVolumeSpecName: "scripts") pod "c132f5da-fe6a-4616-9d2f-bc0c046244a1" (UID: "c132f5da-fe6a-4616-9d2f-bc0c046244a1"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:32:11 crc kubenswrapper[4988]: I1008 18:32:11.025093 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c132f5da-fe6a-4616-9d2f-bc0c046244a1-kube-api-access-dggs9" (OuterVolumeSpecName: "kube-api-access-dggs9") pod "c132f5da-fe6a-4616-9d2f-bc0c046244a1" (UID: "c132f5da-fe6a-4616-9d2f-bc0c046244a1"). InnerVolumeSpecName "kube-api-access-dggs9". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:32:11 crc kubenswrapper[4988]: I1008 18:32:11.043633 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c132f5da-fe6a-4616-9d2f-bc0c046244a1-config-data" (OuterVolumeSpecName: "config-data") pod "c132f5da-fe6a-4616-9d2f-bc0c046244a1" (UID: "c132f5da-fe6a-4616-9d2f-bc0c046244a1"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:32:11 crc kubenswrapper[4988]: I1008 18:32:11.066257 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c132f5da-fe6a-4616-9d2f-bc0c046244a1-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "c132f5da-fe6a-4616-9d2f-bc0c046244a1" (UID: "c132f5da-fe6a-4616-9d2f-bc0c046244a1"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:32:11 crc kubenswrapper[4988]: I1008 18:32:11.114651 4988 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c132f5da-fe6a-4616-9d2f-bc0c046244a1-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 18:32:11 crc kubenswrapper[4988]: I1008 18:32:11.114718 4988 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c132f5da-fe6a-4616-9d2f-bc0c046244a1-scripts\") on node \"crc\" DevicePath \"\"" Oct 08 18:32:11 crc kubenswrapper[4988]: I1008 18:32:11.114731 4988 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c132f5da-fe6a-4616-9d2f-bc0c046244a1-config-data\") on node \"crc\" DevicePath \"\"" Oct 08 18:32:11 crc kubenswrapper[4988]: I1008 18:32:11.114743 4988 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dggs9\" (UniqueName: \"kubernetes.io/projected/c132f5da-fe6a-4616-9d2f-bc0c046244a1-kube-api-access-dggs9\") on node \"crc\" DevicePath \"\"" Oct 08 18:32:11 crc kubenswrapper[4988]: I1008 18:32:11.461820 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-92jjg" event={"ID":"c132f5da-fe6a-4616-9d2f-bc0c046244a1","Type":"ContainerDied","Data":"9a0f75fe7471f48aa300108b7062fdc46cf5df92b5edb0ac5fc5937494c99621"} Oct 08 18:32:11 crc kubenswrapper[4988]: I1008 18:32:11.461869 4988 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9a0f75fe7471f48aa300108b7062fdc46cf5df92b5edb0ac5fc5937494c99621" Oct 08 18:32:11 crc kubenswrapper[4988]: I1008 18:32:11.461881 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-92jjg" Oct 08 18:32:11 crc kubenswrapper[4988]: I1008 18:32:11.664928 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 08 18:32:11 crc kubenswrapper[4988]: I1008 18:32:11.665673 4988 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="38babc76-590a-423e-b104-46e0e58fd3cc" containerName="nova-api-log" containerID="cri-o://2fdad3ce1cf67b244156aa498c775fba81c6157100710cf51356b7373addcde5" gracePeriod=30 Oct 08 18:32:11 crc kubenswrapper[4988]: I1008 18:32:11.665699 4988 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="38babc76-590a-423e-b104-46e0e58fd3cc" containerName="nova-api-api" containerID="cri-o://59b7126406dfff05a4bdc7b64861e9856b465d2f98514267dc274671417640dc" gracePeriod=30 Oct 08 18:32:11 crc kubenswrapper[4988]: I1008 18:32:11.683290 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Oct 08 18:32:11 crc kubenswrapper[4988]: I1008 18:32:11.683559 4988 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="be97711d-1138-4866-a82d-c0ddbcec58c8" containerName="nova-scheduler-scheduler" containerID="cri-o://3f453b3226c2919bd0521ef15a3bf5e939375ac3f298ac9733160689d3fda89b" gracePeriod=30 Oct 08 18:32:11 crc kubenswrapper[4988]: I1008 18:32:11.697056 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Oct 08 18:32:11 crc kubenswrapper[4988]: I1008 18:32:11.697346 4988 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="c51f2710-19e3-457c-ab1b-6c7e19877213" containerName="nova-metadata-log" containerID="cri-o://15a065b5102872a8fc42dc14c1114cd59430e0c88890fc13c4496e3e17a3c6bc" gracePeriod=30 Oct 08 18:32:11 crc kubenswrapper[4988]: I1008 18:32:11.697492 4988 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="c51f2710-19e3-457c-ab1b-6c7e19877213" containerName="nova-metadata-metadata" containerID="cri-o://8673b565fae2659ae04aabd4852445f45c49c3733848e2571f9ff5b470e6a680" gracePeriod=30 Oct 08 18:32:13 crc kubenswrapper[4988]: I1008 18:32:12.239215 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 08 18:32:13 crc kubenswrapper[4988]: I1008 18:32:12.335069 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/38babc76-590a-423e-b104-46e0e58fd3cc-internal-tls-certs\") pod \"38babc76-590a-423e-b104-46e0e58fd3cc\" (UID: \"38babc76-590a-423e-b104-46e0e58fd3cc\") " Oct 08 18:32:13 crc kubenswrapper[4988]: I1008 18:32:12.335246 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/38babc76-590a-423e-b104-46e0e58fd3cc-config-data\") pod \"38babc76-590a-423e-b104-46e0e58fd3cc\" (UID: \"38babc76-590a-423e-b104-46e0e58fd3cc\") " Oct 08 18:32:13 crc kubenswrapper[4988]: I1008 18:32:12.335325 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/38babc76-590a-423e-b104-46e0e58fd3cc-logs\") pod \"38babc76-590a-423e-b104-46e0e58fd3cc\" (UID: \"38babc76-590a-423e-b104-46e0e58fd3cc\") " Oct 08 18:32:13 crc kubenswrapper[4988]: I1008 18:32:12.335469 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/38babc76-590a-423e-b104-46e0e58fd3cc-public-tls-certs\") pod \"38babc76-590a-423e-b104-46e0e58fd3cc\" (UID: \"38babc76-590a-423e-b104-46e0e58fd3cc\") " Oct 08 18:32:13 crc kubenswrapper[4988]: I1008 18:32:12.335509 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/38babc76-590a-423e-b104-46e0e58fd3cc-combined-ca-bundle\") pod \"38babc76-590a-423e-b104-46e0e58fd3cc\" (UID: \"38babc76-590a-423e-b104-46e0e58fd3cc\") " Oct 08 18:32:13 crc kubenswrapper[4988]: I1008 18:32:12.335623 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-45xm7\" (UniqueName: \"kubernetes.io/projected/38babc76-590a-423e-b104-46e0e58fd3cc-kube-api-access-45xm7\") pod \"38babc76-590a-423e-b104-46e0e58fd3cc\" (UID: \"38babc76-590a-423e-b104-46e0e58fd3cc\") " Oct 08 18:32:13 crc kubenswrapper[4988]: I1008 18:32:12.336059 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/38babc76-590a-423e-b104-46e0e58fd3cc-logs" (OuterVolumeSpecName: "logs") pod "38babc76-590a-423e-b104-46e0e58fd3cc" (UID: "38babc76-590a-423e-b104-46e0e58fd3cc"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 18:32:13 crc kubenswrapper[4988]: I1008 18:32:12.336937 4988 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/38babc76-590a-423e-b104-46e0e58fd3cc-logs\") on node \"crc\" DevicePath \"\"" Oct 08 18:32:13 crc kubenswrapper[4988]: I1008 18:32:12.345316 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/38babc76-590a-423e-b104-46e0e58fd3cc-kube-api-access-45xm7" (OuterVolumeSpecName: "kube-api-access-45xm7") pod "38babc76-590a-423e-b104-46e0e58fd3cc" (UID: "38babc76-590a-423e-b104-46e0e58fd3cc"). InnerVolumeSpecName "kube-api-access-45xm7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:32:13 crc kubenswrapper[4988]: I1008 18:32:12.368133 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/38babc76-590a-423e-b104-46e0e58fd3cc-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "38babc76-590a-423e-b104-46e0e58fd3cc" (UID: "38babc76-590a-423e-b104-46e0e58fd3cc"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:32:13 crc kubenswrapper[4988]: I1008 18:32:12.373959 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/38babc76-590a-423e-b104-46e0e58fd3cc-config-data" (OuterVolumeSpecName: "config-data") pod "38babc76-590a-423e-b104-46e0e58fd3cc" (UID: "38babc76-590a-423e-b104-46e0e58fd3cc"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:32:13 crc kubenswrapper[4988]: I1008 18:32:12.394688 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/38babc76-590a-423e-b104-46e0e58fd3cc-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "38babc76-590a-423e-b104-46e0e58fd3cc" (UID: "38babc76-590a-423e-b104-46e0e58fd3cc"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:32:13 crc kubenswrapper[4988]: I1008 18:32:12.403081 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/38babc76-590a-423e-b104-46e0e58fd3cc-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "38babc76-590a-423e-b104-46e0e58fd3cc" (UID: "38babc76-590a-423e-b104-46e0e58fd3cc"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:32:13 crc kubenswrapper[4988]: I1008 18:32:12.439212 4988 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/38babc76-590a-423e-b104-46e0e58fd3cc-public-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 08 18:32:13 crc kubenswrapper[4988]: I1008 18:32:12.439247 4988 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/38babc76-590a-423e-b104-46e0e58fd3cc-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 18:32:13 crc kubenswrapper[4988]: I1008 18:32:12.439258 4988 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-45xm7\" (UniqueName: \"kubernetes.io/projected/38babc76-590a-423e-b104-46e0e58fd3cc-kube-api-access-45xm7\") on node \"crc\" DevicePath \"\"" Oct 08 18:32:13 crc kubenswrapper[4988]: I1008 18:32:12.439272 4988 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/38babc76-590a-423e-b104-46e0e58fd3cc-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 08 18:32:13 crc kubenswrapper[4988]: I1008 18:32:12.439280 4988 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/38babc76-590a-423e-b104-46e0e58fd3cc-config-data\") on node \"crc\" DevicePath \"\"" Oct 08 18:32:13 crc kubenswrapper[4988]: I1008 18:32:12.475458 4988 generic.go:334] "Generic (PLEG): container finished" podID="c51f2710-19e3-457c-ab1b-6c7e19877213" containerID="15a065b5102872a8fc42dc14c1114cd59430e0c88890fc13c4496e3e17a3c6bc" exitCode=143 Oct 08 18:32:13 crc kubenswrapper[4988]: I1008 18:32:12.475540 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"c51f2710-19e3-457c-ab1b-6c7e19877213","Type":"ContainerDied","Data":"15a065b5102872a8fc42dc14c1114cd59430e0c88890fc13c4496e3e17a3c6bc"} Oct 08 18:32:13 crc kubenswrapper[4988]: I1008 18:32:12.478294 4988 generic.go:334] "Generic (PLEG): container finished" podID="38babc76-590a-423e-b104-46e0e58fd3cc" containerID="59b7126406dfff05a4bdc7b64861e9856b465d2f98514267dc274671417640dc" exitCode=0 Oct 08 18:32:13 crc kubenswrapper[4988]: I1008 18:32:12.478316 4988 generic.go:334] "Generic (PLEG): container finished" podID="38babc76-590a-423e-b104-46e0e58fd3cc" containerID="2fdad3ce1cf67b244156aa498c775fba81c6157100710cf51356b7373addcde5" exitCode=143 Oct 08 18:32:13 crc kubenswrapper[4988]: I1008 18:32:12.478338 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"38babc76-590a-423e-b104-46e0e58fd3cc","Type":"ContainerDied","Data":"59b7126406dfff05a4bdc7b64861e9856b465d2f98514267dc274671417640dc"} Oct 08 18:32:13 crc kubenswrapper[4988]: I1008 18:32:12.478371 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"38babc76-590a-423e-b104-46e0e58fd3cc","Type":"ContainerDied","Data":"2fdad3ce1cf67b244156aa498c775fba81c6157100710cf51356b7373addcde5"} Oct 08 18:32:13 crc kubenswrapper[4988]: I1008 18:32:12.478389 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"38babc76-590a-423e-b104-46e0e58fd3cc","Type":"ContainerDied","Data":"958dbdf57bb6d3003dfe38be859959524bf9bccc068bf1735c1789240e600430"} Oct 08 18:32:13 crc kubenswrapper[4988]: I1008 18:32:12.478404 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 08 18:32:13 crc kubenswrapper[4988]: I1008 18:32:12.478425 4988 scope.go:117] "RemoveContainer" containerID="59b7126406dfff05a4bdc7b64861e9856b465d2f98514267dc274671417640dc" Oct 08 18:32:13 crc kubenswrapper[4988]: I1008 18:32:12.521133 4988 scope.go:117] "RemoveContainer" containerID="2fdad3ce1cf67b244156aa498c775fba81c6157100710cf51356b7373addcde5" Oct 08 18:32:13 crc kubenswrapper[4988]: I1008 18:32:12.527695 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 08 18:32:13 crc kubenswrapper[4988]: I1008 18:32:12.538922 4988 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Oct 08 18:32:13 crc kubenswrapper[4988]: I1008 18:32:12.552262 4988 scope.go:117] "RemoveContainer" containerID="59b7126406dfff05a4bdc7b64861e9856b465d2f98514267dc274671417640dc" Oct 08 18:32:13 crc kubenswrapper[4988]: E1008 18:32:12.552713 4988 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"59b7126406dfff05a4bdc7b64861e9856b465d2f98514267dc274671417640dc\": container with ID starting with 59b7126406dfff05a4bdc7b64861e9856b465d2f98514267dc274671417640dc not found: ID does not exist" containerID="59b7126406dfff05a4bdc7b64861e9856b465d2f98514267dc274671417640dc" Oct 08 18:32:13 crc kubenswrapper[4988]: I1008 18:32:12.552744 4988 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"59b7126406dfff05a4bdc7b64861e9856b465d2f98514267dc274671417640dc"} err="failed to get container status \"59b7126406dfff05a4bdc7b64861e9856b465d2f98514267dc274671417640dc\": rpc error: code = NotFound desc = could not find container \"59b7126406dfff05a4bdc7b64861e9856b465d2f98514267dc274671417640dc\": container with ID starting with 59b7126406dfff05a4bdc7b64861e9856b465d2f98514267dc274671417640dc not found: ID does not exist" Oct 08 18:32:13 crc kubenswrapper[4988]: I1008 18:32:12.552766 4988 scope.go:117] "RemoveContainer" containerID="2fdad3ce1cf67b244156aa498c775fba81c6157100710cf51356b7373addcde5" Oct 08 18:32:13 crc kubenswrapper[4988]: E1008 18:32:12.553008 4988 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2fdad3ce1cf67b244156aa498c775fba81c6157100710cf51356b7373addcde5\": container with ID starting with 2fdad3ce1cf67b244156aa498c775fba81c6157100710cf51356b7373addcde5 not found: ID does not exist" containerID="2fdad3ce1cf67b244156aa498c775fba81c6157100710cf51356b7373addcde5" Oct 08 18:32:13 crc kubenswrapper[4988]: I1008 18:32:12.553046 4988 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2fdad3ce1cf67b244156aa498c775fba81c6157100710cf51356b7373addcde5"} err="failed to get container status \"2fdad3ce1cf67b244156aa498c775fba81c6157100710cf51356b7373addcde5\": rpc error: code = NotFound desc = could not find container \"2fdad3ce1cf67b244156aa498c775fba81c6157100710cf51356b7373addcde5\": container with ID starting with 2fdad3ce1cf67b244156aa498c775fba81c6157100710cf51356b7373addcde5 not found: ID does not exist" Oct 08 18:32:13 crc kubenswrapper[4988]: I1008 18:32:12.553059 4988 scope.go:117] "RemoveContainer" containerID="59b7126406dfff05a4bdc7b64861e9856b465d2f98514267dc274671417640dc" Oct 08 18:32:13 crc kubenswrapper[4988]: I1008 18:32:12.554153 4988 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"59b7126406dfff05a4bdc7b64861e9856b465d2f98514267dc274671417640dc"} err="failed to get container status \"59b7126406dfff05a4bdc7b64861e9856b465d2f98514267dc274671417640dc\": rpc error: code = NotFound desc = could not find container \"59b7126406dfff05a4bdc7b64861e9856b465d2f98514267dc274671417640dc\": container with ID starting with 59b7126406dfff05a4bdc7b64861e9856b465d2f98514267dc274671417640dc not found: ID does not exist" Oct 08 18:32:13 crc kubenswrapper[4988]: I1008 18:32:12.554172 4988 scope.go:117] "RemoveContainer" containerID="2fdad3ce1cf67b244156aa498c775fba81c6157100710cf51356b7373addcde5" Oct 08 18:32:13 crc kubenswrapper[4988]: I1008 18:32:12.557538 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Oct 08 18:32:13 crc kubenswrapper[4988]: E1008 18:32:12.558326 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d8520f38-f3e7-4a05-ac90-ed48db1b2ba0" containerName="init" Oct 08 18:32:13 crc kubenswrapper[4988]: I1008 18:32:12.558352 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="d8520f38-f3e7-4a05-ac90-ed48db1b2ba0" containerName="init" Oct 08 18:32:13 crc kubenswrapper[4988]: E1008 18:32:12.558372 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d8520f38-f3e7-4a05-ac90-ed48db1b2ba0" containerName="dnsmasq-dns" Oct 08 18:32:13 crc kubenswrapper[4988]: I1008 18:32:12.558381 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="d8520f38-f3e7-4a05-ac90-ed48db1b2ba0" containerName="dnsmasq-dns" Oct 08 18:32:13 crc kubenswrapper[4988]: E1008 18:32:12.558428 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c132f5da-fe6a-4616-9d2f-bc0c046244a1" containerName="nova-manage" Oct 08 18:32:13 crc kubenswrapper[4988]: I1008 18:32:12.558442 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="c132f5da-fe6a-4616-9d2f-bc0c046244a1" containerName="nova-manage" Oct 08 18:32:13 crc kubenswrapper[4988]: E1008 18:32:12.558474 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="38babc76-590a-423e-b104-46e0e58fd3cc" containerName="nova-api-log" Oct 08 18:32:13 crc kubenswrapper[4988]: I1008 18:32:12.558482 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="38babc76-590a-423e-b104-46e0e58fd3cc" containerName="nova-api-log" Oct 08 18:32:13 crc kubenswrapper[4988]: E1008 18:32:12.558501 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="38babc76-590a-423e-b104-46e0e58fd3cc" containerName="nova-api-api" Oct 08 18:32:13 crc kubenswrapper[4988]: I1008 18:32:12.558508 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="38babc76-590a-423e-b104-46e0e58fd3cc" containerName="nova-api-api" Oct 08 18:32:13 crc kubenswrapper[4988]: I1008 18:32:12.558801 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="38babc76-590a-423e-b104-46e0e58fd3cc" containerName="nova-api-log" Oct 08 18:32:13 crc kubenswrapper[4988]: I1008 18:32:12.558823 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="c132f5da-fe6a-4616-9d2f-bc0c046244a1" containerName="nova-manage" Oct 08 18:32:13 crc kubenswrapper[4988]: I1008 18:32:12.558837 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="38babc76-590a-423e-b104-46e0e58fd3cc" containerName="nova-api-api" Oct 08 18:32:13 crc kubenswrapper[4988]: I1008 18:32:12.558852 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="d8520f38-f3e7-4a05-ac90-ed48db1b2ba0" containerName="dnsmasq-dns" Oct 08 18:32:13 crc kubenswrapper[4988]: I1008 18:32:12.560374 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 08 18:32:13 crc kubenswrapper[4988]: I1008 18:32:12.563118 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Oct 08 18:32:13 crc kubenswrapper[4988]: I1008 18:32:12.563537 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-public-svc" Oct 08 18:32:13 crc kubenswrapper[4988]: I1008 18:32:12.563635 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-internal-svc" Oct 08 18:32:13 crc kubenswrapper[4988]: I1008 18:32:12.567827 4988 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2fdad3ce1cf67b244156aa498c775fba81c6157100710cf51356b7373addcde5"} err="failed to get container status \"2fdad3ce1cf67b244156aa498c775fba81c6157100710cf51356b7373addcde5\": rpc error: code = NotFound desc = could not find container \"2fdad3ce1cf67b244156aa498c775fba81c6157100710cf51356b7373addcde5\": container with ID starting with 2fdad3ce1cf67b244156aa498c775fba81c6157100710cf51356b7373addcde5 not found: ID does not exist" Oct 08 18:32:13 crc kubenswrapper[4988]: I1008 18:32:12.573117 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 08 18:32:13 crc kubenswrapper[4988]: I1008 18:32:12.644056 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0efd1f53-90ad-4eb8-90f2-2ba9fc1593c2-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"0efd1f53-90ad-4eb8-90f2-2ba9fc1593c2\") " pod="openstack/nova-api-0" Oct 08 18:32:13 crc kubenswrapper[4988]: I1008 18:32:12.644239 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0efd1f53-90ad-4eb8-90f2-2ba9fc1593c2-config-data\") pod \"nova-api-0\" (UID: \"0efd1f53-90ad-4eb8-90f2-2ba9fc1593c2\") " pod="openstack/nova-api-0" Oct 08 18:32:13 crc kubenswrapper[4988]: I1008 18:32:12.644449 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/0efd1f53-90ad-4eb8-90f2-2ba9fc1593c2-public-tls-certs\") pod \"nova-api-0\" (UID: \"0efd1f53-90ad-4eb8-90f2-2ba9fc1593c2\") " pod="openstack/nova-api-0" Oct 08 18:32:13 crc kubenswrapper[4988]: I1008 18:32:12.644752 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0efd1f53-90ad-4eb8-90f2-2ba9fc1593c2-logs\") pod \"nova-api-0\" (UID: \"0efd1f53-90ad-4eb8-90f2-2ba9fc1593c2\") " pod="openstack/nova-api-0" Oct 08 18:32:13 crc kubenswrapper[4988]: I1008 18:32:12.644964 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bvpsb\" (UniqueName: \"kubernetes.io/projected/0efd1f53-90ad-4eb8-90f2-2ba9fc1593c2-kube-api-access-bvpsb\") pod \"nova-api-0\" (UID: \"0efd1f53-90ad-4eb8-90f2-2ba9fc1593c2\") " pod="openstack/nova-api-0" Oct 08 18:32:13 crc kubenswrapper[4988]: I1008 18:32:12.645162 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/0efd1f53-90ad-4eb8-90f2-2ba9fc1593c2-internal-tls-certs\") pod \"nova-api-0\" (UID: \"0efd1f53-90ad-4eb8-90f2-2ba9fc1593c2\") " pod="openstack/nova-api-0" Oct 08 18:32:13 crc kubenswrapper[4988]: I1008 18:32:12.747042 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0efd1f53-90ad-4eb8-90f2-2ba9fc1593c2-logs\") pod \"nova-api-0\" (UID: \"0efd1f53-90ad-4eb8-90f2-2ba9fc1593c2\") " pod="openstack/nova-api-0" Oct 08 18:32:13 crc kubenswrapper[4988]: I1008 18:32:12.747131 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bvpsb\" (UniqueName: \"kubernetes.io/projected/0efd1f53-90ad-4eb8-90f2-2ba9fc1593c2-kube-api-access-bvpsb\") pod \"nova-api-0\" (UID: \"0efd1f53-90ad-4eb8-90f2-2ba9fc1593c2\") " pod="openstack/nova-api-0" Oct 08 18:32:13 crc kubenswrapper[4988]: I1008 18:32:12.747193 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/0efd1f53-90ad-4eb8-90f2-2ba9fc1593c2-internal-tls-certs\") pod \"nova-api-0\" (UID: \"0efd1f53-90ad-4eb8-90f2-2ba9fc1593c2\") " pod="openstack/nova-api-0" Oct 08 18:32:13 crc kubenswrapper[4988]: I1008 18:32:12.747228 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0efd1f53-90ad-4eb8-90f2-2ba9fc1593c2-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"0efd1f53-90ad-4eb8-90f2-2ba9fc1593c2\") " pod="openstack/nova-api-0" Oct 08 18:32:13 crc kubenswrapper[4988]: I1008 18:32:12.747287 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0efd1f53-90ad-4eb8-90f2-2ba9fc1593c2-config-data\") pod \"nova-api-0\" (UID: \"0efd1f53-90ad-4eb8-90f2-2ba9fc1593c2\") " pod="openstack/nova-api-0" Oct 08 18:32:13 crc kubenswrapper[4988]: I1008 18:32:12.747368 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/0efd1f53-90ad-4eb8-90f2-2ba9fc1593c2-public-tls-certs\") pod \"nova-api-0\" (UID: \"0efd1f53-90ad-4eb8-90f2-2ba9fc1593c2\") " pod="openstack/nova-api-0" Oct 08 18:32:13 crc kubenswrapper[4988]: I1008 18:32:12.747824 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0efd1f53-90ad-4eb8-90f2-2ba9fc1593c2-logs\") pod \"nova-api-0\" (UID: \"0efd1f53-90ad-4eb8-90f2-2ba9fc1593c2\") " pod="openstack/nova-api-0" Oct 08 18:32:13 crc kubenswrapper[4988]: I1008 18:32:12.751672 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/0efd1f53-90ad-4eb8-90f2-2ba9fc1593c2-public-tls-certs\") pod \"nova-api-0\" (UID: \"0efd1f53-90ad-4eb8-90f2-2ba9fc1593c2\") " pod="openstack/nova-api-0" Oct 08 18:32:13 crc kubenswrapper[4988]: I1008 18:32:12.751924 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/0efd1f53-90ad-4eb8-90f2-2ba9fc1593c2-internal-tls-certs\") pod \"nova-api-0\" (UID: \"0efd1f53-90ad-4eb8-90f2-2ba9fc1593c2\") " pod="openstack/nova-api-0" Oct 08 18:32:13 crc kubenswrapper[4988]: I1008 18:32:12.752298 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0efd1f53-90ad-4eb8-90f2-2ba9fc1593c2-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"0efd1f53-90ad-4eb8-90f2-2ba9fc1593c2\") " pod="openstack/nova-api-0" Oct 08 18:32:13 crc kubenswrapper[4988]: I1008 18:32:12.754089 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0efd1f53-90ad-4eb8-90f2-2ba9fc1593c2-config-data\") pod \"nova-api-0\" (UID: \"0efd1f53-90ad-4eb8-90f2-2ba9fc1593c2\") " pod="openstack/nova-api-0" Oct 08 18:32:13 crc kubenswrapper[4988]: I1008 18:32:12.766593 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bvpsb\" (UniqueName: \"kubernetes.io/projected/0efd1f53-90ad-4eb8-90f2-2ba9fc1593c2-kube-api-access-bvpsb\") pod \"nova-api-0\" (UID: \"0efd1f53-90ad-4eb8-90f2-2ba9fc1593c2\") " pod="openstack/nova-api-0" Oct 08 18:32:13 crc kubenswrapper[4988]: I1008 18:32:12.883984 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 08 18:32:13 crc kubenswrapper[4988]: I1008 18:32:13.257010 4988 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="38babc76-590a-423e-b104-46e0e58fd3cc" path="/var/lib/kubelet/pods/38babc76-590a-423e-b104-46e0e58fd3cc/volumes" Oct 08 18:32:13 crc kubenswrapper[4988]: I1008 18:32:13.493260 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 08 18:32:13 crc kubenswrapper[4988]: I1008 18:32:13.898030 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 08 18:32:14 crc kubenswrapper[4988]: I1008 18:32:14.071313 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/be97711d-1138-4866-a82d-c0ddbcec58c8-config-data\") pod \"be97711d-1138-4866-a82d-c0ddbcec58c8\" (UID: \"be97711d-1138-4866-a82d-c0ddbcec58c8\") " Oct 08 18:32:14 crc kubenswrapper[4988]: I1008 18:32:14.071820 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/be97711d-1138-4866-a82d-c0ddbcec58c8-combined-ca-bundle\") pod \"be97711d-1138-4866-a82d-c0ddbcec58c8\" (UID: \"be97711d-1138-4866-a82d-c0ddbcec58c8\") " Oct 08 18:32:14 crc kubenswrapper[4988]: I1008 18:32:14.071925 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xgfbc\" (UniqueName: \"kubernetes.io/projected/be97711d-1138-4866-a82d-c0ddbcec58c8-kube-api-access-xgfbc\") pod \"be97711d-1138-4866-a82d-c0ddbcec58c8\" (UID: \"be97711d-1138-4866-a82d-c0ddbcec58c8\") " Oct 08 18:32:14 crc kubenswrapper[4988]: I1008 18:32:14.077530 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/be97711d-1138-4866-a82d-c0ddbcec58c8-kube-api-access-xgfbc" (OuterVolumeSpecName: "kube-api-access-xgfbc") pod "be97711d-1138-4866-a82d-c0ddbcec58c8" (UID: "be97711d-1138-4866-a82d-c0ddbcec58c8"). InnerVolumeSpecName "kube-api-access-xgfbc". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:32:14 crc kubenswrapper[4988]: I1008 18:32:14.111886 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/be97711d-1138-4866-a82d-c0ddbcec58c8-config-data" (OuterVolumeSpecName: "config-data") pod "be97711d-1138-4866-a82d-c0ddbcec58c8" (UID: "be97711d-1138-4866-a82d-c0ddbcec58c8"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:32:14 crc kubenswrapper[4988]: I1008 18:32:14.119653 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/be97711d-1138-4866-a82d-c0ddbcec58c8-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "be97711d-1138-4866-a82d-c0ddbcec58c8" (UID: "be97711d-1138-4866-a82d-c0ddbcec58c8"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:32:14 crc kubenswrapper[4988]: I1008 18:32:14.178909 4988 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xgfbc\" (UniqueName: \"kubernetes.io/projected/be97711d-1138-4866-a82d-c0ddbcec58c8-kube-api-access-xgfbc\") on node \"crc\" DevicePath \"\"" Oct 08 18:32:14 crc kubenswrapper[4988]: I1008 18:32:14.178961 4988 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/be97711d-1138-4866-a82d-c0ddbcec58c8-config-data\") on node \"crc\" DevicePath \"\"" Oct 08 18:32:14 crc kubenswrapper[4988]: I1008 18:32:14.178972 4988 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/be97711d-1138-4866-a82d-c0ddbcec58c8-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 18:32:14 crc kubenswrapper[4988]: I1008 18:32:14.504261 4988 generic.go:334] "Generic (PLEG): container finished" podID="be97711d-1138-4866-a82d-c0ddbcec58c8" containerID="3f453b3226c2919bd0521ef15a3bf5e939375ac3f298ac9733160689d3fda89b" exitCode=0 Oct 08 18:32:14 crc kubenswrapper[4988]: I1008 18:32:14.504331 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"be97711d-1138-4866-a82d-c0ddbcec58c8","Type":"ContainerDied","Data":"3f453b3226c2919bd0521ef15a3bf5e939375ac3f298ac9733160689d3fda89b"} Oct 08 18:32:14 crc kubenswrapper[4988]: I1008 18:32:14.504360 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"be97711d-1138-4866-a82d-c0ddbcec58c8","Type":"ContainerDied","Data":"36549afe6c5d0a01f18a6ff16f92b8eb840f2473ae9b0af84b6e4afcaf57e47c"} Oct 08 18:32:14 crc kubenswrapper[4988]: I1008 18:32:14.504378 4988 scope.go:117] "RemoveContainer" containerID="3f453b3226c2919bd0521ef15a3bf5e939375ac3f298ac9733160689d3fda89b" Oct 08 18:32:14 crc kubenswrapper[4988]: I1008 18:32:14.504431 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 08 18:32:14 crc kubenswrapper[4988]: I1008 18:32:14.508022 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"0efd1f53-90ad-4eb8-90f2-2ba9fc1593c2","Type":"ContainerStarted","Data":"a81aba18e5412a571b8ceb649f5e04c6698ca4b21c30f54d96848df1d1a823b6"} Oct 08 18:32:14 crc kubenswrapper[4988]: I1008 18:32:14.508060 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"0efd1f53-90ad-4eb8-90f2-2ba9fc1593c2","Type":"ContainerStarted","Data":"4cdfab290ec25786ba40c3a3d5d3687398ae45094d90768455e35c6642188b15"} Oct 08 18:32:14 crc kubenswrapper[4988]: I1008 18:32:14.508072 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"0efd1f53-90ad-4eb8-90f2-2ba9fc1593c2","Type":"ContainerStarted","Data":"4a380fff05cca0afef17a17e2bbd956d198717c45598f70a6f3af1d5e95b0439"} Oct 08 18:32:14 crc kubenswrapper[4988]: I1008 18:32:14.538940 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.538924022 podStartE2EDuration="2.538924022s" podCreationTimestamp="2025-10-08 18:32:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 18:32:14.538309362 +0000 UTC m=+1279.988152142" watchObservedRunningTime="2025-10-08 18:32:14.538924022 +0000 UTC m=+1279.988766792" Oct 08 18:32:14 crc kubenswrapper[4988]: I1008 18:32:14.540605 4988 scope.go:117] "RemoveContainer" containerID="3f453b3226c2919bd0521ef15a3bf5e939375ac3f298ac9733160689d3fda89b" Oct 08 18:32:14 crc kubenswrapper[4988]: E1008 18:32:14.544961 4988 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3f453b3226c2919bd0521ef15a3bf5e939375ac3f298ac9733160689d3fda89b\": container with ID starting with 3f453b3226c2919bd0521ef15a3bf5e939375ac3f298ac9733160689d3fda89b not found: ID does not exist" containerID="3f453b3226c2919bd0521ef15a3bf5e939375ac3f298ac9733160689d3fda89b" Oct 08 18:32:14 crc kubenswrapper[4988]: I1008 18:32:14.545011 4988 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3f453b3226c2919bd0521ef15a3bf5e939375ac3f298ac9733160689d3fda89b"} err="failed to get container status \"3f453b3226c2919bd0521ef15a3bf5e939375ac3f298ac9733160689d3fda89b\": rpc error: code = NotFound desc = could not find container \"3f453b3226c2919bd0521ef15a3bf5e939375ac3f298ac9733160689d3fda89b\": container with ID starting with 3f453b3226c2919bd0521ef15a3bf5e939375ac3f298ac9733160689d3fda89b not found: ID does not exist" Oct 08 18:32:14 crc kubenswrapper[4988]: I1008 18:32:14.559610 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Oct 08 18:32:14 crc kubenswrapper[4988]: I1008 18:32:14.569931 4988 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Oct 08 18:32:14 crc kubenswrapper[4988]: I1008 18:32:14.580011 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Oct 08 18:32:14 crc kubenswrapper[4988]: E1008 18:32:14.580387 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="be97711d-1138-4866-a82d-c0ddbcec58c8" containerName="nova-scheduler-scheduler" Oct 08 18:32:14 crc kubenswrapper[4988]: I1008 18:32:14.580419 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="be97711d-1138-4866-a82d-c0ddbcec58c8" containerName="nova-scheduler-scheduler" Oct 08 18:32:14 crc kubenswrapper[4988]: I1008 18:32:14.580608 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="be97711d-1138-4866-a82d-c0ddbcec58c8" containerName="nova-scheduler-scheduler" Oct 08 18:32:14 crc kubenswrapper[4988]: I1008 18:32:14.581263 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 08 18:32:14 crc kubenswrapper[4988]: I1008 18:32:14.584753 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Oct 08 18:32:14 crc kubenswrapper[4988]: I1008 18:32:14.593394 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Oct 08 18:32:14 crc kubenswrapper[4988]: I1008 18:32:14.696829 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1b24865e-cac2-490e-9e61-de22a379ff47-config-data\") pod \"nova-scheduler-0\" (UID: \"1b24865e-cac2-490e-9e61-de22a379ff47\") " pod="openstack/nova-scheduler-0" Oct 08 18:32:14 crc kubenswrapper[4988]: I1008 18:32:14.696909 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xctqm\" (UniqueName: \"kubernetes.io/projected/1b24865e-cac2-490e-9e61-de22a379ff47-kube-api-access-xctqm\") pod \"nova-scheduler-0\" (UID: \"1b24865e-cac2-490e-9e61-de22a379ff47\") " pod="openstack/nova-scheduler-0" Oct 08 18:32:14 crc kubenswrapper[4988]: I1008 18:32:14.697130 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1b24865e-cac2-490e-9e61-de22a379ff47-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"1b24865e-cac2-490e-9e61-de22a379ff47\") " pod="openstack/nova-scheduler-0" Oct 08 18:32:14 crc kubenswrapper[4988]: I1008 18:32:14.799319 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1b24865e-cac2-490e-9e61-de22a379ff47-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"1b24865e-cac2-490e-9e61-de22a379ff47\") " pod="openstack/nova-scheduler-0" Oct 08 18:32:14 crc kubenswrapper[4988]: I1008 18:32:14.799467 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1b24865e-cac2-490e-9e61-de22a379ff47-config-data\") pod \"nova-scheduler-0\" (UID: \"1b24865e-cac2-490e-9e61-de22a379ff47\") " pod="openstack/nova-scheduler-0" Oct 08 18:32:14 crc kubenswrapper[4988]: I1008 18:32:14.799540 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xctqm\" (UniqueName: \"kubernetes.io/projected/1b24865e-cac2-490e-9e61-de22a379ff47-kube-api-access-xctqm\") pod \"nova-scheduler-0\" (UID: \"1b24865e-cac2-490e-9e61-de22a379ff47\") " pod="openstack/nova-scheduler-0" Oct 08 18:32:14 crc kubenswrapper[4988]: I1008 18:32:14.806691 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1b24865e-cac2-490e-9e61-de22a379ff47-config-data\") pod \"nova-scheduler-0\" (UID: \"1b24865e-cac2-490e-9e61-de22a379ff47\") " pod="openstack/nova-scheduler-0" Oct 08 18:32:14 crc kubenswrapper[4988]: I1008 18:32:14.807977 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1b24865e-cac2-490e-9e61-de22a379ff47-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"1b24865e-cac2-490e-9e61-de22a379ff47\") " pod="openstack/nova-scheduler-0" Oct 08 18:32:14 crc kubenswrapper[4988]: I1008 18:32:14.819078 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xctqm\" (UniqueName: \"kubernetes.io/projected/1b24865e-cac2-490e-9e61-de22a379ff47-kube-api-access-xctqm\") pod \"nova-scheduler-0\" (UID: \"1b24865e-cac2-490e-9e61-de22a379ff47\") " pod="openstack/nova-scheduler-0" Oct 08 18:32:14 crc kubenswrapper[4988]: I1008 18:32:14.844746 4988 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="c51f2710-19e3-457c-ab1b-6c7e19877213" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.193:8775/\": read tcp 10.217.0.2:52402->10.217.0.193:8775: read: connection reset by peer" Oct 08 18:32:14 crc kubenswrapper[4988]: I1008 18:32:14.845046 4988 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="c51f2710-19e3-457c-ab1b-6c7e19877213" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.193:8775/\": read tcp 10.217.0.2:52386->10.217.0.193:8775: read: connection reset by peer" Oct 08 18:32:14 crc kubenswrapper[4988]: I1008 18:32:14.904678 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 08 18:32:15 crc kubenswrapper[4988]: I1008 18:32:15.246942 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 08 18:32:15 crc kubenswrapper[4988]: I1008 18:32:15.249891 4988 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="be97711d-1138-4866-a82d-c0ddbcec58c8" path="/var/lib/kubelet/pods/be97711d-1138-4866-a82d-c0ddbcec58c8/volumes" Oct 08 18:32:15 crc kubenswrapper[4988]: I1008 18:32:15.409996 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c51f2710-19e3-457c-ab1b-6c7e19877213-config-data\") pod \"c51f2710-19e3-457c-ab1b-6c7e19877213\" (UID: \"c51f2710-19e3-457c-ab1b-6c7e19877213\") " Oct 08 18:32:15 crc kubenswrapper[4988]: I1008 18:32:15.410294 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zk494\" (UniqueName: \"kubernetes.io/projected/c51f2710-19e3-457c-ab1b-6c7e19877213-kube-api-access-zk494\") pod \"c51f2710-19e3-457c-ab1b-6c7e19877213\" (UID: \"c51f2710-19e3-457c-ab1b-6c7e19877213\") " Oct 08 18:32:15 crc kubenswrapper[4988]: I1008 18:32:15.410331 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c51f2710-19e3-457c-ab1b-6c7e19877213-combined-ca-bundle\") pod \"c51f2710-19e3-457c-ab1b-6c7e19877213\" (UID: \"c51f2710-19e3-457c-ab1b-6c7e19877213\") " Oct 08 18:32:15 crc kubenswrapper[4988]: I1008 18:32:15.410437 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c51f2710-19e3-457c-ab1b-6c7e19877213-logs\") pod \"c51f2710-19e3-457c-ab1b-6c7e19877213\" (UID: \"c51f2710-19e3-457c-ab1b-6c7e19877213\") " Oct 08 18:32:15 crc kubenswrapper[4988]: I1008 18:32:15.410844 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c51f2710-19e3-457c-ab1b-6c7e19877213-logs" (OuterVolumeSpecName: "logs") pod "c51f2710-19e3-457c-ab1b-6c7e19877213" (UID: "c51f2710-19e3-457c-ab1b-6c7e19877213"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 18:32:15 crc kubenswrapper[4988]: I1008 18:32:15.411017 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/c51f2710-19e3-457c-ab1b-6c7e19877213-nova-metadata-tls-certs\") pod \"c51f2710-19e3-457c-ab1b-6c7e19877213\" (UID: \"c51f2710-19e3-457c-ab1b-6c7e19877213\") " Oct 08 18:32:15 crc kubenswrapper[4988]: I1008 18:32:15.411673 4988 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c51f2710-19e3-457c-ab1b-6c7e19877213-logs\") on node \"crc\" DevicePath \"\"" Oct 08 18:32:15 crc kubenswrapper[4988]: I1008 18:32:15.414546 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c51f2710-19e3-457c-ab1b-6c7e19877213-kube-api-access-zk494" (OuterVolumeSpecName: "kube-api-access-zk494") pod "c51f2710-19e3-457c-ab1b-6c7e19877213" (UID: "c51f2710-19e3-457c-ab1b-6c7e19877213"). InnerVolumeSpecName "kube-api-access-zk494". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:32:15 crc kubenswrapper[4988]: I1008 18:32:15.439446 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c51f2710-19e3-457c-ab1b-6c7e19877213-config-data" (OuterVolumeSpecName: "config-data") pod "c51f2710-19e3-457c-ab1b-6c7e19877213" (UID: "c51f2710-19e3-457c-ab1b-6c7e19877213"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:32:15 crc kubenswrapper[4988]: I1008 18:32:15.460692 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c51f2710-19e3-457c-ab1b-6c7e19877213-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "c51f2710-19e3-457c-ab1b-6c7e19877213" (UID: "c51f2710-19e3-457c-ab1b-6c7e19877213"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:32:15 crc kubenswrapper[4988]: W1008 18:32:15.468277 4988 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1b24865e_cac2_490e_9e61_de22a379ff47.slice/crio-a05a410cb5097099f45d7708891f938ae729f9de75fcebee924aaf3fbf711bf8 WatchSource:0}: Error finding container a05a410cb5097099f45d7708891f938ae729f9de75fcebee924aaf3fbf711bf8: Status 404 returned error can't find the container with id a05a410cb5097099f45d7708891f938ae729f9de75fcebee924aaf3fbf711bf8 Oct 08 18:32:15 crc kubenswrapper[4988]: I1008 18:32:15.468748 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Oct 08 18:32:15 crc kubenswrapper[4988]: I1008 18:32:15.471773 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c51f2710-19e3-457c-ab1b-6c7e19877213-nova-metadata-tls-certs" (OuterVolumeSpecName: "nova-metadata-tls-certs") pod "c51f2710-19e3-457c-ab1b-6c7e19877213" (UID: "c51f2710-19e3-457c-ab1b-6c7e19877213"). InnerVolumeSpecName "nova-metadata-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:32:15 crc kubenswrapper[4988]: I1008 18:32:15.514358 4988 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/c51f2710-19e3-457c-ab1b-6c7e19877213-nova-metadata-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 08 18:32:15 crc kubenswrapper[4988]: I1008 18:32:15.514433 4988 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c51f2710-19e3-457c-ab1b-6c7e19877213-config-data\") on node \"crc\" DevicePath \"\"" Oct 08 18:32:15 crc kubenswrapper[4988]: I1008 18:32:15.514446 4988 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zk494\" (UniqueName: \"kubernetes.io/projected/c51f2710-19e3-457c-ab1b-6c7e19877213-kube-api-access-zk494\") on node \"crc\" DevicePath \"\"" Oct 08 18:32:15 crc kubenswrapper[4988]: I1008 18:32:15.514459 4988 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c51f2710-19e3-457c-ab1b-6c7e19877213-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 18:32:15 crc kubenswrapper[4988]: I1008 18:32:15.518916 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"1b24865e-cac2-490e-9e61-de22a379ff47","Type":"ContainerStarted","Data":"a05a410cb5097099f45d7708891f938ae729f9de75fcebee924aaf3fbf711bf8"} Oct 08 18:32:15 crc kubenswrapper[4988]: I1008 18:32:15.523297 4988 generic.go:334] "Generic (PLEG): container finished" podID="c51f2710-19e3-457c-ab1b-6c7e19877213" containerID="8673b565fae2659ae04aabd4852445f45c49c3733848e2571f9ff5b470e6a680" exitCode=0 Oct 08 18:32:15 crc kubenswrapper[4988]: I1008 18:32:15.523421 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 08 18:32:15 crc kubenswrapper[4988]: I1008 18:32:15.523437 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"c51f2710-19e3-457c-ab1b-6c7e19877213","Type":"ContainerDied","Data":"8673b565fae2659ae04aabd4852445f45c49c3733848e2571f9ff5b470e6a680"} Oct 08 18:32:15 crc kubenswrapper[4988]: I1008 18:32:15.523500 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"c51f2710-19e3-457c-ab1b-6c7e19877213","Type":"ContainerDied","Data":"854b0b91e2dd7bd74cfd3ea918cb2a1a74151fe5a7d547e3def9a1900ce4f5d7"} Oct 08 18:32:15 crc kubenswrapper[4988]: I1008 18:32:15.523526 4988 scope.go:117] "RemoveContainer" containerID="8673b565fae2659ae04aabd4852445f45c49c3733848e2571f9ff5b470e6a680" Oct 08 18:32:15 crc kubenswrapper[4988]: I1008 18:32:15.565676 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Oct 08 18:32:15 crc kubenswrapper[4988]: I1008 18:32:15.577052 4988 scope.go:117] "RemoveContainer" containerID="15a065b5102872a8fc42dc14c1114cd59430e0c88890fc13c4496e3e17a3c6bc" Oct 08 18:32:15 crc kubenswrapper[4988]: I1008 18:32:15.580099 4988 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Oct 08 18:32:15 crc kubenswrapper[4988]: I1008 18:32:15.587661 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Oct 08 18:32:15 crc kubenswrapper[4988]: E1008 18:32:15.588093 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c51f2710-19e3-457c-ab1b-6c7e19877213" containerName="nova-metadata-metadata" Oct 08 18:32:15 crc kubenswrapper[4988]: I1008 18:32:15.588112 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="c51f2710-19e3-457c-ab1b-6c7e19877213" containerName="nova-metadata-metadata" Oct 08 18:32:15 crc kubenswrapper[4988]: E1008 18:32:15.588149 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c51f2710-19e3-457c-ab1b-6c7e19877213" containerName="nova-metadata-log" Oct 08 18:32:15 crc kubenswrapper[4988]: I1008 18:32:15.588156 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="c51f2710-19e3-457c-ab1b-6c7e19877213" containerName="nova-metadata-log" Oct 08 18:32:15 crc kubenswrapper[4988]: I1008 18:32:15.588342 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="c51f2710-19e3-457c-ab1b-6c7e19877213" containerName="nova-metadata-metadata" Oct 08 18:32:15 crc kubenswrapper[4988]: I1008 18:32:15.588374 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="c51f2710-19e3-457c-ab1b-6c7e19877213" containerName="nova-metadata-log" Oct 08 18:32:15 crc kubenswrapper[4988]: I1008 18:32:15.589575 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 08 18:32:15 crc kubenswrapper[4988]: I1008 18:32:15.593853 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Oct 08 18:32:15 crc kubenswrapper[4988]: I1008 18:32:15.594082 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Oct 08 18:32:15 crc kubenswrapper[4988]: I1008 18:32:15.598928 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 08 18:32:15 crc kubenswrapper[4988]: I1008 18:32:15.606380 4988 scope.go:117] "RemoveContainer" containerID="8673b565fae2659ae04aabd4852445f45c49c3733848e2571f9ff5b470e6a680" Oct 08 18:32:15 crc kubenswrapper[4988]: E1008 18:32:15.616618 4988 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8673b565fae2659ae04aabd4852445f45c49c3733848e2571f9ff5b470e6a680\": container with ID starting with 8673b565fae2659ae04aabd4852445f45c49c3733848e2571f9ff5b470e6a680 not found: ID does not exist" containerID="8673b565fae2659ae04aabd4852445f45c49c3733848e2571f9ff5b470e6a680" Oct 08 18:32:15 crc kubenswrapper[4988]: I1008 18:32:15.616662 4988 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8673b565fae2659ae04aabd4852445f45c49c3733848e2571f9ff5b470e6a680"} err="failed to get container status \"8673b565fae2659ae04aabd4852445f45c49c3733848e2571f9ff5b470e6a680\": rpc error: code = NotFound desc = could not find container \"8673b565fae2659ae04aabd4852445f45c49c3733848e2571f9ff5b470e6a680\": container with ID starting with 8673b565fae2659ae04aabd4852445f45c49c3733848e2571f9ff5b470e6a680 not found: ID does not exist" Oct 08 18:32:15 crc kubenswrapper[4988]: I1008 18:32:15.616686 4988 scope.go:117] "RemoveContainer" containerID="15a065b5102872a8fc42dc14c1114cd59430e0c88890fc13c4496e3e17a3c6bc" Oct 08 18:32:15 crc kubenswrapper[4988]: E1008 18:32:15.617134 4988 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"15a065b5102872a8fc42dc14c1114cd59430e0c88890fc13c4496e3e17a3c6bc\": container with ID starting with 15a065b5102872a8fc42dc14c1114cd59430e0c88890fc13c4496e3e17a3c6bc not found: ID does not exist" containerID="15a065b5102872a8fc42dc14c1114cd59430e0c88890fc13c4496e3e17a3c6bc" Oct 08 18:32:15 crc kubenswrapper[4988]: I1008 18:32:15.617149 4988 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"15a065b5102872a8fc42dc14c1114cd59430e0c88890fc13c4496e3e17a3c6bc"} err="failed to get container status \"15a065b5102872a8fc42dc14c1114cd59430e0c88890fc13c4496e3e17a3c6bc\": rpc error: code = NotFound desc = could not find container \"15a065b5102872a8fc42dc14c1114cd59430e0c88890fc13c4496e3e17a3c6bc\": container with ID starting with 15a065b5102872a8fc42dc14c1114cd59430e0c88890fc13c4496e3e17a3c6bc not found: ID does not exist" Oct 08 18:32:15 crc kubenswrapper[4988]: I1008 18:32:15.719369 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/83ef7610-23ac-4445-b75c-ae7206b4d810-logs\") pod \"nova-metadata-0\" (UID: \"83ef7610-23ac-4445-b75c-ae7206b4d810\") " pod="openstack/nova-metadata-0" Oct 08 18:32:15 crc kubenswrapper[4988]: I1008 18:32:15.719736 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/83ef7610-23ac-4445-b75c-ae7206b4d810-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"83ef7610-23ac-4445-b75c-ae7206b4d810\") " pod="openstack/nova-metadata-0" Oct 08 18:32:15 crc kubenswrapper[4988]: I1008 18:32:15.719809 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lvrb9\" (UniqueName: \"kubernetes.io/projected/83ef7610-23ac-4445-b75c-ae7206b4d810-kube-api-access-lvrb9\") pod \"nova-metadata-0\" (UID: \"83ef7610-23ac-4445-b75c-ae7206b4d810\") " pod="openstack/nova-metadata-0" Oct 08 18:32:15 crc kubenswrapper[4988]: I1008 18:32:15.719943 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/83ef7610-23ac-4445-b75c-ae7206b4d810-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"83ef7610-23ac-4445-b75c-ae7206b4d810\") " pod="openstack/nova-metadata-0" Oct 08 18:32:15 crc kubenswrapper[4988]: I1008 18:32:15.720001 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/83ef7610-23ac-4445-b75c-ae7206b4d810-config-data\") pod \"nova-metadata-0\" (UID: \"83ef7610-23ac-4445-b75c-ae7206b4d810\") " pod="openstack/nova-metadata-0" Oct 08 18:32:15 crc kubenswrapper[4988]: I1008 18:32:15.821601 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/83ef7610-23ac-4445-b75c-ae7206b4d810-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"83ef7610-23ac-4445-b75c-ae7206b4d810\") " pod="openstack/nova-metadata-0" Oct 08 18:32:15 crc kubenswrapper[4988]: I1008 18:32:15.821654 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lvrb9\" (UniqueName: \"kubernetes.io/projected/83ef7610-23ac-4445-b75c-ae7206b4d810-kube-api-access-lvrb9\") pod \"nova-metadata-0\" (UID: \"83ef7610-23ac-4445-b75c-ae7206b4d810\") " pod="openstack/nova-metadata-0" Oct 08 18:32:15 crc kubenswrapper[4988]: I1008 18:32:15.821688 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/83ef7610-23ac-4445-b75c-ae7206b4d810-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"83ef7610-23ac-4445-b75c-ae7206b4d810\") " pod="openstack/nova-metadata-0" Oct 08 18:32:15 crc kubenswrapper[4988]: I1008 18:32:15.821713 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/83ef7610-23ac-4445-b75c-ae7206b4d810-config-data\") pod \"nova-metadata-0\" (UID: \"83ef7610-23ac-4445-b75c-ae7206b4d810\") " pod="openstack/nova-metadata-0" Oct 08 18:32:15 crc kubenswrapper[4988]: I1008 18:32:15.821772 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/83ef7610-23ac-4445-b75c-ae7206b4d810-logs\") pod \"nova-metadata-0\" (UID: \"83ef7610-23ac-4445-b75c-ae7206b4d810\") " pod="openstack/nova-metadata-0" Oct 08 18:32:15 crc kubenswrapper[4988]: I1008 18:32:15.822272 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/83ef7610-23ac-4445-b75c-ae7206b4d810-logs\") pod \"nova-metadata-0\" (UID: \"83ef7610-23ac-4445-b75c-ae7206b4d810\") " pod="openstack/nova-metadata-0" Oct 08 18:32:15 crc kubenswrapper[4988]: I1008 18:32:15.826752 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/83ef7610-23ac-4445-b75c-ae7206b4d810-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"83ef7610-23ac-4445-b75c-ae7206b4d810\") " pod="openstack/nova-metadata-0" Oct 08 18:32:15 crc kubenswrapper[4988]: I1008 18:32:15.827515 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/83ef7610-23ac-4445-b75c-ae7206b4d810-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"83ef7610-23ac-4445-b75c-ae7206b4d810\") " pod="openstack/nova-metadata-0" Oct 08 18:32:15 crc kubenswrapper[4988]: I1008 18:32:15.829548 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/83ef7610-23ac-4445-b75c-ae7206b4d810-config-data\") pod \"nova-metadata-0\" (UID: \"83ef7610-23ac-4445-b75c-ae7206b4d810\") " pod="openstack/nova-metadata-0" Oct 08 18:32:15 crc kubenswrapper[4988]: I1008 18:32:15.837110 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lvrb9\" (UniqueName: \"kubernetes.io/projected/83ef7610-23ac-4445-b75c-ae7206b4d810-kube-api-access-lvrb9\") pod \"nova-metadata-0\" (UID: \"83ef7610-23ac-4445-b75c-ae7206b4d810\") " pod="openstack/nova-metadata-0" Oct 08 18:32:15 crc kubenswrapper[4988]: I1008 18:32:15.954591 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 08 18:32:16 crc kubenswrapper[4988]: W1008 18:32:16.401811 4988 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod83ef7610_23ac_4445_b75c_ae7206b4d810.slice/crio-cce50310cff807710abf94c8f4ae4a3519677bbe1dca06bbe268f6dbc2ae8eec WatchSource:0}: Error finding container cce50310cff807710abf94c8f4ae4a3519677bbe1dca06bbe268f6dbc2ae8eec: Status 404 returned error can't find the container with id cce50310cff807710abf94c8f4ae4a3519677bbe1dca06bbe268f6dbc2ae8eec Oct 08 18:32:16 crc kubenswrapper[4988]: I1008 18:32:16.403801 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 08 18:32:16 crc kubenswrapper[4988]: I1008 18:32:16.539157 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"83ef7610-23ac-4445-b75c-ae7206b4d810","Type":"ContainerStarted","Data":"cce50310cff807710abf94c8f4ae4a3519677bbe1dca06bbe268f6dbc2ae8eec"} Oct 08 18:32:16 crc kubenswrapper[4988]: I1008 18:32:16.542553 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"1b24865e-cac2-490e-9e61-de22a379ff47","Type":"ContainerStarted","Data":"6d8f1c5dc40a53cca8a87fffd5d06729664e7815b666a1eba260acf508457342"} Oct 08 18:32:16 crc kubenswrapper[4988]: I1008 18:32:16.565286 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.564749505 podStartE2EDuration="2.564749505s" podCreationTimestamp="2025-10-08 18:32:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 18:32:16.555678254 +0000 UTC m=+1282.005521034" watchObservedRunningTime="2025-10-08 18:32:16.564749505 +0000 UTC m=+1282.014592295" Oct 08 18:32:17 crc kubenswrapper[4988]: I1008 18:32:17.250218 4988 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c51f2710-19e3-457c-ab1b-6c7e19877213" path="/var/lib/kubelet/pods/c51f2710-19e3-457c-ab1b-6c7e19877213/volumes" Oct 08 18:32:17 crc kubenswrapper[4988]: I1008 18:32:17.563943 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"83ef7610-23ac-4445-b75c-ae7206b4d810","Type":"ContainerStarted","Data":"4b994190cba63b85214c41a6bf0e9b5cca7c29dd236d40d5df3055f98c9ec35b"} Oct 08 18:32:17 crc kubenswrapper[4988]: I1008 18:32:17.564007 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"83ef7610-23ac-4445-b75c-ae7206b4d810","Type":"ContainerStarted","Data":"e504bf27bfc24744eb7b0bdc9918cda7b788dc1c32411c922014b5f7f84ae75d"} Oct 08 18:32:17 crc kubenswrapper[4988]: I1008 18:32:17.593280 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.593260461 podStartE2EDuration="2.593260461s" podCreationTimestamp="2025-10-08 18:32:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 18:32:17.587053272 +0000 UTC m=+1283.036896042" watchObservedRunningTime="2025-10-08 18:32:17.593260461 +0000 UTC m=+1283.043103241" Oct 08 18:32:19 crc kubenswrapper[4988]: I1008 18:32:19.905375 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Oct 08 18:32:20 crc kubenswrapper[4988]: I1008 18:32:20.955618 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Oct 08 18:32:20 crc kubenswrapper[4988]: I1008 18:32:20.955678 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Oct 08 18:32:22 crc kubenswrapper[4988]: I1008 18:32:22.884894 4988 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 08 18:32:22 crc kubenswrapper[4988]: I1008 18:32:22.885305 4988 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 08 18:32:23 crc kubenswrapper[4988]: I1008 18:32:23.903703 4988 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="0efd1f53-90ad-4eb8-90f2-2ba9fc1593c2" containerName="nova-api-api" probeResult="failure" output="Get \"https://10.217.0.203:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 08 18:32:23 crc kubenswrapper[4988]: I1008 18:32:23.903712 4988 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="0efd1f53-90ad-4eb8-90f2-2ba9fc1593c2" containerName="nova-api-log" probeResult="failure" output="Get \"https://10.217.0.203:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 08 18:32:24 crc kubenswrapper[4988]: I1008 18:32:24.905700 4988 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Oct 08 18:32:24 crc kubenswrapper[4988]: I1008 18:32:24.937261 4988 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Oct 08 18:32:25 crc kubenswrapper[4988]: I1008 18:32:25.688612 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Oct 08 18:32:25 crc kubenswrapper[4988]: I1008 18:32:25.955470 4988 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Oct 08 18:32:25 crc kubenswrapper[4988]: I1008 18:32:25.955530 4988 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Oct 08 18:32:26 crc kubenswrapper[4988]: I1008 18:32:26.970737 4988 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="83ef7610-23ac-4445-b75c-ae7206b4d810" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.205:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 08 18:32:26 crc kubenswrapper[4988]: I1008 18:32:26.971018 4988 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="83ef7610-23ac-4445-b75c-ae7206b4d810" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.205:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 08 18:32:31 crc kubenswrapper[4988]: I1008 18:32:31.298291 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Oct 08 18:32:32 crc kubenswrapper[4988]: I1008 18:32:32.893991 4988 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Oct 08 18:32:32 crc kubenswrapper[4988]: I1008 18:32:32.894556 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Oct 08 18:32:32 crc kubenswrapper[4988]: I1008 18:32:32.896581 4988 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Oct 08 18:32:32 crc kubenswrapper[4988]: I1008 18:32:32.907191 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Oct 08 18:32:33 crc kubenswrapper[4988]: I1008 18:32:33.736897 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Oct 08 18:32:33 crc kubenswrapper[4988]: I1008 18:32:33.745229 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Oct 08 18:32:35 crc kubenswrapper[4988]: I1008 18:32:35.961156 4988 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Oct 08 18:32:35 crc kubenswrapper[4988]: I1008 18:32:35.961687 4988 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Oct 08 18:32:35 crc kubenswrapper[4988]: I1008 18:32:35.972692 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Oct 08 18:32:35 crc kubenswrapper[4988]: I1008 18:32:35.974788 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Oct 08 18:32:53 crc kubenswrapper[4988]: I1008 18:32:53.338118 4988 patch_prober.go:28] interesting pod/machine-config-daemon-nm54f container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 18:32:53 crc kubenswrapper[4988]: I1008 18:32:53.338579 4988 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 18:32:56 crc kubenswrapper[4988]: I1008 18:32:56.589805 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/openstackclient"] Oct 08 18:32:56 crc kubenswrapper[4988]: I1008 18:32:56.590550 4988 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/openstackclient" podUID="c97aab35-bbee-41a6-88b2-801b0b097a6d" containerName="openstackclient" containerID="cri-o://47dce63c2014855045e7a231270ccde5910760705845d62023d29e100f929478" gracePeriod=2 Oct 08 18:32:56 crc kubenswrapper[4988]: I1008 18:32:56.646002 4988 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/openstackclient"] Oct 08 18:32:56 crc kubenswrapper[4988]: I1008 18:32:56.950687 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovsdbserver-sb-0"] Oct 08 18:32:56 crc kubenswrapper[4988]: I1008 18:32:56.951050 4988 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ovsdbserver-sb-0" podUID="1f36756f-acb3-439d-b10e-13573cbd252a" containerName="openstack-network-exporter" containerID="cri-o://e8f79299e7fcca754183f5f8a133b9bb867b1fa107190b02e938976ad478eb0b" gracePeriod=300 Oct 08 18:32:56 crc kubenswrapper[4988]: I1008 18:32:56.981666 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 08 18:32:56 crc kubenswrapper[4988]: I1008 18:32:56.981907 4988 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="0ec345a3-1744-4e78-bf13-91f0325296e4" containerName="cinder-scheduler" containerID="cri-o://590ae1792458f3fd70bbade362261ee71cafb4633235b6af45da637905058707" gracePeriod=30 Oct 08 18:32:56 crc kubenswrapper[4988]: I1008 18:32:56.982362 4988 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="0ec345a3-1744-4e78-bf13-91f0325296e4" containerName="probe" containerID="cri-o://19c72634c92910f59b5f356f56971cc3e2867c6c02b8cdd4a997a62cb452e3cb" gracePeriod=30 Oct 08 18:32:57 crc kubenswrapper[4988]: I1008 18:32:57.085311 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 08 18:32:57 crc kubenswrapper[4988]: I1008 18:32:57.199994 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Oct 08 18:32:57 crc kubenswrapper[4988]: I1008 18:32:57.209254 4988 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="99d77eed-40f0-4879-be96-abea3a807cd9" containerName="cinder-api-log" containerID="cri-o://1c61da6cc2966ef990d9bcdbb3737986e369db20fe8b42772368a61ef7dc1f03" gracePeriod=30 Oct 08 18:32:57 crc kubenswrapper[4988]: I1008 18:32:57.209509 4988 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="99d77eed-40f0-4879-be96-abea3a807cd9" containerName="cinder-api" containerID="cri-o://d2face2dfb0693cf7a2c3531e80c082e310d388b04c8005247e18e79ea8bbab3" gracePeriod=30 Oct 08 18:32:57 crc kubenswrapper[4988]: I1008 18:32:57.285689 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement79af-account-delete-hv6j5"] Oct 08 18:32:57 crc kubenswrapper[4988]: E1008 18:32:57.294170 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c97aab35-bbee-41a6-88b2-801b0b097a6d" containerName="openstackclient" Oct 08 18:32:57 crc kubenswrapper[4988]: I1008 18:32:57.294364 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="c97aab35-bbee-41a6-88b2-801b0b097a6d" containerName="openstackclient" Oct 08 18:32:57 crc kubenswrapper[4988]: I1008 18:32:57.294617 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="c97aab35-bbee-41a6-88b2-801b0b097a6d" containerName="openstackclient" Oct 08 18:32:57 crc kubenswrapper[4988]: I1008 18:32:57.295247 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-northd-0"] Oct 08 18:32:57 crc kubenswrapper[4988]: I1008 18:32:57.317475 4988 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ovn-northd-0" podUID="f026ec8f-ffb4-4538-a6fc-3116ce0e7805" containerName="ovn-northd" containerID="cri-o://e9efac4802d1c9a2377dc285da1c498c1363ff5f70695e54b4a257848bdea958" gracePeriod=30 Oct 08 18:32:57 crc kubenswrapper[4988]: I1008 18:32:57.296113 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement79af-account-delete-hv6j5" Oct 08 18:32:57 crc kubenswrapper[4988]: I1008 18:32:57.318834 4988 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ovn-northd-0" podUID="f026ec8f-ffb4-4538-a6fc-3116ce0e7805" containerName="openstack-network-exporter" containerID="cri-o://ada91b938d451d0675a819f062dafde62f4a75b785dc86a5448ce25e246cecbb" gracePeriod=30 Oct 08 18:32:57 crc kubenswrapper[4988]: E1008 18:32:57.298091 4988 configmap.go:193] Couldn't get configMap openstack/rabbitmq-cell1-config-data: configmap "rabbitmq-cell1-config-data" not found Oct 08 18:32:57 crc kubenswrapper[4988]: E1008 18:32:57.323728 4988 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/e5a8f819-c91d-429e-b848-e9c444739623-config-data podName:e5a8f819-c91d-429e-b848-e9c444739623 nodeName:}" failed. No retries permitted until 2025-10-08 18:32:57.823703762 +0000 UTC m=+1323.273546542 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/configmap/e5a8f819-c91d-429e-b848-e9c444739623-config-data") pod "rabbitmq-cell1-server-0" (UID: "e5a8f819-c91d-429e-b848-e9c444739623") : configmap "rabbitmq-cell1-config-data" not found Oct 08 18:32:57 crc kubenswrapper[4988]: I1008 18:32:57.331624 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement79af-account-delete-hv6j5"] Oct 08 18:32:57 crc kubenswrapper[4988]: I1008 18:32:57.358468 4988 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/cinder-api-0" podUID="99d77eed-40f0-4879-be96-abea3a807cd9" containerName="cinder-api" probeResult="failure" output="Get \"https://10.217.0.171:8776/healthcheck\": EOF" Oct 08 18:32:57 crc kubenswrapper[4988]: I1008 18:32:57.358988 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-sync-gwkbl"] Oct 08 18:32:57 crc kubenswrapper[4988]: I1008 18:32:57.393124 4988 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-sync-gwkbl"] Oct 08 18:32:57 crc kubenswrapper[4988]: I1008 18:32:57.397468 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glancec9f7-account-delete-sp22m"] Oct 08 18:32:57 crc kubenswrapper[4988]: I1008 18:32:57.398687 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glancec9f7-account-delete-sp22m" Oct 08 18:32:57 crc kubenswrapper[4988]: I1008 18:32:57.400138 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hjvsd\" (UniqueName: \"kubernetes.io/projected/79cbf065-a458-4dec-b6a7-805f515142e0-kube-api-access-hjvsd\") pod \"placement79af-account-delete-hv6j5\" (UID: \"79cbf065-a458-4dec-b6a7-805f515142e0\") " pod="openstack/placement79af-account-delete-hv6j5" Oct 08 18:32:57 crc kubenswrapper[4988]: I1008 18:32:57.445366 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glancec9f7-account-delete-sp22m"] Oct 08 18:32:57 crc kubenswrapper[4988]: I1008 18:32:57.481512 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican32b1-account-delete-jlhch"] Oct 08 18:32:57 crc kubenswrapper[4988]: I1008 18:32:57.482755 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican32b1-account-delete-jlhch" Oct 08 18:32:57 crc kubenswrapper[4988]: I1008 18:32:57.503266 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican32b1-account-delete-jlhch"] Oct 08 18:32:57 crc kubenswrapper[4988]: I1008 18:32:57.503884 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mlmlm\" (UniqueName: \"kubernetes.io/projected/d7d648e7-3c02-4d26-993d-a901764b3329-kube-api-access-mlmlm\") pod \"glancec9f7-account-delete-sp22m\" (UID: \"d7d648e7-3c02-4d26-993d-a901764b3329\") " pod="openstack/glancec9f7-account-delete-sp22m" Oct 08 18:32:57 crc kubenswrapper[4988]: I1008 18:32:57.503954 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hjvsd\" (UniqueName: \"kubernetes.io/projected/79cbf065-a458-4dec-b6a7-805f515142e0-kube-api-access-hjvsd\") pod \"placement79af-account-delete-hv6j5\" (UID: \"79cbf065-a458-4dec-b6a7-805f515142e0\") " pod="openstack/placement79af-account-delete-hv6j5" Oct 08 18:32:57 crc kubenswrapper[4988]: I1008 18:32:57.607563 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tv2pz\" (UniqueName: \"kubernetes.io/projected/3f7bddd3-86f8-46e9-a38e-1aa5e570d151-kube-api-access-tv2pz\") pod \"barbican32b1-account-delete-jlhch\" (UID: \"3f7bddd3-86f8-46e9-a38e-1aa5e570d151\") " pod="openstack/barbican32b1-account-delete-jlhch" Oct 08 18:32:57 crc kubenswrapper[4988]: I1008 18:32:57.607696 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mlmlm\" (UniqueName: \"kubernetes.io/projected/d7d648e7-3c02-4d26-993d-a901764b3329-kube-api-access-mlmlm\") pod \"glancec9f7-account-delete-sp22m\" (UID: \"d7d648e7-3c02-4d26-993d-a901764b3329\") " pod="openstack/glancec9f7-account-delete-sp22m" Oct 08 18:32:57 crc kubenswrapper[4988]: I1008 18:32:57.612449 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-sync-5v5z9"] Oct 08 18:32:57 crc kubenswrapper[4988]: I1008 18:32:57.617921 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hjvsd\" (UniqueName: \"kubernetes.io/projected/79cbf065-a458-4dec-b6a7-805f515142e0-kube-api-access-hjvsd\") pod \"placement79af-account-delete-hv6j5\" (UID: \"79cbf065-a458-4dec-b6a7-805f515142e0\") " pod="openstack/placement79af-account-delete-hv6j5" Oct 08 18:32:57 crc kubenswrapper[4988]: I1008 18:32:57.664121 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron3849-account-delete-m479b"] Oct 08 18:32:57 crc kubenswrapper[4988]: I1008 18:32:57.665918 4988 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ovsdbserver-sb-0" podUID="1f36756f-acb3-439d-b10e-13573cbd252a" containerName="ovsdbserver-sb" containerID="cri-o://a41e320aa9f3ac4edb764f4ac4f3c561a463787b781e74385ca8530f54e7b28e" gracePeriod=300 Oct 08 18:32:57 crc kubenswrapper[4988]: I1008 18:32:57.681277 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron3849-account-delete-m479b" Oct 08 18:32:57 crc kubenswrapper[4988]: I1008 18:32:57.682217 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mlmlm\" (UniqueName: \"kubernetes.io/projected/d7d648e7-3c02-4d26-993d-a901764b3329-kube-api-access-mlmlm\") pod \"glancec9f7-account-delete-sp22m\" (UID: \"d7d648e7-3c02-4d26-993d-a901764b3329\") " pod="openstack/glancec9f7-account-delete-sp22m" Oct 08 18:32:57 crc kubenswrapper[4988]: I1008 18:32:57.686804 4988 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-sync-5v5z9"] Oct 08 18:32:57 crc kubenswrapper[4988]: I1008 18:32:57.697173 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement79af-account-delete-hv6j5" Oct 08 18:32:57 crc kubenswrapper[4988]: I1008 18:32:57.711585 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tv2pz\" (UniqueName: \"kubernetes.io/projected/3f7bddd3-86f8-46e9-a38e-1aa5e570d151-kube-api-access-tv2pz\") pod \"barbican32b1-account-delete-jlhch\" (UID: \"3f7bddd3-86f8-46e9-a38e-1aa5e570d151\") " pod="openstack/barbican32b1-account-delete-jlhch" Oct 08 18:32:57 crc kubenswrapper[4988]: I1008 18:32:57.723249 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron3849-account-delete-m479b"] Oct 08 18:32:57 crc kubenswrapper[4988]: I1008 18:32:57.761974 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tv2pz\" (UniqueName: \"kubernetes.io/projected/3f7bddd3-86f8-46e9-a38e-1aa5e570d151-kube-api-access-tv2pz\") pod \"barbican32b1-account-delete-jlhch\" (UID: \"3f7bddd3-86f8-46e9-a38e-1aa5e570d151\") " pod="openstack/barbican32b1-account-delete-jlhch" Oct 08 18:32:57 crc kubenswrapper[4988]: I1008 18:32:57.762061 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/novacell02f70-account-delete-cdwsx"] Oct 08 18:32:57 crc kubenswrapper[4988]: I1008 18:32:57.763254 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/novacell02f70-account-delete-cdwsx" Oct 08 18:32:57 crc kubenswrapper[4988]: I1008 18:32:57.774721 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-ovs-kc7fw"] Oct 08 18:32:57 crc kubenswrapper[4988]: I1008 18:32:57.791563 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-gp8f4"] Oct 08 18:32:57 crc kubenswrapper[4988]: I1008 18:32:57.801451 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/novaapiebe4-account-delete-rph2r"] Oct 08 18:32:57 crc kubenswrapper[4988]: I1008 18:32:57.803010 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/novaapiebe4-account-delete-rph2r" Oct 08 18:32:57 crc kubenswrapper[4988]: I1008 18:32:57.810562 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-metrics-w2zhj"] Oct 08 18:32:57 crc kubenswrapper[4988]: I1008 18:32:57.810799 4988 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ovn-controller-metrics-w2zhj" podUID="b92d978b-87c6-4210-9a69-99596b05cc18" containerName="openstack-network-exporter" containerID="cri-o://80822f2e64d8fd8a2c3f08eea237f88c03ba055c04d5bb24498177ab69c61a2e" gracePeriod=30 Oct 08 18:32:57 crc kubenswrapper[4988]: I1008 18:32:57.813484 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f6jr5\" (UniqueName: \"kubernetes.io/projected/ab3315d2-3248-42f1-9834-2bab55fa9935-kube-api-access-f6jr5\") pod \"neutron3849-account-delete-m479b\" (UID: \"ab3315d2-3248-42f1-9834-2bab55fa9935\") " pod="openstack/neutron3849-account-delete-m479b" Oct 08 18:32:57 crc kubenswrapper[4988]: I1008 18:32:57.821989 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glancec9f7-account-delete-sp22m" Oct 08 18:32:57 crc kubenswrapper[4988]: I1008 18:32:57.849730 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/novacell02f70-account-delete-cdwsx"] Oct 08 18:32:57 crc kubenswrapper[4988]: I1008 18:32:57.891818 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican32b1-account-delete-jlhch" Oct 08 18:32:57 crc kubenswrapper[4988]: I1008 18:32:57.896908 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/novaapiebe4-account-delete-rph2r"] Oct 08 18:32:57 crc kubenswrapper[4988]: I1008 18:32:57.942502 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f6q49\" (UniqueName: \"kubernetes.io/projected/3e8c874a-029a-4d6a-bc2e-00aeed1636a7-kube-api-access-f6q49\") pod \"novacell02f70-account-delete-cdwsx\" (UID: \"3e8c874a-029a-4d6a-bc2e-00aeed1636a7\") " pod="openstack/novacell02f70-account-delete-cdwsx" Oct 08 18:32:57 crc kubenswrapper[4988]: I1008 18:32:57.942655 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7xsc4\" (UniqueName: \"kubernetes.io/projected/28f73b7a-bb5e-43ec-a099-13a12cdb33d3-kube-api-access-7xsc4\") pod \"novaapiebe4-account-delete-rph2r\" (UID: \"28f73b7a-bb5e-43ec-a099-13a12cdb33d3\") " pod="openstack/novaapiebe4-account-delete-rph2r" Oct 08 18:32:57 crc kubenswrapper[4988]: I1008 18:32:57.942793 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f6jr5\" (UniqueName: \"kubernetes.io/projected/ab3315d2-3248-42f1-9834-2bab55fa9935-kube-api-access-f6jr5\") pod \"neutron3849-account-delete-m479b\" (UID: \"ab3315d2-3248-42f1-9834-2bab55fa9935\") " pod="openstack/neutron3849-account-delete-m479b" Oct 08 18:32:57 crc kubenswrapper[4988]: E1008 18:32:57.943274 4988 configmap.go:193] Couldn't get configMap openstack/rabbitmq-cell1-config-data: configmap "rabbitmq-cell1-config-data" not found Oct 08 18:32:57 crc kubenswrapper[4988]: E1008 18:32:57.943319 4988 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/e5a8f819-c91d-429e-b848-e9c444739623-config-data podName:e5a8f819-c91d-429e-b848-e9c444739623 nodeName:}" failed. No retries permitted until 2025-10-08 18:32:58.94330419 +0000 UTC m=+1324.393146960 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/configmap/e5a8f819-c91d-429e-b848-e9c444739623-config-data") pod "rabbitmq-cell1-server-0" (UID: "e5a8f819-c91d-429e-b848-e9c444739623") : configmap "rabbitmq-cell1-config-data" not found Oct 08 18:32:57 crc kubenswrapper[4988]: I1008 18:32:57.963613 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/swift-ring-rebalance-9hjmc"] Oct 08 18:32:58 crc kubenswrapper[4988]: I1008 18:32:58.006143 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f6jr5\" (UniqueName: \"kubernetes.io/projected/ab3315d2-3248-42f1-9834-2bab55fa9935-kube-api-access-f6jr5\") pod \"neutron3849-account-delete-m479b\" (UID: \"ab3315d2-3248-42f1-9834-2bab55fa9935\") " pod="openstack/neutron3849-account-delete-m479b" Oct 08 18:32:58 crc kubenswrapper[4988]: I1008 18:32:58.046230 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f6q49\" (UniqueName: \"kubernetes.io/projected/3e8c874a-029a-4d6a-bc2e-00aeed1636a7-kube-api-access-f6q49\") pod \"novacell02f70-account-delete-cdwsx\" (UID: \"3e8c874a-029a-4d6a-bc2e-00aeed1636a7\") " pod="openstack/novacell02f70-account-delete-cdwsx" Oct 08 18:32:58 crc kubenswrapper[4988]: I1008 18:32:58.046405 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7xsc4\" (UniqueName: \"kubernetes.io/projected/28f73b7a-bb5e-43ec-a099-13a12cdb33d3-kube-api-access-7xsc4\") pod \"novaapiebe4-account-delete-rph2r\" (UID: \"28f73b7a-bb5e-43ec-a099-13a12cdb33d3\") " pod="openstack/novaapiebe4-account-delete-rph2r" Oct 08 18:32:58 crc kubenswrapper[4988]: I1008 18:32:58.146398 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7xsc4\" (UniqueName: \"kubernetes.io/projected/28f73b7a-bb5e-43ec-a099-13a12cdb33d3-kube-api-access-7xsc4\") pod \"novaapiebe4-account-delete-rph2r\" (UID: \"28f73b7a-bb5e-43ec-a099-13a12cdb33d3\") " pod="openstack/novaapiebe4-account-delete-rph2r" Oct 08 18:32:58 crc kubenswrapper[4988]: I1008 18:32:58.173146 4988 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/swift-ring-rebalance-9hjmc"] Oct 08 18:32:58 crc kubenswrapper[4988]: I1008 18:32:58.192302 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f6q49\" (UniqueName: \"kubernetes.io/projected/3e8c874a-029a-4d6a-bc2e-00aeed1636a7-kube-api-access-f6q49\") pod \"novacell02f70-account-delete-cdwsx\" (UID: \"3e8c874a-029a-4d6a-bc2e-00aeed1636a7\") " pod="openstack/novacell02f70-account-delete-cdwsx" Oct 08 18:32:58 crc kubenswrapper[4988]: I1008 18:32:58.252405 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/novacell14316-account-delete-gbtkl"] Oct 08 18:32:58 crc kubenswrapper[4988]: I1008 18:32:58.253611 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/novacell14316-account-delete-gbtkl" Oct 08 18:32:58 crc kubenswrapper[4988]: I1008 18:32:58.277270 4988 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_f026ec8f-ffb4-4538-a6fc-3116ce0e7805/ovn-northd/0.log" Oct 08 18:32:58 crc kubenswrapper[4988]: I1008 18:32:58.277405 4988 generic.go:334] "Generic (PLEG): container finished" podID="f026ec8f-ffb4-4538-a6fc-3116ce0e7805" containerID="ada91b938d451d0675a819f062dafde62f4a75b785dc86a5448ce25e246cecbb" exitCode=2 Oct 08 18:32:58 crc kubenswrapper[4988]: I1008 18:32:58.277482 4988 generic.go:334] "Generic (PLEG): container finished" podID="f026ec8f-ffb4-4538-a6fc-3116ce0e7805" containerID="e9efac4802d1c9a2377dc285da1c498c1363ff5f70695e54b4a257848bdea958" exitCode=143 Oct 08 18:32:58 crc kubenswrapper[4988]: I1008 18:32:58.277614 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"f026ec8f-ffb4-4538-a6fc-3116ce0e7805","Type":"ContainerDied","Data":"ada91b938d451d0675a819f062dafde62f4a75b785dc86a5448ce25e246cecbb"} Oct 08 18:32:58 crc kubenswrapper[4988]: I1008 18:32:58.277711 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"f026ec8f-ffb4-4538-a6fc-3116ce0e7805","Type":"ContainerDied","Data":"e9efac4802d1c9a2377dc285da1c498c1363ff5f70695e54b4a257848bdea958"} Oct 08 18:32:58 crc kubenswrapper[4988]: I1008 18:32:58.278153 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron3849-account-delete-m479b" Oct 08 18:32:58 crc kubenswrapper[4988]: I1008 18:32:58.302997 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-sync-dm56p"] Oct 08 18:32:58 crc kubenswrapper[4988]: I1008 18:32:58.305507 4988 generic.go:334] "Generic (PLEG): container finished" podID="99d77eed-40f0-4879-be96-abea3a807cd9" containerID="1c61da6cc2966ef990d9bcdbb3737986e369db20fe8b42772368a61ef7dc1f03" exitCode=143 Oct 08 18:32:58 crc kubenswrapper[4988]: I1008 18:32:58.305574 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"99d77eed-40f0-4879-be96-abea3a807cd9","Type":"ContainerDied","Data":"1c61da6cc2966ef990d9bcdbb3737986e369db20fe8b42772368a61ef7dc1f03"} Oct 08 18:32:58 crc kubenswrapper[4988]: I1008 18:32:58.348290 4988 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-sync-dm56p"] Oct 08 18:32:58 crc kubenswrapper[4988]: I1008 18:32:58.359886 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cjfgk\" (UniqueName: \"kubernetes.io/projected/1127bf6b-65ff-4c59-928c-51e16bb18ce3-kube-api-access-cjfgk\") pod \"novacell14316-account-delete-gbtkl\" (UID: \"1127bf6b-65ff-4c59-928c-51e16bb18ce3\") " pod="openstack/novacell14316-account-delete-gbtkl" Oct 08 18:32:58 crc kubenswrapper[4988]: I1008 18:32:58.399242 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/novacell14316-account-delete-gbtkl"] Oct 08 18:32:58 crc kubenswrapper[4988]: I1008 18:32:58.451374 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-sync-wvjsl"] Oct 08 18:32:58 crc kubenswrapper[4988]: I1008 18:32:58.477501 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cjfgk\" (UniqueName: \"kubernetes.io/projected/1127bf6b-65ff-4c59-928c-51e16bb18ce3-kube-api-access-cjfgk\") pod \"novacell14316-account-delete-gbtkl\" (UID: \"1127bf6b-65ff-4c59-928c-51e16bb18ce3\") " pod="openstack/novacell14316-account-delete-gbtkl" Oct 08 18:32:58 crc kubenswrapper[4988]: I1008 18:32:58.479320 4988 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-sync-wvjsl"] Oct 08 18:32:58 crc kubenswrapper[4988]: E1008 18:32:58.504479 4988 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb92d978b_87c6_4210_9a69_99596b05cc18.slice/crio-80822f2e64d8fd8a2c3f08eea237f88c03ba055c04d5bb24498177ab69c61a2e.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb92d978b_87c6_4210_9a69_99596b05cc18.slice/crio-conmon-80822f2e64d8fd8a2c3f08eea237f88c03ba055c04d5bb24498177ab69c61a2e.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1f36756f_acb3_439d_b10e_13573cbd252a.slice/crio-a41e320aa9f3ac4edb764f4ac4f3c561a463787b781e74385ca8530f54e7b28e.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf026ec8f_ffb4_4538_a6fc_3116ce0e7805.slice/crio-conmon-ada91b938d451d0675a819f062dafde62f4a75b785dc86a5448ce25e246cecbb.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf026ec8f_ffb4_4538_a6fc_3116ce0e7805.slice/crio-e9efac4802d1c9a2377dc285da1c498c1363ff5f70695e54b4a257848bdea958.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod99d77eed_40f0_4879_be96_abea3a807cd9.slice/crio-conmon-1c61da6cc2966ef990d9bcdbb3737986e369db20fe8b42772368a61ef7dc1f03.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf026ec8f_ffb4_4538_a6fc_3116ce0e7805.slice/crio-conmon-e9efac4802d1c9a2377dc285da1c498c1363ff5f70695e54b4a257848bdea958.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod99d77eed_40f0_4879_be96_abea3a807cd9.slice/crio-1c61da6cc2966ef990d9bcdbb3737986e369db20fe8b42772368a61ef7dc1f03.scope\": RecentStats: unable to find data in memory cache]" Oct 08 18:32:58 crc kubenswrapper[4988]: E1008 18:32:58.505784 4988 projected.go:194] Error preparing data for projected volume kube-api-access-cjfgk for pod openstack/novacell14316-account-delete-gbtkl: failed to fetch token: serviceaccounts "galera-openstack-cell1" not found Oct 08 18:32:58 crc kubenswrapper[4988]: E1008 18:32:58.505846 4988 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/1127bf6b-65ff-4c59-928c-51e16bb18ce3-kube-api-access-cjfgk podName:1127bf6b-65ff-4c59-928c-51e16bb18ce3 nodeName:}" failed. No retries permitted until 2025-10-08 18:32:59.005828039 +0000 UTC m=+1324.455670809 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-cjfgk" (UniqueName: "kubernetes.io/projected/1127bf6b-65ff-4c59-928c-51e16bb18ce3-kube-api-access-cjfgk") pod "novacell14316-account-delete-gbtkl" (UID: "1127bf6b-65ff-4c59-928c-51e16bb18ce3") : failed to fetch token: serviceaccounts "galera-openstack-cell1" not found Oct 08 18:32:58 crc kubenswrapper[4988]: I1008 18:32:58.506062 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-c44756996-dfw4k"] Oct 08 18:32:58 crc kubenswrapper[4988]: I1008 18:32:58.506258 4988 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/placement-c44756996-dfw4k" podUID="f4e47608-89b4-45d1-88a6-fe75805d864d" containerName="placement-log" containerID="cri-o://13f8a712b04c6d00db8bc2f724015fda6642a1723ac5c2e59ecbd2400fa35ece" gracePeriod=30 Oct 08 18:32:58 crc kubenswrapper[4988]: I1008 18:32:58.506566 4988 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/placement-c44756996-dfw4k" podUID="f4e47608-89b4-45d1-88a6-fe75805d864d" containerName="placement-api" containerID="cri-o://b226456d2f92fef2d798a8462e29b0c8f7c8a915dd5330879cb7257e0cc73d05" gracePeriod=30 Oct 08 18:32:58 crc kubenswrapper[4988]: I1008 18:32:58.544204 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinderd8da-account-delete-75j2h"] Oct 08 18:32:58 crc kubenswrapper[4988]: I1008 18:32:58.559755 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-cell-mapping-tm2wd"] Oct 08 18:32:58 crc kubenswrapper[4988]: I1008 18:32:58.559856 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinderd8da-account-delete-75j2h" Oct 08 18:32:58 crc kubenswrapper[4988]: I1008 18:32:58.566586 4988 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-metrics-w2zhj_b92d978b-87c6-4210-9a69-99596b05cc18/openstack-network-exporter/0.log" Oct 08 18:32:58 crc kubenswrapper[4988]: I1008 18:32:58.566640 4988 generic.go:334] "Generic (PLEG): container finished" podID="b92d978b-87c6-4210-9a69-99596b05cc18" containerID="80822f2e64d8fd8a2c3f08eea237f88c03ba055c04d5bb24498177ab69c61a2e" exitCode=2 Oct 08 18:32:58 crc kubenswrapper[4988]: I1008 18:32:58.566758 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-w2zhj" event={"ID":"b92d978b-87c6-4210-9a69-99596b05cc18","Type":"ContainerDied","Data":"80822f2e64d8fd8a2c3f08eea237f88c03ba055c04d5bb24498177ab69c61a2e"} Oct 08 18:32:58 crc kubenswrapper[4988]: I1008 18:32:58.613141 4988 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_1f36756f-acb3-439d-b10e-13573cbd252a/ovsdbserver-sb/0.log" Oct 08 18:32:58 crc kubenswrapper[4988]: I1008 18:32:58.613375 4988 generic.go:334] "Generic (PLEG): container finished" podID="1f36756f-acb3-439d-b10e-13573cbd252a" containerID="e8f79299e7fcca754183f5f8a133b9bb867b1fa107190b02e938976ad478eb0b" exitCode=2 Oct 08 18:32:58 crc kubenswrapper[4988]: I1008 18:32:58.613408 4988 generic.go:334] "Generic (PLEG): container finished" podID="1f36756f-acb3-439d-b10e-13573cbd252a" containerID="a41e320aa9f3ac4edb764f4ac4f3c561a463787b781e74385ca8530f54e7b28e" exitCode=143 Oct 08 18:32:58 crc kubenswrapper[4988]: I1008 18:32:58.613430 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"1f36756f-acb3-439d-b10e-13573cbd252a","Type":"ContainerDied","Data":"e8f79299e7fcca754183f5f8a133b9bb867b1fa107190b02e938976ad478eb0b"} Oct 08 18:32:58 crc kubenswrapper[4988]: I1008 18:32:58.613433 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/novacell02f70-account-delete-cdwsx" Oct 08 18:32:58 crc kubenswrapper[4988]: I1008 18:32:58.613456 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"1f36756f-acb3-439d-b10e-13573cbd252a","Type":"ContainerDied","Data":"a41e320aa9f3ac4edb764f4ac4f3c561a463787b781e74385ca8530f54e7b28e"} Oct 08 18:32:58 crc kubenswrapper[4988]: I1008 18:32:58.614987 4988 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-cell-mapping-tm2wd"] Oct 08 18:32:58 crc kubenswrapper[4988]: I1008 18:32:58.638065 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinderd8da-account-delete-75j2h"] Oct 08 18:32:58 crc kubenswrapper[4988]: I1008 18:32:58.638881 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/novaapiebe4-account-delete-rph2r" Oct 08 18:32:58 crc kubenswrapper[4988]: I1008 18:32:58.669600 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 08 18:32:58 crc kubenswrapper[4988]: I1008 18:32:58.691079 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gxjkk\" (UniqueName: \"kubernetes.io/projected/02a37054-4817-4151-a8cb-a9a45867c0d3-kube-api-access-gxjkk\") pod \"cinderd8da-account-delete-75j2h\" (UID: \"02a37054-4817-4151-a8cb-a9a45867c0d3\") " pod="openstack/cinderd8da-account-delete-75j2h" Oct 08 18:32:58 crc kubenswrapper[4988]: I1008 18:32:58.693430 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-cell-mapping-92jjg"] Oct 08 18:32:58 crc kubenswrapper[4988]: I1008 18:32:58.708971 4988 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-cell-mapping-92jjg"] Oct 08 18:32:58 crc kubenswrapper[4988]: I1008 18:32:58.715888 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovsdbserver-nb-0"] Oct 08 18:32:58 crc kubenswrapper[4988]: I1008 18:32:58.716747 4988 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ovsdbserver-nb-0" podUID="b5dc1ee3-4897-45a2-ad00-0b91fbb2ae76" containerName="openstack-network-exporter" containerID="cri-o://b5799148279b22815a3dbd06ed629a27e96e308b0f484f69614b62b235afbb9f" gracePeriod=300 Oct 08 18:32:58 crc kubenswrapper[4988]: I1008 18:32:58.723028 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5967cc9597-r6jrv"] Oct 08 18:32:58 crc kubenswrapper[4988]: I1008 18:32:58.724684 4988 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-5967cc9597-r6jrv" podUID="467e736b-562a-4f31-af64-2fce55e965c8" containerName="dnsmasq-dns" containerID="cri-o://9933c83a6d7a8c9610edd4f70255f0f91bccecf1b14bb8fd883bcb6b9506666f" gracePeriod=10 Oct 08 18:32:58 crc kubenswrapper[4988]: I1008 18:32:58.729317 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-sync-gm2vf"] Oct 08 18:32:58 crc kubenswrapper[4988]: I1008 18:32:58.741611 4988 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-sync-gm2vf"] Oct 08 18:32:58 crc kubenswrapper[4988]: I1008 18:32:58.761557 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 08 18:32:58 crc kubenswrapper[4988]: I1008 18:32:58.761802 4988 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="0fcc9499-a06b-4cd1-a5da-ea8108916050" containerName="glance-log" containerID="cri-o://7d3dbf100ba64891a97f34dd540d20ec061ba6822f245dec78c63c17c5d178fa" gracePeriod=30 Oct 08 18:32:58 crc kubenswrapper[4988]: I1008 18:32:58.762050 4988 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="0fcc9499-a06b-4cd1-a5da-ea8108916050" containerName="glance-httpd" containerID="cri-o://9315fcc87c2b03972c71e57728c4fdbc3b30200edc121e010b7ac2e7d7658e92" gracePeriod=30 Oct 08 18:32:58 crc kubenswrapper[4988]: I1008 18:32:58.773522 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-6bb998f7d5-h4g2x"] Oct 08 18:32:58 crc kubenswrapper[4988]: I1008 18:32:58.773743 4988 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-6bb998f7d5-h4g2x" podUID="6d2df307-23d3-48fd-9bb6-f09e500aaf9e" containerName="neutron-api" containerID="cri-o://ed0f1f9337e028007e93109bea68d88246a9ba24b639073b907414a6728b9835" gracePeriod=30 Oct 08 18:32:58 crc kubenswrapper[4988]: I1008 18:32:58.774081 4988 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-6bb998f7d5-h4g2x" podUID="6d2df307-23d3-48fd-9bb6-f09e500aaf9e" containerName="neutron-httpd" containerID="cri-o://db863ec0ad9c9861590e77462403049f0e352d2f01da17f2db5e4682b149163e" gracePeriod=30 Oct 08 18:32:58 crc kubenswrapper[4988]: I1008 18:32:58.785154 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/swift-storage-0"] Oct 08 18:32:58 crc kubenswrapper[4988]: I1008 18:32:58.785642 4988 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="d04259e5-9595-447b-b804-feba577498d8" containerName="account-server" containerID="cri-o://d0e7528b976d6c731e5ab211ed178a97c7d155611aadc8e3e7dde446d2824719" gracePeriod=30 Oct 08 18:32:58 crc kubenswrapper[4988]: I1008 18:32:58.785674 4988 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="d04259e5-9595-447b-b804-feba577498d8" containerName="object-server" containerID="cri-o://94be3d977f16a3f88c65d1405dd7fd0737c52d3e6c73b44f0651409fc2428746" gracePeriod=30 Oct 08 18:32:58 crc kubenswrapper[4988]: I1008 18:32:58.785785 4988 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="d04259e5-9595-447b-b804-feba577498d8" containerName="container-updater" containerID="cri-o://31e32c3833aea9f83bda1aca6904be1c190d6d671f11ad9f856ce5abce9053fb" gracePeriod=30 Oct 08 18:32:58 crc kubenswrapper[4988]: I1008 18:32:58.785840 4988 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="d04259e5-9595-447b-b804-feba577498d8" containerName="container-auditor" containerID="cri-o://4f7780b6a0c60b8247000b0296848f37576f4cd48b317bfb6bd9807dc977f068" gracePeriod=30 Oct 08 18:32:58 crc kubenswrapper[4988]: I1008 18:32:58.785894 4988 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="d04259e5-9595-447b-b804-feba577498d8" containerName="container-replicator" containerID="cri-o://c1f4256e3ed5eca7a8d3e7c8f01715b4bf6cc7778eac4b717acbddef65864b05" gracePeriod=30 Oct 08 18:32:58 crc kubenswrapper[4988]: I1008 18:32:58.785897 4988 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="d04259e5-9595-447b-b804-feba577498d8" containerName="object-expirer" containerID="cri-o://f6d8e476ef92ad2a5ae0468a14ebdbcb37b6c35a714f5c83ec9ff493a5dd15a3" gracePeriod=30 Oct 08 18:32:58 crc kubenswrapper[4988]: I1008 18:32:58.785940 4988 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="d04259e5-9595-447b-b804-feba577498d8" containerName="container-server" containerID="cri-o://590b8a61b67584c2acbd2bea77795253fad7c7d2a79eb0fb7fde120c857233af" gracePeriod=30 Oct 08 18:32:58 crc kubenswrapper[4988]: I1008 18:32:58.785983 4988 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="d04259e5-9595-447b-b804-feba577498d8" containerName="account-reaper" containerID="cri-o://e8f70c43dc552f28dc00baeb404e73ce329584c5d7563ccfd1bb7e213c973bc7" gracePeriod=30 Oct 08 18:32:58 crc kubenswrapper[4988]: I1008 18:32:58.786014 4988 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="d04259e5-9595-447b-b804-feba577498d8" containerName="swift-recon-cron" containerID="cri-o://b723050435b720195f0b9b9fbdfc7bf95b8dd19f190a2d21120788a8cde0603f" gracePeriod=30 Oct 08 18:32:58 crc kubenswrapper[4988]: I1008 18:32:58.786024 4988 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="d04259e5-9595-447b-b804-feba577498d8" containerName="account-auditor" containerID="cri-o://588710f5625320484d4ef400733b3ae63b0e2f936f619df2ca3b5435aa10a8c9" gracePeriod=30 Oct 08 18:32:58 crc kubenswrapper[4988]: I1008 18:32:58.786051 4988 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="d04259e5-9595-447b-b804-feba577498d8" containerName="rsync" containerID="cri-o://3e8f7969a53f0a0b492bfa6e755ebabd4ef007e6b42af4f280d07a7401c0699d" gracePeriod=30 Oct 08 18:32:58 crc kubenswrapper[4988]: I1008 18:32:58.786053 4988 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="d04259e5-9595-447b-b804-feba577498d8" containerName="object-updater" containerID="cri-o://9935bea63badb3e4cb30261f8985a1224e374fbe56ec5af0a913e1c5fe57db04" gracePeriod=30 Oct 08 18:32:58 crc kubenswrapper[4988]: I1008 18:32:58.786089 4988 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="d04259e5-9595-447b-b804-feba577498d8" containerName="account-replicator" containerID="cri-o://23ee960ca35b0ceb291c74103d14d93f4ff48fb0b2b08acdacc1fd66a0fb6327" gracePeriod=30 Oct 08 18:32:58 crc kubenswrapper[4988]: I1008 18:32:58.786110 4988 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="d04259e5-9595-447b-b804-feba577498d8" containerName="object-auditor" containerID="cri-o://0f92f3d1af5000c50ceb5a04475cb2043188883ce2d85323918a574561017295" gracePeriod=30 Oct 08 18:32:58 crc kubenswrapper[4988]: I1008 18:32:58.786146 4988 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="d04259e5-9595-447b-b804-feba577498d8" containerName="object-replicator" containerID="cri-o://6c62379d0b1b37e153eff91e53dd273d7d1e85bb5f09dd3abdfc7a5b862d3459" gracePeriod=30 Oct 08 18:32:58 crc kubenswrapper[4988]: E1008 18:32:58.790786 4988 upgradeaware.go:427] Error proxying data from client to backend: readfrom tcp 38.102.83.102:45560->38.102.83.102:38495: write tcp 38.102.83.102:45560->38.102.83.102:38495: write: broken pipe Oct 08 18:32:58 crc kubenswrapper[4988]: I1008 18:32:58.796215 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gxjkk\" (UniqueName: \"kubernetes.io/projected/02a37054-4817-4151-a8cb-a9a45867c0d3-kube-api-access-gxjkk\") pod \"cinderd8da-account-delete-75j2h\" (UID: \"02a37054-4817-4151-a8cb-a9a45867c0d3\") " pod="openstack/cinderd8da-account-delete-75j2h" Oct 08 18:32:58 crc kubenswrapper[4988]: E1008 18:32:58.797977 4988 configmap.go:193] Couldn't get configMap openstack/rabbitmq-config-data: configmap "rabbitmq-config-data" not found Oct 08 18:32:58 crc kubenswrapper[4988]: E1008 18:32:58.798036 4988 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/830d6ec5-0015-4a88-b8e8-0fd47f64fd46-config-data podName:830d6ec5-0015-4a88-b8e8-0fd47f64fd46 nodeName:}" failed. No retries permitted until 2025-10-08 18:32:59.298015908 +0000 UTC m=+1324.747858678 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/configmap/830d6ec5-0015-4a88-b8e8-0fd47f64fd46-config-data") pod "rabbitmq-server-0" (UID: "830d6ec5-0015-4a88-b8e8-0fd47f64fd46") : configmap "rabbitmq-config-data" not found Oct 08 18:32:58 crc kubenswrapper[4988]: I1008 18:32:58.798309 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 08 18:32:58 crc kubenswrapper[4988]: I1008 18:32:58.807961 4988 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="56702a03-175d-4a54-b84f-e25cab037431" containerName="glance-log" containerID="cri-o://9f5e2ee884e53117065985ec938aeba8be9ae37755acfba2392c435832c65f0e" gracePeriod=30 Oct 08 18:32:58 crc kubenswrapper[4988]: I1008 18:32:58.808427 4988 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="56702a03-175d-4a54-b84f-e25cab037431" containerName="glance-httpd" containerID="cri-o://23da0ff1f117dc601e37e2bb0a79de17b4662200fa002a471f0c5938e67d484b" gracePeriod=30 Oct 08 18:32:58 crc kubenswrapper[4988]: I1008 18:32:58.808615 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/swift-proxy-7b7cf6bf4f-s6vrs"] Oct 08 18:32:58 crc kubenswrapper[4988]: I1008 18:32:58.808756 4988 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-proxy-7b7cf6bf4f-s6vrs" podUID="857686e3-0c6b-4d08-a8d6-f89e4c832139" containerName="proxy-httpd" containerID="cri-o://42e44fdf6315337e912e8f08c104630a1408f1ece9b5a29dc6c2ddc4b1e15bf2" gracePeriod=30 Oct 08 18:32:58 crc kubenswrapper[4988]: I1008 18:32:58.808815 4988 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-proxy-7b7cf6bf4f-s6vrs" podUID="857686e3-0c6b-4d08-a8d6-f89e4c832139" containerName="proxy-server" containerID="cri-o://ee0c93d867a043a6ace66d0ccc23af52eb47335f2734cbb64f7f04eb41db6e3c" gracePeriod=30 Oct 08 18:32:58 crc kubenswrapper[4988]: I1008 18:32:58.836132 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gxjkk\" (UniqueName: \"kubernetes.io/projected/02a37054-4817-4151-a8cb-a9a45867c0d3-kube-api-access-gxjkk\") pod \"cinderd8da-account-delete-75j2h\" (UID: \"02a37054-4817-4151-a8cb-a9a45867c0d3\") " pod="openstack/cinderd8da-account-delete-75j2h" Oct 08 18:32:58 crc kubenswrapper[4988]: I1008 18:32:58.853129 4988 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ovsdbserver-nb-0" podUID="b5dc1ee3-4897-45a2-ad00-0b91fbb2ae76" containerName="ovsdbserver-nb" containerID="cri-o://74829553075c7054526031a117b72832271514093fbe3617d9278cd92e26275b" gracePeriod=300 Oct 08 18:32:58 crc kubenswrapper[4988]: I1008 18:32:58.903259 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Oct 08 18:32:58 crc kubenswrapper[4988]: I1008 18:32:58.904124 4988 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="83ef7610-23ac-4445-b75c-ae7206b4d810" containerName="nova-metadata-log" containerID="cri-o://e504bf27bfc24744eb7b0bdc9918cda7b788dc1c32411c922014b5f7f84ae75d" gracePeriod=30 Oct 08 18:32:58 crc kubenswrapper[4988]: I1008 18:32:58.904558 4988 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="83ef7610-23ac-4445-b75c-ae7206b4d810" containerName="nova-metadata-metadata" containerID="cri-o://4b994190cba63b85214c41a6bf0e9b5cca7c29dd236d40d5df3055f98c9ec35b" gracePeriod=30 Oct 08 18:32:58 crc kubenswrapper[4988]: I1008 18:32:58.941693 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 08 18:32:58 crc kubenswrapper[4988]: I1008 18:32:58.942136 4988 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="0efd1f53-90ad-4eb8-90f2-2ba9fc1593c2" containerName="nova-api-log" containerID="cri-o://4cdfab290ec25786ba40c3a3d5d3687398ae45094d90768455e35c6642188b15" gracePeriod=30 Oct 08 18:32:58 crc kubenswrapper[4988]: I1008 18:32:58.942609 4988 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="0efd1f53-90ad-4eb8-90f2-2ba9fc1593c2" containerName="nova-api-api" containerID="cri-o://a81aba18e5412a571b8ceb649f5e04c6698ca4b21c30f54d96848df1d1a823b6" gracePeriod=30 Oct 08 18:32:58 crc kubenswrapper[4988]: I1008 18:32:58.982220 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinderd8da-account-delete-75j2h" Oct 08 18:32:58 crc kubenswrapper[4988]: I1008 18:32:58.990808 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/openstack-cell1-galera-0"] Oct 08 18:32:59 crc kubenswrapper[4988]: E1008 18:32:59.001903 4988 configmap.go:193] Couldn't get configMap openstack/rabbitmq-cell1-config-data: configmap "rabbitmq-cell1-config-data" not found Oct 08 18:32:59 crc kubenswrapper[4988]: E1008 18:32:59.001979 4988 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/e5a8f819-c91d-429e-b848-e9c444739623-config-data podName:e5a8f819-c91d-429e-b848-e9c444739623 nodeName:}" failed. No retries permitted until 2025-10-08 18:33:01.001959481 +0000 UTC m=+1326.451802251 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/configmap/e5a8f819-c91d-429e-b848-e9c444739623-config-data") pod "rabbitmq-cell1-server-0" (UID: "e5a8f819-c91d-429e-b848-e9c444739623") : configmap "rabbitmq-cell1-config-data" not found Oct 08 18:32:59 crc kubenswrapper[4988]: I1008 18:32:59.043215 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-worker-7bf697f8df-h54ft"] Oct 08 18:32:59 crc kubenswrapper[4988]: I1008 18:32:59.066303 4988 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-worker-7bf697f8df-h54ft" podUID="9159c915-07a7-4b41-8672-8b9d6a9572ca" containerName="barbican-worker-log" containerID="cri-o://2927fb800c44516806e01d37c729bc04efb6e3981dccab523101ed4405525b47" gracePeriod=30 Oct 08 18:32:59 crc kubenswrapper[4988]: I1008 18:32:59.066721 4988 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-worker-7bf697f8df-h54ft" podUID="9159c915-07a7-4b41-8672-8b9d6a9572ca" containerName="barbican-worker" containerID="cri-o://1f77c788702456fc178f26b1d4a4449a94e1ac2f3f33ad505531b3885e2f8f15" gracePeriod=30 Oct 08 18:32:59 crc kubenswrapper[4988]: I1008 18:32:59.107409 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cjfgk\" (UniqueName: \"kubernetes.io/projected/1127bf6b-65ff-4c59-928c-51e16bb18ce3-kube-api-access-cjfgk\") pod \"novacell14316-account-delete-gbtkl\" (UID: \"1127bf6b-65ff-4c59-928c-51e16bb18ce3\") " pod="openstack/novacell14316-account-delete-gbtkl" Oct 08 18:32:59 crc kubenswrapper[4988]: E1008 18:32:59.112763 4988 projected.go:194] Error preparing data for projected volume kube-api-access-cjfgk for pod openstack/novacell14316-account-delete-gbtkl: failed to fetch token: serviceaccounts "galera-openstack-cell1" not found Oct 08 18:32:59 crc kubenswrapper[4988]: E1008 18:32:59.113012 4988 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/1127bf6b-65ff-4c59-928c-51e16bb18ce3-kube-api-access-cjfgk podName:1127bf6b-65ff-4c59-928c-51e16bb18ce3 nodeName:}" failed. No retries permitted until 2025-10-08 18:33:00.112978738 +0000 UTC m=+1325.562821508 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-cjfgk" (UniqueName: "kubernetes.io/projected/1127bf6b-65ff-4c59-928c-51e16bb18ce3-kube-api-access-cjfgk") pod "novacell14316-account-delete-gbtkl" (UID: "1127bf6b-65ff-4c59-928c-51e16bb18ce3") : failed to fetch token: serviceaccounts "galera-openstack-cell1" not found Oct 08 18:32:59 crc kubenswrapper[4988]: I1008 18:32:59.113312 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-keystone-listener-8694c889b8-lhm4g"] Oct 08 18:32:59 crc kubenswrapper[4988]: I1008 18:32:59.113695 4988 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-keystone-listener-8694c889b8-lhm4g" podUID="c9f172a3-9188-4f5f-bf3a-fbc3b0dc5206" containerName="barbican-keystone-listener-log" containerID="cri-o://c7e441bdd6242790ab99582ae7f776a6bd322dc0c9ca3985c984d45c5cc43a5c" gracePeriod=30 Oct 08 18:32:59 crc kubenswrapper[4988]: I1008 18:32:59.114175 4988 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-keystone-listener-8694c889b8-lhm4g" podUID="c9f172a3-9188-4f5f-bf3a-fbc3b0dc5206" containerName="barbican-keystone-listener" containerID="cri-o://3251e38d412d5c6a4e41496647e54dfc1bc7728f57aa1932956313fe00c0a86b" gracePeriod=30 Oct 08 18:32:59 crc kubenswrapper[4988]: I1008 18:32:59.125224 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-6c486d5744-j86zn"] Oct 08 18:32:59 crc kubenswrapper[4988]: I1008 18:32:59.127443 4988 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-6c486d5744-j86zn" podUID="3a57537b-6115-4249-8d16-3ab0c4b6d21e" containerName="barbican-api-log" containerID="cri-o://8ef71a96a488da8760db257dae569210b2903674028f06127874742bfeb624fd" gracePeriod=30 Oct 08 18:32:59 crc kubenswrapper[4988]: I1008 18:32:59.128446 4988 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-6c486d5744-j86zn" podUID="3a57537b-6115-4249-8d16-3ab0c4b6d21e" containerName="barbican-api" containerID="cri-o://0f23a2df4066a02929dd3edb8dc8d068809211dbfad9bf36b14e3f21916284e7" gracePeriod=30 Oct 08 18:32:59 crc kubenswrapper[4988]: I1008 18:32:59.139363 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 08 18:32:59 crc kubenswrapper[4988]: I1008 18:32:59.146851 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-ebe4-account-create-c64w6"] Oct 08 18:32:59 crc kubenswrapper[4988]: I1008 18:32:59.197259 4988 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-ebe4-account-create-c64w6"] Oct 08 18:32:59 crc kubenswrapper[4988]: I1008 18:32:59.278803 4988 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2e2a1e47-ca67-445c-b358-ad7e5bce08ea" path="/var/lib/kubelet/pods/2e2a1e47-ca67-445c-b358-ad7e5bce08ea/volumes" Oct 08 18:32:59 crc kubenswrapper[4988]: I1008 18:32:59.280170 4988 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="567a67dc-01a7-4720-8081-d60470856578" path="/var/lib/kubelet/pods/567a67dc-01a7-4720-8081-d60470856578/volumes" Oct 08 18:32:59 crc kubenswrapper[4988]: I1008 18:32:59.288115 4988 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5b6f72e9-8dc6-4290-b032-d8a570b91767" path="/var/lib/kubelet/pods/5b6f72e9-8dc6-4290-b032-d8a570b91767/volumes" Oct 08 18:32:59 crc kubenswrapper[4988]: I1008 18:32:59.289100 4988 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="62b5df6a-6198-4608-8cd7-05bfa27bfd75" path="/var/lib/kubelet/pods/62b5df6a-6198-4608-8cd7-05bfa27bfd75/volumes" Oct 08 18:32:59 crc kubenswrapper[4988]: I1008 18:32:59.289673 4988 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c132f5da-fe6a-4616-9d2f-bc0c046244a1" path="/var/lib/kubelet/pods/c132f5da-fe6a-4616-9d2f-bc0c046244a1/volumes" Oct 08 18:32:59 crc kubenswrapper[4988]: I1008 18:32:59.290775 4988 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c1f2511c-e7bc-4433-910b-eb3b94fcf09b" path="/var/lib/kubelet/pods/c1f2511c-e7bc-4433-910b-eb3b94fcf09b/volumes" Oct 08 18:32:59 crc kubenswrapper[4988]: I1008 18:32:59.291723 4988 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d7f4f7aa-ccf2-45e1-8d32-db9684853231" path="/var/lib/kubelet/pods/d7f4f7aa-ccf2-45e1-8d32-db9684853231/volumes" Oct 08 18:32:59 crc kubenswrapper[4988]: I1008 18:32:59.292443 4988 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="de747e73-c280-424f-af48-3135bec91b55" path="/var/lib/kubelet/pods/de747e73-c280-424f-af48-3135bec91b55/volumes" Oct 08 18:32:59 crc kubenswrapper[4988]: I1008 18:32:59.292998 4988 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e3e8eb1b-f0b5-4e56-9ba7-d068435355c5" path="/var/lib/kubelet/pods/e3e8eb1b-f0b5-4e56-9ba7-d068435355c5/volumes" Oct 08 18:32:59 crc kubenswrapper[4988]: I1008 18:32:59.303986 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-db-create-nxrck"] Oct 08 18:32:59 crc kubenswrapper[4988]: I1008 18:32:59.304027 4988 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-db-create-nxrck"] Oct 08 18:32:59 crc kubenswrapper[4988]: I1008 18:32:59.304043 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/novaapiebe4-account-delete-rph2r"] Oct 08 18:32:59 crc kubenswrapper[4988]: E1008 18:32:59.332514 4988 configmap.go:193] Couldn't get configMap openstack/rabbitmq-config-data: configmap "rabbitmq-config-data" not found Oct 08 18:32:59 crc kubenswrapper[4988]: E1008 18:32:59.332725 4988 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/830d6ec5-0015-4a88-b8e8-0fd47f64fd46-config-data podName:830d6ec5-0015-4a88-b8e8-0fd47f64fd46 nodeName:}" failed. No retries permitted until 2025-10-08 18:33:00.332706206 +0000 UTC m=+1325.782548976 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/configmap/830d6ec5-0015-4a88-b8e8-0fd47f64fd46-config-data") pod "rabbitmq-server-0" (UID: "830d6ec5-0015-4a88-b8e8-0fd47f64fd46") : configmap "rabbitmq-config-data" not found Oct 08 18:32:59 crc kubenswrapper[4988]: I1008 18:32:59.337453 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-db-create-bz99d"] Oct 08 18:32:59 crc kubenswrapper[4988]: I1008 18:32:59.385573 4988 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-db-create-bz99d"] Oct 08 18:32:59 crc kubenswrapper[4988]: I1008 18:32:59.395995 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-4316-account-create-5kjmj"] Oct 08 18:32:59 crc kubenswrapper[4988]: I1008 18:32:59.406470 4988 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-4316-account-create-5kjmj"] Oct 08 18:32:59 crc kubenswrapper[4988]: I1008 18:32:59.433539 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/novacell14316-account-delete-gbtkl"] Oct 08 18:32:59 crc kubenswrapper[4988]: E1008 18:32:59.434283 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="unmounted volumes=[kube-api-access-cjfgk], unattached volumes=[], failed to process volumes=[]: context canceled" pod="openstack/novacell14316-account-delete-gbtkl" podUID="1127bf6b-65ff-4c59-928c-51e16bb18ce3" Oct 08 18:32:59 crc kubenswrapper[4988]: I1008 18:32:59.436455 4988 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/openstack-cell1-galera-0" podUID="3503942b-2825-4006-80e8-8c4610b89997" containerName="galera" containerID="cri-o://e1a685c92e7a7b5fad1185fbe08be64ed9f96221d3c3a8c8c5d3364e9a593229" gracePeriod=30 Oct 08 18:32:59 crc kubenswrapper[4988]: I1008 18:32:59.439212 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 08 18:32:59 crc kubenswrapper[4988]: I1008 18:32:59.439447 4988 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell1-novncproxy-0" podUID="816b358d-4428-4fae-b91c-03c0171faa23" containerName="nova-cell1-novncproxy-novncproxy" containerID="cri-o://a947f771405cb73547e24443c664920da93801df8ff9dd6369c101deea715564" gracePeriod=30 Oct 08 18:32:59 crc kubenswrapper[4988]: I1008 18:32:59.445638 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-create-wt7bm"] Oct 08 18:32:59 crc kubenswrapper[4988]: I1008 18:32:59.451503 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-d8da-account-create-fxpc9"] Oct 08 18:32:59 crc kubenswrapper[4988]: I1008 18:32:59.460523 4988 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-create-wt7bm"] Oct 08 18:32:59 crc kubenswrapper[4988]: I1008 18:32:59.482970 4988 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-d8da-account-create-fxpc9"] Oct 08 18:32:59 crc kubenswrapper[4988]: I1008 18:32:59.487229 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinderd8da-account-delete-75j2h"] Oct 08 18:32:59 crc kubenswrapper[4988]: I1008 18:32:59.502988 4988 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_1f36756f-acb3-439d-b10e-13573cbd252a/ovsdbserver-sb/0.log" Oct 08 18:32:59 crc kubenswrapper[4988]: I1008 18:32:59.503084 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Oct 08 18:32:59 crc kubenswrapper[4988]: I1008 18:32:59.504342 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-rx7cw"] Oct 08 18:32:59 crc kubenswrapper[4988]: I1008 18:32:59.514948 4988 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-rx7cw"] Oct 08 18:32:59 crc kubenswrapper[4988]: I1008 18:32:59.530396 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-hzsjj"] Oct 08 18:32:59 crc kubenswrapper[4988]: I1008 18:32:59.538650 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/1f36756f-acb3-439d-b10e-13573cbd252a-ovsdbserver-sb-tls-certs\") pod \"1f36756f-acb3-439d-b10e-13573cbd252a\" (UID: \"1f36756f-acb3-439d-b10e-13573cbd252a\") " Oct 08 18:32:59 crc kubenswrapper[4988]: I1008 18:32:59.538702 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/1f36756f-acb3-439d-b10e-13573cbd252a-metrics-certs-tls-certs\") pod \"1f36756f-acb3-439d-b10e-13573cbd252a\" (UID: \"1f36756f-acb3-439d-b10e-13573cbd252a\") " Oct 08 18:32:59 crc kubenswrapper[4988]: I1008 18:32:59.538836 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1f36756f-acb3-439d-b10e-13573cbd252a-combined-ca-bundle\") pod \"1f36756f-acb3-439d-b10e-13573cbd252a\" (UID: \"1f36756f-acb3-439d-b10e-13573cbd252a\") " Oct 08 18:32:59 crc kubenswrapper[4988]: I1008 18:32:59.538927 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w9pgj\" (UniqueName: \"kubernetes.io/projected/1f36756f-acb3-439d-b10e-13573cbd252a-kube-api-access-w9pgj\") pod \"1f36756f-acb3-439d-b10e-13573cbd252a\" (UID: \"1f36756f-acb3-439d-b10e-13573cbd252a\") " Oct 08 18:32:59 crc kubenswrapper[4988]: I1008 18:32:59.538972 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovndbcluster-sb-etc-ovn\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"1f36756f-acb3-439d-b10e-13573cbd252a\" (UID: \"1f36756f-acb3-439d-b10e-13573cbd252a\") " Oct 08 18:32:59 crc kubenswrapper[4988]: I1008 18:32:59.539010 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/1f36756f-acb3-439d-b10e-13573cbd252a-scripts\") pod \"1f36756f-acb3-439d-b10e-13573cbd252a\" (UID: \"1f36756f-acb3-439d-b10e-13573cbd252a\") " Oct 08 18:32:59 crc kubenswrapper[4988]: I1008 18:32:59.539032 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/1f36756f-acb3-439d-b10e-13573cbd252a-ovsdb-rundir\") pod \"1f36756f-acb3-439d-b10e-13573cbd252a\" (UID: \"1f36756f-acb3-439d-b10e-13573cbd252a\") " Oct 08 18:32:59 crc kubenswrapper[4988]: I1008 18:32:59.539078 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1f36756f-acb3-439d-b10e-13573cbd252a-config\") pod \"1f36756f-acb3-439d-b10e-13573cbd252a\" (UID: \"1f36756f-acb3-439d-b10e-13573cbd252a\") " Oct 08 18:32:59 crc kubenswrapper[4988]: I1008 18:32:59.543519 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-conductor-0"] Oct 08 18:32:59 crc kubenswrapper[4988]: I1008 18:32:59.543778 4988 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell1-conductor-0" podUID="85bc7f8f-fdb3-4abb-88d3-ed0c10982c4c" containerName="nova-cell1-conductor-conductor" containerID="cri-o://e18a49c9de8366d161727792956633e70d61d0fc8c2799f046b7a6f94af795e5" gracePeriod=30 Oct 08 18:32:59 crc kubenswrapper[4988]: I1008 18:32:59.547042 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1f36756f-acb3-439d-b10e-13573cbd252a-scripts" (OuterVolumeSpecName: "scripts") pod "1f36756f-acb3-439d-b10e-13573cbd252a" (UID: "1f36756f-acb3-439d-b10e-13573cbd252a"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 18:32:59 crc kubenswrapper[4988]: I1008 18:32:59.548627 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1f36756f-acb3-439d-b10e-13573cbd252a-config" (OuterVolumeSpecName: "config") pod "1f36756f-acb3-439d-b10e-13573cbd252a" (UID: "1f36756f-acb3-439d-b10e-13573cbd252a"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 18:32:59 crc kubenswrapper[4988]: I1008 18:32:59.549074 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1f36756f-acb3-439d-b10e-13573cbd252a-ovsdb-rundir" (OuterVolumeSpecName: "ovsdb-rundir") pod "1f36756f-acb3-439d-b10e-13573cbd252a" (UID: "1f36756f-acb3-439d-b10e-13573cbd252a"). InnerVolumeSpecName "ovsdb-rundir". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 18:32:59 crc kubenswrapper[4988]: I1008 18:32:59.552939 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1f36756f-acb3-439d-b10e-13573cbd252a-kube-api-access-w9pgj" (OuterVolumeSpecName: "kube-api-access-w9pgj") pod "1f36756f-acb3-439d-b10e-13573cbd252a" (UID: "1f36756f-acb3-439d-b10e-13573cbd252a"). InnerVolumeSpecName "kube-api-access-w9pgj". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:32:59 crc kubenswrapper[4988]: I1008 18:32:59.565987 4988 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-hzsjj"] Oct 08 18:32:59 crc kubenswrapper[4988]: I1008 18:32:59.571186 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage03-crc" (OuterVolumeSpecName: "ovndbcluster-sb-etc-ovn") pod "1f36756f-acb3-439d-b10e-13573cbd252a" (UID: "1f36756f-acb3-439d-b10e-13573cbd252a"). InnerVolumeSpecName "local-storage03-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 08 18:32:59 crc kubenswrapper[4988]: I1008 18:32:59.576732 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-0"] Oct 08 18:32:59 crc kubenswrapper[4988]: I1008 18:32:59.576909 4988 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell0-conductor-0" podUID="bd9cfcdf-4770-4efa-a41e-7867ec85206f" containerName="nova-cell0-conductor-conductor" containerID="cri-o://17f00a030a73692f1b5f14c014e24ee1e9bf7251c877957ee31dd0cb3579f43c" gracePeriod=30 Oct 08 18:32:59 crc kubenswrapper[4988]: I1008 18:32:59.584539 4988 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-cell1-server-0" podUID="e5a8f819-c91d-429e-b848-e9c444739623" containerName="rabbitmq" containerID="cri-o://bb5d4834c48dac50c4e9cebb831a5dc39438bd589b127972d2afef8419d96b2f" gracePeriod=604800 Oct 08 18:32:59 crc kubenswrapper[4988]: I1008 18:32:59.602369 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Oct 08 18:32:59 crc kubenswrapper[4988]: I1008 18:32:59.602623 4988 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="1b24865e-cac2-490e-9e61-de22a379ff47" containerName="nova-scheduler-scheduler" containerID="cri-o://6d8f1c5dc40a53cca8a87fffd5d06729664e7815b666a1eba260acf508457342" gracePeriod=30 Oct 08 18:32:59 crc kubenswrapper[4988]: I1008 18:32:59.602894 4988 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_f026ec8f-ffb4-4538-a6fc-3116ce0e7805/ovn-northd/0.log" Oct 08 18:32:59 crc kubenswrapper[4988]: I1008 18:32:59.602961 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Oct 08 18:32:59 crc kubenswrapper[4988]: I1008 18:32:59.603641 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1f36756f-acb3-439d-b10e-13573cbd252a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "1f36756f-acb3-439d-b10e-13573cbd252a" (UID: "1f36756f-acb3-439d-b10e-13573cbd252a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:32:59 crc kubenswrapper[4988]: I1008 18:32:59.609051 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 08 18:32:59 crc kubenswrapper[4988]: I1008 18:32:59.617083 4988 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-metrics-w2zhj_b92d978b-87c6-4210-9a69-99596b05cc18/openstack-network-exporter/0.log" Oct 08 18:32:59 crc kubenswrapper[4988]: I1008 18:32:59.617186 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-w2zhj" Oct 08 18:32:59 crc kubenswrapper[4988]: I1008 18:32:59.640066 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f026ec8f-ffb4-4538-a6fc-3116ce0e7805-config\") pod \"f026ec8f-ffb4-4538-a6fc-3116ce0e7805\" (UID: \"f026ec8f-ffb4-4538-a6fc-3116ce0e7805\") " Oct 08 18:32:59 crc kubenswrapper[4988]: I1008 18:32:59.640128 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f026ec8f-ffb4-4538-a6fc-3116ce0e7805-scripts\") pod \"f026ec8f-ffb4-4538-a6fc-3116ce0e7805\" (UID: \"f026ec8f-ffb4-4538-a6fc-3116ce0e7805\") " Oct 08 18:32:59 crc kubenswrapper[4988]: I1008 18:32:59.640241 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-765fn\" (UniqueName: \"kubernetes.io/projected/f026ec8f-ffb4-4538-a6fc-3116ce0e7805-kube-api-access-765fn\") pod \"f026ec8f-ffb4-4538-a6fc-3116ce0e7805\" (UID: \"f026ec8f-ffb4-4538-a6fc-3116ce0e7805\") " Oct 08 18:32:59 crc kubenswrapper[4988]: I1008 18:32:59.640318 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f026ec8f-ffb4-4538-a6fc-3116ce0e7805-combined-ca-bundle\") pod \"f026ec8f-ffb4-4538-a6fc-3116ce0e7805\" (UID: \"f026ec8f-ffb4-4538-a6fc-3116ce0e7805\") " Oct 08 18:32:59 crc kubenswrapper[4988]: I1008 18:32:59.640345 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/f026ec8f-ffb4-4538-a6fc-3116ce0e7805-metrics-certs-tls-certs\") pod \"f026ec8f-ffb4-4538-a6fc-3116ce0e7805\" (UID: \"f026ec8f-ffb4-4538-a6fc-3116ce0e7805\") " Oct 08 18:32:59 crc kubenswrapper[4988]: I1008 18:32:59.640422 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/f026ec8f-ffb4-4538-a6fc-3116ce0e7805-ovn-northd-tls-certs\") pod \"f026ec8f-ffb4-4538-a6fc-3116ce0e7805\" (UID: \"f026ec8f-ffb4-4538-a6fc-3116ce0e7805\") " Oct 08 18:32:59 crc kubenswrapper[4988]: I1008 18:32:59.640484 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/f026ec8f-ffb4-4538-a6fc-3116ce0e7805-ovn-rundir\") pod \"f026ec8f-ffb4-4538-a6fc-3116ce0e7805\" (UID: \"f026ec8f-ffb4-4538-a6fc-3116ce0e7805\") " Oct 08 18:32:59 crc kubenswrapper[4988]: I1008 18:32:59.641150 4988 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1f36756f-acb3-439d-b10e-13573cbd252a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 18:32:59 crc kubenswrapper[4988]: I1008 18:32:59.641177 4988 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w9pgj\" (UniqueName: \"kubernetes.io/projected/1f36756f-acb3-439d-b10e-13573cbd252a-kube-api-access-w9pgj\") on node \"crc\" DevicePath \"\"" Oct 08 18:32:59 crc kubenswrapper[4988]: I1008 18:32:59.641204 4988 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") on node \"crc\" " Oct 08 18:32:59 crc kubenswrapper[4988]: I1008 18:32:59.641216 4988 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/1f36756f-acb3-439d-b10e-13573cbd252a-scripts\") on node \"crc\" DevicePath \"\"" Oct 08 18:32:59 crc kubenswrapper[4988]: I1008 18:32:59.641229 4988 reconciler_common.go:293] "Volume detached for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/1f36756f-acb3-439d-b10e-13573cbd252a-ovsdb-rundir\") on node \"crc\" DevicePath \"\"" Oct 08 18:32:59 crc kubenswrapper[4988]: I1008 18:32:59.641240 4988 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1f36756f-acb3-439d-b10e-13573cbd252a-config\") on node \"crc\" DevicePath \"\"" Oct 08 18:32:59 crc kubenswrapper[4988]: I1008 18:32:59.646795 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f026ec8f-ffb4-4538-a6fc-3116ce0e7805-config" (OuterVolumeSpecName: "config") pod "f026ec8f-ffb4-4538-a6fc-3116ce0e7805" (UID: "f026ec8f-ffb4-4538-a6fc-3116ce0e7805"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 18:32:59 crc kubenswrapper[4988]: I1008 18:32:59.647457 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f026ec8f-ffb4-4538-a6fc-3116ce0e7805-ovn-rundir" (OuterVolumeSpecName: "ovn-rundir") pod "f026ec8f-ffb4-4538-a6fc-3116ce0e7805" (UID: "f026ec8f-ffb4-4538-a6fc-3116ce0e7805"). InnerVolumeSpecName "ovn-rundir". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 18:32:59 crc kubenswrapper[4988]: I1008 18:32:59.647632 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f026ec8f-ffb4-4538-a6fc-3116ce0e7805-scripts" (OuterVolumeSpecName: "scripts") pod "f026ec8f-ffb4-4538-a6fc-3116ce0e7805" (UID: "f026ec8f-ffb4-4538-a6fc-3116ce0e7805"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 18:32:59 crc kubenswrapper[4988]: I1008 18:32:59.656289 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f026ec8f-ffb4-4538-a6fc-3116ce0e7805-kube-api-access-765fn" (OuterVolumeSpecName: "kube-api-access-765fn") pod "f026ec8f-ffb4-4538-a6fc-3116ce0e7805" (UID: "f026ec8f-ffb4-4538-a6fc-3116ce0e7805"). InnerVolumeSpecName "kube-api-access-765fn". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:32:59 crc kubenswrapper[4988]: I1008 18:32:59.662948 4988 generic.go:334] "Generic (PLEG): container finished" podID="857686e3-0c6b-4d08-a8d6-f89e4c832139" containerID="42e44fdf6315337e912e8f08c104630a1408f1ece9b5a29dc6c2ddc4b1e15bf2" exitCode=0 Oct 08 18:32:59 crc kubenswrapper[4988]: I1008 18:32:59.663073 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-7b7cf6bf4f-s6vrs" event={"ID":"857686e3-0c6b-4d08-a8d6-f89e4c832139","Type":"ContainerDied","Data":"42e44fdf6315337e912e8f08c104630a1408f1ece9b5a29dc6c2ddc4b1e15bf2"} Oct 08 18:32:59 crc kubenswrapper[4988]: I1008 18:32:59.678297 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement79af-account-delete-hv6j5"] Oct 08 18:32:59 crc kubenswrapper[4988]: I1008 18:32:59.683145 4988 generic.go:334] "Generic (PLEG): container finished" podID="56702a03-175d-4a54-b84f-e25cab037431" containerID="9f5e2ee884e53117065985ec938aeba8be9ae37755acfba2392c435832c65f0e" exitCode=143 Oct 08 18:32:59 crc kubenswrapper[4988]: I1008 18:32:59.683581 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"56702a03-175d-4a54-b84f-e25cab037431","Type":"ContainerDied","Data":"9f5e2ee884e53117065985ec938aeba8be9ae37755acfba2392c435832c65f0e"} Oct 08 18:32:59 crc kubenswrapper[4988]: I1008 18:32:59.690430 4988 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage03-crc" (UniqueName: "kubernetes.io/local-volume/local-storage03-crc") on node "crc" Oct 08 18:32:59 crc kubenswrapper[4988]: I1008 18:32:59.696101 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1f36756f-acb3-439d-b10e-13573cbd252a-metrics-certs-tls-certs" (OuterVolumeSpecName: "metrics-certs-tls-certs") pod "1f36756f-acb3-439d-b10e-13573cbd252a" (UID: "1f36756f-acb3-439d-b10e-13573cbd252a"). InnerVolumeSpecName "metrics-certs-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:32:59 crc kubenswrapper[4988]: I1008 18:32:59.696359 4988 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-metrics-w2zhj_b92d978b-87c6-4210-9a69-99596b05cc18/openstack-network-exporter/0.log" Oct 08 18:32:59 crc kubenswrapper[4988]: I1008 18:32:59.696652 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-w2zhj" Oct 08 18:32:59 crc kubenswrapper[4988]: I1008 18:32:59.696846 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-w2zhj" event={"ID":"b92d978b-87c6-4210-9a69-99596b05cc18","Type":"ContainerDied","Data":"b1e0fd5f4062cfe174f7c631ad5c1f707bc5ae49d66c7afef5bcf92f995195fd"} Oct 08 18:32:59 crc kubenswrapper[4988]: I1008 18:32:59.696895 4988 scope.go:117] "RemoveContainer" containerID="80822f2e64d8fd8a2c3f08eea237f88c03ba055c04d5bb24498177ab69c61a2e" Oct 08 18:32:59 crc kubenswrapper[4988]: I1008 18:32:59.697945 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican32b1-account-delete-jlhch"] Oct 08 18:32:59 crc kubenswrapper[4988]: I1008 18:32:59.707666 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f026ec8f-ffb4-4538-a6fc-3116ce0e7805-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f026ec8f-ffb4-4538-a6fc-3116ce0e7805" (UID: "f026ec8f-ffb4-4538-a6fc-3116ce0e7805"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:32:59 crc kubenswrapper[4988]: I1008 18:32:59.716857 4988 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_b5dc1ee3-4897-45a2-ad00-0b91fbb2ae76/ovsdbserver-nb/0.log" Oct 08 18:32:59 crc kubenswrapper[4988]: I1008 18:32:59.717184 4988 generic.go:334] "Generic (PLEG): container finished" podID="b5dc1ee3-4897-45a2-ad00-0b91fbb2ae76" containerID="b5799148279b22815a3dbd06ed629a27e96e308b0f484f69614b62b235afbb9f" exitCode=2 Oct 08 18:32:59 crc kubenswrapper[4988]: I1008 18:32:59.717319 4988 generic.go:334] "Generic (PLEG): container finished" podID="b5dc1ee3-4897-45a2-ad00-0b91fbb2ae76" containerID="74829553075c7054526031a117b72832271514093fbe3617d9278cd92e26275b" exitCode=143 Oct 08 18:32:59 crc kubenswrapper[4988]: I1008 18:32:59.717942 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"b5dc1ee3-4897-45a2-ad00-0b91fbb2ae76","Type":"ContainerDied","Data":"b5799148279b22815a3dbd06ed629a27e96e308b0f484f69614b62b235afbb9f"} Oct 08 18:32:59 crc kubenswrapper[4988]: I1008 18:32:59.720695 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"b5dc1ee3-4897-45a2-ad00-0b91fbb2ae76","Type":"ContainerDied","Data":"74829553075c7054526031a117b72832271514093fbe3617d9278cd92e26275b"} Oct 08 18:32:59 crc kubenswrapper[4988]: I1008 18:32:59.727122 4988 generic.go:334] "Generic (PLEG): container finished" podID="3a57537b-6115-4249-8d16-3ab0c4b6d21e" containerID="8ef71a96a488da8760db257dae569210b2903674028f06127874742bfeb624fd" exitCode=143 Oct 08 18:32:59 crc kubenswrapper[4988]: I1008 18:32:59.727185 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-6c486d5744-j86zn" event={"ID":"3a57537b-6115-4249-8d16-3ab0c4b6d21e","Type":"ContainerDied","Data":"8ef71a96a488da8760db257dae569210b2903674028f06127874742bfeb624fd"} Oct 08 18:32:59 crc kubenswrapper[4988]: I1008 18:32:59.743971 4988 generic.go:334] "Generic (PLEG): container finished" podID="467e736b-562a-4f31-af64-2fce55e965c8" containerID="9933c83a6d7a8c9610edd4f70255f0f91bccecf1b14bb8fd883bcb6b9506666f" exitCode=0 Oct 08 18:32:59 crc kubenswrapper[4988]: I1008 18:32:59.744051 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5967cc9597-r6jrv" event={"ID":"467e736b-562a-4f31-af64-2fce55e965c8","Type":"ContainerDied","Data":"9933c83a6d7a8c9610edd4f70255f0f91bccecf1b14bb8fd883bcb6b9506666f"} Oct 08 18:32:59 crc kubenswrapper[4988]: I1008 18:32:59.745884 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b9thk\" (UniqueName: \"kubernetes.io/projected/b92d978b-87c6-4210-9a69-99596b05cc18-kube-api-access-b9thk\") pod \"b92d978b-87c6-4210-9a69-99596b05cc18\" (UID: \"b92d978b-87c6-4210-9a69-99596b05cc18\") " Oct 08 18:32:59 crc kubenswrapper[4988]: I1008 18:32:59.746111 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/b92d978b-87c6-4210-9a69-99596b05cc18-ovn-rundir\") pod \"b92d978b-87c6-4210-9a69-99596b05cc18\" (UID: \"b92d978b-87c6-4210-9a69-99596b05cc18\") " Oct 08 18:32:59 crc kubenswrapper[4988]: I1008 18:32:59.746270 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b92d978b-87c6-4210-9a69-99596b05cc18-config\") pod \"b92d978b-87c6-4210-9a69-99596b05cc18\" (UID: \"b92d978b-87c6-4210-9a69-99596b05cc18\") " Oct 08 18:32:59 crc kubenswrapper[4988]: I1008 18:32:59.746425 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/b92d978b-87c6-4210-9a69-99596b05cc18-ovs-rundir\") pod \"b92d978b-87c6-4210-9a69-99596b05cc18\" (UID: \"b92d978b-87c6-4210-9a69-99596b05cc18\") " Oct 08 18:32:59 crc kubenswrapper[4988]: I1008 18:32:59.746527 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b92d978b-87c6-4210-9a69-99596b05cc18-combined-ca-bundle\") pod \"b92d978b-87c6-4210-9a69-99596b05cc18\" (UID: \"b92d978b-87c6-4210-9a69-99596b05cc18\") " Oct 08 18:32:59 crc kubenswrapper[4988]: I1008 18:32:59.746625 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/b92d978b-87c6-4210-9a69-99596b05cc18-metrics-certs-tls-certs\") pod \"b92d978b-87c6-4210-9a69-99596b05cc18\" (UID: \"b92d978b-87c6-4210-9a69-99596b05cc18\") " Oct 08 18:32:59 crc kubenswrapper[4988]: I1008 18:32:59.747241 4988 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f026ec8f-ffb4-4538-a6fc-3116ce0e7805-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 18:32:59 crc kubenswrapper[4988]: I1008 18:32:59.747339 4988 reconciler_common.go:293] "Volume detached for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") on node \"crc\" DevicePath \"\"" Oct 08 18:32:59 crc kubenswrapper[4988]: I1008 18:32:59.747421 4988 reconciler_common.go:293] "Volume detached for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/f026ec8f-ffb4-4538-a6fc-3116ce0e7805-ovn-rundir\") on node \"crc\" DevicePath \"\"" Oct 08 18:32:59 crc kubenswrapper[4988]: I1008 18:32:59.747485 4988 reconciler_common.go:293] "Volume detached for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/1f36756f-acb3-439d-b10e-13573cbd252a-metrics-certs-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 08 18:32:59 crc kubenswrapper[4988]: I1008 18:32:59.747554 4988 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f026ec8f-ffb4-4538-a6fc-3116ce0e7805-config\") on node \"crc\" DevicePath \"\"" Oct 08 18:32:59 crc kubenswrapper[4988]: I1008 18:32:59.747640 4988 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f026ec8f-ffb4-4538-a6fc-3116ce0e7805-scripts\") on node \"crc\" DevicePath \"\"" Oct 08 18:32:59 crc kubenswrapper[4988]: I1008 18:32:59.747710 4988 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-765fn\" (UniqueName: \"kubernetes.io/projected/f026ec8f-ffb4-4538-a6fc-3116ce0e7805-kube-api-access-765fn\") on node \"crc\" DevicePath \"\"" Oct 08 18:32:59 crc kubenswrapper[4988]: I1008 18:32:59.748750 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b92d978b-87c6-4210-9a69-99596b05cc18-ovs-rundir" (OuterVolumeSpecName: "ovs-rundir") pod "b92d978b-87c6-4210-9a69-99596b05cc18" (UID: "b92d978b-87c6-4210-9a69-99596b05cc18"). InnerVolumeSpecName "ovs-rundir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 08 18:32:59 crc kubenswrapper[4988]: I1008 18:32:59.749213 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b92d978b-87c6-4210-9a69-99596b05cc18-config" (OuterVolumeSpecName: "config") pod "b92d978b-87c6-4210-9a69-99596b05cc18" (UID: "b92d978b-87c6-4210-9a69-99596b05cc18"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 18:32:59 crc kubenswrapper[4988]: I1008 18:32:59.749270 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b92d978b-87c6-4210-9a69-99596b05cc18-ovn-rundir" (OuterVolumeSpecName: "ovn-rundir") pod "b92d978b-87c6-4210-9a69-99596b05cc18" (UID: "b92d978b-87c6-4210-9a69-99596b05cc18"). InnerVolumeSpecName "ovn-rundir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 08 18:32:59 crc kubenswrapper[4988]: I1008 18:32:59.755509 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b92d978b-87c6-4210-9a69-99596b05cc18-kube-api-access-b9thk" (OuterVolumeSpecName: "kube-api-access-b9thk") pod "b92d978b-87c6-4210-9a69-99596b05cc18" (UID: "b92d978b-87c6-4210-9a69-99596b05cc18"). InnerVolumeSpecName "kube-api-access-b9thk". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:32:59 crc kubenswrapper[4988]: I1008 18:32:59.758706 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1f36756f-acb3-439d-b10e-13573cbd252a-ovsdbserver-sb-tls-certs" (OuterVolumeSpecName: "ovsdbserver-sb-tls-certs") pod "1f36756f-acb3-439d-b10e-13573cbd252a" (UID: "1f36756f-acb3-439d-b10e-13573cbd252a"). InnerVolumeSpecName "ovsdbserver-sb-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:32:59 crc kubenswrapper[4988]: I1008 18:32:59.775630 4988 generic.go:334] "Generic (PLEG): container finished" podID="6d2df307-23d3-48fd-9bb6-f09e500aaf9e" containerID="db863ec0ad9c9861590e77462403049f0e352d2f01da17f2db5e4682b149163e" exitCode=0 Oct 08 18:32:59 crc kubenswrapper[4988]: I1008 18:32:59.775723 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-6bb998f7d5-h4g2x" event={"ID":"6d2df307-23d3-48fd-9bb6-f09e500aaf9e","Type":"ContainerDied","Data":"db863ec0ad9c9861590e77462403049f0e352d2f01da17f2db5e4682b149163e"} Oct 08 18:32:59 crc kubenswrapper[4988]: I1008 18:32:59.778639 4988 generic.go:334] "Generic (PLEG): container finished" podID="83ef7610-23ac-4445-b75c-ae7206b4d810" containerID="e504bf27bfc24744eb7b0bdc9918cda7b788dc1c32411c922014b5f7f84ae75d" exitCode=143 Oct 08 18:32:59 crc kubenswrapper[4988]: I1008 18:32:59.778703 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"83ef7610-23ac-4445-b75c-ae7206b4d810","Type":"ContainerDied","Data":"e504bf27bfc24744eb7b0bdc9918cda7b788dc1c32411c922014b5f7f84ae75d"} Oct 08 18:32:59 crc kubenswrapper[4988]: I1008 18:32:59.787309 4988 generic.go:334] "Generic (PLEG): container finished" podID="c9f172a3-9188-4f5f-bf3a-fbc3b0dc5206" containerID="c7e441bdd6242790ab99582ae7f776a6bd322dc0c9ca3985c984d45c5cc43a5c" exitCode=143 Oct 08 18:32:59 crc kubenswrapper[4988]: I1008 18:32:59.787402 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-8694c889b8-lhm4g" event={"ID":"c9f172a3-9188-4f5f-bf3a-fbc3b0dc5206","Type":"ContainerDied","Data":"c7e441bdd6242790ab99582ae7f776a6bd322dc0c9ca3985c984d45c5cc43a5c"} Oct 08 18:32:59 crc kubenswrapper[4988]: I1008 18:32:59.802041 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b92d978b-87c6-4210-9a69-99596b05cc18-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "b92d978b-87c6-4210-9a69-99596b05cc18" (UID: "b92d978b-87c6-4210-9a69-99596b05cc18"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:32:59 crc kubenswrapper[4988]: I1008 18:32:59.824012 4988 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ovn-controller-ovs-kc7fw" podUID="f1628013-e0b9-4c31-80c0-91aabaaef0f6" containerName="ovs-vswitchd" containerID="cri-o://8a5332d3d9aa50bb56f12e04b1b32520399363308f9e75d5510ddd01f41593c3" gracePeriod=28 Oct 08 18:32:59 crc kubenswrapper[4988]: I1008 18:32:59.852076 4988 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b92d978b-87c6-4210-9a69-99596b05cc18-config\") on node \"crc\" DevicePath \"\"" Oct 08 18:32:59 crc kubenswrapper[4988]: I1008 18:32:59.852106 4988 reconciler_common.go:293] "Volume detached for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/b92d978b-87c6-4210-9a69-99596b05cc18-ovs-rundir\") on node \"crc\" DevicePath \"\"" Oct 08 18:32:59 crc kubenswrapper[4988]: I1008 18:32:59.852119 4988 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b92d978b-87c6-4210-9a69-99596b05cc18-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 18:32:59 crc kubenswrapper[4988]: I1008 18:32:59.852133 4988 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/1f36756f-acb3-439d-b10e-13573cbd252a-ovsdbserver-sb-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 08 18:32:59 crc kubenswrapper[4988]: I1008 18:32:59.852146 4988 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b9thk\" (UniqueName: \"kubernetes.io/projected/b92d978b-87c6-4210-9a69-99596b05cc18-kube-api-access-b9thk\") on node \"crc\" DevicePath \"\"" Oct 08 18:32:59 crc kubenswrapper[4988]: I1008 18:32:59.852157 4988 reconciler_common.go:293] "Volume detached for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/b92d978b-87c6-4210-9a69-99596b05cc18-ovn-rundir\") on node \"crc\" DevicePath \"\"" Oct 08 18:32:59 crc kubenswrapper[4988]: E1008 18:32:59.855706 4988 handlers.go:78] "Exec lifecycle hook for Container in Pod failed" err=< Oct 08 18:32:59 crc kubenswrapper[4988]: command '/usr/local/bin/container-scripts/stop-ovsdb-server.sh' exited with 137: ++ dirname /usr/local/bin/container-scripts/stop-ovsdb-server.sh Oct 08 18:32:59 crc kubenswrapper[4988]: + source /usr/local/bin/container-scripts/functions Oct 08 18:32:59 crc kubenswrapper[4988]: ++ OVNBridge=br-int Oct 08 18:32:59 crc kubenswrapper[4988]: ++ OVNRemote=tcp:localhost:6642 Oct 08 18:32:59 crc kubenswrapper[4988]: ++ OVNEncapType=geneve Oct 08 18:32:59 crc kubenswrapper[4988]: ++ OVNAvailabilityZones= Oct 08 18:32:59 crc kubenswrapper[4988]: ++ EnableChassisAsGateway=true Oct 08 18:32:59 crc kubenswrapper[4988]: ++ PhysicalNetworks= Oct 08 18:32:59 crc kubenswrapper[4988]: ++ OVNHostName= Oct 08 18:32:59 crc kubenswrapper[4988]: ++ DB_FILE=/etc/openvswitch/conf.db Oct 08 18:32:59 crc kubenswrapper[4988]: ++ ovs_dir=/var/lib/openvswitch Oct 08 18:32:59 crc kubenswrapper[4988]: ++ FLOWS_RESTORE_SCRIPT=/var/lib/openvswitch/flows-script Oct 08 18:32:59 crc kubenswrapper[4988]: ++ FLOWS_RESTORE_DIR=/var/lib/openvswitch/saved-flows Oct 08 18:32:59 crc kubenswrapper[4988]: ++ SAFE_TO_STOP_OVSDB_SERVER_SEMAPHORE=/var/lib/openvswitch/is_safe_to_stop_ovsdb_server Oct 08 18:32:59 crc kubenswrapper[4988]: + '[' '!' -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server ']' Oct 08 18:32:59 crc kubenswrapper[4988]: + sleep 0.5 Oct 08 18:32:59 crc kubenswrapper[4988]: + '[' '!' -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server ']' Oct 08 18:32:59 crc kubenswrapper[4988]: + sleep 0.5 Oct 08 18:32:59 crc kubenswrapper[4988]: + '[' '!' -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server ']' Oct 08 18:32:59 crc kubenswrapper[4988]: + sleep 0.5 Oct 08 18:32:59 crc kubenswrapper[4988]: + '[' '!' -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server ']' Oct 08 18:32:59 crc kubenswrapper[4988]: + cleanup_ovsdb_server_semaphore Oct 08 18:32:59 crc kubenswrapper[4988]: + rm -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server Oct 08 18:32:59 crc kubenswrapper[4988]: + /usr/share/openvswitch/scripts/ovs-ctl stop --no-ovs-vswitchd Oct 08 18:32:59 crc kubenswrapper[4988]: > execCommand=["/usr/local/bin/container-scripts/stop-ovsdb-server.sh"] containerName="ovsdb-server" pod="openstack/ovn-controller-ovs-kc7fw" message=< Oct 08 18:32:59 crc kubenswrapper[4988]: Exiting ovsdb-server (5) [ OK ] Oct 08 18:32:59 crc kubenswrapper[4988]: ++ dirname /usr/local/bin/container-scripts/stop-ovsdb-server.sh Oct 08 18:32:59 crc kubenswrapper[4988]: + source /usr/local/bin/container-scripts/functions Oct 08 18:32:59 crc kubenswrapper[4988]: ++ OVNBridge=br-int Oct 08 18:32:59 crc kubenswrapper[4988]: ++ OVNRemote=tcp:localhost:6642 Oct 08 18:32:59 crc kubenswrapper[4988]: ++ OVNEncapType=geneve Oct 08 18:32:59 crc kubenswrapper[4988]: ++ OVNAvailabilityZones= Oct 08 18:32:59 crc kubenswrapper[4988]: ++ EnableChassisAsGateway=true Oct 08 18:32:59 crc kubenswrapper[4988]: ++ PhysicalNetworks= Oct 08 18:32:59 crc kubenswrapper[4988]: ++ OVNHostName= Oct 08 18:32:59 crc kubenswrapper[4988]: ++ DB_FILE=/etc/openvswitch/conf.db Oct 08 18:32:59 crc kubenswrapper[4988]: ++ ovs_dir=/var/lib/openvswitch Oct 08 18:32:59 crc kubenswrapper[4988]: ++ FLOWS_RESTORE_SCRIPT=/var/lib/openvswitch/flows-script Oct 08 18:32:59 crc kubenswrapper[4988]: ++ FLOWS_RESTORE_DIR=/var/lib/openvswitch/saved-flows Oct 08 18:32:59 crc kubenswrapper[4988]: ++ SAFE_TO_STOP_OVSDB_SERVER_SEMAPHORE=/var/lib/openvswitch/is_safe_to_stop_ovsdb_server Oct 08 18:32:59 crc kubenswrapper[4988]: + '[' '!' -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server ']' Oct 08 18:32:59 crc kubenswrapper[4988]: + sleep 0.5 Oct 08 18:32:59 crc kubenswrapper[4988]: + '[' '!' -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server ']' Oct 08 18:32:59 crc kubenswrapper[4988]: + sleep 0.5 Oct 08 18:32:59 crc kubenswrapper[4988]: + '[' '!' -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server ']' Oct 08 18:32:59 crc kubenswrapper[4988]: + sleep 0.5 Oct 08 18:32:59 crc kubenswrapper[4988]: + '[' '!' -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server ']' Oct 08 18:32:59 crc kubenswrapper[4988]: + cleanup_ovsdb_server_semaphore Oct 08 18:32:59 crc kubenswrapper[4988]: + rm -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server Oct 08 18:32:59 crc kubenswrapper[4988]: + /usr/share/openvswitch/scripts/ovs-ctl stop --no-ovs-vswitchd Oct 08 18:32:59 crc kubenswrapper[4988]: > Oct 08 18:32:59 crc kubenswrapper[4988]: E1008 18:32:59.855746 4988 kuberuntime_container.go:691] "PreStop hook failed" err=< Oct 08 18:32:59 crc kubenswrapper[4988]: command '/usr/local/bin/container-scripts/stop-ovsdb-server.sh' exited with 137: ++ dirname /usr/local/bin/container-scripts/stop-ovsdb-server.sh Oct 08 18:32:59 crc kubenswrapper[4988]: + source /usr/local/bin/container-scripts/functions Oct 08 18:32:59 crc kubenswrapper[4988]: ++ OVNBridge=br-int Oct 08 18:32:59 crc kubenswrapper[4988]: ++ OVNRemote=tcp:localhost:6642 Oct 08 18:32:59 crc kubenswrapper[4988]: ++ OVNEncapType=geneve Oct 08 18:32:59 crc kubenswrapper[4988]: ++ OVNAvailabilityZones= Oct 08 18:32:59 crc kubenswrapper[4988]: ++ EnableChassisAsGateway=true Oct 08 18:32:59 crc kubenswrapper[4988]: ++ PhysicalNetworks= Oct 08 18:32:59 crc kubenswrapper[4988]: ++ OVNHostName= Oct 08 18:32:59 crc kubenswrapper[4988]: ++ DB_FILE=/etc/openvswitch/conf.db Oct 08 18:32:59 crc kubenswrapper[4988]: ++ ovs_dir=/var/lib/openvswitch Oct 08 18:32:59 crc kubenswrapper[4988]: ++ FLOWS_RESTORE_SCRIPT=/var/lib/openvswitch/flows-script Oct 08 18:32:59 crc kubenswrapper[4988]: ++ FLOWS_RESTORE_DIR=/var/lib/openvswitch/saved-flows Oct 08 18:32:59 crc kubenswrapper[4988]: ++ SAFE_TO_STOP_OVSDB_SERVER_SEMAPHORE=/var/lib/openvswitch/is_safe_to_stop_ovsdb_server Oct 08 18:32:59 crc kubenswrapper[4988]: + '[' '!' -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server ']' Oct 08 18:32:59 crc kubenswrapper[4988]: + sleep 0.5 Oct 08 18:32:59 crc kubenswrapper[4988]: + '[' '!' -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server ']' Oct 08 18:32:59 crc kubenswrapper[4988]: + sleep 0.5 Oct 08 18:32:59 crc kubenswrapper[4988]: + '[' '!' -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server ']' Oct 08 18:32:59 crc kubenswrapper[4988]: + sleep 0.5 Oct 08 18:32:59 crc kubenswrapper[4988]: + '[' '!' -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server ']' Oct 08 18:32:59 crc kubenswrapper[4988]: + cleanup_ovsdb_server_semaphore Oct 08 18:32:59 crc kubenswrapper[4988]: + rm -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server Oct 08 18:32:59 crc kubenswrapper[4988]: + /usr/share/openvswitch/scripts/ovs-ctl stop --no-ovs-vswitchd Oct 08 18:32:59 crc kubenswrapper[4988]: > pod="openstack/ovn-controller-ovs-kc7fw" podUID="f1628013-e0b9-4c31-80c0-91aabaaef0f6" containerName="ovsdb-server" containerID="cri-o://863aa8ddeda792a78ba5af321605974ac7082d24fc8fb6f0400148fcf3f06126" Oct 08 18:32:59 crc kubenswrapper[4988]: I1008 18:32:59.855780 4988 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ovn-controller-ovs-kc7fw" podUID="f1628013-e0b9-4c31-80c0-91aabaaef0f6" containerName="ovsdb-server" containerID="cri-o://863aa8ddeda792a78ba5af321605974ac7082d24fc8fb6f0400148fcf3f06126" gracePeriod=28 Oct 08 18:32:59 crc kubenswrapper[4988]: I1008 18:32:59.855934 4988 generic.go:334] "Generic (PLEG): container finished" podID="d04259e5-9595-447b-b804-feba577498d8" containerID="3e8f7969a53f0a0b492bfa6e755ebabd4ef007e6b42af4f280d07a7401c0699d" exitCode=0 Oct 08 18:32:59 crc kubenswrapper[4988]: I1008 18:32:59.855948 4988 generic.go:334] "Generic (PLEG): container finished" podID="d04259e5-9595-447b-b804-feba577498d8" containerID="f6d8e476ef92ad2a5ae0468a14ebdbcb37b6c35a714f5c83ec9ff493a5dd15a3" exitCode=0 Oct 08 18:32:59 crc kubenswrapper[4988]: I1008 18:32:59.855955 4988 generic.go:334] "Generic (PLEG): container finished" podID="d04259e5-9595-447b-b804-feba577498d8" containerID="9935bea63badb3e4cb30261f8985a1224e374fbe56ec5af0a913e1c5fe57db04" exitCode=0 Oct 08 18:32:59 crc kubenswrapper[4988]: I1008 18:32:59.855961 4988 generic.go:334] "Generic (PLEG): container finished" podID="d04259e5-9595-447b-b804-feba577498d8" containerID="0f92f3d1af5000c50ceb5a04475cb2043188883ce2d85323918a574561017295" exitCode=0 Oct 08 18:32:59 crc kubenswrapper[4988]: I1008 18:32:59.855969 4988 generic.go:334] "Generic (PLEG): container finished" podID="d04259e5-9595-447b-b804-feba577498d8" containerID="6c62379d0b1b37e153eff91e53dd273d7d1e85bb5f09dd3abdfc7a5b862d3459" exitCode=0 Oct 08 18:32:59 crc kubenswrapper[4988]: I1008 18:32:59.855975 4988 generic.go:334] "Generic (PLEG): container finished" podID="d04259e5-9595-447b-b804-feba577498d8" containerID="94be3d977f16a3f88c65d1405dd7fd0737c52d3e6c73b44f0651409fc2428746" exitCode=0 Oct 08 18:32:59 crc kubenswrapper[4988]: I1008 18:32:59.855981 4988 generic.go:334] "Generic (PLEG): container finished" podID="d04259e5-9595-447b-b804-feba577498d8" containerID="31e32c3833aea9f83bda1aca6904be1c190d6d671f11ad9f856ce5abce9053fb" exitCode=0 Oct 08 18:32:59 crc kubenswrapper[4988]: I1008 18:32:59.855987 4988 generic.go:334] "Generic (PLEG): container finished" podID="d04259e5-9595-447b-b804-feba577498d8" containerID="4f7780b6a0c60b8247000b0296848f37576f4cd48b317bfb6bd9807dc977f068" exitCode=0 Oct 08 18:32:59 crc kubenswrapper[4988]: I1008 18:32:59.855993 4988 generic.go:334] "Generic (PLEG): container finished" podID="d04259e5-9595-447b-b804-feba577498d8" containerID="c1f4256e3ed5eca7a8d3e7c8f01715b4bf6cc7778eac4b717acbddef65864b05" exitCode=0 Oct 08 18:32:59 crc kubenswrapper[4988]: I1008 18:32:59.855999 4988 generic.go:334] "Generic (PLEG): container finished" podID="d04259e5-9595-447b-b804-feba577498d8" containerID="e8f70c43dc552f28dc00baeb404e73ce329584c5d7563ccfd1bb7e213c973bc7" exitCode=0 Oct 08 18:32:59 crc kubenswrapper[4988]: I1008 18:32:59.856005 4988 generic.go:334] "Generic (PLEG): container finished" podID="d04259e5-9595-447b-b804-feba577498d8" containerID="588710f5625320484d4ef400733b3ae63b0e2f936f619df2ca3b5435aa10a8c9" exitCode=0 Oct 08 18:32:59 crc kubenswrapper[4988]: I1008 18:32:59.856012 4988 generic.go:334] "Generic (PLEG): container finished" podID="d04259e5-9595-447b-b804-feba577498d8" containerID="23ee960ca35b0ceb291c74103d14d93f4ff48fb0b2b08acdacc1fd66a0fb6327" exitCode=0 Oct 08 18:32:59 crc kubenswrapper[4988]: I1008 18:32:59.856019 4988 generic.go:334] "Generic (PLEG): container finished" podID="d04259e5-9595-447b-b804-feba577498d8" containerID="d0e7528b976d6c731e5ab211ed178a97c7d155611aadc8e3e7dde446d2824719" exitCode=0 Oct 08 18:32:59 crc kubenswrapper[4988]: I1008 18:32:59.856065 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"d04259e5-9595-447b-b804-feba577498d8","Type":"ContainerDied","Data":"3e8f7969a53f0a0b492bfa6e755ebabd4ef007e6b42af4f280d07a7401c0699d"} Oct 08 18:32:59 crc kubenswrapper[4988]: I1008 18:32:59.856090 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"d04259e5-9595-447b-b804-feba577498d8","Type":"ContainerDied","Data":"f6d8e476ef92ad2a5ae0468a14ebdbcb37b6c35a714f5c83ec9ff493a5dd15a3"} Oct 08 18:32:59 crc kubenswrapper[4988]: I1008 18:32:59.856103 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"d04259e5-9595-447b-b804-feba577498d8","Type":"ContainerDied","Data":"9935bea63badb3e4cb30261f8985a1224e374fbe56ec5af0a913e1c5fe57db04"} Oct 08 18:32:59 crc kubenswrapper[4988]: I1008 18:32:59.856111 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"d04259e5-9595-447b-b804-feba577498d8","Type":"ContainerDied","Data":"0f92f3d1af5000c50ceb5a04475cb2043188883ce2d85323918a574561017295"} Oct 08 18:32:59 crc kubenswrapper[4988]: I1008 18:32:59.856120 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"d04259e5-9595-447b-b804-feba577498d8","Type":"ContainerDied","Data":"6c62379d0b1b37e153eff91e53dd273d7d1e85bb5f09dd3abdfc7a5b862d3459"} Oct 08 18:32:59 crc kubenswrapper[4988]: I1008 18:32:59.856129 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"d04259e5-9595-447b-b804-feba577498d8","Type":"ContainerDied","Data":"94be3d977f16a3f88c65d1405dd7fd0737c52d3e6c73b44f0651409fc2428746"} Oct 08 18:32:59 crc kubenswrapper[4988]: I1008 18:32:59.856137 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"d04259e5-9595-447b-b804-feba577498d8","Type":"ContainerDied","Data":"31e32c3833aea9f83bda1aca6904be1c190d6d671f11ad9f856ce5abce9053fb"} Oct 08 18:32:59 crc kubenswrapper[4988]: I1008 18:32:59.856146 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"d04259e5-9595-447b-b804-feba577498d8","Type":"ContainerDied","Data":"4f7780b6a0c60b8247000b0296848f37576f4cd48b317bfb6bd9807dc977f068"} Oct 08 18:32:59 crc kubenswrapper[4988]: I1008 18:32:59.856154 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"d04259e5-9595-447b-b804-feba577498d8","Type":"ContainerDied","Data":"c1f4256e3ed5eca7a8d3e7c8f01715b4bf6cc7778eac4b717acbddef65864b05"} Oct 08 18:32:59 crc kubenswrapper[4988]: I1008 18:32:59.856161 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"d04259e5-9595-447b-b804-feba577498d8","Type":"ContainerDied","Data":"e8f70c43dc552f28dc00baeb404e73ce329584c5d7563ccfd1bb7e213c973bc7"} Oct 08 18:32:59 crc kubenswrapper[4988]: I1008 18:32:59.856170 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"d04259e5-9595-447b-b804-feba577498d8","Type":"ContainerDied","Data":"588710f5625320484d4ef400733b3ae63b0e2f936f619df2ca3b5435aa10a8c9"} Oct 08 18:32:59 crc kubenswrapper[4988]: I1008 18:32:59.856177 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"d04259e5-9595-447b-b804-feba577498d8","Type":"ContainerDied","Data":"23ee960ca35b0ceb291c74103d14d93f4ff48fb0b2b08acdacc1fd66a0fb6327"} Oct 08 18:32:59 crc kubenswrapper[4988]: I1008 18:32:59.856185 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"d04259e5-9595-447b-b804-feba577498d8","Type":"ContainerDied","Data":"d0e7528b976d6c731e5ab211ed178a97c7d155611aadc8e3e7dde446d2824719"} Oct 08 18:32:59 crc kubenswrapper[4988]: E1008 18:32:59.865317 4988 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="17f00a030a73692f1b5f14c014e24ee1e9bf7251c877957ee31dd0cb3579f43c" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Oct 08 18:32:59 crc kubenswrapper[4988]: I1008 18:32:59.875338 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f026ec8f-ffb4-4538-a6fc-3116ce0e7805-ovn-northd-tls-certs" (OuterVolumeSpecName: "ovn-northd-tls-certs") pod "f026ec8f-ffb4-4538-a6fc-3116ce0e7805" (UID: "f026ec8f-ffb4-4538-a6fc-3116ce0e7805"). InnerVolumeSpecName "ovn-northd-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:32:59 crc kubenswrapper[4988]: I1008 18:32:59.879147 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f026ec8f-ffb4-4538-a6fc-3116ce0e7805-metrics-certs-tls-certs" (OuterVolumeSpecName: "metrics-certs-tls-certs") pod "f026ec8f-ffb4-4538-a6fc-3116ce0e7805" (UID: "f026ec8f-ffb4-4538-a6fc-3116ce0e7805"). InnerVolumeSpecName "metrics-certs-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:32:59 crc kubenswrapper[4988]: E1008 18:32:59.892587 4988 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="17f00a030a73692f1b5f14c014e24ee1e9bf7251c877957ee31dd0cb3579f43c" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Oct 08 18:32:59 crc kubenswrapper[4988]: I1008 18:32:59.892653 4988 generic.go:334] "Generic (PLEG): container finished" podID="0efd1f53-90ad-4eb8-90f2-2ba9fc1593c2" containerID="4cdfab290ec25786ba40c3a3d5d3687398ae45094d90768455e35c6642188b15" exitCode=143 Oct 08 18:32:59 crc kubenswrapper[4988]: I1008 18:32:59.892772 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"0efd1f53-90ad-4eb8-90f2-2ba9fc1593c2","Type":"ContainerDied","Data":"4cdfab290ec25786ba40c3a3d5d3687398ae45094d90768455e35c6642188b15"} Oct 08 18:32:59 crc kubenswrapper[4988]: I1008 18:32:59.897766 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement79af-account-delete-hv6j5" event={"ID":"79cbf065-a458-4dec-b6a7-805f515142e0","Type":"ContainerStarted","Data":"b89945bf77210eeb7ef560ff52dfcfbca4649687437b5a7f6440340b4c6d51be"} Oct 08 18:32:59 crc kubenswrapper[4988]: E1008 18:32:59.904998 4988 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="17f00a030a73692f1b5f14c014e24ee1e9bf7251c877957ee31dd0cb3579f43c" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Oct 08 18:32:59 crc kubenswrapper[4988]: E1008 18:32:59.905063 4988 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-cell0-conductor-0" podUID="bd9cfcdf-4770-4efa-a41e-7867ec85206f" containerName="nova-cell0-conductor-conductor" Oct 08 18:32:59 crc kubenswrapper[4988]: I1008 18:32:59.906788 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b92d978b-87c6-4210-9a69-99596b05cc18-metrics-certs-tls-certs" (OuterVolumeSpecName: "metrics-certs-tls-certs") pod "b92d978b-87c6-4210-9a69-99596b05cc18" (UID: "b92d978b-87c6-4210-9a69-99596b05cc18"). InnerVolumeSpecName "metrics-certs-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:32:59 crc kubenswrapper[4988]: I1008 18:32:59.910319 4988 generic.go:334] "Generic (PLEG): container finished" podID="9159c915-07a7-4b41-8672-8b9d6a9572ca" containerID="2927fb800c44516806e01d37c729bc04efb6e3981dccab523101ed4405525b47" exitCode=143 Oct 08 18:32:59 crc kubenswrapper[4988]: I1008 18:32:59.910431 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-7bf697f8df-h54ft" event={"ID":"9159c915-07a7-4b41-8672-8b9d6a9572ca","Type":"ContainerDied","Data":"2927fb800c44516806e01d37c729bc04efb6e3981dccab523101ed4405525b47"} Oct 08 18:32:59 crc kubenswrapper[4988]: E1008 18:32:59.915280 4988 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="6d8f1c5dc40a53cca8a87fffd5d06729664e7815b666a1eba260acf508457342" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Oct 08 18:32:59 crc kubenswrapper[4988]: I1008 18:32:59.932370 4988 generic.go:334] "Generic (PLEG): container finished" podID="0ec345a3-1744-4e78-bf13-91f0325296e4" containerID="19c72634c92910f59b5f356f56971cc3e2867c6c02b8cdd4a997a62cb452e3cb" exitCode=0 Oct 08 18:32:59 crc kubenswrapper[4988]: I1008 18:32:59.932488 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"0ec345a3-1744-4e78-bf13-91f0325296e4","Type":"ContainerDied","Data":"19c72634c92910f59b5f356f56971cc3e2867c6c02b8cdd4a997a62cb452e3cb"} Oct 08 18:32:59 crc kubenswrapper[4988]: E1008 18:32:59.938759 4988 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="6d8f1c5dc40a53cca8a87fffd5d06729664e7815b666a1eba260acf508457342" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Oct 08 18:32:59 crc kubenswrapper[4988]: I1008 18:32:59.939574 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/novacell02f70-account-delete-cdwsx"] Oct 08 18:32:59 crc kubenswrapper[4988]: I1008 18:32:59.954482 4988 reconciler_common.go:293] "Volume detached for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/f026ec8f-ffb4-4538-a6fc-3116ce0e7805-metrics-certs-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 08 18:32:59 crc kubenswrapper[4988]: I1008 18:32:59.954526 4988 reconciler_common.go:293] "Volume detached for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/f026ec8f-ffb4-4538-a6fc-3116ce0e7805-ovn-northd-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 08 18:32:59 crc kubenswrapper[4988]: I1008 18:32:59.954541 4988 reconciler_common.go:293] "Volume detached for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/b92d978b-87c6-4210-9a69-99596b05cc18-metrics-certs-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 08 18:32:59 crc kubenswrapper[4988]: I1008 18:32:59.960536 4988 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_1f36756f-acb3-439d-b10e-13573cbd252a/ovsdbserver-sb/0.log" Oct 08 18:32:59 crc kubenswrapper[4988]: I1008 18:32:59.960614 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"1f36756f-acb3-439d-b10e-13573cbd252a","Type":"ContainerDied","Data":"b1a4cd2f1dc8c650e6916cf81b398b6e7853f8d12cb351a87c1894def087333c"} Oct 08 18:32:59 crc kubenswrapper[4988]: I1008 18:32:59.960692 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Oct 08 18:32:59 crc kubenswrapper[4988]: E1008 18:32:59.967810 4988 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="6d8f1c5dc40a53cca8a87fffd5d06729664e7815b666a1eba260acf508457342" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Oct 08 18:32:59 crc kubenswrapper[4988]: E1008 18:32:59.967869 4988 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-scheduler-0" podUID="1b24865e-cac2-490e-9e61-de22a379ff47" containerName="nova-scheduler-scheduler" Oct 08 18:32:59 crc kubenswrapper[4988]: I1008 18:32:59.976079 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Oct 08 18:32:59 crc kubenswrapper[4988]: I1008 18:32:59.976473 4988 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_f026ec8f-ffb4-4538-a6fc-3116ce0e7805/ovn-northd/0.log" Oct 08 18:32:59 crc kubenswrapper[4988]: I1008 18:32:59.976561 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Oct 08 18:32:59 crc kubenswrapper[4988]: I1008 18:32:59.976563 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"f026ec8f-ffb4-4538-a6fc-3116ce0e7805","Type":"ContainerDied","Data":"2c4f9dbb0baa15e9b34f5b39299f44218d04d27e29dfc58039842941ca081b39"} Oct 08 18:32:59 crc kubenswrapper[4988]: I1008 18:32:59.983058 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glancec9f7-account-delete-sp22m"] Oct 08 18:32:59 crc kubenswrapper[4988]: I1008 18:32:59.993800 4988 scope.go:117] "RemoveContainer" containerID="e8f79299e7fcca754183f5f8a133b9bb867b1fa107190b02e938976ad478eb0b" Oct 08 18:33:00 crc kubenswrapper[4988]: I1008 18:33:00.010737 4988 generic.go:334] "Generic (PLEG): container finished" podID="c97aab35-bbee-41a6-88b2-801b0b097a6d" containerID="47dce63c2014855045e7a231270ccde5910760705845d62023d29e100f929478" exitCode=137 Oct 08 18:33:00 crc kubenswrapper[4988]: I1008 18:33:00.010891 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Oct 08 18:33:00 crc kubenswrapper[4988]: I1008 18:33:00.027325 4988 generic.go:334] "Generic (PLEG): container finished" podID="f4e47608-89b4-45d1-88a6-fe75805d864d" containerID="13f8a712b04c6d00db8bc2f724015fda6642a1723ac5c2e59ecbd2400fa35ece" exitCode=143 Oct 08 18:33:00 crc kubenswrapper[4988]: I1008 18:33:00.027449 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-c44756996-dfw4k" event={"ID":"f4e47608-89b4-45d1-88a6-fe75805d864d","Type":"ContainerDied","Data":"13f8a712b04c6d00db8bc2f724015fda6642a1723ac5c2e59ecbd2400fa35ece"} Oct 08 18:33:00 crc kubenswrapper[4988]: I1008 18:33:00.034564 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5967cc9597-r6jrv" Oct 08 18:33:00 crc kubenswrapper[4988]: I1008 18:33:00.037233 4988 generic.go:334] "Generic (PLEG): container finished" podID="0fcc9499-a06b-4cd1-a5da-ea8108916050" containerID="7d3dbf100ba64891a97f34dd540d20ec061ba6822f245dec78c63c17c5d178fa" exitCode=143 Oct 08 18:33:00 crc kubenswrapper[4988]: I1008 18:33:00.037446 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"0fcc9499-a06b-4cd1-a5da-ea8108916050","Type":"ContainerDied","Data":"7d3dbf100ba64891a97f34dd540d20ec061ba6822f245dec78c63c17c5d178fa"} Oct 08 18:33:00 crc kubenswrapper[4988]: I1008 18:33:00.037536 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/novacell14316-account-delete-gbtkl" Oct 08 18:33:00 crc kubenswrapper[4988]: I1008 18:33:00.081301 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pm5r2\" (UniqueName: \"kubernetes.io/projected/c97aab35-bbee-41a6-88b2-801b0b097a6d-kube-api-access-pm5r2\") pod \"c97aab35-bbee-41a6-88b2-801b0b097a6d\" (UID: \"c97aab35-bbee-41a6-88b2-801b0b097a6d\") " Oct 08 18:33:00 crc kubenswrapper[4988]: I1008 18:33:00.081708 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c97aab35-bbee-41a6-88b2-801b0b097a6d-combined-ca-bundle\") pod \"c97aab35-bbee-41a6-88b2-801b0b097a6d\" (UID: \"c97aab35-bbee-41a6-88b2-801b0b097a6d\") " Oct 08 18:33:00 crc kubenswrapper[4988]: I1008 18:33:00.082032 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/c97aab35-bbee-41a6-88b2-801b0b097a6d-openstack-config\") pod \"c97aab35-bbee-41a6-88b2-801b0b097a6d\" (UID: \"c97aab35-bbee-41a6-88b2-801b0b097a6d\") " Oct 08 18:33:00 crc kubenswrapper[4988]: I1008 18:33:00.082127 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/c97aab35-bbee-41a6-88b2-801b0b097a6d-openstack-config-secret\") pod \"c97aab35-bbee-41a6-88b2-801b0b097a6d\" (UID: \"c97aab35-bbee-41a6-88b2-801b0b097a6d\") " Oct 08 18:33:00 crc kubenswrapper[4988]: I1008 18:33:00.088229 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron3849-account-delete-m479b"] Oct 08 18:33:00 crc kubenswrapper[4988]: I1008 18:33:00.099772 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c97aab35-bbee-41a6-88b2-801b0b097a6d-kube-api-access-pm5r2" (OuterVolumeSpecName: "kube-api-access-pm5r2") pod "c97aab35-bbee-41a6-88b2-801b0b097a6d" (UID: "c97aab35-bbee-41a6-88b2-801b0b097a6d"). InnerVolumeSpecName "kube-api-access-pm5r2". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:33:00 crc kubenswrapper[4988]: I1008 18:33:00.150153 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c97aab35-bbee-41a6-88b2-801b0b097a6d-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "c97aab35-bbee-41a6-88b2-801b0b097a6d" (UID: "c97aab35-bbee-41a6-88b2-801b0b097a6d"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:33:00 crc kubenswrapper[4988]: I1008 18:33:00.151908 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c97aab35-bbee-41a6-88b2-801b0b097a6d-openstack-config" (OuterVolumeSpecName: "openstack-config") pod "c97aab35-bbee-41a6-88b2-801b0b097a6d" (UID: "c97aab35-bbee-41a6-88b2-801b0b097a6d"). InnerVolumeSpecName "openstack-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 18:33:00 crc kubenswrapper[4988]: I1008 18:33:00.159735 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/novaapiebe4-account-delete-rph2r"] Oct 08 18:33:00 crc kubenswrapper[4988]: I1008 18:33:00.183957 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/467e736b-562a-4f31-af64-2fce55e965c8-config\") pod \"467e736b-562a-4f31-af64-2fce55e965c8\" (UID: \"467e736b-562a-4f31-af64-2fce55e965c8\") " Oct 08 18:33:00 crc kubenswrapper[4988]: I1008 18:33:00.184072 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/467e736b-562a-4f31-af64-2fce55e965c8-dns-swift-storage-0\") pod \"467e736b-562a-4f31-af64-2fce55e965c8\" (UID: \"467e736b-562a-4f31-af64-2fce55e965c8\") " Oct 08 18:33:00 crc kubenswrapper[4988]: I1008 18:33:00.184116 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/467e736b-562a-4f31-af64-2fce55e965c8-ovsdbserver-nb\") pod \"467e736b-562a-4f31-af64-2fce55e965c8\" (UID: \"467e736b-562a-4f31-af64-2fce55e965c8\") " Oct 08 18:33:00 crc kubenswrapper[4988]: I1008 18:33:00.184233 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m8nnt\" (UniqueName: \"kubernetes.io/projected/467e736b-562a-4f31-af64-2fce55e965c8-kube-api-access-m8nnt\") pod \"467e736b-562a-4f31-af64-2fce55e965c8\" (UID: \"467e736b-562a-4f31-af64-2fce55e965c8\") " Oct 08 18:33:00 crc kubenswrapper[4988]: I1008 18:33:00.184264 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/467e736b-562a-4f31-af64-2fce55e965c8-dns-svc\") pod \"467e736b-562a-4f31-af64-2fce55e965c8\" (UID: \"467e736b-562a-4f31-af64-2fce55e965c8\") " Oct 08 18:33:00 crc kubenswrapper[4988]: I1008 18:33:00.184292 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/467e736b-562a-4f31-af64-2fce55e965c8-ovsdbserver-sb\") pod \"467e736b-562a-4f31-af64-2fce55e965c8\" (UID: \"467e736b-562a-4f31-af64-2fce55e965c8\") " Oct 08 18:33:00 crc kubenswrapper[4988]: I1008 18:33:00.184699 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cjfgk\" (UniqueName: \"kubernetes.io/projected/1127bf6b-65ff-4c59-928c-51e16bb18ce3-kube-api-access-cjfgk\") pod \"novacell14316-account-delete-gbtkl\" (UID: \"1127bf6b-65ff-4c59-928c-51e16bb18ce3\") " pod="openstack/novacell14316-account-delete-gbtkl" Oct 08 18:33:00 crc kubenswrapper[4988]: I1008 18:33:00.184761 4988 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pm5r2\" (UniqueName: \"kubernetes.io/projected/c97aab35-bbee-41a6-88b2-801b0b097a6d-kube-api-access-pm5r2\") on node \"crc\" DevicePath \"\"" Oct 08 18:33:00 crc kubenswrapper[4988]: I1008 18:33:00.184771 4988 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c97aab35-bbee-41a6-88b2-801b0b097a6d-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 18:33:00 crc kubenswrapper[4988]: I1008 18:33:00.184781 4988 reconciler_common.go:293] "Volume detached for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/c97aab35-bbee-41a6-88b2-801b0b097a6d-openstack-config\") on node \"crc\" DevicePath \"\"" Oct 08 18:33:00 crc kubenswrapper[4988]: I1008 18:33:00.186451 4988 scope.go:117] "RemoveContainer" containerID="a41e320aa9f3ac4edb764f4ac4f3c561a463787b781e74385ca8530f54e7b28e" Oct 08 18:33:00 crc kubenswrapper[4988]: E1008 18:33:00.190404 4988 projected.go:194] Error preparing data for projected volume kube-api-access-cjfgk for pod openstack/novacell14316-account-delete-gbtkl: failed to fetch token: serviceaccounts "galera-openstack-cell1" not found Oct 08 18:33:00 crc kubenswrapper[4988]: E1008 18:33:00.190471 4988 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/1127bf6b-65ff-4c59-928c-51e16bb18ce3-kube-api-access-cjfgk podName:1127bf6b-65ff-4c59-928c-51e16bb18ce3 nodeName:}" failed. No retries permitted until 2025-10-08 18:33:02.190449462 +0000 UTC m=+1327.640292242 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-cjfgk" (UniqueName: "kubernetes.io/projected/1127bf6b-65ff-4c59-928c-51e16bb18ce3-kube-api-access-cjfgk") pod "novacell14316-account-delete-gbtkl" (UID: "1127bf6b-65ff-4c59-928c-51e16bb18ce3") : failed to fetch token: serviceaccounts "galera-openstack-cell1" not found Oct 08 18:33:00 crc kubenswrapper[4988]: I1008 18:33:00.190747 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovsdbserver-sb-0"] Oct 08 18:33:00 crc kubenswrapper[4988]: I1008 18:33:00.197374 4988 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovsdbserver-sb-0"] Oct 08 18:33:00 crc kubenswrapper[4988]: I1008 18:33:00.198440 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/467e736b-562a-4f31-af64-2fce55e965c8-kube-api-access-m8nnt" (OuterVolumeSpecName: "kube-api-access-m8nnt") pod "467e736b-562a-4f31-af64-2fce55e965c8" (UID: "467e736b-562a-4f31-af64-2fce55e965c8"). InnerVolumeSpecName "kube-api-access-m8nnt". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:33:00 crc kubenswrapper[4988]: I1008 18:33:00.198605 4988 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-server-0" podUID="830d6ec5-0015-4a88-b8e8-0fd47f64fd46" containerName="rabbitmq" containerID="cri-o://7aeabf72c789f76cad92e2b157ae1d52318b0fdd5da4ef59e8daac0afedb3f6a" gracePeriod=604800 Oct 08 18:33:00 crc kubenswrapper[4988]: I1008 18:33:00.212108 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-northd-0"] Oct 08 18:33:00 crc kubenswrapper[4988]: I1008 18:33:00.217693 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c97aab35-bbee-41a6-88b2-801b0b097a6d-openstack-config-secret" (OuterVolumeSpecName: "openstack-config-secret") pod "c97aab35-bbee-41a6-88b2-801b0b097a6d" (UID: "c97aab35-bbee-41a6-88b2-801b0b097a6d"). InnerVolumeSpecName "openstack-config-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:33:00 crc kubenswrapper[4988]: I1008 18:33:00.222235 4988 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-northd-0"] Oct 08 18:33:00 crc kubenswrapper[4988]: I1008 18:33:00.248579 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/467e736b-562a-4f31-af64-2fce55e965c8-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "467e736b-562a-4f31-af64-2fce55e965c8" (UID: "467e736b-562a-4f31-af64-2fce55e965c8"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 18:33:00 crc kubenswrapper[4988]: I1008 18:33:00.250879 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/467e736b-562a-4f31-af64-2fce55e965c8-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "467e736b-562a-4f31-af64-2fce55e965c8" (UID: "467e736b-562a-4f31-af64-2fce55e965c8"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 18:33:00 crc kubenswrapper[4988]: I1008 18:33:00.272148 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/467e736b-562a-4f31-af64-2fce55e965c8-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "467e736b-562a-4f31-af64-2fce55e965c8" (UID: "467e736b-562a-4f31-af64-2fce55e965c8"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 18:33:00 crc kubenswrapper[4988]: I1008 18:33:00.273922 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/467e736b-562a-4f31-af64-2fce55e965c8-config" (OuterVolumeSpecName: "config") pod "467e736b-562a-4f31-af64-2fce55e965c8" (UID: "467e736b-562a-4f31-af64-2fce55e965c8"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 18:33:00 crc kubenswrapper[4988]: I1008 18:33:00.286688 4988 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-m8nnt\" (UniqueName: \"kubernetes.io/projected/467e736b-562a-4f31-af64-2fce55e965c8-kube-api-access-m8nnt\") on node \"crc\" DevicePath \"\"" Oct 08 18:33:00 crc kubenswrapper[4988]: I1008 18:33:00.286726 4988 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/467e736b-562a-4f31-af64-2fce55e965c8-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 08 18:33:00 crc kubenswrapper[4988]: I1008 18:33:00.286738 4988 reconciler_common.go:293] "Volume detached for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/c97aab35-bbee-41a6-88b2-801b0b097a6d-openstack-config-secret\") on node \"crc\" DevicePath \"\"" Oct 08 18:33:00 crc kubenswrapper[4988]: I1008 18:33:00.286749 4988 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/467e736b-562a-4f31-af64-2fce55e965c8-config\") on node \"crc\" DevicePath \"\"" Oct 08 18:33:00 crc kubenswrapper[4988]: I1008 18:33:00.286760 4988 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/467e736b-562a-4f31-af64-2fce55e965c8-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Oct 08 18:33:00 crc kubenswrapper[4988]: I1008 18:33:00.286771 4988 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/467e736b-562a-4f31-af64-2fce55e965c8-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 08 18:33:00 crc kubenswrapper[4988]: I1008 18:33:00.296027 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/467e736b-562a-4f31-af64-2fce55e965c8-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "467e736b-562a-4f31-af64-2fce55e965c8" (UID: "467e736b-562a-4f31-af64-2fce55e965c8"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 18:33:00 crc kubenswrapper[4988]: I1008 18:33:00.388977 4988 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/467e736b-562a-4f31-af64-2fce55e965c8-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 08 18:33:00 crc kubenswrapper[4988]: E1008 18:33:00.388983 4988 configmap.go:193] Couldn't get configMap openstack/rabbitmq-config-data: configmap "rabbitmq-config-data" not found Oct 08 18:33:00 crc kubenswrapper[4988]: E1008 18:33:00.389087 4988 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/830d6ec5-0015-4a88-b8e8-0fd47f64fd46-config-data podName:830d6ec5-0015-4a88-b8e8-0fd47f64fd46 nodeName:}" failed. No retries permitted until 2025-10-08 18:33:02.389067914 +0000 UTC m=+1327.838910694 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/configmap/830d6ec5-0015-4a88-b8e8-0fd47f64fd46-config-data") pod "rabbitmq-server-0" (UID: "830d6ec5-0015-4a88-b8e8-0fd47f64fd46") : configmap "rabbitmq-config-data" not found Oct 08 18:33:00 crc kubenswrapper[4988]: I1008 18:33:00.584425 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinderd8da-account-delete-75j2h"] Oct 08 18:33:00 crc kubenswrapper[4988]: I1008 18:33:00.621648 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-metrics-w2zhj"] Oct 08 18:33:00 crc kubenswrapper[4988]: I1008 18:33:00.645362 4988 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-metrics-w2zhj"] Oct 08 18:33:00 crc kubenswrapper[4988]: I1008 18:33:00.645742 4988 scope.go:117] "RemoveContainer" containerID="ada91b938d451d0675a819f062dafde62f4a75b785dc86a5448ce25e246cecbb" Oct 08 18:33:00 crc kubenswrapper[4988]: W1008 18:33:00.665986 4988 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod02a37054_4817_4151_a8cb_a9a45867c0d3.slice/crio-bb9a6ad55acd84c2a6010e95ec046d90fe6e67eadd6cc37cb135bacfd262aa56 WatchSource:0}: Error finding container bb9a6ad55acd84c2a6010e95ec046d90fe6e67eadd6cc37cb135bacfd262aa56: Status 404 returned error can't find the container with id bb9a6ad55acd84c2a6010e95ec046d90fe6e67eadd6cc37cb135bacfd262aa56 Oct 08 18:33:00 crc kubenswrapper[4988]: I1008 18:33:00.667675 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/novacell14316-account-delete-gbtkl" Oct 08 18:33:00 crc kubenswrapper[4988]: I1008 18:33:00.687102 4988 scope.go:117] "RemoveContainer" containerID="e9efac4802d1c9a2377dc285da1c498c1363ff5f70695e54b4a257848bdea958" Oct 08 18:33:00 crc kubenswrapper[4988]: I1008 18:33:00.709147 4988 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_b5dc1ee3-4897-45a2-ad00-0b91fbb2ae76/ovsdbserver-nb/0.log" Oct 08 18:33:00 crc kubenswrapper[4988]: I1008 18:33:00.709217 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Oct 08 18:33:00 crc kubenswrapper[4988]: I1008 18:33:00.750968 4988 scope.go:117] "RemoveContainer" containerID="47dce63c2014855045e7a231270ccde5910760705845d62023d29e100f929478" Oct 08 18:33:00 crc kubenswrapper[4988]: I1008 18:33:00.753840 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-7b7cf6bf4f-s6vrs" Oct 08 18:33:00 crc kubenswrapper[4988]: I1008 18:33:00.769440 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Oct 08 18:33:00 crc kubenswrapper[4988]: I1008 18:33:00.797887 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tvmgc\" (UniqueName: \"kubernetes.io/projected/b5dc1ee3-4897-45a2-ad00-0b91fbb2ae76-kube-api-access-tvmgc\") pod \"b5dc1ee3-4897-45a2-ad00-0b91fbb2ae76\" (UID: \"b5dc1ee3-4897-45a2-ad00-0b91fbb2ae76\") " Oct 08 18:33:00 crc kubenswrapper[4988]: I1008 18:33:00.797960 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovndbcluster-nb-etc-ovn\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"b5dc1ee3-4897-45a2-ad00-0b91fbb2ae76\" (UID: \"b5dc1ee3-4897-45a2-ad00-0b91fbb2ae76\") " Oct 08 18:33:00 crc kubenswrapper[4988]: I1008 18:33:00.797997 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b5dc1ee3-4897-45a2-ad00-0b91fbb2ae76-combined-ca-bundle\") pod \"b5dc1ee3-4897-45a2-ad00-0b91fbb2ae76\" (UID: \"b5dc1ee3-4897-45a2-ad00-0b91fbb2ae76\") " Oct 08 18:33:00 crc kubenswrapper[4988]: I1008 18:33:00.798042 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/b5dc1ee3-4897-45a2-ad00-0b91fbb2ae76-scripts\") pod \"b5dc1ee3-4897-45a2-ad00-0b91fbb2ae76\" (UID: \"b5dc1ee3-4897-45a2-ad00-0b91fbb2ae76\") " Oct 08 18:33:00 crc kubenswrapper[4988]: I1008 18:33:00.798091 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b5dc1ee3-4897-45a2-ad00-0b91fbb2ae76-config\") pod \"b5dc1ee3-4897-45a2-ad00-0b91fbb2ae76\" (UID: \"b5dc1ee3-4897-45a2-ad00-0b91fbb2ae76\") " Oct 08 18:33:00 crc kubenswrapper[4988]: I1008 18:33:00.798111 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/b5dc1ee3-4897-45a2-ad00-0b91fbb2ae76-ovsdbserver-nb-tls-certs\") pod \"b5dc1ee3-4897-45a2-ad00-0b91fbb2ae76\" (UID: \"b5dc1ee3-4897-45a2-ad00-0b91fbb2ae76\") " Oct 08 18:33:00 crc kubenswrapper[4988]: I1008 18:33:00.798178 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/b5dc1ee3-4897-45a2-ad00-0b91fbb2ae76-metrics-certs-tls-certs\") pod \"b5dc1ee3-4897-45a2-ad00-0b91fbb2ae76\" (UID: \"b5dc1ee3-4897-45a2-ad00-0b91fbb2ae76\") " Oct 08 18:33:00 crc kubenswrapper[4988]: I1008 18:33:00.798247 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/b5dc1ee3-4897-45a2-ad00-0b91fbb2ae76-ovsdb-rundir\") pod \"b5dc1ee3-4897-45a2-ad00-0b91fbb2ae76\" (UID: \"b5dc1ee3-4897-45a2-ad00-0b91fbb2ae76\") " Oct 08 18:33:00 crc kubenswrapper[4988]: I1008 18:33:00.802498 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b5dc1ee3-4897-45a2-ad00-0b91fbb2ae76-ovsdb-rundir" (OuterVolumeSpecName: "ovsdb-rundir") pod "b5dc1ee3-4897-45a2-ad00-0b91fbb2ae76" (UID: "b5dc1ee3-4897-45a2-ad00-0b91fbb2ae76"). InnerVolumeSpecName "ovsdb-rundir". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 18:33:00 crc kubenswrapper[4988]: I1008 18:33:00.804062 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b5dc1ee3-4897-45a2-ad00-0b91fbb2ae76-config" (OuterVolumeSpecName: "config") pod "b5dc1ee3-4897-45a2-ad00-0b91fbb2ae76" (UID: "b5dc1ee3-4897-45a2-ad00-0b91fbb2ae76"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 18:33:00 crc kubenswrapper[4988]: I1008 18:33:00.804068 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b5dc1ee3-4897-45a2-ad00-0b91fbb2ae76-scripts" (OuterVolumeSpecName: "scripts") pod "b5dc1ee3-4897-45a2-ad00-0b91fbb2ae76" (UID: "b5dc1ee3-4897-45a2-ad00-0b91fbb2ae76"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 18:33:00 crc kubenswrapper[4988]: I1008 18:33:00.822750 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage09-crc" (OuterVolumeSpecName: "ovndbcluster-nb-etc-ovn") pod "b5dc1ee3-4897-45a2-ad00-0b91fbb2ae76" (UID: "b5dc1ee3-4897-45a2-ad00-0b91fbb2ae76"). InnerVolumeSpecName "local-storage09-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 08 18:33:00 crc kubenswrapper[4988]: I1008 18:33:00.823443 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b5dc1ee3-4897-45a2-ad00-0b91fbb2ae76-kube-api-access-tvmgc" (OuterVolumeSpecName: "kube-api-access-tvmgc") pod "b5dc1ee3-4897-45a2-ad00-0b91fbb2ae76" (UID: "b5dc1ee3-4897-45a2-ad00-0b91fbb2ae76"). InnerVolumeSpecName "kube-api-access-tvmgc". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:33:00 crc kubenswrapper[4988]: I1008 18:33:00.870132 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-7bf697f8df-h54ft" Oct 08 18:33:00 crc kubenswrapper[4988]: I1008 18:33:00.903328 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/857686e3-0c6b-4d08-a8d6-f89e4c832139-run-httpd\") pod \"857686e3-0c6b-4d08-a8d6-f89e4c832139\" (UID: \"857686e3-0c6b-4d08-a8d6-f89e4c832139\") " Oct 08 18:33:00 crc kubenswrapper[4988]: I1008 18:33:00.903416 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9d54x\" (UniqueName: \"kubernetes.io/projected/816b358d-4428-4fae-b91c-03c0171faa23-kube-api-access-9d54x\") pod \"816b358d-4428-4fae-b91c-03c0171faa23\" (UID: \"816b358d-4428-4fae-b91c-03c0171faa23\") " Oct 08 18:33:00 crc kubenswrapper[4988]: I1008 18:33:00.903491 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/816b358d-4428-4fae-b91c-03c0171faa23-combined-ca-bundle\") pod \"816b358d-4428-4fae-b91c-03c0171faa23\" (UID: \"816b358d-4428-4fae-b91c-03c0171faa23\") " Oct 08 18:33:00 crc kubenswrapper[4988]: I1008 18:33:00.903583 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/816b358d-4428-4fae-b91c-03c0171faa23-config-data\") pod \"816b358d-4428-4fae-b91c-03c0171faa23\" (UID: \"816b358d-4428-4fae-b91c-03c0171faa23\") " Oct 08 18:33:00 crc kubenswrapper[4988]: I1008 18:33:00.903716 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/857686e3-0c6b-4d08-a8d6-f89e4c832139-internal-tls-certs\") pod \"857686e3-0c6b-4d08-a8d6-f89e4c832139\" (UID: \"857686e3-0c6b-4d08-a8d6-f89e4c832139\") " Oct 08 18:33:00 crc kubenswrapper[4988]: I1008 18:33:00.903803 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/857686e3-0c6b-4d08-a8d6-f89e4c832139-public-tls-certs\") pod \"857686e3-0c6b-4d08-a8d6-f89e4c832139\" (UID: \"857686e3-0c6b-4d08-a8d6-f89e4c832139\") " Oct 08 18:33:00 crc kubenswrapper[4988]: I1008 18:33:00.903883 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/816b358d-4428-4fae-b91c-03c0171faa23-nova-novncproxy-tls-certs\") pod \"816b358d-4428-4fae-b91c-03c0171faa23\" (UID: \"816b358d-4428-4fae-b91c-03c0171faa23\") " Oct 08 18:33:00 crc kubenswrapper[4988]: I1008 18:33:00.903918 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-k5jwg\" (UniqueName: \"kubernetes.io/projected/857686e3-0c6b-4d08-a8d6-f89e4c832139-kube-api-access-k5jwg\") pod \"857686e3-0c6b-4d08-a8d6-f89e4c832139\" (UID: \"857686e3-0c6b-4d08-a8d6-f89e4c832139\") " Oct 08 18:33:00 crc kubenswrapper[4988]: I1008 18:33:00.904212 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/857686e3-0c6b-4d08-a8d6-f89e4c832139-config-data\") pod \"857686e3-0c6b-4d08-a8d6-f89e4c832139\" (UID: \"857686e3-0c6b-4d08-a8d6-f89e4c832139\") " Oct 08 18:33:00 crc kubenswrapper[4988]: I1008 18:33:00.904259 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/857686e3-0c6b-4d08-a8d6-f89e4c832139-log-httpd\") pod \"857686e3-0c6b-4d08-a8d6-f89e4c832139\" (UID: \"857686e3-0c6b-4d08-a8d6-f89e4c832139\") " Oct 08 18:33:00 crc kubenswrapper[4988]: I1008 18:33:00.904326 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/857686e3-0c6b-4d08-a8d6-f89e4c832139-combined-ca-bundle\") pod \"857686e3-0c6b-4d08-a8d6-f89e4c832139\" (UID: \"857686e3-0c6b-4d08-a8d6-f89e4c832139\") " Oct 08 18:33:00 crc kubenswrapper[4988]: I1008 18:33:00.904419 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/857686e3-0c6b-4d08-a8d6-f89e4c832139-etc-swift\") pod \"857686e3-0c6b-4d08-a8d6-f89e4c832139\" (UID: \"857686e3-0c6b-4d08-a8d6-f89e4c832139\") " Oct 08 18:33:00 crc kubenswrapper[4988]: I1008 18:33:00.904441 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/816b358d-4428-4fae-b91c-03c0171faa23-vencrypt-tls-certs\") pod \"816b358d-4428-4fae-b91c-03c0171faa23\" (UID: \"816b358d-4428-4fae-b91c-03c0171faa23\") " Oct 08 18:33:00 crc kubenswrapper[4988]: I1008 18:33:00.904892 4988 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b5dc1ee3-4897-45a2-ad00-0b91fbb2ae76-config\") on node \"crc\" DevicePath \"\"" Oct 08 18:33:00 crc kubenswrapper[4988]: I1008 18:33:00.904909 4988 reconciler_common.go:293] "Volume detached for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/b5dc1ee3-4897-45a2-ad00-0b91fbb2ae76-ovsdb-rundir\") on node \"crc\" DevicePath \"\"" Oct 08 18:33:00 crc kubenswrapper[4988]: I1008 18:33:00.904919 4988 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tvmgc\" (UniqueName: \"kubernetes.io/projected/b5dc1ee3-4897-45a2-ad00-0b91fbb2ae76-kube-api-access-tvmgc\") on node \"crc\" DevicePath \"\"" Oct 08 18:33:00 crc kubenswrapper[4988]: I1008 18:33:00.904937 4988 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") on node \"crc\" " Oct 08 18:33:00 crc kubenswrapper[4988]: I1008 18:33:00.904946 4988 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/b5dc1ee3-4897-45a2-ad00-0b91fbb2ae76-scripts\") on node \"crc\" DevicePath \"\"" Oct 08 18:33:00 crc kubenswrapper[4988]: I1008 18:33:00.905498 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/857686e3-0c6b-4d08-a8d6-f89e4c832139-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "857686e3-0c6b-4d08-a8d6-f89e4c832139" (UID: "857686e3-0c6b-4d08-a8d6-f89e4c832139"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 18:33:00 crc kubenswrapper[4988]: I1008 18:33:00.905703 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-6bb998f7d5-h4g2x" Oct 08 18:33:00 crc kubenswrapper[4988]: I1008 18:33:00.909039 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/857686e3-0c6b-4d08-a8d6-f89e4c832139-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "857686e3-0c6b-4d08-a8d6-f89e4c832139" (UID: "857686e3-0c6b-4d08-a8d6-f89e4c832139"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 18:33:00 crc kubenswrapper[4988]: I1008 18:33:00.913424 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/857686e3-0c6b-4d08-a8d6-f89e4c832139-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "857686e3-0c6b-4d08-a8d6-f89e4c832139" (UID: "857686e3-0c6b-4d08-a8d6-f89e4c832139"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:33:00 crc kubenswrapper[4988]: I1008 18:33:00.924150 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/816b358d-4428-4fae-b91c-03c0171faa23-kube-api-access-9d54x" (OuterVolumeSpecName: "kube-api-access-9d54x") pod "816b358d-4428-4fae-b91c-03c0171faa23" (UID: "816b358d-4428-4fae-b91c-03c0171faa23"). InnerVolumeSpecName "kube-api-access-9d54x". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:33:00 crc kubenswrapper[4988]: I1008 18:33:00.942525 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/857686e3-0c6b-4d08-a8d6-f89e4c832139-kube-api-access-k5jwg" (OuterVolumeSpecName: "kube-api-access-k5jwg") pod "857686e3-0c6b-4d08-a8d6-f89e4c832139" (UID: "857686e3-0c6b-4d08-a8d6-f89e4c832139"). InnerVolumeSpecName "kube-api-access-k5jwg". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:33:00 crc kubenswrapper[4988]: I1008 18:33:00.999495 4988 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage09-crc" (UniqueName: "kubernetes.io/local-volume/local-storage09-crc") on node "crc" Oct 08 18:33:01 crc kubenswrapper[4988]: I1008 18:33:01.008421 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-54jth\" (UniqueName: \"kubernetes.io/projected/9159c915-07a7-4b41-8672-8b9d6a9572ca-kube-api-access-54jth\") pod \"9159c915-07a7-4b41-8672-8b9d6a9572ca\" (UID: \"9159c915-07a7-4b41-8672-8b9d6a9572ca\") " Oct 08 18:33:01 crc kubenswrapper[4988]: I1008 18:33:01.008463 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8gplz\" (UniqueName: \"kubernetes.io/projected/6d2df307-23d3-48fd-9bb6-f09e500aaf9e-kube-api-access-8gplz\") pod \"6d2df307-23d3-48fd-9bb6-f09e500aaf9e\" (UID: \"6d2df307-23d3-48fd-9bb6-f09e500aaf9e\") " Oct 08 18:33:01 crc kubenswrapper[4988]: I1008 18:33:01.008522 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/6d2df307-23d3-48fd-9bb6-f09e500aaf9e-internal-tls-certs\") pod \"6d2df307-23d3-48fd-9bb6-f09e500aaf9e\" (UID: \"6d2df307-23d3-48fd-9bb6-f09e500aaf9e\") " Oct 08 18:33:01 crc kubenswrapper[4988]: I1008 18:33:01.008617 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/6d2df307-23d3-48fd-9bb6-f09e500aaf9e-public-tls-certs\") pod \"6d2df307-23d3-48fd-9bb6-f09e500aaf9e\" (UID: \"6d2df307-23d3-48fd-9bb6-f09e500aaf9e\") " Oct 08 18:33:01 crc kubenswrapper[4988]: I1008 18:33:01.008634 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/6d2df307-23d3-48fd-9bb6-f09e500aaf9e-config\") pod \"6d2df307-23d3-48fd-9bb6-f09e500aaf9e\" (UID: \"6d2df307-23d3-48fd-9bb6-f09e500aaf9e\") " Oct 08 18:33:01 crc kubenswrapper[4988]: I1008 18:33:01.008677 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6d2df307-23d3-48fd-9bb6-f09e500aaf9e-combined-ca-bundle\") pod \"6d2df307-23d3-48fd-9bb6-f09e500aaf9e\" (UID: \"6d2df307-23d3-48fd-9bb6-f09e500aaf9e\") " Oct 08 18:33:01 crc kubenswrapper[4988]: I1008 18:33:01.008715 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/9159c915-07a7-4b41-8672-8b9d6a9572ca-config-data-custom\") pod \"9159c915-07a7-4b41-8672-8b9d6a9572ca\" (UID: \"9159c915-07a7-4b41-8672-8b9d6a9572ca\") " Oct 08 18:33:01 crc kubenswrapper[4988]: I1008 18:33:01.008737 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9159c915-07a7-4b41-8672-8b9d6a9572ca-combined-ca-bundle\") pod \"9159c915-07a7-4b41-8672-8b9d6a9572ca\" (UID: \"9159c915-07a7-4b41-8672-8b9d6a9572ca\") " Oct 08 18:33:01 crc kubenswrapper[4988]: I1008 18:33:01.008851 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9159c915-07a7-4b41-8672-8b9d6a9572ca-logs\") pod \"9159c915-07a7-4b41-8672-8b9d6a9572ca\" (UID: \"9159c915-07a7-4b41-8672-8b9d6a9572ca\") " Oct 08 18:33:01 crc kubenswrapper[4988]: I1008 18:33:01.008877 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9159c915-07a7-4b41-8672-8b9d6a9572ca-config-data\") pod \"9159c915-07a7-4b41-8672-8b9d6a9572ca\" (UID: \"9159c915-07a7-4b41-8672-8b9d6a9572ca\") " Oct 08 18:33:01 crc kubenswrapper[4988]: I1008 18:33:01.008919 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/6d2df307-23d3-48fd-9bb6-f09e500aaf9e-httpd-config\") pod \"6d2df307-23d3-48fd-9bb6-f09e500aaf9e\" (UID: \"6d2df307-23d3-48fd-9bb6-f09e500aaf9e\") " Oct 08 18:33:01 crc kubenswrapper[4988]: I1008 18:33:01.008954 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/6d2df307-23d3-48fd-9bb6-f09e500aaf9e-ovndb-tls-certs\") pod \"6d2df307-23d3-48fd-9bb6-f09e500aaf9e\" (UID: \"6d2df307-23d3-48fd-9bb6-f09e500aaf9e\") " Oct 08 18:33:01 crc kubenswrapper[4988]: I1008 18:33:01.009302 4988 reconciler_common.go:293] "Volume detached for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") on node \"crc\" DevicePath \"\"" Oct 08 18:33:01 crc kubenswrapper[4988]: I1008 18:33:01.009320 4988 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-k5jwg\" (UniqueName: \"kubernetes.io/projected/857686e3-0c6b-4d08-a8d6-f89e4c832139-kube-api-access-k5jwg\") on node \"crc\" DevicePath \"\"" Oct 08 18:33:01 crc kubenswrapper[4988]: I1008 18:33:01.009330 4988 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/857686e3-0c6b-4d08-a8d6-f89e4c832139-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 08 18:33:01 crc kubenswrapper[4988]: I1008 18:33:01.009339 4988 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/857686e3-0c6b-4d08-a8d6-f89e4c832139-etc-swift\") on node \"crc\" DevicePath \"\"" Oct 08 18:33:01 crc kubenswrapper[4988]: I1008 18:33:01.009348 4988 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/857686e3-0c6b-4d08-a8d6-f89e4c832139-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 08 18:33:01 crc kubenswrapper[4988]: I1008 18:33:01.009367 4988 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9d54x\" (UniqueName: \"kubernetes.io/projected/816b358d-4428-4fae-b91c-03c0171faa23-kube-api-access-9d54x\") on node \"crc\" DevicePath \"\"" Oct 08 18:33:01 crc kubenswrapper[4988]: E1008 18:33:01.009458 4988 configmap.go:193] Couldn't get configMap openstack/rabbitmq-cell1-config-data: configmap "rabbitmq-cell1-config-data" not found Oct 08 18:33:01 crc kubenswrapper[4988]: E1008 18:33:01.009515 4988 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/e5a8f819-c91d-429e-b848-e9c444739623-config-data podName:e5a8f819-c91d-429e-b848-e9c444739623 nodeName:}" failed. No retries permitted until 2025-10-08 18:33:05.009499718 +0000 UTC m=+1330.459342488 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/configmap/e5a8f819-c91d-429e-b848-e9c444739623-config-data") pod "rabbitmq-cell1-server-0" (UID: "e5a8f819-c91d-429e-b848-e9c444739623") : configmap "rabbitmq-cell1-config-data" not found Oct 08 18:33:01 crc kubenswrapper[4988]: I1008 18:33:01.010724 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9159c915-07a7-4b41-8672-8b9d6a9572ca-logs" (OuterVolumeSpecName: "logs") pod "9159c915-07a7-4b41-8672-8b9d6a9572ca" (UID: "9159c915-07a7-4b41-8672-8b9d6a9572ca"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 18:33:01 crc kubenswrapper[4988]: I1008 18:33:01.038592 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6d2df307-23d3-48fd-9bb6-f09e500aaf9e-httpd-config" (OuterVolumeSpecName: "httpd-config") pod "6d2df307-23d3-48fd-9bb6-f09e500aaf9e" (UID: "6d2df307-23d3-48fd-9bb6-f09e500aaf9e"). InnerVolumeSpecName "httpd-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:33:01 crc kubenswrapper[4988]: I1008 18:33:01.048561 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9159c915-07a7-4b41-8672-8b9d6a9572ca-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "9159c915-07a7-4b41-8672-8b9d6a9572ca" (UID: "9159c915-07a7-4b41-8672-8b9d6a9572ca"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:33:01 crc kubenswrapper[4988]: I1008 18:33:01.048729 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6d2df307-23d3-48fd-9bb6-f09e500aaf9e-kube-api-access-8gplz" (OuterVolumeSpecName: "kube-api-access-8gplz") pod "6d2df307-23d3-48fd-9bb6-f09e500aaf9e" (UID: "6d2df307-23d3-48fd-9bb6-f09e500aaf9e"). InnerVolumeSpecName "kube-api-access-8gplz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:33:01 crc kubenswrapper[4988]: I1008 18:33:01.055814 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9159c915-07a7-4b41-8672-8b9d6a9572ca-kube-api-access-54jth" (OuterVolumeSpecName: "kube-api-access-54jth") pod "9159c915-07a7-4b41-8672-8b9d6a9572ca" (UID: "9159c915-07a7-4b41-8672-8b9d6a9572ca"). InnerVolumeSpecName "kube-api-access-54jth". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:33:01 crc kubenswrapper[4988]: I1008 18:33:01.061120 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinderd8da-account-delete-75j2h" event={"ID":"02a37054-4817-4151-a8cb-a9a45867c0d3","Type":"ContainerStarted","Data":"bb9a6ad55acd84c2a6010e95ec046d90fe6e67eadd6cc37cb135bacfd262aa56"} Oct 08 18:33:01 crc kubenswrapper[4988]: I1008 18:33:01.063412 4988 generic.go:334] "Generic (PLEG): container finished" podID="816b358d-4428-4fae-b91c-03c0171faa23" containerID="a947f771405cb73547e24443c664920da93801df8ff9dd6369c101deea715564" exitCode=0 Oct 08 18:33:01 crc kubenswrapper[4988]: I1008 18:33:01.063449 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Oct 08 18:33:01 crc kubenswrapper[4988]: I1008 18:33:01.063483 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"816b358d-4428-4fae-b91c-03c0171faa23","Type":"ContainerDied","Data":"a947f771405cb73547e24443c664920da93801df8ff9dd6369c101deea715564"} Oct 08 18:33:01 crc kubenswrapper[4988]: I1008 18:33:01.063513 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"816b358d-4428-4fae-b91c-03c0171faa23","Type":"ContainerDied","Data":"8a452108aab53fec7dc4838c8537c45f71851137a3ecea9fd46ecab51a3db81f"} Oct 08 18:33:01 crc kubenswrapper[4988]: I1008 18:33:01.063529 4988 scope.go:117] "RemoveContainer" containerID="a947f771405cb73547e24443c664920da93801df8ff9dd6369c101deea715564" Oct 08 18:33:01 crc kubenswrapper[4988]: I1008 18:33:01.070999 4988 generic.go:334] "Generic (PLEG): container finished" podID="3e8c874a-029a-4d6a-bc2e-00aeed1636a7" containerID="fb6cef822415bff20531090cd5e76b8c5dae15da1099c2400c9c9b26e238da3c" exitCode=0 Oct 08 18:33:01 crc kubenswrapper[4988]: I1008 18:33:01.071066 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/novacell02f70-account-delete-cdwsx" event={"ID":"3e8c874a-029a-4d6a-bc2e-00aeed1636a7","Type":"ContainerDied","Data":"fb6cef822415bff20531090cd5e76b8c5dae15da1099c2400c9c9b26e238da3c"} Oct 08 18:33:01 crc kubenswrapper[4988]: I1008 18:33:01.071093 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/novacell02f70-account-delete-cdwsx" event={"ID":"3e8c874a-029a-4d6a-bc2e-00aeed1636a7","Type":"ContainerStarted","Data":"cf5297609526e130b9edf694145c045de83390235b5b462b974086d31336cf2d"} Oct 08 18:33:01 crc kubenswrapper[4988]: I1008 18:33:01.074274 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/novaapiebe4-account-delete-rph2r" event={"ID":"28f73b7a-bb5e-43ec-a099-13a12cdb33d3","Type":"ContainerStarted","Data":"1b85ba9999219bc7a7a3b6a0ce7ce2da865b368c7d4d19a8fb8c6910e624f611"} Oct 08 18:33:01 crc kubenswrapper[4988]: I1008 18:33:01.088460 4988 generic.go:334] "Generic (PLEG): container finished" podID="3f7bddd3-86f8-46e9-a38e-1aa5e570d151" containerID="df5e696bea34bc9d704df11826fdc54a7f54a0eaa51859b450ac2001e2f15dda" exitCode=0 Oct 08 18:33:01 crc kubenswrapper[4988]: I1008 18:33:01.088673 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican32b1-account-delete-jlhch" event={"ID":"3f7bddd3-86f8-46e9-a38e-1aa5e570d151","Type":"ContainerDied","Data":"df5e696bea34bc9d704df11826fdc54a7f54a0eaa51859b450ac2001e2f15dda"} Oct 08 18:33:01 crc kubenswrapper[4988]: I1008 18:33:01.088711 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican32b1-account-delete-jlhch" event={"ID":"3f7bddd3-86f8-46e9-a38e-1aa5e570d151","Type":"ContainerStarted","Data":"9d5c29e3c8578e502a01d1e1800e01d52a4abda618cf7cabb3fdf8627cb9f172"} Oct 08 18:33:01 crc kubenswrapper[4988]: I1008 18:33:01.098085 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glancec9f7-account-delete-sp22m" event={"ID":"d7d648e7-3c02-4d26-993d-a901764b3329","Type":"ContainerStarted","Data":"31548197acc0e143742e4fb4cb9299994c2c1dd05e374425c882dac956b9d649"} Oct 08 18:33:01 crc kubenswrapper[4988]: I1008 18:33:01.098138 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glancec9f7-account-delete-sp22m" event={"ID":"d7d648e7-3c02-4d26-993d-a901764b3329","Type":"ContainerStarted","Data":"464217ebf5ac45bfe14b80e19139ccfccd2d00bd4f5783a2e886fa030ab82158"} Oct 08 18:33:01 crc kubenswrapper[4988]: I1008 18:33:01.103338 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron3849-account-delete-m479b" event={"ID":"ab3315d2-3248-42f1-9834-2bab55fa9935","Type":"ContainerStarted","Data":"ee5151c0c831844f2ab3796464c7e52d85522046bd44d8ed010f49dda7a02db4"} Oct 08 18:33:01 crc kubenswrapper[4988]: I1008 18:33:01.103407 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron3849-account-delete-m479b" event={"ID":"ab3315d2-3248-42f1-9834-2bab55fa9935","Type":"ContainerStarted","Data":"d3f029cb7d883d25bc8d0897c54c850a721d92b478323fff209fa6dea5f80250"} Oct 08 18:33:01 crc kubenswrapper[4988]: I1008 18:33:01.111096 4988 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9159c915-07a7-4b41-8672-8b9d6a9572ca-logs\") on node \"crc\" DevicePath \"\"" Oct 08 18:33:01 crc kubenswrapper[4988]: I1008 18:33:01.111130 4988 reconciler_common.go:293] "Volume detached for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/6d2df307-23d3-48fd-9bb6-f09e500aaf9e-httpd-config\") on node \"crc\" DevicePath \"\"" Oct 08 18:33:01 crc kubenswrapper[4988]: I1008 18:33:01.111139 4988 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-54jth\" (UniqueName: \"kubernetes.io/projected/9159c915-07a7-4b41-8672-8b9d6a9572ca-kube-api-access-54jth\") on node \"crc\" DevicePath \"\"" Oct 08 18:33:01 crc kubenswrapper[4988]: I1008 18:33:01.111150 4988 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8gplz\" (UniqueName: \"kubernetes.io/projected/6d2df307-23d3-48fd-9bb6-f09e500aaf9e-kube-api-access-8gplz\") on node \"crc\" DevicePath \"\"" Oct 08 18:33:01 crc kubenswrapper[4988]: I1008 18:33:01.111158 4988 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/9159c915-07a7-4b41-8672-8b9d6a9572ca-config-data-custom\") on node \"crc\" DevicePath \"\"" Oct 08 18:33:01 crc kubenswrapper[4988]: I1008 18:33:01.114065 4988 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_b5dc1ee3-4897-45a2-ad00-0b91fbb2ae76/ovsdbserver-nb/0.log" Oct 08 18:33:01 crc kubenswrapper[4988]: I1008 18:33:01.114152 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"b5dc1ee3-4897-45a2-ad00-0b91fbb2ae76","Type":"ContainerDied","Data":"74fae05a8629c65a280b6c23f9ffb0aee300b29c4f7758a9298dae712985fc5d"} Oct 08 18:33:01 crc kubenswrapper[4988]: I1008 18:33:01.114247 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Oct 08 18:33:01 crc kubenswrapper[4988]: I1008 18:33:01.138423 4988 generic.go:334] "Generic (PLEG): container finished" podID="f1628013-e0b9-4c31-80c0-91aabaaef0f6" containerID="863aa8ddeda792a78ba5af321605974ac7082d24fc8fb6f0400148fcf3f06126" exitCode=0 Oct 08 18:33:01 crc kubenswrapper[4988]: I1008 18:33:01.138687 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-kc7fw" event={"ID":"f1628013-e0b9-4c31-80c0-91aabaaef0f6","Type":"ContainerDied","Data":"863aa8ddeda792a78ba5af321605974ac7082d24fc8fb6f0400148fcf3f06126"} Oct 08 18:33:01 crc kubenswrapper[4988]: I1008 18:33:01.157193 4988 generic.go:334] "Generic (PLEG): container finished" podID="6d2df307-23d3-48fd-9bb6-f09e500aaf9e" containerID="ed0f1f9337e028007e93109bea68d88246a9ba24b639073b907414a6728b9835" exitCode=0 Oct 08 18:33:01 crc kubenswrapper[4988]: I1008 18:33:01.157424 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-6bb998f7d5-h4g2x" event={"ID":"6d2df307-23d3-48fd-9bb6-f09e500aaf9e","Type":"ContainerDied","Data":"ed0f1f9337e028007e93109bea68d88246a9ba24b639073b907414a6728b9835"} Oct 08 18:33:01 crc kubenswrapper[4988]: I1008 18:33:01.157501 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-6bb998f7d5-h4g2x" event={"ID":"6d2df307-23d3-48fd-9bb6-f09e500aaf9e","Type":"ContainerDied","Data":"8dd0d4be4a76c246ed2902066cd503da58a22dc94e2ffe053725437104391724"} Oct 08 18:33:01 crc kubenswrapper[4988]: I1008 18:33:01.157580 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-6bb998f7d5-h4g2x" Oct 08 18:33:01 crc kubenswrapper[4988]: I1008 18:33:01.163074 4988 generic.go:334] "Generic (PLEG): container finished" podID="79cbf065-a458-4dec-b6a7-805f515142e0" containerID="59a272e450f11633e4f5704d41f6545802b926f41dbb29197c7d02ab91df05c0" exitCode=0 Oct 08 18:33:01 crc kubenswrapper[4988]: I1008 18:33:01.163136 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement79af-account-delete-hv6j5" event={"ID":"79cbf065-a458-4dec-b6a7-805f515142e0","Type":"ContainerDied","Data":"59a272e450f11633e4f5704d41f6545802b926f41dbb29197c7d02ab91df05c0"} Oct 08 18:33:01 crc kubenswrapper[4988]: I1008 18:33:01.163807 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b5dc1ee3-4897-45a2-ad00-0b91fbb2ae76-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "b5dc1ee3-4897-45a2-ad00-0b91fbb2ae76" (UID: "b5dc1ee3-4897-45a2-ad00-0b91fbb2ae76"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:33:01 crc kubenswrapper[4988]: I1008 18:33:01.167441 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glancec9f7-account-delete-sp22m" podStartSLOduration=4.167420557 podStartE2EDuration="4.167420557s" podCreationTimestamp="2025-10-08 18:32:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 18:33:01.153203592 +0000 UTC m=+1326.603046382" watchObservedRunningTime="2025-10-08 18:33:01.167420557 +0000 UTC m=+1326.617263347" Oct 08 18:33:01 crc kubenswrapper[4988]: I1008 18:33:01.177557 4988 generic.go:334] "Generic (PLEG): container finished" podID="857686e3-0c6b-4d08-a8d6-f89e4c832139" containerID="ee0c93d867a043a6ace66d0ccc23af52eb47335f2734cbb64f7f04eb41db6e3c" exitCode=0 Oct 08 18:33:01 crc kubenswrapper[4988]: I1008 18:33:01.177629 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-7b7cf6bf4f-s6vrs" event={"ID":"857686e3-0c6b-4d08-a8d6-f89e4c832139","Type":"ContainerDied","Data":"ee0c93d867a043a6ace66d0ccc23af52eb47335f2734cbb64f7f04eb41db6e3c"} Oct 08 18:33:01 crc kubenswrapper[4988]: I1008 18:33:01.177854 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-7b7cf6bf4f-s6vrs" event={"ID":"857686e3-0c6b-4d08-a8d6-f89e4c832139","Type":"ContainerDied","Data":"9bfa3aa2a36b16efd6317af958966d09bd036f9b524035fdf3e77884e6bb79c6"} Oct 08 18:33:01 crc kubenswrapper[4988]: I1008 18:33:01.177629 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-7b7cf6bf4f-s6vrs" Oct 08 18:33:01 crc kubenswrapper[4988]: I1008 18:33:01.179040 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron3849-account-delete-m479b" podStartSLOduration=4.179019259 podStartE2EDuration="4.179019259s" podCreationTimestamp="2025-10-08 18:32:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 18:33:01.173571634 +0000 UTC m=+1326.623414414" watchObservedRunningTime="2025-10-08 18:33:01.179019259 +0000 UTC m=+1326.628862029" Oct 08 18:33:01 crc kubenswrapper[4988]: I1008 18:33:01.180635 4988 generic.go:334] "Generic (PLEG): container finished" podID="9159c915-07a7-4b41-8672-8b9d6a9572ca" containerID="1f77c788702456fc178f26b1d4a4449a94e1ac2f3f33ad505531b3885e2f8f15" exitCode=0 Oct 08 18:33:01 crc kubenswrapper[4988]: I1008 18:33:01.180701 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-7bf697f8df-h54ft" event={"ID":"9159c915-07a7-4b41-8672-8b9d6a9572ca","Type":"ContainerDied","Data":"1f77c788702456fc178f26b1d4a4449a94e1ac2f3f33ad505531b3885e2f8f15"} Oct 08 18:33:01 crc kubenswrapper[4988]: I1008 18:33:01.180721 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-7bf697f8df-h54ft" event={"ID":"9159c915-07a7-4b41-8672-8b9d6a9572ca","Type":"ContainerDied","Data":"fb877c896c174d8672045022e21f1ab69fe359bfbc9c5f531d8c89eae810dc71"} Oct 08 18:33:01 crc kubenswrapper[4988]: I1008 18:33:01.180724 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-7bf697f8df-h54ft" Oct 08 18:33:01 crc kubenswrapper[4988]: I1008 18:33:01.194730 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5967cc9597-r6jrv" event={"ID":"467e736b-562a-4f31-af64-2fce55e965c8","Type":"ContainerDied","Data":"d9b29ee2ea5ef3070f75c3d04f7fcb2db21a70a8b07adf0ecf69890c24328f79"} Oct 08 18:33:01 crc kubenswrapper[4988]: I1008 18:33:01.195072 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5967cc9597-r6jrv" Oct 08 18:33:01 crc kubenswrapper[4988]: I1008 18:33:01.212549 4988 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b5dc1ee3-4897-45a2-ad00-0b91fbb2ae76-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 18:33:01 crc kubenswrapper[4988]: I1008 18:33:01.232916 4988 generic.go:334] "Generic (PLEG): container finished" podID="d04259e5-9595-447b-b804-feba577498d8" containerID="590b8a61b67584c2acbd2bea77795253fad7c7d2a79eb0fb7fde120c857233af" exitCode=0 Oct 08 18:33:01 crc kubenswrapper[4988]: I1008 18:33:01.232987 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"d04259e5-9595-447b-b804-feba577498d8","Type":"ContainerDied","Data":"590b8a61b67584c2acbd2bea77795253fad7c7d2a79eb0fb7fde120c857233af"} Oct 08 18:33:01 crc kubenswrapper[4988]: I1008 18:33:01.233119 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/novacell14316-account-delete-gbtkl" Oct 08 18:33:01 crc kubenswrapper[4988]: I1008 18:33:01.276706 4988 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0fd89f92-b9aa-4a7b-a3ed-130abc4ce161" path="/var/lib/kubelet/pods/0fd89f92-b9aa-4a7b-a3ed-130abc4ce161/volumes" Oct 08 18:33:01 crc kubenswrapper[4988]: I1008 18:33:01.282674 4988 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1f36756f-acb3-439d-b10e-13573cbd252a" path="/var/lib/kubelet/pods/1f36756f-acb3-439d-b10e-13573cbd252a/volumes" Oct 08 18:33:01 crc kubenswrapper[4988]: I1008 18:33:01.283828 4988 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="32722c37-a19b-42e1-be71-432f3499af15" path="/var/lib/kubelet/pods/32722c37-a19b-42e1-be71-432f3499af15/volumes" Oct 08 18:33:01 crc kubenswrapper[4988]: I1008 18:33:01.284584 4988 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="41f6a322-8491-4b4c-9535-4122214d14b5" path="/var/lib/kubelet/pods/41f6a322-8491-4b4c-9535-4122214d14b5/volumes" Oct 08 18:33:01 crc kubenswrapper[4988]: I1008 18:33:01.288261 4988 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7aaaa954-1cc6-4304-8a21-86d9dfe4fd23" path="/var/lib/kubelet/pods/7aaaa954-1cc6-4304-8a21-86d9dfe4fd23/volumes" Oct 08 18:33:01 crc kubenswrapper[4988]: I1008 18:33:01.288868 4988 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a86ec23c-c0e4-4e88-901f-ac5e66f103b7" path="/var/lib/kubelet/pods/a86ec23c-c0e4-4e88-901f-ac5e66f103b7/volumes" Oct 08 18:33:01 crc kubenswrapper[4988]: I1008 18:33:01.289449 4988 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b92d978b-87c6-4210-9a69-99596b05cc18" path="/var/lib/kubelet/pods/b92d978b-87c6-4210-9a69-99596b05cc18/volumes" Oct 08 18:33:01 crc kubenswrapper[4988]: I1008 18:33:01.290632 4988 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c97aab35-bbee-41a6-88b2-801b0b097a6d" path="/var/lib/kubelet/pods/c97aab35-bbee-41a6-88b2-801b0b097a6d/volumes" Oct 08 18:33:01 crc kubenswrapper[4988]: I1008 18:33:01.291198 4988 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ee22c4f7-5dc2-4f3c-99d7-5c1c578c9afe" path="/var/lib/kubelet/pods/ee22c4f7-5dc2-4f3c-99d7-5c1c578c9afe/volumes" Oct 08 18:33:01 crc kubenswrapper[4988]: I1008 18:33:01.291801 4988 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f026ec8f-ffb4-4538-a6fc-3116ce0e7805" path="/var/lib/kubelet/pods/f026ec8f-ffb4-4538-a6fc-3116ce0e7805/volumes" Oct 08 18:33:01 crc kubenswrapper[4988]: I1008 18:33:01.292503 4988 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="feb5b694-6e5c-4778-8adf-6e3d89c0c875" path="/var/lib/kubelet/pods/feb5b694-6e5c-4778-8adf-6e3d89c0c875/volumes" Oct 08 18:33:01 crc kubenswrapper[4988]: I1008 18:33:01.383890 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/816b358d-4428-4fae-b91c-03c0171faa23-vencrypt-tls-certs" (OuterVolumeSpecName: "vencrypt-tls-certs") pod "816b358d-4428-4fae-b91c-03c0171faa23" (UID: "816b358d-4428-4fae-b91c-03c0171faa23"). InnerVolumeSpecName "vencrypt-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:33:01 crc kubenswrapper[4988]: I1008 18:33:01.415207 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/816b358d-4428-4fae-b91c-03c0171faa23-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "816b358d-4428-4fae-b91c-03c0171faa23" (UID: "816b358d-4428-4fae-b91c-03c0171faa23"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:33:01 crc kubenswrapper[4988]: I1008 18:33:01.416681 4988 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/816b358d-4428-4fae-b91c-03c0171faa23-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 18:33:01 crc kubenswrapper[4988]: I1008 18:33:01.416707 4988 reconciler_common.go:293] "Volume detached for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/816b358d-4428-4fae-b91c-03c0171faa23-vencrypt-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 08 18:33:01 crc kubenswrapper[4988]: I1008 18:33:01.423517 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/857686e3-0c6b-4d08-a8d6-f89e4c832139-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "857686e3-0c6b-4d08-a8d6-f89e4c832139" (UID: "857686e3-0c6b-4d08-a8d6-f89e4c832139"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:33:01 crc kubenswrapper[4988]: I1008 18:33:01.455756 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9159c915-07a7-4b41-8672-8b9d6a9572ca-config-data" (OuterVolumeSpecName: "config-data") pod "9159c915-07a7-4b41-8672-8b9d6a9572ca" (UID: "9159c915-07a7-4b41-8672-8b9d6a9572ca"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:33:01 crc kubenswrapper[4988]: I1008 18:33:01.459018 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/857686e3-0c6b-4d08-a8d6-f89e4c832139-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "857686e3-0c6b-4d08-a8d6-f89e4c832139" (UID: "857686e3-0c6b-4d08-a8d6-f89e4c832139"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:33:01 crc kubenswrapper[4988]: I1008 18:33:01.461247 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/857686e3-0c6b-4d08-a8d6-f89e4c832139-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "857686e3-0c6b-4d08-a8d6-f89e4c832139" (UID: "857686e3-0c6b-4d08-a8d6-f89e4c832139"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:33:01 crc kubenswrapper[4988]: I1008 18:33:01.465187 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/857686e3-0c6b-4d08-a8d6-f89e4c832139-config-data" (OuterVolumeSpecName: "config-data") pod "857686e3-0c6b-4d08-a8d6-f89e4c832139" (UID: "857686e3-0c6b-4d08-a8d6-f89e4c832139"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:33:01 crc kubenswrapper[4988]: I1008 18:33:01.469553 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9159c915-07a7-4b41-8672-8b9d6a9572ca-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "9159c915-07a7-4b41-8672-8b9d6a9572ca" (UID: "9159c915-07a7-4b41-8672-8b9d6a9572ca"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:33:01 crc kubenswrapper[4988]: I1008 18:33:01.498039 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/816b358d-4428-4fae-b91c-03c0171faa23-config-data" (OuterVolumeSpecName: "config-data") pod "816b358d-4428-4fae-b91c-03c0171faa23" (UID: "816b358d-4428-4fae-b91c-03c0171faa23"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:33:01 crc kubenswrapper[4988]: I1008 18:33:01.501325 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b5dc1ee3-4897-45a2-ad00-0b91fbb2ae76-metrics-certs-tls-certs" (OuterVolumeSpecName: "metrics-certs-tls-certs") pod "b5dc1ee3-4897-45a2-ad00-0b91fbb2ae76" (UID: "b5dc1ee3-4897-45a2-ad00-0b91fbb2ae76"). InnerVolumeSpecName "metrics-certs-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:33:01 crc kubenswrapper[4988]: I1008 18:33:01.518585 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/816b358d-4428-4fae-b91c-03c0171faa23-nova-novncproxy-tls-certs" (OuterVolumeSpecName: "nova-novncproxy-tls-certs") pod "816b358d-4428-4fae-b91c-03c0171faa23" (UID: "816b358d-4428-4fae-b91c-03c0171faa23"). InnerVolumeSpecName "nova-novncproxy-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:33:01 crc kubenswrapper[4988]: I1008 18:33:01.519422 4988 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/857686e3-0c6b-4d08-a8d6-f89e4c832139-config-data\") on node \"crc\" DevicePath \"\"" Oct 08 18:33:01 crc kubenswrapper[4988]: I1008 18:33:01.519446 4988 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/857686e3-0c6b-4d08-a8d6-f89e4c832139-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 18:33:01 crc kubenswrapper[4988]: I1008 18:33:01.519455 4988 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9159c915-07a7-4b41-8672-8b9d6a9572ca-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 18:33:01 crc kubenswrapper[4988]: I1008 18:33:01.519465 4988 reconciler_common.go:293] "Volume detached for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/b5dc1ee3-4897-45a2-ad00-0b91fbb2ae76-metrics-certs-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 08 18:33:01 crc kubenswrapper[4988]: I1008 18:33:01.519474 4988 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9159c915-07a7-4b41-8672-8b9d6a9572ca-config-data\") on node \"crc\" DevicePath \"\"" Oct 08 18:33:01 crc kubenswrapper[4988]: I1008 18:33:01.519484 4988 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/816b358d-4428-4fae-b91c-03c0171faa23-config-data\") on node \"crc\" DevicePath \"\"" Oct 08 18:33:01 crc kubenswrapper[4988]: I1008 18:33:01.519492 4988 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/857686e3-0c6b-4d08-a8d6-f89e4c832139-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 08 18:33:01 crc kubenswrapper[4988]: I1008 18:33:01.519500 4988 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/857686e3-0c6b-4d08-a8d6-f89e4c832139-public-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 08 18:33:01 crc kubenswrapper[4988]: I1008 18:33:01.519507 4988 reconciler_common.go:293] "Volume detached for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/816b358d-4428-4fae-b91c-03c0171faa23-nova-novncproxy-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 08 18:33:01 crc kubenswrapper[4988]: I1008 18:33:01.525014 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6d2df307-23d3-48fd-9bb6-f09e500aaf9e-config" (OuterVolumeSpecName: "config") pod "6d2df307-23d3-48fd-9bb6-f09e500aaf9e" (UID: "6d2df307-23d3-48fd-9bb6-f09e500aaf9e"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:33:01 crc kubenswrapper[4988]: I1008 18:33:01.528248 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b5dc1ee3-4897-45a2-ad00-0b91fbb2ae76-ovsdbserver-nb-tls-certs" (OuterVolumeSpecName: "ovsdbserver-nb-tls-certs") pod "b5dc1ee3-4897-45a2-ad00-0b91fbb2ae76" (UID: "b5dc1ee3-4897-45a2-ad00-0b91fbb2ae76"). InnerVolumeSpecName "ovsdbserver-nb-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:33:01 crc kubenswrapper[4988]: I1008 18:33:01.535540 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6d2df307-23d3-48fd-9bb6-f09e500aaf9e-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "6d2df307-23d3-48fd-9bb6-f09e500aaf9e" (UID: "6d2df307-23d3-48fd-9bb6-f09e500aaf9e"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:33:01 crc kubenswrapper[4988]: I1008 18:33:01.540854 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6d2df307-23d3-48fd-9bb6-f09e500aaf9e-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "6d2df307-23d3-48fd-9bb6-f09e500aaf9e" (UID: "6d2df307-23d3-48fd-9bb6-f09e500aaf9e"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:33:01 crc kubenswrapper[4988]: I1008 18:33:01.554740 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6d2df307-23d3-48fd-9bb6-f09e500aaf9e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "6d2df307-23d3-48fd-9bb6-f09e500aaf9e" (UID: "6d2df307-23d3-48fd-9bb6-f09e500aaf9e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:33:01 crc kubenswrapper[4988]: I1008 18:33:01.577855 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6d2df307-23d3-48fd-9bb6-f09e500aaf9e-ovndb-tls-certs" (OuterVolumeSpecName: "ovndb-tls-certs") pod "6d2df307-23d3-48fd-9bb6-f09e500aaf9e" (UID: "6d2df307-23d3-48fd-9bb6-f09e500aaf9e"). InnerVolumeSpecName "ovndb-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:33:01 crc kubenswrapper[4988]: I1008 18:33:01.620798 4988 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/6d2df307-23d3-48fd-9bb6-f09e500aaf9e-public-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 08 18:33:01 crc kubenswrapper[4988]: I1008 18:33:01.620843 4988 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/6d2df307-23d3-48fd-9bb6-f09e500aaf9e-config\") on node \"crc\" DevicePath \"\"" Oct 08 18:33:01 crc kubenswrapper[4988]: I1008 18:33:01.620856 4988 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6d2df307-23d3-48fd-9bb6-f09e500aaf9e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 18:33:01 crc kubenswrapper[4988]: I1008 18:33:01.620867 4988 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/b5dc1ee3-4897-45a2-ad00-0b91fbb2ae76-ovsdbserver-nb-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 08 18:33:01 crc kubenswrapper[4988]: I1008 18:33:01.620879 4988 reconciler_common.go:293] "Volume detached for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/6d2df307-23d3-48fd-9bb6-f09e500aaf9e-ovndb-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 08 18:33:01 crc kubenswrapper[4988]: I1008 18:33:01.620889 4988 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/6d2df307-23d3-48fd-9bb6-f09e500aaf9e-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 08 18:33:01 crc kubenswrapper[4988]: I1008 18:33:01.976260 4988 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/cinder-api-0" podUID="99d77eed-40f0-4879-be96-abea3a807cd9" containerName="cinder-api" probeResult="failure" output="Get \"https://10.217.0.171:8776/healthcheck\": dial tcp 10.217.0.171:8776: connect: connection refused" Oct 08 18:33:02 crc kubenswrapper[4988]: I1008 18:33:02.065675 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 08 18:33:02 crc kubenswrapper[4988]: I1008 18:33:02.066014 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 08 18:33:02 crc kubenswrapper[4988]: I1008 18:33:02.066034 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/memcached-0"] Oct 08 18:33:02 crc kubenswrapper[4988]: I1008 18:33:02.066075 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-sync-wzc5c"] Oct 08 18:33:02 crc kubenswrapper[4988]: I1008 18:33:02.066090 4988 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-sync-wzc5c"] Oct 08 18:33:02 crc kubenswrapper[4988]: I1008 18:33:02.066105 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-748qj"] Oct 08 18:33:02 crc kubenswrapper[4988]: I1008 18:33:02.066115 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-6dbc6dc9bd-hk8qw"] Oct 08 18:33:02 crc kubenswrapper[4988]: I1008 18:33:02.066147 4988 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-748qj"] Oct 08 18:33:02 crc kubenswrapper[4988]: I1008 18:33:02.066158 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone3e59-account-delete-vgpc9"] Oct 08 18:33:02 crc kubenswrapper[4988]: E1008 18:33:02.066577 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6d2df307-23d3-48fd-9bb6-f09e500aaf9e" containerName="neutron-httpd" Oct 08 18:33:02 crc kubenswrapper[4988]: I1008 18:33:02.066594 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="6d2df307-23d3-48fd-9bb6-f09e500aaf9e" containerName="neutron-httpd" Oct 08 18:33:02 crc kubenswrapper[4988]: E1008 18:33:02.066613 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9159c915-07a7-4b41-8672-8b9d6a9572ca" containerName="barbican-worker" Oct 08 18:33:02 crc kubenswrapper[4988]: I1008 18:33:02.066619 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="9159c915-07a7-4b41-8672-8b9d6a9572ca" containerName="barbican-worker" Oct 08 18:33:02 crc kubenswrapper[4988]: E1008 18:33:02.066633 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1f36756f-acb3-439d-b10e-13573cbd252a" containerName="ovsdbserver-sb" Oct 08 18:33:02 crc kubenswrapper[4988]: I1008 18:33:02.066639 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="1f36756f-acb3-439d-b10e-13573cbd252a" containerName="ovsdbserver-sb" Oct 08 18:33:02 crc kubenswrapper[4988]: E1008 18:33:02.066648 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9159c915-07a7-4b41-8672-8b9d6a9572ca" containerName="barbican-worker-log" Oct 08 18:33:02 crc kubenswrapper[4988]: I1008 18:33:02.066653 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="9159c915-07a7-4b41-8672-8b9d6a9572ca" containerName="barbican-worker-log" Oct 08 18:33:02 crc kubenswrapper[4988]: E1008 18:33:02.066662 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="857686e3-0c6b-4d08-a8d6-f89e4c832139" containerName="proxy-httpd" Oct 08 18:33:02 crc kubenswrapper[4988]: I1008 18:33:02.066668 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="857686e3-0c6b-4d08-a8d6-f89e4c832139" containerName="proxy-httpd" Oct 08 18:33:02 crc kubenswrapper[4988]: E1008 18:33:02.066681 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="467e736b-562a-4f31-af64-2fce55e965c8" containerName="init" Oct 08 18:33:02 crc kubenswrapper[4988]: I1008 18:33:02.066687 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="467e736b-562a-4f31-af64-2fce55e965c8" containerName="init" Oct 08 18:33:02 crc kubenswrapper[4988]: E1008 18:33:02.066697 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b5dc1ee3-4897-45a2-ad00-0b91fbb2ae76" containerName="ovsdbserver-nb" Oct 08 18:33:02 crc kubenswrapper[4988]: I1008 18:33:02.066703 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="b5dc1ee3-4897-45a2-ad00-0b91fbb2ae76" containerName="ovsdbserver-nb" Oct 08 18:33:02 crc kubenswrapper[4988]: E1008 18:33:02.066714 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="816b358d-4428-4fae-b91c-03c0171faa23" containerName="nova-cell1-novncproxy-novncproxy" Oct 08 18:33:02 crc kubenswrapper[4988]: I1008 18:33:02.066721 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="816b358d-4428-4fae-b91c-03c0171faa23" containerName="nova-cell1-novncproxy-novncproxy" Oct 08 18:33:02 crc kubenswrapper[4988]: E1008 18:33:02.066752 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1f36756f-acb3-439d-b10e-13573cbd252a" containerName="openstack-network-exporter" Oct 08 18:33:02 crc kubenswrapper[4988]: I1008 18:33:02.066758 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="1f36756f-acb3-439d-b10e-13573cbd252a" containerName="openstack-network-exporter" Oct 08 18:33:02 crc kubenswrapper[4988]: E1008 18:33:02.066772 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b5dc1ee3-4897-45a2-ad00-0b91fbb2ae76" containerName="openstack-network-exporter" Oct 08 18:33:02 crc kubenswrapper[4988]: I1008 18:33:02.066778 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="b5dc1ee3-4897-45a2-ad00-0b91fbb2ae76" containerName="openstack-network-exporter" Oct 08 18:33:02 crc kubenswrapper[4988]: E1008 18:33:02.066791 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f026ec8f-ffb4-4538-a6fc-3116ce0e7805" containerName="ovn-northd" Oct 08 18:33:02 crc kubenswrapper[4988]: I1008 18:33:02.066798 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="f026ec8f-ffb4-4538-a6fc-3116ce0e7805" containerName="ovn-northd" Oct 08 18:33:02 crc kubenswrapper[4988]: E1008 18:33:02.066810 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f026ec8f-ffb4-4538-a6fc-3116ce0e7805" containerName="openstack-network-exporter" Oct 08 18:33:02 crc kubenswrapper[4988]: I1008 18:33:02.066816 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="f026ec8f-ffb4-4538-a6fc-3116ce0e7805" containerName="openstack-network-exporter" Oct 08 18:33:02 crc kubenswrapper[4988]: E1008 18:33:02.066831 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="467e736b-562a-4f31-af64-2fce55e965c8" containerName="dnsmasq-dns" Oct 08 18:33:02 crc kubenswrapper[4988]: I1008 18:33:02.066836 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="467e736b-562a-4f31-af64-2fce55e965c8" containerName="dnsmasq-dns" Oct 08 18:33:02 crc kubenswrapper[4988]: E1008 18:33:02.066848 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6d2df307-23d3-48fd-9bb6-f09e500aaf9e" containerName="neutron-api" Oct 08 18:33:02 crc kubenswrapper[4988]: I1008 18:33:02.066853 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="6d2df307-23d3-48fd-9bb6-f09e500aaf9e" containerName="neutron-api" Oct 08 18:33:02 crc kubenswrapper[4988]: E1008 18:33:02.066868 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b92d978b-87c6-4210-9a69-99596b05cc18" containerName="openstack-network-exporter" Oct 08 18:33:02 crc kubenswrapper[4988]: I1008 18:33:02.066873 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="b92d978b-87c6-4210-9a69-99596b05cc18" containerName="openstack-network-exporter" Oct 08 18:33:02 crc kubenswrapper[4988]: E1008 18:33:02.066885 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="857686e3-0c6b-4d08-a8d6-f89e4c832139" containerName="proxy-server" Oct 08 18:33:02 crc kubenswrapper[4988]: I1008 18:33:02.066891 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="857686e3-0c6b-4d08-a8d6-f89e4c832139" containerName="proxy-server" Oct 08 18:33:02 crc kubenswrapper[4988]: I1008 18:33:02.067073 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="b5dc1ee3-4897-45a2-ad00-0b91fbb2ae76" containerName="ovsdbserver-nb" Oct 08 18:33:02 crc kubenswrapper[4988]: I1008 18:33:02.067084 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="b92d978b-87c6-4210-9a69-99596b05cc18" containerName="openstack-network-exporter" Oct 08 18:33:02 crc kubenswrapper[4988]: I1008 18:33:02.067095 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="9159c915-07a7-4b41-8672-8b9d6a9572ca" containerName="barbican-worker" Oct 08 18:33:02 crc kubenswrapper[4988]: I1008 18:33:02.067104 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="9159c915-07a7-4b41-8672-8b9d6a9572ca" containerName="barbican-worker-log" Oct 08 18:33:02 crc kubenswrapper[4988]: I1008 18:33:02.067125 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="6d2df307-23d3-48fd-9bb6-f09e500aaf9e" containerName="neutron-httpd" Oct 08 18:33:02 crc kubenswrapper[4988]: I1008 18:33:02.067136 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="1f36756f-acb3-439d-b10e-13573cbd252a" containerName="ovsdbserver-sb" Oct 08 18:33:02 crc kubenswrapper[4988]: I1008 18:33:02.067142 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="467e736b-562a-4f31-af64-2fce55e965c8" containerName="dnsmasq-dns" Oct 08 18:33:02 crc kubenswrapper[4988]: I1008 18:33:02.067150 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="6d2df307-23d3-48fd-9bb6-f09e500aaf9e" containerName="neutron-api" Oct 08 18:33:02 crc kubenswrapper[4988]: I1008 18:33:02.067157 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="857686e3-0c6b-4d08-a8d6-f89e4c832139" containerName="proxy-httpd" Oct 08 18:33:02 crc kubenswrapper[4988]: I1008 18:33:02.067198 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="1f36756f-acb3-439d-b10e-13573cbd252a" containerName="openstack-network-exporter" Oct 08 18:33:02 crc kubenswrapper[4988]: I1008 18:33:02.067210 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="857686e3-0c6b-4d08-a8d6-f89e4c832139" containerName="proxy-server" Oct 08 18:33:02 crc kubenswrapper[4988]: I1008 18:33:02.067218 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="b5dc1ee3-4897-45a2-ad00-0b91fbb2ae76" containerName="openstack-network-exporter" Oct 08 18:33:02 crc kubenswrapper[4988]: I1008 18:33:02.067228 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="f026ec8f-ffb4-4538-a6fc-3116ce0e7805" containerName="openstack-network-exporter" Oct 08 18:33:02 crc kubenswrapper[4988]: I1008 18:33:02.067237 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="816b358d-4428-4fae-b91c-03c0171faa23" containerName="nova-cell1-novncproxy-novncproxy" Oct 08 18:33:02 crc kubenswrapper[4988]: I1008 18:33:02.067246 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="f026ec8f-ffb4-4538-a6fc-3116ce0e7805" containerName="ovn-northd" Oct 08 18:33:02 crc kubenswrapper[4988]: I1008 18:33:02.067858 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone3e59-account-delete-vgpc9"] Oct 08 18:33:02 crc kubenswrapper[4988]: I1008 18:33:02.067872 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/openstack-galera-0"] Oct 08 18:33:02 crc kubenswrapper[4988]: I1008 18:33:02.073621 4988 scope.go:117] "RemoveContainer" containerID="a947f771405cb73547e24443c664920da93801df8ff9dd6369c101deea715564" Oct 08 18:33:02 crc kubenswrapper[4988]: I1008 18:33:02.073906 4988 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/memcached-0" podUID="32bb7401-6292-4ca6-98f4-15522125c3ee" containerName="memcached" containerID="cri-o://dcd1682ce23dc072a1b6bf3c5aeff64ca69361e18fa5cdcb9712e65170ddc37d" gracePeriod=30 Oct 08 18:33:02 crc kubenswrapper[4988]: I1008 18:33:02.078593 4988 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/kube-state-metrics-0" podUID="b6d1becf-4229-40bf-b53a-d8a9eb4b580e" containerName="kube-state-metrics" containerID="cri-o://0f31eca0db70bc881a635808747404e9ac717d3d38c16e693305f6196d241dac" gracePeriod=30 Oct 08 18:33:02 crc kubenswrapper[4988]: I1008 18:33:02.078934 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone3e59-account-delete-vgpc9" Oct 08 18:33:02 crc kubenswrapper[4988]: I1008 18:33:02.078928 4988 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/keystone-6dbc6dc9bd-hk8qw" podUID="26a3b0a9-6aca-490d-a6fb-08addef86fbb" containerName="keystone-api" containerID="cri-o://62d82328b02c86e078d53d3e4ba279c27a7c17092bf7900027389a74f0ad59d0" gracePeriod=30 Oct 08 18:33:02 crc kubenswrapper[4988]: I1008 18:33:02.079186 4988 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="ccd2f8fa-1c16-4c58-9f87-e60ba3a2f5fb" containerName="ceilometer-central-agent" containerID="cri-o://c7ddcc51fc71ff2377802bb6d1c0c0545ab7211828f994c467b39e3b24eaac16" gracePeriod=30 Oct 08 18:33:02 crc kubenswrapper[4988]: I1008 18:33:02.079208 4988 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="ccd2f8fa-1c16-4c58-9f87-e60ba3a2f5fb" containerName="sg-core" containerID="cri-o://be90eff6d6a94661069208a3760ad429da6e736c3f64018e09b59d748a843d33" gracePeriod=30 Oct 08 18:33:02 crc kubenswrapper[4988]: I1008 18:33:02.079253 4988 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="ccd2f8fa-1c16-4c58-9f87-e60ba3a2f5fb" containerName="ceilometer-notification-agent" containerID="cri-o://06f5039b59f5eef23b916412ccb9a5658d672fcb48cf3d4a8e55e5a55e9519c0" gracePeriod=30 Oct 08 18:33:02 crc kubenswrapper[4988]: I1008 18:33:02.079272 4988 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="ccd2f8fa-1c16-4c58-9f87-e60ba3a2f5fb" containerName="proxy-httpd" containerID="cri-o://f64fdd9df8d4aff7bcd8cfab3af523fa24041d42ea9cfb0addd7bf856026fe76" gracePeriod=30 Oct 08 18:33:02 crc kubenswrapper[4988]: E1008 18:33:02.085963 4988 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a947f771405cb73547e24443c664920da93801df8ff9dd6369c101deea715564\": container with ID starting with a947f771405cb73547e24443c664920da93801df8ff9dd6369c101deea715564 not found: ID does not exist" containerID="a947f771405cb73547e24443c664920da93801df8ff9dd6369c101deea715564" Oct 08 18:33:02 crc kubenswrapper[4988]: I1008 18:33:02.086002 4988 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a947f771405cb73547e24443c664920da93801df8ff9dd6369c101deea715564"} err="failed to get container status \"a947f771405cb73547e24443c664920da93801df8ff9dd6369c101deea715564\": rpc error: code = NotFound desc = could not find container \"a947f771405cb73547e24443c664920da93801df8ff9dd6369c101deea715564\": container with ID starting with a947f771405cb73547e24443c664920da93801df8ff9dd6369c101deea715564 not found: ID does not exist" Oct 08 18:33:02 crc kubenswrapper[4988]: I1008 18:33:02.086027 4988 scope.go:117] "RemoveContainer" containerID="b5799148279b22815a3dbd06ed629a27e96e308b0f484f69614b62b235afbb9f" Oct 08 18:33:02 crc kubenswrapper[4988]: I1008 18:33:02.102110 4988 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="83ef7610-23ac-4445-b75c-ae7206b4d810" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.205:8775/\": read tcp 10.217.0.2:40906->10.217.0.205:8775: read: connection reset by peer" Oct 08 18:33:02 crc kubenswrapper[4988]: I1008 18:33:02.105940 4988 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="83ef7610-23ac-4445-b75c-ae7206b4d810" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.205:8775/\": read tcp 10.217.0.2:40918->10.217.0.205:8775: read: connection reset by peer" Oct 08 18:33:02 crc kubenswrapper[4988]: I1008 18:33:02.131120 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k8mnm\" (UniqueName: \"kubernetes.io/projected/11679d40-6eb6-456c-8996-cd9a946a6db1-kube-api-access-k8mnm\") pod \"keystone3e59-account-delete-vgpc9\" (UID: \"11679d40-6eb6-456c-8996-cd9a946a6db1\") " pod="openstack/keystone3e59-account-delete-vgpc9" Oct 08 18:33:02 crc kubenswrapper[4988]: I1008 18:33:02.192237 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-create-xnv8d"] Oct 08 18:33:02 crc kubenswrapper[4988]: I1008 18:33:02.195922 4988 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-create-xnv8d"] Oct 08 18:33:02 crc kubenswrapper[4988]: I1008 18:33:02.212222 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-79af-account-create-ddfvr"] Oct 08 18:33:02 crc kubenswrapper[4988]: I1008 18:33:02.225222 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement79af-account-delete-hv6j5"] Oct 08 18:33:02 crc kubenswrapper[4988]: I1008 18:33:02.228547 4988 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-79af-account-create-ddfvr"] Oct 08 18:33:02 crc kubenswrapper[4988]: I1008 18:33:02.240412 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cjfgk\" (UniqueName: \"kubernetes.io/projected/1127bf6b-65ff-4c59-928c-51e16bb18ce3-kube-api-access-cjfgk\") pod \"novacell14316-account-delete-gbtkl\" (UID: \"1127bf6b-65ff-4c59-928c-51e16bb18ce3\") " pod="openstack/novacell14316-account-delete-gbtkl" Oct 08 18:33:02 crc kubenswrapper[4988]: I1008 18:33:02.240573 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k8mnm\" (UniqueName: \"kubernetes.io/projected/11679d40-6eb6-456c-8996-cd9a946a6db1-kube-api-access-k8mnm\") pod \"keystone3e59-account-delete-vgpc9\" (UID: \"11679d40-6eb6-456c-8996-cd9a946a6db1\") " pod="openstack/keystone3e59-account-delete-vgpc9" Oct 08 18:33:02 crc kubenswrapper[4988]: I1008 18:33:02.242849 4988 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/glance-default-internal-api-0" podUID="0fcc9499-a06b-4cd1-a5da-ea8108916050" containerName="glance-httpd" probeResult="failure" output="Get \"https://10.217.0.174:9292/healthcheck\": read tcp 10.217.0.2:58672->10.217.0.174:9292: read: connection reset by peer" Oct 08 18:33:02 crc kubenswrapper[4988]: I1008 18:33:02.242884 4988 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/glance-default-internal-api-0" podUID="0fcc9499-a06b-4cd1-a5da-ea8108916050" containerName="glance-log" probeResult="failure" output="Get \"https://10.217.0.174:9292/healthcheck\": read tcp 10.217.0.2:58680->10.217.0.174:9292: read: connection reset by peer" Oct 08 18:33:02 crc kubenswrapper[4988]: E1008 18:33:02.247230 4988 projected.go:194] Error preparing data for projected volume kube-api-access-k8mnm for pod openstack/keystone3e59-account-delete-vgpc9: failed to fetch token: serviceaccounts "galera-openstack" not found Oct 08 18:33:02 crc kubenswrapper[4988]: E1008 18:33:02.247312 4988 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/11679d40-6eb6-456c-8996-cd9a946a6db1-kube-api-access-k8mnm podName:11679d40-6eb6-456c-8996-cd9a946a6db1 nodeName:}" failed. No retries permitted until 2025-10-08 18:33:02.747294058 +0000 UTC m=+1328.197136828 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-k8mnm" (UniqueName: "kubernetes.io/projected/11679d40-6eb6-456c-8996-cd9a946a6db1-kube-api-access-k8mnm") pod "keystone3e59-account-delete-vgpc9" (UID: "11679d40-6eb6-456c-8996-cd9a946a6db1") : failed to fetch token: serviceaccounts "galera-openstack" not found Oct 08 18:33:02 crc kubenswrapper[4988]: E1008 18:33:02.247462 4988 projected.go:194] Error preparing data for projected volume kube-api-access-cjfgk for pod openstack/novacell14316-account-delete-gbtkl: failed to fetch token: serviceaccounts "galera-openstack-cell1" not found Oct 08 18:33:02 crc kubenswrapper[4988]: E1008 18:33:02.247516 4988 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/1127bf6b-65ff-4c59-928c-51e16bb18ce3-kube-api-access-cjfgk podName:1127bf6b-65ff-4c59-928c-51e16bb18ce3 nodeName:}" failed. No retries permitted until 2025-10-08 18:33:06.247498325 +0000 UTC m=+1331.697341085 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cjfgk" (UniqueName: "kubernetes.io/projected/1127bf6b-65ff-4c59-928c-51e16bb18ce3-kube-api-access-cjfgk") pod "novacell14316-account-delete-gbtkl" (UID: "1127bf6b-65ff-4c59-928c-51e16bb18ce3") : failed to fetch token: serviceaccounts "galera-openstack-cell1" not found Oct 08 18:33:02 crc kubenswrapper[4988]: I1008 18:33:02.258210 4988 generic.go:334] "Generic (PLEG): container finished" podID="83ef7610-23ac-4445-b75c-ae7206b4d810" containerID="4b994190cba63b85214c41a6bf0e9b5cca7c29dd236d40d5df3055f98c9ec35b" exitCode=0 Oct 08 18:33:02 crc kubenswrapper[4988]: I1008 18:33:02.258283 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"83ef7610-23ac-4445-b75c-ae7206b4d810","Type":"ContainerDied","Data":"4b994190cba63b85214c41a6bf0e9b5cca7c29dd236d40d5df3055f98c9ec35b"} Oct 08 18:33:02 crc kubenswrapper[4988]: I1008 18:33:02.262809 4988 generic.go:334] "Generic (PLEG): container finished" podID="28f73b7a-bb5e-43ec-a099-13a12cdb33d3" containerID="6d9e1194470201a267d04dcb959b05798998db9b5ddd0f647596f14d4784a516" exitCode=0 Oct 08 18:33:02 crc kubenswrapper[4988]: I1008 18:33:02.262891 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/novaapiebe4-account-delete-rph2r" event={"ID":"28f73b7a-bb5e-43ec-a099-13a12cdb33d3","Type":"ContainerDied","Data":"6d9e1194470201a267d04dcb959b05798998db9b5ddd0f647596f14d4784a516"} Oct 08 18:33:02 crc kubenswrapper[4988]: I1008 18:33:02.267201 4988 generic.go:334] "Generic (PLEG): container finished" podID="ab3315d2-3248-42f1-9834-2bab55fa9935" containerID="ee5151c0c831844f2ab3796464c7e52d85522046bd44d8ed010f49dda7a02db4" exitCode=0 Oct 08 18:33:02 crc kubenswrapper[4988]: I1008 18:33:02.267274 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron3849-account-delete-m479b" event={"ID":"ab3315d2-3248-42f1-9834-2bab55fa9935","Type":"ContainerDied","Data":"ee5151c0c831844f2ab3796464c7e52d85522046bd44d8ed010f49dda7a02db4"} Oct 08 18:33:02 crc kubenswrapper[4988]: I1008 18:33:02.290073 4988 generic.go:334] "Generic (PLEG): container finished" podID="f4e47608-89b4-45d1-88a6-fe75805d864d" containerID="b226456d2f92fef2d798a8462e29b0c8f7c8a915dd5330879cb7257e0cc73d05" exitCode=0 Oct 08 18:33:02 crc kubenswrapper[4988]: I1008 18:33:02.290171 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-c44756996-dfw4k" event={"ID":"f4e47608-89b4-45d1-88a6-fe75805d864d","Type":"ContainerDied","Data":"b226456d2f92fef2d798a8462e29b0c8f7c8a915dd5330879cb7257e0cc73d05"} Oct 08 18:33:02 crc kubenswrapper[4988]: I1008 18:33:02.291780 4988 generic.go:334] "Generic (PLEG): container finished" podID="02a37054-4817-4151-a8cb-a9a45867c0d3" containerID="2b04aad608cf7534875c0cf78abd43e0c25cefee26a5ecebee48c75fd8c10b09" exitCode=0 Oct 08 18:33:02 crc kubenswrapper[4988]: I1008 18:33:02.292034 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinderd8da-account-delete-75j2h" event={"ID":"02a37054-4817-4151-a8cb-a9a45867c0d3","Type":"ContainerDied","Data":"2b04aad608cf7534875c0cf78abd43e0c25cefee26a5ecebee48c75fd8c10b09"} Oct 08 18:33:02 crc kubenswrapper[4988]: I1008 18:33:02.324843 4988 generic.go:334] "Generic (PLEG): container finished" podID="d7d648e7-3c02-4d26-993d-a901764b3329" containerID="31548197acc0e143742e4fb4cb9299994c2c1dd05e374425c882dac956b9d649" exitCode=0 Oct 08 18:33:02 crc kubenswrapper[4988]: I1008 18:33:02.324926 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glancec9f7-account-delete-sp22m" event={"ID":"d7d648e7-3c02-4d26-993d-a901764b3329","Type":"ContainerDied","Data":"31548197acc0e143742e4fb4cb9299994c2c1dd05e374425c882dac956b9d649"} Oct 08 18:33:02 crc kubenswrapper[4988]: I1008 18:33:02.335705 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-create-pspkt"] Oct 08 18:33:02 crc kubenswrapper[4988]: I1008 18:33:02.339592 4988 generic.go:334] "Generic (PLEG): container finished" podID="ccd2f8fa-1c16-4c58-9f87-e60ba3a2f5fb" containerID="be90eff6d6a94661069208a3760ad429da6e736c3f64018e09b59d748a843d33" exitCode=2 Oct 08 18:33:02 crc kubenswrapper[4988]: I1008 18:33:02.339624 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ccd2f8fa-1c16-4c58-9f87-e60ba3a2f5fb","Type":"ContainerDied","Data":"be90eff6d6a94661069208a3760ad429da6e736c3f64018e09b59d748a843d33"} Oct 08 18:33:02 crc kubenswrapper[4988]: I1008 18:33:02.342311 4988 generic.go:334] "Generic (PLEG): container finished" podID="b6d1becf-4229-40bf-b53a-d8a9eb4b580e" containerID="0f31eca0db70bc881a635808747404e9ac717d3d38c16e693305f6196d241dac" exitCode=2 Oct 08 18:33:02 crc kubenswrapper[4988]: I1008 18:33:02.342428 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"b6d1becf-4229-40bf-b53a-d8a9eb4b580e","Type":"ContainerDied","Data":"0f31eca0db70bc881a635808747404e9ac717d3d38c16e693305f6196d241dac"} Oct 08 18:33:02 crc kubenswrapper[4988]: I1008 18:33:02.358816 4988 generic.go:334] "Generic (PLEG): container finished" podID="99d77eed-40f0-4879-be96-abea3a807cd9" containerID="d2face2dfb0693cf7a2c3531e80c082e310d388b04c8005247e18e79ea8bbab3" exitCode=0 Oct 08 18:33:02 crc kubenswrapper[4988]: I1008 18:33:02.358918 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"99d77eed-40f0-4879-be96-abea3a807cd9","Type":"ContainerDied","Data":"d2face2dfb0693cf7a2c3531e80c082e310d388b04c8005247e18e79ea8bbab3"} Oct 08 18:33:02 crc kubenswrapper[4988]: I1008 18:33:02.362917 4988 generic.go:334] "Generic (PLEG): container finished" podID="56702a03-175d-4a54-b84f-e25cab037431" containerID="23da0ff1f117dc601e37e2bb0a79de17b4662200fa002a471f0c5938e67d484b" exitCode=0 Oct 08 18:33:02 crc kubenswrapper[4988]: I1008 18:33:02.363054 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"56702a03-175d-4a54-b84f-e25cab037431","Type":"ContainerDied","Data":"23da0ff1f117dc601e37e2bb0a79de17b4662200fa002a471f0c5938e67d484b"} Oct 08 18:33:02 crc kubenswrapper[4988]: I1008 18:33:02.363261 4988 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/openstack-galera-0" podUID="72c64965-0b57-43e1-8d18-2bce6cabbfa1" containerName="galera" containerID="cri-o://239fe689949862e30d5f5b0021f848adb4380b18bdefdeb68e24c7cd3fb41c24" gracePeriod=30 Oct 08 18:33:02 crc kubenswrapper[4988]: I1008 18:33:02.373288 4988 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-create-pspkt"] Oct 08 18:33:02 crc kubenswrapper[4988]: I1008 18:33:02.374962 4988 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-6c486d5744-j86zn" podUID="3a57537b-6115-4249-8d16-3ab0c4b6d21e" containerName="barbican-api-log" probeResult="failure" output="Get \"https://10.217.0.156:9311/healthcheck\": read tcp 10.217.0.2:44752->10.217.0.156:9311: read: connection reset by peer" Oct 08 18:33:02 crc kubenswrapper[4988]: I1008 18:33:02.375024 4988 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-6c486d5744-j86zn" podUID="3a57537b-6115-4249-8d16-3ab0c4b6d21e" containerName="barbican-api" probeResult="failure" output="Get \"https://10.217.0.156:9311/healthcheck\": read tcp 10.217.0.2:44768->10.217.0.156:9311: read: connection reset by peer" Oct 08 18:33:02 crc kubenswrapper[4988]: I1008 18:33:02.382686 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-c9f7-account-create-p5k9l"] Oct 08 18:33:02 crc kubenswrapper[4988]: I1008 18:33:02.390456 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glancec9f7-account-delete-sp22m"] Oct 08 18:33:02 crc kubenswrapper[4988]: I1008 18:33:02.398875 4988 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-c9f7-account-create-p5k9l"] Oct 08 18:33:02 crc kubenswrapper[4988]: E1008 18:33:02.443166 4988 configmap.go:193] Couldn't get configMap openstack/rabbitmq-config-data: configmap "rabbitmq-config-data" not found Oct 08 18:33:02 crc kubenswrapper[4988]: E1008 18:33:02.443233 4988 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/830d6ec5-0015-4a88-b8e8-0fd47f64fd46-config-data podName:830d6ec5-0015-4a88-b8e8-0fd47f64fd46 nodeName:}" failed. No retries permitted until 2025-10-08 18:33:06.443218454 +0000 UTC m=+1331.893061224 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/configmap/830d6ec5-0015-4a88-b8e8-0fd47f64fd46-config-data") pod "rabbitmq-server-0" (UID: "830d6ec5-0015-4a88-b8e8-0fd47f64fd46") : configmap "rabbitmq-config-data" not found Oct 08 18:33:02 crc kubenswrapper[4988]: I1008 18:33:02.469256 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-create-bt5zq"] Oct 08 18:33:02 crc kubenswrapper[4988]: I1008 18:33:02.476465 4988 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-create-bt5zq"] Oct 08 18:33:02 crc kubenswrapper[4988]: I1008 18:33:02.493946 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-32b1-account-create-tcnwc"] Oct 08 18:33:02 crc kubenswrapper[4988]: I1008 18:33:02.503434 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican32b1-account-delete-jlhch"] Oct 08 18:33:02 crc kubenswrapper[4988]: I1008 18:33:02.515769 4988 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-32b1-account-create-tcnwc"] Oct 08 18:33:02 crc kubenswrapper[4988]: I1008 18:33:02.549345 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-create-8dclb"] Oct 08 18:33:02 crc kubenswrapper[4988]: I1008 18:33:02.559129 4988 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-create-8dclb"] Oct 08 18:33:02 crc kubenswrapper[4988]: I1008 18:33:02.565305 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron3849-account-delete-m479b"] Oct 08 18:33:02 crc kubenswrapper[4988]: I1008 18:33:02.571017 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-3849-account-create-87mb6"] Oct 08 18:33:02 crc kubenswrapper[4988]: I1008 18:33:02.595978 4988 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-3849-account-create-87mb6"] Oct 08 18:33:02 crc kubenswrapper[4988]: I1008 18:33:02.715722 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-db-create-vbbnt"] Oct 08 18:33:02 crc kubenswrapper[4988]: I1008 18:33:02.729264 4988 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-db-create-vbbnt"] Oct 08 18:33:02 crc kubenswrapper[4988]: I1008 18:33:02.737513 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-2f70-account-create-tdkwz"] Oct 08 18:33:02 crc kubenswrapper[4988]: I1008 18:33:02.753233 4988 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-2f70-account-create-tdkwz"] Oct 08 18:33:02 crc kubenswrapper[4988]: I1008 18:33:02.755107 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k8mnm\" (UniqueName: \"kubernetes.io/projected/11679d40-6eb6-456c-8996-cd9a946a6db1-kube-api-access-k8mnm\") pod \"keystone3e59-account-delete-vgpc9\" (UID: \"11679d40-6eb6-456c-8996-cd9a946a6db1\") " pod="openstack/keystone3e59-account-delete-vgpc9" Oct 08 18:33:02 crc kubenswrapper[4988]: E1008 18:33:02.758272 4988 projected.go:194] Error preparing data for projected volume kube-api-access-k8mnm for pod openstack/keystone3e59-account-delete-vgpc9: failed to fetch token: serviceaccounts "galera-openstack" not found Oct 08 18:33:02 crc kubenswrapper[4988]: E1008 18:33:02.758334 4988 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/11679d40-6eb6-456c-8996-cd9a946a6db1-kube-api-access-k8mnm podName:11679d40-6eb6-456c-8996-cd9a946a6db1 nodeName:}" failed. No retries permitted until 2025-10-08 18:33:03.758312068 +0000 UTC m=+1329.208154828 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-k8mnm" (UniqueName: "kubernetes.io/projected/11679d40-6eb6-456c-8996-cd9a946a6db1-kube-api-access-k8mnm") pod "keystone3e59-account-delete-vgpc9" (UID: "11679d40-6eb6-456c-8996-cd9a946a6db1") : failed to fetch token: serviceaccounts "galera-openstack" not found Oct 08 18:33:02 crc kubenswrapper[4988]: I1008 18:33:02.771047 4988 scope.go:117] "RemoveContainer" containerID="74829553075c7054526031a117b72832271514093fbe3617d9278cd92e26275b" Oct 08 18:33:02 crc kubenswrapper[4988]: I1008 18:33:02.781892 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/novacell02f70-account-delete-cdwsx"] Oct 08 18:33:02 crc kubenswrapper[4988]: I1008 18:33:02.818750 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 08 18:33:02 crc kubenswrapper[4988]: I1008 18:33:02.823101 4988 scope.go:117] "RemoveContainer" containerID="db863ec0ad9c9861590e77462403049f0e352d2f01da17f2db5e4682b149163e" Oct 08 18:33:02 crc kubenswrapper[4988]: I1008 18:33:02.866302 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-c44756996-dfw4k" Oct 08 18:33:02 crc kubenswrapper[4988]: I1008 18:33:02.901256 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/novaapiebe4-account-delete-rph2r" Oct 08 18:33:02 crc kubenswrapper[4988]: I1008 18:33:02.918562 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5967cc9597-r6jrv"] Oct 08 18:33:02 crc kubenswrapper[4988]: I1008 18:33:02.925862 4988 scope.go:117] "RemoveContainer" containerID="ed0f1f9337e028007e93109bea68d88246a9ba24b639073b907414a6728b9835" Oct 08 18:33:02 crc kubenswrapper[4988]: I1008 18:33:02.944397 4988 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5967cc9597-r6jrv"] Oct 08 18:33:02 crc kubenswrapper[4988]: I1008 18:33:02.966841 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tt47m\" (UniqueName: \"kubernetes.io/projected/f4e47608-89b4-45d1-88a6-fe75805d864d-kube-api-access-tt47m\") pod \"f4e47608-89b4-45d1-88a6-fe75805d864d\" (UID: \"f4e47608-89b4-45d1-88a6-fe75805d864d\") " Oct 08 18:33:02 crc kubenswrapper[4988]: I1008 18:33:02.966887 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7xsc4\" (UniqueName: \"kubernetes.io/projected/28f73b7a-bb5e-43ec-a099-13a12cdb33d3-kube-api-access-7xsc4\") pod \"28f73b7a-bb5e-43ec-a099-13a12cdb33d3\" (UID: \"28f73b7a-bb5e-43ec-a099-13a12cdb33d3\") " Oct 08 18:33:02 crc kubenswrapper[4988]: I1008 18:33:02.966931 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f4e47608-89b4-45d1-88a6-fe75805d864d-combined-ca-bundle\") pod \"f4e47608-89b4-45d1-88a6-fe75805d864d\" (UID: \"f4e47608-89b4-45d1-88a6-fe75805d864d\") " Oct 08 18:33:02 crc kubenswrapper[4988]: I1008 18:33:02.966961 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/99d77eed-40f0-4879-be96-abea3a807cd9-config-data\") pod \"99d77eed-40f0-4879-be96-abea3a807cd9\" (UID: \"99d77eed-40f0-4879-be96-abea3a807cd9\") " Oct 08 18:33:02 crc kubenswrapper[4988]: I1008 18:33:02.966982 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/99d77eed-40f0-4879-be96-abea3a807cd9-config-data-custom\") pod \"99d77eed-40f0-4879-be96-abea3a807cd9\" (UID: \"99d77eed-40f0-4879-be96-abea3a807cd9\") " Oct 08 18:33:02 crc kubenswrapper[4988]: I1008 18:33:02.967019 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/99d77eed-40f0-4879-be96-abea3a807cd9-scripts\") pod \"99d77eed-40f0-4879-be96-abea3a807cd9\" (UID: \"99d77eed-40f0-4879-be96-abea3a807cd9\") " Oct 08 18:33:02 crc kubenswrapper[4988]: I1008 18:33:02.967093 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f4e47608-89b4-45d1-88a6-fe75805d864d-logs\") pod \"f4e47608-89b4-45d1-88a6-fe75805d864d\" (UID: \"f4e47608-89b4-45d1-88a6-fe75805d864d\") " Oct 08 18:33:02 crc kubenswrapper[4988]: I1008 18:33:02.967165 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/f4e47608-89b4-45d1-88a6-fe75805d864d-internal-tls-certs\") pod \"f4e47608-89b4-45d1-88a6-fe75805d864d\" (UID: \"f4e47608-89b4-45d1-88a6-fe75805d864d\") " Oct 08 18:33:02 crc kubenswrapper[4988]: I1008 18:33:02.967186 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/99d77eed-40f0-4879-be96-abea3a807cd9-public-tls-certs\") pod \"99d77eed-40f0-4879-be96-abea3a807cd9\" (UID: \"99d77eed-40f0-4879-be96-abea3a807cd9\") " Oct 08 18:33:02 crc kubenswrapper[4988]: I1008 18:33:02.967212 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b6krt\" (UniqueName: \"kubernetes.io/projected/99d77eed-40f0-4879-be96-abea3a807cd9-kube-api-access-b6krt\") pod \"99d77eed-40f0-4879-be96-abea3a807cd9\" (UID: \"99d77eed-40f0-4879-be96-abea3a807cd9\") " Oct 08 18:33:02 crc kubenswrapper[4988]: I1008 18:33:02.967252 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/99d77eed-40f0-4879-be96-abea3a807cd9-combined-ca-bundle\") pod \"99d77eed-40f0-4879-be96-abea3a807cd9\" (UID: \"99d77eed-40f0-4879-be96-abea3a807cd9\") " Oct 08 18:33:02 crc kubenswrapper[4988]: I1008 18:33:02.967328 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f4e47608-89b4-45d1-88a6-fe75805d864d-scripts\") pod \"f4e47608-89b4-45d1-88a6-fe75805d864d\" (UID: \"f4e47608-89b4-45d1-88a6-fe75805d864d\") " Oct 08 18:33:02 crc kubenswrapper[4988]: I1008 18:33:02.967345 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/99d77eed-40f0-4879-be96-abea3a807cd9-internal-tls-certs\") pod \"99d77eed-40f0-4879-be96-abea3a807cd9\" (UID: \"99d77eed-40f0-4879-be96-abea3a807cd9\") " Oct 08 18:33:02 crc kubenswrapper[4988]: I1008 18:33:02.967426 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/99d77eed-40f0-4879-be96-abea3a807cd9-logs\") pod \"99d77eed-40f0-4879-be96-abea3a807cd9\" (UID: \"99d77eed-40f0-4879-be96-abea3a807cd9\") " Oct 08 18:33:02 crc kubenswrapper[4988]: I1008 18:33:02.967446 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/99d77eed-40f0-4879-be96-abea3a807cd9-etc-machine-id\") pod \"99d77eed-40f0-4879-be96-abea3a807cd9\" (UID: \"99d77eed-40f0-4879-be96-abea3a807cd9\") " Oct 08 18:33:02 crc kubenswrapper[4988]: I1008 18:33:02.967470 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/f4e47608-89b4-45d1-88a6-fe75805d864d-public-tls-certs\") pod \"f4e47608-89b4-45d1-88a6-fe75805d864d\" (UID: \"f4e47608-89b4-45d1-88a6-fe75805d864d\") " Oct 08 18:33:02 crc kubenswrapper[4988]: I1008 18:33:02.967516 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f4e47608-89b4-45d1-88a6-fe75805d864d-config-data\") pod \"f4e47608-89b4-45d1-88a6-fe75805d864d\" (UID: \"f4e47608-89b4-45d1-88a6-fe75805d864d\") " Oct 08 18:33:02 crc kubenswrapper[4988]: I1008 18:33:02.983659 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f4e47608-89b4-45d1-88a6-fe75805d864d-scripts" (OuterVolumeSpecName: "scripts") pod "f4e47608-89b4-45d1-88a6-fe75805d864d" (UID: "f4e47608-89b4-45d1-88a6-fe75805d864d"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:33:02 crc kubenswrapper[4988]: I1008 18:33:02.986510 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/99d77eed-40f0-4879-be96-abea3a807cd9-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "99d77eed-40f0-4879-be96-abea3a807cd9" (UID: "99d77eed-40f0-4879-be96-abea3a807cd9"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 08 18:33:02 crc kubenswrapper[4988]: I1008 18:33:02.986900 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/99d77eed-40f0-4879-be96-abea3a807cd9-logs" (OuterVolumeSpecName: "logs") pod "99d77eed-40f0-4879-be96-abea3a807cd9" (UID: "99d77eed-40f0-4879-be96-abea3a807cd9"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 18:33:02 crc kubenswrapper[4988]: I1008 18:33:02.989092 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/novacell14316-account-delete-gbtkl"] Oct 08 18:33:02 crc kubenswrapper[4988]: I1008 18:33:02.990218 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinderd8da-account-delete-75j2h" Oct 08 18:33:03 crc kubenswrapper[4988]: I1008 18:33:03.002133 4988 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/novacell14316-account-delete-gbtkl"] Oct 08 18:33:03 crc kubenswrapper[4988]: I1008 18:33:03.002607 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f4e47608-89b4-45d1-88a6-fe75805d864d-logs" (OuterVolumeSpecName: "logs") pod "f4e47608-89b4-45d1-88a6-fe75805d864d" (UID: "f4e47608-89b4-45d1-88a6-fe75805d864d"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 18:33:03 crc kubenswrapper[4988]: I1008 18:33:03.015615 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/28f73b7a-bb5e-43ec-a099-13a12cdb33d3-kube-api-access-7xsc4" (OuterVolumeSpecName: "kube-api-access-7xsc4") pod "28f73b7a-bb5e-43ec-a099-13a12cdb33d3" (UID: "28f73b7a-bb5e-43ec-a099-13a12cdb33d3"). InnerVolumeSpecName "kube-api-access-7xsc4". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:33:03 crc kubenswrapper[4988]: I1008 18:33:03.016195 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/swift-proxy-7b7cf6bf4f-s6vrs"] Oct 08 18:33:03 crc kubenswrapper[4988]: I1008 18:33:03.021346 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/99d77eed-40f0-4879-be96-abea3a807cd9-kube-api-access-b6krt" (OuterVolumeSpecName: "kube-api-access-b6krt") pod "99d77eed-40f0-4879-be96-abea3a807cd9" (UID: "99d77eed-40f0-4879-be96-abea3a807cd9"). InnerVolumeSpecName "kube-api-access-b6krt". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:33:03 crc kubenswrapper[4988]: I1008 18:33:03.026084 4988 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/swift-proxy-7b7cf6bf4f-s6vrs"] Oct 08 18:33:03 crc kubenswrapper[4988]: I1008 18:33:03.039275 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/99d77eed-40f0-4879-be96-abea3a807cd9-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "99d77eed-40f0-4879-be96-abea3a807cd9" (UID: "99d77eed-40f0-4879-be96-abea3a807cd9"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:33:03 crc kubenswrapper[4988]: I1008 18:33:03.039340 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 08 18:33:03 crc kubenswrapper[4988]: I1008 18:33:03.041626 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/99d77eed-40f0-4879-be96-abea3a807cd9-scripts" (OuterVolumeSpecName: "scripts") pod "99d77eed-40f0-4879-be96-abea3a807cd9" (UID: "99d77eed-40f0-4879-be96-abea3a807cd9"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:33:03 crc kubenswrapper[4988]: I1008 18:33:03.045081 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 08 18:33:03 crc kubenswrapper[4988]: I1008 18:33:03.045830 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f4e47608-89b4-45d1-88a6-fe75805d864d-kube-api-access-tt47m" (OuterVolumeSpecName: "kube-api-access-tt47m") pod "f4e47608-89b4-45d1-88a6-fe75805d864d" (UID: "f4e47608-89b4-45d1-88a6-fe75805d864d"). InnerVolumeSpecName "kube-api-access-tt47m". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:33:03 crc kubenswrapper[4988]: I1008 18:33:03.068993 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gxjkk\" (UniqueName: \"kubernetes.io/projected/02a37054-4817-4151-a8cb-a9a45867c0d3-kube-api-access-gxjkk\") pod \"02a37054-4817-4151-a8cb-a9a45867c0d3\" (UID: \"02a37054-4817-4151-a8cb-a9a45867c0d3\") " Oct 08 18:33:03 crc kubenswrapper[4988]: I1008 18:33:03.069852 4988 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tt47m\" (UniqueName: \"kubernetes.io/projected/f4e47608-89b4-45d1-88a6-fe75805d864d-kube-api-access-tt47m\") on node \"crc\" DevicePath \"\"" Oct 08 18:33:03 crc kubenswrapper[4988]: I1008 18:33:03.070096 4988 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7xsc4\" (UniqueName: \"kubernetes.io/projected/28f73b7a-bb5e-43ec-a099-13a12cdb33d3-kube-api-access-7xsc4\") on node \"crc\" DevicePath \"\"" Oct 08 18:33:03 crc kubenswrapper[4988]: I1008 18:33:03.070169 4988 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/99d77eed-40f0-4879-be96-abea3a807cd9-config-data-custom\") on node \"crc\" DevicePath \"\"" Oct 08 18:33:03 crc kubenswrapper[4988]: I1008 18:33:03.070223 4988 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/99d77eed-40f0-4879-be96-abea3a807cd9-scripts\") on node \"crc\" DevicePath \"\"" Oct 08 18:33:03 crc kubenswrapper[4988]: I1008 18:33:03.070273 4988 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f4e47608-89b4-45d1-88a6-fe75805d864d-logs\") on node \"crc\" DevicePath \"\"" Oct 08 18:33:03 crc kubenswrapper[4988]: I1008 18:33:03.070331 4988 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cjfgk\" (UniqueName: \"kubernetes.io/projected/1127bf6b-65ff-4c59-928c-51e16bb18ce3-kube-api-access-cjfgk\") on node \"crc\" DevicePath \"\"" Oct 08 18:33:03 crc kubenswrapper[4988]: I1008 18:33:03.070421 4988 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b6krt\" (UniqueName: \"kubernetes.io/projected/99d77eed-40f0-4879-be96-abea3a807cd9-kube-api-access-b6krt\") on node \"crc\" DevicePath \"\"" Oct 08 18:33:03 crc kubenswrapper[4988]: I1008 18:33:03.070484 4988 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f4e47608-89b4-45d1-88a6-fe75805d864d-scripts\") on node \"crc\" DevicePath \"\"" Oct 08 18:33:03 crc kubenswrapper[4988]: I1008 18:33:03.070547 4988 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/99d77eed-40f0-4879-be96-abea3a807cd9-logs\") on node \"crc\" DevicePath \"\"" Oct 08 18:33:03 crc kubenswrapper[4988]: I1008 18:33:03.070641 4988 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/99d77eed-40f0-4879-be96-abea3a807cd9-etc-machine-id\") on node \"crc\" DevicePath \"\"" Oct 08 18:33:03 crc kubenswrapper[4988]: I1008 18:33:03.084487 4988 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 08 18:33:03 crc kubenswrapper[4988]: I1008 18:33:03.084722 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Oct 08 18:33:03 crc kubenswrapper[4988]: I1008 18:33:03.086597 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/02a37054-4817-4151-a8cb-a9a45867c0d3-kube-api-access-gxjkk" (OuterVolumeSpecName: "kube-api-access-gxjkk") pod "02a37054-4817-4151-a8cb-a9a45867c0d3" (UID: "02a37054-4817-4151-a8cb-a9a45867c0d3"). InnerVolumeSpecName "kube-api-access-gxjkk". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:33:03 crc kubenswrapper[4988]: I1008 18:33:03.104223 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovsdbserver-nb-0"] Oct 08 18:33:03 crc kubenswrapper[4988]: I1008 18:33:03.114008 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/99d77eed-40f0-4879-be96-abea3a807cd9-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "99d77eed-40f0-4879-be96-abea3a807cd9" (UID: "99d77eed-40f0-4879-be96-abea3a807cd9"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:33:03 crc kubenswrapper[4988]: I1008 18:33:03.115981 4988 scope.go:117] "RemoveContainer" containerID="db863ec0ad9c9861590e77462403049f0e352d2f01da17f2db5e4682b149163e" Oct 08 18:33:03 crc kubenswrapper[4988]: E1008 18:33:03.137306 4988 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"db863ec0ad9c9861590e77462403049f0e352d2f01da17f2db5e4682b149163e\": container with ID starting with db863ec0ad9c9861590e77462403049f0e352d2f01da17f2db5e4682b149163e not found: ID does not exist" containerID="db863ec0ad9c9861590e77462403049f0e352d2f01da17f2db5e4682b149163e" Oct 08 18:33:03 crc kubenswrapper[4988]: I1008 18:33:03.137602 4988 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"db863ec0ad9c9861590e77462403049f0e352d2f01da17f2db5e4682b149163e"} err="failed to get container status \"db863ec0ad9c9861590e77462403049f0e352d2f01da17f2db5e4682b149163e\": rpc error: code = NotFound desc = could not find container \"db863ec0ad9c9861590e77462403049f0e352d2f01da17f2db5e4682b149163e\": container with ID starting with db863ec0ad9c9861590e77462403049f0e352d2f01da17f2db5e4682b149163e not found: ID does not exist" Oct 08 18:33:03 crc kubenswrapper[4988]: I1008 18:33:03.137648 4988 scope.go:117] "RemoveContainer" containerID="ed0f1f9337e028007e93109bea68d88246a9ba24b639073b907414a6728b9835" Oct 08 18:33:03 crc kubenswrapper[4988]: E1008 18:33:03.138889 4988 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ed0f1f9337e028007e93109bea68d88246a9ba24b639073b907414a6728b9835\": container with ID starting with ed0f1f9337e028007e93109bea68d88246a9ba24b639073b907414a6728b9835 not found: ID does not exist" containerID="ed0f1f9337e028007e93109bea68d88246a9ba24b639073b907414a6728b9835" Oct 08 18:33:03 crc kubenswrapper[4988]: I1008 18:33:03.138933 4988 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ed0f1f9337e028007e93109bea68d88246a9ba24b639073b907414a6728b9835"} err="failed to get container status \"ed0f1f9337e028007e93109bea68d88246a9ba24b639073b907414a6728b9835\": rpc error: code = NotFound desc = could not find container \"ed0f1f9337e028007e93109bea68d88246a9ba24b639073b907414a6728b9835\": container with ID starting with ed0f1f9337e028007e93109bea68d88246a9ba24b639073b907414a6728b9835 not found: ID does not exist" Oct 08 18:33:03 crc kubenswrapper[4988]: I1008 18:33:03.139187 4988 scope.go:117] "RemoveContainer" containerID="ee0c93d867a043a6ace66d0ccc23af52eb47335f2734cbb64f7f04eb41db6e3c" Oct 08 18:33:03 crc kubenswrapper[4988]: I1008 18:33:03.139998 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican32b1-account-delete-jlhch" Oct 08 18:33:03 crc kubenswrapper[4988]: I1008 18:33:03.151024 4988 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovsdbserver-nb-0"] Oct 08 18:33:03 crc kubenswrapper[4988]: I1008 18:33:03.159375 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement79af-account-delete-hv6j5" Oct 08 18:33:03 crc kubenswrapper[4988]: I1008 18:33:03.166015 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-6bb998f7d5-h4g2x"] Oct 08 18:33:03 crc kubenswrapper[4988]: I1008 18:33:03.175741 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0fcc9499-a06b-4cd1-a5da-ea8108916050-logs\") pod \"0fcc9499-a06b-4cd1-a5da-ea8108916050\" (UID: \"0fcc9499-a06b-4cd1-a5da-ea8108916050\") " Oct 08 18:33:03 crc kubenswrapper[4988]: I1008 18:33:03.175828 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"0fcc9499-a06b-4cd1-a5da-ea8108916050\" (UID: \"0fcc9499-a06b-4cd1-a5da-ea8108916050\") " Oct 08 18:33:03 crc kubenswrapper[4988]: I1008 18:33:03.175849 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/0fcc9499-a06b-4cd1-a5da-ea8108916050-internal-tls-certs\") pod \"0fcc9499-a06b-4cd1-a5da-ea8108916050\" (UID: \"0fcc9499-a06b-4cd1-a5da-ea8108916050\") " Oct 08 18:33:03 crc kubenswrapper[4988]: I1008 18:33:03.175895 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/b6d1becf-4229-40bf-b53a-d8a9eb4b580e-kube-state-metrics-tls-certs\") pod \"b6d1becf-4229-40bf-b53a-d8a9eb4b580e\" (UID: \"b6d1becf-4229-40bf-b53a-d8a9eb4b580e\") " Oct 08 18:33:03 crc kubenswrapper[4988]: I1008 18:33:03.175920 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0fcc9499-a06b-4cd1-a5da-ea8108916050-combined-ca-bundle\") pod \"0fcc9499-a06b-4cd1-a5da-ea8108916050\" (UID: \"0fcc9499-a06b-4cd1-a5da-ea8108916050\") " Oct 08 18:33:03 crc kubenswrapper[4988]: I1008 18:33:03.175959 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2hv4g\" (UniqueName: \"kubernetes.io/projected/b6d1becf-4229-40bf-b53a-d8a9eb4b580e-kube-api-access-2hv4g\") pod \"b6d1becf-4229-40bf-b53a-d8a9eb4b580e\" (UID: \"b6d1becf-4229-40bf-b53a-d8a9eb4b580e\") " Oct 08 18:33:03 crc kubenswrapper[4988]: I1008 18:33:03.176030 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0fcc9499-a06b-4cd1-a5da-ea8108916050-config-data\") pod \"0fcc9499-a06b-4cd1-a5da-ea8108916050\" (UID: \"0fcc9499-a06b-4cd1-a5da-ea8108916050\") " Oct 08 18:33:03 crc kubenswrapper[4988]: I1008 18:33:03.176105 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/b6d1becf-4229-40bf-b53a-d8a9eb4b580e-kube-state-metrics-tls-config\") pod \"b6d1becf-4229-40bf-b53a-d8a9eb4b580e\" (UID: \"b6d1becf-4229-40bf-b53a-d8a9eb4b580e\") " Oct 08 18:33:03 crc kubenswrapper[4988]: I1008 18:33:03.176174 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b6d1becf-4229-40bf-b53a-d8a9eb4b580e-combined-ca-bundle\") pod \"b6d1becf-4229-40bf-b53a-d8a9eb4b580e\" (UID: \"b6d1becf-4229-40bf-b53a-d8a9eb4b580e\") " Oct 08 18:33:03 crc kubenswrapper[4988]: I1008 18:33:03.176204 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/0fcc9499-a06b-4cd1-a5da-ea8108916050-httpd-run\") pod \"0fcc9499-a06b-4cd1-a5da-ea8108916050\" (UID: \"0fcc9499-a06b-4cd1-a5da-ea8108916050\") " Oct 08 18:33:03 crc kubenswrapper[4988]: I1008 18:33:03.176232 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-g9rvl\" (UniqueName: \"kubernetes.io/projected/0fcc9499-a06b-4cd1-a5da-ea8108916050-kube-api-access-g9rvl\") pod \"0fcc9499-a06b-4cd1-a5da-ea8108916050\" (UID: \"0fcc9499-a06b-4cd1-a5da-ea8108916050\") " Oct 08 18:33:03 crc kubenswrapper[4988]: I1008 18:33:03.176264 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0fcc9499-a06b-4cd1-a5da-ea8108916050-scripts\") pod \"0fcc9499-a06b-4cd1-a5da-ea8108916050\" (UID: \"0fcc9499-a06b-4cd1-a5da-ea8108916050\") " Oct 08 18:33:03 crc kubenswrapper[4988]: I1008 18:33:03.178952 4988 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/99d77eed-40f0-4879-be96-abea3a807cd9-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 18:33:03 crc kubenswrapper[4988]: I1008 18:33:03.178977 4988 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gxjkk\" (UniqueName: \"kubernetes.io/projected/02a37054-4817-4151-a8cb-a9a45867c0d3-kube-api-access-gxjkk\") on node \"crc\" DevicePath \"\"" Oct 08 18:33:03 crc kubenswrapper[4988]: I1008 18:33:03.196362 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0fcc9499-a06b-4cd1-a5da-ea8108916050-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "0fcc9499-a06b-4cd1-a5da-ea8108916050" (UID: "0fcc9499-a06b-4cd1-a5da-ea8108916050"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 18:33:03 crc kubenswrapper[4988]: I1008 18:33:03.198137 4988 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-6bb998f7d5-h4g2x"] Oct 08 18:33:03 crc kubenswrapper[4988]: E1008 18:33:03.201267 4988 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="e18a49c9de8366d161727792956633e70d61d0fc8c2799f046b7a6f94af795e5" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Oct 08 18:33:03 crc kubenswrapper[4988]: E1008 18:33:03.204075 4988 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="e18a49c9de8366d161727792956633e70d61d0fc8c2799f046b7a6f94af795e5" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Oct 08 18:33:03 crc kubenswrapper[4988]: I1008 18:33:03.205208 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0fcc9499-a06b-4cd1-a5da-ea8108916050-logs" (OuterVolumeSpecName: "logs") pod "0fcc9499-a06b-4cd1-a5da-ea8108916050" (UID: "0fcc9499-a06b-4cd1-a5da-ea8108916050"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 18:33:03 crc kubenswrapper[4988]: E1008 18:33:03.206463 4988 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="e18a49c9de8366d161727792956633e70d61d0fc8c2799f046b7a6f94af795e5" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Oct 08 18:33:03 crc kubenswrapper[4988]: E1008 18:33:03.206548 4988 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-cell1-conductor-0" podUID="85bc7f8f-fdb3-4abb-88d3-ed0c10982c4c" containerName="nova-cell1-conductor-conductor" Oct 08 18:33:03 crc kubenswrapper[4988]: I1008 18:33:03.208077 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-worker-7bf697f8df-h54ft"] Oct 08 18:33:03 crc kubenswrapper[4988]: I1008 18:33:03.215613 4988 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-worker-7bf697f8df-h54ft"] Oct 08 18:33:03 crc kubenswrapper[4988]: I1008 18:33:03.250140 4988 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1127bf6b-65ff-4c59-928c-51e16bb18ce3" path="/var/lib/kubelet/pods/1127bf6b-65ff-4c59-928c-51e16bb18ce3/volumes" Oct 08 18:33:03 crc kubenswrapper[4988]: I1008 18:33:03.250785 4988 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="434ac32e-3136-4d51-b68e-919adae50ffe" path="/var/lib/kubelet/pods/434ac32e-3136-4d51-b68e-919adae50ffe/volumes" Oct 08 18:33:03 crc kubenswrapper[4988]: I1008 18:33:03.252818 4988 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="467e736b-562a-4f31-af64-2fce55e965c8" path="/var/lib/kubelet/pods/467e736b-562a-4f31-af64-2fce55e965c8/volumes" Oct 08 18:33:03 crc kubenswrapper[4988]: I1008 18:33:03.253314 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0fcc9499-a06b-4cd1-a5da-ea8108916050-scripts" (OuterVolumeSpecName: "scripts") pod "0fcc9499-a06b-4cd1-a5da-ea8108916050" (UID: "0fcc9499-a06b-4cd1-a5da-ea8108916050"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:33:03 crc kubenswrapper[4988]: I1008 18:33:03.255744 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage05-crc" (OuterVolumeSpecName: "glance") pod "0fcc9499-a06b-4cd1-a5da-ea8108916050" (UID: "0fcc9499-a06b-4cd1-a5da-ea8108916050"). InnerVolumeSpecName "local-storage05-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 08 18:33:03 crc kubenswrapper[4988]: I1008 18:33:03.255857 4988 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5f273f27-4991-4aa3-a734-dec827cfb78b" path="/var/lib/kubelet/pods/5f273f27-4991-4aa3-a734-dec827cfb78b/volumes" Oct 08 18:33:03 crc kubenswrapper[4988]: I1008 18:33:03.256624 4988 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="689970ec-f287-451e-bf7d-fde8af627bf1" path="/var/lib/kubelet/pods/689970ec-f287-451e-bf7d-fde8af627bf1/volumes" Oct 08 18:33:03 crc kubenswrapper[4988]: I1008 18:33:03.257583 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/99d77eed-40f0-4879-be96-abea3a807cd9-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "99d77eed-40f0-4879-be96-abea3a807cd9" (UID: "99d77eed-40f0-4879-be96-abea3a807cd9"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:33:03 crc kubenswrapper[4988]: I1008 18:33:03.258713 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6d1becf-4229-40bf-b53a-d8a9eb4b580e-kube-api-access-2hv4g" (OuterVolumeSpecName: "kube-api-access-2hv4g") pod "b6d1becf-4229-40bf-b53a-d8a9eb4b580e" (UID: "b6d1becf-4229-40bf-b53a-d8a9eb4b580e"). InnerVolumeSpecName "kube-api-access-2hv4g". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:33:03 crc kubenswrapper[4988]: I1008 18:33:03.258804 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0fcc9499-a06b-4cd1-a5da-ea8108916050-kube-api-access-g9rvl" (OuterVolumeSpecName: "kube-api-access-g9rvl") pod "0fcc9499-a06b-4cd1-a5da-ea8108916050" (UID: "0fcc9499-a06b-4cd1-a5da-ea8108916050"). InnerVolumeSpecName "kube-api-access-g9rvl". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:33:03 crc kubenswrapper[4988]: I1008 18:33:03.259312 4988 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6d2df307-23d3-48fd-9bb6-f09e500aaf9e" path="/var/lib/kubelet/pods/6d2df307-23d3-48fd-9bb6-f09e500aaf9e/volumes" Oct 08 18:33:03 crc kubenswrapper[4988]: I1008 18:33:03.267611 4988 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="816b358d-4428-4fae-b91c-03c0171faa23" path="/var/lib/kubelet/pods/816b358d-4428-4fae-b91c-03c0171faa23/volumes" Oct 08 18:33:03 crc kubenswrapper[4988]: I1008 18:33:03.271996 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/99d77eed-40f0-4879-be96-abea3a807cd9-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "99d77eed-40f0-4879-be96-abea3a807cd9" (UID: "99d77eed-40f0-4879-be96-abea3a807cd9"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:33:03 crc kubenswrapper[4988]: I1008 18:33:03.273277 4988 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="857686e3-0c6b-4d08-a8d6-f89e4c832139" path="/var/lib/kubelet/pods/857686e3-0c6b-4d08-a8d6-f89e4c832139/volumes" Oct 08 18:33:03 crc kubenswrapper[4988]: I1008 18:33:03.273951 4988 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="881d3e0f-adfd-4b37-a942-6698a6897963" path="/var/lib/kubelet/pods/881d3e0f-adfd-4b37-a942-6698a6897963/volumes" Oct 08 18:33:03 crc kubenswrapper[4988]: I1008 18:33:03.277996 4988 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8f236c17-7f75-40a3-9937-111f670a1aeb" path="/var/lib/kubelet/pods/8f236c17-7f75-40a3-9937-111f670a1aeb/volumes" Oct 08 18:33:03 crc kubenswrapper[4988]: I1008 18:33:03.278782 4988 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9159c915-07a7-4b41-8672-8b9d6a9572ca" path="/var/lib/kubelet/pods/9159c915-07a7-4b41-8672-8b9d6a9572ca/volumes" Oct 08 18:33:03 crc kubenswrapper[4988]: I1008 18:33:03.279446 4988 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="99149e0d-f06b-4136-88da-313ab1705081" path="/var/lib/kubelet/pods/99149e0d-f06b-4136-88da-313ab1705081/volumes" Oct 08 18:33:03 crc kubenswrapper[4988]: I1008 18:33:03.279933 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hjvsd\" (UniqueName: \"kubernetes.io/projected/79cbf065-a458-4dec-b6a7-805f515142e0-kube-api-access-hjvsd\") pod \"79cbf065-a458-4dec-b6a7-805f515142e0\" (UID: \"79cbf065-a458-4dec-b6a7-805f515142e0\") " Oct 08 18:33:03 crc kubenswrapper[4988]: I1008 18:33:03.280148 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tv2pz\" (UniqueName: \"kubernetes.io/projected/3f7bddd3-86f8-46e9-a38e-1aa5e570d151-kube-api-access-tv2pz\") pod \"3f7bddd3-86f8-46e9-a38e-1aa5e570d151\" (UID: \"3f7bddd3-86f8-46e9-a38e-1aa5e570d151\") " Oct 08 18:33:03 crc kubenswrapper[4988]: I1008 18:33:03.280512 4988 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/99d77eed-40f0-4879-be96-abea3a807cd9-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 08 18:33:03 crc kubenswrapper[4988]: I1008 18:33:03.280540 4988 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") on node \"crc\" " Oct 08 18:33:03 crc kubenswrapper[4988]: I1008 18:33:03.280551 4988 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2hv4g\" (UniqueName: \"kubernetes.io/projected/b6d1becf-4229-40bf-b53a-d8a9eb4b580e-kube-api-access-2hv4g\") on node \"crc\" DevicePath \"\"" Oct 08 18:33:03 crc kubenswrapper[4988]: I1008 18:33:03.280561 4988 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/0fcc9499-a06b-4cd1-a5da-ea8108916050-httpd-run\") on node \"crc\" DevicePath \"\"" Oct 08 18:33:03 crc kubenswrapper[4988]: I1008 18:33:03.280568 4988 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-g9rvl\" (UniqueName: \"kubernetes.io/projected/0fcc9499-a06b-4cd1-a5da-ea8108916050-kube-api-access-g9rvl\") on node \"crc\" DevicePath \"\"" Oct 08 18:33:03 crc kubenswrapper[4988]: I1008 18:33:03.280577 4988 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0fcc9499-a06b-4cd1-a5da-ea8108916050-scripts\") on node \"crc\" DevicePath \"\"" Oct 08 18:33:03 crc kubenswrapper[4988]: I1008 18:33:03.280585 4988 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/99d77eed-40f0-4879-be96-abea3a807cd9-public-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 08 18:33:03 crc kubenswrapper[4988]: I1008 18:33:03.280592 4988 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0fcc9499-a06b-4cd1-a5da-ea8108916050-logs\") on node \"crc\" DevicePath \"\"" Oct 08 18:33:03 crc kubenswrapper[4988]: I1008 18:33:03.288070 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3f7bddd3-86f8-46e9-a38e-1aa5e570d151-kube-api-access-tv2pz" (OuterVolumeSpecName: "kube-api-access-tv2pz") pod "3f7bddd3-86f8-46e9-a38e-1aa5e570d151" (UID: "3f7bddd3-86f8-46e9-a38e-1aa5e570d151"). InnerVolumeSpecName "kube-api-access-tv2pz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:33:03 crc kubenswrapper[4988]: I1008 18:33:03.295980 4988 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9b6f6c59-608b-4f5b-b4ef-3cb0e4c85345" path="/var/lib/kubelet/pods/9b6f6c59-608b-4f5b-b4ef-3cb0e4c85345/volumes" Oct 08 18:33:03 crc kubenswrapper[4988]: I1008 18:33:03.296875 4988 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ac19d73a-13bf-4714-aa01-36b72ee937d9" path="/var/lib/kubelet/pods/ac19d73a-13bf-4714-aa01-36b72ee937d9/volumes" Oct 08 18:33:03 crc kubenswrapper[4988]: I1008 18:33:03.297544 4988 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b2e4e6bf-c872-4a45-be09-08decca63284" path="/var/lib/kubelet/pods/b2e4e6bf-c872-4a45-be09-08decca63284/volumes" Oct 08 18:33:03 crc kubenswrapper[4988]: I1008 18:33:03.298329 4988 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b5dc1ee3-4897-45a2-ad00-0b91fbb2ae76" path="/var/lib/kubelet/pods/b5dc1ee3-4897-45a2-ad00-0b91fbb2ae76/volumes" Oct 08 18:33:03 crc kubenswrapper[4988]: I1008 18:33:03.314943 4988 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="caebe15b-a16d-4e0d-9f75-9d78958e6f9a" path="/var/lib/kubelet/pods/caebe15b-a16d-4e0d-9f75-9d78958e6f9a/volumes" Oct 08 18:33:03 crc kubenswrapper[4988]: I1008 18:33:03.315665 4988 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f61ca6e6-56e3-4933-a8aa-5faa53c2b338" path="/var/lib/kubelet/pods/f61ca6e6-56e3-4933-a8aa-5faa53c2b338/volumes" Oct 08 18:33:03 crc kubenswrapper[4988]: I1008 18:33:03.316221 4988 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fa0ac55f-9e87-488b-8626-36024ca51102" path="/var/lib/kubelet/pods/fa0ac55f-9e87-488b-8626-36024ca51102/volumes" Oct 08 18:33:03 crc kubenswrapper[4988]: I1008 18:33:03.326288 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f4e47608-89b4-45d1-88a6-fe75805d864d-config-data" (OuterVolumeSpecName: "config-data") pod "f4e47608-89b4-45d1-88a6-fe75805d864d" (UID: "f4e47608-89b4-45d1-88a6-fe75805d864d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:33:03 crc kubenswrapper[4988]: I1008 18:33:03.327497 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/79cbf065-a458-4dec-b6a7-805f515142e0-kube-api-access-hjvsd" (OuterVolumeSpecName: "kube-api-access-hjvsd") pod "79cbf065-a458-4dec-b6a7-805f515142e0" (UID: "79cbf065-a458-4dec-b6a7-805f515142e0"). InnerVolumeSpecName "kube-api-access-hjvsd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:33:03 crc kubenswrapper[4988]: I1008 18:33:03.383448 4988 generic.go:334] "Generic (PLEG): container finished" podID="0fcc9499-a06b-4cd1-a5da-ea8108916050" containerID="9315fcc87c2b03972c71e57728c4fdbc3b30200edc121e010b7ac2e7d7658e92" exitCode=0 Oct 08 18:33:03 crc kubenswrapper[4988]: I1008 18:33:03.383533 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 08 18:33:03 crc kubenswrapper[4988]: I1008 18:33:03.386509 4988 generic.go:334] "Generic (PLEG): container finished" podID="85bc7f8f-fdb3-4abb-88d3-ed0c10982c4c" containerID="e18a49c9de8366d161727792956633e70d61d0fc8c2799f046b7a6f94af795e5" exitCode=0 Oct 08 18:33:03 crc kubenswrapper[4988]: I1008 18:33:03.388076 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/novaapiebe4-account-delete-rph2r" Oct 08 18:33:03 crc kubenswrapper[4988]: I1008 18:33:03.390282 4988 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hjvsd\" (UniqueName: \"kubernetes.io/projected/79cbf065-a458-4dec-b6a7-805f515142e0-kube-api-access-hjvsd\") on node \"crc\" DevicePath \"\"" Oct 08 18:33:03 crc kubenswrapper[4988]: I1008 18:33:03.390327 4988 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f4e47608-89b4-45d1-88a6-fe75805d864d-config-data\") on node \"crc\" DevicePath \"\"" Oct 08 18:33:03 crc kubenswrapper[4988]: I1008 18:33:03.390338 4988 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tv2pz\" (UniqueName: \"kubernetes.io/projected/3f7bddd3-86f8-46e9-a38e-1aa5e570d151-kube-api-access-tv2pz\") on node \"crc\" DevicePath \"\"" Oct 08 18:33:03 crc kubenswrapper[4988]: I1008 18:33:03.410061 4988 generic.go:334] "Generic (PLEG): container finished" podID="3503942b-2825-4006-80e8-8c4610b89997" containerID="e1a685c92e7a7b5fad1185fbe08be64ed9f96221d3c3a8c8c5d3364e9a593229" exitCode=0 Oct 08 18:33:03 crc kubenswrapper[4988]: I1008 18:33:03.412206 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican32b1-account-delete-jlhch" Oct 08 18:33:03 crc kubenswrapper[4988]: I1008 18:33:03.412698 4988 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage05-crc" (UniqueName: "kubernetes.io/local-volume/local-storage05-crc") on node "crc" Oct 08 18:33:03 crc kubenswrapper[4988]: I1008 18:33:03.434368 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f4e47608-89b4-45d1-88a6-fe75805d864d-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f4e47608-89b4-45d1-88a6-fe75805d864d" (UID: "f4e47608-89b4-45d1-88a6-fe75805d864d"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:33:03 crc kubenswrapper[4988]: I1008 18:33:03.434472 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/99d77eed-40f0-4879-be96-abea3a807cd9-config-data" (OuterVolumeSpecName: "config-data") pod "99d77eed-40f0-4879-be96-abea3a807cd9" (UID: "99d77eed-40f0-4879-be96-abea3a807cd9"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:33:03 crc kubenswrapper[4988]: I1008 18:33:03.434728 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0fcc9499-a06b-4cd1-a5da-ea8108916050-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "0fcc9499-a06b-4cd1-a5da-ea8108916050" (UID: "0fcc9499-a06b-4cd1-a5da-ea8108916050"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:33:03 crc kubenswrapper[4988]: I1008 18:33:03.435574 4988 generic.go:334] "Generic (PLEG): container finished" podID="3a57537b-6115-4249-8d16-3ab0c4b6d21e" containerID="0f23a2df4066a02929dd3edb8dc8d068809211dbfad9bf36b14e3f21916284e7" exitCode=0 Oct 08 18:33:03 crc kubenswrapper[4988]: I1008 18:33:03.443527 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement79af-account-delete-hv6j5" Oct 08 18:33:03 crc kubenswrapper[4988]: I1008 18:33:03.447702 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinderd8da-account-delete-75j2h" Oct 08 18:33:03 crc kubenswrapper[4988]: I1008 18:33:03.450900 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 08 18:33:03 crc kubenswrapper[4988]: I1008 18:33:03.463899 4988 generic.go:334] "Generic (PLEG): container finished" podID="ccd2f8fa-1c16-4c58-9f87-e60ba3a2f5fb" containerID="f64fdd9df8d4aff7bcd8cfab3af523fa24041d42ea9cfb0addd7bf856026fe76" exitCode=0 Oct 08 18:33:03 crc kubenswrapper[4988]: I1008 18:33:03.463935 4988 generic.go:334] "Generic (PLEG): container finished" podID="ccd2f8fa-1c16-4c58-9f87-e60ba3a2f5fb" containerID="c7ddcc51fc71ff2377802bb6d1c0c0545ab7211828f994c467b39e3b24eaac16" exitCode=0 Oct 08 18:33:03 crc kubenswrapper[4988]: I1008 18:33:03.517706 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6d1becf-4229-40bf-b53a-d8a9eb4b580e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "b6d1becf-4229-40bf-b53a-d8a9eb4b580e" (UID: "b6d1becf-4229-40bf-b53a-d8a9eb4b580e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:33:03 crc kubenswrapper[4988]: I1008 18:33:03.518290 4988 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b6d1becf-4229-40bf-b53a-d8a9eb4b580e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 18:33:03 crc kubenswrapper[4988]: I1008 18:33:03.518315 4988 reconciler_common.go:293] "Volume detached for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") on node \"crc\" DevicePath \"\"" Oct 08 18:33:03 crc kubenswrapper[4988]: I1008 18:33:03.518324 4988 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0fcc9499-a06b-4cd1-a5da-ea8108916050-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 18:33:03 crc kubenswrapper[4988]: I1008 18:33:03.518333 4988 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f4e47608-89b4-45d1-88a6-fe75805d864d-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 18:33:03 crc kubenswrapper[4988]: I1008 18:33:03.518343 4988 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/99d77eed-40f0-4879-be96-abea3a807cd9-config-data\") on node \"crc\" DevicePath \"\"" Oct 08 18:33:03 crc kubenswrapper[4988]: I1008 18:33:03.518632 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Oct 08 18:33:03 crc kubenswrapper[4988]: I1008 18:33:03.523977 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-c44756996-dfw4k" Oct 08 18:33:03 crc kubenswrapper[4988]: I1008 18:33:03.528619 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6d1becf-4229-40bf-b53a-d8a9eb4b580e-kube-state-metrics-tls-certs" (OuterVolumeSpecName: "kube-state-metrics-tls-certs") pod "b6d1becf-4229-40bf-b53a-d8a9eb4b580e" (UID: "b6d1becf-4229-40bf-b53a-d8a9eb4b580e"). InnerVolumeSpecName "kube-state-metrics-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:33:03 crc kubenswrapper[4988]: I1008 18:33:03.533639 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0fcc9499-a06b-4cd1-a5da-ea8108916050-config-data" (OuterVolumeSpecName: "config-data") pod "0fcc9499-a06b-4cd1-a5da-ea8108916050" (UID: "0fcc9499-a06b-4cd1-a5da-ea8108916050"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:33:03 crc kubenswrapper[4988]: I1008 18:33:03.534076 4988 generic.go:334] "Generic (PLEG): container finished" podID="32bb7401-6292-4ca6-98f4-15522125c3ee" containerID="dcd1682ce23dc072a1b6bf3c5aeff64ca69361e18fa5cdcb9712e65170ddc37d" exitCode=0 Oct 08 18:33:03 crc kubenswrapper[4988]: I1008 18:33:03.538108 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6d1becf-4229-40bf-b53a-d8a9eb4b580e-kube-state-metrics-tls-config" (OuterVolumeSpecName: "kube-state-metrics-tls-config") pod "b6d1becf-4229-40bf-b53a-d8a9eb4b580e" (UID: "b6d1becf-4229-40bf-b53a-d8a9eb4b580e"). InnerVolumeSpecName "kube-state-metrics-tls-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:33:03 crc kubenswrapper[4988]: I1008 18:33:03.541547 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0fcc9499-a06b-4cd1-a5da-ea8108916050-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "0fcc9499-a06b-4cd1-a5da-ea8108916050" (UID: "0fcc9499-a06b-4cd1-a5da-ea8108916050"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:33:03 crc kubenswrapper[4988]: I1008 18:33:03.542948 4988 generic.go:334] "Generic (PLEG): container finished" podID="0efd1f53-90ad-4eb8-90f2-2ba9fc1593c2" containerID="a81aba18e5412a571b8ceb649f5e04c6698ca4b21c30f54d96848df1d1a823b6" exitCode=0 Oct 08 18:33:03 crc kubenswrapper[4988]: I1008 18:33:03.563570 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f4e47608-89b4-45d1-88a6-fe75805d864d-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "f4e47608-89b4-45d1-88a6-fe75805d864d" (UID: "f4e47608-89b4-45d1-88a6-fe75805d864d"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:33:03 crc kubenswrapper[4988]: I1008 18:33:03.581126 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f4e47608-89b4-45d1-88a6-fe75805d864d-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "f4e47608-89b4-45d1-88a6-fe75805d864d" (UID: "f4e47608-89b4-45d1-88a6-fe75805d864d"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:33:03 crc kubenswrapper[4988]: I1008 18:33:03.620518 4988 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0fcc9499-a06b-4cd1-a5da-ea8108916050-config-data\") on node \"crc\" DevicePath \"\"" Oct 08 18:33:03 crc kubenswrapper[4988]: I1008 18:33:03.620550 4988 reconciler_common.go:293] "Volume detached for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/b6d1becf-4229-40bf-b53a-d8a9eb4b580e-kube-state-metrics-tls-config\") on node \"crc\" DevicePath \"\"" Oct 08 18:33:03 crc kubenswrapper[4988]: I1008 18:33:03.620560 4988 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/f4e47608-89b4-45d1-88a6-fe75805d864d-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 08 18:33:03 crc kubenswrapper[4988]: I1008 18:33:03.620573 4988 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/0fcc9499-a06b-4cd1-a5da-ea8108916050-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 08 18:33:03 crc kubenswrapper[4988]: I1008 18:33:03.620582 4988 reconciler_common.go:293] "Volume detached for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/b6d1becf-4229-40bf-b53a-d8a9eb4b580e-kube-state-metrics-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 08 18:33:03 crc kubenswrapper[4988]: I1008 18:33:03.620591 4988 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/f4e47608-89b4-45d1-88a6-fe75805d864d-public-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 08 18:33:03 crc kubenswrapper[4988]: I1008 18:33:03.694450 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"0fcc9499-a06b-4cd1-a5da-ea8108916050","Type":"ContainerDied","Data":"9315fcc87c2b03972c71e57728c4fdbc3b30200edc121e010b7ac2e7d7658e92"} Oct 08 18:33:03 crc kubenswrapper[4988]: I1008 18:33:03.694490 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"0fcc9499-a06b-4cd1-a5da-ea8108916050","Type":"ContainerDied","Data":"13b42e11464da2a16b26f60e40ed4b477086a4e26e88e8184281465eec0c04c9"} Oct 08 18:33:03 crc kubenswrapper[4988]: I1008 18:33:03.694501 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"85bc7f8f-fdb3-4abb-88d3-ed0c10982c4c","Type":"ContainerDied","Data":"e18a49c9de8366d161727792956633e70d61d0fc8c2799f046b7a6f94af795e5"} Oct 08 18:33:03 crc kubenswrapper[4988]: I1008 18:33:03.694518 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/novaapiebe4-account-delete-rph2r" event={"ID":"28f73b7a-bb5e-43ec-a099-13a12cdb33d3","Type":"ContainerDied","Data":"1b85ba9999219bc7a7a3b6a0ce7ce2da865b368c7d4d19a8fb8c6910e624f611"} Oct 08 18:33:03 crc kubenswrapper[4988]: I1008 18:33:03.694530 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"3503942b-2825-4006-80e8-8c4610b89997","Type":"ContainerDied","Data":"e1a685c92e7a7b5fad1185fbe08be64ed9f96221d3c3a8c8c5d3364e9a593229"} Oct 08 18:33:03 crc kubenswrapper[4988]: I1008 18:33:03.694545 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican32b1-account-delete-jlhch" event={"ID":"3f7bddd3-86f8-46e9-a38e-1aa5e570d151","Type":"ContainerDied","Data":"9d5c29e3c8578e502a01d1e1800e01d52a4abda618cf7cabb3fdf8627cb9f172"} Oct 08 18:33:03 crc kubenswrapper[4988]: I1008 18:33:03.694555 4988 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9d5c29e3c8578e502a01d1e1800e01d52a4abda618cf7cabb3fdf8627cb9f172" Oct 08 18:33:03 crc kubenswrapper[4988]: I1008 18:33:03.694563 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-6c486d5744-j86zn" event={"ID":"3a57537b-6115-4249-8d16-3ab0c4b6d21e","Type":"ContainerDied","Data":"0f23a2df4066a02929dd3edb8dc8d068809211dbfad9bf36b14e3f21916284e7"} Oct 08 18:33:03 crc kubenswrapper[4988]: I1008 18:33:03.694577 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-6c486d5744-j86zn" event={"ID":"3a57537b-6115-4249-8d16-3ab0c4b6d21e","Type":"ContainerDied","Data":"fe2b1b18022fcd542d4117b8811974fd4ee1907095a986d6338c7dd0db4ca3b9"} Oct 08 18:33:03 crc kubenswrapper[4988]: I1008 18:33:03.694585 4988 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fe2b1b18022fcd542d4117b8811974fd4ee1907095a986d6338c7dd0db4ca3b9" Oct 08 18:33:03 crc kubenswrapper[4988]: I1008 18:33:03.694594 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement79af-account-delete-hv6j5" event={"ID":"79cbf065-a458-4dec-b6a7-805f515142e0","Type":"ContainerDied","Data":"b89945bf77210eeb7ef560ff52dfcfbca4649687437b5a7f6440340b4c6d51be"} Oct 08 18:33:03 crc kubenswrapper[4988]: I1008 18:33:03.694602 4988 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b89945bf77210eeb7ef560ff52dfcfbca4649687437b5a7f6440340b4c6d51be" Oct 08 18:33:03 crc kubenswrapper[4988]: I1008 18:33:03.694614 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinderd8da-account-delete-75j2h" event={"ID":"02a37054-4817-4151-a8cb-a9a45867c0d3","Type":"ContainerDied","Data":"bb9a6ad55acd84c2a6010e95ec046d90fe6e67eadd6cc37cb135bacfd262aa56"} Oct 08 18:33:03 crc kubenswrapper[4988]: I1008 18:33:03.694624 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"99d77eed-40f0-4879-be96-abea3a807cd9","Type":"ContainerDied","Data":"26ea215b02c8f513e1a65b6a89053fad73394e2112598a6fe6ab5ff048ac368f"} Oct 08 18:33:03 crc kubenswrapper[4988]: I1008 18:33:03.694637 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ccd2f8fa-1c16-4c58-9f87-e60ba3a2f5fb","Type":"ContainerDied","Data":"f64fdd9df8d4aff7bcd8cfab3af523fa24041d42ea9cfb0addd7bf856026fe76"} Oct 08 18:33:03 crc kubenswrapper[4988]: I1008 18:33:03.694647 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ccd2f8fa-1c16-4c58-9f87-e60ba3a2f5fb","Type":"ContainerDied","Data":"c7ddcc51fc71ff2377802bb6d1c0c0545ab7211828f994c467b39e3b24eaac16"} Oct 08 18:33:03 crc kubenswrapper[4988]: I1008 18:33:03.694658 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"b6d1becf-4229-40bf-b53a-d8a9eb4b580e","Type":"ContainerDied","Data":"f064454d5dc1aaecb4c0413e7e223abb0b5fa88fbea49db04428e68ae3072ec7"} Oct 08 18:33:03 crc kubenswrapper[4988]: I1008 18:33:03.694670 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-c44756996-dfw4k" event={"ID":"f4e47608-89b4-45d1-88a6-fe75805d864d","Type":"ContainerDied","Data":"8cb906ede5c077bda5980c9dc0e8b3c6324dd51624529cfd6fe9c32c2ad8c301"} Oct 08 18:33:03 crc kubenswrapper[4988]: I1008 18:33:03.694680 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"32bb7401-6292-4ca6-98f4-15522125c3ee","Type":"ContainerDied","Data":"dcd1682ce23dc072a1b6bf3c5aeff64ca69361e18fa5cdcb9712e65170ddc37d"} Oct 08 18:33:03 crc kubenswrapper[4988]: I1008 18:33:03.694691 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/novacell02f70-account-delete-cdwsx" event={"ID":"3e8c874a-029a-4d6a-bc2e-00aeed1636a7","Type":"ContainerDied","Data":"cf5297609526e130b9edf694145c045de83390235b5b462b974086d31336cf2d"} Oct 08 18:33:03 crc kubenswrapper[4988]: I1008 18:33:03.694701 4988 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="cf5297609526e130b9edf694145c045de83390235b5b462b974086d31336cf2d" Oct 08 18:33:03 crc kubenswrapper[4988]: I1008 18:33:03.694711 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"0efd1f53-90ad-4eb8-90f2-2ba9fc1593c2","Type":"ContainerDied","Data":"a81aba18e5412a571b8ceb649f5e04c6698ca4b21c30f54d96848df1d1a823b6"} Oct 08 18:33:03 crc kubenswrapper[4988]: I1008 18:33:03.719770 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/novacell02f70-account-delete-cdwsx" Oct 08 18:33:03 crc kubenswrapper[4988]: I1008 18:33:03.721300 4988 scope.go:117] "RemoveContainer" containerID="42e44fdf6315337e912e8f08c104630a1408f1ece9b5a29dc6c2ddc4b1e15bf2" Oct 08 18:33:03 crc kubenswrapper[4988]: I1008 18:33:03.822798 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-f6q49\" (UniqueName: \"kubernetes.io/projected/3e8c874a-029a-4d6a-bc2e-00aeed1636a7-kube-api-access-f6q49\") pod \"3e8c874a-029a-4d6a-bc2e-00aeed1636a7\" (UID: \"3e8c874a-029a-4d6a-bc2e-00aeed1636a7\") " Oct 08 18:33:03 crc kubenswrapper[4988]: I1008 18:33:03.823196 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k8mnm\" (UniqueName: \"kubernetes.io/projected/11679d40-6eb6-456c-8996-cd9a946a6db1-kube-api-access-k8mnm\") pod \"keystone3e59-account-delete-vgpc9\" (UID: \"11679d40-6eb6-456c-8996-cd9a946a6db1\") " pod="openstack/keystone3e59-account-delete-vgpc9" Oct 08 18:33:03 crc kubenswrapper[4988]: E1008 18:33:03.828250 4988 projected.go:194] Error preparing data for projected volume kube-api-access-k8mnm for pod openstack/keystone3e59-account-delete-vgpc9: failed to fetch token: serviceaccounts "galera-openstack" not found Oct 08 18:33:03 crc kubenswrapper[4988]: E1008 18:33:03.828539 4988 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/11679d40-6eb6-456c-8996-cd9a946a6db1-kube-api-access-k8mnm podName:11679d40-6eb6-456c-8996-cd9a946a6db1 nodeName:}" failed. No retries permitted until 2025-10-08 18:33:05.828520229 +0000 UTC m=+1331.278362999 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-k8mnm" (UniqueName: "kubernetes.io/projected/11679d40-6eb6-456c-8996-cd9a946a6db1-kube-api-access-k8mnm") pod "keystone3e59-account-delete-vgpc9" (UID: "11679d40-6eb6-456c-8996-cd9a946a6db1") : failed to fetch token: serviceaccounts "galera-openstack" not found Oct 08 18:33:03 crc kubenswrapper[4988]: I1008 18:33:03.830113 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3e8c874a-029a-4d6a-bc2e-00aeed1636a7-kube-api-access-f6q49" (OuterVolumeSpecName: "kube-api-access-f6q49") pod "3e8c874a-029a-4d6a-bc2e-00aeed1636a7" (UID: "3e8c874a-029a-4d6a-bc2e-00aeed1636a7"). InnerVolumeSpecName "kube-api-access-f6q49". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:33:03 crc kubenswrapper[4988]: I1008 18:33:03.931651 4988 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-f6q49\" (UniqueName: \"kubernetes.io/projected/3e8c874a-029a-4d6a-bc2e-00aeed1636a7-kube-api-access-f6q49\") on node \"crc\" DevicePath \"\"" Oct 08 18:33:04 crc kubenswrapper[4988]: I1008 18:33:04.075272 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-6c486d5744-j86zn" Oct 08 18:33:04 crc kubenswrapper[4988]: I1008 18:33:04.105169 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 08 18:33:04 crc kubenswrapper[4988]: I1008 18:33:04.121088 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinderd8da-account-delete-75j2h"] Oct 08 18:33:04 crc kubenswrapper[4988]: I1008 18:33:04.145637 4988 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinderd8da-account-delete-75j2h"] Oct 08 18:33:04 crc kubenswrapper[4988]: I1008 18:33:04.148980 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 08 18:33:04 crc kubenswrapper[4988]: I1008 18:33:04.154023 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Oct 08 18:33:04 crc kubenswrapper[4988]: I1008 18:33:04.160134 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 08 18:33:04 crc kubenswrapper[4988]: I1008 18:33:04.164757 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican32b1-account-delete-jlhch"] Oct 08 18:33:04 crc kubenswrapper[4988]: I1008 18:33:04.177469 4988 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican32b1-account-delete-jlhch"] Oct 08 18:33:04 crc kubenswrapper[4988]: I1008 18:33:04.177729 4988 scope.go:117] "RemoveContainer" containerID="ee0c93d867a043a6ace66d0ccc23af52eb47335f2734cbb64f7f04eb41db6e3c" Oct 08 18:33:04 crc kubenswrapper[4988]: E1008 18:33:04.178602 4988 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ee0c93d867a043a6ace66d0ccc23af52eb47335f2734cbb64f7f04eb41db6e3c\": container with ID starting with ee0c93d867a043a6ace66d0ccc23af52eb47335f2734cbb64f7f04eb41db6e3c not found: ID does not exist" containerID="ee0c93d867a043a6ace66d0ccc23af52eb47335f2734cbb64f7f04eb41db6e3c" Oct 08 18:33:04 crc kubenswrapper[4988]: I1008 18:33:04.178645 4988 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ee0c93d867a043a6ace66d0ccc23af52eb47335f2734cbb64f7f04eb41db6e3c"} err="failed to get container status \"ee0c93d867a043a6ace66d0ccc23af52eb47335f2734cbb64f7f04eb41db6e3c\": rpc error: code = NotFound desc = could not find container \"ee0c93d867a043a6ace66d0ccc23af52eb47335f2734cbb64f7f04eb41db6e3c\": container with ID starting with ee0c93d867a043a6ace66d0ccc23af52eb47335f2734cbb64f7f04eb41db6e3c not found: ID does not exist" Oct 08 18:33:04 crc kubenswrapper[4988]: I1008 18:33:04.178675 4988 scope.go:117] "RemoveContainer" containerID="42e44fdf6315337e912e8f08c104630a1408f1ece9b5a29dc6c2ddc4b1e15bf2" Oct 08 18:33:04 crc kubenswrapper[4988]: E1008 18:33:04.183849 4988 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"42e44fdf6315337e912e8f08c104630a1408f1ece9b5a29dc6c2ddc4b1e15bf2\": container with ID starting with 42e44fdf6315337e912e8f08c104630a1408f1ece9b5a29dc6c2ddc4b1e15bf2 not found: ID does not exist" containerID="42e44fdf6315337e912e8f08c104630a1408f1ece9b5a29dc6c2ddc4b1e15bf2" Oct 08 18:33:04 crc kubenswrapper[4988]: I1008 18:33:04.183902 4988 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"42e44fdf6315337e912e8f08c104630a1408f1ece9b5a29dc6c2ddc4b1e15bf2"} err="failed to get container status \"42e44fdf6315337e912e8f08c104630a1408f1ece9b5a29dc6c2ddc4b1e15bf2\": rpc error: code = NotFound desc = could not find container \"42e44fdf6315337e912e8f08c104630a1408f1ece9b5a29dc6c2ddc4b1e15bf2\": container with ID starting with 42e44fdf6315337e912e8f08c104630a1408f1ece9b5a29dc6c2ddc4b1e15bf2 not found: ID does not exist" Oct 08 18:33:04 crc kubenswrapper[4988]: I1008 18:33:04.183935 4988 scope.go:117] "RemoveContainer" containerID="1f77c788702456fc178f26b1d4a4449a94e1ac2f3f33ad505531b3885e2f8f15" Oct 08 18:33:04 crc kubenswrapper[4988]: I1008 18:33:04.190469 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/novaapiebe4-account-delete-rph2r"] Oct 08 18:33:04 crc kubenswrapper[4988]: I1008 18:33:04.196323 4988 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/novaapiebe4-account-delete-rph2r"] Oct 08 18:33:04 crc kubenswrapper[4988]: E1008 18:33:04.202891 4988 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 863aa8ddeda792a78ba5af321605974ac7082d24fc8fb6f0400148fcf3f06126 is running failed: container process not found" containerID="863aa8ddeda792a78ba5af321605974ac7082d24fc8fb6f0400148fcf3f06126" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Oct 08 18:33:04 crc kubenswrapper[4988]: E1008 18:33:04.204648 4988 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 863aa8ddeda792a78ba5af321605974ac7082d24fc8fb6f0400148fcf3f06126 is running failed: container process not found" containerID="863aa8ddeda792a78ba5af321605974ac7082d24fc8fb6f0400148fcf3f06126" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Oct 08 18:33:04 crc kubenswrapper[4988]: E1008 18:33:04.205440 4988 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 863aa8ddeda792a78ba5af321605974ac7082d24fc8fb6f0400148fcf3f06126 is running failed: container process not found" containerID="863aa8ddeda792a78ba5af321605974ac7082d24fc8fb6f0400148fcf3f06126" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Oct 08 18:33:04 crc kubenswrapper[4988]: E1008 18:33:04.205543 4988 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 863aa8ddeda792a78ba5af321605974ac7082d24fc8fb6f0400148fcf3f06126 is running failed: container process not found" probeType="Readiness" pod="openstack/ovn-controller-ovs-kc7fw" podUID="f1628013-e0b9-4c31-80c0-91aabaaef0f6" containerName="ovsdb-server" Oct 08 18:33:04 crc kubenswrapper[4988]: E1008 18:33:04.206283 4988 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="8a5332d3d9aa50bb56f12e04b1b32520399363308f9e75d5510ddd01f41593c3" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Oct 08 18:33:04 crc kubenswrapper[4988]: I1008 18:33:04.207025 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Oct 08 18:33:04 crc kubenswrapper[4988]: E1008 18:33:04.210200 4988 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="8a5332d3d9aa50bb56f12e04b1b32520399363308f9e75d5510ddd01f41593c3" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Oct 08 18:33:04 crc kubenswrapper[4988]: I1008 18:33:04.212420 4988 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-api-0"] Oct 08 18:33:04 crc kubenswrapper[4988]: E1008 18:33:04.212805 4988 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="8a5332d3d9aa50bb56f12e04b1b32520399363308f9e75d5510ddd01f41593c3" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Oct 08 18:33:04 crc kubenswrapper[4988]: E1008 18:33:04.212859 4988 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/ovn-controller-ovs-kc7fw" podUID="f1628013-e0b9-4c31-80c0-91aabaaef0f6" containerName="ovs-vswitchd" Oct 08 18:33:04 crc kubenswrapper[4988]: I1008 18:33:04.218005 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 08 18:33:04 crc kubenswrapper[4988]: I1008 18:33:04.228670 4988 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 08 18:33:04 crc kubenswrapper[4988]: I1008 18:33:04.238845 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bvpsb\" (UniqueName: \"kubernetes.io/projected/0efd1f53-90ad-4eb8-90f2-2ba9fc1593c2-kube-api-access-bvpsb\") pod \"0efd1f53-90ad-4eb8-90f2-2ba9fc1593c2\" (UID: \"0efd1f53-90ad-4eb8-90f2-2ba9fc1593c2\") " Oct 08 18:33:04 crc kubenswrapper[4988]: I1008 18:33:04.238888 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3a57537b-6115-4249-8d16-3ab0c4b6d21e-logs\") pod \"3a57537b-6115-4249-8d16-3ab0c4b6d21e\" (UID: \"3a57537b-6115-4249-8d16-3ab0c4b6d21e\") " Oct 08 18:33:04 crc kubenswrapper[4988]: I1008 18:33:04.238961 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3a57537b-6115-4249-8d16-3ab0c4b6d21e-config-data\") pod \"3a57537b-6115-4249-8d16-3ab0c4b6d21e\" (UID: \"3a57537b-6115-4249-8d16-3ab0c4b6d21e\") " Oct 08 18:33:04 crc kubenswrapper[4988]: I1008 18:33:04.238985 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/3a57537b-6115-4249-8d16-3ab0c4b6d21e-internal-tls-certs\") pod \"3a57537b-6115-4249-8d16-3ab0c4b6d21e\" (UID: \"3a57537b-6115-4249-8d16-3ab0c4b6d21e\") " Oct 08 18:33:04 crc kubenswrapper[4988]: I1008 18:33:04.239003 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/0efd1f53-90ad-4eb8-90f2-2ba9fc1593c2-public-tls-certs\") pod \"0efd1f53-90ad-4eb8-90f2-2ba9fc1593c2\" (UID: \"0efd1f53-90ad-4eb8-90f2-2ba9fc1593c2\") " Oct 08 18:33:04 crc kubenswrapper[4988]: I1008 18:33:04.239030 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/3a57537b-6115-4249-8d16-3ab0c4b6d21e-public-tls-certs\") pod \"3a57537b-6115-4249-8d16-3ab0c4b6d21e\" (UID: \"3a57537b-6115-4249-8d16-3ab0c4b6d21e\") " Oct 08 18:33:04 crc kubenswrapper[4988]: I1008 18:33:04.239072 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0efd1f53-90ad-4eb8-90f2-2ba9fc1593c2-config-data\") pod \"0efd1f53-90ad-4eb8-90f2-2ba9fc1593c2\" (UID: \"0efd1f53-90ad-4eb8-90f2-2ba9fc1593c2\") " Oct 08 18:33:04 crc kubenswrapper[4988]: I1008 18:33:04.239097 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3a57537b-6115-4249-8d16-3ab0c4b6d21e-combined-ca-bundle\") pod \"3a57537b-6115-4249-8d16-3ab0c4b6d21e\" (UID: \"3a57537b-6115-4249-8d16-3ab0c4b6d21e\") " Oct 08 18:33:04 crc kubenswrapper[4988]: I1008 18:33:04.239131 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0efd1f53-90ad-4eb8-90f2-2ba9fc1593c2-logs\") pod \"0efd1f53-90ad-4eb8-90f2-2ba9fc1593c2\" (UID: \"0efd1f53-90ad-4eb8-90f2-2ba9fc1593c2\") " Oct 08 18:33:04 crc kubenswrapper[4988]: I1008 18:33:04.239165 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gfb6j\" (UniqueName: \"kubernetes.io/projected/3a57537b-6115-4249-8d16-3ab0c4b6d21e-kube-api-access-gfb6j\") pod \"3a57537b-6115-4249-8d16-3ab0c4b6d21e\" (UID: \"3a57537b-6115-4249-8d16-3ab0c4b6d21e\") " Oct 08 18:33:04 crc kubenswrapper[4988]: I1008 18:33:04.239232 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0efd1f53-90ad-4eb8-90f2-2ba9fc1593c2-combined-ca-bundle\") pod \"0efd1f53-90ad-4eb8-90f2-2ba9fc1593c2\" (UID: \"0efd1f53-90ad-4eb8-90f2-2ba9fc1593c2\") " Oct 08 18:33:04 crc kubenswrapper[4988]: I1008 18:33:04.239248 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/0efd1f53-90ad-4eb8-90f2-2ba9fc1593c2-internal-tls-certs\") pod \"0efd1f53-90ad-4eb8-90f2-2ba9fc1593c2\" (UID: \"0efd1f53-90ad-4eb8-90f2-2ba9fc1593c2\") " Oct 08 18:33:04 crc kubenswrapper[4988]: I1008 18:33:04.239270 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/3a57537b-6115-4249-8d16-3ab0c4b6d21e-config-data-custom\") pod \"3a57537b-6115-4249-8d16-3ab0c4b6d21e\" (UID: \"3a57537b-6115-4249-8d16-3ab0c4b6d21e\") " Oct 08 18:33:04 crc kubenswrapper[4988]: I1008 18:33:04.240840 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3a57537b-6115-4249-8d16-3ab0c4b6d21e-logs" (OuterVolumeSpecName: "logs") pod "3a57537b-6115-4249-8d16-3ab0c4b6d21e" (UID: "3a57537b-6115-4249-8d16-3ab0c4b6d21e"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 18:33:04 crc kubenswrapper[4988]: I1008 18:33:04.242247 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0efd1f53-90ad-4eb8-90f2-2ba9fc1593c2-kube-api-access-bvpsb" (OuterVolumeSpecName: "kube-api-access-bvpsb") pod "0efd1f53-90ad-4eb8-90f2-2ba9fc1593c2" (UID: "0efd1f53-90ad-4eb8-90f2-2ba9fc1593c2"). InnerVolumeSpecName "kube-api-access-bvpsb". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:33:04 crc kubenswrapper[4988]: I1008 18:33:04.244916 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0efd1f53-90ad-4eb8-90f2-2ba9fc1593c2-logs" (OuterVolumeSpecName: "logs") pod "0efd1f53-90ad-4eb8-90f2-2ba9fc1593c2" (UID: "0efd1f53-90ad-4eb8-90f2-2ba9fc1593c2"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 18:33:04 crc kubenswrapper[4988]: I1008 18:33:04.252453 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-c44756996-dfw4k"] Oct 08 18:33:04 crc kubenswrapper[4988]: I1008 18:33:04.252515 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3a57537b-6115-4249-8d16-3ab0c4b6d21e-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "3a57537b-6115-4249-8d16-3ab0c4b6d21e" (UID: "3a57537b-6115-4249-8d16-3ab0c4b6d21e"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:33:04 crc kubenswrapper[4988]: I1008 18:33:04.254638 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3a57537b-6115-4249-8d16-3ab0c4b6d21e-kube-api-access-gfb6j" (OuterVolumeSpecName: "kube-api-access-gfb6j") pod "3a57537b-6115-4249-8d16-3ab0c4b6d21e" (UID: "3a57537b-6115-4249-8d16-3ab0c4b6d21e"). InnerVolumeSpecName "kube-api-access-gfb6j". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:33:04 crc kubenswrapper[4988]: I1008 18:33:04.276205 4988 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-c44756996-dfw4k"] Oct 08 18:33:04 crc kubenswrapper[4988]: I1008 18:33:04.288505 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement79af-account-delete-hv6j5"] Oct 08 18:33:04 crc kubenswrapper[4988]: I1008 18:33:04.294106 4988 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement79af-account-delete-hv6j5"] Oct 08 18:33:04 crc kubenswrapper[4988]: I1008 18:33:04.317711 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3a57537b-6115-4249-8d16-3ab0c4b6d21e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "3a57537b-6115-4249-8d16-3ab0c4b6d21e" (UID: "3a57537b-6115-4249-8d16-3ab0c4b6d21e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:33:04 crc kubenswrapper[4988]: I1008 18:33:04.319306 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0efd1f53-90ad-4eb8-90f2-2ba9fc1593c2-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "0efd1f53-90ad-4eb8-90f2-2ba9fc1593c2" (UID: "0efd1f53-90ad-4eb8-90f2-2ba9fc1593c2"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:33:04 crc kubenswrapper[4988]: I1008 18:33:04.319378 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0efd1f53-90ad-4eb8-90f2-2ba9fc1593c2-config-data" (OuterVolumeSpecName: "config-data") pod "0efd1f53-90ad-4eb8-90f2-2ba9fc1593c2" (UID: "0efd1f53-90ad-4eb8-90f2-2ba9fc1593c2"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:33:04 crc kubenswrapper[4988]: I1008 18:33:04.321006 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0efd1f53-90ad-4eb8-90f2-2ba9fc1593c2-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "0efd1f53-90ad-4eb8-90f2-2ba9fc1593c2" (UID: "0efd1f53-90ad-4eb8-90f2-2ba9fc1593c2"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:33:04 crc kubenswrapper[4988]: I1008 18:33:04.326038 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0efd1f53-90ad-4eb8-90f2-2ba9fc1593c2-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "0efd1f53-90ad-4eb8-90f2-2ba9fc1593c2" (UID: "0efd1f53-90ad-4eb8-90f2-2ba9fc1593c2"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:33:04 crc kubenswrapper[4988]: I1008 18:33:04.326226 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3a57537b-6115-4249-8d16-3ab0c4b6d21e-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "3a57537b-6115-4249-8d16-3ab0c4b6d21e" (UID: "3a57537b-6115-4249-8d16-3ab0c4b6d21e"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:33:04 crc kubenswrapper[4988]: I1008 18:33:04.342086 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3a57537b-6115-4249-8d16-3ab0c4b6d21e-config-data" (OuterVolumeSpecName: "config-data") pod "3a57537b-6115-4249-8d16-3ab0c4b6d21e" (UID: "3a57537b-6115-4249-8d16-3ab0c4b6d21e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:33:04 crc kubenswrapper[4988]: I1008 18:33:04.342546 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/56702a03-175d-4a54-b84f-e25cab037431-httpd-run\") pod \"56702a03-175d-4a54-b84f-e25cab037431\" (UID: \"56702a03-175d-4a54-b84f-e25cab037431\") " Oct 08 18:33:04 crc kubenswrapper[4988]: I1008 18:33:04.342573 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/32bb7401-6292-4ca6-98f4-15522125c3ee-combined-ca-bundle\") pod \"32bb7401-6292-4ca6-98f4-15522125c3ee\" (UID: \"32bb7401-6292-4ca6-98f4-15522125c3ee\") " Oct 08 18:33:04 crc kubenswrapper[4988]: I1008 18:33:04.342609 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n9xkg\" (UniqueName: \"kubernetes.io/projected/32bb7401-6292-4ca6-98f4-15522125c3ee-kube-api-access-n9xkg\") pod \"32bb7401-6292-4ca6-98f4-15522125c3ee\" (UID: \"32bb7401-6292-4ca6-98f4-15522125c3ee\") " Oct 08 18:33:04 crc kubenswrapper[4988]: I1008 18:33:04.342654 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/56702a03-175d-4a54-b84f-e25cab037431-scripts\") pod \"56702a03-175d-4a54-b84f-e25cab037431\" (UID: \"56702a03-175d-4a54-b84f-e25cab037431\") " Oct 08 18:33:04 crc kubenswrapper[4988]: I1008 18:33:04.342674 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/32bb7401-6292-4ca6-98f4-15522125c3ee-memcached-tls-certs\") pod \"32bb7401-6292-4ca6-98f4-15522125c3ee\" (UID: \"32bb7401-6292-4ca6-98f4-15522125c3ee\") " Oct 08 18:33:04 crc kubenswrapper[4988]: I1008 18:33:04.342694 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/32bb7401-6292-4ca6-98f4-15522125c3ee-kolla-config\") pod \"32bb7401-6292-4ca6-98f4-15522125c3ee\" (UID: \"32bb7401-6292-4ca6-98f4-15522125c3ee\") " Oct 08 18:33:04 crc kubenswrapper[4988]: I1008 18:33:04.342712 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"56702a03-175d-4a54-b84f-e25cab037431\" (UID: \"56702a03-175d-4a54-b84f-e25cab037431\") " Oct 08 18:33:04 crc kubenswrapper[4988]: I1008 18:33:04.342758 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/32bb7401-6292-4ca6-98f4-15522125c3ee-config-data\") pod \"32bb7401-6292-4ca6-98f4-15522125c3ee\" (UID: \"32bb7401-6292-4ca6-98f4-15522125c3ee\") " Oct 08 18:33:04 crc kubenswrapper[4988]: I1008 18:33:04.342779 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/83ef7610-23ac-4445-b75c-ae7206b4d810-config-data\") pod \"83ef7610-23ac-4445-b75c-ae7206b4d810\" (UID: \"83ef7610-23ac-4445-b75c-ae7206b4d810\") " Oct 08 18:33:04 crc kubenswrapper[4988]: I1008 18:33:04.342822 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/56702a03-175d-4a54-b84f-e25cab037431-public-tls-certs\") pod \"56702a03-175d-4a54-b84f-e25cab037431\" (UID: \"56702a03-175d-4a54-b84f-e25cab037431\") " Oct 08 18:33:04 crc kubenswrapper[4988]: I1008 18:33:04.343024 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/83ef7610-23ac-4445-b75c-ae7206b4d810-combined-ca-bundle\") pod \"83ef7610-23ac-4445-b75c-ae7206b4d810\" (UID: \"83ef7610-23ac-4445-b75c-ae7206b4d810\") " Oct 08 18:33:04 crc kubenswrapper[4988]: I1008 18:33:04.343042 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lvrb9\" (UniqueName: \"kubernetes.io/projected/83ef7610-23ac-4445-b75c-ae7206b4d810-kube-api-access-lvrb9\") pod \"83ef7610-23ac-4445-b75c-ae7206b4d810\" (UID: \"83ef7610-23ac-4445-b75c-ae7206b4d810\") " Oct 08 18:33:04 crc kubenswrapper[4988]: I1008 18:33:04.343061 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-g74gb\" (UniqueName: \"kubernetes.io/projected/56702a03-175d-4a54-b84f-e25cab037431-kube-api-access-g74gb\") pod \"56702a03-175d-4a54-b84f-e25cab037431\" (UID: \"56702a03-175d-4a54-b84f-e25cab037431\") " Oct 08 18:33:04 crc kubenswrapper[4988]: I1008 18:33:04.343097 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/83ef7610-23ac-4445-b75c-ae7206b4d810-nova-metadata-tls-certs\") pod \"83ef7610-23ac-4445-b75c-ae7206b4d810\" (UID: \"83ef7610-23ac-4445-b75c-ae7206b4d810\") " Oct 08 18:33:04 crc kubenswrapper[4988]: I1008 18:33:04.343116 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/56702a03-175d-4a54-b84f-e25cab037431-combined-ca-bundle\") pod \"56702a03-175d-4a54-b84f-e25cab037431\" (UID: \"56702a03-175d-4a54-b84f-e25cab037431\") " Oct 08 18:33:04 crc kubenswrapper[4988]: I1008 18:33:04.343130 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/83ef7610-23ac-4445-b75c-ae7206b4d810-logs\") pod \"83ef7610-23ac-4445-b75c-ae7206b4d810\" (UID: \"83ef7610-23ac-4445-b75c-ae7206b4d810\") " Oct 08 18:33:04 crc kubenswrapper[4988]: I1008 18:33:04.343154 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/56702a03-175d-4a54-b84f-e25cab037431-config-data\") pod \"56702a03-175d-4a54-b84f-e25cab037431\" (UID: \"56702a03-175d-4a54-b84f-e25cab037431\") " Oct 08 18:33:04 crc kubenswrapper[4988]: I1008 18:33:04.343173 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/56702a03-175d-4a54-b84f-e25cab037431-logs\") pod \"56702a03-175d-4a54-b84f-e25cab037431\" (UID: \"56702a03-175d-4a54-b84f-e25cab037431\") " Oct 08 18:33:04 crc kubenswrapper[4988]: I1008 18:33:04.343588 4988 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0efd1f53-90ad-4eb8-90f2-2ba9fc1593c2-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 18:33:04 crc kubenswrapper[4988]: I1008 18:33:04.343608 4988 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/0efd1f53-90ad-4eb8-90f2-2ba9fc1593c2-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 08 18:33:04 crc kubenswrapper[4988]: I1008 18:33:04.343618 4988 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/3a57537b-6115-4249-8d16-3ab0c4b6d21e-config-data-custom\") on node \"crc\" DevicePath \"\"" Oct 08 18:33:04 crc kubenswrapper[4988]: I1008 18:33:04.343628 4988 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bvpsb\" (UniqueName: \"kubernetes.io/projected/0efd1f53-90ad-4eb8-90f2-2ba9fc1593c2-kube-api-access-bvpsb\") on node \"crc\" DevicePath \"\"" Oct 08 18:33:04 crc kubenswrapper[4988]: I1008 18:33:04.343638 4988 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3a57537b-6115-4249-8d16-3ab0c4b6d21e-logs\") on node \"crc\" DevicePath \"\"" Oct 08 18:33:04 crc kubenswrapper[4988]: I1008 18:33:04.343646 4988 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3a57537b-6115-4249-8d16-3ab0c4b6d21e-config-data\") on node \"crc\" DevicePath \"\"" Oct 08 18:33:04 crc kubenswrapper[4988]: I1008 18:33:04.343654 4988 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/0efd1f53-90ad-4eb8-90f2-2ba9fc1593c2-public-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 08 18:33:04 crc kubenswrapper[4988]: I1008 18:33:04.343663 4988 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/3a57537b-6115-4249-8d16-3ab0c4b6d21e-public-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 08 18:33:04 crc kubenswrapper[4988]: I1008 18:33:04.343672 4988 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0efd1f53-90ad-4eb8-90f2-2ba9fc1593c2-config-data\") on node \"crc\" DevicePath \"\"" Oct 08 18:33:04 crc kubenswrapper[4988]: I1008 18:33:04.343682 4988 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3a57537b-6115-4249-8d16-3ab0c4b6d21e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 18:33:04 crc kubenswrapper[4988]: I1008 18:33:04.343690 4988 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0efd1f53-90ad-4eb8-90f2-2ba9fc1593c2-logs\") on node \"crc\" DevicePath \"\"" Oct 08 18:33:04 crc kubenswrapper[4988]: I1008 18:33:04.343699 4988 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gfb6j\" (UniqueName: \"kubernetes.io/projected/3a57537b-6115-4249-8d16-3ab0c4b6d21e-kube-api-access-gfb6j\") on node \"crc\" DevicePath \"\"" Oct 08 18:33:04 crc kubenswrapper[4988]: I1008 18:33:04.345835 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/83ef7610-23ac-4445-b75c-ae7206b4d810-logs" (OuterVolumeSpecName: "logs") pod "83ef7610-23ac-4445-b75c-ae7206b4d810" (UID: "83ef7610-23ac-4445-b75c-ae7206b4d810"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 18:33:04 crc kubenswrapper[4988]: I1008 18:33:04.346232 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/56702a03-175d-4a54-b84f-e25cab037431-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "56702a03-175d-4a54-b84f-e25cab037431" (UID: "56702a03-175d-4a54-b84f-e25cab037431"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 18:33:04 crc kubenswrapper[4988]: I1008 18:33:04.350132 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/56702a03-175d-4a54-b84f-e25cab037431-logs" (OuterVolumeSpecName: "logs") pod "56702a03-175d-4a54-b84f-e25cab037431" (UID: "56702a03-175d-4a54-b84f-e25cab037431"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 18:33:04 crc kubenswrapper[4988]: I1008 18:33:04.350285 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3a57537b-6115-4249-8d16-3ab0c4b6d21e-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "3a57537b-6115-4249-8d16-3ab0c4b6d21e" (UID: "3a57537b-6115-4249-8d16-3ab0c4b6d21e"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:33:04 crc kubenswrapper[4988]: I1008 18:33:04.353846 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/32bb7401-6292-4ca6-98f4-15522125c3ee-config-data" (OuterVolumeSpecName: "config-data") pod "32bb7401-6292-4ca6-98f4-15522125c3ee" (UID: "32bb7401-6292-4ca6-98f4-15522125c3ee"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 18:33:04 crc kubenswrapper[4988]: I1008 18:33:04.354550 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/32bb7401-6292-4ca6-98f4-15522125c3ee-kolla-config" (OuterVolumeSpecName: "kolla-config") pod "32bb7401-6292-4ca6-98f4-15522125c3ee" (UID: "32bb7401-6292-4ca6-98f4-15522125c3ee"). InnerVolumeSpecName "kolla-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 18:33:04 crc kubenswrapper[4988]: I1008 18:33:04.362934 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/56702a03-175d-4a54-b84f-e25cab037431-kube-api-access-g74gb" (OuterVolumeSpecName: "kube-api-access-g74gb") pod "56702a03-175d-4a54-b84f-e25cab037431" (UID: "56702a03-175d-4a54-b84f-e25cab037431"). InnerVolumeSpecName "kube-api-access-g74gb". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:33:04 crc kubenswrapper[4988]: I1008 18:33:04.364143 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/56702a03-175d-4a54-b84f-e25cab037431-scripts" (OuterVolumeSpecName: "scripts") pod "56702a03-175d-4a54-b84f-e25cab037431" (UID: "56702a03-175d-4a54-b84f-e25cab037431"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:33:04 crc kubenswrapper[4988]: I1008 18:33:04.365546 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage11-crc" (OuterVolumeSpecName: "glance") pod "56702a03-175d-4a54-b84f-e25cab037431" (UID: "56702a03-175d-4a54-b84f-e25cab037431"). InnerVolumeSpecName "local-storage11-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 08 18:33:04 crc kubenswrapper[4988]: I1008 18:33:04.379119 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/83ef7610-23ac-4445-b75c-ae7206b4d810-kube-api-access-lvrb9" (OuterVolumeSpecName: "kube-api-access-lvrb9") pod "83ef7610-23ac-4445-b75c-ae7206b4d810" (UID: "83ef7610-23ac-4445-b75c-ae7206b4d810"). InnerVolumeSpecName "kube-api-access-lvrb9". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:33:04 crc kubenswrapper[4988]: I1008 18:33:04.389765 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Oct 08 18:33:04 crc kubenswrapper[4988]: I1008 18:33:04.391935 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/32bb7401-6292-4ca6-98f4-15522125c3ee-kube-api-access-n9xkg" (OuterVolumeSpecName: "kube-api-access-n9xkg") pod "32bb7401-6292-4ca6-98f4-15522125c3ee" (UID: "32bb7401-6292-4ca6-98f4-15522125c3ee"). InnerVolumeSpecName "kube-api-access-n9xkg". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:33:04 crc kubenswrapper[4988]: I1008 18:33:04.395709 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glancec9f7-account-delete-sp22m" Oct 08 18:33:04 crc kubenswrapper[4988]: I1008 18:33:04.399121 4988 scope.go:117] "RemoveContainer" containerID="2927fb800c44516806e01d37c729bc04efb6e3981dccab523101ed4405525b47" Oct 08 18:33:04 crc kubenswrapper[4988]: I1008 18:33:04.407828 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron3849-account-delete-m479b" Oct 08 18:33:04 crc kubenswrapper[4988]: I1008 18:33:04.422163 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/83ef7610-23ac-4445-b75c-ae7206b4d810-config-data" (OuterVolumeSpecName: "config-data") pod "83ef7610-23ac-4445-b75c-ae7206b4d810" (UID: "83ef7610-23ac-4445-b75c-ae7206b4d810"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:33:04 crc kubenswrapper[4988]: I1008 18:33:04.430239 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Oct 08 18:33:04 crc kubenswrapper[4988]: I1008 18:33:04.433827 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 08 18:33:04 crc kubenswrapper[4988]: I1008 18:33:04.435912 4988 scope.go:117] "RemoveContainer" containerID="1f77c788702456fc178f26b1d4a4449a94e1ac2f3f33ad505531b3885e2f8f15" Oct 08 18:33:04 crc kubenswrapper[4988]: E1008 18:33:04.436575 4988 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1f77c788702456fc178f26b1d4a4449a94e1ac2f3f33ad505531b3885e2f8f15\": container with ID starting with 1f77c788702456fc178f26b1d4a4449a94e1ac2f3f33ad505531b3885e2f8f15 not found: ID does not exist" containerID="1f77c788702456fc178f26b1d4a4449a94e1ac2f3f33ad505531b3885e2f8f15" Oct 08 18:33:04 crc kubenswrapper[4988]: I1008 18:33:04.436615 4988 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1f77c788702456fc178f26b1d4a4449a94e1ac2f3f33ad505531b3885e2f8f15"} err="failed to get container status \"1f77c788702456fc178f26b1d4a4449a94e1ac2f3f33ad505531b3885e2f8f15\": rpc error: code = NotFound desc = could not find container \"1f77c788702456fc178f26b1d4a4449a94e1ac2f3f33ad505531b3885e2f8f15\": container with ID starting with 1f77c788702456fc178f26b1d4a4449a94e1ac2f3f33ad505531b3885e2f8f15 not found: ID does not exist" Oct 08 18:33:04 crc kubenswrapper[4988]: I1008 18:33:04.436643 4988 scope.go:117] "RemoveContainer" containerID="2927fb800c44516806e01d37c729bc04efb6e3981dccab523101ed4405525b47" Oct 08 18:33:04 crc kubenswrapper[4988]: E1008 18:33:04.437124 4988 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2927fb800c44516806e01d37c729bc04efb6e3981dccab523101ed4405525b47\": container with ID starting with 2927fb800c44516806e01d37c729bc04efb6e3981dccab523101ed4405525b47 not found: ID does not exist" containerID="2927fb800c44516806e01d37c729bc04efb6e3981dccab523101ed4405525b47" Oct 08 18:33:04 crc kubenswrapper[4988]: I1008 18:33:04.437156 4988 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2927fb800c44516806e01d37c729bc04efb6e3981dccab523101ed4405525b47"} err="failed to get container status \"2927fb800c44516806e01d37c729bc04efb6e3981dccab523101ed4405525b47\": rpc error: code = NotFound desc = could not find container \"2927fb800c44516806e01d37c729bc04efb6e3981dccab523101ed4405525b47\": container with ID starting with 2927fb800c44516806e01d37c729bc04efb6e3981dccab523101ed4405525b47 not found: ID does not exist" Oct 08 18:33:04 crc kubenswrapper[4988]: I1008 18:33:04.437172 4988 scope.go:117] "RemoveContainer" containerID="9933c83a6d7a8c9610edd4f70255f0f91bccecf1b14bb8fd883bcb6b9506666f" Oct 08 18:33:04 crc kubenswrapper[4988]: I1008 18:33:04.445747 4988 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/56702a03-175d-4a54-b84f-e25cab037431-httpd-run\") on node \"crc\" DevicePath \"\"" Oct 08 18:33:04 crc kubenswrapper[4988]: I1008 18:33:04.445777 4988 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n9xkg\" (UniqueName: \"kubernetes.io/projected/32bb7401-6292-4ca6-98f4-15522125c3ee-kube-api-access-n9xkg\") on node \"crc\" DevicePath \"\"" Oct 08 18:33:04 crc kubenswrapper[4988]: I1008 18:33:04.445787 4988 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/56702a03-175d-4a54-b84f-e25cab037431-scripts\") on node \"crc\" DevicePath \"\"" Oct 08 18:33:04 crc kubenswrapper[4988]: I1008 18:33:04.445799 4988 reconciler_common.go:293] "Volume detached for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/32bb7401-6292-4ca6-98f4-15522125c3ee-kolla-config\") on node \"crc\" DevicePath \"\"" Oct 08 18:33:04 crc kubenswrapper[4988]: I1008 18:33:04.445823 4988 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") on node \"crc\" " Oct 08 18:33:04 crc kubenswrapper[4988]: I1008 18:33:04.445836 4988 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/32bb7401-6292-4ca6-98f4-15522125c3ee-config-data\") on node \"crc\" DevicePath \"\"" Oct 08 18:33:04 crc kubenswrapper[4988]: I1008 18:33:04.445847 4988 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/83ef7610-23ac-4445-b75c-ae7206b4d810-config-data\") on node \"crc\" DevicePath \"\"" Oct 08 18:33:04 crc kubenswrapper[4988]: I1008 18:33:04.446152 4988 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/3a57537b-6115-4249-8d16-3ab0c4b6d21e-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 08 18:33:04 crc kubenswrapper[4988]: I1008 18:33:04.446169 4988 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lvrb9\" (UniqueName: \"kubernetes.io/projected/83ef7610-23ac-4445-b75c-ae7206b4d810-kube-api-access-lvrb9\") on node \"crc\" DevicePath \"\"" Oct 08 18:33:04 crc kubenswrapper[4988]: I1008 18:33:04.446180 4988 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-g74gb\" (UniqueName: \"kubernetes.io/projected/56702a03-175d-4a54-b84f-e25cab037431-kube-api-access-g74gb\") on node \"crc\" DevicePath \"\"" Oct 08 18:33:04 crc kubenswrapper[4988]: I1008 18:33:04.446203 4988 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/83ef7610-23ac-4445-b75c-ae7206b4d810-logs\") on node \"crc\" DevicePath \"\"" Oct 08 18:33:04 crc kubenswrapper[4988]: I1008 18:33:04.446213 4988 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/56702a03-175d-4a54-b84f-e25cab037431-logs\") on node \"crc\" DevicePath \"\"" Oct 08 18:33:04 crc kubenswrapper[4988]: I1008 18:33:04.452609 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Oct 08 18:33:04 crc kubenswrapper[4988]: I1008 18:33:04.458534 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/83ef7610-23ac-4445-b75c-ae7206b4d810-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "83ef7610-23ac-4445-b75c-ae7206b4d810" (UID: "83ef7610-23ac-4445-b75c-ae7206b4d810"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:33:04 crc kubenswrapper[4988]: I1008 18:33:04.460321 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/32bb7401-6292-4ca6-98f4-15522125c3ee-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "32bb7401-6292-4ca6-98f4-15522125c3ee" (UID: "32bb7401-6292-4ca6-98f4-15522125c3ee"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:33:04 crc kubenswrapper[4988]: I1008 18:33:04.460788 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Oct 08 18:33:04 crc kubenswrapper[4988]: I1008 18:33:04.473616 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/83ef7610-23ac-4445-b75c-ae7206b4d810-nova-metadata-tls-certs" (OuterVolumeSpecName: "nova-metadata-tls-certs") pod "83ef7610-23ac-4445-b75c-ae7206b4d810" (UID: "83ef7610-23ac-4445-b75c-ae7206b4d810"). InnerVolumeSpecName "nova-metadata-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:33:04 crc kubenswrapper[4988]: I1008 18:33:04.474698 4988 scope.go:117] "RemoveContainer" containerID="4cfd36bf140608a1fc5f7709dc1e3e6caea4bad2d51f7170f308befa5e5888fd" Oct 08 18:33:04 crc kubenswrapper[4988]: I1008 18:33:04.483490 4988 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage11-crc" (UniqueName: "kubernetes.io/local-volume/local-storage11-crc") on node "crc" Oct 08 18:33:04 crc kubenswrapper[4988]: I1008 18:33:04.487432 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/32bb7401-6292-4ca6-98f4-15522125c3ee-memcached-tls-certs" (OuterVolumeSpecName: "memcached-tls-certs") pod "32bb7401-6292-4ca6-98f4-15522125c3ee" (UID: "32bb7401-6292-4ca6-98f4-15522125c3ee"). InnerVolumeSpecName "memcached-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:33:04 crc kubenswrapper[4988]: I1008 18:33:04.507541 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/56702a03-175d-4a54-b84f-e25cab037431-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "56702a03-175d-4a54-b84f-e25cab037431" (UID: "56702a03-175d-4a54-b84f-e25cab037431"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:33:04 crc kubenswrapper[4988]: I1008 18:33:04.510851 4988 scope.go:117] "RemoveContainer" containerID="9315fcc87c2b03972c71e57728c4fdbc3b30200edc121e010b7ac2e7d7658e92" Oct 08 18:33:04 crc kubenswrapper[4988]: I1008 18:33:04.515122 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/56702a03-175d-4a54-b84f-e25cab037431-config-data" (OuterVolumeSpecName: "config-data") pod "56702a03-175d-4a54-b84f-e25cab037431" (UID: "56702a03-175d-4a54-b84f-e25cab037431"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:33:04 crc kubenswrapper[4988]: I1008 18:33:04.525197 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/56702a03-175d-4a54-b84f-e25cab037431-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "56702a03-175d-4a54-b84f-e25cab037431" (UID: "56702a03-175d-4a54-b84f-e25cab037431"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:33:04 crc kubenswrapper[4988]: I1008 18:33:04.534758 4988 scope.go:117] "RemoveContainer" containerID="7d3dbf100ba64891a97f34dd540d20ec061ba6822f245dec78c63c17c5d178fa" Oct 08 18:33:04 crc kubenswrapper[4988]: I1008 18:33:04.548988 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/3503942b-2825-4006-80e8-8c4610b89997-kolla-config\") pod \"3503942b-2825-4006-80e8-8c4610b89997\" (UID: \"3503942b-2825-4006-80e8-8c4610b89997\") " Oct 08 18:33:04 crc kubenswrapper[4988]: I1008 18:33:04.549035 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/85bc7f8f-fdb3-4abb-88d3-ed0c10982c4c-config-data\") pod \"85bc7f8f-fdb3-4abb-88d3-ed0c10982c4c\" (UID: \"85bc7f8f-fdb3-4abb-88d3-ed0c10982c4c\") " Oct 08 18:33:04 crc kubenswrapper[4988]: I1008 18:33:04.549064 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zrmcr\" (UniqueName: \"kubernetes.io/projected/85bc7f8f-fdb3-4abb-88d3-ed0c10982c4c-kube-api-access-zrmcr\") pod \"85bc7f8f-fdb3-4abb-88d3-ed0c10982c4c\" (UID: \"85bc7f8f-fdb3-4abb-88d3-ed0c10982c4c\") " Oct 08 18:33:04 crc kubenswrapper[4988]: I1008 18:33:04.549116 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/3503942b-2825-4006-80e8-8c4610b89997-secrets\") pod \"3503942b-2825-4006-80e8-8c4610b89997\" (UID: \"3503942b-2825-4006-80e8-8c4610b89997\") " Oct 08 18:33:04 crc kubenswrapper[4988]: I1008 18:33:04.549179 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-f6jr5\" (UniqueName: \"kubernetes.io/projected/ab3315d2-3248-42f1-9834-2bab55fa9935-kube-api-access-f6jr5\") pod \"ab3315d2-3248-42f1-9834-2bab55fa9935\" (UID: \"ab3315d2-3248-42f1-9834-2bab55fa9935\") " Oct 08 18:33:04 crc kubenswrapper[4988]: I1008 18:33:04.549221 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/3503942b-2825-4006-80e8-8c4610b89997-galera-tls-certs\") pod \"3503942b-2825-4006-80e8-8c4610b89997\" (UID: \"3503942b-2825-4006-80e8-8c4610b89997\") " Oct 08 18:33:04 crc kubenswrapper[4988]: I1008 18:33:04.549241 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/3503942b-2825-4006-80e8-8c4610b89997-config-data-default\") pod \"3503942b-2825-4006-80e8-8c4610b89997\" (UID: \"3503942b-2825-4006-80e8-8c4610b89997\") " Oct 08 18:33:04 crc kubenswrapper[4988]: I1008 18:33:04.549279 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mlmlm\" (UniqueName: \"kubernetes.io/projected/d7d648e7-3c02-4d26-993d-a901764b3329-kube-api-access-mlmlm\") pod \"d7d648e7-3c02-4d26-993d-a901764b3329\" (UID: \"d7d648e7-3c02-4d26-993d-a901764b3329\") " Oct 08 18:33:04 crc kubenswrapper[4988]: I1008 18:33:04.549301 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mysql-db\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"3503942b-2825-4006-80e8-8c4610b89997\" (UID: \"3503942b-2825-4006-80e8-8c4610b89997\") " Oct 08 18:33:04 crc kubenswrapper[4988]: I1008 18:33:04.549341 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1b24865e-cac2-490e-9e61-de22a379ff47-config-data\") pod \"1b24865e-cac2-490e-9e61-de22a379ff47\" (UID: \"1b24865e-cac2-490e-9e61-de22a379ff47\") " Oct 08 18:33:04 crc kubenswrapper[4988]: I1008 18:33:04.549371 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1b24865e-cac2-490e-9e61-de22a379ff47-combined-ca-bundle\") pod \"1b24865e-cac2-490e-9e61-de22a379ff47\" (UID: \"1b24865e-cac2-490e-9e61-de22a379ff47\") " Oct 08 18:33:04 crc kubenswrapper[4988]: I1008 18:33:04.549429 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xctqm\" (UniqueName: \"kubernetes.io/projected/1b24865e-cac2-490e-9e61-de22a379ff47-kube-api-access-xctqm\") pod \"1b24865e-cac2-490e-9e61-de22a379ff47\" (UID: \"1b24865e-cac2-490e-9e61-de22a379ff47\") " Oct 08 18:33:04 crc kubenswrapper[4988]: I1008 18:33:04.549453 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gz796\" (UniqueName: \"kubernetes.io/projected/3503942b-2825-4006-80e8-8c4610b89997-kube-api-access-gz796\") pod \"3503942b-2825-4006-80e8-8c4610b89997\" (UID: \"3503942b-2825-4006-80e8-8c4610b89997\") " Oct 08 18:33:04 crc kubenswrapper[4988]: I1008 18:33:04.549471 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3503942b-2825-4006-80e8-8c4610b89997-operator-scripts\") pod \"3503942b-2825-4006-80e8-8c4610b89997\" (UID: \"3503942b-2825-4006-80e8-8c4610b89997\") " Oct 08 18:33:04 crc kubenswrapper[4988]: I1008 18:33:04.549490 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3503942b-2825-4006-80e8-8c4610b89997-combined-ca-bundle\") pod \"3503942b-2825-4006-80e8-8c4610b89997\" (UID: \"3503942b-2825-4006-80e8-8c4610b89997\") " Oct 08 18:33:04 crc kubenswrapper[4988]: I1008 18:33:04.549510 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/85bc7f8f-fdb3-4abb-88d3-ed0c10982c4c-combined-ca-bundle\") pod \"85bc7f8f-fdb3-4abb-88d3-ed0c10982c4c\" (UID: \"85bc7f8f-fdb3-4abb-88d3-ed0c10982c4c\") " Oct 08 18:33:04 crc kubenswrapper[4988]: I1008 18:33:04.549535 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/3503942b-2825-4006-80e8-8c4610b89997-config-data-generated\") pod \"3503942b-2825-4006-80e8-8c4610b89997\" (UID: \"3503942b-2825-4006-80e8-8c4610b89997\") " Oct 08 18:33:04 crc kubenswrapper[4988]: I1008 18:33:04.550140 4988 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/56702a03-175d-4a54-b84f-e25cab037431-public-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 08 18:33:04 crc kubenswrapper[4988]: I1008 18:33:04.550159 4988 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/83ef7610-23ac-4445-b75c-ae7206b4d810-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 18:33:04 crc kubenswrapper[4988]: I1008 18:33:04.550169 4988 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/83ef7610-23ac-4445-b75c-ae7206b4d810-nova-metadata-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 08 18:33:04 crc kubenswrapper[4988]: I1008 18:33:04.550179 4988 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/56702a03-175d-4a54-b84f-e25cab037431-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 18:33:04 crc kubenswrapper[4988]: I1008 18:33:04.550189 4988 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/56702a03-175d-4a54-b84f-e25cab037431-config-data\") on node \"crc\" DevicePath \"\"" Oct 08 18:33:04 crc kubenswrapper[4988]: I1008 18:33:04.550199 4988 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/32bb7401-6292-4ca6-98f4-15522125c3ee-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 18:33:04 crc kubenswrapper[4988]: I1008 18:33:04.550207 4988 reconciler_common.go:293] "Volume detached for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/32bb7401-6292-4ca6-98f4-15522125c3ee-memcached-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 08 18:33:04 crc kubenswrapper[4988]: I1008 18:33:04.550217 4988 reconciler_common.go:293] "Volume detached for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") on node \"crc\" DevicePath \"\"" Oct 08 18:33:04 crc kubenswrapper[4988]: I1008 18:33:04.550566 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3503942b-2825-4006-80e8-8c4610b89997-config-data-generated" (OuterVolumeSpecName: "config-data-generated") pod "3503942b-2825-4006-80e8-8c4610b89997" (UID: "3503942b-2825-4006-80e8-8c4610b89997"). InnerVolumeSpecName "config-data-generated". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 18:33:04 crc kubenswrapper[4988]: I1008 18:33:04.550627 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3503942b-2825-4006-80e8-8c4610b89997-kolla-config" (OuterVolumeSpecName: "kolla-config") pod "3503942b-2825-4006-80e8-8c4610b89997" (UID: "3503942b-2825-4006-80e8-8c4610b89997"). InnerVolumeSpecName "kolla-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 18:33:04 crc kubenswrapper[4988]: I1008 18:33:04.550992 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3503942b-2825-4006-80e8-8c4610b89997-config-data-default" (OuterVolumeSpecName: "config-data-default") pod "3503942b-2825-4006-80e8-8c4610b89997" (UID: "3503942b-2825-4006-80e8-8c4610b89997"). InnerVolumeSpecName "config-data-default". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 18:33:04 crc kubenswrapper[4988]: I1008 18:33:04.551568 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3503942b-2825-4006-80e8-8c4610b89997-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "3503942b-2825-4006-80e8-8c4610b89997" (UID: "3503942b-2825-4006-80e8-8c4610b89997"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 18:33:04 crc kubenswrapper[4988]: I1008 18:33:04.553920 4988 scope.go:117] "RemoveContainer" containerID="9315fcc87c2b03972c71e57728c4fdbc3b30200edc121e010b7ac2e7d7658e92" Oct 08 18:33:04 crc kubenswrapper[4988]: E1008 18:33:04.557043 4988 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9315fcc87c2b03972c71e57728c4fdbc3b30200edc121e010b7ac2e7d7658e92\": container with ID starting with 9315fcc87c2b03972c71e57728c4fdbc3b30200edc121e010b7ac2e7d7658e92 not found: ID does not exist" containerID="9315fcc87c2b03972c71e57728c4fdbc3b30200edc121e010b7ac2e7d7658e92" Oct 08 18:33:04 crc kubenswrapper[4988]: I1008 18:33:04.557077 4988 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9315fcc87c2b03972c71e57728c4fdbc3b30200edc121e010b7ac2e7d7658e92"} err="failed to get container status \"9315fcc87c2b03972c71e57728c4fdbc3b30200edc121e010b7ac2e7d7658e92\": rpc error: code = NotFound desc = could not find container \"9315fcc87c2b03972c71e57728c4fdbc3b30200edc121e010b7ac2e7d7658e92\": container with ID starting with 9315fcc87c2b03972c71e57728c4fdbc3b30200edc121e010b7ac2e7d7658e92 not found: ID does not exist" Oct 08 18:33:04 crc kubenswrapper[4988]: I1008 18:33:04.557100 4988 scope.go:117] "RemoveContainer" containerID="7d3dbf100ba64891a97f34dd540d20ec061ba6822f245dec78c63c17c5d178fa" Oct 08 18:33:04 crc kubenswrapper[4988]: I1008 18:33:04.557185 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1b24865e-cac2-490e-9e61-de22a379ff47-kube-api-access-xctqm" (OuterVolumeSpecName: "kube-api-access-xctqm") pod "1b24865e-cac2-490e-9e61-de22a379ff47" (UID: "1b24865e-cac2-490e-9e61-de22a379ff47"). InnerVolumeSpecName "kube-api-access-xctqm". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:33:04 crc kubenswrapper[4988]: I1008 18:33:04.557323 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ab3315d2-3248-42f1-9834-2bab55fa9935-kube-api-access-f6jr5" (OuterVolumeSpecName: "kube-api-access-f6jr5") pod "ab3315d2-3248-42f1-9834-2bab55fa9935" (UID: "ab3315d2-3248-42f1-9834-2bab55fa9935"). InnerVolumeSpecName "kube-api-access-f6jr5". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:33:04 crc kubenswrapper[4988]: E1008 18:33:04.557835 4988 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7d3dbf100ba64891a97f34dd540d20ec061ba6822f245dec78c63c17c5d178fa\": container with ID starting with 7d3dbf100ba64891a97f34dd540d20ec061ba6822f245dec78c63c17c5d178fa not found: ID does not exist" containerID="7d3dbf100ba64891a97f34dd540d20ec061ba6822f245dec78c63c17c5d178fa" Oct 08 18:33:04 crc kubenswrapper[4988]: I1008 18:33:04.557854 4988 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7d3dbf100ba64891a97f34dd540d20ec061ba6822f245dec78c63c17c5d178fa"} err="failed to get container status \"7d3dbf100ba64891a97f34dd540d20ec061ba6822f245dec78c63c17c5d178fa\": rpc error: code = NotFound desc = could not find container \"7d3dbf100ba64891a97f34dd540d20ec061ba6822f245dec78c63c17c5d178fa\": container with ID starting with 7d3dbf100ba64891a97f34dd540d20ec061ba6822f245dec78c63c17c5d178fa not found: ID does not exist" Oct 08 18:33:04 crc kubenswrapper[4988]: I1008 18:33:04.557869 4988 scope.go:117] "RemoveContainer" containerID="6d9e1194470201a267d04dcb959b05798998db9b5ddd0f647596f14d4784a516" Oct 08 18:33:04 crc kubenswrapper[4988]: I1008 18:33:04.562212 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d7d648e7-3c02-4d26-993d-a901764b3329-kube-api-access-mlmlm" (OuterVolumeSpecName: "kube-api-access-mlmlm") pod "d7d648e7-3c02-4d26-993d-a901764b3329" (UID: "d7d648e7-3c02-4d26-993d-a901764b3329"). InnerVolumeSpecName "kube-api-access-mlmlm". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:33:04 crc kubenswrapper[4988]: I1008 18:33:04.562769 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3503942b-2825-4006-80e8-8c4610b89997-kube-api-access-gz796" (OuterVolumeSpecName: "kube-api-access-gz796") pod "3503942b-2825-4006-80e8-8c4610b89997" (UID: "3503942b-2825-4006-80e8-8c4610b89997"). InnerVolumeSpecName "kube-api-access-gz796". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:33:04 crc kubenswrapper[4988]: I1008 18:33:04.562900 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3503942b-2825-4006-80e8-8c4610b89997-secrets" (OuterVolumeSpecName: "secrets") pod "3503942b-2825-4006-80e8-8c4610b89997" (UID: "3503942b-2825-4006-80e8-8c4610b89997"). InnerVolumeSpecName "secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:33:04 crc kubenswrapper[4988]: I1008 18:33:04.564113 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/85bc7f8f-fdb3-4abb-88d3-ed0c10982c4c-kube-api-access-zrmcr" (OuterVolumeSpecName: "kube-api-access-zrmcr") pod "85bc7f8f-fdb3-4abb-88d3-ed0c10982c4c" (UID: "85bc7f8f-fdb3-4abb-88d3-ed0c10982c4c"). InnerVolumeSpecName "kube-api-access-zrmcr". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:33:04 crc kubenswrapper[4988]: I1008 18:33:04.565051 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"3503942b-2825-4006-80e8-8c4610b89997","Type":"ContainerDied","Data":"f883ee7dc1d600955ad76f9143dec7c4f14a01603d4965aa6075b83d56ed74f4"} Oct 08 18:33:04 crc kubenswrapper[4988]: I1008 18:33:04.565190 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Oct 08 18:33:04 crc kubenswrapper[4988]: I1008 18:33:04.567112 4988 generic.go:334] "Generic (PLEG): container finished" podID="bd9cfcdf-4770-4efa-a41e-7867ec85206f" containerID="17f00a030a73692f1b5f14c014e24ee1e9bf7251c877957ee31dd0cb3579f43c" exitCode=0 Oct 08 18:33:04 crc kubenswrapper[4988]: I1008 18:33:04.567183 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"bd9cfcdf-4770-4efa-a41e-7867ec85206f","Type":"ContainerDied","Data":"17f00a030a73692f1b5f14c014e24ee1e9bf7251c877957ee31dd0cb3579f43c"} Oct 08 18:33:04 crc kubenswrapper[4988]: I1008 18:33:04.567268 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Oct 08 18:33:04 crc kubenswrapper[4988]: I1008 18:33:04.567548 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"bd9cfcdf-4770-4efa-a41e-7867ec85206f","Type":"ContainerDied","Data":"fecc6278563eb7471936c98f9e1f21c2d2390670f6fae45a4a28048909f867b7"} Oct 08 18:33:04 crc kubenswrapper[4988]: I1008 18:33:04.569847 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage01-crc" (OuterVolumeSpecName: "mysql-db") pod "3503942b-2825-4006-80e8-8c4610b89997" (UID: "3503942b-2825-4006-80e8-8c4610b89997"). InnerVolumeSpecName "local-storage01-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 08 18:33:04 crc kubenswrapper[4988]: I1008 18:33:04.570671 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"83ef7610-23ac-4445-b75c-ae7206b4d810","Type":"ContainerDied","Data":"cce50310cff807710abf94c8f4ae4a3519677bbe1dca06bbe268f6dbc2ae8eec"} Oct 08 18:33:04 crc kubenswrapper[4988]: I1008 18:33:04.570749 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 08 18:33:04 crc kubenswrapper[4988]: I1008 18:33:04.576123 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"56702a03-175d-4a54-b84f-e25cab037431","Type":"ContainerDied","Data":"05fc98e1e038eeb4c35b2cae978bb1945dce3667d3c4eaf00c83a5cde3df55e3"} Oct 08 18:33:04 crc kubenswrapper[4988]: I1008 18:33:04.576228 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 08 18:33:04 crc kubenswrapper[4988]: I1008 18:33:04.580853 4988 generic.go:334] "Generic (PLEG): container finished" podID="1b24865e-cac2-490e-9e61-de22a379ff47" containerID="6d8f1c5dc40a53cca8a87fffd5d06729664e7815b666a1eba260acf508457342" exitCode=0 Oct 08 18:33:04 crc kubenswrapper[4988]: I1008 18:33:04.580944 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 08 18:33:04 crc kubenswrapper[4988]: I1008 18:33:04.580934 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"1b24865e-cac2-490e-9e61-de22a379ff47","Type":"ContainerDied","Data":"6d8f1c5dc40a53cca8a87fffd5d06729664e7815b666a1eba260acf508457342"} Oct 08 18:33:04 crc kubenswrapper[4988]: I1008 18:33:04.581043 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"1b24865e-cac2-490e-9e61-de22a379ff47","Type":"ContainerDied","Data":"a05a410cb5097099f45d7708891f938ae729f9de75fcebee924aaf3fbf711bf8"} Oct 08 18:33:04 crc kubenswrapper[4988]: I1008 18:33:04.584742 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron3849-account-delete-m479b" event={"ID":"ab3315d2-3248-42f1-9834-2bab55fa9935","Type":"ContainerDied","Data":"d3f029cb7d883d25bc8d0897c54c850a721d92b478323fff209fa6dea5f80250"} Oct 08 18:33:04 crc kubenswrapper[4988]: I1008 18:33:04.584787 4988 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d3f029cb7d883d25bc8d0897c54c850a721d92b478323fff209fa6dea5f80250" Oct 08 18:33:04 crc kubenswrapper[4988]: I1008 18:33:04.584846 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron3849-account-delete-m479b" Oct 08 18:33:04 crc kubenswrapper[4988]: I1008 18:33:04.589602 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/85bc7f8f-fdb3-4abb-88d3-ed0c10982c4c-config-data" (OuterVolumeSpecName: "config-data") pod "85bc7f8f-fdb3-4abb-88d3-ed0c10982c4c" (UID: "85bc7f8f-fdb3-4abb-88d3-ed0c10982c4c"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:33:04 crc kubenswrapper[4988]: I1008 18:33:04.593304 4988 generic.go:334] "Generic (PLEG): container finished" podID="0ec345a3-1744-4e78-bf13-91f0325296e4" containerID="590ae1792458f3fd70bbade362261ee71cafb4633235b6af45da637905058707" exitCode=0 Oct 08 18:33:04 crc kubenswrapper[4988]: I1008 18:33:04.593380 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"0ec345a3-1744-4e78-bf13-91f0325296e4","Type":"ContainerDied","Data":"590ae1792458f3fd70bbade362261ee71cafb4633235b6af45da637905058707"} Oct 08 18:33:04 crc kubenswrapper[4988]: I1008 18:33:04.593427 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"0ec345a3-1744-4e78-bf13-91f0325296e4","Type":"ContainerDied","Data":"d7041603bf5b31608ccc2efc18fa6bb62e518086af0360055ed352353750dabb"} Oct 08 18:33:04 crc kubenswrapper[4988]: I1008 18:33:04.593588 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Oct 08 18:33:04 crc kubenswrapper[4988]: I1008 18:33:04.595975 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 08 18:33:04 crc kubenswrapper[4988]: I1008 18:33:04.596377 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"0efd1f53-90ad-4eb8-90f2-2ba9fc1593c2","Type":"ContainerDied","Data":"4a380fff05cca0afef17a17e2bbd956d198717c45598f70a6f3af1d5e95b0439"} Oct 08 18:33:04 crc kubenswrapper[4988]: I1008 18:33:04.597412 4988 scope.go:117] "RemoveContainer" containerID="2b04aad608cf7534875c0cf78abd43e0c25cefee26a5ecebee48c75fd8c10b09" Oct 08 18:33:04 crc kubenswrapper[4988]: I1008 18:33:04.610870 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"85bc7f8f-fdb3-4abb-88d3-ed0c10982c4c","Type":"ContainerDied","Data":"20bc7fbba36bae4b2acfad1b694961fa583a6bb472efe5ab62f9bf0a088c1414"} Oct 08 18:33:04 crc kubenswrapper[4988]: I1008 18:33:04.611170 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Oct 08 18:33:04 crc kubenswrapper[4988]: I1008 18:33:04.612903 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/85bc7f8f-fdb3-4abb-88d3-ed0c10982c4c-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "85bc7f8f-fdb3-4abb-88d3-ed0c10982c4c" (UID: "85bc7f8f-fdb3-4abb-88d3-ed0c10982c4c"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:33:04 crc kubenswrapper[4988]: I1008 18:33:04.615280 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glancec9f7-account-delete-sp22m" Oct 08 18:33:04 crc kubenswrapper[4988]: I1008 18:33:04.615279 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glancec9f7-account-delete-sp22m" event={"ID":"d7d648e7-3c02-4d26-993d-a901764b3329","Type":"ContainerDied","Data":"464217ebf5ac45bfe14b80e19139ccfccd2d00bd4f5783a2e886fa030ab82158"} Oct 08 18:33:04 crc kubenswrapper[4988]: I1008 18:33:04.615548 4988 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="464217ebf5ac45bfe14b80e19139ccfccd2d00bd4f5783a2e886fa030ab82158" Oct 08 18:33:04 crc kubenswrapper[4988]: I1008 18:33:04.618888 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 08 18:33:04 crc kubenswrapper[4988]: I1008 18:33:04.626056 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1b24865e-cac2-490e-9e61-de22a379ff47-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "1b24865e-cac2-490e-9e61-de22a379ff47" (UID: "1b24865e-cac2-490e-9e61-de22a379ff47"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:33:04 crc kubenswrapper[4988]: I1008 18:33:04.626747 4988 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 08 18:33:04 crc kubenswrapper[4988]: I1008 18:33:04.628111 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"32bb7401-6292-4ca6-98f4-15522125c3ee","Type":"ContainerDied","Data":"f5dbe430bd2d56992aa14033ff6363ff154185ebd2038a740e4725f0755dc90e"} Oct 08 18:33:04 crc kubenswrapper[4988]: I1008 18:33:04.628177 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-6c486d5744-j86zn" Oct 08 18:33:04 crc kubenswrapper[4988]: I1008 18:33:04.628210 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Oct 08 18:33:04 crc kubenswrapper[4988]: I1008 18:33:04.628263 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/novacell02f70-account-delete-cdwsx" Oct 08 18:33:04 crc kubenswrapper[4988]: I1008 18:33:04.638153 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Oct 08 18:33:04 crc kubenswrapper[4988]: I1008 18:33:04.638607 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3503942b-2825-4006-80e8-8c4610b89997-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "3503942b-2825-4006-80e8-8c4610b89997" (UID: "3503942b-2825-4006-80e8-8c4610b89997"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:33:04 crc kubenswrapper[4988]: I1008 18:33:04.642860 4988 scope.go:117] "RemoveContainer" containerID="d2face2dfb0693cf7a2c3531e80c082e310d388b04c8005247e18e79ea8bbab3" Oct 08 18:33:04 crc kubenswrapper[4988]: I1008 18:33:04.645912 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1b24865e-cac2-490e-9e61-de22a379ff47-config-data" (OuterVolumeSpecName: "config-data") pod "1b24865e-cac2-490e-9e61-de22a379ff47" (UID: "1b24865e-cac2-490e-9e61-de22a379ff47"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:33:04 crc kubenswrapper[4988]: I1008 18:33:04.647471 4988 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Oct 08 18:33:04 crc kubenswrapper[4988]: I1008 18:33:04.651584 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bd9cfcdf-4770-4efa-a41e-7867ec85206f-combined-ca-bundle\") pod \"bd9cfcdf-4770-4efa-a41e-7867ec85206f\" (UID: \"bd9cfcdf-4770-4efa-a41e-7867ec85206f\") " Oct 08 18:33:04 crc kubenswrapper[4988]: I1008 18:33:04.651668 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bd9cfcdf-4770-4efa-a41e-7867ec85206f-config-data\") pod \"bd9cfcdf-4770-4efa-a41e-7867ec85206f\" (UID: \"bd9cfcdf-4770-4efa-a41e-7867ec85206f\") " Oct 08 18:33:04 crc kubenswrapper[4988]: I1008 18:33:04.651692 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9glwn\" (UniqueName: \"kubernetes.io/projected/0ec345a3-1744-4e78-bf13-91f0325296e4-kube-api-access-9glwn\") pod \"0ec345a3-1744-4e78-bf13-91f0325296e4\" (UID: \"0ec345a3-1744-4e78-bf13-91f0325296e4\") " Oct 08 18:33:04 crc kubenswrapper[4988]: I1008 18:33:04.651799 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zb8kd\" (UniqueName: \"kubernetes.io/projected/bd9cfcdf-4770-4efa-a41e-7867ec85206f-kube-api-access-zb8kd\") pod \"bd9cfcdf-4770-4efa-a41e-7867ec85206f\" (UID: \"bd9cfcdf-4770-4efa-a41e-7867ec85206f\") " Oct 08 18:33:04 crc kubenswrapper[4988]: I1008 18:33:04.651886 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0ec345a3-1744-4e78-bf13-91f0325296e4-combined-ca-bundle\") pod \"0ec345a3-1744-4e78-bf13-91f0325296e4\" (UID: \"0ec345a3-1744-4e78-bf13-91f0325296e4\") " Oct 08 18:33:04 crc kubenswrapper[4988]: I1008 18:33:04.651919 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/0ec345a3-1744-4e78-bf13-91f0325296e4-etc-machine-id\") pod \"0ec345a3-1744-4e78-bf13-91f0325296e4\" (UID: \"0ec345a3-1744-4e78-bf13-91f0325296e4\") " Oct 08 18:33:04 crc kubenswrapper[4988]: I1008 18:33:04.651956 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0ec345a3-1744-4e78-bf13-91f0325296e4-scripts\") pod \"0ec345a3-1744-4e78-bf13-91f0325296e4\" (UID: \"0ec345a3-1744-4e78-bf13-91f0325296e4\") " Oct 08 18:33:04 crc kubenswrapper[4988]: I1008 18:33:04.651980 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0ec345a3-1744-4e78-bf13-91f0325296e4-config-data\") pod \"0ec345a3-1744-4e78-bf13-91f0325296e4\" (UID: \"0ec345a3-1744-4e78-bf13-91f0325296e4\") " Oct 08 18:33:04 crc kubenswrapper[4988]: I1008 18:33:04.651999 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/0ec345a3-1744-4e78-bf13-91f0325296e4-config-data-custom\") pod \"0ec345a3-1744-4e78-bf13-91f0325296e4\" (UID: \"0ec345a3-1744-4e78-bf13-91f0325296e4\") " Oct 08 18:33:04 crc kubenswrapper[4988]: I1008 18:33:04.652356 4988 reconciler_common.go:293] "Volume detached for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/3503942b-2825-4006-80e8-8c4610b89997-secrets\") on node \"crc\" DevicePath \"\"" Oct 08 18:33:04 crc kubenswrapper[4988]: I1008 18:33:04.652445 4988 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-f6jr5\" (UniqueName: \"kubernetes.io/projected/ab3315d2-3248-42f1-9834-2bab55fa9935-kube-api-access-f6jr5\") on node \"crc\" DevicePath \"\"" Oct 08 18:33:04 crc kubenswrapper[4988]: I1008 18:33:04.652463 4988 reconciler_common.go:293] "Volume detached for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/3503942b-2825-4006-80e8-8c4610b89997-config-data-default\") on node \"crc\" DevicePath \"\"" Oct 08 18:33:04 crc kubenswrapper[4988]: I1008 18:33:04.652472 4988 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mlmlm\" (UniqueName: \"kubernetes.io/projected/d7d648e7-3c02-4d26-993d-a901764b3329-kube-api-access-mlmlm\") on node \"crc\" DevicePath \"\"" Oct 08 18:33:04 crc kubenswrapper[4988]: I1008 18:33:04.652510 4988 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") on node \"crc\" " Oct 08 18:33:04 crc kubenswrapper[4988]: I1008 18:33:04.652521 4988 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1b24865e-cac2-490e-9e61-de22a379ff47-config-data\") on node \"crc\" DevicePath \"\"" Oct 08 18:33:04 crc kubenswrapper[4988]: I1008 18:33:04.652529 4988 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1b24865e-cac2-490e-9e61-de22a379ff47-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 18:33:04 crc kubenswrapper[4988]: I1008 18:33:04.652537 4988 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xctqm\" (UniqueName: \"kubernetes.io/projected/1b24865e-cac2-490e-9e61-de22a379ff47-kube-api-access-xctqm\") on node \"crc\" DevicePath \"\"" Oct 08 18:33:04 crc kubenswrapper[4988]: I1008 18:33:04.652547 4988 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gz796\" (UniqueName: \"kubernetes.io/projected/3503942b-2825-4006-80e8-8c4610b89997-kube-api-access-gz796\") on node \"crc\" DevicePath \"\"" Oct 08 18:33:04 crc kubenswrapper[4988]: I1008 18:33:04.652556 4988 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3503942b-2825-4006-80e8-8c4610b89997-operator-scripts\") on node \"crc\" DevicePath \"\"" Oct 08 18:33:04 crc kubenswrapper[4988]: I1008 18:33:04.652564 4988 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3503942b-2825-4006-80e8-8c4610b89997-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 18:33:04 crc kubenswrapper[4988]: I1008 18:33:04.652572 4988 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/85bc7f8f-fdb3-4abb-88d3-ed0c10982c4c-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 18:33:04 crc kubenswrapper[4988]: I1008 18:33:04.652580 4988 reconciler_common.go:293] "Volume detached for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/3503942b-2825-4006-80e8-8c4610b89997-config-data-generated\") on node \"crc\" DevicePath \"\"" Oct 08 18:33:04 crc kubenswrapper[4988]: I1008 18:33:04.652587 4988 reconciler_common.go:293] "Volume detached for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/3503942b-2825-4006-80e8-8c4610b89997-kolla-config\") on node \"crc\" DevicePath \"\"" Oct 08 18:33:04 crc kubenswrapper[4988]: I1008 18:33:04.652595 4988 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/85bc7f8f-fdb3-4abb-88d3-ed0c10982c4c-config-data\") on node \"crc\" DevicePath \"\"" Oct 08 18:33:04 crc kubenswrapper[4988]: I1008 18:33:04.652603 4988 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zrmcr\" (UniqueName: \"kubernetes.io/projected/85bc7f8f-fdb3-4abb-88d3-ed0c10982c4c-kube-api-access-zrmcr\") on node \"crc\" DevicePath \"\"" Oct 08 18:33:04 crc kubenswrapper[4988]: I1008 18:33:04.656734 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3503942b-2825-4006-80e8-8c4610b89997-galera-tls-certs" (OuterVolumeSpecName: "galera-tls-certs") pod "3503942b-2825-4006-80e8-8c4610b89997" (UID: "3503942b-2825-4006-80e8-8c4610b89997"). InnerVolumeSpecName "galera-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:33:04 crc kubenswrapper[4988]: I1008 18:33:04.656975 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/0ec345a3-1744-4e78-bf13-91f0325296e4-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "0ec345a3-1744-4e78-bf13-91f0325296e4" (UID: "0ec345a3-1744-4e78-bf13-91f0325296e4"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 08 18:33:04 crc kubenswrapper[4988]: I1008 18:33:04.670060 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron3849-account-delete-m479b"] Oct 08 18:33:04 crc kubenswrapper[4988]: I1008 18:33:04.671834 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0ec345a3-1744-4e78-bf13-91f0325296e4-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "0ec345a3-1744-4e78-bf13-91f0325296e4" (UID: "0ec345a3-1744-4e78-bf13-91f0325296e4"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:33:04 crc kubenswrapper[4988]: I1008 18:33:04.674680 4988 scope.go:117] "RemoveContainer" containerID="1c61da6cc2966ef990d9bcdbb3737986e369db20fe8b42772368a61ef7dc1f03" Oct 08 18:33:04 crc kubenswrapper[4988]: I1008 18:33:04.677682 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bd9cfcdf-4770-4efa-a41e-7867ec85206f-kube-api-access-zb8kd" (OuterVolumeSpecName: "kube-api-access-zb8kd") pod "bd9cfcdf-4770-4efa-a41e-7867ec85206f" (UID: "bd9cfcdf-4770-4efa-a41e-7867ec85206f"). InnerVolumeSpecName "kube-api-access-zb8kd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:33:04 crc kubenswrapper[4988]: I1008 18:33:04.692204 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0ec345a3-1744-4e78-bf13-91f0325296e4-kube-api-access-9glwn" (OuterVolumeSpecName: "kube-api-access-9glwn") pod "0ec345a3-1744-4e78-bf13-91f0325296e4" (UID: "0ec345a3-1744-4e78-bf13-91f0325296e4"). InnerVolumeSpecName "kube-api-access-9glwn". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:33:04 crc kubenswrapper[4988]: I1008 18:33:04.700947 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0ec345a3-1744-4e78-bf13-91f0325296e4-scripts" (OuterVolumeSpecName: "scripts") pod "0ec345a3-1744-4e78-bf13-91f0325296e4" (UID: "0ec345a3-1744-4e78-bf13-91f0325296e4"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:33:04 crc kubenswrapper[4988]: I1008 18:33:04.703014 4988 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage01-crc" (UniqueName: "kubernetes.io/local-volume/local-storage01-crc") on node "crc" Oct 08 18:33:04 crc kubenswrapper[4988]: I1008 18:33:04.706178 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bd9cfcdf-4770-4efa-a41e-7867ec85206f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "bd9cfcdf-4770-4efa-a41e-7867ec85206f" (UID: "bd9cfcdf-4770-4efa-a41e-7867ec85206f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:33:04 crc kubenswrapper[4988]: I1008 18:33:04.713080 4988 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron3849-account-delete-m479b"] Oct 08 18:33:04 crc kubenswrapper[4988]: I1008 18:33:04.726469 4988 scope.go:117] "RemoveContainer" containerID="0f31eca0db70bc881a635808747404e9ac717d3d38c16e693305f6196d241dac" Oct 08 18:33:04 crc kubenswrapper[4988]: I1008 18:33:04.727233 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 08 18:33:04 crc kubenswrapper[4988]: I1008 18:33:04.735531 4988 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Oct 08 18:33:04 crc kubenswrapper[4988]: I1008 18:33:04.740182 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/memcached-0"] Oct 08 18:33:04 crc kubenswrapper[4988]: I1008 18:33:04.740177 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bd9cfcdf-4770-4efa-a41e-7867ec85206f-config-data" (OuterVolumeSpecName: "config-data") pod "bd9cfcdf-4770-4efa-a41e-7867ec85206f" (UID: "bd9cfcdf-4770-4efa-a41e-7867ec85206f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:33:04 crc kubenswrapper[4988]: I1008 18:33:04.744446 4988 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/memcached-0"] Oct 08 18:33:04 crc kubenswrapper[4988]: I1008 18:33:04.748952 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/novacell02f70-account-delete-cdwsx"] Oct 08 18:33:04 crc kubenswrapper[4988]: I1008 18:33:04.754039 4988 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/novacell02f70-account-delete-cdwsx"] Oct 08 18:33:04 crc kubenswrapper[4988]: I1008 18:33:04.755470 4988 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0ec345a3-1744-4e78-bf13-91f0325296e4-scripts\") on node \"crc\" DevicePath \"\"" Oct 08 18:33:04 crc kubenswrapper[4988]: I1008 18:33:04.755496 4988 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/0ec345a3-1744-4e78-bf13-91f0325296e4-config-data-custom\") on node \"crc\" DevicePath \"\"" Oct 08 18:33:04 crc kubenswrapper[4988]: I1008 18:33:04.755505 4988 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bd9cfcdf-4770-4efa-a41e-7867ec85206f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 18:33:04 crc kubenswrapper[4988]: I1008 18:33:04.755514 4988 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bd9cfcdf-4770-4efa-a41e-7867ec85206f-config-data\") on node \"crc\" DevicePath \"\"" Oct 08 18:33:04 crc kubenswrapper[4988]: I1008 18:33:04.755525 4988 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9glwn\" (UniqueName: \"kubernetes.io/projected/0ec345a3-1744-4e78-bf13-91f0325296e4-kube-api-access-9glwn\") on node \"crc\" DevicePath \"\"" Oct 08 18:33:04 crc kubenswrapper[4988]: I1008 18:33:04.755535 4988 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zb8kd\" (UniqueName: \"kubernetes.io/projected/bd9cfcdf-4770-4efa-a41e-7867ec85206f-kube-api-access-zb8kd\") on node \"crc\" DevicePath \"\"" Oct 08 18:33:04 crc kubenswrapper[4988]: I1008 18:33:04.755545 4988 reconciler_common.go:293] "Volume detached for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/3503942b-2825-4006-80e8-8c4610b89997-galera-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 08 18:33:04 crc kubenswrapper[4988]: I1008 18:33:04.755554 4988 reconciler_common.go:293] "Volume detached for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") on node \"crc\" DevicePath \"\"" Oct 08 18:33:04 crc kubenswrapper[4988]: I1008 18:33:04.755574 4988 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/0ec345a3-1744-4e78-bf13-91f0325296e4-etc-machine-id\") on node \"crc\" DevicePath \"\"" Oct 08 18:33:04 crc kubenswrapper[4988]: I1008 18:33:04.757404 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0ec345a3-1744-4e78-bf13-91f0325296e4-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "0ec345a3-1744-4e78-bf13-91f0325296e4" (UID: "0ec345a3-1744-4e78-bf13-91f0325296e4"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:33:04 crc kubenswrapper[4988]: I1008 18:33:04.758607 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glancec9f7-account-delete-sp22m"] Oct 08 18:33:04 crc kubenswrapper[4988]: I1008 18:33:04.760766 4988 scope.go:117] "RemoveContainer" containerID="b226456d2f92fef2d798a8462e29b0c8f7c8a915dd5330879cb7257e0cc73d05" Oct 08 18:33:04 crc kubenswrapper[4988]: I1008 18:33:04.766570 4988 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glancec9f7-account-delete-sp22m"] Oct 08 18:33:04 crc kubenswrapper[4988]: I1008 18:33:04.767261 4988 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-server-0" podUID="830d6ec5-0015-4a88-b8e8-0fd47f64fd46" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.103:5671: connect: connection refused" Oct 08 18:33:04 crc kubenswrapper[4988]: I1008 18:33:04.767408 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0ec345a3-1744-4e78-bf13-91f0325296e4-config-data" (OuterVolumeSpecName: "config-data") pod "0ec345a3-1744-4e78-bf13-91f0325296e4" (UID: "0ec345a3-1744-4e78-bf13-91f0325296e4"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:33:04 crc kubenswrapper[4988]: I1008 18:33:04.778291 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-6c486d5744-j86zn"] Oct 08 18:33:04 crc kubenswrapper[4988]: I1008 18:33:04.783999 4988 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-api-6c486d5744-j86zn"] Oct 08 18:33:04 crc kubenswrapper[4988]: I1008 18:33:04.784261 4988 scope.go:117] "RemoveContainer" containerID="13f8a712b04c6d00db8bc2f724015fda6642a1723ac5c2e59ecbd2400fa35ece" Oct 08 18:33:04 crc kubenswrapper[4988]: I1008 18:33:04.784668 4988 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-gp8f4" podUID="c285e09d-c2f5-4642-9e6b-59b0ed3ed651" containerName="ovn-controller" probeResult="failure" output="command timed out" Oct 08 18:33:04 crc kubenswrapper[4988]: I1008 18:33:04.802376 4988 scope.go:117] "RemoveContainer" containerID="e1a685c92e7a7b5fad1185fbe08be64ed9f96221d3c3a8c8c5d3364e9a593229" Oct 08 18:33:04 crc kubenswrapper[4988]: I1008 18:33:04.815368 4988 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-gp8f4" podUID="c285e09d-c2f5-4642-9e6b-59b0ed3ed651" containerName="ovn-controller" probeResult="failure" output=< Oct 08 18:33:04 crc kubenswrapper[4988]: ERROR - Failed to get connection status from ovn-controller, ovn-appctl exit status: 0 Oct 08 18:33:04 crc kubenswrapper[4988]: > Oct 08 18:33:04 crc kubenswrapper[4988]: I1008 18:33:04.837813 4988 scope.go:117] "RemoveContainer" containerID="cefcf980cfc56a4cdb19408f13a9a6a84705b300fdac96063e98a4090209d49e" Oct 08 18:33:04 crc kubenswrapper[4988]: I1008 18:33:04.856793 4988 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0ec345a3-1744-4e78-bf13-91f0325296e4-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 18:33:04 crc kubenswrapper[4988]: I1008 18:33:04.856829 4988 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0ec345a3-1744-4e78-bf13-91f0325296e4-config-data\") on node \"crc\" DevicePath \"\"" Oct 08 18:33:04 crc kubenswrapper[4988]: I1008 18:33:04.859070 4988 scope.go:117] "RemoveContainer" containerID="17f00a030a73692f1b5f14c014e24ee1e9bf7251c877957ee31dd0cb3579f43c" Oct 08 18:33:04 crc kubenswrapper[4988]: I1008 18:33:04.885412 4988 scope.go:117] "RemoveContainer" containerID="17f00a030a73692f1b5f14c014e24ee1e9bf7251c877957ee31dd0cb3579f43c" Oct 08 18:33:04 crc kubenswrapper[4988]: E1008 18:33:04.885960 4988 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"17f00a030a73692f1b5f14c014e24ee1e9bf7251c877957ee31dd0cb3579f43c\": container with ID starting with 17f00a030a73692f1b5f14c014e24ee1e9bf7251c877957ee31dd0cb3579f43c not found: ID does not exist" containerID="17f00a030a73692f1b5f14c014e24ee1e9bf7251c877957ee31dd0cb3579f43c" Oct 08 18:33:04 crc kubenswrapper[4988]: I1008 18:33:04.886065 4988 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"17f00a030a73692f1b5f14c014e24ee1e9bf7251c877957ee31dd0cb3579f43c"} err="failed to get container status \"17f00a030a73692f1b5f14c014e24ee1e9bf7251c877957ee31dd0cb3579f43c\": rpc error: code = NotFound desc = could not find container \"17f00a030a73692f1b5f14c014e24ee1e9bf7251c877957ee31dd0cb3579f43c\": container with ID starting with 17f00a030a73692f1b5f14c014e24ee1e9bf7251c877957ee31dd0cb3579f43c not found: ID does not exist" Oct 08 18:33:04 crc kubenswrapper[4988]: I1008 18:33:04.886165 4988 scope.go:117] "RemoveContainer" containerID="4b994190cba63b85214c41a6bf0e9b5cca7c29dd236d40d5df3055f98c9ec35b" Oct 08 18:33:04 crc kubenswrapper[4988]: I1008 18:33:04.909406 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/openstack-cell1-galera-0"] Oct 08 18:33:04 crc kubenswrapper[4988]: I1008 18:33:04.924038 4988 scope.go:117] "RemoveContainer" containerID="e504bf27bfc24744eb7b0bdc9918cda7b788dc1c32411c922014b5f7f84ae75d" Oct 08 18:33:04 crc kubenswrapper[4988]: I1008 18:33:04.933964 4988 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/openstack-cell1-galera-0"] Oct 08 18:33:04 crc kubenswrapper[4988]: I1008 18:33:04.951484 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Oct 08 18:33:04 crc kubenswrapper[4988]: I1008 18:33:04.953312 4988 scope.go:117] "RemoveContainer" containerID="23da0ff1f117dc601e37e2bb0a79de17b4662200fa002a471f0c5938e67d484b" Oct 08 18:33:04 crc kubenswrapper[4988]: I1008 18:33:04.966469 4988 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Oct 08 18:33:04 crc kubenswrapper[4988]: I1008 18:33:04.976094 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-0"] Oct 08 18:33:04 crc kubenswrapper[4988]: I1008 18:33:04.978745 4988 scope.go:117] "RemoveContainer" containerID="9f5e2ee884e53117065985ec938aeba8be9ae37755acfba2392c435832c65f0e" Oct 08 18:33:04 crc kubenswrapper[4988]: I1008 18:33:04.981571 4988 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-conductor-0"] Oct 08 18:33:04 crc kubenswrapper[4988]: I1008 18:33:04.988374 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 08 18:33:04 crc kubenswrapper[4988]: I1008 18:33:04.993268 4988 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 08 18:33:04 crc kubenswrapper[4988]: I1008 18:33:04.995731 4988 scope.go:117] "RemoveContainer" containerID="6d8f1c5dc40a53cca8a87fffd5d06729664e7815b666a1eba260acf508457342" Oct 08 18:33:04 crc kubenswrapper[4988]: I1008 18:33:04.997555 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-conductor-0"] Oct 08 18:33:05 crc kubenswrapper[4988]: I1008 18:33:05.001706 4988 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-conductor-0"] Oct 08 18:33:05 crc kubenswrapper[4988]: I1008 18:33:05.015882 4988 scope.go:117] "RemoveContainer" containerID="6d8f1c5dc40a53cca8a87fffd5d06729664e7815b666a1eba260acf508457342" Oct 08 18:33:05 crc kubenswrapper[4988]: E1008 18:33:05.016791 4988 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6d8f1c5dc40a53cca8a87fffd5d06729664e7815b666a1eba260acf508457342\": container with ID starting with 6d8f1c5dc40a53cca8a87fffd5d06729664e7815b666a1eba260acf508457342 not found: ID does not exist" containerID="6d8f1c5dc40a53cca8a87fffd5d06729664e7815b666a1eba260acf508457342" Oct 08 18:33:05 crc kubenswrapper[4988]: I1008 18:33:05.016850 4988 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6d8f1c5dc40a53cca8a87fffd5d06729664e7815b666a1eba260acf508457342"} err="failed to get container status \"6d8f1c5dc40a53cca8a87fffd5d06729664e7815b666a1eba260acf508457342\": rpc error: code = NotFound desc = could not find container \"6d8f1c5dc40a53cca8a87fffd5d06729664e7815b666a1eba260acf508457342\": container with ID starting with 6d8f1c5dc40a53cca8a87fffd5d06729664e7815b666a1eba260acf508457342 not found: ID does not exist" Oct 08 18:33:05 crc kubenswrapper[4988]: I1008 18:33:05.016884 4988 scope.go:117] "RemoveContainer" containerID="19c72634c92910f59b5f356f56971cc3e2867c6c02b8cdd4a997a62cb452e3cb" Oct 08 18:33:05 crc kubenswrapper[4988]: I1008 18:33:05.050757 4988 scope.go:117] "RemoveContainer" containerID="590ae1792458f3fd70bbade362261ee71cafb4633235b6af45da637905058707" Oct 08 18:33:05 crc kubenswrapper[4988]: E1008 18:33:05.059327 4988 configmap.go:193] Couldn't get configMap openstack/rabbitmq-cell1-config-data: configmap "rabbitmq-cell1-config-data" not found Oct 08 18:33:05 crc kubenswrapper[4988]: E1008 18:33:05.059396 4988 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/e5a8f819-c91d-429e-b848-e9c444739623-config-data podName:e5a8f819-c91d-429e-b848-e9c444739623 nodeName:}" failed. No retries permitted until 2025-10-08 18:33:13.059366657 +0000 UTC m=+1338.509209427 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/configmap/e5a8f819-c91d-429e-b848-e9c444739623-config-data") pod "rabbitmq-cell1-server-0" (UID: "e5a8f819-c91d-429e-b848-e9c444739623") : configmap "rabbitmq-cell1-config-data" not found Oct 08 18:33:05 crc kubenswrapper[4988]: I1008 18:33:05.068665 4988 scope.go:117] "RemoveContainer" containerID="19c72634c92910f59b5f356f56971cc3e2867c6c02b8cdd4a997a62cb452e3cb" Oct 08 18:33:05 crc kubenswrapper[4988]: E1008 18:33:05.069026 4988 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"19c72634c92910f59b5f356f56971cc3e2867c6c02b8cdd4a997a62cb452e3cb\": container with ID starting with 19c72634c92910f59b5f356f56971cc3e2867c6c02b8cdd4a997a62cb452e3cb not found: ID does not exist" containerID="19c72634c92910f59b5f356f56971cc3e2867c6c02b8cdd4a997a62cb452e3cb" Oct 08 18:33:05 crc kubenswrapper[4988]: I1008 18:33:05.069056 4988 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"19c72634c92910f59b5f356f56971cc3e2867c6c02b8cdd4a997a62cb452e3cb"} err="failed to get container status \"19c72634c92910f59b5f356f56971cc3e2867c6c02b8cdd4a997a62cb452e3cb\": rpc error: code = NotFound desc = could not find container \"19c72634c92910f59b5f356f56971cc3e2867c6c02b8cdd4a997a62cb452e3cb\": container with ID starting with 19c72634c92910f59b5f356f56971cc3e2867c6c02b8cdd4a997a62cb452e3cb not found: ID does not exist" Oct 08 18:33:05 crc kubenswrapper[4988]: I1008 18:33:05.069079 4988 scope.go:117] "RemoveContainer" containerID="590ae1792458f3fd70bbade362261ee71cafb4633235b6af45da637905058707" Oct 08 18:33:05 crc kubenswrapper[4988]: E1008 18:33:05.069482 4988 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"590ae1792458f3fd70bbade362261ee71cafb4633235b6af45da637905058707\": container with ID starting with 590ae1792458f3fd70bbade362261ee71cafb4633235b6af45da637905058707 not found: ID does not exist" containerID="590ae1792458f3fd70bbade362261ee71cafb4633235b6af45da637905058707" Oct 08 18:33:05 crc kubenswrapper[4988]: I1008 18:33:05.069526 4988 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"590ae1792458f3fd70bbade362261ee71cafb4633235b6af45da637905058707"} err="failed to get container status \"590ae1792458f3fd70bbade362261ee71cafb4633235b6af45da637905058707\": rpc error: code = NotFound desc = could not find container \"590ae1792458f3fd70bbade362261ee71cafb4633235b6af45da637905058707\": container with ID starting with 590ae1792458f3fd70bbade362261ee71cafb4633235b6af45da637905058707 not found: ID does not exist" Oct 08 18:33:05 crc kubenswrapper[4988]: I1008 18:33:05.069552 4988 scope.go:117] "RemoveContainer" containerID="a81aba18e5412a571b8ceb649f5e04c6698ca4b21c30f54d96848df1d1a823b6" Oct 08 18:33:05 crc kubenswrapper[4988]: I1008 18:33:05.087115 4988 scope.go:117] "RemoveContainer" containerID="4cdfab290ec25786ba40c3a3d5d3687398ae45094d90768455e35c6642188b15" Oct 08 18:33:05 crc kubenswrapper[4988]: I1008 18:33:05.106925 4988 scope.go:117] "RemoveContainer" containerID="e18a49c9de8366d161727792956633e70d61d0fc8c2799f046b7a6f94af795e5" Oct 08 18:33:05 crc kubenswrapper[4988]: I1008 18:33:05.123853 4988 scope.go:117] "RemoveContainer" containerID="dcd1682ce23dc072a1b6bf3c5aeff64ca69361e18fa5cdcb9712e65170ddc37d" Oct 08 18:33:05 crc kubenswrapper[4988]: I1008 18:33:05.256370 4988 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="02a37054-4817-4151-a8cb-a9a45867c0d3" path="/var/lib/kubelet/pods/02a37054-4817-4151-a8cb-a9a45867c0d3/volumes" Oct 08 18:33:05 crc kubenswrapper[4988]: I1008 18:33:05.257018 4988 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0ec345a3-1744-4e78-bf13-91f0325296e4" path="/var/lib/kubelet/pods/0ec345a3-1744-4e78-bf13-91f0325296e4/volumes" Oct 08 18:33:05 crc kubenswrapper[4988]: I1008 18:33:05.257663 4988 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0efd1f53-90ad-4eb8-90f2-2ba9fc1593c2" path="/var/lib/kubelet/pods/0efd1f53-90ad-4eb8-90f2-2ba9fc1593c2/volumes" Oct 08 18:33:05 crc kubenswrapper[4988]: I1008 18:33:05.258584 4988 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1b24865e-cac2-490e-9e61-de22a379ff47" path="/var/lib/kubelet/pods/1b24865e-cac2-490e-9e61-de22a379ff47/volumes" Oct 08 18:33:05 crc kubenswrapper[4988]: I1008 18:33:05.258973 4988 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="28f73b7a-bb5e-43ec-a099-13a12cdb33d3" path="/var/lib/kubelet/pods/28f73b7a-bb5e-43ec-a099-13a12cdb33d3/volumes" Oct 08 18:33:05 crc kubenswrapper[4988]: I1008 18:33:05.259473 4988 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="32bb7401-6292-4ca6-98f4-15522125c3ee" path="/var/lib/kubelet/pods/32bb7401-6292-4ca6-98f4-15522125c3ee/volumes" Oct 08 18:33:05 crc kubenswrapper[4988]: I1008 18:33:05.260637 4988 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3503942b-2825-4006-80e8-8c4610b89997" path="/var/lib/kubelet/pods/3503942b-2825-4006-80e8-8c4610b89997/volumes" Oct 08 18:33:05 crc kubenswrapper[4988]: I1008 18:33:05.261180 4988 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3a57537b-6115-4249-8d16-3ab0c4b6d21e" path="/var/lib/kubelet/pods/3a57537b-6115-4249-8d16-3ab0c4b6d21e/volumes" Oct 08 18:33:05 crc kubenswrapper[4988]: I1008 18:33:05.261676 4988 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3e8c874a-029a-4d6a-bc2e-00aeed1636a7" path="/var/lib/kubelet/pods/3e8c874a-029a-4d6a-bc2e-00aeed1636a7/volumes" Oct 08 18:33:05 crc kubenswrapper[4988]: I1008 18:33:05.262452 4988 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3f7bddd3-86f8-46e9-a38e-1aa5e570d151" path="/var/lib/kubelet/pods/3f7bddd3-86f8-46e9-a38e-1aa5e570d151/volumes" Oct 08 18:33:05 crc kubenswrapper[4988]: I1008 18:33:05.262971 4988 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="56702a03-175d-4a54-b84f-e25cab037431" path="/var/lib/kubelet/pods/56702a03-175d-4a54-b84f-e25cab037431/volumes" Oct 08 18:33:05 crc kubenswrapper[4988]: I1008 18:33:05.263603 4988 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="79cbf065-a458-4dec-b6a7-805f515142e0" path="/var/lib/kubelet/pods/79cbf065-a458-4dec-b6a7-805f515142e0/volumes" Oct 08 18:33:05 crc kubenswrapper[4988]: I1008 18:33:05.264504 4988 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="83ef7610-23ac-4445-b75c-ae7206b4d810" path="/var/lib/kubelet/pods/83ef7610-23ac-4445-b75c-ae7206b4d810/volumes" Oct 08 18:33:05 crc kubenswrapper[4988]: I1008 18:33:05.265026 4988 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="85bc7f8f-fdb3-4abb-88d3-ed0c10982c4c" path="/var/lib/kubelet/pods/85bc7f8f-fdb3-4abb-88d3-ed0c10982c4c/volumes" Oct 08 18:33:05 crc kubenswrapper[4988]: I1008 18:33:05.265713 4988 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="99d77eed-40f0-4879-be96-abea3a807cd9" path="/var/lib/kubelet/pods/99d77eed-40f0-4879-be96-abea3a807cd9/volumes" Oct 08 18:33:05 crc kubenswrapper[4988]: I1008 18:33:05.266980 4988 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ab3315d2-3248-42f1-9834-2bab55fa9935" path="/var/lib/kubelet/pods/ab3315d2-3248-42f1-9834-2bab55fa9935/volumes" Oct 08 18:33:05 crc kubenswrapper[4988]: I1008 18:33:05.267743 4988 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6d1becf-4229-40bf-b53a-d8a9eb4b580e" path="/var/lib/kubelet/pods/b6d1becf-4229-40bf-b53a-d8a9eb4b580e/volumes" Oct 08 18:33:05 crc kubenswrapper[4988]: I1008 18:33:05.268558 4988 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bd9cfcdf-4770-4efa-a41e-7867ec85206f" path="/var/lib/kubelet/pods/bd9cfcdf-4770-4efa-a41e-7867ec85206f/volumes" Oct 08 18:33:05 crc kubenswrapper[4988]: I1008 18:33:05.269127 4988 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d7d648e7-3c02-4d26-993d-a901764b3329" path="/var/lib/kubelet/pods/d7d648e7-3c02-4d26-993d-a901764b3329/volumes" Oct 08 18:33:05 crc kubenswrapper[4988]: I1008 18:33:05.270706 4988 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f4e47608-89b4-45d1-88a6-fe75805d864d" path="/var/lib/kubelet/pods/f4e47608-89b4-45d1-88a6-fe75805d864d/volumes" Oct 08 18:33:05 crc kubenswrapper[4988]: I1008 18:33:05.320607 4988 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-cell1-server-0" podUID="e5a8f819-c91d-429e-b848-e9c444739623" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.104:5671: connect: connection refused" Oct 08 18:33:05 crc kubenswrapper[4988]: I1008 18:33:05.870006 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k8mnm\" (UniqueName: \"kubernetes.io/projected/11679d40-6eb6-456c-8996-cd9a946a6db1-kube-api-access-k8mnm\") pod \"keystone3e59-account-delete-vgpc9\" (UID: \"11679d40-6eb6-456c-8996-cd9a946a6db1\") " pod="openstack/keystone3e59-account-delete-vgpc9" Oct 08 18:33:05 crc kubenswrapper[4988]: E1008 18:33:05.873266 4988 projected.go:194] Error preparing data for projected volume kube-api-access-k8mnm for pod openstack/keystone3e59-account-delete-vgpc9: failed to fetch token: serviceaccounts "galera-openstack" not found Oct 08 18:33:05 crc kubenswrapper[4988]: E1008 18:33:05.873345 4988 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/11679d40-6eb6-456c-8996-cd9a946a6db1-kube-api-access-k8mnm podName:11679d40-6eb6-456c-8996-cd9a946a6db1 nodeName:}" failed. No retries permitted until 2025-10-08 18:33:09.87332343 +0000 UTC m=+1335.323166210 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-k8mnm" (UniqueName: "kubernetes.io/projected/11679d40-6eb6-456c-8996-cd9a946a6db1-kube-api-access-k8mnm") pod "keystone3e59-account-delete-vgpc9" (UID: "11679d40-6eb6-456c-8996-cd9a946a6db1") : failed to fetch token: serviceaccounts "galera-openstack" not found Oct 08 18:33:06 crc kubenswrapper[4988]: I1008 18:33:06.072630 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-6dbc6dc9bd-hk8qw" Oct 08 18:33:06 crc kubenswrapper[4988]: I1008 18:33:06.177863 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/26a3b0a9-6aca-490d-a6fb-08addef86fbb-public-tls-certs\") pod \"26a3b0a9-6aca-490d-a6fb-08addef86fbb\" (UID: \"26a3b0a9-6aca-490d-a6fb-08addef86fbb\") " Oct 08 18:33:06 crc kubenswrapper[4988]: I1008 18:33:06.178001 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hczwx\" (UniqueName: \"kubernetes.io/projected/26a3b0a9-6aca-490d-a6fb-08addef86fbb-kube-api-access-hczwx\") pod \"26a3b0a9-6aca-490d-a6fb-08addef86fbb\" (UID: \"26a3b0a9-6aca-490d-a6fb-08addef86fbb\") " Oct 08 18:33:06 crc kubenswrapper[4988]: I1008 18:33:06.178031 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/26a3b0a9-6aca-490d-a6fb-08addef86fbb-fernet-keys\") pod \"26a3b0a9-6aca-490d-a6fb-08addef86fbb\" (UID: \"26a3b0a9-6aca-490d-a6fb-08addef86fbb\") " Oct 08 18:33:06 crc kubenswrapper[4988]: I1008 18:33:06.178056 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/26a3b0a9-6aca-490d-a6fb-08addef86fbb-internal-tls-certs\") pod \"26a3b0a9-6aca-490d-a6fb-08addef86fbb\" (UID: \"26a3b0a9-6aca-490d-a6fb-08addef86fbb\") " Oct 08 18:33:06 crc kubenswrapper[4988]: I1008 18:33:06.178085 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/26a3b0a9-6aca-490d-a6fb-08addef86fbb-config-data\") pod \"26a3b0a9-6aca-490d-a6fb-08addef86fbb\" (UID: \"26a3b0a9-6aca-490d-a6fb-08addef86fbb\") " Oct 08 18:33:06 crc kubenswrapper[4988]: I1008 18:33:06.178189 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/26a3b0a9-6aca-490d-a6fb-08addef86fbb-scripts\") pod \"26a3b0a9-6aca-490d-a6fb-08addef86fbb\" (UID: \"26a3b0a9-6aca-490d-a6fb-08addef86fbb\") " Oct 08 18:33:06 crc kubenswrapper[4988]: I1008 18:33:06.178214 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/26a3b0a9-6aca-490d-a6fb-08addef86fbb-combined-ca-bundle\") pod \"26a3b0a9-6aca-490d-a6fb-08addef86fbb\" (UID: \"26a3b0a9-6aca-490d-a6fb-08addef86fbb\") " Oct 08 18:33:06 crc kubenswrapper[4988]: I1008 18:33:06.178234 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/26a3b0a9-6aca-490d-a6fb-08addef86fbb-credential-keys\") pod \"26a3b0a9-6aca-490d-a6fb-08addef86fbb\" (UID: \"26a3b0a9-6aca-490d-a6fb-08addef86fbb\") " Oct 08 18:33:06 crc kubenswrapper[4988]: I1008 18:33:06.197823 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/26a3b0a9-6aca-490d-a6fb-08addef86fbb-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "26a3b0a9-6aca-490d-a6fb-08addef86fbb" (UID: "26a3b0a9-6aca-490d-a6fb-08addef86fbb"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:33:06 crc kubenswrapper[4988]: I1008 18:33:06.209058 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/26a3b0a9-6aca-490d-a6fb-08addef86fbb-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "26a3b0a9-6aca-490d-a6fb-08addef86fbb" (UID: "26a3b0a9-6aca-490d-a6fb-08addef86fbb"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:33:06 crc kubenswrapper[4988]: I1008 18:33:06.217911 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/26a3b0a9-6aca-490d-a6fb-08addef86fbb-kube-api-access-hczwx" (OuterVolumeSpecName: "kube-api-access-hczwx") pod "26a3b0a9-6aca-490d-a6fb-08addef86fbb" (UID: "26a3b0a9-6aca-490d-a6fb-08addef86fbb"). InnerVolumeSpecName "kube-api-access-hczwx". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:33:06 crc kubenswrapper[4988]: I1008 18:33:06.218014 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/26a3b0a9-6aca-490d-a6fb-08addef86fbb-scripts" (OuterVolumeSpecName: "scripts") pod "26a3b0a9-6aca-490d-a6fb-08addef86fbb" (UID: "26a3b0a9-6aca-490d-a6fb-08addef86fbb"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:33:06 crc kubenswrapper[4988]: I1008 18:33:06.230162 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/26a3b0a9-6aca-490d-a6fb-08addef86fbb-config-data" (OuterVolumeSpecName: "config-data") pod "26a3b0a9-6aca-490d-a6fb-08addef86fbb" (UID: "26a3b0a9-6aca-490d-a6fb-08addef86fbb"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:33:06 crc kubenswrapper[4988]: I1008 18:33:06.249880 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/26a3b0a9-6aca-490d-a6fb-08addef86fbb-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "26a3b0a9-6aca-490d-a6fb-08addef86fbb" (UID: "26a3b0a9-6aca-490d-a6fb-08addef86fbb"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:33:06 crc kubenswrapper[4988]: I1008 18:33:06.261858 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/26a3b0a9-6aca-490d-a6fb-08addef86fbb-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "26a3b0a9-6aca-490d-a6fb-08addef86fbb" (UID: "26a3b0a9-6aca-490d-a6fb-08addef86fbb"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:33:06 crc kubenswrapper[4988]: I1008 18:33:06.274909 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/26a3b0a9-6aca-490d-a6fb-08addef86fbb-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "26a3b0a9-6aca-490d-a6fb-08addef86fbb" (UID: "26a3b0a9-6aca-490d-a6fb-08addef86fbb"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:33:06 crc kubenswrapper[4988]: I1008 18:33:06.280005 4988 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hczwx\" (UniqueName: \"kubernetes.io/projected/26a3b0a9-6aca-490d-a6fb-08addef86fbb-kube-api-access-hczwx\") on node \"crc\" DevicePath \"\"" Oct 08 18:33:06 crc kubenswrapper[4988]: I1008 18:33:06.280216 4988 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/26a3b0a9-6aca-490d-a6fb-08addef86fbb-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 08 18:33:06 crc kubenswrapper[4988]: I1008 18:33:06.280229 4988 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/26a3b0a9-6aca-490d-a6fb-08addef86fbb-fernet-keys\") on node \"crc\" DevicePath \"\"" Oct 08 18:33:06 crc kubenswrapper[4988]: I1008 18:33:06.280238 4988 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/26a3b0a9-6aca-490d-a6fb-08addef86fbb-config-data\") on node \"crc\" DevicePath \"\"" Oct 08 18:33:06 crc kubenswrapper[4988]: I1008 18:33:06.280246 4988 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/26a3b0a9-6aca-490d-a6fb-08addef86fbb-scripts\") on node \"crc\" DevicePath \"\"" Oct 08 18:33:06 crc kubenswrapper[4988]: I1008 18:33:06.280254 4988 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/26a3b0a9-6aca-490d-a6fb-08addef86fbb-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 18:33:06 crc kubenswrapper[4988]: I1008 18:33:06.280262 4988 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/26a3b0a9-6aca-490d-a6fb-08addef86fbb-credential-keys\") on node \"crc\" DevicePath \"\"" Oct 08 18:33:06 crc kubenswrapper[4988]: I1008 18:33:06.280272 4988 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/26a3b0a9-6aca-490d-a6fb-08addef86fbb-public-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 08 18:33:06 crc kubenswrapper[4988]: I1008 18:33:06.355281 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Oct 08 18:33:06 crc kubenswrapper[4988]: I1008 18:33:06.483357 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"e5a8f819-c91d-429e-b848-e9c444739623\" (UID: \"e5a8f819-c91d-429e-b848-e9c444739623\") " Oct 08 18:33:06 crc kubenswrapper[4988]: I1008 18:33:06.483481 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/e5a8f819-c91d-429e-b848-e9c444739623-pod-info\") pod \"e5a8f819-c91d-429e-b848-e9c444739623\" (UID: \"e5a8f819-c91d-429e-b848-e9c444739623\") " Oct 08 18:33:06 crc kubenswrapper[4988]: I1008 18:33:06.483507 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/e5a8f819-c91d-429e-b848-e9c444739623-rabbitmq-plugins\") pod \"e5a8f819-c91d-429e-b848-e9c444739623\" (UID: \"e5a8f819-c91d-429e-b848-e9c444739623\") " Oct 08 18:33:06 crc kubenswrapper[4988]: I1008 18:33:06.483544 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/e5a8f819-c91d-429e-b848-e9c444739623-plugins-conf\") pod \"e5a8f819-c91d-429e-b848-e9c444739623\" (UID: \"e5a8f819-c91d-429e-b848-e9c444739623\") " Oct 08 18:33:06 crc kubenswrapper[4988]: I1008 18:33:06.483590 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/e5a8f819-c91d-429e-b848-e9c444739623-erlang-cookie-secret\") pod \"e5a8f819-c91d-429e-b848-e9c444739623\" (UID: \"e5a8f819-c91d-429e-b848-e9c444739623\") " Oct 08 18:33:06 crc kubenswrapper[4988]: I1008 18:33:06.483629 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/e5a8f819-c91d-429e-b848-e9c444739623-rabbitmq-tls\") pod \"e5a8f819-c91d-429e-b848-e9c444739623\" (UID: \"e5a8f819-c91d-429e-b848-e9c444739623\") " Oct 08 18:33:06 crc kubenswrapper[4988]: I1008 18:33:06.483672 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/e5a8f819-c91d-429e-b848-e9c444739623-server-conf\") pod \"e5a8f819-c91d-429e-b848-e9c444739623\" (UID: \"e5a8f819-c91d-429e-b848-e9c444739623\") " Oct 08 18:33:06 crc kubenswrapper[4988]: I1008 18:33:06.483715 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/e5a8f819-c91d-429e-b848-e9c444739623-config-data\") pod \"e5a8f819-c91d-429e-b848-e9c444739623\" (UID: \"e5a8f819-c91d-429e-b848-e9c444739623\") " Oct 08 18:33:06 crc kubenswrapper[4988]: I1008 18:33:06.483757 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/e5a8f819-c91d-429e-b848-e9c444739623-rabbitmq-erlang-cookie\") pod \"e5a8f819-c91d-429e-b848-e9c444739623\" (UID: \"e5a8f819-c91d-429e-b848-e9c444739623\") " Oct 08 18:33:06 crc kubenswrapper[4988]: I1008 18:33:06.483818 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/e5a8f819-c91d-429e-b848-e9c444739623-rabbitmq-confd\") pod \"e5a8f819-c91d-429e-b848-e9c444739623\" (UID: \"e5a8f819-c91d-429e-b848-e9c444739623\") " Oct 08 18:33:06 crc kubenswrapper[4988]: I1008 18:33:06.483895 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zv5wc\" (UniqueName: \"kubernetes.io/projected/e5a8f819-c91d-429e-b848-e9c444739623-kube-api-access-zv5wc\") pod \"e5a8f819-c91d-429e-b848-e9c444739623\" (UID: \"e5a8f819-c91d-429e-b848-e9c444739623\") " Oct 08 18:33:06 crc kubenswrapper[4988]: I1008 18:33:06.484193 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e5a8f819-c91d-429e-b848-e9c444739623-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "e5a8f819-c91d-429e-b848-e9c444739623" (UID: "e5a8f819-c91d-429e-b848-e9c444739623"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 18:33:06 crc kubenswrapper[4988]: E1008 18:33:06.484293 4988 configmap.go:193] Couldn't get configMap openstack/rabbitmq-config-data: configmap "rabbitmq-config-data" not found Oct 08 18:33:06 crc kubenswrapper[4988]: E1008 18:33:06.484347 4988 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/830d6ec5-0015-4a88-b8e8-0fd47f64fd46-config-data podName:830d6ec5-0015-4a88-b8e8-0fd47f64fd46 nodeName:}" failed. No retries permitted until 2025-10-08 18:33:14.484330933 +0000 UTC m=+1339.934173723 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/configmap/830d6ec5-0015-4a88-b8e8-0fd47f64fd46-config-data") pod "rabbitmq-server-0" (UID: "830d6ec5-0015-4a88-b8e8-0fd47f64fd46") : configmap "rabbitmq-config-data" not found Oct 08 18:33:06 crc kubenswrapper[4988]: I1008 18:33:06.484472 4988 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/e5a8f819-c91d-429e-b848-e9c444739623-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Oct 08 18:33:06 crc kubenswrapper[4988]: I1008 18:33:06.484962 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e5a8f819-c91d-429e-b848-e9c444739623-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "e5a8f819-c91d-429e-b848-e9c444739623" (UID: "e5a8f819-c91d-429e-b848-e9c444739623"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 18:33:06 crc kubenswrapper[4988]: I1008 18:33:06.486014 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e5a8f819-c91d-429e-b848-e9c444739623-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "e5a8f819-c91d-429e-b848-e9c444739623" (UID: "e5a8f819-c91d-429e-b848-e9c444739623"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 18:33:06 crc kubenswrapper[4988]: I1008 18:33:06.486246 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage07-crc" (OuterVolumeSpecName: "persistence") pod "e5a8f819-c91d-429e-b848-e9c444739623" (UID: "e5a8f819-c91d-429e-b848-e9c444739623"). InnerVolumeSpecName "local-storage07-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 08 18:33:06 crc kubenswrapper[4988]: I1008 18:33:06.487992 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e5a8f819-c91d-429e-b848-e9c444739623-rabbitmq-tls" (OuterVolumeSpecName: "rabbitmq-tls") pod "e5a8f819-c91d-429e-b848-e9c444739623" (UID: "e5a8f819-c91d-429e-b848-e9c444739623"). InnerVolumeSpecName "rabbitmq-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:33:06 crc kubenswrapper[4988]: I1008 18:33:06.488337 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e5a8f819-c91d-429e-b848-e9c444739623-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "e5a8f819-c91d-429e-b848-e9c444739623" (UID: "e5a8f819-c91d-429e-b848-e9c444739623"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:33:06 crc kubenswrapper[4988]: I1008 18:33:06.489449 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e5a8f819-c91d-429e-b848-e9c444739623-kube-api-access-zv5wc" (OuterVolumeSpecName: "kube-api-access-zv5wc") pod "e5a8f819-c91d-429e-b848-e9c444739623" (UID: "e5a8f819-c91d-429e-b848-e9c444739623"). InnerVolumeSpecName "kube-api-access-zv5wc". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:33:06 crc kubenswrapper[4988]: I1008 18:33:06.490515 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/e5a8f819-c91d-429e-b848-e9c444739623-pod-info" (OuterVolumeSpecName: "pod-info") pod "e5a8f819-c91d-429e-b848-e9c444739623" (UID: "e5a8f819-c91d-429e-b848-e9c444739623"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Oct 08 18:33:06 crc kubenswrapper[4988]: I1008 18:33:06.506097 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e5a8f819-c91d-429e-b848-e9c444739623-config-data" (OuterVolumeSpecName: "config-data") pod "e5a8f819-c91d-429e-b848-e9c444739623" (UID: "e5a8f819-c91d-429e-b848-e9c444739623"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 18:33:06 crc kubenswrapper[4988]: I1008 18:33:06.525138 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e5a8f819-c91d-429e-b848-e9c444739623-server-conf" (OuterVolumeSpecName: "server-conf") pod "e5a8f819-c91d-429e-b848-e9c444739623" (UID: "e5a8f819-c91d-429e-b848-e9c444739623"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 18:33:06 crc kubenswrapper[4988]: I1008 18:33:06.586554 4988 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zv5wc\" (UniqueName: \"kubernetes.io/projected/e5a8f819-c91d-429e-b848-e9c444739623-kube-api-access-zv5wc\") on node \"crc\" DevicePath \"\"" Oct 08 18:33:06 crc kubenswrapper[4988]: I1008 18:33:06.586632 4988 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") on node \"crc\" " Oct 08 18:33:06 crc kubenswrapper[4988]: I1008 18:33:06.586653 4988 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/e5a8f819-c91d-429e-b848-e9c444739623-pod-info\") on node \"crc\" DevicePath \"\"" Oct 08 18:33:06 crc kubenswrapper[4988]: I1008 18:33:06.586671 4988 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/e5a8f819-c91d-429e-b848-e9c444739623-plugins-conf\") on node \"crc\" DevicePath \"\"" Oct 08 18:33:06 crc kubenswrapper[4988]: I1008 18:33:06.586688 4988 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/e5a8f819-c91d-429e-b848-e9c444739623-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Oct 08 18:33:06 crc kubenswrapper[4988]: I1008 18:33:06.586705 4988 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/e5a8f819-c91d-429e-b848-e9c444739623-rabbitmq-tls\") on node \"crc\" DevicePath \"\"" Oct 08 18:33:06 crc kubenswrapper[4988]: I1008 18:33:06.586722 4988 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/e5a8f819-c91d-429e-b848-e9c444739623-server-conf\") on node \"crc\" DevicePath \"\"" Oct 08 18:33:06 crc kubenswrapper[4988]: I1008 18:33:06.586739 4988 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/e5a8f819-c91d-429e-b848-e9c444739623-config-data\") on node \"crc\" DevicePath \"\"" Oct 08 18:33:06 crc kubenswrapper[4988]: I1008 18:33:06.586757 4988 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/e5a8f819-c91d-429e-b848-e9c444739623-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Oct 08 18:33:06 crc kubenswrapper[4988]: I1008 18:33:06.623049 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e5a8f819-c91d-429e-b848-e9c444739623-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "e5a8f819-c91d-429e-b848-e9c444739623" (UID: "e5a8f819-c91d-429e-b848-e9c444739623"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:33:06 crc kubenswrapper[4988]: I1008 18:33:06.629042 4988 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage07-crc" (UniqueName: "kubernetes.io/local-volume/local-storage07-crc") on node "crc" Oct 08 18:33:06 crc kubenswrapper[4988]: I1008 18:33:06.674946 4988 generic.go:334] "Generic (PLEG): container finished" podID="e5a8f819-c91d-429e-b848-e9c444739623" containerID="bb5d4834c48dac50c4e9cebb831a5dc39438bd589b127972d2afef8419d96b2f" exitCode=0 Oct 08 18:33:06 crc kubenswrapper[4988]: I1008 18:33:06.675007 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"e5a8f819-c91d-429e-b848-e9c444739623","Type":"ContainerDied","Data":"bb5d4834c48dac50c4e9cebb831a5dc39438bd589b127972d2afef8419d96b2f"} Oct 08 18:33:06 crc kubenswrapper[4988]: I1008 18:33:06.675036 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"e5a8f819-c91d-429e-b848-e9c444739623","Type":"ContainerDied","Data":"2a6845add4d60a97c03ae806a22c38e8be0924372a13cb9a5e2ec5177a1ec2dc"} Oct 08 18:33:06 crc kubenswrapper[4988]: I1008 18:33:06.675058 4988 scope.go:117] "RemoveContainer" containerID="bb5d4834c48dac50c4e9cebb831a5dc39438bd589b127972d2afef8419d96b2f" Oct 08 18:33:06 crc kubenswrapper[4988]: I1008 18:33:06.675193 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Oct 08 18:33:06 crc kubenswrapper[4988]: I1008 18:33:06.688029 4988 reconciler_common.go:293] "Volume detached for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") on node \"crc\" DevicePath \"\"" Oct 08 18:33:06 crc kubenswrapper[4988]: I1008 18:33:06.688054 4988 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/e5a8f819-c91d-429e-b848-e9c444739623-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Oct 08 18:33:06 crc kubenswrapper[4988]: I1008 18:33:06.689690 4988 generic.go:334] "Generic (PLEG): container finished" podID="830d6ec5-0015-4a88-b8e8-0fd47f64fd46" containerID="7aeabf72c789f76cad92e2b157ae1d52318b0fdd5da4ef59e8daac0afedb3f6a" exitCode=0 Oct 08 18:33:06 crc kubenswrapper[4988]: I1008 18:33:06.689786 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"830d6ec5-0015-4a88-b8e8-0fd47f64fd46","Type":"ContainerDied","Data":"7aeabf72c789f76cad92e2b157ae1d52318b0fdd5da4ef59e8daac0afedb3f6a"} Oct 08 18:33:06 crc kubenswrapper[4988]: I1008 18:33:06.694544 4988 generic.go:334] "Generic (PLEG): container finished" podID="26a3b0a9-6aca-490d-a6fb-08addef86fbb" containerID="62d82328b02c86e078d53d3e4ba279c27a7c17092bf7900027389a74f0ad59d0" exitCode=0 Oct 08 18:33:06 crc kubenswrapper[4988]: I1008 18:33:06.694594 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-6dbc6dc9bd-hk8qw" event={"ID":"26a3b0a9-6aca-490d-a6fb-08addef86fbb","Type":"ContainerDied","Data":"62d82328b02c86e078d53d3e4ba279c27a7c17092bf7900027389a74f0ad59d0"} Oct 08 18:33:06 crc kubenswrapper[4988]: I1008 18:33:06.694627 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-6dbc6dc9bd-hk8qw" event={"ID":"26a3b0a9-6aca-490d-a6fb-08addef86fbb","Type":"ContainerDied","Data":"7ee10840af798462ce83a151c6143096bcd55ba704c55b5188685444a407217d"} Oct 08 18:33:06 crc kubenswrapper[4988]: I1008 18:33:06.694653 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-6dbc6dc9bd-hk8qw" Oct 08 18:33:06 crc kubenswrapper[4988]: I1008 18:33:06.726870 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 08 18:33:06 crc kubenswrapper[4988]: I1008 18:33:06.730324 4988 scope.go:117] "RemoveContainer" containerID="586d76b71bf326ca8c68b69f484afc42dbe51e835567ee76b7228cb18c8b1384" Oct 08 18:33:06 crc kubenswrapper[4988]: I1008 18:33:06.738211 4988 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 08 18:33:06 crc kubenswrapper[4988]: I1008 18:33:06.757310 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-6dbc6dc9bd-hk8qw"] Oct 08 18:33:06 crc kubenswrapper[4988]: I1008 18:33:06.766139 4988 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-6dbc6dc9bd-hk8qw"] Oct 08 18:33:06 crc kubenswrapper[4988]: E1008 18:33:06.769005 4988 handlers.go:78] "Exec lifecycle hook for Container in Pod failed" err=< Oct 08 18:33:06 crc kubenswrapper[4988]: command '/usr/share/ovn/scripts/ovn-ctl stop_controller' exited with 137: 2025-10-08T18:32:59Z|00001|fatal_signal|WARN|terminating with signal 14 (Alarm clock) Oct 08 18:33:06 crc kubenswrapper[4988]: /etc/init.d/functions: line 589: 400 Alarm clock "$@" Oct 08 18:33:06 crc kubenswrapper[4988]: > execCommand=["/usr/share/ovn/scripts/ovn-ctl","stop_controller"] containerName="ovn-controller" pod="openstack/ovn-controller-gp8f4" message=< Oct 08 18:33:06 crc kubenswrapper[4988]: Exiting ovn-controller (1) [FAILED] Oct 08 18:33:06 crc kubenswrapper[4988]: Killing ovn-controller (1) [ OK ] Oct 08 18:33:06 crc kubenswrapper[4988]: 2025-10-08T18:32:59Z|00001|fatal_signal|WARN|terminating with signal 14 (Alarm clock) Oct 08 18:33:06 crc kubenswrapper[4988]: /etc/init.d/functions: line 589: 400 Alarm clock "$@" Oct 08 18:33:06 crc kubenswrapper[4988]: > Oct 08 18:33:06 crc kubenswrapper[4988]: E1008 18:33:06.769033 4988 kuberuntime_container.go:691] "PreStop hook failed" err=< Oct 08 18:33:06 crc kubenswrapper[4988]: command '/usr/share/ovn/scripts/ovn-ctl stop_controller' exited with 137: 2025-10-08T18:32:59Z|00001|fatal_signal|WARN|terminating with signal 14 (Alarm clock) Oct 08 18:33:06 crc kubenswrapper[4988]: /etc/init.d/functions: line 589: 400 Alarm clock "$@" Oct 08 18:33:06 crc kubenswrapper[4988]: > pod="openstack/ovn-controller-gp8f4" podUID="c285e09d-c2f5-4642-9e6b-59b0ed3ed651" containerName="ovn-controller" containerID="cri-o://f3d6a7298803e1d31ec12101ceb0dd0c2936d1871327a38880d53ef51a7deceb" Oct 08 18:33:06 crc kubenswrapper[4988]: I1008 18:33:06.769075 4988 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ovn-controller-gp8f4" podUID="c285e09d-c2f5-4642-9e6b-59b0ed3ed651" containerName="ovn-controller" containerID="cri-o://f3d6a7298803e1d31ec12101ceb0dd0c2936d1871327a38880d53ef51a7deceb" gracePeriod=22 Oct 08 18:33:06 crc kubenswrapper[4988]: I1008 18:33:06.782555 4988 scope.go:117] "RemoveContainer" containerID="bb5d4834c48dac50c4e9cebb831a5dc39438bd589b127972d2afef8419d96b2f" Oct 08 18:33:06 crc kubenswrapper[4988]: E1008 18:33:06.783024 4988 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bb5d4834c48dac50c4e9cebb831a5dc39438bd589b127972d2afef8419d96b2f\": container with ID starting with bb5d4834c48dac50c4e9cebb831a5dc39438bd589b127972d2afef8419d96b2f not found: ID does not exist" containerID="bb5d4834c48dac50c4e9cebb831a5dc39438bd589b127972d2afef8419d96b2f" Oct 08 18:33:06 crc kubenswrapper[4988]: I1008 18:33:06.783071 4988 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bb5d4834c48dac50c4e9cebb831a5dc39438bd589b127972d2afef8419d96b2f"} err="failed to get container status \"bb5d4834c48dac50c4e9cebb831a5dc39438bd589b127972d2afef8419d96b2f\": rpc error: code = NotFound desc = could not find container \"bb5d4834c48dac50c4e9cebb831a5dc39438bd589b127972d2afef8419d96b2f\": container with ID starting with bb5d4834c48dac50c4e9cebb831a5dc39438bd589b127972d2afef8419d96b2f not found: ID does not exist" Oct 08 18:33:06 crc kubenswrapper[4988]: I1008 18:33:06.783101 4988 scope.go:117] "RemoveContainer" containerID="586d76b71bf326ca8c68b69f484afc42dbe51e835567ee76b7228cb18c8b1384" Oct 08 18:33:06 crc kubenswrapper[4988]: E1008 18:33:06.783479 4988 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"586d76b71bf326ca8c68b69f484afc42dbe51e835567ee76b7228cb18c8b1384\": container with ID starting with 586d76b71bf326ca8c68b69f484afc42dbe51e835567ee76b7228cb18c8b1384 not found: ID does not exist" containerID="586d76b71bf326ca8c68b69f484afc42dbe51e835567ee76b7228cb18c8b1384" Oct 08 18:33:06 crc kubenswrapper[4988]: I1008 18:33:06.783504 4988 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"586d76b71bf326ca8c68b69f484afc42dbe51e835567ee76b7228cb18c8b1384"} err="failed to get container status \"586d76b71bf326ca8c68b69f484afc42dbe51e835567ee76b7228cb18c8b1384\": rpc error: code = NotFound desc = could not find container \"586d76b71bf326ca8c68b69f484afc42dbe51e835567ee76b7228cb18c8b1384\": container with ID starting with 586d76b71bf326ca8c68b69f484afc42dbe51e835567ee76b7228cb18c8b1384 not found: ID does not exist" Oct 08 18:33:06 crc kubenswrapper[4988]: I1008 18:33:06.783524 4988 scope.go:117] "RemoveContainer" containerID="62d82328b02c86e078d53d3e4ba279c27a7c17092bf7900027389a74f0ad59d0" Oct 08 18:33:06 crc kubenswrapper[4988]: I1008 18:33:06.848144 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Oct 08 18:33:06 crc kubenswrapper[4988]: I1008 18:33:06.928189 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-create-fxlld"] Oct 08 18:33:06 crc kubenswrapper[4988]: I1008 18:33:06.935549 4988 scope.go:117] "RemoveContainer" containerID="62d82328b02c86e078d53d3e4ba279c27a7c17092bf7900027389a74f0ad59d0" Oct 08 18:33:06 crc kubenswrapper[4988]: E1008 18:33:06.935929 4988 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"62d82328b02c86e078d53d3e4ba279c27a7c17092bf7900027389a74f0ad59d0\": container with ID starting with 62d82328b02c86e078d53d3e4ba279c27a7c17092bf7900027389a74f0ad59d0 not found: ID does not exist" containerID="62d82328b02c86e078d53d3e4ba279c27a7c17092bf7900027389a74f0ad59d0" Oct 08 18:33:06 crc kubenswrapper[4988]: I1008 18:33:06.935973 4988 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"62d82328b02c86e078d53d3e4ba279c27a7c17092bf7900027389a74f0ad59d0"} err="failed to get container status \"62d82328b02c86e078d53d3e4ba279c27a7c17092bf7900027389a74f0ad59d0\": rpc error: code = NotFound desc = could not find container \"62d82328b02c86e078d53d3e4ba279c27a7c17092bf7900027389a74f0ad59d0\": container with ID starting with 62d82328b02c86e078d53d3e4ba279c27a7c17092bf7900027389a74f0ad59d0 not found: ID does not exist" Oct 08 18:33:06 crc kubenswrapper[4988]: I1008 18:33:06.940267 4988 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-create-fxlld"] Oct 08 18:33:06 crc kubenswrapper[4988]: I1008 18:33:06.946899 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-3e59-account-create-rv4kt"] Oct 08 18:33:06 crc kubenswrapper[4988]: I1008 18:33:06.954529 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone3e59-account-delete-vgpc9"] Oct 08 18:33:06 crc kubenswrapper[4988]: E1008 18:33:06.955210 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="unmounted volumes=[kube-api-access-k8mnm], unattached volumes=[], failed to process volumes=[]: context canceled" pod="openstack/keystone3e59-account-delete-vgpc9" podUID="11679d40-6eb6-456c-8996-cd9a946a6db1" Oct 08 18:33:06 crc kubenswrapper[4988]: I1008 18:33:06.960796 4988 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-3e59-account-create-rv4kt"] Oct 08 18:33:06 crc kubenswrapper[4988]: I1008 18:33:06.991490 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tmhl7\" (UniqueName: \"kubernetes.io/projected/830d6ec5-0015-4a88-b8e8-0fd47f64fd46-kube-api-access-tmhl7\") pod \"830d6ec5-0015-4a88-b8e8-0fd47f64fd46\" (UID: \"830d6ec5-0015-4a88-b8e8-0fd47f64fd46\") " Oct 08 18:33:06 crc kubenswrapper[4988]: I1008 18:33:06.991530 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/830d6ec5-0015-4a88-b8e8-0fd47f64fd46-rabbitmq-tls\") pod \"830d6ec5-0015-4a88-b8e8-0fd47f64fd46\" (UID: \"830d6ec5-0015-4a88-b8e8-0fd47f64fd46\") " Oct 08 18:33:06 crc kubenswrapper[4988]: I1008 18:33:06.991582 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/830d6ec5-0015-4a88-b8e8-0fd47f64fd46-plugins-conf\") pod \"830d6ec5-0015-4a88-b8e8-0fd47f64fd46\" (UID: \"830d6ec5-0015-4a88-b8e8-0fd47f64fd46\") " Oct 08 18:33:06 crc kubenswrapper[4988]: I1008 18:33:06.991605 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/830d6ec5-0015-4a88-b8e8-0fd47f64fd46-rabbitmq-erlang-cookie\") pod \"830d6ec5-0015-4a88-b8e8-0fd47f64fd46\" (UID: \"830d6ec5-0015-4a88-b8e8-0fd47f64fd46\") " Oct 08 18:33:06 crc kubenswrapper[4988]: I1008 18:33:06.991645 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/830d6ec5-0015-4a88-b8e8-0fd47f64fd46-pod-info\") pod \"830d6ec5-0015-4a88-b8e8-0fd47f64fd46\" (UID: \"830d6ec5-0015-4a88-b8e8-0fd47f64fd46\") " Oct 08 18:33:06 crc kubenswrapper[4988]: I1008 18:33:06.991660 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"830d6ec5-0015-4a88-b8e8-0fd47f64fd46\" (UID: \"830d6ec5-0015-4a88-b8e8-0fd47f64fd46\") " Oct 08 18:33:06 crc kubenswrapper[4988]: I1008 18:33:06.991686 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/830d6ec5-0015-4a88-b8e8-0fd47f64fd46-server-conf\") pod \"830d6ec5-0015-4a88-b8e8-0fd47f64fd46\" (UID: \"830d6ec5-0015-4a88-b8e8-0fd47f64fd46\") " Oct 08 18:33:06 crc kubenswrapper[4988]: I1008 18:33:06.991709 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/830d6ec5-0015-4a88-b8e8-0fd47f64fd46-erlang-cookie-secret\") pod \"830d6ec5-0015-4a88-b8e8-0fd47f64fd46\" (UID: \"830d6ec5-0015-4a88-b8e8-0fd47f64fd46\") " Oct 08 18:33:06 crc kubenswrapper[4988]: I1008 18:33:06.991734 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/830d6ec5-0015-4a88-b8e8-0fd47f64fd46-rabbitmq-plugins\") pod \"830d6ec5-0015-4a88-b8e8-0fd47f64fd46\" (UID: \"830d6ec5-0015-4a88-b8e8-0fd47f64fd46\") " Oct 08 18:33:06 crc kubenswrapper[4988]: I1008 18:33:06.991753 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/830d6ec5-0015-4a88-b8e8-0fd47f64fd46-rabbitmq-confd\") pod \"830d6ec5-0015-4a88-b8e8-0fd47f64fd46\" (UID: \"830d6ec5-0015-4a88-b8e8-0fd47f64fd46\") " Oct 08 18:33:06 crc kubenswrapper[4988]: I1008 18:33:06.991813 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/830d6ec5-0015-4a88-b8e8-0fd47f64fd46-config-data\") pod \"830d6ec5-0015-4a88-b8e8-0fd47f64fd46\" (UID: \"830d6ec5-0015-4a88-b8e8-0fd47f64fd46\") " Oct 08 18:33:06 crc kubenswrapper[4988]: I1008 18:33:06.992502 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/830d6ec5-0015-4a88-b8e8-0fd47f64fd46-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "830d6ec5-0015-4a88-b8e8-0fd47f64fd46" (UID: "830d6ec5-0015-4a88-b8e8-0fd47f64fd46"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 18:33:06 crc kubenswrapper[4988]: I1008 18:33:06.992999 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/830d6ec5-0015-4a88-b8e8-0fd47f64fd46-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "830d6ec5-0015-4a88-b8e8-0fd47f64fd46" (UID: "830d6ec5-0015-4a88-b8e8-0fd47f64fd46"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 18:33:06 crc kubenswrapper[4988]: I1008 18:33:06.993052 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/830d6ec5-0015-4a88-b8e8-0fd47f64fd46-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "830d6ec5-0015-4a88-b8e8-0fd47f64fd46" (UID: "830d6ec5-0015-4a88-b8e8-0fd47f64fd46"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 18:33:06 crc kubenswrapper[4988]: I1008 18:33:06.995325 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/830d6ec5-0015-4a88-b8e8-0fd47f64fd46-kube-api-access-tmhl7" (OuterVolumeSpecName: "kube-api-access-tmhl7") pod "830d6ec5-0015-4a88-b8e8-0fd47f64fd46" (UID: "830d6ec5-0015-4a88-b8e8-0fd47f64fd46"). InnerVolumeSpecName "kube-api-access-tmhl7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:33:06 crc kubenswrapper[4988]: I1008 18:33:06.995341 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage04-crc" (OuterVolumeSpecName: "persistence") pod "830d6ec5-0015-4a88-b8e8-0fd47f64fd46" (UID: "830d6ec5-0015-4a88-b8e8-0fd47f64fd46"). InnerVolumeSpecName "local-storage04-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 08 18:33:06 crc kubenswrapper[4988]: I1008 18:33:06.996218 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/830d6ec5-0015-4a88-b8e8-0fd47f64fd46-rabbitmq-tls" (OuterVolumeSpecName: "rabbitmq-tls") pod "830d6ec5-0015-4a88-b8e8-0fd47f64fd46" (UID: "830d6ec5-0015-4a88-b8e8-0fd47f64fd46"). InnerVolumeSpecName "rabbitmq-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:33:06 crc kubenswrapper[4988]: I1008 18:33:06.997164 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/830d6ec5-0015-4a88-b8e8-0fd47f64fd46-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "830d6ec5-0015-4a88-b8e8-0fd47f64fd46" (UID: "830d6ec5-0015-4a88-b8e8-0fd47f64fd46"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:33:07 crc kubenswrapper[4988]: I1008 18:33:07.005555 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/830d6ec5-0015-4a88-b8e8-0fd47f64fd46-pod-info" (OuterVolumeSpecName: "pod-info") pod "830d6ec5-0015-4a88-b8e8-0fd47f64fd46" (UID: "830d6ec5-0015-4a88-b8e8-0fd47f64fd46"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Oct 08 18:33:07 crc kubenswrapper[4988]: I1008 18:33:07.015703 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/830d6ec5-0015-4a88-b8e8-0fd47f64fd46-config-data" (OuterVolumeSpecName: "config-data") pod "830d6ec5-0015-4a88-b8e8-0fd47f64fd46" (UID: "830d6ec5-0015-4a88-b8e8-0fd47f64fd46"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 18:33:07 crc kubenswrapper[4988]: I1008 18:33:07.028272 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/830d6ec5-0015-4a88-b8e8-0fd47f64fd46-server-conf" (OuterVolumeSpecName: "server-conf") pod "830d6ec5-0015-4a88-b8e8-0fd47f64fd46" (UID: "830d6ec5-0015-4a88-b8e8-0fd47f64fd46"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 18:33:07 crc kubenswrapper[4988]: I1008 18:33:07.093885 4988 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/830d6ec5-0015-4a88-b8e8-0fd47f64fd46-plugins-conf\") on node \"crc\" DevicePath \"\"" Oct 08 18:33:07 crc kubenswrapper[4988]: I1008 18:33:07.093931 4988 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/830d6ec5-0015-4a88-b8e8-0fd47f64fd46-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Oct 08 18:33:07 crc kubenswrapper[4988]: I1008 18:33:07.093946 4988 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/830d6ec5-0015-4a88-b8e8-0fd47f64fd46-pod-info\") on node \"crc\" DevicePath \"\"" Oct 08 18:33:07 crc kubenswrapper[4988]: I1008 18:33:07.093979 4988 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") on node \"crc\" " Oct 08 18:33:07 crc kubenswrapper[4988]: I1008 18:33:07.093991 4988 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/830d6ec5-0015-4a88-b8e8-0fd47f64fd46-server-conf\") on node \"crc\" DevicePath \"\"" Oct 08 18:33:07 crc kubenswrapper[4988]: I1008 18:33:07.094007 4988 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/830d6ec5-0015-4a88-b8e8-0fd47f64fd46-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Oct 08 18:33:07 crc kubenswrapper[4988]: I1008 18:33:07.094020 4988 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/830d6ec5-0015-4a88-b8e8-0fd47f64fd46-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Oct 08 18:33:07 crc kubenswrapper[4988]: I1008 18:33:07.094033 4988 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/830d6ec5-0015-4a88-b8e8-0fd47f64fd46-config-data\") on node \"crc\" DevicePath \"\"" Oct 08 18:33:07 crc kubenswrapper[4988]: I1008 18:33:07.094044 4988 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tmhl7\" (UniqueName: \"kubernetes.io/projected/830d6ec5-0015-4a88-b8e8-0fd47f64fd46-kube-api-access-tmhl7\") on node \"crc\" DevicePath \"\"" Oct 08 18:33:07 crc kubenswrapper[4988]: I1008 18:33:07.094056 4988 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/830d6ec5-0015-4a88-b8e8-0fd47f64fd46-rabbitmq-tls\") on node \"crc\" DevicePath \"\"" Oct 08 18:33:07 crc kubenswrapper[4988]: I1008 18:33:07.124832 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/830d6ec5-0015-4a88-b8e8-0fd47f64fd46-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "830d6ec5-0015-4a88-b8e8-0fd47f64fd46" (UID: "830d6ec5-0015-4a88-b8e8-0fd47f64fd46"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:33:07 crc kubenswrapper[4988]: I1008 18:33:07.134564 4988 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage04-crc" (UniqueName: "kubernetes.io/local-volume/local-storage04-crc") on node "crc" Oct 08 18:33:07 crc kubenswrapper[4988]: I1008 18:33:07.195985 4988 reconciler_common.go:293] "Volume detached for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") on node \"crc\" DevicePath \"\"" Oct 08 18:33:07 crc kubenswrapper[4988]: I1008 18:33:07.196021 4988 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/830d6ec5-0015-4a88-b8e8-0fd47f64fd46-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Oct 08 18:33:07 crc kubenswrapper[4988]: I1008 18:33:07.251216 4988 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="26a3b0a9-6aca-490d-a6fb-08addef86fbb" path="/var/lib/kubelet/pods/26a3b0a9-6aca-490d-a6fb-08addef86fbb/volumes" Oct 08 18:33:07 crc kubenswrapper[4988]: I1008 18:33:07.252197 4988 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ad40a4d0-13f5-4e86-ba37-db0add482b46" path="/var/lib/kubelet/pods/ad40a4d0-13f5-4e86-ba37-db0add482b46/volumes" Oct 08 18:33:07 crc kubenswrapper[4988]: I1008 18:33:07.252812 4988 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b1336bd1-4eab-44f9-9b5b-9ae2aa3b379c" path="/var/lib/kubelet/pods/b1336bd1-4eab-44f9-9b5b-9ae2aa3b379c/volumes" Oct 08 18:33:07 crc kubenswrapper[4988]: I1008 18:33:07.253994 4988 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e5a8f819-c91d-429e-b848-e9c444739623" path="/var/lib/kubelet/pods/e5a8f819-c91d-429e-b848-e9c444739623/volumes" Oct 08 18:33:07 crc kubenswrapper[4988]: I1008 18:33:07.333836 4988 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-gp8f4_c285e09d-c2f5-4642-9e6b-59b0ed3ed651/ovn-controller/0.log" Oct 08 18:33:07 crc kubenswrapper[4988]: I1008 18:33:07.333933 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-gp8f4" Oct 08 18:33:07 crc kubenswrapper[4988]: I1008 18:33:07.500743 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/c285e09d-c2f5-4642-9e6b-59b0ed3ed651-ovn-controller-tls-certs\") pod \"c285e09d-c2f5-4642-9e6b-59b0ed3ed651\" (UID: \"c285e09d-c2f5-4642-9e6b-59b0ed3ed651\") " Oct 08 18:33:07 crc kubenswrapper[4988]: I1008 18:33:07.500785 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/c285e09d-c2f5-4642-9e6b-59b0ed3ed651-var-run-ovn\") pod \"c285e09d-c2f5-4642-9e6b-59b0ed3ed651\" (UID: \"c285e09d-c2f5-4642-9e6b-59b0ed3ed651\") " Oct 08 18:33:07 crc kubenswrapper[4988]: I1008 18:33:07.500831 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v2tq6\" (UniqueName: \"kubernetes.io/projected/c285e09d-c2f5-4642-9e6b-59b0ed3ed651-kube-api-access-v2tq6\") pod \"c285e09d-c2f5-4642-9e6b-59b0ed3ed651\" (UID: \"c285e09d-c2f5-4642-9e6b-59b0ed3ed651\") " Oct 08 18:33:07 crc kubenswrapper[4988]: I1008 18:33:07.500941 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/c285e09d-c2f5-4642-9e6b-59b0ed3ed651-var-run\") pod \"c285e09d-c2f5-4642-9e6b-59b0ed3ed651\" (UID: \"c285e09d-c2f5-4642-9e6b-59b0ed3ed651\") " Oct 08 18:33:07 crc kubenswrapper[4988]: I1008 18:33:07.500974 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c285e09d-c2f5-4642-9e6b-59b0ed3ed651-scripts\") pod \"c285e09d-c2f5-4642-9e6b-59b0ed3ed651\" (UID: \"c285e09d-c2f5-4642-9e6b-59b0ed3ed651\") " Oct 08 18:33:07 crc kubenswrapper[4988]: I1008 18:33:07.501004 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/c285e09d-c2f5-4642-9e6b-59b0ed3ed651-var-log-ovn\") pod \"c285e09d-c2f5-4642-9e6b-59b0ed3ed651\" (UID: \"c285e09d-c2f5-4642-9e6b-59b0ed3ed651\") " Oct 08 18:33:07 crc kubenswrapper[4988]: I1008 18:33:07.501023 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c285e09d-c2f5-4642-9e6b-59b0ed3ed651-combined-ca-bundle\") pod \"c285e09d-c2f5-4642-9e6b-59b0ed3ed651\" (UID: \"c285e09d-c2f5-4642-9e6b-59b0ed3ed651\") " Oct 08 18:33:07 crc kubenswrapper[4988]: I1008 18:33:07.501316 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/c285e09d-c2f5-4642-9e6b-59b0ed3ed651-var-run" (OuterVolumeSpecName: "var-run") pod "c285e09d-c2f5-4642-9e6b-59b0ed3ed651" (UID: "c285e09d-c2f5-4642-9e6b-59b0ed3ed651"). InnerVolumeSpecName "var-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 08 18:33:07 crc kubenswrapper[4988]: I1008 18:33:07.501335 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/c285e09d-c2f5-4642-9e6b-59b0ed3ed651-var-run-ovn" (OuterVolumeSpecName: "var-run-ovn") pod "c285e09d-c2f5-4642-9e6b-59b0ed3ed651" (UID: "c285e09d-c2f5-4642-9e6b-59b0ed3ed651"). InnerVolumeSpecName "var-run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 08 18:33:07 crc kubenswrapper[4988]: I1008 18:33:07.501367 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/c285e09d-c2f5-4642-9e6b-59b0ed3ed651-var-log-ovn" (OuterVolumeSpecName: "var-log-ovn") pod "c285e09d-c2f5-4642-9e6b-59b0ed3ed651" (UID: "c285e09d-c2f5-4642-9e6b-59b0ed3ed651"). InnerVolumeSpecName "var-log-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 08 18:33:07 crc kubenswrapper[4988]: I1008 18:33:07.501560 4988 reconciler_common.go:293] "Volume detached for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/c285e09d-c2f5-4642-9e6b-59b0ed3ed651-var-run\") on node \"crc\" DevicePath \"\"" Oct 08 18:33:07 crc kubenswrapper[4988]: I1008 18:33:07.501576 4988 reconciler_common.go:293] "Volume detached for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/c285e09d-c2f5-4642-9e6b-59b0ed3ed651-var-log-ovn\") on node \"crc\" DevicePath \"\"" Oct 08 18:33:07 crc kubenswrapper[4988]: I1008 18:33:07.501589 4988 reconciler_common.go:293] "Volume detached for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/c285e09d-c2f5-4642-9e6b-59b0ed3ed651-var-run-ovn\") on node \"crc\" DevicePath \"\"" Oct 08 18:33:07 crc kubenswrapper[4988]: I1008 18:33:07.503298 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c285e09d-c2f5-4642-9e6b-59b0ed3ed651-scripts" (OuterVolumeSpecName: "scripts") pod "c285e09d-c2f5-4642-9e6b-59b0ed3ed651" (UID: "c285e09d-c2f5-4642-9e6b-59b0ed3ed651"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 18:33:07 crc kubenswrapper[4988]: I1008 18:33:07.517892 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c285e09d-c2f5-4642-9e6b-59b0ed3ed651-kube-api-access-v2tq6" (OuterVolumeSpecName: "kube-api-access-v2tq6") pod "c285e09d-c2f5-4642-9e6b-59b0ed3ed651" (UID: "c285e09d-c2f5-4642-9e6b-59b0ed3ed651"). InnerVolumeSpecName "kube-api-access-v2tq6". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:33:07 crc kubenswrapper[4988]: E1008 18:33:07.518365 4988 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="239fe689949862e30d5f5b0021f848adb4380b18bdefdeb68e24c7cd3fb41c24" cmd=["/bin/bash","/var/lib/operator-scripts/mysql_probe.sh","readiness"] Oct 08 18:33:07 crc kubenswrapper[4988]: E1008 18:33:07.519775 4988 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="239fe689949862e30d5f5b0021f848adb4380b18bdefdeb68e24c7cd3fb41c24" cmd=["/bin/bash","/var/lib/operator-scripts/mysql_probe.sh","readiness"] Oct 08 18:33:07 crc kubenswrapper[4988]: E1008 18:33:07.521490 4988 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="239fe689949862e30d5f5b0021f848adb4380b18bdefdeb68e24c7cd3fb41c24" cmd=["/bin/bash","/var/lib/operator-scripts/mysql_probe.sh","readiness"] Oct 08 18:33:07 crc kubenswrapper[4988]: E1008 18:33:07.521525 4988 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/openstack-galera-0" podUID="72c64965-0b57-43e1-8d18-2bce6cabbfa1" containerName="galera" Oct 08 18:33:07 crc kubenswrapper[4988]: I1008 18:33:07.540004 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c285e09d-c2f5-4642-9e6b-59b0ed3ed651-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "c285e09d-c2f5-4642-9e6b-59b0ed3ed651" (UID: "c285e09d-c2f5-4642-9e6b-59b0ed3ed651"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:33:07 crc kubenswrapper[4988]: I1008 18:33:07.580470 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c285e09d-c2f5-4642-9e6b-59b0ed3ed651-ovn-controller-tls-certs" (OuterVolumeSpecName: "ovn-controller-tls-certs") pod "c285e09d-c2f5-4642-9e6b-59b0ed3ed651" (UID: "c285e09d-c2f5-4642-9e6b-59b0ed3ed651"). InnerVolumeSpecName "ovn-controller-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:33:07 crc kubenswrapper[4988]: I1008 18:33:07.603189 4988 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c285e09d-c2f5-4642-9e6b-59b0ed3ed651-scripts\") on node \"crc\" DevicePath \"\"" Oct 08 18:33:07 crc kubenswrapper[4988]: I1008 18:33:07.603226 4988 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c285e09d-c2f5-4642-9e6b-59b0ed3ed651-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 18:33:07 crc kubenswrapper[4988]: I1008 18:33:07.603239 4988 reconciler_common.go:293] "Volume detached for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/c285e09d-c2f5-4642-9e6b-59b0ed3ed651-ovn-controller-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 08 18:33:07 crc kubenswrapper[4988]: I1008 18:33:07.603250 4988 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v2tq6\" (UniqueName: \"kubernetes.io/projected/c285e09d-c2f5-4642-9e6b-59b0ed3ed651-kube-api-access-v2tq6\") on node \"crc\" DevicePath \"\"" Oct 08 18:33:07 crc kubenswrapper[4988]: I1008 18:33:07.714718 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"830d6ec5-0015-4a88-b8e8-0fd47f64fd46","Type":"ContainerDied","Data":"a0bb4c09203e8a478b42ecb12f92dc6c845688e189426130aedb932153d1239c"} Oct 08 18:33:07 crc kubenswrapper[4988]: I1008 18:33:07.714763 4988 scope.go:117] "RemoveContainer" containerID="7aeabf72c789f76cad92e2b157ae1d52318b0fdd5da4ef59e8daac0afedb3f6a" Oct 08 18:33:07 crc kubenswrapper[4988]: I1008 18:33:07.714842 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Oct 08 18:33:07 crc kubenswrapper[4988]: I1008 18:33:07.722802 4988 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-gp8f4_c285e09d-c2f5-4642-9e6b-59b0ed3ed651/ovn-controller/0.log" Oct 08 18:33:07 crc kubenswrapper[4988]: I1008 18:33:07.722860 4988 generic.go:334] "Generic (PLEG): container finished" podID="c285e09d-c2f5-4642-9e6b-59b0ed3ed651" containerID="f3d6a7298803e1d31ec12101ceb0dd0c2936d1871327a38880d53ef51a7deceb" exitCode=139 Oct 08 18:33:07 crc kubenswrapper[4988]: I1008 18:33:07.722976 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-gp8f4" Oct 08 18:33:07 crc kubenswrapper[4988]: I1008 18:33:07.723075 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-gp8f4" event={"ID":"c285e09d-c2f5-4642-9e6b-59b0ed3ed651","Type":"ContainerDied","Data":"f3d6a7298803e1d31ec12101ceb0dd0c2936d1871327a38880d53ef51a7deceb"} Oct 08 18:33:07 crc kubenswrapper[4988]: I1008 18:33:07.723117 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-gp8f4" event={"ID":"c285e09d-c2f5-4642-9e6b-59b0ed3ed651","Type":"ContainerDied","Data":"3623831a4118466e308ca5fe5cc4eff6cb062697a66654a0488b0c44e1b18285"} Oct 08 18:33:07 crc kubenswrapper[4988]: I1008 18:33:07.730967 4988 generic.go:334] "Generic (PLEG): container finished" podID="c9f172a3-9188-4f5f-bf3a-fbc3b0dc5206" containerID="3251e38d412d5c6a4e41496647e54dfc1bc7728f57aa1932956313fe00c0a86b" exitCode=0 Oct 08 18:33:07 crc kubenswrapper[4988]: I1008 18:33:07.731055 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone3e59-account-delete-vgpc9" Oct 08 18:33:07 crc kubenswrapper[4988]: I1008 18:33:07.731073 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-8694c889b8-lhm4g" event={"ID":"c9f172a3-9188-4f5f-bf3a-fbc3b0dc5206","Type":"ContainerDied","Data":"3251e38d412d5c6a4e41496647e54dfc1bc7728f57aa1932956313fe00c0a86b"} Oct 08 18:33:07 crc kubenswrapper[4988]: I1008 18:33:07.750624 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 08 18:33:07 crc kubenswrapper[4988]: I1008 18:33:07.753790 4988 scope.go:117] "RemoveContainer" containerID="1002767e17c64f7c9060cbd0059630fd4f3dc91d7c2f7558f00143736e0751e0" Oct 08 18:33:07 crc kubenswrapper[4988]: I1008 18:33:07.760613 4988 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 08 18:33:07 crc kubenswrapper[4988]: I1008 18:33:07.777336 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone3e59-account-delete-vgpc9" Oct 08 18:33:07 crc kubenswrapper[4988]: I1008 18:33:07.790396 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-gp8f4"] Oct 08 18:33:07 crc kubenswrapper[4988]: I1008 18:33:07.794882 4988 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-gp8f4"] Oct 08 18:33:07 crc kubenswrapper[4988]: I1008 18:33:07.802419 4988 scope.go:117] "RemoveContainer" containerID="f3d6a7298803e1d31ec12101ceb0dd0c2936d1871327a38880d53ef51a7deceb" Oct 08 18:33:07 crc kubenswrapper[4988]: I1008 18:33:07.827847 4988 scope.go:117] "RemoveContainer" containerID="f3d6a7298803e1d31ec12101ceb0dd0c2936d1871327a38880d53ef51a7deceb" Oct 08 18:33:07 crc kubenswrapper[4988]: E1008 18:33:07.828348 4988 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f3d6a7298803e1d31ec12101ceb0dd0c2936d1871327a38880d53ef51a7deceb\": container with ID starting with f3d6a7298803e1d31ec12101ceb0dd0c2936d1871327a38880d53ef51a7deceb not found: ID does not exist" containerID="f3d6a7298803e1d31ec12101ceb0dd0c2936d1871327a38880d53ef51a7deceb" Oct 08 18:33:07 crc kubenswrapper[4988]: I1008 18:33:07.828379 4988 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f3d6a7298803e1d31ec12101ceb0dd0c2936d1871327a38880d53ef51a7deceb"} err="failed to get container status \"f3d6a7298803e1d31ec12101ceb0dd0c2936d1871327a38880d53ef51a7deceb\": rpc error: code = NotFound desc = could not find container \"f3d6a7298803e1d31ec12101ceb0dd0c2936d1871327a38880d53ef51a7deceb\": container with ID starting with f3d6a7298803e1d31ec12101ceb0dd0c2936d1871327a38880d53ef51a7deceb not found: ID does not exist" Oct 08 18:33:07 crc kubenswrapper[4988]: I1008 18:33:07.902877 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-8694c889b8-lhm4g" Oct 08 18:33:08 crc kubenswrapper[4988]: I1008 18:33:08.011751 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/c9f172a3-9188-4f5f-bf3a-fbc3b0dc5206-config-data-custom\") pod \"c9f172a3-9188-4f5f-bf3a-fbc3b0dc5206\" (UID: \"c9f172a3-9188-4f5f-bf3a-fbc3b0dc5206\") " Oct 08 18:33:08 crc kubenswrapper[4988]: I1008 18:33:08.011822 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d2ttv\" (UniqueName: \"kubernetes.io/projected/c9f172a3-9188-4f5f-bf3a-fbc3b0dc5206-kube-api-access-d2ttv\") pod \"c9f172a3-9188-4f5f-bf3a-fbc3b0dc5206\" (UID: \"c9f172a3-9188-4f5f-bf3a-fbc3b0dc5206\") " Oct 08 18:33:08 crc kubenswrapper[4988]: I1008 18:33:08.011847 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c9f172a3-9188-4f5f-bf3a-fbc3b0dc5206-combined-ca-bundle\") pod \"c9f172a3-9188-4f5f-bf3a-fbc3b0dc5206\" (UID: \"c9f172a3-9188-4f5f-bf3a-fbc3b0dc5206\") " Oct 08 18:33:08 crc kubenswrapper[4988]: I1008 18:33:08.011907 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c9f172a3-9188-4f5f-bf3a-fbc3b0dc5206-config-data\") pod \"c9f172a3-9188-4f5f-bf3a-fbc3b0dc5206\" (UID: \"c9f172a3-9188-4f5f-bf3a-fbc3b0dc5206\") " Oct 08 18:33:08 crc kubenswrapper[4988]: I1008 18:33:08.011961 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c9f172a3-9188-4f5f-bf3a-fbc3b0dc5206-logs\") pod \"c9f172a3-9188-4f5f-bf3a-fbc3b0dc5206\" (UID: \"c9f172a3-9188-4f5f-bf3a-fbc3b0dc5206\") " Oct 08 18:33:08 crc kubenswrapper[4988]: I1008 18:33:08.012603 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c9f172a3-9188-4f5f-bf3a-fbc3b0dc5206-logs" (OuterVolumeSpecName: "logs") pod "c9f172a3-9188-4f5f-bf3a-fbc3b0dc5206" (UID: "c9f172a3-9188-4f5f-bf3a-fbc3b0dc5206"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 18:33:08 crc kubenswrapper[4988]: I1008 18:33:08.017168 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c9f172a3-9188-4f5f-bf3a-fbc3b0dc5206-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "c9f172a3-9188-4f5f-bf3a-fbc3b0dc5206" (UID: "c9f172a3-9188-4f5f-bf3a-fbc3b0dc5206"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:33:08 crc kubenswrapper[4988]: I1008 18:33:08.017696 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c9f172a3-9188-4f5f-bf3a-fbc3b0dc5206-kube-api-access-d2ttv" (OuterVolumeSpecName: "kube-api-access-d2ttv") pod "c9f172a3-9188-4f5f-bf3a-fbc3b0dc5206" (UID: "c9f172a3-9188-4f5f-bf3a-fbc3b0dc5206"). InnerVolumeSpecName "kube-api-access-d2ttv". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:33:08 crc kubenswrapper[4988]: I1008 18:33:08.031866 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c9f172a3-9188-4f5f-bf3a-fbc3b0dc5206-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "c9f172a3-9188-4f5f-bf3a-fbc3b0dc5206" (UID: "c9f172a3-9188-4f5f-bf3a-fbc3b0dc5206"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:33:08 crc kubenswrapper[4988]: I1008 18:33:08.046747 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c9f172a3-9188-4f5f-bf3a-fbc3b0dc5206-config-data" (OuterVolumeSpecName: "config-data") pod "c9f172a3-9188-4f5f-bf3a-fbc3b0dc5206" (UID: "c9f172a3-9188-4f5f-bf3a-fbc3b0dc5206"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:33:08 crc kubenswrapper[4988]: I1008 18:33:08.113369 4988 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c9f172a3-9188-4f5f-bf3a-fbc3b0dc5206-config-data\") on node \"crc\" DevicePath \"\"" Oct 08 18:33:08 crc kubenswrapper[4988]: I1008 18:33:08.113430 4988 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c9f172a3-9188-4f5f-bf3a-fbc3b0dc5206-logs\") on node \"crc\" DevicePath \"\"" Oct 08 18:33:08 crc kubenswrapper[4988]: I1008 18:33:08.113444 4988 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/c9f172a3-9188-4f5f-bf3a-fbc3b0dc5206-config-data-custom\") on node \"crc\" DevicePath \"\"" Oct 08 18:33:08 crc kubenswrapper[4988]: I1008 18:33:08.113458 4988 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d2ttv\" (UniqueName: \"kubernetes.io/projected/c9f172a3-9188-4f5f-bf3a-fbc3b0dc5206-kube-api-access-d2ttv\") on node \"crc\" DevicePath \"\"" Oct 08 18:33:08 crc kubenswrapper[4988]: I1008 18:33:08.113471 4988 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c9f172a3-9188-4f5f-bf3a-fbc3b0dc5206-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 18:33:08 crc kubenswrapper[4988]: I1008 18:33:08.578017 4988 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/memcached-0" podUID="32bb7401-6292-4ca6-98f4-15522125c3ee" containerName="memcached" probeResult="failure" output="dial tcp 10.217.0.107:11211: i/o timeout" Oct 08 18:33:08 crc kubenswrapper[4988]: I1008 18:33:08.745659 4988 generic.go:334] "Generic (PLEG): container finished" podID="ccd2f8fa-1c16-4c58-9f87-e60ba3a2f5fb" containerID="06f5039b59f5eef23b916412ccb9a5658d672fcb48cf3d4a8e55e5a55e9519c0" exitCode=0 Oct 08 18:33:08 crc kubenswrapper[4988]: I1008 18:33:08.745735 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ccd2f8fa-1c16-4c58-9f87-e60ba3a2f5fb","Type":"ContainerDied","Data":"06f5039b59f5eef23b916412ccb9a5658d672fcb48cf3d4a8e55e5a55e9519c0"} Oct 08 18:33:08 crc kubenswrapper[4988]: I1008 18:33:08.748240 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-8694c889b8-lhm4g" event={"ID":"c9f172a3-9188-4f5f-bf3a-fbc3b0dc5206","Type":"ContainerDied","Data":"77cc9e1e0ea4e222419c071fbb94d7b4a3b6fcc421e688d2501dd61fe2fa13ce"} Oct 08 18:33:08 crc kubenswrapper[4988]: I1008 18:33:08.748314 4988 scope.go:117] "RemoveContainer" containerID="3251e38d412d5c6a4e41496647e54dfc1bc7728f57aa1932956313fe00c0a86b" Oct 08 18:33:08 crc kubenswrapper[4988]: I1008 18:33:08.748474 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-8694c889b8-lhm4g" Oct 08 18:33:08 crc kubenswrapper[4988]: I1008 18:33:08.757128 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone3e59-account-delete-vgpc9" Oct 08 18:33:08 crc kubenswrapper[4988]: I1008 18:33:08.780956 4988 scope.go:117] "RemoveContainer" containerID="c7e441bdd6242790ab99582ae7f776a6bd322dc0c9ca3985c984d45c5cc43a5c" Oct 08 18:33:08 crc kubenswrapper[4988]: I1008 18:33:08.817124 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone3e59-account-delete-vgpc9"] Oct 08 18:33:08 crc kubenswrapper[4988]: I1008 18:33:08.821357 4988 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone3e59-account-delete-vgpc9"] Oct 08 18:33:08 crc kubenswrapper[4988]: I1008 18:33:08.830999 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-keystone-listener-8694c889b8-lhm4g"] Oct 08 18:33:08 crc kubenswrapper[4988]: I1008 18:33:08.838560 4988 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-keystone-listener-8694c889b8-lhm4g"] Oct 08 18:33:08 crc kubenswrapper[4988]: E1008 18:33:08.913596 4988 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod11679d40_6eb6_456c_8996_cd9a946a6db1.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc9f172a3_9188_4f5f_bf3a_fbc3b0dc5206.slice/crio-77cc9e1e0ea4e222419c071fbb94d7b4a3b6fcc421e688d2501dd61fe2fa13ce\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc9f172a3_9188_4f5f_bf3a_fbc3b0dc5206.slice\": RecentStats: unable to find data in memory cache]" Oct 08 18:33:08 crc kubenswrapper[4988]: I1008 18:33:08.914595 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 08 18:33:08 crc kubenswrapper[4988]: I1008 18:33:08.928831 4988 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-k8mnm\" (UniqueName: \"kubernetes.io/projected/11679d40-6eb6-456c-8996-cd9a946a6db1-kube-api-access-k8mnm\") on node \"crc\" DevicePath \"\"" Oct 08 18:33:09 crc kubenswrapper[4988]: I1008 18:33:09.030291 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ccd2f8fa-1c16-4c58-9f87-e60ba3a2f5fb-run-httpd\") pod \"ccd2f8fa-1c16-4c58-9f87-e60ba3a2f5fb\" (UID: \"ccd2f8fa-1c16-4c58-9f87-e60ba3a2f5fb\") " Oct 08 18:33:09 crc kubenswrapper[4988]: I1008 18:33:09.030633 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ccd2f8fa-1c16-4c58-9f87-e60ba3a2f5fb-combined-ca-bundle\") pod \"ccd2f8fa-1c16-4c58-9f87-e60ba3a2f5fb\" (UID: \"ccd2f8fa-1c16-4c58-9f87-e60ba3a2f5fb\") " Oct 08 18:33:09 crc kubenswrapper[4988]: I1008 18:33:09.030678 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ccd2f8fa-1c16-4c58-9f87-e60ba3a2f5fb-scripts\") pod \"ccd2f8fa-1c16-4c58-9f87-e60ba3a2f5fb\" (UID: \"ccd2f8fa-1c16-4c58-9f87-e60ba3a2f5fb\") " Oct 08 18:33:09 crc kubenswrapper[4988]: I1008 18:33:09.030790 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ccd2f8fa-1c16-4c58-9f87-e60ba3a2f5fb-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "ccd2f8fa-1c16-4c58-9f87-e60ba3a2f5fb" (UID: "ccd2f8fa-1c16-4c58-9f87-e60ba3a2f5fb"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 18:33:09 crc kubenswrapper[4988]: I1008 18:33:09.030700 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/ccd2f8fa-1c16-4c58-9f87-e60ba3a2f5fb-ceilometer-tls-certs\") pod \"ccd2f8fa-1c16-4c58-9f87-e60ba3a2f5fb\" (UID: \"ccd2f8fa-1c16-4c58-9f87-e60ba3a2f5fb\") " Oct 08 18:33:09 crc kubenswrapper[4988]: I1008 18:33:09.031007 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ccd2f8fa-1c16-4c58-9f87-e60ba3a2f5fb-log-httpd\") pod \"ccd2f8fa-1c16-4c58-9f87-e60ba3a2f5fb\" (UID: \"ccd2f8fa-1c16-4c58-9f87-e60ba3a2f5fb\") " Oct 08 18:33:09 crc kubenswrapper[4988]: I1008 18:33:09.031117 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/ccd2f8fa-1c16-4c58-9f87-e60ba3a2f5fb-sg-core-conf-yaml\") pod \"ccd2f8fa-1c16-4c58-9f87-e60ba3a2f5fb\" (UID: \"ccd2f8fa-1c16-4c58-9f87-e60ba3a2f5fb\") " Oct 08 18:33:09 crc kubenswrapper[4988]: I1008 18:33:09.031153 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8p2rl\" (UniqueName: \"kubernetes.io/projected/ccd2f8fa-1c16-4c58-9f87-e60ba3a2f5fb-kube-api-access-8p2rl\") pod \"ccd2f8fa-1c16-4c58-9f87-e60ba3a2f5fb\" (UID: \"ccd2f8fa-1c16-4c58-9f87-e60ba3a2f5fb\") " Oct 08 18:33:09 crc kubenswrapper[4988]: I1008 18:33:09.031196 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ccd2f8fa-1c16-4c58-9f87-e60ba3a2f5fb-config-data\") pod \"ccd2f8fa-1c16-4c58-9f87-e60ba3a2f5fb\" (UID: \"ccd2f8fa-1c16-4c58-9f87-e60ba3a2f5fb\") " Oct 08 18:33:09 crc kubenswrapper[4988]: I1008 18:33:09.031403 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ccd2f8fa-1c16-4c58-9f87-e60ba3a2f5fb-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "ccd2f8fa-1c16-4c58-9f87-e60ba3a2f5fb" (UID: "ccd2f8fa-1c16-4c58-9f87-e60ba3a2f5fb"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 18:33:09 crc kubenswrapper[4988]: I1008 18:33:09.031723 4988 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ccd2f8fa-1c16-4c58-9f87-e60ba3a2f5fb-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 08 18:33:09 crc kubenswrapper[4988]: I1008 18:33:09.031746 4988 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ccd2f8fa-1c16-4c58-9f87-e60ba3a2f5fb-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 08 18:33:09 crc kubenswrapper[4988]: I1008 18:33:09.034277 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ccd2f8fa-1c16-4c58-9f87-e60ba3a2f5fb-scripts" (OuterVolumeSpecName: "scripts") pod "ccd2f8fa-1c16-4c58-9f87-e60ba3a2f5fb" (UID: "ccd2f8fa-1c16-4c58-9f87-e60ba3a2f5fb"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:33:09 crc kubenswrapper[4988]: I1008 18:33:09.035660 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ccd2f8fa-1c16-4c58-9f87-e60ba3a2f5fb-kube-api-access-8p2rl" (OuterVolumeSpecName: "kube-api-access-8p2rl") pod "ccd2f8fa-1c16-4c58-9f87-e60ba3a2f5fb" (UID: "ccd2f8fa-1c16-4c58-9f87-e60ba3a2f5fb"). InnerVolumeSpecName "kube-api-access-8p2rl". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:33:09 crc kubenswrapper[4988]: I1008 18:33:09.053486 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ccd2f8fa-1c16-4c58-9f87-e60ba3a2f5fb-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "ccd2f8fa-1c16-4c58-9f87-e60ba3a2f5fb" (UID: "ccd2f8fa-1c16-4c58-9f87-e60ba3a2f5fb"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:33:09 crc kubenswrapper[4988]: I1008 18:33:09.064538 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ccd2f8fa-1c16-4c58-9f87-e60ba3a2f5fb-ceilometer-tls-certs" (OuterVolumeSpecName: "ceilometer-tls-certs") pod "ccd2f8fa-1c16-4c58-9f87-e60ba3a2f5fb" (UID: "ccd2f8fa-1c16-4c58-9f87-e60ba3a2f5fb"). InnerVolumeSpecName "ceilometer-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:33:09 crc kubenswrapper[4988]: I1008 18:33:09.084236 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ccd2f8fa-1c16-4c58-9f87-e60ba3a2f5fb-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ccd2f8fa-1c16-4c58-9f87-e60ba3a2f5fb" (UID: "ccd2f8fa-1c16-4c58-9f87-e60ba3a2f5fb"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:33:09 crc kubenswrapper[4988]: I1008 18:33:09.105278 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ccd2f8fa-1c16-4c58-9f87-e60ba3a2f5fb-config-data" (OuterVolumeSpecName: "config-data") pod "ccd2f8fa-1c16-4c58-9f87-e60ba3a2f5fb" (UID: "ccd2f8fa-1c16-4c58-9f87-e60ba3a2f5fb"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:33:09 crc kubenswrapper[4988]: I1008 18:33:09.133328 4988 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/ccd2f8fa-1c16-4c58-9f87-e60ba3a2f5fb-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Oct 08 18:33:09 crc kubenswrapper[4988]: I1008 18:33:09.133365 4988 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8p2rl\" (UniqueName: \"kubernetes.io/projected/ccd2f8fa-1c16-4c58-9f87-e60ba3a2f5fb-kube-api-access-8p2rl\") on node \"crc\" DevicePath \"\"" Oct 08 18:33:09 crc kubenswrapper[4988]: I1008 18:33:09.133407 4988 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ccd2f8fa-1c16-4c58-9f87-e60ba3a2f5fb-config-data\") on node \"crc\" DevicePath \"\"" Oct 08 18:33:09 crc kubenswrapper[4988]: I1008 18:33:09.133419 4988 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ccd2f8fa-1c16-4c58-9f87-e60ba3a2f5fb-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 18:33:09 crc kubenswrapper[4988]: I1008 18:33:09.133431 4988 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ccd2f8fa-1c16-4c58-9f87-e60ba3a2f5fb-scripts\") on node \"crc\" DevicePath \"\"" Oct 08 18:33:09 crc kubenswrapper[4988]: I1008 18:33:09.133441 4988 reconciler_common.go:293] "Volume detached for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/ccd2f8fa-1c16-4c58-9f87-e60ba3a2f5fb-ceilometer-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 08 18:33:09 crc kubenswrapper[4988]: E1008 18:33:09.203661 4988 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 863aa8ddeda792a78ba5af321605974ac7082d24fc8fb6f0400148fcf3f06126 is running failed: container process not found" containerID="863aa8ddeda792a78ba5af321605974ac7082d24fc8fb6f0400148fcf3f06126" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Oct 08 18:33:09 crc kubenswrapper[4988]: E1008 18:33:09.204218 4988 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 863aa8ddeda792a78ba5af321605974ac7082d24fc8fb6f0400148fcf3f06126 is running failed: container process not found" containerID="863aa8ddeda792a78ba5af321605974ac7082d24fc8fb6f0400148fcf3f06126" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Oct 08 18:33:09 crc kubenswrapper[4988]: E1008 18:33:09.204623 4988 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 863aa8ddeda792a78ba5af321605974ac7082d24fc8fb6f0400148fcf3f06126 is running failed: container process not found" containerID="863aa8ddeda792a78ba5af321605974ac7082d24fc8fb6f0400148fcf3f06126" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Oct 08 18:33:09 crc kubenswrapper[4988]: E1008 18:33:09.204667 4988 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 863aa8ddeda792a78ba5af321605974ac7082d24fc8fb6f0400148fcf3f06126 is running failed: container process not found" probeType="Readiness" pod="openstack/ovn-controller-ovs-kc7fw" podUID="f1628013-e0b9-4c31-80c0-91aabaaef0f6" containerName="ovsdb-server" Oct 08 18:33:09 crc kubenswrapper[4988]: E1008 18:33:09.205990 4988 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="8a5332d3d9aa50bb56f12e04b1b32520399363308f9e75d5510ddd01f41593c3" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Oct 08 18:33:09 crc kubenswrapper[4988]: E1008 18:33:09.208702 4988 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="8a5332d3d9aa50bb56f12e04b1b32520399363308f9e75d5510ddd01f41593c3" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Oct 08 18:33:09 crc kubenswrapper[4988]: E1008 18:33:09.210172 4988 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="8a5332d3d9aa50bb56f12e04b1b32520399363308f9e75d5510ddd01f41593c3" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Oct 08 18:33:09 crc kubenswrapper[4988]: E1008 18:33:09.210223 4988 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/ovn-controller-ovs-kc7fw" podUID="f1628013-e0b9-4c31-80c0-91aabaaef0f6" containerName="ovs-vswitchd" Oct 08 18:33:09 crc kubenswrapper[4988]: I1008 18:33:09.250169 4988 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="11679d40-6eb6-456c-8996-cd9a946a6db1" path="/var/lib/kubelet/pods/11679d40-6eb6-456c-8996-cd9a946a6db1/volumes" Oct 08 18:33:09 crc kubenswrapper[4988]: I1008 18:33:09.250796 4988 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="830d6ec5-0015-4a88-b8e8-0fd47f64fd46" path="/var/lib/kubelet/pods/830d6ec5-0015-4a88-b8e8-0fd47f64fd46/volumes" Oct 08 18:33:09 crc kubenswrapper[4988]: I1008 18:33:09.251420 4988 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c285e09d-c2f5-4642-9e6b-59b0ed3ed651" path="/var/lib/kubelet/pods/c285e09d-c2f5-4642-9e6b-59b0ed3ed651/volumes" Oct 08 18:33:09 crc kubenswrapper[4988]: I1008 18:33:09.252574 4988 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c9f172a3-9188-4f5f-bf3a-fbc3b0dc5206" path="/var/lib/kubelet/pods/c9f172a3-9188-4f5f-bf3a-fbc3b0dc5206/volumes" Oct 08 18:33:09 crc kubenswrapper[4988]: I1008 18:33:09.773058 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ccd2f8fa-1c16-4c58-9f87-e60ba3a2f5fb","Type":"ContainerDied","Data":"fd69699740aac8c03e9405c7eb6b657c410818eb3005cb0c61692c403508b677"} Oct 08 18:33:09 crc kubenswrapper[4988]: I1008 18:33:09.773125 4988 scope.go:117] "RemoveContainer" containerID="f64fdd9df8d4aff7bcd8cfab3af523fa24041d42ea9cfb0addd7bf856026fe76" Oct 08 18:33:09 crc kubenswrapper[4988]: I1008 18:33:09.773172 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 08 18:33:09 crc kubenswrapper[4988]: I1008 18:33:09.803331 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 08 18:33:09 crc kubenswrapper[4988]: I1008 18:33:09.804184 4988 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Oct 08 18:33:09 crc kubenswrapper[4988]: I1008 18:33:09.809047 4988 scope.go:117] "RemoveContainer" containerID="be90eff6d6a94661069208a3760ad429da6e736c3f64018e09b59d748a843d33" Oct 08 18:33:09 crc kubenswrapper[4988]: I1008 18:33:09.836456 4988 scope.go:117] "RemoveContainer" containerID="06f5039b59f5eef23b916412ccb9a5658d672fcb48cf3d4a8e55e5a55e9519c0" Oct 08 18:33:09 crc kubenswrapper[4988]: I1008 18:33:09.853903 4988 scope.go:117] "RemoveContainer" containerID="c7ddcc51fc71ff2377802bb6d1c0c0545ab7211828f994c467b39e3b24eaac16" Oct 08 18:33:10 crc kubenswrapper[4988]: I1008 18:33:10.451724 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Oct 08 18:33:10 crc kubenswrapper[4988]: I1008 18:33:10.558224 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/72c64965-0b57-43e1-8d18-2bce6cabbfa1-operator-scripts\") pod \"72c64965-0b57-43e1-8d18-2bce6cabbfa1\" (UID: \"72c64965-0b57-43e1-8d18-2bce6cabbfa1\") " Oct 08 18:33:10 crc kubenswrapper[4988]: I1008 18:33:10.558294 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/72c64965-0b57-43e1-8d18-2bce6cabbfa1-galera-tls-certs\") pod \"72c64965-0b57-43e1-8d18-2bce6cabbfa1\" (UID: \"72c64965-0b57-43e1-8d18-2bce6cabbfa1\") " Oct 08 18:33:10 crc kubenswrapper[4988]: I1008 18:33:10.558325 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/72c64965-0b57-43e1-8d18-2bce6cabbfa1-config-data-default\") pod \"72c64965-0b57-43e1-8d18-2bce6cabbfa1\" (UID: \"72c64965-0b57-43e1-8d18-2bce6cabbfa1\") " Oct 08 18:33:10 crc kubenswrapper[4988]: I1008 18:33:10.558374 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mysql-db\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"72c64965-0b57-43e1-8d18-2bce6cabbfa1\" (UID: \"72c64965-0b57-43e1-8d18-2bce6cabbfa1\") " Oct 08 18:33:10 crc kubenswrapper[4988]: I1008 18:33:10.558456 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/72c64965-0b57-43e1-8d18-2bce6cabbfa1-kolla-config\") pod \"72c64965-0b57-43e1-8d18-2bce6cabbfa1\" (UID: \"72c64965-0b57-43e1-8d18-2bce6cabbfa1\") " Oct 08 18:33:10 crc kubenswrapper[4988]: I1008 18:33:10.558495 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/72c64965-0b57-43e1-8d18-2bce6cabbfa1-config-data-generated\") pod \"72c64965-0b57-43e1-8d18-2bce6cabbfa1\" (UID: \"72c64965-0b57-43e1-8d18-2bce6cabbfa1\") " Oct 08 18:33:10 crc kubenswrapper[4988]: I1008 18:33:10.558565 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/72c64965-0b57-43e1-8d18-2bce6cabbfa1-secrets\") pod \"72c64965-0b57-43e1-8d18-2bce6cabbfa1\" (UID: \"72c64965-0b57-43e1-8d18-2bce6cabbfa1\") " Oct 08 18:33:10 crc kubenswrapper[4988]: I1008 18:33:10.558597 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/72c64965-0b57-43e1-8d18-2bce6cabbfa1-combined-ca-bundle\") pod \"72c64965-0b57-43e1-8d18-2bce6cabbfa1\" (UID: \"72c64965-0b57-43e1-8d18-2bce6cabbfa1\") " Oct 08 18:33:10 crc kubenswrapper[4988]: I1008 18:33:10.558630 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qszql\" (UniqueName: \"kubernetes.io/projected/72c64965-0b57-43e1-8d18-2bce6cabbfa1-kube-api-access-qszql\") pod \"72c64965-0b57-43e1-8d18-2bce6cabbfa1\" (UID: \"72c64965-0b57-43e1-8d18-2bce6cabbfa1\") " Oct 08 18:33:10 crc kubenswrapper[4988]: I1008 18:33:10.559994 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/72c64965-0b57-43e1-8d18-2bce6cabbfa1-kolla-config" (OuterVolumeSpecName: "kolla-config") pod "72c64965-0b57-43e1-8d18-2bce6cabbfa1" (UID: "72c64965-0b57-43e1-8d18-2bce6cabbfa1"). InnerVolumeSpecName "kolla-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 18:33:10 crc kubenswrapper[4988]: I1008 18:33:10.560046 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/72c64965-0b57-43e1-8d18-2bce6cabbfa1-config-data-generated" (OuterVolumeSpecName: "config-data-generated") pod "72c64965-0b57-43e1-8d18-2bce6cabbfa1" (UID: "72c64965-0b57-43e1-8d18-2bce6cabbfa1"). InnerVolumeSpecName "config-data-generated". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 18:33:10 crc kubenswrapper[4988]: I1008 18:33:10.560627 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/72c64965-0b57-43e1-8d18-2bce6cabbfa1-config-data-default" (OuterVolumeSpecName: "config-data-default") pod "72c64965-0b57-43e1-8d18-2bce6cabbfa1" (UID: "72c64965-0b57-43e1-8d18-2bce6cabbfa1"). InnerVolumeSpecName "config-data-default". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 18:33:10 crc kubenswrapper[4988]: I1008 18:33:10.562322 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/72c64965-0b57-43e1-8d18-2bce6cabbfa1-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "72c64965-0b57-43e1-8d18-2bce6cabbfa1" (UID: "72c64965-0b57-43e1-8d18-2bce6cabbfa1"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 18:33:10 crc kubenswrapper[4988]: I1008 18:33:10.565086 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/72c64965-0b57-43e1-8d18-2bce6cabbfa1-secrets" (OuterVolumeSpecName: "secrets") pod "72c64965-0b57-43e1-8d18-2bce6cabbfa1" (UID: "72c64965-0b57-43e1-8d18-2bce6cabbfa1"). InnerVolumeSpecName "secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:33:10 crc kubenswrapper[4988]: I1008 18:33:10.565139 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/72c64965-0b57-43e1-8d18-2bce6cabbfa1-kube-api-access-qszql" (OuterVolumeSpecName: "kube-api-access-qszql") pod "72c64965-0b57-43e1-8d18-2bce6cabbfa1" (UID: "72c64965-0b57-43e1-8d18-2bce6cabbfa1"). InnerVolumeSpecName "kube-api-access-qszql". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:33:10 crc kubenswrapper[4988]: I1008 18:33:10.573972 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage02-crc" (OuterVolumeSpecName: "mysql-db") pod "72c64965-0b57-43e1-8d18-2bce6cabbfa1" (UID: "72c64965-0b57-43e1-8d18-2bce6cabbfa1"). InnerVolumeSpecName "local-storage02-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 08 18:33:10 crc kubenswrapper[4988]: I1008 18:33:10.583945 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/72c64965-0b57-43e1-8d18-2bce6cabbfa1-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "72c64965-0b57-43e1-8d18-2bce6cabbfa1" (UID: "72c64965-0b57-43e1-8d18-2bce6cabbfa1"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:33:10 crc kubenswrapper[4988]: I1008 18:33:10.614771 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/72c64965-0b57-43e1-8d18-2bce6cabbfa1-galera-tls-certs" (OuterVolumeSpecName: "galera-tls-certs") pod "72c64965-0b57-43e1-8d18-2bce6cabbfa1" (UID: "72c64965-0b57-43e1-8d18-2bce6cabbfa1"). InnerVolumeSpecName "galera-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:33:10 crc kubenswrapper[4988]: I1008 18:33:10.660782 4988 reconciler_common.go:293] "Volume detached for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/72c64965-0b57-43e1-8d18-2bce6cabbfa1-galera-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 08 18:33:10 crc kubenswrapper[4988]: I1008 18:33:10.660835 4988 reconciler_common.go:293] "Volume detached for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/72c64965-0b57-43e1-8d18-2bce6cabbfa1-config-data-default\") on node \"crc\" DevicePath \"\"" Oct 08 18:33:10 crc kubenswrapper[4988]: I1008 18:33:10.660890 4988 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") on node \"crc\" " Oct 08 18:33:10 crc kubenswrapper[4988]: I1008 18:33:10.660908 4988 reconciler_common.go:293] "Volume detached for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/72c64965-0b57-43e1-8d18-2bce6cabbfa1-kolla-config\") on node \"crc\" DevicePath \"\"" Oct 08 18:33:10 crc kubenswrapper[4988]: I1008 18:33:10.660926 4988 reconciler_common.go:293] "Volume detached for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/72c64965-0b57-43e1-8d18-2bce6cabbfa1-config-data-generated\") on node \"crc\" DevicePath \"\"" Oct 08 18:33:10 crc kubenswrapper[4988]: I1008 18:33:10.660945 4988 reconciler_common.go:293] "Volume detached for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/72c64965-0b57-43e1-8d18-2bce6cabbfa1-secrets\") on node \"crc\" DevicePath \"\"" Oct 08 18:33:10 crc kubenswrapper[4988]: I1008 18:33:10.660961 4988 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/72c64965-0b57-43e1-8d18-2bce6cabbfa1-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 18:33:10 crc kubenswrapper[4988]: I1008 18:33:10.660976 4988 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qszql\" (UniqueName: \"kubernetes.io/projected/72c64965-0b57-43e1-8d18-2bce6cabbfa1-kube-api-access-qszql\") on node \"crc\" DevicePath \"\"" Oct 08 18:33:10 crc kubenswrapper[4988]: I1008 18:33:10.660991 4988 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/72c64965-0b57-43e1-8d18-2bce6cabbfa1-operator-scripts\") on node \"crc\" DevicePath \"\"" Oct 08 18:33:10 crc kubenswrapper[4988]: I1008 18:33:10.680311 4988 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage02-crc" (UniqueName: "kubernetes.io/local-volume/local-storage02-crc") on node "crc" Oct 08 18:33:10 crc kubenswrapper[4988]: I1008 18:33:10.761730 4988 reconciler_common.go:293] "Volume detached for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") on node \"crc\" DevicePath \"\"" Oct 08 18:33:10 crc kubenswrapper[4988]: I1008 18:33:10.789558 4988 generic.go:334] "Generic (PLEG): container finished" podID="72c64965-0b57-43e1-8d18-2bce6cabbfa1" containerID="239fe689949862e30d5f5b0021f848adb4380b18bdefdeb68e24c7cd3fb41c24" exitCode=0 Oct 08 18:33:10 crc kubenswrapper[4988]: I1008 18:33:10.789609 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"72c64965-0b57-43e1-8d18-2bce6cabbfa1","Type":"ContainerDied","Data":"239fe689949862e30d5f5b0021f848adb4380b18bdefdeb68e24c7cd3fb41c24"} Oct 08 18:33:10 crc kubenswrapper[4988]: I1008 18:33:10.789617 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Oct 08 18:33:10 crc kubenswrapper[4988]: I1008 18:33:10.789652 4988 scope.go:117] "RemoveContainer" containerID="239fe689949862e30d5f5b0021f848adb4380b18bdefdeb68e24c7cd3fb41c24" Oct 08 18:33:10 crc kubenswrapper[4988]: I1008 18:33:10.789640 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"72c64965-0b57-43e1-8d18-2bce6cabbfa1","Type":"ContainerDied","Data":"a0f4596eebd8fb2c500b21efcd1ffce02fd0b60bfebe08354d1aa52b412471be"} Oct 08 18:33:10 crc kubenswrapper[4988]: I1008 18:33:10.813825 4988 scope.go:117] "RemoveContainer" containerID="4a95ce795323840905212e6edf0cc0a3e009ff2a132cb0c21d30fe459fa29c5d" Oct 08 18:33:10 crc kubenswrapper[4988]: I1008 18:33:10.838829 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/openstack-galera-0"] Oct 08 18:33:10 crc kubenswrapper[4988]: I1008 18:33:10.845539 4988 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/openstack-galera-0"] Oct 08 18:33:10 crc kubenswrapper[4988]: I1008 18:33:10.863237 4988 scope.go:117] "RemoveContainer" containerID="239fe689949862e30d5f5b0021f848adb4380b18bdefdeb68e24c7cd3fb41c24" Oct 08 18:33:10 crc kubenswrapper[4988]: E1008 18:33:10.870990 4988 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"239fe689949862e30d5f5b0021f848adb4380b18bdefdeb68e24c7cd3fb41c24\": container with ID starting with 239fe689949862e30d5f5b0021f848adb4380b18bdefdeb68e24c7cd3fb41c24 not found: ID does not exist" containerID="239fe689949862e30d5f5b0021f848adb4380b18bdefdeb68e24c7cd3fb41c24" Oct 08 18:33:10 crc kubenswrapper[4988]: I1008 18:33:10.871037 4988 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"239fe689949862e30d5f5b0021f848adb4380b18bdefdeb68e24c7cd3fb41c24"} err="failed to get container status \"239fe689949862e30d5f5b0021f848adb4380b18bdefdeb68e24c7cd3fb41c24\": rpc error: code = NotFound desc = could not find container \"239fe689949862e30d5f5b0021f848adb4380b18bdefdeb68e24c7cd3fb41c24\": container with ID starting with 239fe689949862e30d5f5b0021f848adb4380b18bdefdeb68e24c7cd3fb41c24 not found: ID does not exist" Oct 08 18:33:10 crc kubenswrapper[4988]: I1008 18:33:10.871062 4988 scope.go:117] "RemoveContainer" containerID="4a95ce795323840905212e6edf0cc0a3e009ff2a132cb0c21d30fe459fa29c5d" Oct 08 18:33:10 crc kubenswrapper[4988]: E1008 18:33:10.871290 4988 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4a95ce795323840905212e6edf0cc0a3e009ff2a132cb0c21d30fe459fa29c5d\": container with ID starting with 4a95ce795323840905212e6edf0cc0a3e009ff2a132cb0c21d30fe459fa29c5d not found: ID does not exist" containerID="4a95ce795323840905212e6edf0cc0a3e009ff2a132cb0c21d30fe459fa29c5d" Oct 08 18:33:10 crc kubenswrapper[4988]: I1008 18:33:10.871312 4988 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4a95ce795323840905212e6edf0cc0a3e009ff2a132cb0c21d30fe459fa29c5d"} err="failed to get container status \"4a95ce795323840905212e6edf0cc0a3e009ff2a132cb0c21d30fe459fa29c5d\": rpc error: code = NotFound desc = could not find container \"4a95ce795323840905212e6edf0cc0a3e009ff2a132cb0c21d30fe459fa29c5d\": container with ID starting with 4a95ce795323840905212e6edf0cc0a3e009ff2a132cb0c21d30fe459fa29c5d not found: ID does not exist" Oct 08 18:33:11 crc kubenswrapper[4988]: I1008 18:33:11.250000 4988 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="72c64965-0b57-43e1-8d18-2bce6cabbfa1" path="/var/lib/kubelet/pods/72c64965-0b57-43e1-8d18-2bce6cabbfa1/volumes" Oct 08 18:33:11 crc kubenswrapper[4988]: I1008 18:33:11.250836 4988 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ccd2f8fa-1c16-4c58-9f87-e60ba3a2f5fb" path="/var/lib/kubelet/pods/ccd2f8fa-1c16-4c58-9f87-e60ba3a2f5fb/volumes" Oct 08 18:33:14 crc kubenswrapper[4988]: E1008 18:33:14.205576 4988 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 863aa8ddeda792a78ba5af321605974ac7082d24fc8fb6f0400148fcf3f06126 is running failed: container process not found" containerID="863aa8ddeda792a78ba5af321605974ac7082d24fc8fb6f0400148fcf3f06126" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Oct 08 18:33:14 crc kubenswrapper[4988]: E1008 18:33:14.206644 4988 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="8a5332d3d9aa50bb56f12e04b1b32520399363308f9e75d5510ddd01f41593c3" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Oct 08 18:33:14 crc kubenswrapper[4988]: E1008 18:33:14.207042 4988 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 863aa8ddeda792a78ba5af321605974ac7082d24fc8fb6f0400148fcf3f06126 is running failed: container process not found" containerID="863aa8ddeda792a78ba5af321605974ac7082d24fc8fb6f0400148fcf3f06126" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Oct 08 18:33:14 crc kubenswrapper[4988]: E1008 18:33:14.207544 4988 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 863aa8ddeda792a78ba5af321605974ac7082d24fc8fb6f0400148fcf3f06126 is running failed: container process not found" containerID="863aa8ddeda792a78ba5af321605974ac7082d24fc8fb6f0400148fcf3f06126" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Oct 08 18:33:14 crc kubenswrapper[4988]: E1008 18:33:14.207594 4988 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 863aa8ddeda792a78ba5af321605974ac7082d24fc8fb6f0400148fcf3f06126 is running failed: container process not found" probeType="Readiness" pod="openstack/ovn-controller-ovs-kc7fw" podUID="f1628013-e0b9-4c31-80c0-91aabaaef0f6" containerName="ovsdb-server" Oct 08 18:33:14 crc kubenswrapper[4988]: E1008 18:33:14.208943 4988 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="8a5332d3d9aa50bb56f12e04b1b32520399363308f9e75d5510ddd01f41593c3" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Oct 08 18:33:14 crc kubenswrapper[4988]: E1008 18:33:14.212505 4988 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="8a5332d3d9aa50bb56f12e04b1b32520399363308f9e75d5510ddd01f41593c3" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Oct 08 18:33:14 crc kubenswrapper[4988]: E1008 18:33:14.212557 4988 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/ovn-controller-ovs-kc7fw" podUID="f1628013-e0b9-4c31-80c0-91aabaaef0f6" containerName="ovs-vswitchd" Oct 08 18:33:19 crc kubenswrapper[4988]: E1008 18:33:19.202777 4988 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 863aa8ddeda792a78ba5af321605974ac7082d24fc8fb6f0400148fcf3f06126 is running failed: container process not found" containerID="863aa8ddeda792a78ba5af321605974ac7082d24fc8fb6f0400148fcf3f06126" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Oct 08 18:33:19 crc kubenswrapper[4988]: E1008 18:33:19.203700 4988 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 863aa8ddeda792a78ba5af321605974ac7082d24fc8fb6f0400148fcf3f06126 is running failed: container process not found" containerID="863aa8ddeda792a78ba5af321605974ac7082d24fc8fb6f0400148fcf3f06126" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Oct 08 18:33:19 crc kubenswrapper[4988]: E1008 18:33:19.204151 4988 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 863aa8ddeda792a78ba5af321605974ac7082d24fc8fb6f0400148fcf3f06126 is running failed: container process not found" containerID="863aa8ddeda792a78ba5af321605974ac7082d24fc8fb6f0400148fcf3f06126" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Oct 08 18:33:19 crc kubenswrapper[4988]: E1008 18:33:19.204197 4988 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 863aa8ddeda792a78ba5af321605974ac7082d24fc8fb6f0400148fcf3f06126 is running failed: container process not found" probeType="Readiness" pod="openstack/ovn-controller-ovs-kc7fw" podUID="f1628013-e0b9-4c31-80c0-91aabaaef0f6" containerName="ovsdb-server" Oct 08 18:33:19 crc kubenswrapper[4988]: E1008 18:33:19.204837 4988 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="8a5332d3d9aa50bb56f12e04b1b32520399363308f9e75d5510ddd01f41593c3" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Oct 08 18:33:19 crc kubenswrapper[4988]: E1008 18:33:19.206600 4988 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="8a5332d3d9aa50bb56f12e04b1b32520399363308f9e75d5510ddd01f41593c3" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Oct 08 18:33:19 crc kubenswrapper[4988]: E1008 18:33:19.208757 4988 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="8a5332d3d9aa50bb56f12e04b1b32520399363308f9e75d5510ddd01f41593c3" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Oct 08 18:33:19 crc kubenswrapper[4988]: E1008 18:33:19.208874 4988 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/ovn-controller-ovs-kc7fw" podUID="f1628013-e0b9-4c31-80c0-91aabaaef0f6" containerName="ovs-vswitchd" Oct 08 18:33:23 crc kubenswrapper[4988]: I1008 18:33:23.337805 4988 patch_prober.go:28] interesting pod/machine-config-daemon-nm54f container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 18:33:23 crc kubenswrapper[4988]: I1008 18:33:23.338092 4988 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 18:33:24 crc kubenswrapper[4988]: E1008 18:33:24.203046 4988 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 863aa8ddeda792a78ba5af321605974ac7082d24fc8fb6f0400148fcf3f06126 is running failed: container process not found" containerID="863aa8ddeda792a78ba5af321605974ac7082d24fc8fb6f0400148fcf3f06126" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Oct 08 18:33:24 crc kubenswrapper[4988]: E1008 18:33:24.203571 4988 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="8a5332d3d9aa50bb56f12e04b1b32520399363308f9e75d5510ddd01f41593c3" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Oct 08 18:33:24 crc kubenswrapper[4988]: E1008 18:33:24.203608 4988 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 863aa8ddeda792a78ba5af321605974ac7082d24fc8fb6f0400148fcf3f06126 is running failed: container process not found" containerID="863aa8ddeda792a78ba5af321605974ac7082d24fc8fb6f0400148fcf3f06126" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Oct 08 18:33:24 crc kubenswrapper[4988]: E1008 18:33:24.204785 4988 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 863aa8ddeda792a78ba5af321605974ac7082d24fc8fb6f0400148fcf3f06126 is running failed: container process not found" containerID="863aa8ddeda792a78ba5af321605974ac7082d24fc8fb6f0400148fcf3f06126" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Oct 08 18:33:24 crc kubenswrapper[4988]: E1008 18:33:24.204860 4988 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 863aa8ddeda792a78ba5af321605974ac7082d24fc8fb6f0400148fcf3f06126 is running failed: container process not found" probeType="Readiness" pod="openstack/ovn-controller-ovs-kc7fw" podUID="f1628013-e0b9-4c31-80c0-91aabaaef0f6" containerName="ovsdb-server" Oct 08 18:33:24 crc kubenswrapper[4988]: E1008 18:33:24.205126 4988 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="8a5332d3d9aa50bb56f12e04b1b32520399363308f9e75d5510ddd01f41593c3" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Oct 08 18:33:24 crc kubenswrapper[4988]: E1008 18:33:24.206209 4988 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="8a5332d3d9aa50bb56f12e04b1b32520399363308f9e75d5510ddd01f41593c3" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Oct 08 18:33:24 crc kubenswrapper[4988]: E1008 18:33:24.206249 4988 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/ovn-controller-ovs-kc7fw" podUID="f1628013-e0b9-4c31-80c0-91aabaaef0f6" containerName="ovs-vswitchd" Oct 08 18:33:27 crc kubenswrapper[4988]: I1008 18:33:27.959201 4988 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-kc7fw_f1628013-e0b9-4c31-80c0-91aabaaef0f6/ovs-vswitchd/0.log" Oct 08 18:33:27 crc kubenswrapper[4988]: I1008 18:33:27.960788 4988 generic.go:334] "Generic (PLEG): container finished" podID="f1628013-e0b9-4c31-80c0-91aabaaef0f6" containerID="8a5332d3d9aa50bb56f12e04b1b32520399363308f9e75d5510ddd01f41593c3" exitCode=137 Oct 08 18:33:27 crc kubenswrapper[4988]: I1008 18:33:27.960837 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-kc7fw" event={"ID":"f1628013-e0b9-4c31-80c0-91aabaaef0f6","Type":"ContainerDied","Data":"8a5332d3d9aa50bb56f12e04b1b32520399363308f9e75d5510ddd01f41593c3"} Oct 08 18:33:28 crc kubenswrapper[4988]: I1008 18:33:28.377829 4988 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-kc7fw_f1628013-e0b9-4c31-80c0-91aabaaef0f6/ovs-vswitchd/0.log" Oct 08 18:33:28 crc kubenswrapper[4988]: I1008 18:33:28.379207 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-kc7fw" Oct 08 18:33:28 crc kubenswrapper[4988]: I1008 18:33:28.522083 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/f1628013-e0b9-4c31-80c0-91aabaaef0f6-var-run\") pod \"f1628013-e0b9-4c31-80c0-91aabaaef0f6\" (UID: \"f1628013-e0b9-4c31-80c0-91aabaaef0f6\") " Oct 08 18:33:28 crc kubenswrapper[4988]: I1008 18:33:28.522182 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f1628013-e0b9-4c31-80c0-91aabaaef0f6-var-run" (OuterVolumeSpecName: "var-run") pod "f1628013-e0b9-4c31-80c0-91aabaaef0f6" (UID: "f1628013-e0b9-4c31-80c0-91aabaaef0f6"). InnerVolumeSpecName "var-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 08 18:33:28 crc kubenswrapper[4988]: I1008 18:33:28.522221 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f1628013-e0b9-4c31-80c0-91aabaaef0f6-scripts\") pod \"f1628013-e0b9-4c31-80c0-91aabaaef0f6\" (UID: \"f1628013-e0b9-4c31-80c0-91aabaaef0f6\") " Oct 08 18:33:28 crc kubenswrapper[4988]: I1008 18:33:28.522303 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/f1628013-e0b9-4c31-80c0-91aabaaef0f6-etc-ovs\") pod \"f1628013-e0b9-4c31-80c0-91aabaaef0f6\" (UID: \"f1628013-e0b9-4c31-80c0-91aabaaef0f6\") " Oct 08 18:33:28 crc kubenswrapper[4988]: I1008 18:33:28.522657 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f1628013-e0b9-4c31-80c0-91aabaaef0f6-etc-ovs" (OuterVolumeSpecName: "etc-ovs") pod "f1628013-e0b9-4c31-80c0-91aabaaef0f6" (UID: "f1628013-e0b9-4c31-80c0-91aabaaef0f6"). InnerVolumeSpecName "etc-ovs". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 08 18:33:28 crc kubenswrapper[4988]: I1008 18:33:28.522869 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/f1628013-e0b9-4c31-80c0-91aabaaef0f6-var-lib\") pod \"f1628013-e0b9-4c31-80c0-91aabaaef0f6\" (UID: \"f1628013-e0b9-4c31-80c0-91aabaaef0f6\") " Oct 08 18:33:28 crc kubenswrapper[4988]: I1008 18:33:28.522938 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f1628013-e0b9-4c31-80c0-91aabaaef0f6-var-log\") pod \"f1628013-e0b9-4c31-80c0-91aabaaef0f6\" (UID: \"f1628013-e0b9-4c31-80c0-91aabaaef0f6\") " Oct 08 18:33:28 crc kubenswrapper[4988]: I1008 18:33:28.522984 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f1628013-e0b9-4c31-80c0-91aabaaef0f6-var-log" (OuterVolumeSpecName: "var-log") pod "f1628013-e0b9-4c31-80c0-91aabaaef0f6" (UID: "f1628013-e0b9-4c31-80c0-91aabaaef0f6"). InnerVolumeSpecName "var-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 08 18:33:28 crc kubenswrapper[4988]: I1008 18:33:28.523040 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mhxmj\" (UniqueName: \"kubernetes.io/projected/f1628013-e0b9-4c31-80c0-91aabaaef0f6-kube-api-access-mhxmj\") pod \"f1628013-e0b9-4c31-80c0-91aabaaef0f6\" (UID: \"f1628013-e0b9-4c31-80c0-91aabaaef0f6\") " Oct 08 18:33:28 crc kubenswrapper[4988]: I1008 18:33:28.523131 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f1628013-e0b9-4c31-80c0-91aabaaef0f6-var-lib" (OuterVolumeSpecName: "var-lib") pod "f1628013-e0b9-4c31-80c0-91aabaaef0f6" (UID: "f1628013-e0b9-4c31-80c0-91aabaaef0f6"). InnerVolumeSpecName "var-lib". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 08 18:33:28 crc kubenswrapper[4988]: I1008 18:33:28.523446 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f1628013-e0b9-4c31-80c0-91aabaaef0f6-scripts" (OuterVolumeSpecName: "scripts") pod "f1628013-e0b9-4c31-80c0-91aabaaef0f6" (UID: "f1628013-e0b9-4c31-80c0-91aabaaef0f6"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 18:33:28 crc kubenswrapper[4988]: I1008 18:33:28.523529 4988 reconciler_common.go:293] "Volume detached for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/f1628013-e0b9-4c31-80c0-91aabaaef0f6-var-run\") on node \"crc\" DevicePath \"\"" Oct 08 18:33:28 crc kubenswrapper[4988]: I1008 18:33:28.523544 4988 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f1628013-e0b9-4c31-80c0-91aabaaef0f6-scripts\") on node \"crc\" DevicePath \"\"" Oct 08 18:33:28 crc kubenswrapper[4988]: I1008 18:33:28.523555 4988 reconciler_common.go:293] "Volume detached for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/f1628013-e0b9-4c31-80c0-91aabaaef0f6-etc-ovs\") on node \"crc\" DevicePath \"\"" Oct 08 18:33:28 crc kubenswrapper[4988]: I1008 18:33:28.523566 4988 reconciler_common.go:293] "Volume detached for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/f1628013-e0b9-4c31-80c0-91aabaaef0f6-var-lib\") on node \"crc\" DevicePath \"\"" Oct 08 18:33:28 crc kubenswrapper[4988]: I1008 18:33:28.523575 4988 reconciler_common.go:293] "Volume detached for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f1628013-e0b9-4c31-80c0-91aabaaef0f6-var-log\") on node \"crc\" DevicePath \"\"" Oct 08 18:33:28 crc kubenswrapper[4988]: I1008 18:33:28.528137 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f1628013-e0b9-4c31-80c0-91aabaaef0f6-kube-api-access-mhxmj" (OuterVolumeSpecName: "kube-api-access-mhxmj") pod "f1628013-e0b9-4c31-80c0-91aabaaef0f6" (UID: "f1628013-e0b9-4c31-80c0-91aabaaef0f6"). InnerVolumeSpecName "kube-api-access-mhxmj". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:33:28 crc kubenswrapper[4988]: I1008 18:33:28.624901 4988 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mhxmj\" (UniqueName: \"kubernetes.io/projected/f1628013-e0b9-4c31-80c0-91aabaaef0f6-kube-api-access-mhxmj\") on node \"crc\" DevicePath \"\"" Oct 08 18:33:28 crc kubenswrapper[4988]: I1008 18:33:28.976842 4988 generic.go:334] "Generic (PLEG): container finished" podID="d04259e5-9595-447b-b804-feba577498d8" containerID="b723050435b720195f0b9b9fbdfc7bf95b8dd19f190a2d21120788a8cde0603f" exitCode=137 Oct 08 18:33:28 crc kubenswrapper[4988]: I1008 18:33:28.976910 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"d04259e5-9595-447b-b804-feba577498d8","Type":"ContainerDied","Data":"b723050435b720195f0b9b9fbdfc7bf95b8dd19f190a2d21120788a8cde0603f"} Oct 08 18:33:28 crc kubenswrapper[4988]: I1008 18:33:28.979169 4988 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-kc7fw_f1628013-e0b9-4c31-80c0-91aabaaef0f6/ovs-vswitchd/0.log" Oct 08 18:33:28 crc kubenswrapper[4988]: I1008 18:33:28.979987 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-kc7fw" event={"ID":"f1628013-e0b9-4c31-80c0-91aabaaef0f6","Type":"ContainerDied","Data":"8464f2b7c6120e931f4c293370cfc0d2b039813f7d469e2e29ad85a2d9185c98"} Oct 08 18:33:28 crc kubenswrapper[4988]: I1008 18:33:28.980046 4988 scope.go:117] "RemoveContainer" containerID="8a5332d3d9aa50bb56f12e04b1b32520399363308f9e75d5510ddd01f41593c3" Oct 08 18:33:28 crc kubenswrapper[4988]: I1008 18:33:28.980059 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-kc7fw" Oct 08 18:33:29 crc kubenswrapper[4988]: I1008 18:33:29.008744 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-ovs-kc7fw"] Oct 08 18:33:29 crc kubenswrapper[4988]: I1008 18:33:29.014169 4988 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-ovs-kc7fw"] Oct 08 18:33:29 crc kubenswrapper[4988]: I1008 18:33:29.016104 4988 scope.go:117] "RemoveContainer" containerID="863aa8ddeda792a78ba5af321605974ac7082d24fc8fb6f0400148fcf3f06126" Oct 08 18:33:29 crc kubenswrapper[4988]: I1008 18:33:29.040323 4988 scope.go:117] "RemoveContainer" containerID="8d3bc1f9ff6541935bebe925efdef3e65cd202a230aed96556f734e8974eaf13" Oct 08 18:33:29 crc kubenswrapper[4988]: I1008 18:33:29.253379 4988 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f1628013-e0b9-4c31-80c0-91aabaaef0f6" path="/var/lib/kubelet/pods/f1628013-e0b9-4c31-80c0-91aabaaef0f6/volumes" Oct 08 18:33:29 crc kubenswrapper[4988]: I1008 18:33:29.262229 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Oct 08 18:33:29 crc kubenswrapper[4988]: I1008 18:33:29.333832 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vfz4q\" (UniqueName: \"kubernetes.io/projected/d04259e5-9595-447b-b804-feba577498d8-kube-api-access-vfz4q\") pod \"d04259e5-9595-447b-b804-feba577498d8\" (UID: \"d04259e5-9595-447b-b804-feba577498d8\") " Oct 08 18:33:29 crc kubenswrapper[4988]: I1008 18:33:29.333907 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/d04259e5-9595-447b-b804-feba577498d8-etc-swift\") pod \"d04259e5-9595-447b-b804-feba577498d8\" (UID: \"d04259e5-9595-447b-b804-feba577498d8\") " Oct 08 18:33:29 crc kubenswrapper[4988]: I1008 18:33:29.333937 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/d04259e5-9595-447b-b804-feba577498d8-cache\") pod \"d04259e5-9595-447b-b804-feba577498d8\" (UID: \"d04259e5-9595-447b-b804-feba577498d8\") " Oct 08 18:33:29 crc kubenswrapper[4988]: I1008 18:33:29.334044 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/d04259e5-9595-447b-b804-feba577498d8-lock\") pod \"d04259e5-9595-447b-b804-feba577498d8\" (UID: \"d04259e5-9595-447b-b804-feba577498d8\") " Oct 08 18:33:29 crc kubenswrapper[4988]: I1008 18:33:29.334061 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swift\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"d04259e5-9595-447b-b804-feba577498d8\" (UID: \"d04259e5-9595-447b-b804-feba577498d8\") " Oct 08 18:33:29 crc kubenswrapper[4988]: I1008 18:33:29.335981 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d04259e5-9595-447b-b804-feba577498d8-lock" (OuterVolumeSpecName: "lock") pod "d04259e5-9595-447b-b804-feba577498d8" (UID: "d04259e5-9595-447b-b804-feba577498d8"). InnerVolumeSpecName "lock". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 18:33:29 crc kubenswrapper[4988]: I1008 18:33:29.336478 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d04259e5-9595-447b-b804-feba577498d8-cache" (OuterVolumeSpecName: "cache") pod "d04259e5-9595-447b-b804-feba577498d8" (UID: "d04259e5-9595-447b-b804-feba577498d8"). InnerVolumeSpecName "cache". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 18:33:29 crc kubenswrapper[4988]: I1008 18:33:29.340204 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d04259e5-9595-447b-b804-feba577498d8-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "d04259e5-9595-447b-b804-feba577498d8" (UID: "d04259e5-9595-447b-b804-feba577498d8"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:33:29 crc kubenswrapper[4988]: I1008 18:33:29.341037 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage06-crc" (OuterVolumeSpecName: "swift") pod "d04259e5-9595-447b-b804-feba577498d8" (UID: "d04259e5-9595-447b-b804-feba577498d8"). InnerVolumeSpecName "local-storage06-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 08 18:33:29 crc kubenswrapper[4988]: I1008 18:33:29.341555 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d04259e5-9595-447b-b804-feba577498d8-kube-api-access-vfz4q" (OuterVolumeSpecName: "kube-api-access-vfz4q") pod "d04259e5-9595-447b-b804-feba577498d8" (UID: "d04259e5-9595-447b-b804-feba577498d8"). InnerVolumeSpecName "kube-api-access-vfz4q". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:33:29 crc kubenswrapper[4988]: I1008 18:33:29.435426 4988 reconciler_common.go:293] "Volume detached for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/d04259e5-9595-447b-b804-feba577498d8-lock\") on node \"crc\" DevicePath \"\"" Oct 08 18:33:29 crc kubenswrapper[4988]: I1008 18:33:29.435759 4988 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") on node \"crc\" " Oct 08 18:33:29 crc kubenswrapper[4988]: I1008 18:33:29.435904 4988 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vfz4q\" (UniqueName: \"kubernetes.io/projected/d04259e5-9595-447b-b804-feba577498d8-kube-api-access-vfz4q\") on node \"crc\" DevicePath \"\"" Oct 08 18:33:29 crc kubenswrapper[4988]: I1008 18:33:29.436000 4988 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/d04259e5-9595-447b-b804-feba577498d8-etc-swift\") on node \"crc\" DevicePath \"\"" Oct 08 18:33:29 crc kubenswrapper[4988]: I1008 18:33:29.436076 4988 reconciler_common.go:293] "Volume detached for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/d04259e5-9595-447b-b804-feba577498d8-cache\") on node \"crc\" DevicePath \"\"" Oct 08 18:33:29 crc kubenswrapper[4988]: I1008 18:33:29.449816 4988 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage06-crc" (UniqueName: "kubernetes.io/local-volume/local-storage06-crc") on node "crc" Oct 08 18:33:29 crc kubenswrapper[4988]: I1008 18:33:29.537460 4988 reconciler_common.go:293] "Volume detached for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") on node \"crc\" DevicePath \"\"" Oct 08 18:33:30 crc kubenswrapper[4988]: I1008 18:33:30.001426 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"d04259e5-9595-447b-b804-feba577498d8","Type":"ContainerDied","Data":"635706ae6f49645d3fc33dca27b64639e03636df81aa81c90a13bafd1aca31df"} Oct 08 18:33:30 crc kubenswrapper[4988]: I1008 18:33:30.001499 4988 scope.go:117] "RemoveContainer" containerID="b723050435b720195f0b9b9fbdfc7bf95b8dd19f190a2d21120788a8cde0603f" Oct 08 18:33:30 crc kubenswrapper[4988]: I1008 18:33:30.001602 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Oct 08 18:33:30 crc kubenswrapper[4988]: I1008 18:33:30.050074 4988 scope.go:117] "RemoveContainer" containerID="3e8f7969a53f0a0b492bfa6e755ebabd4ef007e6b42af4f280d07a7401c0699d" Oct 08 18:33:30 crc kubenswrapper[4988]: I1008 18:33:30.056407 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/swift-storage-0"] Oct 08 18:33:30 crc kubenswrapper[4988]: I1008 18:33:30.066203 4988 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/swift-storage-0"] Oct 08 18:33:30 crc kubenswrapper[4988]: I1008 18:33:30.098035 4988 scope.go:117] "RemoveContainer" containerID="f6d8e476ef92ad2a5ae0468a14ebdbcb37b6c35a714f5c83ec9ff493a5dd15a3" Oct 08 18:33:30 crc kubenswrapper[4988]: I1008 18:33:30.125546 4988 scope.go:117] "RemoveContainer" containerID="9935bea63badb3e4cb30261f8985a1224e374fbe56ec5af0a913e1c5fe57db04" Oct 08 18:33:30 crc kubenswrapper[4988]: I1008 18:33:30.146380 4988 scope.go:117] "RemoveContainer" containerID="0f92f3d1af5000c50ceb5a04475cb2043188883ce2d85323918a574561017295" Oct 08 18:33:30 crc kubenswrapper[4988]: I1008 18:33:30.165238 4988 scope.go:117] "RemoveContainer" containerID="6c62379d0b1b37e153eff91e53dd273d7d1e85bb5f09dd3abdfc7a5b862d3459" Oct 08 18:33:30 crc kubenswrapper[4988]: I1008 18:33:30.180677 4988 scope.go:117] "RemoveContainer" containerID="94be3d977f16a3f88c65d1405dd7fd0737c52d3e6c73b44f0651409fc2428746" Oct 08 18:33:30 crc kubenswrapper[4988]: I1008 18:33:30.195353 4988 scope.go:117] "RemoveContainer" containerID="31e32c3833aea9f83bda1aca6904be1c190d6d671f11ad9f856ce5abce9053fb" Oct 08 18:33:30 crc kubenswrapper[4988]: I1008 18:33:30.215781 4988 scope.go:117] "RemoveContainer" containerID="4f7780b6a0c60b8247000b0296848f37576f4cd48b317bfb6bd9807dc977f068" Oct 08 18:33:30 crc kubenswrapper[4988]: I1008 18:33:30.234766 4988 scope.go:117] "RemoveContainer" containerID="c1f4256e3ed5eca7a8d3e7c8f01715b4bf6cc7778eac4b717acbddef65864b05" Oct 08 18:33:30 crc kubenswrapper[4988]: I1008 18:33:30.254101 4988 scope.go:117] "RemoveContainer" containerID="590b8a61b67584c2acbd2bea77795253fad7c7d2a79eb0fb7fde120c857233af" Oct 08 18:33:30 crc kubenswrapper[4988]: I1008 18:33:30.273608 4988 scope.go:117] "RemoveContainer" containerID="e8f70c43dc552f28dc00baeb404e73ce329584c5d7563ccfd1bb7e213c973bc7" Oct 08 18:33:30 crc kubenswrapper[4988]: I1008 18:33:30.291453 4988 scope.go:117] "RemoveContainer" containerID="588710f5625320484d4ef400733b3ae63b0e2f936f619df2ca3b5435aa10a8c9" Oct 08 18:33:30 crc kubenswrapper[4988]: I1008 18:33:30.314082 4988 scope.go:117] "RemoveContainer" containerID="23ee960ca35b0ceb291c74103d14d93f4ff48fb0b2b08acdacc1fd66a0fb6327" Oct 08 18:33:30 crc kubenswrapper[4988]: I1008 18:33:30.341627 4988 scope.go:117] "RemoveContainer" containerID="d0e7528b976d6c731e5ab211ed178a97c7d155611aadc8e3e7dde446d2824719" Oct 08 18:33:31 crc kubenswrapper[4988]: I1008 18:33:31.246304 4988 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d04259e5-9595-447b-b804-feba577498d8" path="/var/lib/kubelet/pods/d04259e5-9595-447b-b804-feba577498d8/volumes" Oct 08 18:33:33 crc kubenswrapper[4988]: I1008 18:33:33.693344 4988 pod_container_manager_linux.go:210] "Failed to delete cgroup paths" cgroupName=["kubepods","besteffort","pod28f73b7a-bb5e-43ec-a099-13a12cdb33d3"] err="unable to destroy cgroup paths for cgroup [kubepods besteffort pod28f73b7a-bb5e-43ec-a099-13a12cdb33d3] : Timed out while waiting for systemd to remove kubepods-besteffort-pod28f73b7a_bb5e_43ec_a099_13a12cdb33d3.slice" Oct 08 18:33:34 crc kubenswrapper[4988]: I1008 18:33:34.077799 4988 pod_container_manager_linux.go:210] "Failed to delete cgroup paths" cgroupName=["kubepods","besteffort","pod0fcc9499-a06b-4cd1-a5da-ea8108916050"] err="unable to destroy cgroup paths for cgroup [kubepods besteffort pod0fcc9499-a06b-4cd1-a5da-ea8108916050] : Timed out while waiting for systemd to remove kubepods-besteffort-pod0fcc9499_a06b_4cd1_a5da_ea8108916050.slice" Oct 08 18:33:34 crc kubenswrapper[4988]: E1008 18:33:34.077900 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to delete cgroup paths for [kubepods besteffort pod0fcc9499-a06b-4cd1-a5da-ea8108916050] : unable to destroy cgroup paths for cgroup [kubepods besteffort pod0fcc9499-a06b-4cd1-a5da-ea8108916050] : Timed out while waiting for systemd to remove kubepods-besteffort-pod0fcc9499_a06b_4cd1_a5da_ea8108916050.slice" pod="openstack/glance-default-internal-api-0" podUID="0fcc9499-a06b-4cd1-a5da-ea8108916050" Oct 08 18:33:35 crc kubenswrapper[4988]: I1008 18:33:35.062871 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 08 18:33:35 crc kubenswrapper[4988]: I1008 18:33:35.095564 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 08 18:33:35 crc kubenswrapper[4988]: I1008 18:33:35.103949 4988 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 08 18:33:35 crc kubenswrapper[4988]: I1008 18:33:35.253700 4988 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0fcc9499-a06b-4cd1-a5da-ea8108916050" path="/var/lib/kubelet/pods/0fcc9499-a06b-4cd1-a5da-ea8108916050/volumes" Oct 08 18:33:53 crc kubenswrapper[4988]: I1008 18:33:53.338575 4988 patch_prober.go:28] interesting pod/machine-config-daemon-nm54f container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 18:33:53 crc kubenswrapper[4988]: I1008 18:33:53.339564 4988 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 18:33:53 crc kubenswrapper[4988]: I1008 18:33:53.339644 4988 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" Oct 08 18:33:53 crc kubenswrapper[4988]: I1008 18:33:53.340743 4988 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"e0a9bc6112323fbe5f5506d9ba72b002a9b9faef1191582e4a39898dc43fcbbb"} pod="openshift-machine-config-operator/machine-config-daemon-nm54f" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 08 18:33:53 crc kubenswrapper[4988]: I1008 18:33:53.340878 4988 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" containerName="machine-config-daemon" containerID="cri-o://e0a9bc6112323fbe5f5506d9ba72b002a9b9faef1191582e4a39898dc43fcbbb" gracePeriod=600 Oct 08 18:33:54 crc kubenswrapper[4988]: I1008 18:33:54.288577 4988 generic.go:334] "Generic (PLEG): container finished" podID="f328a20b-5283-407e-8a93-2ab2da599f18" containerID="e0a9bc6112323fbe5f5506d9ba72b002a9b9faef1191582e4a39898dc43fcbbb" exitCode=0 Oct 08 18:33:54 crc kubenswrapper[4988]: I1008 18:33:54.288651 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" event={"ID":"f328a20b-5283-407e-8a93-2ab2da599f18","Type":"ContainerDied","Data":"e0a9bc6112323fbe5f5506d9ba72b002a9b9faef1191582e4a39898dc43fcbbb"} Oct 08 18:33:54 crc kubenswrapper[4988]: I1008 18:33:54.288946 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" event={"ID":"f328a20b-5283-407e-8a93-2ab2da599f18","Type":"ContainerStarted","Data":"1e3c2ea55d3c49c9fa802ab01ed9324a6c941b8da678ffaccb80377e3f159aee"} Oct 08 18:33:54 crc kubenswrapper[4988]: I1008 18:33:54.288973 4988 scope.go:117] "RemoveContainer" containerID="33dcd4d530d69056c83e9e7f7862766d54a3df5b03043465d12788d2876eac58" Oct 08 18:33:55 crc kubenswrapper[4988]: I1008 18:33:55.634020 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-xbvnj"] Oct 08 18:33:55 crc kubenswrapper[4988]: E1008 18:33:55.634827 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d04259e5-9595-447b-b804-feba577498d8" containerName="container-replicator" Oct 08 18:33:55 crc kubenswrapper[4988]: I1008 18:33:55.634844 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="d04259e5-9595-447b-b804-feba577498d8" containerName="container-replicator" Oct 08 18:33:55 crc kubenswrapper[4988]: E1008 18:33:55.634870 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1b24865e-cac2-490e-9e61-de22a379ff47" containerName="nova-scheduler-scheduler" Oct 08 18:33:55 crc kubenswrapper[4988]: I1008 18:33:55.634879 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="1b24865e-cac2-490e-9e61-de22a379ff47" containerName="nova-scheduler-scheduler" Oct 08 18:33:55 crc kubenswrapper[4988]: E1008 18:33:55.634889 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e5a8f819-c91d-429e-b848-e9c444739623" containerName="rabbitmq" Oct 08 18:33:55 crc kubenswrapper[4988]: I1008 18:33:55.634897 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="e5a8f819-c91d-429e-b848-e9c444739623" containerName="rabbitmq" Oct 08 18:33:55 crc kubenswrapper[4988]: E1008 18:33:55.634908 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d04259e5-9595-447b-b804-feba577498d8" containerName="account-reaper" Oct 08 18:33:55 crc kubenswrapper[4988]: I1008 18:33:55.634917 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="d04259e5-9595-447b-b804-feba577498d8" containerName="account-reaper" Oct 08 18:33:55 crc kubenswrapper[4988]: E1008 18:33:55.634928 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3e8c874a-029a-4d6a-bc2e-00aeed1636a7" containerName="mariadb-account-delete" Oct 08 18:33:55 crc kubenswrapper[4988]: I1008 18:33:55.634935 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="3e8c874a-029a-4d6a-bc2e-00aeed1636a7" containerName="mariadb-account-delete" Oct 08 18:33:55 crc kubenswrapper[4988]: E1008 18:33:55.634945 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b6d1becf-4229-40bf-b53a-d8a9eb4b580e" containerName="kube-state-metrics" Oct 08 18:33:55 crc kubenswrapper[4988]: I1008 18:33:55.634952 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="b6d1becf-4229-40bf-b53a-d8a9eb4b580e" containerName="kube-state-metrics" Oct 08 18:33:55 crc kubenswrapper[4988]: E1008 18:33:55.634971 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3f7bddd3-86f8-46e9-a38e-1aa5e570d151" containerName="mariadb-account-delete" Oct 08 18:33:55 crc kubenswrapper[4988]: I1008 18:33:55.634981 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="3f7bddd3-86f8-46e9-a38e-1aa5e570d151" containerName="mariadb-account-delete" Oct 08 18:33:55 crc kubenswrapper[4988]: E1008 18:33:55.634996 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c9f172a3-9188-4f5f-bf3a-fbc3b0dc5206" containerName="barbican-keystone-listener" Oct 08 18:33:55 crc kubenswrapper[4988]: I1008 18:33:55.635004 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="c9f172a3-9188-4f5f-bf3a-fbc3b0dc5206" containerName="barbican-keystone-listener" Oct 08 18:33:55 crc kubenswrapper[4988]: E1008 18:33:55.635014 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3a57537b-6115-4249-8d16-3ab0c4b6d21e" containerName="barbican-api" Oct 08 18:33:55 crc kubenswrapper[4988]: I1008 18:33:55.635023 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="3a57537b-6115-4249-8d16-3ab0c4b6d21e" containerName="barbican-api" Oct 08 18:33:55 crc kubenswrapper[4988]: E1008 18:33:55.635034 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4e47608-89b4-45d1-88a6-fe75805d864d" containerName="placement-log" Oct 08 18:33:55 crc kubenswrapper[4988]: I1008 18:33:55.635041 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4e47608-89b4-45d1-88a6-fe75805d864d" containerName="placement-log" Oct 08 18:33:55 crc kubenswrapper[4988]: E1008 18:33:55.635059 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0fcc9499-a06b-4cd1-a5da-ea8108916050" containerName="glance-httpd" Oct 08 18:33:55 crc kubenswrapper[4988]: I1008 18:33:55.635067 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="0fcc9499-a06b-4cd1-a5da-ea8108916050" containerName="glance-httpd" Oct 08 18:33:55 crc kubenswrapper[4988]: E1008 18:33:55.635079 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d04259e5-9595-447b-b804-feba577498d8" containerName="object-updater" Oct 08 18:33:55 crc kubenswrapper[4988]: I1008 18:33:55.635087 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="d04259e5-9595-447b-b804-feba577498d8" containerName="object-updater" Oct 08 18:33:55 crc kubenswrapper[4988]: E1008 18:33:55.635094 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ab3315d2-3248-42f1-9834-2bab55fa9935" containerName="mariadb-account-delete" Oct 08 18:33:55 crc kubenswrapper[4988]: I1008 18:33:55.635102 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="ab3315d2-3248-42f1-9834-2bab55fa9935" containerName="mariadb-account-delete" Oct 08 18:33:55 crc kubenswrapper[4988]: E1008 18:33:55.635116 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ccd2f8fa-1c16-4c58-9f87-e60ba3a2f5fb" containerName="proxy-httpd" Oct 08 18:33:55 crc kubenswrapper[4988]: I1008 18:33:55.635123 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="ccd2f8fa-1c16-4c58-9f87-e60ba3a2f5fb" containerName="proxy-httpd" Oct 08 18:33:55 crc kubenswrapper[4988]: E1008 18:33:55.635139 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3503942b-2825-4006-80e8-8c4610b89997" containerName="galera" Oct 08 18:33:55 crc kubenswrapper[4988]: I1008 18:33:55.635148 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="3503942b-2825-4006-80e8-8c4610b89997" containerName="galera" Oct 08 18:33:55 crc kubenswrapper[4988]: E1008 18:33:55.635156 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d04259e5-9595-447b-b804-feba577498d8" containerName="object-replicator" Oct 08 18:33:55 crc kubenswrapper[4988]: I1008 18:33:55.635164 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="d04259e5-9595-447b-b804-feba577498d8" containerName="object-replicator" Oct 08 18:33:55 crc kubenswrapper[4988]: E1008 18:33:55.635172 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d7d648e7-3c02-4d26-993d-a901764b3329" containerName="mariadb-account-delete" Oct 08 18:33:55 crc kubenswrapper[4988]: I1008 18:33:55.635180 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="d7d648e7-3c02-4d26-993d-a901764b3329" containerName="mariadb-account-delete" Oct 08 18:33:55 crc kubenswrapper[4988]: E1008 18:33:55.635189 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="28f73b7a-bb5e-43ec-a099-13a12cdb33d3" containerName="mariadb-account-delete" Oct 08 18:33:55 crc kubenswrapper[4988]: I1008 18:33:55.635199 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="28f73b7a-bb5e-43ec-a099-13a12cdb33d3" containerName="mariadb-account-delete" Oct 08 18:33:55 crc kubenswrapper[4988]: E1008 18:33:55.635215 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d04259e5-9595-447b-b804-feba577498d8" containerName="container-server" Oct 08 18:33:55 crc kubenswrapper[4988]: I1008 18:33:55.635223 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="d04259e5-9595-447b-b804-feba577498d8" containerName="container-server" Oct 08 18:33:55 crc kubenswrapper[4988]: E1008 18:33:55.635233 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="79cbf065-a458-4dec-b6a7-805f515142e0" containerName="mariadb-account-delete" Oct 08 18:33:55 crc kubenswrapper[4988]: I1008 18:33:55.635240 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="79cbf065-a458-4dec-b6a7-805f515142e0" containerName="mariadb-account-delete" Oct 08 18:33:55 crc kubenswrapper[4988]: E1008 18:33:55.635254 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d04259e5-9595-447b-b804-feba577498d8" containerName="swift-recon-cron" Oct 08 18:33:55 crc kubenswrapper[4988]: I1008 18:33:55.635263 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="d04259e5-9595-447b-b804-feba577498d8" containerName="swift-recon-cron" Oct 08 18:33:55 crc kubenswrapper[4988]: E1008 18:33:55.635327 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="72c64965-0b57-43e1-8d18-2bce6cabbfa1" containerName="mysql-bootstrap" Oct 08 18:33:55 crc kubenswrapper[4988]: I1008 18:33:55.635338 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="72c64965-0b57-43e1-8d18-2bce6cabbfa1" containerName="mysql-bootstrap" Oct 08 18:33:55 crc kubenswrapper[4988]: E1008 18:33:55.635350 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d04259e5-9595-447b-b804-feba577498d8" containerName="object-server" Oct 08 18:33:55 crc kubenswrapper[4988]: I1008 18:33:55.635358 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="d04259e5-9595-447b-b804-feba577498d8" containerName="object-server" Oct 08 18:33:55 crc kubenswrapper[4988]: E1008 18:33:55.635375 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d04259e5-9595-447b-b804-feba577498d8" containerName="object-expirer" Oct 08 18:33:55 crc kubenswrapper[4988]: I1008 18:33:55.635397 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="d04259e5-9595-447b-b804-feba577498d8" containerName="object-expirer" Oct 08 18:33:55 crc kubenswrapper[4988]: E1008 18:33:55.635408 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0ec345a3-1744-4e78-bf13-91f0325296e4" containerName="cinder-scheduler" Oct 08 18:33:55 crc kubenswrapper[4988]: I1008 18:33:55.635416 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="0ec345a3-1744-4e78-bf13-91f0325296e4" containerName="cinder-scheduler" Oct 08 18:33:55 crc kubenswrapper[4988]: E1008 18:33:55.635429 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d04259e5-9595-447b-b804-feba577498d8" containerName="object-auditor" Oct 08 18:33:55 crc kubenswrapper[4988]: I1008 18:33:55.635439 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="d04259e5-9595-447b-b804-feba577498d8" containerName="object-auditor" Oct 08 18:33:55 crc kubenswrapper[4988]: E1008 18:33:55.635450 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0fcc9499-a06b-4cd1-a5da-ea8108916050" containerName="glance-log" Oct 08 18:33:55 crc kubenswrapper[4988]: I1008 18:33:55.635457 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="0fcc9499-a06b-4cd1-a5da-ea8108916050" containerName="glance-log" Oct 08 18:33:55 crc kubenswrapper[4988]: E1008 18:33:55.635469 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d04259e5-9595-447b-b804-feba577498d8" containerName="account-server" Oct 08 18:33:55 crc kubenswrapper[4988]: I1008 18:33:55.635478 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="d04259e5-9595-447b-b804-feba577498d8" containerName="account-server" Oct 08 18:33:55 crc kubenswrapper[4988]: E1008 18:33:55.635494 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3a57537b-6115-4249-8d16-3ab0c4b6d21e" containerName="barbican-api-log" Oct 08 18:33:55 crc kubenswrapper[4988]: I1008 18:33:55.635501 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="3a57537b-6115-4249-8d16-3ab0c4b6d21e" containerName="barbican-api-log" Oct 08 18:33:55 crc kubenswrapper[4988]: E1008 18:33:55.635514 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e5a8f819-c91d-429e-b848-e9c444739623" containerName="setup-container" Oct 08 18:33:55 crc kubenswrapper[4988]: I1008 18:33:55.635521 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="e5a8f819-c91d-429e-b848-e9c444739623" containerName="setup-container" Oct 08 18:33:55 crc kubenswrapper[4988]: E1008 18:33:55.635534 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ccd2f8fa-1c16-4c58-9f87-e60ba3a2f5fb" containerName="sg-core" Oct 08 18:33:55 crc kubenswrapper[4988]: I1008 18:33:55.635543 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="ccd2f8fa-1c16-4c58-9f87-e60ba3a2f5fb" containerName="sg-core" Oct 08 18:33:55 crc kubenswrapper[4988]: E1008 18:33:55.635557 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="56702a03-175d-4a54-b84f-e25cab037431" containerName="glance-httpd" Oct 08 18:33:55 crc kubenswrapper[4988]: I1008 18:33:55.635564 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="56702a03-175d-4a54-b84f-e25cab037431" containerName="glance-httpd" Oct 08 18:33:55 crc kubenswrapper[4988]: E1008 18:33:55.635573 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0ec345a3-1744-4e78-bf13-91f0325296e4" containerName="probe" Oct 08 18:33:55 crc kubenswrapper[4988]: I1008 18:33:55.635580 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="0ec345a3-1744-4e78-bf13-91f0325296e4" containerName="probe" Oct 08 18:33:55 crc kubenswrapper[4988]: E1008 18:33:55.635589 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ccd2f8fa-1c16-4c58-9f87-e60ba3a2f5fb" containerName="ceilometer-central-agent" Oct 08 18:33:55 crc kubenswrapper[4988]: I1008 18:33:55.635599 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="ccd2f8fa-1c16-4c58-9f87-e60ba3a2f5fb" containerName="ceilometer-central-agent" Oct 08 18:33:55 crc kubenswrapper[4988]: E1008 18:33:55.635610 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c9f172a3-9188-4f5f-bf3a-fbc3b0dc5206" containerName="barbican-keystone-listener-log" Oct 08 18:33:55 crc kubenswrapper[4988]: I1008 18:33:55.635619 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="c9f172a3-9188-4f5f-bf3a-fbc3b0dc5206" containerName="barbican-keystone-listener-log" Oct 08 18:33:55 crc kubenswrapper[4988]: E1008 18:33:55.635631 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="99d77eed-40f0-4879-be96-abea3a807cd9" containerName="cinder-api" Oct 08 18:33:55 crc kubenswrapper[4988]: I1008 18:33:55.635665 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="99d77eed-40f0-4879-be96-abea3a807cd9" containerName="cinder-api" Oct 08 18:33:55 crc kubenswrapper[4988]: E1008 18:33:55.635679 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="830d6ec5-0015-4a88-b8e8-0fd47f64fd46" containerName="setup-container" Oct 08 18:33:55 crc kubenswrapper[4988]: I1008 18:33:55.635686 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="830d6ec5-0015-4a88-b8e8-0fd47f64fd46" containerName="setup-container" Oct 08 18:33:55 crc kubenswrapper[4988]: E1008 18:33:55.635698 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bd9cfcdf-4770-4efa-a41e-7867ec85206f" containerName="nova-cell0-conductor-conductor" Oct 08 18:33:55 crc kubenswrapper[4988]: I1008 18:33:55.635783 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="bd9cfcdf-4770-4efa-a41e-7867ec85206f" containerName="nova-cell0-conductor-conductor" Oct 08 18:33:55 crc kubenswrapper[4988]: E1008 18:33:55.635799 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f1628013-e0b9-4c31-80c0-91aabaaef0f6" containerName="ovsdb-server" Oct 08 18:33:55 crc kubenswrapper[4988]: I1008 18:33:55.635808 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="f1628013-e0b9-4c31-80c0-91aabaaef0f6" containerName="ovsdb-server" Oct 08 18:33:55 crc kubenswrapper[4988]: E1008 18:33:55.635822 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d04259e5-9595-447b-b804-feba577498d8" containerName="rsync" Oct 08 18:33:55 crc kubenswrapper[4988]: I1008 18:33:55.635829 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="d04259e5-9595-447b-b804-feba577498d8" containerName="rsync" Oct 08 18:33:55 crc kubenswrapper[4988]: E1008 18:33:55.635841 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="83ef7610-23ac-4445-b75c-ae7206b4d810" containerName="nova-metadata-metadata" Oct 08 18:33:55 crc kubenswrapper[4988]: I1008 18:33:55.635850 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="83ef7610-23ac-4445-b75c-ae7206b4d810" containerName="nova-metadata-metadata" Oct 08 18:33:55 crc kubenswrapper[4988]: E1008 18:33:55.635865 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d04259e5-9595-447b-b804-feba577498d8" containerName="container-auditor" Oct 08 18:33:55 crc kubenswrapper[4988]: I1008 18:33:55.635872 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="d04259e5-9595-447b-b804-feba577498d8" containerName="container-auditor" Oct 08 18:33:55 crc kubenswrapper[4988]: E1008 18:33:55.635882 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ccd2f8fa-1c16-4c58-9f87-e60ba3a2f5fb" containerName="ceilometer-notification-agent" Oct 08 18:33:55 crc kubenswrapper[4988]: I1008 18:33:55.635889 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="ccd2f8fa-1c16-4c58-9f87-e60ba3a2f5fb" containerName="ceilometer-notification-agent" Oct 08 18:33:55 crc kubenswrapper[4988]: E1008 18:33:55.635919 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="32bb7401-6292-4ca6-98f4-15522125c3ee" containerName="memcached" Oct 08 18:33:55 crc kubenswrapper[4988]: I1008 18:33:55.635932 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="32bb7401-6292-4ca6-98f4-15522125c3ee" containerName="memcached" Oct 08 18:33:55 crc kubenswrapper[4988]: E1008 18:33:55.635945 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="85bc7f8f-fdb3-4abb-88d3-ed0c10982c4c" containerName="nova-cell1-conductor-conductor" Oct 08 18:33:55 crc kubenswrapper[4988]: I1008 18:33:55.635953 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="85bc7f8f-fdb3-4abb-88d3-ed0c10982c4c" containerName="nova-cell1-conductor-conductor" Oct 08 18:33:55 crc kubenswrapper[4988]: E1008 18:33:55.635970 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d04259e5-9595-447b-b804-feba577498d8" containerName="account-auditor" Oct 08 18:33:55 crc kubenswrapper[4988]: I1008 18:33:55.635979 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="d04259e5-9595-447b-b804-feba577498d8" containerName="account-auditor" Oct 08 18:33:55 crc kubenswrapper[4988]: E1008 18:33:55.635993 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d04259e5-9595-447b-b804-feba577498d8" containerName="account-replicator" Oct 08 18:33:55 crc kubenswrapper[4988]: I1008 18:33:55.636001 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="d04259e5-9595-447b-b804-feba577498d8" containerName="account-replicator" Oct 08 18:33:55 crc kubenswrapper[4988]: E1008 18:33:55.636014 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4e47608-89b4-45d1-88a6-fe75805d864d" containerName="placement-api" Oct 08 18:33:55 crc kubenswrapper[4988]: I1008 18:33:55.636023 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4e47608-89b4-45d1-88a6-fe75805d864d" containerName="placement-api" Oct 08 18:33:55 crc kubenswrapper[4988]: E1008 18:33:55.636036 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f1628013-e0b9-4c31-80c0-91aabaaef0f6" containerName="ovsdb-server-init" Oct 08 18:33:55 crc kubenswrapper[4988]: I1008 18:33:55.636044 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="f1628013-e0b9-4c31-80c0-91aabaaef0f6" containerName="ovsdb-server-init" Oct 08 18:33:55 crc kubenswrapper[4988]: E1008 18:33:55.636058 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="830d6ec5-0015-4a88-b8e8-0fd47f64fd46" containerName="rabbitmq" Oct 08 18:33:55 crc kubenswrapper[4988]: I1008 18:33:55.636065 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="830d6ec5-0015-4a88-b8e8-0fd47f64fd46" containerName="rabbitmq" Oct 08 18:33:55 crc kubenswrapper[4988]: E1008 18:33:55.636076 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="26a3b0a9-6aca-490d-a6fb-08addef86fbb" containerName="keystone-api" Oct 08 18:33:55 crc kubenswrapper[4988]: I1008 18:33:55.636084 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="26a3b0a9-6aca-490d-a6fb-08addef86fbb" containerName="keystone-api" Oct 08 18:33:55 crc kubenswrapper[4988]: E1008 18:33:55.636094 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="83ef7610-23ac-4445-b75c-ae7206b4d810" containerName="nova-metadata-log" Oct 08 18:33:55 crc kubenswrapper[4988]: I1008 18:33:55.636101 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="83ef7610-23ac-4445-b75c-ae7206b4d810" containerName="nova-metadata-log" Oct 08 18:33:55 crc kubenswrapper[4988]: E1008 18:33:55.636114 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f1628013-e0b9-4c31-80c0-91aabaaef0f6" containerName="ovs-vswitchd" Oct 08 18:33:55 crc kubenswrapper[4988]: I1008 18:33:55.636122 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="f1628013-e0b9-4c31-80c0-91aabaaef0f6" containerName="ovs-vswitchd" Oct 08 18:33:55 crc kubenswrapper[4988]: E1008 18:33:55.636138 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d04259e5-9595-447b-b804-feba577498d8" containerName="container-updater" Oct 08 18:33:55 crc kubenswrapper[4988]: I1008 18:33:55.636146 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="d04259e5-9595-447b-b804-feba577498d8" containerName="container-updater" Oct 08 18:33:55 crc kubenswrapper[4988]: E1008 18:33:55.636159 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="02a37054-4817-4151-a8cb-a9a45867c0d3" containerName="mariadb-account-delete" Oct 08 18:33:55 crc kubenswrapper[4988]: I1008 18:33:55.636166 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="02a37054-4817-4151-a8cb-a9a45867c0d3" containerName="mariadb-account-delete" Oct 08 18:33:55 crc kubenswrapper[4988]: E1008 18:33:55.636177 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="56702a03-175d-4a54-b84f-e25cab037431" containerName="glance-log" Oct 08 18:33:55 crc kubenswrapper[4988]: I1008 18:33:55.636185 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="56702a03-175d-4a54-b84f-e25cab037431" containerName="glance-log" Oct 08 18:33:55 crc kubenswrapper[4988]: E1008 18:33:55.636198 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0efd1f53-90ad-4eb8-90f2-2ba9fc1593c2" containerName="nova-api-api" Oct 08 18:33:55 crc kubenswrapper[4988]: I1008 18:33:55.636205 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="0efd1f53-90ad-4eb8-90f2-2ba9fc1593c2" containerName="nova-api-api" Oct 08 18:33:55 crc kubenswrapper[4988]: E1008 18:33:55.636217 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c285e09d-c2f5-4642-9e6b-59b0ed3ed651" containerName="ovn-controller" Oct 08 18:33:55 crc kubenswrapper[4988]: I1008 18:33:55.636224 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="c285e09d-c2f5-4642-9e6b-59b0ed3ed651" containerName="ovn-controller" Oct 08 18:33:55 crc kubenswrapper[4988]: E1008 18:33:55.636233 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="72c64965-0b57-43e1-8d18-2bce6cabbfa1" containerName="galera" Oct 08 18:33:55 crc kubenswrapper[4988]: I1008 18:33:55.636240 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="72c64965-0b57-43e1-8d18-2bce6cabbfa1" containerName="galera" Oct 08 18:33:55 crc kubenswrapper[4988]: E1008 18:33:55.636250 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0efd1f53-90ad-4eb8-90f2-2ba9fc1593c2" containerName="nova-api-log" Oct 08 18:33:55 crc kubenswrapper[4988]: I1008 18:33:55.636259 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="0efd1f53-90ad-4eb8-90f2-2ba9fc1593c2" containerName="nova-api-log" Oct 08 18:33:55 crc kubenswrapper[4988]: E1008 18:33:55.636270 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3503942b-2825-4006-80e8-8c4610b89997" containerName="mysql-bootstrap" Oct 08 18:33:55 crc kubenswrapper[4988]: I1008 18:33:55.636278 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="3503942b-2825-4006-80e8-8c4610b89997" containerName="mysql-bootstrap" Oct 08 18:33:55 crc kubenswrapper[4988]: E1008 18:33:55.636287 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="99d77eed-40f0-4879-be96-abea3a807cd9" containerName="cinder-api-log" Oct 08 18:33:55 crc kubenswrapper[4988]: I1008 18:33:55.636294 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="99d77eed-40f0-4879-be96-abea3a807cd9" containerName="cinder-api-log" Oct 08 18:33:55 crc kubenswrapper[4988]: I1008 18:33:55.636475 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="c9f172a3-9188-4f5f-bf3a-fbc3b0dc5206" containerName="barbican-keystone-listener" Oct 08 18:33:55 crc kubenswrapper[4988]: I1008 18:33:55.636494 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="d7d648e7-3c02-4d26-993d-a901764b3329" containerName="mariadb-account-delete" Oct 08 18:33:55 crc kubenswrapper[4988]: I1008 18:33:55.636509 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="83ef7610-23ac-4445-b75c-ae7206b4d810" containerName="nova-metadata-log" Oct 08 18:33:55 crc kubenswrapper[4988]: I1008 18:33:55.636524 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="1b24865e-cac2-490e-9e61-de22a379ff47" containerName="nova-scheduler-scheduler" Oct 08 18:33:55 crc kubenswrapper[4988]: I1008 18:33:55.636531 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="d04259e5-9595-447b-b804-feba577498d8" containerName="account-auditor" Oct 08 18:33:55 crc kubenswrapper[4988]: I1008 18:33:55.636543 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="f1628013-e0b9-4c31-80c0-91aabaaef0f6" containerName="ovsdb-server" Oct 08 18:33:55 crc kubenswrapper[4988]: I1008 18:33:55.636561 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="ab3315d2-3248-42f1-9834-2bab55fa9935" containerName="mariadb-account-delete" Oct 08 18:33:55 crc kubenswrapper[4988]: I1008 18:33:55.636573 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="ccd2f8fa-1c16-4c58-9f87-e60ba3a2f5fb" containerName="proxy-httpd" Oct 08 18:33:55 crc kubenswrapper[4988]: I1008 18:33:55.636585 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="28f73b7a-bb5e-43ec-a099-13a12cdb33d3" containerName="mariadb-account-delete" Oct 08 18:33:55 crc kubenswrapper[4988]: I1008 18:33:55.636597 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="3e8c874a-029a-4d6a-bc2e-00aeed1636a7" containerName="mariadb-account-delete" Oct 08 18:33:55 crc kubenswrapper[4988]: I1008 18:33:55.636615 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="3a57537b-6115-4249-8d16-3ab0c4b6d21e" containerName="barbican-api" Oct 08 18:33:55 crc kubenswrapper[4988]: I1008 18:33:55.636625 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="83ef7610-23ac-4445-b75c-ae7206b4d810" containerName="nova-metadata-metadata" Oct 08 18:33:55 crc kubenswrapper[4988]: I1008 18:33:55.636638 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="c285e09d-c2f5-4642-9e6b-59b0ed3ed651" containerName="ovn-controller" Oct 08 18:33:55 crc kubenswrapper[4988]: I1008 18:33:55.636645 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="79cbf065-a458-4dec-b6a7-805f515142e0" containerName="mariadb-account-delete" Oct 08 18:33:55 crc kubenswrapper[4988]: I1008 18:33:55.636655 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="02a37054-4817-4151-a8cb-a9a45867c0d3" containerName="mariadb-account-delete" Oct 08 18:33:55 crc kubenswrapper[4988]: I1008 18:33:55.636671 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="72c64965-0b57-43e1-8d18-2bce6cabbfa1" containerName="galera" Oct 08 18:33:55 crc kubenswrapper[4988]: I1008 18:33:55.636680 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="ccd2f8fa-1c16-4c58-9f87-e60ba3a2f5fb" containerName="ceilometer-notification-agent" Oct 08 18:33:55 crc kubenswrapper[4988]: I1008 18:33:55.636694 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="26a3b0a9-6aca-490d-a6fb-08addef86fbb" containerName="keystone-api" Oct 08 18:33:55 crc kubenswrapper[4988]: I1008 18:33:55.636707 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="d04259e5-9595-447b-b804-feba577498d8" containerName="object-server" Oct 08 18:33:55 crc kubenswrapper[4988]: I1008 18:33:55.636719 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="99d77eed-40f0-4879-be96-abea3a807cd9" containerName="cinder-api-log" Oct 08 18:33:55 crc kubenswrapper[4988]: I1008 18:33:55.636732 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="b6d1becf-4229-40bf-b53a-d8a9eb4b580e" containerName="kube-state-metrics" Oct 08 18:33:55 crc kubenswrapper[4988]: I1008 18:33:55.636747 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="ccd2f8fa-1c16-4c58-9f87-e60ba3a2f5fb" containerName="ceilometer-central-agent" Oct 08 18:33:55 crc kubenswrapper[4988]: I1008 18:33:55.636757 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="d04259e5-9595-447b-b804-feba577498d8" containerName="object-updater" Oct 08 18:33:55 crc kubenswrapper[4988]: I1008 18:33:55.636765 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="32bb7401-6292-4ca6-98f4-15522125c3ee" containerName="memcached" Oct 08 18:33:55 crc kubenswrapper[4988]: I1008 18:33:55.636774 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4e47608-89b4-45d1-88a6-fe75805d864d" containerName="placement-api" Oct 08 18:33:55 crc kubenswrapper[4988]: I1008 18:33:55.636785 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="3a57537b-6115-4249-8d16-3ab0c4b6d21e" containerName="barbican-api-log" Oct 08 18:33:55 crc kubenswrapper[4988]: I1008 18:33:55.636794 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="d04259e5-9595-447b-b804-feba577498d8" containerName="container-updater" Oct 08 18:33:55 crc kubenswrapper[4988]: I1008 18:33:55.636807 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="d04259e5-9595-447b-b804-feba577498d8" containerName="object-replicator" Oct 08 18:33:55 crc kubenswrapper[4988]: I1008 18:33:55.636816 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="d04259e5-9595-447b-b804-feba577498d8" containerName="swift-recon-cron" Oct 08 18:33:55 crc kubenswrapper[4988]: I1008 18:33:55.636825 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="bd9cfcdf-4770-4efa-a41e-7867ec85206f" containerName="nova-cell0-conductor-conductor" Oct 08 18:33:55 crc kubenswrapper[4988]: I1008 18:33:55.636835 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="ccd2f8fa-1c16-4c58-9f87-e60ba3a2f5fb" containerName="sg-core" Oct 08 18:33:55 crc kubenswrapper[4988]: I1008 18:33:55.636843 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="0ec345a3-1744-4e78-bf13-91f0325296e4" containerName="cinder-scheduler" Oct 08 18:33:55 crc kubenswrapper[4988]: I1008 18:33:55.636856 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="85bc7f8f-fdb3-4abb-88d3-ed0c10982c4c" containerName="nova-cell1-conductor-conductor" Oct 08 18:33:55 crc kubenswrapper[4988]: I1008 18:33:55.636867 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="d04259e5-9595-447b-b804-feba577498d8" containerName="container-replicator" Oct 08 18:33:55 crc kubenswrapper[4988]: I1008 18:33:55.636877 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="d04259e5-9595-447b-b804-feba577498d8" containerName="container-auditor" Oct 08 18:33:55 crc kubenswrapper[4988]: I1008 18:33:55.636891 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="f1628013-e0b9-4c31-80c0-91aabaaef0f6" containerName="ovs-vswitchd" Oct 08 18:33:55 crc kubenswrapper[4988]: I1008 18:33:55.636898 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="d04259e5-9595-447b-b804-feba577498d8" containerName="object-auditor" Oct 08 18:33:55 crc kubenswrapper[4988]: I1008 18:33:55.636905 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="c9f172a3-9188-4f5f-bf3a-fbc3b0dc5206" containerName="barbican-keystone-listener-log" Oct 08 18:33:55 crc kubenswrapper[4988]: I1008 18:33:55.636917 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="d04259e5-9595-447b-b804-feba577498d8" containerName="rsync" Oct 08 18:33:55 crc kubenswrapper[4988]: I1008 18:33:55.636930 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="0ec345a3-1744-4e78-bf13-91f0325296e4" containerName="probe" Oct 08 18:33:55 crc kubenswrapper[4988]: I1008 18:33:55.636938 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="e5a8f819-c91d-429e-b848-e9c444739623" containerName="rabbitmq" Oct 08 18:33:55 crc kubenswrapper[4988]: I1008 18:33:55.636948 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4e47608-89b4-45d1-88a6-fe75805d864d" containerName="placement-log" Oct 08 18:33:55 crc kubenswrapper[4988]: I1008 18:33:55.636958 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="0fcc9499-a06b-4cd1-a5da-ea8108916050" containerName="glance-httpd" Oct 08 18:33:55 crc kubenswrapper[4988]: I1008 18:33:55.636969 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="56702a03-175d-4a54-b84f-e25cab037431" containerName="glance-log" Oct 08 18:33:55 crc kubenswrapper[4988]: I1008 18:33:55.636980 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="d04259e5-9595-447b-b804-feba577498d8" containerName="account-replicator" Oct 08 18:33:55 crc kubenswrapper[4988]: I1008 18:33:55.636992 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="3503942b-2825-4006-80e8-8c4610b89997" containerName="galera" Oct 08 18:33:55 crc kubenswrapper[4988]: I1008 18:33:55.636999 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="0fcc9499-a06b-4cd1-a5da-ea8108916050" containerName="glance-log" Oct 08 18:33:55 crc kubenswrapper[4988]: I1008 18:33:55.637013 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="d04259e5-9595-447b-b804-feba577498d8" containerName="account-server" Oct 08 18:33:55 crc kubenswrapper[4988]: I1008 18:33:55.637023 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="d04259e5-9595-447b-b804-feba577498d8" containerName="container-server" Oct 08 18:33:55 crc kubenswrapper[4988]: I1008 18:33:55.637033 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="d04259e5-9595-447b-b804-feba577498d8" containerName="object-expirer" Oct 08 18:33:55 crc kubenswrapper[4988]: I1008 18:33:55.637045 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="d04259e5-9595-447b-b804-feba577498d8" containerName="account-reaper" Oct 08 18:33:55 crc kubenswrapper[4988]: I1008 18:33:55.637056 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="0efd1f53-90ad-4eb8-90f2-2ba9fc1593c2" containerName="nova-api-log" Oct 08 18:33:55 crc kubenswrapper[4988]: I1008 18:33:55.637066 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="0efd1f53-90ad-4eb8-90f2-2ba9fc1593c2" containerName="nova-api-api" Oct 08 18:33:55 crc kubenswrapper[4988]: I1008 18:33:55.637078 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="56702a03-175d-4a54-b84f-e25cab037431" containerName="glance-httpd" Oct 08 18:33:55 crc kubenswrapper[4988]: I1008 18:33:55.637088 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="830d6ec5-0015-4a88-b8e8-0fd47f64fd46" containerName="rabbitmq" Oct 08 18:33:55 crc kubenswrapper[4988]: I1008 18:33:55.637098 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="99d77eed-40f0-4879-be96-abea3a807cd9" containerName="cinder-api" Oct 08 18:33:55 crc kubenswrapper[4988]: I1008 18:33:55.637109 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="3f7bddd3-86f8-46e9-a38e-1aa5e570d151" containerName="mariadb-account-delete" Oct 08 18:33:55 crc kubenswrapper[4988]: I1008 18:33:55.639078 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-xbvnj" Oct 08 18:33:55 crc kubenswrapper[4988]: I1008 18:33:55.644915 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-xbvnj"] Oct 08 18:33:55 crc kubenswrapper[4988]: I1008 18:33:55.758047 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2fc0a957-e540-465a-ae01-39bd7f0489c3-utilities\") pod \"certified-operators-xbvnj\" (UID: \"2fc0a957-e540-465a-ae01-39bd7f0489c3\") " pod="openshift-marketplace/certified-operators-xbvnj" Oct 08 18:33:55 crc kubenswrapper[4988]: I1008 18:33:55.758101 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ljskw\" (UniqueName: \"kubernetes.io/projected/2fc0a957-e540-465a-ae01-39bd7f0489c3-kube-api-access-ljskw\") pod \"certified-operators-xbvnj\" (UID: \"2fc0a957-e540-465a-ae01-39bd7f0489c3\") " pod="openshift-marketplace/certified-operators-xbvnj" Oct 08 18:33:55 crc kubenswrapper[4988]: I1008 18:33:55.758191 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2fc0a957-e540-465a-ae01-39bd7f0489c3-catalog-content\") pod \"certified-operators-xbvnj\" (UID: \"2fc0a957-e540-465a-ae01-39bd7f0489c3\") " pod="openshift-marketplace/certified-operators-xbvnj" Oct 08 18:33:55 crc kubenswrapper[4988]: I1008 18:33:55.860110 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2fc0a957-e540-465a-ae01-39bd7f0489c3-catalog-content\") pod \"certified-operators-xbvnj\" (UID: \"2fc0a957-e540-465a-ae01-39bd7f0489c3\") " pod="openshift-marketplace/certified-operators-xbvnj" Oct 08 18:33:55 crc kubenswrapper[4988]: I1008 18:33:55.860236 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2fc0a957-e540-465a-ae01-39bd7f0489c3-utilities\") pod \"certified-operators-xbvnj\" (UID: \"2fc0a957-e540-465a-ae01-39bd7f0489c3\") " pod="openshift-marketplace/certified-operators-xbvnj" Oct 08 18:33:55 crc kubenswrapper[4988]: I1008 18:33:55.860275 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ljskw\" (UniqueName: \"kubernetes.io/projected/2fc0a957-e540-465a-ae01-39bd7f0489c3-kube-api-access-ljskw\") pod \"certified-operators-xbvnj\" (UID: \"2fc0a957-e540-465a-ae01-39bd7f0489c3\") " pod="openshift-marketplace/certified-operators-xbvnj" Oct 08 18:33:55 crc kubenswrapper[4988]: I1008 18:33:55.860818 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2fc0a957-e540-465a-ae01-39bd7f0489c3-utilities\") pod \"certified-operators-xbvnj\" (UID: \"2fc0a957-e540-465a-ae01-39bd7f0489c3\") " pod="openshift-marketplace/certified-operators-xbvnj" Oct 08 18:33:55 crc kubenswrapper[4988]: I1008 18:33:55.861124 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2fc0a957-e540-465a-ae01-39bd7f0489c3-catalog-content\") pod \"certified-operators-xbvnj\" (UID: \"2fc0a957-e540-465a-ae01-39bd7f0489c3\") " pod="openshift-marketplace/certified-operators-xbvnj" Oct 08 18:33:55 crc kubenswrapper[4988]: I1008 18:33:55.888595 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ljskw\" (UniqueName: \"kubernetes.io/projected/2fc0a957-e540-465a-ae01-39bd7f0489c3-kube-api-access-ljskw\") pod \"certified-operators-xbvnj\" (UID: \"2fc0a957-e540-465a-ae01-39bd7f0489c3\") " pod="openshift-marketplace/certified-operators-xbvnj" Oct 08 18:33:55 crc kubenswrapper[4988]: I1008 18:33:55.966845 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-xbvnj" Oct 08 18:33:56 crc kubenswrapper[4988]: I1008 18:33:56.464728 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-xbvnj"] Oct 08 18:33:57 crc kubenswrapper[4988]: I1008 18:33:57.329055 4988 generic.go:334] "Generic (PLEG): container finished" podID="2fc0a957-e540-465a-ae01-39bd7f0489c3" containerID="3a4d90ccc7d4dc6a98f54453c1a0e36d862fa9f4ebda431f99351411c32091c5" exitCode=0 Oct 08 18:33:57 crc kubenswrapper[4988]: I1008 18:33:57.329130 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xbvnj" event={"ID":"2fc0a957-e540-465a-ae01-39bd7f0489c3","Type":"ContainerDied","Data":"3a4d90ccc7d4dc6a98f54453c1a0e36d862fa9f4ebda431f99351411c32091c5"} Oct 08 18:33:57 crc kubenswrapper[4988]: I1008 18:33:57.329337 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xbvnj" event={"ID":"2fc0a957-e540-465a-ae01-39bd7f0489c3","Type":"ContainerStarted","Data":"9e56cda9ae87610641beb972ec7782e48cea5bba8cdce827ce38bfe4ac144f65"} Oct 08 18:33:57 crc kubenswrapper[4988]: I1008 18:33:57.331786 4988 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 08 18:33:58 crc kubenswrapper[4988]: I1008 18:33:58.340770 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xbvnj" event={"ID":"2fc0a957-e540-465a-ae01-39bd7f0489c3","Type":"ContainerStarted","Data":"868bd5b7e2b8dc63b882f1300d37efc8f0cd6cbaa602a70d1c1b554641e3b8dc"} Oct 08 18:33:59 crc kubenswrapper[4988]: I1008 18:33:59.358662 4988 generic.go:334] "Generic (PLEG): container finished" podID="2fc0a957-e540-465a-ae01-39bd7f0489c3" containerID="868bd5b7e2b8dc63b882f1300d37efc8f0cd6cbaa602a70d1c1b554641e3b8dc" exitCode=0 Oct 08 18:33:59 crc kubenswrapper[4988]: I1008 18:33:59.358712 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xbvnj" event={"ID":"2fc0a957-e540-465a-ae01-39bd7f0489c3","Type":"ContainerDied","Data":"868bd5b7e2b8dc63b882f1300d37efc8f0cd6cbaa602a70d1c1b554641e3b8dc"} Oct 08 18:34:00 crc kubenswrapper[4988]: I1008 18:34:00.370455 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xbvnj" event={"ID":"2fc0a957-e540-465a-ae01-39bd7f0489c3","Type":"ContainerStarted","Data":"75d1e71628dc60a6374b794b5784fb25e2858585761cf426b5b4d57757a322ca"} Oct 08 18:34:00 crc kubenswrapper[4988]: I1008 18:34:00.393094 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-xbvnj" podStartSLOduration=2.863141395 podStartE2EDuration="5.39307666s" podCreationTimestamp="2025-10-08 18:33:55 +0000 UTC" firstStartedPulling="2025-10-08 18:33:57.331309381 +0000 UTC m=+1382.781152201" lastFinishedPulling="2025-10-08 18:33:59.861244686 +0000 UTC m=+1385.311087466" observedRunningTime="2025-10-08 18:34:00.392669197 +0000 UTC m=+1385.842512007" watchObservedRunningTime="2025-10-08 18:34:00.39307666 +0000 UTC m=+1385.842919450" Oct 08 18:34:05 crc kubenswrapper[4988]: I1008 18:34:05.967671 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-xbvnj" Oct 08 18:34:05 crc kubenswrapper[4988]: I1008 18:34:05.968208 4988 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-xbvnj" Oct 08 18:34:06 crc kubenswrapper[4988]: I1008 18:34:06.063789 4988 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-xbvnj" Oct 08 18:34:06 crc kubenswrapper[4988]: I1008 18:34:06.464804 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-xbvnj" Oct 08 18:34:06 crc kubenswrapper[4988]: I1008 18:34:06.504082 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-xbvnj"] Oct 08 18:34:08 crc kubenswrapper[4988]: I1008 18:34:08.446644 4988 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-xbvnj" podUID="2fc0a957-e540-465a-ae01-39bd7f0489c3" containerName="registry-server" containerID="cri-o://75d1e71628dc60a6374b794b5784fb25e2858585761cf426b5b4d57757a322ca" gracePeriod=2 Oct 08 18:34:08 crc kubenswrapper[4988]: I1008 18:34:08.876909 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-xbvnj" Oct 08 18:34:09 crc kubenswrapper[4988]: I1008 18:34:09.071314 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ljskw\" (UniqueName: \"kubernetes.io/projected/2fc0a957-e540-465a-ae01-39bd7f0489c3-kube-api-access-ljskw\") pod \"2fc0a957-e540-465a-ae01-39bd7f0489c3\" (UID: \"2fc0a957-e540-465a-ae01-39bd7f0489c3\") " Oct 08 18:34:09 crc kubenswrapper[4988]: I1008 18:34:09.071416 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2fc0a957-e540-465a-ae01-39bd7f0489c3-utilities\") pod \"2fc0a957-e540-465a-ae01-39bd7f0489c3\" (UID: \"2fc0a957-e540-465a-ae01-39bd7f0489c3\") " Oct 08 18:34:09 crc kubenswrapper[4988]: I1008 18:34:09.071462 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2fc0a957-e540-465a-ae01-39bd7f0489c3-catalog-content\") pod \"2fc0a957-e540-465a-ae01-39bd7f0489c3\" (UID: \"2fc0a957-e540-465a-ae01-39bd7f0489c3\") " Oct 08 18:34:09 crc kubenswrapper[4988]: I1008 18:34:09.072813 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2fc0a957-e540-465a-ae01-39bd7f0489c3-utilities" (OuterVolumeSpecName: "utilities") pod "2fc0a957-e540-465a-ae01-39bd7f0489c3" (UID: "2fc0a957-e540-465a-ae01-39bd7f0489c3"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 18:34:09 crc kubenswrapper[4988]: I1008 18:34:09.078674 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2fc0a957-e540-465a-ae01-39bd7f0489c3-kube-api-access-ljskw" (OuterVolumeSpecName: "kube-api-access-ljskw") pod "2fc0a957-e540-465a-ae01-39bd7f0489c3" (UID: "2fc0a957-e540-465a-ae01-39bd7f0489c3"). InnerVolumeSpecName "kube-api-access-ljskw". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:34:09 crc kubenswrapper[4988]: I1008 18:34:09.172443 4988 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ljskw\" (UniqueName: \"kubernetes.io/projected/2fc0a957-e540-465a-ae01-39bd7f0489c3-kube-api-access-ljskw\") on node \"crc\" DevicePath \"\"" Oct 08 18:34:09 crc kubenswrapper[4988]: I1008 18:34:09.172663 4988 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2fc0a957-e540-465a-ae01-39bd7f0489c3-utilities\") on node \"crc\" DevicePath \"\"" Oct 08 18:34:09 crc kubenswrapper[4988]: I1008 18:34:09.290359 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2fc0a957-e540-465a-ae01-39bd7f0489c3-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "2fc0a957-e540-465a-ae01-39bd7f0489c3" (UID: "2fc0a957-e540-465a-ae01-39bd7f0489c3"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 18:34:09 crc kubenswrapper[4988]: I1008 18:34:09.375249 4988 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2fc0a957-e540-465a-ae01-39bd7f0489c3-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 08 18:34:09 crc kubenswrapper[4988]: I1008 18:34:09.456376 4988 generic.go:334] "Generic (PLEG): container finished" podID="2fc0a957-e540-465a-ae01-39bd7f0489c3" containerID="75d1e71628dc60a6374b794b5784fb25e2858585761cf426b5b4d57757a322ca" exitCode=0 Oct 08 18:34:09 crc kubenswrapper[4988]: I1008 18:34:09.456440 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xbvnj" event={"ID":"2fc0a957-e540-465a-ae01-39bd7f0489c3","Type":"ContainerDied","Data":"75d1e71628dc60a6374b794b5784fb25e2858585761cf426b5b4d57757a322ca"} Oct 08 18:34:09 crc kubenswrapper[4988]: I1008 18:34:09.456476 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-xbvnj" Oct 08 18:34:09 crc kubenswrapper[4988]: I1008 18:34:09.456487 4988 scope.go:117] "RemoveContainer" containerID="75d1e71628dc60a6374b794b5784fb25e2858585761cf426b5b4d57757a322ca" Oct 08 18:34:09 crc kubenswrapper[4988]: I1008 18:34:09.456475 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xbvnj" event={"ID":"2fc0a957-e540-465a-ae01-39bd7f0489c3","Type":"ContainerDied","Data":"9e56cda9ae87610641beb972ec7782e48cea5bba8cdce827ce38bfe4ac144f65"} Oct 08 18:34:09 crc kubenswrapper[4988]: I1008 18:34:09.501637 4988 scope.go:117] "RemoveContainer" containerID="868bd5b7e2b8dc63b882f1300d37efc8f0cd6cbaa602a70d1c1b554641e3b8dc" Oct 08 18:34:09 crc kubenswrapper[4988]: I1008 18:34:09.521152 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-xbvnj"] Oct 08 18:34:09 crc kubenswrapper[4988]: I1008 18:34:09.529637 4988 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-xbvnj"] Oct 08 18:34:09 crc kubenswrapper[4988]: I1008 18:34:09.558288 4988 scope.go:117] "RemoveContainer" containerID="3a4d90ccc7d4dc6a98f54453c1a0e36d862fa9f4ebda431f99351411c32091c5" Oct 08 18:34:09 crc kubenswrapper[4988]: I1008 18:34:09.576771 4988 scope.go:117] "RemoveContainer" containerID="75d1e71628dc60a6374b794b5784fb25e2858585761cf426b5b4d57757a322ca" Oct 08 18:34:09 crc kubenswrapper[4988]: E1008 18:34:09.577590 4988 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"75d1e71628dc60a6374b794b5784fb25e2858585761cf426b5b4d57757a322ca\": container with ID starting with 75d1e71628dc60a6374b794b5784fb25e2858585761cf426b5b4d57757a322ca not found: ID does not exist" containerID="75d1e71628dc60a6374b794b5784fb25e2858585761cf426b5b4d57757a322ca" Oct 08 18:34:09 crc kubenswrapper[4988]: I1008 18:34:09.577648 4988 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"75d1e71628dc60a6374b794b5784fb25e2858585761cf426b5b4d57757a322ca"} err="failed to get container status \"75d1e71628dc60a6374b794b5784fb25e2858585761cf426b5b4d57757a322ca\": rpc error: code = NotFound desc = could not find container \"75d1e71628dc60a6374b794b5784fb25e2858585761cf426b5b4d57757a322ca\": container with ID starting with 75d1e71628dc60a6374b794b5784fb25e2858585761cf426b5b4d57757a322ca not found: ID does not exist" Oct 08 18:34:09 crc kubenswrapper[4988]: I1008 18:34:09.577681 4988 scope.go:117] "RemoveContainer" containerID="868bd5b7e2b8dc63b882f1300d37efc8f0cd6cbaa602a70d1c1b554641e3b8dc" Oct 08 18:34:09 crc kubenswrapper[4988]: E1008 18:34:09.578081 4988 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"868bd5b7e2b8dc63b882f1300d37efc8f0cd6cbaa602a70d1c1b554641e3b8dc\": container with ID starting with 868bd5b7e2b8dc63b882f1300d37efc8f0cd6cbaa602a70d1c1b554641e3b8dc not found: ID does not exist" containerID="868bd5b7e2b8dc63b882f1300d37efc8f0cd6cbaa602a70d1c1b554641e3b8dc" Oct 08 18:34:09 crc kubenswrapper[4988]: I1008 18:34:09.578114 4988 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"868bd5b7e2b8dc63b882f1300d37efc8f0cd6cbaa602a70d1c1b554641e3b8dc"} err="failed to get container status \"868bd5b7e2b8dc63b882f1300d37efc8f0cd6cbaa602a70d1c1b554641e3b8dc\": rpc error: code = NotFound desc = could not find container \"868bd5b7e2b8dc63b882f1300d37efc8f0cd6cbaa602a70d1c1b554641e3b8dc\": container with ID starting with 868bd5b7e2b8dc63b882f1300d37efc8f0cd6cbaa602a70d1c1b554641e3b8dc not found: ID does not exist" Oct 08 18:34:09 crc kubenswrapper[4988]: I1008 18:34:09.578134 4988 scope.go:117] "RemoveContainer" containerID="3a4d90ccc7d4dc6a98f54453c1a0e36d862fa9f4ebda431f99351411c32091c5" Oct 08 18:34:09 crc kubenswrapper[4988]: E1008 18:34:09.578379 4988 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3a4d90ccc7d4dc6a98f54453c1a0e36d862fa9f4ebda431f99351411c32091c5\": container with ID starting with 3a4d90ccc7d4dc6a98f54453c1a0e36d862fa9f4ebda431f99351411c32091c5 not found: ID does not exist" containerID="3a4d90ccc7d4dc6a98f54453c1a0e36d862fa9f4ebda431f99351411c32091c5" Oct 08 18:34:09 crc kubenswrapper[4988]: I1008 18:34:09.578460 4988 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3a4d90ccc7d4dc6a98f54453c1a0e36d862fa9f4ebda431f99351411c32091c5"} err="failed to get container status \"3a4d90ccc7d4dc6a98f54453c1a0e36d862fa9f4ebda431f99351411c32091c5\": rpc error: code = NotFound desc = could not find container \"3a4d90ccc7d4dc6a98f54453c1a0e36d862fa9f4ebda431f99351411c32091c5\": container with ID starting with 3a4d90ccc7d4dc6a98f54453c1a0e36d862fa9f4ebda431f99351411c32091c5 not found: ID does not exist" Oct 08 18:34:11 crc kubenswrapper[4988]: I1008 18:34:11.248994 4988 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2fc0a957-e540-465a-ae01-39bd7f0489c3" path="/var/lib/kubelet/pods/2fc0a957-e540-465a-ae01-39bd7f0489c3/volumes" Oct 08 18:34:42 crc kubenswrapper[4988]: I1008 18:34:42.396960 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-svlvr"] Oct 08 18:34:42 crc kubenswrapper[4988]: E1008 18:34:42.397885 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2fc0a957-e540-465a-ae01-39bd7f0489c3" containerName="registry-server" Oct 08 18:34:42 crc kubenswrapper[4988]: I1008 18:34:42.397903 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="2fc0a957-e540-465a-ae01-39bd7f0489c3" containerName="registry-server" Oct 08 18:34:42 crc kubenswrapper[4988]: E1008 18:34:42.397916 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2fc0a957-e540-465a-ae01-39bd7f0489c3" containerName="extract-utilities" Oct 08 18:34:42 crc kubenswrapper[4988]: I1008 18:34:42.397923 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="2fc0a957-e540-465a-ae01-39bd7f0489c3" containerName="extract-utilities" Oct 08 18:34:42 crc kubenswrapper[4988]: E1008 18:34:42.397944 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2fc0a957-e540-465a-ae01-39bd7f0489c3" containerName="extract-content" Oct 08 18:34:42 crc kubenswrapper[4988]: I1008 18:34:42.397953 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="2fc0a957-e540-465a-ae01-39bd7f0489c3" containerName="extract-content" Oct 08 18:34:42 crc kubenswrapper[4988]: I1008 18:34:42.398126 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="2fc0a957-e540-465a-ae01-39bd7f0489c3" containerName="registry-server" Oct 08 18:34:42 crc kubenswrapper[4988]: I1008 18:34:42.399722 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-svlvr" Oct 08 18:34:42 crc kubenswrapper[4988]: I1008 18:34:42.415132 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-svlvr"] Oct 08 18:34:42 crc kubenswrapper[4988]: I1008 18:34:42.461195 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3b90bb85-7199-431e-9659-82a7642f923b-catalog-content\") pod \"community-operators-svlvr\" (UID: \"3b90bb85-7199-431e-9659-82a7642f923b\") " pod="openshift-marketplace/community-operators-svlvr" Oct 08 18:34:42 crc kubenswrapper[4988]: I1008 18:34:42.461375 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m9nws\" (UniqueName: \"kubernetes.io/projected/3b90bb85-7199-431e-9659-82a7642f923b-kube-api-access-m9nws\") pod \"community-operators-svlvr\" (UID: \"3b90bb85-7199-431e-9659-82a7642f923b\") " pod="openshift-marketplace/community-operators-svlvr" Oct 08 18:34:42 crc kubenswrapper[4988]: I1008 18:34:42.461488 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3b90bb85-7199-431e-9659-82a7642f923b-utilities\") pod \"community-operators-svlvr\" (UID: \"3b90bb85-7199-431e-9659-82a7642f923b\") " pod="openshift-marketplace/community-operators-svlvr" Oct 08 18:34:42 crc kubenswrapper[4988]: I1008 18:34:42.562511 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m9nws\" (UniqueName: \"kubernetes.io/projected/3b90bb85-7199-431e-9659-82a7642f923b-kube-api-access-m9nws\") pod \"community-operators-svlvr\" (UID: \"3b90bb85-7199-431e-9659-82a7642f923b\") " pod="openshift-marketplace/community-operators-svlvr" Oct 08 18:34:42 crc kubenswrapper[4988]: I1008 18:34:42.562581 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3b90bb85-7199-431e-9659-82a7642f923b-utilities\") pod \"community-operators-svlvr\" (UID: \"3b90bb85-7199-431e-9659-82a7642f923b\") " pod="openshift-marketplace/community-operators-svlvr" Oct 08 18:34:42 crc kubenswrapper[4988]: I1008 18:34:42.562631 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3b90bb85-7199-431e-9659-82a7642f923b-catalog-content\") pod \"community-operators-svlvr\" (UID: \"3b90bb85-7199-431e-9659-82a7642f923b\") " pod="openshift-marketplace/community-operators-svlvr" Oct 08 18:34:42 crc kubenswrapper[4988]: I1008 18:34:42.563165 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3b90bb85-7199-431e-9659-82a7642f923b-catalog-content\") pod \"community-operators-svlvr\" (UID: \"3b90bb85-7199-431e-9659-82a7642f923b\") " pod="openshift-marketplace/community-operators-svlvr" Oct 08 18:34:42 crc kubenswrapper[4988]: I1008 18:34:42.563320 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3b90bb85-7199-431e-9659-82a7642f923b-utilities\") pod \"community-operators-svlvr\" (UID: \"3b90bb85-7199-431e-9659-82a7642f923b\") " pod="openshift-marketplace/community-operators-svlvr" Oct 08 18:34:42 crc kubenswrapper[4988]: I1008 18:34:42.585027 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m9nws\" (UniqueName: \"kubernetes.io/projected/3b90bb85-7199-431e-9659-82a7642f923b-kube-api-access-m9nws\") pod \"community-operators-svlvr\" (UID: \"3b90bb85-7199-431e-9659-82a7642f923b\") " pod="openshift-marketplace/community-operators-svlvr" Oct 08 18:34:42 crc kubenswrapper[4988]: I1008 18:34:42.743726 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-svlvr" Oct 08 18:34:43 crc kubenswrapper[4988]: I1008 18:34:43.262418 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-svlvr"] Oct 08 18:34:43 crc kubenswrapper[4988]: I1008 18:34:43.758445 4988 generic.go:334] "Generic (PLEG): container finished" podID="3b90bb85-7199-431e-9659-82a7642f923b" containerID="ee18b819689372d8c516ceb778cb11b21f100c1c5be2487e055dea859b91307f" exitCode=0 Oct 08 18:34:43 crc kubenswrapper[4988]: I1008 18:34:43.758502 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-svlvr" event={"ID":"3b90bb85-7199-431e-9659-82a7642f923b","Type":"ContainerDied","Data":"ee18b819689372d8c516ceb778cb11b21f100c1c5be2487e055dea859b91307f"} Oct 08 18:34:43 crc kubenswrapper[4988]: I1008 18:34:43.758716 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-svlvr" event={"ID":"3b90bb85-7199-431e-9659-82a7642f923b","Type":"ContainerStarted","Data":"ec88bbb3f99667c9a4a76d54f99e9f3099e197c6cade95fa7f2f70094db6d302"} Oct 08 18:34:45 crc kubenswrapper[4988]: I1008 18:34:45.777221 4988 generic.go:334] "Generic (PLEG): container finished" podID="3b90bb85-7199-431e-9659-82a7642f923b" containerID="ee613c1e76c22dfff499d58975a1f55587ed11eba6d5e56d261ff224557c8c00" exitCode=0 Oct 08 18:34:45 crc kubenswrapper[4988]: I1008 18:34:45.777355 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-svlvr" event={"ID":"3b90bb85-7199-431e-9659-82a7642f923b","Type":"ContainerDied","Data":"ee613c1e76c22dfff499d58975a1f55587ed11eba6d5e56d261ff224557c8c00"} Oct 08 18:34:46 crc kubenswrapper[4988]: I1008 18:34:46.792510 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-svlvr" event={"ID":"3b90bb85-7199-431e-9659-82a7642f923b","Type":"ContainerStarted","Data":"1b1efe1540538fa4c1d2013e099e0af9cadbee348f9f5a6f27c64a9f39a895ee"} Oct 08 18:34:52 crc kubenswrapper[4988]: I1008 18:34:52.744834 4988 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-svlvr" Oct 08 18:34:52 crc kubenswrapper[4988]: I1008 18:34:52.745777 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-svlvr" Oct 08 18:34:52 crc kubenswrapper[4988]: I1008 18:34:52.812721 4988 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-svlvr" Oct 08 18:34:52 crc kubenswrapper[4988]: I1008 18:34:52.847592 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-svlvr" podStartSLOduration=8.321197237 podStartE2EDuration="10.84755705s" podCreationTimestamp="2025-10-08 18:34:42 +0000 UTC" firstStartedPulling="2025-10-08 18:34:43.760187564 +0000 UTC m=+1429.210030334" lastFinishedPulling="2025-10-08 18:34:46.286547377 +0000 UTC m=+1431.736390147" observedRunningTime="2025-10-08 18:34:46.827680221 +0000 UTC m=+1432.277523031" watchObservedRunningTime="2025-10-08 18:34:52.84755705 +0000 UTC m=+1438.297399860" Oct 08 18:34:52 crc kubenswrapper[4988]: I1008 18:34:52.921537 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-svlvr" Oct 08 18:34:53 crc kubenswrapper[4988]: I1008 18:34:53.049751 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-svlvr"] Oct 08 18:34:54 crc kubenswrapper[4988]: I1008 18:34:54.880429 4988 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-svlvr" podUID="3b90bb85-7199-431e-9659-82a7642f923b" containerName="registry-server" containerID="cri-o://1b1efe1540538fa4c1d2013e099e0af9cadbee348f9f5a6f27c64a9f39a895ee" gracePeriod=2 Oct 08 18:34:55 crc kubenswrapper[4988]: I1008 18:34:55.302765 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-svlvr" Oct 08 18:34:55 crc kubenswrapper[4988]: I1008 18:34:55.455743 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3b90bb85-7199-431e-9659-82a7642f923b-catalog-content\") pod \"3b90bb85-7199-431e-9659-82a7642f923b\" (UID: \"3b90bb85-7199-431e-9659-82a7642f923b\") " Oct 08 18:34:55 crc kubenswrapper[4988]: I1008 18:34:55.455832 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3b90bb85-7199-431e-9659-82a7642f923b-utilities\") pod \"3b90bb85-7199-431e-9659-82a7642f923b\" (UID: \"3b90bb85-7199-431e-9659-82a7642f923b\") " Oct 08 18:34:55 crc kubenswrapper[4988]: I1008 18:34:55.455904 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m9nws\" (UniqueName: \"kubernetes.io/projected/3b90bb85-7199-431e-9659-82a7642f923b-kube-api-access-m9nws\") pod \"3b90bb85-7199-431e-9659-82a7642f923b\" (UID: \"3b90bb85-7199-431e-9659-82a7642f923b\") " Oct 08 18:34:55 crc kubenswrapper[4988]: I1008 18:34:55.456510 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3b90bb85-7199-431e-9659-82a7642f923b-utilities" (OuterVolumeSpecName: "utilities") pod "3b90bb85-7199-431e-9659-82a7642f923b" (UID: "3b90bb85-7199-431e-9659-82a7642f923b"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 18:34:55 crc kubenswrapper[4988]: I1008 18:34:55.460728 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3b90bb85-7199-431e-9659-82a7642f923b-kube-api-access-m9nws" (OuterVolumeSpecName: "kube-api-access-m9nws") pod "3b90bb85-7199-431e-9659-82a7642f923b" (UID: "3b90bb85-7199-431e-9659-82a7642f923b"). InnerVolumeSpecName "kube-api-access-m9nws". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:34:55 crc kubenswrapper[4988]: I1008 18:34:55.503230 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3b90bb85-7199-431e-9659-82a7642f923b-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "3b90bb85-7199-431e-9659-82a7642f923b" (UID: "3b90bb85-7199-431e-9659-82a7642f923b"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 18:34:55 crc kubenswrapper[4988]: I1008 18:34:55.557121 4988 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3b90bb85-7199-431e-9659-82a7642f923b-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 08 18:34:55 crc kubenswrapper[4988]: I1008 18:34:55.557160 4988 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3b90bb85-7199-431e-9659-82a7642f923b-utilities\") on node \"crc\" DevicePath \"\"" Oct 08 18:34:55 crc kubenswrapper[4988]: I1008 18:34:55.557175 4988 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-m9nws\" (UniqueName: \"kubernetes.io/projected/3b90bb85-7199-431e-9659-82a7642f923b-kube-api-access-m9nws\") on node \"crc\" DevicePath \"\"" Oct 08 18:34:55 crc kubenswrapper[4988]: I1008 18:34:55.894108 4988 generic.go:334] "Generic (PLEG): container finished" podID="3b90bb85-7199-431e-9659-82a7642f923b" containerID="1b1efe1540538fa4c1d2013e099e0af9cadbee348f9f5a6f27c64a9f39a895ee" exitCode=0 Oct 08 18:34:55 crc kubenswrapper[4988]: I1008 18:34:55.894146 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-svlvr" event={"ID":"3b90bb85-7199-431e-9659-82a7642f923b","Type":"ContainerDied","Data":"1b1efe1540538fa4c1d2013e099e0af9cadbee348f9f5a6f27c64a9f39a895ee"} Oct 08 18:34:55 crc kubenswrapper[4988]: I1008 18:34:55.894171 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-svlvr" event={"ID":"3b90bb85-7199-431e-9659-82a7642f923b","Type":"ContainerDied","Data":"ec88bbb3f99667c9a4a76d54f99e9f3099e197c6cade95fa7f2f70094db6d302"} Oct 08 18:34:55 crc kubenswrapper[4988]: I1008 18:34:55.894191 4988 scope.go:117] "RemoveContainer" containerID="1b1efe1540538fa4c1d2013e099e0af9cadbee348f9f5a6f27c64a9f39a895ee" Oct 08 18:34:55 crc kubenswrapper[4988]: I1008 18:34:55.894194 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-svlvr" Oct 08 18:34:55 crc kubenswrapper[4988]: I1008 18:34:55.933973 4988 scope.go:117] "RemoveContainer" containerID="ee613c1e76c22dfff499d58975a1f55587ed11eba6d5e56d261ff224557c8c00" Oct 08 18:34:55 crc kubenswrapper[4988]: I1008 18:34:55.939883 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-svlvr"] Oct 08 18:34:55 crc kubenswrapper[4988]: I1008 18:34:55.951048 4988 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-svlvr"] Oct 08 18:34:55 crc kubenswrapper[4988]: I1008 18:34:55.956257 4988 scope.go:117] "RemoveContainer" containerID="ee18b819689372d8c516ceb778cb11b21f100c1c5be2487e055dea859b91307f" Oct 08 18:34:56 crc kubenswrapper[4988]: I1008 18:34:56.001968 4988 scope.go:117] "RemoveContainer" containerID="1b1efe1540538fa4c1d2013e099e0af9cadbee348f9f5a6f27c64a9f39a895ee" Oct 08 18:34:56 crc kubenswrapper[4988]: E1008 18:34:56.002537 4988 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1b1efe1540538fa4c1d2013e099e0af9cadbee348f9f5a6f27c64a9f39a895ee\": container with ID starting with 1b1efe1540538fa4c1d2013e099e0af9cadbee348f9f5a6f27c64a9f39a895ee not found: ID does not exist" containerID="1b1efe1540538fa4c1d2013e099e0af9cadbee348f9f5a6f27c64a9f39a895ee" Oct 08 18:34:56 crc kubenswrapper[4988]: I1008 18:34:56.002577 4988 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1b1efe1540538fa4c1d2013e099e0af9cadbee348f9f5a6f27c64a9f39a895ee"} err="failed to get container status \"1b1efe1540538fa4c1d2013e099e0af9cadbee348f9f5a6f27c64a9f39a895ee\": rpc error: code = NotFound desc = could not find container \"1b1efe1540538fa4c1d2013e099e0af9cadbee348f9f5a6f27c64a9f39a895ee\": container with ID starting with 1b1efe1540538fa4c1d2013e099e0af9cadbee348f9f5a6f27c64a9f39a895ee not found: ID does not exist" Oct 08 18:34:56 crc kubenswrapper[4988]: I1008 18:34:56.002606 4988 scope.go:117] "RemoveContainer" containerID="ee613c1e76c22dfff499d58975a1f55587ed11eba6d5e56d261ff224557c8c00" Oct 08 18:34:56 crc kubenswrapper[4988]: E1008 18:34:56.003520 4988 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ee613c1e76c22dfff499d58975a1f55587ed11eba6d5e56d261ff224557c8c00\": container with ID starting with ee613c1e76c22dfff499d58975a1f55587ed11eba6d5e56d261ff224557c8c00 not found: ID does not exist" containerID="ee613c1e76c22dfff499d58975a1f55587ed11eba6d5e56d261ff224557c8c00" Oct 08 18:34:56 crc kubenswrapper[4988]: I1008 18:34:56.003760 4988 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ee613c1e76c22dfff499d58975a1f55587ed11eba6d5e56d261ff224557c8c00"} err="failed to get container status \"ee613c1e76c22dfff499d58975a1f55587ed11eba6d5e56d261ff224557c8c00\": rpc error: code = NotFound desc = could not find container \"ee613c1e76c22dfff499d58975a1f55587ed11eba6d5e56d261ff224557c8c00\": container with ID starting with ee613c1e76c22dfff499d58975a1f55587ed11eba6d5e56d261ff224557c8c00 not found: ID does not exist" Oct 08 18:34:56 crc kubenswrapper[4988]: I1008 18:34:56.003848 4988 scope.go:117] "RemoveContainer" containerID="ee18b819689372d8c516ceb778cb11b21f100c1c5be2487e055dea859b91307f" Oct 08 18:34:56 crc kubenswrapper[4988]: E1008 18:34:56.004244 4988 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ee18b819689372d8c516ceb778cb11b21f100c1c5be2487e055dea859b91307f\": container with ID starting with ee18b819689372d8c516ceb778cb11b21f100c1c5be2487e055dea859b91307f not found: ID does not exist" containerID="ee18b819689372d8c516ceb778cb11b21f100c1c5be2487e055dea859b91307f" Oct 08 18:34:56 crc kubenswrapper[4988]: I1008 18:34:56.004287 4988 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ee18b819689372d8c516ceb778cb11b21f100c1c5be2487e055dea859b91307f"} err="failed to get container status \"ee18b819689372d8c516ceb778cb11b21f100c1c5be2487e055dea859b91307f\": rpc error: code = NotFound desc = could not find container \"ee18b819689372d8c516ceb778cb11b21f100c1c5be2487e055dea859b91307f\": container with ID starting with ee18b819689372d8c516ceb778cb11b21f100c1c5be2487e055dea859b91307f not found: ID does not exist" Oct 08 18:34:57 crc kubenswrapper[4988]: I1008 18:34:57.246668 4988 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3b90bb85-7199-431e-9659-82a7642f923b" path="/var/lib/kubelet/pods/3b90bb85-7199-431e-9659-82a7642f923b/volumes" Oct 08 18:34:59 crc kubenswrapper[4988]: I1008 18:34:59.292644 4988 scope.go:117] "RemoveContainer" containerID="b870a785f013fafbe59c8f5824890a66f0b0d43937a44d2f40d45448ece61d72" Oct 08 18:34:59 crc kubenswrapper[4988]: I1008 18:34:59.326716 4988 scope.go:117] "RemoveContainer" containerID="26ce2b8666b3f9444f2693312e4cfc9b72728425b1ed4dff0ae989f452ccc809" Oct 08 18:34:59 crc kubenswrapper[4988]: I1008 18:34:59.352485 4988 scope.go:117] "RemoveContainer" containerID="69d3e16a97ce13679cc0f4e17a4f51c18734213323f046dd107262854a6f6bb8" Oct 08 18:34:59 crc kubenswrapper[4988]: I1008 18:34:59.381098 4988 scope.go:117] "RemoveContainer" containerID="adcb01b7b9d0f727f2f5ff67b48f5ae3862d402862564f8b83cd5f630ad8e178" Oct 08 18:34:59 crc kubenswrapper[4988]: I1008 18:34:59.442502 4988 scope.go:117] "RemoveContainer" containerID="8623084f7aa4582501018d0249b0c75a282e2e977b5c5edd610997795d930f58" Oct 08 18:34:59 crc kubenswrapper[4988]: I1008 18:34:59.464673 4988 scope.go:117] "RemoveContainer" containerID="a57d3af529ac7b94dd865e100104b122dc8c82e5f1730d8857445b107d9a020d" Oct 08 18:34:59 crc kubenswrapper[4988]: I1008 18:34:59.491080 4988 scope.go:117] "RemoveContainer" containerID="0bcd9f7810eb0aa145aba595b3f54d8a03770e0a5e29ab50c4eff8c3b0166474" Oct 08 18:34:59 crc kubenswrapper[4988]: I1008 18:34:59.511849 4988 scope.go:117] "RemoveContainer" containerID="0f84577cfc3de7bfdb97674747df46d7d0dd3b852e078530650e529f2fc757bb" Oct 08 18:34:59 crc kubenswrapper[4988]: I1008 18:34:59.545303 4988 scope.go:117] "RemoveContainer" containerID="632678ba07992badaec60f5c80d6cf410b41634603190807315e2c76b85f8b5c" Oct 08 18:34:59 crc kubenswrapper[4988]: I1008 18:34:59.566377 4988 scope.go:117] "RemoveContainer" containerID="768f2a566b566e4f387ca37429cbe36110b0ea648df5d258af5492e387e65734" Oct 08 18:34:59 crc kubenswrapper[4988]: I1008 18:34:59.594548 4988 scope.go:117] "RemoveContainer" containerID="d8856aa4d121b65fbe8cb4f1653c4695099ac2b3a12881a31fcab24a31fa9319" Oct 08 18:35:37 crc kubenswrapper[4988]: I1008 18:35:37.731433 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-q776s"] Oct 08 18:35:37 crc kubenswrapper[4988]: E1008 18:35:37.732258 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3b90bb85-7199-431e-9659-82a7642f923b" containerName="registry-server" Oct 08 18:35:37 crc kubenswrapper[4988]: I1008 18:35:37.732276 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="3b90bb85-7199-431e-9659-82a7642f923b" containerName="registry-server" Oct 08 18:35:37 crc kubenswrapper[4988]: E1008 18:35:37.732302 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3b90bb85-7199-431e-9659-82a7642f923b" containerName="extract-content" Oct 08 18:35:37 crc kubenswrapper[4988]: I1008 18:35:37.732310 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="3b90bb85-7199-431e-9659-82a7642f923b" containerName="extract-content" Oct 08 18:35:37 crc kubenswrapper[4988]: E1008 18:35:37.732324 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3b90bb85-7199-431e-9659-82a7642f923b" containerName="extract-utilities" Oct 08 18:35:37 crc kubenswrapper[4988]: I1008 18:35:37.732332 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="3b90bb85-7199-431e-9659-82a7642f923b" containerName="extract-utilities" Oct 08 18:35:37 crc kubenswrapper[4988]: I1008 18:35:37.732516 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="3b90bb85-7199-431e-9659-82a7642f923b" containerName="registry-server" Oct 08 18:35:37 crc kubenswrapper[4988]: I1008 18:35:37.733668 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-q776s" Oct 08 18:35:37 crc kubenswrapper[4988]: I1008 18:35:37.745111 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-q776s"] Oct 08 18:35:37 crc kubenswrapper[4988]: I1008 18:35:37.782282 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/77acd7a4-4ca3-4626-9c00-0cfad17e709d-utilities\") pod \"redhat-marketplace-q776s\" (UID: \"77acd7a4-4ca3-4626-9c00-0cfad17e709d\") " pod="openshift-marketplace/redhat-marketplace-q776s" Oct 08 18:35:37 crc kubenswrapper[4988]: I1008 18:35:37.782337 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/77acd7a4-4ca3-4626-9c00-0cfad17e709d-catalog-content\") pod \"redhat-marketplace-q776s\" (UID: \"77acd7a4-4ca3-4626-9c00-0cfad17e709d\") " pod="openshift-marketplace/redhat-marketplace-q776s" Oct 08 18:35:37 crc kubenswrapper[4988]: I1008 18:35:37.782370 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-54tb8\" (UniqueName: \"kubernetes.io/projected/77acd7a4-4ca3-4626-9c00-0cfad17e709d-kube-api-access-54tb8\") pod \"redhat-marketplace-q776s\" (UID: \"77acd7a4-4ca3-4626-9c00-0cfad17e709d\") " pod="openshift-marketplace/redhat-marketplace-q776s" Oct 08 18:35:37 crc kubenswrapper[4988]: I1008 18:35:37.883285 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/77acd7a4-4ca3-4626-9c00-0cfad17e709d-catalog-content\") pod \"redhat-marketplace-q776s\" (UID: \"77acd7a4-4ca3-4626-9c00-0cfad17e709d\") " pod="openshift-marketplace/redhat-marketplace-q776s" Oct 08 18:35:37 crc kubenswrapper[4988]: I1008 18:35:37.883357 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-54tb8\" (UniqueName: \"kubernetes.io/projected/77acd7a4-4ca3-4626-9c00-0cfad17e709d-kube-api-access-54tb8\") pod \"redhat-marketplace-q776s\" (UID: \"77acd7a4-4ca3-4626-9c00-0cfad17e709d\") " pod="openshift-marketplace/redhat-marketplace-q776s" Oct 08 18:35:37 crc kubenswrapper[4988]: I1008 18:35:37.883445 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/77acd7a4-4ca3-4626-9c00-0cfad17e709d-utilities\") pod \"redhat-marketplace-q776s\" (UID: \"77acd7a4-4ca3-4626-9c00-0cfad17e709d\") " pod="openshift-marketplace/redhat-marketplace-q776s" Oct 08 18:35:37 crc kubenswrapper[4988]: I1008 18:35:37.883949 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/77acd7a4-4ca3-4626-9c00-0cfad17e709d-catalog-content\") pod \"redhat-marketplace-q776s\" (UID: \"77acd7a4-4ca3-4626-9c00-0cfad17e709d\") " pod="openshift-marketplace/redhat-marketplace-q776s" Oct 08 18:35:37 crc kubenswrapper[4988]: I1008 18:35:37.883976 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/77acd7a4-4ca3-4626-9c00-0cfad17e709d-utilities\") pod \"redhat-marketplace-q776s\" (UID: \"77acd7a4-4ca3-4626-9c00-0cfad17e709d\") " pod="openshift-marketplace/redhat-marketplace-q776s" Oct 08 18:35:37 crc kubenswrapper[4988]: I1008 18:35:37.905174 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-54tb8\" (UniqueName: \"kubernetes.io/projected/77acd7a4-4ca3-4626-9c00-0cfad17e709d-kube-api-access-54tb8\") pod \"redhat-marketplace-q776s\" (UID: \"77acd7a4-4ca3-4626-9c00-0cfad17e709d\") " pod="openshift-marketplace/redhat-marketplace-q776s" Oct 08 18:35:38 crc kubenswrapper[4988]: I1008 18:35:38.054951 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-q776s" Oct 08 18:35:38 crc kubenswrapper[4988]: I1008 18:35:38.535542 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-q776s"] Oct 08 18:35:38 crc kubenswrapper[4988]: W1008 18:35:38.542160 4988 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod77acd7a4_4ca3_4626_9c00_0cfad17e709d.slice/crio-7c0c307766502bcdd3874977efee594173c09b870716ef7e0f4fce616554c456 WatchSource:0}: Error finding container 7c0c307766502bcdd3874977efee594173c09b870716ef7e0f4fce616554c456: Status 404 returned error can't find the container with id 7c0c307766502bcdd3874977efee594173c09b870716ef7e0f4fce616554c456 Oct 08 18:35:39 crc kubenswrapper[4988]: I1008 18:35:39.288012 4988 generic.go:334] "Generic (PLEG): container finished" podID="77acd7a4-4ca3-4626-9c00-0cfad17e709d" containerID="7697185d0f17dc33f2070358a051d6aaff7e1b139c2b23d81c5904bf7d5fefe9" exitCode=0 Oct 08 18:35:39 crc kubenswrapper[4988]: I1008 18:35:39.288084 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-q776s" event={"ID":"77acd7a4-4ca3-4626-9c00-0cfad17e709d","Type":"ContainerDied","Data":"7697185d0f17dc33f2070358a051d6aaff7e1b139c2b23d81c5904bf7d5fefe9"} Oct 08 18:35:39 crc kubenswrapper[4988]: I1008 18:35:39.288293 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-q776s" event={"ID":"77acd7a4-4ca3-4626-9c00-0cfad17e709d","Type":"ContainerStarted","Data":"7c0c307766502bcdd3874977efee594173c09b870716ef7e0f4fce616554c456"} Oct 08 18:35:40 crc kubenswrapper[4988]: I1008 18:35:40.301417 4988 generic.go:334] "Generic (PLEG): container finished" podID="77acd7a4-4ca3-4626-9c00-0cfad17e709d" containerID="950d3d4ae30536f1b271874d9b7fee07713bb49bda7f24286d9b5cac0007e7d1" exitCode=0 Oct 08 18:35:40 crc kubenswrapper[4988]: I1008 18:35:40.301473 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-q776s" event={"ID":"77acd7a4-4ca3-4626-9c00-0cfad17e709d","Type":"ContainerDied","Data":"950d3d4ae30536f1b271874d9b7fee07713bb49bda7f24286d9b5cac0007e7d1"} Oct 08 18:35:41 crc kubenswrapper[4988]: I1008 18:35:41.313529 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-q776s" event={"ID":"77acd7a4-4ca3-4626-9c00-0cfad17e709d","Type":"ContainerStarted","Data":"e1d1d32e8ae84fc4492072ff29863bdab54301ebaa54b656baadfbb8d31156b6"} Oct 08 18:35:41 crc kubenswrapper[4988]: I1008 18:35:41.334025 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-q776s" podStartSLOduration=2.651490307 podStartE2EDuration="4.333994258s" podCreationTimestamp="2025-10-08 18:35:37 +0000 UTC" firstStartedPulling="2025-10-08 18:35:39.290582453 +0000 UTC m=+1484.740425243" lastFinishedPulling="2025-10-08 18:35:40.973086424 +0000 UTC m=+1486.422929194" observedRunningTime="2025-10-08 18:35:41.331056117 +0000 UTC m=+1486.780898897" watchObservedRunningTime="2025-10-08 18:35:41.333994258 +0000 UTC m=+1486.783837078" Oct 08 18:35:48 crc kubenswrapper[4988]: I1008 18:35:48.056126 4988 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-q776s" Oct 08 18:35:48 crc kubenswrapper[4988]: I1008 18:35:48.057016 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-q776s" Oct 08 18:35:48 crc kubenswrapper[4988]: I1008 18:35:48.158255 4988 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-q776s" Oct 08 18:35:48 crc kubenswrapper[4988]: I1008 18:35:48.415094 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-q776s" Oct 08 18:35:48 crc kubenswrapper[4988]: I1008 18:35:48.455790 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-q776s"] Oct 08 18:35:50 crc kubenswrapper[4988]: I1008 18:35:50.395255 4988 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-q776s" podUID="77acd7a4-4ca3-4626-9c00-0cfad17e709d" containerName="registry-server" containerID="cri-o://e1d1d32e8ae84fc4492072ff29863bdab54301ebaa54b656baadfbb8d31156b6" gracePeriod=2 Oct 08 18:35:50 crc kubenswrapper[4988]: I1008 18:35:50.895723 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-q776s" Oct 08 18:35:51 crc kubenswrapper[4988]: I1008 18:35:51.072147 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/77acd7a4-4ca3-4626-9c00-0cfad17e709d-catalog-content\") pod \"77acd7a4-4ca3-4626-9c00-0cfad17e709d\" (UID: \"77acd7a4-4ca3-4626-9c00-0cfad17e709d\") " Oct 08 18:35:51 crc kubenswrapper[4988]: I1008 18:35:51.072557 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/77acd7a4-4ca3-4626-9c00-0cfad17e709d-utilities\") pod \"77acd7a4-4ca3-4626-9c00-0cfad17e709d\" (UID: \"77acd7a4-4ca3-4626-9c00-0cfad17e709d\") " Oct 08 18:35:51 crc kubenswrapper[4988]: I1008 18:35:51.072589 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-54tb8\" (UniqueName: \"kubernetes.io/projected/77acd7a4-4ca3-4626-9c00-0cfad17e709d-kube-api-access-54tb8\") pod \"77acd7a4-4ca3-4626-9c00-0cfad17e709d\" (UID: \"77acd7a4-4ca3-4626-9c00-0cfad17e709d\") " Oct 08 18:35:51 crc kubenswrapper[4988]: I1008 18:35:51.073419 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/77acd7a4-4ca3-4626-9c00-0cfad17e709d-utilities" (OuterVolumeSpecName: "utilities") pod "77acd7a4-4ca3-4626-9c00-0cfad17e709d" (UID: "77acd7a4-4ca3-4626-9c00-0cfad17e709d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 18:35:51 crc kubenswrapper[4988]: I1008 18:35:51.078477 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/77acd7a4-4ca3-4626-9c00-0cfad17e709d-kube-api-access-54tb8" (OuterVolumeSpecName: "kube-api-access-54tb8") pod "77acd7a4-4ca3-4626-9c00-0cfad17e709d" (UID: "77acd7a4-4ca3-4626-9c00-0cfad17e709d"). InnerVolumeSpecName "kube-api-access-54tb8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:35:51 crc kubenswrapper[4988]: I1008 18:35:51.091852 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/77acd7a4-4ca3-4626-9c00-0cfad17e709d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "77acd7a4-4ca3-4626-9c00-0cfad17e709d" (UID: "77acd7a4-4ca3-4626-9c00-0cfad17e709d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 18:35:51 crc kubenswrapper[4988]: I1008 18:35:51.174450 4988 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/77acd7a4-4ca3-4626-9c00-0cfad17e709d-utilities\") on node \"crc\" DevicePath \"\"" Oct 08 18:35:51 crc kubenswrapper[4988]: I1008 18:35:51.174483 4988 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-54tb8\" (UniqueName: \"kubernetes.io/projected/77acd7a4-4ca3-4626-9c00-0cfad17e709d-kube-api-access-54tb8\") on node \"crc\" DevicePath \"\"" Oct 08 18:35:51 crc kubenswrapper[4988]: I1008 18:35:51.174492 4988 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/77acd7a4-4ca3-4626-9c00-0cfad17e709d-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 08 18:35:51 crc kubenswrapper[4988]: I1008 18:35:51.411487 4988 generic.go:334] "Generic (PLEG): container finished" podID="77acd7a4-4ca3-4626-9c00-0cfad17e709d" containerID="e1d1d32e8ae84fc4492072ff29863bdab54301ebaa54b656baadfbb8d31156b6" exitCode=0 Oct 08 18:35:51 crc kubenswrapper[4988]: I1008 18:35:51.411545 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-q776s" event={"ID":"77acd7a4-4ca3-4626-9c00-0cfad17e709d","Type":"ContainerDied","Data":"e1d1d32e8ae84fc4492072ff29863bdab54301ebaa54b656baadfbb8d31156b6"} Oct 08 18:35:51 crc kubenswrapper[4988]: I1008 18:35:51.411584 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-q776s" event={"ID":"77acd7a4-4ca3-4626-9c00-0cfad17e709d","Type":"ContainerDied","Data":"7c0c307766502bcdd3874977efee594173c09b870716ef7e0f4fce616554c456"} Oct 08 18:35:51 crc kubenswrapper[4988]: I1008 18:35:51.411591 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-q776s" Oct 08 18:35:51 crc kubenswrapper[4988]: I1008 18:35:51.411612 4988 scope.go:117] "RemoveContainer" containerID="e1d1d32e8ae84fc4492072ff29863bdab54301ebaa54b656baadfbb8d31156b6" Oct 08 18:35:51 crc kubenswrapper[4988]: I1008 18:35:51.469640 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-q776s"] Oct 08 18:35:51 crc kubenswrapper[4988]: I1008 18:35:51.474821 4988 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-q776s"] Oct 08 18:35:51 crc kubenswrapper[4988]: I1008 18:35:51.485915 4988 scope.go:117] "RemoveContainer" containerID="950d3d4ae30536f1b271874d9b7fee07713bb49bda7f24286d9b5cac0007e7d1" Oct 08 18:35:51 crc kubenswrapper[4988]: I1008 18:35:51.517703 4988 scope.go:117] "RemoveContainer" containerID="7697185d0f17dc33f2070358a051d6aaff7e1b139c2b23d81c5904bf7d5fefe9" Oct 08 18:35:51 crc kubenswrapper[4988]: I1008 18:35:51.540111 4988 scope.go:117] "RemoveContainer" containerID="e1d1d32e8ae84fc4492072ff29863bdab54301ebaa54b656baadfbb8d31156b6" Oct 08 18:35:51 crc kubenswrapper[4988]: E1008 18:35:51.540646 4988 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e1d1d32e8ae84fc4492072ff29863bdab54301ebaa54b656baadfbb8d31156b6\": container with ID starting with e1d1d32e8ae84fc4492072ff29863bdab54301ebaa54b656baadfbb8d31156b6 not found: ID does not exist" containerID="e1d1d32e8ae84fc4492072ff29863bdab54301ebaa54b656baadfbb8d31156b6" Oct 08 18:35:51 crc kubenswrapper[4988]: I1008 18:35:51.540716 4988 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e1d1d32e8ae84fc4492072ff29863bdab54301ebaa54b656baadfbb8d31156b6"} err="failed to get container status \"e1d1d32e8ae84fc4492072ff29863bdab54301ebaa54b656baadfbb8d31156b6\": rpc error: code = NotFound desc = could not find container \"e1d1d32e8ae84fc4492072ff29863bdab54301ebaa54b656baadfbb8d31156b6\": container with ID starting with e1d1d32e8ae84fc4492072ff29863bdab54301ebaa54b656baadfbb8d31156b6 not found: ID does not exist" Oct 08 18:35:51 crc kubenswrapper[4988]: I1008 18:35:51.540761 4988 scope.go:117] "RemoveContainer" containerID="950d3d4ae30536f1b271874d9b7fee07713bb49bda7f24286d9b5cac0007e7d1" Oct 08 18:35:51 crc kubenswrapper[4988]: E1008 18:35:51.541133 4988 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"950d3d4ae30536f1b271874d9b7fee07713bb49bda7f24286d9b5cac0007e7d1\": container with ID starting with 950d3d4ae30536f1b271874d9b7fee07713bb49bda7f24286d9b5cac0007e7d1 not found: ID does not exist" containerID="950d3d4ae30536f1b271874d9b7fee07713bb49bda7f24286d9b5cac0007e7d1" Oct 08 18:35:51 crc kubenswrapper[4988]: I1008 18:35:51.541179 4988 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"950d3d4ae30536f1b271874d9b7fee07713bb49bda7f24286d9b5cac0007e7d1"} err="failed to get container status \"950d3d4ae30536f1b271874d9b7fee07713bb49bda7f24286d9b5cac0007e7d1\": rpc error: code = NotFound desc = could not find container \"950d3d4ae30536f1b271874d9b7fee07713bb49bda7f24286d9b5cac0007e7d1\": container with ID starting with 950d3d4ae30536f1b271874d9b7fee07713bb49bda7f24286d9b5cac0007e7d1 not found: ID does not exist" Oct 08 18:35:51 crc kubenswrapper[4988]: I1008 18:35:51.541205 4988 scope.go:117] "RemoveContainer" containerID="7697185d0f17dc33f2070358a051d6aaff7e1b139c2b23d81c5904bf7d5fefe9" Oct 08 18:35:51 crc kubenswrapper[4988]: E1008 18:35:51.541521 4988 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7697185d0f17dc33f2070358a051d6aaff7e1b139c2b23d81c5904bf7d5fefe9\": container with ID starting with 7697185d0f17dc33f2070358a051d6aaff7e1b139c2b23d81c5904bf7d5fefe9 not found: ID does not exist" containerID="7697185d0f17dc33f2070358a051d6aaff7e1b139c2b23d81c5904bf7d5fefe9" Oct 08 18:35:51 crc kubenswrapper[4988]: I1008 18:35:51.541558 4988 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7697185d0f17dc33f2070358a051d6aaff7e1b139c2b23d81c5904bf7d5fefe9"} err="failed to get container status \"7697185d0f17dc33f2070358a051d6aaff7e1b139c2b23d81c5904bf7d5fefe9\": rpc error: code = NotFound desc = could not find container \"7697185d0f17dc33f2070358a051d6aaff7e1b139c2b23d81c5904bf7d5fefe9\": container with ID starting with 7697185d0f17dc33f2070358a051d6aaff7e1b139c2b23d81c5904bf7d5fefe9 not found: ID does not exist" Oct 08 18:35:53 crc kubenswrapper[4988]: I1008 18:35:53.253246 4988 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="77acd7a4-4ca3-4626-9c00-0cfad17e709d" path="/var/lib/kubelet/pods/77acd7a4-4ca3-4626-9c00-0cfad17e709d/volumes" Oct 08 18:35:53 crc kubenswrapper[4988]: I1008 18:35:53.338780 4988 patch_prober.go:28] interesting pod/machine-config-daemon-nm54f container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 18:35:53 crc kubenswrapper[4988]: I1008 18:35:53.338878 4988 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 18:35:59 crc kubenswrapper[4988]: I1008 18:35:59.851056 4988 scope.go:117] "RemoveContainer" containerID="46ab4b0ce428b8f21cfe1456c1a0cac36aef5111febc6ab77c67ce8016a55c70" Oct 08 18:35:59 crc kubenswrapper[4988]: I1008 18:35:59.903597 4988 scope.go:117] "RemoveContainer" containerID="c6a4733dbf41fe0410a80b6de0ff01b8123be53bae21bb88ca041b937dd29455" Oct 08 18:35:59 crc kubenswrapper[4988]: I1008 18:35:59.937270 4988 scope.go:117] "RemoveContainer" containerID="ee38f29c71febdb6cdc5a350ab90d5ad4cfc0fbe576a87a90cc2de98e24cac3d" Oct 08 18:35:59 crc kubenswrapper[4988]: I1008 18:35:59.962330 4988 scope.go:117] "RemoveContainer" containerID="e0c58c7d3f2cbaa16eb7f1b22d366459ec6d6b096a0336a9554a9d97a33841f1" Oct 08 18:36:00 crc kubenswrapper[4988]: I1008 18:36:00.015776 4988 scope.go:117] "RemoveContainer" containerID="8418ddc980b5b490b168e615b996b3b3baa5d93df3d598cef9398a441f9533d0" Oct 08 18:36:00 crc kubenswrapper[4988]: I1008 18:36:00.053415 4988 scope.go:117] "RemoveContainer" containerID="f82c2c46fdece62aac8a106f15e5fddfb69696f72a1f951b75f376ae4f50e07b" Oct 08 18:36:00 crc kubenswrapper[4988]: I1008 18:36:00.103588 4988 scope.go:117] "RemoveContainer" containerID="f3dcb22d4b89fb2ad16ba29f989be77152f28df772bd76ecfc8b14528c658bd2" Oct 08 18:36:00 crc kubenswrapper[4988]: I1008 18:36:00.122719 4988 scope.go:117] "RemoveContainer" containerID="57dcd881b0a41e0c7d3247b6644d30ca7dd361b249e7b89b3a8cd1fe5df4b6c9" Oct 08 18:36:00 crc kubenswrapper[4988]: I1008 18:36:00.160907 4988 scope.go:117] "RemoveContainer" containerID="8ef71a96a488da8760db257dae569210b2903674028f06127874742bfeb624fd" Oct 08 18:36:00 crc kubenswrapper[4988]: I1008 18:36:00.192451 4988 scope.go:117] "RemoveContainer" containerID="e634b0ebf42763a7900bb4a166f36326fba120e119d166d4e1c0843990c5daf2" Oct 08 18:36:00 crc kubenswrapper[4988]: I1008 18:36:00.225603 4988 scope.go:117] "RemoveContainer" containerID="0f23a2df4066a02929dd3edb8dc8d068809211dbfad9bf36b14e3f21916284e7" Oct 08 18:36:00 crc kubenswrapper[4988]: I1008 18:36:00.247208 4988 scope.go:117] "RemoveContainer" containerID="75ad7b164d49b05e8fa5d093445ac8c6b637d6459bbbacb342ee57317efae57a" Oct 08 18:36:23 crc kubenswrapper[4988]: I1008 18:36:23.338787 4988 patch_prober.go:28] interesting pod/machine-config-daemon-nm54f container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 18:36:23 crc kubenswrapper[4988]: I1008 18:36:23.339347 4988 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 18:36:53 crc kubenswrapper[4988]: I1008 18:36:53.337646 4988 patch_prober.go:28] interesting pod/machine-config-daemon-nm54f container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 18:36:53 crc kubenswrapper[4988]: I1008 18:36:53.338575 4988 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 18:36:53 crc kubenswrapper[4988]: I1008 18:36:53.338736 4988 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" Oct 08 18:36:53 crc kubenswrapper[4988]: I1008 18:36:53.339887 4988 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"1e3c2ea55d3c49c9fa802ab01ed9324a6c941b8da678ffaccb80377e3f159aee"} pod="openshift-machine-config-operator/machine-config-daemon-nm54f" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 08 18:36:53 crc kubenswrapper[4988]: I1008 18:36:53.340169 4988 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" containerName="machine-config-daemon" containerID="cri-o://1e3c2ea55d3c49c9fa802ab01ed9324a6c941b8da678ffaccb80377e3f159aee" gracePeriod=600 Oct 08 18:36:53 crc kubenswrapper[4988]: E1008 18:36:53.471736 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nm54f_openshift-machine-config-operator(f328a20b-5283-407e-8a93-2ab2da599f18)\"" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" Oct 08 18:36:53 crc kubenswrapper[4988]: I1008 18:36:53.953881 4988 generic.go:334] "Generic (PLEG): container finished" podID="f328a20b-5283-407e-8a93-2ab2da599f18" containerID="1e3c2ea55d3c49c9fa802ab01ed9324a6c941b8da678ffaccb80377e3f159aee" exitCode=0 Oct 08 18:36:53 crc kubenswrapper[4988]: I1008 18:36:53.954000 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" event={"ID":"f328a20b-5283-407e-8a93-2ab2da599f18","Type":"ContainerDied","Data":"1e3c2ea55d3c49c9fa802ab01ed9324a6c941b8da678ffaccb80377e3f159aee"} Oct 08 18:36:53 crc kubenswrapper[4988]: I1008 18:36:53.954063 4988 scope.go:117] "RemoveContainer" containerID="e0a9bc6112323fbe5f5506d9ba72b002a9b9faef1191582e4a39898dc43fcbbb" Oct 08 18:36:53 crc kubenswrapper[4988]: I1008 18:36:53.955627 4988 scope.go:117] "RemoveContainer" containerID="1e3c2ea55d3c49c9fa802ab01ed9324a6c941b8da678ffaccb80377e3f159aee" Oct 08 18:36:53 crc kubenswrapper[4988]: E1008 18:36:53.957796 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nm54f_openshift-machine-config-operator(f328a20b-5283-407e-8a93-2ab2da599f18)\"" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" Oct 08 18:37:00 crc kubenswrapper[4988]: I1008 18:37:00.450253 4988 scope.go:117] "RemoveContainer" containerID="5c2684680583eebae32531e937601c8903ed775d6ac916679e0120825b5358f8" Oct 08 18:37:00 crc kubenswrapper[4988]: I1008 18:37:00.480061 4988 scope.go:117] "RemoveContainer" containerID="415de07802af7873cf01a1c532b1c1de922fc28c756bb42991542444c0b9ed64" Oct 08 18:37:00 crc kubenswrapper[4988]: I1008 18:37:00.505962 4988 scope.go:117] "RemoveContainer" containerID="d8e23c92099ddad91fb908ab60379b8040c04ad905cb12b8b115b0fe8da69e06" Oct 08 18:37:00 crc kubenswrapper[4988]: I1008 18:37:00.528765 4988 scope.go:117] "RemoveContainer" containerID="b532a4185e833359abeaceb113ab2b9b1e3c1758139d29685926ea848c821a01" Oct 08 18:37:00 crc kubenswrapper[4988]: I1008 18:37:00.557549 4988 scope.go:117] "RemoveContainer" containerID="22daa187f81f4dead5f47bc9340c19470b641fcc7b4541232f44df8458409f63" Oct 08 18:37:00 crc kubenswrapper[4988]: I1008 18:37:00.583552 4988 scope.go:117] "RemoveContainer" containerID="3bb3e6305405075980f0e6b13cc2d64214327a9883596c00b4d0ce337b183212" Oct 08 18:37:00 crc kubenswrapper[4988]: I1008 18:37:00.625870 4988 scope.go:117] "RemoveContainer" containerID="ffe051d7124cf2f0d968449026dd9267afba1ef3e9305140464f3494f98986b8" Oct 08 18:37:00 crc kubenswrapper[4988]: I1008 18:37:00.672188 4988 scope.go:117] "RemoveContainer" containerID="bd8657ab54dc617e4190cb1a864b28e10b3355e63b38eb13c2d1d5db69eb55ac" Oct 08 18:37:09 crc kubenswrapper[4988]: I1008 18:37:09.237719 4988 scope.go:117] "RemoveContainer" containerID="1e3c2ea55d3c49c9fa802ab01ed9324a6c941b8da678ffaccb80377e3f159aee" Oct 08 18:37:09 crc kubenswrapper[4988]: E1008 18:37:09.238644 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nm54f_openshift-machine-config-operator(f328a20b-5283-407e-8a93-2ab2da599f18)\"" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" Oct 08 18:37:20 crc kubenswrapper[4988]: I1008 18:37:20.237848 4988 scope.go:117] "RemoveContainer" containerID="1e3c2ea55d3c49c9fa802ab01ed9324a6c941b8da678ffaccb80377e3f159aee" Oct 08 18:37:20 crc kubenswrapper[4988]: E1008 18:37:20.238956 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nm54f_openshift-machine-config-operator(f328a20b-5283-407e-8a93-2ab2da599f18)\"" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" Oct 08 18:37:34 crc kubenswrapper[4988]: I1008 18:37:34.237614 4988 scope.go:117] "RemoveContainer" containerID="1e3c2ea55d3c49c9fa802ab01ed9324a6c941b8da678ffaccb80377e3f159aee" Oct 08 18:37:34 crc kubenswrapper[4988]: E1008 18:37:34.238988 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nm54f_openshift-machine-config-operator(f328a20b-5283-407e-8a93-2ab2da599f18)\"" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" Oct 08 18:37:47 crc kubenswrapper[4988]: I1008 18:37:47.239780 4988 scope.go:117] "RemoveContainer" containerID="1e3c2ea55d3c49c9fa802ab01ed9324a6c941b8da678ffaccb80377e3f159aee" Oct 08 18:37:47 crc kubenswrapper[4988]: E1008 18:37:47.240879 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nm54f_openshift-machine-config-operator(f328a20b-5283-407e-8a93-2ab2da599f18)\"" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" Oct 08 18:38:00 crc kubenswrapper[4988]: I1008 18:38:00.238290 4988 scope.go:117] "RemoveContainer" containerID="1e3c2ea55d3c49c9fa802ab01ed9324a6c941b8da678ffaccb80377e3f159aee" Oct 08 18:38:00 crc kubenswrapper[4988]: E1008 18:38:00.239001 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nm54f_openshift-machine-config-operator(f328a20b-5283-407e-8a93-2ab2da599f18)\"" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" Oct 08 18:38:00 crc kubenswrapper[4988]: I1008 18:38:00.788348 4988 scope.go:117] "RemoveContainer" containerID="f1e67556c910c6012c3493e80adf12b19a788da6c247b9ef602743fd99cb3b7a" Oct 08 18:38:00 crc kubenswrapper[4988]: I1008 18:38:00.820903 4988 scope.go:117] "RemoveContainer" containerID="8187466829ece57145a6f580c49fd29ced64f87be389192fb463c718a5acae13" Oct 08 18:38:00 crc kubenswrapper[4988]: I1008 18:38:00.867116 4988 scope.go:117] "RemoveContainer" containerID="8b312e4e6552af848a89843736fde9707853545f97ff16dd7eb3c47b05d0c32d" Oct 08 18:38:00 crc kubenswrapper[4988]: I1008 18:38:00.889366 4988 scope.go:117] "RemoveContainer" containerID="170776f5aecd85fd95ec5552397ec9b090c5fec990524b633f8d6b1d51a1c74e" Oct 08 18:38:11 crc kubenswrapper[4988]: I1008 18:38:11.238766 4988 scope.go:117] "RemoveContainer" containerID="1e3c2ea55d3c49c9fa802ab01ed9324a6c941b8da678ffaccb80377e3f159aee" Oct 08 18:38:11 crc kubenswrapper[4988]: E1008 18:38:11.239858 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nm54f_openshift-machine-config-operator(f328a20b-5283-407e-8a93-2ab2da599f18)\"" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" Oct 08 18:38:26 crc kubenswrapper[4988]: I1008 18:38:26.238110 4988 scope.go:117] "RemoveContainer" containerID="1e3c2ea55d3c49c9fa802ab01ed9324a6c941b8da678ffaccb80377e3f159aee" Oct 08 18:38:26 crc kubenswrapper[4988]: E1008 18:38:26.239248 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nm54f_openshift-machine-config-operator(f328a20b-5283-407e-8a93-2ab2da599f18)\"" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" Oct 08 18:38:40 crc kubenswrapper[4988]: I1008 18:38:40.237949 4988 scope.go:117] "RemoveContainer" containerID="1e3c2ea55d3c49c9fa802ab01ed9324a6c941b8da678ffaccb80377e3f159aee" Oct 08 18:38:40 crc kubenswrapper[4988]: E1008 18:38:40.238616 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nm54f_openshift-machine-config-operator(f328a20b-5283-407e-8a93-2ab2da599f18)\"" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" Oct 08 18:38:53 crc kubenswrapper[4988]: I1008 18:38:53.238440 4988 scope.go:117] "RemoveContainer" containerID="1e3c2ea55d3c49c9fa802ab01ed9324a6c941b8da678ffaccb80377e3f159aee" Oct 08 18:38:53 crc kubenswrapper[4988]: E1008 18:38:53.239121 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nm54f_openshift-machine-config-operator(f328a20b-5283-407e-8a93-2ab2da599f18)\"" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" Oct 08 18:39:00 crc kubenswrapper[4988]: I1008 18:39:00.977616 4988 scope.go:117] "RemoveContainer" containerID="31548197acc0e143742e4fb4cb9299994c2c1dd05e374425c882dac956b9d649" Oct 08 18:39:01 crc kubenswrapper[4988]: I1008 18:39:01.007492 4988 scope.go:117] "RemoveContainer" containerID="59a272e450f11633e4f5704d41f6545802b926f41dbb29197c7d02ab91df05c0" Oct 08 18:39:01 crc kubenswrapper[4988]: I1008 18:39:01.031063 4988 scope.go:117] "RemoveContainer" containerID="ee5151c0c831844f2ab3796464c7e52d85522046bd44d8ed010f49dda7a02db4" Oct 08 18:39:01 crc kubenswrapper[4988]: I1008 18:39:01.052333 4988 scope.go:117] "RemoveContainer" containerID="fb6cef822415bff20531090cd5e76b8c5dae15da1099c2400c9c9b26e238da3c" Oct 08 18:39:01 crc kubenswrapper[4988]: I1008 18:39:01.073864 4988 scope.go:117] "RemoveContainer" containerID="df5e696bea34bc9d704df11826fdc54a7f54a0eaa51859b450ac2001e2f15dda" Oct 08 18:39:01 crc kubenswrapper[4988]: I1008 18:39:01.095998 4988 scope.go:117] "RemoveContainer" containerID="85a5febf4bb240fccaef0fc62af632ef9fb88c921ce282445ce2e5cfef1289b0" Oct 08 18:39:06 crc kubenswrapper[4988]: I1008 18:39:06.237194 4988 scope.go:117] "RemoveContainer" containerID="1e3c2ea55d3c49c9fa802ab01ed9324a6c941b8da678ffaccb80377e3f159aee" Oct 08 18:39:06 crc kubenswrapper[4988]: E1008 18:39:06.237731 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nm54f_openshift-machine-config-operator(f328a20b-5283-407e-8a93-2ab2da599f18)\"" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" Oct 08 18:39:18 crc kubenswrapper[4988]: I1008 18:39:18.237770 4988 scope.go:117] "RemoveContainer" containerID="1e3c2ea55d3c49c9fa802ab01ed9324a6c941b8da678ffaccb80377e3f159aee" Oct 08 18:39:18 crc kubenswrapper[4988]: E1008 18:39:18.238670 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nm54f_openshift-machine-config-operator(f328a20b-5283-407e-8a93-2ab2da599f18)\"" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" Oct 08 18:39:31 crc kubenswrapper[4988]: I1008 18:39:31.237508 4988 scope.go:117] "RemoveContainer" containerID="1e3c2ea55d3c49c9fa802ab01ed9324a6c941b8da678ffaccb80377e3f159aee" Oct 08 18:39:31 crc kubenswrapper[4988]: E1008 18:39:31.238662 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nm54f_openshift-machine-config-operator(f328a20b-5283-407e-8a93-2ab2da599f18)\"" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" Oct 08 18:39:46 crc kubenswrapper[4988]: I1008 18:39:46.237252 4988 scope.go:117] "RemoveContainer" containerID="1e3c2ea55d3c49c9fa802ab01ed9324a6c941b8da678ffaccb80377e3f159aee" Oct 08 18:39:46 crc kubenswrapper[4988]: E1008 18:39:46.237958 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nm54f_openshift-machine-config-operator(f328a20b-5283-407e-8a93-2ab2da599f18)\"" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" Oct 08 18:40:00 crc kubenswrapper[4988]: I1008 18:40:00.237987 4988 scope.go:117] "RemoveContainer" containerID="1e3c2ea55d3c49c9fa802ab01ed9324a6c941b8da678ffaccb80377e3f159aee" Oct 08 18:40:00 crc kubenswrapper[4988]: E1008 18:40:00.239224 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nm54f_openshift-machine-config-operator(f328a20b-5283-407e-8a93-2ab2da599f18)\"" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" Oct 08 18:40:12 crc kubenswrapper[4988]: I1008 18:40:12.238811 4988 scope.go:117] "RemoveContainer" containerID="1e3c2ea55d3c49c9fa802ab01ed9324a6c941b8da678ffaccb80377e3f159aee" Oct 08 18:40:12 crc kubenswrapper[4988]: E1008 18:40:12.240185 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nm54f_openshift-machine-config-operator(f328a20b-5283-407e-8a93-2ab2da599f18)\"" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" Oct 08 18:40:25 crc kubenswrapper[4988]: I1008 18:40:25.241873 4988 scope.go:117] "RemoveContainer" containerID="1e3c2ea55d3c49c9fa802ab01ed9324a6c941b8da678ffaccb80377e3f159aee" Oct 08 18:40:25 crc kubenswrapper[4988]: E1008 18:40:25.242818 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nm54f_openshift-machine-config-operator(f328a20b-5283-407e-8a93-2ab2da599f18)\"" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" Oct 08 18:40:38 crc kubenswrapper[4988]: I1008 18:40:38.238174 4988 scope.go:117] "RemoveContainer" containerID="1e3c2ea55d3c49c9fa802ab01ed9324a6c941b8da678ffaccb80377e3f159aee" Oct 08 18:40:38 crc kubenswrapper[4988]: E1008 18:40:38.239131 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nm54f_openshift-machine-config-operator(f328a20b-5283-407e-8a93-2ab2da599f18)\"" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" Oct 08 18:40:49 crc kubenswrapper[4988]: I1008 18:40:49.238323 4988 scope.go:117] "RemoveContainer" containerID="1e3c2ea55d3c49c9fa802ab01ed9324a6c941b8da678ffaccb80377e3f159aee" Oct 08 18:40:49 crc kubenswrapper[4988]: E1008 18:40:49.239580 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nm54f_openshift-machine-config-operator(f328a20b-5283-407e-8a93-2ab2da599f18)\"" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" Oct 08 18:41:01 crc kubenswrapper[4988]: I1008 18:41:01.237638 4988 scope.go:117] "RemoveContainer" containerID="1e3c2ea55d3c49c9fa802ab01ed9324a6c941b8da678ffaccb80377e3f159aee" Oct 08 18:41:01 crc kubenswrapper[4988]: E1008 18:41:01.238547 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nm54f_openshift-machine-config-operator(f328a20b-5283-407e-8a93-2ab2da599f18)\"" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" Oct 08 18:41:01 crc kubenswrapper[4988]: I1008 18:41:01.897673 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-vxx5r"] Oct 08 18:41:01 crc kubenswrapper[4988]: E1008 18:41:01.898260 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="77acd7a4-4ca3-4626-9c00-0cfad17e709d" containerName="extract-content" Oct 08 18:41:01 crc kubenswrapper[4988]: I1008 18:41:01.898293 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="77acd7a4-4ca3-4626-9c00-0cfad17e709d" containerName="extract-content" Oct 08 18:41:01 crc kubenswrapper[4988]: E1008 18:41:01.898339 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="77acd7a4-4ca3-4626-9c00-0cfad17e709d" containerName="extract-utilities" Oct 08 18:41:01 crc kubenswrapper[4988]: I1008 18:41:01.898353 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="77acd7a4-4ca3-4626-9c00-0cfad17e709d" containerName="extract-utilities" Oct 08 18:41:01 crc kubenswrapper[4988]: E1008 18:41:01.898377 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="77acd7a4-4ca3-4626-9c00-0cfad17e709d" containerName="registry-server" Oct 08 18:41:01 crc kubenswrapper[4988]: I1008 18:41:01.898528 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="77acd7a4-4ca3-4626-9c00-0cfad17e709d" containerName="registry-server" Oct 08 18:41:01 crc kubenswrapper[4988]: I1008 18:41:01.898849 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="77acd7a4-4ca3-4626-9c00-0cfad17e709d" containerName="registry-server" Oct 08 18:41:01 crc kubenswrapper[4988]: I1008 18:41:01.900758 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-vxx5r" Oct 08 18:41:01 crc kubenswrapper[4988]: I1008 18:41:01.925262 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-vxx5r"] Oct 08 18:41:02 crc kubenswrapper[4988]: I1008 18:41:02.092730 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3b9bf303-f9fc-4f51-8b3b-31627d7ea93e-utilities\") pod \"redhat-operators-vxx5r\" (UID: \"3b9bf303-f9fc-4f51-8b3b-31627d7ea93e\") " pod="openshift-marketplace/redhat-operators-vxx5r" Oct 08 18:41:02 crc kubenswrapper[4988]: I1008 18:41:02.092874 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3b9bf303-f9fc-4f51-8b3b-31627d7ea93e-catalog-content\") pod \"redhat-operators-vxx5r\" (UID: \"3b9bf303-f9fc-4f51-8b3b-31627d7ea93e\") " pod="openshift-marketplace/redhat-operators-vxx5r" Oct 08 18:41:02 crc kubenswrapper[4988]: I1008 18:41:02.093019 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nwg5d\" (UniqueName: \"kubernetes.io/projected/3b9bf303-f9fc-4f51-8b3b-31627d7ea93e-kube-api-access-nwg5d\") pod \"redhat-operators-vxx5r\" (UID: \"3b9bf303-f9fc-4f51-8b3b-31627d7ea93e\") " pod="openshift-marketplace/redhat-operators-vxx5r" Oct 08 18:41:02 crc kubenswrapper[4988]: I1008 18:41:02.194606 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nwg5d\" (UniqueName: \"kubernetes.io/projected/3b9bf303-f9fc-4f51-8b3b-31627d7ea93e-kube-api-access-nwg5d\") pod \"redhat-operators-vxx5r\" (UID: \"3b9bf303-f9fc-4f51-8b3b-31627d7ea93e\") " pod="openshift-marketplace/redhat-operators-vxx5r" Oct 08 18:41:02 crc kubenswrapper[4988]: I1008 18:41:02.194692 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3b9bf303-f9fc-4f51-8b3b-31627d7ea93e-utilities\") pod \"redhat-operators-vxx5r\" (UID: \"3b9bf303-f9fc-4f51-8b3b-31627d7ea93e\") " pod="openshift-marketplace/redhat-operators-vxx5r" Oct 08 18:41:02 crc kubenswrapper[4988]: I1008 18:41:02.194759 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3b9bf303-f9fc-4f51-8b3b-31627d7ea93e-catalog-content\") pod \"redhat-operators-vxx5r\" (UID: \"3b9bf303-f9fc-4f51-8b3b-31627d7ea93e\") " pod="openshift-marketplace/redhat-operators-vxx5r" Oct 08 18:41:02 crc kubenswrapper[4988]: I1008 18:41:02.195242 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3b9bf303-f9fc-4f51-8b3b-31627d7ea93e-catalog-content\") pod \"redhat-operators-vxx5r\" (UID: \"3b9bf303-f9fc-4f51-8b3b-31627d7ea93e\") " pod="openshift-marketplace/redhat-operators-vxx5r" Oct 08 18:41:02 crc kubenswrapper[4988]: I1008 18:41:02.195341 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3b9bf303-f9fc-4f51-8b3b-31627d7ea93e-utilities\") pod \"redhat-operators-vxx5r\" (UID: \"3b9bf303-f9fc-4f51-8b3b-31627d7ea93e\") " pod="openshift-marketplace/redhat-operators-vxx5r" Oct 08 18:41:02 crc kubenswrapper[4988]: I1008 18:41:02.214780 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nwg5d\" (UniqueName: \"kubernetes.io/projected/3b9bf303-f9fc-4f51-8b3b-31627d7ea93e-kube-api-access-nwg5d\") pod \"redhat-operators-vxx5r\" (UID: \"3b9bf303-f9fc-4f51-8b3b-31627d7ea93e\") " pod="openshift-marketplace/redhat-operators-vxx5r" Oct 08 18:41:02 crc kubenswrapper[4988]: I1008 18:41:02.229694 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-vxx5r" Oct 08 18:41:02 crc kubenswrapper[4988]: I1008 18:41:02.694499 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-vxx5r"] Oct 08 18:41:03 crc kubenswrapper[4988]: I1008 18:41:03.120950 4988 generic.go:334] "Generic (PLEG): container finished" podID="3b9bf303-f9fc-4f51-8b3b-31627d7ea93e" containerID="571e18487db11ac5ae0e1e35244c3df0a01ddaec71ace4c503a2c633d3344492" exitCode=0 Oct 08 18:41:03 crc kubenswrapper[4988]: I1008 18:41:03.120999 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vxx5r" event={"ID":"3b9bf303-f9fc-4f51-8b3b-31627d7ea93e","Type":"ContainerDied","Data":"571e18487db11ac5ae0e1e35244c3df0a01ddaec71ace4c503a2c633d3344492"} Oct 08 18:41:03 crc kubenswrapper[4988]: I1008 18:41:03.121031 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vxx5r" event={"ID":"3b9bf303-f9fc-4f51-8b3b-31627d7ea93e","Type":"ContainerStarted","Data":"3e8b0bd8997bc70abacad122e9a655c0e1e874ad77715660494c2806c5272358"} Oct 08 18:41:03 crc kubenswrapper[4988]: I1008 18:41:03.123119 4988 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 08 18:41:04 crc kubenswrapper[4988]: I1008 18:41:04.128361 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vxx5r" event={"ID":"3b9bf303-f9fc-4f51-8b3b-31627d7ea93e","Type":"ContainerStarted","Data":"92c3afb329ddda76f56e5352ae62526deb3a39b33c8a8e3c9b685809d760d2ae"} Oct 08 18:41:05 crc kubenswrapper[4988]: I1008 18:41:05.143080 4988 generic.go:334] "Generic (PLEG): container finished" podID="3b9bf303-f9fc-4f51-8b3b-31627d7ea93e" containerID="92c3afb329ddda76f56e5352ae62526deb3a39b33c8a8e3c9b685809d760d2ae" exitCode=0 Oct 08 18:41:05 crc kubenswrapper[4988]: I1008 18:41:05.143136 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vxx5r" event={"ID":"3b9bf303-f9fc-4f51-8b3b-31627d7ea93e","Type":"ContainerDied","Data":"92c3afb329ddda76f56e5352ae62526deb3a39b33c8a8e3c9b685809d760d2ae"} Oct 08 18:41:06 crc kubenswrapper[4988]: I1008 18:41:06.152886 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vxx5r" event={"ID":"3b9bf303-f9fc-4f51-8b3b-31627d7ea93e","Type":"ContainerStarted","Data":"d9c2595dadbe85e76e7f6663054832800467b3b6b99baa5bb75c1f59b134151d"} Oct 08 18:41:06 crc kubenswrapper[4988]: I1008 18:41:06.174972 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-vxx5r" podStartSLOduration=2.652370028 podStartE2EDuration="5.174955755s" podCreationTimestamp="2025-10-08 18:41:01 +0000 UTC" firstStartedPulling="2025-10-08 18:41:03.122650879 +0000 UTC m=+1808.572493689" lastFinishedPulling="2025-10-08 18:41:05.645236636 +0000 UTC m=+1811.095079416" observedRunningTime="2025-10-08 18:41:06.17222012 +0000 UTC m=+1811.622062910" watchObservedRunningTime="2025-10-08 18:41:06.174955755 +0000 UTC m=+1811.624798525" Oct 08 18:41:12 crc kubenswrapper[4988]: I1008 18:41:12.230616 4988 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-vxx5r" Oct 08 18:41:12 crc kubenswrapper[4988]: I1008 18:41:12.231194 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-vxx5r" Oct 08 18:41:12 crc kubenswrapper[4988]: I1008 18:41:12.237554 4988 scope.go:117] "RemoveContainer" containerID="1e3c2ea55d3c49c9fa802ab01ed9324a6c941b8da678ffaccb80377e3f159aee" Oct 08 18:41:12 crc kubenswrapper[4988]: E1008 18:41:12.237814 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nm54f_openshift-machine-config-operator(f328a20b-5283-407e-8a93-2ab2da599f18)\"" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" Oct 08 18:41:12 crc kubenswrapper[4988]: I1008 18:41:12.273957 4988 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-vxx5r" Oct 08 18:41:13 crc kubenswrapper[4988]: I1008 18:41:13.285047 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-vxx5r" Oct 08 18:41:13 crc kubenswrapper[4988]: I1008 18:41:13.339613 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-vxx5r"] Oct 08 18:41:15 crc kubenswrapper[4988]: I1008 18:41:15.224936 4988 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-vxx5r" podUID="3b9bf303-f9fc-4f51-8b3b-31627d7ea93e" containerName="registry-server" containerID="cri-o://d9c2595dadbe85e76e7f6663054832800467b3b6b99baa5bb75c1f59b134151d" gracePeriod=2 Oct 08 18:41:16 crc kubenswrapper[4988]: I1008 18:41:16.138321 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-vxx5r" Oct 08 18:41:16 crc kubenswrapper[4988]: I1008 18:41:16.236705 4988 generic.go:334] "Generic (PLEG): container finished" podID="3b9bf303-f9fc-4f51-8b3b-31627d7ea93e" containerID="d9c2595dadbe85e76e7f6663054832800467b3b6b99baa5bb75c1f59b134151d" exitCode=0 Oct 08 18:41:16 crc kubenswrapper[4988]: I1008 18:41:16.236791 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-vxx5r" Oct 08 18:41:16 crc kubenswrapper[4988]: I1008 18:41:16.237045 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vxx5r" event={"ID":"3b9bf303-f9fc-4f51-8b3b-31627d7ea93e","Type":"ContainerDied","Data":"d9c2595dadbe85e76e7f6663054832800467b3b6b99baa5bb75c1f59b134151d"} Oct 08 18:41:16 crc kubenswrapper[4988]: I1008 18:41:16.237137 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vxx5r" event={"ID":"3b9bf303-f9fc-4f51-8b3b-31627d7ea93e","Type":"ContainerDied","Data":"3e8b0bd8997bc70abacad122e9a655c0e1e874ad77715660494c2806c5272358"} Oct 08 18:41:16 crc kubenswrapper[4988]: I1008 18:41:16.237164 4988 scope.go:117] "RemoveContainer" containerID="d9c2595dadbe85e76e7f6663054832800467b3b6b99baa5bb75c1f59b134151d" Oct 08 18:41:16 crc kubenswrapper[4988]: I1008 18:41:16.281614 4988 scope.go:117] "RemoveContainer" containerID="92c3afb329ddda76f56e5352ae62526deb3a39b33c8a8e3c9b685809d760d2ae" Oct 08 18:41:16 crc kubenswrapper[4988]: I1008 18:41:16.298743 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3b9bf303-f9fc-4f51-8b3b-31627d7ea93e-utilities\") pod \"3b9bf303-f9fc-4f51-8b3b-31627d7ea93e\" (UID: \"3b9bf303-f9fc-4f51-8b3b-31627d7ea93e\") " Oct 08 18:41:16 crc kubenswrapper[4988]: I1008 18:41:16.298793 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nwg5d\" (UniqueName: \"kubernetes.io/projected/3b9bf303-f9fc-4f51-8b3b-31627d7ea93e-kube-api-access-nwg5d\") pod \"3b9bf303-f9fc-4f51-8b3b-31627d7ea93e\" (UID: \"3b9bf303-f9fc-4f51-8b3b-31627d7ea93e\") " Oct 08 18:41:16 crc kubenswrapper[4988]: I1008 18:41:16.298812 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3b9bf303-f9fc-4f51-8b3b-31627d7ea93e-catalog-content\") pod \"3b9bf303-f9fc-4f51-8b3b-31627d7ea93e\" (UID: \"3b9bf303-f9fc-4f51-8b3b-31627d7ea93e\") " Oct 08 18:41:16 crc kubenswrapper[4988]: I1008 18:41:16.299957 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3b9bf303-f9fc-4f51-8b3b-31627d7ea93e-utilities" (OuterVolumeSpecName: "utilities") pod "3b9bf303-f9fc-4f51-8b3b-31627d7ea93e" (UID: "3b9bf303-f9fc-4f51-8b3b-31627d7ea93e"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 18:41:16 crc kubenswrapper[4988]: I1008 18:41:16.301897 4988 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3b9bf303-f9fc-4f51-8b3b-31627d7ea93e-utilities\") on node \"crc\" DevicePath \"\"" Oct 08 18:41:16 crc kubenswrapper[4988]: I1008 18:41:16.306076 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3b9bf303-f9fc-4f51-8b3b-31627d7ea93e-kube-api-access-nwg5d" (OuterVolumeSpecName: "kube-api-access-nwg5d") pod "3b9bf303-f9fc-4f51-8b3b-31627d7ea93e" (UID: "3b9bf303-f9fc-4f51-8b3b-31627d7ea93e"). InnerVolumeSpecName "kube-api-access-nwg5d". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:41:16 crc kubenswrapper[4988]: I1008 18:41:16.314131 4988 scope.go:117] "RemoveContainer" containerID="571e18487db11ac5ae0e1e35244c3df0a01ddaec71ace4c503a2c633d3344492" Oct 08 18:41:16 crc kubenswrapper[4988]: I1008 18:41:16.381585 4988 scope.go:117] "RemoveContainer" containerID="d9c2595dadbe85e76e7f6663054832800467b3b6b99baa5bb75c1f59b134151d" Oct 08 18:41:16 crc kubenswrapper[4988]: E1008 18:41:16.382162 4988 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d9c2595dadbe85e76e7f6663054832800467b3b6b99baa5bb75c1f59b134151d\": container with ID starting with d9c2595dadbe85e76e7f6663054832800467b3b6b99baa5bb75c1f59b134151d not found: ID does not exist" containerID="d9c2595dadbe85e76e7f6663054832800467b3b6b99baa5bb75c1f59b134151d" Oct 08 18:41:16 crc kubenswrapper[4988]: I1008 18:41:16.382352 4988 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d9c2595dadbe85e76e7f6663054832800467b3b6b99baa5bb75c1f59b134151d"} err="failed to get container status \"d9c2595dadbe85e76e7f6663054832800467b3b6b99baa5bb75c1f59b134151d\": rpc error: code = NotFound desc = could not find container \"d9c2595dadbe85e76e7f6663054832800467b3b6b99baa5bb75c1f59b134151d\": container with ID starting with d9c2595dadbe85e76e7f6663054832800467b3b6b99baa5bb75c1f59b134151d not found: ID does not exist" Oct 08 18:41:16 crc kubenswrapper[4988]: I1008 18:41:16.382577 4988 scope.go:117] "RemoveContainer" containerID="92c3afb329ddda76f56e5352ae62526deb3a39b33c8a8e3c9b685809d760d2ae" Oct 08 18:41:16 crc kubenswrapper[4988]: E1008 18:41:16.383059 4988 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"92c3afb329ddda76f56e5352ae62526deb3a39b33c8a8e3c9b685809d760d2ae\": container with ID starting with 92c3afb329ddda76f56e5352ae62526deb3a39b33c8a8e3c9b685809d760d2ae not found: ID does not exist" containerID="92c3afb329ddda76f56e5352ae62526deb3a39b33c8a8e3c9b685809d760d2ae" Oct 08 18:41:16 crc kubenswrapper[4988]: I1008 18:41:16.383091 4988 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"92c3afb329ddda76f56e5352ae62526deb3a39b33c8a8e3c9b685809d760d2ae"} err="failed to get container status \"92c3afb329ddda76f56e5352ae62526deb3a39b33c8a8e3c9b685809d760d2ae\": rpc error: code = NotFound desc = could not find container \"92c3afb329ddda76f56e5352ae62526deb3a39b33c8a8e3c9b685809d760d2ae\": container with ID starting with 92c3afb329ddda76f56e5352ae62526deb3a39b33c8a8e3c9b685809d760d2ae not found: ID does not exist" Oct 08 18:41:16 crc kubenswrapper[4988]: I1008 18:41:16.383110 4988 scope.go:117] "RemoveContainer" containerID="571e18487db11ac5ae0e1e35244c3df0a01ddaec71ace4c503a2c633d3344492" Oct 08 18:41:16 crc kubenswrapper[4988]: E1008 18:41:16.383724 4988 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"571e18487db11ac5ae0e1e35244c3df0a01ddaec71ace4c503a2c633d3344492\": container with ID starting with 571e18487db11ac5ae0e1e35244c3df0a01ddaec71ace4c503a2c633d3344492 not found: ID does not exist" containerID="571e18487db11ac5ae0e1e35244c3df0a01ddaec71ace4c503a2c633d3344492" Oct 08 18:41:16 crc kubenswrapper[4988]: I1008 18:41:16.383791 4988 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"571e18487db11ac5ae0e1e35244c3df0a01ddaec71ace4c503a2c633d3344492"} err="failed to get container status \"571e18487db11ac5ae0e1e35244c3df0a01ddaec71ace4c503a2c633d3344492\": rpc error: code = NotFound desc = could not find container \"571e18487db11ac5ae0e1e35244c3df0a01ddaec71ace4c503a2c633d3344492\": container with ID starting with 571e18487db11ac5ae0e1e35244c3df0a01ddaec71ace4c503a2c633d3344492 not found: ID does not exist" Oct 08 18:41:16 crc kubenswrapper[4988]: I1008 18:41:16.395509 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3b9bf303-f9fc-4f51-8b3b-31627d7ea93e-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "3b9bf303-f9fc-4f51-8b3b-31627d7ea93e" (UID: "3b9bf303-f9fc-4f51-8b3b-31627d7ea93e"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 18:41:16 crc kubenswrapper[4988]: I1008 18:41:16.405461 4988 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nwg5d\" (UniqueName: \"kubernetes.io/projected/3b9bf303-f9fc-4f51-8b3b-31627d7ea93e-kube-api-access-nwg5d\") on node \"crc\" DevicePath \"\"" Oct 08 18:41:16 crc kubenswrapper[4988]: I1008 18:41:16.405520 4988 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3b9bf303-f9fc-4f51-8b3b-31627d7ea93e-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 08 18:41:16 crc kubenswrapper[4988]: I1008 18:41:16.582269 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-vxx5r"] Oct 08 18:41:16 crc kubenswrapper[4988]: I1008 18:41:16.588961 4988 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-vxx5r"] Oct 08 18:41:17 crc kubenswrapper[4988]: I1008 18:41:17.257548 4988 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3b9bf303-f9fc-4f51-8b3b-31627d7ea93e" path="/var/lib/kubelet/pods/3b9bf303-f9fc-4f51-8b3b-31627d7ea93e/volumes" Oct 08 18:41:24 crc kubenswrapper[4988]: I1008 18:41:24.237257 4988 scope.go:117] "RemoveContainer" containerID="1e3c2ea55d3c49c9fa802ab01ed9324a6c941b8da678ffaccb80377e3f159aee" Oct 08 18:41:24 crc kubenswrapper[4988]: E1008 18:41:24.238039 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nm54f_openshift-machine-config-operator(f328a20b-5283-407e-8a93-2ab2da599f18)\"" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" Oct 08 18:41:35 crc kubenswrapper[4988]: I1008 18:41:35.246887 4988 scope.go:117] "RemoveContainer" containerID="1e3c2ea55d3c49c9fa802ab01ed9324a6c941b8da678ffaccb80377e3f159aee" Oct 08 18:41:35 crc kubenswrapper[4988]: E1008 18:41:35.248162 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nm54f_openshift-machine-config-operator(f328a20b-5283-407e-8a93-2ab2da599f18)\"" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" Oct 08 18:41:46 crc kubenswrapper[4988]: I1008 18:41:46.238216 4988 scope.go:117] "RemoveContainer" containerID="1e3c2ea55d3c49c9fa802ab01ed9324a6c941b8da678ffaccb80377e3f159aee" Oct 08 18:41:46 crc kubenswrapper[4988]: E1008 18:41:46.239466 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nm54f_openshift-machine-config-operator(f328a20b-5283-407e-8a93-2ab2da599f18)\"" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" Oct 08 18:41:59 crc kubenswrapper[4988]: I1008 18:41:59.238734 4988 scope.go:117] "RemoveContainer" containerID="1e3c2ea55d3c49c9fa802ab01ed9324a6c941b8da678ffaccb80377e3f159aee" Oct 08 18:41:59 crc kubenswrapper[4988]: I1008 18:41:59.639922 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" event={"ID":"f328a20b-5283-407e-8a93-2ab2da599f18","Type":"ContainerStarted","Data":"c9447394f2908e13ce0cc9fbb2e1c8c9953530f5820e7b144ac3da28722ce971"} Oct 08 18:44:16 crc kubenswrapper[4988]: I1008 18:44:16.940760 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-w95sg"] Oct 08 18:44:16 crc kubenswrapper[4988]: E1008 18:44:16.942320 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3b9bf303-f9fc-4f51-8b3b-31627d7ea93e" containerName="extract-utilities" Oct 08 18:44:16 crc kubenswrapper[4988]: I1008 18:44:16.942349 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="3b9bf303-f9fc-4f51-8b3b-31627d7ea93e" containerName="extract-utilities" Oct 08 18:44:16 crc kubenswrapper[4988]: E1008 18:44:16.942426 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3b9bf303-f9fc-4f51-8b3b-31627d7ea93e" containerName="registry-server" Oct 08 18:44:16 crc kubenswrapper[4988]: I1008 18:44:16.942445 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="3b9bf303-f9fc-4f51-8b3b-31627d7ea93e" containerName="registry-server" Oct 08 18:44:16 crc kubenswrapper[4988]: E1008 18:44:16.942482 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3b9bf303-f9fc-4f51-8b3b-31627d7ea93e" containerName="extract-content" Oct 08 18:44:16 crc kubenswrapper[4988]: I1008 18:44:16.942501 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="3b9bf303-f9fc-4f51-8b3b-31627d7ea93e" containerName="extract-content" Oct 08 18:44:16 crc kubenswrapper[4988]: I1008 18:44:16.942860 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="3b9bf303-f9fc-4f51-8b3b-31627d7ea93e" containerName="registry-server" Oct 08 18:44:16 crc kubenswrapper[4988]: I1008 18:44:16.946075 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-w95sg" Oct 08 18:44:16 crc kubenswrapper[4988]: I1008 18:44:16.966491 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-w95sg"] Oct 08 18:44:17 crc kubenswrapper[4988]: I1008 18:44:17.090685 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/72ec298a-45fa-4fe7-9ad8-c07446170868-utilities\") pod \"certified-operators-w95sg\" (UID: \"72ec298a-45fa-4fe7-9ad8-c07446170868\") " pod="openshift-marketplace/certified-operators-w95sg" Oct 08 18:44:17 crc kubenswrapper[4988]: I1008 18:44:17.090721 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vnzqs\" (UniqueName: \"kubernetes.io/projected/72ec298a-45fa-4fe7-9ad8-c07446170868-kube-api-access-vnzqs\") pod \"certified-operators-w95sg\" (UID: \"72ec298a-45fa-4fe7-9ad8-c07446170868\") " pod="openshift-marketplace/certified-operators-w95sg" Oct 08 18:44:17 crc kubenswrapper[4988]: I1008 18:44:17.090752 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/72ec298a-45fa-4fe7-9ad8-c07446170868-catalog-content\") pod \"certified-operators-w95sg\" (UID: \"72ec298a-45fa-4fe7-9ad8-c07446170868\") " pod="openshift-marketplace/certified-operators-w95sg" Oct 08 18:44:17 crc kubenswrapper[4988]: I1008 18:44:17.191973 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/72ec298a-45fa-4fe7-9ad8-c07446170868-utilities\") pod \"certified-operators-w95sg\" (UID: \"72ec298a-45fa-4fe7-9ad8-c07446170868\") " pod="openshift-marketplace/certified-operators-w95sg" Oct 08 18:44:17 crc kubenswrapper[4988]: I1008 18:44:17.192080 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vnzqs\" (UniqueName: \"kubernetes.io/projected/72ec298a-45fa-4fe7-9ad8-c07446170868-kube-api-access-vnzqs\") pod \"certified-operators-w95sg\" (UID: \"72ec298a-45fa-4fe7-9ad8-c07446170868\") " pod="openshift-marketplace/certified-operators-w95sg" Oct 08 18:44:17 crc kubenswrapper[4988]: I1008 18:44:17.192144 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/72ec298a-45fa-4fe7-9ad8-c07446170868-catalog-content\") pod \"certified-operators-w95sg\" (UID: \"72ec298a-45fa-4fe7-9ad8-c07446170868\") " pod="openshift-marketplace/certified-operators-w95sg" Oct 08 18:44:17 crc kubenswrapper[4988]: I1008 18:44:17.192614 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/72ec298a-45fa-4fe7-9ad8-c07446170868-catalog-content\") pod \"certified-operators-w95sg\" (UID: \"72ec298a-45fa-4fe7-9ad8-c07446170868\") " pod="openshift-marketplace/certified-operators-w95sg" Oct 08 18:44:17 crc kubenswrapper[4988]: I1008 18:44:17.192614 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/72ec298a-45fa-4fe7-9ad8-c07446170868-utilities\") pod \"certified-operators-w95sg\" (UID: \"72ec298a-45fa-4fe7-9ad8-c07446170868\") " pod="openshift-marketplace/certified-operators-w95sg" Oct 08 18:44:17 crc kubenswrapper[4988]: I1008 18:44:17.218725 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vnzqs\" (UniqueName: \"kubernetes.io/projected/72ec298a-45fa-4fe7-9ad8-c07446170868-kube-api-access-vnzqs\") pod \"certified-operators-w95sg\" (UID: \"72ec298a-45fa-4fe7-9ad8-c07446170868\") " pod="openshift-marketplace/certified-operators-w95sg" Oct 08 18:44:17 crc kubenswrapper[4988]: I1008 18:44:17.264411 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-w95sg" Oct 08 18:44:17 crc kubenswrapper[4988]: I1008 18:44:17.648113 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-w95sg"] Oct 08 18:44:17 crc kubenswrapper[4988]: I1008 18:44:17.904772 4988 generic.go:334] "Generic (PLEG): container finished" podID="72ec298a-45fa-4fe7-9ad8-c07446170868" containerID="d921d93f9a7249fb3046a645d4f788ee0fe89c8e0d472b78ee2ee79f2cf974f7" exitCode=0 Oct 08 18:44:17 crc kubenswrapper[4988]: I1008 18:44:17.904813 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-w95sg" event={"ID":"72ec298a-45fa-4fe7-9ad8-c07446170868","Type":"ContainerDied","Data":"d921d93f9a7249fb3046a645d4f788ee0fe89c8e0d472b78ee2ee79f2cf974f7"} Oct 08 18:44:17 crc kubenswrapper[4988]: I1008 18:44:17.904836 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-w95sg" event={"ID":"72ec298a-45fa-4fe7-9ad8-c07446170868","Type":"ContainerStarted","Data":"57e429e2103d6483074418308288b0b1ae885456db18b616e1fae5f9c73ce637"} Oct 08 18:44:18 crc kubenswrapper[4988]: I1008 18:44:18.913662 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-w95sg" event={"ID":"72ec298a-45fa-4fe7-9ad8-c07446170868","Type":"ContainerStarted","Data":"eef53e6cee57f898c395084aedc4a6fb80f8f55140b16ab54c16b056b54a6f34"} Oct 08 18:44:19 crc kubenswrapper[4988]: I1008 18:44:19.922631 4988 generic.go:334] "Generic (PLEG): container finished" podID="72ec298a-45fa-4fe7-9ad8-c07446170868" containerID="eef53e6cee57f898c395084aedc4a6fb80f8f55140b16ab54c16b056b54a6f34" exitCode=0 Oct 08 18:44:19 crc kubenswrapper[4988]: I1008 18:44:19.922731 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-w95sg" event={"ID":"72ec298a-45fa-4fe7-9ad8-c07446170868","Type":"ContainerDied","Data":"eef53e6cee57f898c395084aedc4a6fb80f8f55140b16ab54c16b056b54a6f34"} Oct 08 18:44:20 crc kubenswrapper[4988]: I1008 18:44:20.931903 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-w95sg" event={"ID":"72ec298a-45fa-4fe7-9ad8-c07446170868","Type":"ContainerStarted","Data":"1f2b253fb647d1ee9976528e334bbcf160c9ecbfdebe29eb8c5cb8eddee1bca9"} Oct 08 18:44:20 crc kubenswrapper[4988]: I1008 18:44:20.950137 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-w95sg" podStartSLOduration=2.489967676 podStartE2EDuration="4.95011693s" podCreationTimestamp="2025-10-08 18:44:16 +0000 UTC" firstStartedPulling="2025-10-08 18:44:17.905917391 +0000 UTC m=+2003.355760161" lastFinishedPulling="2025-10-08 18:44:20.366066645 +0000 UTC m=+2005.815909415" observedRunningTime="2025-10-08 18:44:20.947607586 +0000 UTC m=+2006.397450376" watchObservedRunningTime="2025-10-08 18:44:20.95011693 +0000 UTC m=+2006.399959710" Oct 08 18:44:23 crc kubenswrapper[4988]: I1008 18:44:23.338594 4988 patch_prober.go:28] interesting pod/machine-config-daemon-nm54f container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 18:44:23 crc kubenswrapper[4988]: I1008 18:44:23.338912 4988 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 18:44:27 crc kubenswrapper[4988]: I1008 18:44:27.264689 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-w95sg" Oct 08 18:44:27 crc kubenswrapper[4988]: I1008 18:44:27.265014 4988 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-w95sg" Oct 08 18:44:27 crc kubenswrapper[4988]: I1008 18:44:27.316771 4988 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-w95sg" Oct 08 18:44:28 crc kubenswrapper[4988]: I1008 18:44:28.060774 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-w95sg" Oct 08 18:44:28 crc kubenswrapper[4988]: I1008 18:44:28.109822 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-w95sg"] Oct 08 18:44:30 crc kubenswrapper[4988]: I1008 18:44:30.007696 4988 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-w95sg" podUID="72ec298a-45fa-4fe7-9ad8-c07446170868" containerName="registry-server" containerID="cri-o://1f2b253fb647d1ee9976528e334bbcf160c9ecbfdebe29eb8c5cb8eddee1bca9" gracePeriod=2 Oct 08 18:44:30 crc kubenswrapper[4988]: I1008 18:44:30.424529 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-w95sg" Oct 08 18:44:30 crc kubenswrapper[4988]: I1008 18:44:30.512737 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vnzqs\" (UniqueName: \"kubernetes.io/projected/72ec298a-45fa-4fe7-9ad8-c07446170868-kube-api-access-vnzqs\") pod \"72ec298a-45fa-4fe7-9ad8-c07446170868\" (UID: \"72ec298a-45fa-4fe7-9ad8-c07446170868\") " Oct 08 18:44:30 crc kubenswrapper[4988]: I1008 18:44:30.512928 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/72ec298a-45fa-4fe7-9ad8-c07446170868-catalog-content\") pod \"72ec298a-45fa-4fe7-9ad8-c07446170868\" (UID: \"72ec298a-45fa-4fe7-9ad8-c07446170868\") " Oct 08 18:44:30 crc kubenswrapper[4988]: I1008 18:44:30.512961 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/72ec298a-45fa-4fe7-9ad8-c07446170868-utilities\") pod \"72ec298a-45fa-4fe7-9ad8-c07446170868\" (UID: \"72ec298a-45fa-4fe7-9ad8-c07446170868\") " Oct 08 18:44:30 crc kubenswrapper[4988]: I1008 18:44:30.513761 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/72ec298a-45fa-4fe7-9ad8-c07446170868-utilities" (OuterVolumeSpecName: "utilities") pod "72ec298a-45fa-4fe7-9ad8-c07446170868" (UID: "72ec298a-45fa-4fe7-9ad8-c07446170868"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 18:44:30 crc kubenswrapper[4988]: I1008 18:44:30.515798 4988 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/72ec298a-45fa-4fe7-9ad8-c07446170868-utilities\") on node \"crc\" DevicePath \"\"" Oct 08 18:44:30 crc kubenswrapper[4988]: I1008 18:44:30.519734 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/72ec298a-45fa-4fe7-9ad8-c07446170868-kube-api-access-vnzqs" (OuterVolumeSpecName: "kube-api-access-vnzqs") pod "72ec298a-45fa-4fe7-9ad8-c07446170868" (UID: "72ec298a-45fa-4fe7-9ad8-c07446170868"). InnerVolumeSpecName "kube-api-access-vnzqs". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:44:30 crc kubenswrapper[4988]: I1008 18:44:30.559528 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/72ec298a-45fa-4fe7-9ad8-c07446170868-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "72ec298a-45fa-4fe7-9ad8-c07446170868" (UID: "72ec298a-45fa-4fe7-9ad8-c07446170868"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 18:44:30 crc kubenswrapper[4988]: I1008 18:44:30.618016 4988 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vnzqs\" (UniqueName: \"kubernetes.io/projected/72ec298a-45fa-4fe7-9ad8-c07446170868-kube-api-access-vnzqs\") on node \"crc\" DevicePath \"\"" Oct 08 18:44:30 crc kubenswrapper[4988]: I1008 18:44:30.618080 4988 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/72ec298a-45fa-4fe7-9ad8-c07446170868-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 08 18:44:31 crc kubenswrapper[4988]: I1008 18:44:31.018684 4988 generic.go:334] "Generic (PLEG): container finished" podID="72ec298a-45fa-4fe7-9ad8-c07446170868" containerID="1f2b253fb647d1ee9976528e334bbcf160c9ecbfdebe29eb8c5cb8eddee1bca9" exitCode=0 Oct 08 18:44:31 crc kubenswrapper[4988]: I1008 18:44:31.018733 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-w95sg" event={"ID":"72ec298a-45fa-4fe7-9ad8-c07446170868","Type":"ContainerDied","Data":"1f2b253fb647d1ee9976528e334bbcf160c9ecbfdebe29eb8c5cb8eddee1bca9"} Oct 08 18:44:31 crc kubenswrapper[4988]: I1008 18:44:31.018766 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-w95sg" event={"ID":"72ec298a-45fa-4fe7-9ad8-c07446170868","Type":"ContainerDied","Data":"57e429e2103d6483074418308288b0b1ae885456db18b616e1fae5f9c73ce637"} Oct 08 18:44:31 crc kubenswrapper[4988]: I1008 18:44:31.018789 4988 scope.go:117] "RemoveContainer" containerID="1f2b253fb647d1ee9976528e334bbcf160c9ecbfdebe29eb8c5cb8eddee1bca9" Oct 08 18:44:31 crc kubenswrapper[4988]: I1008 18:44:31.018969 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-w95sg" Oct 08 18:44:31 crc kubenswrapper[4988]: I1008 18:44:31.069446 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-w95sg"] Oct 08 18:44:31 crc kubenswrapper[4988]: I1008 18:44:31.073584 4988 scope.go:117] "RemoveContainer" containerID="eef53e6cee57f898c395084aedc4a6fb80f8f55140b16ab54c16b056b54a6f34" Oct 08 18:44:31 crc kubenswrapper[4988]: I1008 18:44:31.074438 4988 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-w95sg"] Oct 08 18:44:31 crc kubenswrapper[4988]: I1008 18:44:31.111060 4988 scope.go:117] "RemoveContainer" containerID="d921d93f9a7249fb3046a645d4f788ee0fe89c8e0d472b78ee2ee79f2cf974f7" Oct 08 18:44:31 crc kubenswrapper[4988]: I1008 18:44:31.161517 4988 scope.go:117] "RemoveContainer" containerID="1f2b253fb647d1ee9976528e334bbcf160c9ecbfdebe29eb8c5cb8eddee1bca9" Oct 08 18:44:31 crc kubenswrapper[4988]: E1008 18:44:31.165976 4988 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1f2b253fb647d1ee9976528e334bbcf160c9ecbfdebe29eb8c5cb8eddee1bca9\": container with ID starting with 1f2b253fb647d1ee9976528e334bbcf160c9ecbfdebe29eb8c5cb8eddee1bca9 not found: ID does not exist" containerID="1f2b253fb647d1ee9976528e334bbcf160c9ecbfdebe29eb8c5cb8eddee1bca9" Oct 08 18:44:31 crc kubenswrapper[4988]: I1008 18:44:31.166011 4988 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1f2b253fb647d1ee9976528e334bbcf160c9ecbfdebe29eb8c5cb8eddee1bca9"} err="failed to get container status \"1f2b253fb647d1ee9976528e334bbcf160c9ecbfdebe29eb8c5cb8eddee1bca9\": rpc error: code = NotFound desc = could not find container \"1f2b253fb647d1ee9976528e334bbcf160c9ecbfdebe29eb8c5cb8eddee1bca9\": container with ID starting with 1f2b253fb647d1ee9976528e334bbcf160c9ecbfdebe29eb8c5cb8eddee1bca9 not found: ID does not exist" Oct 08 18:44:31 crc kubenswrapper[4988]: I1008 18:44:31.166034 4988 scope.go:117] "RemoveContainer" containerID="eef53e6cee57f898c395084aedc4a6fb80f8f55140b16ab54c16b056b54a6f34" Oct 08 18:44:31 crc kubenswrapper[4988]: E1008 18:44:31.168805 4988 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"eef53e6cee57f898c395084aedc4a6fb80f8f55140b16ab54c16b056b54a6f34\": container with ID starting with eef53e6cee57f898c395084aedc4a6fb80f8f55140b16ab54c16b056b54a6f34 not found: ID does not exist" containerID="eef53e6cee57f898c395084aedc4a6fb80f8f55140b16ab54c16b056b54a6f34" Oct 08 18:44:31 crc kubenswrapper[4988]: I1008 18:44:31.168828 4988 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"eef53e6cee57f898c395084aedc4a6fb80f8f55140b16ab54c16b056b54a6f34"} err="failed to get container status \"eef53e6cee57f898c395084aedc4a6fb80f8f55140b16ab54c16b056b54a6f34\": rpc error: code = NotFound desc = could not find container \"eef53e6cee57f898c395084aedc4a6fb80f8f55140b16ab54c16b056b54a6f34\": container with ID starting with eef53e6cee57f898c395084aedc4a6fb80f8f55140b16ab54c16b056b54a6f34 not found: ID does not exist" Oct 08 18:44:31 crc kubenswrapper[4988]: I1008 18:44:31.168841 4988 scope.go:117] "RemoveContainer" containerID="d921d93f9a7249fb3046a645d4f788ee0fe89c8e0d472b78ee2ee79f2cf974f7" Oct 08 18:44:31 crc kubenswrapper[4988]: E1008 18:44:31.174065 4988 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d921d93f9a7249fb3046a645d4f788ee0fe89c8e0d472b78ee2ee79f2cf974f7\": container with ID starting with d921d93f9a7249fb3046a645d4f788ee0fe89c8e0d472b78ee2ee79f2cf974f7 not found: ID does not exist" containerID="d921d93f9a7249fb3046a645d4f788ee0fe89c8e0d472b78ee2ee79f2cf974f7" Oct 08 18:44:31 crc kubenswrapper[4988]: I1008 18:44:31.174112 4988 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d921d93f9a7249fb3046a645d4f788ee0fe89c8e0d472b78ee2ee79f2cf974f7"} err="failed to get container status \"d921d93f9a7249fb3046a645d4f788ee0fe89c8e0d472b78ee2ee79f2cf974f7\": rpc error: code = NotFound desc = could not find container \"d921d93f9a7249fb3046a645d4f788ee0fe89c8e0d472b78ee2ee79f2cf974f7\": container with ID starting with d921d93f9a7249fb3046a645d4f788ee0fe89c8e0d472b78ee2ee79f2cf974f7 not found: ID does not exist" Oct 08 18:44:31 crc kubenswrapper[4988]: I1008 18:44:31.270274 4988 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="72ec298a-45fa-4fe7-9ad8-c07446170868" path="/var/lib/kubelet/pods/72ec298a-45fa-4fe7-9ad8-c07446170868/volumes" Oct 08 18:44:53 crc kubenswrapper[4988]: I1008 18:44:53.338018 4988 patch_prober.go:28] interesting pod/machine-config-daemon-nm54f container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 18:44:53 crc kubenswrapper[4988]: I1008 18:44:53.339833 4988 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 18:45:00 crc kubenswrapper[4988]: I1008 18:45:00.157100 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29332485-n65qn"] Oct 08 18:45:00 crc kubenswrapper[4988]: E1008 18:45:00.157897 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="72ec298a-45fa-4fe7-9ad8-c07446170868" containerName="extract-content" Oct 08 18:45:00 crc kubenswrapper[4988]: I1008 18:45:00.157920 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="72ec298a-45fa-4fe7-9ad8-c07446170868" containerName="extract-content" Oct 08 18:45:00 crc kubenswrapper[4988]: E1008 18:45:00.157963 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="72ec298a-45fa-4fe7-9ad8-c07446170868" containerName="registry-server" Oct 08 18:45:00 crc kubenswrapper[4988]: I1008 18:45:00.157974 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="72ec298a-45fa-4fe7-9ad8-c07446170868" containerName="registry-server" Oct 08 18:45:00 crc kubenswrapper[4988]: E1008 18:45:00.157995 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="72ec298a-45fa-4fe7-9ad8-c07446170868" containerName="extract-utilities" Oct 08 18:45:00 crc kubenswrapper[4988]: I1008 18:45:00.158006 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="72ec298a-45fa-4fe7-9ad8-c07446170868" containerName="extract-utilities" Oct 08 18:45:00 crc kubenswrapper[4988]: I1008 18:45:00.158246 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="72ec298a-45fa-4fe7-9ad8-c07446170868" containerName="registry-server" Oct 08 18:45:00 crc kubenswrapper[4988]: I1008 18:45:00.159017 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29332485-n65qn" Oct 08 18:45:00 crc kubenswrapper[4988]: I1008 18:45:00.161686 4988 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 08 18:45:00 crc kubenswrapper[4988]: I1008 18:45:00.162830 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 08 18:45:00 crc kubenswrapper[4988]: I1008 18:45:00.181027 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29332485-n65qn"] Oct 08 18:45:00 crc kubenswrapper[4988]: I1008 18:45:00.276559 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6vg4b\" (UniqueName: \"kubernetes.io/projected/58a39bcb-ac94-4290-b2cb-0bea4547a67c-kube-api-access-6vg4b\") pod \"collect-profiles-29332485-n65qn\" (UID: \"58a39bcb-ac94-4290-b2cb-0bea4547a67c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29332485-n65qn" Oct 08 18:45:00 crc kubenswrapper[4988]: I1008 18:45:00.276655 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/58a39bcb-ac94-4290-b2cb-0bea4547a67c-secret-volume\") pod \"collect-profiles-29332485-n65qn\" (UID: \"58a39bcb-ac94-4290-b2cb-0bea4547a67c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29332485-n65qn" Oct 08 18:45:00 crc kubenswrapper[4988]: I1008 18:45:00.276825 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/58a39bcb-ac94-4290-b2cb-0bea4547a67c-config-volume\") pod \"collect-profiles-29332485-n65qn\" (UID: \"58a39bcb-ac94-4290-b2cb-0bea4547a67c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29332485-n65qn" Oct 08 18:45:00 crc kubenswrapper[4988]: I1008 18:45:00.379145 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/58a39bcb-ac94-4290-b2cb-0bea4547a67c-config-volume\") pod \"collect-profiles-29332485-n65qn\" (UID: \"58a39bcb-ac94-4290-b2cb-0bea4547a67c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29332485-n65qn" Oct 08 18:45:00 crc kubenswrapper[4988]: I1008 18:45:00.379490 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6vg4b\" (UniqueName: \"kubernetes.io/projected/58a39bcb-ac94-4290-b2cb-0bea4547a67c-kube-api-access-6vg4b\") pod \"collect-profiles-29332485-n65qn\" (UID: \"58a39bcb-ac94-4290-b2cb-0bea4547a67c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29332485-n65qn" Oct 08 18:45:00 crc kubenswrapper[4988]: I1008 18:45:00.379582 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/58a39bcb-ac94-4290-b2cb-0bea4547a67c-secret-volume\") pod \"collect-profiles-29332485-n65qn\" (UID: \"58a39bcb-ac94-4290-b2cb-0bea4547a67c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29332485-n65qn" Oct 08 18:45:00 crc kubenswrapper[4988]: I1008 18:45:00.381081 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/58a39bcb-ac94-4290-b2cb-0bea4547a67c-config-volume\") pod \"collect-profiles-29332485-n65qn\" (UID: \"58a39bcb-ac94-4290-b2cb-0bea4547a67c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29332485-n65qn" Oct 08 18:45:00 crc kubenswrapper[4988]: I1008 18:45:00.394109 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/58a39bcb-ac94-4290-b2cb-0bea4547a67c-secret-volume\") pod \"collect-profiles-29332485-n65qn\" (UID: \"58a39bcb-ac94-4290-b2cb-0bea4547a67c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29332485-n65qn" Oct 08 18:45:00 crc kubenswrapper[4988]: I1008 18:45:00.412560 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6vg4b\" (UniqueName: \"kubernetes.io/projected/58a39bcb-ac94-4290-b2cb-0bea4547a67c-kube-api-access-6vg4b\") pod \"collect-profiles-29332485-n65qn\" (UID: \"58a39bcb-ac94-4290-b2cb-0bea4547a67c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29332485-n65qn" Oct 08 18:45:00 crc kubenswrapper[4988]: I1008 18:45:00.482062 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29332485-n65qn" Oct 08 18:45:00 crc kubenswrapper[4988]: I1008 18:45:00.938675 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29332485-n65qn"] Oct 08 18:45:01 crc kubenswrapper[4988]: I1008 18:45:01.291996 4988 generic.go:334] "Generic (PLEG): container finished" podID="58a39bcb-ac94-4290-b2cb-0bea4547a67c" containerID="035f001a51642541a361b47eefe00a39e503f22f29697eb33feb5c4a611ce3e3" exitCode=0 Oct 08 18:45:01 crc kubenswrapper[4988]: I1008 18:45:01.292115 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29332485-n65qn" event={"ID":"58a39bcb-ac94-4290-b2cb-0bea4547a67c","Type":"ContainerDied","Data":"035f001a51642541a361b47eefe00a39e503f22f29697eb33feb5c4a611ce3e3"} Oct 08 18:45:01 crc kubenswrapper[4988]: I1008 18:45:01.292346 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29332485-n65qn" event={"ID":"58a39bcb-ac94-4290-b2cb-0bea4547a67c","Type":"ContainerStarted","Data":"8fd817a2cddd61cbc9f32f456103cda7e21dd36a988f791180b8b34d639f64e3"} Oct 08 18:45:02 crc kubenswrapper[4988]: I1008 18:45:02.624666 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29332485-n65qn" Oct 08 18:45:02 crc kubenswrapper[4988]: I1008 18:45:02.715781 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/58a39bcb-ac94-4290-b2cb-0bea4547a67c-config-volume\") pod \"58a39bcb-ac94-4290-b2cb-0bea4547a67c\" (UID: \"58a39bcb-ac94-4290-b2cb-0bea4547a67c\") " Oct 08 18:45:02 crc kubenswrapper[4988]: I1008 18:45:02.715893 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6vg4b\" (UniqueName: \"kubernetes.io/projected/58a39bcb-ac94-4290-b2cb-0bea4547a67c-kube-api-access-6vg4b\") pod \"58a39bcb-ac94-4290-b2cb-0bea4547a67c\" (UID: \"58a39bcb-ac94-4290-b2cb-0bea4547a67c\") " Oct 08 18:45:02 crc kubenswrapper[4988]: I1008 18:45:02.715931 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/58a39bcb-ac94-4290-b2cb-0bea4547a67c-secret-volume\") pod \"58a39bcb-ac94-4290-b2cb-0bea4547a67c\" (UID: \"58a39bcb-ac94-4290-b2cb-0bea4547a67c\") " Oct 08 18:45:02 crc kubenswrapper[4988]: I1008 18:45:02.716940 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/58a39bcb-ac94-4290-b2cb-0bea4547a67c-config-volume" (OuterVolumeSpecName: "config-volume") pod "58a39bcb-ac94-4290-b2cb-0bea4547a67c" (UID: "58a39bcb-ac94-4290-b2cb-0bea4547a67c"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 18:45:02 crc kubenswrapper[4988]: I1008 18:45:02.727897 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/58a39bcb-ac94-4290-b2cb-0bea4547a67c-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "58a39bcb-ac94-4290-b2cb-0bea4547a67c" (UID: "58a39bcb-ac94-4290-b2cb-0bea4547a67c"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:45:02 crc kubenswrapper[4988]: I1008 18:45:02.727928 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/58a39bcb-ac94-4290-b2cb-0bea4547a67c-kube-api-access-6vg4b" (OuterVolumeSpecName: "kube-api-access-6vg4b") pod "58a39bcb-ac94-4290-b2cb-0bea4547a67c" (UID: "58a39bcb-ac94-4290-b2cb-0bea4547a67c"). InnerVolumeSpecName "kube-api-access-6vg4b". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:45:02 crc kubenswrapper[4988]: I1008 18:45:02.818155 4988 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6vg4b\" (UniqueName: \"kubernetes.io/projected/58a39bcb-ac94-4290-b2cb-0bea4547a67c-kube-api-access-6vg4b\") on node \"crc\" DevicePath \"\"" Oct 08 18:45:02 crc kubenswrapper[4988]: I1008 18:45:02.818192 4988 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/58a39bcb-ac94-4290-b2cb-0bea4547a67c-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 08 18:45:02 crc kubenswrapper[4988]: I1008 18:45:02.818204 4988 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/58a39bcb-ac94-4290-b2cb-0bea4547a67c-config-volume\") on node \"crc\" DevicePath \"\"" Oct 08 18:45:03 crc kubenswrapper[4988]: I1008 18:45:03.312479 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29332485-n65qn" event={"ID":"58a39bcb-ac94-4290-b2cb-0bea4547a67c","Type":"ContainerDied","Data":"8fd817a2cddd61cbc9f32f456103cda7e21dd36a988f791180b8b34d639f64e3"} Oct 08 18:45:03 crc kubenswrapper[4988]: I1008 18:45:03.312865 4988 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8fd817a2cddd61cbc9f32f456103cda7e21dd36a988f791180b8b34d639f64e3" Oct 08 18:45:03 crc kubenswrapper[4988]: I1008 18:45:03.312550 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29332485-n65qn" Oct 08 18:45:03 crc kubenswrapper[4988]: I1008 18:45:03.714405 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29332440-qxt8r"] Oct 08 18:45:03 crc kubenswrapper[4988]: I1008 18:45:03.720562 4988 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29332440-qxt8r"] Oct 08 18:45:05 crc kubenswrapper[4988]: I1008 18:45:05.255192 4988 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c08e9af6-fc25-4c32-bd25-ac82aebb828e" path="/var/lib/kubelet/pods/c08e9af6-fc25-4c32-bd25-ac82aebb828e/volumes" Oct 08 18:45:23 crc kubenswrapper[4988]: I1008 18:45:23.338496 4988 patch_prober.go:28] interesting pod/machine-config-daemon-nm54f container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 18:45:23 crc kubenswrapper[4988]: I1008 18:45:23.339202 4988 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 18:45:23 crc kubenswrapper[4988]: I1008 18:45:23.339276 4988 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" Oct 08 18:45:23 crc kubenswrapper[4988]: I1008 18:45:23.340326 4988 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"c9447394f2908e13ce0cc9fbb2e1c8c9953530f5820e7b144ac3da28722ce971"} pod="openshift-machine-config-operator/machine-config-daemon-nm54f" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 08 18:45:23 crc kubenswrapper[4988]: I1008 18:45:23.340459 4988 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" containerName="machine-config-daemon" containerID="cri-o://c9447394f2908e13ce0cc9fbb2e1c8c9953530f5820e7b144ac3da28722ce971" gracePeriod=600 Oct 08 18:45:23 crc kubenswrapper[4988]: I1008 18:45:23.517355 4988 generic.go:334] "Generic (PLEG): container finished" podID="f328a20b-5283-407e-8a93-2ab2da599f18" containerID="c9447394f2908e13ce0cc9fbb2e1c8c9953530f5820e7b144ac3da28722ce971" exitCode=0 Oct 08 18:45:23 crc kubenswrapper[4988]: I1008 18:45:23.517432 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" event={"ID":"f328a20b-5283-407e-8a93-2ab2da599f18","Type":"ContainerDied","Data":"c9447394f2908e13ce0cc9fbb2e1c8c9953530f5820e7b144ac3da28722ce971"} Oct 08 18:45:23 crc kubenswrapper[4988]: I1008 18:45:23.517512 4988 scope.go:117] "RemoveContainer" containerID="1e3c2ea55d3c49c9fa802ab01ed9324a6c941b8da678ffaccb80377e3f159aee" Oct 08 18:45:24 crc kubenswrapper[4988]: I1008 18:45:24.533641 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" event={"ID":"f328a20b-5283-407e-8a93-2ab2da599f18","Type":"ContainerStarted","Data":"2d7aa899cb7029f3e0b408a19acc6a33d067b6fecc0276f108f0b0335f638129"} Oct 08 18:46:01 crc kubenswrapper[4988]: I1008 18:46:01.325634 4988 scope.go:117] "RemoveContainer" containerID="313c57df12ab0a4da3188d7fc8368593246ef417dbcc4914e60937f26494b1b0" Oct 08 18:46:11 crc kubenswrapper[4988]: I1008 18:46:11.170824 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-g4454"] Oct 08 18:46:11 crc kubenswrapper[4988]: E1008 18:46:11.171807 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="58a39bcb-ac94-4290-b2cb-0bea4547a67c" containerName="collect-profiles" Oct 08 18:46:11 crc kubenswrapper[4988]: I1008 18:46:11.171831 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="58a39bcb-ac94-4290-b2cb-0bea4547a67c" containerName="collect-profiles" Oct 08 18:46:11 crc kubenswrapper[4988]: I1008 18:46:11.172112 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="58a39bcb-ac94-4290-b2cb-0bea4547a67c" containerName="collect-profiles" Oct 08 18:46:11 crc kubenswrapper[4988]: I1008 18:46:11.173903 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-g4454" Oct 08 18:46:11 crc kubenswrapper[4988]: I1008 18:46:11.188751 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-g4454"] Oct 08 18:46:11 crc kubenswrapper[4988]: I1008 18:46:11.327347 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2a473369-437d-4530-9bd7-4040c1cb7c94-catalog-content\") pod \"community-operators-g4454\" (UID: \"2a473369-437d-4530-9bd7-4040c1cb7c94\") " pod="openshift-marketplace/community-operators-g4454" Oct 08 18:46:11 crc kubenswrapper[4988]: I1008 18:46:11.327434 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wmhln\" (UniqueName: \"kubernetes.io/projected/2a473369-437d-4530-9bd7-4040c1cb7c94-kube-api-access-wmhln\") pod \"community-operators-g4454\" (UID: \"2a473369-437d-4530-9bd7-4040c1cb7c94\") " pod="openshift-marketplace/community-operators-g4454" Oct 08 18:46:11 crc kubenswrapper[4988]: I1008 18:46:11.327634 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2a473369-437d-4530-9bd7-4040c1cb7c94-utilities\") pod \"community-operators-g4454\" (UID: \"2a473369-437d-4530-9bd7-4040c1cb7c94\") " pod="openshift-marketplace/community-operators-g4454" Oct 08 18:46:11 crc kubenswrapper[4988]: I1008 18:46:11.429527 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2a473369-437d-4530-9bd7-4040c1cb7c94-utilities\") pod \"community-operators-g4454\" (UID: \"2a473369-437d-4530-9bd7-4040c1cb7c94\") " pod="openshift-marketplace/community-operators-g4454" Oct 08 18:46:11 crc kubenswrapper[4988]: I1008 18:46:11.429701 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2a473369-437d-4530-9bd7-4040c1cb7c94-catalog-content\") pod \"community-operators-g4454\" (UID: \"2a473369-437d-4530-9bd7-4040c1cb7c94\") " pod="openshift-marketplace/community-operators-g4454" Oct 08 18:46:11 crc kubenswrapper[4988]: I1008 18:46:11.429750 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wmhln\" (UniqueName: \"kubernetes.io/projected/2a473369-437d-4530-9bd7-4040c1cb7c94-kube-api-access-wmhln\") pod \"community-operators-g4454\" (UID: \"2a473369-437d-4530-9bd7-4040c1cb7c94\") " pod="openshift-marketplace/community-operators-g4454" Oct 08 18:46:11 crc kubenswrapper[4988]: I1008 18:46:11.430156 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2a473369-437d-4530-9bd7-4040c1cb7c94-utilities\") pod \"community-operators-g4454\" (UID: \"2a473369-437d-4530-9bd7-4040c1cb7c94\") " pod="openshift-marketplace/community-operators-g4454" Oct 08 18:46:11 crc kubenswrapper[4988]: I1008 18:46:11.430156 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2a473369-437d-4530-9bd7-4040c1cb7c94-catalog-content\") pod \"community-operators-g4454\" (UID: \"2a473369-437d-4530-9bd7-4040c1cb7c94\") " pod="openshift-marketplace/community-operators-g4454" Oct 08 18:46:11 crc kubenswrapper[4988]: I1008 18:46:11.449367 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wmhln\" (UniqueName: \"kubernetes.io/projected/2a473369-437d-4530-9bd7-4040c1cb7c94-kube-api-access-wmhln\") pod \"community-operators-g4454\" (UID: \"2a473369-437d-4530-9bd7-4040c1cb7c94\") " pod="openshift-marketplace/community-operators-g4454" Oct 08 18:46:11 crc kubenswrapper[4988]: I1008 18:46:11.510230 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-g4454" Oct 08 18:46:12 crc kubenswrapper[4988]: I1008 18:46:12.003771 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-g4454"] Oct 08 18:46:12 crc kubenswrapper[4988]: I1008 18:46:12.960534 4988 generic.go:334] "Generic (PLEG): container finished" podID="2a473369-437d-4530-9bd7-4040c1cb7c94" containerID="ddd2c060902f2e87207ddcf4589d63d4fdb5e57fd15e6007a0f03454234c8599" exitCode=0 Oct 08 18:46:12 crc kubenswrapper[4988]: I1008 18:46:12.960702 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-g4454" event={"ID":"2a473369-437d-4530-9bd7-4040c1cb7c94","Type":"ContainerDied","Data":"ddd2c060902f2e87207ddcf4589d63d4fdb5e57fd15e6007a0f03454234c8599"} Oct 08 18:46:12 crc kubenswrapper[4988]: I1008 18:46:12.963637 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-g4454" event={"ID":"2a473369-437d-4530-9bd7-4040c1cb7c94","Type":"ContainerStarted","Data":"3b04de587145096978ad1b1367035de4dcd917429510fc1ebabf197fdd45a1fb"} Oct 08 18:46:12 crc kubenswrapper[4988]: I1008 18:46:12.963937 4988 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 08 18:46:13 crc kubenswrapper[4988]: I1008 18:46:13.974523 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-g4454" event={"ID":"2a473369-437d-4530-9bd7-4040c1cb7c94","Type":"ContainerStarted","Data":"90be9f9698639c65e0730b63b709bbea229d93fbcac131e5be774c0f474cf33b"} Oct 08 18:46:14 crc kubenswrapper[4988]: I1008 18:46:14.985896 4988 generic.go:334] "Generic (PLEG): container finished" podID="2a473369-437d-4530-9bd7-4040c1cb7c94" containerID="90be9f9698639c65e0730b63b709bbea229d93fbcac131e5be774c0f474cf33b" exitCode=0 Oct 08 18:46:14 crc kubenswrapper[4988]: I1008 18:46:14.986204 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-g4454" event={"ID":"2a473369-437d-4530-9bd7-4040c1cb7c94","Type":"ContainerDied","Data":"90be9f9698639c65e0730b63b709bbea229d93fbcac131e5be774c0f474cf33b"} Oct 08 18:46:15 crc kubenswrapper[4988]: I1008 18:46:15.996304 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-g4454" event={"ID":"2a473369-437d-4530-9bd7-4040c1cb7c94","Type":"ContainerStarted","Data":"1b1bfb1d3a8f3136d0b01e27e9475169bb5c83c9d3e497b24c26296a1d52c852"} Oct 08 18:46:16 crc kubenswrapper[4988]: I1008 18:46:16.018738 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-g4454" podStartSLOduration=2.454336623 podStartE2EDuration="5.018713257s" podCreationTimestamp="2025-10-08 18:46:11 +0000 UTC" firstStartedPulling="2025-10-08 18:46:12.963617232 +0000 UTC m=+2118.413460012" lastFinishedPulling="2025-10-08 18:46:15.527993876 +0000 UTC m=+2120.977836646" observedRunningTime="2025-10-08 18:46:16.017198348 +0000 UTC m=+2121.467041118" watchObservedRunningTime="2025-10-08 18:46:16.018713257 +0000 UTC m=+2121.468556067" Oct 08 18:46:21 crc kubenswrapper[4988]: I1008 18:46:21.510669 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-g4454" Oct 08 18:46:21 crc kubenswrapper[4988]: I1008 18:46:21.511209 4988 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-g4454" Oct 08 18:46:21 crc kubenswrapper[4988]: I1008 18:46:21.569469 4988 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-g4454" Oct 08 18:46:22 crc kubenswrapper[4988]: I1008 18:46:22.138186 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-g4454" Oct 08 18:46:22 crc kubenswrapper[4988]: I1008 18:46:22.200089 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-g4454"] Oct 08 18:46:24 crc kubenswrapper[4988]: I1008 18:46:24.069900 4988 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-g4454" podUID="2a473369-437d-4530-9bd7-4040c1cb7c94" containerName="registry-server" containerID="cri-o://1b1bfb1d3a8f3136d0b01e27e9475169bb5c83c9d3e497b24c26296a1d52c852" gracePeriod=2 Oct 08 18:46:24 crc kubenswrapper[4988]: I1008 18:46:24.535219 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-g4454" Oct 08 18:46:24 crc kubenswrapper[4988]: I1008 18:46:24.633125 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2a473369-437d-4530-9bd7-4040c1cb7c94-catalog-content\") pod \"2a473369-437d-4530-9bd7-4040c1cb7c94\" (UID: \"2a473369-437d-4530-9bd7-4040c1cb7c94\") " Oct 08 18:46:24 crc kubenswrapper[4988]: I1008 18:46:24.633203 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2a473369-437d-4530-9bd7-4040c1cb7c94-utilities\") pod \"2a473369-437d-4530-9bd7-4040c1cb7c94\" (UID: \"2a473369-437d-4530-9bd7-4040c1cb7c94\") " Oct 08 18:46:24 crc kubenswrapper[4988]: I1008 18:46:24.633354 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wmhln\" (UniqueName: \"kubernetes.io/projected/2a473369-437d-4530-9bd7-4040c1cb7c94-kube-api-access-wmhln\") pod \"2a473369-437d-4530-9bd7-4040c1cb7c94\" (UID: \"2a473369-437d-4530-9bd7-4040c1cb7c94\") " Oct 08 18:46:24 crc kubenswrapper[4988]: I1008 18:46:24.634639 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2a473369-437d-4530-9bd7-4040c1cb7c94-utilities" (OuterVolumeSpecName: "utilities") pod "2a473369-437d-4530-9bd7-4040c1cb7c94" (UID: "2a473369-437d-4530-9bd7-4040c1cb7c94"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 18:46:24 crc kubenswrapper[4988]: I1008 18:46:24.642007 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2a473369-437d-4530-9bd7-4040c1cb7c94-kube-api-access-wmhln" (OuterVolumeSpecName: "kube-api-access-wmhln") pod "2a473369-437d-4530-9bd7-4040c1cb7c94" (UID: "2a473369-437d-4530-9bd7-4040c1cb7c94"). InnerVolumeSpecName "kube-api-access-wmhln". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:46:24 crc kubenswrapper[4988]: I1008 18:46:24.698413 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2a473369-437d-4530-9bd7-4040c1cb7c94-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "2a473369-437d-4530-9bd7-4040c1cb7c94" (UID: "2a473369-437d-4530-9bd7-4040c1cb7c94"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 18:46:24 crc kubenswrapper[4988]: I1008 18:46:24.735848 4988 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2a473369-437d-4530-9bd7-4040c1cb7c94-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 08 18:46:24 crc kubenswrapper[4988]: I1008 18:46:24.735896 4988 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2a473369-437d-4530-9bd7-4040c1cb7c94-utilities\") on node \"crc\" DevicePath \"\"" Oct 08 18:46:24 crc kubenswrapper[4988]: I1008 18:46:24.735916 4988 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wmhln\" (UniqueName: \"kubernetes.io/projected/2a473369-437d-4530-9bd7-4040c1cb7c94-kube-api-access-wmhln\") on node \"crc\" DevicePath \"\"" Oct 08 18:46:25 crc kubenswrapper[4988]: I1008 18:46:25.080423 4988 generic.go:334] "Generic (PLEG): container finished" podID="2a473369-437d-4530-9bd7-4040c1cb7c94" containerID="1b1bfb1d3a8f3136d0b01e27e9475169bb5c83c9d3e497b24c26296a1d52c852" exitCode=0 Oct 08 18:46:25 crc kubenswrapper[4988]: I1008 18:46:25.080470 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-g4454" event={"ID":"2a473369-437d-4530-9bd7-4040c1cb7c94","Type":"ContainerDied","Data":"1b1bfb1d3a8f3136d0b01e27e9475169bb5c83c9d3e497b24c26296a1d52c852"} Oct 08 18:46:25 crc kubenswrapper[4988]: I1008 18:46:25.080498 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-g4454" event={"ID":"2a473369-437d-4530-9bd7-4040c1cb7c94","Type":"ContainerDied","Data":"3b04de587145096978ad1b1367035de4dcd917429510fc1ebabf197fdd45a1fb"} Oct 08 18:46:25 crc kubenswrapper[4988]: I1008 18:46:25.080500 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-g4454" Oct 08 18:46:25 crc kubenswrapper[4988]: I1008 18:46:25.080518 4988 scope.go:117] "RemoveContainer" containerID="1b1bfb1d3a8f3136d0b01e27e9475169bb5c83c9d3e497b24c26296a1d52c852" Oct 08 18:46:25 crc kubenswrapper[4988]: I1008 18:46:25.119710 4988 scope.go:117] "RemoveContainer" containerID="90be9f9698639c65e0730b63b709bbea229d93fbcac131e5be774c0f474cf33b" Oct 08 18:46:25 crc kubenswrapper[4988]: I1008 18:46:25.125761 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-g4454"] Oct 08 18:46:25 crc kubenswrapper[4988]: I1008 18:46:25.132579 4988 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-g4454"] Oct 08 18:46:25 crc kubenswrapper[4988]: I1008 18:46:25.142871 4988 scope.go:117] "RemoveContainer" containerID="ddd2c060902f2e87207ddcf4589d63d4fdb5e57fd15e6007a0f03454234c8599" Oct 08 18:46:25 crc kubenswrapper[4988]: I1008 18:46:25.183298 4988 scope.go:117] "RemoveContainer" containerID="1b1bfb1d3a8f3136d0b01e27e9475169bb5c83c9d3e497b24c26296a1d52c852" Oct 08 18:46:25 crc kubenswrapper[4988]: E1008 18:46:25.183838 4988 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1b1bfb1d3a8f3136d0b01e27e9475169bb5c83c9d3e497b24c26296a1d52c852\": container with ID starting with 1b1bfb1d3a8f3136d0b01e27e9475169bb5c83c9d3e497b24c26296a1d52c852 not found: ID does not exist" containerID="1b1bfb1d3a8f3136d0b01e27e9475169bb5c83c9d3e497b24c26296a1d52c852" Oct 08 18:46:25 crc kubenswrapper[4988]: I1008 18:46:25.183972 4988 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1b1bfb1d3a8f3136d0b01e27e9475169bb5c83c9d3e497b24c26296a1d52c852"} err="failed to get container status \"1b1bfb1d3a8f3136d0b01e27e9475169bb5c83c9d3e497b24c26296a1d52c852\": rpc error: code = NotFound desc = could not find container \"1b1bfb1d3a8f3136d0b01e27e9475169bb5c83c9d3e497b24c26296a1d52c852\": container with ID starting with 1b1bfb1d3a8f3136d0b01e27e9475169bb5c83c9d3e497b24c26296a1d52c852 not found: ID does not exist" Oct 08 18:46:25 crc kubenswrapper[4988]: I1008 18:46:25.184014 4988 scope.go:117] "RemoveContainer" containerID="90be9f9698639c65e0730b63b709bbea229d93fbcac131e5be774c0f474cf33b" Oct 08 18:46:25 crc kubenswrapper[4988]: E1008 18:46:25.184551 4988 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"90be9f9698639c65e0730b63b709bbea229d93fbcac131e5be774c0f474cf33b\": container with ID starting with 90be9f9698639c65e0730b63b709bbea229d93fbcac131e5be774c0f474cf33b not found: ID does not exist" containerID="90be9f9698639c65e0730b63b709bbea229d93fbcac131e5be774c0f474cf33b" Oct 08 18:46:25 crc kubenswrapper[4988]: I1008 18:46:25.184589 4988 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"90be9f9698639c65e0730b63b709bbea229d93fbcac131e5be774c0f474cf33b"} err="failed to get container status \"90be9f9698639c65e0730b63b709bbea229d93fbcac131e5be774c0f474cf33b\": rpc error: code = NotFound desc = could not find container \"90be9f9698639c65e0730b63b709bbea229d93fbcac131e5be774c0f474cf33b\": container with ID starting with 90be9f9698639c65e0730b63b709bbea229d93fbcac131e5be774c0f474cf33b not found: ID does not exist" Oct 08 18:46:25 crc kubenswrapper[4988]: I1008 18:46:25.184620 4988 scope.go:117] "RemoveContainer" containerID="ddd2c060902f2e87207ddcf4589d63d4fdb5e57fd15e6007a0f03454234c8599" Oct 08 18:46:25 crc kubenswrapper[4988]: E1008 18:46:25.184924 4988 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ddd2c060902f2e87207ddcf4589d63d4fdb5e57fd15e6007a0f03454234c8599\": container with ID starting with ddd2c060902f2e87207ddcf4589d63d4fdb5e57fd15e6007a0f03454234c8599 not found: ID does not exist" containerID="ddd2c060902f2e87207ddcf4589d63d4fdb5e57fd15e6007a0f03454234c8599" Oct 08 18:46:25 crc kubenswrapper[4988]: I1008 18:46:25.184954 4988 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ddd2c060902f2e87207ddcf4589d63d4fdb5e57fd15e6007a0f03454234c8599"} err="failed to get container status \"ddd2c060902f2e87207ddcf4589d63d4fdb5e57fd15e6007a0f03454234c8599\": rpc error: code = NotFound desc = could not find container \"ddd2c060902f2e87207ddcf4589d63d4fdb5e57fd15e6007a0f03454234c8599\": container with ID starting with ddd2c060902f2e87207ddcf4589d63d4fdb5e57fd15e6007a0f03454234c8599 not found: ID does not exist" Oct 08 18:46:25 crc kubenswrapper[4988]: I1008 18:46:25.252587 4988 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2a473369-437d-4530-9bd7-4040c1cb7c94" path="/var/lib/kubelet/pods/2a473369-437d-4530-9bd7-4040c1cb7c94/volumes" Oct 08 18:46:58 crc kubenswrapper[4988]: I1008 18:46:58.554690 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-tgzts"] Oct 08 18:46:58 crc kubenswrapper[4988]: E1008 18:46:58.555751 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2a473369-437d-4530-9bd7-4040c1cb7c94" containerName="extract-utilities" Oct 08 18:46:58 crc kubenswrapper[4988]: I1008 18:46:58.555771 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="2a473369-437d-4530-9bd7-4040c1cb7c94" containerName="extract-utilities" Oct 08 18:46:58 crc kubenswrapper[4988]: E1008 18:46:58.555816 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2a473369-437d-4530-9bd7-4040c1cb7c94" containerName="registry-server" Oct 08 18:46:58 crc kubenswrapper[4988]: I1008 18:46:58.555828 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="2a473369-437d-4530-9bd7-4040c1cb7c94" containerName="registry-server" Oct 08 18:46:58 crc kubenswrapper[4988]: E1008 18:46:58.555852 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2a473369-437d-4530-9bd7-4040c1cb7c94" containerName="extract-content" Oct 08 18:46:58 crc kubenswrapper[4988]: I1008 18:46:58.555863 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="2a473369-437d-4530-9bd7-4040c1cb7c94" containerName="extract-content" Oct 08 18:46:58 crc kubenswrapper[4988]: I1008 18:46:58.556111 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="2a473369-437d-4530-9bd7-4040c1cb7c94" containerName="registry-server" Oct 08 18:46:58 crc kubenswrapper[4988]: I1008 18:46:58.557699 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-tgzts" Oct 08 18:46:58 crc kubenswrapper[4988]: I1008 18:46:58.572049 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-tgzts"] Oct 08 18:46:58 crc kubenswrapper[4988]: I1008 18:46:58.744997 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c4c49be9-e58e-4887-a207-33d6048f649f-catalog-content\") pod \"redhat-marketplace-tgzts\" (UID: \"c4c49be9-e58e-4887-a207-33d6048f649f\") " pod="openshift-marketplace/redhat-marketplace-tgzts" Oct 08 18:46:58 crc kubenswrapper[4988]: I1008 18:46:58.745050 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c4c49be9-e58e-4887-a207-33d6048f649f-utilities\") pod \"redhat-marketplace-tgzts\" (UID: \"c4c49be9-e58e-4887-a207-33d6048f649f\") " pod="openshift-marketplace/redhat-marketplace-tgzts" Oct 08 18:46:58 crc kubenswrapper[4988]: I1008 18:46:58.745095 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cxs8h\" (UniqueName: \"kubernetes.io/projected/c4c49be9-e58e-4887-a207-33d6048f649f-kube-api-access-cxs8h\") pod \"redhat-marketplace-tgzts\" (UID: \"c4c49be9-e58e-4887-a207-33d6048f649f\") " pod="openshift-marketplace/redhat-marketplace-tgzts" Oct 08 18:46:58 crc kubenswrapper[4988]: I1008 18:46:58.846234 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c4c49be9-e58e-4887-a207-33d6048f649f-catalog-content\") pod \"redhat-marketplace-tgzts\" (UID: \"c4c49be9-e58e-4887-a207-33d6048f649f\") " pod="openshift-marketplace/redhat-marketplace-tgzts" Oct 08 18:46:58 crc kubenswrapper[4988]: I1008 18:46:58.846285 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c4c49be9-e58e-4887-a207-33d6048f649f-utilities\") pod \"redhat-marketplace-tgzts\" (UID: \"c4c49be9-e58e-4887-a207-33d6048f649f\") " pod="openshift-marketplace/redhat-marketplace-tgzts" Oct 08 18:46:58 crc kubenswrapper[4988]: I1008 18:46:58.846318 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cxs8h\" (UniqueName: \"kubernetes.io/projected/c4c49be9-e58e-4887-a207-33d6048f649f-kube-api-access-cxs8h\") pod \"redhat-marketplace-tgzts\" (UID: \"c4c49be9-e58e-4887-a207-33d6048f649f\") " pod="openshift-marketplace/redhat-marketplace-tgzts" Oct 08 18:46:58 crc kubenswrapper[4988]: I1008 18:46:58.846796 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c4c49be9-e58e-4887-a207-33d6048f649f-catalog-content\") pod \"redhat-marketplace-tgzts\" (UID: \"c4c49be9-e58e-4887-a207-33d6048f649f\") " pod="openshift-marketplace/redhat-marketplace-tgzts" Oct 08 18:46:58 crc kubenswrapper[4988]: I1008 18:46:58.846892 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c4c49be9-e58e-4887-a207-33d6048f649f-utilities\") pod \"redhat-marketplace-tgzts\" (UID: \"c4c49be9-e58e-4887-a207-33d6048f649f\") " pod="openshift-marketplace/redhat-marketplace-tgzts" Oct 08 18:46:58 crc kubenswrapper[4988]: I1008 18:46:58.864467 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cxs8h\" (UniqueName: \"kubernetes.io/projected/c4c49be9-e58e-4887-a207-33d6048f649f-kube-api-access-cxs8h\") pod \"redhat-marketplace-tgzts\" (UID: \"c4c49be9-e58e-4887-a207-33d6048f649f\") " pod="openshift-marketplace/redhat-marketplace-tgzts" Oct 08 18:46:58 crc kubenswrapper[4988]: I1008 18:46:58.923712 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-tgzts" Oct 08 18:46:59 crc kubenswrapper[4988]: I1008 18:46:59.362788 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-tgzts"] Oct 08 18:46:59 crc kubenswrapper[4988]: I1008 18:46:59.389000 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-tgzts" event={"ID":"c4c49be9-e58e-4887-a207-33d6048f649f","Type":"ContainerStarted","Data":"239e43083799e3b9d90aca47d72a85924919dadc192a96e2d903c215dd1baa6a"} Oct 08 18:47:00 crc kubenswrapper[4988]: I1008 18:47:00.404172 4988 generic.go:334] "Generic (PLEG): container finished" podID="c4c49be9-e58e-4887-a207-33d6048f649f" containerID="09e9658c7646e271821eeaccfcfc30354b448be8dad7e92aa9c3067708501184" exitCode=0 Oct 08 18:47:00 crc kubenswrapper[4988]: I1008 18:47:00.404238 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-tgzts" event={"ID":"c4c49be9-e58e-4887-a207-33d6048f649f","Type":"ContainerDied","Data":"09e9658c7646e271821eeaccfcfc30354b448be8dad7e92aa9c3067708501184"} Oct 08 18:47:01 crc kubenswrapper[4988]: I1008 18:47:01.424486 4988 generic.go:334] "Generic (PLEG): container finished" podID="c4c49be9-e58e-4887-a207-33d6048f649f" containerID="9d8508afc6e8142817ab9b8ac7d80e65a60ad470f7615e118fb6243f808c0fe6" exitCode=0 Oct 08 18:47:01 crc kubenswrapper[4988]: I1008 18:47:01.424562 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-tgzts" event={"ID":"c4c49be9-e58e-4887-a207-33d6048f649f","Type":"ContainerDied","Data":"9d8508afc6e8142817ab9b8ac7d80e65a60ad470f7615e118fb6243f808c0fe6"} Oct 08 18:47:02 crc kubenswrapper[4988]: I1008 18:47:02.441180 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-tgzts" event={"ID":"c4c49be9-e58e-4887-a207-33d6048f649f","Type":"ContainerStarted","Data":"0aeb2ff5ef4537fa1343ba27591da7172ef3f3f88881fbf45540f8d6dc1f9a66"} Oct 08 18:47:02 crc kubenswrapper[4988]: I1008 18:47:02.484352 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-tgzts" podStartSLOduration=3.039972015 podStartE2EDuration="4.484332125s" podCreationTimestamp="2025-10-08 18:46:58 +0000 UTC" firstStartedPulling="2025-10-08 18:47:00.406550127 +0000 UTC m=+2165.856392947" lastFinishedPulling="2025-10-08 18:47:01.850910287 +0000 UTC m=+2167.300753057" observedRunningTime="2025-10-08 18:47:02.482753913 +0000 UTC m=+2167.932596743" watchObservedRunningTime="2025-10-08 18:47:02.484332125 +0000 UTC m=+2167.934174905" Oct 08 18:47:08 crc kubenswrapper[4988]: I1008 18:47:08.924285 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-tgzts" Oct 08 18:47:08 crc kubenswrapper[4988]: I1008 18:47:08.925078 4988 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-tgzts" Oct 08 18:47:08 crc kubenswrapper[4988]: I1008 18:47:08.987700 4988 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-tgzts" Oct 08 18:47:09 crc kubenswrapper[4988]: I1008 18:47:09.563542 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-tgzts" Oct 08 18:47:09 crc kubenswrapper[4988]: I1008 18:47:09.609781 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-tgzts"] Oct 08 18:47:11 crc kubenswrapper[4988]: I1008 18:47:11.520594 4988 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-tgzts" podUID="c4c49be9-e58e-4887-a207-33d6048f649f" containerName="registry-server" containerID="cri-o://0aeb2ff5ef4537fa1343ba27591da7172ef3f3f88881fbf45540f8d6dc1f9a66" gracePeriod=2 Oct 08 18:47:12 crc kubenswrapper[4988]: I1008 18:47:12.006871 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-tgzts" Oct 08 18:47:12 crc kubenswrapper[4988]: I1008 18:47:12.204299 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c4c49be9-e58e-4887-a207-33d6048f649f-catalog-content\") pod \"c4c49be9-e58e-4887-a207-33d6048f649f\" (UID: \"c4c49be9-e58e-4887-a207-33d6048f649f\") " Oct 08 18:47:12 crc kubenswrapper[4988]: I1008 18:47:12.204423 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c4c49be9-e58e-4887-a207-33d6048f649f-utilities\") pod \"c4c49be9-e58e-4887-a207-33d6048f649f\" (UID: \"c4c49be9-e58e-4887-a207-33d6048f649f\") " Oct 08 18:47:12 crc kubenswrapper[4988]: I1008 18:47:12.204468 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cxs8h\" (UniqueName: \"kubernetes.io/projected/c4c49be9-e58e-4887-a207-33d6048f649f-kube-api-access-cxs8h\") pod \"c4c49be9-e58e-4887-a207-33d6048f649f\" (UID: \"c4c49be9-e58e-4887-a207-33d6048f649f\") " Oct 08 18:47:12 crc kubenswrapper[4988]: I1008 18:47:12.207315 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c4c49be9-e58e-4887-a207-33d6048f649f-utilities" (OuterVolumeSpecName: "utilities") pod "c4c49be9-e58e-4887-a207-33d6048f649f" (UID: "c4c49be9-e58e-4887-a207-33d6048f649f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 18:47:12 crc kubenswrapper[4988]: I1008 18:47:12.223575 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c4c49be9-e58e-4887-a207-33d6048f649f-kube-api-access-cxs8h" (OuterVolumeSpecName: "kube-api-access-cxs8h") pod "c4c49be9-e58e-4887-a207-33d6048f649f" (UID: "c4c49be9-e58e-4887-a207-33d6048f649f"). InnerVolumeSpecName "kube-api-access-cxs8h". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:47:12 crc kubenswrapper[4988]: I1008 18:47:12.231345 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c4c49be9-e58e-4887-a207-33d6048f649f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "c4c49be9-e58e-4887-a207-33d6048f649f" (UID: "c4c49be9-e58e-4887-a207-33d6048f649f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 18:47:12 crc kubenswrapper[4988]: I1008 18:47:12.306597 4988 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c4c49be9-e58e-4887-a207-33d6048f649f-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 08 18:47:12 crc kubenswrapper[4988]: I1008 18:47:12.306653 4988 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c4c49be9-e58e-4887-a207-33d6048f649f-utilities\") on node \"crc\" DevicePath \"\"" Oct 08 18:47:12 crc kubenswrapper[4988]: I1008 18:47:12.306671 4988 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cxs8h\" (UniqueName: \"kubernetes.io/projected/c4c49be9-e58e-4887-a207-33d6048f649f-kube-api-access-cxs8h\") on node \"crc\" DevicePath \"\"" Oct 08 18:47:12 crc kubenswrapper[4988]: I1008 18:47:12.533937 4988 generic.go:334] "Generic (PLEG): container finished" podID="c4c49be9-e58e-4887-a207-33d6048f649f" containerID="0aeb2ff5ef4537fa1343ba27591da7172ef3f3f88881fbf45540f8d6dc1f9a66" exitCode=0 Oct 08 18:47:12 crc kubenswrapper[4988]: I1008 18:47:12.534005 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-tgzts" event={"ID":"c4c49be9-e58e-4887-a207-33d6048f649f","Type":"ContainerDied","Data":"0aeb2ff5ef4537fa1343ba27591da7172ef3f3f88881fbf45540f8d6dc1f9a66"} Oct 08 18:47:12 crc kubenswrapper[4988]: I1008 18:47:12.534051 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-tgzts" event={"ID":"c4c49be9-e58e-4887-a207-33d6048f649f","Type":"ContainerDied","Data":"239e43083799e3b9d90aca47d72a85924919dadc192a96e2d903c215dd1baa6a"} Oct 08 18:47:12 crc kubenswrapper[4988]: I1008 18:47:12.534079 4988 scope.go:117] "RemoveContainer" containerID="0aeb2ff5ef4537fa1343ba27591da7172ef3f3f88881fbf45540f8d6dc1f9a66" Oct 08 18:47:12 crc kubenswrapper[4988]: I1008 18:47:12.534086 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-tgzts" Oct 08 18:47:12 crc kubenswrapper[4988]: I1008 18:47:12.560267 4988 scope.go:117] "RemoveContainer" containerID="9d8508afc6e8142817ab9b8ac7d80e65a60ad470f7615e118fb6243f808c0fe6" Oct 08 18:47:12 crc kubenswrapper[4988]: I1008 18:47:12.589497 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-tgzts"] Oct 08 18:47:12 crc kubenswrapper[4988]: I1008 18:47:12.597226 4988 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-tgzts"] Oct 08 18:47:12 crc kubenswrapper[4988]: I1008 18:47:12.599082 4988 scope.go:117] "RemoveContainer" containerID="09e9658c7646e271821eeaccfcfc30354b448be8dad7e92aa9c3067708501184" Oct 08 18:47:12 crc kubenswrapper[4988]: I1008 18:47:12.641674 4988 scope.go:117] "RemoveContainer" containerID="0aeb2ff5ef4537fa1343ba27591da7172ef3f3f88881fbf45540f8d6dc1f9a66" Oct 08 18:47:12 crc kubenswrapper[4988]: E1008 18:47:12.643404 4988 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0aeb2ff5ef4537fa1343ba27591da7172ef3f3f88881fbf45540f8d6dc1f9a66\": container with ID starting with 0aeb2ff5ef4537fa1343ba27591da7172ef3f3f88881fbf45540f8d6dc1f9a66 not found: ID does not exist" containerID="0aeb2ff5ef4537fa1343ba27591da7172ef3f3f88881fbf45540f8d6dc1f9a66" Oct 08 18:47:12 crc kubenswrapper[4988]: I1008 18:47:12.643440 4988 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0aeb2ff5ef4537fa1343ba27591da7172ef3f3f88881fbf45540f8d6dc1f9a66"} err="failed to get container status \"0aeb2ff5ef4537fa1343ba27591da7172ef3f3f88881fbf45540f8d6dc1f9a66\": rpc error: code = NotFound desc = could not find container \"0aeb2ff5ef4537fa1343ba27591da7172ef3f3f88881fbf45540f8d6dc1f9a66\": container with ID starting with 0aeb2ff5ef4537fa1343ba27591da7172ef3f3f88881fbf45540f8d6dc1f9a66 not found: ID does not exist" Oct 08 18:47:12 crc kubenswrapper[4988]: I1008 18:47:12.643465 4988 scope.go:117] "RemoveContainer" containerID="9d8508afc6e8142817ab9b8ac7d80e65a60ad470f7615e118fb6243f808c0fe6" Oct 08 18:47:12 crc kubenswrapper[4988]: E1008 18:47:12.643718 4988 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9d8508afc6e8142817ab9b8ac7d80e65a60ad470f7615e118fb6243f808c0fe6\": container with ID starting with 9d8508afc6e8142817ab9b8ac7d80e65a60ad470f7615e118fb6243f808c0fe6 not found: ID does not exist" containerID="9d8508afc6e8142817ab9b8ac7d80e65a60ad470f7615e118fb6243f808c0fe6" Oct 08 18:47:12 crc kubenswrapper[4988]: I1008 18:47:12.643746 4988 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9d8508afc6e8142817ab9b8ac7d80e65a60ad470f7615e118fb6243f808c0fe6"} err="failed to get container status \"9d8508afc6e8142817ab9b8ac7d80e65a60ad470f7615e118fb6243f808c0fe6\": rpc error: code = NotFound desc = could not find container \"9d8508afc6e8142817ab9b8ac7d80e65a60ad470f7615e118fb6243f808c0fe6\": container with ID starting with 9d8508afc6e8142817ab9b8ac7d80e65a60ad470f7615e118fb6243f808c0fe6 not found: ID does not exist" Oct 08 18:47:12 crc kubenswrapper[4988]: I1008 18:47:12.643764 4988 scope.go:117] "RemoveContainer" containerID="09e9658c7646e271821eeaccfcfc30354b448be8dad7e92aa9c3067708501184" Oct 08 18:47:12 crc kubenswrapper[4988]: E1008 18:47:12.643982 4988 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"09e9658c7646e271821eeaccfcfc30354b448be8dad7e92aa9c3067708501184\": container with ID starting with 09e9658c7646e271821eeaccfcfc30354b448be8dad7e92aa9c3067708501184 not found: ID does not exist" containerID="09e9658c7646e271821eeaccfcfc30354b448be8dad7e92aa9c3067708501184" Oct 08 18:47:12 crc kubenswrapper[4988]: I1008 18:47:12.644007 4988 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"09e9658c7646e271821eeaccfcfc30354b448be8dad7e92aa9c3067708501184"} err="failed to get container status \"09e9658c7646e271821eeaccfcfc30354b448be8dad7e92aa9c3067708501184\": rpc error: code = NotFound desc = could not find container \"09e9658c7646e271821eeaccfcfc30354b448be8dad7e92aa9c3067708501184\": container with ID starting with 09e9658c7646e271821eeaccfcfc30354b448be8dad7e92aa9c3067708501184 not found: ID does not exist" Oct 08 18:47:13 crc kubenswrapper[4988]: I1008 18:47:13.250416 4988 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c4c49be9-e58e-4887-a207-33d6048f649f" path="/var/lib/kubelet/pods/c4c49be9-e58e-4887-a207-33d6048f649f/volumes" Oct 08 18:47:23 crc kubenswrapper[4988]: I1008 18:47:23.338276 4988 patch_prober.go:28] interesting pod/machine-config-daemon-nm54f container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 18:47:23 crc kubenswrapper[4988]: I1008 18:47:23.338902 4988 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 18:47:53 crc kubenswrapper[4988]: I1008 18:47:53.337682 4988 patch_prober.go:28] interesting pod/machine-config-daemon-nm54f container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 18:47:53 crc kubenswrapper[4988]: I1008 18:47:53.338340 4988 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 18:48:23 crc kubenswrapper[4988]: I1008 18:48:23.338434 4988 patch_prober.go:28] interesting pod/machine-config-daemon-nm54f container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 18:48:23 crc kubenswrapper[4988]: I1008 18:48:23.338967 4988 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 18:48:23 crc kubenswrapper[4988]: I1008 18:48:23.339014 4988 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" Oct 08 18:48:23 crc kubenswrapper[4988]: I1008 18:48:23.339643 4988 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"2d7aa899cb7029f3e0b408a19acc6a33d067b6fecc0276f108f0b0335f638129"} pod="openshift-machine-config-operator/machine-config-daemon-nm54f" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 08 18:48:23 crc kubenswrapper[4988]: I1008 18:48:23.339698 4988 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" containerName="machine-config-daemon" containerID="cri-o://2d7aa899cb7029f3e0b408a19acc6a33d067b6fecc0276f108f0b0335f638129" gracePeriod=600 Oct 08 18:48:23 crc kubenswrapper[4988]: E1008 18:48:23.464740 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nm54f_openshift-machine-config-operator(f328a20b-5283-407e-8a93-2ab2da599f18)\"" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" Oct 08 18:48:24 crc kubenswrapper[4988]: I1008 18:48:24.212957 4988 generic.go:334] "Generic (PLEG): container finished" podID="f328a20b-5283-407e-8a93-2ab2da599f18" containerID="2d7aa899cb7029f3e0b408a19acc6a33d067b6fecc0276f108f0b0335f638129" exitCode=0 Oct 08 18:48:24 crc kubenswrapper[4988]: I1008 18:48:24.213022 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" event={"ID":"f328a20b-5283-407e-8a93-2ab2da599f18","Type":"ContainerDied","Data":"2d7aa899cb7029f3e0b408a19acc6a33d067b6fecc0276f108f0b0335f638129"} Oct 08 18:48:24 crc kubenswrapper[4988]: I1008 18:48:24.213110 4988 scope.go:117] "RemoveContainer" containerID="c9447394f2908e13ce0cc9fbb2e1c8c9953530f5820e7b144ac3da28722ce971" Oct 08 18:48:24 crc kubenswrapper[4988]: I1008 18:48:24.213735 4988 scope.go:117] "RemoveContainer" containerID="2d7aa899cb7029f3e0b408a19acc6a33d067b6fecc0276f108f0b0335f638129" Oct 08 18:48:24 crc kubenswrapper[4988]: E1008 18:48:24.214057 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nm54f_openshift-machine-config-operator(f328a20b-5283-407e-8a93-2ab2da599f18)\"" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" Oct 08 18:48:39 crc kubenswrapper[4988]: I1008 18:48:39.237956 4988 scope.go:117] "RemoveContainer" containerID="2d7aa899cb7029f3e0b408a19acc6a33d067b6fecc0276f108f0b0335f638129" Oct 08 18:48:39 crc kubenswrapper[4988]: E1008 18:48:39.239171 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nm54f_openshift-machine-config-operator(f328a20b-5283-407e-8a93-2ab2da599f18)\"" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" Oct 08 18:48:50 crc kubenswrapper[4988]: I1008 18:48:50.238419 4988 scope.go:117] "RemoveContainer" containerID="2d7aa899cb7029f3e0b408a19acc6a33d067b6fecc0276f108f0b0335f638129" Oct 08 18:48:50 crc kubenswrapper[4988]: E1008 18:48:50.239413 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nm54f_openshift-machine-config-operator(f328a20b-5283-407e-8a93-2ab2da599f18)\"" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" Oct 08 18:49:05 crc kubenswrapper[4988]: I1008 18:49:05.242020 4988 scope.go:117] "RemoveContainer" containerID="2d7aa899cb7029f3e0b408a19acc6a33d067b6fecc0276f108f0b0335f638129" Oct 08 18:49:05 crc kubenswrapper[4988]: E1008 18:49:05.242777 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nm54f_openshift-machine-config-operator(f328a20b-5283-407e-8a93-2ab2da599f18)\"" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" Oct 08 18:49:19 crc kubenswrapper[4988]: I1008 18:49:19.239205 4988 scope.go:117] "RemoveContainer" containerID="2d7aa899cb7029f3e0b408a19acc6a33d067b6fecc0276f108f0b0335f638129" Oct 08 18:49:19 crc kubenswrapper[4988]: E1008 18:49:19.241300 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nm54f_openshift-machine-config-operator(f328a20b-5283-407e-8a93-2ab2da599f18)\"" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" Oct 08 18:49:33 crc kubenswrapper[4988]: I1008 18:49:33.238354 4988 scope.go:117] "RemoveContainer" containerID="2d7aa899cb7029f3e0b408a19acc6a33d067b6fecc0276f108f0b0335f638129" Oct 08 18:49:33 crc kubenswrapper[4988]: E1008 18:49:33.239184 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nm54f_openshift-machine-config-operator(f328a20b-5283-407e-8a93-2ab2da599f18)\"" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" Oct 08 18:49:48 crc kubenswrapper[4988]: I1008 18:49:48.238211 4988 scope.go:117] "RemoveContainer" containerID="2d7aa899cb7029f3e0b408a19acc6a33d067b6fecc0276f108f0b0335f638129" Oct 08 18:49:48 crc kubenswrapper[4988]: E1008 18:49:48.239183 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nm54f_openshift-machine-config-operator(f328a20b-5283-407e-8a93-2ab2da599f18)\"" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" Oct 08 18:50:01 crc kubenswrapper[4988]: I1008 18:50:01.237622 4988 scope.go:117] "RemoveContainer" containerID="2d7aa899cb7029f3e0b408a19acc6a33d067b6fecc0276f108f0b0335f638129" Oct 08 18:50:01 crc kubenswrapper[4988]: E1008 18:50:01.238483 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nm54f_openshift-machine-config-operator(f328a20b-5283-407e-8a93-2ab2da599f18)\"" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" Oct 08 18:50:14 crc kubenswrapper[4988]: I1008 18:50:14.238021 4988 scope.go:117] "RemoveContainer" containerID="2d7aa899cb7029f3e0b408a19acc6a33d067b6fecc0276f108f0b0335f638129" Oct 08 18:50:14 crc kubenswrapper[4988]: E1008 18:50:14.239120 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nm54f_openshift-machine-config-operator(f328a20b-5283-407e-8a93-2ab2da599f18)\"" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" Oct 08 18:50:29 crc kubenswrapper[4988]: I1008 18:50:29.247482 4988 scope.go:117] "RemoveContainer" containerID="2d7aa899cb7029f3e0b408a19acc6a33d067b6fecc0276f108f0b0335f638129" Oct 08 18:50:29 crc kubenswrapper[4988]: E1008 18:50:29.248882 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nm54f_openshift-machine-config-operator(f328a20b-5283-407e-8a93-2ab2da599f18)\"" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" Oct 08 18:50:41 crc kubenswrapper[4988]: I1008 18:50:41.238084 4988 scope.go:117] "RemoveContainer" containerID="2d7aa899cb7029f3e0b408a19acc6a33d067b6fecc0276f108f0b0335f638129" Oct 08 18:50:41 crc kubenswrapper[4988]: E1008 18:50:41.240047 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nm54f_openshift-machine-config-operator(f328a20b-5283-407e-8a93-2ab2da599f18)\"" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" Oct 08 18:50:55 crc kubenswrapper[4988]: I1008 18:50:55.248119 4988 scope.go:117] "RemoveContainer" containerID="2d7aa899cb7029f3e0b408a19acc6a33d067b6fecc0276f108f0b0335f638129" Oct 08 18:50:55 crc kubenswrapper[4988]: E1008 18:50:55.249122 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nm54f_openshift-machine-config-operator(f328a20b-5283-407e-8a93-2ab2da599f18)\"" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" Oct 08 18:51:09 crc kubenswrapper[4988]: I1008 18:51:09.238503 4988 scope.go:117] "RemoveContainer" containerID="2d7aa899cb7029f3e0b408a19acc6a33d067b6fecc0276f108f0b0335f638129" Oct 08 18:51:09 crc kubenswrapper[4988]: E1008 18:51:09.240971 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nm54f_openshift-machine-config-operator(f328a20b-5283-407e-8a93-2ab2da599f18)\"" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" Oct 08 18:51:24 crc kubenswrapper[4988]: I1008 18:51:24.238462 4988 scope.go:117] "RemoveContainer" containerID="2d7aa899cb7029f3e0b408a19acc6a33d067b6fecc0276f108f0b0335f638129" Oct 08 18:51:24 crc kubenswrapper[4988]: E1008 18:51:24.239470 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nm54f_openshift-machine-config-operator(f328a20b-5283-407e-8a93-2ab2da599f18)\"" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" Oct 08 18:51:33 crc kubenswrapper[4988]: I1008 18:51:33.085237 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-8jmkl"] Oct 08 18:51:33 crc kubenswrapper[4988]: E1008 18:51:33.087261 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c4c49be9-e58e-4887-a207-33d6048f649f" containerName="extract-content" Oct 08 18:51:33 crc kubenswrapper[4988]: I1008 18:51:33.087342 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="c4c49be9-e58e-4887-a207-33d6048f649f" containerName="extract-content" Oct 08 18:51:33 crc kubenswrapper[4988]: E1008 18:51:33.087441 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c4c49be9-e58e-4887-a207-33d6048f649f" containerName="registry-server" Oct 08 18:51:33 crc kubenswrapper[4988]: I1008 18:51:33.087510 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="c4c49be9-e58e-4887-a207-33d6048f649f" containerName="registry-server" Oct 08 18:51:33 crc kubenswrapper[4988]: E1008 18:51:33.087583 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c4c49be9-e58e-4887-a207-33d6048f649f" containerName="extract-utilities" Oct 08 18:51:33 crc kubenswrapper[4988]: I1008 18:51:33.087664 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="c4c49be9-e58e-4887-a207-33d6048f649f" containerName="extract-utilities" Oct 08 18:51:33 crc kubenswrapper[4988]: I1008 18:51:33.089092 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="c4c49be9-e58e-4887-a207-33d6048f649f" containerName="registry-server" Oct 08 18:51:33 crc kubenswrapper[4988]: I1008 18:51:33.090856 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-8jmkl" Oct 08 18:51:33 crc kubenswrapper[4988]: I1008 18:51:33.123190 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-8jmkl"] Oct 08 18:51:33 crc kubenswrapper[4988]: I1008 18:51:33.214484 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4e2b4013-5365-4a49-8cee-2771c0eb7518-utilities\") pod \"redhat-operators-8jmkl\" (UID: \"4e2b4013-5365-4a49-8cee-2771c0eb7518\") " pod="openshift-marketplace/redhat-operators-8jmkl" Oct 08 18:51:33 crc kubenswrapper[4988]: I1008 18:51:33.214575 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4e2b4013-5365-4a49-8cee-2771c0eb7518-catalog-content\") pod \"redhat-operators-8jmkl\" (UID: \"4e2b4013-5365-4a49-8cee-2771c0eb7518\") " pod="openshift-marketplace/redhat-operators-8jmkl" Oct 08 18:51:33 crc kubenswrapper[4988]: I1008 18:51:33.214602 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vldhs\" (UniqueName: \"kubernetes.io/projected/4e2b4013-5365-4a49-8cee-2771c0eb7518-kube-api-access-vldhs\") pod \"redhat-operators-8jmkl\" (UID: \"4e2b4013-5365-4a49-8cee-2771c0eb7518\") " pod="openshift-marketplace/redhat-operators-8jmkl" Oct 08 18:51:33 crc kubenswrapper[4988]: I1008 18:51:33.315697 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4e2b4013-5365-4a49-8cee-2771c0eb7518-utilities\") pod \"redhat-operators-8jmkl\" (UID: \"4e2b4013-5365-4a49-8cee-2771c0eb7518\") " pod="openshift-marketplace/redhat-operators-8jmkl" Oct 08 18:51:33 crc kubenswrapper[4988]: I1008 18:51:33.315783 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4e2b4013-5365-4a49-8cee-2771c0eb7518-catalog-content\") pod \"redhat-operators-8jmkl\" (UID: \"4e2b4013-5365-4a49-8cee-2771c0eb7518\") " pod="openshift-marketplace/redhat-operators-8jmkl" Oct 08 18:51:33 crc kubenswrapper[4988]: I1008 18:51:33.315825 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vldhs\" (UniqueName: \"kubernetes.io/projected/4e2b4013-5365-4a49-8cee-2771c0eb7518-kube-api-access-vldhs\") pod \"redhat-operators-8jmkl\" (UID: \"4e2b4013-5365-4a49-8cee-2771c0eb7518\") " pod="openshift-marketplace/redhat-operators-8jmkl" Oct 08 18:51:33 crc kubenswrapper[4988]: I1008 18:51:33.316214 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4e2b4013-5365-4a49-8cee-2771c0eb7518-utilities\") pod \"redhat-operators-8jmkl\" (UID: \"4e2b4013-5365-4a49-8cee-2771c0eb7518\") " pod="openshift-marketplace/redhat-operators-8jmkl" Oct 08 18:51:33 crc kubenswrapper[4988]: I1008 18:51:33.316263 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4e2b4013-5365-4a49-8cee-2771c0eb7518-catalog-content\") pod \"redhat-operators-8jmkl\" (UID: \"4e2b4013-5365-4a49-8cee-2771c0eb7518\") " pod="openshift-marketplace/redhat-operators-8jmkl" Oct 08 18:51:33 crc kubenswrapper[4988]: I1008 18:51:33.334004 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vldhs\" (UniqueName: \"kubernetes.io/projected/4e2b4013-5365-4a49-8cee-2771c0eb7518-kube-api-access-vldhs\") pod \"redhat-operators-8jmkl\" (UID: \"4e2b4013-5365-4a49-8cee-2771c0eb7518\") " pod="openshift-marketplace/redhat-operators-8jmkl" Oct 08 18:51:33 crc kubenswrapper[4988]: I1008 18:51:33.419204 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-8jmkl" Oct 08 18:51:33 crc kubenswrapper[4988]: I1008 18:51:33.898161 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-8jmkl"] Oct 08 18:51:33 crc kubenswrapper[4988]: W1008 18:51:33.901157 4988 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4e2b4013_5365_4a49_8cee_2771c0eb7518.slice/crio-1aaa02d5f7a9026f64598eae91d393203e68533069e302dc1f6c96cdb1f25b4a WatchSource:0}: Error finding container 1aaa02d5f7a9026f64598eae91d393203e68533069e302dc1f6c96cdb1f25b4a: Status 404 returned error can't find the container with id 1aaa02d5f7a9026f64598eae91d393203e68533069e302dc1f6c96cdb1f25b4a Oct 08 18:51:34 crc kubenswrapper[4988]: I1008 18:51:34.852771 4988 generic.go:334] "Generic (PLEG): container finished" podID="4e2b4013-5365-4a49-8cee-2771c0eb7518" containerID="a6facc038324f85c5daeaa0e1041936fb6924bb4bf18c8355283d8e3284c603c" exitCode=0 Oct 08 18:51:34 crc kubenswrapper[4988]: I1008 18:51:34.852863 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-8jmkl" event={"ID":"4e2b4013-5365-4a49-8cee-2771c0eb7518","Type":"ContainerDied","Data":"a6facc038324f85c5daeaa0e1041936fb6924bb4bf18c8355283d8e3284c603c"} Oct 08 18:51:34 crc kubenswrapper[4988]: I1008 18:51:34.853290 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-8jmkl" event={"ID":"4e2b4013-5365-4a49-8cee-2771c0eb7518","Type":"ContainerStarted","Data":"1aaa02d5f7a9026f64598eae91d393203e68533069e302dc1f6c96cdb1f25b4a"} Oct 08 18:51:34 crc kubenswrapper[4988]: I1008 18:51:34.856987 4988 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 08 18:51:36 crc kubenswrapper[4988]: I1008 18:51:36.238278 4988 scope.go:117] "RemoveContainer" containerID="2d7aa899cb7029f3e0b408a19acc6a33d067b6fecc0276f108f0b0335f638129" Oct 08 18:51:36 crc kubenswrapper[4988]: E1008 18:51:36.239016 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nm54f_openshift-machine-config-operator(f328a20b-5283-407e-8a93-2ab2da599f18)\"" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" Oct 08 18:51:36 crc kubenswrapper[4988]: I1008 18:51:36.873228 4988 generic.go:334] "Generic (PLEG): container finished" podID="4e2b4013-5365-4a49-8cee-2771c0eb7518" containerID="35efadab288d3a9e5de8b2bb8bfc93706b1337d4de0d59bb3941d2f37decf241" exitCode=0 Oct 08 18:51:36 crc kubenswrapper[4988]: I1008 18:51:36.873303 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-8jmkl" event={"ID":"4e2b4013-5365-4a49-8cee-2771c0eb7518","Type":"ContainerDied","Data":"35efadab288d3a9e5de8b2bb8bfc93706b1337d4de0d59bb3941d2f37decf241"} Oct 08 18:51:37 crc kubenswrapper[4988]: I1008 18:51:37.883765 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-8jmkl" event={"ID":"4e2b4013-5365-4a49-8cee-2771c0eb7518","Type":"ContainerStarted","Data":"8b9e831b9fbe883bce227a77fbcf516c02e66b101654f15bbaa52e71a143c481"} Oct 08 18:51:37 crc kubenswrapper[4988]: I1008 18:51:37.912416 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-8jmkl" podStartSLOduration=2.318396775 podStartE2EDuration="4.912379002s" podCreationTimestamp="2025-10-08 18:51:33 +0000 UTC" firstStartedPulling="2025-10-08 18:51:34.856430159 +0000 UTC m=+2440.306272969" lastFinishedPulling="2025-10-08 18:51:37.450412426 +0000 UTC m=+2442.900255196" observedRunningTime="2025-10-08 18:51:37.906286137 +0000 UTC m=+2443.356128947" watchObservedRunningTime="2025-10-08 18:51:37.912379002 +0000 UTC m=+2443.362221772" Oct 08 18:51:43 crc kubenswrapper[4988]: I1008 18:51:43.420375 4988 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-8jmkl" Oct 08 18:51:43 crc kubenswrapper[4988]: I1008 18:51:43.420847 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-8jmkl" Oct 08 18:51:43 crc kubenswrapper[4988]: I1008 18:51:43.487763 4988 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-8jmkl" Oct 08 18:51:44 crc kubenswrapper[4988]: I1008 18:51:44.014189 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-8jmkl" Oct 08 18:51:44 crc kubenswrapper[4988]: I1008 18:51:44.070021 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-8jmkl"] Oct 08 18:51:45 crc kubenswrapper[4988]: I1008 18:51:45.954018 4988 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-8jmkl" podUID="4e2b4013-5365-4a49-8cee-2771c0eb7518" containerName="registry-server" containerID="cri-o://8b9e831b9fbe883bce227a77fbcf516c02e66b101654f15bbaa52e71a143c481" gracePeriod=2 Oct 08 18:51:46 crc kubenswrapper[4988]: I1008 18:51:46.390888 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-8jmkl" Oct 08 18:51:46 crc kubenswrapper[4988]: I1008 18:51:46.534590 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4e2b4013-5365-4a49-8cee-2771c0eb7518-utilities\") pod \"4e2b4013-5365-4a49-8cee-2771c0eb7518\" (UID: \"4e2b4013-5365-4a49-8cee-2771c0eb7518\") " Oct 08 18:51:46 crc kubenswrapper[4988]: I1008 18:51:46.534635 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4e2b4013-5365-4a49-8cee-2771c0eb7518-catalog-content\") pod \"4e2b4013-5365-4a49-8cee-2771c0eb7518\" (UID: \"4e2b4013-5365-4a49-8cee-2771c0eb7518\") " Oct 08 18:51:46 crc kubenswrapper[4988]: I1008 18:51:46.534739 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vldhs\" (UniqueName: \"kubernetes.io/projected/4e2b4013-5365-4a49-8cee-2771c0eb7518-kube-api-access-vldhs\") pod \"4e2b4013-5365-4a49-8cee-2771c0eb7518\" (UID: \"4e2b4013-5365-4a49-8cee-2771c0eb7518\") " Oct 08 18:51:46 crc kubenswrapper[4988]: I1008 18:51:46.535627 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4e2b4013-5365-4a49-8cee-2771c0eb7518-utilities" (OuterVolumeSpecName: "utilities") pod "4e2b4013-5365-4a49-8cee-2771c0eb7518" (UID: "4e2b4013-5365-4a49-8cee-2771c0eb7518"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 18:51:46 crc kubenswrapper[4988]: I1008 18:51:46.536441 4988 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4e2b4013-5365-4a49-8cee-2771c0eb7518-utilities\") on node \"crc\" DevicePath \"\"" Oct 08 18:51:46 crc kubenswrapper[4988]: I1008 18:51:46.540250 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4e2b4013-5365-4a49-8cee-2771c0eb7518-kube-api-access-vldhs" (OuterVolumeSpecName: "kube-api-access-vldhs") pod "4e2b4013-5365-4a49-8cee-2771c0eb7518" (UID: "4e2b4013-5365-4a49-8cee-2771c0eb7518"). InnerVolumeSpecName "kube-api-access-vldhs". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:51:46 crc kubenswrapper[4988]: I1008 18:51:46.637002 4988 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vldhs\" (UniqueName: \"kubernetes.io/projected/4e2b4013-5365-4a49-8cee-2771c0eb7518-kube-api-access-vldhs\") on node \"crc\" DevicePath \"\"" Oct 08 18:51:46 crc kubenswrapper[4988]: I1008 18:51:46.969288 4988 generic.go:334] "Generic (PLEG): container finished" podID="4e2b4013-5365-4a49-8cee-2771c0eb7518" containerID="8b9e831b9fbe883bce227a77fbcf516c02e66b101654f15bbaa52e71a143c481" exitCode=0 Oct 08 18:51:46 crc kubenswrapper[4988]: I1008 18:51:46.969351 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-8jmkl" event={"ID":"4e2b4013-5365-4a49-8cee-2771c0eb7518","Type":"ContainerDied","Data":"8b9e831b9fbe883bce227a77fbcf516c02e66b101654f15bbaa52e71a143c481"} Oct 08 18:51:46 crc kubenswrapper[4988]: I1008 18:51:46.969438 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-8jmkl" Oct 08 18:51:46 crc kubenswrapper[4988]: I1008 18:51:46.969465 4988 scope.go:117] "RemoveContainer" containerID="8b9e831b9fbe883bce227a77fbcf516c02e66b101654f15bbaa52e71a143c481" Oct 08 18:51:46 crc kubenswrapper[4988]: I1008 18:51:46.969445 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-8jmkl" event={"ID":"4e2b4013-5365-4a49-8cee-2771c0eb7518","Type":"ContainerDied","Data":"1aaa02d5f7a9026f64598eae91d393203e68533069e302dc1f6c96cdb1f25b4a"} Oct 08 18:51:46 crc kubenswrapper[4988]: I1008 18:51:46.999880 4988 scope.go:117] "RemoveContainer" containerID="35efadab288d3a9e5de8b2bb8bfc93706b1337d4de0d59bb3941d2f37decf241" Oct 08 18:51:47 crc kubenswrapper[4988]: I1008 18:51:47.030731 4988 scope.go:117] "RemoveContainer" containerID="a6facc038324f85c5daeaa0e1041936fb6924bb4bf18c8355283d8e3284c603c" Oct 08 18:51:47 crc kubenswrapper[4988]: I1008 18:51:47.080668 4988 scope.go:117] "RemoveContainer" containerID="8b9e831b9fbe883bce227a77fbcf516c02e66b101654f15bbaa52e71a143c481" Oct 08 18:51:47 crc kubenswrapper[4988]: E1008 18:51:47.084020 4988 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8b9e831b9fbe883bce227a77fbcf516c02e66b101654f15bbaa52e71a143c481\": container with ID starting with 8b9e831b9fbe883bce227a77fbcf516c02e66b101654f15bbaa52e71a143c481 not found: ID does not exist" containerID="8b9e831b9fbe883bce227a77fbcf516c02e66b101654f15bbaa52e71a143c481" Oct 08 18:51:47 crc kubenswrapper[4988]: I1008 18:51:47.084068 4988 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8b9e831b9fbe883bce227a77fbcf516c02e66b101654f15bbaa52e71a143c481"} err="failed to get container status \"8b9e831b9fbe883bce227a77fbcf516c02e66b101654f15bbaa52e71a143c481\": rpc error: code = NotFound desc = could not find container \"8b9e831b9fbe883bce227a77fbcf516c02e66b101654f15bbaa52e71a143c481\": container with ID starting with 8b9e831b9fbe883bce227a77fbcf516c02e66b101654f15bbaa52e71a143c481 not found: ID does not exist" Oct 08 18:51:47 crc kubenswrapper[4988]: I1008 18:51:47.084100 4988 scope.go:117] "RemoveContainer" containerID="35efadab288d3a9e5de8b2bb8bfc93706b1337d4de0d59bb3941d2f37decf241" Oct 08 18:51:47 crc kubenswrapper[4988]: E1008 18:51:47.084740 4988 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"35efadab288d3a9e5de8b2bb8bfc93706b1337d4de0d59bb3941d2f37decf241\": container with ID starting with 35efadab288d3a9e5de8b2bb8bfc93706b1337d4de0d59bb3941d2f37decf241 not found: ID does not exist" containerID="35efadab288d3a9e5de8b2bb8bfc93706b1337d4de0d59bb3941d2f37decf241" Oct 08 18:51:47 crc kubenswrapper[4988]: I1008 18:51:47.084815 4988 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"35efadab288d3a9e5de8b2bb8bfc93706b1337d4de0d59bb3941d2f37decf241"} err="failed to get container status \"35efadab288d3a9e5de8b2bb8bfc93706b1337d4de0d59bb3941d2f37decf241\": rpc error: code = NotFound desc = could not find container \"35efadab288d3a9e5de8b2bb8bfc93706b1337d4de0d59bb3941d2f37decf241\": container with ID starting with 35efadab288d3a9e5de8b2bb8bfc93706b1337d4de0d59bb3941d2f37decf241 not found: ID does not exist" Oct 08 18:51:47 crc kubenswrapper[4988]: I1008 18:51:47.084865 4988 scope.go:117] "RemoveContainer" containerID="a6facc038324f85c5daeaa0e1041936fb6924bb4bf18c8355283d8e3284c603c" Oct 08 18:51:47 crc kubenswrapper[4988]: E1008 18:51:47.085555 4988 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a6facc038324f85c5daeaa0e1041936fb6924bb4bf18c8355283d8e3284c603c\": container with ID starting with a6facc038324f85c5daeaa0e1041936fb6924bb4bf18c8355283d8e3284c603c not found: ID does not exist" containerID="a6facc038324f85c5daeaa0e1041936fb6924bb4bf18c8355283d8e3284c603c" Oct 08 18:51:47 crc kubenswrapper[4988]: I1008 18:51:47.085594 4988 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a6facc038324f85c5daeaa0e1041936fb6924bb4bf18c8355283d8e3284c603c"} err="failed to get container status \"a6facc038324f85c5daeaa0e1041936fb6924bb4bf18c8355283d8e3284c603c\": rpc error: code = NotFound desc = could not find container \"a6facc038324f85c5daeaa0e1041936fb6924bb4bf18c8355283d8e3284c603c\": container with ID starting with a6facc038324f85c5daeaa0e1041936fb6924bb4bf18c8355283d8e3284c603c not found: ID does not exist" Oct 08 18:51:47 crc kubenswrapper[4988]: I1008 18:51:47.239102 4988 scope.go:117] "RemoveContainer" containerID="2d7aa899cb7029f3e0b408a19acc6a33d067b6fecc0276f108f0b0335f638129" Oct 08 18:51:47 crc kubenswrapper[4988]: E1008 18:51:47.240055 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nm54f_openshift-machine-config-operator(f328a20b-5283-407e-8a93-2ab2da599f18)\"" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" Oct 08 18:51:47 crc kubenswrapper[4988]: I1008 18:51:47.360399 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4e2b4013-5365-4a49-8cee-2771c0eb7518-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "4e2b4013-5365-4a49-8cee-2771c0eb7518" (UID: "4e2b4013-5365-4a49-8cee-2771c0eb7518"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 18:51:47 crc kubenswrapper[4988]: I1008 18:51:47.449519 4988 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4e2b4013-5365-4a49-8cee-2771c0eb7518-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 08 18:51:47 crc kubenswrapper[4988]: I1008 18:51:47.621495 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-8jmkl"] Oct 08 18:51:47 crc kubenswrapper[4988]: I1008 18:51:47.634257 4988 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-8jmkl"] Oct 08 18:51:49 crc kubenswrapper[4988]: I1008 18:51:49.276263 4988 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4e2b4013-5365-4a49-8cee-2771c0eb7518" path="/var/lib/kubelet/pods/4e2b4013-5365-4a49-8cee-2771c0eb7518/volumes" Oct 08 18:52:01 crc kubenswrapper[4988]: I1008 18:52:01.239511 4988 scope.go:117] "RemoveContainer" containerID="2d7aa899cb7029f3e0b408a19acc6a33d067b6fecc0276f108f0b0335f638129" Oct 08 18:52:01 crc kubenswrapper[4988]: E1008 18:52:01.240593 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nm54f_openshift-machine-config-operator(f328a20b-5283-407e-8a93-2ab2da599f18)\"" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" Oct 08 18:52:15 crc kubenswrapper[4988]: I1008 18:52:15.243517 4988 scope.go:117] "RemoveContainer" containerID="2d7aa899cb7029f3e0b408a19acc6a33d067b6fecc0276f108f0b0335f638129" Oct 08 18:52:15 crc kubenswrapper[4988]: E1008 18:52:15.244270 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nm54f_openshift-machine-config-operator(f328a20b-5283-407e-8a93-2ab2da599f18)\"" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" Oct 08 18:52:26 crc kubenswrapper[4988]: I1008 18:52:26.237898 4988 scope.go:117] "RemoveContainer" containerID="2d7aa899cb7029f3e0b408a19acc6a33d067b6fecc0276f108f0b0335f638129" Oct 08 18:52:26 crc kubenswrapper[4988]: E1008 18:52:26.238938 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nm54f_openshift-machine-config-operator(f328a20b-5283-407e-8a93-2ab2da599f18)\"" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" Oct 08 18:52:38 crc kubenswrapper[4988]: I1008 18:52:38.238933 4988 scope.go:117] "RemoveContainer" containerID="2d7aa899cb7029f3e0b408a19acc6a33d067b6fecc0276f108f0b0335f638129" Oct 08 18:52:38 crc kubenswrapper[4988]: E1008 18:52:38.239693 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nm54f_openshift-machine-config-operator(f328a20b-5283-407e-8a93-2ab2da599f18)\"" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" Oct 08 18:52:49 crc kubenswrapper[4988]: I1008 18:52:49.239078 4988 scope.go:117] "RemoveContainer" containerID="2d7aa899cb7029f3e0b408a19acc6a33d067b6fecc0276f108f0b0335f638129" Oct 08 18:52:49 crc kubenswrapper[4988]: E1008 18:52:49.239856 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nm54f_openshift-machine-config-operator(f328a20b-5283-407e-8a93-2ab2da599f18)\"" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" Oct 08 18:53:00 crc kubenswrapper[4988]: I1008 18:53:00.237941 4988 scope.go:117] "RemoveContainer" containerID="2d7aa899cb7029f3e0b408a19acc6a33d067b6fecc0276f108f0b0335f638129" Oct 08 18:53:00 crc kubenswrapper[4988]: E1008 18:53:00.238754 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nm54f_openshift-machine-config-operator(f328a20b-5283-407e-8a93-2ab2da599f18)\"" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" Oct 08 18:53:12 crc kubenswrapper[4988]: I1008 18:53:12.238147 4988 scope.go:117] "RemoveContainer" containerID="2d7aa899cb7029f3e0b408a19acc6a33d067b6fecc0276f108f0b0335f638129" Oct 08 18:53:12 crc kubenswrapper[4988]: E1008 18:53:12.238864 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nm54f_openshift-machine-config-operator(f328a20b-5283-407e-8a93-2ab2da599f18)\"" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" Oct 08 18:53:23 crc kubenswrapper[4988]: I1008 18:53:23.238548 4988 scope.go:117] "RemoveContainer" containerID="2d7aa899cb7029f3e0b408a19acc6a33d067b6fecc0276f108f0b0335f638129" Oct 08 18:53:23 crc kubenswrapper[4988]: E1008 18:53:23.239524 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nm54f_openshift-machine-config-operator(f328a20b-5283-407e-8a93-2ab2da599f18)\"" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" Oct 08 18:53:37 crc kubenswrapper[4988]: I1008 18:53:37.237537 4988 scope.go:117] "RemoveContainer" containerID="2d7aa899cb7029f3e0b408a19acc6a33d067b6fecc0276f108f0b0335f638129" Oct 08 18:53:37 crc kubenswrapper[4988]: I1008 18:53:37.945160 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" event={"ID":"f328a20b-5283-407e-8a93-2ab2da599f18","Type":"ContainerStarted","Data":"df9f51685b2cfae41c8322202dd543ae485c4f24e54c4b9380bef89dae14c52e"} Oct 08 18:54:43 crc kubenswrapper[4988]: I1008 18:54:43.439424 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-d7zdq"] Oct 08 18:54:43 crc kubenswrapper[4988]: E1008 18:54:43.440256 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4e2b4013-5365-4a49-8cee-2771c0eb7518" containerName="extract-utilities" Oct 08 18:54:43 crc kubenswrapper[4988]: I1008 18:54:43.440269 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="4e2b4013-5365-4a49-8cee-2771c0eb7518" containerName="extract-utilities" Oct 08 18:54:43 crc kubenswrapper[4988]: E1008 18:54:43.440290 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4e2b4013-5365-4a49-8cee-2771c0eb7518" containerName="registry-server" Oct 08 18:54:43 crc kubenswrapper[4988]: I1008 18:54:43.440296 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="4e2b4013-5365-4a49-8cee-2771c0eb7518" containerName="registry-server" Oct 08 18:54:43 crc kubenswrapper[4988]: E1008 18:54:43.440308 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4e2b4013-5365-4a49-8cee-2771c0eb7518" containerName="extract-content" Oct 08 18:54:43 crc kubenswrapper[4988]: I1008 18:54:43.440314 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="4e2b4013-5365-4a49-8cee-2771c0eb7518" containerName="extract-content" Oct 08 18:54:43 crc kubenswrapper[4988]: I1008 18:54:43.440465 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="4e2b4013-5365-4a49-8cee-2771c0eb7518" containerName="registry-server" Oct 08 18:54:43 crc kubenswrapper[4988]: I1008 18:54:43.441797 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-d7zdq" Oct 08 18:54:43 crc kubenswrapper[4988]: I1008 18:54:43.451889 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-d7zdq"] Oct 08 18:54:43 crc kubenswrapper[4988]: I1008 18:54:43.589517 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1a45a1cf-0219-4f7f-b2b1-99003dc3ce39-utilities\") pod \"certified-operators-d7zdq\" (UID: \"1a45a1cf-0219-4f7f-b2b1-99003dc3ce39\") " pod="openshift-marketplace/certified-operators-d7zdq" Oct 08 18:54:43 crc kubenswrapper[4988]: I1008 18:54:43.589802 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1a45a1cf-0219-4f7f-b2b1-99003dc3ce39-catalog-content\") pod \"certified-operators-d7zdq\" (UID: \"1a45a1cf-0219-4f7f-b2b1-99003dc3ce39\") " pod="openshift-marketplace/certified-operators-d7zdq" Oct 08 18:54:43 crc kubenswrapper[4988]: I1008 18:54:43.589820 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dv8nm\" (UniqueName: \"kubernetes.io/projected/1a45a1cf-0219-4f7f-b2b1-99003dc3ce39-kube-api-access-dv8nm\") pod \"certified-operators-d7zdq\" (UID: \"1a45a1cf-0219-4f7f-b2b1-99003dc3ce39\") " pod="openshift-marketplace/certified-operators-d7zdq" Oct 08 18:54:43 crc kubenswrapper[4988]: I1008 18:54:43.690946 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1a45a1cf-0219-4f7f-b2b1-99003dc3ce39-utilities\") pod \"certified-operators-d7zdq\" (UID: \"1a45a1cf-0219-4f7f-b2b1-99003dc3ce39\") " pod="openshift-marketplace/certified-operators-d7zdq" Oct 08 18:54:43 crc kubenswrapper[4988]: I1008 18:54:43.691003 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1a45a1cf-0219-4f7f-b2b1-99003dc3ce39-catalog-content\") pod \"certified-operators-d7zdq\" (UID: \"1a45a1cf-0219-4f7f-b2b1-99003dc3ce39\") " pod="openshift-marketplace/certified-operators-d7zdq" Oct 08 18:54:43 crc kubenswrapper[4988]: I1008 18:54:43.691026 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dv8nm\" (UniqueName: \"kubernetes.io/projected/1a45a1cf-0219-4f7f-b2b1-99003dc3ce39-kube-api-access-dv8nm\") pod \"certified-operators-d7zdq\" (UID: \"1a45a1cf-0219-4f7f-b2b1-99003dc3ce39\") " pod="openshift-marketplace/certified-operators-d7zdq" Oct 08 18:54:43 crc kubenswrapper[4988]: I1008 18:54:43.691827 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1a45a1cf-0219-4f7f-b2b1-99003dc3ce39-utilities\") pod \"certified-operators-d7zdq\" (UID: \"1a45a1cf-0219-4f7f-b2b1-99003dc3ce39\") " pod="openshift-marketplace/certified-operators-d7zdq" Oct 08 18:54:43 crc kubenswrapper[4988]: I1008 18:54:43.692010 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1a45a1cf-0219-4f7f-b2b1-99003dc3ce39-catalog-content\") pod \"certified-operators-d7zdq\" (UID: \"1a45a1cf-0219-4f7f-b2b1-99003dc3ce39\") " pod="openshift-marketplace/certified-operators-d7zdq" Oct 08 18:54:43 crc kubenswrapper[4988]: I1008 18:54:43.716459 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dv8nm\" (UniqueName: \"kubernetes.io/projected/1a45a1cf-0219-4f7f-b2b1-99003dc3ce39-kube-api-access-dv8nm\") pod \"certified-operators-d7zdq\" (UID: \"1a45a1cf-0219-4f7f-b2b1-99003dc3ce39\") " pod="openshift-marketplace/certified-operators-d7zdq" Oct 08 18:54:43 crc kubenswrapper[4988]: I1008 18:54:43.791592 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-d7zdq" Oct 08 18:54:44 crc kubenswrapper[4988]: I1008 18:54:44.280670 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-d7zdq"] Oct 08 18:54:44 crc kubenswrapper[4988]: I1008 18:54:44.522629 4988 generic.go:334] "Generic (PLEG): container finished" podID="1a45a1cf-0219-4f7f-b2b1-99003dc3ce39" containerID="f9b46a9c3325f4524f4079cc48a1c73a896a360c8631847163970cfab5c3c27d" exitCode=0 Oct 08 18:54:44 crc kubenswrapper[4988]: I1008 18:54:44.522683 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-d7zdq" event={"ID":"1a45a1cf-0219-4f7f-b2b1-99003dc3ce39","Type":"ContainerDied","Data":"f9b46a9c3325f4524f4079cc48a1c73a896a360c8631847163970cfab5c3c27d"} Oct 08 18:54:44 crc kubenswrapper[4988]: I1008 18:54:44.522987 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-d7zdq" event={"ID":"1a45a1cf-0219-4f7f-b2b1-99003dc3ce39","Type":"ContainerStarted","Data":"46c732294c2cf00a7a549ed2cd0d7663e0aee71acd556a464cf85bcdcfa99c39"} Oct 08 18:54:45 crc kubenswrapper[4988]: I1008 18:54:45.530958 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-d7zdq" event={"ID":"1a45a1cf-0219-4f7f-b2b1-99003dc3ce39","Type":"ContainerStarted","Data":"f17eb9e7523ad9837192c79f54417c2e602695cad2e29c6dcc5bbccf234ed36c"} Oct 08 18:54:46 crc kubenswrapper[4988]: I1008 18:54:46.538501 4988 generic.go:334] "Generic (PLEG): container finished" podID="1a45a1cf-0219-4f7f-b2b1-99003dc3ce39" containerID="f17eb9e7523ad9837192c79f54417c2e602695cad2e29c6dcc5bbccf234ed36c" exitCode=0 Oct 08 18:54:46 crc kubenswrapper[4988]: I1008 18:54:46.538547 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-d7zdq" event={"ID":"1a45a1cf-0219-4f7f-b2b1-99003dc3ce39","Type":"ContainerDied","Data":"f17eb9e7523ad9837192c79f54417c2e602695cad2e29c6dcc5bbccf234ed36c"} Oct 08 18:54:47 crc kubenswrapper[4988]: I1008 18:54:47.555840 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-d7zdq" event={"ID":"1a45a1cf-0219-4f7f-b2b1-99003dc3ce39","Type":"ContainerStarted","Data":"033ac97fed776f112001636b1386c46a241bf715604422b8a3f5520e4b55f505"} Oct 08 18:54:47 crc kubenswrapper[4988]: I1008 18:54:47.578835 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-d7zdq" podStartSLOduration=2.013511814 podStartE2EDuration="4.57881041s" podCreationTimestamp="2025-10-08 18:54:43 +0000 UTC" firstStartedPulling="2025-10-08 18:54:44.525426052 +0000 UTC m=+2629.975268822" lastFinishedPulling="2025-10-08 18:54:47.090724648 +0000 UTC m=+2632.540567418" observedRunningTime="2025-10-08 18:54:47.572772047 +0000 UTC m=+2633.022614877" watchObservedRunningTime="2025-10-08 18:54:47.57881041 +0000 UTC m=+2633.028653220" Oct 08 18:54:53 crc kubenswrapper[4988]: I1008 18:54:53.792886 4988 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-d7zdq" Oct 08 18:54:53 crc kubenswrapper[4988]: I1008 18:54:53.793423 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-d7zdq" Oct 08 18:54:53 crc kubenswrapper[4988]: I1008 18:54:53.845230 4988 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-d7zdq" Oct 08 18:54:54 crc kubenswrapper[4988]: I1008 18:54:54.676665 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-d7zdq" Oct 08 18:54:55 crc kubenswrapper[4988]: I1008 18:54:55.609290 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-d7zdq"] Oct 08 18:54:56 crc kubenswrapper[4988]: I1008 18:54:56.620853 4988 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-d7zdq" podUID="1a45a1cf-0219-4f7f-b2b1-99003dc3ce39" containerName="registry-server" containerID="cri-o://033ac97fed776f112001636b1386c46a241bf715604422b8a3f5520e4b55f505" gracePeriod=2 Oct 08 18:54:57 crc kubenswrapper[4988]: I1008 18:54:57.585931 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-d7zdq" Oct 08 18:54:57 crc kubenswrapper[4988]: I1008 18:54:57.628505 4988 generic.go:334] "Generic (PLEG): container finished" podID="1a45a1cf-0219-4f7f-b2b1-99003dc3ce39" containerID="033ac97fed776f112001636b1386c46a241bf715604422b8a3f5520e4b55f505" exitCode=0 Oct 08 18:54:57 crc kubenswrapper[4988]: I1008 18:54:57.628551 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-d7zdq" event={"ID":"1a45a1cf-0219-4f7f-b2b1-99003dc3ce39","Type":"ContainerDied","Data":"033ac97fed776f112001636b1386c46a241bf715604422b8a3f5520e4b55f505"} Oct 08 18:54:57 crc kubenswrapper[4988]: I1008 18:54:57.628582 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-d7zdq" Oct 08 18:54:57 crc kubenswrapper[4988]: I1008 18:54:57.628692 4988 scope.go:117] "RemoveContainer" containerID="033ac97fed776f112001636b1386c46a241bf715604422b8a3f5520e4b55f505" Oct 08 18:54:57 crc kubenswrapper[4988]: I1008 18:54:57.628601 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-d7zdq" event={"ID":"1a45a1cf-0219-4f7f-b2b1-99003dc3ce39","Type":"ContainerDied","Data":"46c732294c2cf00a7a549ed2cd0d7663e0aee71acd556a464cf85bcdcfa99c39"} Oct 08 18:54:57 crc kubenswrapper[4988]: I1008 18:54:57.646720 4988 scope.go:117] "RemoveContainer" containerID="f17eb9e7523ad9837192c79f54417c2e602695cad2e29c6dcc5bbccf234ed36c" Oct 08 18:54:57 crc kubenswrapper[4988]: I1008 18:54:57.671617 4988 scope.go:117] "RemoveContainer" containerID="f9b46a9c3325f4524f4079cc48a1c73a896a360c8631847163970cfab5c3c27d" Oct 08 18:54:57 crc kubenswrapper[4988]: I1008 18:54:57.690712 4988 scope.go:117] "RemoveContainer" containerID="033ac97fed776f112001636b1386c46a241bf715604422b8a3f5520e4b55f505" Oct 08 18:54:57 crc kubenswrapper[4988]: E1008 18:54:57.691214 4988 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"033ac97fed776f112001636b1386c46a241bf715604422b8a3f5520e4b55f505\": container with ID starting with 033ac97fed776f112001636b1386c46a241bf715604422b8a3f5520e4b55f505 not found: ID does not exist" containerID="033ac97fed776f112001636b1386c46a241bf715604422b8a3f5520e4b55f505" Oct 08 18:54:57 crc kubenswrapper[4988]: I1008 18:54:57.691261 4988 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"033ac97fed776f112001636b1386c46a241bf715604422b8a3f5520e4b55f505"} err="failed to get container status \"033ac97fed776f112001636b1386c46a241bf715604422b8a3f5520e4b55f505\": rpc error: code = NotFound desc = could not find container \"033ac97fed776f112001636b1386c46a241bf715604422b8a3f5520e4b55f505\": container with ID starting with 033ac97fed776f112001636b1386c46a241bf715604422b8a3f5520e4b55f505 not found: ID does not exist" Oct 08 18:54:57 crc kubenswrapper[4988]: I1008 18:54:57.691289 4988 scope.go:117] "RemoveContainer" containerID="f17eb9e7523ad9837192c79f54417c2e602695cad2e29c6dcc5bbccf234ed36c" Oct 08 18:54:57 crc kubenswrapper[4988]: E1008 18:54:57.691718 4988 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f17eb9e7523ad9837192c79f54417c2e602695cad2e29c6dcc5bbccf234ed36c\": container with ID starting with f17eb9e7523ad9837192c79f54417c2e602695cad2e29c6dcc5bbccf234ed36c not found: ID does not exist" containerID="f17eb9e7523ad9837192c79f54417c2e602695cad2e29c6dcc5bbccf234ed36c" Oct 08 18:54:57 crc kubenswrapper[4988]: I1008 18:54:57.691759 4988 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f17eb9e7523ad9837192c79f54417c2e602695cad2e29c6dcc5bbccf234ed36c"} err="failed to get container status \"f17eb9e7523ad9837192c79f54417c2e602695cad2e29c6dcc5bbccf234ed36c\": rpc error: code = NotFound desc = could not find container \"f17eb9e7523ad9837192c79f54417c2e602695cad2e29c6dcc5bbccf234ed36c\": container with ID starting with f17eb9e7523ad9837192c79f54417c2e602695cad2e29c6dcc5bbccf234ed36c not found: ID does not exist" Oct 08 18:54:57 crc kubenswrapper[4988]: I1008 18:54:57.691787 4988 scope.go:117] "RemoveContainer" containerID="f9b46a9c3325f4524f4079cc48a1c73a896a360c8631847163970cfab5c3c27d" Oct 08 18:54:57 crc kubenswrapper[4988]: E1008 18:54:57.692189 4988 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f9b46a9c3325f4524f4079cc48a1c73a896a360c8631847163970cfab5c3c27d\": container with ID starting with f9b46a9c3325f4524f4079cc48a1c73a896a360c8631847163970cfab5c3c27d not found: ID does not exist" containerID="f9b46a9c3325f4524f4079cc48a1c73a896a360c8631847163970cfab5c3c27d" Oct 08 18:54:57 crc kubenswrapper[4988]: I1008 18:54:57.692224 4988 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f9b46a9c3325f4524f4079cc48a1c73a896a360c8631847163970cfab5c3c27d"} err="failed to get container status \"f9b46a9c3325f4524f4079cc48a1c73a896a360c8631847163970cfab5c3c27d\": rpc error: code = NotFound desc = could not find container \"f9b46a9c3325f4524f4079cc48a1c73a896a360c8631847163970cfab5c3c27d\": container with ID starting with f9b46a9c3325f4524f4079cc48a1c73a896a360c8631847163970cfab5c3c27d not found: ID does not exist" Oct 08 18:54:57 crc kubenswrapper[4988]: I1008 18:54:57.694784 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dv8nm\" (UniqueName: \"kubernetes.io/projected/1a45a1cf-0219-4f7f-b2b1-99003dc3ce39-kube-api-access-dv8nm\") pod \"1a45a1cf-0219-4f7f-b2b1-99003dc3ce39\" (UID: \"1a45a1cf-0219-4f7f-b2b1-99003dc3ce39\") " Oct 08 18:54:57 crc kubenswrapper[4988]: I1008 18:54:57.694838 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1a45a1cf-0219-4f7f-b2b1-99003dc3ce39-catalog-content\") pod \"1a45a1cf-0219-4f7f-b2b1-99003dc3ce39\" (UID: \"1a45a1cf-0219-4f7f-b2b1-99003dc3ce39\") " Oct 08 18:54:57 crc kubenswrapper[4988]: I1008 18:54:57.694866 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1a45a1cf-0219-4f7f-b2b1-99003dc3ce39-utilities\") pod \"1a45a1cf-0219-4f7f-b2b1-99003dc3ce39\" (UID: \"1a45a1cf-0219-4f7f-b2b1-99003dc3ce39\") " Oct 08 18:54:57 crc kubenswrapper[4988]: I1008 18:54:57.696335 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1a45a1cf-0219-4f7f-b2b1-99003dc3ce39-utilities" (OuterVolumeSpecName: "utilities") pod "1a45a1cf-0219-4f7f-b2b1-99003dc3ce39" (UID: "1a45a1cf-0219-4f7f-b2b1-99003dc3ce39"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 18:54:57 crc kubenswrapper[4988]: I1008 18:54:57.703661 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1a45a1cf-0219-4f7f-b2b1-99003dc3ce39-kube-api-access-dv8nm" (OuterVolumeSpecName: "kube-api-access-dv8nm") pod "1a45a1cf-0219-4f7f-b2b1-99003dc3ce39" (UID: "1a45a1cf-0219-4f7f-b2b1-99003dc3ce39"). InnerVolumeSpecName "kube-api-access-dv8nm". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:54:57 crc kubenswrapper[4988]: I1008 18:54:57.747962 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1a45a1cf-0219-4f7f-b2b1-99003dc3ce39-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1a45a1cf-0219-4f7f-b2b1-99003dc3ce39" (UID: "1a45a1cf-0219-4f7f-b2b1-99003dc3ce39"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 18:54:57 crc kubenswrapper[4988]: I1008 18:54:57.796305 4988 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dv8nm\" (UniqueName: \"kubernetes.io/projected/1a45a1cf-0219-4f7f-b2b1-99003dc3ce39-kube-api-access-dv8nm\") on node \"crc\" DevicePath \"\"" Oct 08 18:54:57 crc kubenswrapper[4988]: I1008 18:54:57.796349 4988 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1a45a1cf-0219-4f7f-b2b1-99003dc3ce39-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 08 18:54:57 crc kubenswrapper[4988]: I1008 18:54:57.796362 4988 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1a45a1cf-0219-4f7f-b2b1-99003dc3ce39-utilities\") on node \"crc\" DevicePath \"\"" Oct 08 18:54:57 crc kubenswrapper[4988]: I1008 18:54:57.960099 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-d7zdq"] Oct 08 18:54:57 crc kubenswrapper[4988]: I1008 18:54:57.968945 4988 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-d7zdq"] Oct 08 18:54:59 crc kubenswrapper[4988]: I1008 18:54:59.245552 4988 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1a45a1cf-0219-4f7f-b2b1-99003dc3ce39" path="/var/lib/kubelet/pods/1a45a1cf-0219-4f7f-b2b1-99003dc3ce39/volumes" Oct 08 18:55:53 crc kubenswrapper[4988]: I1008 18:55:53.338213 4988 patch_prober.go:28] interesting pod/machine-config-daemon-nm54f container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 18:55:53 crc kubenswrapper[4988]: I1008 18:55:53.339044 4988 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 18:56:17 crc kubenswrapper[4988]: I1008 18:56:17.475431 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-gvv7b"] Oct 08 18:56:17 crc kubenswrapper[4988]: E1008 18:56:17.478661 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1a45a1cf-0219-4f7f-b2b1-99003dc3ce39" containerName="extract-content" Oct 08 18:56:17 crc kubenswrapper[4988]: I1008 18:56:17.478865 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="1a45a1cf-0219-4f7f-b2b1-99003dc3ce39" containerName="extract-content" Oct 08 18:56:17 crc kubenswrapper[4988]: E1008 18:56:17.479031 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1a45a1cf-0219-4f7f-b2b1-99003dc3ce39" containerName="registry-server" Oct 08 18:56:17 crc kubenswrapper[4988]: I1008 18:56:17.479176 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="1a45a1cf-0219-4f7f-b2b1-99003dc3ce39" containerName="registry-server" Oct 08 18:56:17 crc kubenswrapper[4988]: E1008 18:56:17.479361 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1a45a1cf-0219-4f7f-b2b1-99003dc3ce39" containerName="extract-utilities" Oct 08 18:56:17 crc kubenswrapper[4988]: I1008 18:56:17.479576 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="1a45a1cf-0219-4f7f-b2b1-99003dc3ce39" containerName="extract-utilities" Oct 08 18:56:17 crc kubenswrapper[4988]: I1008 18:56:17.479977 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="1a45a1cf-0219-4f7f-b2b1-99003dc3ce39" containerName="registry-server" Oct 08 18:56:17 crc kubenswrapper[4988]: I1008 18:56:17.482544 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-gvv7b" Oct 08 18:56:17 crc kubenswrapper[4988]: I1008 18:56:17.494406 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-gvv7b"] Oct 08 18:56:17 crc kubenswrapper[4988]: I1008 18:56:17.581946 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f002b931-d433-4323-ae50-f51c7dde5170-utilities\") pod \"community-operators-gvv7b\" (UID: \"f002b931-d433-4323-ae50-f51c7dde5170\") " pod="openshift-marketplace/community-operators-gvv7b" Oct 08 18:56:17 crc kubenswrapper[4988]: I1008 18:56:17.582081 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-26x9p\" (UniqueName: \"kubernetes.io/projected/f002b931-d433-4323-ae50-f51c7dde5170-kube-api-access-26x9p\") pod \"community-operators-gvv7b\" (UID: \"f002b931-d433-4323-ae50-f51c7dde5170\") " pod="openshift-marketplace/community-operators-gvv7b" Oct 08 18:56:17 crc kubenswrapper[4988]: I1008 18:56:17.582124 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f002b931-d433-4323-ae50-f51c7dde5170-catalog-content\") pod \"community-operators-gvv7b\" (UID: \"f002b931-d433-4323-ae50-f51c7dde5170\") " pod="openshift-marketplace/community-operators-gvv7b" Oct 08 18:56:17 crc kubenswrapper[4988]: I1008 18:56:17.683157 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f002b931-d433-4323-ae50-f51c7dde5170-catalog-content\") pod \"community-operators-gvv7b\" (UID: \"f002b931-d433-4323-ae50-f51c7dde5170\") " pod="openshift-marketplace/community-operators-gvv7b" Oct 08 18:56:17 crc kubenswrapper[4988]: I1008 18:56:17.683290 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f002b931-d433-4323-ae50-f51c7dde5170-utilities\") pod \"community-operators-gvv7b\" (UID: \"f002b931-d433-4323-ae50-f51c7dde5170\") " pod="openshift-marketplace/community-operators-gvv7b" Oct 08 18:56:17 crc kubenswrapper[4988]: I1008 18:56:17.683742 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f002b931-d433-4323-ae50-f51c7dde5170-catalog-content\") pod \"community-operators-gvv7b\" (UID: \"f002b931-d433-4323-ae50-f51c7dde5170\") " pod="openshift-marketplace/community-operators-gvv7b" Oct 08 18:56:17 crc kubenswrapper[4988]: I1008 18:56:17.683768 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f002b931-d433-4323-ae50-f51c7dde5170-utilities\") pod \"community-operators-gvv7b\" (UID: \"f002b931-d433-4323-ae50-f51c7dde5170\") " pod="openshift-marketplace/community-operators-gvv7b" Oct 08 18:56:17 crc kubenswrapper[4988]: I1008 18:56:17.683922 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-26x9p\" (UniqueName: \"kubernetes.io/projected/f002b931-d433-4323-ae50-f51c7dde5170-kube-api-access-26x9p\") pod \"community-operators-gvv7b\" (UID: \"f002b931-d433-4323-ae50-f51c7dde5170\") " pod="openshift-marketplace/community-operators-gvv7b" Oct 08 18:56:17 crc kubenswrapper[4988]: I1008 18:56:17.713686 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-26x9p\" (UniqueName: \"kubernetes.io/projected/f002b931-d433-4323-ae50-f51c7dde5170-kube-api-access-26x9p\") pod \"community-operators-gvv7b\" (UID: \"f002b931-d433-4323-ae50-f51c7dde5170\") " pod="openshift-marketplace/community-operators-gvv7b" Oct 08 18:56:17 crc kubenswrapper[4988]: I1008 18:56:17.810139 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-gvv7b" Oct 08 18:56:18 crc kubenswrapper[4988]: I1008 18:56:18.364984 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-gvv7b"] Oct 08 18:56:19 crc kubenswrapper[4988]: I1008 18:56:19.282335 4988 generic.go:334] "Generic (PLEG): container finished" podID="f002b931-d433-4323-ae50-f51c7dde5170" containerID="9c95df213fd1a013780661956c68e45c0badd55f6f7d57cc570d951f07b73ce6" exitCode=0 Oct 08 18:56:19 crc kubenswrapper[4988]: I1008 18:56:19.282438 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gvv7b" event={"ID":"f002b931-d433-4323-ae50-f51c7dde5170","Type":"ContainerDied","Data":"9c95df213fd1a013780661956c68e45c0badd55f6f7d57cc570d951f07b73ce6"} Oct 08 18:56:19 crc kubenswrapper[4988]: I1008 18:56:19.282695 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gvv7b" event={"ID":"f002b931-d433-4323-ae50-f51c7dde5170","Type":"ContainerStarted","Data":"123a415126c46d59500b806e6ab90ffd8d6a0eadbdf8d9ee7ea9fdab5ef3f4a9"} Oct 08 18:56:20 crc kubenswrapper[4988]: I1008 18:56:20.293048 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gvv7b" event={"ID":"f002b931-d433-4323-ae50-f51c7dde5170","Type":"ContainerStarted","Data":"020516d076a7bd8a4fdd2383efbaff26fa799620ccba2970dfa8b145ab5943ab"} Oct 08 18:56:21 crc kubenswrapper[4988]: I1008 18:56:21.305941 4988 generic.go:334] "Generic (PLEG): container finished" podID="f002b931-d433-4323-ae50-f51c7dde5170" containerID="020516d076a7bd8a4fdd2383efbaff26fa799620ccba2970dfa8b145ab5943ab" exitCode=0 Oct 08 18:56:21 crc kubenswrapper[4988]: I1008 18:56:21.306003 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gvv7b" event={"ID":"f002b931-d433-4323-ae50-f51c7dde5170","Type":"ContainerDied","Data":"020516d076a7bd8a4fdd2383efbaff26fa799620ccba2970dfa8b145ab5943ab"} Oct 08 18:56:22 crc kubenswrapper[4988]: I1008 18:56:22.316275 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gvv7b" event={"ID":"f002b931-d433-4323-ae50-f51c7dde5170","Type":"ContainerStarted","Data":"27e0525bec7cc0108aea60e3e67671d33916ea528a46e45064717c48ed27bbd8"} Oct 08 18:56:22 crc kubenswrapper[4988]: I1008 18:56:22.336471 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-gvv7b" podStartSLOduration=2.674842997 podStartE2EDuration="5.33645182s" podCreationTimestamp="2025-10-08 18:56:17 +0000 UTC" firstStartedPulling="2025-10-08 18:56:19.283544326 +0000 UTC m=+2724.733387096" lastFinishedPulling="2025-10-08 18:56:21.945153109 +0000 UTC m=+2727.394995919" observedRunningTime="2025-10-08 18:56:22.330311683 +0000 UTC m=+2727.780154473" watchObservedRunningTime="2025-10-08 18:56:22.33645182 +0000 UTC m=+2727.786294600" Oct 08 18:56:23 crc kubenswrapper[4988]: I1008 18:56:23.337705 4988 patch_prober.go:28] interesting pod/machine-config-daemon-nm54f container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 18:56:23 crc kubenswrapper[4988]: I1008 18:56:23.337777 4988 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 18:56:27 crc kubenswrapper[4988]: I1008 18:56:27.811369 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-gvv7b" Oct 08 18:56:27 crc kubenswrapper[4988]: I1008 18:56:27.811942 4988 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-gvv7b" Oct 08 18:56:27 crc kubenswrapper[4988]: I1008 18:56:27.884844 4988 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-gvv7b" Oct 08 18:56:28 crc kubenswrapper[4988]: I1008 18:56:28.435354 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-gvv7b" Oct 08 18:56:28 crc kubenswrapper[4988]: I1008 18:56:28.481027 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-gvv7b"] Oct 08 18:56:30 crc kubenswrapper[4988]: I1008 18:56:30.379544 4988 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-gvv7b" podUID="f002b931-d433-4323-ae50-f51c7dde5170" containerName="registry-server" containerID="cri-o://27e0525bec7cc0108aea60e3e67671d33916ea528a46e45064717c48ed27bbd8" gracePeriod=2 Oct 08 18:56:30 crc kubenswrapper[4988]: I1008 18:56:30.880926 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-gvv7b" Oct 08 18:56:30 crc kubenswrapper[4988]: I1008 18:56:30.997672 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f002b931-d433-4323-ae50-f51c7dde5170-catalog-content\") pod \"f002b931-d433-4323-ae50-f51c7dde5170\" (UID: \"f002b931-d433-4323-ae50-f51c7dde5170\") " Oct 08 18:56:30 crc kubenswrapper[4988]: I1008 18:56:30.997719 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-26x9p\" (UniqueName: \"kubernetes.io/projected/f002b931-d433-4323-ae50-f51c7dde5170-kube-api-access-26x9p\") pod \"f002b931-d433-4323-ae50-f51c7dde5170\" (UID: \"f002b931-d433-4323-ae50-f51c7dde5170\") " Oct 08 18:56:30 crc kubenswrapper[4988]: I1008 18:56:30.997791 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f002b931-d433-4323-ae50-f51c7dde5170-utilities\") pod \"f002b931-d433-4323-ae50-f51c7dde5170\" (UID: \"f002b931-d433-4323-ae50-f51c7dde5170\") " Oct 08 18:56:30 crc kubenswrapper[4988]: I1008 18:56:30.998760 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f002b931-d433-4323-ae50-f51c7dde5170-utilities" (OuterVolumeSpecName: "utilities") pod "f002b931-d433-4323-ae50-f51c7dde5170" (UID: "f002b931-d433-4323-ae50-f51c7dde5170"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 18:56:31 crc kubenswrapper[4988]: I1008 18:56:31.004269 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f002b931-d433-4323-ae50-f51c7dde5170-kube-api-access-26x9p" (OuterVolumeSpecName: "kube-api-access-26x9p") pod "f002b931-d433-4323-ae50-f51c7dde5170" (UID: "f002b931-d433-4323-ae50-f51c7dde5170"). InnerVolumeSpecName "kube-api-access-26x9p". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:56:31 crc kubenswrapper[4988]: I1008 18:56:31.088792 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f002b931-d433-4323-ae50-f51c7dde5170-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "f002b931-d433-4323-ae50-f51c7dde5170" (UID: "f002b931-d433-4323-ae50-f51c7dde5170"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 18:56:31 crc kubenswrapper[4988]: I1008 18:56:31.099770 4988 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f002b931-d433-4323-ae50-f51c7dde5170-utilities\") on node \"crc\" DevicePath \"\"" Oct 08 18:56:31 crc kubenswrapper[4988]: I1008 18:56:31.099833 4988 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f002b931-d433-4323-ae50-f51c7dde5170-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 08 18:56:31 crc kubenswrapper[4988]: I1008 18:56:31.099860 4988 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-26x9p\" (UniqueName: \"kubernetes.io/projected/f002b931-d433-4323-ae50-f51c7dde5170-kube-api-access-26x9p\") on node \"crc\" DevicePath \"\"" Oct 08 18:56:31 crc kubenswrapper[4988]: I1008 18:56:31.392633 4988 generic.go:334] "Generic (PLEG): container finished" podID="f002b931-d433-4323-ae50-f51c7dde5170" containerID="27e0525bec7cc0108aea60e3e67671d33916ea528a46e45064717c48ed27bbd8" exitCode=0 Oct 08 18:56:31 crc kubenswrapper[4988]: I1008 18:56:31.392716 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gvv7b" event={"ID":"f002b931-d433-4323-ae50-f51c7dde5170","Type":"ContainerDied","Data":"27e0525bec7cc0108aea60e3e67671d33916ea528a46e45064717c48ed27bbd8"} Oct 08 18:56:31 crc kubenswrapper[4988]: I1008 18:56:31.392748 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-gvv7b" Oct 08 18:56:31 crc kubenswrapper[4988]: I1008 18:56:31.393059 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gvv7b" event={"ID":"f002b931-d433-4323-ae50-f51c7dde5170","Type":"ContainerDied","Data":"123a415126c46d59500b806e6ab90ffd8d6a0eadbdf8d9ee7ea9fdab5ef3f4a9"} Oct 08 18:56:31 crc kubenswrapper[4988]: I1008 18:56:31.393105 4988 scope.go:117] "RemoveContainer" containerID="27e0525bec7cc0108aea60e3e67671d33916ea528a46e45064717c48ed27bbd8" Oct 08 18:56:31 crc kubenswrapper[4988]: I1008 18:56:31.419850 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-gvv7b"] Oct 08 18:56:31 crc kubenswrapper[4988]: I1008 18:56:31.427006 4988 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-gvv7b"] Oct 08 18:56:31 crc kubenswrapper[4988]: I1008 18:56:31.431542 4988 scope.go:117] "RemoveContainer" containerID="020516d076a7bd8a4fdd2383efbaff26fa799620ccba2970dfa8b145ab5943ab" Oct 08 18:56:31 crc kubenswrapper[4988]: I1008 18:56:31.451022 4988 scope.go:117] "RemoveContainer" containerID="9c95df213fd1a013780661956c68e45c0badd55f6f7d57cc570d951f07b73ce6" Oct 08 18:56:31 crc kubenswrapper[4988]: I1008 18:56:31.475567 4988 scope.go:117] "RemoveContainer" containerID="27e0525bec7cc0108aea60e3e67671d33916ea528a46e45064717c48ed27bbd8" Oct 08 18:56:31 crc kubenswrapper[4988]: E1008 18:56:31.476072 4988 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"27e0525bec7cc0108aea60e3e67671d33916ea528a46e45064717c48ed27bbd8\": container with ID starting with 27e0525bec7cc0108aea60e3e67671d33916ea528a46e45064717c48ed27bbd8 not found: ID does not exist" containerID="27e0525bec7cc0108aea60e3e67671d33916ea528a46e45064717c48ed27bbd8" Oct 08 18:56:31 crc kubenswrapper[4988]: I1008 18:56:31.476114 4988 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"27e0525bec7cc0108aea60e3e67671d33916ea528a46e45064717c48ed27bbd8"} err="failed to get container status \"27e0525bec7cc0108aea60e3e67671d33916ea528a46e45064717c48ed27bbd8\": rpc error: code = NotFound desc = could not find container \"27e0525bec7cc0108aea60e3e67671d33916ea528a46e45064717c48ed27bbd8\": container with ID starting with 27e0525bec7cc0108aea60e3e67671d33916ea528a46e45064717c48ed27bbd8 not found: ID does not exist" Oct 08 18:56:31 crc kubenswrapper[4988]: I1008 18:56:31.476144 4988 scope.go:117] "RemoveContainer" containerID="020516d076a7bd8a4fdd2383efbaff26fa799620ccba2970dfa8b145ab5943ab" Oct 08 18:56:31 crc kubenswrapper[4988]: E1008 18:56:31.476601 4988 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"020516d076a7bd8a4fdd2383efbaff26fa799620ccba2970dfa8b145ab5943ab\": container with ID starting with 020516d076a7bd8a4fdd2383efbaff26fa799620ccba2970dfa8b145ab5943ab not found: ID does not exist" containerID="020516d076a7bd8a4fdd2383efbaff26fa799620ccba2970dfa8b145ab5943ab" Oct 08 18:56:31 crc kubenswrapper[4988]: I1008 18:56:31.476669 4988 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"020516d076a7bd8a4fdd2383efbaff26fa799620ccba2970dfa8b145ab5943ab"} err="failed to get container status \"020516d076a7bd8a4fdd2383efbaff26fa799620ccba2970dfa8b145ab5943ab\": rpc error: code = NotFound desc = could not find container \"020516d076a7bd8a4fdd2383efbaff26fa799620ccba2970dfa8b145ab5943ab\": container with ID starting with 020516d076a7bd8a4fdd2383efbaff26fa799620ccba2970dfa8b145ab5943ab not found: ID does not exist" Oct 08 18:56:31 crc kubenswrapper[4988]: I1008 18:56:31.476713 4988 scope.go:117] "RemoveContainer" containerID="9c95df213fd1a013780661956c68e45c0badd55f6f7d57cc570d951f07b73ce6" Oct 08 18:56:31 crc kubenswrapper[4988]: E1008 18:56:31.477039 4988 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9c95df213fd1a013780661956c68e45c0badd55f6f7d57cc570d951f07b73ce6\": container with ID starting with 9c95df213fd1a013780661956c68e45c0badd55f6f7d57cc570d951f07b73ce6 not found: ID does not exist" containerID="9c95df213fd1a013780661956c68e45c0badd55f6f7d57cc570d951f07b73ce6" Oct 08 18:56:31 crc kubenswrapper[4988]: I1008 18:56:31.477083 4988 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9c95df213fd1a013780661956c68e45c0badd55f6f7d57cc570d951f07b73ce6"} err="failed to get container status \"9c95df213fd1a013780661956c68e45c0badd55f6f7d57cc570d951f07b73ce6\": rpc error: code = NotFound desc = could not find container \"9c95df213fd1a013780661956c68e45c0badd55f6f7d57cc570d951f07b73ce6\": container with ID starting with 9c95df213fd1a013780661956c68e45c0badd55f6f7d57cc570d951f07b73ce6 not found: ID does not exist" Oct 08 18:56:33 crc kubenswrapper[4988]: I1008 18:56:33.249447 4988 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f002b931-d433-4323-ae50-f51c7dde5170" path="/var/lib/kubelet/pods/f002b931-d433-4323-ae50-f51c7dde5170/volumes" Oct 08 18:56:53 crc kubenswrapper[4988]: I1008 18:56:53.338118 4988 patch_prober.go:28] interesting pod/machine-config-daemon-nm54f container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 18:56:53 crc kubenswrapper[4988]: I1008 18:56:53.338859 4988 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 18:56:53 crc kubenswrapper[4988]: I1008 18:56:53.338916 4988 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" Oct 08 18:56:53 crc kubenswrapper[4988]: I1008 18:56:53.339703 4988 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"df9f51685b2cfae41c8322202dd543ae485c4f24e54c4b9380bef89dae14c52e"} pod="openshift-machine-config-operator/machine-config-daemon-nm54f" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 08 18:56:53 crc kubenswrapper[4988]: I1008 18:56:53.339801 4988 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" containerName="machine-config-daemon" containerID="cri-o://df9f51685b2cfae41c8322202dd543ae485c4f24e54c4b9380bef89dae14c52e" gracePeriod=600 Oct 08 18:56:53 crc kubenswrapper[4988]: I1008 18:56:53.595108 4988 generic.go:334] "Generic (PLEG): container finished" podID="f328a20b-5283-407e-8a93-2ab2da599f18" containerID="df9f51685b2cfae41c8322202dd543ae485c4f24e54c4b9380bef89dae14c52e" exitCode=0 Oct 08 18:56:53 crc kubenswrapper[4988]: I1008 18:56:53.595254 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" event={"ID":"f328a20b-5283-407e-8a93-2ab2da599f18","Type":"ContainerDied","Data":"df9f51685b2cfae41c8322202dd543ae485c4f24e54c4b9380bef89dae14c52e"} Oct 08 18:56:53 crc kubenswrapper[4988]: I1008 18:56:53.595799 4988 scope.go:117] "RemoveContainer" containerID="2d7aa899cb7029f3e0b408a19acc6a33d067b6fecc0276f108f0b0335f638129" Oct 08 18:56:54 crc kubenswrapper[4988]: I1008 18:56:54.607900 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" event={"ID":"f328a20b-5283-407e-8a93-2ab2da599f18","Type":"ContainerStarted","Data":"fea33312e34e76e64c798028f94adde7baabd667aa354082b2aebdb7e50779da"} Oct 08 18:58:05 crc kubenswrapper[4988]: I1008 18:58:05.020771 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-7ffjt"] Oct 08 18:58:05 crc kubenswrapper[4988]: E1008 18:58:05.021477 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f002b931-d433-4323-ae50-f51c7dde5170" containerName="extract-utilities" Oct 08 18:58:05 crc kubenswrapper[4988]: I1008 18:58:05.021489 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="f002b931-d433-4323-ae50-f51c7dde5170" containerName="extract-utilities" Oct 08 18:58:05 crc kubenswrapper[4988]: E1008 18:58:05.021505 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f002b931-d433-4323-ae50-f51c7dde5170" containerName="extract-content" Oct 08 18:58:05 crc kubenswrapper[4988]: I1008 18:58:05.021510 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="f002b931-d433-4323-ae50-f51c7dde5170" containerName="extract-content" Oct 08 18:58:05 crc kubenswrapper[4988]: E1008 18:58:05.021531 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f002b931-d433-4323-ae50-f51c7dde5170" containerName="registry-server" Oct 08 18:58:05 crc kubenswrapper[4988]: I1008 18:58:05.021537 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="f002b931-d433-4323-ae50-f51c7dde5170" containerName="registry-server" Oct 08 18:58:05 crc kubenswrapper[4988]: I1008 18:58:05.021669 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="f002b931-d433-4323-ae50-f51c7dde5170" containerName="registry-server" Oct 08 18:58:05 crc kubenswrapper[4988]: I1008 18:58:05.029318 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-7ffjt" Oct 08 18:58:05 crc kubenswrapper[4988]: I1008 18:58:05.036119 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-7ffjt"] Oct 08 18:58:05 crc kubenswrapper[4988]: I1008 18:58:05.081494 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cv9rx\" (UniqueName: \"kubernetes.io/projected/8fe66bb3-6940-4535-b1a9-f6e09074b9b4-kube-api-access-cv9rx\") pod \"redhat-marketplace-7ffjt\" (UID: \"8fe66bb3-6940-4535-b1a9-f6e09074b9b4\") " pod="openshift-marketplace/redhat-marketplace-7ffjt" Oct 08 18:58:05 crc kubenswrapper[4988]: I1008 18:58:05.081564 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8fe66bb3-6940-4535-b1a9-f6e09074b9b4-catalog-content\") pod \"redhat-marketplace-7ffjt\" (UID: \"8fe66bb3-6940-4535-b1a9-f6e09074b9b4\") " pod="openshift-marketplace/redhat-marketplace-7ffjt" Oct 08 18:58:05 crc kubenswrapper[4988]: I1008 18:58:05.081594 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8fe66bb3-6940-4535-b1a9-f6e09074b9b4-utilities\") pod \"redhat-marketplace-7ffjt\" (UID: \"8fe66bb3-6940-4535-b1a9-f6e09074b9b4\") " pod="openshift-marketplace/redhat-marketplace-7ffjt" Oct 08 18:58:05 crc kubenswrapper[4988]: I1008 18:58:05.182426 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cv9rx\" (UniqueName: \"kubernetes.io/projected/8fe66bb3-6940-4535-b1a9-f6e09074b9b4-kube-api-access-cv9rx\") pod \"redhat-marketplace-7ffjt\" (UID: \"8fe66bb3-6940-4535-b1a9-f6e09074b9b4\") " pod="openshift-marketplace/redhat-marketplace-7ffjt" Oct 08 18:58:05 crc kubenswrapper[4988]: I1008 18:58:05.182480 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8fe66bb3-6940-4535-b1a9-f6e09074b9b4-catalog-content\") pod \"redhat-marketplace-7ffjt\" (UID: \"8fe66bb3-6940-4535-b1a9-f6e09074b9b4\") " pod="openshift-marketplace/redhat-marketplace-7ffjt" Oct 08 18:58:05 crc kubenswrapper[4988]: I1008 18:58:05.182498 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8fe66bb3-6940-4535-b1a9-f6e09074b9b4-utilities\") pod \"redhat-marketplace-7ffjt\" (UID: \"8fe66bb3-6940-4535-b1a9-f6e09074b9b4\") " pod="openshift-marketplace/redhat-marketplace-7ffjt" Oct 08 18:58:05 crc kubenswrapper[4988]: I1008 18:58:05.182908 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8fe66bb3-6940-4535-b1a9-f6e09074b9b4-catalog-content\") pod \"redhat-marketplace-7ffjt\" (UID: \"8fe66bb3-6940-4535-b1a9-f6e09074b9b4\") " pod="openshift-marketplace/redhat-marketplace-7ffjt" Oct 08 18:58:05 crc kubenswrapper[4988]: I1008 18:58:05.182941 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8fe66bb3-6940-4535-b1a9-f6e09074b9b4-utilities\") pod \"redhat-marketplace-7ffjt\" (UID: \"8fe66bb3-6940-4535-b1a9-f6e09074b9b4\") " pod="openshift-marketplace/redhat-marketplace-7ffjt" Oct 08 18:58:05 crc kubenswrapper[4988]: I1008 18:58:05.205487 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cv9rx\" (UniqueName: \"kubernetes.io/projected/8fe66bb3-6940-4535-b1a9-f6e09074b9b4-kube-api-access-cv9rx\") pod \"redhat-marketplace-7ffjt\" (UID: \"8fe66bb3-6940-4535-b1a9-f6e09074b9b4\") " pod="openshift-marketplace/redhat-marketplace-7ffjt" Oct 08 18:58:05 crc kubenswrapper[4988]: I1008 18:58:05.349510 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-7ffjt" Oct 08 18:58:05 crc kubenswrapper[4988]: I1008 18:58:05.767821 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-7ffjt"] Oct 08 18:58:06 crc kubenswrapper[4988]: I1008 18:58:06.140278 4988 generic.go:334] "Generic (PLEG): container finished" podID="8fe66bb3-6940-4535-b1a9-f6e09074b9b4" containerID="5016e99fb3491854a8ffb5899b1274a9ec58ee4613d8359bd3ae34f8503f8aa7" exitCode=0 Oct 08 18:58:06 crc kubenswrapper[4988]: I1008 18:58:06.140446 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-7ffjt" event={"ID":"8fe66bb3-6940-4535-b1a9-f6e09074b9b4","Type":"ContainerDied","Data":"5016e99fb3491854a8ffb5899b1274a9ec58ee4613d8359bd3ae34f8503f8aa7"} Oct 08 18:58:06 crc kubenswrapper[4988]: I1008 18:58:06.140701 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-7ffjt" event={"ID":"8fe66bb3-6940-4535-b1a9-f6e09074b9b4","Type":"ContainerStarted","Data":"1fc3c75c5072bced1041592b875bea40118b1f0fd1a5c9e4052f221460581103"} Oct 08 18:58:06 crc kubenswrapper[4988]: I1008 18:58:06.142675 4988 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 08 18:58:07 crc kubenswrapper[4988]: I1008 18:58:07.154462 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-7ffjt" event={"ID":"8fe66bb3-6940-4535-b1a9-f6e09074b9b4","Type":"ContainerStarted","Data":"8f1be56309184bc0d1691b5cd9754f6729069b530e9a5ca228ffa0801e7cebfe"} Oct 08 18:58:08 crc kubenswrapper[4988]: I1008 18:58:08.167874 4988 generic.go:334] "Generic (PLEG): container finished" podID="8fe66bb3-6940-4535-b1a9-f6e09074b9b4" containerID="8f1be56309184bc0d1691b5cd9754f6729069b530e9a5ca228ffa0801e7cebfe" exitCode=0 Oct 08 18:58:08 crc kubenswrapper[4988]: I1008 18:58:08.167925 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-7ffjt" event={"ID":"8fe66bb3-6940-4535-b1a9-f6e09074b9b4","Type":"ContainerDied","Data":"8f1be56309184bc0d1691b5cd9754f6729069b530e9a5ca228ffa0801e7cebfe"} Oct 08 18:58:09 crc kubenswrapper[4988]: I1008 18:58:09.178307 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-7ffjt" event={"ID":"8fe66bb3-6940-4535-b1a9-f6e09074b9b4","Type":"ContainerStarted","Data":"9ace3c07f9a1e873ac4b66c75cd550f975e1459572f9df000a17407d39d74dfe"} Oct 08 18:58:09 crc kubenswrapper[4988]: I1008 18:58:09.208694 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-7ffjt" podStartSLOduration=2.768612287 podStartE2EDuration="5.208671904s" podCreationTimestamp="2025-10-08 18:58:04 +0000 UTC" firstStartedPulling="2025-10-08 18:58:06.142241652 +0000 UTC m=+2831.592084462" lastFinishedPulling="2025-10-08 18:58:08.582301309 +0000 UTC m=+2834.032144079" observedRunningTime="2025-10-08 18:58:09.195033618 +0000 UTC m=+2834.644876398" watchObservedRunningTime="2025-10-08 18:58:09.208671904 +0000 UTC m=+2834.658514674" Oct 08 18:58:15 crc kubenswrapper[4988]: I1008 18:58:15.350624 4988 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-7ffjt" Oct 08 18:58:15 crc kubenswrapper[4988]: I1008 18:58:15.352747 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-7ffjt" Oct 08 18:58:15 crc kubenswrapper[4988]: I1008 18:58:15.419910 4988 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-7ffjt" Oct 08 18:58:16 crc kubenswrapper[4988]: I1008 18:58:16.303881 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-7ffjt" Oct 08 18:58:16 crc kubenswrapper[4988]: I1008 18:58:16.363562 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-7ffjt"] Oct 08 18:58:18 crc kubenswrapper[4988]: I1008 18:58:18.254845 4988 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-7ffjt" podUID="8fe66bb3-6940-4535-b1a9-f6e09074b9b4" containerName="registry-server" containerID="cri-o://9ace3c07f9a1e873ac4b66c75cd550f975e1459572f9df000a17407d39d74dfe" gracePeriod=2 Oct 08 18:58:18 crc kubenswrapper[4988]: I1008 18:58:18.670985 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-7ffjt" Oct 08 18:58:18 crc kubenswrapper[4988]: I1008 18:58:18.787557 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8fe66bb3-6940-4535-b1a9-f6e09074b9b4-utilities\") pod \"8fe66bb3-6940-4535-b1a9-f6e09074b9b4\" (UID: \"8fe66bb3-6940-4535-b1a9-f6e09074b9b4\") " Oct 08 18:58:18 crc kubenswrapper[4988]: I1008 18:58:18.787712 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cv9rx\" (UniqueName: \"kubernetes.io/projected/8fe66bb3-6940-4535-b1a9-f6e09074b9b4-kube-api-access-cv9rx\") pod \"8fe66bb3-6940-4535-b1a9-f6e09074b9b4\" (UID: \"8fe66bb3-6940-4535-b1a9-f6e09074b9b4\") " Oct 08 18:58:18 crc kubenswrapper[4988]: I1008 18:58:18.787808 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8fe66bb3-6940-4535-b1a9-f6e09074b9b4-catalog-content\") pod \"8fe66bb3-6940-4535-b1a9-f6e09074b9b4\" (UID: \"8fe66bb3-6940-4535-b1a9-f6e09074b9b4\") " Oct 08 18:58:18 crc kubenswrapper[4988]: I1008 18:58:18.788891 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8fe66bb3-6940-4535-b1a9-f6e09074b9b4-utilities" (OuterVolumeSpecName: "utilities") pod "8fe66bb3-6940-4535-b1a9-f6e09074b9b4" (UID: "8fe66bb3-6940-4535-b1a9-f6e09074b9b4"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 18:58:18 crc kubenswrapper[4988]: I1008 18:58:18.793892 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8fe66bb3-6940-4535-b1a9-f6e09074b9b4-kube-api-access-cv9rx" (OuterVolumeSpecName: "kube-api-access-cv9rx") pod "8fe66bb3-6940-4535-b1a9-f6e09074b9b4" (UID: "8fe66bb3-6940-4535-b1a9-f6e09074b9b4"). InnerVolumeSpecName "kube-api-access-cv9rx". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:58:18 crc kubenswrapper[4988]: I1008 18:58:18.801726 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8fe66bb3-6940-4535-b1a9-f6e09074b9b4-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "8fe66bb3-6940-4535-b1a9-f6e09074b9b4" (UID: "8fe66bb3-6940-4535-b1a9-f6e09074b9b4"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 18:58:18 crc kubenswrapper[4988]: I1008 18:58:18.888938 4988 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8fe66bb3-6940-4535-b1a9-f6e09074b9b4-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 08 18:58:18 crc kubenswrapper[4988]: I1008 18:58:18.888968 4988 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8fe66bb3-6940-4535-b1a9-f6e09074b9b4-utilities\") on node \"crc\" DevicePath \"\"" Oct 08 18:58:18 crc kubenswrapper[4988]: I1008 18:58:18.888982 4988 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cv9rx\" (UniqueName: \"kubernetes.io/projected/8fe66bb3-6940-4535-b1a9-f6e09074b9b4-kube-api-access-cv9rx\") on node \"crc\" DevicePath \"\"" Oct 08 18:58:19 crc kubenswrapper[4988]: I1008 18:58:19.266640 4988 generic.go:334] "Generic (PLEG): container finished" podID="8fe66bb3-6940-4535-b1a9-f6e09074b9b4" containerID="9ace3c07f9a1e873ac4b66c75cd550f975e1459572f9df000a17407d39d74dfe" exitCode=0 Oct 08 18:58:19 crc kubenswrapper[4988]: I1008 18:58:19.266688 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-7ffjt" event={"ID":"8fe66bb3-6940-4535-b1a9-f6e09074b9b4","Type":"ContainerDied","Data":"9ace3c07f9a1e873ac4b66c75cd550f975e1459572f9df000a17407d39d74dfe"} Oct 08 18:58:19 crc kubenswrapper[4988]: I1008 18:58:19.266716 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-7ffjt" Oct 08 18:58:19 crc kubenswrapper[4988]: I1008 18:58:19.266726 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-7ffjt" event={"ID":"8fe66bb3-6940-4535-b1a9-f6e09074b9b4","Type":"ContainerDied","Data":"1fc3c75c5072bced1041592b875bea40118b1f0fd1a5c9e4052f221460581103"} Oct 08 18:58:19 crc kubenswrapper[4988]: I1008 18:58:19.266748 4988 scope.go:117] "RemoveContainer" containerID="9ace3c07f9a1e873ac4b66c75cd550f975e1459572f9df000a17407d39d74dfe" Oct 08 18:58:19 crc kubenswrapper[4988]: I1008 18:58:19.312800 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-7ffjt"] Oct 08 18:58:19 crc kubenswrapper[4988]: I1008 18:58:19.314063 4988 scope.go:117] "RemoveContainer" containerID="8f1be56309184bc0d1691b5cd9754f6729069b530e9a5ca228ffa0801e7cebfe" Oct 08 18:58:19 crc kubenswrapper[4988]: I1008 18:58:19.318069 4988 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-7ffjt"] Oct 08 18:58:19 crc kubenswrapper[4988]: I1008 18:58:19.333594 4988 scope.go:117] "RemoveContainer" containerID="5016e99fb3491854a8ffb5899b1274a9ec58ee4613d8359bd3ae34f8503f8aa7" Oct 08 18:58:19 crc kubenswrapper[4988]: I1008 18:58:19.356334 4988 scope.go:117] "RemoveContainer" containerID="9ace3c07f9a1e873ac4b66c75cd550f975e1459572f9df000a17407d39d74dfe" Oct 08 18:58:19 crc kubenswrapper[4988]: E1008 18:58:19.356981 4988 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9ace3c07f9a1e873ac4b66c75cd550f975e1459572f9df000a17407d39d74dfe\": container with ID starting with 9ace3c07f9a1e873ac4b66c75cd550f975e1459572f9df000a17407d39d74dfe not found: ID does not exist" containerID="9ace3c07f9a1e873ac4b66c75cd550f975e1459572f9df000a17407d39d74dfe" Oct 08 18:58:19 crc kubenswrapper[4988]: I1008 18:58:19.357030 4988 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9ace3c07f9a1e873ac4b66c75cd550f975e1459572f9df000a17407d39d74dfe"} err="failed to get container status \"9ace3c07f9a1e873ac4b66c75cd550f975e1459572f9df000a17407d39d74dfe\": rpc error: code = NotFound desc = could not find container \"9ace3c07f9a1e873ac4b66c75cd550f975e1459572f9df000a17407d39d74dfe\": container with ID starting with 9ace3c07f9a1e873ac4b66c75cd550f975e1459572f9df000a17407d39d74dfe not found: ID does not exist" Oct 08 18:58:19 crc kubenswrapper[4988]: I1008 18:58:19.357066 4988 scope.go:117] "RemoveContainer" containerID="8f1be56309184bc0d1691b5cd9754f6729069b530e9a5ca228ffa0801e7cebfe" Oct 08 18:58:19 crc kubenswrapper[4988]: E1008 18:58:19.357633 4988 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8f1be56309184bc0d1691b5cd9754f6729069b530e9a5ca228ffa0801e7cebfe\": container with ID starting with 8f1be56309184bc0d1691b5cd9754f6729069b530e9a5ca228ffa0801e7cebfe not found: ID does not exist" containerID="8f1be56309184bc0d1691b5cd9754f6729069b530e9a5ca228ffa0801e7cebfe" Oct 08 18:58:19 crc kubenswrapper[4988]: I1008 18:58:19.357688 4988 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8f1be56309184bc0d1691b5cd9754f6729069b530e9a5ca228ffa0801e7cebfe"} err="failed to get container status \"8f1be56309184bc0d1691b5cd9754f6729069b530e9a5ca228ffa0801e7cebfe\": rpc error: code = NotFound desc = could not find container \"8f1be56309184bc0d1691b5cd9754f6729069b530e9a5ca228ffa0801e7cebfe\": container with ID starting with 8f1be56309184bc0d1691b5cd9754f6729069b530e9a5ca228ffa0801e7cebfe not found: ID does not exist" Oct 08 18:58:19 crc kubenswrapper[4988]: I1008 18:58:19.357720 4988 scope.go:117] "RemoveContainer" containerID="5016e99fb3491854a8ffb5899b1274a9ec58ee4613d8359bd3ae34f8503f8aa7" Oct 08 18:58:19 crc kubenswrapper[4988]: E1008 18:58:19.358020 4988 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5016e99fb3491854a8ffb5899b1274a9ec58ee4613d8359bd3ae34f8503f8aa7\": container with ID starting with 5016e99fb3491854a8ffb5899b1274a9ec58ee4613d8359bd3ae34f8503f8aa7 not found: ID does not exist" containerID="5016e99fb3491854a8ffb5899b1274a9ec58ee4613d8359bd3ae34f8503f8aa7" Oct 08 18:58:19 crc kubenswrapper[4988]: I1008 18:58:19.358049 4988 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5016e99fb3491854a8ffb5899b1274a9ec58ee4613d8359bd3ae34f8503f8aa7"} err="failed to get container status \"5016e99fb3491854a8ffb5899b1274a9ec58ee4613d8359bd3ae34f8503f8aa7\": rpc error: code = NotFound desc = could not find container \"5016e99fb3491854a8ffb5899b1274a9ec58ee4613d8359bd3ae34f8503f8aa7\": container with ID starting with 5016e99fb3491854a8ffb5899b1274a9ec58ee4613d8359bd3ae34f8503f8aa7 not found: ID does not exist" Oct 08 18:58:21 crc kubenswrapper[4988]: I1008 18:58:21.249287 4988 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8fe66bb3-6940-4535-b1a9-f6e09074b9b4" path="/var/lib/kubelet/pods/8fe66bb3-6940-4535-b1a9-f6e09074b9b4/volumes" Oct 08 18:58:53 crc kubenswrapper[4988]: I1008 18:58:53.338669 4988 patch_prober.go:28] interesting pod/machine-config-daemon-nm54f container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 18:58:53 crc kubenswrapper[4988]: I1008 18:58:53.339273 4988 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 18:59:23 crc kubenswrapper[4988]: I1008 18:59:23.338508 4988 patch_prober.go:28] interesting pod/machine-config-daemon-nm54f container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 18:59:23 crc kubenswrapper[4988]: I1008 18:59:23.339176 4988 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 18:59:53 crc kubenswrapper[4988]: I1008 18:59:53.338595 4988 patch_prober.go:28] interesting pod/machine-config-daemon-nm54f container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 18:59:53 crc kubenswrapper[4988]: I1008 18:59:53.339294 4988 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 18:59:53 crc kubenswrapper[4988]: I1008 18:59:53.339372 4988 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" Oct 08 18:59:53 crc kubenswrapper[4988]: I1008 18:59:53.340523 4988 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"fea33312e34e76e64c798028f94adde7baabd667aa354082b2aebdb7e50779da"} pod="openshift-machine-config-operator/machine-config-daemon-nm54f" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 08 18:59:53 crc kubenswrapper[4988]: I1008 18:59:53.340631 4988 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" containerName="machine-config-daemon" containerID="cri-o://fea33312e34e76e64c798028f94adde7baabd667aa354082b2aebdb7e50779da" gracePeriod=600 Oct 08 18:59:53 crc kubenswrapper[4988]: E1008 18:59:53.475975 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nm54f_openshift-machine-config-operator(f328a20b-5283-407e-8a93-2ab2da599f18)\"" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" Oct 08 18:59:54 crc kubenswrapper[4988]: I1008 18:59:54.128934 4988 generic.go:334] "Generic (PLEG): container finished" podID="f328a20b-5283-407e-8a93-2ab2da599f18" containerID="fea33312e34e76e64c798028f94adde7baabd667aa354082b2aebdb7e50779da" exitCode=0 Oct 08 18:59:54 crc kubenswrapper[4988]: I1008 18:59:54.128950 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" event={"ID":"f328a20b-5283-407e-8a93-2ab2da599f18","Type":"ContainerDied","Data":"fea33312e34e76e64c798028f94adde7baabd667aa354082b2aebdb7e50779da"} Oct 08 18:59:54 crc kubenswrapper[4988]: I1008 18:59:54.129989 4988 scope.go:117] "RemoveContainer" containerID="df9f51685b2cfae41c8322202dd543ae485c4f24e54c4b9380bef89dae14c52e" Oct 08 18:59:54 crc kubenswrapper[4988]: I1008 18:59:54.130457 4988 scope.go:117] "RemoveContainer" containerID="fea33312e34e76e64c798028f94adde7baabd667aa354082b2aebdb7e50779da" Oct 08 18:59:54 crc kubenswrapper[4988]: E1008 18:59:54.130721 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nm54f_openshift-machine-config-operator(f328a20b-5283-407e-8a93-2ab2da599f18)\"" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" Oct 08 19:00:00 crc kubenswrapper[4988]: I1008 19:00:00.201515 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29332500-85wnq"] Oct 08 19:00:00 crc kubenswrapper[4988]: E1008 19:00:00.202094 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8fe66bb3-6940-4535-b1a9-f6e09074b9b4" containerName="registry-server" Oct 08 19:00:00 crc kubenswrapper[4988]: I1008 19:00:00.202107 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="8fe66bb3-6940-4535-b1a9-f6e09074b9b4" containerName="registry-server" Oct 08 19:00:00 crc kubenswrapper[4988]: E1008 19:00:00.202144 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8fe66bb3-6940-4535-b1a9-f6e09074b9b4" containerName="extract-content" Oct 08 19:00:00 crc kubenswrapper[4988]: I1008 19:00:00.202152 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="8fe66bb3-6940-4535-b1a9-f6e09074b9b4" containerName="extract-content" Oct 08 19:00:00 crc kubenswrapper[4988]: E1008 19:00:00.202173 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8fe66bb3-6940-4535-b1a9-f6e09074b9b4" containerName="extract-utilities" Oct 08 19:00:00 crc kubenswrapper[4988]: I1008 19:00:00.202183 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="8fe66bb3-6940-4535-b1a9-f6e09074b9b4" containerName="extract-utilities" Oct 08 19:00:00 crc kubenswrapper[4988]: I1008 19:00:00.202353 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="8fe66bb3-6940-4535-b1a9-f6e09074b9b4" containerName="registry-server" Oct 08 19:00:00 crc kubenswrapper[4988]: I1008 19:00:00.202848 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29332500-85wnq" Oct 08 19:00:00 crc kubenswrapper[4988]: I1008 19:00:00.205039 4988 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 08 19:00:00 crc kubenswrapper[4988]: I1008 19:00:00.210000 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29332500-85wnq"] Oct 08 19:00:00 crc kubenswrapper[4988]: I1008 19:00:00.217151 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 08 19:00:00 crc kubenswrapper[4988]: I1008 19:00:00.289073 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/8a70d02b-f876-4063-9dbc-a7159e12d8d8-config-volume\") pod \"collect-profiles-29332500-85wnq\" (UID: \"8a70d02b-f876-4063-9dbc-a7159e12d8d8\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29332500-85wnq" Oct 08 19:00:00 crc kubenswrapper[4988]: I1008 19:00:00.289335 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4hjb4\" (UniqueName: \"kubernetes.io/projected/8a70d02b-f876-4063-9dbc-a7159e12d8d8-kube-api-access-4hjb4\") pod \"collect-profiles-29332500-85wnq\" (UID: \"8a70d02b-f876-4063-9dbc-a7159e12d8d8\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29332500-85wnq" Oct 08 19:00:00 crc kubenswrapper[4988]: I1008 19:00:00.289435 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/8a70d02b-f876-4063-9dbc-a7159e12d8d8-secret-volume\") pod \"collect-profiles-29332500-85wnq\" (UID: \"8a70d02b-f876-4063-9dbc-a7159e12d8d8\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29332500-85wnq" Oct 08 19:00:00 crc kubenswrapper[4988]: I1008 19:00:00.390551 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/8a70d02b-f876-4063-9dbc-a7159e12d8d8-config-volume\") pod \"collect-profiles-29332500-85wnq\" (UID: \"8a70d02b-f876-4063-9dbc-a7159e12d8d8\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29332500-85wnq" Oct 08 19:00:00 crc kubenswrapper[4988]: I1008 19:00:00.390603 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4hjb4\" (UniqueName: \"kubernetes.io/projected/8a70d02b-f876-4063-9dbc-a7159e12d8d8-kube-api-access-4hjb4\") pod \"collect-profiles-29332500-85wnq\" (UID: \"8a70d02b-f876-4063-9dbc-a7159e12d8d8\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29332500-85wnq" Oct 08 19:00:00 crc kubenswrapper[4988]: I1008 19:00:00.390623 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/8a70d02b-f876-4063-9dbc-a7159e12d8d8-secret-volume\") pod \"collect-profiles-29332500-85wnq\" (UID: \"8a70d02b-f876-4063-9dbc-a7159e12d8d8\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29332500-85wnq" Oct 08 19:00:00 crc kubenswrapper[4988]: I1008 19:00:00.392011 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/8a70d02b-f876-4063-9dbc-a7159e12d8d8-config-volume\") pod \"collect-profiles-29332500-85wnq\" (UID: \"8a70d02b-f876-4063-9dbc-a7159e12d8d8\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29332500-85wnq" Oct 08 19:00:00 crc kubenswrapper[4988]: I1008 19:00:00.399147 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/8a70d02b-f876-4063-9dbc-a7159e12d8d8-secret-volume\") pod \"collect-profiles-29332500-85wnq\" (UID: \"8a70d02b-f876-4063-9dbc-a7159e12d8d8\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29332500-85wnq" Oct 08 19:00:00 crc kubenswrapper[4988]: I1008 19:00:00.412830 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4hjb4\" (UniqueName: \"kubernetes.io/projected/8a70d02b-f876-4063-9dbc-a7159e12d8d8-kube-api-access-4hjb4\") pod \"collect-profiles-29332500-85wnq\" (UID: \"8a70d02b-f876-4063-9dbc-a7159e12d8d8\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29332500-85wnq" Oct 08 19:00:00 crc kubenswrapper[4988]: I1008 19:00:00.520880 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29332500-85wnq" Oct 08 19:00:00 crc kubenswrapper[4988]: I1008 19:00:00.798563 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29332500-85wnq"] Oct 08 19:00:00 crc kubenswrapper[4988]: W1008 19:00:00.807672 4988 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8a70d02b_f876_4063_9dbc_a7159e12d8d8.slice/crio-8f259e6c756992cb23063daf4e7c677b8d33b0b8d501ac2d0b0c0c7db749c1c2 WatchSource:0}: Error finding container 8f259e6c756992cb23063daf4e7c677b8d33b0b8d501ac2d0b0c0c7db749c1c2: Status 404 returned error can't find the container with id 8f259e6c756992cb23063daf4e7c677b8d33b0b8d501ac2d0b0c0c7db749c1c2 Oct 08 19:00:01 crc kubenswrapper[4988]: I1008 19:00:01.194466 4988 generic.go:334] "Generic (PLEG): container finished" podID="8a70d02b-f876-4063-9dbc-a7159e12d8d8" containerID="9ddccb0f49cf424fcbc1f7bf2729999a66d10acdc7d77d5760f1c39fa7f1fd0e" exitCode=0 Oct 08 19:00:01 crc kubenswrapper[4988]: I1008 19:00:01.194562 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29332500-85wnq" event={"ID":"8a70d02b-f876-4063-9dbc-a7159e12d8d8","Type":"ContainerDied","Data":"9ddccb0f49cf424fcbc1f7bf2729999a66d10acdc7d77d5760f1c39fa7f1fd0e"} Oct 08 19:00:01 crc kubenswrapper[4988]: I1008 19:00:01.194742 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29332500-85wnq" event={"ID":"8a70d02b-f876-4063-9dbc-a7159e12d8d8","Type":"ContainerStarted","Data":"8f259e6c756992cb23063daf4e7c677b8d33b0b8d501ac2d0b0c0c7db749c1c2"} Oct 08 19:00:02 crc kubenswrapper[4988]: I1008 19:00:02.486514 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29332500-85wnq" Oct 08 19:00:02 crc kubenswrapper[4988]: I1008 19:00:02.621898 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4hjb4\" (UniqueName: \"kubernetes.io/projected/8a70d02b-f876-4063-9dbc-a7159e12d8d8-kube-api-access-4hjb4\") pod \"8a70d02b-f876-4063-9dbc-a7159e12d8d8\" (UID: \"8a70d02b-f876-4063-9dbc-a7159e12d8d8\") " Oct 08 19:00:02 crc kubenswrapper[4988]: I1008 19:00:02.622037 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/8a70d02b-f876-4063-9dbc-a7159e12d8d8-secret-volume\") pod \"8a70d02b-f876-4063-9dbc-a7159e12d8d8\" (UID: \"8a70d02b-f876-4063-9dbc-a7159e12d8d8\") " Oct 08 19:00:02 crc kubenswrapper[4988]: I1008 19:00:02.622113 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/8a70d02b-f876-4063-9dbc-a7159e12d8d8-config-volume\") pod \"8a70d02b-f876-4063-9dbc-a7159e12d8d8\" (UID: \"8a70d02b-f876-4063-9dbc-a7159e12d8d8\") " Oct 08 19:00:02 crc kubenswrapper[4988]: I1008 19:00:02.623155 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8a70d02b-f876-4063-9dbc-a7159e12d8d8-config-volume" (OuterVolumeSpecName: "config-volume") pod "8a70d02b-f876-4063-9dbc-a7159e12d8d8" (UID: "8a70d02b-f876-4063-9dbc-a7159e12d8d8"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 19:00:02 crc kubenswrapper[4988]: I1008 19:00:02.631704 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8a70d02b-f876-4063-9dbc-a7159e12d8d8-kube-api-access-4hjb4" (OuterVolumeSpecName: "kube-api-access-4hjb4") pod "8a70d02b-f876-4063-9dbc-a7159e12d8d8" (UID: "8a70d02b-f876-4063-9dbc-a7159e12d8d8"). InnerVolumeSpecName "kube-api-access-4hjb4". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 19:00:02 crc kubenswrapper[4988]: I1008 19:00:02.633062 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8a70d02b-f876-4063-9dbc-a7159e12d8d8-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "8a70d02b-f876-4063-9dbc-a7159e12d8d8" (UID: "8a70d02b-f876-4063-9dbc-a7159e12d8d8"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 19:00:02 crc kubenswrapper[4988]: I1008 19:00:02.723577 4988 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/8a70d02b-f876-4063-9dbc-a7159e12d8d8-config-volume\") on node \"crc\" DevicePath \"\"" Oct 08 19:00:02 crc kubenswrapper[4988]: I1008 19:00:02.723922 4988 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4hjb4\" (UniqueName: \"kubernetes.io/projected/8a70d02b-f876-4063-9dbc-a7159e12d8d8-kube-api-access-4hjb4\") on node \"crc\" DevicePath \"\"" Oct 08 19:00:02 crc kubenswrapper[4988]: I1008 19:00:02.724051 4988 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/8a70d02b-f876-4063-9dbc-a7159e12d8d8-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 08 19:00:03 crc kubenswrapper[4988]: I1008 19:00:03.212631 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29332500-85wnq" event={"ID":"8a70d02b-f876-4063-9dbc-a7159e12d8d8","Type":"ContainerDied","Data":"8f259e6c756992cb23063daf4e7c677b8d33b0b8d501ac2d0b0c0c7db749c1c2"} Oct 08 19:00:03 crc kubenswrapper[4988]: I1008 19:00:03.213114 4988 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8f259e6c756992cb23063daf4e7c677b8d33b0b8d501ac2d0b0c0c7db749c1c2" Oct 08 19:00:03 crc kubenswrapper[4988]: I1008 19:00:03.212685 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29332500-85wnq" Oct 08 19:00:03 crc kubenswrapper[4988]: I1008 19:00:03.583238 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29332455-v2b7c"] Oct 08 19:00:03 crc kubenswrapper[4988]: I1008 19:00:03.589859 4988 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29332455-v2b7c"] Oct 08 19:00:05 crc kubenswrapper[4988]: I1008 19:00:05.248083 4988 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3cd39683-3b34-4c0e-9fb1-f63668fe5e4a" path="/var/lib/kubelet/pods/3cd39683-3b34-4c0e-9fb1-f63668fe5e4a/volumes" Oct 08 19:00:07 crc kubenswrapper[4988]: I1008 19:00:07.237947 4988 scope.go:117] "RemoveContainer" containerID="fea33312e34e76e64c798028f94adde7baabd667aa354082b2aebdb7e50779da" Oct 08 19:00:07 crc kubenswrapper[4988]: E1008 19:00:07.238867 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nm54f_openshift-machine-config-operator(f328a20b-5283-407e-8a93-2ab2da599f18)\"" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" Oct 08 19:00:19 crc kubenswrapper[4988]: I1008 19:00:19.239565 4988 scope.go:117] "RemoveContainer" containerID="fea33312e34e76e64c798028f94adde7baabd667aa354082b2aebdb7e50779da" Oct 08 19:00:19 crc kubenswrapper[4988]: E1008 19:00:19.240742 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nm54f_openshift-machine-config-operator(f328a20b-5283-407e-8a93-2ab2da599f18)\"" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" Oct 08 19:00:32 crc kubenswrapper[4988]: I1008 19:00:32.256654 4988 scope.go:117] "RemoveContainer" containerID="fea33312e34e76e64c798028f94adde7baabd667aa354082b2aebdb7e50779da" Oct 08 19:00:32 crc kubenswrapper[4988]: E1008 19:00:32.257540 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nm54f_openshift-machine-config-operator(f328a20b-5283-407e-8a93-2ab2da599f18)\"" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" Oct 08 19:00:44 crc kubenswrapper[4988]: I1008 19:00:44.238238 4988 scope.go:117] "RemoveContainer" containerID="fea33312e34e76e64c798028f94adde7baabd667aa354082b2aebdb7e50779da" Oct 08 19:00:44 crc kubenswrapper[4988]: E1008 19:00:44.239442 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nm54f_openshift-machine-config-operator(f328a20b-5283-407e-8a93-2ab2da599f18)\"" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" Oct 08 19:00:57 crc kubenswrapper[4988]: I1008 19:00:57.238054 4988 scope.go:117] "RemoveContainer" containerID="fea33312e34e76e64c798028f94adde7baabd667aa354082b2aebdb7e50779da" Oct 08 19:00:57 crc kubenswrapper[4988]: E1008 19:00:57.238887 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nm54f_openshift-machine-config-operator(f328a20b-5283-407e-8a93-2ab2da599f18)\"" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" Oct 08 19:01:01 crc kubenswrapper[4988]: I1008 19:01:01.692774 4988 scope.go:117] "RemoveContainer" containerID="09aeb58123957fe3fad0f83bd19a7f72bda0b4f670d430867d190556638609ce" Oct 08 19:01:09 crc kubenswrapper[4988]: I1008 19:01:09.238475 4988 scope.go:117] "RemoveContainer" containerID="fea33312e34e76e64c798028f94adde7baabd667aa354082b2aebdb7e50779da" Oct 08 19:01:09 crc kubenswrapper[4988]: E1008 19:01:09.239486 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nm54f_openshift-machine-config-operator(f328a20b-5283-407e-8a93-2ab2da599f18)\"" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" Oct 08 19:01:24 crc kubenswrapper[4988]: I1008 19:01:24.237897 4988 scope.go:117] "RemoveContainer" containerID="fea33312e34e76e64c798028f94adde7baabd667aa354082b2aebdb7e50779da" Oct 08 19:01:24 crc kubenswrapper[4988]: E1008 19:01:24.238811 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nm54f_openshift-machine-config-operator(f328a20b-5283-407e-8a93-2ab2da599f18)\"" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" Oct 08 19:01:39 crc kubenswrapper[4988]: I1008 19:01:39.238692 4988 scope.go:117] "RemoveContainer" containerID="fea33312e34e76e64c798028f94adde7baabd667aa354082b2aebdb7e50779da" Oct 08 19:01:39 crc kubenswrapper[4988]: E1008 19:01:39.240070 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nm54f_openshift-machine-config-operator(f328a20b-5283-407e-8a93-2ab2da599f18)\"" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" Oct 08 19:01:53 crc kubenswrapper[4988]: I1008 19:01:53.242419 4988 scope.go:117] "RemoveContainer" containerID="fea33312e34e76e64c798028f94adde7baabd667aa354082b2aebdb7e50779da" Oct 08 19:01:53 crc kubenswrapper[4988]: E1008 19:01:53.243310 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nm54f_openshift-machine-config-operator(f328a20b-5283-407e-8a93-2ab2da599f18)\"" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" Oct 08 19:02:07 crc kubenswrapper[4988]: I1008 19:02:07.238051 4988 scope.go:117] "RemoveContainer" containerID="fea33312e34e76e64c798028f94adde7baabd667aa354082b2aebdb7e50779da" Oct 08 19:02:07 crc kubenswrapper[4988]: E1008 19:02:07.238953 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nm54f_openshift-machine-config-operator(f328a20b-5283-407e-8a93-2ab2da599f18)\"" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" Oct 08 19:02:20 crc kubenswrapper[4988]: I1008 19:02:20.237738 4988 scope.go:117] "RemoveContainer" containerID="fea33312e34e76e64c798028f94adde7baabd667aa354082b2aebdb7e50779da" Oct 08 19:02:20 crc kubenswrapper[4988]: E1008 19:02:20.238488 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nm54f_openshift-machine-config-operator(f328a20b-5283-407e-8a93-2ab2da599f18)\"" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" Oct 08 19:02:32 crc kubenswrapper[4988]: I1008 19:02:32.237887 4988 scope.go:117] "RemoveContainer" containerID="fea33312e34e76e64c798028f94adde7baabd667aa354082b2aebdb7e50779da" Oct 08 19:02:32 crc kubenswrapper[4988]: E1008 19:02:32.238943 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nm54f_openshift-machine-config-operator(f328a20b-5283-407e-8a93-2ab2da599f18)\"" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" Oct 08 19:02:47 crc kubenswrapper[4988]: I1008 19:02:47.237592 4988 scope.go:117] "RemoveContainer" containerID="fea33312e34e76e64c798028f94adde7baabd667aa354082b2aebdb7e50779da" Oct 08 19:02:47 crc kubenswrapper[4988]: E1008 19:02:47.238381 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nm54f_openshift-machine-config-operator(f328a20b-5283-407e-8a93-2ab2da599f18)\"" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" Oct 08 19:02:58 crc kubenswrapper[4988]: I1008 19:02:58.238452 4988 scope.go:117] "RemoveContainer" containerID="fea33312e34e76e64c798028f94adde7baabd667aa354082b2aebdb7e50779da" Oct 08 19:02:58 crc kubenswrapper[4988]: E1008 19:02:58.239158 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nm54f_openshift-machine-config-operator(f328a20b-5283-407e-8a93-2ab2da599f18)\"" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" Oct 08 19:03:11 crc kubenswrapper[4988]: I1008 19:03:11.238270 4988 scope.go:117] "RemoveContainer" containerID="fea33312e34e76e64c798028f94adde7baabd667aa354082b2aebdb7e50779da" Oct 08 19:03:11 crc kubenswrapper[4988]: E1008 19:03:11.239034 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nm54f_openshift-machine-config-operator(f328a20b-5283-407e-8a93-2ab2da599f18)\"" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" Oct 08 19:03:24 crc kubenswrapper[4988]: I1008 19:03:24.238158 4988 scope.go:117] "RemoveContainer" containerID="fea33312e34e76e64c798028f94adde7baabd667aa354082b2aebdb7e50779da" Oct 08 19:03:24 crc kubenswrapper[4988]: E1008 19:03:24.238908 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nm54f_openshift-machine-config-operator(f328a20b-5283-407e-8a93-2ab2da599f18)\"" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" Oct 08 19:03:37 crc kubenswrapper[4988]: I1008 19:03:37.238074 4988 scope.go:117] "RemoveContainer" containerID="fea33312e34e76e64c798028f94adde7baabd667aa354082b2aebdb7e50779da" Oct 08 19:03:37 crc kubenswrapper[4988]: E1008 19:03:37.238833 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nm54f_openshift-machine-config-operator(f328a20b-5283-407e-8a93-2ab2da599f18)\"" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" Oct 08 19:03:48 crc kubenswrapper[4988]: I1008 19:03:48.238406 4988 scope.go:117] "RemoveContainer" containerID="fea33312e34e76e64c798028f94adde7baabd667aa354082b2aebdb7e50779da" Oct 08 19:03:48 crc kubenswrapper[4988]: E1008 19:03:48.239208 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nm54f_openshift-machine-config-operator(f328a20b-5283-407e-8a93-2ab2da599f18)\"" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" Oct 08 19:04:00 crc kubenswrapper[4988]: I1008 19:04:00.238103 4988 scope.go:117] "RemoveContainer" containerID="fea33312e34e76e64c798028f94adde7baabd667aa354082b2aebdb7e50779da" Oct 08 19:04:00 crc kubenswrapper[4988]: E1008 19:04:00.239271 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nm54f_openshift-machine-config-operator(f328a20b-5283-407e-8a93-2ab2da599f18)\"" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" Oct 08 19:04:11 crc kubenswrapper[4988]: I1008 19:04:11.238483 4988 scope.go:117] "RemoveContainer" containerID="fea33312e34e76e64c798028f94adde7baabd667aa354082b2aebdb7e50779da" Oct 08 19:04:11 crc kubenswrapper[4988]: E1008 19:04:11.239669 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nm54f_openshift-machine-config-operator(f328a20b-5283-407e-8a93-2ab2da599f18)\"" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" Oct 08 19:04:24 crc kubenswrapper[4988]: I1008 19:04:24.238027 4988 scope.go:117] "RemoveContainer" containerID="fea33312e34e76e64c798028f94adde7baabd667aa354082b2aebdb7e50779da" Oct 08 19:04:24 crc kubenswrapper[4988]: E1008 19:04:24.239133 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nm54f_openshift-machine-config-operator(f328a20b-5283-407e-8a93-2ab2da599f18)\"" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" Oct 08 19:04:39 crc kubenswrapper[4988]: I1008 19:04:39.238692 4988 scope.go:117] "RemoveContainer" containerID="fea33312e34e76e64c798028f94adde7baabd667aa354082b2aebdb7e50779da" Oct 08 19:04:39 crc kubenswrapper[4988]: E1008 19:04:39.239270 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nm54f_openshift-machine-config-operator(f328a20b-5283-407e-8a93-2ab2da599f18)\"" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" Oct 08 19:04:50 crc kubenswrapper[4988]: I1008 19:04:50.237926 4988 scope.go:117] "RemoveContainer" containerID="fea33312e34e76e64c798028f94adde7baabd667aa354082b2aebdb7e50779da" Oct 08 19:04:50 crc kubenswrapper[4988]: E1008 19:04:50.239050 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nm54f_openshift-machine-config-operator(f328a20b-5283-407e-8a93-2ab2da599f18)\"" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" Oct 08 19:05:02 crc kubenswrapper[4988]: I1008 19:05:02.237664 4988 scope.go:117] "RemoveContainer" containerID="fea33312e34e76e64c798028f94adde7baabd667aa354082b2aebdb7e50779da" Oct 08 19:05:02 crc kubenswrapper[4988]: I1008 19:05:02.648018 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" event={"ID":"f328a20b-5283-407e-8a93-2ab2da599f18","Type":"ContainerStarted","Data":"482cb9ffb14e123461aee7ddc866cf73ccfa365d4521f4ee6abf4f22170a86b2"} Oct 08 19:07:07 crc kubenswrapper[4988]: I1008 19:07:07.733858 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-sz5l9"] Oct 08 19:07:07 crc kubenswrapper[4988]: E1008 19:07:07.734811 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8a70d02b-f876-4063-9dbc-a7159e12d8d8" containerName="collect-profiles" Oct 08 19:07:07 crc kubenswrapper[4988]: I1008 19:07:07.734832 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="8a70d02b-f876-4063-9dbc-a7159e12d8d8" containerName="collect-profiles" Oct 08 19:07:07 crc kubenswrapper[4988]: I1008 19:07:07.735132 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="8a70d02b-f876-4063-9dbc-a7159e12d8d8" containerName="collect-profiles" Oct 08 19:07:07 crc kubenswrapper[4988]: I1008 19:07:07.736935 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-sz5l9" Oct 08 19:07:07 crc kubenswrapper[4988]: I1008 19:07:07.763744 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-sz5l9"] Oct 08 19:07:07 crc kubenswrapper[4988]: I1008 19:07:07.907895 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6b4d243b-e683-4688-8044-a729a62a2cb3-utilities\") pod \"community-operators-sz5l9\" (UID: \"6b4d243b-e683-4688-8044-a729a62a2cb3\") " pod="openshift-marketplace/community-operators-sz5l9" Oct 08 19:07:07 crc kubenswrapper[4988]: I1008 19:07:07.907961 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6b4d243b-e683-4688-8044-a729a62a2cb3-catalog-content\") pod \"community-operators-sz5l9\" (UID: \"6b4d243b-e683-4688-8044-a729a62a2cb3\") " pod="openshift-marketplace/community-operators-sz5l9" Oct 08 19:07:07 crc kubenswrapper[4988]: I1008 19:07:07.908493 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4brx8\" (UniqueName: \"kubernetes.io/projected/6b4d243b-e683-4688-8044-a729a62a2cb3-kube-api-access-4brx8\") pod \"community-operators-sz5l9\" (UID: \"6b4d243b-e683-4688-8044-a729a62a2cb3\") " pod="openshift-marketplace/community-operators-sz5l9" Oct 08 19:07:08 crc kubenswrapper[4988]: I1008 19:07:08.009997 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6b4d243b-e683-4688-8044-a729a62a2cb3-catalog-content\") pod \"community-operators-sz5l9\" (UID: \"6b4d243b-e683-4688-8044-a729a62a2cb3\") " pod="openshift-marketplace/community-operators-sz5l9" Oct 08 19:07:08 crc kubenswrapper[4988]: I1008 19:07:08.010120 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4brx8\" (UniqueName: \"kubernetes.io/projected/6b4d243b-e683-4688-8044-a729a62a2cb3-kube-api-access-4brx8\") pod \"community-operators-sz5l9\" (UID: \"6b4d243b-e683-4688-8044-a729a62a2cb3\") " pod="openshift-marketplace/community-operators-sz5l9" Oct 08 19:07:08 crc kubenswrapper[4988]: I1008 19:07:08.010187 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6b4d243b-e683-4688-8044-a729a62a2cb3-utilities\") pod \"community-operators-sz5l9\" (UID: \"6b4d243b-e683-4688-8044-a729a62a2cb3\") " pod="openshift-marketplace/community-operators-sz5l9" Oct 08 19:07:08 crc kubenswrapper[4988]: I1008 19:07:08.010713 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6b4d243b-e683-4688-8044-a729a62a2cb3-catalog-content\") pod \"community-operators-sz5l9\" (UID: \"6b4d243b-e683-4688-8044-a729a62a2cb3\") " pod="openshift-marketplace/community-operators-sz5l9" Oct 08 19:07:08 crc kubenswrapper[4988]: I1008 19:07:08.010739 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6b4d243b-e683-4688-8044-a729a62a2cb3-utilities\") pod \"community-operators-sz5l9\" (UID: \"6b4d243b-e683-4688-8044-a729a62a2cb3\") " pod="openshift-marketplace/community-operators-sz5l9" Oct 08 19:07:08 crc kubenswrapper[4988]: I1008 19:07:08.036450 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4brx8\" (UniqueName: \"kubernetes.io/projected/6b4d243b-e683-4688-8044-a729a62a2cb3-kube-api-access-4brx8\") pod \"community-operators-sz5l9\" (UID: \"6b4d243b-e683-4688-8044-a729a62a2cb3\") " pod="openshift-marketplace/community-operators-sz5l9" Oct 08 19:07:08 crc kubenswrapper[4988]: I1008 19:07:08.074572 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-sz5l9" Oct 08 19:07:08 crc kubenswrapper[4988]: I1008 19:07:08.551207 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-sz5l9"] Oct 08 19:07:08 crc kubenswrapper[4988]: I1008 19:07:08.760052 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-sz5l9" event={"ID":"6b4d243b-e683-4688-8044-a729a62a2cb3","Type":"ContainerStarted","Data":"34a6dc939250917cd74b220199d805d3b6253b5b055ffec3b954c30f128708e3"} Oct 08 19:07:08 crc kubenswrapper[4988]: I1008 19:07:08.760098 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-sz5l9" event={"ID":"6b4d243b-e683-4688-8044-a729a62a2cb3","Type":"ContainerStarted","Data":"fec35c168b02bd82ba9a055f953b335328c9158f5906982c6742434cb9dbf91f"} Oct 08 19:07:09 crc kubenswrapper[4988]: I1008 19:07:09.771241 4988 generic.go:334] "Generic (PLEG): container finished" podID="6b4d243b-e683-4688-8044-a729a62a2cb3" containerID="34a6dc939250917cd74b220199d805d3b6253b5b055ffec3b954c30f128708e3" exitCode=0 Oct 08 19:07:09 crc kubenswrapper[4988]: I1008 19:07:09.771297 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-sz5l9" event={"ID":"6b4d243b-e683-4688-8044-a729a62a2cb3","Type":"ContainerDied","Data":"34a6dc939250917cd74b220199d805d3b6253b5b055ffec3b954c30f128708e3"} Oct 08 19:07:09 crc kubenswrapper[4988]: I1008 19:07:09.774241 4988 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 08 19:07:10 crc kubenswrapper[4988]: I1008 19:07:10.788538 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-sz5l9" event={"ID":"6b4d243b-e683-4688-8044-a729a62a2cb3","Type":"ContainerStarted","Data":"90d8a1960cc20709b55e8398ac082ad4c14e737a1b2e2db6b420dcab9d888e7d"} Oct 08 19:07:11 crc kubenswrapper[4988]: I1008 19:07:11.798106 4988 generic.go:334] "Generic (PLEG): container finished" podID="6b4d243b-e683-4688-8044-a729a62a2cb3" containerID="90d8a1960cc20709b55e8398ac082ad4c14e737a1b2e2db6b420dcab9d888e7d" exitCode=0 Oct 08 19:07:11 crc kubenswrapper[4988]: I1008 19:07:11.798153 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-sz5l9" event={"ID":"6b4d243b-e683-4688-8044-a729a62a2cb3","Type":"ContainerDied","Data":"90d8a1960cc20709b55e8398ac082ad4c14e737a1b2e2db6b420dcab9d888e7d"} Oct 08 19:07:12 crc kubenswrapper[4988]: I1008 19:07:12.811060 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-sz5l9" event={"ID":"6b4d243b-e683-4688-8044-a729a62a2cb3","Type":"ContainerStarted","Data":"74fe37312dd863ad7ba7e6afa5ad23adff128a4bdea75e1f6d13e020061d4bb4"} Oct 08 19:07:12 crc kubenswrapper[4988]: I1008 19:07:12.850715 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-sz5l9" podStartSLOduration=3.241481181 podStartE2EDuration="5.850691301s" podCreationTimestamp="2025-10-08 19:07:07 +0000 UTC" firstStartedPulling="2025-10-08 19:07:09.774035865 +0000 UTC m=+3375.223878645" lastFinishedPulling="2025-10-08 19:07:12.383245955 +0000 UTC m=+3377.833088765" observedRunningTime="2025-10-08 19:07:12.847430817 +0000 UTC m=+3378.297273657" watchObservedRunningTime="2025-10-08 19:07:12.850691301 +0000 UTC m=+3378.300534081" Oct 08 19:07:18 crc kubenswrapper[4988]: I1008 19:07:18.075241 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-sz5l9" Oct 08 19:07:18 crc kubenswrapper[4988]: I1008 19:07:18.075939 4988 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-sz5l9" Oct 08 19:07:18 crc kubenswrapper[4988]: I1008 19:07:18.136553 4988 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-sz5l9" Oct 08 19:07:18 crc kubenswrapper[4988]: I1008 19:07:18.930767 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-sz5l9" Oct 08 19:07:18 crc kubenswrapper[4988]: I1008 19:07:18.991652 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-sz5l9"] Oct 08 19:07:20 crc kubenswrapper[4988]: I1008 19:07:20.882171 4988 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-sz5l9" podUID="6b4d243b-e683-4688-8044-a729a62a2cb3" containerName="registry-server" containerID="cri-o://74fe37312dd863ad7ba7e6afa5ad23adff128a4bdea75e1f6d13e020061d4bb4" gracePeriod=2 Oct 08 19:07:21 crc kubenswrapper[4988]: I1008 19:07:21.367224 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-sz5l9" Oct 08 19:07:21 crc kubenswrapper[4988]: I1008 19:07:21.517599 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4brx8\" (UniqueName: \"kubernetes.io/projected/6b4d243b-e683-4688-8044-a729a62a2cb3-kube-api-access-4brx8\") pod \"6b4d243b-e683-4688-8044-a729a62a2cb3\" (UID: \"6b4d243b-e683-4688-8044-a729a62a2cb3\") " Oct 08 19:07:21 crc kubenswrapper[4988]: I1008 19:07:21.518703 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6b4d243b-e683-4688-8044-a729a62a2cb3-catalog-content\") pod \"6b4d243b-e683-4688-8044-a729a62a2cb3\" (UID: \"6b4d243b-e683-4688-8044-a729a62a2cb3\") " Oct 08 19:07:21 crc kubenswrapper[4988]: I1008 19:07:21.518751 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6b4d243b-e683-4688-8044-a729a62a2cb3-utilities\") pod \"6b4d243b-e683-4688-8044-a729a62a2cb3\" (UID: \"6b4d243b-e683-4688-8044-a729a62a2cb3\") " Oct 08 19:07:21 crc kubenswrapper[4988]: I1008 19:07:21.519796 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6b4d243b-e683-4688-8044-a729a62a2cb3-utilities" (OuterVolumeSpecName: "utilities") pod "6b4d243b-e683-4688-8044-a729a62a2cb3" (UID: "6b4d243b-e683-4688-8044-a729a62a2cb3"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 19:07:21 crc kubenswrapper[4988]: I1008 19:07:21.525720 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6b4d243b-e683-4688-8044-a729a62a2cb3-kube-api-access-4brx8" (OuterVolumeSpecName: "kube-api-access-4brx8") pod "6b4d243b-e683-4688-8044-a729a62a2cb3" (UID: "6b4d243b-e683-4688-8044-a729a62a2cb3"). InnerVolumeSpecName "kube-api-access-4brx8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 19:07:21 crc kubenswrapper[4988]: I1008 19:07:21.601813 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6b4d243b-e683-4688-8044-a729a62a2cb3-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "6b4d243b-e683-4688-8044-a729a62a2cb3" (UID: "6b4d243b-e683-4688-8044-a729a62a2cb3"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 19:07:21 crc kubenswrapper[4988]: I1008 19:07:21.619832 4988 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4brx8\" (UniqueName: \"kubernetes.io/projected/6b4d243b-e683-4688-8044-a729a62a2cb3-kube-api-access-4brx8\") on node \"crc\" DevicePath \"\"" Oct 08 19:07:21 crc kubenswrapper[4988]: I1008 19:07:21.619903 4988 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6b4d243b-e683-4688-8044-a729a62a2cb3-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 08 19:07:21 crc kubenswrapper[4988]: I1008 19:07:21.619921 4988 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6b4d243b-e683-4688-8044-a729a62a2cb3-utilities\") on node \"crc\" DevicePath \"\"" Oct 08 19:07:21 crc kubenswrapper[4988]: I1008 19:07:21.895169 4988 generic.go:334] "Generic (PLEG): container finished" podID="6b4d243b-e683-4688-8044-a729a62a2cb3" containerID="74fe37312dd863ad7ba7e6afa5ad23adff128a4bdea75e1f6d13e020061d4bb4" exitCode=0 Oct 08 19:07:21 crc kubenswrapper[4988]: I1008 19:07:21.895224 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-sz5l9" event={"ID":"6b4d243b-e683-4688-8044-a729a62a2cb3","Type":"ContainerDied","Data":"74fe37312dd863ad7ba7e6afa5ad23adff128a4bdea75e1f6d13e020061d4bb4"} Oct 08 19:07:21 crc kubenswrapper[4988]: I1008 19:07:21.895259 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-sz5l9" event={"ID":"6b4d243b-e683-4688-8044-a729a62a2cb3","Type":"ContainerDied","Data":"fec35c168b02bd82ba9a055f953b335328c9158f5906982c6742434cb9dbf91f"} Oct 08 19:07:21 crc kubenswrapper[4988]: I1008 19:07:21.895286 4988 scope.go:117] "RemoveContainer" containerID="74fe37312dd863ad7ba7e6afa5ad23adff128a4bdea75e1f6d13e020061d4bb4" Oct 08 19:07:21 crc kubenswrapper[4988]: I1008 19:07:21.895365 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-sz5l9" Oct 08 19:07:21 crc kubenswrapper[4988]: I1008 19:07:21.925221 4988 scope.go:117] "RemoveContainer" containerID="90d8a1960cc20709b55e8398ac082ad4c14e737a1b2e2db6b420dcab9d888e7d" Oct 08 19:07:21 crc kubenswrapper[4988]: I1008 19:07:21.950726 4988 scope.go:117] "RemoveContainer" containerID="34a6dc939250917cd74b220199d805d3b6253b5b055ffec3b954c30f128708e3" Oct 08 19:07:22 crc kubenswrapper[4988]: I1008 19:07:22.005586 4988 scope.go:117] "RemoveContainer" containerID="74fe37312dd863ad7ba7e6afa5ad23adff128a4bdea75e1f6d13e020061d4bb4" Oct 08 19:07:22 crc kubenswrapper[4988]: E1008 19:07:22.007497 4988 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"74fe37312dd863ad7ba7e6afa5ad23adff128a4bdea75e1f6d13e020061d4bb4\": container with ID starting with 74fe37312dd863ad7ba7e6afa5ad23adff128a4bdea75e1f6d13e020061d4bb4 not found: ID does not exist" containerID="74fe37312dd863ad7ba7e6afa5ad23adff128a4bdea75e1f6d13e020061d4bb4" Oct 08 19:07:22 crc kubenswrapper[4988]: I1008 19:07:22.007587 4988 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"74fe37312dd863ad7ba7e6afa5ad23adff128a4bdea75e1f6d13e020061d4bb4"} err="failed to get container status \"74fe37312dd863ad7ba7e6afa5ad23adff128a4bdea75e1f6d13e020061d4bb4\": rpc error: code = NotFound desc = could not find container \"74fe37312dd863ad7ba7e6afa5ad23adff128a4bdea75e1f6d13e020061d4bb4\": container with ID starting with 74fe37312dd863ad7ba7e6afa5ad23adff128a4bdea75e1f6d13e020061d4bb4 not found: ID does not exist" Oct 08 19:07:22 crc kubenswrapper[4988]: I1008 19:07:22.007616 4988 scope.go:117] "RemoveContainer" containerID="90d8a1960cc20709b55e8398ac082ad4c14e737a1b2e2db6b420dcab9d888e7d" Oct 08 19:07:22 crc kubenswrapper[4988]: E1008 19:07:22.008018 4988 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"90d8a1960cc20709b55e8398ac082ad4c14e737a1b2e2db6b420dcab9d888e7d\": container with ID starting with 90d8a1960cc20709b55e8398ac082ad4c14e737a1b2e2db6b420dcab9d888e7d not found: ID does not exist" containerID="90d8a1960cc20709b55e8398ac082ad4c14e737a1b2e2db6b420dcab9d888e7d" Oct 08 19:07:22 crc kubenswrapper[4988]: I1008 19:07:22.008097 4988 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"90d8a1960cc20709b55e8398ac082ad4c14e737a1b2e2db6b420dcab9d888e7d"} err="failed to get container status \"90d8a1960cc20709b55e8398ac082ad4c14e737a1b2e2db6b420dcab9d888e7d\": rpc error: code = NotFound desc = could not find container \"90d8a1960cc20709b55e8398ac082ad4c14e737a1b2e2db6b420dcab9d888e7d\": container with ID starting with 90d8a1960cc20709b55e8398ac082ad4c14e737a1b2e2db6b420dcab9d888e7d not found: ID does not exist" Oct 08 19:07:22 crc kubenswrapper[4988]: I1008 19:07:22.008151 4988 scope.go:117] "RemoveContainer" containerID="34a6dc939250917cd74b220199d805d3b6253b5b055ffec3b954c30f128708e3" Oct 08 19:07:22 crc kubenswrapper[4988]: E1008 19:07:22.008816 4988 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"34a6dc939250917cd74b220199d805d3b6253b5b055ffec3b954c30f128708e3\": container with ID starting with 34a6dc939250917cd74b220199d805d3b6253b5b055ffec3b954c30f128708e3 not found: ID does not exist" containerID="34a6dc939250917cd74b220199d805d3b6253b5b055ffec3b954c30f128708e3" Oct 08 19:07:22 crc kubenswrapper[4988]: I1008 19:07:22.008865 4988 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"34a6dc939250917cd74b220199d805d3b6253b5b055ffec3b954c30f128708e3"} err="failed to get container status \"34a6dc939250917cd74b220199d805d3b6253b5b055ffec3b954c30f128708e3\": rpc error: code = NotFound desc = could not find container \"34a6dc939250917cd74b220199d805d3b6253b5b055ffec3b954c30f128708e3\": container with ID starting with 34a6dc939250917cd74b220199d805d3b6253b5b055ffec3b954c30f128708e3 not found: ID does not exist" Oct 08 19:07:22 crc kubenswrapper[4988]: I1008 19:07:22.009718 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-sz5l9"] Oct 08 19:07:22 crc kubenswrapper[4988]: I1008 19:07:22.015669 4988 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-sz5l9"] Oct 08 19:07:23 crc kubenswrapper[4988]: I1008 19:07:23.257311 4988 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6b4d243b-e683-4688-8044-a729a62a2cb3" path="/var/lib/kubelet/pods/6b4d243b-e683-4688-8044-a729a62a2cb3/volumes" Oct 08 19:07:23 crc kubenswrapper[4988]: I1008 19:07:23.338518 4988 patch_prober.go:28] interesting pod/machine-config-daemon-nm54f container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 19:07:23 crc kubenswrapper[4988]: I1008 19:07:23.338605 4988 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 19:07:53 crc kubenswrapper[4988]: I1008 19:07:53.338187 4988 patch_prober.go:28] interesting pod/machine-config-daemon-nm54f container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 19:07:53 crc kubenswrapper[4988]: I1008 19:07:53.338872 4988 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 19:08:23 crc kubenswrapper[4988]: I1008 19:08:23.337671 4988 patch_prober.go:28] interesting pod/machine-config-daemon-nm54f container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 19:08:23 crc kubenswrapper[4988]: I1008 19:08:23.338438 4988 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 19:08:23 crc kubenswrapper[4988]: I1008 19:08:23.338517 4988 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" Oct 08 19:08:23 crc kubenswrapper[4988]: I1008 19:08:23.340823 4988 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"482cb9ffb14e123461aee7ddc866cf73ccfa365d4521f4ee6abf4f22170a86b2"} pod="openshift-machine-config-operator/machine-config-daemon-nm54f" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 08 19:08:23 crc kubenswrapper[4988]: I1008 19:08:23.340981 4988 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" containerName="machine-config-daemon" containerID="cri-o://482cb9ffb14e123461aee7ddc866cf73ccfa365d4521f4ee6abf4f22170a86b2" gracePeriod=600 Oct 08 19:08:23 crc kubenswrapper[4988]: I1008 19:08:23.493812 4988 generic.go:334] "Generic (PLEG): container finished" podID="f328a20b-5283-407e-8a93-2ab2da599f18" containerID="482cb9ffb14e123461aee7ddc866cf73ccfa365d4521f4ee6abf4f22170a86b2" exitCode=0 Oct 08 19:08:23 crc kubenswrapper[4988]: I1008 19:08:23.493904 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" event={"ID":"f328a20b-5283-407e-8a93-2ab2da599f18","Type":"ContainerDied","Data":"482cb9ffb14e123461aee7ddc866cf73ccfa365d4521f4ee6abf4f22170a86b2"} Oct 08 19:08:23 crc kubenswrapper[4988]: I1008 19:08:23.494215 4988 scope.go:117] "RemoveContainer" containerID="fea33312e34e76e64c798028f94adde7baabd667aa354082b2aebdb7e50779da" Oct 08 19:08:24 crc kubenswrapper[4988]: I1008 19:08:24.505036 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" event={"ID":"f328a20b-5283-407e-8a93-2ab2da599f18","Type":"ContainerStarted","Data":"af2043d46f0bbf33ed02542990cdf54feefa63a7c8331b6a70288efbb7b7a3ae"} Oct 08 19:08:38 crc kubenswrapper[4988]: I1008 19:08:38.246084 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-ksltc"] Oct 08 19:08:38 crc kubenswrapper[4988]: E1008 19:08:38.247247 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6b4d243b-e683-4688-8044-a729a62a2cb3" containerName="extract-utilities" Oct 08 19:08:38 crc kubenswrapper[4988]: I1008 19:08:38.247271 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="6b4d243b-e683-4688-8044-a729a62a2cb3" containerName="extract-utilities" Oct 08 19:08:38 crc kubenswrapper[4988]: E1008 19:08:38.247295 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6b4d243b-e683-4688-8044-a729a62a2cb3" containerName="extract-content" Oct 08 19:08:38 crc kubenswrapper[4988]: I1008 19:08:38.247308 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="6b4d243b-e683-4688-8044-a729a62a2cb3" containerName="extract-content" Oct 08 19:08:38 crc kubenswrapper[4988]: E1008 19:08:38.247321 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6b4d243b-e683-4688-8044-a729a62a2cb3" containerName="registry-server" Oct 08 19:08:38 crc kubenswrapper[4988]: I1008 19:08:38.247334 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="6b4d243b-e683-4688-8044-a729a62a2cb3" containerName="registry-server" Oct 08 19:08:38 crc kubenswrapper[4988]: I1008 19:08:38.247622 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="6b4d243b-e683-4688-8044-a729a62a2cb3" containerName="registry-server" Oct 08 19:08:38 crc kubenswrapper[4988]: I1008 19:08:38.249448 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-ksltc" Oct 08 19:08:38 crc kubenswrapper[4988]: I1008 19:08:38.266095 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-ksltc"] Oct 08 19:08:38 crc kubenswrapper[4988]: I1008 19:08:38.385275 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e3b1b754-1ed3-465e-9479-3b22a506f8d0-catalog-content\") pod \"redhat-marketplace-ksltc\" (UID: \"e3b1b754-1ed3-465e-9479-3b22a506f8d0\") " pod="openshift-marketplace/redhat-marketplace-ksltc" Oct 08 19:08:38 crc kubenswrapper[4988]: I1008 19:08:38.385534 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tlmkb\" (UniqueName: \"kubernetes.io/projected/e3b1b754-1ed3-465e-9479-3b22a506f8d0-kube-api-access-tlmkb\") pod \"redhat-marketplace-ksltc\" (UID: \"e3b1b754-1ed3-465e-9479-3b22a506f8d0\") " pod="openshift-marketplace/redhat-marketplace-ksltc" Oct 08 19:08:38 crc kubenswrapper[4988]: I1008 19:08:38.387289 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e3b1b754-1ed3-465e-9479-3b22a506f8d0-utilities\") pod \"redhat-marketplace-ksltc\" (UID: \"e3b1b754-1ed3-465e-9479-3b22a506f8d0\") " pod="openshift-marketplace/redhat-marketplace-ksltc" Oct 08 19:08:38 crc kubenswrapper[4988]: I1008 19:08:38.489799 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e3b1b754-1ed3-465e-9479-3b22a506f8d0-utilities\") pod \"redhat-marketplace-ksltc\" (UID: \"e3b1b754-1ed3-465e-9479-3b22a506f8d0\") " pod="openshift-marketplace/redhat-marketplace-ksltc" Oct 08 19:08:38 crc kubenswrapper[4988]: I1008 19:08:38.489954 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e3b1b754-1ed3-465e-9479-3b22a506f8d0-catalog-content\") pod \"redhat-marketplace-ksltc\" (UID: \"e3b1b754-1ed3-465e-9479-3b22a506f8d0\") " pod="openshift-marketplace/redhat-marketplace-ksltc" Oct 08 19:08:38 crc kubenswrapper[4988]: I1008 19:08:38.490034 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tlmkb\" (UniqueName: \"kubernetes.io/projected/e3b1b754-1ed3-465e-9479-3b22a506f8d0-kube-api-access-tlmkb\") pod \"redhat-marketplace-ksltc\" (UID: \"e3b1b754-1ed3-465e-9479-3b22a506f8d0\") " pod="openshift-marketplace/redhat-marketplace-ksltc" Oct 08 19:08:38 crc kubenswrapper[4988]: I1008 19:08:38.490511 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e3b1b754-1ed3-465e-9479-3b22a506f8d0-utilities\") pod \"redhat-marketplace-ksltc\" (UID: \"e3b1b754-1ed3-465e-9479-3b22a506f8d0\") " pod="openshift-marketplace/redhat-marketplace-ksltc" Oct 08 19:08:38 crc kubenswrapper[4988]: I1008 19:08:38.490563 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e3b1b754-1ed3-465e-9479-3b22a506f8d0-catalog-content\") pod \"redhat-marketplace-ksltc\" (UID: \"e3b1b754-1ed3-465e-9479-3b22a506f8d0\") " pod="openshift-marketplace/redhat-marketplace-ksltc" Oct 08 19:08:38 crc kubenswrapper[4988]: I1008 19:08:38.513519 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tlmkb\" (UniqueName: \"kubernetes.io/projected/e3b1b754-1ed3-465e-9479-3b22a506f8d0-kube-api-access-tlmkb\") pod \"redhat-marketplace-ksltc\" (UID: \"e3b1b754-1ed3-465e-9479-3b22a506f8d0\") " pod="openshift-marketplace/redhat-marketplace-ksltc" Oct 08 19:08:38 crc kubenswrapper[4988]: I1008 19:08:38.594912 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-ksltc" Oct 08 19:08:39 crc kubenswrapper[4988]: I1008 19:08:39.029815 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-ksltc"] Oct 08 19:08:39 crc kubenswrapper[4988]: I1008 19:08:39.646763 4988 generic.go:334] "Generic (PLEG): container finished" podID="e3b1b754-1ed3-465e-9479-3b22a506f8d0" containerID="c05e989d6f9b2fecfd02990af7edf428231eb619746759121ede6fa1c5427cc1" exitCode=0 Oct 08 19:08:39 crc kubenswrapper[4988]: I1008 19:08:39.646835 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-ksltc" event={"ID":"e3b1b754-1ed3-465e-9479-3b22a506f8d0","Type":"ContainerDied","Data":"c05e989d6f9b2fecfd02990af7edf428231eb619746759121ede6fa1c5427cc1"} Oct 08 19:08:39 crc kubenswrapper[4988]: I1008 19:08:39.646944 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-ksltc" event={"ID":"e3b1b754-1ed3-465e-9479-3b22a506f8d0","Type":"ContainerStarted","Data":"901dbbf8594a1f0468c88419685b998bc237c473795c0eb3b9911dae0a10186a"} Oct 08 19:08:40 crc kubenswrapper[4988]: I1008 19:08:40.040475 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-72nzl"] Oct 08 19:08:40 crc kubenswrapper[4988]: I1008 19:08:40.046122 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-72nzl" Oct 08 19:08:40 crc kubenswrapper[4988]: I1008 19:08:40.053492 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-72nzl"] Oct 08 19:08:40 crc kubenswrapper[4988]: I1008 19:08:40.216588 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/70935226-3356-4142-8e07-27da09ba58fa-utilities\") pod \"redhat-operators-72nzl\" (UID: \"70935226-3356-4142-8e07-27da09ba58fa\") " pod="openshift-marketplace/redhat-operators-72nzl" Oct 08 19:08:40 crc kubenswrapper[4988]: I1008 19:08:40.216715 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/70935226-3356-4142-8e07-27da09ba58fa-catalog-content\") pod \"redhat-operators-72nzl\" (UID: \"70935226-3356-4142-8e07-27da09ba58fa\") " pod="openshift-marketplace/redhat-operators-72nzl" Oct 08 19:08:40 crc kubenswrapper[4988]: I1008 19:08:40.216811 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8q54h\" (UniqueName: \"kubernetes.io/projected/70935226-3356-4142-8e07-27da09ba58fa-kube-api-access-8q54h\") pod \"redhat-operators-72nzl\" (UID: \"70935226-3356-4142-8e07-27da09ba58fa\") " pod="openshift-marketplace/redhat-operators-72nzl" Oct 08 19:08:40 crc kubenswrapper[4988]: I1008 19:08:40.318103 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/70935226-3356-4142-8e07-27da09ba58fa-catalog-content\") pod \"redhat-operators-72nzl\" (UID: \"70935226-3356-4142-8e07-27da09ba58fa\") " pod="openshift-marketplace/redhat-operators-72nzl" Oct 08 19:08:40 crc kubenswrapper[4988]: I1008 19:08:40.318214 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8q54h\" (UniqueName: \"kubernetes.io/projected/70935226-3356-4142-8e07-27da09ba58fa-kube-api-access-8q54h\") pod \"redhat-operators-72nzl\" (UID: \"70935226-3356-4142-8e07-27da09ba58fa\") " pod="openshift-marketplace/redhat-operators-72nzl" Oct 08 19:08:40 crc kubenswrapper[4988]: I1008 19:08:40.319199 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/70935226-3356-4142-8e07-27da09ba58fa-utilities\") pod \"redhat-operators-72nzl\" (UID: \"70935226-3356-4142-8e07-27da09ba58fa\") " pod="openshift-marketplace/redhat-operators-72nzl" Oct 08 19:08:40 crc kubenswrapper[4988]: I1008 19:08:40.319309 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/70935226-3356-4142-8e07-27da09ba58fa-utilities\") pod \"redhat-operators-72nzl\" (UID: \"70935226-3356-4142-8e07-27da09ba58fa\") " pod="openshift-marketplace/redhat-operators-72nzl" Oct 08 19:08:40 crc kubenswrapper[4988]: I1008 19:08:40.319927 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/70935226-3356-4142-8e07-27da09ba58fa-catalog-content\") pod \"redhat-operators-72nzl\" (UID: \"70935226-3356-4142-8e07-27da09ba58fa\") " pod="openshift-marketplace/redhat-operators-72nzl" Oct 08 19:08:40 crc kubenswrapper[4988]: I1008 19:08:40.347171 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8q54h\" (UniqueName: \"kubernetes.io/projected/70935226-3356-4142-8e07-27da09ba58fa-kube-api-access-8q54h\") pod \"redhat-operators-72nzl\" (UID: \"70935226-3356-4142-8e07-27da09ba58fa\") " pod="openshift-marketplace/redhat-operators-72nzl" Oct 08 19:08:40 crc kubenswrapper[4988]: I1008 19:08:40.400400 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-72nzl" Oct 08 19:08:40 crc kubenswrapper[4988]: I1008 19:08:40.651153 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-72nzl"] Oct 08 19:08:40 crc kubenswrapper[4988]: I1008 19:08:40.660211 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-ksltc" event={"ID":"e3b1b754-1ed3-465e-9479-3b22a506f8d0","Type":"ContainerStarted","Data":"d69531e1a9d302c4d54bfdc7135bdce6364d7fe00e1edb0a8bc932550c7ff0fb"} Oct 08 19:08:40 crc kubenswrapper[4988]: W1008 19:08:40.710497 4988 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod70935226_3356_4142_8e07_27da09ba58fa.slice/crio-3bb60e0c522433fa06f47d049e2cfa8ca1ccdaca15af110839fb9fc0da7d303a WatchSource:0}: Error finding container 3bb60e0c522433fa06f47d049e2cfa8ca1ccdaca15af110839fb9fc0da7d303a: Status 404 returned error can't find the container with id 3bb60e0c522433fa06f47d049e2cfa8ca1ccdaca15af110839fb9fc0da7d303a Oct 08 19:08:41 crc kubenswrapper[4988]: I1008 19:08:41.437483 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-8wmdb"] Oct 08 19:08:41 crc kubenswrapper[4988]: I1008 19:08:41.441269 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-8wmdb" Oct 08 19:08:41 crc kubenswrapper[4988]: I1008 19:08:41.451335 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-8wmdb"] Oct 08 19:08:41 crc kubenswrapper[4988]: I1008 19:08:41.538129 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a9451896-2cc4-43a2-bae2-fbebae3e96da-catalog-content\") pod \"certified-operators-8wmdb\" (UID: \"a9451896-2cc4-43a2-bae2-fbebae3e96da\") " pod="openshift-marketplace/certified-operators-8wmdb" Oct 08 19:08:41 crc kubenswrapper[4988]: I1008 19:08:41.538437 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a9451896-2cc4-43a2-bae2-fbebae3e96da-utilities\") pod \"certified-operators-8wmdb\" (UID: \"a9451896-2cc4-43a2-bae2-fbebae3e96da\") " pod="openshift-marketplace/certified-operators-8wmdb" Oct 08 19:08:41 crc kubenswrapper[4988]: I1008 19:08:41.538571 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wmzv8\" (UniqueName: \"kubernetes.io/projected/a9451896-2cc4-43a2-bae2-fbebae3e96da-kube-api-access-wmzv8\") pod \"certified-operators-8wmdb\" (UID: \"a9451896-2cc4-43a2-bae2-fbebae3e96da\") " pod="openshift-marketplace/certified-operators-8wmdb" Oct 08 19:08:41 crc kubenswrapper[4988]: I1008 19:08:41.639521 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a9451896-2cc4-43a2-bae2-fbebae3e96da-utilities\") pod \"certified-operators-8wmdb\" (UID: \"a9451896-2cc4-43a2-bae2-fbebae3e96da\") " pod="openshift-marketplace/certified-operators-8wmdb" Oct 08 19:08:41 crc kubenswrapper[4988]: I1008 19:08:41.639603 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wmzv8\" (UniqueName: \"kubernetes.io/projected/a9451896-2cc4-43a2-bae2-fbebae3e96da-kube-api-access-wmzv8\") pod \"certified-operators-8wmdb\" (UID: \"a9451896-2cc4-43a2-bae2-fbebae3e96da\") " pod="openshift-marketplace/certified-operators-8wmdb" Oct 08 19:08:41 crc kubenswrapper[4988]: I1008 19:08:41.639659 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a9451896-2cc4-43a2-bae2-fbebae3e96da-catalog-content\") pod \"certified-operators-8wmdb\" (UID: \"a9451896-2cc4-43a2-bae2-fbebae3e96da\") " pod="openshift-marketplace/certified-operators-8wmdb" Oct 08 19:08:41 crc kubenswrapper[4988]: I1008 19:08:41.640202 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a9451896-2cc4-43a2-bae2-fbebae3e96da-catalog-content\") pod \"certified-operators-8wmdb\" (UID: \"a9451896-2cc4-43a2-bae2-fbebae3e96da\") " pod="openshift-marketplace/certified-operators-8wmdb" Oct 08 19:08:41 crc kubenswrapper[4988]: I1008 19:08:41.640205 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a9451896-2cc4-43a2-bae2-fbebae3e96da-utilities\") pod \"certified-operators-8wmdb\" (UID: \"a9451896-2cc4-43a2-bae2-fbebae3e96da\") " pod="openshift-marketplace/certified-operators-8wmdb" Oct 08 19:08:41 crc kubenswrapper[4988]: I1008 19:08:41.675186 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wmzv8\" (UniqueName: \"kubernetes.io/projected/a9451896-2cc4-43a2-bae2-fbebae3e96da-kube-api-access-wmzv8\") pod \"certified-operators-8wmdb\" (UID: \"a9451896-2cc4-43a2-bae2-fbebae3e96da\") " pod="openshift-marketplace/certified-operators-8wmdb" Oct 08 19:08:41 crc kubenswrapper[4988]: I1008 19:08:41.677685 4988 generic.go:334] "Generic (PLEG): container finished" podID="e3b1b754-1ed3-465e-9479-3b22a506f8d0" containerID="d69531e1a9d302c4d54bfdc7135bdce6364d7fe00e1edb0a8bc932550c7ff0fb" exitCode=0 Oct 08 19:08:41 crc kubenswrapper[4988]: I1008 19:08:41.677745 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-ksltc" event={"ID":"e3b1b754-1ed3-465e-9479-3b22a506f8d0","Type":"ContainerDied","Data":"d69531e1a9d302c4d54bfdc7135bdce6364d7fe00e1edb0a8bc932550c7ff0fb"} Oct 08 19:08:41 crc kubenswrapper[4988]: I1008 19:08:41.677769 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-ksltc" event={"ID":"e3b1b754-1ed3-465e-9479-3b22a506f8d0","Type":"ContainerStarted","Data":"6be42789d4b7b7c8c6243ccbd24d48a2fcf3c77bea2f5a8164d4496e6179c218"} Oct 08 19:08:41 crc kubenswrapper[4988]: I1008 19:08:41.686833 4988 generic.go:334] "Generic (PLEG): container finished" podID="70935226-3356-4142-8e07-27da09ba58fa" containerID="159519ad6ff6c1f641f5a9fc5031c17ee4eb0eb634b9bbd864e4a2cec62f9cae" exitCode=0 Oct 08 19:08:41 crc kubenswrapper[4988]: I1008 19:08:41.686860 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-72nzl" event={"ID":"70935226-3356-4142-8e07-27da09ba58fa","Type":"ContainerDied","Data":"159519ad6ff6c1f641f5a9fc5031c17ee4eb0eb634b9bbd864e4a2cec62f9cae"} Oct 08 19:08:41 crc kubenswrapper[4988]: I1008 19:08:41.686876 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-72nzl" event={"ID":"70935226-3356-4142-8e07-27da09ba58fa","Type":"ContainerStarted","Data":"3bb60e0c522433fa06f47d049e2cfa8ca1ccdaca15af110839fb9fc0da7d303a"} Oct 08 19:08:41 crc kubenswrapper[4988]: I1008 19:08:41.735020 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-ksltc" podStartSLOduration=2.080051102 podStartE2EDuration="3.735005122s" podCreationTimestamp="2025-10-08 19:08:38 +0000 UTC" firstStartedPulling="2025-10-08 19:08:39.649079628 +0000 UTC m=+3465.098922428" lastFinishedPulling="2025-10-08 19:08:41.304033678 +0000 UTC m=+3466.753876448" observedRunningTime="2025-10-08 19:08:41.709686625 +0000 UTC m=+3467.159529435" watchObservedRunningTime="2025-10-08 19:08:41.735005122 +0000 UTC m=+3467.184847892" Oct 08 19:08:41 crc kubenswrapper[4988]: I1008 19:08:41.828136 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-8wmdb" Oct 08 19:08:42 crc kubenswrapper[4988]: I1008 19:08:42.248951 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-8wmdb"] Oct 08 19:08:42 crc kubenswrapper[4988]: W1008 19:08:42.259362 4988 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda9451896_2cc4_43a2_bae2_fbebae3e96da.slice/crio-346dc3b09ae9e65f7fe182dc7ba09429496fd28d20eb943ca6d32908afa4dfbe WatchSource:0}: Error finding container 346dc3b09ae9e65f7fe182dc7ba09429496fd28d20eb943ca6d32908afa4dfbe: Status 404 returned error can't find the container with id 346dc3b09ae9e65f7fe182dc7ba09429496fd28d20eb943ca6d32908afa4dfbe Oct 08 19:08:42 crc kubenswrapper[4988]: I1008 19:08:42.696948 4988 generic.go:334] "Generic (PLEG): container finished" podID="a9451896-2cc4-43a2-bae2-fbebae3e96da" containerID="69cf88d50d661821115675be8c45acb6fc92a3aba9c3ef689b73a82409020485" exitCode=0 Oct 08 19:08:42 crc kubenswrapper[4988]: I1008 19:08:42.697064 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8wmdb" event={"ID":"a9451896-2cc4-43a2-bae2-fbebae3e96da","Type":"ContainerDied","Data":"69cf88d50d661821115675be8c45acb6fc92a3aba9c3ef689b73a82409020485"} Oct 08 19:08:42 crc kubenswrapper[4988]: I1008 19:08:42.697130 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8wmdb" event={"ID":"a9451896-2cc4-43a2-bae2-fbebae3e96da","Type":"ContainerStarted","Data":"346dc3b09ae9e65f7fe182dc7ba09429496fd28d20eb943ca6d32908afa4dfbe"} Oct 08 19:08:42 crc kubenswrapper[4988]: I1008 19:08:42.707164 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-72nzl" event={"ID":"70935226-3356-4142-8e07-27da09ba58fa","Type":"ContainerStarted","Data":"d68d69603719a32351931f815e28cf592fbc1239c42dcb89cc8b6925170d9fcb"} Oct 08 19:08:43 crc kubenswrapper[4988]: I1008 19:08:43.719291 4988 generic.go:334] "Generic (PLEG): container finished" podID="70935226-3356-4142-8e07-27da09ba58fa" containerID="d68d69603719a32351931f815e28cf592fbc1239c42dcb89cc8b6925170d9fcb" exitCode=0 Oct 08 19:08:43 crc kubenswrapper[4988]: I1008 19:08:43.719454 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-72nzl" event={"ID":"70935226-3356-4142-8e07-27da09ba58fa","Type":"ContainerDied","Data":"d68d69603719a32351931f815e28cf592fbc1239c42dcb89cc8b6925170d9fcb"} Oct 08 19:08:43 crc kubenswrapper[4988]: I1008 19:08:43.723299 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8wmdb" event={"ID":"a9451896-2cc4-43a2-bae2-fbebae3e96da","Type":"ContainerStarted","Data":"3605aa64ce8219bf64a4f6fa833fd5bd86a35f7344ce8ca728bf727e823e6926"} Oct 08 19:08:43 crc kubenswrapper[4988]: E1008 19:08:43.916404 4988 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda9451896_2cc4_43a2_bae2_fbebae3e96da.slice/crio-conmon-3605aa64ce8219bf64a4f6fa833fd5bd86a35f7344ce8ca728bf727e823e6926.scope\": RecentStats: unable to find data in memory cache]" Oct 08 19:08:44 crc kubenswrapper[4988]: I1008 19:08:44.734981 4988 generic.go:334] "Generic (PLEG): container finished" podID="a9451896-2cc4-43a2-bae2-fbebae3e96da" containerID="3605aa64ce8219bf64a4f6fa833fd5bd86a35f7344ce8ca728bf727e823e6926" exitCode=0 Oct 08 19:08:44 crc kubenswrapper[4988]: I1008 19:08:44.735119 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8wmdb" event={"ID":"a9451896-2cc4-43a2-bae2-fbebae3e96da","Type":"ContainerDied","Data":"3605aa64ce8219bf64a4f6fa833fd5bd86a35f7344ce8ca728bf727e823e6926"} Oct 08 19:08:44 crc kubenswrapper[4988]: I1008 19:08:44.739146 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-72nzl" event={"ID":"70935226-3356-4142-8e07-27da09ba58fa","Type":"ContainerStarted","Data":"0d637b0843bf4651c3078ccbb0db3645594129aa4d52b0f547150c3775a27592"} Oct 08 19:08:44 crc kubenswrapper[4988]: I1008 19:08:44.784022 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-72nzl" podStartSLOduration=2.224248462 podStartE2EDuration="4.783999087s" podCreationTimestamp="2025-10-08 19:08:40 +0000 UTC" firstStartedPulling="2025-10-08 19:08:41.688709786 +0000 UTC m=+3467.138552556" lastFinishedPulling="2025-10-08 19:08:44.248460371 +0000 UTC m=+3469.698303181" observedRunningTime="2025-10-08 19:08:44.778960107 +0000 UTC m=+3470.228802887" watchObservedRunningTime="2025-10-08 19:08:44.783999087 +0000 UTC m=+3470.233841877" Oct 08 19:08:45 crc kubenswrapper[4988]: I1008 19:08:45.753554 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8wmdb" event={"ID":"a9451896-2cc4-43a2-bae2-fbebae3e96da","Type":"ContainerStarted","Data":"b46ed966cab813489a1473b07aaa164bc514b78fdf99ebdb9e4e66d6a5afcba4"} Oct 08 19:08:45 crc kubenswrapper[4988]: I1008 19:08:45.778841 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-8wmdb" podStartSLOduration=2.094522277 podStartE2EDuration="4.77882461s" podCreationTimestamp="2025-10-08 19:08:41 +0000 UTC" firstStartedPulling="2025-10-08 19:08:42.701504063 +0000 UTC m=+3468.151346843" lastFinishedPulling="2025-10-08 19:08:45.385806386 +0000 UTC m=+3470.835649176" observedRunningTime="2025-10-08 19:08:45.773465259 +0000 UTC m=+3471.223308049" watchObservedRunningTime="2025-10-08 19:08:45.77882461 +0000 UTC m=+3471.228667380" Oct 08 19:08:48 crc kubenswrapper[4988]: I1008 19:08:48.595549 4988 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-ksltc" Oct 08 19:08:48 crc kubenswrapper[4988]: I1008 19:08:48.595601 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-ksltc" Oct 08 19:08:48 crc kubenswrapper[4988]: I1008 19:08:48.680661 4988 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-ksltc" Oct 08 19:08:48 crc kubenswrapper[4988]: I1008 19:08:48.832543 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-ksltc" Oct 08 19:08:50 crc kubenswrapper[4988]: I1008 19:08:50.027023 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-ksltc"] Oct 08 19:08:50 crc kubenswrapper[4988]: I1008 19:08:50.401630 4988 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-72nzl" Oct 08 19:08:50 crc kubenswrapper[4988]: I1008 19:08:50.401751 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-72nzl" Oct 08 19:08:50 crc kubenswrapper[4988]: I1008 19:08:50.478620 4988 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-72nzl" Oct 08 19:08:50 crc kubenswrapper[4988]: I1008 19:08:50.802934 4988 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-ksltc" podUID="e3b1b754-1ed3-465e-9479-3b22a506f8d0" containerName="registry-server" containerID="cri-o://6be42789d4b7b7c8c6243ccbd24d48a2fcf3c77bea2f5a8164d4496e6179c218" gracePeriod=2 Oct 08 19:08:50 crc kubenswrapper[4988]: I1008 19:08:50.905250 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-72nzl" Oct 08 19:08:51 crc kubenswrapper[4988]: I1008 19:08:51.236554 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-ksltc" Oct 08 19:08:51 crc kubenswrapper[4988]: I1008 19:08:51.290067 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tlmkb\" (UniqueName: \"kubernetes.io/projected/e3b1b754-1ed3-465e-9479-3b22a506f8d0-kube-api-access-tlmkb\") pod \"e3b1b754-1ed3-465e-9479-3b22a506f8d0\" (UID: \"e3b1b754-1ed3-465e-9479-3b22a506f8d0\") " Oct 08 19:08:51 crc kubenswrapper[4988]: I1008 19:08:51.290199 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e3b1b754-1ed3-465e-9479-3b22a506f8d0-utilities\") pod \"e3b1b754-1ed3-465e-9479-3b22a506f8d0\" (UID: \"e3b1b754-1ed3-465e-9479-3b22a506f8d0\") " Oct 08 19:08:51 crc kubenswrapper[4988]: I1008 19:08:51.290255 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e3b1b754-1ed3-465e-9479-3b22a506f8d0-catalog-content\") pod \"e3b1b754-1ed3-465e-9479-3b22a506f8d0\" (UID: \"e3b1b754-1ed3-465e-9479-3b22a506f8d0\") " Oct 08 19:08:51 crc kubenswrapper[4988]: I1008 19:08:51.292536 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e3b1b754-1ed3-465e-9479-3b22a506f8d0-utilities" (OuterVolumeSpecName: "utilities") pod "e3b1b754-1ed3-465e-9479-3b22a506f8d0" (UID: "e3b1b754-1ed3-465e-9479-3b22a506f8d0"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 19:08:51 crc kubenswrapper[4988]: I1008 19:08:51.296921 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e3b1b754-1ed3-465e-9479-3b22a506f8d0-kube-api-access-tlmkb" (OuterVolumeSpecName: "kube-api-access-tlmkb") pod "e3b1b754-1ed3-465e-9479-3b22a506f8d0" (UID: "e3b1b754-1ed3-465e-9479-3b22a506f8d0"). InnerVolumeSpecName "kube-api-access-tlmkb". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 19:08:51 crc kubenswrapper[4988]: I1008 19:08:51.302802 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e3b1b754-1ed3-465e-9479-3b22a506f8d0-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "e3b1b754-1ed3-465e-9479-3b22a506f8d0" (UID: "e3b1b754-1ed3-465e-9479-3b22a506f8d0"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 19:08:51 crc kubenswrapper[4988]: I1008 19:08:51.391806 4988 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e3b1b754-1ed3-465e-9479-3b22a506f8d0-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 08 19:08:51 crc kubenswrapper[4988]: I1008 19:08:51.391857 4988 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tlmkb\" (UniqueName: \"kubernetes.io/projected/e3b1b754-1ed3-465e-9479-3b22a506f8d0-kube-api-access-tlmkb\") on node \"crc\" DevicePath \"\"" Oct 08 19:08:51 crc kubenswrapper[4988]: I1008 19:08:51.391868 4988 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e3b1b754-1ed3-465e-9479-3b22a506f8d0-utilities\") on node \"crc\" DevicePath \"\"" Oct 08 19:08:51 crc kubenswrapper[4988]: I1008 19:08:51.816153 4988 generic.go:334] "Generic (PLEG): container finished" podID="e3b1b754-1ed3-465e-9479-3b22a506f8d0" containerID="6be42789d4b7b7c8c6243ccbd24d48a2fcf3c77bea2f5a8164d4496e6179c218" exitCode=0 Oct 08 19:08:51 crc kubenswrapper[4988]: I1008 19:08:51.816252 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-ksltc" Oct 08 19:08:51 crc kubenswrapper[4988]: I1008 19:08:51.816346 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-ksltc" event={"ID":"e3b1b754-1ed3-465e-9479-3b22a506f8d0","Type":"ContainerDied","Data":"6be42789d4b7b7c8c6243ccbd24d48a2fcf3c77bea2f5a8164d4496e6179c218"} Oct 08 19:08:51 crc kubenswrapper[4988]: I1008 19:08:51.816427 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-ksltc" event={"ID":"e3b1b754-1ed3-465e-9479-3b22a506f8d0","Type":"ContainerDied","Data":"901dbbf8594a1f0468c88419685b998bc237c473795c0eb3b9911dae0a10186a"} Oct 08 19:08:51 crc kubenswrapper[4988]: I1008 19:08:51.816456 4988 scope.go:117] "RemoveContainer" containerID="6be42789d4b7b7c8c6243ccbd24d48a2fcf3c77bea2f5a8164d4496e6179c218" Oct 08 19:08:51 crc kubenswrapper[4988]: I1008 19:08:51.828995 4988 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-8wmdb" Oct 08 19:08:51 crc kubenswrapper[4988]: I1008 19:08:51.829037 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-8wmdb" Oct 08 19:08:51 crc kubenswrapper[4988]: I1008 19:08:51.856737 4988 scope.go:117] "RemoveContainer" containerID="d69531e1a9d302c4d54bfdc7135bdce6364d7fe00e1edb0a8bc932550c7ff0fb" Oct 08 19:08:51 crc kubenswrapper[4988]: I1008 19:08:51.865233 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-ksltc"] Oct 08 19:08:51 crc kubenswrapper[4988]: I1008 19:08:51.875070 4988 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-ksltc"] Oct 08 19:08:51 crc kubenswrapper[4988]: I1008 19:08:51.923709 4988 scope.go:117] "RemoveContainer" containerID="c05e989d6f9b2fecfd02990af7edf428231eb619746759121ede6fa1c5427cc1" Oct 08 19:08:51 crc kubenswrapper[4988]: I1008 19:08:51.925655 4988 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-8wmdb" Oct 08 19:08:51 crc kubenswrapper[4988]: I1008 19:08:51.957540 4988 scope.go:117] "RemoveContainer" containerID="6be42789d4b7b7c8c6243ccbd24d48a2fcf3c77bea2f5a8164d4496e6179c218" Oct 08 19:08:51 crc kubenswrapper[4988]: E1008 19:08:51.958326 4988 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6be42789d4b7b7c8c6243ccbd24d48a2fcf3c77bea2f5a8164d4496e6179c218\": container with ID starting with 6be42789d4b7b7c8c6243ccbd24d48a2fcf3c77bea2f5a8164d4496e6179c218 not found: ID does not exist" containerID="6be42789d4b7b7c8c6243ccbd24d48a2fcf3c77bea2f5a8164d4496e6179c218" Oct 08 19:08:51 crc kubenswrapper[4988]: I1008 19:08:51.958369 4988 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6be42789d4b7b7c8c6243ccbd24d48a2fcf3c77bea2f5a8164d4496e6179c218"} err="failed to get container status \"6be42789d4b7b7c8c6243ccbd24d48a2fcf3c77bea2f5a8164d4496e6179c218\": rpc error: code = NotFound desc = could not find container \"6be42789d4b7b7c8c6243ccbd24d48a2fcf3c77bea2f5a8164d4496e6179c218\": container with ID starting with 6be42789d4b7b7c8c6243ccbd24d48a2fcf3c77bea2f5a8164d4496e6179c218 not found: ID does not exist" Oct 08 19:08:51 crc kubenswrapper[4988]: I1008 19:08:51.958408 4988 scope.go:117] "RemoveContainer" containerID="d69531e1a9d302c4d54bfdc7135bdce6364d7fe00e1edb0a8bc932550c7ff0fb" Oct 08 19:08:51 crc kubenswrapper[4988]: E1008 19:08:51.958977 4988 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d69531e1a9d302c4d54bfdc7135bdce6364d7fe00e1edb0a8bc932550c7ff0fb\": container with ID starting with d69531e1a9d302c4d54bfdc7135bdce6364d7fe00e1edb0a8bc932550c7ff0fb not found: ID does not exist" containerID="d69531e1a9d302c4d54bfdc7135bdce6364d7fe00e1edb0a8bc932550c7ff0fb" Oct 08 19:08:51 crc kubenswrapper[4988]: I1008 19:08:51.959029 4988 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d69531e1a9d302c4d54bfdc7135bdce6364d7fe00e1edb0a8bc932550c7ff0fb"} err="failed to get container status \"d69531e1a9d302c4d54bfdc7135bdce6364d7fe00e1edb0a8bc932550c7ff0fb\": rpc error: code = NotFound desc = could not find container \"d69531e1a9d302c4d54bfdc7135bdce6364d7fe00e1edb0a8bc932550c7ff0fb\": container with ID starting with d69531e1a9d302c4d54bfdc7135bdce6364d7fe00e1edb0a8bc932550c7ff0fb not found: ID does not exist" Oct 08 19:08:51 crc kubenswrapper[4988]: I1008 19:08:51.959068 4988 scope.go:117] "RemoveContainer" containerID="c05e989d6f9b2fecfd02990af7edf428231eb619746759121ede6fa1c5427cc1" Oct 08 19:08:51 crc kubenswrapper[4988]: E1008 19:08:51.959601 4988 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c05e989d6f9b2fecfd02990af7edf428231eb619746759121ede6fa1c5427cc1\": container with ID starting with c05e989d6f9b2fecfd02990af7edf428231eb619746759121ede6fa1c5427cc1 not found: ID does not exist" containerID="c05e989d6f9b2fecfd02990af7edf428231eb619746759121ede6fa1c5427cc1" Oct 08 19:08:51 crc kubenswrapper[4988]: I1008 19:08:51.959626 4988 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c05e989d6f9b2fecfd02990af7edf428231eb619746759121ede6fa1c5427cc1"} err="failed to get container status \"c05e989d6f9b2fecfd02990af7edf428231eb619746759121ede6fa1c5427cc1\": rpc error: code = NotFound desc = could not find container \"c05e989d6f9b2fecfd02990af7edf428231eb619746759121ede6fa1c5427cc1\": container with ID starting with c05e989d6f9b2fecfd02990af7edf428231eb619746759121ede6fa1c5427cc1 not found: ID does not exist" Oct 08 19:08:52 crc kubenswrapper[4988]: I1008 19:08:52.830987 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-72nzl"] Oct 08 19:08:52 crc kubenswrapper[4988]: I1008 19:08:52.831332 4988 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-72nzl" podUID="70935226-3356-4142-8e07-27da09ba58fa" containerName="registry-server" containerID="cri-o://0d637b0843bf4651c3078ccbb0db3645594129aa4d52b0f547150c3775a27592" gracePeriod=2 Oct 08 19:08:52 crc kubenswrapper[4988]: I1008 19:08:52.898759 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-8wmdb" Oct 08 19:08:53 crc kubenswrapper[4988]: I1008 19:08:53.253483 4988 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e3b1b754-1ed3-465e-9479-3b22a506f8d0" path="/var/lib/kubelet/pods/e3b1b754-1ed3-465e-9479-3b22a506f8d0/volumes" Oct 08 19:08:54 crc kubenswrapper[4988]: I1008 19:08:54.259728 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-72nzl" Oct 08 19:08:54 crc kubenswrapper[4988]: I1008 19:08:54.361341 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/70935226-3356-4142-8e07-27da09ba58fa-utilities\") pod \"70935226-3356-4142-8e07-27da09ba58fa\" (UID: \"70935226-3356-4142-8e07-27da09ba58fa\") " Oct 08 19:08:54 crc kubenswrapper[4988]: I1008 19:08:54.361418 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8q54h\" (UniqueName: \"kubernetes.io/projected/70935226-3356-4142-8e07-27da09ba58fa-kube-api-access-8q54h\") pod \"70935226-3356-4142-8e07-27da09ba58fa\" (UID: \"70935226-3356-4142-8e07-27da09ba58fa\") " Oct 08 19:08:54 crc kubenswrapper[4988]: I1008 19:08:54.361496 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/70935226-3356-4142-8e07-27da09ba58fa-catalog-content\") pod \"70935226-3356-4142-8e07-27da09ba58fa\" (UID: \"70935226-3356-4142-8e07-27da09ba58fa\") " Oct 08 19:08:54 crc kubenswrapper[4988]: I1008 19:08:54.362570 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/70935226-3356-4142-8e07-27da09ba58fa-utilities" (OuterVolumeSpecName: "utilities") pod "70935226-3356-4142-8e07-27da09ba58fa" (UID: "70935226-3356-4142-8e07-27da09ba58fa"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 19:08:54 crc kubenswrapper[4988]: I1008 19:08:54.367126 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/70935226-3356-4142-8e07-27da09ba58fa-kube-api-access-8q54h" (OuterVolumeSpecName: "kube-api-access-8q54h") pod "70935226-3356-4142-8e07-27da09ba58fa" (UID: "70935226-3356-4142-8e07-27da09ba58fa"). InnerVolumeSpecName "kube-api-access-8q54h". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 19:08:54 crc kubenswrapper[4988]: I1008 19:08:54.462929 4988 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/70935226-3356-4142-8e07-27da09ba58fa-utilities\") on node \"crc\" DevicePath \"\"" Oct 08 19:08:54 crc kubenswrapper[4988]: I1008 19:08:54.462963 4988 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8q54h\" (UniqueName: \"kubernetes.io/projected/70935226-3356-4142-8e07-27da09ba58fa-kube-api-access-8q54h\") on node \"crc\" DevicePath \"\"" Oct 08 19:08:54 crc kubenswrapper[4988]: I1008 19:08:54.475687 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/70935226-3356-4142-8e07-27da09ba58fa-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "70935226-3356-4142-8e07-27da09ba58fa" (UID: "70935226-3356-4142-8e07-27da09ba58fa"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 19:08:54 crc kubenswrapper[4988]: I1008 19:08:54.564836 4988 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/70935226-3356-4142-8e07-27da09ba58fa-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 08 19:08:54 crc kubenswrapper[4988]: I1008 19:08:54.853477 4988 generic.go:334] "Generic (PLEG): container finished" podID="70935226-3356-4142-8e07-27da09ba58fa" containerID="0d637b0843bf4651c3078ccbb0db3645594129aa4d52b0f547150c3775a27592" exitCode=0 Oct 08 19:08:54 crc kubenswrapper[4988]: I1008 19:08:54.853540 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-72nzl" event={"ID":"70935226-3356-4142-8e07-27da09ba58fa","Type":"ContainerDied","Data":"0d637b0843bf4651c3078ccbb0db3645594129aa4d52b0f547150c3775a27592"} Oct 08 19:08:54 crc kubenswrapper[4988]: I1008 19:08:54.853580 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-72nzl" event={"ID":"70935226-3356-4142-8e07-27da09ba58fa","Type":"ContainerDied","Data":"3bb60e0c522433fa06f47d049e2cfa8ca1ccdaca15af110839fb9fc0da7d303a"} Oct 08 19:08:54 crc kubenswrapper[4988]: I1008 19:08:54.853608 4988 scope.go:117] "RemoveContainer" containerID="0d637b0843bf4651c3078ccbb0db3645594129aa4d52b0f547150c3775a27592" Oct 08 19:08:54 crc kubenswrapper[4988]: I1008 19:08:54.853622 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-72nzl" Oct 08 19:08:54 crc kubenswrapper[4988]: I1008 19:08:54.884792 4988 scope.go:117] "RemoveContainer" containerID="d68d69603719a32351931f815e28cf592fbc1239c42dcb89cc8b6925170d9fcb" Oct 08 19:08:54 crc kubenswrapper[4988]: I1008 19:08:54.903094 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-72nzl"] Oct 08 19:08:54 crc kubenswrapper[4988]: I1008 19:08:54.912985 4988 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-72nzl"] Oct 08 19:08:54 crc kubenswrapper[4988]: I1008 19:08:54.926470 4988 scope.go:117] "RemoveContainer" containerID="159519ad6ff6c1f641f5a9fc5031c17ee4eb0eb634b9bbd864e4a2cec62f9cae" Oct 08 19:08:54 crc kubenswrapper[4988]: I1008 19:08:54.950853 4988 scope.go:117] "RemoveContainer" containerID="0d637b0843bf4651c3078ccbb0db3645594129aa4d52b0f547150c3775a27592" Oct 08 19:08:54 crc kubenswrapper[4988]: E1008 19:08:54.951493 4988 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0d637b0843bf4651c3078ccbb0db3645594129aa4d52b0f547150c3775a27592\": container with ID starting with 0d637b0843bf4651c3078ccbb0db3645594129aa4d52b0f547150c3775a27592 not found: ID does not exist" containerID="0d637b0843bf4651c3078ccbb0db3645594129aa4d52b0f547150c3775a27592" Oct 08 19:08:54 crc kubenswrapper[4988]: I1008 19:08:54.951538 4988 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0d637b0843bf4651c3078ccbb0db3645594129aa4d52b0f547150c3775a27592"} err="failed to get container status \"0d637b0843bf4651c3078ccbb0db3645594129aa4d52b0f547150c3775a27592\": rpc error: code = NotFound desc = could not find container \"0d637b0843bf4651c3078ccbb0db3645594129aa4d52b0f547150c3775a27592\": container with ID starting with 0d637b0843bf4651c3078ccbb0db3645594129aa4d52b0f547150c3775a27592 not found: ID does not exist" Oct 08 19:08:54 crc kubenswrapper[4988]: I1008 19:08:54.951572 4988 scope.go:117] "RemoveContainer" containerID="d68d69603719a32351931f815e28cf592fbc1239c42dcb89cc8b6925170d9fcb" Oct 08 19:08:54 crc kubenswrapper[4988]: E1008 19:08:54.951935 4988 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d68d69603719a32351931f815e28cf592fbc1239c42dcb89cc8b6925170d9fcb\": container with ID starting with d68d69603719a32351931f815e28cf592fbc1239c42dcb89cc8b6925170d9fcb not found: ID does not exist" containerID="d68d69603719a32351931f815e28cf592fbc1239c42dcb89cc8b6925170d9fcb" Oct 08 19:08:54 crc kubenswrapper[4988]: I1008 19:08:54.951970 4988 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d68d69603719a32351931f815e28cf592fbc1239c42dcb89cc8b6925170d9fcb"} err="failed to get container status \"d68d69603719a32351931f815e28cf592fbc1239c42dcb89cc8b6925170d9fcb\": rpc error: code = NotFound desc = could not find container \"d68d69603719a32351931f815e28cf592fbc1239c42dcb89cc8b6925170d9fcb\": container with ID starting with d68d69603719a32351931f815e28cf592fbc1239c42dcb89cc8b6925170d9fcb not found: ID does not exist" Oct 08 19:08:54 crc kubenswrapper[4988]: I1008 19:08:54.951999 4988 scope.go:117] "RemoveContainer" containerID="159519ad6ff6c1f641f5a9fc5031c17ee4eb0eb634b9bbd864e4a2cec62f9cae" Oct 08 19:08:54 crc kubenswrapper[4988]: E1008 19:08:54.952420 4988 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"159519ad6ff6c1f641f5a9fc5031c17ee4eb0eb634b9bbd864e4a2cec62f9cae\": container with ID starting with 159519ad6ff6c1f641f5a9fc5031c17ee4eb0eb634b9bbd864e4a2cec62f9cae not found: ID does not exist" containerID="159519ad6ff6c1f641f5a9fc5031c17ee4eb0eb634b9bbd864e4a2cec62f9cae" Oct 08 19:08:54 crc kubenswrapper[4988]: I1008 19:08:54.952462 4988 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"159519ad6ff6c1f641f5a9fc5031c17ee4eb0eb634b9bbd864e4a2cec62f9cae"} err="failed to get container status \"159519ad6ff6c1f641f5a9fc5031c17ee4eb0eb634b9bbd864e4a2cec62f9cae\": rpc error: code = NotFound desc = could not find container \"159519ad6ff6c1f641f5a9fc5031c17ee4eb0eb634b9bbd864e4a2cec62f9cae\": container with ID starting with 159519ad6ff6c1f641f5a9fc5031c17ee4eb0eb634b9bbd864e4a2cec62f9cae not found: ID does not exist" Oct 08 19:08:55 crc kubenswrapper[4988]: I1008 19:08:55.223521 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-8wmdb"] Oct 08 19:08:55 crc kubenswrapper[4988]: I1008 19:08:55.223756 4988 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-8wmdb" podUID="a9451896-2cc4-43a2-bae2-fbebae3e96da" containerName="registry-server" containerID="cri-o://b46ed966cab813489a1473b07aaa164bc514b78fdf99ebdb9e4e66d6a5afcba4" gracePeriod=2 Oct 08 19:08:55 crc kubenswrapper[4988]: I1008 19:08:55.263055 4988 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="70935226-3356-4142-8e07-27da09ba58fa" path="/var/lib/kubelet/pods/70935226-3356-4142-8e07-27da09ba58fa/volumes" Oct 08 19:08:55 crc kubenswrapper[4988]: I1008 19:08:55.650867 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-8wmdb" Oct 08 19:08:55 crc kubenswrapper[4988]: I1008 19:08:55.785227 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wmzv8\" (UniqueName: \"kubernetes.io/projected/a9451896-2cc4-43a2-bae2-fbebae3e96da-kube-api-access-wmzv8\") pod \"a9451896-2cc4-43a2-bae2-fbebae3e96da\" (UID: \"a9451896-2cc4-43a2-bae2-fbebae3e96da\") " Oct 08 19:08:55 crc kubenswrapper[4988]: I1008 19:08:55.785345 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a9451896-2cc4-43a2-bae2-fbebae3e96da-utilities\") pod \"a9451896-2cc4-43a2-bae2-fbebae3e96da\" (UID: \"a9451896-2cc4-43a2-bae2-fbebae3e96da\") " Oct 08 19:08:55 crc kubenswrapper[4988]: I1008 19:08:55.785502 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a9451896-2cc4-43a2-bae2-fbebae3e96da-catalog-content\") pod \"a9451896-2cc4-43a2-bae2-fbebae3e96da\" (UID: \"a9451896-2cc4-43a2-bae2-fbebae3e96da\") " Oct 08 19:08:55 crc kubenswrapper[4988]: I1008 19:08:55.786331 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a9451896-2cc4-43a2-bae2-fbebae3e96da-utilities" (OuterVolumeSpecName: "utilities") pod "a9451896-2cc4-43a2-bae2-fbebae3e96da" (UID: "a9451896-2cc4-43a2-bae2-fbebae3e96da"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 19:08:55 crc kubenswrapper[4988]: I1008 19:08:55.791928 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a9451896-2cc4-43a2-bae2-fbebae3e96da-kube-api-access-wmzv8" (OuterVolumeSpecName: "kube-api-access-wmzv8") pod "a9451896-2cc4-43a2-bae2-fbebae3e96da" (UID: "a9451896-2cc4-43a2-bae2-fbebae3e96da"). InnerVolumeSpecName "kube-api-access-wmzv8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 19:08:55 crc kubenswrapper[4988]: I1008 19:08:55.839215 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a9451896-2cc4-43a2-bae2-fbebae3e96da-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "a9451896-2cc4-43a2-bae2-fbebae3e96da" (UID: "a9451896-2cc4-43a2-bae2-fbebae3e96da"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 19:08:55 crc kubenswrapper[4988]: I1008 19:08:55.862072 4988 generic.go:334] "Generic (PLEG): container finished" podID="a9451896-2cc4-43a2-bae2-fbebae3e96da" containerID="b46ed966cab813489a1473b07aaa164bc514b78fdf99ebdb9e4e66d6a5afcba4" exitCode=0 Oct 08 19:08:55 crc kubenswrapper[4988]: I1008 19:08:55.862133 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-8wmdb" Oct 08 19:08:55 crc kubenswrapper[4988]: I1008 19:08:55.862173 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8wmdb" event={"ID":"a9451896-2cc4-43a2-bae2-fbebae3e96da","Type":"ContainerDied","Data":"b46ed966cab813489a1473b07aaa164bc514b78fdf99ebdb9e4e66d6a5afcba4"} Oct 08 19:08:55 crc kubenswrapper[4988]: I1008 19:08:55.862210 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8wmdb" event={"ID":"a9451896-2cc4-43a2-bae2-fbebae3e96da","Type":"ContainerDied","Data":"346dc3b09ae9e65f7fe182dc7ba09429496fd28d20eb943ca6d32908afa4dfbe"} Oct 08 19:08:55 crc kubenswrapper[4988]: I1008 19:08:55.862236 4988 scope.go:117] "RemoveContainer" containerID="b46ed966cab813489a1473b07aaa164bc514b78fdf99ebdb9e4e66d6a5afcba4" Oct 08 19:08:55 crc kubenswrapper[4988]: I1008 19:08:55.881751 4988 scope.go:117] "RemoveContainer" containerID="3605aa64ce8219bf64a4f6fa833fd5bd86a35f7344ce8ca728bf727e823e6926" Oct 08 19:08:55 crc kubenswrapper[4988]: I1008 19:08:55.892166 4988 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a9451896-2cc4-43a2-bae2-fbebae3e96da-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 08 19:08:55 crc kubenswrapper[4988]: I1008 19:08:55.892214 4988 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wmzv8\" (UniqueName: \"kubernetes.io/projected/a9451896-2cc4-43a2-bae2-fbebae3e96da-kube-api-access-wmzv8\") on node \"crc\" DevicePath \"\"" Oct 08 19:08:55 crc kubenswrapper[4988]: I1008 19:08:55.892230 4988 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a9451896-2cc4-43a2-bae2-fbebae3e96da-utilities\") on node \"crc\" DevicePath \"\"" Oct 08 19:08:55 crc kubenswrapper[4988]: I1008 19:08:55.896651 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-8wmdb"] Oct 08 19:08:55 crc kubenswrapper[4988]: I1008 19:08:55.905538 4988 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-8wmdb"] Oct 08 19:08:55 crc kubenswrapper[4988]: I1008 19:08:55.918220 4988 scope.go:117] "RemoveContainer" containerID="69cf88d50d661821115675be8c45acb6fc92a3aba9c3ef689b73a82409020485" Oct 08 19:08:55 crc kubenswrapper[4988]: I1008 19:08:55.940239 4988 scope.go:117] "RemoveContainer" containerID="b46ed966cab813489a1473b07aaa164bc514b78fdf99ebdb9e4e66d6a5afcba4" Oct 08 19:08:55 crc kubenswrapper[4988]: E1008 19:08:55.940748 4988 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b46ed966cab813489a1473b07aaa164bc514b78fdf99ebdb9e4e66d6a5afcba4\": container with ID starting with b46ed966cab813489a1473b07aaa164bc514b78fdf99ebdb9e4e66d6a5afcba4 not found: ID does not exist" containerID="b46ed966cab813489a1473b07aaa164bc514b78fdf99ebdb9e4e66d6a5afcba4" Oct 08 19:08:55 crc kubenswrapper[4988]: I1008 19:08:55.940789 4988 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b46ed966cab813489a1473b07aaa164bc514b78fdf99ebdb9e4e66d6a5afcba4"} err="failed to get container status \"b46ed966cab813489a1473b07aaa164bc514b78fdf99ebdb9e4e66d6a5afcba4\": rpc error: code = NotFound desc = could not find container \"b46ed966cab813489a1473b07aaa164bc514b78fdf99ebdb9e4e66d6a5afcba4\": container with ID starting with b46ed966cab813489a1473b07aaa164bc514b78fdf99ebdb9e4e66d6a5afcba4 not found: ID does not exist" Oct 08 19:08:55 crc kubenswrapper[4988]: I1008 19:08:55.940817 4988 scope.go:117] "RemoveContainer" containerID="3605aa64ce8219bf64a4f6fa833fd5bd86a35f7344ce8ca728bf727e823e6926" Oct 08 19:08:55 crc kubenswrapper[4988]: E1008 19:08:55.941223 4988 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3605aa64ce8219bf64a4f6fa833fd5bd86a35f7344ce8ca728bf727e823e6926\": container with ID starting with 3605aa64ce8219bf64a4f6fa833fd5bd86a35f7344ce8ca728bf727e823e6926 not found: ID does not exist" containerID="3605aa64ce8219bf64a4f6fa833fd5bd86a35f7344ce8ca728bf727e823e6926" Oct 08 19:08:55 crc kubenswrapper[4988]: I1008 19:08:55.941258 4988 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3605aa64ce8219bf64a4f6fa833fd5bd86a35f7344ce8ca728bf727e823e6926"} err="failed to get container status \"3605aa64ce8219bf64a4f6fa833fd5bd86a35f7344ce8ca728bf727e823e6926\": rpc error: code = NotFound desc = could not find container \"3605aa64ce8219bf64a4f6fa833fd5bd86a35f7344ce8ca728bf727e823e6926\": container with ID starting with 3605aa64ce8219bf64a4f6fa833fd5bd86a35f7344ce8ca728bf727e823e6926 not found: ID does not exist" Oct 08 19:08:55 crc kubenswrapper[4988]: I1008 19:08:55.941287 4988 scope.go:117] "RemoveContainer" containerID="69cf88d50d661821115675be8c45acb6fc92a3aba9c3ef689b73a82409020485" Oct 08 19:08:55 crc kubenswrapper[4988]: E1008 19:08:55.941595 4988 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"69cf88d50d661821115675be8c45acb6fc92a3aba9c3ef689b73a82409020485\": container with ID starting with 69cf88d50d661821115675be8c45acb6fc92a3aba9c3ef689b73a82409020485 not found: ID does not exist" containerID="69cf88d50d661821115675be8c45acb6fc92a3aba9c3ef689b73a82409020485" Oct 08 19:08:55 crc kubenswrapper[4988]: I1008 19:08:55.941631 4988 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"69cf88d50d661821115675be8c45acb6fc92a3aba9c3ef689b73a82409020485"} err="failed to get container status \"69cf88d50d661821115675be8c45acb6fc92a3aba9c3ef689b73a82409020485\": rpc error: code = NotFound desc = could not find container \"69cf88d50d661821115675be8c45acb6fc92a3aba9c3ef689b73a82409020485\": container with ID starting with 69cf88d50d661821115675be8c45acb6fc92a3aba9c3ef689b73a82409020485 not found: ID does not exist" Oct 08 19:08:57 crc kubenswrapper[4988]: I1008 19:08:57.246662 4988 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a9451896-2cc4-43a2-bae2-fbebae3e96da" path="/var/lib/kubelet/pods/a9451896-2cc4-43a2-bae2-fbebae3e96da/volumes" Oct 08 19:09:04 crc kubenswrapper[4988]: E1008 19:09:04.287859 4988 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda9451896_2cc4_43a2_bae2_fbebae3e96da.slice/crio-conmon-b46ed966cab813489a1473b07aaa164bc514b78fdf99ebdb9e4e66d6a5afcba4.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda9451896_2cc4_43a2_bae2_fbebae3e96da.slice/crio-b46ed966cab813489a1473b07aaa164bc514b78fdf99ebdb9e4e66d6a5afcba4.scope\": RecentStats: unable to find data in memory cache]" Oct 08 19:09:14 crc kubenswrapper[4988]: E1008 19:09:14.548122 4988 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda9451896_2cc4_43a2_bae2_fbebae3e96da.slice/crio-conmon-b46ed966cab813489a1473b07aaa164bc514b78fdf99ebdb9e4e66d6a5afcba4.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda9451896_2cc4_43a2_bae2_fbebae3e96da.slice/crio-b46ed966cab813489a1473b07aaa164bc514b78fdf99ebdb9e4e66d6a5afcba4.scope\": RecentStats: unable to find data in memory cache]" Oct 08 19:09:24 crc kubenswrapper[4988]: E1008 19:09:24.782401 4988 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda9451896_2cc4_43a2_bae2_fbebae3e96da.slice/crio-b46ed966cab813489a1473b07aaa164bc514b78fdf99ebdb9e4e66d6a5afcba4.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda9451896_2cc4_43a2_bae2_fbebae3e96da.slice/crio-conmon-b46ed966cab813489a1473b07aaa164bc514b78fdf99ebdb9e4e66d6a5afcba4.scope\": RecentStats: unable to find data in memory cache]" Oct 08 19:09:35 crc kubenswrapper[4988]: E1008 19:09:35.018264 4988 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda9451896_2cc4_43a2_bae2_fbebae3e96da.slice/crio-b46ed966cab813489a1473b07aaa164bc514b78fdf99ebdb9e4e66d6a5afcba4.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda9451896_2cc4_43a2_bae2_fbebae3e96da.slice/crio-conmon-b46ed966cab813489a1473b07aaa164bc514b78fdf99ebdb9e4e66d6a5afcba4.scope\": RecentStats: unable to find data in memory cache]" Oct 08 19:09:45 crc kubenswrapper[4988]: E1008 19:09:45.227457 4988 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda9451896_2cc4_43a2_bae2_fbebae3e96da.slice/crio-conmon-b46ed966cab813489a1473b07aaa164bc514b78fdf99ebdb9e4e66d6a5afcba4.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda9451896_2cc4_43a2_bae2_fbebae3e96da.slice/crio-b46ed966cab813489a1473b07aaa164bc514b78fdf99ebdb9e4e66d6a5afcba4.scope\": RecentStats: unable to find data in memory cache]" Oct 08 19:10:23 crc kubenswrapper[4988]: I1008 19:10:23.340963 4988 patch_prober.go:28] interesting pod/machine-config-daemon-nm54f container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 19:10:23 crc kubenswrapper[4988]: I1008 19:10:23.341566 4988 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 19:10:53 crc kubenswrapper[4988]: I1008 19:10:53.338053 4988 patch_prober.go:28] interesting pod/machine-config-daemon-nm54f container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 19:10:53 crc kubenswrapper[4988]: I1008 19:10:53.338686 4988 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 19:11:23 crc kubenswrapper[4988]: I1008 19:11:23.337648 4988 patch_prober.go:28] interesting pod/machine-config-daemon-nm54f container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 19:11:23 crc kubenswrapper[4988]: I1008 19:11:23.338205 4988 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 19:11:23 crc kubenswrapper[4988]: I1008 19:11:23.338264 4988 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" Oct 08 19:11:23 crc kubenswrapper[4988]: I1008 19:11:23.339018 4988 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"af2043d46f0bbf33ed02542990cdf54feefa63a7c8331b6a70288efbb7b7a3ae"} pod="openshift-machine-config-operator/machine-config-daemon-nm54f" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 08 19:11:23 crc kubenswrapper[4988]: I1008 19:11:23.339117 4988 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" containerName="machine-config-daemon" containerID="cri-o://af2043d46f0bbf33ed02542990cdf54feefa63a7c8331b6a70288efbb7b7a3ae" gracePeriod=600 Oct 08 19:11:23 crc kubenswrapper[4988]: E1008 19:11:23.493285 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nm54f_openshift-machine-config-operator(f328a20b-5283-407e-8a93-2ab2da599f18)\"" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" Oct 08 19:11:24 crc kubenswrapper[4988]: I1008 19:11:24.253198 4988 generic.go:334] "Generic (PLEG): container finished" podID="f328a20b-5283-407e-8a93-2ab2da599f18" containerID="af2043d46f0bbf33ed02542990cdf54feefa63a7c8331b6a70288efbb7b7a3ae" exitCode=0 Oct 08 19:11:24 crc kubenswrapper[4988]: I1008 19:11:24.253296 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" event={"ID":"f328a20b-5283-407e-8a93-2ab2da599f18","Type":"ContainerDied","Data":"af2043d46f0bbf33ed02542990cdf54feefa63a7c8331b6a70288efbb7b7a3ae"} Oct 08 19:11:24 crc kubenswrapper[4988]: I1008 19:11:24.254104 4988 scope.go:117] "RemoveContainer" containerID="482cb9ffb14e123461aee7ddc866cf73ccfa365d4521f4ee6abf4f22170a86b2" Oct 08 19:11:24 crc kubenswrapper[4988]: I1008 19:11:24.254520 4988 scope.go:117] "RemoveContainer" containerID="af2043d46f0bbf33ed02542990cdf54feefa63a7c8331b6a70288efbb7b7a3ae" Oct 08 19:11:24 crc kubenswrapper[4988]: E1008 19:11:24.254797 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nm54f_openshift-machine-config-operator(f328a20b-5283-407e-8a93-2ab2da599f18)\"" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" Oct 08 19:11:36 crc kubenswrapper[4988]: I1008 19:11:36.238831 4988 scope.go:117] "RemoveContainer" containerID="af2043d46f0bbf33ed02542990cdf54feefa63a7c8331b6a70288efbb7b7a3ae" Oct 08 19:11:36 crc kubenswrapper[4988]: E1008 19:11:36.239981 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nm54f_openshift-machine-config-operator(f328a20b-5283-407e-8a93-2ab2da599f18)\"" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" Oct 08 19:11:51 crc kubenswrapper[4988]: I1008 19:11:51.238893 4988 scope.go:117] "RemoveContainer" containerID="af2043d46f0bbf33ed02542990cdf54feefa63a7c8331b6a70288efbb7b7a3ae" Oct 08 19:11:51 crc kubenswrapper[4988]: E1008 19:11:51.240105 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nm54f_openshift-machine-config-operator(f328a20b-5283-407e-8a93-2ab2da599f18)\"" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" Oct 08 19:12:02 crc kubenswrapper[4988]: I1008 19:12:02.238547 4988 scope.go:117] "RemoveContainer" containerID="af2043d46f0bbf33ed02542990cdf54feefa63a7c8331b6a70288efbb7b7a3ae" Oct 08 19:12:02 crc kubenswrapper[4988]: E1008 19:12:02.239622 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nm54f_openshift-machine-config-operator(f328a20b-5283-407e-8a93-2ab2da599f18)\"" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" Oct 08 19:12:13 crc kubenswrapper[4988]: I1008 19:12:13.239324 4988 scope.go:117] "RemoveContainer" containerID="af2043d46f0bbf33ed02542990cdf54feefa63a7c8331b6a70288efbb7b7a3ae" Oct 08 19:12:13 crc kubenswrapper[4988]: E1008 19:12:13.240271 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nm54f_openshift-machine-config-operator(f328a20b-5283-407e-8a93-2ab2da599f18)\"" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" Oct 08 19:12:24 crc kubenswrapper[4988]: I1008 19:12:24.237945 4988 scope.go:117] "RemoveContainer" containerID="af2043d46f0bbf33ed02542990cdf54feefa63a7c8331b6a70288efbb7b7a3ae" Oct 08 19:12:24 crc kubenswrapper[4988]: E1008 19:12:24.238863 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nm54f_openshift-machine-config-operator(f328a20b-5283-407e-8a93-2ab2da599f18)\"" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" Oct 08 19:12:37 crc kubenswrapper[4988]: I1008 19:12:37.237579 4988 scope.go:117] "RemoveContainer" containerID="af2043d46f0bbf33ed02542990cdf54feefa63a7c8331b6a70288efbb7b7a3ae" Oct 08 19:12:37 crc kubenswrapper[4988]: E1008 19:12:37.238421 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nm54f_openshift-machine-config-operator(f328a20b-5283-407e-8a93-2ab2da599f18)\"" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" Oct 08 19:12:52 crc kubenswrapper[4988]: I1008 19:12:52.237969 4988 scope.go:117] "RemoveContainer" containerID="af2043d46f0bbf33ed02542990cdf54feefa63a7c8331b6a70288efbb7b7a3ae" Oct 08 19:12:52 crc kubenswrapper[4988]: E1008 19:12:52.239016 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nm54f_openshift-machine-config-operator(f328a20b-5283-407e-8a93-2ab2da599f18)\"" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" Oct 08 19:13:07 crc kubenswrapper[4988]: I1008 19:13:07.238630 4988 scope.go:117] "RemoveContainer" containerID="af2043d46f0bbf33ed02542990cdf54feefa63a7c8331b6a70288efbb7b7a3ae" Oct 08 19:13:07 crc kubenswrapper[4988]: E1008 19:13:07.240671 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nm54f_openshift-machine-config-operator(f328a20b-5283-407e-8a93-2ab2da599f18)\"" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" Oct 08 19:13:20 crc kubenswrapper[4988]: I1008 19:13:20.238277 4988 scope.go:117] "RemoveContainer" containerID="af2043d46f0bbf33ed02542990cdf54feefa63a7c8331b6a70288efbb7b7a3ae" Oct 08 19:13:20 crc kubenswrapper[4988]: E1008 19:13:20.239491 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nm54f_openshift-machine-config-operator(f328a20b-5283-407e-8a93-2ab2da599f18)\"" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" Oct 08 19:13:31 crc kubenswrapper[4988]: I1008 19:13:31.240254 4988 scope.go:117] "RemoveContainer" containerID="af2043d46f0bbf33ed02542990cdf54feefa63a7c8331b6a70288efbb7b7a3ae" Oct 08 19:13:31 crc kubenswrapper[4988]: E1008 19:13:31.241024 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nm54f_openshift-machine-config-operator(f328a20b-5283-407e-8a93-2ab2da599f18)\"" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" Oct 08 19:13:45 crc kubenswrapper[4988]: I1008 19:13:45.245537 4988 scope.go:117] "RemoveContainer" containerID="af2043d46f0bbf33ed02542990cdf54feefa63a7c8331b6a70288efbb7b7a3ae" Oct 08 19:13:45 crc kubenswrapper[4988]: E1008 19:13:45.247133 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nm54f_openshift-machine-config-operator(f328a20b-5283-407e-8a93-2ab2da599f18)\"" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" Oct 08 19:13:56 crc kubenswrapper[4988]: I1008 19:13:56.238826 4988 scope.go:117] "RemoveContainer" containerID="af2043d46f0bbf33ed02542990cdf54feefa63a7c8331b6a70288efbb7b7a3ae" Oct 08 19:13:56 crc kubenswrapper[4988]: E1008 19:13:56.239878 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nm54f_openshift-machine-config-operator(f328a20b-5283-407e-8a93-2ab2da599f18)\"" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" Oct 08 19:14:09 crc kubenswrapper[4988]: I1008 19:14:09.238572 4988 scope.go:117] "RemoveContainer" containerID="af2043d46f0bbf33ed02542990cdf54feefa63a7c8331b6a70288efbb7b7a3ae" Oct 08 19:14:09 crc kubenswrapper[4988]: E1008 19:14:09.239822 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nm54f_openshift-machine-config-operator(f328a20b-5283-407e-8a93-2ab2da599f18)\"" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" Oct 08 19:14:23 crc kubenswrapper[4988]: I1008 19:14:23.238246 4988 scope.go:117] "RemoveContainer" containerID="af2043d46f0bbf33ed02542990cdf54feefa63a7c8331b6a70288efbb7b7a3ae" Oct 08 19:14:23 crc kubenswrapper[4988]: E1008 19:14:23.239255 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nm54f_openshift-machine-config-operator(f328a20b-5283-407e-8a93-2ab2da599f18)\"" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" Oct 08 19:14:38 crc kubenswrapper[4988]: I1008 19:14:38.239074 4988 scope.go:117] "RemoveContainer" containerID="af2043d46f0bbf33ed02542990cdf54feefa63a7c8331b6a70288efbb7b7a3ae" Oct 08 19:14:38 crc kubenswrapper[4988]: E1008 19:14:38.240256 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nm54f_openshift-machine-config-operator(f328a20b-5283-407e-8a93-2ab2da599f18)\"" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" Oct 08 19:14:53 crc kubenswrapper[4988]: I1008 19:14:53.238424 4988 scope.go:117] "RemoveContainer" containerID="af2043d46f0bbf33ed02542990cdf54feefa63a7c8331b6a70288efbb7b7a3ae" Oct 08 19:14:53 crc kubenswrapper[4988]: E1008 19:14:53.239376 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nm54f_openshift-machine-config-operator(f328a20b-5283-407e-8a93-2ab2da599f18)\"" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" Oct 08 19:15:00 crc kubenswrapper[4988]: I1008 19:15:00.160757 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29332515-zjccp"] Oct 08 19:15:00 crc kubenswrapper[4988]: E1008 19:15:00.162744 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e3b1b754-1ed3-465e-9479-3b22a506f8d0" containerName="extract-utilities" Oct 08 19:15:00 crc kubenswrapper[4988]: I1008 19:15:00.162794 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="e3b1b754-1ed3-465e-9479-3b22a506f8d0" containerName="extract-utilities" Oct 08 19:15:00 crc kubenswrapper[4988]: E1008 19:15:00.162821 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="70935226-3356-4142-8e07-27da09ba58fa" containerName="registry-server" Oct 08 19:15:00 crc kubenswrapper[4988]: I1008 19:15:00.162833 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="70935226-3356-4142-8e07-27da09ba58fa" containerName="registry-server" Oct 08 19:15:00 crc kubenswrapper[4988]: E1008 19:15:00.162853 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e3b1b754-1ed3-465e-9479-3b22a506f8d0" containerName="registry-server" Oct 08 19:15:00 crc kubenswrapper[4988]: I1008 19:15:00.162867 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="e3b1b754-1ed3-465e-9479-3b22a506f8d0" containerName="registry-server" Oct 08 19:15:00 crc kubenswrapper[4988]: E1008 19:15:00.162887 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e3b1b754-1ed3-465e-9479-3b22a506f8d0" containerName="extract-content" Oct 08 19:15:00 crc kubenswrapper[4988]: I1008 19:15:00.162899 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="e3b1b754-1ed3-465e-9479-3b22a506f8d0" containerName="extract-content" Oct 08 19:15:00 crc kubenswrapper[4988]: E1008 19:15:00.162914 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a9451896-2cc4-43a2-bae2-fbebae3e96da" containerName="extract-utilities" Oct 08 19:15:00 crc kubenswrapper[4988]: I1008 19:15:00.162925 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="a9451896-2cc4-43a2-bae2-fbebae3e96da" containerName="extract-utilities" Oct 08 19:15:00 crc kubenswrapper[4988]: E1008 19:15:00.162942 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a9451896-2cc4-43a2-bae2-fbebae3e96da" containerName="registry-server" Oct 08 19:15:00 crc kubenswrapper[4988]: I1008 19:15:00.162954 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="a9451896-2cc4-43a2-bae2-fbebae3e96da" containerName="registry-server" Oct 08 19:15:00 crc kubenswrapper[4988]: E1008 19:15:00.162980 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="70935226-3356-4142-8e07-27da09ba58fa" containerName="extract-content" Oct 08 19:15:00 crc kubenswrapper[4988]: I1008 19:15:00.162992 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="70935226-3356-4142-8e07-27da09ba58fa" containerName="extract-content" Oct 08 19:15:00 crc kubenswrapper[4988]: E1008 19:15:00.163019 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="70935226-3356-4142-8e07-27da09ba58fa" containerName="extract-utilities" Oct 08 19:15:00 crc kubenswrapper[4988]: I1008 19:15:00.163031 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="70935226-3356-4142-8e07-27da09ba58fa" containerName="extract-utilities" Oct 08 19:15:00 crc kubenswrapper[4988]: E1008 19:15:00.163059 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a9451896-2cc4-43a2-bae2-fbebae3e96da" containerName="extract-content" Oct 08 19:15:00 crc kubenswrapper[4988]: I1008 19:15:00.163071 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="a9451896-2cc4-43a2-bae2-fbebae3e96da" containerName="extract-content" Oct 08 19:15:00 crc kubenswrapper[4988]: I1008 19:15:00.163294 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="e3b1b754-1ed3-465e-9479-3b22a506f8d0" containerName="registry-server" Oct 08 19:15:00 crc kubenswrapper[4988]: I1008 19:15:00.163316 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="70935226-3356-4142-8e07-27da09ba58fa" containerName="registry-server" Oct 08 19:15:00 crc kubenswrapper[4988]: I1008 19:15:00.163339 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="a9451896-2cc4-43a2-bae2-fbebae3e96da" containerName="registry-server" Oct 08 19:15:00 crc kubenswrapper[4988]: I1008 19:15:00.164174 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29332515-zjccp" Oct 08 19:15:00 crc kubenswrapper[4988]: I1008 19:15:00.169095 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 08 19:15:00 crc kubenswrapper[4988]: I1008 19:15:00.169507 4988 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 08 19:15:00 crc kubenswrapper[4988]: I1008 19:15:00.178874 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29332515-zjccp"] Oct 08 19:15:00 crc kubenswrapper[4988]: I1008 19:15:00.298708 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/4f49819e-fea9-4f31-bd01-f34ddaa425cf-config-volume\") pod \"collect-profiles-29332515-zjccp\" (UID: \"4f49819e-fea9-4f31-bd01-f34ddaa425cf\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29332515-zjccp" Oct 08 19:15:00 crc kubenswrapper[4988]: I1008 19:15:00.298758 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/4f49819e-fea9-4f31-bd01-f34ddaa425cf-secret-volume\") pod \"collect-profiles-29332515-zjccp\" (UID: \"4f49819e-fea9-4f31-bd01-f34ddaa425cf\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29332515-zjccp" Oct 08 19:15:00 crc kubenswrapper[4988]: I1008 19:15:00.298794 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gfss5\" (UniqueName: \"kubernetes.io/projected/4f49819e-fea9-4f31-bd01-f34ddaa425cf-kube-api-access-gfss5\") pod \"collect-profiles-29332515-zjccp\" (UID: \"4f49819e-fea9-4f31-bd01-f34ddaa425cf\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29332515-zjccp" Oct 08 19:15:00 crc kubenswrapper[4988]: I1008 19:15:00.399884 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/4f49819e-fea9-4f31-bd01-f34ddaa425cf-config-volume\") pod \"collect-profiles-29332515-zjccp\" (UID: \"4f49819e-fea9-4f31-bd01-f34ddaa425cf\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29332515-zjccp" Oct 08 19:15:00 crc kubenswrapper[4988]: I1008 19:15:00.399939 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/4f49819e-fea9-4f31-bd01-f34ddaa425cf-secret-volume\") pod \"collect-profiles-29332515-zjccp\" (UID: \"4f49819e-fea9-4f31-bd01-f34ddaa425cf\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29332515-zjccp" Oct 08 19:15:00 crc kubenswrapper[4988]: I1008 19:15:00.399983 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gfss5\" (UniqueName: \"kubernetes.io/projected/4f49819e-fea9-4f31-bd01-f34ddaa425cf-kube-api-access-gfss5\") pod \"collect-profiles-29332515-zjccp\" (UID: \"4f49819e-fea9-4f31-bd01-f34ddaa425cf\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29332515-zjccp" Oct 08 19:15:00 crc kubenswrapper[4988]: I1008 19:15:00.401993 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/4f49819e-fea9-4f31-bd01-f34ddaa425cf-config-volume\") pod \"collect-profiles-29332515-zjccp\" (UID: \"4f49819e-fea9-4f31-bd01-f34ddaa425cf\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29332515-zjccp" Oct 08 19:15:00 crc kubenswrapper[4988]: I1008 19:15:00.424161 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gfss5\" (UniqueName: \"kubernetes.io/projected/4f49819e-fea9-4f31-bd01-f34ddaa425cf-kube-api-access-gfss5\") pod \"collect-profiles-29332515-zjccp\" (UID: \"4f49819e-fea9-4f31-bd01-f34ddaa425cf\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29332515-zjccp" Oct 08 19:15:00 crc kubenswrapper[4988]: I1008 19:15:00.435983 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/4f49819e-fea9-4f31-bd01-f34ddaa425cf-secret-volume\") pod \"collect-profiles-29332515-zjccp\" (UID: \"4f49819e-fea9-4f31-bd01-f34ddaa425cf\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29332515-zjccp" Oct 08 19:15:00 crc kubenswrapper[4988]: I1008 19:15:00.495284 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29332515-zjccp" Oct 08 19:15:00 crc kubenswrapper[4988]: I1008 19:15:00.929411 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29332515-zjccp"] Oct 08 19:15:01 crc kubenswrapper[4988]: I1008 19:15:01.332180 4988 generic.go:334] "Generic (PLEG): container finished" podID="4f49819e-fea9-4f31-bd01-f34ddaa425cf" containerID="b4459fd38af0cc29ba01f889cf7c7cfa630d4be8878f893f3431416f7fa443d3" exitCode=0 Oct 08 19:15:01 crc kubenswrapper[4988]: I1008 19:15:01.332258 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29332515-zjccp" event={"ID":"4f49819e-fea9-4f31-bd01-f34ddaa425cf","Type":"ContainerDied","Data":"b4459fd38af0cc29ba01f889cf7c7cfa630d4be8878f893f3431416f7fa443d3"} Oct 08 19:15:01 crc kubenswrapper[4988]: I1008 19:15:01.333561 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29332515-zjccp" event={"ID":"4f49819e-fea9-4f31-bd01-f34ddaa425cf","Type":"ContainerStarted","Data":"0dca32c1061208e9bf2295176691764b0a8747b1fdbcee54eec7f71db621b3a2"} Oct 08 19:15:02 crc kubenswrapper[4988]: I1008 19:15:02.736322 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29332515-zjccp" Oct 08 19:15:02 crc kubenswrapper[4988]: I1008 19:15:02.834376 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/4f49819e-fea9-4f31-bd01-f34ddaa425cf-secret-volume\") pod \"4f49819e-fea9-4f31-bd01-f34ddaa425cf\" (UID: \"4f49819e-fea9-4f31-bd01-f34ddaa425cf\") " Oct 08 19:15:02 crc kubenswrapper[4988]: I1008 19:15:02.834495 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/4f49819e-fea9-4f31-bd01-f34ddaa425cf-config-volume\") pod \"4f49819e-fea9-4f31-bd01-f34ddaa425cf\" (UID: \"4f49819e-fea9-4f31-bd01-f34ddaa425cf\") " Oct 08 19:15:02 crc kubenswrapper[4988]: I1008 19:15:02.834534 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gfss5\" (UniqueName: \"kubernetes.io/projected/4f49819e-fea9-4f31-bd01-f34ddaa425cf-kube-api-access-gfss5\") pod \"4f49819e-fea9-4f31-bd01-f34ddaa425cf\" (UID: \"4f49819e-fea9-4f31-bd01-f34ddaa425cf\") " Oct 08 19:15:02 crc kubenswrapper[4988]: I1008 19:15:02.835805 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4f49819e-fea9-4f31-bd01-f34ddaa425cf-config-volume" (OuterVolumeSpecName: "config-volume") pod "4f49819e-fea9-4f31-bd01-f34ddaa425cf" (UID: "4f49819e-fea9-4f31-bd01-f34ddaa425cf"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 19:15:02 crc kubenswrapper[4988]: I1008 19:15:02.840305 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4f49819e-fea9-4f31-bd01-f34ddaa425cf-kube-api-access-gfss5" (OuterVolumeSpecName: "kube-api-access-gfss5") pod "4f49819e-fea9-4f31-bd01-f34ddaa425cf" (UID: "4f49819e-fea9-4f31-bd01-f34ddaa425cf"). InnerVolumeSpecName "kube-api-access-gfss5". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 19:15:02 crc kubenswrapper[4988]: I1008 19:15:02.843112 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4f49819e-fea9-4f31-bd01-f34ddaa425cf-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "4f49819e-fea9-4f31-bd01-f34ddaa425cf" (UID: "4f49819e-fea9-4f31-bd01-f34ddaa425cf"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 19:15:02 crc kubenswrapper[4988]: I1008 19:15:02.936463 4988 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/4f49819e-fea9-4f31-bd01-f34ddaa425cf-config-volume\") on node \"crc\" DevicePath \"\"" Oct 08 19:15:02 crc kubenswrapper[4988]: I1008 19:15:02.936511 4988 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gfss5\" (UniqueName: \"kubernetes.io/projected/4f49819e-fea9-4f31-bd01-f34ddaa425cf-kube-api-access-gfss5\") on node \"crc\" DevicePath \"\"" Oct 08 19:15:02 crc kubenswrapper[4988]: I1008 19:15:02.936528 4988 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/4f49819e-fea9-4f31-bd01-f34ddaa425cf-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 08 19:15:03 crc kubenswrapper[4988]: I1008 19:15:03.354987 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29332515-zjccp" event={"ID":"4f49819e-fea9-4f31-bd01-f34ddaa425cf","Type":"ContainerDied","Data":"0dca32c1061208e9bf2295176691764b0a8747b1fdbcee54eec7f71db621b3a2"} Oct 08 19:15:03 crc kubenswrapper[4988]: I1008 19:15:03.355190 4988 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0dca32c1061208e9bf2295176691764b0a8747b1fdbcee54eec7f71db621b3a2" Oct 08 19:15:03 crc kubenswrapper[4988]: I1008 19:15:03.355302 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29332515-zjccp" Oct 08 19:15:03 crc kubenswrapper[4988]: I1008 19:15:03.826411 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29332470-lbpdq"] Oct 08 19:15:03 crc kubenswrapper[4988]: I1008 19:15:03.830795 4988 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29332470-lbpdq"] Oct 08 19:15:04 crc kubenswrapper[4988]: I1008 19:15:04.238096 4988 scope.go:117] "RemoveContainer" containerID="af2043d46f0bbf33ed02542990cdf54feefa63a7c8331b6a70288efbb7b7a3ae" Oct 08 19:15:04 crc kubenswrapper[4988]: E1008 19:15:04.238310 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nm54f_openshift-machine-config-operator(f328a20b-5283-407e-8a93-2ab2da599f18)\"" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" Oct 08 19:15:05 crc kubenswrapper[4988]: I1008 19:15:05.258021 4988 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b039de59-9a58-4b20-b279-1447415e8843" path="/var/lib/kubelet/pods/b039de59-9a58-4b20-b279-1447415e8843/volumes" Oct 08 19:15:17 crc kubenswrapper[4988]: I1008 19:15:17.237546 4988 scope.go:117] "RemoveContainer" containerID="af2043d46f0bbf33ed02542990cdf54feefa63a7c8331b6a70288efbb7b7a3ae" Oct 08 19:15:17 crc kubenswrapper[4988]: E1008 19:15:17.238327 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nm54f_openshift-machine-config-operator(f328a20b-5283-407e-8a93-2ab2da599f18)\"" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" Oct 08 19:15:31 crc kubenswrapper[4988]: I1008 19:15:31.237446 4988 scope.go:117] "RemoveContainer" containerID="af2043d46f0bbf33ed02542990cdf54feefa63a7c8331b6a70288efbb7b7a3ae" Oct 08 19:15:31 crc kubenswrapper[4988]: E1008 19:15:31.238216 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nm54f_openshift-machine-config-operator(f328a20b-5283-407e-8a93-2ab2da599f18)\"" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" Oct 08 19:15:44 crc kubenswrapper[4988]: I1008 19:15:44.237322 4988 scope.go:117] "RemoveContainer" containerID="af2043d46f0bbf33ed02542990cdf54feefa63a7c8331b6a70288efbb7b7a3ae" Oct 08 19:15:44 crc kubenswrapper[4988]: E1008 19:15:44.239218 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nm54f_openshift-machine-config-operator(f328a20b-5283-407e-8a93-2ab2da599f18)\"" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" Oct 08 19:15:58 crc kubenswrapper[4988]: I1008 19:15:58.237563 4988 scope.go:117] "RemoveContainer" containerID="af2043d46f0bbf33ed02542990cdf54feefa63a7c8331b6a70288efbb7b7a3ae" Oct 08 19:15:58 crc kubenswrapper[4988]: E1008 19:15:58.238633 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nm54f_openshift-machine-config-operator(f328a20b-5283-407e-8a93-2ab2da599f18)\"" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" Oct 08 19:16:02 crc kubenswrapper[4988]: I1008 19:16:02.064675 4988 scope.go:117] "RemoveContainer" containerID="75398b89f53af48fd5e5ebbacf47ccffc023704db78076e20861c7bf6ee6b5ee" Oct 08 19:16:13 crc kubenswrapper[4988]: I1008 19:16:13.238734 4988 scope.go:117] "RemoveContainer" containerID="af2043d46f0bbf33ed02542990cdf54feefa63a7c8331b6a70288efbb7b7a3ae" Oct 08 19:16:13 crc kubenswrapper[4988]: E1008 19:16:13.239670 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nm54f_openshift-machine-config-operator(f328a20b-5283-407e-8a93-2ab2da599f18)\"" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" Oct 08 19:16:25 crc kubenswrapper[4988]: I1008 19:16:25.247019 4988 scope.go:117] "RemoveContainer" containerID="af2043d46f0bbf33ed02542990cdf54feefa63a7c8331b6a70288efbb7b7a3ae" Oct 08 19:16:26 crc kubenswrapper[4988]: I1008 19:16:26.103752 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" event={"ID":"f328a20b-5283-407e-8a93-2ab2da599f18","Type":"ContainerStarted","Data":"eca8e1f95f09ab5d917be9719293edda80592194583b0c97c98c1d0e281973e8"} Oct 08 19:17:09 crc kubenswrapper[4988]: I1008 19:17:09.131756 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-r2b4h"] Oct 08 19:17:09 crc kubenswrapper[4988]: E1008 19:17:09.133050 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4f49819e-fea9-4f31-bd01-f34ddaa425cf" containerName="collect-profiles" Oct 08 19:17:09 crc kubenswrapper[4988]: I1008 19:17:09.133082 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="4f49819e-fea9-4f31-bd01-f34ddaa425cf" containerName="collect-profiles" Oct 08 19:17:09 crc kubenswrapper[4988]: I1008 19:17:09.133506 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="4f49819e-fea9-4f31-bd01-f34ddaa425cf" containerName="collect-profiles" Oct 08 19:17:09 crc kubenswrapper[4988]: I1008 19:17:09.135838 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-r2b4h" Oct 08 19:17:09 crc kubenswrapper[4988]: I1008 19:17:09.148341 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-r2b4h"] Oct 08 19:17:09 crc kubenswrapper[4988]: I1008 19:17:09.298933 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1f41a3d7-55ff-49b4-8da0-4777210cdd5c-catalog-content\") pod \"community-operators-r2b4h\" (UID: \"1f41a3d7-55ff-49b4-8da0-4777210cdd5c\") " pod="openshift-marketplace/community-operators-r2b4h" Oct 08 19:17:09 crc kubenswrapper[4988]: I1008 19:17:09.299037 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hhk28\" (UniqueName: \"kubernetes.io/projected/1f41a3d7-55ff-49b4-8da0-4777210cdd5c-kube-api-access-hhk28\") pod \"community-operators-r2b4h\" (UID: \"1f41a3d7-55ff-49b4-8da0-4777210cdd5c\") " pod="openshift-marketplace/community-operators-r2b4h" Oct 08 19:17:09 crc kubenswrapper[4988]: I1008 19:17:09.299121 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1f41a3d7-55ff-49b4-8da0-4777210cdd5c-utilities\") pod \"community-operators-r2b4h\" (UID: \"1f41a3d7-55ff-49b4-8da0-4777210cdd5c\") " pod="openshift-marketplace/community-operators-r2b4h" Oct 08 19:17:09 crc kubenswrapper[4988]: I1008 19:17:09.399857 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hhk28\" (UniqueName: \"kubernetes.io/projected/1f41a3d7-55ff-49b4-8da0-4777210cdd5c-kube-api-access-hhk28\") pod \"community-operators-r2b4h\" (UID: \"1f41a3d7-55ff-49b4-8da0-4777210cdd5c\") " pod="openshift-marketplace/community-operators-r2b4h" Oct 08 19:17:09 crc kubenswrapper[4988]: I1008 19:17:09.399982 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1f41a3d7-55ff-49b4-8da0-4777210cdd5c-utilities\") pod \"community-operators-r2b4h\" (UID: \"1f41a3d7-55ff-49b4-8da0-4777210cdd5c\") " pod="openshift-marketplace/community-operators-r2b4h" Oct 08 19:17:09 crc kubenswrapper[4988]: I1008 19:17:09.400024 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1f41a3d7-55ff-49b4-8da0-4777210cdd5c-catalog-content\") pod \"community-operators-r2b4h\" (UID: \"1f41a3d7-55ff-49b4-8da0-4777210cdd5c\") " pod="openshift-marketplace/community-operators-r2b4h" Oct 08 19:17:09 crc kubenswrapper[4988]: I1008 19:17:09.400963 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1f41a3d7-55ff-49b4-8da0-4777210cdd5c-catalog-content\") pod \"community-operators-r2b4h\" (UID: \"1f41a3d7-55ff-49b4-8da0-4777210cdd5c\") " pod="openshift-marketplace/community-operators-r2b4h" Oct 08 19:17:09 crc kubenswrapper[4988]: I1008 19:17:09.401161 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1f41a3d7-55ff-49b4-8da0-4777210cdd5c-utilities\") pod \"community-operators-r2b4h\" (UID: \"1f41a3d7-55ff-49b4-8da0-4777210cdd5c\") " pod="openshift-marketplace/community-operators-r2b4h" Oct 08 19:17:09 crc kubenswrapper[4988]: I1008 19:17:09.430955 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hhk28\" (UniqueName: \"kubernetes.io/projected/1f41a3d7-55ff-49b4-8da0-4777210cdd5c-kube-api-access-hhk28\") pod \"community-operators-r2b4h\" (UID: \"1f41a3d7-55ff-49b4-8da0-4777210cdd5c\") " pod="openshift-marketplace/community-operators-r2b4h" Oct 08 19:17:09 crc kubenswrapper[4988]: I1008 19:17:09.497415 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-r2b4h" Oct 08 19:17:10 crc kubenswrapper[4988]: I1008 19:17:10.000510 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-r2b4h"] Oct 08 19:17:10 crc kubenswrapper[4988]: W1008 19:17:10.006284 4988 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1f41a3d7_55ff_49b4_8da0_4777210cdd5c.slice/crio-a6e0e190342e7b3e118b9a874dfad6875c14982ba92b70abb8c75593ca877c0f WatchSource:0}: Error finding container a6e0e190342e7b3e118b9a874dfad6875c14982ba92b70abb8c75593ca877c0f: Status 404 returned error can't find the container with id a6e0e190342e7b3e118b9a874dfad6875c14982ba92b70abb8c75593ca877c0f Oct 08 19:17:10 crc kubenswrapper[4988]: I1008 19:17:10.501028 4988 generic.go:334] "Generic (PLEG): container finished" podID="1f41a3d7-55ff-49b4-8da0-4777210cdd5c" containerID="79af207fc0034c0c23ff4128f0c93601df28378dbd34b25bf7952caf8da6a7c4" exitCode=0 Oct 08 19:17:10 crc kubenswrapper[4988]: I1008 19:17:10.501160 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-r2b4h" event={"ID":"1f41a3d7-55ff-49b4-8da0-4777210cdd5c","Type":"ContainerDied","Data":"79af207fc0034c0c23ff4128f0c93601df28378dbd34b25bf7952caf8da6a7c4"} Oct 08 19:17:10 crc kubenswrapper[4988]: I1008 19:17:10.501540 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-r2b4h" event={"ID":"1f41a3d7-55ff-49b4-8da0-4777210cdd5c","Type":"ContainerStarted","Data":"a6e0e190342e7b3e118b9a874dfad6875c14982ba92b70abb8c75593ca877c0f"} Oct 08 19:17:10 crc kubenswrapper[4988]: I1008 19:17:10.504742 4988 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 08 19:17:12 crc kubenswrapper[4988]: I1008 19:17:12.523072 4988 generic.go:334] "Generic (PLEG): container finished" podID="1f41a3d7-55ff-49b4-8da0-4777210cdd5c" containerID="3bc127f9d2c1cbb820f0beb3774da0fcd2882fb1c65c1bbabfb25f50eda4918c" exitCode=0 Oct 08 19:17:12 crc kubenswrapper[4988]: I1008 19:17:12.523166 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-r2b4h" event={"ID":"1f41a3d7-55ff-49b4-8da0-4777210cdd5c","Type":"ContainerDied","Data":"3bc127f9d2c1cbb820f0beb3774da0fcd2882fb1c65c1bbabfb25f50eda4918c"} Oct 08 19:17:13 crc kubenswrapper[4988]: I1008 19:17:13.538589 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-r2b4h" event={"ID":"1f41a3d7-55ff-49b4-8da0-4777210cdd5c","Type":"ContainerStarted","Data":"4afa5a68688a65806e6f6f7e355838efe7bf287e47bb3e47e1da7967ef3a85ca"} Oct 08 19:17:13 crc kubenswrapper[4988]: I1008 19:17:13.568707 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-r2b4h" podStartSLOduration=2.060585751 podStartE2EDuration="4.5686805s" podCreationTimestamp="2025-10-08 19:17:09 +0000 UTC" firstStartedPulling="2025-10-08 19:17:10.50409872 +0000 UTC m=+3975.953941520" lastFinishedPulling="2025-10-08 19:17:13.012193489 +0000 UTC m=+3978.462036269" observedRunningTime="2025-10-08 19:17:13.566225101 +0000 UTC m=+3979.016067911" watchObservedRunningTime="2025-10-08 19:17:13.5686805 +0000 UTC m=+3979.018523310" Oct 08 19:17:19 crc kubenswrapper[4988]: I1008 19:17:19.498276 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-r2b4h" Oct 08 19:17:19 crc kubenswrapper[4988]: I1008 19:17:19.499097 4988 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-r2b4h" Oct 08 19:17:19 crc kubenswrapper[4988]: I1008 19:17:19.557492 4988 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-r2b4h" Oct 08 19:17:19 crc kubenswrapper[4988]: I1008 19:17:19.666153 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-r2b4h" Oct 08 19:17:19 crc kubenswrapper[4988]: I1008 19:17:19.802104 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-r2b4h"] Oct 08 19:17:21 crc kubenswrapper[4988]: I1008 19:17:21.620272 4988 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-r2b4h" podUID="1f41a3d7-55ff-49b4-8da0-4777210cdd5c" containerName="registry-server" containerID="cri-o://4afa5a68688a65806e6f6f7e355838efe7bf287e47bb3e47e1da7967ef3a85ca" gracePeriod=2 Oct 08 19:17:22 crc kubenswrapper[4988]: I1008 19:17:22.112843 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-r2b4h" Oct 08 19:17:22 crc kubenswrapper[4988]: I1008 19:17:22.194108 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hhk28\" (UniqueName: \"kubernetes.io/projected/1f41a3d7-55ff-49b4-8da0-4777210cdd5c-kube-api-access-hhk28\") pod \"1f41a3d7-55ff-49b4-8da0-4777210cdd5c\" (UID: \"1f41a3d7-55ff-49b4-8da0-4777210cdd5c\") " Oct 08 19:17:22 crc kubenswrapper[4988]: I1008 19:17:22.194167 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1f41a3d7-55ff-49b4-8da0-4777210cdd5c-utilities\") pod \"1f41a3d7-55ff-49b4-8da0-4777210cdd5c\" (UID: \"1f41a3d7-55ff-49b4-8da0-4777210cdd5c\") " Oct 08 19:17:22 crc kubenswrapper[4988]: I1008 19:17:22.194305 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1f41a3d7-55ff-49b4-8da0-4777210cdd5c-catalog-content\") pod \"1f41a3d7-55ff-49b4-8da0-4777210cdd5c\" (UID: \"1f41a3d7-55ff-49b4-8da0-4777210cdd5c\") " Oct 08 19:17:22 crc kubenswrapper[4988]: I1008 19:17:22.195289 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1f41a3d7-55ff-49b4-8da0-4777210cdd5c-utilities" (OuterVolumeSpecName: "utilities") pod "1f41a3d7-55ff-49b4-8da0-4777210cdd5c" (UID: "1f41a3d7-55ff-49b4-8da0-4777210cdd5c"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 19:17:22 crc kubenswrapper[4988]: I1008 19:17:22.196149 4988 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1f41a3d7-55ff-49b4-8da0-4777210cdd5c-utilities\") on node \"crc\" DevicePath \"\"" Oct 08 19:17:22 crc kubenswrapper[4988]: I1008 19:17:22.200610 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1f41a3d7-55ff-49b4-8da0-4777210cdd5c-kube-api-access-hhk28" (OuterVolumeSpecName: "kube-api-access-hhk28") pod "1f41a3d7-55ff-49b4-8da0-4777210cdd5c" (UID: "1f41a3d7-55ff-49b4-8da0-4777210cdd5c"). InnerVolumeSpecName "kube-api-access-hhk28". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 19:17:22 crc kubenswrapper[4988]: I1008 19:17:22.250177 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1f41a3d7-55ff-49b4-8da0-4777210cdd5c-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1f41a3d7-55ff-49b4-8da0-4777210cdd5c" (UID: "1f41a3d7-55ff-49b4-8da0-4777210cdd5c"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 19:17:22 crc kubenswrapper[4988]: I1008 19:17:22.298423 4988 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1f41a3d7-55ff-49b4-8da0-4777210cdd5c-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 08 19:17:22 crc kubenswrapper[4988]: I1008 19:17:22.298511 4988 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hhk28\" (UniqueName: \"kubernetes.io/projected/1f41a3d7-55ff-49b4-8da0-4777210cdd5c-kube-api-access-hhk28\") on node \"crc\" DevicePath \"\"" Oct 08 19:17:22 crc kubenswrapper[4988]: I1008 19:17:22.633591 4988 generic.go:334] "Generic (PLEG): container finished" podID="1f41a3d7-55ff-49b4-8da0-4777210cdd5c" containerID="4afa5a68688a65806e6f6f7e355838efe7bf287e47bb3e47e1da7967ef3a85ca" exitCode=0 Oct 08 19:17:22 crc kubenswrapper[4988]: I1008 19:17:22.633654 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-r2b4h" event={"ID":"1f41a3d7-55ff-49b4-8da0-4777210cdd5c","Type":"ContainerDied","Data":"4afa5a68688a65806e6f6f7e355838efe7bf287e47bb3e47e1da7967ef3a85ca"} Oct 08 19:17:22 crc kubenswrapper[4988]: I1008 19:17:22.634025 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-r2b4h" event={"ID":"1f41a3d7-55ff-49b4-8da0-4777210cdd5c","Type":"ContainerDied","Data":"a6e0e190342e7b3e118b9a874dfad6875c14982ba92b70abb8c75593ca877c0f"} Oct 08 19:17:22 crc kubenswrapper[4988]: I1008 19:17:22.634063 4988 scope.go:117] "RemoveContainer" containerID="4afa5a68688a65806e6f6f7e355838efe7bf287e47bb3e47e1da7967ef3a85ca" Oct 08 19:17:22 crc kubenswrapper[4988]: I1008 19:17:22.633807 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-r2b4h" Oct 08 19:17:22 crc kubenswrapper[4988]: I1008 19:17:22.683217 4988 scope.go:117] "RemoveContainer" containerID="3bc127f9d2c1cbb820f0beb3774da0fcd2882fb1c65c1bbabfb25f50eda4918c" Oct 08 19:17:22 crc kubenswrapper[4988]: I1008 19:17:22.686246 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-r2b4h"] Oct 08 19:17:22 crc kubenswrapper[4988]: I1008 19:17:22.696572 4988 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-r2b4h"] Oct 08 19:17:22 crc kubenswrapper[4988]: I1008 19:17:22.733253 4988 scope.go:117] "RemoveContainer" containerID="79af207fc0034c0c23ff4128f0c93601df28378dbd34b25bf7952caf8da6a7c4" Oct 08 19:17:22 crc kubenswrapper[4988]: I1008 19:17:22.755041 4988 scope.go:117] "RemoveContainer" containerID="4afa5a68688a65806e6f6f7e355838efe7bf287e47bb3e47e1da7967ef3a85ca" Oct 08 19:17:22 crc kubenswrapper[4988]: E1008 19:17:22.755489 4988 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4afa5a68688a65806e6f6f7e355838efe7bf287e47bb3e47e1da7967ef3a85ca\": container with ID starting with 4afa5a68688a65806e6f6f7e355838efe7bf287e47bb3e47e1da7967ef3a85ca not found: ID does not exist" containerID="4afa5a68688a65806e6f6f7e355838efe7bf287e47bb3e47e1da7967ef3a85ca" Oct 08 19:17:22 crc kubenswrapper[4988]: I1008 19:17:22.755527 4988 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4afa5a68688a65806e6f6f7e355838efe7bf287e47bb3e47e1da7967ef3a85ca"} err="failed to get container status \"4afa5a68688a65806e6f6f7e355838efe7bf287e47bb3e47e1da7967ef3a85ca\": rpc error: code = NotFound desc = could not find container \"4afa5a68688a65806e6f6f7e355838efe7bf287e47bb3e47e1da7967ef3a85ca\": container with ID starting with 4afa5a68688a65806e6f6f7e355838efe7bf287e47bb3e47e1da7967ef3a85ca not found: ID does not exist" Oct 08 19:17:22 crc kubenswrapper[4988]: I1008 19:17:22.755553 4988 scope.go:117] "RemoveContainer" containerID="3bc127f9d2c1cbb820f0beb3774da0fcd2882fb1c65c1bbabfb25f50eda4918c" Oct 08 19:17:22 crc kubenswrapper[4988]: E1008 19:17:22.755900 4988 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3bc127f9d2c1cbb820f0beb3774da0fcd2882fb1c65c1bbabfb25f50eda4918c\": container with ID starting with 3bc127f9d2c1cbb820f0beb3774da0fcd2882fb1c65c1bbabfb25f50eda4918c not found: ID does not exist" containerID="3bc127f9d2c1cbb820f0beb3774da0fcd2882fb1c65c1bbabfb25f50eda4918c" Oct 08 19:17:22 crc kubenswrapper[4988]: I1008 19:17:22.755985 4988 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3bc127f9d2c1cbb820f0beb3774da0fcd2882fb1c65c1bbabfb25f50eda4918c"} err="failed to get container status \"3bc127f9d2c1cbb820f0beb3774da0fcd2882fb1c65c1bbabfb25f50eda4918c\": rpc error: code = NotFound desc = could not find container \"3bc127f9d2c1cbb820f0beb3774da0fcd2882fb1c65c1bbabfb25f50eda4918c\": container with ID starting with 3bc127f9d2c1cbb820f0beb3774da0fcd2882fb1c65c1bbabfb25f50eda4918c not found: ID does not exist" Oct 08 19:17:22 crc kubenswrapper[4988]: I1008 19:17:22.756041 4988 scope.go:117] "RemoveContainer" containerID="79af207fc0034c0c23ff4128f0c93601df28378dbd34b25bf7952caf8da6a7c4" Oct 08 19:17:22 crc kubenswrapper[4988]: E1008 19:17:22.756757 4988 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"79af207fc0034c0c23ff4128f0c93601df28378dbd34b25bf7952caf8da6a7c4\": container with ID starting with 79af207fc0034c0c23ff4128f0c93601df28378dbd34b25bf7952caf8da6a7c4 not found: ID does not exist" containerID="79af207fc0034c0c23ff4128f0c93601df28378dbd34b25bf7952caf8da6a7c4" Oct 08 19:17:22 crc kubenswrapper[4988]: I1008 19:17:22.756812 4988 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"79af207fc0034c0c23ff4128f0c93601df28378dbd34b25bf7952caf8da6a7c4"} err="failed to get container status \"79af207fc0034c0c23ff4128f0c93601df28378dbd34b25bf7952caf8da6a7c4\": rpc error: code = NotFound desc = could not find container \"79af207fc0034c0c23ff4128f0c93601df28378dbd34b25bf7952caf8da6a7c4\": container with ID starting with 79af207fc0034c0c23ff4128f0c93601df28378dbd34b25bf7952caf8da6a7c4 not found: ID does not exist" Oct 08 19:17:23 crc kubenswrapper[4988]: I1008 19:17:23.253836 4988 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1f41a3d7-55ff-49b4-8da0-4777210cdd5c" path="/var/lib/kubelet/pods/1f41a3d7-55ff-49b4-8da0-4777210cdd5c/volumes" Oct 08 19:18:53 crc kubenswrapper[4988]: I1008 19:18:53.338035 4988 patch_prober.go:28] interesting pod/machine-config-daemon-nm54f container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 19:18:53 crc kubenswrapper[4988]: I1008 19:18:53.338703 4988 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 19:18:55 crc kubenswrapper[4988]: I1008 19:18:55.295429 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-6v8w9"] Oct 08 19:18:55 crc kubenswrapper[4988]: E1008 19:18:55.296066 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1f41a3d7-55ff-49b4-8da0-4777210cdd5c" containerName="extract-utilities" Oct 08 19:18:55 crc kubenswrapper[4988]: I1008 19:18:55.296095 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="1f41a3d7-55ff-49b4-8da0-4777210cdd5c" containerName="extract-utilities" Oct 08 19:18:55 crc kubenswrapper[4988]: E1008 19:18:55.296121 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1f41a3d7-55ff-49b4-8da0-4777210cdd5c" containerName="extract-content" Oct 08 19:18:55 crc kubenswrapper[4988]: I1008 19:18:55.296137 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="1f41a3d7-55ff-49b4-8da0-4777210cdd5c" containerName="extract-content" Oct 08 19:18:55 crc kubenswrapper[4988]: E1008 19:18:55.296164 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1f41a3d7-55ff-49b4-8da0-4777210cdd5c" containerName="registry-server" Oct 08 19:18:55 crc kubenswrapper[4988]: I1008 19:18:55.296185 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="1f41a3d7-55ff-49b4-8da0-4777210cdd5c" containerName="registry-server" Oct 08 19:18:55 crc kubenswrapper[4988]: I1008 19:18:55.296608 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="1f41a3d7-55ff-49b4-8da0-4777210cdd5c" containerName="registry-server" Oct 08 19:18:55 crc kubenswrapper[4988]: I1008 19:18:55.306607 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-6v8w9" Oct 08 19:18:55 crc kubenswrapper[4988]: I1008 19:18:55.307576 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-6v8w9"] Oct 08 19:18:55 crc kubenswrapper[4988]: I1008 19:18:55.356867 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9275b003-e46f-4a5e-9608-e35f1117acd5-catalog-content\") pod \"redhat-operators-6v8w9\" (UID: \"9275b003-e46f-4a5e-9608-e35f1117acd5\") " pod="openshift-marketplace/redhat-operators-6v8w9" Oct 08 19:18:55 crc kubenswrapper[4988]: I1008 19:18:55.356939 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7fn26\" (UniqueName: \"kubernetes.io/projected/9275b003-e46f-4a5e-9608-e35f1117acd5-kube-api-access-7fn26\") pod \"redhat-operators-6v8w9\" (UID: \"9275b003-e46f-4a5e-9608-e35f1117acd5\") " pod="openshift-marketplace/redhat-operators-6v8w9" Oct 08 19:18:55 crc kubenswrapper[4988]: I1008 19:18:55.356995 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9275b003-e46f-4a5e-9608-e35f1117acd5-utilities\") pod \"redhat-operators-6v8w9\" (UID: \"9275b003-e46f-4a5e-9608-e35f1117acd5\") " pod="openshift-marketplace/redhat-operators-6v8w9" Oct 08 19:18:55 crc kubenswrapper[4988]: I1008 19:18:55.458720 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9275b003-e46f-4a5e-9608-e35f1117acd5-catalog-content\") pod \"redhat-operators-6v8w9\" (UID: \"9275b003-e46f-4a5e-9608-e35f1117acd5\") " pod="openshift-marketplace/redhat-operators-6v8w9" Oct 08 19:18:55 crc kubenswrapper[4988]: I1008 19:18:55.458790 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7fn26\" (UniqueName: \"kubernetes.io/projected/9275b003-e46f-4a5e-9608-e35f1117acd5-kube-api-access-7fn26\") pod \"redhat-operators-6v8w9\" (UID: \"9275b003-e46f-4a5e-9608-e35f1117acd5\") " pod="openshift-marketplace/redhat-operators-6v8w9" Oct 08 19:18:55 crc kubenswrapper[4988]: I1008 19:18:55.458836 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9275b003-e46f-4a5e-9608-e35f1117acd5-utilities\") pod \"redhat-operators-6v8w9\" (UID: \"9275b003-e46f-4a5e-9608-e35f1117acd5\") " pod="openshift-marketplace/redhat-operators-6v8w9" Oct 08 19:18:55 crc kubenswrapper[4988]: I1008 19:18:55.459305 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9275b003-e46f-4a5e-9608-e35f1117acd5-catalog-content\") pod \"redhat-operators-6v8w9\" (UID: \"9275b003-e46f-4a5e-9608-e35f1117acd5\") " pod="openshift-marketplace/redhat-operators-6v8w9" Oct 08 19:18:55 crc kubenswrapper[4988]: I1008 19:18:55.459321 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9275b003-e46f-4a5e-9608-e35f1117acd5-utilities\") pod \"redhat-operators-6v8w9\" (UID: \"9275b003-e46f-4a5e-9608-e35f1117acd5\") " pod="openshift-marketplace/redhat-operators-6v8w9" Oct 08 19:18:55 crc kubenswrapper[4988]: I1008 19:18:55.482665 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7fn26\" (UniqueName: \"kubernetes.io/projected/9275b003-e46f-4a5e-9608-e35f1117acd5-kube-api-access-7fn26\") pod \"redhat-operators-6v8w9\" (UID: \"9275b003-e46f-4a5e-9608-e35f1117acd5\") " pod="openshift-marketplace/redhat-operators-6v8w9" Oct 08 19:18:55 crc kubenswrapper[4988]: I1008 19:18:55.641157 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-6v8w9" Oct 08 19:18:56 crc kubenswrapper[4988]: I1008 19:18:56.085952 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-6v8w9"] Oct 08 19:18:56 crc kubenswrapper[4988]: I1008 19:18:56.507713 4988 generic.go:334] "Generic (PLEG): container finished" podID="9275b003-e46f-4a5e-9608-e35f1117acd5" containerID="a7667ad79e395bd9f9732984f47af311fb8021665bfff8939b2645cce32fcfc3" exitCode=0 Oct 08 19:18:56 crc kubenswrapper[4988]: I1008 19:18:56.507788 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-6v8w9" event={"ID":"9275b003-e46f-4a5e-9608-e35f1117acd5","Type":"ContainerDied","Data":"a7667ad79e395bd9f9732984f47af311fb8021665bfff8939b2645cce32fcfc3"} Oct 08 19:18:56 crc kubenswrapper[4988]: I1008 19:18:56.507830 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-6v8w9" event={"ID":"9275b003-e46f-4a5e-9608-e35f1117acd5","Type":"ContainerStarted","Data":"76484a1f0d99f9841d981460ea00aab007f771c5f71c1e471c76c37ed6f2b370"} Oct 08 19:18:57 crc kubenswrapper[4988]: I1008 19:18:57.522390 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-6v8w9" event={"ID":"9275b003-e46f-4a5e-9608-e35f1117acd5","Type":"ContainerStarted","Data":"3a77ff28c9e10c52ed17fda5c3e591514ad5ae9ca5fa8c18214d0d07a9083b05"} Oct 08 19:18:58 crc kubenswrapper[4988]: I1008 19:18:58.531420 4988 generic.go:334] "Generic (PLEG): container finished" podID="9275b003-e46f-4a5e-9608-e35f1117acd5" containerID="3a77ff28c9e10c52ed17fda5c3e591514ad5ae9ca5fa8c18214d0d07a9083b05" exitCode=0 Oct 08 19:18:58 crc kubenswrapper[4988]: I1008 19:18:58.531778 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-6v8w9" event={"ID":"9275b003-e46f-4a5e-9608-e35f1117acd5","Type":"ContainerDied","Data":"3a77ff28c9e10c52ed17fda5c3e591514ad5ae9ca5fa8c18214d0d07a9083b05"} Oct 08 19:18:59 crc kubenswrapper[4988]: I1008 19:18:59.540994 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-6v8w9" event={"ID":"9275b003-e46f-4a5e-9608-e35f1117acd5","Type":"ContainerStarted","Data":"097ec63252beddd3d56cb780a7587417770508a2c845a65a2c93540890f09f60"} Oct 08 19:18:59 crc kubenswrapper[4988]: I1008 19:18:59.560378 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-6v8w9" podStartSLOduration=2.09328347 podStartE2EDuration="4.560359817s" podCreationTimestamp="2025-10-08 19:18:55 +0000 UTC" firstStartedPulling="2025-10-08 19:18:56.510052523 +0000 UTC m=+4081.959895323" lastFinishedPulling="2025-10-08 19:18:58.97712844 +0000 UTC m=+4084.426971670" observedRunningTime="2025-10-08 19:18:59.555542785 +0000 UTC m=+4085.005385565" watchObservedRunningTime="2025-10-08 19:18:59.560359817 +0000 UTC m=+4085.010202587" Oct 08 19:19:05 crc kubenswrapper[4988]: I1008 19:19:05.642006 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-6v8w9" Oct 08 19:19:05 crc kubenswrapper[4988]: I1008 19:19:05.642581 4988 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-6v8w9" Oct 08 19:19:05 crc kubenswrapper[4988]: I1008 19:19:05.698528 4988 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-6v8w9" Oct 08 19:19:06 crc kubenswrapper[4988]: I1008 19:19:06.674809 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-6v8w9" Oct 08 19:19:06 crc kubenswrapper[4988]: I1008 19:19:06.762932 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-6v8w9"] Oct 08 19:19:08 crc kubenswrapper[4988]: I1008 19:19:08.618503 4988 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-6v8w9" podUID="9275b003-e46f-4a5e-9608-e35f1117acd5" containerName="registry-server" containerID="cri-o://097ec63252beddd3d56cb780a7587417770508a2c845a65a2c93540890f09f60" gracePeriod=2 Oct 08 19:19:09 crc kubenswrapper[4988]: I1008 19:19:09.498519 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-6v8w9" Oct 08 19:19:09 crc kubenswrapper[4988]: I1008 19:19:09.597027 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9275b003-e46f-4a5e-9608-e35f1117acd5-catalog-content\") pod \"9275b003-e46f-4a5e-9608-e35f1117acd5\" (UID: \"9275b003-e46f-4a5e-9608-e35f1117acd5\") " Oct 08 19:19:09 crc kubenswrapper[4988]: I1008 19:19:09.597235 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7fn26\" (UniqueName: \"kubernetes.io/projected/9275b003-e46f-4a5e-9608-e35f1117acd5-kube-api-access-7fn26\") pod \"9275b003-e46f-4a5e-9608-e35f1117acd5\" (UID: \"9275b003-e46f-4a5e-9608-e35f1117acd5\") " Oct 08 19:19:09 crc kubenswrapper[4988]: I1008 19:19:09.597336 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9275b003-e46f-4a5e-9608-e35f1117acd5-utilities\") pod \"9275b003-e46f-4a5e-9608-e35f1117acd5\" (UID: \"9275b003-e46f-4a5e-9608-e35f1117acd5\") " Oct 08 19:19:09 crc kubenswrapper[4988]: I1008 19:19:09.598873 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9275b003-e46f-4a5e-9608-e35f1117acd5-utilities" (OuterVolumeSpecName: "utilities") pod "9275b003-e46f-4a5e-9608-e35f1117acd5" (UID: "9275b003-e46f-4a5e-9608-e35f1117acd5"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 19:19:09 crc kubenswrapper[4988]: I1008 19:19:09.603849 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9275b003-e46f-4a5e-9608-e35f1117acd5-kube-api-access-7fn26" (OuterVolumeSpecName: "kube-api-access-7fn26") pod "9275b003-e46f-4a5e-9608-e35f1117acd5" (UID: "9275b003-e46f-4a5e-9608-e35f1117acd5"). InnerVolumeSpecName "kube-api-access-7fn26". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 19:19:09 crc kubenswrapper[4988]: I1008 19:19:09.629231 4988 generic.go:334] "Generic (PLEG): container finished" podID="9275b003-e46f-4a5e-9608-e35f1117acd5" containerID="097ec63252beddd3d56cb780a7587417770508a2c845a65a2c93540890f09f60" exitCode=0 Oct 08 19:19:09 crc kubenswrapper[4988]: I1008 19:19:09.629301 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-6v8w9" event={"ID":"9275b003-e46f-4a5e-9608-e35f1117acd5","Type":"ContainerDied","Data":"097ec63252beddd3d56cb780a7587417770508a2c845a65a2c93540890f09f60"} Oct 08 19:19:09 crc kubenswrapper[4988]: I1008 19:19:09.629353 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-6v8w9" event={"ID":"9275b003-e46f-4a5e-9608-e35f1117acd5","Type":"ContainerDied","Data":"76484a1f0d99f9841d981460ea00aab007f771c5f71c1e471c76c37ed6f2b370"} Oct 08 19:19:09 crc kubenswrapper[4988]: I1008 19:19:09.629375 4988 scope.go:117] "RemoveContainer" containerID="097ec63252beddd3d56cb780a7587417770508a2c845a65a2c93540890f09f60" Oct 08 19:19:09 crc kubenswrapper[4988]: I1008 19:19:09.629608 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-6v8w9" Oct 08 19:19:09 crc kubenswrapper[4988]: I1008 19:19:09.659308 4988 scope.go:117] "RemoveContainer" containerID="3a77ff28c9e10c52ed17fda5c3e591514ad5ae9ca5fa8c18214d0d07a9083b05" Oct 08 19:19:09 crc kubenswrapper[4988]: I1008 19:19:09.682333 4988 scope.go:117] "RemoveContainer" containerID="a7667ad79e395bd9f9732984f47af311fb8021665bfff8939b2645cce32fcfc3" Oct 08 19:19:09 crc kubenswrapper[4988]: I1008 19:19:09.699682 4988 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7fn26\" (UniqueName: \"kubernetes.io/projected/9275b003-e46f-4a5e-9608-e35f1117acd5-kube-api-access-7fn26\") on node \"crc\" DevicePath \"\"" Oct 08 19:19:09 crc kubenswrapper[4988]: I1008 19:19:09.699791 4988 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9275b003-e46f-4a5e-9608-e35f1117acd5-utilities\") on node \"crc\" DevicePath \"\"" Oct 08 19:19:09 crc kubenswrapper[4988]: I1008 19:19:09.715569 4988 scope.go:117] "RemoveContainer" containerID="097ec63252beddd3d56cb780a7587417770508a2c845a65a2c93540890f09f60" Oct 08 19:19:09 crc kubenswrapper[4988]: E1008 19:19:09.716116 4988 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"097ec63252beddd3d56cb780a7587417770508a2c845a65a2c93540890f09f60\": container with ID starting with 097ec63252beddd3d56cb780a7587417770508a2c845a65a2c93540890f09f60 not found: ID does not exist" containerID="097ec63252beddd3d56cb780a7587417770508a2c845a65a2c93540890f09f60" Oct 08 19:19:09 crc kubenswrapper[4988]: I1008 19:19:09.716148 4988 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"097ec63252beddd3d56cb780a7587417770508a2c845a65a2c93540890f09f60"} err="failed to get container status \"097ec63252beddd3d56cb780a7587417770508a2c845a65a2c93540890f09f60\": rpc error: code = NotFound desc = could not find container \"097ec63252beddd3d56cb780a7587417770508a2c845a65a2c93540890f09f60\": container with ID starting with 097ec63252beddd3d56cb780a7587417770508a2c845a65a2c93540890f09f60 not found: ID does not exist" Oct 08 19:19:09 crc kubenswrapper[4988]: I1008 19:19:09.716174 4988 scope.go:117] "RemoveContainer" containerID="3a77ff28c9e10c52ed17fda5c3e591514ad5ae9ca5fa8c18214d0d07a9083b05" Oct 08 19:19:09 crc kubenswrapper[4988]: E1008 19:19:09.716519 4988 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3a77ff28c9e10c52ed17fda5c3e591514ad5ae9ca5fa8c18214d0d07a9083b05\": container with ID starting with 3a77ff28c9e10c52ed17fda5c3e591514ad5ae9ca5fa8c18214d0d07a9083b05 not found: ID does not exist" containerID="3a77ff28c9e10c52ed17fda5c3e591514ad5ae9ca5fa8c18214d0d07a9083b05" Oct 08 19:19:09 crc kubenswrapper[4988]: I1008 19:19:09.716603 4988 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3a77ff28c9e10c52ed17fda5c3e591514ad5ae9ca5fa8c18214d0d07a9083b05"} err="failed to get container status \"3a77ff28c9e10c52ed17fda5c3e591514ad5ae9ca5fa8c18214d0d07a9083b05\": rpc error: code = NotFound desc = could not find container \"3a77ff28c9e10c52ed17fda5c3e591514ad5ae9ca5fa8c18214d0d07a9083b05\": container with ID starting with 3a77ff28c9e10c52ed17fda5c3e591514ad5ae9ca5fa8c18214d0d07a9083b05 not found: ID does not exist" Oct 08 19:19:09 crc kubenswrapper[4988]: I1008 19:19:09.716621 4988 scope.go:117] "RemoveContainer" containerID="a7667ad79e395bd9f9732984f47af311fb8021665bfff8939b2645cce32fcfc3" Oct 08 19:19:09 crc kubenswrapper[4988]: E1008 19:19:09.717089 4988 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a7667ad79e395bd9f9732984f47af311fb8021665bfff8939b2645cce32fcfc3\": container with ID starting with a7667ad79e395bd9f9732984f47af311fb8021665bfff8939b2645cce32fcfc3 not found: ID does not exist" containerID="a7667ad79e395bd9f9732984f47af311fb8021665bfff8939b2645cce32fcfc3" Oct 08 19:19:09 crc kubenswrapper[4988]: I1008 19:19:09.717139 4988 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a7667ad79e395bd9f9732984f47af311fb8021665bfff8939b2645cce32fcfc3"} err="failed to get container status \"a7667ad79e395bd9f9732984f47af311fb8021665bfff8939b2645cce32fcfc3\": rpc error: code = NotFound desc = could not find container \"a7667ad79e395bd9f9732984f47af311fb8021665bfff8939b2645cce32fcfc3\": container with ID starting with a7667ad79e395bd9f9732984f47af311fb8021665bfff8939b2645cce32fcfc3 not found: ID does not exist" Oct 08 19:19:09 crc kubenswrapper[4988]: I1008 19:19:09.724230 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9275b003-e46f-4a5e-9608-e35f1117acd5-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "9275b003-e46f-4a5e-9608-e35f1117acd5" (UID: "9275b003-e46f-4a5e-9608-e35f1117acd5"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 19:19:09 crc kubenswrapper[4988]: I1008 19:19:09.801011 4988 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9275b003-e46f-4a5e-9608-e35f1117acd5-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 08 19:19:09 crc kubenswrapper[4988]: I1008 19:19:09.963998 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-6v8w9"] Oct 08 19:19:09 crc kubenswrapper[4988]: I1008 19:19:09.969669 4988 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-6v8w9"] Oct 08 19:19:11 crc kubenswrapper[4988]: I1008 19:19:11.266495 4988 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9275b003-e46f-4a5e-9608-e35f1117acd5" path="/var/lib/kubelet/pods/9275b003-e46f-4a5e-9608-e35f1117acd5/volumes" Oct 08 19:19:23 crc kubenswrapper[4988]: I1008 19:19:23.338701 4988 patch_prober.go:28] interesting pod/machine-config-daemon-nm54f container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 19:19:23 crc kubenswrapper[4988]: I1008 19:19:23.339209 4988 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 19:19:41 crc kubenswrapper[4988]: I1008 19:19:41.907370 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-glkmr"] Oct 08 19:19:41 crc kubenswrapper[4988]: E1008 19:19:41.908512 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9275b003-e46f-4a5e-9608-e35f1117acd5" containerName="registry-server" Oct 08 19:19:41 crc kubenswrapper[4988]: I1008 19:19:41.908537 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="9275b003-e46f-4a5e-9608-e35f1117acd5" containerName="registry-server" Oct 08 19:19:41 crc kubenswrapper[4988]: E1008 19:19:41.908580 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9275b003-e46f-4a5e-9608-e35f1117acd5" containerName="extract-content" Oct 08 19:19:41 crc kubenswrapper[4988]: I1008 19:19:41.908599 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="9275b003-e46f-4a5e-9608-e35f1117acd5" containerName="extract-content" Oct 08 19:19:41 crc kubenswrapper[4988]: E1008 19:19:41.908657 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9275b003-e46f-4a5e-9608-e35f1117acd5" containerName="extract-utilities" Oct 08 19:19:41 crc kubenswrapper[4988]: I1008 19:19:41.908673 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="9275b003-e46f-4a5e-9608-e35f1117acd5" containerName="extract-utilities" Oct 08 19:19:41 crc kubenswrapper[4988]: I1008 19:19:41.908940 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="9275b003-e46f-4a5e-9608-e35f1117acd5" containerName="registry-server" Oct 08 19:19:41 crc kubenswrapper[4988]: I1008 19:19:41.911152 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-glkmr" Oct 08 19:19:41 crc kubenswrapper[4988]: I1008 19:19:41.921168 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-glkmr"] Oct 08 19:19:42 crc kubenswrapper[4988]: I1008 19:19:42.020525 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zk8cf\" (UniqueName: \"kubernetes.io/projected/416bd9da-ad07-4798-93f2-304d65f24a82-kube-api-access-zk8cf\") pod \"certified-operators-glkmr\" (UID: \"416bd9da-ad07-4798-93f2-304d65f24a82\") " pod="openshift-marketplace/certified-operators-glkmr" Oct 08 19:19:42 crc kubenswrapper[4988]: I1008 19:19:42.020606 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/416bd9da-ad07-4798-93f2-304d65f24a82-catalog-content\") pod \"certified-operators-glkmr\" (UID: \"416bd9da-ad07-4798-93f2-304d65f24a82\") " pod="openshift-marketplace/certified-operators-glkmr" Oct 08 19:19:42 crc kubenswrapper[4988]: I1008 19:19:42.020658 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/416bd9da-ad07-4798-93f2-304d65f24a82-utilities\") pod \"certified-operators-glkmr\" (UID: \"416bd9da-ad07-4798-93f2-304d65f24a82\") " pod="openshift-marketplace/certified-operators-glkmr" Oct 08 19:19:42 crc kubenswrapper[4988]: I1008 19:19:42.122577 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zk8cf\" (UniqueName: \"kubernetes.io/projected/416bd9da-ad07-4798-93f2-304d65f24a82-kube-api-access-zk8cf\") pod \"certified-operators-glkmr\" (UID: \"416bd9da-ad07-4798-93f2-304d65f24a82\") " pod="openshift-marketplace/certified-operators-glkmr" Oct 08 19:19:42 crc kubenswrapper[4988]: I1008 19:19:42.122633 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/416bd9da-ad07-4798-93f2-304d65f24a82-catalog-content\") pod \"certified-operators-glkmr\" (UID: \"416bd9da-ad07-4798-93f2-304d65f24a82\") " pod="openshift-marketplace/certified-operators-glkmr" Oct 08 19:19:42 crc kubenswrapper[4988]: I1008 19:19:42.122672 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/416bd9da-ad07-4798-93f2-304d65f24a82-utilities\") pod \"certified-operators-glkmr\" (UID: \"416bd9da-ad07-4798-93f2-304d65f24a82\") " pod="openshift-marketplace/certified-operators-glkmr" Oct 08 19:19:42 crc kubenswrapper[4988]: I1008 19:19:42.123066 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/416bd9da-ad07-4798-93f2-304d65f24a82-utilities\") pod \"certified-operators-glkmr\" (UID: \"416bd9da-ad07-4798-93f2-304d65f24a82\") " pod="openshift-marketplace/certified-operators-glkmr" Oct 08 19:19:42 crc kubenswrapper[4988]: I1008 19:19:42.123277 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/416bd9da-ad07-4798-93f2-304d65f24a82-catalog-content\") pod \"certified-operators-glkmr\" (UID: \"416bd9da-ad07-4798-93f2-304d65f24a82\") " pod="openshift-marketplace/certified-operators-glkmr" Oct 08 19:19:42 crc kubenswrapper[4988]: I1008 19:19:42.146019 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zk8cf\" (UniqueName: \"kubernetes.io/projected/416bd9da-ad07-4798-93f2-304d65f24a82-kube-api-access-zk8cf\") pod \"certified-operators-glkmr\" (UID: \"416bd9da-ad07-4798-93f2-304d65f24a82\") " pod="openshift-marketplace/certified-operators-glkmr" Oct 08 19:19:42 crc kubenswrapper[4988]: I1008 19:19:42.231977 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-glkmr" Oct 08 19:19:42 crc kubenswrapper[4988]: I1008 19:19:42.512005 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-glkmr"] Oct 08 19:19:42 crc kubenswrapper[4988]: I1008 19:19:42.953437 4988 generic.go:334] "Generic (PLEG): container finished" podID="416bd9da-ad07-4798-93f2-304d65f24a82" containerID="3fce47776133cf6fd9a06083c075eb2e8f68780bdec3eaae542476e0ca3d453e" exitCode=0 Oct 08 19:19:42 crc kubenswrapper[4988]: I1008 19:19:42.953482 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-glkmr" event={"ID":"416bd9da-ad07-4798-93f2-304d65f24a82","Type":"ContainerDied","Data":"3fce47776133cf6fd9a06083c075eb2e8f68780bdec3eaae542476e0ca3d453e"} Oct 08 19:19:42 crc kubenswrapper[4988]: I1008 19:19:42.953513 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-glkmr" event={"ID":"416bd9da-ad07-4798-93f2-304d65f24a82","Type":"ContainerStarted","Data":"098c19a3097751aab9b81e44ff1890fbaed17ebdb080ff012819e32e4e061fb5"} Oct 08 19:19:43 crc kubenswrapper[4988]: I1008 19:19:43.961044 4988 generic.go:334] "Generic (PLEG): container finished" podID="416bd9da-ad07-4798-93f2-304d65f24a82" containerID="bc6ed572d8da8b957053390f336758a694c0bb508383f3fa97d47d6c1fd382ff" exitCode=0 Oct 08 19:19:43 crc kubenswrapper[4988]: I1008 19:19:43.961092 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-glkmr" event={"ID":"416bd9da-ad07-4798-93f2-304d65f24a82","Type":"ContainerDied","Data":"bc6ed572d8da8b957053390f336758a694c0bb508383f3fa97d47d6c1fd382ff"} Oct 08 19:19:44 crc kubenswrapper[4988]: I1008 19:19:44.974332 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-glkmr" event={"ID":"416bd9da-ad07-4798-93f2-304d65f24a82","Type":"ContainerStarted","Data":"74fc57f64764c2b11d7a0cc9f8394e8b0594783115879fa52aa0b86889b415cd"} Oct 08 19:19:44 crc kubenswrapper[4988]: I1008 19:19:44.998344 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-glkmr" podStartSLOduration=2.607390056 podStartE2EDuration="3.998323327s" podCreationTimestamp="2025-10-08 19:19:41 +0000 UTC" firstStartedPulling="2025-10-08 19:19:42.960298942 +0000 UTC m=+4128.410141732" lastFinishedPulling="2025-10-08 19:19:44.351232223 +0000 UTC m=+4129.801075003" observedRunningTime="2025-10-08 19:19:44.992251815 +0000 UTC m=+4130.442094595" watchObservedRunningTime="2025-10-08 19:19:44.998323327 +0000 UTC m=+4130.448166107" Oct 08 19:19:52 crc kubenswrapper[4988]: I1008 19:19:52.232890 4988 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-glkmr" Oct 08 19:19:52 crc kubenswrapper[4988]: I1008 19:19:52.233565 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-glkmr" Oct 08 19:19:52 crc kubenswrapper[4988]: I1008 19:19:52.313306 4988 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-glkmr" Oct 08 19:19:53 crc kubenswrapper[4988]: I1008 19:19:53.124105 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-glkmr" Oct 08 19:19:53 crc kubenswrapper[4988]: I1008 19:19:53.174594 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-glkmr"] Oct 08 19:19:53 crc kubenswrapper[4988]: I1008 19:19:53.338500 4988 patch_prober.go:28] interesting pod/machine-config-daemon-nm54f container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 19:19:53 crc kubenswrapper[4988]: I1008 19:19:53.338581 4988 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 19:19:53 crc kubenswrapper[4988]: I1008 19:19:53.338637 4988 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" Oct 08 19:19:53 crc kubenswrapper[4988]: I1008 19:19:53.339480 4988 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"eca8e1f95f09ab5d917be9719293edda80592194583b0c97c98c1d0e281973e8"} pod="openshift-machine-config-operator/machine-config-daemon-nm54f" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 08 19:19:53 crc kubenswrapper[4988]: I1008 19:19:53.339559 4988 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" containerName="machine-config-daemon" containerID="cri-o://eca8e1f95f09ab5d917be9719293edda80592194583b0c97c98c1d0e281973e8" gracePeriod=600 Oct 08 19:19:54 crc kubenswrapper[4988]: I1008 19:19:54.056754 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" event={"ID":"f328a20b-5283-407e-8a93-2ab2da599f18","Type":"ContainerDied","Data":"eca8e1f95f09ab5d917be9719293edda80592194583b0c97c98c1d0e281973e8"} Oct 08 19:19:54 crc kubenswrapper[4988]: I1008 19:19:54.057128 4988 scope.go:117] "RemoveContainer" containerID="af2043d46f0bbf33ed02542990cdf54feefa63a7c8331b6a70288efbb7b7a3ae" Oct 08 19:19:54 crc kubenswrapper[4988]: I1008 19:19:54.057431 4988 generic.go:334] "Generic (PLEG): container finished" podID="f328a20b-5283-407e-8a93-2ab2da599f18" containerID="eca8e1f95f09ab5d917be9719293edda80592194583b0c97c98c1d0e281973e8" exitCode=0 Oct 08 19:19:55 crc kubenswrapper[4988]: I1008 19:19:55.071120 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" event={"ID":"f328a20b-5283-407e-8a93-2ab2da599f18","Type":"ContainerStarted","Data":"20477a4d0d4c3448d84d69456b20adede3fe2d7c9fa57e53eb98b6ad7aabc27d"} Oct 08 19:19:55 crc kubenswrapper[4988]: I1008 19:19:55.071171 4988 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-glkmr" podUID="416bd9da-ad07-4798-93f2-304d65f24a82" containerName="registry-server" containerID="cri-o://74fc57f64764c2b11d7a0cc9f8394e8b0594783115879fa52aa0b86889b415cd" gracePeriod=2 Oct 08 19:19:55 crc kubenswrapper[4988]: I1008 19:19:55.559046 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-glkmr" Oct 08 19:19:55 crc kubenswrapper[4988]: I1008 19:19:55.611537 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/416bd9da-ad07-4798-93f2-304d65f24a82-catalog-content\") pod \"416bd9da-ad07-4798-93f2-304d65f24a82\" (UID: \"416bd9da-ad07-4798-93f2-304d65f24a82\") " Oct 08 19:19:55 crc kubenswrapper[4988]: I1008 19:19:55.611602 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zk8cf\" (UniqueName: \"kubernetes.io/projected/416bd9da-ad07-4798-93f2-304d65f24a82-kube-api-access-zk8cf\") pod \"416bd9da-ad07-4798-93f2-304d65f24a82\" (UID: \"416bd9da-ad07-4798-93f2-304d65f24a82\") " Oct 08 19:19:55 crc kubenswrapper[4988]: I1008 19:19:55.611758 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/416bd9da-ad07-4798-93f2-304d65f24a82-utilities\") pod \"416bd9da-ad07-4798-93f2-304d65f24a82\" (UID: \"416bd9da-ad07-4798-93f2-304d65f24a82\") " Oct 08 19:19:55 crc kubenswrapper[4988]: I1008 19:19:55.612867 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/416bd9da-ad07-4798-93f2-304d65f24a82-utilities" (OuterVolumeSpecName: "utilities") pod "416bd9da-ad07-4798-93f2-304d65f24a82" (UID: "416bd9da-ad07-4798-93f2-304d65f24a82"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 19:19:55 crc kubenswrapper[4988]: I1008 19:19:55.619529 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/416bd9da-ad07-4798-93f2-304d65f24a82-kube-api-access-zk8cf" (OuterVolumeSpecName: "kube-api-access-zk8cf") pod "416bd9da-ad07-4798-93f2-304d65f24a82" (UID: "416bd9da-ad07-4798-93f2-304d65f24a82"). InnerVolumeSpecName "kube-api-access-zk8cf". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 19:19:55 crc kubenswrapper[4988]: I1008 19:19:55.666090 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/416bd9da-ad07-4798-93f2-304d65f24a82-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "416bd9da-ad07-4798-93f2-304d65f24a82" (UID: "416bd9da-ad07-4798-93f2-304d65f24a82"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 19:19:55 crc kubenswrapper[4988]: I1008 19:19:55.713536 4988 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/416bd9da-ad07-4798-93f2-304d65f24a82-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 08 19:19:55 crc kubenswrapper[4988]: I1008 19:19:55.713576 4988 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zk8cf\" (UniqueName: \"kubernetes.io/projected/416bd9da-ad07-4798-93f2-304d65f24a82-kube-api-access-zk8cf\") on node \"crc\" DevicePath \"\"" Oct 08 19:19:55 crc kubenswrapper[4988]: I1008 19:19:55.713592 4988 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/416bd9da-ad07-4798-93f2-304d65f24a82-utilities\") on node \"crc\" DevicePath \"\"" Oct 08 19:19:56 crc kubenswrapper[4988]: I1008 19:19:56.083634 4988 generic.go:334] "Generic (PLEG): container finished" podID="416bd9da-ad07-4798-93f2-304d65f24a82" containerID="74fc57f64764c2b11d7a0cc9f8394e8b0594783115879fa52aa0b86889b415cd" exitCode=0 Oct 08 19:19:56 crc kubenswrapper[4988]: I1008 19:19:56.083720 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-glkmr" Oct 08 19:19:56 crc kubenswrapper[4988]: I1008 19:19:56.083709 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-glkmr" event={"ID":"416bd9da-ad07-4798-93f2-304d65f24a82","Type":"ContainerDied","Data":"74fc57f64764c2b11d7a0cc9f8394e8b0594783115879fa52aa0b86889b415cd"} Oct 08 19:19:56 crc kubenswrapper[4988]: I1008 19:19:56.084220 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-glkmr" event={"ID":"416bd9da-ad07-4798-93f2-304d65f24a82","Type":"ContainerDied","Data":"098c19a3097751aab9b81e44ff1890fbaed17ebdb080ff012819e32e4e061fb5"} Oct 08 19:19:56 crc kubenswrapper[4988]: I1008 19:19:56.084260 4988 scope.go:117] "RemoveContainer" containerID="74fc57f64764c2b11d7a0cc9f8394e8b0594783115879fa52aa0b86889b415cd" Oct 08 19:19:56 crc kubenswrapper[4988]: I1008 19:19:56.119780 4988 scope.go:117] "RemoveContainer" containerID="bc6ed572d8da8b957053390f336758a694c0bb508383f3fa97d47d6c1fd382ff" Oct 08 19:19:56 crc kubenswrapper[4988]: I1008 19:19:56.143167 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-glkmr"] Oct 08 19:19:56 crc kubenswrapper[4988]: I1008 19:19:56.151495 4988 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-glkmr"] Oct 08 19:19:56 crc kubenswrapper[4988]: I1008 19:19:56.169724 4988 scope.go:117] "RemoveContainer" containerID="3fce47776133cf6fd9a06083c075eb2e8f68780bdec3eaae542476e0ca3d453e" Oct 08 19:19:56 crc kubenswrapper[4988]: I1008 19:19:56.204672 4988 scope.go:117] "RemoveContainer" containerID="74fc57f64764c2b11d7a0cc9f8394e8b0594783115879fa52aa0b86889b415cd" Oct 08 19:19:56 crc kubenswrapper[4988]: E1008 19:19:56.205100 4988 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"74fc57f64764c2b11d7a0cc9f8394e8b0594783115879fa52aa0b86889b415cd\": container with ID starting with 74fc57f64764c2b11d7a0cc9f8394e8b0594783115879fa52aa0b86889b415cd not found: ID does not exist" containerID="74fc57f64764c2b11d7a0cc9f8394e8b0594783115879fa52aa0b86889b415cd" Oct 08 19:19:56 crc kubenswrapper[4988]: I1008 19:19:56.205143 4988 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"74fc57f64764c2b11d7a0cc9f8394e8b0594783115879fa52aa0b86889b415cd"} err="failed to get container status \"74fc57f64764c2b11d7a0cc9f8394e8b0594783115879fa52aa0b86889b415cd\": rpc error: code = NotFound desc = could not find container \"74fc57f64764c2b11d7a0cc9f8394e8b0594783115879fa52aa0b86889b415cd\": container with ID starting with 74fc57f64764c2b11d7a0cc9f8394e8b0594783115879fa52aa0b86889b415cd not found: ID does not exist" Oct 08 19:19:56 crc kubenswrapper[4988]: I1008 19:19:56.205170 4988 scope.go:117] "RemoveContainer" containerID="bc6ed572d8da8b957053390f336758a694c0bb508383f3fa97d47d6c1fd382ff" Oct 08 19:19:56 crc kubenswrapper[4988]: E1008 19:19:56.205657 4988 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bc6ed572d8da8b957053390f336758a694c0bb508383f3fa97d47d6c1fd382ff\": container with ID starting with bc6ed572d8da8b957053390f336758a694c0bb508383f3fa97d47d6c1fd382ff not found: ID does not exist" containerID="bc6ed572d8da8b957053390f336758a694c0bb508383f3fa97d47d6c1fd382ff" Oct 08 19:19:56 crc kubenswrapper[4988]: I1008 19:19:56.205692 4988 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bc6ed572d8da8b957053390f336758a694c0bb508383f3fa97d47d6c1fd382ff"} err="failed to get container status \"bc6ed572d8da8b957053390f336758a694c0bb508383f3fa97d47d6c1fd382ff\": rpc error: code = NotFound desc = could not find container \"bc6ed572d8da8b957053390f336758a694c0bb508383f3fa97d47d6c1fd382ff\": container with ID starting with bc6ed572d8da8b957053390f336758a694c0bb508383f3fa97d47d6c1fd382ff not found: ID does not exist" Oct 08 19:19:56 crc kubenswrapper[4988]: I1008 19:19:56.205718 4988 scope.go:117] "RemoveContainer" containerID="3fce47776133cf6fd9a06083c075eb2e8f68780bdec3eaae542476e0ca3d453e" Oct 08 19:19:56 crc kubenswrapper[4988]: E1008 19:19:56.206107 4988 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3fce47776133cf6fd9a06083c075eb2e8f68780bdec3eaae542476e0ca3d453e\": container with ID starting with 3fce47776133cf6fd9a06083c075eb2e8f68780bdec3eaae542476e0ca3d453e not found: ID does not exist" containerID="3fce47776133cf6fd9a06083c075eb2e8f68780bdec3eaae542476e0ca3d453e" Oct 08 19:19:56 crc kubenswrapper[4988]: I1008 19:19:56.206133 4988 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3fce47776133cf6fd9a06083c075eb2e8f68780bdec3eaae542476e0ca3d453e"} err="failed to get container status \"3fce47776133cf6fd9a06083c075eb2e8f68780bdec3eaae542476e0ca3d453e\": rpc error: code = NotFound desc = could not find container \"3fce47776133cf6fd9a06083c075eb2e8f68780bdec3eaae542476e0ca3d453e\": container with ID starting with 3fce47776133cf6fd9a06083c075eb2e8f68780bdec3eaae542476e0ca3d453e not found: ID does not exist" Oct 08 19:19:57 crc kubenswrapper[4988]: I1008 19:19:57.253314 4988 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="416bd9da-ad07-4798-93f2-304d65f24a82" path="/var/lib/kubelet/pods/416bd9da-ad07-4798-93f2-304d65f24a82/volumes" Oct 08 19:22:23 crc kubenswrapper[4988]: I1008 19:22:23.338556 4988 patch_prober.go:28] interesting pod/machine-config-daemon-nm54f container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 19:22:23 crc kubenswrapper[4988]: I1008 19:22:23.339167 4988 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 19:22:27 crc kubenswrapper[4988]: I1008 19:22:27.408720 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-cbs84"] Oct 08 19:22:27 crc kubenswrapper[4988]: E1008 19:22:27.409504 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="416bd9da-ad07-4798-93f2-304d65f24a82" containerName="registry-server" Oct 08 19:22:27 crc kubenswrapper[4988]: I1008 19:22:27.409544 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="416bd9da-ad07-4798-93f2-304d65f24a82" containerName="registry-server" Oct 08 19:22:27 crc kubenswrapper[4988]: E1008 19:22:27.409568 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="416bd9da-ad07-4798-93f2-304d65f24a82" containerName="extract-content" Oct 08 19:22:27 crc kubenswrapper[4988]: I1008 19:22:27.409579 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="416bd9da-ad07-4798-93f2-304d65f24a82" containerName="extract-content" Oct 08 19:22:27 crc kubenswrapper[4988]: E1008 19:22:27.409602 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="416bd9da-ad07-4798-93f2-304d65f24a82" containerName="extract-utilities" Oct 08 19:22:27 crc kubenswrapper[4988]: I1008 19:22:27.409616 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="416bd9da-ad07-4798-93f2-304d65f24a82" containerName="extract-utilities" Oct 08 19:22:27 crc kubenswrapper[4988]: I1008 19:22:27.409883 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="416bd9da-ad07-4798-93f2-304d65f24a82" containerName="registry-server" Oct 08 19:22:27 crc kubenswrapper[4988]: I1008 19:22:27.411221 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-cbs84" Oct 08 19:22:27 crc kubenswrapper[4988]: I1008 19:22:27.428882 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-cbs84"] Oct 08 19:22:27 crc kubenswrapper[4988]: I1008 19:22:27.463669 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3961388f-9e20-4a53-97fc-36bdd716e679-utilities\") pod \"redhat-marketplace-cbs84\" (UID: \"3961388f-9e20-4a53-97fc-36bdd716e679\") " pod="openshift-marketplace/redhat-marketplace-cbs84" Oct 08 19:22:27 crc kubenswrapper[4988]: I1008 19:22:27.463837 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r6vmd\" (UniqueName: \"kubernetes.io/projected/3961388f-9e20-4a53-97fc-36bdd716e679-kube-api-access-r6vmd\") pod \"redhat-marketplace-cbs84\" (UID: \"3961388f-9e20-4a53-97fc-36bdd716e679\") " pod="openshift-marketplace/redhat-marketplace-cbs84" Oct 08 19:22:27 crc kubenswrapper[4988]: I1008 19:22:27.464173 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3961388f-9e20-4a53-97fc-36bdd716e679-catalog-content\") pod \"redhat-marketplace-cbs84\" (UID: \"3961388f-9e20-4a53-97fc-36bdd716e679\") " pod="openshift-marketplace/redhat-marketplace-cbs84" Oct 08 19:22:27 crc kubenswrapper[4988]: I1008 19:22:27.564937 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3961388f-9e20-4a53-97fc-36bdd716e679-utilities\") pod \"redhat-marketplace-cbs84\" (UID: \"3961388f-9e20-4a53-97fc-36bdd716e679\") " pod="openshift-marketplace/redhat-marketplace-cbs84" Oct 08 19:22:27 crc kubenswrapper[4988]: I1008 19:22:27.565225 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r6vmd\" (UniqueName: \"kubernetes.io/projected/3961388f-9e20-4a53-97fc-36bdd716e679-kube-api-access-r6vmd\") pod \"redhat-marketplace-cbs84\" (UID: \"3961388f-9e20-4a53-97fc-36bdd716e679\") " pod="openshift-marketplace/redhat-marketplace-cbs84" Oct 08 19:22:27 crc kubenswrapper[4988]: I1008 19:22:27.565377 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3961388f-9e20-4a53-97fc-36bdd716e679-catalog-content\") pod \"redhat-marketplace-cbs84\" (UID: \"3961388f-9e20-4a53-97fc-36bdd716e679\") " pod="openshift-marketplace/redhat-marketplace-cbs84" Oct 08 19:22:27 crc kubenswrapper[4988]: I1008 19:22:27.565489 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3961388f-9e20-4a53-97fc-36bdd716e679-utilities\") pod \"redhat-marketplace-cbs84\" (UID: \"3961388f-9e20-4a53-97fc-36bdd716e679\") " pod="openshift-marketplace/redhat-marketplace-cbs84" Oct 08 19:22:27 crc kubenswrapper[4988]: I1008 19:22:27.566178 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3961388f-9e20-4a53-97fc-36bdd716e679-catalog-content\") pod \"redhat-marketplace-cbs84\" (UID: \"3961388f-9e20-4a53-97fc-36bdd716e679\") " pod="openshift-marketplace/redhat-marketplace-cbs84" Oct 08 19:22:27 crc kubenswrapper[4988]: I1008 19:22:27.590134 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r6vmd\" (UniqueName: \"kubernetes.io/projected/3961388f-9e20-4a53-97fc-36bdd716e679-kube-api-access-r6vmd\") pod \"redhat-marketplace-cbs84\" (UID: \"3961388f-9e20-4a53-97fc-36bdd716e679\") " pod="openshift-marketplace/redhat-marketplace-cbs84" Oct 08 19:22:27 crc kubenswrapper[4988]: I1008 19:22:27.738055 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-cbs84" Oct 08 19:22:28 crc kubenswrapper[4988]: I1008 19:22:28.194470 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-cbs84"] Oct 08 19:22:28 crc kubenswrapper[4988]: W1008 19:22:28.209698 4988 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3961388f_9e20_4a53_97fc_36bdd716e679.slice/crio-e7ce49a5c1793ffa74d8163ba255470fa3b56511ba6b2488a1e92d4694bbb6ce WatchSource:0}: Error finding container e7ce49a5c1793ffa74d8163ba255470fa3b56511ba6b2488a1e92d4694bbb6ce: Status 404 returned error can't find the container with id e7ce49a5c1793ffa74d8163ba255470fa3b56511ba6b2488a1e92d4694bbb6ce Oct 08 19:22:28 crc kubenswrapper[4988]: I1008 19:22:28.532589 4988 generic.go:334] "Generic (PLEG): container finished" podID="3961388f-9e20-4a53-97fc-36bdd716e679" containerID="d6b21b601d16472dce4c53aeffceaee8eddce3d845ccb7ea9714192f083bb85b" exitCode=0 Oct 08 19:22:28 crc kubenswrapper[4988]: I1008 19:22:28.532694 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-cbs84" event={"ID":"3961388f-9e20-4a53-97fc-36bdd716e679","Type":"ContainerDied","Data":"d6b21b601d16472dce4c53aeffceaee8eddce3d845ccb7ea9714192f083bb85b"} Oct 08 19:22:28 crc kubenswrapper[4988]: I1008 19:22:28.533020 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-cbs84" event={"ID":"3961388f-9e20-4a53-97fc-36bdd716e679","Type":"ContainerStarted","Data":"e7ce49a5c1793ffa74d8163ba255470fa3b56511ba6b2488a1e92d4694bbb6ce"} Oct 08 19:22:28 crc kubenswrapper[4988]: I1008 19:22:28.534523 4988 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 08 19:22:29 crc kubenswrapper[4988]: I1008 19:22:29.545817 4988 generic.go:334] "Generic (PLEG): container finished" podID="3961388f-9e20-4a53-97fc-36bdd716e679" containerID="8684b7542c8479bf77c19289214625e4ca323fa04e81a4d595f9ea6b83e57a5e" exitCode=0 Oct 08 19:22:29 crc kubenswrapper[4988]: I1008 19:22:29.545929 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-cbs84" event={"ID":"3961388f-9e20-4a53-97fc-36bdd716e679","Type":"ContainerDied","Data":"8684b7542c8479bf77c19289214625e4ca323fa04e81a4d595f9ea6b83e57a5e"} Oct 08 19:22:30 crc kubenswrapper[4988]: I1008 19:22:30.562395 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-cbs84" event={"ID":"3961388f-9e20-4a53-97fc-36bdd716e679","Type":"ContainerStarted","Data":"f6fe63cab862dc5d27b31154801d6ad92861ec92fe4017c137d8c8cde4ab7061"} Oct 08 19:22:30 crc kubenswrapper[4988]: I1008 19:22:30.585310 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-cbs84" podStartSLOduration=1.94383704 podStartE2EDuration="3.585263137s" podCreationTimestamp="2025-10-08 19:22:27 +0000 UTC" firstStartedPulling="2025-10-08 19:22:28.534106537 +0000 UTC m=+4293.983949317" lastFinishedPulling="2025-10-08 19:22:30.175532634 +0000 UTC m=+4295.625375414" observedRunningTime="2025-10-08 19:22:30.578327377 +0000 UTC m=+4296.028170157" watchObservedRunningTime="2025-10-08 19:22:30.585263137 +0000 UTC m=+4296.035105907" Oct 08 19:22:37 crc kubenswrapper[4988]: I1008 19:22:37.738895 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-cbs84" Oct 08 19:22:37 crc kubenswrapper[4988]: I1008 19:22:37.739207 4988 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-cbs84" Oct 08 19:22:37 crc kubenswrapper[4988]: I1008 19:22:37.936656 4988 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-cbs84" Oct 08 19:22:38 crc kubenswrapper[4988]: I1008 19:22:38.704270 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-cbs84" Oct 08 19:22:38 crc kubenswrapper[4988]: I1008 19:22:38.750917 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-cbs84"] Oct 08 19:22:40 crc kubenswrapper[4988]: I1008 19:22:40.657852 4988 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-cbs84" podUID="3961388f-9e20-4a53-97fc-36bdd716e679" containerName="registry-server" containerID="cri-o://f6fe63cab862dc5d27b31154801d6ad92861ec92fe4017c137d8c8cde4ab7061" gracePeriod=2 Oct 08 19:22:41 crc kubenswrapper[4988]: I1008 19:22:41.135009 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-cbs84" Oct 08 19:22:41 crc kubenswrapper[4988]: I1008 19:22:41.294679 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3961388f-9e20-4a53-97fc-36bdd716e679-catalog-content\") pod \"3961388f-9e20-4a53-97fc-36bdd716e679\" (UID: \"3961388f-9e20-4a53-97fc-36bdd716e679\") " Oct 08 19:22:41 crc kubenswrapper[4988]: I1008 19:22:41.295269 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3961388f-9e20-4a53-97fc-36bdd716e679-utilities\") pod \"3961388f-9e20-4a53-97fc-36bdd716e679\" (UID: \"3961388f-9e20-4a53-97fc-36bdd716e679\") " Oct 08 19:22:41 crc kubenswrapper[4988]: I1008 19:22:41.295382 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r6vmd\" (UniqueName: \"kubernetes.io/projected/3961388f-9e20-4a53-97fc-36bdd716e679-kube-api-access-r6vmd\") pod \"3961388f-9e20-4a53-97fc-36bdd716e679\" (UID: \"3961388f-9e20-4a53-97fc-36bdd716e679\") " Oct 08 19:22:41 crc kubenswrapper[4988]: I1008 19:22:41.297783 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3961388f-9e20-4a53-97fc-36bdd716e679-utilities" (OuterVolumeSpecName: "utilities") pod "3961388f-9e20-4a53-97fc-36bdd716e679" (UID: "3961388f-9e20-4a53-97fc-36bdd716e679"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 19:22:41 crc kubenswrapper[4988]: I1008 19:22:41.303867 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3961388f-9e20-4a53-97fc-36bdd716e679-kube-api-access-r6vmd" (OuterVolumeSpecName: "kube-api-access-r6vmd") pod "3961388f-9e20-4a53-97fc-36bdd716e679" (UID: "3961388f-9e20-4a53-97fc-36bdd716e679"). InnerVolumeSpecName "kube-api-access-r6vmd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 19:22:41 crc kubenswrapper[4988]: I1008 19:22:41.307107 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3961388f-9e20-4a53-97fc-36bdd716e679-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "3961388f-9e20-4a53-97fc-36bdd716e679" (UID: "3961388f-9e20-4a53-97fc-36bdd716e679"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 19:22:41 crc kubenswrapper[4988]: I1008 19:22:41.397293 4988 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3961388f-9e20-4a53-97fc-36bdd716e679-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 08 19:22:41 crc kubenswrapper[4988]: I1008 19:22:41.397355 4988 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3961388f-9e20-4a53-97fc-36bdd716e679-utilities\") on node \"crc\" DevicePath \"\"" Oct 08 19:22:41 crc kubenswrapper[4988]: I1008 19:22:41.397415 4988 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r6vmd\" (UniqueName: \"kubernetes.io/projected/3961388f-9e20-4a53-97fc-36bdd716e679-kube-api-access-r6vmd\") on node \"crc\" DevicePath \"\"" Oct 08 19:22:41 crc kubenswrapper[4988]: I1008 19:22:41.668881 4988 generic.go:334] "Generic (PLEG): container finished" podID="3961388f-9e20-4a53-97fc-36bdd716e679" containerID="f6fe63cab862dc5d27b31154801d6ad92861ec92fe4017c137d8c8cde4ab7061" exitCode=0 Oct 08 19:22:41 crc kubenswrapper[4988]: I1008 19:22:41.668937 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-cbs84" Oct 08 19:22:41 crc kubenswrapper[4988]: I1008 19:22:41.668961 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-cbs84" event={"ID":"3961388f-9e20-4a53-97fc-36bdd716e679","Type":"ContainerDied","Data":"f6fe63cab862dc5d27b31154801d6ad92861ec92fe4017c137d8c8cde4ab7061"} Oct 08 19:22:41 crc kubenswrapper[4988]: I1008 19:22:41.669021 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-cbs84" event={"ID":"3961388f-9e20-4a53-97fc-36bdd716e679","Type":"ContainerDied","Data":"e7ce49a5c1793ffa74d8163ba255470fa3b56511ba6b2488a1e92d4694bbb6ce"} Oct 08 19:22:41 crc kubenswrapper[4988]: I1008 19:22:41.669062 4988 scope.go:117] "RemoveContainer" containerID="f6fe63cab862dc5d27b31154801d6ad92861ec92fe4017c137d8c8cde4ab7061" Oct 08 19:22:41 crc kubenswrapper[4988]: I1008 19:22:41.711689 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-cbs84"] Oct 08 19:22:41 crc kubenswrapper[4988]: I1008 19:22:41.714119 4988 scope.go:117] "RemoveContainer" containerID="8684b7542c8479bf77c19289214625e4ca323fa04e81a4d595f9ea6b83e57a5e" Oct 08 19:22:41 crc kubenswrapper[4988]: I1008 19:22:41.720848 4988 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-cbs84"] Oct 08 19:22:41 crc kubenswrapper[4988]: I1008 19:22:41.734051 4988 scope.go:117] "RemoveContainer" containerID="d6b21b601d16472dce4c53aeffceaee8eddce3d845ccb7ea9714192f083bb85b" Oct 08 19:22:41 crc kubenswrapper[4988]: I1008 19:22:41.764239 4988 scope.go:117] "RemoveContainer" containerID="f6fe63cab862dc5d27b31154801d6ad92861ec92fe4017c137d8c8cde4ab7061" Oct 08 19:22:41 crc kubenswrapper[4988]: E1008 19:22:41.764829 4988 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f6fe63cab862dc5d27b31154801d6ad92861ec92fe4017c137d8c8cde4ab7061\": container with ID starting with f6fe63cab862dc5d27b31154801d6ad92861ec92fe4017c137d8c8cde4ab7061 not found: ID does not exist" containerID="f6fe63cab862dc5d27b31154801d6ad92861ec92fe4017c137d8c8cde4ab7061" Oct 08 19:22:41 crc kubenswrapper[4988]: I1008 19:22:41.764946 4988 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f6fe63cab862dc5d27b31154801d6ad92861ec92fe4017c137d8c8cde4ab7061"} err="failed to get container status \"f6fe63cab862dc5d27b31154801d6ad92861ec92fe4017c137d8c8cde4ab7061\": rpc error: code = NotFound desc = could not find container \"f6fe63cab862dc5d27b31154801d6ad92861ec92fe4017c137d8c8cde4ab7061\": container with ID starting with f6fe63cab862dc5d27b31154801d6ad92861ec92fe4017c137d8c8cde4ab7061 not found: ID does not exist" Oct 08 19:22:41 crc kubenswrapper[4988]: I1008 19:22:41.765047 4988 scope.go:117] "RemoveContainer" containerID="8684b7542c8479bf77c19289214625e4ca323fa04e81a4d595f9ea6b83e57a5e" Oct 08 19:22:41 crc kubenswrapper[4988]: E1008 19:22:41.765691 4988 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8684b7542c8479bf77c19289214625e4ca323fa04e81a4d595f9ea6b83e57a5e\": container with ID starting with 8684b7542c8479bf77c19289214625e4ca323fa04e81a4d595f9ea6b83e57a5e not found: ID does not exist" containerID="8684b7542c8479bf77c19289214625e4ca323fa04e81a4d595f9ea6b83e57a5e" Oct 08 19:22:41 crc kubenswrapper[4988]: I1008 19:22:41.765740 4988 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8684b7542c8479bf77c19289214625e4ca323fa04e81a4d595f9ea6b83e57a5e"} err="failed to get container status \"8684b7542c8479bf77c19289214625e4ca323fa04e81a4d595f9ea6b83e57a5e\": rpc error: code = NotFound desc = could not find container \"8684b7542c8479bf77c19289214625e4ca323fa04e81a4d595f9ea6b83e57a5e\": container with ID starting with 8684b7542c8479bf77c19289214625e4ca323fa04e81a4d595f9ea6b83e57a5e not found: ID does not exist" Oct 08 19:22:41 crc kubenswrapper[4988]: I1008 19:22:41.765778 4988 scope.go:117] "RemoveContainer" containerID="d6b21b601d16472dce4c53aeffceaee8eddce3d845ccb7ea9714192f083bb85b" Oct 08 19:22:41 crc kubenswrapper[4988]: E1008 19:22:41.766418 4988 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d6b21b601d16472dce4c53aeffceaee8eddce3d845ccb7ea9714192f083bb85b\": container with ID starting with d6b21b601d16472dce4c53aeffceaee8eddce3d845ccb7ea9714192f083bb85b not found: ID does not exist" containerID="d6b21b601d16472dce4c53aeffceaee8eddce3d845ccb7ea9714192f083bb85b" Oct 08 19:22:41 crc kubenswrapper[4988]: I1008 19:22:41.766528 4988 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d6b21b601d16472dce4c53aeffceaee8eddce3d845ccb7ea9714192f083bb85b"} err="failed to get container status \"d6b21b601d16472dce4c53aeffceaee8eddce3d845ccb7ea9714192f083bb85b\": rpc error: code = NotFound desc = could not find container \"d6b21b601d16472dce4c53aeffceaee8eddce3d845ccb7ea9714192f083bb85b\": container with ID starting with d6b21b601d16472dce4c53aeffceaee8eddce3d845ccb7ea9714192f083bb85b not found: ID does not exist" Oct 08 19:22:43 crc kubenswrapper[4988]: I1008 19:22:43.255657 4988 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3961388f-9e20-4a53-97fc-36bdd716e679" path="/var/lib/kubelet/pods/3961388f-9e20-4a53-97fc-36bdd716e679/volumes" Oct 08 19:22:53 crc kubenswrapper[4988]: I1008 19:22:53.338712 4988 patch_prober.go:28] interesting pod/machine-config-daemon-nm54f container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 19:22:53 crc kubenswrapper[4988]: I1008 19:22:53.339271 4988 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 19:23:23 crc kubenswrapper[4988]: I1008 19:23:23.337803 4988 patch_prober.go:28] interesting pod/machine-config-daemon-nm54f container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 19:23:23 crc kubenswrapper[4988]: I1008 19:23:23.338546 4988 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 19:23:23 crc kubenswrapper[4988]: I1008 19:23:23.338626 4988 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" Oct 08 19:23:23 crc kubenswrapper[4988]: I1008 19:23:23.339472 4988 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"20477a4d0d4c3448d84d69456b20adede3fe2d7c9fa57e53eb98b6ad7aabc27d"} pod="openshift-machine-config-operator/machine-config-daemon-nm54f" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 08 19:23:23 crc kubenswrapper[4988]: I1008 19:23:23.339578 4988 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" containerName="machine-config-daemon" containerID="cri-o://20477a4d0d4c3448d84d69456b20adede3fe2d7c9fa57e53eb98b6ad7aabc27d" gracePeriod=600 Oct 08 19:23:23 crc kubenswrapper[4988]: E1008 19:23:23.467217 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nm54f_openshift-machine-config-operator(f328a20b-5283-407e-8a93-2ab2da599f18)\"" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" Oct 08 19:23:24 crc kubenswrapper[4988]: I1008 19:23:24.048436 4988 generic.go:334] "Generic (PLEG): container finished" podID="f328a20b-5283-407e-8a93-2ab2da599f18" containerID="20477a4d0d4c3448d84d69456b20adede3fe2d7c9fa57e53eb98b6ad7aabc27d" exitCode=0 Oct 08 19:23:24 crc kubenswrapper[4988]: I1008 19:23:24.048672 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" event={"ID":"f328a20b-5283-407e-8a93-2ab2da599f18","Type":"ContainerDied","Data":"20477a4d0d4c3448d84d69456b20adede3fe2d7c9fa57e53eb98b6ad7aabc27d"} Oct 08 19:23:24 crc kubenswrapper[4988]: I1008 19:23:24.048891 4988 scope.go:117] "RemoveContainer" containerID="eca8e1f95f09ab5d917be9719293edda80592194583b0c97c98c1d0e281973e8" Oct 08 19:23:24 crc kubenswrapper[4988]: I1008 19:23:24.049398 4988 scope.go:117] "RemoveContainer" containerID="20477a4d0d4c3448d84d69456b20adede3fe2d7c9fa57e53eb98b6ad7aabc27d" Oct 08 19:23:24 crc kubenswrapper[4988]: E1008 19:23:24.049679 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nm54f_openshift-machine-config-operator(f328a20b-5283-407e-8a93-2ab2da599f18)\"" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" Oct 08 19:23:36 crc kubenswrapper[4988]: I1008 19:23:36.237689 4988 scope.go:117] "RemoveContainer" containerID="20477a4d0d4c3448d84d69456b20adede3fe2d7c9fa57e53eb98b6ad7aabc27d" Oct 08 19:23:36 crc kubenswrapper[4988]: E1008 19:23:36.238430 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nm54f_openshift-machine-config-operator(f328a20b-5283-407e-8a93-2ab2da599f18)\"" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" Oct 08 19:23:49 crc kubenswrapper[4988]: I1008 19:23:49.237815 4988 scope.go:117] "RemoveContainer" containerID="20477a4d0d4c3448d84d69456b20adede3fe2d7c9fa57e53eb98b6ad7aabc27d" Oct 08 19:23:49 crc kubenswrapper[4988]: E1008 19:23:49.239652 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nm54f_openshift-machine-config-operator(f328a20b-5283-407e-8a93-2ab2da599f18)\"" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" Oct 08 19:24:00 crc kubenswrapper[4988]: I1008 19:24:00.237592 4988 scope.go:117] "RemoveContainer" containerID="20477a4d0d4c3448d84d69456b20adede3fe2d7c9fa57e53eb98b6ad7aabc27d" Oct 08 19:24:00 crc kubenswrapper[4988]: E1008 19:24:00.238343 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nm54f_openshift-machine-config-operator(f328a20b-5283-407e-8a93-2ab2da599f18)\"" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" Oct 08 19:24:14 crc kubenswrapper[4988]: I1008 19:24:14.238036 4988 scope.go:117] "RemoveContainer" containerID="20477a4d0d4c3448d84d69456b20adede3fe2d7c9fa57e53eb98b6ad7aabc27d" Oct 08 19:24:14 crc kubenswrapper[4988]: E1008 19:24:14.238984 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nm54f_openshift-machine-config-operator(f328a20b-5283-407e-8a93-2ab2da599f18)\"" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" Oct 08 19:24:25 crc kubenswrapper[4988]: I1008 19:24:25.242569 4988 scope.go:117] "RemoveContainer" containerID="20477a4d0d4c3448d84d69456b20adede3fe2d7c9fa57e53eb98b6ad7aabc27d" Oct 08 19:24:25 crc kubenswrapper[4988]: E1008 19:24:25.243266 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nm54f_openshift-machine-config-operator(f328a20b-5283-407e-8a93-2ab2da599f18)\"" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" Oct 08 19:24:37 crc kubenswrapper[4988]: I1008 19:24:37.237691 4988 scope.go:117] "RemoveContainer" containerID="20477a4d0d4c3448d84d69456b20adede3fe2d7c9fa57e53eb98b6ad7aabc27d" Oct 08 19:24:37 crc kubenswrapper[4988]: E1008 19:24:37.242050 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nm54f_openshift-machine-config-operator(f328a20b-5283-407e-8a93-2ab2da599f18)\"" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" Oct 08 19:24:50 crc kubenswrapper[4988]: I1008 19:24:50.238213 4988 scope.go:117] "RemoveContainer" containerID="20477a4d0d4c3448d84d69456b20adede3fe2d7c9fa57e53eb98b6ad7aabc27d" Oct 08 19:24:50 crc kubenswrapper[4988]: E1008 19:24:50.238988 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nm54f_openshift-machine-config-operator(f328a20b-5283-407e-8a93-2ab2da599f18)\"" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" Oct 08 19:25:05 crc kubenswrapper[4988]: I1008 19:25:05.244484 4988 scope.go:117] "RemoveContainer" containerID="20477a4d0d4c3448d84d69456b20adede3fe2d7c9fa57e53eb98b6ad7aabc27d" Oct 08 19:25:05 crc kubenswrapper[4988]: E1008 19:25:05.245597 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nm54f_openshift-machine-config-operator(f328a20b-5283-407e-8a93-2ab2da599f18)\"" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" Oct 08 19:25:19 crc kubenswrapper[4988]: I1008 19:25:19.238553 4988 scope.go:117] "RemoveContainer" containerID="20477a4d0d4c3448d84d69456b20adede3fe2d7c9fa57e53eb98b6ad7aabc27d" Oct 08 19:25:19 crc kubenswrapper[4988]: E1008 19:25:19.239701 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nm54f_openshift-machine-config-operator(f328a20b-5283-407e-8a93-2ab2da599f18)\"" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" Oct 08 19:25:31 crc kubenswrapper[4988]: I1008 19:25:31.237715 4988 scope.go:117] "RemoveContainer" containerID="20477a4d0d4c3448d84d69456b20adede3fe2d7c9fa57e53eb98b6ad7aabc27d" Oct 08 19:25:31 crc kubenswrapper[4988]: E1008 19:25:31.239048 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nm54f_openshift-machine-config-operator(f328a20b-5283-407e-8a93-2ab2da599f18)\"" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" Oct 08 19:25:42 crc kubenswrapper[4988]: I1008 19:25:42.238818 4988 scope.go:117] "RemoveContainer" containerID="20477a4d0d4c3448d84d69456b20adede3fe2d7c9fa57e53eb98b6ad7aabc27d" Oct 08 19:25:42 crc kubenswrapper[4988]: E1008 19:25:42.240055 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nm54f_openshift-machine-config-operator(f328a20b-5283-407e-8a93-2ab2da599f18)\"" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" Oct 08 19:25:57 crc kubenswrapper[4988]: I1008 19:25:57.238431 4988 scope.go:117] "RemoveContainer" containerID="20477a4d0d4c3448d84d69456b20adede3fe2d7c9fa57e53eb98b6ad7aabc27d" Oct 08 19:25:57 crc kubenswrapper[4988]: E1008 19:25:57.239398 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nm54f_openshift-machine-config-operator(f328a20b-5283-407e-8a93-2ab2da599f18)\"" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" Oct 08 19:26:10 crc kubenswrapper[4988]: I1008 19:26:10.238338 4988 scope.go:117] "RemoveContainer" containerID="20477a4d0d4c3448d84d69456b20adede3fe2d7c9fa57e53eb98b6ad7aabc27d" Oct 08 19:26:10 crc kubenswrapper[4988]: E1008 19:26:10.239447 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nm54f_openshift-machine-config-operator(f328a20b-5283-407e-8a93-2ab2da599f18)\"" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" Oct 08 19:26:24 crc kubenswrapper[4988]: I1008 19:26:24.238103 4988 scope.go:117] "RemoveContainer" containerID="20477a4d0d4c3448d84d69456b20adede3fe2d7c9fa57e53eb98b6ad7aabc27d" Oct 08 19:26:24 crc kubenswrapper[4988]: E1008 19:26:24.240470 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nm54f_openshift-machine-config-operator(f328a20b-5283-407e-8a93-2ab2da599f18)\"" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" Oct 08 19:26:35 crc kubenswrapper[4988]: I1008 19:26:35.245833 4988 scope.go:117] "RemoveContainer" containerID="20477a4d0d4c3448d84d69456b20adede3fe2d7c9fa57e53eb98b6ad7aabc27d" Oct 08 19:26:35 crc kubenswrapper[4988]: E1008 19:26:35.246899 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nm54f_openshift-machine-config-operator(f328a20b-5283-407e-8a93-2ab2da599f18)\"" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" Oct 08 19:26:47 crc kubenswrapper[4988]: I1008 19:26:47.237955 4988 scope.go:117] "RemoveContainer" containerID="20477a4d0d4c3448d84d69456b20adede3fe2d7c9fa57e53eb98b6ad7aabc27d" Oct 08 19:26:47 crc kubenswrapper[4988]: E1008 19:26:47.239258 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nm54f_openshift-machine-config-operator(f328a20b-5283-407e-8a93-2ab2da599f18)\"" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" Oct 08 19:27:02 crc kubenswrapper[4988]: I1008 19:27:02.237804 4988 scope.go:117] "RemoveContainer" containerID="20477a4d0d4c3448d84d69456b20adede3fe2d7c9fa57e53eb98b6ad7aabc27d" Oct 08 19:27:02 crc kubenswrapper[4988]: E1008 19:27:02.239083 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nm54f_openshift-machine-config-operator(f328a20b-5283-407e-8a93-2ab2da599f18)\"" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" Oct 08 19:27:14 crc kubenswrapper[4988]: I1008 19:27:14.237908 4988 scope.go:117] "RemoveContainer" containerID="20477a4d0d4c3448d84d69456b20adede3fe2d7c9fa57e53eb98b6ad7aabc27d" Oct 08 19:27:14 crc kubenswrapper[4988]: E1008 19:27:14.238694 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nm54f_openshift-machine-config-operator(f328a20b-5283-407e-8a93-2ab2da599f18)\"" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" Oct 08 19:27:21 crc kubenswrapper[4988]: I1008 19:27:21.320617 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-4jzcq"] Oct 08 19:27:21 crc kubenswrapper[4988]: E1008 19:27:21.321721 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3961388f-9e20-4a53-97fc-36bdd716e679" containerName="extract-content" Oct 08 19:27:21 crc kubenswrapper[4988]: I1008 19:27:21.321744 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="3961388f-9e20-4a53-97fc-36bdd716e679" containerName="extract-content" Oct 08 19:27:21 crc kubenswrapper[4988]: E1008 19:27:21.321778 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3961388f-9e20-4a53-97fc-36bdd716e679" containerName="registry-server" Oct 08 19:27:21 crc kubenswrapper[4988]: I1008 19:27:21.321787 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="3961388f-9e20-4a53-97fc-36bdd716e679" containerName="registry-server" Oct 08 19:27:21 crc kubenswrapper[4988]: E1008 19:27:21.321807 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3961388f-9e20-4a53-97fc-36bdd716e679" containerName="extract-utilities" Oct 08 19:27:21 crc kubenswrapper[4988]: I1008 19:27:21.321816 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="3961388f-9e20-4a53-97fc-36bdd716e679" containerName="extract-utilities" Oct 08 19:27:21 crc kubenswrapper[4988]: I1008 19:27:21.322001 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="3961388f-9e20-4a53-97fc-36bdd716e679" containerName="registry-server" Oct 08 19:27:21 crc kubenswrapper[4988]: I1008 19:27:21.323373 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-4jzcq" Oct 08 19:27:21 crc kubenswrapper[4988]: I1008 19:27:21.333278 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-4jzcq"] Oct 08 19:27:21 crc kubenswrapper[4988]: I1008 19:27:21.423327 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0e1eef88-0680-4fba-9b85-765d6fbe8b99-catalog-content\") pod \"community-operators-4jzcq\" (UID: \"0e1eef88-0680-4fba-9b85-765d6fbe8b99\") " pod="openshift-marketplace/community-operators-4jzcq" Oct 08 19:27:21 crc kubenswrapper[4988]: I1008 19:27:21.423445 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lgnq2\" (UniqueName: \"kubernetes.io/projected/0e1eef88-0680-4fba-9b85-765d6fbe8b99-kube-api-access-lgnq2\") pod \"community-operators-4jzcq\" (UID: \"0e1eef88-0680-4fba-9b85-765d6fbe8b99\") " pod="openshift-marketplace/community-operators-4jzcq" Oct 08 19:27:21 crc kubenswrapper[4988]: I1008 19:27:21.423501 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0e1eef88-0680-4fba-9b85-765d6fbe8b99-utilities\") pod \"community-operators-4jzcq\" (UID: \"0e1eef88-0680-4fba-9b85-765d6fbe8b99\") " pod="openshift-marketplace/community-operators-4jzcq" Oct 08 19:27:21 crc kubenswrapper[4988]: I1008 19:27:21.525090 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0e1eef88-0680-4fba-9b85-765d6fbe8b99-catalog-content\") pod \"community-operators-4jzcq\" (UID: \"0e1eef88-0680-4fba-9b85-765d6fbe8b99\") " pod="openshift-marketplace/community-operators-4jzcq" Oct 08 19:27:21 crc kubenswrapper[4988]: I1008 19:27:21.525532 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lgnq2\" (UniqueName: \"kubernetes.io/projected/0e1eef88-0680-4fba-9b85-765d6fbe8b99-kube-api-access-lgnq2\") pod \"community-operators-4jzcq\" (UID: \"0e1eef88-0680-4fba-9b85-765d6fbe8b99\") " pod="openshift-marketplace/community-operators-4jzcq" Oct 08 19:27:21 crc kubenswrapper[4988]: I1008 19:27:21.525721 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0e1eef88-0680-4fba-9b85-765d6fbe8b99-utilities\") pod \"community-operators-4jzcq\" (UID: \"0e1eef88-0680-4fba-9b85-765d6fbe8b99\") " pod="openshift-marketplace/community-operators-4jzcq" Oct 08 19:27:21 crc kubenswrapper[4988]: I1008 19:27:21.526207 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0e1eef88-0680-4fba-9b85-765d6fbe8b99-catalog-content\") pod \"community-operators-4jzcq\" (UID: \"0e1eef88-0680-4fba-9b85-765d6fbe8b99\") " pod="openshift-marketplace/community-operators-4jzcq" Oct 08 19:27:21 crc kubenswrapper[4988]: I1008 19:27:21.526215 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0e1eef88-0680-4fba-9b85-765d6fbe8b99-utilities\") pod \"community-operators-4jzcq\" (UID: \"0e1eef88-0680-4fba-9b85-765d6fbe8b99\") " pod="openshift-marketplace/community-operators-4jzcq" Oct 08 19:27:21 crc kubenswrapper[4988]: I1008 19:27:21.552515 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lgnq2\" (UniqueName: \"kubernetes.io/projected/0e1eef88-0680-4fba-9b85-765d6fbe8b99-kube-api-access-lgnq2\") pod \"community-operators-4jzcq\" (UID: \"0e1eef88-0680-4fba-9b85-765d6fbe8b99\") " pod="openshift-marketplace/community-operators-4jzcq" Oct 08 19:27:21 crc kubenswrapper[4988]: I1008 19:27:21.650151 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-4jzcq" Oct 08 19:27:22 crc kubenswrapper[4988]: I1008 19:27:22.169790 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-4jzcq"] Oct 08 19:27:22 crc kubenswrapper[4988]: I1008 19:27:22.240826 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-4jzcq" event={"ID":"0e1eef88-0680-4fba-9b85-765d6fbe8b99","Type":"ContainerStarted","Data":"6719070c0a61f0464bfb082d669793c3c6732dba77655379e4256456bc96c0bd"} Oct 08 19:27:23 crc kubenswrapper[4988]: I1008 19:27:23.253969 4988 generic.go:334] "Generic (PLEG): container finished" podID="0e1eef88-0680-4fba-9b85-765d6fbe8b99" containerID="e9be172ed6bbf4f9017c2c60e467d15e66a9dd7913bd23a8d4f0b71560896818" exitCode=0 Oct 08 19:27:23 crc kubenswrapper[4988]: I1008 19:27:23.254067 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-4jzcq" event={"ID":"0e1eef88-0680-4fba-9b85-765d6fbe8b99","Type":"ContainerDied","Data":"e9be172ed6bbf4f9017c2c60e467d15e66a9dd7913bd23a8d4f0b71560896818"} Oct 08 19:27:24 crc kubenswrapper[4988]: I1008 19:27:24.262841 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-4jzcq" event={"ID":"0e1eef88-0680-4fba-9b85-765d6fbe8b99","Type":"ContainerStarted","Data":"3b8ae4b758f2e13aeb149d47edfc1fe01c3046e68a3791f01694697eec9c9c54"} Oct 08 19:27:25 crc kubenswrapper[4988]: I1008 19:27:25.273594 4988 generic.go:334] "Generic (PLEG): container finished" podID="0e1eef88-0680-4fba-9b85-765d6fbe8b99" containerID="3b8ae4b758f2e13aeb149d47edfc1fe01c3046e68a3791f01694697eec9c9c54" exitCode=0 Oct 08 19:27:25 crc kubenswrapper[4988]: I1008 19:27:25.273646 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-4jzcq" event={"ID":"0e1eef88-0680-4fba-9b85-765d6fbe8b99","Type":"ContainerDied","Data":"3b8ae4b758f2e13aeb149d47edfc1fe01c3046e68a3791f01694697eec9c9c54"} Oct 08 19:27:26 crc kubenswrapper[4988]: I1008 19:27:26.285671 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-4jzcq" event={"ID":"0e1eef88-0680-4fba-9b85-765d6fbe8b99","Type":"ContainerStarted","Data":"f2a69ac8624ffd8a6bbca07a7ae35b98c4e177bd44a7c0daef9a4d7a3d3a764e"} Oct 08 19:27:26 crc kubenswrapper[4988]: I1008 19:27:26.312846 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-4jzcq" podStartSLOduration=2.8499756339999998 podStartE2EDuration="5.31282304s" podCreationTimestamp="2025-10-08 19:27:21 +0000 UTC" firstStartedPulling="2025-10-08 19:27:23.259183698 +0000 UTC m=+4588.709026468" lastFinishedPulling="2025-10-08 19:27:25.722031114 +0000 UTC m=+4591.171873874" observedRunningTime="2025-10-08 19:27:26.309142943 +0000 UTC m=+4591.758985713" watchObservedRunningTime="2025-10-08 19:27:26.31282304 +0000 UTC m=+4591.762665840" Oct 08 19:27:28 crc kubenswrapper[4988]: I1008 19:27:28.238451 4988 scope.go:117] "RemoveContainer" containerID="20477a4d0d4c3448d84d69456b20adede3fe2d7c9fa57e53eb98b6ad7aabc27d" Oct 08 19:27:28 crc kubenswrapper[4988]: E1008 19:27:28.239050 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nm54f_openshift-machine-config-operator(f328a20b-5283-407e-8a93-2ab2da599f18)\"" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" Oct 08 19:27:31 crc kubenswrapper[4988]: I1008 19:27:31.651545 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-4jzcq" Oct 08 19:27:31 crc kubenswrapper[4988]: I1008 19:27:31.651615 4988 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-4jzcq" Oct 08 19:27:31 crc kubenswrapper[4988]: I1008 19:27:31.710016 4988 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-4jzcq" Oct 08 19:27:32 crc kubenswrapper[4988]: I1008 19:27:32.399565 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-4jzcq" Oct 08 19:27:32 crc kubenswrapper[4988]: I1008 19:27:32.454049 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-4jzcq"] Oct 08 19:27:34 crc kubenswrapper[4988]: I1008 19:27:34.356630 4988 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-4jzcq" podUID="0e1eef88-0680-4fba-9b85-765d6fbe8b99" containerName="registry-server" containerID="cri-o://f2a69ac8624ffd8a6bbca07a7ae35b98c4e177bd44a7c0daef9a4d7a3d3a764e" gracePeriod=2 Oct 08 19:27:34 crc kubenswrapper[4988]: I1008 19:27:34.832525 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-4jzcq" Oct 08 19:27:34 crc kubenswrapper[4988]: I1008 19:27:34.926358 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0e1eef88-0680-4fba-9b85-765d6fbe8b99-catalog-content\") pod \"0e1eef88-0680-4fba-9b85-765d6fbe8b99\" (UID: \"0e1eef88-0680-4fba-9b85-765d6fbe8b99\") " Oct 08 19:27:34 crc kubenswrapper[4988]: I1008 19:27:34.926700 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0e1eef88-0680-4fba-9b85-765d6fbe8b99-utilities\") pod \"0e1eef88-0680-4fba-9b85-765d6fbe8b99\" (UID: \"0e1eef88-0680-4fba-9b85-765d6fbe8b99\") " Oct 08 19:27:34 crc kubenswrapper[4988]: I1008 19:27:34.926824 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lgnq2\" (UniqueName: \"kubernetes.io/projected/0e1eef88-0680-4fba-9b85-765d6fbe8b99-kube-api-access-lgnq2\") pod \"0e1eef88-0680-4fba-9b85-765d6fbe8b99\" (UID: \"0e1eef88-0680-4fba-9b85-765d6fbe8b99\") " Oct 08 19:27:34 crc kubenswrapper[4988]: I1008 19:27:34.928146 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0e1eef88-0680-4fba-9b85-765d6fbe8b99-utilities" (OuterVolumeSpecName: "utilities") pod "0e1eef88-0680-4fba-9b85-765d6fbe8b99" (UID: "0e1eef88-0680-4fba-9b85-765d6fbe8b99"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 19:27:34 crc kubenswrapper[4988]: I1008 19:27:34.939499 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0e1eef88-0680-4fba-9b85-765d6fbe8b99-kube-api-access-lgnq2" (OuterVolumeSpecName: "kube-api-access-lgnq2") pod "0e1eef88-0680-4fba-9b85-765d6fbe8b99" (UID: "0e1eef88-0680-4fba-9b85-765d6fbe8b99"). InnerVolumeSpecName "kube-api-access-lgnq2". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 19:27:35 crc kubenswrapper[4988]: I1008 19:27:35.028293 4988 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lgnq2\" (UniqueName: \"kubernetes.io/projected/0e1eef88-0680-4fba-9b85-765d6fbe8b99-kube-api-access-lgnq2\") on node \"crc\" DevicePath \"\"" Oct 08 19:27:35 crc kubenswrapper[4988]: I1008 19:27:35.028324 4988 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0e1eef88-0680-4fba-9b85-765d6fbe8b99-utilities\") on node \"crc\" DevicePath \"\"" Oct 08 19:27:35 crc kubenswrapper[4988]: I1008 19:27:35.186023 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0e1eef88-0680-4fba-9b85-765d6fbe8b99-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "0e1eef88-0680-4fba-9b85-765d6fbe8b99" (UID: "0e1eef88-0680-4fba-9b85-765d6fbe8b99"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 19:27:35 crc kubenswrapper[4988]: I1008 19:27:35.231296 4988 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0e1eef88-0680-4fba-9b85-765d6fbe8b99-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 08 19:27:35 crc kubenswrapper[4988]: I1008 19:27:35.367926 4988 generic.go:334] "Generic (PLEG): container finished" podID="0e1eef88-0680-4fba-9b85-765d6fbe8b99" containerID="f2a69ac8624ffd8a6bbca07a7ae35b98c4e177bd44a7c0daef9a4d7a3d3a764e" exitCode=0 Oct 08 19:27:35 crc kubenswrapper[4988]: I1008 19:27:35.367996 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-4jzcq" event={"ID":"0e1eef88-0680-4fba-9b85-765d6fbe8b99","Type":"ContainerDied","Data":"f2a69ac8624ffd8a6bbca07a7ae35b98c4e177bd44a7c0daef9a4d7a3d3a764e"} Oct 08 19:27:35 crc kubenswrapper[4988]: I1008 19:27:35.368072 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-4jzcq" event={"ID":"0e1eef88-0680-4fba-9b85-765d6fbe8b99","Type":"ContainerDied","Data":"6719070c0a61f0464bfb082d669793c3c6732dba77655379e4256456bc96c0bd"} Oct 08 19:27:35 crc kubenswrapper[4988]: I1008 19:27:35.368078 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-4jzcq" Oct 08 19:27:35 crc kubenswrapper[4988]: I1008 19:27:35.368103 4988 scope.go:117] "RemoveContainer" containerID="f2a69ac8624ffd8a6bbca07a7ae35b98c4e177bd44a7c0daef9a4d7a3d3a764e" Oct 08 19:27:35 crc kubenswrapper[4988]: I1008 19:27:35.401377 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-4jzcq"] Oct 08 19:27:35 crc kubenswrapper[4988]: I1008 19:27:35.407313 4988 scope.go:117] "RemoveContainer" containerID="3b8ae4b758f2e13aeb149d47edfc1fe01c3046e68a3791f01694697eec9c9c54" Oct 08 19:27:35 crc kubenswrapper[4988]: I1008 19:27:35.412264 4988 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-4jzcq"] Oct 08 19:27:35 crc kubenswrapper[4988]: I1008 19:27:35.429580 4988 scope.go:117] "RemoveContainer" containerID="e9be172ed6bbf4f9017c2c60e467d15e66a9dd7913bd23a8d4f0b71560896818" Oct 08 19:27:35 crc kubenswrapper[4988]: I1008 19:27:35.461644 4988 scope.go:117] "RemoveContainer" containerID="f2a69ac8624ffd8a6bbca07a7ae35b98c4e177bd44a7c0daef9a4d7a3d3a764e" Oct 08 19:27:35 crc kubenswrapper[4988]: E1008 19:27:35.462043 4988 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f2a69ac8624ffd8a6bbca07a7ae35b98c4e177bd44a7c0daef9a4d7a3d3a764e\": container with ID starting with f2a69ac8624ffd8a6bbca07a7ae35b98c4e177bd44a7c0daef9a4d7a3d3a764e not found: ID does not exist" containerID="f2a69ac8624ffd8a6bbca07a7ae35b98c4e177bd44a7c0daef9a4d7a3d3a764e" Oct 08 19:27:35 crc kubenswrapper[4988]: I1008 19:27:35.462077 4988 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f2a69ac8624ffd8a6bbca07a7ae35b98c4e177bd44a7c0daef9a4d7a3d3a764e"} err="failed to get container status \"f2a69ac8624ffd8a6bbca07a7ae35b98c4e177bd44a7c0daef9a4d7a3d3a764e\": rpc error: code = NotFound desc = could not find container \"f2a69ac8624ffd8a6bbca07a7ae35b98c4e177bd44a7c0daef9a4d7a3d3a764e\": container with ID starting with f2a69ac8624ffd8a6bbca07a7ae35b98c4e177bd44a7c0daef9a4d7a3d3a764e not found: ID does not exist" Oct 08 19:27:35 crc kubenswrapper[4988]: I1008 19:27:35.462103 4988 scope.go:117] "RemoveContainer" containerID="3b8ae4b758f2e13aeb149d47edfc1fe01c3046e68a3791f01694697eec9c9c54" Oct 08 19:27:35 crc kubenswrapper[4988]: E1008 19:27:35.462537 4988 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3b8ae4b758f2e13aeb149d47edfc1fe01c3046e68a3791f01694697eec9c9c54\": container with ID starting with 3b8ae4b758f2e13aeb149d47edfc1fe01c3046e68a3791f01694697eec9c9c54 not found: ID does not exist" containerID="3b8ae4b758f2e13aeb149d47edfc1fe01c3046e68a3791f01694697eec9c9c54" Oct 08 19:27:35 crc kubenswrapper[4988]: I1008 19:27:35.462571 4988 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3b8ae4b758f2e13aeb149d47edfc1fe01c3046e68a3791f01694697eec9c9c54"} err="failed to get container status \"3b8ae4b758f2e13aeb149d47edfc1fe01c3046e68a3791f01694697eec9c9c54\": rpc error: code = NotFound desc = could not find container \"3b8ae4b758f2e13aeb149d47edfc1fe01c3046e68a3791f01694697eec9c9c54\": container with ID starting with 3b8ae4b758f2e13aeb149d47edfc1fe01c3046e68a3791f01694697eec9c9c54 not found: ID does not exist" Oct 08 19:27:35 crc kubenswrapper[4988]: I1008 19:27:35.462588 4988 scope.go:117] "RemoveContainer" containerID="e9be172ed6bbf4f9017c2c60e467d15e66a9dd7913bd23a8d4f0b71560896818" Oct 08 19:27:35 crc kubenswrapper[4988]: E1008 19:27:35.462872 4988 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e9be172ed6bbf4f9017c2c60e467d15e66a9dd7913bd23a8d4f0b71560896818\": container with ID starting with e9be172ed6bbf4f9017c2c60e467d15e66a9dd7913bd23a8d4f0b71560896818 not found: ID does not exist" containerID="e9be172ed6bbf4f9017c2c60e467d15e66a9dd7913bd23a8d4f0b71560896818" Oct 08 19:27:35 crc kubenswrapper[4988]: I1008 19:27:35.462967 4988 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e9be172ed6bbf4f9017c2c60e467d15e66a9dd7913bd23a8d4f0b71560896818"} err="failed to get container status \"e9be172ed6bbf4f9017c2c60e467d15e66a9dd7913bd23a8d4f0b71560896818\": rpc error: code = NotFound desc = could not find container \"e9be172ed6bbf4f9017c2c60e467d15e66a9dd7913bd23a8d4f0b71560896818\": container with ID starting with e9be172ed6bbf4f9017c2c60e467d15e66a9dd7913bd23a8d4f0b71560896818 not found: ID does not exist" Oct 08 19:27:37 crc kubenswrapper[4988]: I1008 19:27:37.253090 4988 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0e1eef88-0680-4fba-9b85-765d6fbe8b99" path="/var/lib/kubelet/pods/0e1eef88-0680-4fba-9b85-765d6fbe8b99/volumes" Oct 08 19:27:43 crc kubenswrapper[4988]: I1008 19:27:43.238551 4988 scope.go:117] "RemoveContainer" containerID="20477a4d0d4c3448d84d69456b20adede3fe2d7c9fa57e53eb98b6ad7aabc27d" Oct 08 19:27:43 crc kubenswrapper[4988]: E1008 19:27:43.239446 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nm54f_openshift-machine-config-operator(f328a20b-5283-407e-8a93-2ab2da599f18)\"" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" Oct 08 19:27:57 crc kubenswrapper[4988]: I1008 19:27:57.238127 4988 scope.go:117] "RemoveContainer" containerID="20477a4d0d4c3448d84d69456b20adede3fe2d7c9fa57e53eb98b6ad7aabc27d" Oct 08 19:27:57 crc kubenswrapper[4988]: E1008 19:27:57.238704 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nm54f_openshift-machine-config-operator(f328a20b-5283-407e-8a93-2ab2da599f18)\"" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" Oct 08 19:28:12 crc kubenswrapper[4988]: I1008 19:28:12.238976 4988 scope.go:117] "RemoveContainer" containerID="20477a4d0d4c3448d84d69456b20adede3fe2d7c9fa57e53eb98b6ad7aabc27d" Oct 08 19:28:12 crc kubenswrapper[4988]: E1008 19:28:12.239788 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nm54f_openshift-machine-config-operator(f328a20b-5283-407e-8a93-2ab2da599f18)\"" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" Oct 08 19:28:24 crc kubenswrapper[4988]: I1008 19:28:24.238032 4988 scope.go:117] "RemoveContainer" containerID="20477a4d0d4c3448d84d69456b20adede3fe2d7c9fa57e53eb98b6ad7aabc27d" Oct 08 19:28:24 crc kubenswrapper[4988]: I1008 19:28:24.856271 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" event={"ID":"f328a20b-5283-407e-8a93-2ab2da599f18","Type":"ContainerStarted","Data":"6615b98ff2af8d9695912c635a4216452e43b30637b6fb36a988358158ee4055"} Oct 08 19:30:00 crc kubenswrapper[4988]: I1008 19:30:00.153842 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29332530-rgqrs"] Oct 08 19:30:00 crc kubenswrapper[4988]: E1008 19:30:00.154716 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0e1eef88-0680-4fba-9b85-765d6fbe8b99" containerName="registry-server" Oct 08 19:30:00 crc kubenswrapper[4988]: I1008 19:30:00.154732 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="0e1eef88-0680-4fba-9b85-765d6fbe8b99" containerName="registry-server" Oct 08 19:30:00 crc kubenswrapper[4988]: E1008 19:30:00.154748 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0e1eef88-0680-4fba-9b85-765d6fbe8b99" containerName="extract-utilities" Oct 08 19:30:00 crc kubenswrapper[4988]: I1008 19:30:00.154756 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="0e1eef88-0680-4fba-9b85-765d6fbe8b99" containerName="extract-utilities" Oct 08 19:30:00 crc kubenswrapper[4988]: E1008 19:30:00.154789 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0e1eef88-0680-4fba-9b85-765d6fbe8b99" containerName="extract-content" Oct 08 19:30:00 crc kubenswrapper[4988]: I1008 19:30:00.154797 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="0e1eef88-0680-4fba-9b85-765d6fbe8b99" containerName="extract-content" Oct 08 19:30:00 crc kubenswrapper[4988]: I1008 19:30:00.155000 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="0e1eef88-0680-4fba-9b85-765d6fbe8b99" containerName="registry-server" Oct 08 19:30:00 crc kubenswrapper[4988]: I1008 19:30:00.155648 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29332530-rgqrs" Oct 08 19:30:00 crc kubenswrapper[4988]: I1008 19:30:00.158695 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 08 19:30:00 crc kubenswrapper[4988]: I1008 19:30:00.159041 4988 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 08 19:30:00 crc kubenswrapper[4988]: I1008 19:30:00.160682 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29332530-rgqrs"] Oct 08 19:30:00 crc kubenswrapper[4988]: I1008 19:30:00.314589 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/501d2535-94b2-41bf-93e1-262f008d88a6-config-volume\") pod \"collect-profiles-29332530-rgqrs\" (UID: \"501d2535-94b2-41bf-93e1-262f008d88a6\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29332530-rgqrs" Oct 08 19:30:00 crc kubenswrapper[4988]: I1008 19:30:00.314909 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/501d2535-94b2-41bf-93e1-262f008d88a6-secret-volume\") pod \"collect-profiles-29332530-rgqrs\" (UID: \"501d2535-94b2-41bf-93e1-262f008d88a6\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29332530-rgqrs" Oct 08 19:30:00 crc kubenswrapper[4988]: I1008 19:30:00.314966 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wnttp\" (UniqueName: \"kubernetes.io/projected/501d2535-94b2-41bf-93e1-262f008d88a6-kube-api-access-wnttp\") pod \"collect-profiles-29332530-rgqrs\" (UID: \"501d2535-94b2-41bf-93e1-262f008d88a6\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29332530-rgqrs" Oct 08 19:30:00 crc kubenswrapper[4988]: I1008 19:30:00.416289 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/501d2535-94b2-41bf-93e1-262f008d88a6-config-volume\") pod \"collect-profiles-29332530-rgqrs\" (UID: \"501d2535-94b2-41bf-93e1-262f008d88a6\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29332530-rgqrs" Oct 08 19:30:00 crc kubenswrapper[4988]: I1008 19:30:00.416377 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/501d2535-94b2-41bf-93e1-262f008d88a6-secret-volume\") pod \"collect-profiles-29332530-rgqrs\" (UID: \"501d2535-94b2-41bf-93e1-262f008d88a6\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29332530-rgqrs" Oct 08 19:30:00 crc kubenswrapper[4988]: I1008 19:30:00.416508 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wnttp\" (UniqueName: \"kubernetes.io/projected/501d2535-94b2-41bf-93e1-262f008d88a6-kube-api-access-wnttp\") pod \"collect-profiles-29332530-rgqrs\" (UID: \"501d2535-94b2-41bf-93e1-262f008d88a6\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29332530-rgqrs" Oct 08 19:30:00 crc kubenswrapper[4988]: I1008 19:30:00.418019 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/501d2535-94b2-41bf-93e1-262f008d88a6-config-volume\") pod \"collect-profiles-29332530-rgqrs\" (UID: \"501d2535-94b2-41bf-93e1-262f008d88a6\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29332530-rgqrs" Oct 08 19:30:00 crc kubenswrapper[4988]: I1008 19:30:00.427817 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/501d2535-94b2-41bf-93e1-262f008d88a6-secret-volume\") pod \"collect-profiles-29332530-rgqrs\" (UID: \"501d2535-94b2-41bf-93e1-262f008d88a6\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29332530-rgqrs" Oct 08 19:30:00 crc kubenswrapper[4988]: I1008 19:30:00.441742 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wnttp\" (UniqueName: \"kubernetes.io/projected/501d2535-94b2-41bf-93e1-262f008d88a6-kube-api-access-wnttp\") pod \"collect-profiles-29332530-rgqrs\" (UID: \"501d2535-94b2-41bf-93e1-262f008d88a6\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29332530-rgqrs" Oct 08 19:30:00 crc kubenswrapper[4988]: I1008 19:30:00.479377 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29332530-rgqrs" Oct 08 19:30:00 crc kubenswrapper[4988]: I1008 19:30:00.936936 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29332530-rgqrs"] Oct 08 19:30:01 crc kubenswrapper[4988]: I1008 19:30:01.769043 4988 generic.go:334] "Generic (PLEG): container finished" podID="501d2535-94b2-41bf-93e1-262f008d88a6" containerID="5416125fd85a323f78420b1ef7049c539da1430e9d5b0a9596588cb33b829447" exitCode=0 Oct 08 19:30:01 crc kubenswrapper[4988]: I1008 19:30:01.769226 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29332530-rgqrs" event={"ID":"501d2535-94b2-41bf-93e1-262f008d88a6","Type":"ContainerDied","Data":"5416125fd85a323f78420b1ef7049c539da1430e9d5b0a9596588cb33b829447"} Oct 08 19:30:01 crc kubenswrapper[4988]: I1008 19:30:01.769417 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29332530-rgqrs" event={"ID":"501d2535-94b2-41bf-93e1-262f008d88a6","Type":"ContainerStarted","Data":"d5d6a09f5ca55a22a354255de7b8e1512843983d0015359891b8f24aaaad7a89"} Oct 08 19:30:03 crc kubenswrapper[4988]: I1008 19:30:03.204720 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29332530-rgqrs" Oct 08 19:30:03 crc kubenswrapper[4988]: I1008 19:30:03.360992 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wnttp\" (UniqueName: \"kubernetes.io/projected/501d2535-94b2-41bf-93e1-262f008d88a6-kube-api-access-wnttp\") pod \"501d2535-94b2-41bf-93e1-262f008d88a6\" (UID: \"501d2535-94b2-41bf-93e1-262f008d88a6\") " Oct 08 19:30:03 crc kubenswrapper[4988]: I1008 19:30:03.361082 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/501d2535-94b2-41bf-93e1-262f008d88a6-config-volume\") pod \"501d2535-94b2-41bf-93e1-262f008d88a6\" (UID: \"501d2535-94b2-41bf-93e1-262f008d88a6\") " Oct 08 19:30:03 crc kubenswrapper[4988]: I1008 19:30:03.361127 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/501d2535-94b2-41bf-93e1-262f008d88a6-secret-volume\") pod \"501d2535-94b2-41bf-93e1-262f008d88a6\" (UID: \"501d2535-94b2-41bf-93e1-262f008d88a6\") " Oct 08 19:30:03 crc kubenswrapper[4988]: I1008 19:30:03.361959 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/501d2535-94b2-41bf-93e1-262f008d88a6-config-volume" (OuterVolumeSpecName: "config-volume") pod "501d2535-94b2-41bf-93e1-262f008d88a6" (UID: "501d2535-94b2-41bf-93e1-262f008d88a6"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 19:30:03 crc kubenswrapper[4988]: I1008 19:30:03.366627 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/501d2535-94b2-41bf-93e1-262f008d88a6-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "501d2535-94b2-41bf-93e1-262f008d88a6" (UID: "501d2535-94b2-41bf-93e1-262f008d88a6"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 19:30:03 crc kubenswrapper[4988]: I1008 19:30:03.366693 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/501d2535-94b2-41bf-93e1-262f008d88a6-kube-api-access-wnttp" (OuterVolumeSpecName: "kube-api-access-wnttp") pod "501d2535-94b2-41bf-93e1-262f008d88a6" (UID: "501d2535-94b2-41bf-93e1-262f008d88a6"). InnerVolumeSpecName "kube-api-access-wnttp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 19:30:03 crc kubenswrapper[4988]: I1008 19:30:03.462892 4988 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wnttp\" (UniqueName: \"kubernetes.io/projected/501d2535-94b2-41bf-93e1-262f008d88a6-kube-api-access-wnttp\") on node \"crc\" DevicePath \"\"" Oct 08 19:30:03 crc kubenswrapper[4988]: I1008 19:30:03.462945 4988 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/501d2535-94b2-41bf-93e1-262f008d88a6-config-volume\") on node \"crc\" DevicePath \"\"" Oct 08 19:30:03 crc kubenswrapper[4988]: I1008 19:30:03.462968 4988 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/501d2535-94b2-41bf-93e1-262f008d88a6-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 08 19:30:03 crc kubenswrapper[4988]: I1008 19:30:03.657298 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-77tcl"] Oct 08 19:30:03 crc kubenswrapper[4988]: E1008 19:30:03.657923 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="501d2535-94b2-41bf-93e1-262f008d88a6" containerName="collect-profiles" Oct 08 19:30:03 crc kubenswrapper[4988]: I1008 19:30:03.657967 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="501d2535-94b2-41bf-93e1-262f008d88a6" containerName="collect-profiles" Oct 08 19:30:03 crc kubenswrapper[4988]: I1008 19:30:03.658478 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="501d2535-94b2-41bf-93e1-262f008d88a6" containerName="collect-profiles" Oct 08 19:30:03 crc kubenswrapper[4988]: I1008 19:30:03.660879 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-77tcl" Oct 08 19:30:03 crc kubenswrapper[4988]: I1008 19:30:03.665563 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5b8bb56c-d81e-4e7e-bd92-363604cd99d9-catalog-content\") pod \"redhat-operators-77tcl\" (UID: \"5b8bb56c-d81e-4e7e-bd92-363604cd99d9\") " pod="openshift-marketplace/redhat-operators-77tcl" Oct 08 19:30:03 crc kubenswrapper[4988]: I1008 19:30:03.665626 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5b8bb56c-d81e-4e7e-bd92-363604cd99d9-utilities\") pod \"redhat-operators-77tcl\" (UID: \"5b8bb56c-d81e-4e7e-bd92-363604cd99d9\") " pod="openshift-marketplace/redhat-operators-77tcl" Oct 08 19:30:03 crc kubenswrapper[4988]: I1008 19:30:03.665722 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7z6s9\" (UniqueName: \"kubernetes.io/projected/5b8bb56c-d81e-4e7e-bd92-363604cd99d9-kube-api-access-7z6s9\") pod \"redhat-operators-77tcl\" (UID: \"5b8bb56c-d81e-4e7e-bd92-363604cd99d9\") " pod="openshift-marketplace/redhat-operators-77tcl" Oct 08 19:30:03 crc kubenswrapper[4988]: I1008 19:30:03.669796 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-77tcl"] Oct 08 19:30:03 crc kubenswrapper[4988]: I1008 19:30:03.767291 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5b8bb56c-d81e-4e7e-bd92-363604cd99d9-catalog-content\") pod \"redhat-operators-77tcl\" (UID: \"5b8bb56c-d81e-4e7e-bd92-363604cd99d9\") " pod="openshift-marketplace/redhat-operators-77tcl" Oct 08 19:30:03 crc kubenswrapper[4988]: I1008 19:30:03.767355 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5b8bb56c-d81e-4e7e-bd92-363604cd99d9-utilities\") pod \"redhat-operators-77tcl\" (UID: \"5b8bb56c-d81e-4e7e-bd92-363604cd99d9\") " pod="openshift-marketplace/redhat-operators-77tcl" Oct 08 19:30:03 crc kubenswrapper[4988]: I1008 19:30:03.767418 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7z6s9\" (UniqueName: \"kubernetes.io/projected/5b8bb56c-d81e-4e7e-bd92-363604cd99d9-kube-api-access-7z6s9\") pod \"redhat-operators-77tcl\" (UID: \"5b8bb56c-d81e-4e7e-bd92-363604cd99d9\") " pod="openshift-marketplace/redhat-operators-77tcl" Oct 08 19:30:03 crc kubenswrapper[4988]: I1008 19:30:03.768218 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5b8bb56c-d81e-4e7e-bd92-363604cd99d9-utilities\") pod \"redhat-operators-77tcl\" (UID: \"5b8bb56c-d81e-4e7e-bd92-363604cd99d9\") " pod="openshift-marketplace/redhat-operators-77tcl" Oct 08 19:30:03 crc kubenswrapper[4988]: I1008 19:30:03.768289 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5b8bb56c-d81e-4e7e-bd92-363604cd99d9-catalog-content\") pod \"redhat-operators-77tcl\" (UID: \"5b8bb56c-d81e-4e7e-bd92-363604cd99d9\") " pod="openshift-marketplace/redhat-operators-77tcl" Oct 08 19:30:03 crc kubenswrapper[4988]: I1008 19:30:03.791716 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29332530-rgqrs" event={"ID":"501d2535-94b2-41bf-93e1-262f008d88a6","Type":"ContainerDied","Data":"d5d6a09f5ca55a22a354255de7b8e1512843983d0015359891b8f24aaaad7a89"} Oct 08 19:30:03 crc kubenswrapper[4988]: I1008 19:30:03.791772 4988 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d5d6a09f5ca55a22a354255de7b8e1512843983d0015359891b8f24aaaad7a89" Oct 08 19:30:03 crc kubenswrapper[4988]: I1008 19:30:03.791775 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29332530-rgqrs" Oct 08 19:30:03 crc kubenswrapper[4988]: I1008 19:30:03.803582 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7z6s9\" (UniqueName: \"kubernetes.io/projected/5b8bb56c-d81e-4e7e-bd92-363604cd99d9-kube-api-access-7z6s9\") pod \"redhat-operators-77tcl\" (UID: \"5b8bb56c-d81e-4e7e-bd92-363604cd99d9\") " pod="openshift-marketplace/redhat-operators-77tcl" Oct 08 19:30:03 crc kubenswrapper[4988]: I1008 19:30:03.984201 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-77tcl" Oct 08 19:30:04 crc kubenswrapper[4988]: I1008 19:30:04.283522 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29332485-n65qn"] Oct 08 19:30:04 crc kubenswrapper[4988]: I1008 19:30:04.285421 4988 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29332485-n65qn"] Oct 08 19:30:04 crc kubenswrapper[4988]: I1008 19:30:04.431555 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-77tcl"] Oct 08 19:30:04 crc kubenswrapper[4988]: I1008 19:30:04.798999 4988 generic.go:334] "Generic (PLEG): container finished" podID="5b8bb56c-d81e-4e7e-bd92-363604cd99d9" containerID="b470cc83ea952b8031b3fa465312c711439539461fd2d847334839c950431f37" exitCode=0 Oct 08 19:30:04 crc kubenswrapper[4988]: I1008 19:30:04.799065 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-77tcl" event={"ID":"5b8bb56c-d81e-4e7e-bd92-363604cd99d9","Type":"ContainerDied","Data":"b470cc83ea952b8031b3fa465312c711439539461fd2d847334839c950431f37"} Oct 08 19:30:04 crc kubenswrapper[4988]: I1008 19:30:04.799288 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-77tcl" event={"ID":"5b8bb56c-d81e-4e7e-bd92-363604cd99d9","Type":"ContainerStarted","Data":"130e6b13761e7658bcccf0a8855d4e23bf3b0aa2e73323863749a3161eb32912"} Oct 08 19:30:04 crc kubenswrapper[4988]: I1008 19:30:04.801910 4988 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 08 19:30:05 crc kubenswrapper[4988]: I1008 19:30:05.254177 4988 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="58a39bcb-ac94-4290-b2cb-0bea4547a67c" path="/var/lib/kubelet/pods/58a39bcb-ac94-4290-b2cb-0bea4547a67c/volumes" Oct 08 19:30:06 crc kubenswrapper[4988]: I1008 19:30:06.824653 4988 generic.go:334] "Generic (PLEG): container finished" podID="5b8bb56c-d81e-4e7e-bd92-363604cd99d9" containerID="1b5126e2126fb24d834d88c6f0aac73caedb5e7fa2654212fb12a22432632578" exitCode=0 Oct 08 19:30:06 crc kubenswrapper[4988]: I1008 19:30:06.825121 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-77tcl" event={"ID":"5b8bb56c-d81e-4e7e-bd92-363604cd99d9","Type":"ContainerDied","Data":"1b5126e2126fb24d834d88c6f0aac73caedb5e7fa2654212fb12a22432632578"} Oct 08 19:30:07 crc kubenswrapper[4988]: I1008 19:30:07.835696 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-77tcl" event={"ID":"5b8bb56c-d81e-4e7e-bd92-363604cd99d9","Type":"ContainerStarted","Data":"e1cbd338b4106a215339166c1512aa94d8cb9f19b5179204833c75f46da7b776"} Oct 08 19:30:07 crc kubenswrapper[4988]: I1008 19:30:07.862309 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-77tcl" podStartSLOduration=2.4108522580000002 podStartE2EDuration="4.8622923s" podCreationTimestamp="2025-10-08 19:30:03 +0000 UTC" firstStartedPulling="2025-10-08 19:30:04.801530012 +0000 UTC m=+4750.251372802" lastFinishedPulling="2025-10-08 19:30:07.252970054 +0000 UTC m=+4752.702812844" observedRunningTime="2025-10-08 19:30:07.859630335 +0000 UTC m=+4753.309473135" watchObservedRunningTime="2025-10-08 19:30:07.8622923 +0000 UTC m=+4753.312135080" Oct 08 19:30:13 crc kubenswrapper[4988]: I1008 19:30:13.984835 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-77tcl" Oct 08 19:30:13 crc kubenswrapper[4988]: I1008 19:30:13.986866 4988 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-77tcl" Oct 08 19:30:14 crc kubenswrapper[4988]: I1008 19:30:14.060576 4988 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-77tcl" Oct 08 19:30:14 crc kubenswrapper[4988]: I1008 19:30:14.984009 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-77tcl" Oct 08 19:30:15 crc kubenswrapper[4988]: I1008 19:30:15.058304 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-77tcl"] Oct 08 19:30:16 crc kubenswrapper[4988]: I1008 19:30:16.926053 4988 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-77tcl" podUID="5b8bb56c-d81e-4e7e-bd92-363604cd99d9" containerName="registry-server" containerID="cri-o://e1cbd338b4106a215339166c1512aa94d8cb9f19b5179204833c75f46da7b776" gracePeriod=2 Oct 08 19:30:17 crc kubenswrapper[4988]: I1008 19:30:17.934285 4988 generic.go:334] "Generic (PLEG): container finished" podID="5b8bb56c-d81e-4e7e-bd92-363604cd99d9" containerID="e1cbd338b4106a215339166c1512aa94d8cb9f19b5179204833c75f46da7b776" exitCode=0 Oct 08 19:30:17 crc kubenswrapper[4988]: I1008 19:30:17.934497 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-77tcl" event={"ID":"5b8bb56c-d81e-4e7e-bd92-363604cd99d9","Type":"ContainerDied","Data":"e1cbd338b4106a215339166c1512aa94d8cb9f19b5179204833c75f46da7b776"} Oct 08 19:30:18 crc kubenswrapper[4988]: I1008 19:30:18.475591 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-77tcl" Oct 08 19:30:18 crc kubenswrapper[4988]: I1008 19:30:18.611951 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5b8bb56c-d81e-4e7e-bd92-363604cd99d9-utilities\") pod \"5b8bb56c-d81e-4e7e-bd92-363604cd99d9\" (UID: \"5b8bb56c-d81e-4e7e-bd92-363604cd99d9\") " Oct 08 19:30:18 crc kubenswrapper[4988]: I1008 19:30:18.612062 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5b8bb56c-d81e-4e7e-bd92-363604cd99d9-catalog-content\") pod \"5b8bb56c-d81e-4e7e-bd92-363604cd99d9\" (UID: \"5b8bb56c-d81e-4e7e-bd92-363604cd99d9\") " Oct 08 19:30:18 crc kubenswrapper[4988]: I1008 19:30:18.612082 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7z6s9\" (UniqueName: \"kubernetes.io/projected/5b8bb56c-d81e-4e7e-bd92-363604cd99d9-kube-api-access-7z6s9\") pod \"5b8bb56c-d81e-4e7e-bd92-363604cd99d9\" (UID: \"5b8bb56c-d81e-4e7e-bd92-363604cd99d9\") " Oct 08 19:30:18 crc kubenswrapper[4988]: I1008 19:30:18.614651 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5b8bb56c-d81e-4e7e-bd92-363604cd99d9-utilities" (OuterVolumeSpecName: "utilities") pod "5b8bb56c-d81e-4e7e-bd92-363604cd99d9" (UID: "5b8bb56c-d81e-4e7e-bd92-363604cd99d9"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 19:30:18 crc kubenswrapper[4988]: I1008 19:30:18.620152 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5b8bb56c-d81e-4e7e-bd92-363604cd99d9-kube-api-access-7z6s9" (OuterVolumeSpecName: "kube-api-access-7z6s9") pod "5b8bb56c-d81e-4e7e-bd92-363604cd99d9" (UID: "5b8bb56c-d81e-4e7e-bd92-363604cd99d9"). InnerVolumeSpecName "kube-api-access-7z6s9". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 19:30:18 crc kubenswrapper[4988]: I1008 19:30:18.710521 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5b8bb56c-d81e-4e7e-bd92-363604cd99d9-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5b8bb56c-d81e-4e7e-bd92-363604cd99d9" (UID: "5b8bb56c-d81e-4e7e-bd92-363604cd99d9"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 19:30:18 crc kubenswrapper[4988]: I1008 19:30:18.713844 4988 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5b8bb56c-d81e-4e7e-bd92-363604cd99d9-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 08 19:30:18 crc kubenswrapper[4988]: I1008 19:30:18.713867 4988 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7z6s9\" (UniqueName: \"kubernetes.io/projected/5b8bb56c-d81e-4e7e-bd92-363604cd99d9-kube-api-access-7z6s9\") on node \"crc\" DevicePath \"\"" Oct 08 19:30:18 crc kubenswrapper[4988]: I1008 19:30:18.713876 4988 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5b8bb56c-d81e-4e7e-bd92-363604cd99d9-utilities\") on node \"crc\" DevicePath \"\"" Oct 08 19:30:18 crc kubenswrapper[4988]: I1008 19:30:18.947591 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-77tcl" event={"ID":"5b8bb56c-d81e-4e7e-bd92-363604cd99d9","Type":"ContainerDied","Data":"130e6b13761e7658bcccf0a8855d4e23bf3b0aa2e73323863749a3161eb32912"} Oct 08 19:30:18 crc kubenswrapper[4988]: I1008 19:30:18.947656 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-77tcl" Oct 08 19:30:18 crc kubenswrapper[4988]: I1008 19:30:18.947661 4988 scope.go:117] "RemoveContainer" containerID="e1cbd338b4106a215339166c1512aa94d8cb9f19b5179204833c75f46da7b776" Oct 08 19:30:18 crc kubenswrapper[4988]: I1008 19:30:18.980125 4988 scope.go:117] "RemoveContainer" containerID="1b5126e2126fb24d834d88c6f0aac73caedb5e7fa2654212fb12a22432632578" Oct 08 19:30:18 crc kubenswrapper[4988]: I1008 19:30:18.999541 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-77tcl"] Oct 08 19:30:19 crc kubenswrapper[4988]: I1008 19:30:19.008704 4988 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-77tcl"] Oct 08 19:30:19 crc kubenswrapper[4988]: I1008 19:30:19.018221 4988 scope.go:117] "RemoveContainer" containerID="b470cc83ea952b8031b3fa465312c711439539461fd2d847334839c950431f37" Oct 08 19:30:19 crc kubenswrapper[4988]: I1008 19:30:19.248885 4988 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5b8bb56c-d81e-4e7e-bd92-363604cd99d9" path="/var/lib/kubelet/pods/5b8bb56c-d81e-4e7e-bd92-363604cd99d9/volumes" Oct 08 19:30:27 crc kubenswrapper[4988]: I1008 19:30:27.321999 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-rqzbh"] Oct 08 19:30:27 crc kubenswrapper[4988]: E1008 19:30:27.322808 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5b8bb56c-d81e-4e7e-bd92-363604cd99d9" containerName="extract-content" Oct 08 19:30:27 crc kubenswrapper[4988]: I1008 19:30:27.322826 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="5b8bb56c-d81e-4e7e-bd92-363604cd99d9" containerName="extract-content" Oct 08 19:30:27 crc kubenswrapper[4988]: E1008 19:30:27.322858 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5b8bb56c-d81e-4e7e-bd92-363604cd99d9" containerName="extract-utilities" Oct 08 19:30:27 crc kubenswrapper[4988]: I1008 19:30:27.322867 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="5b8bb56c-d81e-4e7e-bd92-363604cd99d9" containerName="extract-utilities" Oct 08 19:30:27 crc kubenswrapper[4988]: E1008 19:30:27.322879 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5b8bb56c-d81e-4e7e-bd92-363604cd99d9" containerName="registry-server" Oct 08 19:30:27 crc kubenswrapper[4988]: I1008 19:30:27.322886 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="5b8bb56c-d81e-4e7e-bd92-363604cd99d9" containerName="registry-server" Oct 08 19:30:27 crc kubenswrapper[4988]: I1008 19:30:27.323059 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="5b8bb56c-d81e-4e7e-bd92-363604cd99d9" containerName="registry-server" Oct 08 19:30:27 crc kubenswrapper[4988]: I1008 19:30:27.324162 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-rqzbh" Oct 08 19:30:27 crc kubenswrapper[4988]: I1008 19:30:27.339897 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-rqzbh"] Oct 08 19:30:27 crc kubenswrapper[4988]: I1008 19:30:27.474948 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cfe2e92b-2da7-4b1b-999e-c94c4b0ce438-utilities\") pod \"certified-operators-rqzbh\" (UID: \"cfe2e92b-2da7-4b1b-999e-c94c4b0ce438\") " pod="openshift-marketplace/certified-operators-rqzbh" Oct 08 19:30:27 crc kubenswrapper[4988]: I1008 19:30:27.475216 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wpwqc\" (UniqueName: \"kubernetes.io/projected/cfe2e92b-2da7-4b1b-999e-c94c4b0ce438-kube-api-access-wpwqc\") pod \"certified-operators-rqzbh\" (UID: \"cfe2e92b-2da7-4b1b-999e-c94c4b0ce438\") " pod="openshift-marketplace/certified-operators-rqzbh" Oct 08 19:30:27 crc kubenswrapper[4988]: I1008 19:30:27.475321 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cfe2e92b-2da7-4b1b-999e-c94c4b0ce438-catalog-content\") pod \"certified-operators-rqzbh\" (UID: \"cfe2e92b-2da7-4b1b-999e-c94c4b0ce438\") " pod="openshift-marketplace/certified-operators-rqzbh" Oct 08 19:30:27 crc kubenswrapper[4988]: I1008 19:30:27.576874 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wpwqc\" (UniqueName: \"kubernetes.io/projected/cfe2e92b-2da7-4b1b-999e-c94c4b0ce438-kube-api-access-wpwqc\") pod \"certified-operators-rqzbh\" (UID: \"cfe2e92b-2da7-4b1b-999e-c94c4b0ce438\") " pod="openshift-marketplace/certified-operators-rqzbh" Oct 08 19:30:27 crc kubenswrapper[4988]: I1008 19:30:27.576953 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cfe2e92b-2da7-4b1b-999e-c94c4b0ce438-catalog-content\") pod \"certified-operators-rqzbh\" (UID: \"cfe2e92b-2da7-4b1b-999e-c94c4b0ce438\") " pod="openshift-marketplace/certified-operators-rqzbh" Oct 08 19:30:27 crc kubenswrapper[4988]: I1008 19:30:27.576993 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cfe2e92b-2da7-4b1b-999e-c94c4b0ce438-utilities\") pod \"certified-operators-rqzbh\" (UID: \"cfe2e92b-2da7-4b1b-999e-c94c4b0ce438\") " pod="openshift-marketplace/certified-operators-rqzbh" Oct 08 19:30:27 crc kubenswrapper[4988]: I1008 19:30:27.577417 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cfe2e92b-2da7-4b1b-999e-c94c4b0ce438-catalog-content\") pod \"certified-operators-rqzbh\" (UID: \"cfe2e92b-2da7-4b1b-999e-c94c4b0ce438\") " pod="openshift-marketplace/certified-operators-rqzbh" Oct 08 19:30:27 crc kubenswrapper[4988]: I1008 19:30:27.577548 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cfe2e92b-2da7-4b1b-999e-c94c4b0ce438-utilities\") pod \"certified-operators-rqzbh\" (UID: \"cfe2e92b-2da7-4b1b-999e-c94c4b0ce438\") " pod="openshift-marketplace/certified-operators-rqzbh" Oct 08 19:30:27 crc kubenswrapper[4988]: I1008 19:30:27.598271 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wpwqc\" (UniqueName: \"kubernetes.io/projected/cfe2e92b-2da7-4b1b-999e-c94c4b0ce438-kube-api-access-wpwqc\") pod \"certified-operators-rqzbh\" (UID: \"cfe2e92b-2da7-4b1b-999e-c94c4b0ce438\") " pod="openshift-marketplace/certified-operators-rqzbh" Oct 08 19:30:27 crc kubenswrapper[4988]: I1008 19:30:27.694975 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-rqzbh" Oct 08 19:30:28 crc kubenswrapper[4988]: I1008 19:30:28.143936 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-rqzbh"] Oct 08 19:30:29 crc kubenswrapper[4988]: I1008 19:30:29.032496 4988 generic.go:334] "Generic (PLEG): container finished" podID="cfe2e92b-2da7-4b1b-999e-c94c4b0ce438" containerID="3e2832a2e4daba22b298091b390f5d7c95cfabadadea1b6b1590e75d82d03a39" exitCode=0 Oct 08 19:30:29 crc kubenswrapper[4988]: I1008 19:30:29.032605 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rqzbh" event={"ID":"cfe2e92b-2da7-4b1b-999e-c94c4b0ce438","Type":"ContainerDied","Data":"3e2832a2e4daba22b298091b390f5d7c95cfabadadea1b6b1590e75d82d03a39"} Oct 08 19:30:29 crc kubenswrapper[4988]: I1008 19:30:29.032918 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rqzbh" event={"ID":"cfe2e92b-2da7-4b1b-999e-c94c4b0ce438","Type":"ContainerStarted","Data":"3876ac86545950ee1395e3f9f03d45c0f693f8d9dcdb11ae578d6cf4f1f35fad"} Oct 08 19:30:31 crc kubenswrapper[4988]: I1008 19:30:31.055336 4988 generic.go:334] "Generic (PLEG): container finished" podID="cfe2e92b-2da7-4b1b-999e-c94c4b0ce438" containerID="aead4b266d8e9ba77bf8cc3b74eafa8c0edebae665361da2b2a5cf3e6caaaf7c" exitCode=0 Oct 08 19:30:31 crc kubenswrapper[4988]: I1008 19:30:31.055652 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rqzbh" event={"ID":"cfe2e92b-2da7-4b1b-999e-c94c4b0ce438","Type":"ContainerDied","Data":"aead4b266d8e9ba77bf8cc3b74eafa8c0edebae665361da2b2a5cf3e6caaaf7c"} Oct 08 19:30:32 crc kubenswrapper[4988]: I1008 19:30:32.071000 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rqzbh" event={"ID":"cfe2e92b-2da7-4b1b-999e-c94c4b0ce438","Type":"ContainerStarted","Data":"494e80f378b948058c13e4b943bcfc7c7755b6f76a923855376ed819e2c70263"} Oct 08 19:30:32 crc kubenswrapper[4988]: I1008 19:30:32.099962 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-rqzbh" podStartSLOduration=2.405658178 podStartE2EDuration="5.099938319s" podCreationTimestamp="2025-10-08 19:30:27 +0000 UTC" firstStartedPulling="2025-10-08 19:30:29.037697833 +0000 UTC m=+4774.487540633" lastFinishedPulling="2025-10-08 19:30:31.731978004 +0000 UTC m=+4777.181820774" observedRunningTime="2025-10-08 19:30:32.091459279 +0000 UTC m=+4777.541302119" watchObservedRunningTime="2025-10-08 19:30:32.099938319 +0000 UTC m=+4777.549781119" Oct 08 19:30:37 crc kubenswrapper[4988]: I1008 19:30:37.695405 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-rqzbh" Oct 08 19:30:37 crc kubenswrapper[4988]: I1008 19:30:37.696171 4988 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-rqzbh" Oct 08 19:30:37 crc kubenswrapper[4988]: I1008 19:30:37.790000 4988 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-rqzbh" Oct 08 19:30:38 crc kubenswrapper[4988]: I1008 19:30:38.180899 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-rqzbh" Oct 08 19:30:39 crc kubenswrapper[4988]: I1008 19:30:39.310477 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-rqzbh"] Oct 08 19:30:40 crc kubenswrapper[4988]: I1008 19:30:40.149241 4988 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-rqzbh" podUID="cfe2e92b-2da7-4b1b-999e-c94c4b0ce438" containerName="registry-server" containerID="cri-o://494e80f378b948058c13e4b943bcfc7c7755b6f76a923855376ed819e2c70263" gracePeriod=2 Oct 08 19:30:40 crc kubenswrapper[4988]: I1008 19:30:40.586487 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-rqzbh" Oct 08 19:30:40 crc kubenswrapper[4988]: I1008 19:30:40.729013 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cfe2e92b-2da7-4b1b-999e-c94c4b0ce438-catalog-content\") pod \"cfe2e92b-2da7-4b1b-999e-c94c4b0ce438\" (UID: \"cfe2e92b-2da7-4b1b-999e-c94c4b0ce438\") " Oct 08 19:30:40 crc kubenswrapper[4988]: I1008 19:30:40.729109 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cfe2e92b-2da7-4b1b-999e-c94c4b0ce438-utilities\") pod \"cfe2e92b-2da7-4b1b-999e-c94c4b0ce438\" (UID: \"cfe2e92b-2da7-4b1b-999e-c94c4b0ce438\") " Oct 08 19:30:40 crc kubenswrapper[4988]: I1008 19:30:40.729162 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wpwqc\" (UniqueName: \"kubernetes.io/projected/cfe2e92b-2da7-4b1b-999e-c94c4b0ce438-kube-api-access-wpwqc\") pod \"cfe2e92b-2da7-4b1b-999e-c94c4b0ce438\" (UID: \"cfe2e92b-2da7-4b1b-999e-c94c4b0ce438\") " Oct 08 19:30:40 crc kubenswrapper[4988]: I1008 19:30:40.732778 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cfe2e92b-2da7-4b1b-999e-c94c4b0ce438-utilities" (OuterVolumeSpecName: "utilities") pod "cfe2e92b-2da7-4b1b-999e-c94c4b0ce438" (UID: "cfe2e92b-2da7-4b1b-999e-c94c4b0ce438"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 19:30:40 crc kubenswrapper[4988]: I1008 19:30:40.736692 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cfe2e92b-2da7-4b1b-999e-c94c4b0ce438-kube-api-access-wpwqc" (OuterVolumeSpecName: "kube-api-access-wpwqc") pod "cfe2e92b-2da7-4b1b-999e-c94c4b0ce438" (UID: "cfe2e92b-2da7-4b1b-999e-c94c4b0ce438"). InnerVolumeSpecName "kube-api-access-wpwqc". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 19:30:40 crc kubenswrapper[4988]: I1008 19:30:40.786736 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cfe2e92b-2da7-4b1b-999e-c94c4b0ce438-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "cfe2e92b-2da7-4b1b-999e-c94c4b0ce438" (UID: "cfe2e92b-2da7-4b1b-999e-c94c4b0ce438"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 19:30:40 crc kubenswrapper[4988]: I1008 19:30:40.830810 4988 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wpwqc\" (UniqueName: \"kubernetes.io/projected/cfe2e92b-2da7-4b1b-999e-c94c4b0ce438-kube-api-access-wpwqc\") on node \"crc\" DevicePath \"\"" Oct 08 19:30:40 crc kubenswrapper[4988]: I1008 19:30:40.830843 4988 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cfe2e92b-2da7-4b1b-999e-c94c4b0ce438-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 08 19:30:40 crc kubenswrapper[4988]: I1008 19:30:40.830852 4988 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cfe2e92b-2da7-4b1b-999e-c94c4b0ce438-utilities\") on node \"crc\" DevicePath \"\"" Oct 08 19:30:41 crc kubenswrapper[4988]: I1008 19:30:41.160330 4988 generic.go:334] "Generic (PLEG): container finished" podID="cfe2e92b-2da7-4b1b-999e-c94c4b0ce438" containerID="494e80f378b948058c13e4b943bcfc7c7755b6f76a923855376ed819e2c70263" exitCode=0 Oct 08 19:30:41 crc kubenswrapper[4988]: I1008 19:30:41.160414 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rqzbh" event={"ID":"cfe2e92b-2da7-4b1b-999e-c94c4b0ce438","Type":"ContainerDied","Data":"494e80f378b948058c13e4b943bcfc7c7755b6f76a923855376ed819e2c70263"} Oct 08 19:30:41 crc kubenswrapper[4988]: I1008 19:30:41.160442 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-rqzbh" Oct 08 19:30:41 crc kubenswrapper[4988]: I1008 19:30:41.160462 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rqzbh" event={"ID":"cfe2e92b-2da7-4b1b-999e-c94c4b0ce438","Type":"ContainerDied","Data":"3876ac86545950ee1395e3f9f03d45c0f693f8d9dcdb11ae578d6cf4f1f35fad"} Oct 08 19:30:41 crc kubenswrapper[4988]: I1008 19:30:41.160490 4988 scope.go:117] "RemoveContainer" containerID="494e80f378b948058c13e4b943bcfc7c7755b6f76a923855376ed819e2c70263" Oct 08 19:30:41 crc kubenswrapper[4988]: I1008 19:30:41.184962 4988 scope.go:117] "RemoveContainer" containerID="aead4b266d8e9ba77bf8cc3b74eafa8c0edebae665361da2b2a5cf3e6caaaf7c" Oct 08 19:30:41 crc kubenswrapper[4988]: I1008 19:30:41.218089 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-rqzbh"] Oct 08 19:30:41 crc kubenswrapper[4988]: I1008 19:30:41.226856 4988 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-rqzbh"] Oct 08 19:30:41 crc kubenswrapper[4988]: I1008 19:30:41.250046 4988 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cfe2e92b-2da7-4b1b-999e-c94c4b0ce438" path="/var/lib/kubelet/pods/cfe2e92b-2da7-4b1b-999e-c94c4b0ce438/volumes" Oct 08 19:30:41 crc kubenswrapper[4988]: I1008 19:30:41.603739 4988 scope.go:117] "RemoveContainer" containerID="3e2832a2e4daba22b298091b390f5d7c95cfabadadea1b6b1590e75d82d03a39" Oct 08 19:30:41 crc kubenswrapper[4988]: I1008 19:30:41.650974 4988 scope.go:117] "RemoveContainer" containerID="494e80f378b948058c13e4b943bcfc7c7755b6f76a923855376ed819e2c70263" Oct 08 19:30:41 crc kubenswrapper[4988]: E1008 19:30:41.652483 4988 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"494e80f378b948058c13e4b943bcfc7c7755b6f76a923855376ed819e2c70263\": container with ID starting with 494e80f378b948058c13e4b943bcfc7c7755b6f76a923855376ed819e2c70263 not found: ID does not exist" containerID="494e80f378b948058c13e4b943bcfc7c7755b6f76a923855376ed819e2c70263" Oct 08 19:30:41 crc kubenswrapper[4988]: I1008 19:30:41.652554 4988 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"494e80f378b948058c13e4b943bcfc7c7755b6f76a923855376ed819e2c70263"} err="failed to get container status \"494e80f378b948058c13e4b943bcfc7c7755b6f76a923855376ed819e2c70263\": rpc error: code = NotFound desc = could not find container \"494e80f378b948058c13e4b943bcfc7c7755b6f76a923855376ed819e2c70263\": container with ID starting with 494e80f378b948058c13e4b943bcfc7c7755b6f76a923855376ed819e2c70263 not found: ID does not exist" Oct 08 19:30:41 crc kubenswrapper[4988]: I1008 19:30:41.652598 4988 scope.go:117] "RemoveContainer" containerID="aead4b266d8e9ba77bf8cc3b74eafa8c0edebae665361da2b2a5cf3e6caaaf7c" Oct 08 19:30:41 crc kubenswrapper[4988]: E1008 19:30:41.654248 4988 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"aead4b266d8e9ba77bf8cc3b74eafa8c0edebae665361da2b2a5cf3e6caaaf7c\": container with ID starting with aead4b266d8e9ba77bf8cc3b74eafa8c0edebae665361da2b2a5cf3e6caaaf7c not found: ID does not exist" containerID="aead4b266d8e9ba77bf8cc3b74eafa8c0edebae665361da2b2a5cf3e6caaaf7c" Oct 08 19:30:41 crc kubenswrapper[4988]: I1008 19:30:41.655358 4988 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"aead4b266d8e9ba77bf8cc3b74eafa8c0edebae665361da2b2a5cf3e6caaaf7c"} err="failed to get container status \"aead4b266d8e9ba77bf8cc3b74eafa8c0edebae665361da2b2a5cf3e6caaaf7c\": rpc error: code = NotFound desc = could not find container \"aead4b266d8e9ba77bf8cc3b74eafa8c0edebae665361da2b2a5cf3e6caaaf7c\": container with ID starting with aead4b266d8e9ba77bf8cc3b74eafa8c0edebae665361da2b2a5cf3e6caaaf7c not found: ID does not exist" Oct 08 19:30:41 crc kubenswrapper[4988]: I1008 19:30:41.655460 4988 scope.go:117] "RemoveContainer" containerID="3e2832a2e4daba22b298091b390f5d7c95cfabadadea1b6b1590e75d82d03a39" Oct 08 19:30:41 crc kubenswrapper[4988]: E1008 19:30:41.657583 4988 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3e2832a2e4daba22b298091b390f5d7c95cfabadadea1b6b1590e75d82d03a39\": container with ID starting with 3e2832a2e4daba22b298091b390f5d7c95cfabadadea1b6b1590e75d82d03a39 not found: ID does not exist" containerID="3e2832a2e4daba22b298091b390f5d7c95cfabadadea1b6b1590e75d82d03a39" Oct 08 19:30:41 crc kubenswrapper[4988]: I1008 19:30:41.657656 4988 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3e2832a2e4daba22b298091b390f5d7c95cfabadadea1b6b1590e75d82d03a39"} err="failed to get container status \"3e2832a2e4daba22b298091b390f5d7c95cfabadadea1b6b1590e75d82d03a39\": rpc error: code = NotFound desc = could not find container \"3e2832a2e4daba22b298091b390f5d7c95cfabadadea1b6b1590e75d82d03a39\": container with ID starting with 3e2832a2e4daba22b298091b390f5d7c95cfabadadea1b6b1590e75d82d03a39 not found: ID does not exist" Oct 08 19:30:52 crc kubenswrapper[4988]: I1008 19:30:52.709494 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["crc-storage/crc-storage-crc-4s445"] Oct 08 19:30:52 crc kubenswrapper[4988]: I1008 19:30:52.714928 4988 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["crc-storage/crc-storage-crc-4s445"] Oct 08 19:30:52 crc kubenswrapper[4988]: I1008 19:30:52.864942 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["crc-storage/crc-storage-crc-8qc9q"] Oct 08 19:30:52 crc kubenswrapper[4988]: E1008 19:30:52.866184 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cfe2e92b-2da7-4b1b-999e-c94c4b0ce438" containerName="extract-content" Oct 08 19:30:52 crc kubenswrapper[4988]: I1008 19:30:52.866235 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="cfe2e92b-2da7-4b1b-999e-c94c4b0ce438" containerName="extract-content" Oct 08 19:30:52 crc kubenswrapper[4988]: E1008 19:30:52.866273 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cfe2e92b-2da7-4b1b-999e-c94c4b0ce438" containerName="registry-server" Oct 08 19:30:52 crc kubenswrapper[4988]: I1008 19:30:52.866291 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="cfe2e92b-2da7-4b1b-999e-c94c4b0ce438" containerName="registry-server" Oct 08 19:30:52 crc kubenswrapper[4988]: E1008 19:30:52.866355 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cfe2e92b-2da7-4b1b-999e-c94c4b0ce438" containerName="extract-utilities" Oct 08 19:30:52 crc kubenswrapper[4988]: I1008 19:30:52.866378 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="cfe2e92b-2da7-4b1b-999e-c94c4b0ce438" containerName="extract-utilities" Oct 08 19:30:52 crc kubenswrapper[4988]: I1008 19:30:52.866762 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="cfe2e92b-2da7-4b1b-999e-c94c4b0ce438" containerName="registry-server" Oct 08 19:30:52 crc kubenswrapper[4988]: I1008 19:30:52.867996 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-8qc9q" Oct 08 19:30:52 crc kubenswrapper[4988]: I1008 19:30:52.871086 4988 reflector.go:368] Caches populated for *v1.Secret from object-"crc-storage"/"crc-storage-dockercfg-frtmh" Oct 08 19:30:52 crc kubenswrapper[4988]: I1008 19:30:52.871178 4988 reflector.go:368] Caches populated for *v1.ConfigMap from object-"crc-storage"/"crc-storage" Oct 08 19:30:52 crc kubenswrapper[4988]: I1008 19:30:52.871252 4988 reflector.go:368] Caches populated for *v1.ConfigMap from object-"crc-storage"/"kube-root-ca.crt" Oct 08 19:30:52 crc kubenswrapper[4988]: I1008 19:30:52.871289 4988 reflector.go:368] Caches populated for *v1.ConfigMap from object-"crc-storage"/"openshift-service-ca.crt" Oct 08 19:30:52 crc kubenswrapper[4988]: I1008 19:30:52.875102 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["crc-storage/crc-storage-crc-8qc9q"] Oct 08 19:30:53 crc kubenswrapper[4988]: I1008 19:30:53.026719 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/49eb6a32-0d05-491c-a103-08147a25636c-node-mnt\") pod \"crc-storage-crc-8qc9q\" (UID: \"49eb6a32-0d05-491c-a103-08147a25636c\") " pod="crc-storage/crc-storage-crc-8qc9q" Oct 08 19:30:53 crc kubenswrapper[4988]: I1008 19:30:53.027381 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/49eb6a32-0d05-491c-a103-08147a25636c-crc-storage\") pod \"crc-storage-crc-8qc9q\" (UID: \"49eb6a32-0d05-491c-a103-08147a25636c\") " pod="crc-storage/crc-storage-crc-8qc9q" Oct 08 19:30:53 crc kubenswrapper[4988]: I1008 19:30:53.027616 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x4dpt\" (UniqueName: \"kubernetes.io/projected/49eb6a32-0d05-491c-a103-08147a25636c-kube-api-access-x4dpt\") pod \"crc-storage-crc-8qc9q\" (UID: \"49eb6a32-0d05-491c-a103-08147a25636c\") " pod="crc-storage/crc-storage-crc-8qc9q" Oct 08 19:30:53 crc kubenswrapper[4988]: I1008 19:30:53.130095 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/49eb6a32-0d05-491c-a103-08147a25636c-node-mnt\") pod \"crc-storage-crc-8qc9q\" (UID: \"49eb6a32-0d05-491c-a103-08147a25636c\") " pod="crc-storage/crc-storage-crc-8qc9q" Oct 08 19:30:53 crc kubenswrapper[4988]: I1008 19:30:53.129757 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/49eb6a32-0d05-491c-a103-08147a25636c-node-mnt\") pod \"crc-storage-crc-8qc9q\" (UID: \"49eb6a32-0d05-491c-a103-08147a25636c\") " pod="crc-storage/crc-storage-crc-8qc9q" Oct 08 19:30:53 crc kubenswrapper[4988]: I1008 19:30:53.130699 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/49eb6a32-0d05-491c-a103-08147a25636c-crc-storage\") pod \"crc-storage-crc-8qc9q\" (UID: \"49eb6a32-0d05-491c-a103-08147a25636c\") " pod="crc-storage/crc-storage-crc-8qc9q" Oct 08 19:30:53 crc kubenswrapper[4988]: I1008 19:30:53.130761 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x4dpt\" (UniqueName: \"kubernetes.io/projected/49eb6a32-0d05-491c-a103-08147a25636c-kube-api-access-x4dpt\") pod \"crc-storage-crc-8qc9q\" (UID: \"49eb6a32-0d05-491c-a103-08147a25636c\") " pod="crc-storage/crc-storage-crc-8qc9q" Oct 08 19:30:53 crc kubenswrapper[4988]: I1008 19:30:53.132014 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/49eb6a32-0d05-491c-a103-08147a25636c-crc-storage\") pod \"crc-storage-crc-8qc9q\" (UID: \"49eb6a32-0d05-491c-a103-08147a25636c\") " pod="crc-storage/crc-storage-crc-8qc9q" Oct 08 19:30:53 crc kubenswrapper[4988]: I1008 19:30:53.168897 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x4dpt\" (UniqueName: \"kubernetes.io/projected/49eb6a32-0d05-491c-a103-08147a25636c-kube-api-access-x4dpt\") pod \"crc-storage-crc-8qc9q\" (UID: \"49eb6a32-0d05-491c-a103-08147a25636c\") " pod="crc-storage/crc-storage-crc-8qc9q" Oct 08 19:30:53 crc kubenswrapper[4988]: I1008 19:30:53.205702 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-8qc9q" Oct 08 19:30:53 crc kubenswrapper[4988]: I1008 19:30:53.259444 4988 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3cf2d054-6c91-4e4c-829b-1e038c477c41" path="/var/lib/kubelet/pods/3cf2d054-6c91-4e4c-829b-1e038c477c41/volumes" Oct 08 19:30:53 crc kubenswrapper[4988]: I1008 19:30:53.338119 4988 patch_prober.go:28] interesting pod/machine-config-daemon-nm54f container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 19:30:53 crc kubenswrapper[4988]: I1008 19:30:53.338254 4988 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 19:30:53 crc kubenswrapper[4988]: I1008 19:30:53.724866 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["crc-storage/crc-storage-crc-8qc9q"] Oct 08 19:30:54 crc kubenswrapper[4988]: I1008 19:30:54.311298 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="crc-storage/crc-storage-crc-8qc9q" event={"ID":"49eb6a32-0d05-491c-a103-08147a25636c","Type":"ContainerStarted","Data":"36db859cc4800d31db2604b822420f5d147d447f177b709c26d93488ef7c40b1"} Oct 08 19:30:55 crc kubenswrapper[4988]: I1008 19:30:55.321491 4988 generic.go:334] "Generic (PLEG): container finished" podID="49eb6a32-0d05-491c-a103-08147a25636c" containerID="958238b7f2a84803fc5d7c9871650e2109caba0937a5700fc4a4e0599c4a5569" exitCode=0 Oct 08 19:30:55 crc kubenswrapper[4988]: I1008 19:30:55.321597 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="crc-storage/crc-storage-crc-8qc9q" event={"ID":"49eb6a32-0d05-491c-a103-08147a25636c","Type":"ContainerDied","Data":"958238b7f2a84803fc5d7c9871650e2109caba0937a5700fc4a4e0599c4a5569"} Oct 08 19:30:56 crc kubenswrapper[4988]: I1008 19:30:56.745976 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-8qc9q" Oct 08 19:30:56 crc kubenswrapper[4988]: I1008 19:30:56.889651 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/49eb6a32-0d05-491c-a103-08147a25636c-crc-storage\") pod \"49eb6a32-0d05-491c-a103-08147a25636c\" (UID: \"49eb6a32-0d05-491c-a103-08147a25636c\") " Oct 08 19:30:56 crc kubenswrapper[4988]: I1008 19:30:56.889777 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/49eb6a32-0d05-491c-a103-08147a25636c-node-mnt\") pod \"49eb6a32-0d05-491c-a103-08147a25636c\" (UID: \"49eb6a32-0d05-491c-a103-08147a25636c\") " Oct 08 19:30:56 crc kubenswrapper[4988]: I1008 19:30:56.889832 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x4dpt\" (UniqueName: \"kubernetes.io/projected/49eb6a32-0d05-491c-a103-08147a25636c-kube-api-access-x4dpt\") pod \"49eb6a32-0d05-491c-a103-08147a25636c\" (UID: \"49eb6a32-0d05-491c-a103-08147a25636c\") " Oct 08 19:30:56 crc kubenswrapper[4988]: I1008 19:30:56.889945 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/49eb6a32-0d05-491c-a103-08147a25636c-node-mnt" (OuterVolumeSpecName: "node-mnt") pod "49eb6a32-0d05-491c-a103-08147a25636c" (UID: "49eb6a32-0d05-491c-a103-08147a25636c"). InnerVolumeSpecName "node-mnt". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 08 19:30:56 crc kubenswrapper[4988]: I1008 19:30:56.890308 4988 reconciler_common.go:293] "Volume detached for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/49eb6a32-0d05-491c-a103-08147a25636c-node-mnt\") on node \"crc\" DevicePath \"\"" Oct 08 19:30:56 crc kubenswrapper[4988]: I1008 19:30:56.895569 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49eb6a32-0d05-491c-a103-08147a25636c-kube-api-access-x4dpt" (OuterVolumeSpecName: "kube-api-access-x4dpt") pod "49eb6a32-0d05-491c-a103-08147a25636c" (UID: "49eb6a32-0d05-491c-a103-08147a25636c"). InnerVolumeSpecName "kube-api-access-x4dpt". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 19:30:56 crc kubenswrapper[4988]: I1008 19:30:56.910955 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49eb6a32-0d05-491c-a103-08147a25636c-crc-storage" (OuterVolumeSpecName: "crc-storage") pod "49eb6a32-0d05-491c-a103-08147a25636c" (UID: "49eb6a32-0d05-491c-a103-08147a25636c"). InnerVolumeSpecName "crc-storage". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 19:30:56 crc kubenswrapper[4988]: I1008 19:30:56.992250 4988 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x4dpt\" (UniqueName: \"kubernetes.io/projected/49eb6a32-0d05-491c-a103-08147a25636c-kube-api-access-x4dpt\") on node \"crc\" DevicePath \"\"" Oct 08 19:30:56 crc kubenswrapper[4988]: I1008 19:30:56.992292 4988 reconciler_common.go:293] "Volume detached for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/49eb6a32-0d05-491c-a103-08147a25636c-crc-storage\") on node \"crc\" DevicePath \"\"" Oct 08 19:30:57 crc kubenswrapper[4988]: I1008 19:30:57.344552 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="crc-storage/crc-storage-crc-8qc9q" event={"ID":"49eb6a32-0d05-491c-a103-08147a25636c","Type":"ContainerDied","Data":"36db859cc4800d31db2604b822420f5d147d447f177b709c26d93488ef7c40b1"} Oct 08 19:30:57 crc kubenswrapper[4988]: I1008 19:30:57.344611 4988 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="36db859cc4800d31db2604b822420f5d147d447f177b709c26d93488ef7c40b1" Oct 08 19:30:57 crc kubenswrapper[4988]: I1008 19:30:57.344633 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-8qc9q" Oct 08 19:30:59 crc kubenswrapper[4988]: I1008 19:30:59.018887 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["crc-storage/crc-storage-crc-8qc9q"] Oct 08 19:30:59 crc kubenswrapper[4988]: I1008 19:30:59.027192 4988 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["crc-storage/crc-storage-crc-8qc9q"] Oct 08 19:30:59 crc kubenswrapper[4988]: I1008 19:30:59.157824 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["crc-storage/crc-storage-crc-k6z2l"] Oct 08 19:30:59 crc kubenswrapper[4988]: E1008 19:30:59.158327 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="49eb6a32-0d05-491c-a103-08147a25636c" containerName="storage" Oct 08 19:30:59 crc kubenswrapper[4988]: I1008 19:30:59.158371 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="49eb6a32-0d05-491c-a103-08147a25636c" containerName="storage" Oct 08 19:30:59 crc kubenswrapper[4988]: I1008 19:30:59.158800 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="49eb6a32-0d05-491c-a103-08147a25636c" containerName="storage" Oct 08 19:30:59 crc kubenswrapper[4988]: I1008 19:30:59.159822 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-k6z2l" Oct 08 19:30:59 crc kubenswrapper[4988]: I1008 19:30:59.185297 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["crc-storage/crc-storage-crc-k6z2l"] Oct 08 19:30:59 crc kubenswrapper[4988]: I1008 19:30:59.195753 4988 reflector.go:368] Caches populated for *v1.ConfigMap from object-"crc-storage"/"crc-storage" Oct 08 19:30:59 crc kubenswrapper[4988]: I1008 19:30:59.196042 4988 reflector.go:368] Caches populated for *v1.ConfigMap from object-"crc-storage"/"openshift-service-ca.crt" Oct 08 19:30:59 crc kubenswrapper[4988]: I1008 19:30:59.196132 4988 reflector.go:368] Caches populated for *v1.ConfigMap from object-"crc-storage"/"kube-root-ca.crt" Oct 08 19:30:59 crc kubenswrapper[4988]: I1008 19:30:59.196256 4988 reflector.go:368] Caches populated for *v1.Secret from object-"crc-storage"/"crc-storage-dockercfg-frtmh" Oct 08 19:30:59 crc kubenswrapper[4988]: I1008 19:30:59.261573 4988 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49eb6a32-0d05-491c-a103-08147a25636c" path="/var/lib/kubelet/pods/49eb6a32-0d05-491c-a103-08147a25636c/volumes" Oct 08 19:30:59 crc kubenswrapper[4988]: I1008 19:30:59.328353 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/44a8b5df-f874-456f-b786-819ccaadd236-crc-storage\") pod \"crc-storage-crc-k6z2l\" (UID: \"44a8b5df-f874-456f-b786-819ccaadd236\") " pod="crc-storage/crc-storage-crc-k6z2l" Oct 08 19:30:59 crc kubenswrapper[4988]: I1008 19:30:59.328459 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-87r9x\" (UniqueName: \"kubernetes.io/projected/44a8b5df-f874-456f-b786-819ccaadd236-kube-api-access-87r9x\") pod \"crc-storage-crc-k6z2l\" (UID: \"44a8b5df-f874-456f-b786-819ccaadd236\") " pod="crc-storage/crc-storage-crc-k6z2l" Oct 08 19:30:59 crc kubenswrapper[4988]: I1008 19:30:59.328532 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/44a8b5df-f874-456f-b786-819ccaadd236-node-mnt\") pod \"crc-storage-crc-k6z2l\" (UID: \"44a8b5df-f874-456f-b786-819ccaadd236\") " pod="crc-storage/crc-storage-crc-k6z2l" Oct 08 19:30:59 crc kubenswrapper[4988]: I1008 19:30:59.430786 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-87r9x\" (UniqueName: \"kubernetes.io/projected/44a8b5df-f874-456f-b786-819ccaadd236-kube-api-access-87r9x\") pod \"crc-storage-crc-k6z2l\" (UID: \"44a8b5df-f874-456f-b786-819ccaadd236\") " pod="crc-storage/crc-storage-crc-k6z2l" Oct 08 19:30:59 crc kubenswrapper[4988]: I1008 19:30:59.430845 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/44a8b5df-f874-456f-b786-819ccaadd236-node-mnt\") pod \"crc-storage-crc-k6z2l\" (UID: \"44a8b5df-f874-456f-b786-819ccaadd236\") " pod="crc-storage/crc-storage-crc-k6z2l" Oct 08 19:30:59 crc kubenswrapper[4988]: I1008 19:30:59.430899 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/44a8b5df-f874-456f-b786-819ccaadd236-crc-storage\") pod \"crc-storage-crc-k6z2l\" (UID: \"44a8b5df-f874-456f-b786-819ccaadd236\") " pod="crc-storage/crc-storage-crc-k6z2l" Oct 08 19:30:59 crc kubenswrapper[4988]: I1008 19:30:59.431125 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/44a8b5df-f874-456f-b786-819ccaadd236-node-mnt\") pod \"crc-storage-crc-k6z2l\" (UID: \"44a8b5df-f874-456f-b786-819ccaadd236\") " pod="crc-storage/crc-storage-crc-k6z2l" Oct 08 19:30:59 crc kubenswrapper[4988]: I1008 19:30:59.431530 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/44a8b5df-f874-456f-b786-819ccaadd236-crc-storage\") pod \"crc-storage-crc-k6z2l\" (UID: \"44a8b5df-f874-456f-b786-819ccaadd236\") " pod="crc-storage/crc-storage-crc-k6z2l" Oct 08 19:30:59 crc kubenswrapper[4988]: I1008 19:30:59.450293 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-87r9x\" (UniqueName: \"kubernetes.io/projected/44a8b5df-f874-456f-b786-819ccaadd236-kube-api-access-87r9x\") pod \"crc-storage-crc-k6z2l\" (UID: \"44a8b5df-f874-456f-b786-819ccaadd236\") " pod="crc-storage/crc-storage-crc-k6z2l" Oct 08 19:30:59 crc kubenswrapper[4988]: I1008 19:30:59.528468 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-k6z2l" Oct 08 19:31:00 crc kubenswrapper[4988]: I1008 19:31:00.041844 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["crc-storage/crc-storage-crc-k6z2l"] Oct 08 19:31:00 crc kubenswrapper[4988]: I1008 19:31:00.379591 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="crc-storage/crc-storage-crc-k6z2l" event={"ID":"44a8b5df-f874-456f-b786-819ccaadd236","Type":"ContainerStarted","Data":"c625fcc8e71f0d04f5ddf02a9d40a66a9e6c03e33ea3f5b8d26b2e3aa0ca98b2"} Oct 08 19:31:01 crc kubenswrapper[4988]: I1008 19:31:01.391203 4988 generic.go:334] "Generic (PLEG): container finished" podID="44a8b5df-f874-456f-b786-819ccaadd236" containerID="5735878f5d07c49094cb988269b80ae008bcdb2fc4711d669287d19984b9a329" exitCode=0 Oct 08 19:31:01 crc kubenswrapper[4988]: I1008 19:31:01.391302 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="crc-storage/crc-storage-crc-k6z2l" event={"ID":"44a8b5df-f874-456f-b786-819ccaadd236","Type":"ContainerDied","Data":"5735878f5d07c49094cb988269b80ae008bcdb2fc4711d669287d19984b9a329"} Oct 08 19:31:02 crc kubenswrapper[4988]: I1008 19:31:02.477416 4988 scope.go:117] "RemoveContainer" containerID="035f001a51642541a361b47eefe00a39e503f22f29697eb33feb5c4a611ce3e3" Oct 08 19:31:02 crc kubenswrapper[4988]: I1008 19:31:02.510186 4988 scope.go:117] "RemoveContainer" containerID="2cc88615496f18e32e9c6b61e125c82fc20f3b526f8ead63c6e79da3b05461a8" Oct 08 19:31:02 crc kubenswrapper[4988]: I1008 19:31:02.781839 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-k6z2l" Oct 08 19:31:02 crc kubenswrapper[4988]: I1008 19:31:02.911473 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/44a8b5df-f874-456f-b786-819ccaadd236-crc-storage\") pod \"44a8b5df-f874-456f-b786-819ccaadd236\" (UID: \"44a8b5df-f874-456f-b786-819ccaadd236\") " Oct 08 19:31:02 crc kubenswrapper[4988]: I1008 19:31:02.911568 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-87r9x\" (UniqueName: \"kubernetes.io/projected/44a8b5df-f874-456f-b786-819ccaadd236-kube-api-access-87r9x\") pod \"44a8b5df-f874-456f-b786-819ccaadd236\" (UID: \"44a8b5df-f874-456f-b786-819ccaadd236\") " Oct 08 19:31:02 crc kubenswrapper[4988]: I1008 19:31:02.911622 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/44a8b5df-f874-456f-b786-819ccaadd236-node-mnt\") pod \"44a8b5df-f874-456f-b786-819ccaadd236\" (UID: \"44a8b5df-f874-456f-b786-819ccaadd236\") " Oct 08 19:31:02 crc kubenswrapper[4988]: I1008 19:31:02.911966 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/44a8b5df-f874-456f-b786-819ccaadd236-node-mnt" (OuterVolumeSpecName: "node-mnt") pod "44a8b5df-f874-456f-b786-819ccaadd236" (UID: "44a8b5df-f874-456f-b786-819ccaadd236"). InnerVolumeSpecName "node-mnt". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 08 19:31:02 crc kubenswrapper[4988]: I1008 19:31:02.918606 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/44a8b5df-f874-456f-b786-819ccaadd236-kube-api-access-87r9x" (OuterVolumeSpecName: "kube-api-access-87r9x") pod "44a8b5df-f874-456f-b786-819ccaadd236" (UID: "44a8b5df-f874-456f-b786-819ccaadd236"). InnerVolumeSpecName "kube-api-access-87r9x". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 19:31:02 crc kubenswrapper[4988]: I1008 19:31:02.939018 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/44a8b5df-f874-456f-b786-819ccaadd236-crc-storage" (OuterVolumeSpecName: "crc-storage") pod "44a8b5df-f874-456f-b786-819ccaadd236" (UID: "44a8b5df-f874-456f-b786-819ccaadd236"). InnerVolumeSpecName "crc-storage". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 19:31:03 crc kubenswrapper[4988]: I1008 19:31:03.013682 4988 reconciler_common.go:293] "Volume detached for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/44a8b5df-f874-456f-b786-819ccaadd236-crc-storage\") on node \"crc\" DevicePath \"\"" Oct 08 19:31:03 crc kubenswrapper[4988]: I1008 19:31:03.014091 4988 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-87r9x\" (UniqueName: \"kubernetes.io/projected/44a8b5df-f874-456f-b786-819ccaadd236-kube-api-access-87r9x\") on node \"crc\" DevicePath \"\"" Oct 08 19:31:03 crc kubenswrapper[4988]: I1008 19:31:03.014164 4988 reconciler_common.go:293] "Volume detached for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/44a8b5df-f874-456f-b786-819ccaadd236-node-mnt\") on node \"crc\" DevicePath \"\"" Oct 08 19:31:03 crc kubenswrapper[4988]: I1008 19:31:03.410926 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="crc-storage/crc-storage-crc-k6z2l" event={"ID":"44a8b5df-f874-456f-b786-819ccaadd236","Type":"ContainerDied","Data":"c625fcc8e71f0d04f5ddf02a9d40a66a9e6c03e33ea3f5b8d26b2e3aa0ca98b2"} Oct 08 19:31:03 crc kubenswrapper[4988]: I1008 19:31:03.410984 4988 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c625fcc8e71f0d04f5ddf02a9d40a66a9e6c03e33ea3f5b8d26b2e3aa0ca98b2" Oct 08 19:31:03 crc kubenswrapper[4988]: I1008 19:31:03.410989 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-k6z2l" Oct 08 19:31:23 crc kubenswrapper[4988]: I1008 19:31:23.338107 4988 patch_prober.go:28] interesting pod/machine-config-daemon-nm54f container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 19:31:23 crc kubenswrapper[4988]: I1008 19:31:23.338661 4988 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 19:31:53 crc kubenswrapper[4988]: I1008 19:31:53.340056 4988 patch_prober.go:28] interesting pod/machine-config-daemon-nm54f container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 19:31:53 crc kubenswrapper[4988]: I1008 19:31:53.340710 4988 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 19:31:53 crc kubenswrapper[4988]: I1008 19:31:53.340775 4988 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" Oct 08 19:31:53 crc kubenswrapper[4988]: I1008 19:31:53.341643 4988 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"6615b98ff2af8d9695912c635a4216452e43b30637b6fb36a988358158ee4055"} pod="openshift-machine-config-operator/machine-config-daemon-nm54f" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 08 19:31:53 crc kubenswrapper[4988]: I1008 19:31:53.341730 4988 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" containerName="machine-config-daemon" containerID="cri-o://6615b98ff2af8d9695912c635a4216452e43b30637b6fb36a988358158ee4055" gracePeriod=600 Oct 08 19:31:53 crc kubenswrapper[4988]: I1008 19:31:53.936042 4988 generic.go:334] "Generic (PLEG): container finished" podID="f328a20b-5283-407e-8a93-2ab2da599f18" containerID="6615b98ff2af8d9695912c635a4216452e43b30637b6fb36a988358158ee4055" exitCode=0 Oct 08 19:31:53 crc kubenswrapper[4988]: I1008 19:31:53.936102 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" event={"ID":"f328a20b-5283-407e-8a93-2ab2da599f18","Type":"ContainerDied","Data":"6615b98ff2af8d9695912c635a4216452e43b30637b6fb36a988358158ee4055"} Oct 08 19:31:53 crc kubenswrapper[4988]: I1008 19:31:53.936407 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" event={"ID":"f328a20b-5283-407e-8a93-2ab2da599f18","Type":"ContainerStarted","Data":"3cd52d2ff674f037f4b111ff08caa426d4d7977052a05e385f05362baabb8da9"} Oct 08 19:31:53 crc kubenswrapper[4988]: I1008 19:31:53.936425 4988 scope.go:117] "RemoveContainer" containerID="20477a4d0d4c3448d84d69456b20adede3fe2d7c9fa57e53eb98b6ad7aabc27d" Oct 08 19:33:01 crc kubenswrapper[4988]: I1008 19:33:01.398319 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-678578b8df-87kwv"] Oct 08 19:33:01 crc kubenswrapper[4988]: E1008 19:33:01.399361 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="44a8b5df-f874-456f-b786-819ccaadd236" containerName="storage" Oct 08 19:33:01 crc kubenswrapper[4988]: I1008 19:33:01.399377 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="44a8b5df-f874-456f-b786-819ccaadd236" containerName="storage" Oct 08 19:33:01 crc kubenswrapper[4988]: I1008 19:33:01.399570 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="44a8b5df-f874-456f-b786-819ccaadd236" containerName="storage" Oct 08 19:33:01 crc kubenswrapper[4988]: I1008 19:33:01.400540 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-678578b8df-87kwv" Oct 08 19:33:01 crc kubenswrapper[4988]: I1008 19:33:01.404094 4988 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"kube-root-ca.crt" Oct 08 19:33:01 crc kubenswrapper[4988]: I1008 19:33:01.404382 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dnsmasq-dns-dockercfg-jbr9d" Oct 08 19:33:01 crc kubenswrapper[4988]: I1008 19:33:01.404875 4988 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns" Oct 08 19:33:01 crc kubenswrapper[4988]: I1008 19:33:01.405101 4988 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openshift-service-ca.crt" Oct 08 19:33:01 crc kubenswrapper[4988]: I1008 19:33:01.409963 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qwcqh\" (UniqueName: \"kubernetes.io/projected/ed8779b3-c8aa-4a12-90db-e0f09722d942-kube-api-access-qwcqh\") pod \"dnsmasq-dns-678578b8df-87kwv\" (UID: \"ed8779b3-c8aa-4a12-90db-e0f09722d942\") " pod="openstack/dnsmasq-dns-678578b8df-87kwv" Oct 08 19:33:01 crc kubenswrapper[4988]: I1008 19:33:01.410028 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ed8779b3-c8aa-4a12-90db-e0f09722d942-config\") pod \"dnsmasq-dns-678578b8df-87kwv\" (UID: \"ed8779b3-c8aa-4a12-90db-e0f09722d942\") " pod="openstack/dnsmasq-dns-678578b8df-87kwv" Oct 08 19:33:01 crc kubenswrapper[4988]: I1008 19:33:01.411003 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-6b8f87f5c5-79zbp"] Oct 08 19:33:01 crc kubenswrapper[4988]: I1008 19:33:01.414961 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6b8f87f5c5-79zbp" Oct 08 19:33:01 crc kubenswrapper[4988]: I1008 19:33:01.419806 4988 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns-svc" Oct 08 19:33:01 crc kubenswrapper[4988]: I1008 19:33:01.431601 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-678578b8df-87kwv"] Oct 08 19:33:01 crc kubenswrapper[4988]: I1008 19:33:01.443653 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6b8f87f5c5-79zbp"] Oct 08 19:33:01 crc kubenswrapper[4988]: I1008 19:33:01.511544 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f160a4ae-d7d0-4f4d-89a7-c2d8bd07bf4d-config\") pod \"dnsmasq-dns-6b8f87f5c5-79zbp\" (UID: \"f160a4ae-d7d0-4f4d-89a7-c2d8bd07bf4d\") " pod="openstack/dnsmasq-dns-6b8f87f5c5-79zbp" Oct 08 19:33:01 crc kubenswrapper[4988]: I1008 19:33:01.511614 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qwcqh\" (UniqueName: \"kubernetes.io/projected/ed8779b3-c8aa-4a12-90db-e0f09722d942-kube-api-access-qwcqh\") pod \"dnsmasq-dns-678578b8df-87kwv\" (UID: \"ed8779b3-c8aa-4a12-90db-e0f09722d942\") " pod="openstack/dnsmasq-dns-678578b8df-87kwv" Oct 08 19:33:01 crc kubenswrapper[4988]: I1008 19:33:01.511668 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s5qdw\" (UniqueName: \"kubernetes.io/projected/f160a4ae-d7d0-4f4d-89a7-c2d8bd07bf4d-kube-api-access-s5qdw\") pod \"dnsmasq-dns-6b8f87f5c5-79zbp\" (UID: \"f160a4ae-d7d0-4f4d-89a7-c2d8bd07bf4d\") " pod="openstack/dnsmasq-dns-6b8f87f5c5-79zbp" Oct 08 19:33:01 crc kubenswrapper[4988]: I1008 19:33:01.511706 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ed8779b3-c8aa-4a12-90db-e0f09722d942-config\") pod \"dnsmasq-dns-678578b8df-87kwv\" (UID: \"ed8779b3-c8aa-4a12-90db-e0f09722d942\") " pod="openstack/dnsmasq-dns-678578b8df-87kwv" Oct 08 19:33:01 crc kubenswrapper[4988]: I1008 19:33:01.511741 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f160a4ae-d7d0-4f4d-89a7-c2d8bd07bf4d-dns-svc\") pod \"dnsmasq-dns-6b8f87f5c5-79zbp\" (UID: \"f160a4ae-d7d0-4f4d-89a7-c2d8bd07bf4d\") " pod="openstack/dnsmasq-dns-6b8f87f5c5-79zbp" Oct 08 19:33:01 crc kubenswrapper[4988]: I1008 19:33:01.512602 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ed8779b3-c8aa-4a12-90db-e0f09722d942-config\") pod \"dnsmasq-dns-678578b8df-87kwv\" (UID: \"ed8779b3-c8aa-4a12-90db-e0f09722d942\") " pod="openstack/dnsmasq-dns-678578b8df-87kwv" Oct 08 19:33:01 crc kubenswrapper[4988]: I1008 19:33:01.613307 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f160a4ae-d7d0-4f4d-89a7-c2d8bd07bf4d-config\") pod \"dnsmasq-dns-6b8f87f5c5-79zbp\" (UID: \"f160a4ae-d7d0-4f4d-89a7-c2d8bd07bf4d\") " pod="openstack/dnsmasq-dns-6b8f87f5c5-79zbp" Oct 08 19:33:01 crc kubenswrapper[4988]: I1008 19:33:01.613367 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s5qdw\" (UniqueName: \"kubernetes.io/projected/f160a4ae-d7d0-4f4d-89a7-c2d8bd07bf4d-kube-api-access-s5qdw\") pod \"dnsmasq-dns-6b8f87f5c5-79zbp\" (UID: \"f160a4ae-d7d0-4f4d-89a7-c2d8bd07bf4d\") " pod="openstack/dnsmasq-dns-6b8f87f5c5-79zbp" Oct 08 19:33:01 crc kubenswrapper[4988]: I1008 19:33:01.613421 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f160a4ae-d7d0-4f4d-89a7-c2d8bd07bf4d-dns-svc\") pod \"dnsmasq-dns-6b8f87f5c5-79zbp\" (UID: \"f160a4ae-d7d0-4f4d-89a7-c2d8bd07bf4d\") " pod="openstack/dnsmasq-dns-6b8f87f5c5-79zbp" Oct 08 19:33:01 crc kubenswrapper[4988]: I1008 19:33:01.614283 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f160a4ae-d7d0-4f4d-89a7-c2d8bd07bf4d-dns-svc\") pod \"dnsmasq-dns-6b8f87f5c5-79zbp\" (UID: \"f160a4ae-d7d0-4f4d-89a7-c2d8bd07bf4d\") " pod="openstack/dnsmasq-dns-6b8f87f5c5-79zbp" Oct 08 19:33:01 crc kubenswrapper[4988]: I1008 19:33:01.614561 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f160a4ae-d7d0-4f4d-89a7-c2d8bd07bf4d-config\") pod \"dnsmasq-dns-6b8f87f5c5-79zbp\" (UID: \"f160a4ae-d7d0-4f4d-89a7-c2d8bd07bf4d\") " pod="openstack/dnsmasq-dns-6b8f87f5c5-79zbp" Oct 08 19:33:01 crc kubenswrapper[4988]: I1008 19:33:01.680352 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s5qdw\" (UniqueName: \"kubernetes.io/projected/f160a4ae-d7d0-4f4d-89a7-c2d8bd07bf4d-kube-api-access-s5qdw\") pod \"dnsmasq-dns-6b8f87f5c5-79zbp\" (UID: \"f160a4ae-d7d0-4f4d-89a7-c2d8bd07bf4d\") " pod="openstack/dnsmasq-dns-6b8f87f5c5-79zbp" Oct 08 19:33:01 crc kubenswrapper[4988]: I1008 19:33:01.693065 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qwcqh\" (UniqueName: \"kubernetes.io/projected/ed8779b3-c8aa-4a12-90db-e0f09722d942-kube-api-access-qwcqh\") pod \"dnsmasq-dns-678578b8df-87kwv\" (UID: \"ed8779b3-c8aa-4a12-90db-e0f09722d942\") " pod="openstack/dnsmasq-dns-678578b8df-87kwv" Oct 08 19:33:01 crc kubenswrapper[4988]: I1008 19:33:01.715808 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-678578b8df-87kwv" Oct 08 19:33:01 crc kubenswrapper[4988]: I1008 19:33:01.729523 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6b8f87f5c5-79zbp" Oct 08 19:33:02 crc kubenswrapper[4988]: I1008 19:33:02.050120 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-678578b8df-87kwv"] Oct 08 19:33:02 crc kubenswrapper[4988]: I1008 19:33:02.075683 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-85f98b87f9-gvdgx"] Oct 08 19:33:02 crc kubenswrapper[4988]: I1008 19:33:02.078243 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-85f98b87f9-gvdgx" Oct 08 19:33:02 crc kubenswrapper[4988]: I1008 19:33:02.090361 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-85f98b87f9-gvdgx"] Oct 08 19:33:02 crc kubenswrapper[4988]: I1008 19:33:02.190457 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-678578b8df-87kwv"] Oct 08 19:33:02 crc kubenswrapper[4988]: I1008 19:33:02.221249 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a63f21fb-1703-48d0-b2f3-2f9aaa3b0e95-config\") pod \"dnsmasq-dns-85f98b87f9-gvdgx\" (UID: \"a63f21fb-1703-48d0-b2f3-2f9aaa3b0e95\") " pod="openstack/dnsmasq-dns-85f98b87f9-gvdgx" Oct 08 19:33:02 crc kubenswrapper[4988]: I1008 19:33:02.221322 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a63f21fb-1703-48d0-b2f3-2f9aaa3b0e95-dns-svc\") pod \"dnsmasq-dns-85f98b87f9-gvdgx\" (UID: \"a63f21fb-1703-48d0-b2f3-2f9aaa3b0e95\") " pod="openstack/dnsmasq-dns-85f98b87f9-gvdgx" Oct 08 19:33:02 crc kubenswrapper[4988]: I1008 19:33:02.221374 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nzwq2\" (UniqueName: \"kubernetes.io/projected/a63f21fb-1703-48d0-b2f3-2f9aaa3b0e95-kube-api-access-nzwq2\") pod \"dnsmasq-dns-85f98b87f9-gvdgx\" (UID: \"a63f21fb-1703-48d0-b2f3-2f9aaa3b0e95\") " pod="openstack/dnsmasq-dns-85f98b87f9-gvdgx" Oct 08 19:33:02 crc kubenswrapper[4988]: W1008 19:33:02.299351 4988 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf160a4ae_d7d0_4f4d_89a7_c2d8bd07bf4d.slice/crio-b948f349580df45a21f29d2c07ef1eef512f6599297b50aade002907958c03c4 WatchSource:0}: Error finding container b948f349580df45a21f29d2c07ef1eef512f6599297b50aade002907958c03c4: Status 404 returned error can't find the container with id b948f349580df45a21f29d2c07ef1eef512f6599297b50aade002907958c03c4 Oct 08 19:33:02 crc kubenswrapper[4988]: I1008 19:33:02.299820 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6b8f87f5c5-79zbp"] Oct 08 19:33:02 crc kubenswrapper[4988]: I1008 19:33:02.322533 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a63f21fb-1703-48d0-b2f3-2f9aaa3b0e95-dns-svc\") pod \"dnsmasq-dns-85f98b87f9-gvdgx\" (UID: \"a63f21fb-1703-48d0-b2f3-2f9aaa3b0e95\") " pod="openstack/dnsmasq-dns-85f98b87f9-gvdgx" Oct 08 19:33:02 crc kubenswrapper[4988]: I1008 19:33:02.322663 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nzwq2\" (UniqueName: \"kubernetes.io/projected/a63f21fb-1703-48d0-b2f3-2f9aaa3b0e95-kube-api-access-nzwq2\") pod \"dnsmasq-dns-85f98b87f9-gvdgx\" (UID: \"a63f21fb-1703-48d0-b2f3-2f9aaa3b0e95\") " pod="openstack/dnsmasq-dns-85f98b87f9-gvdgx" Oct 08 19:33:02 crc kubenswrapper[4988]: I1008 19:33:02.322782 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a63f21fb-1703-48d0-b2f3-2f9aaa3b0e95-config\") pod \"dnsmasq-dns-85f98b87f9-gvdgx\" (UID: \"a63f21fb-1703-48d0-b2f3-2f9aaa3b0e95\") " pod="openstack/dnsmasq-dns-85f98b87f9-gvdgx" Oct 08 19:33:02 crc kubenswrapper[4988]: I1008 19:33:02.323970 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a63f21fb-1703-48d0-b2f3-2f9aaa3b0e95-config\") pod \"dnsmasq-dns-85f98b87f9-gvdgx\" (UID: \"a63f21fb-1703-48d0-b2f3-2f9aaa3b0e95\") " pod="openstack/dnsmasq-dns-85f98b87f9-gvdgx" Oct 08 19:33:02 crc kubenswrapper[4988]: I1008 19:33:02.324798 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a63f21fb-1703-48d0-b2f3-2f9aaa3b0e95-dns-svc\") pod \"dnsmasq-dns-85f98b87f9-gvdgx\" (UID: \"a63f21fb-1703-48d0-b2f3-2f9aaa3b0e95\") " pod="openstack/dnsmasq-dns-85f98b87f9-gvdgx" Oct 08 19:33:02 crc kubenswrapper[4988]: I1008 19:33:02.346225 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nzwq2\" (UniqueName: \"kubernetes.io/projected/a63f21fb-1703-48d0-b2f3-2f9aaa3b0e95-kube-api-access-nzwq2\") pod \"dnsmasq-dns-85f98b87f9-gvdgx\" (UID: \"a63f21fb-1703-48d0-b2f3-2f9aaa3b0e95\") " pod="openstack/dnsmasq-dns-85f98b87f9-gvdgx" Oct 08 19:33:02 crc kubenswrapper[4988]: I1008 19:33:02.356716 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6b8f87f5c5-79zbp"] Oct 08 19:33:02 crc kubenswrapper[4988]: I1008 19:33:02.384180 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-67d9f7fb89-lf6gm"] Oct 08 19:33:02 crc kubenswrapper[4988]: I1008 19:33:02.385632 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-67d9f7fb89-lf6gm" Oct 08 19:33:02 crc kubenswrapper[4988]: I1008 19:33:02.396046 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-67d9f7fb89-lf6gm"] Oct 08 19:33:02 crc kubenswrapper[4988]: I1008 19:33:02.403034 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-85f98b87f9-gvdgx" Oct 08 19:33:02 crc kubenswrapper[4988]: I1008 19:33:02.433512 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-klrcv\" (UniqueName: \"kubernetes.io/projected/fa2476ec-d5d4-49e5-8ef0-8ae8ccfa3993-kube-api-access-klrcv\") pod \"dnsmasq-dns-67d9f7fb89-lf6gm\" (UID: \"fa2476ec-d5d4-49e5-8ef0-8ae8ccfa3993\") " pod="openstack/dnsmasq-dns-67d9f7fb89-lf6gm" Oct 08 19:33:02 crc kubenswrapper[4988]: I1008 19:33:02.433595 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/fa2476ec-d5d4-49e5-8ef0-8ae8ccfa3993-dns-svc\") pod \"dnsmasq-dns-67d9f7fb89-lf6gm\" (UID: \"fa2476ec-d5d4-49e5-8ef0-8ae8ccfa3993\") " pod="openstack/dnsmasq-dns-67d9f7fb89-lf6gm" Oct 08 19:33:02 crc kubenswrapper[4988]: I1008 19:33:02.433686 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fa2476ec-d5d4-49e5-8ef0-8ae8ccfa3993-config\") pod \"dnsmasq-dns-67d9f7fb89-lf6gm\" (UID: \"fa2476ec-d5d4-49e5-8ef0-8ae8ccfa3993\") " pod="openstack/dnsmasq-dns-67d9f7fb89-lf6gm" Oct 08 19:33:02 crc kubenswrapper[4988]: I1008 19:33:02.534483 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-klrcv\" (UniqueName: \"kubernetes.io/projected/fa2476ec-d5d4-49e5-8ef0-8ae8ccfa3993-kube-api-access-klrcv\") pod \"dnsmasq-dns-67d9f7fb89-lf6gm\" (UID: \"fa2476ec-d5d4-49e5-8ef0-8ae8ccfa3993\") " pod="openstack/dnsmasq-dns-67d9f7fb89-lf6gm" Oct 08 19:33:02 crc kubenswrapper[4988]: I1008 19:33:02.534555 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/fa2476ec-d5d4-49e5-8ef0-8ae8ccfa3993-dns-svc\") pod \"dnsmasq-dns-67d9f7fb89-lf6gm\" (UID: \"fa2476ec-d5d4-49e5-8ef0-8ae8ccfa3993\") " pod="openstack/dnsmasq-dns-67d9f7fb89-lf6gm" Oct 08 19:33:02 crc kubenswrapper[4988]: I1008 19:33:02.534587 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fa2476ec-d5d4-49e5-8ef0-8ae8ccfa3993-config\") pod \"dnsmasq-dns-67d9f7fb89-lf6gm\" (UID: \"fa2476ec-d5d4-49e5-8ef0-8ae8ccfa3993\") " pod="openstack/dnsmasq-dns-67d9f7fb89-lf6gm" Oct 08 19:33:02 crc kubenswrapper[4988]: I1008 19:33:02.535483 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fa2476ec-d5d4-49e5-8ef0-8ae8ccfa3993-config\") pod \"dnsmasq-dns-67d9f7fb89-lf6gm\" (UID: \"fa2476ec-d5d4-49e5-8ef0-8ae8ccfa3993\") " pod="openstack/dnsmasq-dns-67d9f7fb89-lf6gm" Oct 08 19:33:02 crc kubenswrapper[4988]: I1008 19:33:02.536612 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/fa2476ec-d5d4-49e5-8ef0-8ae8ccfa3993-dns-svc\") pod \"dnsmasq-dns-67d9f7fb89-lf6gm\" (UID: \"fa2476ec-d5d4-49e5-8ef0-8ae8ccfa3993\") " pod="openstack/dnsmasq-dns-67d9f7fb89-lf6gm" Oct 08 19:33:02 crc kubenswrapper[4988]: I1008 19:33:02.554153 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-klrcv\" (UniqueName: \"kubernetes.io/projected/fa2476ec-d5d4-49e5-8ef0-8ae8ccfa3993-kube-api-access-klrcv\") pod \"dnsmasq-dns-67d9f7fb89-lf6gm\" (UID: \"fa2476ec-d5d4-49e5-8ef0-8ae8ccfa3993\") " pod="openstack/dnsmasq-dns-67d9f7fb89-lf6gm" Oct 08 19:33:02 crc kubenswrapper[4988]: I1008 19:33:02.610509 4988 generic.go:334] "Generic (PLEG): container finished" podID="ed8779b3-c8aa-4a12-90db-e0f09722d942" containerID="fa36918e83b554078c6e8798f77e08f6e2cfea53a13d2d6c7ce66c679009747c" exitCode=0 Oct 08 19:33:02 crc kubenswrapper[4988]: I1008 19:33:02.610580 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-678578b8df-87kwv" event={"ID":"ed8779b3-c8aa-4a12-90db-e0f09722d942","Type":"ContainerDied","Data":"fa36918e83b554078c6e8798f77e08f6e2cfea53a13d2d6c7ce66c679009747c"} Oct 08 19:33:02 crc kubenswrapper[4988]: I1008 19:33:02.610608 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-678578b8df-87kwv" event={"ID":"ed8779b3-c8aa-4a12-90db-e0f09722d942","Type":"ContainerStarted","Data":"be4ea4be85ba1d831ca897cc7f86b6fade22fe3639e9643fe70f1d5e554cab2f"} Oct 08 19:33:02 crc kubenswrapper[4988]: I1008 19:33:02.616532 4988 generic.go:334] "Generic (PLEG): container finished" podID="f160a4ae-d7d0-4f4d-89a7-c2d8bd07bf4d" containerID="60f208a89fbe4889fd85f95fb5d29fbb0a7974e56e041ffb5b9b0a2cb4b12506" exitCode=0 Oct 08 19:33:02 crc kubenswrapper[4988]: I1008 19:33:02.616576 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6b8f87f5c5-79zbp" event={"ID":"f160a4ae-d7d0-4f4d-89a7-c2d8bd07bf4d","Type":"ContainerDied","Data":"60f208a89fbe4889fd85f95fb5d29fbb0a7974e56e041ffb5b9b0a2cb4b12506"} Oct 08 19:33:02 crc kubenswrapper[4988]: I1008 19:33:02.616605 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6b8f87f5c5-79zbp" event={"ID":"f160a4ae-d7d0-4f4d-89a7-c2d8bd07bf4d","Type":"ContainerStarted","Data":"b948f349580df45a21f29d2c07ef1eef512f6599297b50aade002907958c03c4"} Oct 08 19:33:02 crc kubenswrapper[4988]: I1008 19:33:02.740908 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-67d9f7fb89-lf6gm" Oct 08 19:33:02 crc kubenswrapper[4988]: I1008 19:33:02.837522 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-85f98b87f9-gvdgx"] Oct 08 19:33:03 crc kubenswrapper[4988]: I1008 19:33:03.032726 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6b8f87f5c5-79zbp" Oct 08 19:33:03 crc kubenswrapper[4988]: I1008 19:33:03.056496 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-678578b8df-87kwv" Oct 08 19:33:03 crc kubenswrapper[4988]: I1008 19:33:03.146509 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f160a4ae-d7d0-4f4d-89a7-c2d8bd07bf4d-dns-svc\") pod \"f160a4ae-d7d0-4f4d-89a7-c2d8bd07bf4d\" (UID: \"f160a4ae-d7d0-4f4d-89a7-c2d8bd07bf4d\") " Oct 08 19:33:03 crc kubenswrapper[4988]: I1008 19:33:03.146844 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f160a4ae-d7d0-4f4d-89a7-c2d8bd07bf4d-config\") pod \"f160a4ae-d7d0-4f4d-89a7-c2d8bd07bf4d\" (UID: \"f160a4ae-d7d0-4f4d-89a7-c2d8bd07bf4d\") " Oct 08 19:33:03 crc kubenswrapper[4988]: I1008 19:33:03.146868 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ed8779b3-c8aa-4a12-90db-e0f09722d942-config\") pod \"ed8779b3-c8aa-4a12-90db-e0f09722d942\" (UID: \"ed8779b3-c8aa-4a12-90db-e0f09722d942\") " Oct 08 19:33:03 crc kubenswrapper[4988]: I1008 19:33:03.146917 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s5qdw\" (UniqueName: \"kubernetes.io/projected/f160a4ae-d7d0-4f4d-89a7-c2d8bd07bf4d-kube-api-access-s5qdw\") pod \"f160a4ae-d7d0-4f4d-89a7-c2d8bd07bf4d\" (UID: \"f160a4ae-d7d0-4f4d-89a7-c2d8bd07bf4d\") " Oct 08 19:33:03 crc kubenswrapper[4988]: I1008 19:33:03.150614 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f160a4ae-d7d0-4f4d-89a7-c2d8bd07bf4d-kube-api-access-s5qdw" (OuterVolumeSpecName: "kube-api-access-s5qdw") pod "f160a4ae-d7d0-4f4d-89a7-c2d8bd07bf4d" (UID: "f160a4ae-d7d0-4f4d-89a7-c2d8bd07bf4d"). InnerVolumeSpecName "kube-api-access-s5qdw". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 19:33:03 crc kubenswrapper[4988]: I1008 19:33:03.161713 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f160a4ae-d7d0-4f4d-89a7-c2d8bd07bf4d-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "f160a4ae-d7d0-4f4d-89a7-c2d8bd07bf4d" (UID: "f160a4ae-d7d0-4f4d-89a7-c2d8bd07bf4d"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 19:33:03 crc kubenswrapper[4988]: I1008 19:33:03.163323 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ed8779b3-c8aa-4a12-90db-e0f09722d942-config" (OuterVolumeSpecName: "config") pod "ed8779b3-c8aa-4a12-90db-e0f09722d942" (UID: "ed8779b3-c8aa-4a12-90db-e0f09722d942"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 19:33:03 crc kubenswrapper[4988]: I1008 19:33:03.163882 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f160a4ae-d7d0-4f4d-89a7-c2d8bd07bf4d-config" (OuterVolumeSpecName: "config") pod "f160a4ae-d7d0-4f4d-89a7-c2d8bd07bf4d" (UID: "f160a4ae-d7d0-4f4d-89a7-c2d8bd07bf4d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 19:33:03 crc kubenswrapper[4988]: I1008 19:33:03.223600 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-0"] Oct 08 19:33:03 crc kubenswrapper[4988]: E1008 19:33:03.223883 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ed8779b3-c8aa-4a12-90db-e0f09722d942" containerName="init" Oct 08 19:33:03 crc kubenswrapper[4988]: I1008 19:33:03.223898 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="ed8779b3-c8aa-4a12-90db-e0f09722d942" containerName="init" Oct 08 19:33:03 crc kubenswrapper[4988]: E1008 19:33:03.223913 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f160a4ae-d7d0-4f4d-89a7-c2d8bd07bf4d" containerName="init" Oct 08 19:33:03 crc kubenswrapper[4988]: I1008 19:33:03.223919 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="f160a4ae-d7d0-4f4d-89a7-c2d8bd07bf4d" containerName="init" Oct 08 19:33:03 crc kubenswrapper[4988]: I1008 19:33:03.224041 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="f160a4ae-d7d0-4f4d-89a7-c2d8bd07bf4d" containerName="init" Oct 08 19:33:03 crc kubenswrapper[4988]: I1008 19:33:03.224055 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="ed8779b3-c8aa-4a12-90db-e0f09722d942" containerName="init" Oct 08 19:33:03 crc kubenswrapper[4988]: I1008 19:33:03.224792 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Oct 08 19:33:03 crc kubenswrapper[4988]: I1008 19:33:03.227019 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-erlang-cookie" Oct 08 19:33:03 crc kubenswrapper[4988]: I1008 19:33:03.227055 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-default-user" Oct 08 19:33:03 crc kubenswrapper[4988]: I1008 19:33:03.227666 4988 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-server-conf" Oct 08 19:33:03 crc kubenswrapper[4988]: I1008 19:33:03.228368 4988 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-config-data" Oct 08 19:33:03 crc kubenswrapper[4988]: I1008 19:33:03.228453 4988 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-plugins-conf" Oct 08 19:33:03 crc kubenswrapper[4988]: I1008 19:33:03.229845 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-svc" Oct 08 19:33:03 crc kubenswrapper[4988]: I1008 19:33:03.231174 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-server-dockercfg-djzh7" Oct 08 19:33:03 crc kubenswrapper[4988]: I1008 19:33:03.248131 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qwcqh\" (UniqueName: \"kubernetes.io/projected/ed8779b3-c8aa-4a12-90db-e0f09722d942-kube-api-access-qwcqh\") pod \"ed8779b3-c8aa-4a12-90db-e0f09722d942\" (UID: \"ed8779b3-c8aa-4a12-90db-e0f09722d942\") " Oct 08 19:33:03 crc kubenswrapper[4988]: I1008 19:33:03.248577 4988 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f160a4ae-d7d0-4f4d-89a7-c2d8bd07bf4d-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 08 19:33:03 crc kubenswrapper[4988]: I1008 19:33:03.248595 4988 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f160a4ae-d7d0-4f4d-89a7-c2d8bd07bf4d-config\") on node \"crc\" DevicePath \"\"" Oct 08 19:33:03 crc kubenswrapper[4988]: I1008 19:33:03.254492 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ed8779b3-c8aa-4a12-90db-e0f09722d942-kube-api-access-qwcqh" (OuterVolumeSpecName: "kube-api-access-qwcqh") pod "ed8779b3-c8aa-4a12-90db-e0f09722d942" (UID: "ed8779b3-c8aa-4a12-90db-e0f09722d942"). InnerVolumeSpecName "kube-api-access-qwcqh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 19:33:03 crc kubenswrapper[4988]: I1008 19:33:03.254847 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-67d9f7fb89-lf6gm"] Oct 08 19:33:03 crc kubenswrapper[4988]: I1008 19:33:03.254891 4988 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ed8779b3-c8aa-4a12-90db-e0f09722d942-config\") on node \"crc\" DevicePath \"\"" Oct 08 19:33:03 crc kubenswrapper[4988]: I1008 19:33:03.256367 4988 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s5qdw\" (UniqueName: \"kubernetes.io/projected/f160a4ae-d7d0-4f4d-89a7-c2d8bd07bf4d-kube-api-access-s5qdw\") on node \"crc\" DevicePath \"\"" Oct 08 19:33:03 crc kubenswrapper[4988]: I1008 19:33:03.282953 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 08 19:33:03 crc kubenswrapper[4988]: I1008 19:33:03.357650 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/c754aa11-b846-4860-a9c7-9e3a12be054e-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"c754aa11-b846-4860-a9c7-9e3a12be054e\") " pod="openstack/rabbitmq-server-0" Oct 08 19:33:03 crc kubenswrapper[4988]: I1008 19:33:03.357712 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/c754aa11-b846-4860-a9c7-9e3a12be054e-pod-info\") pod \"rabbitmq-server-0\" (UID: \"c754aa11-b846-4860-a9c7-9e3a12be054e\") " pod="openstack/rabbitmq-server-0" Oct 08 19:33:03 crc kubenswrapper[4988]: I1008 19:33:03.357732 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n4gpj\" (UniqueName: \"kubernetes.io/projected/c754aa11-b846-4860-a9c7-9e3a12be054e-kube-api-access-n4gpj\") pod \"rabbitmq-server-0\" (UID: \"c754aa11-b846-4860-a9c7-9e3a12be054e\") " pod="openstack/rabbitmq-server-0" Oct 08 19:33:03 crc kubenswrapper[4988]: I1008 19:33:03.357793 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/c754aa11-b846-4860-a9c7-9e3a12be054e-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"c754aa11-b846-4860-a9c7-9e3a12be054e\") " pod="openstack/rabbitmq-server-0" Oct 08 19:33:03 crc kubenswrapper[4988]: I1008 19:33:03.357810 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/c754aa11-b846-4860-a9c7-9e3a12be054e-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"c754aa11-b846-4860-a9c7-9e3a12be054e\") " pod="openstack/rabbitmq-server-0" Oct 08 19:33:03 crc kubenswrapper[4988]: I1008 19:33:03.357933 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/c754aa11-b846-4860-a9c7-9e3a12be054e-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"c754aa11-b846-4860-a9c7-9e3a12be054e\") " pod="openstack/rabbitmq-server-0" Oct 08 19:33:03 crc kubenswrapper[4988]: I1008 19:33:03.357982 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-a769d776-019d-49d3-8817-18554b6efd94\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-a769d776-019d-49d3-8817-18554b6efd94\") pod \"rabbitmq-server-0\" (UID: \"c754aa11-b846-4860-a9c7-9e3a12be054e\") " pod="openstack/rabbitmq-server-0" Oct 08 19:33:03 crc kubenswrapper[4988]: I1008 19:33:03.358055 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/c754aa11-b846-4860-a9c7-9e3a12be054e-server-conf\") pod \"rabbitmq-server-0\" (UID: \"c754aa11-b846-4860-a9c7-9e3a12be054e\") " pod="openstack/rabbitmq-server-0" Oct 08 19:33:03 crc kubenswrapper[4988]: I1008 19:33:03.358122 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/c754aa11-b846-4860-a9c7-9e3a12be054e-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"c754aa11-b846-4860-a9c7-9e3a12be054e\") " pod="openstack/rabbitmq-server-0" Oct 08 19:33:03 crc kubenswrapper[4988]: I1008 19:33:03.358156 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/c754aa11-b846-4860-a9c7-9e3a12be054e-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"c754aa11-b846-4860-a9c7-9e3a12be054e\") " pod="openstack/rabbitmq-server-0" Oct 08 19:33:03 crc kubenswrapper[4988]: I1008 19:33:03.358177 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/c754aa11-b846-4860-a9c7-9e3a12be054e-config-data\") pod \"rabbitmq-server-0\" (UID: \"c754aa11-b846-4860-a9c7-9e3a12be054e\") " pod="openstack/rabbitmq-server-0" Oct 08 19:33:03 crc kubenswrapper[4988]: I1008 19:33:03.358261 4988 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qwcqh\" (UniqueName: \"kubernetes.io/projected/ed8779b3-c8aa-4a12-90db-e0f09722d942-kube-api-access-qwcqh\") on node \"crc\" DevicePath \"\"" Oct 08 19:33:03 crc kubenswrapper[4988]: I1008 19:33:03.459892 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/c754aa11-b846-4860-a9c7-9e3a12be054e-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"c754aa11-b846-4860-a9c7-9e3a12be054e\") " pod="openstack/rabbitmq-server-0" Oct 08 19:33:03 crc kubenswrapper[4988]: I1008 19:33:03.459935 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/c754aa11-b846-4860-a9c7-9e3a12be054e-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"c754aa11-b846-4860-a9c7-9e3a12be054e\") " pod="openstack/rabbitmq-server-0" Oct 08 19:33:03 crc kubenswrapper[4988]: I1008 19:33:03.459969 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/c754aa11-b846-4860-a9c7-9e3a12be054e-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"c754aa11-b846-4860-a9c7-9e3a12be054e\") " pod="openstack/rabbitmq-server-0" Oct 08 19:33:03 crc kubenswrapper[4988]: I1008 19:33:03.459988 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-a769d776-019d-49d3-8817-18554b6efd94\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-a769d776-019d-49d3-8817-18554b6efd94\") pod \"rabbitmq-server-0\" (UID: \"c754aa11-b846-4860-a9c7-9e3a12be054e\") " pod="openstack/rabbitmq-server-0" Oct 08 19:33:03 crc kubenswrapper[4988]: I1008 19:33:03.460022 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/c754aa11-b846-4860-a9c7-9e3a12be054e-server-conf\") pod \"rabbitmq-server-0\" (UID: \"c754aa11-b846-4860-a9c7-9e3a12be054e\") " pod="openstack/rabbitmq-server-0" Oct 08 19:33:03 crc kubenswrapper[4988]: I1008 19:33:03.460053 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/c754aa11-b846-4860-a9c7-9e3a12be054e-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"c754aa11-b846-4860-a9c7-9e3a12be054e\") " pod="openstack/rabbitmq-server-0" Oct 08 19:33:03 crc kubenswrapper[4988]: I1008 19:33:03.460070 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/c754aa11-b846-4860-a9c7-9e3a12be054e-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"c754aa11-b846-4860-a9c7-9e3a12be054e\") " pod="openstack/rabbitmq-server-0" Oct 08 19:33:03 crc kubenswrapper[4988]: I1008 19:33:03.460085 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/c754aa11-b846-4860-a9c7-9e3a12be054e-config-data\") pod \"rabbitmq-server-0\" (UID: \"c754aa11-b846-4860-a9c7-9e3a12be054e\") " pod="openstack/rabbitmq-server-0" Oct 08 19:33:03 crc kubenswrapper[4988]: I1008 19:33:03.460109 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/c754aa11-b846-4860-a9c7-9e3a12be054e-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"c754aa11-b846-4860-a9c7-9e3a12be054e\") " pod="openstack/rabbitmq-server-0" Oct 08 19:33:03 crc kubenswrapper[4988]: I1008 19:33:03.460126 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/c754aa11-b846-4860-a9c7-9e3a12be054e-pod-info\") pod \"rabbitmq-server-0\" (UID: \"c754aa11-b846-4860-a9c7-9e3a12be054e\") " pod="openstack/rabbitmq-server-0" Oct 08 19:33:03 crc kubenswrapper[4988]: I1008 19:33:03.460139 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n4gpj\" (UniqueName: \"kubernetes.io/projected/c754aa11-b846-4860-a9c7-9e3a12be054e-kube-api-access-n4gpj\") pod \"rabbitmq-server-0\" (UID: \"c754aa11-b846-4860-a9c7-9e3a12be054e\") " pod="openstack/rabbitmq-server-0" Oct 08 19:33:03 crc kubenswrapper[4988]: I1008 19:33:03.460733 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/c754aa11-b846-4860-a9c7-9e3a12be054e-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"c754aa11-b846-4860-a9c7-9e3a12be054e\") " pod="openstack/rabbitmq-server-0" Oct 08 19:33:03 crc kubenswrapper[4988]: I1008 19:33:03.461313 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/c754aa11-b846-4860-a9c7-9e3a12be054e-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"c754aa11-b846-4860-a9c7-9e3a12be054e\") " pod="openstack/rabbitmq-server-0" Oct 08 19:33:03 crc kubenswrapper[4988]: I1008 19:33:03.462167 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/c754aa11-b846-4860-a9c7-9e3a12be054e-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"c754aa11-b846-4860-a9c7-9e3a12be054e\") " pod="openstack/rabbitmq-server-0" Oct 08 19:33:03 crc kubenswrapper[4988]: I1008 19:33:03.462602 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/c754aa11-b846-4860-a9c7-9e3a12be054e-config-data\") pod \"rabbitmq-server-0\" (UID: \"c754aa11-b846-4860-a9c7-9e3a12be054e\") " pod="openstack/rabbitmq-server-0" Oct 08 19:33:03 crc kubenswrapper[4988]: I1008 19:33:03.462644 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/c754aa11-b846-4860-a9c7-9e3a12be054e-server-conf\") pod \"rabbitmq-server-0\" (UID: \"c754aa11-b846-4860-a9c7-9e3a12be054e\") " pod="openstack/rabbitmq-server-0" Oct 08 19:33:03 crc kubenswrapper[4988]: I1008 19:33:03.464664 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/c754aa11-b846-4860-a9c7-9e3a12be054e-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"c754aa11-b846-4860-a9c7-9e3a12be054e\") " pod="openstack/rabbitmq-server-0" Oct 08 19:33:03 crc kubenswrapper[4988]: I1008 19:33:03.464960 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/c754aa11-b846-4860-a9c7-9e3a12be054e-pod-info\") pod \"rabbitmq-server-0\" (UID: \"c754aa11-b846-4860-a9c7-9e3a12be054e\") " pod="openstack/rabbitmq-server-0" Oct 08 19:33:03 crc kubenswrapper[4988]: I1008 19:33:03.465445 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/c754aa11-b846-4860-a9c7-9e3a12be054e-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"c754aa11-b846-4860-a9c7-9e3a12be054e\") " pod="openstack/rabbitmq-server-0" Oct 08 19:33:03 crc kubenswrapper[4988]: I1008 19:33:03.465593 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/c754aa11-b846-4860-a9c7-9e3a12be054e-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"c754aa11-b846-4860-a9c7-9e3a12be054e\") " pod="openstack/rabbitmq-server-0" Oct 08 19:33:03 crc kubenswrapper[4988]: I1008 19:33:03.466558 4988 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Oct 08 19:33:03 crc kubenswrapper[4988]: I1008 19:33:03.466590 4988 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-a769d776-019d-49d3-8817-18554b6efd94\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-a769d776-019d-49d3-8817-18554b6efd94\") pod \"rabbitmq-server-0\" (UID: \"c754aa11-b846-4860-a9c7-9e3a12be054e\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/7468fb34cd0cef4ea54b9c37ebe6b505e744898f63936b052a7832fbe4f2b464/globalmount\"" pod="openstack/rabbitmq-server-0" Oct 08 19:33:03 crc kubenswrapper[4988]: I1008 19:33:03.490167 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n4gpj\" (UniqueName: \"kubernetes.io/projected/c754aa11-b846-4860-a9c7-9e3a12be054e-kube-api-access-n4gpj\") pod \"rabbitmq-server-0\" (UID: \"c754aa11-b846-4860-a9c7-9e3a12be054e\") " pod="openstack/rabbitmq-server-0" Oct 08 19:33:03 crc kubenswrapper[4988]: I1008 19:33:03.507573 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-a769d776-019d-49d3-8817-18554b6efd94\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-a769d776-019d-49d3-8817-18554b6efd94\") pod \"rabbitmq-server-0\" (UID: \"c754aa11-b846-4860-a9c7-9e3a12be054e\") " pod="openstack/rabbitmq-server-0" Oct 08 19:33:03 crc kubenswrapper[4988]: I1008 19:33:03.513554 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 08 19:33:03 crc kubenswrapper[4988]: I1008 19:33:03.516537 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Oct 08 19:33:03 crc kubenswrapper[4988]: I1008 19:33:03.522263 4988 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-plugins-conf" Oct 08 19:33:03 crc kubenswrapper[4988]: I1008 19:33:03.522569 4988 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-server-conf" Oct 08 19:33:03 crc kubenswrapper[4988]: I1008 19:33:03.522261 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-default-user" Oct 08 19:33:03 crc kubenswrapper[4988]: I1008 19:33:03.522695 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-cell1-svc" Oct 08 19:33:03 crc kubenswrapper[4988]: I1008 19:33:03.522728 4988 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-config-data" Oct 08 19:33:03 crc kubenswrapper[4988]: I1008 19:33:03.522771 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-server-dockercfg-k7shm" Oct 08 19:33:03 crc kubenswrapper[4988]: I1008 19:33:03.522914 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-erlang-cookie" Oct 08 19:33:03 crc kubenswrapper[4988]: I1008 19:33:03.528542 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 08 19:33:03 crc kubenswrapper[4988]: I1008 19:33:03.547724 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Oct 08 19:33:03 crc kubenswrapper[4988]: I1008 19:33:03.624655 4988 generic.go:334] "Generic (PLEG): container finished" podID="a63f21fb-1703-48d0-b2f3-2f9aaa3b0e95" containerID="57a11453c9957b0585dee8ccb4454ce79d70d58eab6edcd79dd3937bf6d0a59c" exitCode=0 Oct 08 19:33:03 crc kubenswrapper[4988]: I1008 19:33:03.624719 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-85f98b87f9-gvdgx" event={"ID":"a63f21fb-1703-48d0-b2f3-2f9aaa3b0e95","Type":"ContainerDied","Data":"57a11453c9957b0585dee8ccb4454ce79d70d58eab6edcd79dd3937bf6d0a59c"} Oct 08 19:33:03 crc kubenswrapper[4988]: I1008 19:33:03.624762 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-85f98b87f9-gvdgx" event={"ID":"a63f21fb-1703-48d0-b2f3-2f9aaa3b0e95","Type":"ContainerStarted","Data":"8f212e2ec4b4172dfc103dc9399e104dae8f4e7ed97c332127f19f5b1c56689f"} Oct 08 19:33:03 crc kubenswrapper[4988]: I1008 19:33:03.633143 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6b8f87f5c5-79zbp" event={"ID":"f160a4ae-d7d0-4f4d-89a7-c2d8bd07bf4d","Type":"ContainerDied","Data":"b948f349580df45a21f29d2c07ef1eef512f6599297b50aade002907958c03c4"} Oct 08 19:33:03 crc kubenswrapper[4988]: I1008 19:33:03.633190 4988 scope.go:117] "RemoveContainer" containerID="60f208a89fbe4889fd85f95fb5d29fbb0a7974e56e041ffb5b9b0a2cb4b12506" Oct 08 19:33:03 crc kubenswrapper[4988]: I1008 19:33:03.633309 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6b8f87f5c5-79zbp" Oct 08 19:33:03 crc kubenswrapper[4988]: I1008 19:33:03.640050 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-678578b8df-87kwv" event={"ID":"ed8779b3-c8aa-4a12-90db-e0f09722d942","Type":"ContainerDied","Data":"be4ea4be85ba1d831ca897cc7f86b6fade22fe3639e9643fe70f1d5e554cab2f"} Oct 08 19:33:03 crc kubenswrapper[4988]: I1008 19:33:03.640133 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-678578b8df-87kwv" Oct 08 19:33:03 crc kubenswrapper[4988]: I1008 19:33:03.652643 4988 generic.go:334] "Generic (PLEG): container finished" podID="fa2476ec-d5d4-49e5-8ef0-8ae8ccfa3993" containerID="b9e0aee791a48280c1b9ca382d7c0eed368e87e4c349d8ba9e39177851280fac" exitCode=0 Oct 08 19:33:03 crc kubenswrapper[4988]: I1008 19:33:03.652914 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-67d9f7fb89-lf6gm" event={"ID":"fa2476ec-d5d4-49e5-8ef0-8ae8ccfa3993","Type":"ContainerDied","Data":"b9e0aee791a48280c1b9ca382d7c0eed368e87e4c349d8ba9e39177851280fac"} Oct 08 19:33:03 crc kubenswrapper[4988]: I1008 19:33:03.653781 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-67d9f7fb89-lf6gm" event={"ID":"fa2476ec-d5d4-49e5-8ef0-8ae8ccfa3993","Type":"ContainerStarted","Data":"05262201f8b1cab31c28ca9335c65bc376d371025cccabf49b1493da4092d10e"} Oct 08 19:33:03 crc kubenswrapper[4988]: I1008 19:33:03.663524 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/111b6cdd-1e77-4992-a331-c66bc9e95a90-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"111b6cdd-1e77-4992-a331-c66bc9e95a90\") " pod="openstack/rabbitmq-cell1-server-0" Oct 08 19:33:03 crc kubenswrapper[4988]: I1008 19:33:03.663572 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/111b6cdd-1e77-4992-a331-c66bc9e95a90-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"111b6cdd-1e77-4992-a331-c66bc9e95a90\") " pod="openstack/rabbitmq-cell1-server-0" Oct 08 19:33:03 crc kubenswrapper[4988]: I1008 19:33:03.663602 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hqxvn\" (UniqueName: \"kubernetes.io/projected/111b6cdd-1e77-4992-a331-c66bc9e95a90-kube-api-access-hqxvn\") pod \"rabbitmq-cell1-server-0\" (UID: \"111b6cdd-1e77-4992-a331-c66bc9e95a90\") " pod="openstack/rabbitmq-cell1-server-0" Oct 08 19:33:03 crc kubenswrapper[4988]: I1008 19:33:03.663793 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/111b6cdd-1e77-4992-a331-c66bc9e95a90-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"111b6cdd-1e77-4992-a331-c66bc9e95a90\") " pod="openstack/rabbitmq-cell1-server-0" Oct 08 19:33:03 crc kubenswrapper[4988]: I1008 19:33:03.663857 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/111b6cdd-1e77-4992-a331-c66bc9e95a90-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"111b6cdd-1e77-4992-a331-c66bc9e95a90\") " pod="openstack/rabbitmq-cell1-server-0" Oct 08 19:33:03 crc kubenswrapper[4988]: I1008 19:33:03.663890 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-ee25b6c1-7b38-49fa-97b1-836588148f84\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-ee25b6c1-7b38-49fa-97b1-836588148f84\") pod \"rabbitmq-cell1-server-0\" (UID: \"111b6cdd-1e77-4992-a331-c66bc9e95a90\") " pod="openstack/rabbitmq-cell1-server-0" Oct 08 19:33:03 crc kubenswrapper[4988]: I1008 19:33:03.663918 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/111b6cdd-1e77-4992-a331-c66bc9e95a90-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"111b6cdd-1e77-4992-a331-c66bc9e95a90\") " pod="openstack/rabbitmq-cell1-server-0" Oct 08 19:33:03 crc kubenswrapper[4988]: I1008 19:33:03.663970 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/111b6cdd-1e77-4992-a331-c66bc9e95a90-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"111b6cdd-1e77-4992-a331-c66bc9e95a90\") " pod="openstack/rabbitmq-cell1-server-0" Oct 08 19:33:03 crc kubenswrapper[4988]: I1008 19:33:03.663997 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/111b6cdd-1e77-4992-a331-c66bc9e95a90-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"111b6cdd-1e77-4992-a331-c66bc9e95a90\") " pod="openstack/rabbitmq-cell1-server-0" Oct 08 19:33:03 crc kubenswrapper[4988]: I1008 19:33:03.664027 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/111b6cdd-1e77-4992-a331-c66bc9e95a90-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"111b6cdd-1e77-4992-a331-c66bc9e95a90\") " pod="openstack/rabbitmq-cell1-server-0" Oct 08 19:33:03 crc kubenswrapper[4988]: I1008 19:33:03.664051 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/111b6cdd-1e77-4992-a331-c66bc9e95a90-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"111b6cdd-1e77-4992-a331-c66bc9e95a90\") " pod="openstack/rabbitmq-cell1-server-0" Oct 08 19:33:03 crc kubenswrapper[4988]: I1008 19:33:03.702728 4988 scope.go:117] "RemoveContainer" containerID="fa36918e83b554078c6e8798f77e08f6e2cfea53a13d2d6c7ce66c679009747c" Oct 08 19:33:03 crc kubenswrapper[4988]: I1008 19:33:03.721181 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6b8f87f5c5-79zbp"] Oct 08 19:33:03 crc kubenswrapper[4988]: I1008 19:33:03.741488 4988 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-6b8f87f5c5-79zbp"] Oct 08 19:33:03 crc kubenswrapper[4988]: I1008 19:33:03.762887 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-678578b8df-87kwv"] Oct 08 19:33:03 crc kubenswrapper[4988]: I1008 19:33:03.765113 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hqxvn\" (UniqueName: \"kubernetes.io/projected/111b6cdd-1e77-4992-a331-c66bc9e95a90-kube-api-access-hqxvn\") pod \"rabbitmq-cell1-server-0\" (UID: \"111b6cdd-1e77-4992-a331-c66bc9e95a90\") " pod="openstack/rabbitmq-cell1-server-0" Oct 08 19:33:03 crc kubenswrapper[4988]: I1008 19:33:03.765154 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/111b6cdd-1e77-4992-a331-c66bc9e95a90-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"111b6cdd-1e77-4992-a331-c66bc9e95a90\") " pod="openstack/rabbitmq-cell1-server-0" Oct 08 19:33:03 crc kubenswrapper[4988]: I1008 19:33:03.765223 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/111b6cdd-1e77-4992-a331-c66bc9e95a90-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"111b6cdd-1e77-4992-a331-c66bc9e95a90\") " pod="openstack/rabbitmq-cell1-server-0" Oct 08 19:33:03 crc kubenswrapper[4988]: I1008 19:33:03.765248 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-ee25b6c1-7b38-49fa-97b1-836588148f84\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-ee25b6c1-7b38-49fa-97b1-836588148f84\") pod \"rabbitmq-cell1-server-0\" (UID: \"111b6cdd-1e77-4992-a331-c66bc9e95a90\") " pod="openstack/rabbitmq-cell1-server-0" Oct 08 19:33:03 crc kubenswrapper[4988]: I1008 19:33:03.765278 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/111b6cdd-1e77-4992-a331-c66bc9e95a90-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"111b6cdd-1e77-4992-a331-c66bc9e95a90\") " pod="openstack/rabbitmq-cell1-server-0" Oct 08 19:33:03 crc kubenswrapper[4988]: I1008 19:33:03.765330 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/111b6cdd-1e77-4992-a331-c66bc9e95a90-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"111b6cdd-1e77-4992-a331-c66bc9e95a90\") " pod="openstack/rabbitmq-cell1-server-0" Oct 08 19:33:03 crc kubenswrapper[4988]: I1008 19:33:03.765351 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/111b6cdd-1e77-4992-a331-c66bc9e95a90-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"111b6cdd-1e77-4992-a331-c66bc9e95a90\") " pod="openstack/rabbitmq-cell1-server-0" Oct 08 19:33:03 crc kubenswrapper[4988]: I1008 19:33:03.765370 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/111b6cdd-1e77-4992-a331-c66bc9e95a90-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"111b6cdd-1e77-4992-a331-c66bc9e95a90\") " pod="openstack/rabbitmq-cell1-server-0" Oct 08 19:33:03 crc kubenswrapper[4988]: I1008 19:33:03.765400 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/111b6cdd-1e77-4992-a331-c66bc9e95a90-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"111b6cdd-1e77-4992-a331-c66bc9e95a90\") " pod="openstack/rabbitmq-cell1-server-0" Oct 08 19:33:03 crc kubenswrapper[4988]: I1008 19:33:03.765437 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/111b6cdd-1e77-4992-a331-c66bc9e95a90-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"111b6cdd-1e77-4992-a331-c66bc9e95a90\") " pod="openstack/rabbitmq-cell1-server-0" Oct 08 19:33:03 crc kubenswrapper[4988]: I1008 19:33:03.765454 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/111b6cdd-1e77-4992-a331-c66bc9e95a90-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"111b6cdd-1e77-4992-a331-c66bc9e95a90\") " pod="openstack/rabbitmq-cell1-server-0" Oct 08 19:33:03 crc kubenswrapper[4988]: I1008 19:33:03.766695 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/111b6cdd-1e77-4992-a331-c66bc9e95a90-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"111b6cdd-1e77-4992-a331-c66bc9e95a90\") " pod="openstack/rabbitmq-cell1-server-0" Oct 08 19:33:03 crc kubenswrapper[4988]: I1008 19:33:03.766720 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/111b6cdd-1e77-4992-a331-c66bc9e95a90-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"111b6cdd-1e77-4992-a331-c66bc9e95a90\") " pod="openstack/rabbitmq-cell1-server-0" Oct 08 19:33:03 crc kubenswrapper[4988]: I1008 19:33:03.766763 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/111b6cdd-1e77-4992-a331-c66bc9e95a90-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"111b6cdd-1e77-4992-a331-c66bc9e95a90\") " pod="openstack/rabbitmq-cell1-server-0" Oct 08 19:33:03 crc kubenswrapper[4988]: I1008 19:33:03.767534 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/111b6cdd-1e77-4992-a331-c66bc9e95a90-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"111b6cdd-1e77-4992-a331-c66bc9e95a90\") " pod="openstack/rabbitmq-cell1-server-0" Oct 08 19:33:03 crc kubenswrapper[4988]: I1008 19:33:03.774156 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/111b6cdd-1e77-4992-a331-c66bc9e95a90-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"111b6cdd-1e77-4992-a331-c66bc9e95a90\") " pod="openstack/rabbitmq-cell1-server-0" Oct 08 19:33:03 crc kubenswrapper[4988]: I1008 19:33:03.774639 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/111b6cdd-1e77-4992-a331-c66bc9e95a90-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"111b6cdd-1e77-4992-a331-c66bc9e95a90\") " pod="openstack/rabbitmq-cell1-server-0" Oct 08 19:33:03 crc kubenswrapper[4988]: I1008 19:33:03.775178 4988 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Oct 08 19:33:03 crc kubenswrapper[4988]: I1008 19:33:03.775204 4988 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-ee25b6c1-7b38-49fa-97b1-836588148f84\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-ee25b6c1-7b38-49fa-97b1-836588148f84\") pod \"rabbitmq-cell1-server-0\" (UID: \"111b6cdd-1e77-4992-a331-c66bc9e95a90\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1c9155ea9cd9063941368b7285cdba7000bafae9c9b6d32fa454bc55cae46f02/globalmount\"" pod="openstack/rabbitmq-cell1-server-0" Oct 08 19:33:03 crc kubenswrapper[4988]: I1008 19:33:03.775890 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/111b6cdd-1e77-4992-a331-c66bc9e95a90-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"111b6cdd-1e77-4992-a331-c66bc9e95a90\") " pod="openstack/rabbitmq-cell1-server-0" Oct 08 19:33:03 crc kubenswrapper[4988]: I1008 19:33:03.776184 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/111b6cdd-1e77-4992-a331-c66bc9e95a90-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"111b6cdd-1e77-4992-a331-c66bc9e95a90\") " pod="openstack/rabbitmq-cell1-server-0" Oct 08 19:33:03 crc kubenswrapper[4988]: I1008 19:33:03.776307 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/111b6cdd-1e77-4992-a331-c66bc9e95a90-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"111b6cdd-1e77-4992-a331-c66bc9e95a90\") " pod="openstack/rabbitmq-cell1-server-0" Oct 08 19:33:03 crc kubenswrapper[4988]: I1008 19:33:03.778494 4988 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-678578b8df-87kwv"] Oct 08 19:33:03 crc kubenswrapper[4988]: I1008 19:33:03.788332 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hqxvn\" (UniqueName: \"kubernetes.io/projected/111b6cdd-1e77-4992-a331-c66bc9e95a90-kube-api-access-hqxvn\") pod \"rabbitmq-cell1-server-0\" (UID: \"111b6cdd-1e77-4992-a331-c66bc9e95a90\") " pod="openstack/rabbitmq-cell1-server-0" Oct 08 19:33:03 crc kubenswrapper[4988]: I1008 19:33:03.827341 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-ee25b6c1-7b38-49fa-97b1-836588148f84\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-ee25b6c1-7b38-49fa-97b1-836588148f84\") pod \"rabbitmq-cell1-server-0\" (UID: \"111b6cdd-1e77-4992-a331-c66bc9e95a90\") " pod="openstack/rabbitmq-cell1-server-0" Oct 08 19:33:03 crc kubenswrapper[4988]: I1008 19:33:03.856822 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Oct 08 19:33:04 crc kubenswrapper[4988]: I1008 19:33:04.033268 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 08 19:33:04 crc kubenswrapper[4988]: I1008 19:33:04.328706 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 08 19:33:04 crc kubenswrapper[4988]: W1008 19:33:04.331518 4988 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod111b6cdd_1e77_4992_a331_c66bc9e95a90.slice/crio-73c6f78f7d2284833ade391bb3f6d275c9404bd39d59c930323a575ad3e096ad WatchSource:0}: Error finding container 73c6f78f7d2284833ade391bb3f6d275c9404bd39d59c930323a575ad3e096ad: Status 404 returned error can't find the container with id 73c6f78f7d2284833ade391bb3f6d275c9404bd39d59c930323a575ad3e096ad Oct 08 19:33:04 crc kubenswrapper[4988]: I1008 19:33:04.363707 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-galera-0"] Oct 08 19:33:04 crc kubenswrapper[4988]: I1008 19:33:04.364861 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Oct 08 19:33:04 crc kubenswrapper[4988]: I1008 19:33:04.377742 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Oct 08 19:33:04 crc kubenswrapper[4988]: I1008 19:33:04.378297 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-svc" Oct 08 19:33:04 crc kubenswrapper[4988]: I1008 19:33:04.378326 4988 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-scripts" Oct 08 19:33:04 crc kubenswrapper[4988]: I1008 19:33:04.378439 4988 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config-data" Oct 08 19:33:04 crc kubenswrapper[4988]: I1008 19:33:04.378575 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-dockercfg-6glfr" Oct 08 19:33:04 crc kubenswrapper[4988]: I1008 19:33:04.387013 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Oct 08 19:33:04 crc kubenswrapper[4988]: I1008 19:33:04.401344 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"combined-ca-bundle" Oct 08 19:33:04 crc kubenswrapper[4988]: I1008 19:33:04.476064 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/689b2a7d-38cb-4476-a0ea-27bbd614fb2b-secrets\") pod \"openstack-galera-0\" (UID: \"689b2a7d-38cb-4476-a0ea-27bbd614fb2b\") " pod="openstack/openstack-galera-0" Oct 08 19:33:04 crc kubenswrapper[4988]: I1008 19:33:04.476335 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-b1918327-07b3-4af6-acef-5e39d1edde8c\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b1918327-07b3-4af6-acef-5e39d1edde8c\") pod \"openstack-galera-0\" (UID: \"689b2a7d-38cb-4476-a0ea-27bbd614fb2b\") " pod="openstack/openstack-galera-0" Oct 08 19:33:04 crc kubenswrapper[4988]: I1008 19:33:04.476368 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/689b2a7d-38cb-4476-a0ea-27bbd614fb2b-kolla-config\") pod \"openstack-galera-0\" (UID: \"689b2a7d-38cb-4476-a0ea-27bbd614fb2b\") " pod="openstack/openstack-galera-0" Oct 08 19:33:04 crc kubenswrapper[4988]: I1008 19:33:04.476411 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/689b2a7d-38cb-4476-a0ea-27bbd614fb2b-operator-scripts\") pod \"openstack-galera-0\" (UID: \"689b2a7d-38cb-4476-a0ea-27bbd614fb2b\") " pod="openstack/openstack-galera-0" Oct 08 19:33:04 crc kubenswrapper[4988]: I1008 19:33:04.476441 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/689b2a7d-38cb-4476-a0ea-27bbd614fb2b-config-data-generated\") pod \"openstack-galera-0\" (UID: \"689b2a7d-38cb-4476-a0ea-27bbd614fb2b\") " pod="openstack/openstack-galera-0" Oct 08 19:33:04 crc kubenswrapper[4988]: I1008 19:33:04.476615 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/689b2a7d-38cb-4476-a0ea-27bbd614fb2b-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"689b2a7d-38cb-4476-a0ea-27bbd614fb2b\") " pod="openstack/openstack-galera-0" Oct 08 19:33:04 crc kubenswrapper[4988]: I1008 19:33:04.476677 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/689b2a7d-38cb-4476-a0ea-27bbd614fb2b-config-data-default\") pod \"openstack-galera-0\" (UID: \"689b2a7d-38cb-4476-a0ea-27bbd614fb2b\") " pod="openstack/openstack-galera-0" Oct 08 19:33:04 crc kubenswrapper[4988]: I1008 19:33:04.476719 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/689b2a7d-38cb-4476-a0ea-27bbd614fb2b-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"689b2a7d-38cb-4476-a0ea-27bbd614fb2b\") " pod="openstack/openstack-galera-0" Oct 08 19:33:04 crc kubenswrapper[4988]: I1008 19:33:04.476768 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8lp5b\" (UniqueName: \"kubernetes.io/projected/689b2a7d-38cb-4476-a0ea-27bbd614fb2b-kube-api-access-8lp5b\") pod \"openstack-galera-0\" (UID: \"689b2a7d-38cb-4476-a0ea-27bbd614fb2b\") " pod="openstack/openstack-galera-0" Oct 08 19:33:04 crc kubenswrapper[4988]: I1008 19:33:04.577762 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/689b2a7d-38cb-4476-a0ea-27bbd614fb2b-operator-scripts\") pod \"openstack-galera-0\" (UID: \"689b2a7d-38cb-4476-a0ea-27bbd614fb2b\") " pod="openstack/openstack-galera-0" Oct 08 19:33:04 crc kubenswrapper[4988]: I1008 19:33:04.577829 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/689b2a7d-38cb-4476-a0ea-27bbd614fb2b-config-data-generated\") pod \"openstack-galera-0\" (UID: \"689b2a7d-38cb-4476-a0ea-27bbd614fb2b\") " pod="openstack/openstack-galera-0" Oct 08 19:33:04 crc kubenswrapper[4988]: I1008 19:33:04.577906 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/689b2a7d-38cb-4476-a0ea-27bbd614fb2b-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"689b2a7d-38cb-4476-a0ea-27bbd614fb2b\") " pod="openstack/openstack-galera-0" Oct 08 19:33:04 crc kubenswrapper[4988]: I1008 19:33:04.577936 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/689b2a7d-38cb-4476-a0ea-27bbd614fb2b-config-data-default\") pod \"openstack-galera-0\" (UID: \"689b2a7d-38cb-4476-a0ea-27bbd614fb2b\") " pod="openstack/openstack-galera-0" Oct 08 19:33:04 crc kubenswrapper[4988]: I1008 19:33:04.577965 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/689b2a7d-38cb-4476-a0ea-27bbd614fb2b-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"689b2a7d-38cb-4476-a0ea-27bbd614fb2b\") " pod="openstack/openstack-galera-0" Oct 08 19:33:04 crc kubenswrapper[4988]: I1008 19:33:04.577996 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8lp5b\" (UniqueName: \"kubernetes.io/projected/689b2a7d-38cb-4476-a0ea-27bbd614fb2b-kube-api-access-8lp5b\") pod \"openstack-galera-0\" (UID: \"689b2a7d-38cb-4476-a0ea-27bbd614fb2b\") " pod="openstack/openstack-galera-0" Oct 08 19:33:04 crc kubenswrapper[4988]: I1008 19:33:04.578036 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/689b2a7d-38cb-4476-a0ea-27bbd614fb2b-secrets\") pod \"openstack-galera-0\" (UID: \"689b2a7d-38cb-4476-a0ea-27bbd614fb2b\") " pod="openstack/openstack-galera-0" Oct 08 19:33:04 crc kubenswrapper[4988]: I1008 19:33:04.578080 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-b1918327-07b3-4af6-acef-5e39d1edde8c\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b1918327-07b3-4af6-acef-5e39d1edde8c\") pod \"openstack-galera-0\" (UID: \"689b2a7d-38cb-4476-a0ea-27bbd614fb2b\") " pod="openstack/openstack-galera-0" Oct 08 19:33:04 crc kubenswrapper[4988]: I1008 19:33:04.578117 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/689b2a7d-38cb-4476-a0ea-27bbd614fb2b-kolla-config\") pod \"openstack-galera-0\" (UID: \"689b2a7d-38cb-4476-a0ea-27bbd614fb2b\") " pod="openstack/openstack-galera-0" Oct 08 19:33:04 crc kubenswrapper[4988]: I1008 19:33:04.578319 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/689b2a7d-38cb-4476-a0ea-27bbd614fb2b-config-data-generated\") pod \"openstack-galera-0\" (UID: \"689b2a7d-38cb-4476-a0ea-27bbd614fb2b\") " pod="openstack/openstack-galera-0" Oct 08 19:33:04 crc kubenswrapper[4988]: I1008 19:33:04.578860 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/689b2a7d-38cb-4476-a0ea-27bbd614fb2b-config-data-default\") pod \"openstack-galera-0\" (UID: \"689b2a7d-38cb-4476-a0ea-27bbd614fb2b\") " pod="openstack/openstack-galera-0" Oct 08 19:33:04 crc kubenswrapper[4988]: I1008 19:33:04.578972 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/689b2a7d-38cb-4476-a0ea-27bbd614fb2b-kolla-config\") pod \"openstack-galera-0\" (UID: \"689b2a7d-38cb-4476-a0ea-27bbd614fb2b\") " pod="openstack/openstack-galera-0" Oct 08 19:33:04 crc kubenswrapper[4988]: I1008 19:33:04.579710 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/689b2a7d-38cb-4476-a0ea-27bbd614fb2b-operator-scripts\") pod \"openstack-galera-0\" (UID: \"689b2a7d-38cb-4476-a0ea-27bbd614fb2b\") " pod="openstack/openstack-galera-0" Oct 08 19:33:04 crc kubenswrapper[4988]: I1008 19:33:04.580833 4988 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Oct 08 19:33:04 crc kubenswrapper[4988]: I1008 19:33:04.580871 4988 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-b1918327-07b3-4af6-acef-5e39d1edde8c\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b1918327-07b3-4af6-acef-5e39d1edde8c\") pod \"openstack-galera-0\" (UID: \"689b2a7d-38cb-4476-a0ea-27bbd614fb2b\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/c148bb08cefb2d1b6bac92a69bdc785a5927ea12e6f28236516185bc355211b7/globalmount\"" pod="openstack/openstack-galera-0" Oct 08 19:33:04 crc kubenswrapper[4988]: I1008 19:33:04.659906 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"111b6cdd-1e77-4992-a331-c66bc9e95a90","Type":"ContainerStarted","Data":"73c6f78f7d2284833ade391bb3f6d275c9404bd39d59c930323a575ad3e096ad"} Oct 08 19:33:04 crc kubenswrapper[4988]: I1008 19:33:04.660801 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"c754aa11-b846-4860-a9c7-9e3a12be054e","Type":"ContainerStarted","Data":"409780b0c81127be49c2b132287e43ad5221b0ee4c461ceb727454e4e1a41bec"} Oct 08 19:33:04 crc kubenswrapper[4988]: I1008 19:33:04.663655 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-67d9f7fb89-lf6gm" event={"ID":"fa2476ec-d5d4-49e5-8ef0-8ae8ccfa3993","Type":"ContainerStarted","Data":"595680900cdc256c43ae6fa59ae84429550508e7fbf5148fcce9c333973e9db1"} Oct 08 19:33:04 crc kubenswrapper[4988]: I1008 19:33:04.663779 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-67d9f7fb89-lf6gm" Oct 08 19:33:04 crc kubenswrapper[4988]: I1008 19:33:04.665692 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-85f98b87f9-gvdgx" event={"ID":"a63f21fb-1703-48d0-b2f3-2f9aaa3b0e95","Type":"ContainerStarted","Data":"121035b9c5e9e2a7f2ccc2ed29356a72aa767268cf19f124327fddfca3dcf584"} Oct 08 19:33:04 crc kubenswrapper[4988]: I1008 19:33:04.665832 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-85f98b87f9-gvdgx" Oct 08 19:33:04 crc kubenswrapper[4988]: I1008 19:33:04.677907 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/689b2a7d-38cb-4476-a0ea-27bbd614fb2b-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"689b2a7d-38cb-4476-a0ea-27bbd614fb2b\") " pod="openstack/openstack-galera-0" Oct 08 19:33:04 crc kubenswrapper[4988]: I1008 19:33:04.677945 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/689b2a7d-38cb-4476-a0ea-27bbd614fb2b-secrets\") pod \"openstack-galera-0\" (UID: \"689b2a7d-38cb-4476-a0ea-27bbd614fb2b\") " pod="openstack/openstack-galera-0" Oct 08 19:33:04 crc kubenswrapper[4988]: I1008 19:33:04.678287 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/689b2a7d-38cb-4476-a0ea-27bbd614fb2b-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"689b2a7d-38cb-4476-a0ea-27bbd614fb2b\") " pod="openstack/openstack-galera-0" Oct 08 19:33:04 crc kubenswrapper[4988]: I1008 19:33:04.678509 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8lp5b\" (UniqueName: \"kubernetes.io/projected/689b2a7d-38cb-4476-a0ea-27bbd614fb2b-kube-api-access-8lp5b\") pod \"openstack-galera-0\" (UID: \"689b2a7d-38cb-4476-a0ea-27bbd614fb2b\") " pod="openstack/openstack-galera-0" Oct 08 19:33:04 crc kubenswrapper[4988]: I1008 19:33:04.688042 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-67d9f7fb89-lf6gm" podStartSLOduration=2.688020041 podStartE2EDuration="2.688020041s" podCreationTimestamp="2025-10-08 19:33:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 19:33:04.682607429 +0000 UTC m=+4930.132450199" watchObservedRunningTime="2025-10-08 19:33:04.688020041 +0000 UTC m=+4930.137862811" Oct 08 19:33:04 crc kubenswrapper[4988]: I1008 19:33:04.699198 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-85f98b87f9-gvdgx" podStartSLOduration=2.699181706 podStartE2EDuration="2.699181706s" podCreationTimestamp="2025-10-08 19:33:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 19:33:04.695992745 +0000 UTC m=+4930.145835535" watchObservedRunningTime="2025-10-08 19:33:04.699181706 +0000 UTC m=+4930.149024476" Oct 08 19:33:04 crc kubenswrapper[4988]: I1008 19:33:04.928959 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-b1918327-07b3-4af6-acef-5e39d1edde8c\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b1918327-07b3-4af6-acef-5e39d1edde8c\") pod \"openstack-galera-0\" (UID: \"689b2a7d-38cb-4476-a0ea-27bbd614fb2b\") " pod="openstack/openstack-galera-0" Oct 08 19:33:04 crc kubenswrapper[4988]: I1008 19:33:04.988367 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Oct 08 19:33:05 crc kubenswrapper[4988]: I1008 19:33:05.278406 4988 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ed8779b3-c8aa-4a12-90db-e0f09722d942" path="/var/lib/kubelet/pods/ed8779b3-c8aa-4a12-90db-e0f09722d942/volumes" Oct 08 19:33:05 crc kubenswrapper[4988]: I1008 19:33:05.279269 4988 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f160a4ae-d7d0-4f4d-89a7-c2d8bd07bf4d" path="/var/lib/kubelet/pods/f160a4ae-d7d0-4f4d-89a7-c2d8bd07bf4d/volumes" Oct 08 19:33:05 crc kubenswrapper[4988]: I1008 19:33:05.551586 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Oct 08 19:33:05 crc kubenswrapper[4988]: W1008 19:33:05.554869 4988 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod689b2a7d_38cb_4476_a0ea_27bbd614fb2b.slice/crio-62c3fa4608275db772822f47bfd752c511d2289c1daea5f7208e01e183da2f4f WatchSource:0}: Error finding container 62c3fa4608275db772822f47bfd752c511d2289c1daea5f7208e01e183da2f4f: Status 404 returned error can't find the container with id 62c3fa4608275db772822f47bfd752c511d2289c1daea5f7208e01e183da2f4f Oct 08 19:33:05 crc kubenswrapper[4988]: I1008 19:33:05.657381 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-cell1-galera-0"] Oct 08 19:33:05 crc kubenswrapper[4988]: I1008 19:33:05.660005 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Oct 08 19:33:05 crc kubenswrapper[4988]: I1008 19:33:05.664326 4988 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-config-data" Oct 08 19:33:05 crc kubenswrapper[4988]: I1008 19:33:05.664686 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-cell1-svc" Oct 08 19:33:05 crc kubenswrapper[4988]: I1008 19:33:05.664710 4988 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-scripts" Oct 08 19:33:05 crc kubenswrapper[4988]: I1008 19:33:05.665210 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-cell1-dockercfg-4n58l" Oct 08 19:33:05 crc kubenswrapper[4988]: I1008 19:33:05.669503 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Oct 08 19:33:05 crc kubenswrapper[4988]: I1008 19:33:05.690203 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"689b2a7d-38cb-4476-a0ea-27bbd614fb2b","Type":"ContainerStarted","Data":"62c3fa4608275db772822f47bfd752c511d2289c1daea5f7208e01e183da2f4f"} Oct 08 19:33:05 crc kubenswrapper[4988]: I1008 19:33:05.691580 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"111b6cdd-1e77-4992-a331-c66bc9e95a90","Type":"ContainerStarted","Data":"6b0b30e0590363081acab72ac2c5ea5643a1db58b78fa1169b9902c4781121cb"} Oct 08 19:33:05 crc kubenswrapper[4988]: I1008 19:33:05.694316 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"c754aa11-b846-4860-a9c7-9e3a12be054e","Type":"ContainerStarted","Data":"9345a690467ea890e581ad94aa2b8e47eeb1adbd92e069d7d3fad8698f98fc51"} Oct 08 19:33:05 crc kubenswrapper[4988]: I1008 19:33:05.793630 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/bbe1c39b-8f0b-41a4-87a3-20672272af49-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"bbe1c39b-8f0b-41a4-87a3-20672272af49\") " pod="openstack/openstack-cell1-galera-0" Oct 08 19:33:05 crc kubenswrapper[4988]: I1008 19:33:05.793700 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/bbe1c39b-8f0b-41a4-87a3-20672272af49-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"bbe1c39b-8f0b-41a4-87a3-20672272af49\") " pod="openstack/openstack-cell1-galera-0" Oct 08 19:33:05 crc kubenswrapper[4988]: I1008 19:33:05.793799 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/bbe1c39b-8f0b-41a4-87a3-20672272af49-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"bbe1c39b-8f0b-41a4-87a3-20672272af49\") " pod="openstack/openstack-cell1-galera-0" Oct 08 19:33:05 crc kubenswrapper[4988]: I1008 19:33:05.793838 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/bbe1c39b-8f0b-41a4-87a3-20672272af49-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"bbe1c39b-8f0b-41a4-87a3-20672272af49\") " pod="openstack/openstack-cell1-galera-0" Oct 08 19:33:05 crc kubenswrapper[4988]: I1008 19:33:05.793890 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/bbe1c39b-8f0b-41a4-87a3-20672272af49-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"bbe1c39b-8f0b-41a4-87a3-20672272af49\") " pod="openstack/openstack-cell1-galera-0" Oct 08 19:33:05 crc kubenswrapper[4988]: I1008 19:33:05.793906 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bbe1c39b-8f0b-41a4-87a3-20672272af49-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"bbe1c39b-8f0b-41a4-87a3-20672272af49\") " pod="openstack/openstack-cell1-galera-0" Oct 08 19:33:05 crc kubenswrapper[4988]: I1008 19:33:05.793944 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/bbe1c39b-8f0b-41a4-87a3-20672272af49-secrets\") pod \"openstack-cell1-galera-0\" (UID: \"bbe1c39b-8f0b-41a4-87a3-20672272af49\") " pod="openstack/openstack-cell1-galera-0" Oct 08 19:33:05 crc kubenswrapper[4988]: I1008 19:33:05.794074 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-klvsr\" (UniqueName: \"kubernetes.io/projected/bbe1c39b-8f0b-41a4-87a3-20672272af49-kube-api-access-klvsr\") pod \"openstack-cell1-galera-0\" (UID: \"bbe1c39b-8f0b-41a4-87a3-20672272af49\") " pod="openstack/openstack-cell1-galera-0" Oct 08 19:33:05 crc kubenswrapper[4988]: I1008 19:33:05.794154 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-0020677f-6522-4086-9a15-9d1dae7b1119\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-0020677f-6522-4086-9a15-9d1dae7b1119\") pod \"openstack-cell1-galera-0\" (UID: \"bbe1c39b-8f0b-41a4-87a3-20672272af49\") " pod="openstack/openstack-cell1-galera-0" Oct 08 19:33:05 crc kubenswrapper[4988]: I1008 19:33:05.905845 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/bbe1c39b-8f0b-41a4-87a3-20672272af49-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"bbe1c39b-8f0b-41a4-87a3-20672272af49\") " pod="openstack/openstack-cell1-galera-0" Oct 08 19:33:05 crc kubenswrapper[4988]: I1008 19:33:05.905910 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/bbe1c39b-8f0b-41a4-87a3-20672272af49-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"bbe1c39b-8f0b-41a4-87a3-20672272af49\") " pod="openstack/openstack-cell1-galera-0" Oct 08 19:33:05 crc kubenswrapper[4988]: I1008 19:33:05.905942 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/bbe1c39b-8f0b-41a4-87a3-20672272af49-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"bbe1c39b-8f0b-41a4-87a3-20672272af49\") " pod="openstack/openstack-cell1-galera-0" Oct 08 19:33:05 crc kubenswrapper[4988]: I1008 19:33:05.905959 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bbe1c39b-8f0b-41a4-87a3-20672272af49-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"bbe1c39b-8f0b-41a4-87a3-20672272af49\") " pod="openstack/openstack-cell1-galera-0" Oct 08 19:33:05 crc kubenswrapper[4988]: I1008 19:33:05.906016 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/bbe1c39b-8f0b-41a4-87a3-20672272af49-secrets\") pod \"openstack-cell1-galera-0\" (UID: \"bbe1c39b-8f0b-41a4-87a3-20672272af49\") " pod="openstack/openstack-cell1-galera-0" Oct 08 19:33:05 crc kubenswrapper[4988]: I1008 19:33:05.906111 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-klvsr\" (UniqueName: \"kubernetes.io/projected/bbe1c39b-8f0b-41a4-87a3-20672272af49-kube-api-access-klvsr\") pod \"openstack-cell1-galera-0\" (UID: \"bbe1c39b-8f0b-41a4-87a3-20672272af49\") " pod="openstack/openstack-cell1-galera-0" Oct 08 19:33:05 crc kubenswrapper[4988]: I1008 19:33:05.906144 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-0020677f-6522-4086-9a15-9d1dae7b1119\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-0020677f-6522-4086-9a15-9d1dae7b1119\") pod \"openstack-cell1-galera-0\" (UID: \"bbe1c39b-8f0b-41a4-87a3-20672272af49\") " pod="openstack/openstack-cell1-galera-0" Oct 08 19:33:05 crc kubenswrapper[4988]: I1008 19:33:05.906189 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/bbe1c39b-8f0b-41a4-87a3-20672272af49-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"bbe1c39b-8f0b-41a4-87a3-20672272af49\") " pod="openstack/openstack-cell1-galera-0" Oct 08 19:33:05 crc kubenswrapper[4988]: I1008 19:33:05.906225 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/bbe1c39b-8f0b-41a4-87a3-20672272af49-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"bbe1c39b-8f0b-41a4-87a3-20672272af49\") " pod="openstack/openstack-cell1-galera-0" Oct 08 19:33:05 crc kubenswrapper[4988]: I1008 19:33:05.906664 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/bbe1c39b-8f0b-41a4-87a3-20672272af49-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"bbe1c39b-8f0b-41a4-87a3-20672272af49\") " pod="openstack/openstack-cell1-galera-0" Oct 08 19:33:05 crc kubenswrapper[4988]: I1008 19:33:05.906950 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/bbe1c39b-8f0b-41a4-87a3-20672272af49-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"bbe1c39b-8f0b-41a4-87a3-20672272af49\") " pod="openstack/openstack-cell1-galera-0" Oct 08 19:33:05 crc kubenswrapper[4988]: I1008 19:33:05.907665 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/bbe1c39b-8f0b-41a4-87a3-20672272af49-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"bbe1c39b-8f0b-41a4-87a3-20672272af49\") " pod="openstack/openstack-cell1-galera-0" Oct 08 19:33:05 crc kubenswrapper[4988]: I1008 19:33:05.909947 4988 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Oct 08 19:33:05 crc kubenswrapper[4988]: I1008 19:33:05.909980 4988 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-0020677f-6522-4086-9a15-9d1dae7b1119\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-0020677f-6522-4086-9a15-9d1dae7b1119\") pod \"openstack-cell1-galera-0\" (UID: \"bbe1c39b-8f0b-41a4-87a3-20672272af49\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/4bf396003bb69568d29add85001905f6048d124eeff1dbee4e0fb552663d85e4/globalmount\"" pod="openstack/openstack-cell1-galera-0" Oct 08 19:33:05 crc kubenswrapper[4988]: I1008 19:33:05.910049 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/bbe1c39b-8f0b-41a4-87a3-20672272af49-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"bbe1c39b-8f0b-41a4-87a3-20672272af49\") " pod="openstack/openstack-cell1-galera-0" Oct 08 19:33:05 crc kubenswrapper[4988]: I1008 19:33:05.911880 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/bbe1c39b-8f0b-41a4-87a3-20672272af49-secrets\") pod \"openstack-cell1-galera-0\" (UID: \"bbe1c39b-8f0b-41a4-87a3-20672272af49\") " pod="openstack/openstack-cell1-galera-0" Oct 08 19:33:05 crc kubenswrapper[4988]: I1008 19:33:05.911882 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/bbe1c39b-8f0b-41a4-87a3-20672272af49-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"bbe1c39b-8f0b-41a4-87a3-20672272af49\") " pod="openstack/openstack-cell1-galera-0" Oct 08 19:33:05 crc kubenswrapper[4988]: I1008 19:33:05.932100 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-klvsr\" (UniqueName: \"kubernetes.io/projected/bbe1c39b-8f0b-41a4-87a3-20672272af49-kube-api-access-klvsr\") pod \"openstack-cell1-galera-0\" (UID: \"bbe1c39b-8f0b-41a4-87a3-20672272af49\") " pod="openstack/openstack-cell1-galera-0" Oct 08 19:33:05 crc kubenswrapper[4988]: I1008 19:33:05.932952 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bbe1c39b-8f0b-41a4-87a3-20672272af49-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"bbe1c39b-8f0b-41a4-87a3-20672272af49\") " pod="openstack/openstack-cell1-galera-0" Oct 08 19:33:05 crc kubenswrapper[4988]: I1008 19:33:05.953616 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-0020677f-6522-4086-9a15-9d1dae7b1119\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-0020677f-6522-4086-9a15-9d1dae7b1119\") pod \"openstack-cell1-galera-0\" (UID: \"bbe1c39b-8f0b-41a4-87a3-20672272af49\") " pod="openstack/openstack-cell1-galera-0" Oct 08 19:33:06 crc kubenswrapper[4988]: I1008 19:33:06.003943 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/memcached-0"] Oct 08 19:33:06 crc kubenswrapper[4988]: I1008 19:33:06.005209 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Oct 08 19:33:06 crc kubenswrapper[4988]: I1008 19:33:06.009339 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-memcached-svc" Oct 08 19:33:06 crc kubenswrapper[4988]: I1008 19:33:06.009339 4988 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"memcached-config-data" Oct 08 19:33:06 crc kubenswrapper[4988]: I1008 19:33:06.009368 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"memcached-memcached-dockercfg-jl7dz" Oct 08 19:33:06 crc kubenswrapper[4988]: I1008 19:33:06.023116 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Oct 08 19:33:06 crc kubenswrapper[4988]: I1008 19:33:06.047522 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Oct 08 19:33:06 crc kubenswrapper[4988]: I1008 19:33:06.108671 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/a44ce926-29f2-4851-aabf-65cc7ffd62bc-memcached-tls-certs\") pod \"memcached-0\" (UID: \"a44ce926-29f2-4851-aabf-65cc7ffd62bc\") " pod="openstack/memcached-0" Oct 08 19:33:06 crc kubenswrapper[4988]: I1008 19:33:06.108915 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/a44ce926-29f2-4851-aabf-65cc7ffd62bc-kolla-config\") pod \"memcached-0\" (UID: \"a44ce926-29f2-4851-aabf-65cc7ffd62bc\") " pod="openstack/memcached-0" Oct 08 19:33:06 crc kubenswrapper[4988]: I1008 19:33:06.108945 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/a44ce926-29f2-4851-aabf-65cc7ffd62bc-config-data\") pod \"memcached-0\" (UID: \"a44ce926-29f2-4851-aabf-65cc7ffd62bc\") " pod="openstack/memcached-0" Oct 08 19:33:06 crc kubenswrapper[4988]: I1008 19:33:06.108975 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a44ce926-29f2-4851-aabf-65cc7ffd62bc-combined-ca-bundle\") pod \"memcached-0\" (UID: \"a44ce926-29f2-4851-aabf-65cc7ffd62bc\") " pod="openstack/memcached-0" Oct 08 19:33:06 crc kubenswrapper[4988]: I1008 19:33:06.109000 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jx5rc\" (UniqueName: \"kubernetes.io/projected/a44ce926-29f2-4851-aabf-65cc7ffd62bc-kube-api-access-jx5rc\") pod \"memcached-0\" (UID: \"a44ce926-29f2-4851-aabf-65cc7ffd62bc\") " pod="openstack/memcached-0" Oct 08 19:33:06 crc kubenswrapper[4988]: I1008 19:33:06.210650 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/a44ce926-29f2-4851-aabf-65cc7ffd62bc-memcached-tls-certs\") pod \"memcached-0\" (UID: \"a44ce926-29f2-4851-aabf-65cc7ffd62bc\") " pod="openstack/memcached-0" Oct 08 19:33:06 crc kubenswrapper[4988]: I1008 19:33:06.210699 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/a44ce926-29f2-4851-aabf-65cc7ffd62bc-kolla-config\") pod \"memcached-0\" (UID: \"a44ce926-29f2-4851-aabf-65cc7ffd62bc\") " pod="openstack/memcached-0" Oct 08 19:33:06 crc kubenswrapper[4988]: I1008 19:33:06.210735 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/a44ce926-29f2-4851-aabf-65cc7ffd62bc-config-data\") pod \"memcached-0\" (UID: \"a44ce926-29f2-4851-aabf-65cc7ffd62bc\") " pod="openstack/memcached-0" Oct 08 19:33:06 crc kubenswrapper[4988]: I1008 19:33:06.210779 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a44ce926-29f2-4851-aabf-65cc7ffd62bc-combined-ca-bundle\") pod \"memcached-0\" (UID: \"a44ce926-29f2-4851-aabf-65cc7ffd62bc\") " pod="openstack/memcached-0" Oct 08 19:33:06 crc kubenswrapper[4988]: I1008 19:33:06.210816 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jx5rc\" (UniqueName: \"kubernetes.io/projected/a44ce926-29f2-4851-aabf-65cc7ffd62bc-kube-api-access-jx5rc\") pod \"memcached-0\" (UID: \"a44ce926-29f2-4851-aabf-65cc7ffd62bc\") " pod="openstack/memcached-0" Oct 08 19:33:06 crc kubenswrapper[4988]: I1008 19:33:06.211802 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/a44ce926-29f2-4851-aabf-65cc7ffd62bc-kolla-config\") pod \"memcached-0\" (UID: \"a44ce926-29f2-4851-aabf-65cc7ffd62bc\") " pod="openstack/memcached-0" Oct 08 19:33:06 crc kubenswrapper[4988]: I1008 19:33:06.211981 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/a44ce926-29f2-4851-aabf-65cc7ffd62bc-config-data\") pod \"memcached-0\" (UID: \"a44ce926-29f2-4851-aabf-65cc7ffd62bc\") " pod="openstack/memcached-0" Oct 08 19:33:06 crc kubenswrapper[4988]: I1008 19:33:06.214685 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/a44ce926-29f2-4851-aabf-65cc7ffd62bc-memcached-tls-certs\") pod \"memcached-0\" (UID: \"a44ce926-29f2-4851-aabf-65cc7ffd62bc\") " pod="openstack/memcached-0" Oct 08 19:33:06 crc kubenswrapper[4988]: I1008 19:33:06.215051 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a44ce926-29f2-4851-aabf-65cc7ffd62bc-combined-ca-bundle\") pod \"memcached-0\" (UID: \"a44ce926-29f2-4851-aabf-65cc7ffd62bc\") " pod="openstack/memcached-0" Oct 08 19:33:06 crc kubenswrapper[4988]: I1008 19:33:06.227724 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jx5rc\" (UniqueName: \"kubernetes.io/projected/a44ce926-29f2-4851-aabf-65cc7ffd62bc-kube-api-access-jx5rc\") pod \"memcached-0\" (UID: \"a44ce926-29f2-4851-aabf-65cc7ffd62bc\") " pod="openstack/memcached-0" Oct 08 19:33:06 crc kubenswrapper[4988]: I1008 19:33:06.363357 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Oct 08 19:33:06 crc kubenswrapper[4988]: I1008 19:33:06.496126 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Oct 08 19:33:06 crc kubenswrapper[4988]: W1008 19:33:06.499491 4988 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podbbe1c39b_8f0b_41a4_87a3_20672272af49.slice/crio-f9c1a5b3977fdbb90a64c1a2b0286ba6cc221aab5080d117497d4d8c020f902f WatchSource:0}: Error finding container f9c1a5b3977fdbb90a64c1a2b0286ba6cc221aab5080d117497d4d8c020f902f: Status 404 returned error can't find the container with id f9c1a5b3977fdbb90a64c1a2b0286ba6cc221aab5080d117497d4d8c020f902f Oct 08 19:33:06 crc kubenswrapper[4988]: I1008 19:33:06.590032 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Oct 08 19:33:06 crc kubenswrapper[4988]: W1008 19:33:06.682928 4988 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda44ce926_29f2_4851_aabf_65cc7ffd62bc.slice/crio-09d930dc885232ec4c55e5f50580bf504330a0e3bc22845d6a0125361ec6ee6c WatchSource:0}: Error finding container 09d930dc885232ec4c55e5f50580bf504330a0e3bc22845d6a0125361ec6ee6c: Status 404 returned error can't find the container with id 09d930dc885232ec4c55e5f50580bf504330a0e3bc22845d6a0125361ec6ee6c Oct 08 19:33:06 crc kubenswrapper[4988]: I1008 19:33:06.702216 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"689b2a7d-38cb-4476-a0ea-27bbd614fb2b","Type":"ContainerStarted","Data":"eca33a88e54407e293c7e533977055c54ba2c63973f13d665e1d020a91a20ede"} Oct 08 19:33:06 crc kubenswrapper[4988]: I1008 19:33:06.703338 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"bbe1c39b-8f0b-41a4-87a3-20672272af49","Type":"ContainerStarted","Data":"f9c1a5b3977fdbb90a64c1a2b0286ba6cc221aab5080d117497d4d8c020f902f"} Oct 08 19:33:06 crc kubenswrapper[4988]: I1008 19:33:06.706114 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"a44ce926-29f2-4851-aabf-65cc7ffd62bc","Type":"ContainerStarted","Data":"09d930dc885232ec4c55e5f50580bf504330a0e3bc22845d6a0125361ec6ee6c"} Oct 08 19:33:07 crc kubenswrapper[4988]: I1008 19:33:07.720182 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"a44ce926-29f2-4851-aabf-65cc7ffd62bc","Type":"ContainerStarted","Data":"b249b9768364f79ee9501933a149141d7e7525c7ba09cada0a65c0bc967989a6"} Oct 08 19:33:07 crc kubenswrapper[4988]: I1008 19:33:07.720789 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/memcached-0" Oct 08 19:33:07 crc kubenswrapper[4988]: I1008 19:33:07.723936 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"bbe1c39b-8f0b-41a4-87a3-20672272af49","Type":"ContainerStarted","Data":"012d46d52da728696876e7fa0671048ba75858be083919a6e00a0cade2ccab74"} Oct 08 19:33:07 crc kubenswrapper[4988]: I1008 19:33:07.747641 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/memcached-0" podStartSLOduration=2.747610082 podStartE2EDuration="2.747610082s" podCreationTimestamp="2025-10-08 19:33:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 19:33:07.743665797 +0000 UTC m=+4933.193508607" watchObservedRunningTime="2025-10-08 19:33:07.747610082 +0000 UTC m=+4933.197452892" Oct 08 19:33:09 crc kubenswrapper[4988]: I1008 19:33:09.747122 4988 generic.go:334] "Generic (PLEG): container finished" podID="689b2a7d-38cb-4476-a0ea-27bbd614fb2b" containerID="eca33a88e54407e293c7e533977055c54ba2c63973f13d665e1d020a91a20ede" exitCode=0 Oct 08 19:33:09 crc kubenswrapper[4988]: I1008 19:33:09.747193 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"689b2a7d-38cb-4476-a0ea-27bbd614fb2b","Type":"ContainerDied","Data":"eca33a88e54407e293c7e533977055c54ba2c63973f13d665e1d020a91a20ede"} Oct 08 19:33:10 crc kubenswrapper[4988]: I1008 19:33:10.762697 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"689b2a7d-38cb-4476-a0ea-27bbd614fb2b","Type":"ContainerStarted","Data":"0ee7ebc07c3057f41e216d3d35f86c399c796cbd25d420c44c9cb0a35684e338"} Oct 08 19:33:10 crc kubenswrapper[4988]: I1008 19:33:10.767152 4988 generic.go:334] "Generic (PLEG): container finished" podID="bbe1c39b-8f0b-41a4-87a3-20672272af49" containerID="012d46d52da728696876e7fa0671048ba75858be083919a6e00a0cade2ccab74" exitCode=0 Oct 08 19:33:10 crc kubenswrapper[4988]: I1008 19:33:10.767204 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"bbe1c39b-8f0b-41a4-87a3-20672272af49","Type":"ContainerDied","Data":"012d46d52da728696876e7fa0671048ba75858be083919a6e00a0cade2ccab74"} Oct 08 19:33:10 crc kubenswrapper[4988]: I1008 19:33:10.801297 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-galera-0" podStartSLOduration=7.801266515 podStartE2EDuration="7.801266515s" podCreationTimestamp="2025-10-08 19:33:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 19:33:10.794952834 +0000 UTC m=+4936.244795644" watchObservedRunningTime="2025-10-08 19:33:10.801266515 +0000 UTC m=+4936.251109325" Oct 08 19:33:11 crc kubenswrapper[4988]: I1008 19:33:11.365639 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/memcached-0" Oct 08 19:33:11 crc kubenswrapper[4988]: I1008 19:33:11.774705 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"bbe1c39b-8f0b-41a4-87a3-20672272af49","Type":"ContainerStarted","Data":"071975dc0bf39868da9bd405198e196e81c61524ad23d5801d8fa539ffbe2532"} Oct 08 19:33:11 crc kubenswrapper[4988]: I1008 19:33:11.807342 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-cell1-galera-0" podStartSLOduration=7.8073212 podStartE2EDuration="7.8073212s" podCreationTimestamp="2025-10-08 19:33:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 19:33:11.801052181 +0000 UTC m=+4937.250894961" watchObservedRunningTime="2025-10-08 19:33:11.8073212 +0000 UTC m=+4937.257163970" Oct 08 19:33:12 crc kubenswrapper[4988]: I1008 19:33:12.406641 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-85f98b87f9-gvdgx" Oct 08 19:33:12 crc kubenswrapper[4988]: I1008 19:33:12.743358 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-67d9f7fb89-lf6gm" Oct 08 19:33:12 crc kubenswrapper[4988]: I1008 19:33:12.789304 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-85f98b87f9-gvdgx"] Oct 08 19:33:12 crc kubenswrapper[4988]: I1008 19:33:12.789568 4988 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-85f98b87f9-gvdgx" podUID="a63f21fb-1703-48d0-b2f3-2f9aaa3b0e95" containerName="dnsmasq-dns" containerID="cri-o://121035b9c5e9e2a7f2ccc2ed29356a72aa767268cf19f124327fddfca3dcf584" gracePeriod=10 Oct 08 19:33:13 crc kubenswrapper[4988]: I1008 19:33:13.210397 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-85f98b87f9-gvdgx" Oct 08 19:33:13 crc kubenswrapper[4988]: I1008 19:33:13.331096 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a63f21fb-1703-48d0-b2f3-2f9aaa3b0e95-dns-svc\") pod \"a63f21fb-1703-48d0-b2f3-2f9aaa3b0e95\" (UID: \"a63f21fb-1703-48d0-b2f3-2f9aaa3b0e95\") " Oct 08 19:33:13 crc kubenswrapper[4988]: I1008 19:33:13.331239 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nzwq2\" (UniqueName: \"kubernetes.io/projected/a63f21fb-1703-48d0-b2f3-2f9aaa3b0e95-kube-api-access-nzwq2\") pod \"a63f21fb-1703-48d0-b2f3-2f9aaa3b0e95\" (UID: \"a63f21fb-1703-48d0-b2f3-2f9aaa3b0e95\") " Oct 08 19:33:13 crc kubenswrapper[4988]: I1008 19:33:13.331312 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a63f21fb-1703-48d0-b2f3-2f9aaa3b0e95-config\") pod \"a63f21fb-1703-48d0-b2f3-2f9aaa3b0e95\" (UID: \"a63f21fb-1703-48d0-b2f3-2f9aaa3b0e95\") " Oct 08 19:33:13 crc kubenswrapper[4988]: I1008 19:33:13.337538 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a63f21fb-1703-48d0-b2f3-2f9aaa3b0e95-kube-api-access-nzwq2" (OuterVolumeSpecName: "kube-api-access-nzwq2") pod "a63f21fb-1703-48d0-b2f3-2f9aaa3b0e95" (UID: "a63f21fb-1703-48d0-b2f3-2f9aaa3b0e95"). InnerVolumeSpecName "kube-api-access-nzwq2". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 19:33:13 crc kubenswrapper[4988]: I1008 19:33:13.372906 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a63f21fb-1703-48d0-b2f3-2f9aaa3b0e95-config" (OuterVolumeSpecName: "config") pod "a63f21fb-1703-48d0-b2f3-2f9aaa3b0e95" (UID: "a63f21fb-1703-48d0-b2f3-2f9aaa3b0e95"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 19:33:13 crc kubenswrapper[4988]: I1008 19:33:13.373079 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a63f21fb-1703-48d0-b2f3-2f9aaa3b0e95-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "a63f21fb-1703-48d0-b2f3-2f9aaa3b0e95" (UID: "a63f21fb-1703-48d0-b2f3-2f9aaa3b0e95"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 19:33:13 crc kubenswrapper[4988]: I1008 19:33:13.433741 4988 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a63f21fb-1703-48d0-b2f3-2f9aaa3b0e95-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 08 19:33:13 crc kubenswrapper[4988]: I1008 19:33:13.433792 4988 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nzwq2\" (UniqueName: \"kubernetes.io/projected/a63f21fb-1703-48d0-b2f3-2f9aaa3b0e95-kube-api-access-nzwq2\") on node \"crc\" DevicePath \"\"" Oct 08 19:33:13 crc kubenswrapper[4988]: I1008 19:33:13.433814 4988 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a63f21fb-1703-48d0-b2f3-2f9aaa3b0e95-config\") on node \"crc\" DevicePath \"\"" Oct 08 19:33:13 crc kubenswrapper[4988]: I1008 19:33:13.790740 4988 generic.go:334] "Generic (PLEG): container finished" podID="a63f21fb-1703-48d0-b2f3-2f9aaa3b0e95" containerID="121035b9c5e9e2a7f2ccc2ed29356a72aa767268cf19f124327fddfca3dcf584" exitCode=0 Oct 08 19:33:13 crc kubenswrapper[4988]: I1008 19:33:13.790785 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-85f98b87f9-gvdgx" event={"ID":"a63f21fb-1703-48d0-b2f3-2f9aaa3b0e95","Type":"ContainerDied","Data":"121035b9c5e9e2a7f2ccc2ed29356a72aa767268cf19f124327fddfca3dcf584"} Oct 08 19:33:13 crc kubenswrapper[4988]: I1008 19:33:13.790817 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-85f98b87f9-gvdgx" event={"ID":"a63f21fb-1703-48d0-b2f3-2f9aaa3b0e95","Type":"ContainerDied","Data":"8f212e2ec4b4172dfc103dc9399e104dae8f4e7ed97c332127f19f5b1c56689f"} Oct 08 19:33:13 crc kubenswrapper[4988]: I1008 19:33:13.790825 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-85f98b87f9-gvdgx" Oct 08 19:33:13 crc kubenswrapper[4988]: I1008 19:33:13.790836 4988 scope.go:117] "RemoveContainer" containerID="121035b9c5e9e2a7f2ccc2ed29356a72aa767268cf19f124327fddfca3dcf584" Oct 08 19:33:13 crc kubenswrapper[4988]: I1008 19:33:13.815534 4988 scope.go:117] "RemoveContainer" containerID="57a11453c9957b0585dee8ccb4454ce79d70d58eab6edcd79dd3937bf6d0a59c" Oct 08 19:33:13 crc kubenswrapper[4988]: I1008 19:33:13.833147 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-85f98b87f9-gvdgx"] Oct 08 19:33:13 crc kubenswrapper[4988]: I1008 19:33:13.840644 4988 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-85f98b87f9-gvdgx"] Oct 08 19:33:13 crc kubenswrapper[4988]: I1008 19:33:13.852193 4988 scope.go:117] "RemoveContainer" containerID="121035b9c5e9e2a7f2ccc2ed29356a72aa767268cf19f124327fddfca3dcf584" Oct 08 19:33:13 crc kubenswrapper[4988]: E1008 19:33:13.852754 4988 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"121035b9c5e9e2a7f2ccc2ed29356a72aa767268cf19f124327fddfca3dcf584\": container with ID starting with 121035b9c5e9e2a7f2ccc2ed29356a72aa767268cf19f124327fddfca3dcf584 not found: ID does not exist" containerID="121035b9c5e9e2a7f2ccc2ed29356a72aa767268cf19f124327fddfca3dcf584" Oct 08 19:33:13 crc kubenswrapper[4988]: I1008 19:33:13.852802 4988 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"121035b9c5e9e2a7f2ccc2ed29356a72aa767268cf19f124327fddfca3dcf584"} err="failed to get container status \"121035b9c5e9e2a7f2ccc2ed29356a72aa767268cf19f124327fddfca3dcf584\": rpc error: code = NotFound desc = could not find container \"121035b9c5e9e2a7f2ccc2ed29356a72aa767268cf19f124327fddfca3dcf584\": container with ID starting with 121035b9c5e9e2a7f2ccc2ed29356a72aa767268cf19f124327fddfca3dcf584 not found: ID does not exist" Oct 08 19:33:13 crc kubenswrapper[4988]: I1008 19:33:13.852837 4988 scope.go:117] "RemoveContainer" containerID="57a11453c9957b0585dee8ccb4454ce79d70d58eab6edcd79dd3937bf6d0a59c" Oct 08 19:33:13 crc kubenswrapper[4988]: E1008 19:33:13.853369 4988 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"57a11453c9957b0585dee8ccb4454ce79d70d58eab6edcd79dd3937bf6d0a59c\": container with ID starting with 57a11453c9957b0585dee8ccb4454ce79d70d58eab6edcd79dd3937bf6d0a59c not found: ID does not exist" containerID="57a11453c9957b0585dee8ccb4454ce79d70d58eab6edcd79dd3937bf6d0a59c" Oct 08 19:33:13 crc kubenswrapper[4988]: I1008 19:33:13.853464 4988 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"57a11453c9957b0585dee8ccb4454ce79d70d58eab6edcd79dd3937bf6d0a59c"} err="failed to get container status \"57a11453c9957b0585dee8ccb4454ce79d70d58eab6edcd79dd3937bf6d0a59c\": rpc error: code = NotFound desc = could not find container \"57a11453c9957b0585dee8ccb4454ce79d70d58eab6edcd79dd3937bf6d0a59c\": container with ID starting with 57a11453c9957b0585dee8ccb4454ce79d70d58eab6edcd79dd3937bf6d0a59c not found: ID does not exist" Oct 08 19:33:14 crc kubenswrapper[4988]: I1008 19:33:14.988709 4988 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-galera-0" Oct 08 19:33:14 crc kubenswrapper[4988]: I1008 19:33:14.989123 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-galera-0" Oct 08 19:33:15 crc kubenswrapper[4988]: I1008 19:33:15.272816 4988 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a63f21fb-1703-48d0-b2f3-2f9aaa3b0e95" path="/var/lib/kubelet/pods/a63f21fb-1703-48d0-b2f3-2f9aaa3b0e95/volumes" Oct 08 19:33:15 crc kubenswrapper[4988]: E1008 19:33:15.628062 4988 upgradeaware.go:427] Error proxying data from client to backend: readfrom tcp 38.102.83.102:50680->38.102.83.102:38495: write tcp 38.102.83.102:50680->38.102.83.102:38495: write: broken pipe Oct 08 19:33:16 crc kubenswrapper[4988]: I1008 19:33:16.048316 4988 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-cell1-galera-0" Oct 08 19:33:16 crc kubenswrapper[4988]: I1008 19:33:16.048374 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-cell1-galera-0" Oct 08 19:33:17 crc kubenswrapper[4988]: I1008 19:33:17.074900 4988 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-galera-0" Oct 08 19:33:17 crc kubenswrapper[4988]: I1008 19:33:17.153570 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-galera-0" Oct 08 19:33:18 crc kubenswrapper[4988]: I1008 19:33:18.140647 4988 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-cell1-galera-0" Oct 08 19:33:18 crc kubenswrapper[4988]: I1008 19:33:18.191626 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-cell1-galera-0" Oct 08 19:33:38 crc kubenswrapper[4988]: I1008 19:33:38.046808 4988 generic.go:334] "Generic (PLEG): container finished" podID="c754aa11-b846-4860-a9c7-9e3a12be054e" containerID="9345a690467ea890e581ad94aa2b8e47eeb1adbd92e069d7d3fad8698f98fc51" exitCode=0 Oct 08 19:33:38 crc kubenswrapper[4988]: I1008 19:33:38.046882 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"c754aa11-b846-4860-a9c7-9e3a12be054e","Type":"ContainerDied","Data":"9345a690467ea890e581ad94aa2b8e47eeb1adbd92e069d7d3fad8698f98fc51"} Oct 08 19:33:38 crc kubenswrapper[4988]: I1008 19:33:38.051317 4988 generic.go:334] "Generic (PLEG): container finished" podID="111b6cdd-1e77-4992-a331-c66bc9e95a90" containerID="6b0b30e0590363081acab72ac2c5ea5643a1db58b78fa1169b9902c4781121cb" exitCode=0 Oct 08 19:33:38 crc kubenswrapper[4988]: I1008 19:33:38.051360 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"111b6cdd-1e77-4992-a331-c66bc9e95a90","Type":"ContainerDied","Data":"6b0b30e0590363081acab72ac2c5ea5643a1db58b78fa1169b9902c4781121cb"} Oct 08 19:33:39 crc kubenswrapper[4988]: I1008 19:33:39.062282 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"111b6cdd-1e77-4992-a331-c66bc9e95a90","Type":"ContainerStarted","Data":"3315a80b1321e1fb678ae4d9697fad7f24942f895c83ae0783014fc111a3463c"} Oct 08 19:33:39 crc kubenswrapper[4988]: I1008 19:33:39.062942 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Oct 08 19:33:39 crc kubenswrapper[4988]: I1008 19:33:39.064854 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"c754aa11-b846-4860-a9c7-9e3a12be054e","Type":"ContainerStarted","Data":"28d7c34187c55210cc93cc0372a468595ac8a8f5f7dfde181ec13d9210f0055a"} Oct 08 19:33:39 crc kubenswrapper[4988]: I1008 19:33:39.065434 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Oct 08 19:33:39 crc kubenswrapper[4988]: I1008 19:33:39.088745 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-cell1-server-0" podStartSLOduration=37.088728816 podStartE2EDuration="37.088728816s" podCreationTimestamp="2025-10-08 19:33:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 19:33:39.086302749 +0000 UTC m=+4964.536145529" watchObservedRunningTime="2025-10-08 19:33:39.088728816 +0000 UTC m=+4964.538571576" Oct 08 19:33:39 crc kubenswrapper[4988]: I1008 19:33:39.128199 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-0" podStartSLOduration=37.12816973 podStartE2EDuration="37.12816973s" podCreationTimestamp="2025-10-08 19:33:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 19:33:39.120167346 +0000 UTC m=+4964.570010156" watchObservedRunningTime="2025-10-08 19:33:39.12816973 +0000 UTC m=+4964.578012510" Oct 08 19:33:53 crc kubenswrapper[4988]: I1008 19:33:53.338200 4988 patch_prober.go:28] interesting pod/machine-config-daemon-nm54f container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 19:33:53 crc kubenswrapper[4988]: I1008 19:33:53.338874 4988 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 19:33:53 crc kubenswrapper[4988]: I1008 19:33:53.552927 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-0" Oct 08 19:33:53 crc kubenswrapper[4988]: I1008 19:33:53.861702 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-cell1-server-0" Oct 08 19:33:58 crc kubenswrapper[4988]: I1008 19:33:58.576655 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5fdc957c47-mtggm"] Oct 08 19:33:58 crc kubenswrapper[4988]: E1008 19:33:58.579161 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a63f21fb-1703-48d0-b2f3-2f9aaa3b0e95" containerName="init" Oct 08 19:33:58 crc kubenswrapper[4988]: I1008 19:33:58.579200 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="a63f21fb-1703-48d0-b2f3-2f9aaa3b0e95" containerName="init" Oct 08 19:33:58 crc kubenswrapper[4988]: E1008 19:33:58.579250 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a63f21fb-1703-48d0-b2f3-2f9aaa3b0e95" containerName="dnsmasq-dns" Oct 08 19:33:58 crc kubenswrapper[4988]: I1008 19:33:58.579263 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="a63f21fb-1703-48d0-b2f3-2f9aaa3b0e95" containerName="dnsmasq-dns" Oct 08 19:33:58 crc kubenswrapper[4988]: I1008 19:33:58.579551 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="a63f21fb-1703-48d0-b2f3-2f9aaa3b0e95" containerName="dnsmasq-dns" Oct 08 19:33:58 crc kubenswrapper[4988]: I1008 19:33:58.581023 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5fdc957c47-mtggm" Oct 08 19:33:58 crc kubenswrapper[4988]: I1008 19:33:58.600345 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5fdc957c47-mtggm"] Oct 08 19:33:58 crc kubenswrapper[4988]: I1008 19:33:58.729542 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/14aa686b-2873-45e4-8c67-ae59733b51a6-config\") pod \"dnsmasq-dns-5fdc957c47-mtggm\" (UID: \"14aa686b-2873-45e4-8c67-ae59733b51a6\") " pod="openstack/dnsmasq-dns-5fdc957c47-mtggm" Oct 08 19:33:58 crc kubenswrapper[4988]: I1008 19:33:58.729839 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zqckg\" (UniqueName: \"kubernetes.io/projected/14aa686b-2873-45e4-8c67-ae59733b51a6-kube-api-access-zqckg\") pod \"dnsmasq-dns-5fdc957c47-mtggm\" (UID: \"14aa686b-2873-45e4-8c67-ae59733b51a6\") " pod="openstack/dnsmasq-dns-5fdc957c47-mtggm" Oct 08 19:33:58 crc kubenswrapper[4988]: I1008 19:33:58.729971 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/14aa686b-2873-45e4-8c67-ae59733b51a6-dns-svc\") pod \"dnsmasq-dns-5fdc957c47-mtggm\" (UID: \"14aa686b-2873-45e4-8c67-ae59733b51a6\") " pod="openstack/dnsmasq-dns-5fdc957c47-mtggm" Oct 08 19:33:58 crc kubenswrapper[4988]: I1008 19:33:58.831764 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/14aa686b-2873-45e4-8c67-ae59733b51a6-config\") pod \"dnsmasq-dns-5fdc957c47-mtggm\" (UID: \"14aa686b-2873-45e4-8c67-ae59733b51a6\") " pod="openstack/dnsmasq-dns-5fdc957c47-mtggm" Oct 08 19:33:58 crc kubenswrapper[4988]: I1008 19:33:58.831832 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zqckg\" (UniqueName: \"kubernetes.io/projected/14aa686b-2873-45e4-8c67-ae59733b51a6-kube-api-access-zqckg\") pod \"dnsmasq-dns-5fdc957c47-mtggm\" (UID: \"14aa686b-2873-45e4-8c67-ae59733b51a6\") " pod="openstack/dnsmasq-dns-5fdc957c47-mtggm" Oct 08 19:33:58 crc kubenswrapper[4988]: I1008 19:33:58.831863 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/14aa686b-2873-45e4-8c67-ae59733b51a6-dns-svc\") pod \"dnsmasq-dns-5fdc957c47-mtggm\" (UID: \"14aa686b-2873-45e4-8c67-ae59733b51a6\") " pod="openstack/dnsmasq-dns-5fdc957c47-mtggm" Oct 08 19:33:58 crc kubenswrapper[4988]: I1008 19:33:58.832845 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/14aa686b-2873-45e4-8c67-ae59733b51a6-config\") pod \"dnsmasq-dns-5fdc957c47-mtggm\" (UID: \"14aa686b-2873-45e4-8c67-ae59733b51a6\") " pod="openstack/dnsmasq-dns-5fdc957c47-mtggm" Oct 08 19:33:58 crc kubenswrapper[4988]: I1008 19:33:58.833647 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/14aa686b-2873-45e4-8c67-ae59733b51a6-dns-svc\") pod \"dnsmasq-dns-5fdc957c47-mtggm\" (UID: \"14aa686b-2873-45e4-8c67-ae59733b51a6\") " pod="openstack/dnsmasq-dns-5fdc957c47-mtggm" Oct 08 19:33:58 crc kubenswrapper[4988]: I1008 19:33:58.852929 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zqckg\" (UniqueName: \"kubernetes.io/projected/14aa686b-2873-45e4-8c67-ae59733b51a6-kube-api-access-zqckg\") pod \"dnsmasq-dns-5fdc957c47-mtggm\" (UID: \"14aa686b-2873-45e4-8c67-ae59733b51a6\") " pod="openstack/dnsmasq-dns-5fdc957c47-mtggm" Oct 08 19:33:58 crc kubenswrapper[4988]: I1008 19:33:58.909848 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5fdc957c47-mtggm" Oct 08 19:33:59 crc kubenswrapper[4988]: W1008 19:33:59.317898 4988 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod14aa686b_2873_45e4_8c67_ae59733b51a6.slice/crio-db28ff1ba831b237c6929dd2b10e2395e9578409ebf0ed76ff11f73956df8c82 WatchSource:0}: Error finding container db28ff1ba831b237c6929dd2b10e2395e9578409ebf0ed76ff11f73956df8c82: Status 404 returned error can't find the container with id db28ff1ba831b237c6929dd2b10e2395e9578409ebf0ed76ff11f73956df8c82 Oct 08 19:33:59 crc kubenswrapper[4988]: I1008 19:33:59.319608 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5fdc957c47-mtggm"] Oct 08 19:33:59 crc kubenswrapper[4988]: I1008 19:33:59.359877 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 08 19:34:00 crc kubenswrapper[4988]: I1008 19:34:00.020852 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 08 19:34:00 crc kubenswrapper[4988]: I1008 19:34:00.252421 4988 generic.go:334] "Generic (PLEG): container finished" podID="14aa686b-2873-45e4-8c67-ae59733b51a6" containerID="11f2c8c855b5ae429ee4690e77070660f55fed41beb95245096e21790bb4d327" exitCode=0 Oct 08 19:34:00 crc kubenswrapper[4988]: I1008 19:34:00.252459 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5fdc957c47-mtggm" event={"ID":"14aa686b-2873-45e4-8c67-ae59733b51a6","Type":"ContainerDied","Data":"11f2c8c855b5ae429ee4690e77070660f55fed41beb95245096e21790bb4d327"} Oct 08 19:34:00 crc kubenswrapper[4988]: I1008 19:34:00.252706 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5fdc957c47-mtggm" event={"ID":"14aa686b-2873-45e4-8c67-ae59733b51a6","Type":"ContainerStarted","Data":"db28ff1ba831b237c6929dd2b10e2395e9578409ebf0ed76ff11f73956df8c82"} Oct 08 19:34:01 crc kubenswrapper[4988]: I1008 19:34:01.264764 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5fdc957c47-mtggm" event={"ID":"14aa686b-2873-45e4-8c67-ae59733b51a6","Type":"ContainerStarted","Data":"1b9e6f6534d0081b7d90abb344b16fbd0a10f4c080bd83ee4f8bb225f6645fe3"} Oct 08 19:34:01 crc kubenswrapper[4988]: I1008 19:34:01.265208 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-5fdc957c47-mtggm" Oct 08 19:34:01 crc kubenswrapper[4988]: I1008 19:34:01.291062 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-5fdc957c47-mtggm" podStartSLOduration=3.2910414980000002 podStartE2EDuration="3.291041498s" podCreationTimestamp="2025-10-08 19:33:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 19:34:01.289575612 +0000 UTC m=+4986.739418452" watchObservedRunningTime="2025-10-08 19:34:01.291041498 +0000 UTC m=+4986.740884268" Oct 08 19:34:03 crc kubenswrapper[4988]: I1008 19:34:03.548555 4988 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-server-0" podUID="c754aa11-b846-4860-a9c7-9e3a12be054e" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.247:5671: connect: connection refused" Oct 08 19:34:03 crc kubenswrapper[4988]: I1008 19:34:03.560466 4988 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-server-0" podUID="c754aa11-b846-4860-a9c7-9e3a12be054e" containerName="rabbitmq" containerID="cri-o://28d7c34187c55210cc93cc0372a468595ac8a8f5f7dfde181ec13d9210f0055a" gracePeriod=604796 Oct 08 19:34:04 crc kubenswrapper[4988]: I1008 19:34:04.633446 4988 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-cell1-server-0" podUID="111b6cdd-1e77-4992-a331-c66bc9e95a90" containerName="rabbitmq" containerID="cri-o://3315a80b1321e1fb678ae4d9697fad7f24942f895c83ae0783014fc111a3463c" gracePeriod=604796 Oct 08 19:34:08 crc kubenswrapper[4988]: I1008 19:34:08.912607 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-5fdc957c47-mtggm" Oct 08 19:34:08 crc kubenswrapper[4988]: I1008 19:34:08.971010 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-67d9f7fb89-lf6gm"] Oct 08 19:34:08 crc kubenswrapper[4988]: I1008 19:34:08.971286 4988 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-67d9f7fb89-lf6gm" podUID="fa2476ec-d5d4-49e5-8ef0-8ae8ccfa3993" containerName="dnsmasq-dns" containerID="cri-o://595680900cdc256c43ae6fa59ae84429550508e7fbf5148fcce9c333973e9db1" gracePeriod=10 Oct 08 19:34:09 crc kubenswrapper[4988]: I1008 19:34:09.334228 4988 generic.go:334] "Generic (PLEG): container finished" podID="fa2476ec-d5d4-49e5-8ef0-8ae8ccfa3993" containerID="595680900cdc256c43ae6fa59ae84429550508e7fbf5148fcce9c333973e9db1" exitCode=0 Oct 08 19:34:09 crc kubenswrapper[4988]: I1008 19:34:09.334456 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-67d9f7fb89-lf6gm" event={"ID":"fa2476ec-d5d4-49e5-8ef0-8ae8ccfa3993","Type":"ContainerDied","Data":"595680900cdc256c43ae6fa59ae84429550508e7fbf5148fcce9c333973e9db1"} Oct 08 19:34:09 crc kubenswrapper[4988]: I1008 19:34:09.438050 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-67d9f7fb89-lf6gm" Oct 08 19:34:09 crc kubenswrapper[4988]: I1008 19:34:09.611455 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/fa2476ec-d5d4-49e5-8ef0-8ae8ccfa3993-dns-svc\") pod \"fa2476ec-d5d4-49e5-8ef0-8ae8ccfa3993\" (UID: \"fa2476ec-d5d4-49e5-8ef0-8ae8ccfa3993\") " Oct 08 19:34:09 crc kubenswrapper[4988]: I1008 19:34:09.611498 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-klrcv\" (UniqueName: \"kubernetes.io/projected/fa2476ec-d5d4-49e5-8ef0-8ae8ccfa3993-kube-api-access-klrcv\") pod \"fa2476ec-d5d4-49e5-8ef0-8ae8ccfa3993\" (UID: \"fa2476ec-d5d4-49e5-8ef0-8ae8ccfa3993\") " Oct 08 19:34:09 crc kubenswrapper[4988]: I1008 19:34:09.611570 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fa2476ec-d5d4-49e5-8ef0-8ae8ccfa3993-config\") pod \"fa2476ec-d5d4-49e5-8ef0-8ae8ccfa3993\" (UID: \"fa2476ec-d5d4-49e5-8ef0-8ae8ccfa3993\") " Oct 08 19:34:09 crc kubenswrapper[4988]: I1008 19:34:09.619863 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fa2476ec-d5d4-49e5-8ef0-8ae8ccfa3993-kube-api-access-klrcv" (OuterVolumeSpecName: "kube-api-access-klrcv") pod "fa2476ec-d5d4-49e5-8ef0-8ae8ccfa3993" (UID: "fa2476ec-d5d4-49e5-8ef0-8ae8ccfa3993"). InnerVolumeSpecName "kube-api-access-klrcv". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 19:34:09 crc kubenswrapper[4988]: I1008 19:34:09.666619 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fa2476ec-d5d4-49e5-8ef0-8ae8ccfa3993-config" (OuterVolumeSpecName: "config") pod "fa2476ec-d5d4-49e5-8ef0-8ae8ccfa3993" (UID: "fa2476ec-d5d4-49e5-8ef0-8ae8ccfa3993"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 19:34:09 crc kubenswrapper[4988]: I1008 19:34:09.682310 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fa2476ec-d5d4-49e5-8ef0-8ae8ccfa3993-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "fa2476ec-d5d4-49e5-8ef0-8ae8ccfa3993" (UID: "fa2476ec-d5d4-49e5-8ef0-8ae8ccfa3993"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 19:34:09 crc kubenswrapper[4988]: I1008 19:34:09.714360 4988 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fa2476ec-d5d4-49e5-8ef0-8ae8ccfa3993-config\") on node \"crc\" DevicePath \"\"" Oct 08 19:34:09 crc kubenswrapper[4988]: I1008 19:34:09.715184 4988 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/fa2476ec-d5d4-49e5-8ef0-8ae8ccfa3993-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 08 19:34:09 crc kubenswrapper[4988]: I1008 19:34:09.715434 4988 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-klrcv\" (UniqueName: \"kubernetes.io/projected/fa2476ec-d5d4-49e5-8ef0-8ae8ccfa3993-kube-api-access-klrcv\") on node \"crc\" DevicePath \"\"" Oct 08 19:34:10 crc kubenswrapper[4988]: I1008 19:34:10.328256 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Oct 08 19:34:10 crc kubenswrapper[4988]: I1008 19:34:10.400415 4988 generic.go:334] "Generic (PLEG): container finished" podID="c754aa11-b846-4860-a9c7-9e3a12be054e" containerID="28d7c34187c55210cc93cc0372a468595ac8a8f5f7dfde181ec13d9210f0055a" exitCode=0 Oct 08 19:34:10 crc kubenswrapper[4988]: I1008 19:34:10.400499 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Oct 08 19:34:10 crc kubenswrapper[4988]: I1008 19:34:10.400515 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"c754aa11-b846-4860-a9c7-9e3a12be054e","Type":"ContainerDied","Data":"28d7c34187c55210cc93cc0372a468595ac8a8f5f7dfde181ec13d9210f0055a"} Oct 08 19:34:10 crc kubenswrapper[4988]: I1008 19:34:10.400571 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"c754aa11-b846-4860-a9c7-9e3a12be054e","Type":"ContainerDied","Data":"409780b0c81127be49c2b132287e43ad5221b0ee4c461ceb727454e4e1a41bec"} Oct 08 19:34:10 crc kubenswrapper[4988]: I1008 19:34:10.400593 4988 scope.go:117] "RemoveContainer" containerID="28d7c34187c55210cc93cc0372a468595ac8a8f5f7dfde181ec13d9210f0055a" Oct 08 19:34:10 crc kubenswrapper[4988]: I1008 19:34:10.405203 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-67d9f7fb89-lf6gm" event={"ID":"fa2476ec-d5d4-49e5-8ef0-8ae8ccfa3993","Type":"ContainerDied","Data":"05262201f8b1cab31c28ca9335c65bc376d371025cccabf49b1493da4092d10e"} Oct 08 19:34:10 crc kubenswrapper[4988]: I1008 19:34:10.405362 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-67d9f7fb89-lf6gm" Oct 08 19:34:10 crc kubenswrapper[4988]: I1008 19:34:10.427230 4988 scope.go:117] "RemoveContainer" containerID="9345a690467ea890e581ad94aa2b8e47eeb1adbd92e069d7d3fad8698f98fc51" Oct 08 19:34:10 crc kubenswrapper[4988]: I1008 19:34:10.453227 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-67d9f7fb89-lf6gm"] Oct 08 19:34:10 crc kubenswrapper[4988]: I1008 19:34:10.457756 4988 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-67d9f7fb89-lf6gm"] Oct 08 19:34:10 crc kubenswrapper[4988]: I1008 19:34:10.479058 4988 scope.go:117] "RemoveContainer" containerID="28d7c34187c55210cc93cc0372a468595ac8a8f5f7dfde181ec13d9210f0055a" Oct 08 19:34:10 crc kubenswrapper[4988]: E1008 19:34:10.479560 4988 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"28d7c34187c55210cc93cc0372a468595ac8a8f5f7dfde181ec13d9210f0055a\": container with ID starting with 28d7c34187c55210cc93cc0372a468595ac8a8f5f7dfde181ec13d9210f0055a not found: ID does not exist" containerID="28d7c34187c55210cc93cc0372a468595ac8a8f5f7dfde181ec13d9210f0055a" Oct 08 19:34:10 crc kubenswrapper[4988]: I1008 19:34:10.479592 4988 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"28d7c34187c55210cc93cc0372a468595ac8a8f5f7dfde181ec13d9210f0055a"} err="failed to get container status \"28d7c34187c55210cc93cc0372a468595ac8a8f5f7dfde181ec13d9210f0055a\": rpc error: code = NotFound desc = could not find container \"28d7c34187c55210cc93cc0372a468595ac8a8f5f7dfde181ec13d9210f0055a\": container with ID starting with 28d7c34187c55210cc93cc0372a468595ac8a8f5f7dfde181ec13d9210f0055a not found: ID does not exist" Oct 08 19:34:10 crc kubenswrapper[4988]: I1008 19:34:10.479615 4988 scope.go:117] "RemoveContainer" containerID="9345a690467ea890e581ad94aa2b8e47eeb1adbd92e069d7d3fad8698f98fc51" Oct 08 19:34:10 crc kubenswrapper[4988]: E1008 19:34:10.479820 4988 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9345a690467ea890e581ad94aa2b8e47eeb1adbd92e069d7d3fad8698f98fc51\": container with ID starting with 9345a690467ea890e581ad94aa2b8e47eeb1adbd92e069d7d3fad8698f98fc51 not found: ID does not exist" containerID="9345a690467ea890e581ad94aa2b8e47eeb1adbd92e069d7d3fad8698f98fc51" Oct 08 19:34:10 crc kubenswrapper[4988]: I1008 19:34:10.479854 4988 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9345a690467ea890e581ad94aa2b8e47eeb1adbd92e069d7d3fad8698f98fc51"} err="failed to get container status \"9345a690467ea890e581ad94aa2b8e47eeb1adbd92e069d7d3fad8698f98fc51\": rpc error: code = NotFound desc = could not find container \"9345a690467ea890e581ad94aa2b8e47eeb1adbd92e069d7d3fad8698f98fc51\": container with ID starting with 9345a690467ea890e581ad94aa2b8e47eeb1adbd92e069d7d3fad8698f98fc51 not found: ID does not exist" Oct 08 19:34:10 crc kubenswrapper[4988]: I1008 19:34:10.479872 4988 scope.go:117] "RemoveContainer" containerID="595680900cdc256c43ae6fa59ae84429550508e7fbf5148fcce9c333973e9db1" Oct 08 19:34:10 crc kubenswrapper[4988]: I1008 19:34:10.501001 4988 scope.go:117] "RemoveContainer" containerID="b9e0aee791a48280c1b9ca382d7c0eed368e87e4c349d8ba9e39177851280fac" Oct 08 19:34:10 crc kubenswrapper[4988]: I1008 19:34:10.529099 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/c754aa11-b846-4860-a9c7-9e3a12be054e-rabbitmq-plugins\") pod \"c754aa11-b846-4860-a9c7-9e3a12be054e\" (UID: \"c754aa11-b846-4860-a9c7-9e3a12be054e\") " Oct 08 19:34:10 crc kubenswrapper[4988]: I1008 19:34:10.529145 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/c754aa11-b846-4860-a9c7-9e3a12be054e-rabbitmq-confd\") pod \"c754aa11-b846-4860-a9c7-9e3a12be054e\" (UID: \"c754aa11-b846-4860-a9c7-9e3a12be054e\") " Oct 08 19:34:10 crc kubenswrapper[4988]: I1008 19:34:10.529181 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/c754aa11-b846-4860-a9c7-9e3a12be054e-rabbitmq-erlang-cookie\") pod \"c754aa11-b846-4860-a9c7-9e3a12be054e\" (UID: \"c754aa11-b846-4860-a9c7-9e3a12be054e\") " Oct 08 19:34:10 crc kubenswrapper[4988]: I1008 19:34:10.529234 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n4gpj\" (UniqueName: \"kubernetes.io/projected/c754aa11-b846-4860-a9c7-9e3a12be054e-kube-api-access-n4gpj\") pod \"c754aa11-b846-4860-a9c7-9e3a12be054e\" (UID: \"c754aa11-b846-4860-a9c7-9e3a12be054e\") " Oct 08 19:34:10 crc kubenswrapper[4988]: I1008 19:34:10.529304 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/c754aa11-b846-4860-a9c7-9e3a12be054e-plugins-conf\") pod \"c754aa11-b846-4860-a9c7-9e3a12be054e\" (UID: \"c754aa11-b846-4860-a9c7-9e3a12be054e\") " Oct 08 19:34:10 crc kubenswrapper[4988]: I1008 19:34:10.529328 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/c754aa11-b846-4860-a9c7-9e3a12be054e-erlang-cookie-secret\") pod \"c754aa11-b846-4860-a9c7-9e3a12be054e\" (UID: \"c754aa11-b846-4860-a9c7-9e3a12be054e\") " Oct 08 19:34:10 crc kubenswrapper[4988]: I1008 19:34:10.529359 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/c754aa11-b846-4860-a9c7-9e3a12be054e-server-conf\") pod \"c754aa11-b846-4860-a9c7-9e3a12be054e\" (UID: \"c754aa11-b846-4860-a9c7-9e3a12be054e\") " Oct 08 19:34:10 crc kubenswrapper[4988]: I1008 19:34:10.529516 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-a769d776-019d-49d3-8817-18554b6efd94\") pod \"c754aa11-b846-4860-a9c7-9e3a12be054e\" (UID: \"c754aa11-b846-4860-a9c7-9e3a12be054e\") " Oct 08 19:34:10 crc kubenswrapper[4988]: I1008 19:34:10.529560 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/c754aa11-b846-4860-a9c7-9e3a12be054e-config-data\") pod \"c754aa11-b846-4860-a9c7-9e3a12be054e\" (UID: \"c754aa11-b846-4860-a9c7-9e3a12be054e\") " Oct 08 19:34:10 crc kubenswrapper[4988]: I1008 19:34:10.529586 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/c754aa11-b846-4860-a9c7-9e3a12be054e-rabbitmq-tls\") pod \"c754aa11-b846-4860-a9c7-9e3a12be054e\" (UID: \"c754aa11-b846-4860-a9c7-9e3a12be054e\") " Oct 08 19:34:10 crc kubenswrapper[4988]: I1008 19:34:10.529619 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/c754aa11-b846-4860-a9c7-9e3a12be054e-pod-info\") pod \"c754aa11-b846-4860-a9c7-9e3a12be054e\" (UID: \"c754aa11-b846-4860-a9c7-9e3a12be054e\") " Oct 08 19:34:10 crc kubenswrapper[4988]: I1008 19:34:10.529732 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c754aa11-b846-4860-a9c7-9e3a12be054e-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "c754aa11-b846-4860-a9c7-9e3a12be054e" (UID: "c754aa11-b846-4860-a9c7-9e3a12be054e"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 19:34:10 crc kubenswrapper[4988]: I1008 19:34:10.529932 4988 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/c754aa11-b846-4860-a9c7-9e3a12be054e-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Oct 08 19:34:10 crc kubenswrapper[4988]: I1008 19:34:10.531291 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c754aa11-b846-4860-a9c7-9e3a12be054e-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "c754aa11-b846-4860-a9c7-9e3a12be054e" (UID: "c754aa11-b846-4860-a9c7-9e3a12be054e"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 19:34:10 crc kubenswrapper[4988]: I1008 19:34:10.533193 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c754aa11-b846-4860-a9c7-9e3a12be054e-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "c754aa11-b846-4860-a9c7-9e3a12be054e" (UID: "c754aa11-b846-4860-a9c7-9e3a12be054e"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 19:34:10 crc kubenswrapper[4988]: I1008 19:34:10.537520 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c754aa11-b846-4860-a9c7-9e3a12be054e-kube-api-access-n4gpj" (OuterVolumeSpecName: "kube-api-access-n4gpj") pod "c754aa11-b846-4860-a9c7-9e3a12be054e" (UID: "c754aa11-b846-4860-a9c7-9e3a12be054e"). InnerVolumeSpecName "kube-api-access-n4gpj". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 19:34:10 crc kubenswrapper[4988]: I1008 19:34:10.538236 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c754aa11-b846-4860-a9c7-9e3a12be054e-rabbitmq-tls" (OuterVolumeSpecName: "rabbitmq-tls") pod "c754aa11-b846-4860-a9c7-9e3a12be054e" (UID: "c754aa11-b846-4860-a9c7-9e3a12be054e"). InnerVolumeSpecName "rabbitmq-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 19:34:10 crc kubenswrapper[4988]: I1008 19:34:10.539941 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/c754aa11-b846-4860-a9c7-9e3a12be054e-pod-info" (OuterVolumeSpecName: "pod-info") pod "c754aa11-b846-4860-a9c7-9e3a12be054e" (UID: "c754aa11-b846-4860-a9c7-9e3a12be054e"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Oct 08 19:34:10 crc kubenswrapper[4988]: I1008 19:34:10.540602 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c754aa11-b846-4860-a9c7-9e3a12be054e-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "c754aa11-b846-4860-a9c7-9e3a12be054e" (UID: "c754aa11-b846-4860-a9c7-9e3a12be054e"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 19:34:10 crc kubenswrapper[4988]: I1008 19:34:10.548831 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-a769d776-019d-49d3-8817-18554b6efd94" (OuterVolumeSpecName: "persistence") pod "c754aa11-b846-4860-a9c7-9e3a12be054e" (UID: "c754aa11-b846-4860-a9c7-9e3a12be054e"). InnerVolumeSpecName "pvc-a769d776-019d-49d3-8817-18554b6efd94". PluginName "kubernetes.io/csi", VolumeGidValue "" Oct 08 19:34:10 crc kubenswrapper[4988]: I1008 19:34:10.559467 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c754aa11-b846-4860-a9c7-9e3a12be054e-config-data" (OuterVolumeSpecName: "config-data") pod "c754aa11-b846-4860-a9c7-9e3a12be054e" (UID: "c754aa11-b846-4860-a9c7-9e3a12be054e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 19:34:10 crc kubenswrapper[4988]: I1008 19:34:10.573598 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c754aa11-b846-4860-a9c7-9e3a12be054e-server-conf" (OuterVolumeSpecName: "server-conf") pod "c754aa11-b846-4860-a9c7-9e3a12be054e" (UID: "c754aa11-b846-4860-a9c7-9e3a12be054e"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 19:34:10 crc kubenswrapper[4988]: I1008 19:34:10.607450 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c754aa11-b846-4860-a9c7-9e3a12be054e-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "c754aa11-b846-4860-a9c7-9e3a12be054e" (UID: "c754aa11-b846-4860-a9c7-9e3a12be054e"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 19:34:10 crc kubenswrapper[4988]: I1008 19:34:10.631349 4988 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/c754aa11-b846-4860-a9c7-9e3a12be054e-rabbitmq-tls\") on node \"crc\" DevicePath \"\"" Oct 08 19:34:10 crc kubenswrapper[4988]: I1008 19:34:10.631421 4988 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/c754aa11-b846-4860-a9c7-9e3a12be054e-pod-info\") on node \"crc\" DevicePath \"\"" Oct 08 19:34:10 crc kubenswrapper[4988]: I1008 19:34:10.631433 4988 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/c754aa11-b846-4860-a9c7-9e3a12be054e-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Oct 08 19:34:10 crc kubenswrapper[4988]: I1008 19:34:10.631449 4988 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/c754aa11-b846-4860-a9c7-9e3a12be054e-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Oct 08 19:34:10 crc kubenswrapper[4988]: I1008 19:34:10.631461 4988 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n4gpj\" (UniqueName: \"kubernetes.io/projected/c754aa11-b846-4860-a9c7-9e3a12be054e-kube-api-access-n4gpj\") on node \"crc\" DevicePath \"\"" Oct 08 19:34:10 crc kubenswrapper[4988]: I1008 19:34:10.631479 4988 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/c754aa11-b846-4860-a9c7-9e3a12be054e-plugins-conf\") on node \"crc\" DevicePath \"\"" Oct 08 19:34:10 crc kubenswrapper[4988]: I1008 19:34:10.631490 4988 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/c754aa11-b846-4860-a9c7-9e3a12be054e-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Oct 08 19:34:10 crc kubenswrapper[4988]: I1008 19:34:10.631501 4988 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/c754aa11-b846-4860-a9c7-9e3a12be054e-server-conf\") on node \"crc\" DevicePath \"\"" Oct 08 19:34:10 crc kubenswrapper[4988]: I1008 19:34:10.631558 4988 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"pvc-a769d776-019d-49d3-8817-18554b6efd94\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-a769d776-019d-49d3-8817-18554b6efd94\") on node \"crc\" " Oct 08 19:34:10 crc kubenswrapper[4988]: I1008 19:34:10.631574 4988 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/c754aa11-b846-4860-a9c7-9e3a12be054e-config-data\") on node \"crc\" DevicePath \"\"" Oct 08 19:34:10 crc kubenswrapper[4988]: I1008 19:34:10.651783 4988 csi_attacher.go:630] kubernetes.io/csi: attacher.UnmountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping UnmountDevice... Oct 08 19:34:10 crc kubenswrapper[4988]: I1008 19:34:10.651965 4988 operation_generator.go:917] UnmountDevice succeeded for volume "pvc-a769d776-019d-49d3-8817-18554b6efd94" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-a769d776-019d-49d3-8817-18554b6efd94") on node "crc" Oct 08 19:34:10 crc kubenswrapper[4988]: I1008 19:34:10.734590 4988 reconciler_common.go:293] "Volume detached for volume \"pvc-a769d776-019d-49d3-8817-18554b6efd94\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-a769d776-019d-49d3-8817-18554b6efd94\") on node \"crc\" DevicePath \"\"" Oct 08 19:34:10 crc kubenswrapper[4988]: I1008 19:34:10.744096 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 08 19:34:10 crc kubenswrapper[4988]: I1008 19:34:10.751070 4988 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 08 19:34:10 crc kubenswrapper[4988]: I1008 19:34:10.767890 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-0"] Oct 08 19:34:10 crc kubenswrapper[4988]: E1008 19:34:10.770188 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fa2476ec-d5d4-49e5-8ef0-8ae8ccfa3993" containerName="init" Oct 08 19:34:10 crc kubenswrapper[4988]: I1008 19:34:10.770222 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="fa2476ec-d5d4-49e5-8ef0-8ae8ccfa3993" containerName="init" Oct 08 19:34:10 crc kubenswrapper[4988]: E1008 19:34:10.770248 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c754aa11-b846-4860-a9c7-9e3a12be054e" containerName="setup-container" Oct 08 19:34:10 crc kubenswrapper[4988]: I1008 19:34:10.770255 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="c754aa11-b846-4860-a9c7-9e3a12be054e" containerName="setup-container" Oct 08 19:34:10 crc kubenswrapper[4988]: E1008 19:34:10.770277 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fa2476ec-d5d4-49e5-8ef0-8ae8ccfa3993" containerName="dnsmasq-dns" Oct 08 19:34:10 crc kubenswrapper[4988]: I1008 19:34:10.770284 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="fa2476ec-d5d4-49e5-8ef0-8ae8ccfa3993" containerName="dnsmasq-dns" Oct 08 19:34:10 crc kubenswrapper[4988]: E1008 19:34:10.770293 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c754aa11-b846-4860-a9c7-9e3a12be054e" containerName="rabbitmq" Oct 08 19:34:10 crc kubenswrapper[4988]: I1008 19:34:10.770299 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="c754aa11-b846-4860-a9c7-9e3a12be054e" containerName="rabbitmq" Oct 08 19:34:10 crc kubenswrapper[4988]: I1008 19:34:10.770469 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="fa2476ec-d5d4-49e5-8ef0-8ae8ccfa3993" containerName="dnsmasq-dns" Oct 08 19:34:10 crc kubenswrapper[4988]: I1008 19:34:10.770480 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="c754aa11-b846-4860-a9c7-9e3a12be054e" containerName="rabbitmq" Oct 08 19:34:10 crc kubenswrapper[4988]: I1008 19:34:10.771506 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Oct 08 19:34:10 crc kubenswrapper[4988]: I1008 19:34:10.774514 4988 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-plugins-conf" Oct 08 19:34:10 crc kubenswrapper[4988]: I1008 19:34:10.774845 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-default-user" Oct 08 19:34:10 crc kubenswrapper[4988]: I1008 19:34:10.775563 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-svc" Oct 08 19:34:10 crc kubenswrapper[4988]: I1008 19:34:10.776024 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-server-dockercfg-djzh7" Oct 08 19:34:10 crc kubenswrapper[4988]: I1008 19:34:10.776234 4988 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-server-conf" Oct 08 19:34:10 crc kubenswrapper[4988]: I1008 19:34:10.776502 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-erlang-cookie" Oct 08 19:34:10 crc kubenswrapper[4988]: I1008 19:34:10.776697 4988 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-config-data" Oct 08 19:34:10 crc kubenswrapper[4988]: I1008 19:34:10.790704 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 08 19:34:10 crc kubenswrapper[4988]: I1008 19:34:10.938498 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/9d6674d9-eacf-4d44-91db-5ace88fa9a01-server-conf\") pod \"rabbitmq-server-0\" (UID: \"9d6674d9-eacf-4d44-91db-5ace88fa9a01\") " pod="openstack/rabbitmq-server-0" Oct 08 19:34:10 crc kubenswrapper[4988]: I1008 19:34:10.939261 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/9d6674d9-eacf-4d44-91db-5ace88fa9a01-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"9d6674d9-eacf-4d44-91db-5ace88fa9a01\") " pod="openstack/rabbitmq-server-0" Oct 08 19:34:10 crc kubenswrapper[4988]: I1008 19:34:10.939318 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/9d6674d9-eacf-4d44-91db-5ace88fa9a01-pod-info\") pod \"rabbitmq-server-0\" (UID: \"9d6674d9-eacf-4d44-91db-5ace88fa9a01\") " pod="openstack/rabbitmq-server-0" Oct 08 19:34:10 crc kubenswrapper[4988]: I1008 19:34:10.939498 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/9d6674d9-eacf-4d44-91db-5ace88fa9a01-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"9d6674d9-eacf-4d44-91db-5ace88fa9a01\") " pod="openstack/rabbitmq-server-0" Oct 08 19:34:10 crc kubenswrapper[4988]: I1008 19:34:10.939578 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/9d6674d9-eacf-4d44-91db-5ace88fa9a01-config-data\") pod \"rabbitmq-server-0\" (UID: \"9d6674d9-eacf-4d44-91db-5ace88fa9a01\") " pod="openstack/rabbitmq-server-0" Oct 08 19:34:10 crc kubenswrapper[4988]: I1008 19:34:10.939681 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-a769d776-019d-49d3-8817-18554b6efd94\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-a769d776-019d-49d3-8817-18554b6efd94\") pod \"rabbitmq-server-0\" (UID: \"9d6674d9-eacf-4d44-91db-5ace88fa9a01\") " pod="openstack/rabbitmq-server-0" Oct 08 19:34:10 crc kubenswrapper[4988]: I1008 19:34:10.939745 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/9d6674d9-eacf-4d44-91db-5ace88fa9a01-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"9d6674d9-eacf-4d44-91db-5ace88fa9a01\") " pod="openstack/rabbitmq-server-0" Oct 08 19:34:10 crc kubenswrapper[4988]: I1008 19:34:10.939832 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bfzhd\" (UniqueName: \"kubernetes.io/projected/9d6674d9-eacf-4d44-91db-5ace88fa9a01-kube-api-access-bfzhd\") pod \"rabbitmq-server-0\" (UID: \"9d6674d9-eacf-4d44-91db-5ace88fa9a01\") " pod="openstack/rabbitmq-server-0" Oct 08 19:34:10 crc kubenswrapper[4988]: I1008 19:34:10.939889 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/9d6674d9-eacf-4d44-91db-5ace88fa9a01-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"9d6674d9-eacf-4d44-91db-5ace88fa9a01\") " pod="openstack/rabbitmq-server-0" Oct 08 19:34:10 crc kubenswrapper[4988]: I1008 19:34:10.939992 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/9d6674d9-eacf-4d44-91db-5ace88fa9a01-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"9d6674d9-eacf-4d44-91db-5ace88fa9a01\") " pod="openstack/rabbitmq-server-0" Oct 08 19:34:10 crc kubenswrapper[4988]: I1008 19:34:10.940106 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/9d6674d9-eacf-4d44-91db-5ace88fa9a01-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"9d6674d9-eacf-4d44-91db-5ace88fa9a01\") " pod="openstack/rabbitmq-server-0" Oct 08 19:34:11 crc kubenswrapper[4988]: I1008 19:34:11.048987 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/9d6674d9-eacf-4d44-91db-5ace88fa9a01-server-conf\") pod \"rabbitmq-server-0\" (UID: \"9d6674d9-eacf-4d44-91db-5ace88fa9a01\") " pod="openstack/rabbitmq-server-0" Oct 08 19:34:11 crc kubenswrapper[4988]: I1008 19:34:11.049054 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/9d6674d9-eacf-4d44-91db-5ace88fa9a01-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"9d6674d9-eacf-4d44-91db-5ace88fa9a01\") " pod="openstack/rabbitmq-server-0" Oct 08 19:34:11 crc kubenswrapper[4988]: I1008 19:34:11.049075 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/9d6674d9-eacf-4d44-91db-5ace88fa9a01-pod-info\") pod \"rabbitmq-server-0\" (UID: \"9d6674d9-eacf-4d44-91db-5ace88fa9a01\") " pod="openstack/rabbitmq-server-0" Oct 08 19:34:11 crc kubenswrapper[4988]: I1008 19:34:11.049130 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/9d6674d9-eacf-4d44-91db-5ace88fa9a01-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"9d6674d9-eacf-4d44-91db-5ace88fa9a01\") " pod="openstack/rabbitmq-server-0" Oct 08 19:34:11 crc kubenswrapper[4988]: I1008 19:34:11.049158 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/9d6674d9-eacf-4d44-91db-5ace88fa9a01-config-data\") pod \"rabbitmq-server-0\" (UID: \"9d6674d9-eacf-4d44-91db-5ace88fa9a01\") " pod="openstack/rabbitmq-server-0" Oct 08 19:34:11 crc kubenswrapper[4988]: I1008 19:34:11.049198 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-a769d776-019d-49d3-8817-18554b6efd94\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-a769d776-019d-49d3-8817-18554b6efd94\") pod \"rabbitmq-server-0\" (UID: \"9d6674d9-eacf-4d44-91db-5ace88fa9a01\") " pod="openstack/rabbitmq-server-0" Oct 08 19:34:11 crc kubenswrapper[4988]: I1008 19:34:11.049218 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/9d6674d9-eacf-4d44-91db-5ace88fa9a01-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"9d6674d9-eacf-4d44-91db-5ace88fa9a01\") " pod="openstack/rabbitmq-server-0" Oct 08 19:34:11 crc kubenswrapper[4988]: I1008 19:34:11.049259 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bfzhd\" (UniqueName: \"kubernetes.io/projected/9d6674d9-eacf-4d44-91db-5ace88fa9a01-kube-api-access-bfzhd\") pod \"rabbitmq-server-0\" (UID: \"9d6674d9-eacf-4d44-91db-5ace88fa9a01\") " pod="openstack/rabbitmq-server-0" Oct 08 19:34:11 crc kubenswrapper[4988]: I1008 19:34:11.049275 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/9d6674d9-eacf-4d44-91db-5ace88fa9a01-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"9d6674d9-eacf-4d44-91db-5ace88fa9a01\") " pod="openstack/rabbitmq-server-0" Oct 08 19:34:11 crc kubenswrapper[4988]: I1008 19:34:11.049314 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/9d6674d9-eacf-4d44-91db-5ace88fa9a01-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"9d6674d9-eacf-4d44-91db-5ace88fa9a01\") " pod="openstack/rabbitmq-server-0" Oct 08 19:34:11 crc kubenswrapper[4988]: I1008 19:34:11.049368 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/9d6674d9-eacf-4d44-91db-5ace88fa9a01-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"9d6674d9-eacf-4d44-91db-5ace88fa9a01\") " pod="openstack/rabbitmq-server-0" Oct 08 19:34:11 crc kubenswrapper[4988]: I1008 19:34:11.051014 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/9d6674d9-eacf-4d44-91db-5ace88fa9a01-config-data\") pod \"rabbitmq-server-0\" (UID: \"9d6674d9-eacf-4d44-91db-5ace88fa9a01\") " pod="openstack/rabbitmq-server-0" Oct 08 19:34:11 crc kubenswrapper[4988]: I1008 19:34:11.052107 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/9d6674d9-eacf-4d44-91db-5ace88fa9a01-server-conf\") pod \"rabbitmq-server-0\" (UID: \"9d6674d9-eacf-4d44-91db-5ace88fa9a01\") " pod="openstack/rabbitmq-server-0" Oct 08 19:34:11 crc kubenswrapper[4988]: I1008 19:34:11.053887 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/9d6674d9-eacf-4d44-91db-5ace88fa9a01-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"9d6674d9-eacf-4d44-91db-5ace88fa9a01\") " pod="openstack/rabbitmq-server-0" Oct 08 19:34:11 crc kubenswrapper[4988]: I1008 19:34:11.054056 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/9d6674d9-eacf-4d44-91db-5ace88fa9a01-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"9d6674d9-eacf-4d44-91db-5ace88fa9a01\") " pod="openstack/rabbitmq-server-0" Oct 08 19:34:11 crc kubenswrapper[4988]: I1008 19:34:11.054266 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/9d6674d9-eacf-4d44-91db-5ace88fa9a01-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"9d6674d9-eacf-4d44-91db-5ace88fa9a01\") " pod="openstack/rabbitmq-server-0" Oct 08 19:34:11 crc kubenswrapper[4988]: I1008 19:34:11.055136 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/9d6674d9-eacf-4d44-91db-5ace88fa9a01-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"9d6674d9-eacf-4d44-91db-5ace88fa9a01\") " pod="openstack/rabbitmq-server-0" Oct 08 19:34:11 crc kubenswrapper[4988]: I1008 19:34:11.057769 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/9d6674d9-eacf-4d44-91db-5ace88fa9a01-pod-info\") pod \"rabbitmq-server-0\" (UID: \"9d6674d9-eacf-4d44-91db-5ace88fa9a01\") " pod="openstack/rabbitmq-server-0" Oct 08 19:34:11 crc kubenswrapper[4988]: I1008 19:34:11.060347 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/9d6674d9-eacf-4d44-91db-5ace88fa9a01-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"9d6674d9-eacf-4d44-91db-5ace88fa9a01\") " pod="openstack/rabbitmq-server-0" Oct 08 19:34:11 crc kubenswrapper[4988]: I1008 19:34:11.063271 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/9d6674d9-eacf-4d44-91db-5ace88fa9a01-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"9d6674d9-eacf-4d44-91db-5ace88fa9a01\") " pod="openstack/rabbitmq-server-0" Oct 08 19:34:11 crc kubenswrapper[4988]: I1008 19:34:11.063809 4988 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Oct 08 19:34:11 crc kubenswrapper[4988]: I1008 19:34:11.063842 4988 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-a769d776-019d-49d3-8817-18554b6efd94\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-a769d776-019d-49d3-8817-18554b6efd94\") pod \"rabbitmq-server-0\" (UID: \"9d6674d9-eacf-4d44-91db-5ace88fa9a01\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/7468fb34cd0cef4ea54b9c37ebe6b505e744898f63936b052a7832fbe4f2b464/globalmount\"" pod="openstack/rabbitmq-server-0" Oct 08 19:34:11 crc kubenswrapper[4988]: I1008 19:34:11.072992 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bfzhd\" (UniqueName: \"kubernetes.io/projected/9d6674d9-eacf-4d44-91db-5ace88fa9a01-kube-api-access-bfzhd\") pod \"rabbitmq-server-0\" (UID: \"9d6674d9-eacf-4d44-91db-5ace88fa9a01\") " pod="openstack/rabbitmq-server-0" Oct 08 19:34:11 crc kubenswrapper[4988]: I1008 19:34:11.102103 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-a769d776-019d-49d3-8817-18554b6efd94\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-a769d776-019d-49d3-8817-18554b6efd94\") pod \"rabbitmq-server-0\" (UID: \"9d6674d9-eacf-4d44-91db-5ace88fa9a01\") " pod="openstack/rabbitmq-server-0" Oct 08 19:34:11 crc kubenswrapper[4988]: I1008 19:34:11.172192 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Oct 08 19:34:11 crc kubenswrapper[4988]: I1008 19:34:11.256724 4988 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c754aa11-b846-4860-a9c7-9e3a12be054e" path="/var/lib/kubelet/pods/c754aa11-b846-4860-a9c7-9e3a12be054e/volumes" Oct 08 19:34:11 crc kubenswrapper[4988]: I1008 19:34:11.258042 4988 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fa2476ec-d5d4-49e5-8ef0-8ae8ccfa3993" path="/var/lib/kubelet/pods/fa2476ec-d5d4-49e5-8ef0-8ae8ccfa3993/volumes" Oct 08 19:34:11 crc kubenswrapper[4988]: I1008 19:34:11.353334 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/111b6cdd-1e77-4992-a331-c66bc9e95a90-rabbitmq-erlang-cookie\") pod \"111b6cdd-1e77-4992-a331-c66bc9e95a90\" (UID: \"111b6cdd-1e77-4992-a331-c66bc9e95a90\") " Oct 08 19:34:11 crc kubenswrapper[4988]: I1008 19:34:11.353435 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/111b6cdd-1e77-4992-a331-c66bc9e95a90-rabbitmq-plugins\") pod \"111b6cdd-1e77-4992-a331-c66bc9e95a90\" (UID: \"111b6cdd-1e77-4992-a331-c66bc9e95a90\") " Oct 08 19:34:11 crc kubenswrapper[4988]: I1008 19:34:11.353479 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/111b6cdd-1e77-4992-a331-c66bc9e95a90-rabbitmq-tls\") pod \"111b6cdd-1e77-4992-a331-c66bc9e95a90\" (UID: \"111b6cdd-1e77-4992-a331-c66bc9e95a90\") " Oct 08 19:34:11 crc kubenswrapper[4988]: I1008 19:34:11.353522 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hqxvn\" (UniqueName: \"kubernetes.io/projected/111b6cdd-1e77-4992-a331-c66bc9e95a90-kube-api-access-hqxvn\") pod \"111b6cdd-1e77-4992-a331-c66bc9e95a90\" (UID: \"111b6cdd-1e77-4992-a331-c66bc9e95a90\") " Oct 08 19:34:11 crc kubenswrapper[4988]: I1008 19:34:11.353559 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/111b6cdd-1e77-4992-a331-c66bc9e95a90-pod-info\") pod \"111b6cdd-1e77-4992-a331-c66bc9e95a90\" (UID: \"111b6cdd-1e77-4992-a331-c66bc9e95a90\") " Oct 08 19:34:11 crc kubenswrapper[4988]: I1008 19:34:11.353678 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-ee25b6c1-7b38-49fa-97b1-836588148f84\") pod \"111b6cdd-1e77-4992-a331-c66bc9e95a90\" (UID: \"111b6cdd-1e77-4992-a331-c66bc9e95a90\") " Oct 08 19:34:11 crc kubenswrapper[4988]: I1008 19:34:11.353703 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/111b6cdd-1e77-4992-a331-c66bc9e95a90-config-data\") pod \"111b6cdd-1e77-4992-a331-c66bc9e95a90\" (UID: \"111b6cdd-1e77-4992-a331-c66bc9e95a90\") " Oct 08 19:34:11 crc kubenswrapper[4988]: I1008 19:34:11.353745 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/111b6cdd-1e77-4992-a331-c66bc9e95a90-rabbitmq-confd\") pod \"111b6cdd-1e77-4992-a331-c66bc9e95a90\" (UID: \"111b6cdd-1e77-4992-a331-c66bc9e95a90\") " Oct 08 19:34:11 crc kubenswrapper[4988]: I1008 19:34:11.353767 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/111b6cdd-1e77-4992-a331-c66bc9e95a90-erlang-cookie-secret\") pod \"111b6cdd-1e77-4992-a331-c66bc9e95a90\" (UID: \"111b6cdd-1e77-4992-a331-c66bc9e95a90\") " Oct 08 19:34:11 crc kubenswrapper[4988]: I1008 19:34:11.353803 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/111b6cdd-1e77-4992-a331-c66bc9e95a90-plugins-conf\") pod \"111b6cdd-1e77-4992-a331-c66bc9e95a90\" (UID: \"111b6cdd-1e77-4992-a331-c66bc9e95a90\") " Oct 08 19:34:11 crc kubenswrapper[4988]: I1008 19:34:11.353833 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/111b6cdd-1e77-4992-a331-c66bc9e95a90-server-conf\") pod \"111b6cdd-1e77-4992-a331-c66bc9e95a90\" (UID: \"111b6cdd-1e77-4992-a331-c66bc9e95a90\") " Oct 08 19:34:11 crc kubenswrapper[4988]: I1008 19:34:11.354033 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/111b6cdd-1e77-4992-a331-c66bc9e95a90-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "111b6cdd-1e77-4992-a331-c66bc9e95a90" (UID: "111b6cdd-1e77-4992-a331-c66bc9e95a90"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 19:34:11 crc kubenswrapper[4988]: I1008 19:34:11.354225 4988 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/111b6cdd-1e77-4992-a331-c66bc9e95a90-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Oct 08 19:34:11 crc kubenswrapper[4988]: I1008 19:34:11.354303 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/111b6cdd-1e77-4992-a331-c66bc9e95a90-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "111b6cdd-1e77-4992-a331-c66bc9e95a90" (UID: "111b6cdd-1e77-4992-a331-c66bc9e95a90"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 19:34:11 crc kubenswrapper[4988]: I1008 19:34:11.355022 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/111b6cdd-1e77-4992-a331-c66bc9e95a90-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "111b6cdd-1e77-4992-a331-c66bc9e95a90" (UID: "111b6cdd-1e77-4992-a331-c66bc9e95a90"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 19:34:11 crc kubenswrapper[4988]: I1008 19:34:11.357748 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/111b6cdd-1e77-4992-a331-c66bc9e95a90-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "111b6cdd-1e77-4992-a331-c66bc9e95a90" (UID: "111b6cdd-1e77-4992-a331-c66bc9e95a90"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 19:34:11 crc kubenswrapper[4988]: I1008 19:34:11.357775 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/111b6cdd-1e77-4992-a331-c66bc9e95a90-kube-api-access-hqxvn" (OuterVolumeSpecName: "kube-api-access-hqxvn") pod "111b6cdd-1e77-4992-a331-c66bc9e95a90" (UID: "111b6cdd-1e77-4992-a331-c66bc9e95a90"). InnerVolumeSpecName "kube-api-access-hqxvn". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 19:34:11 crc kubenswrapper[4988]: I1008 19:34:11.359071 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/111b6cdd-1e77-4992-a331-c66bc9e95a90-rabbitmq-tls" (OuterVolumeSpecName: "rabbitmq-tls") pod "111b6cdd-1e77-4992-a331-c66bc9e95a90" (UID: "111b6cdd-1e77-4992-a331-c66bc9e95a90"). InnerVolumeSpecName "rabbitmq-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 19:34:11 crc kubenswrapper[4988]: I1008 19:34:11.361598 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/111b6cdd-1e77-4992-a331-c66bc9e95a90-pod-info" (OuterVolumeSpecName: "pod-info") pod "111b6cdd-1e77-4992-a331-c66bc9e95a90" (UID: "111b6cdd-1e77-4992-a331-c66bc9e95a90"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Oct 08 19:34:11 crc kubenswrapper[4988]: I1008 19:34:11.370568 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-ee25b6c1-7b38-49fa-97b1-836588148f84" (OuterVolumeSpecName: "persistence") pod "111b6cdd-1e77-4992-a331-c66bc9e95a90" (UID: "111b6cdd-1e77-4992-a331-c66bc9e95a90"). InnerVolumeSpecName "pvc-ee25b6c1-7b38-49fa-97b1-836588148f84". PluginName "kubernetes.io/csi", VolumeGidValue "" Oct 08 19:34:11 crc kubenswrapper[4988]: I1008 19:34:11.383359 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/111b6cdd-1e77-4992-a331-c66bc9e95a90-config-data" (OuterVolumeSpecName: "config-data") pod "111b6cdd-1e77-4992-a331-c66bc9e95a90" (UID: "111b6cdd-1e77-4992-a331-c66bc9e95a90"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 19:34:11 crc kubenswrapper[4988]: I1008 19:34:11.390337 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Oct 08 19:34:11 crc kubenswrapper[4988]: I1008 19:34:11.407701 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/111b6cdd-1e77-4992-a331-c66bc9e95a90-server-conf" (OuterVolumeSpecName: "server-conf") pod "111b6cdd-1e77-4992-a331-c66bc9e95a90" (UID: "111b6cdd-1e77-4992-a331-c66bc9e95a90"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 19:34:11 crc kubenswrapper[4988]: I1008 19:34:11.418775 4988 generic.go:334] "Generic (PLEG): container finished" podID="111b6cdd-1e77-4992-a331-c66bc9e95a90" containerID="3315a80b1321e1fb678ae4d9697fad7f24942f895c83ae0783014fc111a3463c" exitCode=0 Oct 08 19:34:11 crc kubenswrapper[4988]: I1008 19:34:11.418867 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Oct 08 19:34:11 crc kubenswrapper[4988]: I1008 19:34:11.418927 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"111b6cdd-1e77-4992-a331-c66bc9e95a90","Type":"ContainerDied","Data":"3315a80b1321e1fb678ae4d9697fad7f24942f895c83ae0783014fc111a3463c"} Oct 08 19:34:11 crc kubenswrapper[4988]: I1008 19:34:11.419287 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"111b6cdd-1e77-4992-a331-c66bc9e95a90","Type":"ContainerDied","Data":"73c6f78f7d2284833ade391bb3f6d275c9404bd39d59c930323a575ad3e096ad"} Oct 08 19:34:11 crc kubenswrapper[4988]: I1008 19:34:11.419319 4988 scope.go:117] "RemoveContainer" containerID="3315a80b1321e1fb678ae4d9697fad7f24942f895c83ae0783014fc111a3463c" Oct 08 19:34:11 crc kubenswrapper[4988]: I1008 19:34:11.452933 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/111b6cdd-1e77-4992-a331-c66bc9e95a90-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "111b6cdd-1e77-4992-a331-c66bc9e95a90" (UID: "111b6cdd-1e77-4992-a331-c66bc9e95a90"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 19:34:11 crc kubenswrapper[4988]: I1008 19:34:11.455912 4988 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/111b6cdd-1e77-4992-a331-c66bc9e95a90-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Oct 08 19:34:11 crc kubenswrapper[4988]: I1008 19:34:11.455942 4988 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/111b6cdd-1e77-4992-a331-c66bc9e95a90-rabbitmq-tls\") on node \"crc\" DevicePath \"\"" Oct 08 19:34:11 crc kubenswrapper[4988]: I1008 19:34:11.455953 4988 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hqxvn\" (UniqueName: \"kubernetes.io/projected/111b6cdd-1e77-4992-a331-c66bc9e95a90-kube-api-access-hqxvn\") on node \"crc\" DevicePath \"\"" Oct 08 19:34:11 crc kubenswrapper[4988]: I1008 19:34:11.455962 4988 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/111b6cdd-1e77-4992-a331-c66bc9e95a90-pod-info\") on node \"crc\" DevicePath \"\"" Oct 08 19:34:11 crc kubenswrapper[4988]: I1008 19:34:11.455997 4988 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"pvc-ee25b6c1-7b38-49fa-97b1-836588148f84\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-ee25b6c1-7b38-49fa-97b1-836588148f84\") on node \"crc\" " Oct 08 19:34:11 crc kubenswrapper[4988]: I1008 19:34:11.456081 4988 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/111b6cdd-1e77-4992-a331-c66bc9e95a90-config-data\") on node \"crc\" DevicePath \"\"" Oct 08 19:34:11 crc kubenswrapper[4988]: I1008 19:34:11.456097 4988 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/111b6cdd-1e77-4992-a331-c66bc9e95a90-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Oct 08 19:34:11 crc kubenswrapper[4988]: I1008 19:34:11.456113 4988 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/111b6cdd-1e77-4992-a331-c66bc9e95a90-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Oct 08 19:34:11 crc kubenswrapper[4988]: I1008 19:34:11.456122 4988 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/111b6cdd-1e77-4992-a331-c66bc9e95a90-plugins-conf\") on node \"crc\" DevicePath \"\"" Oct 08 19:34:11 crc kubenswrapper[4988]: I1008 19:34:11.456130 4988 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/111b6cdd-1e77-4992-a331-c66bc9e95a90-server-conf\") on node \"crc\" DevicePath \"\"" Oct 08 19:34:11 crc kubenswrapper[4988]: I1008 19:34:11.473584 4988 scope.go:117] "RemoveContainer" containerID="6b0b30e0590363081acab72ac2c5ea5643a1db58b78fa1169b9902c4781121cb" Oct 08 19:34:11 crc kubenswrapper[4988]: I1008 19:34:11.474019 4988 csi_attacher.go:630] kubernetes.io/csi: attacher.UnmountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping UnmountDevice... Oct 08 19:34:11 crc kubenswrapper[4988]: I1008 19:34:11.474262 4988 operation_generator.go:917] UnmountDevice succeeded for volume "pvc-ee25b6c1-7b38-49fa-97b1-836588148f84" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-ee25b6c1-7b38-49fa-97b1-836588148f84") on node "crc" Oct 08 19:34:11 crc kubenswrapper[4988]: I1008 19:34:11.502992 4988 scope.go:117] "RemoveContainer" containerID="3315a80b1321e1fb678ae4d9697fad7f24942f895c83ae0783014fc111a3463c" Oct 08 19:34:11 crc kubenswrapper[4988]: E1008 19:34:11.505859 4988 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3315a80b1321e1fb678ae4d9697fad7f24942f895c83ae0783014fc111a3463c\": container with ID starting with 3315a80b1321e1fb678ae4d9697fad7f24942f895c83ae0783014fc111a3463c not found: ID does not exist" containerID="3315a80b1321e1fb678ae4d9697fad7f24942f895c83ae0783014fc111a3463c" Oct 08 19:34:11 crc kubenswrapper[4988]: I1008 19:34:11.505900 4988 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3315a80b1321e1fb678ae4d9697fad7f24942f895c83ae0783014fc111a3463c"} err="failed to get container status \"3315a80b1321e1fb678ae4d9697fad7f24942f895c83ae0783014fc111a3463c\": rpc error: code = NotFound desc = could not find container \"3315a80b1321e1fb678ae4d9697fad7f24942f895c83ae0783014fc111a3463c\": container with ID starting with 3315a80b1321e1fb678ae4d9697fad7f24942f895c83ae0783014fc111a3463c not found: ID does not exist" Oct 08 19:34:11 crc kubenswrapper[4988]: I1008 19:34:11.505932 4988 scope.go:117] "RemoveContainer" containerID="6b0b30e0590363081acab72ac2c5ea5643a1db58b78fa1169b9902c4781121cb" Oct 08 19:34:11 crc kubenswrapper[4988]: E1008 19:34:11.506368 4988 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6b0b30e0590363081acab72ac2c5ea5643a1db58b78fa1169b9902c4781121cb\": container with ID starting with 6b0b30e0590363081acab72ac2c5ea5643a1db58b78fa1169b9902c4781121cb not found: ID does not exist" containerID="6b0b30e0590363081acab72ac2c5ea5643a1db58b78fa1169b9902c4781121cb" Oct 08 19:34:11 crc kubenswrapper[4988]: I1008 19:34:11.506422 4988 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6b0b30e0590363081acab72ac2c5ea5643a1db58b78fa1169b9902c4781121cb"} err="failed to get container status \"6b0b30e0590363081acab72ac2c5ea5643a1db58b78fa1169b9902c4781121cb\": rpc error: code = NotFound desc = could not find container \"6b0b30e0590363081acab72ac2c5ea5643a1db58b78fa1169b9902c4781121cb\": container with ID starting with 6b0b30e0590363081acab72ac2c5ea5643a1db58b78fa1169b9902c4781121cb not found: ID does not exist" Oct 08 19:34:11 crc kubenswrapper[4988]: I1008 19:34:11.557610 4988 reconciler_common.go:293] "Volume detached for volume \"pvc-ee25b6c1-7b38-49fa-97b1-836588148f84\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-ee25b6c1-7b38-49fa-97b1-836588148f84\") on node \"crc\" DevicePath \"\"" Oct 08 19:34:11 crc kubenswrapper[4988]: I1008 19:34:11.749178 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 08 19:34:11 crc kubenswrapper[4988]: I1008 19:34:11.753706 4988 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 08 19:34:11 crc kubenswrapper[4988]: I1008 19:34:11.784309 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 08 19:34:11 crc kubenswrapper[4988]: E1008 19:34:11.784719 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="111b6cdd-1e77-4992-a331-c66bc9e95a90" containerName="setup-container" Oct 08 19:34:11 crc kubenswrapper[4988]: I1008 19:34:11.784743 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="111b6cdd-1e77-4992-a331-c66bc9e95a90" containerName="setup-container" Oct 08 19:34:11 crc kubenswrapper[4988]: E1008 19:34:11.784756 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="111b6cdd-1e77-4992-a331-c66bc9e95a90" containerName="rabbitmq" Oct 08 19:34:11 crc kubenswrapper[4988]: I1008 19:34:11.784764 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="111b6cdd-1e77-4992-a331-c66bc9e95a90" containerName="rabbitmq" Oct 08 19:34:11 crc kubenswrapper[4988]: I1008 19:34:11.784955 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="111b6cdd-1e77-4992-a331-c66bc9e95a90" containerName="rabbitmq" Oct 08 19:34:11 crc kubenswrapper[4988]: I1008 19:34:11.786001 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Oct 08 19:34:11 crc kubenswrapper[4988]: I1008 19:34:11.787918 4988 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-plugins-conf" Oct 08 19:34:11 crc kubenswrapper[4988]: I1008 19:34:11.787936 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-default-user" Oct 08 19:34:11 crc kubenswrapper[4988]: I1008 19:34:11.788024 4988 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-config-data" Oct 08 19:34:11 crc kubenswrapper[4988]: I1008 19:34:11.788233 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-erlang-cookie" Oct 08 19:34:11 crc kubenswrapper[4988]: I1008 19:34:11.788368 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-cell1-svc" Oct 08 19:34:11 crc kubenswrapper[4988]: I1008 19:34:11.788429 4988 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-server-conf" Oct 08 19:34:11 crc kubenswrapper[4988]: I1008 19:34:11.788528 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-server-dockercfg-k7shm" Oct 08 19:34:11 crc kubenswrapper[4988]: I1008 19:34:11.808703 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 08 19:34:11 crc kubenswrapper[4988]: I1008 19:34:11.902713 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 08 19:34:11 crc kubenswrapper[4988]: W1008 19:34:11.908610 4988 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9d6674d9_eacf_4d44_91db_5ace88fa9a01.slice/crio-71faf5970b3021b09ffc78d6eb5009dfc9bb01c57fc5a052e3cfb70ffb2518a6 WatchSource:0}: Error finding container 71faf5970b3021b09ffc78d6eb5009dfc9bb01c57fc5a052e3cfb70ffb2518a6: Status 404 returned error can't find the container with id 71faf5970b3021b09ffc78d6eb5009dfc9bb01c57fc5a052e3cfb70ffb2518a6 Oct 08 19:34:11 crc kubenswrapper[4988]: I1008 19:34:11.962333 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/16348dd8-8383-45c0-910a-fe3cf4abcab0-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"16348dd8-8383-45c0-910a-fe3cf4abcab0\") " pod="openstack/rabbitmq-cell1-server-0" Oct 08 19:34:11 crc kubenswrapper[4988]: I1008 19:34:11.962379 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/16348dd8-8383-45c0-910a-fe3cf4abcab0-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"16348dd8-8383-45c0-910a-fe3cf4abcab0\") " pod="openstack/rabbitmq-cell1-server-0" Oct 08 19:34:11 crc kubenswrapper[4988]: I1008 19:34:11.962420 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/16348dd8-8383-45c0-910a-fe3cf4abcab0-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"16348dd8-8383-45c0-910a-fe3cf4abcab0\") " pod="openstack/rabbitmq-cell1-server-0" Oct 08 19:34:11 crc kubenswrapper[4988]: I1008 19:34:11.962438 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/16348dd8-8383-45c0-910a-fe3cf4abcab0-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"16348dd8-8383-45c0-910a-fe3cf4abcab0\") " pod="openstack/rabbitmq-cell1-server-0" Oct 08 19:34:11 crc kubenswrapper[4988]: I1008 19:34:11.962458 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/16348dd8-8383-45c0-910a-fe3cf4abcab0-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"16348dd8-8383-45c0-910a-fe3cf4abcab0\") " pod="openstack/rabbitmq-cell1-server-0" Oct 08 19:34:11 crc kubenswrapper[4988]: I1008 19:34:11.962478 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/16348dd8-8383-45c0-910a-fe3cf4abcab0-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"16348dd8-8383-45c0-910a-fe3cf4abcab0\") " pod="openstack/rabbitmq-cell1-server-0" Oct 08 19:34:11 crc kubenswrapper[4988]: I1008 19:34:11.962508 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-ee25b6c1-7b38-49fa-97b1-836588148f84\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-ee25b6c1-7b38-49fa-97b1-836588148f84\") pod \"rabbitmq-cell1-server-0\" (UID: \"16348dd8-8383-45c0-910a-fe3cf4abcab0\") " pod="openstack/rabbitmq-cell1-server-0" Oct 08 19:34:11 crc kubenswrapper[4988]: I1008 19:34:11.962537 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/16348dd8-8383-45c0-910a-fe3cf4abcab0-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"16348dd8-8383-45c0-910a-fe3cf4abcab0\") " pod="openstack/rabbitmq-cell1-server-0" Oct 08 19:34:11 crc kubenswrapper[4988]: I1008 19:34:11.962582 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/16348dd8-8383-45c0-910a-fe3cf4abcab0-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"16348dd8-8383-45c0-910a-fe3cf4abcab0\") " pod="openstack/rabbitmq-cell1-server-0" Oct 08 19:34:11 crc kubenswrapper[4988]: I1008 19:34:11.962598 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4xd76\" (UniqueName: \"kubernetes.io/projected/16348dd8-8383-45c0-910a-fe3cf4abcab0-kube-api-access-4xd76\") pod \"rabbitmq-cell1-server-0\" (UID: \"16348dd8-8383-45c0-910a-fe3cf4abcab0\") " pod="openstack/rabbitmq-cell1-server-0" Oct 08 19:34:11 crc kubenswrapper[4988]: I1008 19:34:11.962618 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/16348dd8-8383-45c0-910a-fe3cf4abcab0-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"16348dd8-8383-45c0-910a-fe3cf4abcab0\") " pod="openstack/rabbitmq-cell1-server-0" Oct 08 19:34:12 crc kubenswrapper[4988]: I1008 19:34:12.063717 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-ee25b6c1-7b38-49fa-97b1-836588148f84\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-ee25b6c1-7b38-49fa-97b1-836588148f84\") pod \"rabbitmq-cell1-server-0\" (UID: \"16348dd8-8383-45c0-910a-fe3cf4abcab0\") " pod="openstack/rabbitmq-cell1-server-0" Oct 08 19:34:12 crc kubenswrapper[4988]: I1008 19:34:12.064439 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/16348dd8-8383-45c0-910a-fe3cf4abcab0-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"16348dd8-8383-45c0-910a-fe3cf4abcab0\") " pod="openstack/rabbitmq-cell1-server-0" Oct 08 19:34:12 crc kubenswrapper[4988]: I1008 19:34:12.064588 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/16348dd8-8383-45c0-910a-fe3cf4abcab0-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"16348dd8-8383-45c0-910a-fe3cf4abcab0\") " pod="openstack/rabbitmq-cell1-server-0" Oct 08 19:34:12 crc kubenswrapper[4988]: I1008 19:34:12.064647 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4xd76\" (UniqueName: \"kubernetes.io/projected/16348dd8-8383-45c0-910a-fe3cf4abcab0-kube-api-access-4xd76\") pod \"rabbitmq-cell1-server-0\" (UID: \"16348dd8-8383-45c0-910a-fe3cf4abcab0\") " pod="openstack/rabbitmq-cell1-server-0" Oct 08 19:34:12 crc kubenswrapper[4988]: I1008 19:34:12.064717 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/16348dd8-8383-45c0-910a-fe3cf4abcab0-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"16348dd8-8383-45c0-910a-fe3cf4abcab0\") " pod="openstack/rabbitmq-cell1-server-0" Oct 08 19:34:12 crc kubenswrapper[4988]: I1008 19:34:12.064824 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/16348dd8-8383-45c0-910a-fe3cf4abcab0-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"16348dd8-8383-45c0-910a-fe3cf4abcab0\") " pod="openstack/rabbitmq-cell1-server-0" Oct 08 19:34:12 crc kubenswrapper[4988]: I1008 19:34:12.064881 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/16348dd8-8383-45c0-910a-fe3cf4abcab0-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"16348dd8-8383-45c0-910a-fe3cf4abcab0\") " pod="openstack/rabbitmq-cell1-server-0" Oct 08 19:34:12 crc kubenswrapper[4988]: I1008 19:34:12.064945 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/16348dd8-8383-45c0-910a-fe3cf4abcab0-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"16348dd8-8383-45c0-910a-fe3cf4abcab0\") " pod="openstack/rabbitmq-cell1-server-0" Oct 08 19:34:12 crc kubenswrapper[4988]: I1008 19:34:12.064998 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/16348dd8-8383-45c0-910a-fe3cf4abcab0-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"16348dd8-8383-45c0-910a-fe3cf4abcab0\") " pod="openstack/rabbitmq-cell1-server-0" Oct 08 19:34:12 crc kubenswrapper[4988]: I1008 19:34:12.065086 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/16348dd8-8383-45c0-910a-fe3cf4abcab0-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"16348dd8-8383-45c0-910a-fe3cf4abcab0\") " pod="openstack/rabbitmq-cell1-server-0" Oct 08 19:34:12 crc kubenswrapper[4988]: I1008 19:34:12.065136 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/16348dd8-8383-45c0-910a-fe3cf4abcab0-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"16348dd8-8383-45c0-910a-fe3cf4abcab0\") " pod="openstack/rabbitmq-cell1-server-0" Oct 08 19:34:12 crc kubenswrapper[4988]: I1008 19:34:12.066370 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/16348dd8-8383-45c0-910a-fe3cf4abcab0-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"16348dd8-8383-45c0-910a-fe3cf4abcab0\") " pod="openstack/rabbitmq-cell1-server-0" Oct 08 19:34:12 crc kubenswrapper[4988]: I1008 19:34:12.066487 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/16348dd8-8383-45c0-910a-fe3cf4abcab0-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"16348dd8-8383-45c0-910a-fe3cf4abcab0\") " pod="openstack/rabbitmq-cell1-server-0" Oct 08 19:34:12 crc kubenswrapper[4988]: I1008 19:34:12.066634 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/16348dd8-8383-45c0-910a-fe3cf4abcab0-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"16348dd8-8383-45c0-910a-fe3cf4abcab0\") " pod="openstack/rabbitmq-cell1-server-0" Oct 08 19:34:12 crc kubenswrapper[4988]: I1008 19:34:12.067022 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/16348dd8-8383-45c0-910a-fe3cf4abcab0-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"16348dd8-8383-45c0-910a-fe3cf4abcab0\") " pod="openstack/rabbitmq-cell1-server-0" Oct 08 19:34:12 crc kubenswrapper[4988]: I1008 19:34:12.067063 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/16348dd8-8383-45c0-910a-fe3cf4abcab0-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"16348dd8-8383-45c0-910a-fe3cf4abcab0\") " pod="openstack/rabbitmq-cell1-server-0" Oct 08 19:34:12 crc kubenswrapper[4988]: I1008 19:34:12.067904 4988 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Oct 08 19:34:12 crc kubenswrapper[4988]: I1008 19:34:12.068018 4988 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-ee25b6c1-7b38-49fa-97b1-836588148f84\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-ee25b6c1-7b38-49fa-97b1-836588148f84\") pod \"rabbitmq-cell1-server-0\" (UID: \"16348dd8-8383-45c0-910a-fe3cf4abcab0\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1c9155ea9cd9063941368b7285cdba7000bafae9c9b6d32fa454bc55cae46f02/globalmount\"" pod="openstack/rabbitmq-cell1-server-0" Oct 08 19:34:12 crc kubenswrapper[4988]: I1008 19:34:12.070224 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/16348dd8-8383-45c0-910a-fe3cf4abcab0-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"16348dd8-8383-45c0-910a-fe3cf4abcab0\") " pod="openstack/rabbitmq-cell1-server-0" Oct 08 19:34:12 crc kubenswrapper[4988]: I1008 19:34:12.070908 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/16348dd8-8383-45c0-910a-fe3cf4abcab0-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"16348dd8-8383-45c0-910a-fe3cf4abcab0\") " pod="openstack/rabbitmq-cell1-server-0" Oct 08 19:34:12 crc kubenswrapper[4988]: I1008 19:34:12.072886 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/16348dd8-8383-45c0-910a-fe3cf4abcab0-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"16348dd8-8383-45c0-910a-fe3cf4abcab0\") " pod="openstack/rabbitmq-cell1-server-0" Oct 08 19:34:12 crc kubenswrapper[4988]: I1008 19:34:12.074937 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/16348dd8-8383-45c0-910a-fe3cf4abcab0-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"16348dd8-8383-45c0-910a-fe3cf4abcab0\") " pod="openstack/rabbitmq-cell1-server-0" Oct 08 19:34:12 crc kubenswrapper[4988]: I1008 19:34:12.099024 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4xd76\" (UniqueName: \"kubernetes.io/projected/16348dd8-8383-45c0-910a-fe3cf4abcab0-kube-api-access-4xd76\") pod \"rabbitmq-cell1-server-0\" (UID: \"16348dd8-8383-45c0-910a-fe3cf4abcab0\") " pod="openstack/rabbitmq-cell1-server-0" Oct 08 19:34:12 crc kubenswrapper[4988]: I1008 19:34:12.124929 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-ee25b6c1-7b38-49fa-97b1-836588148f84\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-ee25b6c1-7b38-49fa-97b1-836588148f84\") pod \"rabbitmq-cell1-server-0\" (UID: \"16348dd8-8383-45c0-910a-fe3cf4abcab0\") " pod="openstack/rabbitmq-cell1-server-0" Oct 08 19:34:12 crc kubenswrapper[4988]: I1008 19:34:12.411195 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Oct 08 19:34:12 crc kubenswrapper[4988]: I1008 19:34:12.442655 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"9d6674d9-eacf-4d44-91db-5ace88fa9a01","Type":"ContainerStarted","Data":"71faf5970b3021b09ffc78d6eb5009dfc9bb01c57fc5a052e3cfb70ffb2518a6"} Oct 08 19:34:12 crc kubenswrapper[4988]: I1008 19:34:12.648990 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 08 19:34:12 crc kubenswrapper[4988]: W1008 19:34:12.659562 4988 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod16348dd8_8383_45c0_910a_fe3cf4abcab0.slice/crio-3941e009d90e276cf1e0994f391ae5dc4b7309a34b299df59dbd59dab1511da6 WatchSource:0}: Error finding container 3941e009d90e276cf1e0994f391ae5dc4b7309a34b299df59dbd59dab1511da6: Status 404 returned error can't find the container with id 3941e009d90e276cf1e0994f391ae5dc4b7309a34b299df59dbd59dab1511da6 Oct 08 19:34:13 crc kubenswrapper[4988]: I1008 19:34:13.255805 4988 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="111b6cdd-1e77-4992-a331-c66bc9e95a90" path="/var/lib/kubelet/pods/111b6cdd-1e77-4992-a331-c66bc9e95a90/volumes" Oct 08 19:34:13 crc kubenswrapper[4988]: I1008 19:34:13.451110 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"16348dd8-8383-45c0-910a-fe3cf4abcab0","Type":"ContainerStarted","Data":"3941e009d90e276cf1e0994f391ae5dc4b7309a34b299df59dbd59dab1511da6"} Oct 08 19:34:14 crc kubenswrapper[4988]: I1008 19:34:14.464377 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"9d6674d9-eacf-4d44-91db-5ace88fa9a01","Type":"ContainerStarted","Data":"cdfa85d4b9111580afb5eca2c6118b9ae4ef26865fe2989958803eae0db64e9d"} Oct 08 19:34:14 crc kubenswrapper[4988]: I1008 19:34:14.467234 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"16348dd8-8383-45c0-910a-fe3cf4abcab0","Type":"ContainerStarted","Data":"0c8194bb8fac4c2232f34d5c0bca79ced5949b0e7faf25bc8364c8e0baa97db0"} Oct 08 19:34:23 crc kubenswrapper[4988]: I1008 19:34:23.337769 4988 patch_prober.go:28] interesting pod/machine-config-daemon-nm54f container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 19:34:23 crc kubenswrapper[4988]: I1008 19:34:23.338415 4988 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 19:34:46 crc kubenswrapper[4988]: I1008 19:34:46.782658 4988 generic.go:334] "Generic (PLEG): container finished" podID="16348dd8-8383-45c0-910a-fe3cf4abcab0" containerID="0c8194bb8fac4c2232f34d5c0bca79ced5949b0e7faf25bc8364c8e0baa97db0" exitCode=0 Oct 08 19:34:46 crc kubenswrapper[4988]: I1008 19:34:46.782753 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"16348dd8-8383-45c0-910a-fe3cf4abcab0","Type":"ContainerDied","Data":"0c8194bb8fac4c2232f34d5c0bca79ced5949b0e7faf25bc8364c8e0baa97db0"} Oct 08 19:34:47 crc kubenswrapper[4988]: I1008 19:34:47.796581 4988 generic.go:334] "Generic (PLEG): container finished" podID="9d6674d9-eacf-4d44-91db-5ace88fa9a01" containerID="cdfa85d4b9111580afb5eca2c6118b9ae4ef26865fe2989958803eae0db64e9d" exitCode=0 Oct 08 19:34:47 crc kubenswrapper[4988]: I1008 19:34:47.796695 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"9d6674d9-eacf-4d44-91db-5ace88fa9a01","Type":"ContainerDied","Data":"cdfa85d4b9111580afb5eca2c6118b9ae4ef26865fe2989958803eae0db64e9d"} Oct 08 19:34:47 crc kubenswrapper[4988]: I1008 19:34:47.803073 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"16348dd8-8383-45c0-910a-fe3cf4abcab0","Type":"ContainerStarted","Data":"1f40e9fe2997c48676930c1e40a9b7bbba06ae2aaddfd591b99292ee1d60238b"} Oct 08 19:34:47 crc kubenswrapper[4988]: I1008 19:34:47.804113 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Oct 08 19:34:47 crc kubenswrapper[4988]: I1008 19:34:47.885346 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-cell1-server-0" podStartSLOduration=36.885330296 podStartE2EDuration="36.885330296s" podCreationTimestamp="2025-10-08 19:34:11 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 19:34:47.881904767 +0000 UTC m=+5033.331747537" watchObservedRunningTime="2025-10-08 19:34:47.885330296 +0000 UTC m=+5033.335173066" Oct 08 19:34:48 crc kubenswrapper[4988]: I1008 19:34:48.811289 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"9d6674d9-eacf-4d44-91db-5ace88fa9a01","Type":"ContainerStarted","Data":"d87df08c67601c994a987c94ff9c2d96e4e3eef0839a8c87ddcfbdb31519f8a0"} Oct 08 19:34:48 crc kubenswrapper[4988]: I1008 19:34:48.812069 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Oct 08 19:34:48 crc kubenswrapper[4988]: I1008 19:34:48.836889 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-0" podStartSLOduration=38.836873188 podStartE2EDuration="38.836873188s" podCreationTimestamp="2025-10-08 19:34:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 19:34:48.835951258 +0000 UTC m=+5034.285794048" watchObservedRunningTime="2025-10-08 19:34:48.836873188 +0000 UTC m=+5034.286715958" Oct 08 19:34:53 crc kubenswrapper[4988]: I1008 19:34:53.338530 4988 patch_prober.go:28] interesting pod/machine-config-daemon-nm54f container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 19:34:53 crc kubenswrapper[4988]: I1008 19:34:53.339068 4988 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 19:34:53 crc kubenswrapper[4988]: I1008 19:34:53.339135 4988 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" Oct 08 19:34:53 crc kubenswrapper[4988]: I1008 19:34:53.340010 4988 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"3cd52d2ff674f037f4b111ff08caa426d4d7977052a05e385f05362baabb8da9"} pod="openshift-machine-config-operator/machine-config-daemon-nm54f" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 08 19:34:53 crc kubenswrapper[4988]: I1008 19:34:53.340112 4988 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" containerName="machine-config-daemon" containerID="cri-o://3cd52d2ff674f037f4b111ff08caa426d4d7977052a05e385f05362baabb8da9" gracePeriod=600 Oct 08 19:34:53 crc kubenswrapper[4988]: E1008 19:34:53.485840 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nm54f_openshift-machine-config-operator(f328a20b-5283-407e-8a93-2ab2da599f18)\"" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" Oct 08 19:34:53 crc kubenswrapper[4988]: I1008 19:34:53.856698 4988 generic.go:334] "Generic (PLEG): container finished" podID="f328a20b-5283-407e-8a93-2ab2da599f18" containerID="3cd52d2ff674f037f4b111ff08caa426d4d7977052a05e385f05362baabb8da9" exitCode=0 Oct 08 19:34:53 crc kubenswrapper[4988]: I1008 19:34:53.856744 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" event={"ID":"f328a20b-5283-407e-8a93-2ab2da599f18","Type":"ContainerDied","Data":"3cd52d2ff674f037f4b111ff08caa426d4d7977052a05e385f05362baabb8da9"} Oct 08 19:34:53 crc kubenswrapper[4988]: I1008 19:34:53.857040 4988 scope.go:117] "RemoveContainer" containerID="6615b98ff2af8d9695912c635a4216452e43b30637b6fb36a988358158ee4055" Oct 08 19:34:53 crc kubenswrapper[4988]: I1008 19:34:53.857884 4988 scope.go:117] "RemoveContainer" containerID="3cd52d2ff674f037f4b111ff08caa426d4d7977052a05e385f05362baabb8da9" Oct 08 19:34:53 crc kubenswrapper[4988]: E1008 19:34:53.858378 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nm54f_openshift-machine-config-operator(f328a20b-5283-407e-8a93-2ab2da599f18)\"" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" Oct 08 19:35:01 crc kubenswrapper[4988]: I1008 19:35:01.394756 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-0" Oct 08 19:35:02 crc kubenswrapper[4988]: I1008 19:35:02.414584 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-cell1-server-0" Oct 08 19:35:05 crc kubenswrapper[4988]: I1008 19:35:05.247242 4988 scope.go:117] "RemoveContainer" containerID="3cd52d2ff674f037f4b111ff08caa426d4d7977052a05e385f05362baabb8da9" Oct 08 19:35:05 crc kubenswrapper[4988]: E1008 19:35:05.248549 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nm54f_openshift-machine-config-operator(f328a20b-5283-407e-8a93-2ab2da599f18)\"" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" Oct 08 19:35:06 crc kubenswrapper[4988]: I1008 19:35:06.333436 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/mariadb-client-1-default"] Oct 08 19:35:06 crc kubenswrapper[4988]: I1008 19:35:06.335495 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-1-default" Oct 08 19:35:06 crc kubenswrapper[4988]: I1008 19:35:06.338852 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"default-dockercfg-cmd56" Oct 08 19:35:06 crc kubenswrapper[4988]: I1008 19:35:06.340781 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client-1-default"] Oct 08 19:35:06 crc kubenswrapper[4988]: I1008 19:35:06.440126 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qbqpq\" (UniqueName: \"kubernetes.io/projected/98f94d54-0c36-4247-b896-d7cb16de7195-kube-api-access-qbqpq\") pod \"mariadb-client-1-default\" (UID: \"98f94d54-0c36-4247-b896-d7cb16de7195\") " pod="openstack/mariadb-client-1-default" Oct 08 19:35:06 crc kubenswrapper[4988]: I1008 19:35:06.541803 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qbqpq\" (UniqueName: \"kubernetes.io/projected/98f94d54-0c36-4247-b896-d7cb16de7195-kube-api-access-qbqpq\") pod \"mariadb-client-1-default\" (UID: \"98f94d54-0c36-4247-b896-d7cb16de7195\") " pod="openstack/mariadb-client-1-default" Oct 08 19:35:06 crc kubenswrapper[4988]: I1008 19:35:06.572842 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qbqpq\" (UniqueName: \"kubernetes.io/projected/98f94d54-0c36-4247-b896-d7cb16de7195-kube-api-access-qbqpq\") pod \"mariadb-client-1-default\" (UID: \"98f94d54-0c36-4247-b896-d7cb16de7195\") " pod="openstack/mariadb-client-1-default" Oct 08 19:35:06 crc kubenswrapper[4988]: I1008 19:35:06.670728 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-1-default" Oct 08 19:35:07 crc kubenswrapper[4988]: I1008 19:35:07.268201 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client-1-default"] Oct 08 19:35:07 crc kubenswrapper[4988]: I1008 19:35:07.271583 4988 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 08 19:35:07 crc kubenswrapper[4988]: I1008 19:35:07.987828 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-1-default" event={"ID":"98f94d54-0c36-4247-b896-d7cb16de7195","Type":"ContainerStarted","Data":"c41a47b67dda9707005d6116e8388c5500ddbea3516a8c1057381c54c0d811f1"} Oct 08 19:35:08 crc kubenswrapper[4988]: I1008 19:35:08.999199 4988 generic.go:334] "Generic (PLEG): container finished" podID="98f94d54-0c36-4247-b896-d7cb16de7195" containerID="9214b02e8010153a4c65cd88a96915ebf2fe9de3d5325d6c59491c347b8fbfb4" exitCode=0 Oct 08 19:35:08 crc kubenswrapper[4988]: I1008 19:35:08.999261 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-1-default" event={"ID":"98f94d54-0c36-4247-b896-d7cb16de7195","Type":"ContainerDied","Data":"9214b02e8010153a4c65cd88a96915ebf2fe9de3d5325d6c59491c347b8fbfb4"} Oct 08 19:35:10 crc kubenswrapper[4988]: I1008 19:35:10.428697 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-1-default" Oct 08 19:35:10 crc kubenswrapper[4988]: I1008 19:35:10.463806 4988 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_mariadb-client-1-default_98f94d54-0c36-4247-b896-d7cb16de7195/mariadb-client-1-default/0.log" Oct 08 19:35:10 crc kubenswrapper[4988]: I1008 19:35:10.491940 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mariadb-client-1-default"] Oct 08 19:35:10 crc kubenswrapper[4988]: I1008 19:35:10.496171 4988 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/mariadb-client-1-default"] Oct 08 19:35:10 crc kubenswrapper[4988]: I1008 19:35:10.617779 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qbqpq\" (UniqueName: \"kubernetes.io/projected/98f94d54-0c36-4247-b896-d7cb16de7195-kube-api-access-qbqpq\") pod \"98f94d54-0c36-4247-b896-d7cb16de7195\" (UID: \"98f94d54-0c36-4247-b896-d7cb16de7195\") " Oct 08 19:35:10 crc kubenswrapper[4988]: I1008 19:35:10.628912 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/98f94d54-0c36-4247-b896-d7cb16de7195-kube-api-access-qbqpq" (OuterVolumeSpecName: "kube-api-access-qbqpq") pod "98f94d54-0c36-4247-b896-d7cb16de7195" (UID: "98f94d54-0c36-4247-b896-d7cb16de7195"). InnerVolumeSpecName "kube-api-access-qbqpq". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 19:35:10 crc kubenswrapper[4988]: I1008 19:35:10.720155 4988 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qbqpq\" (UniqueName: \"kubernetes.io/projected/98f94d54-0c36-4247-b896-d7cb16de7195-kube-api-access-qbqpq\") on node \"crc\" DevicePath \"\"" Oct 08 19:35:10 crc kubenswrapper[4988]: I1008 19:35:10.981188 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/mariadb-client-2-default"] Oct 08 19:35:10 crc kubenswrapper[4988]: E1008 19:35:10.981578 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="98f94d54-0c36-4247-b896-d7cb16de7195" containerName="mariadb-client-1-default" Oct 08 19:35:10 crc kubenswrapper[4988]: I1008 19:35:10.981600 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="98f94d54-0c36-4247-b896-d7cb16de7195" containerName="mariadb-client-1-default" Oct 08 19:35:10 crc kubenswrapper[4988]: I1008 19:35:10.981791 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="98f94d54-0c36-4247-b896-d7cb16de7195" containerName="mariadb-client-1-default" Oct 08 19:35:10 crc kubenswrapper[4988]: I1008 19:35:10.982520 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-2-default" Oct 08 19:35:10 crc kubenswrapper[4988]: I1008 19:35:10.994990 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client-2-default"] Oct 08 19:35:11 crc kubenswrapper[4988]: I1008 19:35:11.029891 4988 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c41a47b67dda9707005d6116e8388c5500ddbea3516a8c1057381c54c0d811f1" Oct 08 19:35:11 crc kubenswrapper[4988]: I1008 19:35:11.029994 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-1-default" Oct 08 19:35:11 crc kubenswrapper[4988]: I1008 19:35:11.127199 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d22b9\" (UniqueName: \"kubernetes.io/projected/fc4ef802-5de7-4734-a546-59a17d9705b9-kube-api-access-d22b9\") pod \"mariadb-client-2-default\" (UID: \"fc4ef802-5de7-4734-a546-59a17d9705b9\") " pod="openstack/mariadb-client-2-default" Oct 08 19:35:11 crc kubenswrapper[4988]: I1008 19:35:11.229307 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d22b9\" (UniqueName: \"kubernetes.io/projected/fc4ef802-5de7-4734-a546-59a17d9705b9-kube-api-access-d22b9\") pod \"mariadb-client-2-default\" (UID: \"fc4ef802-5de7-4734-a546-59a17d9705b9\") " pod="openstack/mariadb-client-2-default" Oct 08 19:35:11 crc kubenswrapper[4988]: I1008 19:35:11.251993 4988 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="98f94d54-0c36-4247-b896-d7cb16de7195" path="/var/lib/kubelet/pods/98f94d54-0c36-4247-b896-d7cb16de7195/volumes" Oct 08 19:35:11 crc kubenswrapper[4988]: I1008 19:35:11.256741 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d22b9\" (UniqueName: \"kubernetes.io/projected/fc4ef802-5de7-4734-a546-59a17d9705b9-kube-api-access-d22b9\") pod \"mariadb-client-2-default\" (UID: \"fc4ef802-5de7-4734-a546-59a17d9705b9\") " pod="openstack/mariadb-client-2-default" Oct 08 19:35:11 crc kubenswrapper[4988]: I1008 19:35:11.316690 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-2-default" Oct 08 19:35:11 crc kubenswrapper[4988]: I1008 19:35:11.716055 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client-2-default"] Oct 08 19:35:12 crc kubenswrapper[4988]: I1008 19:35:12.039427 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-2-default" event={"ID":"fc4ef802-5de7-4734-a546-59a17d9705b9","Type":"ContainerStarted","Data":"a85b0052d2b128dbf00387b13b812baf59c4018713676419ff62ff83fef60d3e"} Oct 08 19:35:13 crc kubenswrapper[4988]: I1008 19:35:13.050324 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-2-default" event={"ID":"fc4ef802-5de7-4734-a546-59a17d9705b9","Type":"ContainerStarted","Data":"a9930599f243f306a3e5c3973da6b1f9e04bcffe18286169854acaa0a07ef6f1"} Oct 08 19:35:13 crc kubenswrapper[4988]: I1008 19:35:13.072186 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/mariadb-client-2-default" podStartSLOduration=3.072153621 podStartE2EDuration="3.072153621s" podCreationTimestamp="2025-10-08 19:35:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 19:35:13.068116263 +0000 UTC m=+5058.517959043" watchObservedRunningTime="2025-10-08 19:35:13.072153621 +0000 UTC m=+5058.521996421" Oct 08 19:35:13 crc kubenswrapper[4988]: I1008 19:35:13.188672 4988 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_mariadb-client-2-default_fc4ef802-5de7-4734-a546-59a17d9705b9/mariadb-client-2-default/0.log" Oct 08 19:35:14 crc kubenswrapper[4988]: I1008 19:35:14.061112 4988 generic.go:334] "Generic (PLEG): container finished" podID="fc4ef802-5de7-4734-a546-59a17d9705b9" containerID="a9930599f243f306a3e5c3973da6b1f9e04bcffe18286169854acaa0a07ef6f1" exitCode=0 Oct 08 19:35:14 crc kubenswrapper[4988]: I1008 19:35:14.061163 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-2-default" event={"ID":"fc4ef802-5de7-4734-a546-59a17d9705b9","Type":"ContainerDied","Data":"a9930599f243f306a3e5c3973da6b1f9e04bcffe18286169854acaa0a07ef6f1"} Oct 08 19:35:15 crc kubenswrapper[4988]: I1008 19:35:15.561856 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-2-default" Oct 08 19:35:15 crc kubenswrapper[4988]: I1008 19:35:15.605434 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mariadb-client-2-default"] Oct 08 19:35:15 crc kubenswrapper[4988]: I1008 19:35:15.611504 4988 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/mariadb-client-2-default"] Oct 08 19:35:15 crc kubenswrapper[4988]: I1008 19:35:15.707528 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d22b9\" (UniqueName: \"kubernetes.io/projected/fc4ef802-5de7-4734-a546-59a17d9705b9-kube-api-access-d22b9\") pod \"fc4ef802-5de7-4734-a546-59a17d9705b9\" (UID: \"fc4ef802-5de7-4734-a546-59a17d9705b9\") " Oct 08 19:35:15 crc kubenswrapper[4988]: I1008 19:35:15.716493 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fc4ef802-5de7-4734-a546-59a17d9705b9-kube-api-access-d22b9" (OuterVolumeSpecName: "kube-api-access-d22b9") pod "fc4ef802-5de7-4734-a546-59a17d9705b9" (UID: "fc4ef802-5de7-4734-a546-59a17d9705b9"). InnerVolumeSpecName "kube-api-access-d22b9". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 19:35:15 crc kubenswrapper[4988]: I1008 19:35:15.809927 4988 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d22b9\" (UniqueName: \"kubernetes.io/projected/fc4ef802-5de7-4734-a546-59a17d9705b9-kube-api-access-d22b9\") on node \"crc\" DevicePath \"\"" Oct 08 19:35:16 crc kubenswrapper[4988]: I1008 19:35:16.084178 4988 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a85b0052d2b128dbf00387b13b812baf59c4018713676419ff62ff83fef60d3e" Oct 08 19:35:16 crc kubenswrapper[4988]: I1008 19:35:16.084271 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-2-default" Oct 08 19:35:16 crc kubenswrapper[4988]: I1008 19:35:16.112544 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/mariadb-client-1"] Oct 08 19:35:16 crc kubenswrapper[4988]: E1008 19:35:16.113278 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fc4ef802-5de7-4734-a546-59a17d9705b9" containerName="mariadb-client-2-default" Oct 08 19:35:16 crc kubenswrapper[4988]: I1008 19:35:16.113299 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="fc4ef802-5de7-4734-a546-59a17d9705b9" containerName="mariadb-client-2-default" Oct 08 19:35:16 crc kubenswrapper[4988]: I1008 19:35:16.113559 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="fc4ef802-5de7-4734-a546-59a17d9705b9" containerName="mariadb-client-2-default" Oct 08 19:35:16 crc kubenswrapper[4988]: I1008 19:35:16.114207 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-1" Oct 08 19:35:16 crc kubenswrapper[4988]: I1008 19:35:16.119543 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client-1"] Oct 08 19:35:16 crc kubenswrapper[4988]: I1008 19:35:16.121810 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"default-dockercfg-cmd56" Oct 08 19:35:16 crc kubenswrapper[4988]: I1008 19:35:16.217571 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vcsvz\" (UniqueName: \"kubernetes.io/projected/8831cd59-aabc-4ad3-9825-8abcca9a07d3-kube-api-access-vcsvz\") pod \"mariadb-client-1\" (UID: \"8831cd59-aabc-4ad3-9825-8abcca9a07d3\") " pod="openstack/mariadb-client-1" Oct 08 19:35:16 crc kubenswrapper[4988]: I1008 19:35:16.237606 4988 scope.go:117] "RemoveContainer" containerID="3cd52d2ff674f037f4b111ff08caa426d4d7977052a05e385f05362baabb8da9" Oct 08 19:35:16 crc kubenswrapper[4988]: E1008 19:35:16.237951 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nm54f_openshift-machine-config-operator(f328a20b-5283-407e-8a93-2ab2da599f18)\"" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" Oct 08 19:35:16 crc kubenswrapper[4988]: I1008 19:35:16.318675 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vcsvz\" (UniqueName: \"kubernetes.io/projected/8831cd59-aabc-4ad3-9825-8abcca9a07d3-kube-api-access-vcsvz\") pod \"mariadb-client-1\" (UID: \"8831cd59-aabc-4ad3-9825-8abcca9a07d3\") " pod="openstack/mariadb-client-1" Oct 08 19:35:16 crc kubenswrapper[4988]: I1008 19:35:16.342084 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vcsvz\" (UniqueName: \"kubernetes.io/projected/8831cd59-aabc-4ad3-9825-8abcca9a07d3-kube-api-access-vcsvz\") pod \"mariadb-client-1\" (UID: \"8831cd59-aabc-4ad3-9825-8abcca9a07d3\") " pod="openstack/mariadb-client-1" Oct 08 19:35:16 crc kubenswrapper[4988]: I1008 19:35:16.449810 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-1" Oct 08 19:35:16 crc kubenswrapper[4988]: I1008 19:35:16.769107 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client-1"] Oct 08 19:35:16 crc kubenswrapper[4988]: W1008 19:35:16.774706 4988 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8831cd59_aabc_4ad3_9825_8abcca9a07d3.slice/crio-de19b5f412f4d6ebc575923c9ec7d7f22782ed29619584c228a0b63f16997748 WatchSource:0}: Error finding container de19b5f412f4d6ebc575923c9ec7d7f22782ed29619584c228a0b63f16997748: Status 404 returned error can't find the container with id de19b5f412f4d6ebc575923c9ec7d7f22782ed29619584c228a0b63f16997748 Oct 08 19:35:17 crc kubenswrapper[4988]: I1008 19:35:17.106530 4988 generic.go:334] "Generic (PLEG): container finished" podID="8831cd59-aabc-4ad3-9825-8abcca9a07d3" containerID="147b1b278899f1d5b3702571f4c2b413ed486c9aa9a015474beba8f457183f9a" exitCode=0 Oct 08 19:35:17 crc kubenswrapper[4988]: I1008 19:35:17.106596 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-1" event={"ID":"8831cd59-aabc-4ad3-9825-8abcca9a07d3","Type":"ContainerDied","Data":"147b1b278899f1d5b3702571f4c2b413ed486c9aa9a015474beba8f457183f9a"} Oct 08 19:35:17 crc kubenswrapper[4988]: I1008 19:35:17.106634 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-1" event={"ID":"8831cd59-aabc-4ad3-9825-8abcca9a07d3","Type":"ContainerStarted","Data":"de19b5f412f4d6ebc575923c9ec7d7f22782ed29619584c228a0b63f16997748"} Oct 08 19:35:17 crc kubenswrapper[4988]: I1008 19:35:17.254912 4988 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fc4ef802-5de7-4734-a546-59a17d9705b9" path="/var/lib/kubelet/pods/fc4ef802-5de7-4734-a546-59a17d9705b9/volumes" Oct 08 19:35:18 crc kubenswrapper[4988]: I1008 19:35:18.563616 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-1" Oct 08 19:35:18 crc kubenswrapper[4988]: I1008 19:35:18.583311 4988 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_mariadb-client-1_8831cd59-aabc-4ad3-9825-8abcca9a07d3/mariadb-client-1/0.log" Oct 08 19:35:18 crc kubenswrapper[4988]: I1008 19:35:18.622454 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mariadb-client-1"] Oct 08 19:35:18 crc kubenswrapper[4988]: I1008 19:35:18.630084 4988 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/mariadb-client-1"] Oct 08 19:35:18 crc kubenswrapper[4988]: I1008 19:35:18.661432 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vcsvz\" (UniqueName: \"kubernetes.io/projected/8831cd59-aabc-4ad3-9825-8abcca9a07d3-kube-api-access-vcsvz\") pod \"8831cd59-aabc-4ad3-9825-8abcca9a07d3\" (UID: \"8831cd59-aabc-4ad3-9825-8abcca9a07d3\") " Oct 08 19:35:18 crc kubenswrapper[4988]: I1008 19:35:18.668579 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8831cd59-aabc-4ad3-9825-8abcca9a07d3-kube-api-access-vcsvz" (OuterVolumeSpecName: "kube-api-access-vcsvz") pod "8831cd59-aabc-4ad3-9825-8abcca9a07d3" (UID: "8831cd59-aabc-4ad3-9825-8abcca9a07d3"). InnerVolumeSpecName "kube-api-access-vcsvz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 19:35:18 crc kubenswrapper[4988]: I1008 19:35:18.838030 4988 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vcsvz\" (UniqueName: \"kubernetes.io/projected/8831cd59-aabc-4ad3-9825-8abcca9a07d3-kube-api-access-vcsvz\") on node \"crc\" DevicePath \"\"" Oct 08 19:35:19 crc kubenswrapper[4988]: I1008 19:35:19.052977 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/mariadb-client-4-default"] Oct 08 19:35:19 crc kubenswrapper[4988]: E1008 19:35:19.053582 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8831cd59-aabc-4ad3-9825-8abcca9a07d3" containerName="mariadb-client-1" Oct 08 19:35:19 crc kubenswrapper[4988]: I1008 19:35:19.053614 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="8831cd59-aabc-4ad3-9825-8abcca9a07d3" containerName="mariadb-client-1" Oct 08 19:35:19 crc kubenswrapper[4988]: I1008 19:35:19.053908 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="8831cd59-aabc-4ad3-9825-8abcca9a07d3" containerName="mariadb-client-1" Oct 08 19:35:19 crc kubenswrapper[4988]: I1008 19:35:19.054835 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-4-default" Oct 08 19:35:19 crc kubenswrapper[4988]: I1008 19:35:19.063879 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client-4-default"] Oct 08 19:35:19 crc kubenswrapper[4988]: I1008 19:35:19.128026 4988 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="de19b5f412f4d6ebc575923c9ec7d7f22782ed29619584c228a0b63f16997748" Oct 08 19:35:19 crc kubenswrapper[4988]: I1008 19:35:19.128062 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-1" Oct 08 19:35:19 crc kubenswrapper[4988]: I1008 19:35:19.141885 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-69sjk\" (UniqueName: \"kubernetes.io/projected/32324671-0560-4d85-a511-484a448ae294-kube-api-access-69sjk\") pod \"mariadb-client-4-default\" (UID: \"32324671-0560-4d85-a511-484a448ae294\") " pod="openstack/mariadb-client-4-default" Oct 08 19:35:19 crc kubenswrapper[4988]: I1008 19:35:19.243634 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-69sjk\" (UniqueName: \"kubernetes.io/projected/32324671-0560-4d85-a511-484a448ae294-kube-api-access-69sjk\") pod \"mariadb-client-4-default\" (UID: \"32324671-0560-4d85-a511-484a448ae294\") " pod="openstack/mariadb-client-4-default" Oct 08 19:35:19 crc kubenswrapper[4988]: I1008 19:35:19.260089 4988 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8831cd59-aabc-4ad3-9825-8abcca9a07d3" path="/var/lib/kubelet/pods/8831cd59-aabc-4ad3-9825-8abcca9a07d3/volumes" Oct 08 19:35:19 crc kubenswrapper[4988]: I1008 19:35:19.282283 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-69sjk\" (UniqueName: \"kubernetes.io/projected/32324671-0560-4d85-a511-484a448ae294-kube-api-access-69sjk\") pod \"mariadb-client-4-default\" (UID: \"32324671-0560-4d85-a511-484a448ae294\") " pod="openstack/mariadb-client-4-default" Oct 08 19:35:19 crc kubenswrapper[4988]: I1008 19:35:19.381485 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-4-default" Oct 08 19:35:19 crc kubenswrapper[4988]: I1008 19:35:19.895655 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client-4-default"] Oct 08 19:35:20 crc kubenswrapper[4988]: I1008 19:35:20.138228 4988 generic.go:334] "Generic (PLEG): container finished" podID="32324671-0560-4d85-a511-484a448ae294" containerID="509d7603080340ee6579c4843868b1b26c3851e047d18904944d7e4c09fd834d" exitCode=0 Oct 08 19:35:20 crc kubenswrapper[4988]: I1008 19:35:20.138280 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-4-default" event={"ID":"32324671-0560-4d85-a511-484a448ae294","Type":"ContainerDied","Data":"509d7603080340ee6579c4843868b1b26c3851e047d18904944d7e4c09fd834d"} Oct 08 19:35:20 crc kubenswrapper[4988]: I1008 19:35:20.138311 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-4-default" event={"ID":"32324671-0560-4d85-a511-484a448ae294","Type":"ContainerStarted","Data":"84af949f3ec7cbb2911b9b7c9e92fa65cd70dea7abbd911f2ac8d2d70c511b47"} Oct 08 19:35:21 crc kubenswrapper[4988]: I1008 19:35:21.611338 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-4-default" Oct 08 19:35:21 crc kubenswrapper[4988]: I1008 19:35:21.633363 4988 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_mariadb-client-4-default_32324671-0560-4d85-a511-484a448ae294/mariadb-client-4-default/0.log" Oct 08 19:35:21 crc kubenswrapper[4988]: I1008 19:35:21.669811 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mariadb-client-4-default"] Oct 08 19:35:21 crc kubenswrapper[4988]: I1008 19:35:21.679572 4988 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/mariadb-client-4-default"] Oct 08 19:35:21 crc kubenswrapper[4988]: I1008 19:35:21.789169 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-69sjk\" (UniqueName: \"kubernetes.io/projected/32324671-0560-4d85-a511-484a448ae294-kube-api-access-69sjk\") pod \"32324671-0560-4d85-a511-484a448ae294\" (UID: \"32324671-0560-4d85-a511-484a448ae294\") " Oct 08 19:35:21 crc kubenswrapper[4988]: I1008 19:35:21.796711 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/32324671-0560-4d85-a511-484a448ae294-kube-api-access-69sjk" (OuterVolumeSpecName: "kube-api-access-69sjk") pod "32324671-0560-4d85-a511-484a448ae294" (UID: "32324671-0560-4d85-a511-484a448ae294"). InnerVolumeSpecName "kube-api-access-69sjk". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 19:35:21 crc kubenswrapper[4988]: I1008 19:35:21.891553 4988 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-69sjk\" (UniqueName: \"kubernetes.io/projected/32324671-0560-4d85-a511-484a448ae294-kube-api-access-69sjk\") on node \"crc\" DevicePath \"\"" Oct 08 19:35:22 crc kubenswrapper[4988]: I1008 19:35:22.194143 4988 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="84af949f3ec7cbb2911b9b7c9e92fa65cd70dea7abbd911f2ac8d2d70c511b47" Oct 08 19:35:22 crc kubenswrapper[4988]: I1008 19:35:22.194246 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-4-default" Oct 08 19:35:23 crc kubenswrapper[4988]: I1008 19:35:23.252419 4988 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="32324671-0560-4d85-a511-484a448ae294" path="/var/lib/kubelet/pods/32324671-0560-4d85-a511-484a448ae294/volumes" Oct 08 19:35:26 crc kubenswrapper[4988]: I1008 19:35:26.628717 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/mariadb-client-5-default"] Oct 08 19:35:26 crc kubenswrapper[4988]: E1008 19:35:26.629368 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="32324671-0560-4d85-a511-484a448ae294" containerName="mariadb-client-4-default" Oct 08 19:35:26 crc kubenswrapper[4988]: I1008 19:35:26.629407 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="32324671-0560-4d85-a511-484a448ae294" containerName="mariadb-client-4-default" Oct 08 19:35:26 crc kubenswrapper[4988]: I1008 19:35:26.629617 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="32324671-0560-4d85-a511-484a448ae294" containerName="mariadb-client-4-default" Oct 08 19:35:26 crc kubenswrapper[4988]: I1008 19:35:26.630198 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-5-default" Oct 08 19:35:26 crc kubenswrapper[4988]: I1008 19:35:26.633879 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"default-dockercfg-cmd56" Oct 08 19:35:26 crc kubenswrapper[4988]: I1008 19:35:26.673149 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client-5-default"] Oct 08 19:35:26 crc kubenswrapper[4988]: I1008 19:35:26.769153 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fgc92\" (UniqueName: \"kubernetes.io/projected/56fad65a-577b-4227-a3fa-fe286ef09a16-kube-api-access-fgc92\") pod \"mariadb-client-5-default\" (UID: \"56fad65a-577b-4227-a3fa-fe286ef09a16\") " pod="openstack/mariadb-client-5-default" Oct 08 19:35:26 crc kubenswrapper[4988]: I1008 19:35:26.871309 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fgc92\" (UniqueName: \"kubernetes.io/projected/56fad65a-577b-4227-a3fa-fe286ef09a16-kube-api-access-fgc92\") pod \"mariadb-client-5-default\" (UID: \"56fad65a-577b-4227-a3fa-fe286ef09a16\") " pod="openstack/mariadb-client-5-default" Oct 08 19:35:26 crc kubenswrapper[4988]: I1008 19:35:26.900005 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fgc92\" (UniqueName: \"kubernetes.io/projected/56fad65a-577b-4227-a3fa-fe286ef09a16-kube-api-access-fgc92\") pod \"mariadb-client-5-default\" (UID: \"56fad65a-577b-4227-a3fa-fe286ef09a16\") " pod="openstack/mariadb-client-5-default" Oct 08 19:35:26 crc kubenswrapper[4988]: I1008 19:35:26.973694 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-5-default" Oct 08 19:35:27 crc kubenswrapper[4988]: I1008 19:35:27.545429 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client-5-default"] Oct 08 19:35:27 crc kubenswrapper[4988]: W1008 19:35:27.558631 4988 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod56fad65a_577b_4227_a3fa_fe286ef09a16.slice/crio-d133be4f76344b50471853406cf2ce27c3a233e8926a2d19f70a721984ecefee WatchSource:0}: Error finding container d133be4f76344b50471853406cf2ce27c3a233e8926a2d19f70a721984ecefee: Status 404 returned error can't find the container with id d133be4f76344b50471853406cf2ce27c3a233e8926a2d19f70a721984ecefee Oct 08 19:35:28 crc kubenswrapper[4988]: I1008 19:35:28.258728 4988 generic.go:334] "Generic (PLEG): container finished" podID="56fad65a-577b-4227-a3fa-fe286ef09a16" containerID="dcf87a67647e61ae96ca317046aa60dbfe871da46ec2397732c9a83ed077e54c" exitCode=0 Oct 08 19:35:28 crc kubenswrapper[4988]: I1008 19:35:28.258829 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-5-default" event={"ID":"56fad65a-577b-4227-a3fa-fe286ef09a16","Type":"ContainerDied","Data":"dcf87a67647e61ae96ca317046aa60dbfe871da46ec2397732c9a83ed077e54c"} Oct 08 19:35:28 crc kubenswrapper[4988]: I1008 19:35:28.259059 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-5-default" event={"ID":"56fad65a-577b-4227-a3fa-fe286ef09a16","Type":"ContainerStarted","Data":"d133be4f76344b50471853406cf2ce27c3a233e8926a2d19f70a721984ecefee"} Oct 08 19:35:29 crc kubenswrapper[4988]: I1008 19:35:29.238245 4988 scope.go:117] "RemoveContainer" containerID="3cd52d2ff674f037f4b111ff08caa426d4d7977052a05e385f05362baabb8da9" Oct 08 19:35:29 crc kubenswrapper[4988]: E1008 19:35:29.238595 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nm54f_openshift-machine-config-operator(f328a20b-5283-407e-8a93-2ab2da599f18)\"" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" Oct 08 19:35:29 crc kubenswrapper[4988]: I1008 19:35:29.696780 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-5-default" Oct 08 19:35:29 crc kubenswrapper[4988]: I1008 19:35:29.737188 4988 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_mariadb-client-5-default_56fad65a-577b-4227-a3fa-fe286ef09a16/mariadb-client-5-default/0.log" Oct 08 19:35:29 crc kubenswrapper[4988]: I1008 19:35:29.770335 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mariadb-client-5-default"] Oct 08 19:35:29 crc kubenswrapper[4988]: I1008 19:35:29.780741 4988 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/mariadb-client-5-default"] Oct 08 19:35:29 crc kubenswrapper[4988]: I1008 19:35:29.831811 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fgc92\" (UniqueName: \"kubernetes.io/projected/56fad65a-577b-4227-a3fa-fe286ef09a16-kube-api-access-fgc92\") pod \"56fad65a-577b-4227-a3fa-fe286ef09a16\" (UID: \"56fad65a-577b-4227-a3fa-fe286ef09a16\") " Oct 08 19:35:29 crc kubenswrapper[4988]: I1008 19:35:29.840891 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/56fad65a-577b-4227-a3fa-fe286ef09a16-kube-api-access-fgc92" (OuterVolumeSpecName: "kube-api-access-fgc92") pod "56fad65a-577b-4227-a3fa-fe286ef09a16" (UID: "56fad65a-577b-4227-a3fa-fe286ef09a16"). InnerVolumeSpecName "kube-api-access-fgc92". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 19:35:29 crc kubenswrapper[4988]: I1008 19:35:29.900411 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/mariadb-client-6-default"] Oct 08 19:35:29 crc kubenswrapper[4988]: E1008 19:35:29.900989 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="56fad65a-577b-4227-a3fa-fe286ef09a16" containerName="mariadb-client-5-default" Oct 08 19:35:29 crc kubenswrapper[4988]: I1008 19:35:29.901018 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="56fad65a-577b-4227-a3fa-fe286ef09a16" containerName="mariadb-client-5-default" Oct 08 19:35:29 crc kubenswrapper[4988]: I1008 19:35:29.901298 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="56fad65a-577b-4227-a3fa-fe286ef09a16" containerName="mariadb-client-5-default" Oct 08 19:35:29 crc kubenswrapper[4988]: I1008 19:35:29.902116 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-6-default" Oct 08 19:35:29 crc kubenswrapper[4988]: I1008 19:35:29.909607 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client-6-default"] Oct 08 19:35:29 crc kubenswrapper[4988]: I1008 19:35:29.935129 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ff74m\" (UniqueName: \"kubernetes.io/projected/30972cbc-cdb8-4379-a21d-4d727074ab7c-kube-api-access-ff74m\") pod \"mariadb-client-6-default\" (UID: \"30972cbc-cdb8-4379-a21d-4d727074ab7c\") " pod="openstack/mariadb-client-6-default" Oct 08 19:35:29 crc kubenswrapper[4988]: I1008 19:35:29.935556 4988 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fgc92\" (UniqueName: \"kubernetes.io/projected/56fad65a-577b-4227-a3fa-fe286ef09a16-kube-api-access-fgc92\") on node \"crc\" DevicePath \"\"" Oct 08 19:35:30 crc kubenswrapper[4988]: I1008 19:35:30.037050 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ff74m\" (UniqueName: \"kubernetes.io/projected/30972cbc-cdb8-4379-a21d-4d727074ab7c-kube-api-access-ff74m\") pod \"mariadb-client-6-default\" (UID: \"30972cbc-cdb8-4379-a21d-4d727074ab7c\") " pod="openstack/mariadb-client-6-default" Oct 08 19:35:30 crc kubenswrapper[4988]: I1008 19:35:30.059858 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ff74m\" (UniqueName: \"kubernetes.io/projected/30972cbc-cdb8-4379-a21d-4d727074ab7c-kube-api-access-ff74m\") pod \"mariadb-client-6-default\" (UID: \"30972cbc-cdb8-4379-a21d-4d727074ab7c\") " pod="openstack/mariadb-client-6-default" Oct 08 19:35:30 crc kubenswrapper[4988]: I1008 19:35:30.221209 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-6-default" Oct 08 19:35:30 crc kubenswrapper[4988]: I1008 19:35:30.279842 4988 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d133be4f76344b50471853406cf2ce27c3a233e8926a2d19f70a721984ecefee" Oct 08 19:35:30 crc kubenswrapper[4988]: I1008 19:35:30.280018 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-5-default" Oct 08 19:35:30 crc kubenswrapper[4988]: I1008 19:35:30.615754 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client-6-default"] Oct 08 19:35:30 crc kubenswrapper[4988]: W1008 19:35:30.619456 4988 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod30972cbc_cdb8_4379_a21d_4d727074ab7c.slice/crio-2fc7eba33623f8e582288851af2e1a0bde3095469bacccb36843c7b32567b568 WatchSource:0}: Error finding container 2fc7eba33623f8e582288851af2e1a0bde3095469bacccb36843c7b32567b568: Status 404 returned error can't find the container with id 2fc7eba33623f8e582288851af2e1a0bde3095469bacccb36843c7b32567b568 Oct 08 19:35:31 crc kubenswrapper[4988]: I1008 19:35:31.248625 4988 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="56fad65a-577b-4227-a3fa-fe286ef09a16" path="/var/lib/kubelet/pods/56fad65a-577b-4227-a3fa-fe286ef09a16/volumes" Oct 08 19:35:31 crc kubenswrapper[4988]: I1008 19:35:31.291915 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-6-default" event={"ID":"30972cbc-cdb8-4379-a21d-4d727074ab7c","Type":"ContainerStarted","Data":"bd8e78de51a3113df5ef108a4b3654ce67ad603f688627031a2267254b61c993"} Oct 08 19:35:31 crc kubenswrapper[4988]: I1008 19:35:31.291982 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-6-default" event={"ID":"30972cbc-cdb8-4379-a21d-4d727074ab7c","Type":"ContainerStarted","Data":"2fc7eba33623f8e582288851af2e1a0bde3095469bacccb36843c7b32567b568"} Oct 08 19:35:31 crc kubenswrapper[4988]: I1008 19:35:31.320215 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/mariadb-client-6-default" podStartSLOduration=2.320188968 podStartE2EDuration="2.320188968s" podCreationTimestamp="2025-10-08 19:35:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 19:35:31.315681315 +0000 UTC m=+5076.765524156" watchObservedRunningTime="2025-10-08 19:35:31.320188968 +0000 UTC m=+5076.770031778" Oct 08 19:35:31 crc kubenswrapper[4988]: I1008 19:35:31.963788 4988 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_mariadb-client-6-default_30972cbc-cdb8-4379-a21d-4d727074ab7c/mariadb-client-6-default/0.log" Oct 08 19:35:32 crc kubenswrapper[4988]: I1008 19:35:32.305729 4988 generic.go:334] "Generic (PLEG): container finished" podID="30972cbc-cdb8-4379-a21d-4d727074ab7c" containerID="bd8e78de51a3113df5ef108a4b3654ce67ad603f688627031a2267254b61c993" exitCode=0 Oct 08 19:35:32 crc kubenswrapper[4988]: I1008 19:35:32.305814 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-6-default" event={"ID":"30972cbc-cdb8-4379-a21d-4d727074ab7c","Type":"ContainerDied","Data":"bd8e78de51a3113df5ef108a4b3654ce67ad603f688627031a2267254b61c993"} Oct 08 19:35:33 crc kubenswrapper[4988]: I1008 19:35:33.804772 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-6-default" Oct 08 19:35:33 crc kubenswrapper[4988]: I1008 19:35:33.847706 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mariadb-client-6-default"] Oct 08 19:35:33 crc kubenswrapper[4988]: I1008 19:35:33.857736 4988 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/mariadb-client-6-default"] Oct 08 19:35:33 crc kubenswrapper[4988]: I1008 19:35:33.899286 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ff74m\" (UniqueName: \"kubernetes.io/projected/30972cbc-cdb8-4379-a21d-4d727074ab7c-kube-api-access-ff74m\") pod \"30972cbc-cdb8-4379-a21d-4d727074ab7c\" (UID: \"30972cbc-cdb8-4379-a21d-4d727074ab7c\") " Oct 08 19:35:33 crc kubenswrapper[4988]: I1008 19:35:33.904225 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/30972cbc-cdb8-4379-a21d-4d727074ab7c-kube-api-access-ff74m" (OuterVolumeSpecName: "kube-api-access-ff74m") pod "30972cbc-cdb8-4379-a21d-4d727074ab7c" (UID: "30972cbc-cdb8-4379-a21d-4d727074ab7c"). InnerVolumeSpecName "kube-api-access-ff74m". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 19:35:33 crc kubenswrapper[4988]: I1008 19:35:33.995089 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/mariadb-client-7-default"] Oct 08 19:35:33 crc kubenswrapper[4988]: E1008 19:35:33.995406 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="30972cbc-cdb8-4379-a21d-4d727074ab7c" containerName="mariadb-client-6-default" Oct 08 19:35:33 crc kubenswrapper[4988]: I1008 19:35:33.995422 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="30972cbc-cdb8-4379-a21d-4d727074ab7c" containerName="mariadb-client-6-default" Oct 08 19:35:33 crc kubenswrapper[4988]: I1008 19:35:33.995589 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="30972cbc-cdb8-4379-a21d-4d727074ab7c" containerName="mariadb-client-6-default" Oct 08 19:35:33 crc kubenswrapper[4988]: I1008 19:35:33.996245 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-7-default" Oct 08 19:35:34 crc kubenswrapper[4988]: I1008 19:35:34.000364 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-st6vr\" (UniqueName: \"kubernetes.io/projected/48879998-4a1e-4e00-920d-c8589ea660d8-kube-api-access-st6vr\") pod \"mariadb-client-7-default\" (UID: \"48879998-4a1e-4e00-920d-c8589ea660d8\") " pod="openstack/mariadb-client-7-default" Oct 08 19:35:34 crc kubenswrapper[4988]: I1008 19:35:34.001077 4988 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ff74m\" (UniqueName: \"kubernetes.io/projected/30972cbc-cdb8-4379-a21d-4d727074ab7c-kube-api-access-ff74m\") on node \"crc\" DevicePath \"\"" Oct 08 19:35:34 crc kubenswrapper[4988]: I1008 19:35:34.006777 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client-7-default"] Oct 08 19:35:34 crc kubenswrapper[4988]: I1008 19:35:34.101738 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-st6vr\" (UniqueName: \"kubernetes.io/projected/48879998-4a1e-4e00-920d-c8589ea660d8-kube-api-access-st6vr\") pod \"mariadb-client-7-default\" (UID: \"48879998-4a1e-4e00-920d-c8589ea660d8\") " pod="openstack/mariadb-client-7-default" Oct 08 19:35:34 crc kubenswrapper[4988]: I1008 19:35:34.133073 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-st6vr\" (UniqueName: \"kubernetes.io/projected/48879998-4a1e-4e00-920d-c8589ea660d8-kube-api-access-st6vr\") pod \"mariadb-client-7-default\" (UID: \"48879998-4a1e-4e00-920d-c8589ea660d8\") " pod="openstack/mariadb-client-7-default" Oct 08 19:35:34 crc kubenswrapper[4988]: I1008 19:35:34.317369 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-7-default" Oct 08 19:35:34 crc kubenswrapper[4988]: I1008 19:35:34.334456 4988 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2fc7eba33623f8e582288851af2e1a0bde3095469bacccb36843c7b32567b568" Oct 08 19:35:34 crc kubenswrapper[4988]: I1008 19:35:34.334559 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-6-default" Oct 08 19:35:34 crc kubenswrapper[4988]: I1008 19:35:34.877055 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client-7-default"] Oct 08 19:35:35 crc kubenswrapper[4988]: I1008 19:35:35.255897 4988 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="30972cbc-cdb8-4379-a21d-4d727074ab7c" path="/var/lib/kubelet/pods/30972cbc-cdb8-4379-a21d-4d727074ab7c/volumes" Oct 08 19:35:35 crc kubenswrapper[4988]: I1008 19:35:35.347594 4988 generic.go:334] "Generic (PLEG): container finished" podID="48879998-4a1e-4e00-920d-c8589ea660d8" containerID="442acb8dc7a01aa4df3c793d24fc54425b1084aa512057dd2c64d1eecff0790b" exitCode=0 Oct 08 19:35:35 crc kubenswrapper[4988]: I1008 19:35:35.347662 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-7-default" event={"ID":"48879998-4a1e-4e00-920d-c8589ea660d8","Type":"ContainerDied","Data":"442acb8dc7a01aa4df3c793d24fc54425b1084aa512057dd2c64d1eecff0790b"} Oct 08 19:35:35 crc kubenswrapper[4988]: I1008 19:35:35.347702 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-7-default" event={"ID":"48879998-4a1e-4e00-920d-c8589ea660d8","Type":"ContainerStarted","Data":"773cf77a17b5fe0ca4aca114b8489aa99c20886cb51105ef6f7da6d7ef756160"} Oct 08 19:35:36 crc kubenswrapper[4988]: I1008 19:35:36.904040 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-7-default" Oct 08 19:35:36 crc kubenswrapper[4988]: I1008 19:35:36.921478 4988 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_mariadb-client-7-default_48879998-4a1e-4e00-920d-c8589ea660d8/mariadb-client-7-default/0.log" Oct 08 19:35:36 crc kubenswrapper[4988]: I1008 19:35:36.953138 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mariadb-client-7-default"] Oct 08 19:35:36 crc kubenswrapper[4988]: I1008 19:35:36.958233 4988 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/mariadb-client-7-default"] Oct 08 19:35:37 crc kubenswrapper[4988]: I1008 19:35:37.052072 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-st6vr\" (UniqueName: \"kubernetes.io/projected/48879998-4a1e-4e00-920d-c8589ea660d8-kube-api-access-st6vr\") pod \"48879998-4a1e-4e00-920d-c8589ea660d8\" (UID: \"48879998-4a1e-4e00-920d-c8589ea660d8\") " Oct 08 19:35:37 crc kubenswrapper[4988]: I1008 19:35:37.060832 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/48879998-4a1e-4e00-920d-c8589ea660d8-kube-api-access-st6vr" (OuterVolumeSpecName: "kube-api-access-st6vr") pod "48879998-4a1e-4e00-920d-c8589ea660d8" (UID: "48879998-4a1e-4e00-920d-c8589ea660d8"). InnerVolumeSpecName "kube-api-access-st6vr". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 19:35:37 crc kubenswrapper[4988]: I1008 19:35:37.144747 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/mariadb-client-2"] Oct 08 19:35:37 crc kubenswrapper[4988]: E1008 19:35:37.145522 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="48879998-4a1e-4e00-920d-c8589ea660d8" containerName="mariadb-client-7-default" Oct 08 19:35:37 crc kubenswrapper[4988]: I1008 19:35:37.145547 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="48879998-4a1e-4e00-920d-c8589ea660d8" containerName="mariadb-client-7-default" Oct 08 19:35:37 crc kubenswrapper[4988]: I1008 19:35:37.145789 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="48879998-4a1e-4e00-920d-c8589ea660d8" containerName="mariadb-client-7-default" Oct 08 19:35:37 crc kubenswrapper[4988]: I1008 19:35:37.146793 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-2" Oct 08 19:35:37 crc kubenswrapper[4988]: I1008 19:35:37.154240 4988 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-st6vr\" (UniqueName: \"kubernetes.io/projected/48879998-4a1e-4e00-920d-c8589ea660d8-kube-api-access-st6vr\") on node \"crc\" DevicePath \"\"" Oct 08 19:35:37 crc kubenswrapper[4988]: I1008 19:35:37.154353 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client-2"] Oct 08 19:35:37 crc kubenswrapper[4988]: I1008 19:35:37.248230 4988 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="48879998-4a1e-4e00-920d-c8589ea660d8" path="/var/lib/kubelet/pods/48879998-4a1e-4e00-920d-c8589ea660d8/volumes" Oct 08 19:35:37 crc kubenswrapper[4988]: I1008 19:35:37.255981 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jg7hn\" (UniqueName: \"kubernetes.io/projected/e90080b2-813b-493a-bf34-9403347a33e0-kube-api-access-jg7hn\") pod \"mariadb-client-2\" (UID: \"e90080b2-813b-493a-bf34-9403347a33e0\") " pod="openstack/mariadb-client-2" Oct 08 19:35:37 crc kubenswrapper[4988]: I1008 19:35:37.358632 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jg7hn\" (UniqueName: \"kubernetes.io/projected/e90080b2-813b-493a-bf34-9403347a33e0-kube-api-access-jg7hn\") pod \"mariadb-client-2\" (UID: \"e90080b2-813b-493a-bf34-9403347a33e0\") " pod="openstack/mariadb-client-2" Oct 08 19:35:37 crc kubenswrapper[4988]: I1008 19:35:37.367816 4988 scope.go:117] "RemoveContainer" containerID="442acb8dc7a01aa4df3c793d24fc54425b1084aa512057dd2c64d1eecff0790b" Oct 08 19:35:37 crc kubenswrapper[4988]: I1008 19:35:37.367886 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-7-default" Oct 08 19:35:37 crc kubenswrapper[4988]: I1008 19:35:37.385015 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jg7hn\" (UniqueName: \"kubernetes.io/projected/e90080b2-813b-493a-bf34-9403347a33e0-kube-api-access-jg7hn\") pod \"mariadb-client-2\" (UID: \"e90080b2-813b-493a-bf34-9403347a33e0\") " pod="openstack/mariadb-client-2" Oct 08 19:35:37 crc kubenswrapper[4988]: I1008 19:35:37.470840 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-2" Oct 08 19:35:38 crc kubenswrapper[4988]: I1008 19:35:38.113872 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client-2"] Oct 08 19:35:38 crc kubenswrapper[4988]: W1008 19:35:38.122979 4988 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode90080b2_813b_493a_bf34_9403347a33e0.slice/crio-2d9131ea6588c686b452b125656b0067222e991fa558e01d90aa438bf82c7d21 WatchSource:0}: Error finding container 2d9131ea6588c686b452b125656b0067222e991fa558e01d90aa438bf82c7d21: Status 404 returned error can't find the container with id 2d9131ea6588c686b452b125656b0067222e991fa558e01d90aa438bf82c7d21 Oct 08 19:35:38 crc kubenswrapper[4988]: I1008 19:35:38.383937 4988 generic.go:334] "Generic (PLEG): container finished" podID="e90080b2-813b-493a-bf34-9403347a33e0" containerID="e9f20141a2a09213dbeecf8fa028433b50c47f2435e762d362eae8fa7625d987" exitCode=0 Oct 08 19:35:38 crc kubenswrapper[4988]: I1008 19:35:38.384036 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-2" event={"ID":"e90080b2-813b-493a-bf34-9403347a33e0","Type":"ContainerDied","Data":"e9f20141a2a09213dbeecf8fa028433b50c47f2435e762d362eae8fa7625d987"} Oct 08 19:35:38 crc kubenswrapper[4988]: I1008 19:35:38.384072 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-2" event={"ID":"e90080b2-813b-493a-bf34-9403347a33e0","Type":"ContainerStarted","Data":"2d9131ea6588c686b452b125656b0067222e991fa558e01d90aa438bf82c7d21"} Oct 08 19:35:39 crc kubenswrapper[4988]: I1008 19:35:39.770910 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-2" Oct 08 19:35:39 crc kubenswrapper[4988]: I1008 19:35:39.797587 4988 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_mariadb-client-2_e90080b2-813b-493a-bf34-9403347a33e0/mariadb-client-2/0.log" Oct 08 19:35:39 crc kubenswrapper[4988]: I1008 19:35:39.826428 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mariadb-client-2"] Oct 08 19:35:39 crc kubenswrapper[4988]: I1008 19:35:39.836177 4988 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/mariadb-client-2"] Oct 08 19:35:39 crc kubenswrapper[4988]: I1008 19:35:39.897499 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jg7hn\" (UniqueName: \"kubernetes.io/projected/e90080b2-813b-493a-bf34-9403347a33e0-kube-api-access-jg7hn\") pod \"e90080b2-813b-493a-bf34-9403347a33e0\" (UID: \"e90080b2-813b-493a-bf34-9403347a33e0\") " Oct 08 19:35:39 crc kubenswrapper[4988]: I1008 19:35:39.904465 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e90080b2-813b-493a-bf34-9403347a33e0-kube-api-access-jg7hn" (OuterVolumeSpecName: "kube-api-access-jg7hn") pod "e90080b2-813b-493a-bf34-9403347a33e0" (UID: "e90080b2-813b-493a-bf34-9403347a33e0"). InnerVolumeSpecName "kube-api-access-jg7hn". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 19:35:39 crc kubenswrapper[4988]: I1008 19:35:39.999846 4988 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jg7hn\" (UniqueName: \"kubernetes.io/projected/e90080b2-813b-493a-bf34-9403347a33e0-kube-api-access-jg7hn\") on node \"crc\" DevicePath \"\"" Oct 08 19:35:40 crc kubenswrapper[4988]: I1008 19:35:40.408156 4988 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2d9131ea6588c686b452b125656b0067222e991fa558e01d90aa438bf82c7d21" Oct 08 19:35:40 crc kubenswrapper[4988]: I1008 19:35:40.408238 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-2" Oct 08 19:35:41 crc kubenswrapper[4988]: I1008 19:35:41.237762 4988 scope.go:117] "RemoveContainer" containerID="3cd52d2ff674f037f4b111ff08caa426d4d7977052a05e385f05362baabb8da9" Oct 08 19:35:41 crc kubenswrapper[4988]: E1008 19:35:41.238449 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nm54f_openshift-machine-config-operator(f328a20b-5283-407e-8a93-2ab2da599f18)\"" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" Oct 08 19:35:41 crc kubenswrapper[4988]: I1008 19:35:41.255997 4988 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e90080b2-813b-493a-bf34-9403347a33e0" path="/var/lib/kubelet/pods/e90080b2-813b-493a-bf34-9403347a33e0/volumes" Oct 08 19:35:52 crc kubenswrapper[4988]: I1008 19:35:52.238246 4988 scope.go:117] "RemoveContainer" containerID="3cd52d2ff674f037f4b111ff08caa426d4d7977052a05e385f05362baabb8da9" Oct 08 19:35:52 crc kubenswrapper[4988]: E1008 19:35:52.239355 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nm54f_openshift-machine-config-operator(f328a20b-5283-407e-8a93-2ab2da599f18)\"" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" Oct 08 19:36:07 crc kubenswrapper[4988]: I1008 19:36:07.239350 4988 scope.go:117] "RemoveContainer" containerID="3cd52d2ff674f037f4b111ff08caa426d4d7977052a05e385f05362baabb8da9" Oct 08 19:36:07 crc kubenswrapper[4988]: E1008 19:36:07.240545 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nm54f_openshift-machine-config-operator(f328a20b-5283-407e-8a93-2ab2da599f18)\"" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" Oct 08 19:36:16 crc kubenswrapper[4988]: I1008 19:36:16.969758 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-9tsnv"] Oct 08 19:36:16 crc kubenswrapper[4988]: E1008 19:36:16.970597 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e90080b2-813b-493a-bf34-9403347a33e0" containerName="mariadb-client-2" Oct 08 19:36:16 crc kubenswrapper[4988]: I1008 19:36:16.970610 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="e90080b2-813b-493a-bf34-9403347a33e0" containerName="mariadb-client-2" Oct 08 19:36:16 crc kubenswrapper[4988]: I1008 19:36:16.970759 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="e90080b2-813b-493a-bf34-9403347a33e0" containerName="mariadb-client-2" Oct 08 19:36:16 crc kubenswrapper[4988]: I1008 19:36:16.971814 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-9tsnv" Oct 08 19:36:17 crc kubenswrapper[4988]: I1008 19:36:17.023957 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-9tsnv"] Oct 08 19:36:17 crc kubenswrapper[4988]: I1008 19:36:17.043661 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/80b6de22-319e-4d60-894a-616ea20ac81d-utilities\") pod \"redhat-marketplace-9tsnv\" (UID: \"80b6de22-319e-4d60-894a-616ea20ac81d\") " pod="openshift-marketplace/redhat-marketplace-9tsnv" Oct 08 19:36:17 crc kubenswrapper[4988]: I1008 19:36:17.043783 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/80b6de22-319e-4d60-894a-616ea20ac81d-catalog-content\") pod \"redhat-marketplace-9tsnv\" (UID: \"80b6de22-319e-4d60-894a-616ea20ac81d\") " pod="openshift-marketplace/redhat-marketplace-9tsnv" Oct 08 19:36:17 crc kubenswrapper[4988]: I1008 19:36:17.043843 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zlqx4\" (UniqueName: \"kubernetes.io/projected/80b6de22-319e-4d60-894a-616ea20ac81d-kube-api-access-zlqx4\") pod \"redhat-marketplace-9tsnv\" (UID: \"80b6de22-319e-4d60-894a-616ea20ac81d\") " pod="openshift-marketplace/redhat-marketplace-9tsnv" Oct 08 19:36:17 crc kubenswrapper[4988]: I1008 19:36:17.144765 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/80b6de22-319e-4d60-894a-616ea20ac81d-catalog-content\") pod \"redhat-marketplace-9tsnv\" (UID: \"80b6de22-319e-4d60-894a-616ea20ac81d\") " pod="openshift-marketplace/redhat-marketplace-9tsnv" Oct 08 19:36:17 crc kubenswrapper[4988]: I1008 19:36:17.144827 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zlqx4\" (UniqueName: \"kubernetes.io/projected/80b6de22-319e-4d60-894a-616ea20ac81d-kube-api-access-zlqx4\") pod \"redhat-marketplace-9tsnv\" (UID: \"80b6de22-319e-4d60-894a-616ea20ac81d\") " pod="openshift-marketplace/redhat-marketplace-9tsnv" Oct 08 19:36:17 crc kubenswrapper[4988]: I1008 19:36:17.144899 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/80b6de22-319e-4d60-894a-616ea20ac81d-utilities\") pod \"redhat-marketplace-9tsnv\" (UID: \"80b6de22-319e-4d60-894a-616ea20ac81d\") " pod="openshift-marketplace/redhat-marketplace-9tsnv" Oct 08 19:36:17 crc kubenswrapper[4988]: I1008 19:36:17.145340 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/80b6de22-319e-4d60-894a-616ea20ac81d-utilities\") pod \"redhat-marketplace-9tsnv\" (UID: \"80b6de22-319e-4d60-894a-616ea20ac81d\") " pod="openshift-marketplace/redhat-marketplace-9tsnv" Oct 08 19:36:17 crc kubenswrapper[4988]: I1008 19:36:17.145861 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/80b6de22-319e-4d60-894a-616ea20ac81d-catalog-content\") pod \"redhat-marketplace-9tsnv\" (UID: \"80b6de22-319e-4d60-894a-616ea20ac81d\") " pod="openshift-marketplace/redhat-marketplace-9tsnv" Oct 08 19:36:17 crc kubenswrapper[4988]: I1008 19:36:17.174799 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zlqx4\" (UniqueName: \"kubernetes.io/projected/80b6de22-319e-4d60-894a-616ea20ac81d-kube-api-access-zlqx4\") pod \"redhat-marketplace-9tsnv\" (UID: \"80b6de22-319e-4d60-894a-616ea20ac81d\") " pod="openshift-marketplace/redhat-marketplace-9tsnv" Oct 08 19:36:17 crc kubenswrapper[4988]: I1008 19:36:17.287954 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-9tsnv" Oct 08 19:36:17 crc kubenswrapper[4988]: I1008 19:36:17.721284 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-9tsnv"] Oct 08 19:36:17 crc kubenswrapper[4988]: W1008 19:36:17.727828 4988 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod80b6de22_319e_4d60_894a_616ea20ac81d.slice/crio-a2be18b20d604279570cdc2a4febf578de1e452926b11baadcea9da0b0d4a4f4 WatchSource:0}: Error finding container a2be18b20d604279570cdc2a4febf578de1e452926b11baadcea9da0b0d4a4f4: Status 404 returned error can't find the container with id a2be18b20d604279570cdc2a4febf578de1e452926b11baadcea9da0b0d4a4f4 Oct 08 19:36:17 crc kubenswrapper[4988]: I1008 19:36:17.817214 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-9tsnv" event={"ID":"80b6de22-319e-4d60-894a-616ea20ac81d","Type":"ContainerStarted","Data":"a2be18b20d604279570cdc2a4febf578de1e452926b11baadcea9da0b0d4a4f4"} Oct 08 19:36:18 crc kubenswrapper[4988]: I1008 19:36:18.829435 4988 generic.go:334] "Generic (PLEG): container finished" podID="80b6de22-319e-4d60-894a-616ea20ac81d" containerID="897ac8c560f958708decbc1cd2966ec6fc86910487b6c64a4b0c3665325aa543" exitCode=0 Oct 08 19:36:18 crc kubenswrapper[4988]: I1008 19:36:18.829505 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-9tsnv" event={"ID":"80b6de22-319e-4d60-894a-616ea20ac81d","Type":"ContainerDied","Data":"897ac8c560f958708decbc1cd2966ec6fc86910487b6c64a4b0c3665325aa543"} Oct 08 19:36:19 crc kubenswrapper[4988]: I1008 19:36:19.843828 4988 generic.go:334] "Generic (PLEG): container finished" podID="80b6de22-319e-4d60-894a-616ea20ac81d" containerID="7aa8d788a98678c56fdf0ea009a5fb8236e5ed2f0a29d1f504ec743465fa23fb" exitCode=0 Oct 08 19:36:19 crc kubenswrapper[4988]: I1008 19:36:19.843891 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-9tsnv" event={"ID":"80b6de22-319e-4d60-894a-616ea20ac81d","Type":"ContainerDied","Data":"7aa8d788a98678c56fdf0ea009a5fb8236e5ed2f0a29d1f504ec743465fa23fb"} Oct 08 19:36:20 crc kubenswrapper[4988]: I1008 19:36:20.238142 4988 scope.go:117] "RemoveContainer" containerID="3cd52d2ff674f037f4b111ff08caa426d4d7977052a05e385f05362baabb8da9" Oct 08 19:36:20 crc kubenswrapper[4988]: E1008 19:36:20.238505 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nm54f_openshift-machine-config-operator(f328a20b-5283-407e-8a93-2ab2da599f18)\"" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" Oct 08 19:36:20 crc kubenswrapper[4988]: I1008 19:36:20.855958 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-9tsnv" event={"ID":"80b6de22-319e-4d60-894a-616ea20ac81d","Type":"ContainerStarted","Data":"7636cad6b316779da704dbc370ffbe266b2808e0acfe2ad5286926899a7d91d9"} Oct 08 19:36:20 crc kubenswrapper[4988]: I1008 19:36:20.876263 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-9tsnv" podStartSLOduration=3.314951988 podStartE2EDuration="4.876242328s" podCreationTimestamp="2025-10-08 19:36:16 +0000 UTC" firstStartedPulling="2025-10-08 19:36:18.832842183 +0000 UTC m=+5124.282684963" lastFinishedPulling="2025-10-08 19:36:20.394132523 +0000 UTC m=+5125.843975303" observedRunningTime="2025-10-08 19:36:20.873633166 +0000 UTC m=+5126.323475936" watchObservedRunningTime="2025-10-08 19:36:20.876242328 +0000 UTC m=+5126.326085108" Oct 08 19:36:27 crc kubenswrapper[4988]: I1008 19:36:27.288797 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-9tsnv" Oct 08 19:36:27 crc kubenswrapper[4988]: I1008 19:36:27.290599 4988 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-9tsnv" Oct 08 19:36:27 crc kubenswrapper[4988]: I1008 19:36:27.357799 4988 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-9tsnv" Oct 08 19:36:27 crc kubenswrapper[4988]: I1008 19:36:27.983269 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-9tsnv" Oct 08 19:36:28 crc kubenswrapper[4988]: I1008 19:36:28.051610 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-9tsnv"] Oct 08 19:36:29 crc kubenswrapper[4988]: I1008 19:36:29.946805 4988 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-9tsnv" podUID="80b6de22-319e-4d60-894a-616ea20ac81d" containerName="registry-server" containerID="cri-o://7636cad6b316779da704dbc370ffbe266b2808e0acfe2ad5286926899a7d91d9" gracePeriod=2 Oct 08 19:36:30 crc kubenswrapper[4988]: I1008 19:36:30.381466 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-9tsnv" Oct 08 19:36:30 crc kubenswrapper[4988]: I1008 19:36:30.472314 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zlqx4\" (UniqueName: \"kubernetes.io/projected/80b6de22-319e-4d60-894a-616ea20ac81d-kube-api-access-zlqx4\") pod \"80b6de22-319e-4d60-894a-616ea20ac81d\" (UID: \"80b6de22-319e-4d60-894a-616ea20ac81d\") " Oct 08 19:36:30 crc kubenswrapper[4988]: I1008 19:36:30.472486 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/80b6de22-319e-4d60-894a-616ea20ac81d-catalog-content\") pod \"80b6de22-319e-4d60-894a-616ea20ac81d\" (UID: \"80b6de22-319e-4d60-894a-616ea20ac81d\") " Oct 08 19:36:30 crc kubenswrapper[4988]: I1008 19:36:30.472595 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/80b6de22-319e-4d60-894a-616ea20ac81d-utilities\") pod \"80b6de22-319e-4d60-894a-616ea20ac81d\" (UID: \"80b6de22-319e-4d60-894a-616ea20ac81d\") " Oct 08 19:36:30 crc kubenswrapper[4988]: I1008 19:36:30.473351 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/80b6de22-319e-4d60-894a-616ea20ac81d-utilities" (OuterVolumeSpecName: "utilities") pod "80b6de22-319e-4d60-894a-616ea20ac81d" (UID: "80b6de22-319e-4d60-894a-616ea20ac81d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 19:36:30 crc kubenswrapper[4988]: I1008 19:36:30.477671 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/80b6de22-319e-4d60-894a-616ea20ac81d-kube-api-access-zlqx4" (OuterVolumeSpecName: "kube-api-access-zlqx4") pod "80b6de22-319e-4d60-894a-616ea20ac81d" (UID: "80b6de22-319e-4d60-894a-616ea20ac81d"). InnerVolumeSpecName "kube-api-access-zlqx4". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 19:36:30 crc kubenswrapper[4988]: I1008 19:36:30.490200 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/80b6de22-319e-4d60-894a-616ea20ac81d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "80b6de22-319e-4d60-894a-616ea20ac81d" (UID: "80b6de22-319e-4d60-894a-616ea20ac81d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 19:36:30 crc kubenswrapper[4988]: I1008 19:36:30.574501 4988 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zlqx4\" (UniqueName: \"kubernetes.io/projected/80b6de22-319e-4d60-894a-616ea20ac81d-kube-api-access-zlqx4\") on node \"crc\" DevicePath \"\"" Oct 08 19:36:30 crc kubenswrapper[4988]: I1008 19:36:30.574556 4988 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/80b6de22-319e-4d60-894a-616ea20ac81d-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 08 19:36:30 crc kubenswrapper[4988]: I1008 19:36:30.574576 4988 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/80b6de22-319e-4d60-894a-616ea20ac81d-utilities\") on node \"crc\" DevicePath \"\"" Oct 08 19:36:30 crc kubenswrapper[4988]: I1008 19:36:30.957334 4988 generic.go:334] "Generic (PLEG): container finished" podID="80b6de22-319e-4d60-894a-616ea20ac81d" containerID="7636cad6b316779da704dbc370ffbe266b2808e0acfe2ad5286926899a7d91d9" exitCode=0 Oct 08 19:36:30 crc kubenswrapper[4988]: I1008 19:36:30.957433 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-9tsnv" Oct 08 19:36:30 crc kubenswrapper[4988]: I1008 19:36:30.957483 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-9tsnv" event={"ID":"80b6de22-319e-4d60-894a-616ea20ac81d","Type":"ContainerDied","Data":"7636cad6b316779da704dbc370ffbe266b2808e0acfe2ad5286926899a7d91d9"} Oct 08 19:36:30 crc kubenswrapper[4988]: I1008 19:36:30.957968 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-9tsnv" event={"ID":"80b6de22-319e-4d60-894a-616ea20ac81d","Type":"ContainerDied","Data":"a2be18b20d604279570cdc2a4febf578de1e452926b11baadcea9da0b0d4a4f4"} Oct 08 19:36:30 crc kubenswrapper[4988]: I1008 19:36:30.958091 4988 scope.go:117] "RemoveContainer" containerID="7636cad6b316779da704dbc370ffbe266b2808e0acfe2ad5286926899a7d91d9" Oct 08 19:36:30 crc kubenswrapper[4988]: I1008 19:36:30.987738 4988 scope.go:117] "RemoveContainer" containerID="7aa8d788a98678c56fdf0ea009a5fb8236e5ed2f0a29d1f504ec743465fa23fb" Oct 08 19:36:30 crc kubenswrapper[4988]: I1008 19:36:30.999026 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-9tsnv"] Oct 08 19:36:31 crc kubenswrapper[4988]: I1008 19:36:31.007347 4988 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-9tsnv"] Oct 08 19:36:31 crc kubenswrapper[4988]: I1008 19:36:31.028057 4988 scope.go:117] "RemoveContainer" containerID="897ac8c560f958708decbc1cd2966ec6fc86910487b6c64a4b0c3665325aa543" Oct 08 19:36:31 crc kubenswrapper[4988]: I1008 19:36:31.060730 4988 scope.go:117] "RemoveContainer" containerID="7636cad6b316779da704dbc370ffbe266b2808e0acfe2ad5286926899a7d91d9" Oct 08 19:36:31 crc kubenswrapper[4988]: E1008 19:36:31.061499 4988 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7636cad6b316779da704dbc370ffbe266b2808e0acfe2ad5286926899a7d91d9\": container with ID starting with 7636cad6b316779da704dbc370ffbe266b2808e0acfe2ad5286926899a7d91d9 not found: ID does not exist" containerID="7636cad6b316779da704dbc370ffbe266b2808e0acfe2ad5286926899a7d91d9" Oct 08 19:36:31 crc kubenswrapper[4988]: I1008 19:36:31.061533 4988 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7636cad6b316779da704dbc370ffbe266b2808e0acfe2ad5286926899a7d91d9"} err="failed to get container status \"7636cad6b316779da704dbc370ffbe266b2808e0acfe2ad5286926899a7d91d9\": rpc error: code = NotFound desc = could not find container \"7636cad6b316779da704dbc370ffbe266b2808e0acfe2ad5286926899a7d91d9\": container with ID starting with 7636cad6b316779da704dbc370ffbe266b2808e0acfe2ad5286926899a7d91d9 not found: ID does not exist" Oct 08 19:36:31 crc kubenswrapper[4988]: I1008 19:36:31.061570 4988 scope.go:117] "RemoveContainer" containerID="7aa8d788a98678c56fdf0ea009a5fb8236e5ed2f0a29d1f504ec743465fa23fb" Oct 08 19:36:31 crc kubenswrapper[4988]: E1008 19:36:31.062010 4988 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7aa8d788a98678c56fdf0ea009a5fb8236e5ed2f0a29d1f504ec743465fa23fb\": container with ID starting with 7aa8d788a98678c56fdf0ea009a5fb8236e5ed2f0a29d1f504ec743465fa23fb not found: ID does not exist" containerID="7aa8d788a98678c56fdf0ea009a5fb8236e5ed2f0a29d1f504ec743465fa23fb" Oct 08 19:36:31 crc kubenswrapper[4988]: I1008 19:36:31.062084 4988 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7aa8d788a98678c56fdf0ea009a5fb8236e5ed2f0a29d1f504ec743465fa23fb"} err="failed to get container status \"7aa8d788a98678c56fdf0ea009a5fb8236e5ed2f0a29d1f504ec743465fa23fb\": rpc error: code = NotFound desc = could not find container \"7aa8d788a98678c56fdf0ea009a5fb8236e5ed2f0a29d1f504ec743465fa23fb\": container with ID starting with 7aa8d788a98678c56fdf0ea009a5fb8236e5ed2f0a29d1f504ec743465fa23fb not found: ID does not exist" Oct 08 19:36:31 crc kubenswrapper[4988]: I1008 19:36:31.062130 4988 scope.go:117] "RemoveContainer" containerID="897ac8c560f958708decbc1cd2966ec6fc86910487b6c64a4b0c3665325aa543" Oct 08 19:36:31 crc kubenswrapper[4988]: E1008 19:36:31.062530 4988 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"897ac8c560f958708decbc1cd2966ec6fc86910487b6c64a4b0c3665325aa543\": container with ID starting with 897ac8c560f958708decbc1cd2966ec6fc86910487b6c64a4b0c3665325aa543 not found: ID does not exist" containerID="897ac8c560f958708decbc1cd2966ec6fc86910487b6c64a4b0c3665325aa543" Oct 08 19:36:31 crc kubenswrapper[4988]: I1008 19:36:31.062573 4988 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"897ac8c560f958708decbc1cd2966ec6fc86910487b6c64a4b0c3665325aa543"} err="failed to get container status \"897ac8c560f958708decbc1cd2966ec6fc86910487b6c64a4b0c3665325aa543\": rpc error: code = NotFound desc = could not find container \"897ac8c560f958708decbc1cd2966ec6fc86910487b6c64a4b0c3665325aa543\": container with ID starting with 897ac8c560f958708decbc1cd2966ec6fc86910487b6c64a4b0c3665325aa543 not found: ID does not exist" Oct 08 19:36:31 crc kubenswrapper[4988]: I1008 19:36:31.256223 4988 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="80b6de22-319e-4d60-894a-616ea20ac81d" path="/var/lib/kubelet/pods/80b6de22-319e-4d60-894a-616ea20ac81d/volumes" Oct 08 19:36:34 crc kubenswrapper[4988]: I1008 19:36:34.238552 4988 scope.go:117] "RemoveContainer" containerID="3cd52d2ff674f037f4b111ff08caa426d4d7977052a05e385f05362baabb8da9" Oct 08 19:36:34 crc kubenswrapper[4988]: E1008 19:36:34.239439 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nm54f_openshift-machine-config-operator(f328a20b-5283-407e-8a93-2ab2da599f18)\"" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" Oct 08 19:36:48 crc kubenswrapper[4988]: I1008 19:36:48.238048 4988 scope.go:117] "RemoveContainer" containerID="3cd52d2ff674f037f4b111ff08caa426d4d7977052a05e385f05362baabb8da9" Oct 08 19:36:48 crc kubenswrapper[4988]: E1008 19:36:48.238883 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nm54f_openshift-machine-config-operator(f328a20b-5283-407e-8a93-2ab2da599f18)\"" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" Oct 08 19:37:00 crc kubenswrapper[4988]: I1008 19:37:00.238112 4988 scope.go:117] "RemoveContainer" containerID="3cd52d2ff674f037f4b111ff08caa426d4d7977052a05e385f05362baabb8da9" Oct 08 19:37:00 crc kubenswrapper[4988]: E1008 19:37:00.239286 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nm54f_openshift-machine-config-operator(f328a20b-5283-407e-8a93-2ab2da599f18)\"" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" Oct 08 19:37:02 crc kubenswrapper[4988]: I1008 19:37:02.845030 4988 scope.go:117] "RemoveContainer" containerID="958238b7f2a84803fc5d7c9871650e2109caba0937a5700fc4a4e0599c4a5569" Oct 08 19:37:11 crc kubenswrapper[4988]: I1008 19:37:11.238021 4988 scope.go:117] "RemoveContainer" containerID="3cd52d2ff674f037f4b111ff08caa426d4d7977052a05e385f05362baabb8da9" Oct 08 19:37:11 crc kubenswrapper[4988]: E1008 19:37:11.238740 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nm54f_openshift-machine-config-operator(f328a20b-5283-407e-8a93-2ab2da599f18)\"" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" Oct 08 19:37:22 crc kubenswrapper[4988]: I1008 19:37:22.238722 4988 scope.go:117] "RemoveContainer" containerID="3cd52d2ff674f037f4b111ff08caa426d4d7977052a05e385f05362baabb8da9" Oct 08 19:37:22 crc kubenswrapper[4988]: E1008 19:37:22.239282 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nm54f_openshift-machine-config-operator(f328a20b-5283-407e-8a93-2ab2da599f18)\"" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" Oct 08 19:37:35 crc kubenswrapper[4988]: I1008 19:37:35.248503 4988 scope.go:117] "RemoveContainer" containerID="3cd52d2ff674f037f4b111ff08caa426d4d7977052a05e385f05362baabb8da9" Oct 08 19:37:35 crc kubenswrapper[4988]: E1008 19:37:35.249887 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nm54f_openshift-machine-config-operator(f328a20b-5283-407e-8a93-2ab2da599f18)\"" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" Oct 08 19:37:50 crc kubenswrapper[4988]: I1008 19:37:50.238261 4988 scope.go:117] "RemoveContainer" containerID="3cd52d2ff674f037f4b111ff08caa426d4d7977052a05e385f05362baabb8da9" Oct 08 19:37:50 crc kubenswrapper[4988]: E1008 19:37:50.239160 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nm54f_openshift-machine-config-operator(f328a20b-5283-407e-8a93-2ab2da599f18)\"" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" Oct 08 19:38:04 crc kubenswrapper[4988]: I1008 19:38:04.239665 4988 scope.go:117] "RemoveContainer" containerID="3cd52d2ff674f037f4b111ff08caa426d4d7977052a05e385f05362baabb8da9" Oct 08 19:38:04 crc kubenswrapper[4988]: E1008 19:38:04.240262 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nm54f_openshift-machine-config-operator(f328a20b-5283-407e-8a93-2ab2da599f18)\"" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" Oct 08 19:38:17 crc kubenswrapper[4988]: I1008 19:38:17.238842 4988 scope.go:117] "RemoveContainer" containerID="3cd52d2ff674f037f4b111ff08caa426d4d7977052a05e385f05362baabb8da9" Oct 08 19:38:17 crc kubenswrapper[4988]: E1008 19:38:17.239852 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nm54f_openshift-machine-config-operator(f328a20b-5283-407e-8a93-2ab2da599f18)\"" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" Oct 08 19:38:28 crc kubenswrapper[4988]: I1008 19:38:28.237906 4988 scope.go:117] "RemoveContainer" containerID="3cd52d2ff674f037f4b111ff08caa426d4d7977052a05e385f05362baabb8da9" Oct 08 19:38:28 crc kubenswrapper[4988]: E1008 19:38:28.238885 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nm54f_openshift-machine-config-operator(f328a20b-5283-407e-8a93-2ab2da599f18)\"" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" Oct 08 19:38:42 crc kubenswrapper[4988]: I1008 19:38:42.239026 4988 scope.go:117] "RemoveContainer" containerID="3cd52d2ff674f037f4b111ff08caa426d4d7977052a05e385f05362baabb8da9" Oct 08 19:38:42 crc kubenswrapper[4988]: E1008 19:38:42.240186 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nm54f_openshift-machine-config-operator(f328a20b-5283-407e-8a93-2ab2da599f18)\"" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" Oct 08 19:38:55 crc kubenswrapper[4988]: I1008 19:38:55.244738 4988 scope.go:117] "RemoveContainer" containerID="3cd52d2ff674f037f4b111ff08caa426d4d7977052a05e385f05362baabb8da9" Oct 08 19:38:55 crc kubenswrapper[4988]: E1008 19:38:55.245765 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nm54f_openshift-machine-config-operator(f328a20b-5283-407e-8a93-2ab2da599f18)\"" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" Oct 08 19:39:10 crc kubenswrapper[4988]: I1008 19:39:10.237733 4988 scope.go:117] "RemoveContainer" containerID="3cd52d2ff674f037f4b111ff08caa426d4d7977052a05e385f05362baabb8da9" Oct 08 19:39:10 crc kubenswrapper[4988]: E1008 19:39:10.238555 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nm54f_openshift-machine-config-operator(f328a20b-5283-407e-8a93-2ab2da599f18)\"" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" Oct 08 19:39:22 crc kubenswrapper[4988]: I1008 19:39:22.238473 4988 scope.go:117] "RemoveContainer" containerID="3cd52d2ff674f037f4b111ff08caa426d4d7977052a05e385f05362baabb8da9" Oct 08 19:39:22 crc kubenswrapper[4988]: E1008 19:39:22.239343 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nm54f_openshift-machine-config-operator(f328a20b-5283-407e-8a93-2ab2da599f18)\"" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" Oct 08 19:39:29 crc kubenswrapper[4988]: I1008 19:39:29.333431 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/mariadb-copy-data"] Oct 08 19:39:29 crc kubenswrapper[4988]: E1008 19:39:29.334470 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="80b6de22-319e-4d60-894a-616ea20ac81d" containerName="extract-content" Oct 08 19:39:29 crc kubenswrapper[4988]: I1008 19:39:29.334492 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="80b6de22-319e-4d60-894a-616ea20ac81d" containerName="extract-content" Oct 08 19:39:29 crc kubenswrapper[4988]: E1008 19:39:29.334507 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="80b6de22-319e-4d60-894a-616ea20ac81d" containerName="extract-utilities" Oct 08 19:39:29 crc kubenswrapper[4988]: I1008 19:39:29.334519 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="80b6de22-319e-4d60-894a-616ea20ac81d" containerName="extract-utilities" Oct 08 19:39:29 crc kubenswrapper[4988]: E1008 19:39:29.334572 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="80b6de22-319e-4d60-894a-616ea20ac81d" containerName="registry-server" Oct 08 19:39:29 crc kubenswrapper[4988]: I1008 19:39:29.334585 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="80b6de22-319e-4d60-894a-616ea20ac81d" containerName="registry-server" Oct 08 19:39:29 crc kubenswrapper[4988]: I1008 19:39:29.334849 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="80b6de22-319e-4d60-894a-616ea20ac81d" containerName="registry-server" Oct 08 19:39:29 crc kubenswrapper[4988]: I1008 19:39:29.335717 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-copy-data" Oct 08 19:39:29 crc kubenswrapper[4988]: I1008 19:39:29.338965 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"default-dockercfg-cmd56" Oct 08 19:39:29 crc kubenswrapper[4988]: I1008 19:39:29.377033 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-copy-data"] Oct 08 19:39:29 crc kubenswrapper[4988]: I1008 19:39:29.449029 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jw4zs\" (UniqueName: \"kubernetes.io/projected/9e312e3a-e0e2-490b-bb43-cc3b7191c109-kube-api-access-jw4zs\") pod \"mariadb-copy-data\" (UID: \"9e312e3a-e0e2-490b-bb43-cc3b7191c109\") " pod="openstack/mariadb-copy-data" Oct 08 19:39:29 crc kubenswrapper[4988]: I1008 19:39:29.449090 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-81095d3c-890a-4460-a497-531b459f10ce\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-81095d3c-890a-4460-a497-531b459f10ce\") pod \"mariadb-copy-data\" (UID: \"9e312e3a-e0e2-490b-bb43-cc3b7191c109\") " pod="openstack/mariadb-copy-data" Oct 08 19:39:29 crc kubenswrapper[4988]: I1008 19:39:29.550985 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jw4zs\" (UniqueName: \"kubernetes.io/projected/9e312e3a-e0e2-490b-bb43-cc3b7191c109-kube-api-access-jw4zs\") pod \"mariadb-copy-data\" (UID: \"9e312e3a-e0e2-490b-bb43-cc3b7191c109\") " pod="openstack/mariadb-copy-data" Oct 08 19:39:29 crc kubenswrapper[4988]: I1008 19:39:29.551068 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-81095d3c-890a-4460-a497-531b459f10ce\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-81095d3c-890a-4460-a497-531b459f10ce\") pod \"mariadb-copy-data\" (UID: \"9e312e3a-e0e2-490b-bb43-cc3b7191c109\") " pod="openstack/mariadb-copy-data" Oct 08 19:39:29 crc kubenswrapper[4988]: I1008 19:39:29.555451 4988 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Oct 08 19:39:29 crc kubenswrapper[4988]: I1008 19:39:29.555503 4988 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-81095d3c-890a-4460-a497-531b459f10ce\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-81095d3c-890a-4460-a497-531b459f10ce\") pod \"mariadb-copy-data\" (UID: \"9e312e3a-e0e2-490b-bb43-cc3b7191c109\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/8b0d18385f86e57330e0d14e030f86f560f31bf1519c68baa398f62e3e8af0b3/globalmount\"" pod="openstack/mariadb-copy-data" Oct 08 19:39:29 crc kubenswrapper[4988]: I1008 19:39:29.580752 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jw4zs\" (UniqueName: \"kubernetes.io/projected/9e312e3a-e0e2-490b-bb43-cc3b7191c109-kube-api-access-jw4zs\") pod \"mariadb-copy-data\" (UID: \"9e312e3a-e0e2-490b-bb43-cc3b7191c109\") " pod="openstack/mariadb-copy-data" Oct 08 19:39:29 crc kubenswrapper[4988]: I1008 19:39:29.598207 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-81095d3c-890a-4460-a497-531b459f10ce\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-81095d3c-890a-4460-a497-531b459f10ce\") pod \"mariadb-copy-data\" (UID: \"9e312e3a-e0e2-490b-bb43-cc3b7191c109\") " pod="openstack/mariadb-copy-data" Oct 08 19:39:29 crc kubenswrapper[4988]: I1008 19:39:29.674975 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-copy-data" Oct 08 19:39:30 crc kubenswrapper[4988]: I1008 19:39:30.077282 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-copy-data"] Oct 08 19:39:30 crc kubenswrapper[4988]: I1008 19:39:30.773616 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-copy-data" event={"ID":"9e312e3a-e0e2-490b-bb43-cc3b7191c109","Type":"ContainerStarted","Data":"9f72418b4bb8bbb8075246185311d8f58cdc5981b8ff6897ca7c1eb0dedfa988"} Oct 08 19:39:30 crc kubenswrapper[4988]: I1008 19:39:30.773659 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-copy-data" event={"ID":"9e312e3a-e0e2-490b-bb43-cc3b7191c109","Type":"ContainerStarted","Data":"2c41cfac11e6123f9fbbd7a1a71110c6bb83d0cde4a428538571ef8fe94ffa50"} Oct 08 19:39:30 crc kubenswrapper[4988]: I1008 19:39:30.798549 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/mariadb-copy-data" podStartSLOduration=2.798300081 podStartE2EDuration="2.798300081s" podCreationTimestamp="2025-10-08 19:39:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 19:39:30.795834603 +0000 UTC m=+5316.245677453" watchObservedRunningTime="2025-10-08 19:39:30.798300081 +0000 UTC m=+5316.248142871" Oct 08 19:39:31 crc kubenswrapper[4988]: E1008 19:39:31.477997 4988 upgradeaware.go:427] Error proxying data from client to backend: readfrom tcp 38.102.83.102:60204->38.102.83.102:38495: write tcp 38.102.83.102:60204->38.102.83.102:38495: write: connection reset by peer Oct 08 19:39:32 crc kubenswrapper[4988]: I1008 19:39:32.998436 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/mariadb-client"] Oct 08 19:39:33 crc kubenswrapper[4988]: I1008 19:39:33.003378 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client" Oct 08 19:39:33 crc kubenswrapper[4988]: I1008 19:39:33.011325 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client"] Oct 08 19:39:33 crc kubenswrapper[4988]: I1008 19:39:33.116374 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tlls5\" (UniqueName: \"kubernetes.io/projected/001372c3-4c1d-4cfa-9659-02dadde4018a-kube-api-access-tlls5\") pod \"mariadb-client\" (UID: \"001372c3-4c1d-4cfa-9659-02dadde4018a\") " pod="openstack/mariadb-client" Oct 08 19:39:33 crc kubenswrapper[4988]: I1008 19:39:33.218052 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tlls5\" (UniqueName: \"kubernetes.io/projected/001372c3-4c1d-4cfa-9659-02dadde4018a-kube-api-access-tlls5\") pod \"mariadb-client\" (UID: \"001372c3-4c1d-4cfa-9659-02dadde4018a\") " pod="openstack/mariadb-client" Oct 08 19:39:33 crc kubenswrapper[4988]: I1008 19:39:33.251445 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tlls5\" (UniqueName: \"kubernetes.io/projected/001372c3-4c1d-4cfa-9659-02dadde4018a-kube-api-access-tlls5\") pod \"mariadb-client\" (UID: \"001372c3-4c1d-4cfa-9659-02dadde4018a\") " pod="openstack/mariadb-client" Oct 08 19:39:33 crc kubenswrapper[4988]: I1008 19:39:33.336728 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client" Oct 08 19:39:33 crc kubenswrapper[4988]: I1008 19:39:33.639928 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client"] Oct 08 19:39:33 crc kubenswrapper[4988]: W1008 19:39:33.647613 4988 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod001372c3_4c1d_4cfa_9659_02dadde4018a.slice/crio-a5fe27c9acec6fbd0bdd79a5192e059f84a4f5f8bd135f0ad666e50368318b33 WatchSource:0}: Error finding container a5fe27c9acec6fbd0bdd79a5192e059f84a4f5f8bd135f0ad666e50368318b33: Status 404 returned error can't find the container with id a5fe27c9acec6fbd0bdd79a5192e059f84a4f5f8bd135f0ad666e50368318b33 Oct 08 19:39:33 crc kubenswrapper[4988]: I1008 19:39:33.828177 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client" event={"ID":"001372c3-4c1d-4cfa-9659-02dadde4018a","Type":"ContainerStarted","Data":"a5fe27c9acec6fbd0bdd79a5192e059f84a4f5f8bd135f0ad666e50368318b33"} Oct 08 19:39:34 crc kubenswrapper[4988]: I1008 19:39:34.843027 4988 generic.go:334] "Generic (PLEG): container finished" podID="001372c3-4c1d-4cfa-9659-02dadde4018a" containerID="d3f24e4ba303a1481781dd4658f349d18349f68f11a5f157594e62045a1fcf90" exitCode=0 Oct 08 19:39:34 crc kubenswrapper[4988]: I1008 19:39:34.843424 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client" event={"ID":"001372c3-4c1d-4cfa-9659-02dadde4018a","Type":"ContainerDied","Data":"d3f24e4ba303a1481781dd4658f349d18349f68f11a5f157594e62045a1fcf90"} Oct 08 19:39:36 crc kubenswrapper[4988]: I1008 19:39:36.205875 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client" Oct 08 19:39:36 crc kubenswrapper[4988]: I1008 19:39:36.232117 4988 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_mariadb-client_001372c3-4c1d-4cfa-9659-02dadde4018a/mariadb-client/0.log" Oct 08 19:39:36 crc kubenswrapper[4988]: I1008 19:39:36.263774 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mariadb-client"] Oct 08 19:39:36 crc kubenswrapper[4988]: I1008 19:39:36.266617 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tlls5\" (UniqueName: \"kubernetes.io/projected/001372c3-4c1d-4cfa-9659-02dadde4018a-kube-api-access-tlls5\") pod \"001372c3-4c1d-4cfa-9659-02dadde4018a\" (UID: \"001372c3-4c1d-4cfa-9659-02dadde4018a\") " Oct 08 19:39:36 crc kubenswrapper[4988]: I1008 19:39:36.272452 4988 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/mariadb-client"] Oct 08 19:39:36 crc kubenswrapper[4988]: I1008 19:39:36.273420 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/001372c3-4c1d-4cfa-9659-02dadde4018a-kube-api-access-tlls5" (OuterVolumeSpecName: "kube-api-access-tlls5") pod "001372c3-4c1d-4cfa-9659-02dadde4018a" (UID: "001372c3-4c1d-4cfa-9659-02dadde4018a"). InnerVolumeSpecName "kube-api-access-tlls5". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 19:39:36 crc kubenswrapper[4988]: I1008 19:39:36.369443 4988 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tlls5\" (UniqueName: \"kubernetes.io/projected/001372c3-4c1d-4cfa-9659-02dadde4018a-kube-api-access-tlls5\") on node \"crc\" DevicePath \"\"" Oct 08 19:39:36 crc kubenswrapper[4988]: I1008 19:39:36.451903 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/mariadb-client"] Oct 08 19:39:36 crc kubenswrapper[4988]: E1008 19:39:36.452604 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="001372c3-4c1d-4cfa-9659-02dadde4018a" containerName="mariadb-client" Oct 08 19:39:36 crc kubenswrapper[4988]: I1008 19:39:36.452637 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="001372c3-4c1d-4cfa-9659-02dadde4018a" containerName="mariadb-client" Oct 08 19:39:36 crc kubenswrapper[4988]: I1008 19:39:36.453026 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="001372c3-4c1d-4cfa-9659-02dadde4018a" containerName="mariadb-client" Oct 08 19:39:36 crc kubenswrapper[4988]: I1008 19:39:36.454252 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client" Oct 08 19:39:36 crc kubenswrapper[4988]: I1008 19:39:36.465897 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client"] Oct 08 19:39:36 crc kubenswrapper[4988]: I1008 19:39:36.572658 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bnt2f\" (UniqueName: \"kubernetes.io/projected/6009df01-495c-4da1-82dc-e69c7875b5c1-kube-api-access-bnt2f\") pod \"mariadb-client\" (UID: \"6009df01-495c-4da1-82dc-e69c7875b5c1\") " pod="openstack/mariadb-client" Oct 08 19:39:36 crc kubenswrapper[4988]: I1008 19:39:36.674774 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bnt2f\" (UniqueName: \"kubernetes.io/projected/6009df01-495c-4da1-82dc-e69c7875b5c1-kube-api-access-bnt2f\") pod \"mariadb-client\" (UID: \"6009df01-495c-4da1-82dc-e69c7875b5c1\") " pod="openstack/mariadb-client" Oct 08 19:39:36 crc kubenswrapper[4988]: I1008 19:39:36.705369 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bnt2f\" (UniqueName: \"kubernetes.io/projected/6009df01-495c-4da1-82dc-e69c7875b5c1-kube-api-access-bnt2f\") pod \"mariadb-client\" (UID: \"6009df01-495c-4da1-82dc-e69c7875b5c1\") " pod="openstack/mariadb-client" Oct 08 19:39:36 crc kubenswrapper[4988]: I1008 19:39:36.786166 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client" Oct 08 19:39:36 crc kubenswrapper[4988]: I1008 19:39:36.863198 4988 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a5fe27c9acec6fbd0bdd79a5192e059f84a4f5f8bd135f0ad666e50368318b33" Oct 08 19:39:36 crc kubenswrapper[4988]: I1008 19:39:36.863299 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client" Oct 08 19:39:36 crc kubenswrapper[4988]: I1008 19:39:36.886413 4988 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openstack/mariadb-client" oldPodUID="001372c3-4c1d-4cfa-9659-02dadde4018a" podUID="6009df01-495c-4da1-82dc-e69c7875b5c1" Oct 08 19:39:37 crc kubenswrapper[4988]: I1008 19:39:37.010714 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client"] Oct 08 19:39:37 crc kubenswrapper[4988]: W1008 19:39:37.021373 4988 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6009df01_495c_4da1_82dc_e69c7875b5c1.slice/crio-d9548dd8d5214575d884710d2d84074dc8f857b2db5b66b848485218060ef49b WatchSource:0}: Error finding container d9548dd8d5214575d884710d2d84074dc8f857b2db5b66b848485218060ef49b: Status 404 returned error can't find the container with id d9548dd8d5214575d884710d2d84074dc8f857b2db5b66b848485218060ef49b Oct 08 19:39:37 crc kubenswrapper[4988]: I1008 19:39:37.238321 4988 scope.go:117] "RemoveContainer" containerID="3cd52d2ff674f037f4b111ff08caa426d4d7977052a05e385f05362baabb8da9" Oct 08 19:39:37 crc kubenswrapper[4988]: E1008 19:39:37.238711 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nm54f_openshift-machine-config-operator(f328a20b-5283-407e-8a93-2ab2da599f18)\"" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" Oct 08 19:39:37 crc kubenswrapper[4988]: I1008 19:39:37.265176 4988 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="001372c3-4c1d-4cfa-9659-02dadde4018a" path="/var/lib/kubelet/pods/001372c3-4c1d-4cfa-9659-02dadde4018a/volumes" Oct 08 19:39:37 crc kubenswrapper[4988]: I1008 19:39:37.881002 4988 generic.go:334] "Generic (PLEG): container finished" podID="6009df01-495c-4da1-82dc-e69c7875b5c1" containerID="0a1aef9772ec4c79f7c96f4d80291f1758dec92998f6d480f86c70553420632e" exitCode=0 Oct 08 19:39:37 crc kubenswrapper[4988]: I1008 19:39:37.881096 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client" event={"ID":"6009df01-495c-4da1-82dc-e69c7875b5c1","Type":"ContainerDied","Data":"0a1aef9772ec4c79f7c96f4d80291f1758dec92998f6d480f86c70553420632e"} Oct 08 19:39:37 crc kubenswrapper[4988]: I1008 19:39:37.881423 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client" event={"ID":"6009df01-495c-4da1-82dc-e69c7875b5c1","Type":"ContainerStarted","Data":"d9548dd8d5214575d884710d2d84074dc8f857b2db5b66b848485218060ef49b"} Oct 08 19:39:39 crc kubenswrapper[4988]: I1008 19:39:39.247647 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client" Oct 08 19:39:39 crc kubenswrapper[4988]: I1008 19:39:39.271719 4988 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_mariadb-client_6009df01-495c-4da1-82dc-e69c7875b5c1/mariadb-client/0.log" Oct 08 19:39:39 crc kubenswrapper[4988]: I1008 19:39:39.307430 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mariadb-client"] Oct 08 19:39:39 crc kubenswrapper[4988]: I1008 19:39:39.315474 4988 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/mariadb-client"] Oct 08 19:39:39 crc kubenswrapper[4988]: I1008 19:39:39.328877 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bnt2f\" (UniqueName: \"kubernetes.io/projected/6009df01-495c-4da1-82dc-e69c7875b5c1-kube-api-access-bnt2f\") pod \"6009df01-495c-4da1-82dc-e69c7875b5c1\" (UID: \"6009df01-495c-4da1-82dc-e69c7875b5c1\") " Oct 08 19:39:39 crc kubenswrapper[4988]: I1008 19:39:39.337913 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6009df01-495c-4da1-82dc-e69c7875b5c1-kube-api-access-bnt2f" (OuterVolumeSpecName: "kube-api-access-bnt2f") pod "6009df01-495c-4da1-82dc-e69c7875b5c1" (UID: "6009df01-495c-4da1-82dc-e69c7875b5c1"). InnerVolumeSpecName "kube-api-access-bnt2f". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 19:39:39 crc kubenswrapper[4988]: I1008 19:39:39.432762 4988 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bnt2f\" (UniqueName: \"kubernetes.io/projected/6009df01-495c-4da1-82dc-e69c7875b5c1-kube-api-access-bnt2f\") on node \"crc\" DevicePath \"\"" Oct 08 19:39:39 crc kubenswrapper[4988]: I1008 19:39:39.905589 4988 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d9548dd8d5214575d884710d2d84074dc8f857b2db5b66b848485218060ef49b" Oct 08 19:39:39 crc kubenswrapper[4988]: I1008 19:39:39.905629 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client" Oct 08 19:39:41 crc kubenswrapper[4988]: I1008 19:39:41.254179 4988 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6009df01-495c-4da1-82dc-e69c7875b5c1" path="/var/lib/kubelet/pods/6009df01-495c-4da1-82dc-e69c7875b5c1/volumes" Oct 08 19:39:50 crc kubenswrapper[4988]: I1008 19:39:50.238572 4988 scope.go:117] "RemoveContainer" containerID="3cd52d2ff674f037f4b111ff08caa426d4d7977052a05e385f05362baabb8da9" Oct 08 19:39:50 crc kubenswrapper[4988]: E1008 19:39:50.239408 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nm54f_openshift-machine-config-operator(f328a20b-5283-407e-8a93-2ab2da599f18)\"" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" Oct 08 19:40:04 crc kubenswrapper[4988]: I1008 19:40:04.237962 4988 scope.go:117] "RemoveContainer" containerID="3cd52d2ff674f037f4b111ff08caa426d4d7977052a05e385f05362baabb8da9" Oct 08 19:40:05 crc kubenswrapper[4988]: I1008 19:40:05.159405 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" event={"ID":"f328a20b-5283-407e-8a93-2ab2da599f18","Type":"ContainerStarted","Data":"bfa76a99ab92f7c2c936649e27734f81c2af70cdaf33874dac1d856a4b92f5f7"} Oct 08 19:40:12 crc kubenswrapper[4988]: I1008 19:40:12.647102 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-nb-0"] Oct 08 19:40:12 crc kubenswrapper[4988]: E1008 19:40:12.647951 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6009df01-495c-4da1-82dc-e69c7875b5c1" containerName="mariadb-client" Oct 08 19:40:12 crc kubenswrapper[4988]: I1008 19:40:12.647967 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="6009df01-495c-4da1-82dc-e69c7875b5c1" containerName="mariadb-client" Oct 08 19:40:12 crc kubenswrapper[4988]: I1008 19:40:12.648163 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="6009df01-495c-4da1-82dc-e69c7875b5c1" containerName="mariadb-client" Oct 08 19:40:12 crc kubenswrapper[4988]: I1008 19:40:12.649133 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Oct 08 19:40:12 crc kubenswrapper[4988]: I1008 19:40:12.652969 4988 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-scripts" Oct 08 19:40:12 crc kubenswrapper[4988]: I1008 19:40:12.653061 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-nb-dockercfg-m4hmx" Oct 08 19:40:12 crc kubenswrapper[4988]: I1008 19:40:12.653097 4988 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-config" Oct 08 19:40:12 crc kubenswrapper[4988]: I1008 19:40:12.653236 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovn-metrics" Oct 08 19:40:12 crc kubenswrapper[4988]: I1008 19:40:12.653611 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovndbcluster-nb-ovndbs" Oct 08 19:40:12 crc kubenswrapper[4988]: I1008 19:40:12.670448 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Oct 08 19:40:12 crc kubenswrapper[4988]: I1008 19:40:12.693807 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-nb-2"] Oct 08 19:40:12 crc kubenswrapper[4988]: I1008 19:40:12.704494 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-2" Oct 08 19:40:12 crc kubenswrapper[4988]: I1008 19:40:12.707449 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-nb-1"] Oct 08 19:40:12 crc kubenswrapper[4988]: I1008 19:40:12.715227 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-1" Oct 08 19:40:12 crc kubenswrapper[4988]: I1008 19:40:12.744024 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-1"] Oct 08 19:40:12 crc kubenswrapper[4988]: I1008 19:40:12.751241 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/e00c8be6-0ec3-4045-9a25-9f397cc78dc9-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"e00c8be6-0ec3-4045-9a25-9f397cc78dc9\") " pod="openstack/ovsdbserver-nb-0" Oct 08 19:40:12 crc kubenswrapper[4988]: I1008 19:40:12.751314 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/e00c8be6-0ec3-4045-9a25-9f397cc78dc9-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"e00c8be6-0ec3-4045-9a25-9f397cc78dc9\") " pod="openstack/ovsdbserver-nb-0" Oct 08 19:40:12 crc kubenswrapper[4988]: I1008 19:40:12.751965 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e00c8be6-0ec3-4045-9a25-9f397cc78dc9-config\") pod \"ovsdbserver-nb-0\" (UID: \"e00c8be6-0ec3-4045-9a25-9f397cc78dc9\") " pod="openstack/ovsdbserver-nb-0" Oct 08 19:40:12 crc kubenswrapper[4988]: I1008 19:40:12.752042 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-530540f4-d9d8-4485-99f8-d43f11861f63\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-530540f4-d9d8-4485-99f8-d43f11861f63\") pod \"ovsdbserver-nb-0\" (UID: \"e00c8be6-0ec3-4045-9a25-9f397cc78dc9\") " pod="openstack/ovsdbserver-nb-0" Oct 08 19:40:12 crc kubenswrapper[4988]: I1008 19:40:12.752072 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9k8p4\" (UniqueName: \"kubernetes.io/projected/e00c8be6-0ec3-4045-9a25-9f397cc78dc9-kube-api-access-9k8p4\") pod \"ovsdbserver-nb-0\" (UID: \"e00c8be6-0ec3-4045-9a25-9f397cc78dc9\") " pod="openstack/ovsdbserver-nb-0" Oct 08 19:40:12 crc kubenswrapper[4988]: I1008 19:40:12.752092 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/e00c8be6-0ec3-4045-9a25-9f397cc78dc9-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"e00c8be6-0ec3-4045-9a25-9f397cc78dc9\") " pod="openstack/ovsdbserver-nb-0" Oct 08 19:40:12 crc kubenswrapper[4988]: I1008 19:40:12.752119 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/e00c8be6-0ec3-4045-9a25-9f397cc78dc9-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"e00c8be6-0ec3-4045-9a25-9f397cc78dc9\") " pod="openstack/ovsdbserver-nb-0" Oct 08 19:40:12 crc kubenswrapper[4988]: I1008 19:40:12.752199 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e00c8be6-0ec3-4045-9a25-9f397cc78dc9-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"e00c8be6-0ec3-4045-9a25-9f397cc78dc9\") " pod="openstack/ovsdbserver-nb-0" Oct 08 19:40:12 crc kubenswrapper[4988]: I1008 19:40:12.760358 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-2"] Oct 08 19:40:12 crc kubenswrapper[4988]: I1008 19:40:12.854005 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-530540f4-d9d8-4485-99f8-d43f11861f63\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-530540f4-d9d8-4485-99f8-d43f11861f63\") pod \"ovsdbserver-nb-0\" (UID: \"e00c8be6-0ec3-4045-9a25-9f397cc78dc9\") " pod="openstack/ovsdbserver-nb-0" Oct 08 19:40:12 crc kubenswrapper[4988]: I1008 19:40:12.855961 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9k8p4\" (UniqueName: \"kubernetes.io/projected/e00c8be6-0ec3-4045-9a25-9f397cc78dc9-kube-api-access-9k8p4\") pod \"ovsdbserver-nb-0\" (UID: \"e00c8be6-0ec3-4045-9a25-9f397cc78dc9\") " pod="openstack/ovsdbserver-nb-0" Oct 08 19:40:12 crc kubenswrapper[4988]: I1008 19:40:12.856198 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-4ada9da7-06d8-48a6-9767-336007c53ec8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-4ada9da7-06d8-48a6-9767-336007c53ec8\") pod \"ovsdbserver-nb-1\" (UID: \"cbef3512-4ad7-45f0-9e87-d01ea0177708\") " pod="openstack/ovsdbserver-nb-1" Oct 08 19:40:12 crc kubenswrapper[4988]: I1008 19:40:12.856411 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/e00c8be6-0ec3-4045-9a25-9f397cc78dc9-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"e00c8be6-0ec3-4045-9a25-9f397cc78dc9\") " pod="openstack/ovsdbserver-nb-0" Oct 08 19:40:12 crc kubenswrapper[4988]: I1008 19:40:12.856796 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/e00c8be6-0ec3-4045-9a25-9f397cc78dc9-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"e00c8be6-0ec3-4045-9a25-9f397cc78dc9\") " pod="openstack/ovsdbserver-nb-0" Oct 08 19:40:12 crc kubenswrapper[4988]: I1008 19:40:12.856956 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/40f9f201-93e1-4d27-842f-77040d48d28d-scripts\") pod \"ovsdbserver-nb-2\" (UID: \"40f9f201-93e1-4d27-842f-77040d48d28d\") " pod="openstack/ovsdbserver-nb-2" Oct 08 19:40:12 crc kubenswrapper[4988]: I1008 19:40:12.857122 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/cbef3512-4ad7-45f0-9e87-d01ea0177708-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-1\" (UID: \"cbef3512-4ad7-45f0-9e87-d01ea0177708\") " pod="openstack/ovsdbserver-nb-1" Oct 08 19:40:12 crc kubenswrapper[4988]: I1008 19:40:12.857292 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/cbef3512-4ad7-45f0-9e87-d01ea0177708-scripts\") pod \"ovsdbserver-nb-1\" (UID: \"cbef3512-4ad7-45f0-9e87-d01ea0177708\") " pod="openstack/ovsdbserver-nb-1" Oct 08 19:40:12 crc kubenswrapper[4988]: I1008 19:40:12.857499 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/cbef3512-4ad7-45f0-9e87-d01ea0177708-ovsdb-rundir\") pod \"ovsdbserver-nb-1\" (UID: \"cbef3512-4ad7-45f0-9e87-d01ea0177708\") " pod="openstack/ovsdbserver-nb-1" Oct 08 19:40:12 crc kubenswrapper[4988]: I1008 19:40:12.857665 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cbef3512-4ad7-45f0-9e87-d01ea0177708-config\") pod \"ovsdbserver-nb-1\" (UID: \"cbef3512-4ad7-45f0-9e87-d01ea0177708\") " pod="openstack/ovsdbserver-nb-1" Oct 08 19:40:12 crc kubenswrapper[4988]: I1008 19:40:12.857854 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e00c8be6-0ec3-4045-9a25-9f397cc78dc9-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"e00c8be6-0ec3-4045-9a25-9f397cc78dc9\") " pod="openstack/ovsdbserver-nb-0" Oct 08 19:40:12 crc kubenswrapper[4988]: I1008 19:40:12.858008 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cbef3512-4ad7-45f0-9e87-d01ea0177708-combined-ca-bundle\") pod \"ovsdbserver-nb-1\" (UID: \"cbef3512-4ad7-45f0-9e87-d01ea0177708\") " pod="openstack/ovsdbserver-nb-1" Oct 08 19:40:12 crc kubenswrapper[4988]: I1008 19:40:12.858124 4988 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Oct 08 19:40:12 crc kubenswrapper[4988]: I1008 19:40:12.858175 4988 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-530540f4-d9d8-4485-99f8-d43f11861f63\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-530540f4-d9d8-4485-99f8-d43f11861f63\") pod \"ovsdbserver-nb-0\" (UID: \"e00c8be6-0ec3-4045-9a25-9f397cc78dc9\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/ae79c665820bf39d78a5471543af3ea64aea4c1ff62941e353f5d0e7f26f2049/globalmount\"" pod="openstack/ovsdbserver-nb-0" Oct 08 19:40:12 crc kubenswrapper[4988]: I1008 19:40:12.858008 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/e00c8be6-0ec3-4045-9a25-9f397cc78dc9-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"e00c8be6-0ec3-4045-9a25-9f397cc78dc9\") " pod="openstack/ovsdbserver-nb-0" Oct 08 19:40:12 crc kubenswrapper[4988]: I1008 19:40:12.858510 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/40f9f201-93e1-4d27-842f-77040d48d28d-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-2\" (UID: \"40f9f201-93e1-4d27-842f-77040d48d28d\") " pod="openstack/ovsdbserver-nb-2" Oct 08 19:40:12 crc kubenswrapper[4988]: I1008 19:40:12.858620 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/40f9f201-93e1-4d27-842f-77040d48d28d-ovsdb-rundir\") pod \"ovsdbserver-nb-2\" (UID: \"40f9f201-93e1-4d27-842f-77040d48d28d\") " pod="openstack/ovsdbserver-nb-2" Oct 08 19:40:12 crc kubenswrapper[4988]: I1008 19:40:12.858723 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/40f9f201-93e1-4d27-842f-77040d48d28d-combined-ca-bundle\") pod \"ovsdbserver-nb-2\" (UID: \"40f9f201-93e1-4d27-842f-77040d48d28d\") " pod="openstack/ovsdbserver-nb-2" Oct 08 19:40:12 crc kubenswrapper[4988]: I1008 19:40:12.858830 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-29e2f68f-2339-4acf-a9b4-8fd64cfe7ffc\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-29e2f68f-2339-4acf-a9b4-8fd64cfe7ffc\") pod \"ovsdbserver-nb-2\" (UID: \"40f9f201-93e1-4d27-842f-77040d48d28d\") " pod="openstack/ovsdbserver-nb-2" Oct 08 19:40:12 crc kubenswrapper[4988]: I1008 19:40:12.858947 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/cbef3512-4ad7-45f0-9e87-d01ea0177708-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-1\" (UID: \"cbef3512-4ad7-45f0-9e87-d01ea0177708\") " pod="openstack/ovsdbserver-nb-1" Oct 08 19:40:12 crc kubenswrapper[4988]: I1008 19:40:12.859049 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kncn5\" (UniqueName: \"kubernetes.io/projected/40f9f201-93e1-4d27-842f-77040d48d28d-kube-api-access-kncn5\") pod \"ovsdbserver-nb-2\" (UID: \"40f9f201-93e1-4d27-842f-77040d48d28d\") " pod="openstack/ovsdbserver-nb-2" Oct 08 19:40:12 crc kubenswrapper[4988]: I1008 19:40:12.859156 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/e00c8be6-0ec3-4045-9a25-9f397cc78dc9-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"e00c8be6-0ec3-4045-9a25-9f397cc78dc9\") " pod="openstack/ovsdbserver-nb-0" Oct 08 19:40:12 crc kubenswrapper[4988]: I1008 19:40:12.859253 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/e00c8be6-0ec3-4045-9a25-9f397cc78dc9-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"e00c8be6-0ec3-4045-9a25-9f397cc78dc9\") " pod="openstack/ovsdbserver-nb-0" Oct 08 19:40:12 crc kubenswrapper[4988]: I1008 19:40:12.859349 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9b4d6\" (UniqueName: \"kubernetes.io/projected/cbef3512-4ad7-45f0-9e87-d01ea0177708-kube-api-access-9b4d6\") pod \"ovsdbserver-nb-1\" (UID: \"cbef3512-4ad7-45f0-9e87-d01ea0177708\") " pod="openstack/ovsdbserver-nb-1" Oct 08 19:40:12 crc kubenswrapper[4988]: I1008 19:40:12.859502 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e00c8be6-0ec3-4045-9a25-9f397cc78dc9-config\") pod \"ovsdbserver-nb-0\" (UID: \"e00c8be6-0ec3-4045-9a25-9f397cc78dc9\") " pod="openstack/ovsdbserver-nb-0" Oct 08 19:40:12 crc kubenswrapper[4988]: I1008 19:40:12.859646 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/40f9f201-93e1-4d27-842f-77040d48d28d-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-2\" (UID: \"40f9f201-93e1-4d27-842f-77040d48d28d\") " pod="openstack/ovsdbserver-nb-2" Oct 08 19:40:12 crc kubenswrapper[4988]: I1008 19:40:12.859883 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/40f9f201-93e1-4d27-842f-77040d48d28d-config\") pod \"ovsdbserver-nb-2\" (UID: \"40f9f201-93e1-4d27-842f-77040d48d28d\") " pod="openstack/ovsdbserver-nb-2" Oct 08 19:40:12 crc kubenswrapper[4988]: I1008 19:40:12.860218 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/e00c8be6-0ec3-4045-9a25-9f397cc78dc9-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"e00c8be6-0ec3-4045-9a25-9f397cc78dc9\") " pod="openstack/ovsdbserver-nb-0" Oct 08 19:40:12 crc kubenswrapper[4988]: I1008 19:40:12.861327 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e00c8be6-0ec3-4045-9a25-9f397cc78dc9-config\") pod \"ovsdbserver-nb-0\" (UID: \"e00c8be6-0ec3-4045-9a25-9f397cc78dc9\") " pod="openstack/ovsdbserver-nb-0" Oct 08 19:40:12 crc kubenswrapper[4988]: I1008 19:40:12.868733 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e00c8be6-0ec3-4045-9a25-9f397cc78dc9-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"e00c8be6-0ec3-4045-9a25-9f397cc78dc9\") " pod="openstack/ovsdbserver-nb-0" Oct 08 19:40:12 crc kubenswrapper[4988]: I1008 19:40:12.868912 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/e00c8be6-0ec3-4045-9a25-9f397cc78dc9-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"e00c8be6-0ec3-4045-9a25-9f397cc78dc9\") " pod="openstack/ovsdbserver-nb-0" Oct 08 19:40:12 crc kubenswrapper[4988]: I1008 19:40:12.868986 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/e00c8be6-0ec3-4045-9a25-9f397cc78dc9-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"e00c8be6-0ec3-4045-9a25-9f397cc78dc9\") " pod="openstack/ovsdbserver-nb-0" Oct 08 19:40:12 crc kubenswrapper[4988]: I1008 19:40:12.880467 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9k8p4\" (UniqueName: \"kubernetes.io/projected/e00c8be6-0ec3-4045-9a25-9f397cc78dc9-kube-api-access-9k8p4\") pod \"ovsdbserver-nb-0\" (UID: \"e00c8be6-0ec3-4045-9a25-9f397cc78dc9\") " pod="openstack/ovsdbserver-nb-0" Oct 08 19:40:12 crc kubenswrapper[4988]: I1008 19:40:12.895941 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-530540f4-d9d8-4485-99f8-d43f11861f63\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-530540f4-d9d8-4485-99f8-d43f11861f63\") pod \"ovsdbserver-nb-0\" (UID: \"e00c8be6-0ec3-4045-9a25-9f397cc78dc9\") " pod="openstack/ovsdbserver-nb-0" Oct 08 19:40:12 crc kubenswrapper[4988]: I1008 19:40:12.961338 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/40f9f201-93e1-4d27-842f-77040d48d28d-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-2\" (UID: \"40f9f201-93e1-4d27-842f-77040d48d28d\") " pod="openstack/ovsdbserver-nb-2" Oct 08 19:40:12 crc kubenswrapper[4988]: I1008 19:40:12.961650 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/40f9f201-93e1-4d27-842f-77040d48d28d-ovsdb-rundir\") pod \"ovsdbserver-nb-2\" (UID: \"40f9f201-93e1-4d27-842f-77040d48d28d\") " pod="openstack/ovsdbserver-nb-2" Oct 08 19:40:12 crc kubenswrapper[4988]: I1008 19:40:12.961669 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/40f9f201-93e1-4d27-842f-77040d48d28d-combined-ca-bundle\") pod \"ovsdbserver-nb-2\" (UID: \"40f9f201-93e1-4d27-842f-77040d48d28d\") " pod="openstack/ovsdbserver-nb-2" Oct 08 19:40:12 crc kubenswrapper[4988]: I1008 19:40:12.961690 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-29e2f68f-2339-4acf-a9b4-8fd64cfe7ffc\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-29e2f68f-2339-4acf-a9b4-8fd64cfe7ffc\") pod \"ovsdbserver-nb-2\" (UID: \"40f9f201-93e1-4d27-842f-77040d48d28d\") " pod="openstack/ovsdbserver-nb-2" Oct 08 19:40:12 crc kubenswrapper[4988]: I1008 19:40:12.961715 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/cbef3512-4ad7-45f0-9e87-d01ea0177708-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-1\" (UID: \"cbef3512-4ad7-45f0-9e87-d01ea0177708\") " pod="openstack/ovsdbserver-nb-1" Oct 08 19:40:12 crc kubenswrapper[4988]: I1008 19:40:12.961729 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kncn5\" (UniqueName: \"kubernetes.io/projected/40f9f201-93e1-4d27-842f-77040d48d28d-kube-api-access-kncn5\") pod \"ovsdbserver-nb-2\" (UID: \"40f9f201-93e1-4d27-842f-77040d48d28d\") " pod="openstack/ovsdbserver-nb-2" Oct 08 19:40:12 crc kubenswrapper[4988]: I1008 19:40:12.961744 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9b4d6\" (UniqueName: \"kubernetes.io/projected/cbef3512-4ad7-45f0-9e87-d01ea0177708-kube-api-access-9b4d6\") pod \"ovsdbserver-nb-1\" (UID: \"cbef3512-4ad7-45f0-9e87-d01ea0177708\") " pod="openstack/ovsdbserver-nb-1" Oct 08 19:40:12 crc kubenswrapper[4988]: I1008 19:40:12.961774 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/40f9f201-93e1-4d27-842f-77040d48d28d-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-2\" (UID: \"40f9f201-93e1-4d27-842f-77040d48d28d\") " pod="openstack/ovsdbserver-nb-2" Oct 08 19:40:12 crc kubenswrapper[4988]: I1008 19:40:12.961795 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/40f9f201-93e1-4d27-842f-77040d48d28d-config\") pod \"ovsdbserver-nb-2\" (UID: \"40f9f201-93e1-4d27-842f-77040d48d28d\") " pod="openstack/ovsdbserver-nb-2" Oct 08 19:40:12 crc kubenswrapper[4988]: I1008 19:40:12.961820 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-4ada9da7-06d8-48a6-9767-336007c53ec8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-4ada9da7-06d8-48a6-9767-336007c53ec8\") pod \"ovsdbserver-nb-1\" (UID: \"cbef3512-4ad7-45f0-9e87-d01ea0177708\") " pod="openstack/ovsdbserver-nb-1" Oct 08 19:40:12 crc kubenswrapper[4988]: I1008 19:40:12.961847 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/40f9f201-93e1-4d27-842f-77040d48d28d-scripts\") pod \"ovsdbserver-nb-2\" (UID: \"40f9f201-93e1-4d27-842f-77040d48d28d\") " pod="openstack/ovsdbserver-nb-2" Oct 08 19:40:12 crc kubenswrapper[4988]: I1008 19:40:12.961869 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/cbef3512-4ad7-45f0-9e87-d01ea0177708-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-1\" (UID: \"cbef3512-4ad7-45f0-9e87-d01ea0177708\") " pod="openstack/ovsdbserver-nb-1" Oct 08 19:40:12 crc kubenswrapper[4988]: I1008 19:40:12.961887 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/cbef3512-4ad7-45f0-9e87-d01ea0177708-scripts\") pod \"ovsdbserver-nb-1\" (UID: \"cbef3512-4ad7-45f0-9e87-d01ea0177708\") " pod="openstack/ovsdbserver-nb-1" Oct 08 19:40:12 crc kubenswrapper[4988]: I1008 19:40:12.961906 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/cbef3512-4ad7-45f0-9e87-d01ea0177708-ovsdb-rundir\") pod \"ovsdbserver-nb-1\" (UID: \"cbef3512-4ad7-45f0-9e87-d01ea0177708\") " pod="openstack/ovsdbserver-nb-1" Oct 08 19:40:12 crc kubenswrapper[4988]: I1008 19:40:12.961928 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cbef3512-4ad7-45f0-9e87-d01ea0177708-config\") pod \"ovsdbserver-nb-1\" (UID: \"cbef3512-4ad7-45f0-9e87-d01ea0177708\") " pod="openstack/ovsdbserver-nb-1" Oct 08 19:40:12 crc kubenswrapper[4988]: I1008 19:40:12.961947 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cbef3512-4ad7-45f0-9e87-d01ea0177708-combined-ca-bundle\") pod \"ovsdbserver-nb-1\" (UID: \"cbef3512-4ad7-45f0-9e87-d01ea0177708\") " pod="openstack/ovsdbserver-nb-1" Oct 08 19:40:12 crc kubenswrapper[4988]: I1008 19:40:12.963359 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/40f9f201-93e1-4d27-842f-77040d48d28d-ovsdb-rundir\") pod \"ovsdbserver-nb-2\" (UID: \"40f9f201-93e1-4d27-842f-77040d48d28d\") " pod="openstack/ovsdbserver-nb-2" Oct 08 19:40:12 crc kubenswrapper[4988]: I1008 19:40:12.963758 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cbef3512-4ad7-45f0-9e87-d01ea0177708-config\") pod \"ovsdbserver-nb-1\" (UID: \"cbef3512-4ad7-45f0-9e87-d01ea0177708\") " pod="openstack/ovsdbserver-nb-1" Oct 08 19:40:12 crc kubenswrapper[4988]: I1008 19:40:12.963369 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/cbef3512-4ad7-45f0-9e87-d01ea0177708-ovsdb-rundir\") pod \"ovsdbserver-nb-1\" (UID: \"cbef3512-4ad7-45f0-9e87-d01ea0177708\") " pod="openstack/ovsdbserver-nb-1" Oct 08 19:40:12 crc kubenswrapper[4988]: I1008 19:40:12.964635 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/40f9f201-93e1-4d27-842f-77040d48d28d-scripts\") pod \"ovsdbserver-nb-2\" (UID: \"40f9f201-93e1-4d27-842f-77040d48d28d\") " pod="openstack/ovsdbserver-nb-2" Oct 08 19:40:12 crc kubenswrapper[4988]: I1008 19:40:12.964917 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/40f9f201-93e1-4d27-842f-77040d48d28d-config\") pod \"ovsdbserver-nb-2\" (UID: \"40f9f201-93e1-4d27-842f-77040d48d28d\") " pod="openstack/ovsdbserver-nb-2" Oct 08 19:40:12 crc kubenswrapper[4988]: I1008 19:40:12.965520 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/40f9f201-93e1-4d27-842f-77040d48d28d-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-2\" (UID: \"40f9f201-93e1-4d27-842f-77040d48d28d\") " pod="openstack/ovsdbserver-nb-2" Oct 08 19:40:12 crc kubenswrapper[4988]: I1008 19:40:12.966525 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/40f9f201-93e1-4d27-842f-77040d48d28d-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-2\" (UID: \"40f9f201-93e1-4d27-842f-77040d48d28d\") " pod="openstack/ovsdbserver-nb-2" Oct 08 19:40:12 crc kubenswrapper[4988]: I1008 19:40:12.966595 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/cbef3512-4ad7-45f0-9e87-d01ea0177708-scripts\") pod \"ovsdbserver-nb-1\" (UID: \"cbef3512-4ad7-45f0-9e87-d01ea0177708\") " pod="openstack/ovsdbserver-nb-1" Oct 08 19:40:12 crc kubenswrapper[4988]: I1008 19:40:12.967781 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/cbef3512-4ad7-45f0-9e87-d01ea0177708-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-1\" (UID: \"cbef3512-4ad7-45f0-9e87-d01ea0177708\") " pod="openstack/ovsdbserver-nb-1" Oct 08 19:40:12 crc kubenswrapper[4988]: I1008 19:40:12.968002 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/cbef3512-4ad7-45f0-9e87-d01ea0177708-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-1\" (UID: \"cbef3512-4ad7-45f0-9e87-d01ea0177708\") " pod="openstack/ovsdbserver-nb-1" Oct 08 19:40:12 crc kubenswrapper[4988]: I1008 19:40:12.968158 4988 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Oct 08 19:40:12 crc kubenswrapper[4988]: I1008 19:40:12.968179 4988 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-29e2f68f-2339-4acf-a9b4-8fd64cfe7ffc\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-29e2f68f-2339-4acf-a9b4-8fd64cfe7ffc\") pod \"ovsdbserver-nb-2\" (UID: \"40f9f201-93e1-4d27-842f-77040d48d28d\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/f9d22fdbf8b2a34e33f53a6770f991b26c14589f90eeef4264e75fec67fa2f8e/globalmount\"" pod="openstack/ovsdbserver-nb-2" Oct 08 19:40:12 crc kubenswrapper[4988]: I1008 19:40:12.968904 4988 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Oct 08 19:40:12 crc kubenswrapper[4988]: I1008 19:40:12.969083 4988 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-4ada9da7-06d8-48a6-9767-336007c53ec8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-4ada9da7-06d8-48a6-9767-336007c53ec8\") pod \"ovsdbserver-nb-1\" (UID: \"cbef3512-4ad7-45f0-9e87-d01ea0177708\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/944bd5f18b7adb9820b7f3b549c554ef1e4d447799377677054e515c8d3fe1eb/globalmount\"" pod="openstack/ovsdbserver-nb-1" Oct 08 19:40:12 crc kubenswrapper[4988]: I1008 19:40:12.969323 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/40f9f201-93e1-4d27-842f-77040d48d28d-combined-ca-bundle\") pod \"ovsdbserver-nb-2\" (UID: \"40f9f201-93e1-4d27-842f-77040d48d28d\") " pod="openstack/ovsdbserver-nb-2" Oct 08 19:40:12 crc kubenswrapper[4988]: I1008 19:40:12.972143 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cbef3512-4ad7-45f0-9e87-d01ea0177708-combined-ca-bundle\") pod \"ovsdbserver-nb-1\" (UID: \"cbef3512-4ad7-45f0-9e87-d01ea0177708\") " pod="openstack/ovsdbserver-nb-1" Oct 08 19:40:12 crc kubenswrapper[4988]: I1008 19:40:12.978931 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Oct 08 19:40:12 crc kubenswrapper[4988]: I1008 19:40:12.982321 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kncn5\" (UniqueName: \"kubernetes.io/projected/40f9f201-93e1-4d27-842f-77040d48d28d-kube-api-access-kncn5\") pod \"ovsdbserver-nb-2\" (UID: \"40f9f201-93e1-4d27-842f-77040d48d28d\") " pod="openstack/ovsdbserver-nb-2" Oct 08 19:40:12 crc kubenswrapper[4988]: I1008 19:40:12.985719 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9b4d6\" (UniqueName: \"kubernetes.io/projected/cbef3512-4ad7-45f0-9e87-d01ea0177708-kube-api-access-9b4d6\") pod \"ovsdbserver-nb-1\" (UID: \"cbef3512-4ad7-45f0-9e87-d01ea0177708\") " pod="openstack/ovsdbserver-nb-1" Oct 08 19:40:12 crc kubenswrapper[4988]: I1008 19:40:12.999648 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-4ada9da7-06d8-48a6-9767-336007c53ec8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-4ada9da7-06d8-48a6-9767-336007c53ec8\") pod \"ovsdbserver-nb-1\" (UID: \"cbef3512-4ad7-45f0-9e87-d01ea0177708\") " pod="openstack/ovsdbserver-nb-1" Oct 08 19:40:13 crc kubenswrapper[4988]: I1008 19:40:13.013051 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-29e2f68f-2339-4acf-a9b4-8fd64cfe7ffc\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-29e2f68f-2339-4acf-a9b4-8fd64cfe7ffc\") pod \"ovsdbserver-nb-2\" (UID: \"40f9f201-93e1-4d27-842f-77040d48d28d\") " pod="openstack/ovsdbserver-nb-2" Oct 08 19:40:13 crc kubenswrapper[4988]: I1008 19:40:13.039793 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-2" Oct 08 19:40:13 crc kubenswrapper[4988]: I1008 19:40:13.045567 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-1" Oct 08 19:40:13 crc kubenswrapper[4988]: I1008 19:40:13.210421 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-sb-0"] Oct 08 19:40:13 crc kubenswrapper[4988]: I1008 19:40:13.217924 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Oct 08 19:40:13 crc kubenswrapper[4988]: I1008 19:40:13.218015 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Oct 08 19:40:13 crc kubenswrapper[4988]: I1008 19:40:13.220021 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-sb-dockercfg-qr85h" Oct 08 19:40:13 crc kubenswrapper[4988]: I1008 19:40:13.222184 4988 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-config" Oct 08 19:40:13 crc kubenswrapper[4988]: I1008 19:40:13.222437 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovndbcluster-sb-ovndbs" Oct 08 19:40:13 crc kubenswrapper[4988]: I1008 19:40:13.222602 4988 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-scripts" Oct 08 19:40:13 crc kubenswrapper[4988]: I1008 19:40:13.250789 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-sb-1"] Oct 08 19:40:13 crc kubenswrapper[4988]: I1008 19:40:13.252093 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-1" Oct 08 19:40:13 crc kubenswrapper[4988]: I1008 19:40:13.252373 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-sb-2"] Oct 08 19:40:13 crc kubenswrapper[4988]: I1008 19:40:13.253767 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-2" Oct 08 19:40:13 crc kubenswrapper[4988]: I1008 19:40:13.267487 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/ba4c95b7-edad-454a-86a5-d2edb9a8908d-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"ba4c95b7-edad-454a-86a5-d2edb9a8908d\") " pod="openstack/ovsdbserver-sb-0" Oct 08 19:40:13 crc kubenswrapper[4988]: I1008 19:40:13.267556 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-f1e00901-a757-4ad2-9991-57c6f0861a28\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-f1e00901-a757-4ad2-9991-57c6f0861a28\") pod \"ovsdbserver-sb-0\" (UID: \"ba4c95b7-edad-454a-86a5-d2edb9a8908d\") " pod="openstack/ovsdbserver-sb-0" Oct 08 19:40:13 crc kubenswrapper[4988]: I1008 19:40:13.267621 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ba4c95b7-edad-454a-86a5-d2edb9a8908d-config\") pod \"ovsdbserver-sb-0\" (UID: \"ba4c95b7-edad-454a-86a5-d2edb9a8908d\") " pod="openstack/ovsdbserver-sb-0" Oct 08 19:40:13 crc kubenswrapper[4988]: I1008 19:40:13.267647 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v66s8\" (UniqueName: \"kubernetes.io/projected/ba4c95b7-edad-454a-86a5-d2edb9a8908d-kube-api-access-v66s8\") pod \"ovsdbserver-sb-0\" (UID: \"ba4c95b7-edad-454a-86a5-d2edb9a8908d\") " pod="openstack/ovsdbserver-sb-0" Oct 08 19:40:13 crc kubenswrapper[4988]: I1008 19:40:13.267680 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ba4c95b7-edad-454a-86a5-d2edb9a8908d-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"ba4c95b7-edad-454a-86a5-d2edb9a8908d\") " pod="openstack/ovsdbserver-sb-0" Oct 08 19:40:13 crc kubenswrapper[4988]: I1008 19:40:13.267731 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ba4c95b7-edad-454a-86a5-d2edb9a8908d-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"ba4c95b7-edad-454a-86a5-d2edb9a8908d\") " pod="openstack/ovsdbserver-sb-0" Oct 08 19:40:13 crc kubenswrapper[4988]: I1008 19:40:13.267766 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/ba4c95b7-edad-454a-86a5-d2edb9a8908d-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"ba4c95b7-edad-454a-86a5-d2edb9a8908d\") " pod="openstack/ovsdbserver-sb-0" Oct 08 19:40:13 crc kubenswrapper[4988]: I1008 19:40:13.267788 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/ba4c95b7-edad-454a-86a5-d2edb9a8908d-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"ba4c95b7-edad-454a-86a5-d2edb9a8908d\") " pod="openstack/ovsdbserver-sb-0" Oct 08 19:40:13 crc kubenswrapper[4988]: I1008 19:40:13.272933 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-1"] Oct 08 19:40:13 crc kubenswrapper[4988]: I1008 19:40:13.286211 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-2"] Oct 08 19:40:13 crc kubenswrapper[4988]: I1008 19:40:13.369199 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/40504de9-4567-46fb-a598-6bbfb722b367-combined-ca-bundle\") pod \"ovsdbserver-sb-2\" (UID: \"40504de9-4567-46fb-a598-6bbfb722b367\") " pod="openstack/ovsdbserver-sb-2" Oct 08 19:40:13 crc kubenswrapper[4988]: I1008 19:40:13.369266 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ba4c95b7-edad-454a-86a5-d2edb9a8908d-config\") pod \"ovsdbserver-sb-0\" (UID: \"ba4c95b7-edad-454a-86a5-d2edb9a8908d\") " pod="openstack/ovsdbserver-sb-0" Oct 08 19:40:13 crc kubenswrapper[4988]: I1008 19:40:13.369295 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v66s8\" (UniqueName: \"kubernetes.io/projected/ba4c95b7-edad-454a-86a5-d2edb9a8908d-kube-api-access-v66s8\") pod \"ovsdbserver-sb-0\" (UID: \"ba4c95b7-edad-454a-86a5-d2edb9a8908d\") " pod="openstack/ovsdbserver-sb-0" Oct 08 19:40:13 crc kubenswrapper[4988]: I1008 19:40:13.369349 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ba4c95b7-edad-454a-86a5-d2edb9a8908d-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"ba4c95b7-edad-454a-86a5-d2edb9a8908d\") " pod="openstack/ovsdbserver-sb-0" Oct 08 19:40:13 crc kubenswrapper[4988]: I1008 19:40:13.369378 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dad7c6b4-98d7-490b-b17a-c4f13a12cd92-config\") pod \"ovsdbserver-sb-1\" (UID: \"dad7c6b4-98d7-490b-b17a-c4f13a12cd92\") " pod="openstack/ovsdbserver-sb-1" Oct 08 19:40:13 crc kubenswrapper[4988]: I1008 19:40:13.369444 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ba4c95b7-edad-454a-86a5-d2edb9a8908d-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"ba4c95b7-edad-454a-86a5-d2edb9a8908d\") " pod="openstack/ovsdbserver-sb-0" Oct 08 19:40:13 crc kubenswrapper[4988]: I1008 19:40:13.369470 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/dad7c6b4-98d7-490b-b17a-c4f13a12cd92-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-1\" (UID: \"dad7c6b4-98d7-490b-b17a-c4f13a12cd92\") " pod="openstack/ovsdbserver-sb-1" Oct 08 19:40:13 crc kubenswrapper[4988]: I1008 19:40:13.369508 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/40504de9-4567-46fb-a598-6bbfb722b367-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-2\" (UID: \"40504de9-4567-46fb-a598-6bbfb722b367\") " pod="openstack/ovsdbserver-sb-2" Oct 08 19:40:13 crc kubenswrapper[4988]: I1008 19:40:13.369530 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dad7c6b4-98d7-490b-b17a-c4f13a12cd92-combined-ca-bundle\") pod \"ovsdbserver-sb-1\" (UID: \"dad7c6b4-98d7-490b-b17a-c4f13a12cd92\") " pod="openstack/ovsdbserver-sb-1" Oct 08 19:40:13 crc kubenswrapper[4988]: I1008 19:40:13.369559 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/ba4c95b7-edad-454a-86a5-d2edb9a8908d-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"ba4c95b7-edad-454a-86a5-d2edb9a8908d\") " pod="openstack/ovsdbserver-sb-0" Oct 08 19:40:13 crc kubenswrapper[4988]: I1008 19:40:13.369584 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/ba4c95b7-edad-454a-86a5-d2edb9a8908d-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"ba4c95b7-edad-454a-86a5-d2edb9a8908d\") " pod="openstack/ovsdbserver-sb-0" Oct 08 19:40:13 crc kubenswrapper[4988]: I1008 19:40:13.369615 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b66dj\" (UniqueName: \"kubernetes.io/projected/dad7c6b4-98d7-490b-b17a-c4f13a12cd92-kube-api-access-b66dj\") pod \"ovsdbserver-sb-1\" (UID: \"dad7c6b4-98d7-490b-b17a-c4f13a12cd92\") " pod="openstack/ovsdbserver-sb-1" Oct 08 19:40:13 crc kubenswrapper[4988]: I1008 19:40:13.369657 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6rw6d\" (UniqueName: \"kubernetes.io/projected/40504de9-4567-46fb-a598-6bbfb722b367-kube-api-access-6rw6d\") pod \"ovsdbserver-sb-2\" (UID: \"40504de9-4567-46fb-a598-6bbfb722b367\") " pod="openstack/ovsdbserver-sb-2" Oct 08 19:40:13 crc kubenswrapper[4988]: I1008 19:40:13.369682 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-69d98e5e-5f9b-4ab8-ae80-3bf659bff3b1\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-69d98e5e-5f9b-4ab8-ae80-3bf659bff3b1\") pod \"ovsdbserver-sb-1\" (UID: \"dad7c6b4-98d7-490b-b17a-c4f13a12cd92\") " pod="openstack/ovsdbserver-sb-1" Oct 08 19:40:13 crc kubenswrapper[4988]: I1008 19:40:13.369708 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-066c3d91-7d28-4dfa-a5e3-871eaa4b03a5\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-066c3d91-7d28-4dfa-a5e3-871eaa4b03a5\") pod \"ovsdbserver-sb-2\" (UID: \"40504de9-4567-46fb-a598-6bbfb722b367\") " pod="openstack/ovsdbserver-sb-2" Oct 08 19:40:13 crc kubenswrapper[4988]: I1008 19:40:13.369738 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/40504de9-4567-46fb-a598-6bbfb722b367-scripts\") pod \"ovsdbserver-sb-2\" (UID: \"40504de9-4567-46fb-a598-6bbfb722b367\") " pod="openstack/ovsdbserver-sb-2" Oct 08 19:40:13 crc kubenswrapper[4988]: I1008 19:40:13.369766 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/40504de9-4567-46fb-a598-6bbfb722b367-ovsdb-rundir\") pod \"ovsdbserver-sb-2\" (UID: \"40504de9-4567-46fb-a598-6bbfb722b367\") " pod="openstack/ovsdbserver-sb-2" Oct 08 19:40:13 crc kubenswrapper[4988]: I1008 19:40:13.369792 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/40504de9-4567-46fb-a598-6bbfb722b367-config\") pod \"ovsdbserver-sb-2\" (UID: \"40504de9-4567-46fb-a598-6bbfb722b367\") " pod="openstack/ovsdbserver-sb-2" Oct 08 19:40:13 crc kubenswrapper[4988]: I1008 19:40:13.370118 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/dad7c6b4-98d7-490b-b17a-c4f13a12cd92-scripts\") pod \"ovsdbserver-sb-1\" (UID: \"dad7c6b4-98d7-490b-b17a-c4f13a12cd92\") " pod="openstack/ovsdbserver-sb-1" Oct 08 19:40:13 crc kubenswrapper[4988]: I1008 19:40:13.370183 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/dad7c6b4-98d7-490b-b17a-c4f13a12cd92-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-1\" (UID: \"dad7c6b4-98d7-490b-b17a-c4f13a12cd92\") " pod="openstack/ovsdbserver-sb-1" Oct 08 19:40:13 crc kubenswrapper[4988]: I1008 19:40:13.370288 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/ba4c95b7-edad-454a-86a5-d2edb9a8908d-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"ba4c95b7-edad-454a-86a5-d2edb9a8908d\") " pod="openstack/ovsdbserver-sb-0" Oct 08 19:40:13 crc kubenswrapper[4988]: I1008 19:40:13.370406 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-f1e00901-a757-4ad2-9991-57c6f0861a28\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-f1e00901-a757-4ad2-9991-57c6f0861a28\") pod \"ovsdbserver-sb-0\" (UID: \"ba4c95b7-edad-454a-86a5-d2edb9a8908d\") " pod="openstack/ovsdbserver-sb-0" Oct 08 19:40:13 crc kubenswrapper[4988]: I1008 19:40:13.370452 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/dad7c6b4-98d7-490b-b17a-c4f13a12cd92-ovsdb-rundir\") pod \"ovsdbserver-sb-1\" (UID: \"dad7c6b4-98d7-490b-b17a-c4f13a12cd92\") " pod="openstack/ovsdbserver-sb-1" Oct 08 19:40:13 crc kubenswrapper[4988]: I1008 19:40:13.370477 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/40504de9-4567-46fb-a598-6bbfb722b367-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-2\" (UID: \"40504de9-4567-46fb-a598-6bbfb722b367\") " pod="openstack/ovsdbserver-sb-2" Oct 08 19:40:13 crc kubenswrapper[4988]: I1008 19:40:13.370745 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ba4c95b7-edad-454a-86a5-d2edb9a8908d-config\") pod \"ovsdbserver-sb-0\" (UID: \"ba4c95b7-edad-454a-86a5-d2edb9a8908d\") " pod="openstack/ovsdbserver-sb-0" Oct 08 19:40:13 crc kubenswrapper[4988]: I1008 19:40:13.371034 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/ba4c95b7-edad-454a-86a5-d2edb9a8908d-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"ba4c95b7-edad-454a-86a5-d2edb9a8908d\") " pod="openstack/ovsdbserver-sb-0" Oct 08 19:40:13 crc kubenswrapper[4988]: I1008 19:40:13.372110 4988 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Oct 08 19:40:13 crc kubenswrapper[4988]: I1008 19:40:13.372132 4988 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-f1e00901-a757-4ad2-9991-57c6f0861a28\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-f1e00901-a757-4ad2-9991-57c6f0861a28\") pod \"ovsdbserver-sb-0\" (UID: \"ba4c95b7-edad-454a-86a5-d2edb9a8908d\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/c760b4cd48bfe234ce539e3ac0ce564bc404d7720116d320f370d3dd156333c1/globalmount\"" pod="openstack/ovsdbserver-sb-0" Oct 08 19:40:13 crc kubenswrapper[4988]: I1008 19:40:13.372332 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ba4c95b7-edad-454a-86a5-d2edb9a8908d-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"ba4c95b7-edad-454a-86a5-d2edb9a8908d\") " pod="openstack/ovsdbserver-sb-0" Oct 08 19:40:13 crc kubenswrapper[4988]: I1008 19:40:13.375589 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/ba4c95b7-edad-454a-86a5-d2edb9a8908d-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"ba4c95b7-edad-454a-86a5-d2edb9a8908d\") " pod="openstack/ovsdbserver-sb-0" Oct 08 19:40:13 crc kubenswrapper[4988]: I1008 19:40:13.375981 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ba4c95b7-edad-454a-86a5-d2edb9a8908d-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"ba4c95b7-edad-454a-86a5-d2edb9a8908d\") " pod="openstack/ovsdbserver-sb-0" Oct 08 19:40:13 crc kubenswrapper[4988]: I1008 19:40:13.376221 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/ba4c95b7-edad-454a-86a5-d2edb9a8908d-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"ba4c95b7-edad-454a-86a5-d2edb9a8908d\") " pod="openstack/ovsdbserver-sb-0" Oct 08 19:40:13 crc kubenswrapper[4988]: I1008 19:40:13.389765 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v66s8\" (UniqueName: \"kubernetes.io/projected/ba4c95b7-edad-454a-86a5-d2edb9a8908d-kube-api-access-v66s8\") pod \"ovsdbserver-sb-0\" (UID: \"ba4c95b7-edad-454a-86a5-d2edb9a8908d\") " pod="openstack/ovsdbserver-sb-0" Oct 08 19:40:13 crc kubenswrapper[4988]: I1008 19:40:13.410848 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-f1e00901-a757-4ad2-9991-57c6f0861a28\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-f1e00901-a757-4ad2-9991-57c6f0861a28\") pod \"ovsdbserver-sb-0\" (UID: \"ba4c95b7-edad-454a-86a5-d2edb9a8908d\") " pod="openstack/ovsdbserver-sb-0" Oct 08 19:40:13 crc kubenswrapper[4988]: I1008 19:40:13.472551 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/dad7c6b4-98d7-490b-b17a-c4f13a12cd92-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-1\" (UID: \"dad7c6b4-98d7-490b-b17a-c4f13a12cd92\") " pod="openstack/ovsdbserver-sb-1" Oct 08 19:40:13 crc kubenswrapper[4988]: I1008 19:40:13.472605 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/40504de9-4567-46fb-a598-6bbfb722b367-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-2\" (UID: \"40504de9-4567-46fb-a598-6bbfb722b367\") " pod="openstack/ovsdbserver-sb-2" Oct 08 19:40:13 crc kubenswrapper[4988]: I1008 19:40:13.472632 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dad7c6b4-98d7-490b-b17a-c4f13a12cd92-combined-ca-bundle\") pod \"ovsdbserver-sb-1\" (UID: \"dad7c6b4-98d7-490b-b17a-c4f13a12cd92\") " pod="openstack/ovsdbserver-sb-1" Oct 08 19:40:13 crc kubenswrapper[4988]: I1008 19:40:13.472673 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b66dj\" (UniqueName: \"kubernetes.io/projected/dad7c6b4-98d7-490b-b17a-c4f13a12cd92-kube-api-access-b66dj\") pod \"ovsdbserver-sb-1\" (UID: \"dad7c6b4-98d7-490b-b17a-c4f13a12cd92\") " pod="openstack/ovsdbserver-sb-1" Oct 08 19:40:13 crc kubenswrapper[4988]: I1008 19:40:13.472718 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6rw6d\" (UniqueName: \"kubernetes.io/projected/40504de9-4567-46fb-a598-6bbfb722b367-kube-api-access-6rw6d\") pod \"ovsdbserver-sb-2\" (UID: \"40504de9-4567-46fb-a598-6bbfb722b367\") " pod="openstack/ovsdbserver-sb-2" Oct 08 19:40:13 crc kubenswrapper[4988]: I1008 19:40:13.472745 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-69d98e5e-5f9b-4ab8-ae80-3bf659bff3b1\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-69d98e5e-5f9b-4ab8-ae80-3bf659bff3b1\") pod \"ovsdbserver-sb-1\" (UID: \"dad7c6b4-98d7-490b-b17a-c4f13a12cd92\") " pod="openstack/ovsdbserver-sb-1" Oct 08 19:40:13 crc kubenswrapper[4988]: I1008 19:40:13.472773 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-066c3d91-7d28-4dfa-a5e3-871eaa4b03a5\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-066c3d91-7d28-4dfa-a5e3-871eaa4b03a5\") pod \"ovsdbserver-sb-2\" (UID: \"40504de9-4567-46fb-a598-6bbfb722b367\") " pod="openstack/ovsdbserver-sb-2" Oct 08 19:40:13 crc kubenswrapper[4988]: I1008 19:40:13.472829 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/40504de9-4567-46fb-a598-6bbfb722b367-scripts\") pod \"ovsdbserver-sb-2\" (UID: \"40504de9-4567-46fb-a598-6bbfb722b367\") " pod="openstack/ovsdbserver-sb-2" Oct 08 19:40:13 crc kubenswrapper[4988]: I1008 19:40:13.472856 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/40504de9-4567-46fb-a598-6bbfb722b367-ovsdb-rundir\") pod \"ovsdbserver-sb-2\" (UID: \"40504de9-4567-46fb-a598-6bbfb722b367\") " pod="openstack/ovsdbserver-sb-2" Oct 08 19:40:13 crc kubenswrapper[4988]: I1008 19:40:13.472893 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/40504de9-4567-46fb-a598-6bbfb722b367-config\") pod \"ovsdbserver-sb-2\" (UID: \"40504de9-4567-46fb-a598-6bbfb722b367\") " pod="openstack/ovsdbserver-sb-2" Oct 08 19:40:13 crc kubenswrapper[4988]: I1008 19:40:13.472921 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/dad7c6b4-98d7-490b-b17a-c4f13a12cd92-scripts\") pod \"ovsdbserver-sb-1\" (UID: \"dad7c6b4-98d7-490b-b17a-c4f13a12cd92\") " pod="openstack/ovsdbserver-sb-1" Oct 08 19:40:13 crc kubenswrapper[4988]: I1008 19:40:13.472944 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/dad7c6b4-98d7-490b-b17a-c4f13a12cd92-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-1\" (UID: \"dad7c6b4-98d7-490b-b17a-c4f13a12cd92\") " pod="openstack/ovsdbserver-sb-1" Oct 08 19:40:13 crc kubenswrapper[4988]: I1008 19:40:13.473007 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/dad7c6b4-98d7-490b-b17a-c4f13a12cd92-ovsdb-rundir\") pod \"ovsdbserver-sb-1\" (UID: \"dad7c6b4-98d7-490b-b17a-c4f13a12cd92\") " pod="openstack/ovsdbserver-sb-1" Oct 08 19:40:13 crc kubenswrapper[4988]: I1008 19:40:13.473036 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/40504de9-4567-46fb-a598-6bbfb722b367-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-2\" (UID: \"40504de9-4567-46fb-a598-6bbfb722b367\") " pod="openstack/ovsdbserver-sb-2" Oct 08 19:40:13 crc kubenswrapper[4988]: I1008 19:40:13.473085 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/40504de9-4567-46fb-a598-6bbfb722b367-combined-ca-bundle\") pod \"ovsdbserver-sb-2\" (UID: \"40504de9-4567-46fb-a598-6bbfb722b367\") " pod="openstack/ovsdbserver-sb-2" Oct 08 19:40:13 crc kubenswrapper[4988]: I1008 19:40:13.473128 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dad7c6b4-98d7-490b-b17a-c4f13a12cd92-config\") pod \"ovsdbserver-sb-1\" (UID: \"dad7c6b4-98d7-490b-b17a-c4f13a12cd92\") " pod="openstack/ovsdbserver-sb-1" Oct 08 19:40:13 crc kubenswrapper[4988]: I1008 19:40:13.473975 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/40504de9-4567-46fb-a598-6bbfb722b367-config\") pod \"ovsdbserver-sb-2\" (UID: \"40504de9-4567-46fb-a598-6bbfb722b367\") " pod="openstack/ovsdbserver-sb-2" Oct 08 19:40:13 crc kubenswrapper[4988]: I1008 19:40:13.474486 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/40504de9-4567-46fb-a598-6bbfb722b367-ovsdb-rundir\") pod \"ovsdbserver-sb-2\" (UID: \"40504de9-4567-46fb-a598-6bbfb722b367\") " pod="openstack/ovsdbserver-sb-2" Oct 08 19:40:13 crc kubenswrapper[4988]: I1008 19:40:13.474486 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/dad7c6b4-98d7-490b-b17a-c4f13a12cd92-ovsdb-rundir\") pod \"ovsdbserver-sb-1\" (UID: \"dad7c6b4-98d7-490b-b17a-c4f13a12cd92\") " pod="openstack/ovsdbserver-sb-1" Oct 08 19:40:13 crc kubenswrapper[4988]: I1008 19:40:13.474865 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/40504de9-4567-46fb-a598-6bbfb722b367-scripts\") pod \"ovsdbserver-sb-2\" (UID: \"40504de9-4567-46fb-a598-6bbfb722b367\") " pod="openstack/ovsdbserver-sb-2" Oct 08 19:40:13 crc kubenswrapper[4988]: I1008 19:40:13.474975 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dad7c6b4-98d7-490b-b17a-c4f13a12cd92-config\") pod \"ovsdbserver-sb-1\" (UID: \"dad7c6b4-98d7-490b-b17a-c4f13a12cd92\") " pod="openstack/ovsdbserver-sb-1" Oct 08 19:40:13 crc kubenswrapper[4988]: I1008 19:40:13.475914 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/dad7c6b4-98d7-490b-b17a-c4f13a12cd92-scripts\") pod \"ovsdbserver-sb-1\" (UID: \"dad7c6b4-98d7-490b-b17a-c4f13a12cd92\") " pod="openstack/ovsdbserver-sb-1" Oct 08 19:40:13 crc kubenswrapper[4988]: I1008 19:40:13.476817 4988 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Oct 08 19:40:13 crc kubenswrapper[4988]: I1008 19:40:13.476845 4988 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-69d98e5e-5f9b-4ab8-ae80-3bf659bff3b1\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-69d98e5e-5f9b-4ab8-ae80-3bf659bff3b1\") pod \"ovsdbserver-sb-1\" (UID: \"dad7c6b4-98d7-490b-b17a-c4f13a12cd92\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/03d06c60787758d7d0c9da8bc78a13cb2f7825cf2da68dc62aef0a8e94a0ec53/globalmount\"" pod="openstack/ovsdbserver-sb-1" Oct 08 19:40:13 crc kubenswrapper[4988]: I1008 19:40:13.476969 4988 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Oct 08 19:40:13 crc kubenswrapper[4988]: I1008 19:40:13.477000 4988 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-066c3d91-7d28-4dfa-a5e3-871eaa4b03a5\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-066c3d91-7d28-4dfa-a5e3-871eaa4b03a5\") pod \"ovsdbserver-sb-2\" (UID: \"40504de9-4567-46fb-a598-6bbfb722b367\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/930529ea9f306c3df84fd60c22f7b3b6d64f80d09cfc1dacc1fd218a51fbaac7/globalmount\"" pod="openstack/ovsdbserver-sb-2" Oct 08 19:40:13 crc kubenswrapper[4988]: I1008 19:40:13.477090 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dad7c6b4-98d7-490b-b17a-c4f13a12cd92-combined-ca-bundle\") pod \"ovsdbserver-sb-1\" (UID: \"dad7c6b4-98d7-490b-b17a-c4f13a12cd92\") " pod="openstack/ovsdbserver-sb-1" Oct 08 19:40:13 crc kubenswrapper[4988]: I1008 19:40:13.477951 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/40504de9-4567-46fb-a598-6bbfb722b367-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-2\" (UID: \"40504de9-4567-46fb-a598-6bbfb722b367\") " pod="openstack/ovsdbserver-sb-2" Oct 08 19:40:13 crc kubenswrapper[4988]: I1008 19:40:13.478486 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/40504de9-4567-46fb-a598-6bbfb722b367-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-2\" (UID: \"40504de9-4567-46fb-a598-6bbfb722b367\") " pod="openstack/ovsdbserver-sb-2" Oct 08 19:40:13 crc kubenswrapper[4988]: I1008 19:40:13.479093 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/40504de9-4567-46fb-a598-6bbfb722b367-combined-ca-bundle\") pod \"ovsdbserver-sb-2\" (UID: \"40504de9-4567-46fb-a598-6bbfb722b367\") " pod="openstack/ovsdbserver-sb-2" Oct 08 19:40:13 crc kubenswrapper[4988]: I1008 19:40:13.479115 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/dad7c6b4-98d7-490b-b17a-c4f13a12cd92-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-1\" (UID: \"dad7c6b4-98d7-490b-b17a-c4f13a12cd92\") " pod="openstack/ovsdbserver-sb-1" Oct 08 19:40:13 crc kubenswrapper[4988]: I1008 19:40:13.480465 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/dad7c6b4-98d7-490b-b17a-c4f13a12cd92-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-1\" (UID: \"dad7c6b4-98d7-490b-b17a-c4f13a12cd92\") " pod="openstack/ovsdbserver-sb-1" Oct 08 19:40:13 crc kubenswrapper[4988]: I1008 19:40:13.492516 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b66dj\" (UniqueName: \"kubernetes.io/projected/dad7c6b4-98d7-490b-b17a-c4f13a12cd92-kube-api-access-b66dj\") pod \"ovsdbserver-sb-1\" (UID: \"dad7c6b4-98d7-490b-b17a-c4f13a12cd92\") " pod="openstack/ovsdbserver-sb-1" Oct 08 19:40:13 crc kubenswrapper[4988]: I1008 19:40:13.492868 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6rw6d\" (UniqueName: \"kubernetes.io/projected/40504de9-4567-46fb-a598-6bbfb722b367-kube-api-access-6rw6d\") pod \"ovsdbserver-sb-2\" (UID: \"40504de9-4567-46fb-a598-6bbfb722b367\") " pod="openstack/ovsdbserver-sb-2" Oct 08 19:40:13 crc kubenswrapper[4988]: I1008 19:40:13.515552 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Oct 08 19:40:13 crc kubenswrapper[4988]: I1008 19:40:13.524057 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-066c3d91-7d28-4dfa-a5e3-871eaa4b03a5\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-066c3d91-7d28-4dfa-a5e3-871eaa4b03a5\") pod \"ovsdbserver-sb-2\" (UID: \"40504de9-4567-46fb-a598-6bbfb722b367\") " pod="openstack/ovsdbserver-sb-2" Oct 08 19:40:13 crc kubenswrapper[4988]: I1008 19:40:13.529766 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-69d98e5e-5f9b-4ab8-ae80-3bf659bff3b1\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-69d98e5e-5f9b-4ab8-ae80-3bf659bff3b1\") pod \"ovsdbserver-sb-1\" (UID: \"dad7c6b4-98d7-490b-b17a-c4f13a12cd92\") " pod="openstack/ovsdbserver-sb-1" Oct 08 19:40:13 crc kubenswrapper[4988]: I1008 19:40:13.553607 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Oct 08 19:40:13 crc kubenswrapper[4988]: I1008 19:40:13.576737 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-1" Oct 08 19:40:13 crc kubenswrapper[4988]: I1008 19:40:13.585890 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-2" Oct 08 19:40:13 crc kubenswrapper[4988]: I1008 19:40:13.623705 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-2"] Oct 08 19:40:13 crc kubenswrapper[4988]: I1008 19:40:13.734436 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-1"] Oct 08 19:40:14 crc kubenswrapper[4988]: I1008 19:40:14.056623 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Oct 08 19:40:14 crc kubenswrapper[4988]: W1008 19:40:14.062773 4988 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podba4c95b7_edad_454a_86a5_d2edb9a8908d.slice/crio-17ea9b8604535a0479b0ec69a5d6343c085a08d46edb4ce7850ae81e040c1f12 WatchSource:0}: Error finding container 17ea9b8604535a0479b0ec69a5d6343c085a08d46edb4ce7850ae81e040c1f12: Status 404 returned error can't find the container with id 17ea9b8604535a0479b0ec69a5d6343c085a08d46edb4ce7850ae81e040c1f12 Oct 08 19:40:14 crc kubenswrapper[4988]: I1008 19:40:14.231245 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-1"] Oct 08 19:40:14 crc kubenswrapper[4988]: I1008 19:40:14.238205 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"e00c8be6-0ec3-4045-9a25-9f397cc78dc9","Type":"ContainerStarted","Data":"92b7c945e3dd159afe0e37c45ab27c3d63847a6ab9a65af7c118e1dc11359c63"} Oct 08 19:40:14 crc kubenswrapper[4988]: I1008 19:40:14.238248 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"e00c8be6-0ec3-4045-9a25-9f397cc78dc9","Type":"ContainerStarted","Data":"6c9746db92170fdd697f8a0c574876f8a592609bab70b2c4e0aa97bd15aa4f98"} Oct 08 19:40:14 crc kubenswrapper[4988]: I1008 19:40:14.238263 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"e00c8be6-0ec3-4045-9a25-9f397cc78dc9","Type":"ContainerStarted","Data":"b27671eef151e4300b213cdeb6dc987698eec23bd41686438e38b3d4cfc5643f"} Oct 08 19:40:14 crc kubenswrapper[4988]: W1008 19:40:14.239222 4988 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poddad7c6b4_98d7_490b_b17a_c4f13a12cd92.slice/crio-ad9a815b82bec5ac0be0891c3a835a3612b89958952ab54cf7d1461ce800007c WatchSource:0}: Error finding container ad9a815b82bec5ac0be0891c3a835a3612b89958952ab54cf7d1461ce800007c: Status 404 returned error can't find the container with id ad9a815b82bec5ac0be0891c3a835a3612b89958952ab54cf7d1461ce800007c Oct 08 19:40:14 crc kubenswrapper[4988]: I1008 19:40:14.241077 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"ba4c95b7-edad-454a-86a5-d2edb9a8908d","Type":"ContainerStarted","Data":"edc23422051b241da43ec4e4c0df27a8ec69cae7ef116cd60ee5f8a2a924d5c3"} Oct 08 19:40:14 crc kubenswrapper[4988]: I1008 19:40:14.241253 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"ba4c95b7-edad-454a-86a5-d2edb9a8908d","Type":"ContainerStarted","Data":"17ea9b8604535a0479b0ec69a5d6343c085a08d46edb4ce7850ae81e040c1f12"} Oct 08 19:40:14 crc kubenswrapper[4988]: I1008 19:40:14.243068 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-2" event={"ID":"40f9f201-93e1-4d27-842f-77040d48d28d","Type":"ContainerStarted","Data":"615e4622f01fd4a2d4abdf98fa0a97ce45f33de2892400cefe7b8ad9510e9b71"} Oct 08 19:40:14 crc kubenswrapper[4988]: I1008 19:40:14.243101 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-2" event={"ID":"40f9f201-93e1-4d27-842f-77040d48d28d","Type":"ContainerStarted","Data":"51aa1cdc22a9be40b64a78a6ff181b42f3829a410d1e5c123942438f83ed5a3b"} Oct 08 19:40:14 crc kubenswrapper[4988]: I1008 19:40:14.243110 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-2" event={"ID":"40f9f201-93e1-4d27-842f-77040d48d28d","Type":"ContainerStarted","Data":"19036cee712e347e369281b2410b8a0fb38aa630787317f6293326a4e3b7a135"} Oct 08 19:40:14 crc kubenswrapper[4988]: I1008 19:40:14.245034 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-1" event={"ID":"cbef3512-4ad7-45f0-9e87-d01ea0177708","Type":"ContainerStarted","Data":"4b5df6ff20be013d96bb461b331c1d4c7bd5828b226e70bc40550631056fd0b3"} Oct 08 19:40:14 crc kubenswrapper[4988]: I1008 19:40:14.245059 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-1" event={"ID":"cbef3512-4ad7-45f0-9e87-d01ea0177708","Type":"ContainerStarted","Data":"d98d2df4abe4863c753f465585e72067567056b6d4d0dc98a5bfd6665b00f33f"} Oct 08 19:40:14 crc kubenswrapper[4988]: I1008 19:40:14.257289 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-nb-0" podStartSLOduration=3.257265554 podStartE2EDuration="3.257265554s" podCreationTimestamp="2025-10-08 19:40:11 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 19:40:14.25276263 +0000 UTC m=+5359.702605400" watchObservedRunningTime="2025-10-08 19:40:14.257265554 +0000 UTC m=+5359.707108324" Oct 08 19:40:14 crc kubenswrapper[4988]: I1008 19:40:14.279219 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-nb-2" podStartSLOduration=3.279195371 podStartE2EDuration="3.279195371s" podCreationTimestamp="2025-10-08 19:40:11 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 19:40:14.273673426 +0000 UTC m=+5359.723516196" watchObservedRunningTime="2025-10-08 19:40:14.279195371 +0000 UTC m=+5359.729038141" Oct 08 19:40:14 crc kubenswrapper[4988]: I1008 19:40:14.315496 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-2"] Oct 08 19:40:15 crc kubenswrapper[4988]: I1008 19:40:15.259945 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-1" event={"ID":"cbef3512-4ad7-45f0-9e87-d01ea0177708","Type":"ContainerStarted","Data":"dd616b7f13ca7fd6749172ff9a01ce2e826a8d0a7285231f37c73f83c039af91"} Oct 08 19:40:15 crc kubenswrapper[4988]: I1008 19:40:15.261464 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-2" event={"ID":"40504de9-4567-46fb-a598-6bbfb722b367","Type":"ContainerStarted","Data":"db3053c9f065567d9fb004ba681f609d855c9d3e78b92e3587249dfbc5165a47"} Oct 08 19:40:15 crc kubenswrapper[4988]: I1008 19:40:15.261516 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-2" event={"ID":"40504de9-4567-46fb-a598-6bbfb722b367","Type":"ContainerStarted","Data":"4a116423548bf26bfaf690fb166a178d04e81dc2d04db202751748871c1f9b3c"} Oct 08 19:40:15 crc kubenswrapper[4988]: I1008 19:40:15.261535 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-2" event={"ID":"40504de9-4567-46fb-a598-6bbfb722b367","Type":"ContainerStarted","Data":"3a646c63f56ed69a0ba392ff7a346f4e8dda3ed993ade990af2afd5a030519e3"} Oct 08 19:40:15 crc kubenswrapper[4988]: I1008 19:40:15.263939 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-1" event={"ID":"dad7c6b4-98d7-490b-b17a-c4f13a12cd92","Type":"ContainerStarted","Data":"ac1884b2967b8286c2645703edd4d1acc2005adfe57b4b6919c694b29e725187"} Oct 08 19:40:15 crc kubenswrapper[4988]: I1008 19:40:15.263988 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-1" event={"ID":"dad7c6b4-98d7-490b-b17a-c4f13a12cd92","Type":"ContainerStarted","Data":"1cd9d77ac2ad9b25c189d6900adb91fb0db24dab4494d62137c86f7dfbae20b5"} Oct 08 19:40:15 crc kubenswrapper[4988]: I1008 19:40:15.264016 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-1" event={"ID":"dad7c6b4-98d7-490b-b17a-c4f13a12cd92","Type":"ContainerStarted","Data":"ad9a815b82bec5ac0be0891c3a835a3612b89958952ab54cf7d1461ce800007c"} Oct 08 19:40:15 crc kubenswrapper[4988]: I1008 19:40:15.267200 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"ba4c95b7-edad-454a-86a5-d2edb9a8908d","Type":"ContainerStarted","Data":"c54ae1524a2d6593ccdccd6673d25395ba55c72e4800dc7c76cd3a352fc48b79"} Oct 08 19:40:15 crc kubenswrapper[4988]: I1008 19:40:15.396189 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-sb-0" podStartSLOduration=3.396049653 podStartE2EDuration="3.396049653s" podCreationTimestamp="2025-10-08 19:40:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 19:40:15.384958081 +0000 UTC m=+5360.834800891" watchObservedRunningTime="2025-10-08 19:40:15.396049653 +0000 UTC m=+5360.845892453" Oct 08 19:40:15 crc kubenswrapper[4988]: I1008 19:40:15.418127 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-sb-1" podStartSLOduration=3.418102814 podStartE2EDuration="3.418102814s" podCreationTimestamp="2025-10-08 19:40:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 19:40:15.414423267 +0000 UTC m=+5360.864266077" watchObservedRunningTime="2025-10-08 19:40:15.418102814 +0000 UTC m=+5360.867945674" Oct 08 19:40:15 crc kubenswrapper[4988]: I1008 19:40:15.440436 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-sb-2" podStartSLOduration=3.440413713 podStartE2EDuration="3.440413713s" podCreationTimestamp="2025-10-08 19:40:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 19:40:15.433812243 +0000 UTC m=+5360.883655023" watchObservedRunningTime="2025-10-08 19:40:15.440413713 +0000 UTC m=+5360.890256494" Oct 08 19:40:15 crc kubenswrapper[4988]: I1008 19:40:15.472973 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-nb-1" podStartSLOduration=4.472954969 podStartE2EDuration="4.472954969s" podCreationTimestamp="2025-10-08 19:40:11 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 19:40:15.470379236 +0000 UTC m=+5360.920222016" watchObservedRunningTime="2025-10-08 19:40:15.472954969 +0000 UTC m=+5360.922797749" Oct 08 19:40:15 crc kubenswrapper[4988]: I1008 19:40:15.979736 4988 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-nb-0" Oct 08 19:40:16 crc kubenswrapper[4988]: I1008 19:40:16.040561 4988 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-nb-2" Oct 08 19:40:16 crc kubenswrapper[4988]: I1008 19:40:16.046271 4988 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-nb-1" Oct 08 19:40:16 crc kubenswrapper[4988]: I1008 19:40:16.099524 4988 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-nb-2" Oct 08 19:40:16 crc kubenswrapper[4988]: I1008 19:40:16.275461 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-nb-2" Oct 08 19:40:16 crc kubenswrapper[4988]: I1008 19:40:16.554290 4988 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-sb-0" Oct 08 19:40:16 crc kubenswrapper[4988]: I1008 19:40:16.577894 4988 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-sb-1" Oct 08 19:40:16 crc kubenswrapper[4988]: I1008 19:40:16.586040 4988 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-sb-2" Oct 08 19:40:17 crc kubenswrapper[4988]: I1008 19:40:17.979561 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-nb-0" Oct 08 19:40:18 crc kubenswrapper[4988]: I1008 19:40:18.045694 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-nb-1" Oct 08 19:40:18 crc kubenswrapper[4988]: I1008 19:40:18.116994 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-nb-2" Oct 08 19:40:18 crc kubenswrapper[4988]: I1008 19:40:18.441776 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-6444955555-ssqb7"] Oct 08 19:40:18 crc kubenswrapper[4988]: I1008 19:40:18.442997 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6444955555-ssqb7" Oct 08 19:40:18 crc kubenswrapper[4988]: I1008 19:40:18.446097 4988 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-nb" Oct 08 19:40:18 crc kubenswrapper[4988]: I1008 19:40:18.470405 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6444955555-ssqb7"] Oct 08 19:40:18 crc kubenswrapper[4988]: I1008 19:40:18.476728 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2c6cd276-44be-4ae8-af3e-272844e7b27e-config\") pod \"dnsmasq-dns-6444955555-ssqb7\" (UID: \"2c6cd276-44be-4ae8-af3e-272844e7b27e\") " pod="openstack/dnsmasq-dns-6444955555-ssqb7" Oct 08 19:40:18 crc kubenswrapper[4988]: I1008 19:40:18.476815 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/2c6cd276-44be-4ae8-af3e-272844e7b27e-ovsdbserver-nb\") pod \"dnsmasq-dns-6444955555-ssqb7\" (UID: \"2c6cd276-44be-4ae8-af3e-272844e7b27e\") " pod="openstack/dnsmasq-dns-6444955555-ssqb7" Oct 08 19:40:18 crc kubenswrapper[4988]: I1008 19:40:18.476891 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xhfpf\" (UniqueName: \"kubernetes.io/projected/2c6cd276-44be-4ae8-af3e-272844e7b27e-kube-api-access-xhfpf\") pod \"dnsmasq-dns-6444955555-ssqb7\" (UID: \"2c6cd276-44be-4ae8-af3e-272844e7b27e\") " pod="openstack/dnsmasq-dns-6444955555-ssqb7" Oct 08 19:40:18 crc kubenswrapper[4988]: I1008 19:40:18.476943 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2c6cd276-44be-4ae8-af3e-272844e7b27e-dns-svc\") pod \"dnsmasq-dns-6444955555-ssqb7\" (UID: \"2c6cd276-44be-4ae8-af3e-272844e7b27e\") " pod="openstack/dnsmasq-dns-6444955555-ssqb7" Oct 08 19:40:18 crc kubenswrapper[4988]: I1008 19:40:18.554019 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-sb-0" Oct 08 19:40:18 crc kubenswrapper[4988]: I1008 19:40:18.576900 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-sb-1" Oct 08 19:40:18 crc kubenswrapper[4988]: I1008 19:40:18.577976 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2c6cd276-44be-4ae8-af3e-272844e7b27e-dns-svc\") pod \"dnsmasq-dns-6444955555-ssqb7\" (UID: \"2c6cd276-44be-4ae8-af3e-272844e7b27e\") " pod="openstack/dnsmasq-dns-6444955555-ssqb7" Oct 08 19:40:18 crc kubenswrapper[4988]: I1008 19:40:18.578084 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2c6cd276-44be-4ae8-af3e-272844e7b27e-config\") pod \"dnsmasq-dns-6444955555-ssqb7\" (UID: \"2c6cd276-44be-4ae8-af3e-272844e7b27e\") " pod="openstack/dnsmasq-dns-6444955555-ssqb7" Oct 08 19:40:18 crc kubenswrapper[4988]: I1008 19:40:18.578141 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/2c6cd276-44be-4ae8-af3e-272844e7b27e-ovsdbserver-nb\") pod \"dnsmasq-dns-6444955555-ssqb7\" (UID: \"2c6cd276-44be-4ae8-af3e-272844e7b27e\") " pod="openstack/dnsmasq-dns-6444955555-ssqb7" Oct 08 19:40:18 crc kubenswrapper[4988]: I1008 19:40:18.578188 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xhfpf\" (UniqueName: \"kubernetes.io/projected/2c6cd276-44be-4ae8-af3e-272844e7b27e-kube-api-access-xhfpf\") pod \"dnsmasq-dns-6444955555-ssqb7\" (UID: \"2c6cd276-44be-4ae8-af3e-272844e7b27e\") " pod="openstack/dnsmasq-dns-6444955555-ssqb7" Oct 08 19:40:18 crc kubenswrapper[4988]: I1008 19:40:18.578849 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2c6cd276-44be-4ae8-af3e-272844e7b27e-dns-svc\") pod \"dnsmasq-dns-6444955555-ssqb7\" (UID: \"2c6cd276-44be-4ae8-af3e-272844e7b27e\") " pod="openstack/dnsmasq-dns-6444955555-ssqb7" Oct 08 19:40:18 crc kubenswrapper[4988]: I1008 19:40:18.579207 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/2c6cd276-44be-4ae8-af3e-272844e7b27e-ovsdbserver-nb\") pod \"dnsmasq-dns-6444955555-ssqb7\" (UID: \"2c6cd276-44be-4ae8-af3e-272844e7b27e\") " pod="openstack/dnsmasq-dns-6444955555-ssqb7" Oct 08 19:40:18 crc kubenswrapper[4988]: I1008 19:40:18.579473 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2c6cd276-44be-4ae8-af3e-272844e7b27e-config\") pod \"dnsmasq-dns-6444955555-ssqb7\" (UID: \"2c6cd276-44be-4ae8-af3e-272844e7b27e\") " pod="openstack/dnsmasq-dns-6444955555-ssqb7" Oct 08 19:40:18 crc kubenswrapper[4988]: I1008 19:40:18.586874 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-sb-2" Oct 08 19:40:18 crc kubenswrapper[4988]: I1008 19:40:18.599239 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xhfpf\" (UniqueName: \"kubernetes.io/projected/2c6cd276-44be-4ae8-af3e-272844e7b27e-kube-api-access-xhfpf\") pod \"dnsmasq-dns-6444955555-ssqb7\" (UID: \"2c6cd276-44be-4ae8-af3e-272844e7b27e\") " pod="openstack/dnsmasq-dns-6444955555-ssqb7" Oct 08 19:40:18 crc kubenswrapper[4988]: I1008 19:40:18.771874 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6444955555-ssqb7" Oct 08 19:40:19 crc kubenswrapper[4988]: I1008 19:40:19.034852 4988 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-nb-0" Oct 08 19:40:19 crc kubenswrapper[4988]: I1008 19:40:19.087458 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-nb-0" Oct 08 19:40:19 crc kubenswrapper[4988]: I1008 19:40:19.093742 4988 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-nb-1" Oct 08 19:40:19 crc kubenswrapper[4988]: I1008 19:40:19.326532 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6444955555-ssqb7"] Oct 08 19:40:19 crc kubenswrapper[4988]: I1008 19:40:19.352469 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-nb-1" Oct 08 19:40:19 crc kubenswrapper[4988]: I1008 19:40:19.601454 4988 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-sb-0" Oct 08 19:40:19 crc kubenswrapper[4988]: I1008 19:40:19.636127 4988 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-sb-2" Oct 08 19:40:19 crc kubenswrapper[4988]: I1008 19:40:19.646829 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-sb-0" Oct 08 19:40:19 crc kubenswrapper[4988]: I1008 19:40:19.648888 4988 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-sb-1" Oct 08 19:40:19 crc kubenswrapper[4988]: I1008 19:40:19.689657 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-sb-2" Oct 08 19:40:19 crc kubenswrapper[4988]: I1008 19:40:19.705221 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-sb-1" Oct 08 19:40:19 crc kubenswrapper[4988]: I1008 19:40:19.866758 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6444955555-ssqb7"] Oct 08 19:40:19 crc kubenswrapper[4988]: I1008 19:40:19.887502 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-69864b896c-pw7dl"] Oct 08 19:40:19 crc kubenswrapper[4988]: I1008 19:40:19.888806 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-69864b896c-pw7dl" Oct 08 19:40:19 crc kubenswrapper[4988]: I1008 19:40:19.890912 4988 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-sb" Oct 08 19:40:19 crc kubenswrapper[4988]: I1008 19:40:19.907900 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-69864b896c-pw7dl"] Oct 08 19:40:20 crc kubenswrapper[4988]: I1008 19:40:20.000069 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4babecb5-3754-4372-8322-6bdddbc06931-dns-svc\") pod \"dnsmasq-dns-69864b896c-pw7dl\" (UID: \"4babecb5-3754-4372-8322-6bdddbc06931\") " pod="openstack/dnsmasq-dns-69864b896c-pw7dl" Oct 08 19:40:20 crc kubenswrapper[4988]: I1008 19:40:20.000313 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s7zwt\" (UniqueName: \"kubernetes.io/projected/4babecb5-3754-4372-8322-6bdddbc06931-kube-api-access-s7zwt\") pod \"dnsmasq-dns-69864b896c-pw7dl\" (UID: \"4babecb5-3754-4372-8322-6bdddbc06931\") " pod="openstack/dnsmasq-dns-69864b896c-pw7dl" Oct 08 19:40:20 crc kubenswrapper[4988]: I1008 19:40:20.000362 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/4babecb5-3754-4372-8322-6bdddbc06931-ovsdbserver-sb\") pod \"dnsmasq-dns-69864b896c-pw7dl\" (UID: \"4babecb5-3754-4372-8322-6bdddbc06931\") " pod="openstack/dnsmasq-dns-69864b896c-pw7dl" Oct 08 19:40:20 crc kubenswrapper[4988]: I1008 19:40:20.000521 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4babecb5-3754-4372-8322-6bdddbc06931-config\") pod \"dnsmasq-dns-69864b896c-pw7dl\" (UID: \"4babecb5-3754-4372-8322-6bdddbc06931\") " pod="openstack/dnsmasq-dns-69864b896c-pw7dl" Oct 08 19:40:20 crc kubenswrapper[4988]: I1008 19:40:20.000581 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4babecb5-3754-4372-8322-6bdddbc06931-ovsdbserver-nb\") pod \"dnsmasq-dns-69864b896c-pw7dl\" (UID: \"4babecb5-3754-4372-8322-6bdddbc06931\") " pod="openstack/dnsmasq-dns-69864b896c-pw7dl" Oct 08 19:40:20 crc kubenswrapper[4988]: I1008 19:40:20.101998 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4babecb5-3754-4372-8322-6bdddbc06931-dns-svc\") pod \"dnsmasq-dns-69864b896c-pw7dl\" (UID: \"4babecb5-3754-4372-8322-6bdddbc06931\") " pod="openstack/dnsmasq-dns-69864b896c-pw7dl" Oct 08 19:40:20 crc kubenswrapper[4988]: I1008 19:40:20.102159 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s7zwt\" (UniqueName: \"kubernetes.io/projected/4babecb5-3754-4372-8322-6bdddbc06931-kube-api-access-s7zwt\") pod \"dnsmasq-dns-69864b896c-pw7dl\" (UID: \"4babecb5-3754-4372-8322-6bdddbc06931\") " pod="openstack/dnsmasq-dns-69864b896c-pw7dl" Oct 08 19:40:20 crc kubenswrapper[4988]: I1008 19:40:20.102187 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/4babecb5-3754-4372-8322-6bdddbc06931-ovsdbserver-sb\") pod \"dnsmasq-dns-69864b896c-pw7dl\" (UID: \"4babecb5-3754-4372-8322-6bdddbc06931\") " pod="openstack/dnsmasq-dns-69864b896c-pw7dl" Oct 08 19:40:20 crc kubenswrapper[4988]: I1008 19:40:20.102235 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4babecb5-3754-4372-8322-6bdddbc06931-config\") pod \"dnsmasq-dns-69864b896c-pw7dl\" (UID: \"4babecb5-3754-4372-8322-6bdddbc06931\") " pod="openstack/dnsmasq-dns-69864b896c-pw7dl" Oct 08 19:40:20 crc kubenswrapper[4988]: I1008 19:40:20.102266 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4babecb5-3754-4372-8322-6bdddbc06931-ovsdbserver-nb\") pod \"dnsmasq-dns-69864b896c-pw7dl\" (UID: \"4babecb5-3754-4372-8322-6bdddbc06931\") " pod="openstack/dnsmasq-dns-69864b896c-pw7dl" Oct 08 19:40:20 crc kubenswrapper[4988]: I1008 19:40:20.103010 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4babecb5-3754-4372-8322-6bdddbc06931-dns-svc\") pod \"dnsmasq-dns-69864b896c-pw7dl\" (UID: \"4babecb5-3754-4372-8322-6bdddbc06931\") " pod="openstack/dnsmasq-dns-69864b896c-pw7dl" Oct 08 19:40:20 crc kubenswrapper[4988]: I1008 19:40:20.103067 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4babecb5-3754-4372-8322-6bdddbc06931-config\") pod \"dnsmasq-dns-69864b896c-pw7dl\" (UID: \"4babecb5-3754-4372-8322-6bdddbc06931\") " pod="openstack/dnsmasq-dns-69864b896c-pw7dl" Oct 08 19:40:20 crc kubenswrapper[4988]: I1008 19:40:20.103136 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4babecb5-3754-4372-8322-6bdddbc06931-ovsdbserver-nb\") pod \"dnsmasq-dns-69864b896c-pw7dl\" (UID: \"4babecb5-3754-4372-8322-6bdddbc06931\") " pod="openstack/dnsmasq-dns-69864b896c-pw7dl" Oct 08 19:40:20 crc kubenswrapper[4988]: I1008 19:40:20.103512 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/4babecb5-3754-4372-8322-6bdddbc06931-ovsdbserver-sb\") pod \"dnsmasq-dns-69864b896c-pw7dl\" (UID: \"4babecb5-3754-4372-8322-6bdddbc06931\") " pod="openstack/dnsmasq-dns-69864b896c-pw7dl" Oct 08 19:40:20 crc kubenswrapper[4988]: I1008 19:40:20.132430 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s7zwt\" (UniqueName: \"kubernetes.io/projected/4babecb5-3754-4372-8322-6bdddbc06931-kube-api-access-s7zwt\") pod \"dnsmasq-dns-69864b896c-pw7dl\" (UID: \"4babecb5-3754-4372-8322-6bdddbc06931\") " pod="openstack/dnsmasq-dns-69864b896c-pw7dl" Oct 08 19:40:20 crc kubenswrapper[4988]: I1008 19:40:20.209302 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-69864b896c-pw7dl" Oct 08 19:40:20 crc kubenswrapper[4988]: I1008 19:40:20.312334 4988 generic.go:334] "Generic (PLEG): container finished" podID="2c6cd276-44be-4ae8-af3e-272844e7b27e" containerID="00fef83e4152d6831ab20bd852c969e97aa6d96313b582774360b2a013cb1124" exitCode=0 Oct 08 19:40:20 crc kubenswrapper[4988]: I1008 19:40:20.312509 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6444955555-ssqb7" event={"ID":"2c6cd276-44be-4ae8-af3e-272844e7b27e","Type":"ContainerDied","Data":"00fef83e4152d6831ab20bd852c969e97aa6d96313b582774360b2a013cb1124"} Oct 08 19:40:20 crc kubenswrapper[4988]: I1008 19:40:20.312552 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6444955555-ssqb7" event={"ID":"2c6cd276-44be-4ae8-af3e-272844e7b27e","Type":"ContainerStarted","Data":"284b30656ca881193f912ed9b5033ff7015728ecbcfffb11ac4023d7fb4bf984"} Oct 08 19:40:20 crc kubenswrapper[4988]: I1008 19:40:20.648719 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-69864b896c-pw7dl"] Oct 08 19:40:21 crc kubenswrapper[4988]: I1008 19:40:21.321455 4988 generic.go:334] "Generic (PLEG): container finished" podID="4babecb5-3754-4372-8322-6bdddbc06931" containerID="238814049da025e35cf5de0e82820d2f2b2e7db3b9e15fe4e489117316b60cdc" exitCode=0 Oct 08 19:40:21 crc kubenswrapper[4988]: I1008 19:40:21.321589 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-69864b896c-pw7dl" event={"ID":"4babecb5-3754-4372-8322-6bdddbc06931","Type":"ContainerDied","Data":"238814049da025e35cf5de0e82820d2f2b2e7db3b9e15fe4e489117316b60cdc"} Oct 08 19:40:21 crc kubenswrapper[4988]: I1008 19:40:21.321864 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-69864b896c-pw7dl" event={"ID":"4babecb5-3754-4372-8322-6bdddbc06931","Type":"ContainerStarted","Data":"5d181cce6c5543682377478f69f2fdd9f09e4180f5e47455d0f0fd28384e7b0e"} Oct 08 19:40:21 crc kubenswrapper[4988]: I1008 19:40:21.326795 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6444955555-ssqb7" event={"ID":"2c6cd276-44be-4ae8-af3e-272844e7b27e","Type":"ContainerStarted","Data":"0c51531936f6f5c38ab873e494a3bd07271c87a721a0e4cc0f57a3d3466fb363"} Oct 08 19:40:21 crc kubenswrapper[4988]: I1008 19:40:21.326938 4988 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-6444955555-ssqb7" podUID="2c6cd276-44be-4ae8-af3e-272844e7b27e" containerName="dnsmasq-dns" containerID="cri-o://0c51531936f6f5c38ab873e494a3bd07271c87a721a0e4cc0f57a3d3466fb363" gracePeriod=10 Oct 08 19:40:21 crc kubenswrapper[4988]: I1008 19:40:21.326996 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-6444955555-ssqb7" Oct 08 19:40:21 crc kubenswrapper[4988]: I1008 19:40:21.375539 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-6444955555-ssqb7" podStartSLOduration=3.375524088 podStartE2EDuration="3.375524088s" podCreationTimestamp="2025-10-08 19:40:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 19:40:21.372310005 +0000 UTC m=+5366.822152775" watchObservedRunningTime="2025-10-08 19:40:21.375524088 +0000 UTC m=+5366.825366858" Oct 08 19:40:21 crc kubenswrapper[4988]: I1008 19:40:21.718132 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6444955555-ssqb7" Oct 08 19:40:21 crc kubenswrapper[4988]: I1008 19:40:21.837733 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/2c6cd276-44be-4ae8-af3e-272844e7b27e-ovsdbserver-nb\") pod \"2c6cd276-44be-4ae8-af3e-272844e7b27e\" (UID: \"2c6cd276-44be-4ae8-af3e-272844e7b27e\") " Oct 08 19:40:21 crc kubenswrapper[4988]: I1008 19:40:21.837893 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2c6cd276-44be-4ae8-af3e-272844e7b27e-config\") pod \"2c6cd276-44be-4ae8-af3e-272844e7b27e\" (UID: \"2c6cd276-44be-4ae8-af3e-272844e7b27e\") " Oct 08 19:40:21 crc kubenswrapper[4988]: I1008 19:40:21.837941 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2c6cd276-44be-4ae8-af3e-272844e7b27e-dns-svc\") pod \"2c6cd276-44be-4ae8-af3e-272844e7b27e\" (UID: \"2c6cd276-44be-4ae8-af3e-272844e7b27e\") " Oct 08 19:40:21 crc kubenswrapper[4988]: I1008 19:40:21.837964 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xhfpf\" (UniqueName: \"kubernetes.io/projected/2c6cd276-44be-4ae8-af3e-272844e7b27e-kube-api-access-xhfpf\") pod \"2c6cd276-44be-4ae8-af3e-272844e7b27e\" (UID: \"2c6cd276-44be-4ae8-af3e-272844e7b27e\") " Oct 08 19:40:21 crc kubenswrapper[4988]: I1008 19:40:21.843603 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2c6cd276-44be-4ae8-af3e-272844e7b27e-kube-api-access-xhfpf" (OuterVolumeSpecName: "kube-api-access-xhfpf") pod "2c6cd276-44be-4ae8-af3e-272844e7b27e" (UID: "2c6cd276-44be-4ae8-af3e-272844e7b27e"). InnerVolumeSpecName "kube-api-access-xhfpf". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 19:40:21 crc kubenswrapper[4988]: I1008 19:40:21.874014 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2c6cd276-44be-4ae8-af3e-272844e7b27e-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "2c6cd276-44be-4ae8-af3e-272844e7b27e" (UID: "2c6cd276-44be-4ae8-af3e-272844e7b27e"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 19:40:21 crc kubenswrapper[4988]: I1008 19:40:21.877505 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2c6cd276-44be-4ae8-af3e-272844e7b27e-config" (OuterVolumeSpecName: "config") pod "2c6cd276-44be-4ae8-af3e-272844e7b27e" (UID: "2c6cd276-44be-4ae8-af3e-272844e7b27e"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 19:40:21 crc kubenswrapper[4988]: I1008 19:40:21.902969 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2c6cd276-44be-4ae8-af3e-272844e7b27e-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "2c6cd276-44be-4ae8-af3e-272844e7b27e" (UID: "2c6cd276-44be-4ae8-af3e-272844e7b27e"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 19:40:21 crc kubenswrapper[4988]: I1008 19:40:21.940225 4988 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/2c6cd276-44be-4ae8-af3e-272844e7b27e-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 08 19:40:21 crc kubenswrapper[4988]: I1008 19:40:21.940264 4988 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2c6cd276-44be-4ae8-af3e-272844e7b27e-config\") on node \"crc\" DevicePath \"\"" Oct 08 19:40:21 crc kubenswrapper[4988]: I1008 19:40:21.940278 4988 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2c6cd276-44be-4ae8-af3e-272844e7b27e-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 08 19:40:21 crc kubenswrapper[4988]: I1008 19:40:21.940291 4988 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xhfpf\" (UniqueName: \"kubernetes.io/projected/2c6cd276-44be-4ae8-af3e-272844e7b27e-kube-api-access-xhfpf\") on node \"crc\" DevicePath \"\"" Oct 08 19:40:22 crc kubenswrapper[4988]: I1008 19:40:22.186401 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-copy-data"] Oct 08 19:40:22 crc kubenswrapper[4988]: E1008 19:40:22.186766 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2c6cd276-44be-4ae8-af3e-272844e7b27e" containerName="dnsmasq-dns" Oct 08 19:40:22 crc kubenswrapper[4988]: I1008 19:40:22.186780 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="2c6cd276-44be-4ae8-af3e-272844e7b27e" containerName="dnsmasq-dns" Oct 08 19:40:22 crc kubenswrapper[4988]: E1008 19:40:22.186802 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2c6cd276-44be-4ae8-af3e-272844e7b27e" containerName="init" Oct 08 19:40:22 crc kubenswrapper[4988]: I1008 19:40:22.186810 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="2c6cd276-44be-4ae8-af3e-272844e7b27e" containerName="init" Oct 08 19:40:22 crc kubenswrapper[4988]: I1008 19:40:22.186993 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="2c6cd276-44be-4ae8-af3e-272844e7b27e" containerName="dnsmasq-dns" Oct 08 19:40:22 crc kubenswrapper[4988]: I1008 19:40:22.187652 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-copy-data" Oct 08 19:40:22 crc kubenswrapper[4988]: I1008 19:40:22.196115 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovn-data-cert" Oct 08 19:40:22 crc kubenswrapper[4988]: I1008 19:40:22.198570 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-copy-data"] Oct 08 19:40:22 crc kubenswrapper[4988]: I1008 19:40:22.247654 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-data-cert\" (UniqueName: \"kubernetes.io/secret/bc523d09-e017-4e91-8ba0-45ca290bd102-ovn-data-cert\") pod \"ovn-copy-data\" (UID: \"bc523d09-e017-4e91-8ba0-45ca290bd102\") " pod="openstack/ovn-copy-data" Oct 08 19:40:22 crc kubenswrapper[4988]: I1008 19:40:22.247729 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-c3809aad-6fef-4719-a8e9-1427666bd89a\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-c3809aad-6fef-4719-a8e9-1427666bd89a\") pod \"ovn-copy-data\" (UID: \"bc523d09-e017-4e91-8ba0-45ca290bd102\") " pod="openstack/ovn-copy-data" Oct 08 19:40:22 crc kubenswrapper[4988]: I1008 19:40:22.247762 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dfqbt\" (UniqueName: \"kubernetes.io/projected/bc523d09-e017-4e91-8ba0-45ca290bd102-kube-api-access-dfqbt\") pod \"ovn-copy-data\" (UID: \"bc523d09-e017-4e91-8ba0-45ca290bd102\") " pod="openstack/ovn-copy-data" Oct 08 19:40:22 crc kubenswrapper[4988]: I1008 19:40:22.349137 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-data-cert\" (UniqueName: \"kubernetes.io/secret/bc523d09-e017-4e91-8ba0-45ca290bd102-ovn-data-cert\") pod \"ovn-copy-data\" (UID: \"bc523d09-e017-4e91-8ba0-45ca290bd102\") " pod="openstack/ovn-copy-data" Oct 08 19:40:22 crc kubenswrapper[4988]: I1008 19:40:22.349275 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-c3809aad-6fef-4719-a8e9-1427666bd89a\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-c3809aad-6fef-4719-a8e9-1427666bd89a\") pod \"ovn-copy-data\" (UID: \"bc523d09-e017-4e91-8ba0-45ca290bd102\") " pod="openstack/ovn-copy-data" Oct 08 19:40:22 crc kubenswrapper[4988]: I1008 19:40:22.349341 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dfqbt\" (UniqueName: \"kubernetes.io/projected/bc523d09-e017-4e91-8ba0-45ca290bd102-kube-api-access-dfqbt\") pod \"ovn-copy-data\" (UID: \"bc523d09-e017-4e91-8ba0-45ca290bd102\") " pod="openstack/ovn-copy-data" Oct 08 19:40:22 crc kubenswrapper[4988]: I1008 19:40:22.351075 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-69864b896c-pw7dl" event={"ID":"4babecb5-3754-4372-8322-6bdddbc06931","Type":"ContainerStarted","Data":"d1805129113610bb5b101c52737138b7d3b595ac13a200fb458d8b1748808654"} Oct 08 19:40:22 crc kubenswrapper[4988]: I1008 19:40:22.351133 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-69864b896c-pw7dl" Oct 08 19:40:22 crc kubenswrapper[4988]: I1008 19:40:22.353629 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-data-cert\" (UniqueName: \"kubernetes.io/secret/bc523d09-e017-4e91-8ba0-45ca290bd102-ovn-data-cert\") pod \"ovn-copy-data\" (UID: \"bc523d09-e017-4e91-8ba0-45ca290bd102\") " pod="openstack/ovn-copy-data" Oct 08 19:40:22 crc kubenswrapper[4988]: I1008 19:40:22.353652 4988 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Oct 08 19:40:22 crc kubenswrapper[4988]: I1008 19:40:22.353679 4988 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-c3809aad-6fef-4719-a8e9-1427666bd89a\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-c3809aad-6fef-4719-a8e9-1427666bd89a\") pod \"ovn-copy-data\" (UID: \"bc523d09-e017-4e91-8ba0-45ca290bd102\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/09b0bc38088ab07694ac79f341904bb028473a748d217ce4a6331bb9e3191de3/globalmount\"" pod="openstack/ovn-copy-data" Oct 08 19:40:22 crc kubenswrapper[4988]: I1008 19:40:22.356799 4988 generic.go:334] "Generic (PLEG): container finished" podID="2c6cd276-44be-4ae8-af3e-272844e7b27e" containerID="0c51531936f6f5c38ab873e494a3bd07271c87a721a0e4cc0f57a3d3466fb363" exitCode=0 Oct 08 19:40:22 crc kubenswrapper[4988]: I1008 19:40:22.356836 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6444955555-ssqb7" event={"ID":"2c6cd276-44be-4ae8-af3e-272844e7b27e","Type":"ContainerDied","Data":"0c51531936f6f5c38ab873e494a3bd07271c87a721a0e4cc0f57a3d3466fb363"} Oct 08 19:40:22 crc kubenswrapper[4988]: I1008 19:40:22.356863 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6444955555-ssqb7" event={"ID":"2c6cd276-44be-4ae8-af3e-272844e7b27e","Type":"ContainerDied","Data":"284b30656ca881193f912ed9b5033ff7015728ecbcfffb11ac4023d7fb4bf984"} Oct 08 19:40:22 crc kubenswrapper[4988]: I1008 19:40:22.356882 4988 scope.go:117] "RemoveContainer" containerID="0c51531936f6f5c38ab873e494a3bd07271c87a721a0e4cc0f57a3d3466fb363" Oct 08 19:40:22 crc kubenswrapper[4988]: I1008 19:40:22.357008 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6444955555-ssqb7" Oct 08 19:40:22 crc kubenswrapper[4988]: I1008 19:40:22.369756 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dfqbt\" (UniqueName: \"kubernetes.io/projected/bc523d09-e017-4e91-8ba0-45ca290bd102-kube-api-access-dfqbt\") pod \"ovn-copy-data\" (UID: \"bc523d09-e017-4e91-8ba0-45ca290bd102\") " pod="openstack/ovn-copy-data" Oct 08 19:40:22 crc kubenswrapper[4988]: I1008 19:40:22.371863 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-69864b896c-pw7dl" podStartSLOduration=3.371850389 podStartE2EDuration="3.371850389s" podCreationTimestamp="2025-10-08 19:40:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 19:40:22.36967798 +0000 UTC m=+5367.819520750" watchObservedRunningTime="2025-10-08 19:40:22.371850389 +0000 UTC m=+5367.821693159" Oct 08 19:40:22 crc kubenswrapper[4988]: I1008 19:40:22.397965 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-c3809aad-6fef-4719-a8e9-1427666bd89a\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-c3809aad-6fef-4719-a8e9-1427666bd89a\") pod \"ovn-copy-data\" (UID: \"bc523d09-e017-4e91-8ba0-45ca290bd102\") " pod="openstack/ovn-copy-data" Oct 08 19:40:22 crc kubenswrapper[4988]: I1008 19:40:22.415369 4988 scope.go:117] "RemoveContainer" containerID="00fef83e4152d6831ab20bd852c969e97aa6d96313b582774360b2a013cb1124" Oct 08 19:40:22 crc kubenswrapper[4988]: I1008 19:40:22.416882 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6444955555-ssqb7"] Oct 08 19:40:22 crc kubenswrapper[4988]: I1008 19:40:22.423275 4988 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-6444955555-ssqb7"] Oct 08 19:40:22 crc kubenswrapper[4988]: I1008 19:40:22.431080 4988 scope.go:117] "RemoveContainer" containerID="0c51531936f6f5c38ab873e494a3bd07271c87a721a0e4cc0f57a3d3466fb363" Oct 08 19:40:22 crc kubenswrapper[4988]: E1008 19:40:22.431497 4988 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0c51531936f6f5c38ab873e494a3bd07271c87a721a0e4cc0f57a3d3466fb363\": container with ID starting with 0c51531936f6f5c38ab873e494a3bd07271c87a721a0e4cc0f57a3d3466fb363 not found: ID does not exist" containerID="0c51531936f6f5c38ab873e494a3bd07271c87a721a0e4cc0f57a3d3466fb363" Oct 08 19:40:22 crc kubenswrapper[4988]: I1008 19:40:22.431525 4988 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0c51531936f6f5c38ab873e494a3bd07271c87a721a0e4cc0f57a3d3466fb363"} err="failed to get container status \"0c51531936f6f5c38ab873e494a3bd07271c87a721a0e4cc0f57a3d3466fb363\": rpc error: code = NotFound desc = could not find container \"0c51531936f6f5c38ab873e494a3bd07271c87a721a0e4cc0f57a3d3466fb363\": container with ID starting with 0c51531936f6f5c38ab873e494a3bd07271c87a721a0e4cc0f57a3d3466fb363 not found: ID does not exist" Oct 08 19:40:22 crc kubenswrapper[4988]: I1008 19:40:22.431544 4988 scope.go:117] "RemoveContainer" containerID="00fef83e4152d6831ab20bd852c969e97aa6d96313b582774360b2a013cb1124" Oct 08 19:40:22 crc kubenswrapper[4988]: E1008 19:40:22.431856 4988 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"00fef83e4152d6831ab20bd852c969e97aa6d96313b582774360b2a013cb1124\": container with ID starting with 00fef83e4152d6831ab20bd852c969e97aa6d96313b582774360b2a013cb1124 not found: ID does not exist" containerID="00fef83e4152d6831ab20bd852c969e97aa6d96313b582774360b2a013cb1124" Oct 08 19:40:22 crc kubenswrapper[4988]: I1008 19:40:22.431893 4988 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"00fef83e4152d6831ab20bd852c969e97aa6d96313b582774360b2a013cb1124"} err="failed to get container status \"00fef83e4152d6831ab20bd852c969e97aa6d96313b582774360b2a013cb1124\": rpc error: code = NotFound desc = could not find container \"00fef83e4152d6831ab20bd852c969e97aa6d96313b582774360b2a013cb1124\": container with ID starting with 00fef83e4152d6831ab20bd852c969e97aa6d96313b582774360b2a013cb1124 not found: ID does not exist" Oct 08 19:40:22 crc kubenswrapper[4988]: I1008 19:40:22.516043 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-copy-data" Oct 08 19:40:23 crc kubenswrapper[4988]: I1008 19:40:23.130897 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-copy-data"] Oct 08 19:40:23 crc kubenswrapper[4988]: W1008 19:40:23.139884 4988 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podbc523d09_e017_4e91_8ba0_45ca290bd102.slice/crio-6f022cf84222d9a831a3fdf675980d24133f09be02587a2014db51b0ae62dc24 WatchSource:0}: Error finding container 6f022cf84222d9a831a3fdf675980d24133f09be02587a2014db51b0ae62dc24: Status 404 returned error can't find the container with id 6f022cf84222d9a831a3fdf675980d24133f09be02587a2014db51b0ae62dc24 Oct 08 19:40:23 crc kubenswrapper[4988]: I1008 19:40:23.143521 4988 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 08 19:40:23 crc kubenswrapper[4988]: I1008 19:40:23.251041 4988 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2c6cd276-44be-4ae8-af3e-272844e7b27e" path="/var/lib/kubelet/pods/2c6cd276-44be-4ae8-af3e-272844e7b27e/volumes" Oct 08 19:40:23 crc kubenswrapper[4988]: I1008 19:40:23.367975 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-copy-data" event={"ID":"bc523d09-e017-4e91-8ba0-45ca290bd102","Type":"ContainerStarted","Data":"6f022cf84222d9a831a3fdf675980d24133f09be02587a2014db51b0ae62dc24"} Oct 08 19:40:24 crc kubenswrapper[4988]: I1008 19:40:24.381777 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-copy-data" event={"ID":"bc523d09-e017-4e91-8ba0-45ca290bd102","Type":"ContainerStarted","Data":"ec84a95c1676eda81cf66f0edcee8ec4cd2b7b456c53f4e63d241ae3a94c3fd2"} Oct 08 19:40:24 crc kubenswrapper[4988]: I1008 19:40:24.408620 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-copy-data" podStartSLOduration=2.855541502 podStartE2EDuration="3.408595843s" podCreationTimestamp="2025-10-08 19:40:21 +0000 UTC" firstStartedPulling="2025-10-08 19:40:23.142206817 +0000 UTC m=+5368.592049597" lastFinishedPulling="2025-10-08 19:40:23.695261168 +0000 UTC m=+5369.145103938" observedRunningTime="2025-10-08 19:40:24.405280347 +0000 UTC m=+5369.855123187" watchObservedRunningTime="2025-10-08 19:40:24.408595843 +0000 UTC m=+5369.858438653" Oct 08 19:40:24 crc kubenswrapper[4988]: I1008 19:40:24.666373 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-7mrpf"] Oct 08 19:40:24 crc kubenswrapper[4988]: I1008 19:40:24.670154 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-7mrpf" Oct 08 19:40:24 crc kubenswrapper[4988]: I1008 19:40:24.693966 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-7mrpf"] Oct 08 19:40:24 crc kubenswrapper[4988]: I1008 19:40:24.794704 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f011ae53-f461-4a61-a75e-1aca0a10910e-utilities\") pod \"redhat-operators-7mrpf\" (UID: \"f011ae53-f461-4a61-a75e-1aca0a10910e\") " pod="openshift-marketplace/redhat-operators-7mrpf" Oct 08 19:40:24 crc kubenswrapper[4988]: I1008 19:40:24.794939 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f011ae53-f461-4a61-a75e-1aca0a10910e-catalog-content\") pod \"redhat-operators-7mrpf\" (UID: \"f011ae53-f461-4a61-a75e-1aca0a10910e\") " pod="openshift-marketplace/redhat-operators-7mrpf" Oct 08 19:40:24 crc kubenswrapper[4988]: I1008 19:40:24.795037 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-84952\" (UniqueName: \"kubernetes.io/projected/f011ae53-f461-4a61-a75e-1aca0a10910e-kube-api-access-84952\") pod \"redhat-operators-7mrpf\" (UID: \"f011ae53-f461-4a61-a75e-1aca0a10910e\") " pod="openshift-marketplace/redhat-operators-7mrpf" Oct 08 19:40:24 crc kubenswrapper[4988]: I1008 19:40:24.896769 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f011ae53-f461-4a61-a75e-1aca0a10910e-catalog-content\") pod \"redhat-operators-7mrpf\" (UID: \"f011ae53-f461-4a61-a75e-1aca0a10910e\") " pod="openshift-marketplace/redhat-operators-7mrpf" Oct 08 19:40:24 crc kubenswrapper[4988]: I1008 19:40:24.896876 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-84952\" (UniqueName: \"kubernetes.io/projected/f011ae53-f461-4a61-a75e-1aca0a10910e-kube-api-access-84952\") pod \"redhat-operators-7mrpf\" (UID: \"f011ae53-f461-4a61-a75e-1aca0a10910e\") " pod="openshift-marketplace/redhat-operators-7mrpf" Oct 08 19:40:24 crc kubenswrapper[4988]: I1008 19:40:24.896943 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f011ae53-f461-4a61-a75e-1aca0a10910e-utilities\") pod \"redhat-operators-7mrpf\" (UID: \"f011ae53-f461-4a61-a75e-1aca0a10910e\") " pod="openshift-marketplace/redhat-operators-7mrpf" Oct 08 19:40:24 crc kubenswrapper[4988]: I1008 19:40:24.897480 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f011ae53-f461-4a61-a75e-1aca0a10910e-catalog-content\") pod \"redhat-operators-7mrpf\" (UID: \"f011ae53-f461-4a61-a75e-1aca0a10910e\") " pod="openshift-marketplace/redhat-operators-7mrpf" Oct 08 19:40:24 crc kubenswrapper[4988]: I1008 19:40:24.897510 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f011ae53-f461-4a61-a75e-1aca0a10910e-utilities\") pod \"redhat-operators-7mrpf\" (UID: \"f011ae53-f461-4a61-a75e-1aca0a10910e\") " pod="openshift-marketplace/redhat-operators-7mrpf" Oct 08 19:40:24 crc kubenswrapper[4988]: I1008 19:40:24.928943 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-84952\" (UniqueName: \"kubernetes.io/projected/f011ae53-f461-4a61-a75e-1aca0a10910e-kube-api-access-84952\") pod \"redhat-operators-7mrpf\" (UID: \"f011ae53-f461-4a61-a75e-1aca0a10910e\") " pod="openshift-marketplace/redhat-operators-7mrpf" Oct 08 19:40:25 crc kubenswrapper[4988]: I1008 19:40:25.005200 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-7mrpf" Oct 08 19:40:25 crc kubenswrapper[4988]: I1008 19:40:25.454307 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-7mrpf"] Oct 08 19:40:25 crc kubenswrapper[4988]: W1008 19:40:25.458996 4988 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf011ae53_f461_4a61_a75e_1aca0a10910e.slice/crio-169ce692280b4823f18bda9386a88f7af11e41925ba75cf116a2c481c416e39c WatchSource:0}: Error finding container 169ce692280b4823f18bda9386a88f7af11e41925ba75cf116a2c481c416e39c: Status 404 returned error can't find the container with id 169ce692280b4823f18bda9386a88f7af11e41925ba75cf116a2c481c416e39c Oct 08 19:40:26 crc kubenswrapper[4988]: I1008 19:40:26.420820 4988 generic.go:334] "Generic (PLEG): container finished" podID="f011ae53-f461-4a61-a75e-1aca0a10910e" containerID="12c32255518dddc429f60e1433d5ed358d7cb532bf4b64b275a52763db6b9b3f" exitCode=0 Oct 08 19:40:26 crc kubenswrapper[4988]: I1008 19:40:26.420919 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7mrpf" event={"ID":"f011ae53-f461-4a61-a75e-1aca0a10910e","Type":"ContainerDied","Data":"12c32255518dddc429f60e1433d5ed358d7cb532bf4b64b275a52763db6b9b3f"} Oct 08 19:40:26 crc kubenswrapper[4988]: I1008 19:40:26.421096 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7mrpf" event={"ID":"f011ae53-f461-4a61-a75e-1aca0a10910e","Type":"ContainerStarted","Data":"169ce692280b4823f18bda9386a88f7af11e41925ba75cf116a2c481c416e39c"} Oct 08 19:40:28 crc kubenswrapper[4988]: I1008 19:40:28.447487 4988 generic.go:334] "Generic (PLEG): container finished" podID="f011ae53-f461-4a61-a75e-1aca0a10910e" containerID="541c0d85840e9497ac964a70f86a258c1f597dbfb4cd0ae3f150ea0239e0e76e" exitCode=0 Oct 08 19:40:28 crc kubenswrapper[4988]: I1008 19:40:28.447565 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7mrpf" event={"ID":"f011ae53-f461-4a61-a75e-1aca0a10910e","Type":"ContainerDied","Data":"541c0d85840e9497ac964a70f86a258c1f597dbfb4cd0ae3f150ea0239e0e76e"} Oct 08 19:40:29 crc kubenswrapper[4988]: I1008 19:40:29.461342 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7mrpf" event={"ID":"f011ae53-f461-4a61-a75e-1aca0a10910e","Type":"ContainerStarted","Data":"79840c34b6e676b7f4b6e987238e6c0e33867561a11a83478d01694d3ddf24ed"} Oct 08 19:40:29 crc kubenswrapper[4988]: I1008 19:40:29.483661 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-7mrpf" podStartSLOduration=2.855195005 podStartE2EDuration="5.483624677s" podCreationTimestamp="2025-10-08 19:40:24 +0000 UTC" firstStartedPulling="2025-10-08 19:40:26.423544424 +0000 UTC m=+5371.873387194" lastFinishedPulling="2025-10-08 19:40:29.051974056 +0000 UTC m=+5374.501816866" observedRunningTime="2025-10-08 19:40:29.480638582 +0000 UTC m=+5374.930481382" watchObservedRunningTime="2025-10-08 19:40:29.483624677 +0000 UTC m=+5374.933467477" Oct 08 19:40:29 crc kubenswrapper[4988]: I1008 19:40:29.765293 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-northd-0"] Oct 08 19:40:29 crc kubenswrapper[4988]: I1008 19:40:29.769242 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Oct 08 19:40:29 crc kubenswrapper[4988]: I1008 19:40:29.774356 4988 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-scripts" Oct 08 19:40:29 crc kubenswrapper[4988]: I1008 19:40:29.774666 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovnnorthd-ovndbs" Oct 08 19:40:29 crc kubenswrapper[4988]: I1008 19:40:29.774911 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovnnorthd-ovnnorthd-dockercfg-txbz5" Oct 08 19:40:29 crc kubenswrapper[4988]: I1008 19:40:29.783285 4988 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-config" Oct 08 19:40:29 crc kubenswrapper[4988]: I1008 19:40:29.784260 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/442d8773-cc55-4b79-945a-7d8c05431385-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"442d8773-cc55-4b79-945a-7d8c05431385\") " pod="openstack/ovn-northd-0" Oct 08 19:40:29 crc kubenswrapper[4988]: I1008 19:40:29.784307 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/442d8773-cc55-4b79-945a-7d8c05431385-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"442d8773-cc55-4b79-945a-7d8c05431385\") " pod="openstack/ovn-northd-0" Oct 08 19:40:29 crc kubenswrapper[4988]: I1008 19:40:29.784354 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/442d8773-cc55-4b79-945a-7d8c05431385-config\") pod \"ovn-northd-0\" (UID: \"442d8773-cc55-4b79-945a-7d8c05431385\") " pod="openstack/ovn-northd-0" Oct 08 19:40:29 crc kubenswrapper[4988]: I1008 19:40:29.784380 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/442d8773-cc55-4b79-945a-7d8c05431385-scripts\") pod \"ovn-northd-0\" (UID: \"442d8773-cc55-4b79-945a-7d8c05431385\") " pod="openstack/ovn-northd-0" Oct 08 19:40:29 crc kubenswrapper[4988]: I1008 19:40:29.784430 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/442d8773-cc55-4b79-945a-7d8c05431385-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"442d8773-cc55-4b79-945a-7d8c05431385\") " pod="openstack/ovn-northd-0" Oct 08 19:40:29 crc kubenswrapper[4988]: I1008 19:40:29.784473 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/442d8773-cc55-4b79-945a-7d8c05431385-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"442d8773-cc55-4b79-945a-7d8c05431385\") " pod="openstack/ovn-northd-0" Oct 08 19:40:29 crc kubenswrapper[4988]: I1008 19:40:29.784500 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9mbk2\" (UniqueName: \"kubernetes.io/projected/442d8773-cc55-4b79-945a-7d8c05431385-kube-api-access-9mbk2\") pod \"ovn-northd-0\" (UID: \"442d8773-cc55-4b79-945a-7d8c05431385\") " pod="openstack/ovn-northd-0" Oct 08 19:40:29 crc kubenswrapper[4988]: I1008 19:40:29.792878 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Oct 08 19:40:29 crc kubenswrapper[4988]: I1008 19:40:29.885642 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/442d8773-cc55-4b79-945a-7d8c05431385-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"442d8773-cc55-4b79-945a-7d8c05431385\") " pod="openstack/ovn-northd-0" Oct 08 19:40:29 crc kubenswrapper[4988]: I1008 19:40:29.885700 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9mbk2\" (UniqueName: \"kubernetes.io/projected/442d8773-cc55-4b79-945a-7d8c05431385-kube-api-access-9mbk2\") pod \"ovn-northd-0\" (UID: \"442d8773-cc55-4b79-945a-7d8c05431385\") " pod="openstack/ovn-northd-0" Oct 08 19:40:29 crc kubenswrapper[4988]: I1008 19:40:29.885737 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/442d8773-cc55-4b79-945a-7d8c05431385-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"442d8773-cc55-4b79-945a-7d8c05431385\") " pod="openstack/ovn-northd-0" Oct 08 19:40:29 crc kubenswrapper[4988]: I1008 19:40:29.885766 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/442d8773-cc55-4b79-945a-7d8c05431385-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"442d8773-cc55-4b79-945a-7d8c05431385\") " pod="openstack/ovn-northd-0" Oct 08 19:40:29 crc kubenswrapper[4988]: I1008 19:40:29.885796 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/442d8773-cc55-4b79-945a-7d8c05431385-config\") pod \"ovn-northd-0\" (UID: \"442d8773-cc55-4b79-945a-7d8c05431385\") " pod="openstack/ovn-northd-0" Oct 08 19:40:29 crc kubenswrapper[4988]: I1008 19:40:29.885823 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/442d8773-cc55-4b79-945a-7d8c05431385-scripts\") pod \"ovn-northd-0\" (UID: \"442d8773-cc55-4b79-945a-7d8c05431385\") " pod="openstack/ovn-northd-0" Oct 08 19:40:29 crc kubenswrapper[4988]: I1008 19:40:29.885854 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/442d8773-cc55-4b79-945a-7d8c05431385-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"442d8773-cc55-4b79-945a-7d8c05431385\") " pod="openstack/ovn-northd-0" Oct 08 19:40:29 crc kubenswrapper[4988]: I1008 19:40:29.888077 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/442d8773-cc55-4b79-945a-7d8c05431385-config\") pod \"ovn-northd-0\" (UID: \"442d8773-cc55-4b79-945a-7d8c05431385\") " pod="openstack/ovn-northd-0" Oct 08 19:40:29 crc kubenswrapper[4988]: I1008 19:40:29.888102 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/442d8773-cc55-4b79-945a-7d8c05431385-scripts\") pod \"ovn-northd-0\" (UID: \"442d8773-cc55-4b79-945a-7d8c05431385\") " pod="openstack/ovn-northd-0" Oct 08 19:40:29 crc kubenswrapper[4988]: I1008 19:40:29.888366 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/442d8773-cc55-4b79-945a-7d8c05431385-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"442d8773-cc55-4b79-945a-7d8c05431385\") " pod="openstack/ovn-northd-0" Oct 08 19:40:29 crc kubenswrapper[4988]: I1008 19:40:29.893149 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/442d8773-cc55-4b79-945a-7d8c05431385-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"442d8773-cc55-4b79-945a-7d8c05431385\") " pod="openstack/ovn-northd-0" Oct 08 19:40:29 crc kubenswrapper[4988]: I1008 19:40:29.893708 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/442d8773-cc55-4b79-945a-7d8c05431385-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"442d8773-cc55-4b79-945a-7d8c05431385\") " pod="openstack/ovn-northd-0" Oct 08 19:40:29 crc kubenswrapper[4988]: I1008 19:40:29.897092 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/442d8773-cc55-4b79-945a-7d8c05431385-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"442d8773-cc55-4b79-945a-7d8c05431385\") " pod="openstack/ovn-northd-0" Oct 08 19:40:29 crc kubenswrapper[4988]: I1008 19:40:29.905373 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9mbk2\" (UniqueName: \"kubernetes.io/projected/442d8773-cc55-4b79-945a-7d8c05431385-kube-api-access-9mbk2\") pod \"ovn-northd-0\" (UID: \"442d8773-cc55-4b79-945a-7d8c05431385\") " pod="openstack/ovn-northd-0" Oct 08 19:40:30 crc kubenswrapper[4988]: I1008 19:40:30.103837 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Oct 08 19:40:30 crc kubenswrapper[4988]: I1008 19:40:30.214262 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-69864b896c-pw7dl" Oct 08 19:40:30 crc kubenswrapper[4988]: I1008 19:40:30.273027 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5fdc957c47-mtggm"] Oct 08 19:40:30 crc kubenswrapper[4988]: I1008 19:40:30.273300 4988 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-5fdc957c47-mtggm" podUID="14aa686b-2873-45e4-8c67-ae59733b51a6" containerName="dnsmasq-dns" containerID="cri-o://1b9e6f6534d0081b7d90abb344b16fbd0a10f4c080bd83ee4f8bb225f6645fe3" gracePeriod=10 Oct 08 19:40:30 crc kubenswrapper[4988]: I1008 19:40:30.478248 4988 generic.go:334] "Generic (PLEG): container finished" podID="14aa686b-2873-45e4-8c67-ae59733b51a6" containerID="1b9e6f6534d0081b7d90abb344b16fbd0a10f4c080bd83ee4f8bb225f6645fe3" exitCode=0 Oct 08 19:40:30 crc kubenswrapper[4988]: I1008 19:40:30.478330 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5fdc957c47-mtggm" event={"ID":"14aa686b-2873-45e4-8c67-ae59733b51a6","Type":"ContainerDied","Data":"1b9e6f6534d0081b7d90abb344b16fbd0a10f4c080bd83ee4f8bb225f6645fe3"} Oct 08 19:40:30 crc kubenswrapper[4988]: I1008 19:40:30.613188 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Oct 08 19:40:30 crc kubenswrapper[4988]: I1008 19:40:30.745002 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5fdc957c47-mtggm" Oct 08 19:40:30 crc kubenswrapper[4988]: I1008 19:40:30.803536 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/14aa686b-2873-45e4-8c67-ae59733b51a6-dns-svc\") pod \"14aa686b-2873-45e4-8c67-ae59733b51a6\" (UID: \"14aa686b-2873-45e4-8c67-ae59733b51a6\") " Oct 08 19:40:30 crc kubenswrapper[4988]: I1008 19:40:30.803601 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/14aa686b-2873-45e4-8c67-ae59733b51a6-config\") pod \"14aa686b-2873-45e4-8c67-ae59733b51a6\" (UID: \"14aa686b-2873-45e4-8c67-ae59733b51a6\") " Oct 08 19:40:30 crc kubenswrapper[4988]: I1008 19:40:30.803634 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zqckg\" (UniqueName: \"kubernetes.io/projected/14aa686b-2873-45e4-8c67-ae59733b51a6-kube-api-access-zqckg\") pod \"14aa686b-2873-45e4-8c67-ae59733b51a6\" (UID: \"14aa686b-2873-45e4-8c67-ae59733b51a6\") " Oct 08 19:40:30 crc kubenswrapper[4988]: I1008 19:40:30.808677 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/14aa686b-2873-45e4-8c67-ae59733b51a6-kube-api-access-zqckg" (OuterVolumeSpecName: "kube-api-access-zqckg") pod "14aa686b-2873-45e4-8c67-ae59733b51a6" (UID: "14aa686b-2873-45e4-8c67-ae59733b51a6"). InnerVolumeSpecName "kube-api-access-zqckg". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 19:40:30 crc kubenswrapper[4988]: I1008 19:40:30.847438 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/14aa686b-2873-45e4-8c67-ae59733b51a6-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "14aa686b-2873-45e4-8c67-ae59733b51a6" (UID: "14aa686b-2873-45e4-8c67-ae59733b51a6"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 19:40:30 crc kubenswrapper[4988]: I1008 19:40:30.852245 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/14aa686b-2873-45e4-8c67-ae59733b51a6-config" (OuterVolumeSpecName: "config") pod "14aa686b-2873-45e4-8c67-ae59733b51a6" (UID: "14aa686b-2873-45e4-8c67-ae59733b51a6"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 19:40:30 crc kubenswrapper[4988]: I1008 19:40:30.905230 4988 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/14aa686b-2873-45e4-8c67-ae59733b51a6-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 08 19:40:30 crc kubenswrapper[4988]: I1008 19:40:30.905368 4988 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/14aa686b-2873-45e4-8c67-ae59733b51a6-config\") on node \"crc\" DevicePath \"\"" Oct 08 19:40:30 crc kubenswrapper[4988]: I1008 19:40:30.905497 4988 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zqckg\" (UniqueName: \"kubernetes.io/projected/14aa686b-2873-45e4-8c67-ae59733b51a6-kube-api-access-zqckg\") on node \"crc\" DevicePath \"\"" Oct 08 19:40:31 crc kubenswrapper[4988]: I1008 19:40:31.490638 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5fdc957c47-mtggm" event={"ID":"14aa686b-2873-45e4-8c67-ae59733b51a6","Type":"ContainerDied","Data":"db28ff1ba831b237c6929dd2b10e2395e9578409ebf0ed76ff11f73956df8c82"} Oct 08 19:40:31 crc kubenswrapper[4988]: I1008 19:40:31.490691 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5fdc957c47-mtggm" Oct 08 19:40:31 crc kubenswrapper[4988]: I1008 19:40:31.490722 4988 scope.go:117] "RemoveContainer" containerID="1b9e6f6534d0081b7d90abb344b16fbd0a10f4c080bd83ee4f8bb225f6645fe3" Oct 08 19:40:31 crc kubenswrapper[4988]: I1008 19:40:31.494357 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"442d8773-cc55-4b79-945a-7d8c05431385","Type":"ContainerStarted","Data":"f2c46c9153e07bf62e1d9138caa17658a2bb66a55f79523a71cf47cbb8f016df"} Oct 08 19:40:31 crc kubenswrapper[4988]: I1008 19:40:31.494746 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"442d8773-cc55-4b79-945a-7d8c05431385","Type":"ContainerStarted","Data":"f5158bda801dd954ecc406cf259884ab07fbcd1e997ddb3e5cc8669e697a33e8"} Oct 08 19:40:31 crc kubenswrapper[4988]: I1008 19:40:31.494764 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"442d8773-cc55-4b79-945a-7d8c05431385","Type":"ContainerStarted","Data":"6d546428b477846a6d9827969f6e589d00f8c7047968886d42c369001ad23e19"} Oct 08 19:40:31 crc kubenswrapper[4988]: I1008 19:40:31.494948 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-northd-0" Oct 08 19:40:31 crc kubenswrapper[4988]: I1008 19:40:31.521193 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-northd-0" podStartSLOduration=2.521171096 podStartE2EDuration="2.521171096s" podCreationTimestamp="2025-10-08 19:40:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 19:40:31.518051447 +0000 UTC m=+5376.967894247" watchObservedRunningTime="2025-10-08 19:40:31.521171096 +0000 UTC m=+5376.971013906" Oct 08 19:40:31 crc kubenswrapper[4988]: I1008 19:40:31.532208 4988 scope.go:117] "RemoveContainer" containerID="11f2c8c855b5ae429ee4690e77070660f55fed41beb95245096e21790bb4d327" Oct 08 19:40:31 crc kubenswrapper[4988]: I1008 19:40:31.551522 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5fdc957c47-mtggm"] Oct 08 19:40:31 crc kubenswrapper[4988]: I1008 19:40:31.561343 4988 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5fdc957c47-mtggm"] Oct 08 19:40:33 crc kubenswrapper[4988]: I1008 19:40:33.253564 4988 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="14aa686b-2873-45e4-8c67-ae59733b51a6" path="/var/lib/kubelet/pods/14aa686b-2873-45e4-8c67-ae59733b51a6/volumes" Oct 08 19:40:35 crc kubenswrapper[4988]: I1008 19:40:35.005353 4988 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-7mrpf" Oct 08 19:40:35 crc kubenswrapper[4988]: I1008 19:40:35.005663 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-7mrpf" Oct 08 19:40:35 crc kubenswrapper[4988]: I1008 19:40:35.085123 4988 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-7mrpf" Oct 08 19:40:35 crc kubenswrapper[4988]: I1008 19:40:35.348756 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-create-zk7zh"] Oct 08 19:40:35 crc kubenswrapper[4988]: E1008 19:40:35.349302 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="14aa686b-2873-45e4-8c67-ae59733b51a6" containerName="dnsmasq-dns" Oct 08 19:40:35 crc kubenswrapper[4988]: I1008 19:40:35.349329 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="14aa686b-2873-45e4-8c67-ae59733b51a6" containerName="dnsmasq-dns" Oct 08 19:40:35 crc kubenswrapper[4988]: E1008 19:40:35.349365 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="14aa686b-2873-45e4-8c67-ae59733b51a6" containerName="init" Oct 08 19:40:35 crc kubenswrapper[4988]: I1008 19:40:35.349378 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="14aa686b-2873-45e4-8c67-ae59733b51a6" containerName="init" Oct 08 19:40:35 crc kubenswrapper[4988]: I1008 19:40:35.349661 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="14aa686b-2873-45e4-8c67-ae59733b51a6" containerName="dnsmasq-dns" Oct 08 19:40:35 crc kubenswrapper[4988]: I1008 19:40:35.350600 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-zk7zh" Oct 08 19:40:35 crc kubenswrapper[4988]: I1008 19:40:35.356338 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-zk7zh"] Oct 08 19:40:35 crc kubenswrapper[4988]: I1008 19:40:35.392426 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vdw7v\" (UniqueName: \"kubernetes.io/projected/e6a11d4a-8b71-47bf-a169-f814d7c5455c-kube-api-access-vdw7v\") pod \"keystone-db-create-zk7zh\" (UID: \"e6a11d4a-8b71-47bf-a169-f814d7c5455c\") " pod="openstack/keystone-db-create-zk7zh" Oct 08 19:40:35 crc kubenswrapper[4988]: I1008 19:40:35.493761 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vdw7v\" (UniqueName: \"kubernetes.io/projected/e6a11d4a-8b71-47bf-a169-f814d7c5455c-kube-api-access-vdw7v\") pod \"keystone-db-create-zk7zh\" (UID: \"e6a11d4a-8b71-47bf-a169-f814d7c5455c\") " pod="openstack/keystone-db-create-zk7zh" Oct 08 19:40:35 crc kubenswrapper[4988]: I1008 19:40:35.520477 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vdw7v\" (UniqueName: \"kubernetes.io/projected/e6a11d4a-8b71-47bf-a169-f814d7c5455c-kube-api-access-vdw7v\") pod \"keystone-db-create-zk7zh\" (UID: \"e6a11d4a-8b71-47bf-a169-f814d7c5455c\") " pod="openstack/keystone-db-create-zk7zh" Oct 08 19:40:35 crc kubenswrapper[4988]: I1008 19:40:35.595100 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-7mrpf" Oct 08 19:40:35 crc kubenswrapper[4988]: I1008 19:40:35.653153 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-7mrpf"] Oct 08 19:40:35 crc kubenswrapper[4988]: I1008 19:40:35.695512 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-zk7zh" Oct 08 19:40:36 crc kubenswrapper[4988]: I1008 19:40:36.171341 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-zk7zh"] Oct 08 19:40:36 crc kubenswrapper[4988]: W1008 19:40:36.173042 4988 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode6a11d4a_8b71_47bf_a169_f814d7c5455c.slice/crio-2697305fa30740c4ed2df5b0d1ab5df9b7193647fc800a67b584df073c7738df WatchSource:0}: Error finding container 2697305fa30740c4ed2df5b0d1ab5df9b7193647fc800a67b584df073c7738df: Status 404 returned error can't find the container with id 2697305fa30740c4ed2df5b0d1ab5df9b7193647fc800a67b584df073c7738df Oct 08 19:40:36 crc kubenswrapper[4988]: I1008 19:40:36.555117 4988 generic.go:334] "Generic (PLEG): container finished" podID="e6a11d4a-8b71-47bf-a169-f814d7c5455c" containerID="24b1cee3957790a1ad65dc1e2719eba944956a68854fff075295d74d45fe7f8d" exitCode=0 Oct 08 19:40:36 crc kubenswrapper[4988]: I1008 19:40:36.555244 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-zk7zh" event={"ID":"e6a11d4a-8b71-47bf-a169-f814d7c5455c","Type":"ContainerDied","Data":"24b1cee3957790a1ad65dc1e2719eba944956a68854fff075295d74d45fe7f8d"} Oct 08 19:40:36 crc kubenswrapper[4988]: I1008 19:40:36.555327 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-zk7zh" event={"ID":"e6a11d4a-8b71-47bf-a169-f814d7c5455c","Type":"ContainerStarted","Data":"2697305fa30740c4ed2df5b0d1ab5df9b7193647fc800a67b584df073c7738df"} Oct 08 19:40:37 crc kubenswrapper[4988]: I1008 19:40:37.566188 4988 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-7mrpf" podUID="f011ae53-f461-4a61-a75e-1aca0a10910e" containerName="registry-server" containerID="cri-o://79840c34b6e676b7f4b6e987238e6c0e33867561a11a83478d01694d3ddf24ed" gracePeriod=2 Oct 08 19:40:37 crc kubenswrapper[4988]: I1008 19:40:37.962075 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-zk7zh" Oct 08 19:40:38 crc kubenswrapper[4988]: I1008 19:40:38.049001 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vdw7v\" (UniqueName: \"kubernetes.io/projected/e6a11d4a-8b71-47bf-a169-f814d7c5455c-kube-api-access-vdw7v\") pod \"e6a11d4a-8b71-47bf-a169-f814d7c5455c\" (UID: \"e6a11d4a-8b71-47bf-a169-f814d7c5455c\") " Oct 08 19:40:38 crc kubenswrapper[4988]: I1008 19:40:38.051222 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-7mrpf" Oct 08 19:40:38 crc kubenswrapper[4988]: I1008 19:40:38.055946 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e6a11d4a-8b71-47bf-a169-f814d7c5455c-kube-api-access-vdw7v" (OuterVolumeSpecName: "kube-api-access-vdw7v") pod "e6a11d4a-8b71-47bf-a169-f814d7c5455c" (UID: "e6a11d4a-8b71-47bf-a169-f814d7c5455c"). InnerVolumeSpecName "kube-api-access-vdw7v". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 19:40:38 crc kubenswrapper[4988]: I1008 19:40:38.151898 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f011ae53-f461-4a61-a75e-1aca0a10910e-utilities\") pod \"f011ae53-f461-4a61-a75e-1aca0a10910e\" (UID: \"f011ae53-f461-4a61-a75e-1aca0a10910e\") " Oct 08 19:40:38 crc kubenswrapper[4988]: I1008 19:40:38.152081 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f011ae53-f461-4a61-a75e-1aca0a10910e-catalog-content\") pod \"f011ae53-f461-4a61-a75e-1aca0a10910e\" (UID: \"f011ae53-f461-4a61-a75e-1aca0a10910e\") " Oct 08 19:40:38 crc kubenswrapper[4988]: I1008 19:40:38.152129 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-84952\" (UniqueName: \"kubernetes.io/projected/f011ae53-f461-4a61-a75e-1aca0a10910e-kube-api-access-84952\") pod \"f011ae53-f461-4a61-a75e-1aca0a10910e\" (UID: \"f011ae53-f461-4a61-a75e-1aca0a10910e\") " Oct 08 19:40:38 crc kubenswrapper[4988]: I1008 19:40:38.152480 4988 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vdw7v\" (UniqueName: \"kubernetes.io/projected/e6a11d4a-8b71-47bf-a169-f814d7c5455c-kube-api-access-vdw7v\") on node \"crc\" DevicePath \"\"" Oct 08 19:40:38 crc kubenswrapper[4988]: I1008 19:40:38.153189 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f011ae53-f461-4a61-a75e-1aca0a10910e-utilities" (OuterVolumeSpecName: "utilities") pod "f011ae53-f461-4a61-a75e-1aca0a10910e" (UID: "f011ae53-f461-4a61-a75e-1aca0a10910e"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 19:40:38 crc kubenswrapper[4988]: I1008 19:40:38.155403 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f011ae53-f461-4a61-a75e-1aca0a10910e-kube-api-access-84952" (OuterVolumeSpecName: "kube-api-access-84952") pod "f011ae53-f461-4a61-a75e-1aca0a10910e" (UID: "f011ae53-f461-4a61-a75e-1aca0a10910e"). InnerVolumeSpecName "kube-api-access-84952". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 19:40:38 crc kubenswrapper[4988]: I1008 19:40:38.253612 4988 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-84952\" (UniqueName: \"kubernetes.io/projected/f011ae53-f461-4a61-a75e-1aca0a10910e-kube-api-access-84952\") on node \"crc\" DevicePath \"\"" Oct 08 19:40:38 crc kubenswrapper[4988]: I1008 19:40:38.253649 4988 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f011ae53-f461-4a61-a75e-1aca0a10910e-utilities\") on node \"crc\" DevicePath \"\"" Oct 08 19:40:38 crc kubenswrapper[4988]: I1008 19:40:38.581104 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-zk7zh" event={"ID":"e6a11d4a-8b71-47bf-a169-f814d7c5455c","Type":"ContainerDied","Data":"2697305fa30740c4ed2df5b0d1ab5df9b7193647fc800a67b584df073c7738df"} Oct 08 19:40:38 crc kubenswrapper[4988]: I1008 19:40:38.581154 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-zk7zh" Oct 08 19:40:38 crc kubenswrapper[4988]: I1008 19:40:38.581172 4988 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2697305fa30740c4ed2df5b0d1ab5df9b7193647fc800a67b584df073c7738df" Oct 08 19:40:38 crc kubenswrapper[4988]: I1008 19:40:38.585451 4988 generic.go:334] "Generic (PLEG): container finished" podID="f011ae53-f461-4a61-a75e-1aca0a10910e" containerID="79840c34b6e676b7f4b6e987238e6c0e33867561a11a83478d01694d3ddf24ed" exitCode=0 Oct 08 19:40:38 crc kubenswrapper[4988]: I1008 19:40:38.585512 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7mrpf" event={"ID":"f011ae53-f461-4a61-a75e-1aca0a10910e","Type":"ContainerDied","Data":"79840c34b6e676b7f4b6e987238e6c0e33867561a11a83478d01694d3ddf24ed"} Oct 08 19:40:38 crc kubenswrapper[4988]: I1008 19:40:38.585551 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7mrpf" event={"ID":"f011ae53-f461-4a61-a75e-1aca0a10910e","Type":"ContainerDied","Data":"169ce692280b4823f18bda9386a88f7af11e41925ba75cf116a2c481c416e39c"} Oct 08 19:40:38 crc kubenswrapper[4988]: I1008 19:40:38.585556 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-7mrpf" Oct 08 19:40:38 crc kubenswrapper[4988]: I1008 19:40:38.585580 4988 scope.go:117] "RemoveContainer" containerID="79840c34b6e676b7f4b6e987238e6c0e33867561a11a83478d01694d3ddf24ed" Oct 08 19:40:38 crc kubenswrapper[4988]: I1008 19:40:38.623838 4988 scope.go:117] "RemoveContainer" containerID="541c0d85840e9497ac964a70f86a258c1f597dbfb4cd0ae3f150ea0239e0e76e" Oct 08 19:40:38 crc kubenswrapper[4988]: I1008 19:40:38.662129 4988 scope.go:117] "RemoveContainer" containerID="12c32255518dddc429f60e1433d5ed358d7cb532bf4b64b275a52763db6b9b3f" Oct 08 19:40:38 crc kubenswrapper[4988]: I1008 19:40:38.686502 4988 scope.go:117] "RemoveContainer" containerID="79840c34b6e676b7f4b6e987238e6c0e33867561a11a83478d01694d3ddf24ed" Oct 08 19:40:38 crc kubenswrapper[4988]: E1008 19:40:38.687081 4988 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"79840c34b6e676b7f4b6e987238e6c0e33867561a11a83478d01694d3ddf24ed\": container with ID starting with 79840c34b6e676b7f4b6e987238e6c0e33867561a11a83478d01694d3ddf24ed not found: ID does not exist" containerID="79840c34b6e676b7f4b6e987238e6c0e33867561a11a83478d01694d3ddf24ed" Oct 08 19:40:38 crc kubenswrapper[4988]: I1008 19:40:38.687143 4988 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"79840c34b6e676b7f4b6e987238e6c0e33867561a11a83478d01694d3ddf24ed"} err="failed to get container status \"79840c34b6e676b7f4b6e987238e6c0e33867561a11a83478d01694d3ddf24ed\": rpc error: code = NotFound desc = could not find container \"79840c34b6e676b7f4b6e987238e6c0e33867561a11a83478d01694d3ddf24ed\": container with ID starting with 79840c34b6e676b7f4b6e987238e6c0e33867561a11a83478d01694d3ddf24ed not found: ID does not exist" Oct 08 19:40:38 crc kubenswrapper[4988]: I1008 19:40:38.687182 4988 scope.go:117] "RemoveContainer" containerID="541c0d85840e9497ac964a70f86a258c1f597dbfb4cd0ae3f150ea0239e0e76e" Oct 08 19:40:38 crc kubenswrapper[4988]: E1008 19:40:38.687840 4988 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"541c0d85840e9497ac964a70f86a258c1f597dbfb4cd0ae3f150ea0239e0e76e\": container with ID starting with 541c0d85840e9497ac964a70f86a258c1f597dbfb4cd0ae3f150ea0239e0e76e not found: ID does not exist" containerID="541c0d85840e9497ac964a70f86a258c1f597dbfb4cd0ae3f150ea0239e0e76e" Oct 08 19:40:38 crc kubenswrapper[4988]: I1008 19:40:38.687893 4988 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"541c0d85840e9497ac964a70f86a258c1f597dbfb4cd0ae3f150ea0239e0e76e"} err="failed to get container status \"541c0d85840e9497ac964a70f86a258c1f597dbfb4cd0ae3f150ea0239e0e76e\": rpc error: code = NotFound desc = could not find container \"541c0d85840e9497ac964a70f86a258c1f597dbfb4cd0ae3f150ea0239e0e76e\": container with ID starting with 541c0d85840e9497ac964a70f86a258c1f597dbfb4cd0ae3f150ea0239e0e76e not found: ID does not exist" Oct 08 19:40:38 crc kubenswrapper[4988]: I1008 19:40:38.687925 4988 scope.go:117] "RemoveContainer" containerID="12c32255518dddc429f60e1433d5ed358d7cb532bf4b64b275a52763db6b9b3f" Oct 08 19:40:38 crc kubenswrapper[4988]: E1008 19:40:38.688233 4988 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"12c32255518dddc429f60e1433d5ed358d7cb532bf4b64b275a52763db6b9b3f\": container with ID starting with 12c32255518dddc429f60e1433d5ed358d7cb532bf4b64b275a52763db6b9b3f not found: ID does not exist" containerID="12c32255518dddc429f60e1433d5ed358d7cb532bf4b64b275a52763db6b9b3f" Oct 08 19:40:38 crc kubenswrapper[4988]: I1008 19:40:38.688291 4988 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"12c32255518dddc429f60e1433d5ed358d7cb532bf4b64b275a52763db6b9b3f"} err="failed to get container status \"12c32255518dddc429f60e1433d5ed358d7cb532bf4b64b275a52763db6b9b3f\": rpc error: code = NotFound desc = could not find container \"12c32255518dddc429f60e1433d5ed358d7cb532bf4b64b275a52763db6b9b3f\": container with ID starting with 12c32255518dddc429f60e1433d5ed358d7cb532bf4b64b275a52763db6b9b3f not found: ID does not exist" Oct 08 19:40:39 crc kubenswrapper[4988]: I1008 19:40:39.220961 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f011ae53-f461-4a61-a75e-1aca0a10910e-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "f011ae53-f461-4a61-a75e-1aca0a10910e" (UID: "f011ae53-f461-4a61-a75e-1aca0a10910e"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 19:40:39 crc kubenswrapper[4988]: I1008 19:40:39.273048 4988 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f011ae53-f461-4a61-a75e-1aca0a10910e-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 08 19:40:39 crc kubenswrapper[4988]: I1008 19:40:39.521536 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-7mrpf"] Oct 08 19:40:39 crc kubenswrapper[4988]: I1008 19:40:39.537379 4988 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-7mrpf"] Oct 08 19:40:40 crc kubenswrapper[4988]: I1008 19:40:40.181998 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-northd-0" Oct 08 19:40:41 crc kubenswrapper[4988]: I1008 19:40:41.250081 4988 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f011ae53-f461-4a61-a75e-1aca0a10910e" path="/var/lib/kubelet/pods/f011ae53-f461-4a61-a75e-1aca0a10910e/volumes" Oct 08 19:40:45 crc kubenswrapper[4988]: I1008 19:40:45.454514 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-70ac-account-create-wmgnd"] Oct 08 19:40:45 crc kubenswrapper[4988]: E1008 19:40:45.455881 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f011ae53-f461-4a61-a75e-1aca0a10910e" containerName="extract-content" Oct 08 19:40:45 crc kubenswrapper[4988]: I1008 19:40:45.455913 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="f011ae53-f461-4a61-a75e-1aca0a10910e" containerName="extract-content" Oct 08 19:40:45 crc kubenswrapper[4988]: E1008 19:40:45.455984 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e6a11d4a-8b71-47bf-a169-f814d7c5455c" containerName="mariadb-database-create" Oct 08 19:40:45 crc kubenswrapper[4988]: I1008 19:40:45.456008 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="e6a11d4a-8b71-47bf-a169-f814d7c5455c" containerName="mariadb-database-create" Oct 08 19:40:45 crc kubenswrapper[4988]: E1008 19:40:45.456058 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f011ae53-f461-4a61-a75e-1aca0a10910e" containerName="extract-utilities" Oct 08 19:40:45 crc kubenswrapper[4988]: I1008 19:40:45.456108 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="f011ae53-f461-4a61-a75e-1aca0a10910e" containerName="extract-utilities" Oct 08 19:40:45 crc kubenswrapper[4988]: E1008 19:40:45.456139 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f011ae53-f461-4a61-a75e-1aca0a10910e" containerName="registry-server" Oct 08 19:40:45 crc kubenswrapper[4988]: I1008 19:40:45.456156 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="f011ae53-f461-4a61-a75e-1aca0a10910e" containerName="registry-server" Oct 08 19:40:45 crc kubenswrapper[4988]: I1008 19:40:45.456580 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="f011ae53-f461-4a61-a75e-1aca0a10910e" containerName="registry-server" Oct 08 19:40:45 crc kubenswrapper[4988]: I1008 19:40:45.456614 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="e6a11d4a-8b71-47bf-a169-f814d7c5455c" containerName="mariadb-database-create" Oct 08 19:40:45 crc kubenswrapper[4988]: I1008 19:40:45.457919 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-70ac-account-create-wmgnd" Oct 08 19:40:45 crc kubenswrapper[4988]: I1008 19:40:45.460763 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-db-secret" Oct 08 19:40:45 crc kubenswrapper[4988]: I1008 19:40:45.462029 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-70ac-account-create-wmgnd"] Oct 08 19:40:45 crc kubenswrapper[4988]: I1008 19:40:45.496633 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hzcdh\" (UniqueName: \"kubernetes.io/projected/175bafdf-d19e-4c77-9296-6a150c92a009-kube-api-access-hzcdh\") pod \"keystone-70ac-account-create-wmgnd\" (UID: \"175bafdf-d19e-4c77-9296-6a150c92a009\") " pod="openstack/keystone-70ac-account-create-wmgnd" Oct 08 19:40:45 crc kubenswrapper[4988]: I1008 19:40:45.605339 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hzcdh\" (UniqueName: \"kubernetes.io/projected/175bafdf-d19e-4c77-9296-6a150c92a009-kube-api-access-hzcdh\") pod \"keystone-70ac-account-create-wmgnd\" (UID: \"175bafdf-d19e-4c77-9296-6a150c92a009\") " pod="openstack/keystone-70ac-account-create-wmgnd" Oct 08 19:40:45 crc kubenswrapper[4988]: I1008 19:40:45.626901 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hzcdh\" (UniqueName: \"kubernetes.io/projected/175bafdf-d19e-4c77-9296-6a150c92a009-kube-api-access-hzcdh\") pod \"keystone-70ac-account-create-wmgnd\" (UID: \"175bafdf-d19e-4c77-9296-6a150c92a009\") " pod="openstack/keystone-70ac-account-create-wmgnd" Oct 08 19:40:45 crc kubenswrapper[4988]: I1008 19:40:45.794284 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-70ac-account-create-wmgnd" Oct 08 19:40:46 crc kubenswrapper[4988]: I1008 19:40:46.248594 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-70ac-account-create-wmgnd"] Oct 08 19:40:46 crc kubenswrapper[4988]: W1008 19:40:46.257059 4988 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod175bafdf_d19e_4c77_9296_6a150c92a009.slice/crio-622b48a1717b09ecdfcc5208600ade692b5c44825878472266bb567efb585794 WatchSource:0}: Error finding container 622b48a1717b09ecdfcc5208600ade692b5c44825878472266bb567efb585794: Status 404 returned error can't find the container with id 622b48a1717b09ecdfcc5208600ade692b5c44825878472266bb567efb585794 Oct 08 19:40:46 crc kubenswrapper[4988]: I1008 19:40:46.688785 4988 generic.go:334] "Generic (PLEG): container finished" podID="175bafdf-d19e-4c77-9296-6a150c92a009" containerID="16f993453e17b5da7222fd493e51ced51bce9012568fdd229415c647988b5f60" exitCode=0 Oct 08 19:40:46 crc kubenswrapper[4988]: I1008 19:40:46.688868 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-70ac-account-create-wmgnd" event={"ID":"175bafdf-d19e-4c77-9296-6a150c92a009","Type":"ContainerDied","Data":"16f993453e17b5da7222fd493e51ced51bce9012568fdd229415c647988b5f60"} Oct 08 19:40:46 crc kubenswrapper[4988]: I1008 19:40:46.689057 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-70ac-account-create-wmgnd" event={"ID":"175bafdf-d19e-4c77-9296-6a150c92a009","Type":"ContainerStarted","Data":"622b48a1717b09ecdfcc5208600ade692b5c44825878472266bb567efb585794"} Oct 08 19:40:48 crc kubenswrapper[4988]: I1008 19:40:48.154366 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-70ac-account-create-wmgnd" Oct 08 19:40:48 crc kubenswrapper[4988]: I1008 19:40:48.247004 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hzcdh\" (UniqueName: \"kubernetes.io/projected/175bafdf-d19e-4c77-9296-6a150c92a009-kube-api-access-hzcdh\") pod \"175bafdf-d19e-4c77-9296-6a150c92a009\" (UID: \"175bafdf-d19e-4c77-9296-6a150c92a009\") " Oct 08 19:40:48 crc kubenswrapper[4988]: I1008 19:40:48.255406 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/175bafdf-d19e-4c77-9296-6a150c92a009-kube-api-access-hzcdh" (OuterVolumeSpecName: "kube-api-access-hzcdh") pod "175bafdf-d19e-4c77-9296-6a150c92a009" (UID: "175bafdf-d19e-4c77-9296-6a150c92a009"). InnerVolumeSpecName "kube-api-access-hzcdh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 19:40:48 crc kubenswrapper[4988]: I1008 19:40:48.349297 4988 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hzcdh\" (UniqueName: \"kubernetes.io/projected/175bafdf-d19e-4c77-9296-6a150c92a009-kube-api-access-hzcdh\") on node \"crc\" DevicePath \"\"" Oct 08 19:40:48 crc kubenswrapper[4988]: I1008 19:40:48.708067 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-70ac-account-create-wmgnd" event={"ID":"175bafdf-d19e-4c77-9296-6a150c92a009","Type":"ContainerDied","Data":"622b48a1717b09ecdfcc5208600ade692b5c44825878472266bb567efb585794"} Oct 08 19:40:48 crc kubenswrapper[4988]: I1008 19:40:48.708123 4988 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="622b48a1717b09ecdfcc5208600ade692b5c44825878472266bb567efb585794" Oct 08 19:40:48 crc kubenswrapper[4988]: I1008 19:40:48.708134 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-70ac-account-create-wmgnd" Oct 08 19:40:50 crc kubenswrapper[4988]: I1008 19:40:50.826915 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-sync-m5jld"] Oct 08 19:40:50 crc kubenswrapper[4988]: E1008 19:40:50.827626 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="175bafdf-d19e-4c77-9296-6a150c92a009" containerName="mariadb-account-create" Oct 08 19:40:50 crc kubenswrapper[4988]: I1008 19:40:50.827644 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="175bafdf-d19e-4c77-9296-6a150c92a009" containerName="mariadb-account-create" Oct 08 19:40:50 crc kubenswrapper[4988]: I1008 19:40:50.827873 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="175bafdf-d19e-4c77-9296-6a150c92a009" containerName="mariadb-account-create" Oct 08 19:40:50 crc kubenswrapper[4988]: I1008 19:40:50.828465 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-m5jld" Oct 08 19:40:50 crc kubenswrapper[4988]: I1008 19:40:50.830671 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Oct 08 19:40:50 crc kubenswrapper[4988]: I1008 19:40:50.832294 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Oct 08 19:40:50 crc kubenswrapper[4988]: I1008 19:40:50.832310 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-992fc" Oct 08 19:40:50 crc kubenswrapper[4988]: I1008 19:40:50.833808 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Oct 08 19:40:50 crc kubenswrapper[4988]: I1008 19:40:50.849595 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-m5jld"] Oct 08 19:40:50 crc kubenswrapper[4988]: I1008 19:40:50.897795 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3c634709-46fb-46fd-a43b-7c6c0fc8e844-config-data\") pod \"keystone-db-sync-m5jld\" (UID: \"3c634709-46fb-46fd-a43b-7c6c0fc8e844\") " pod="openstack/keystone-db-sync-m5jld" Oct 08 19:40:50 crc kubenswrapper[4988]: I1008 19:40:50.897873 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3c634709-46fb-46fd-a43b-7c6c0fc8e844-combined-ca-bundle\") pod \"keystone-db-sync-m5jld\" (UID: \"3c634709-46fb-46fd-a43b-7c6c0fc8e844\") " pod="openstack/keystone-db-sync-m5jld" Oct 08 19:40:50 crc kubenswrapper[4988]: I1008 19:40:50.897916 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f962k\" (UniqueName: \"kubernetes.io/projected/3c634709-46fb-46fd-a43b-7c6c0fc8e844-kube-api-access-f962k\") pod \"keystone-db-sync-m5jld\" (UID: \"3c634709-46fb-46fd-a43b-7c6c0fc8e844\") " pod="openstack/keystone-db-sync-m5jld" Oct 08 19:40:50 crc kubenswrapper[4988]: I1008 19:40:50.999145 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3c634709-46fb-46fd-a43b-7c6c0fc8e844-config-data\") pod \"keystone-db-sync-m5jld\" (UID: \"3c634709-46fb-46fd-a43b-7c6c0fc8e844\") " pod="openstack/keystone-db-sync-m5jld" Oct 08 19:40:50 crc kubenswrapper[4988]: I1008 19:40:50.999245 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3c634709-46fb-46fd-a43b-7c6c0fc8e844-combined-ca-bundle\") pod \"keystone-db-sync-m5jld\" (UID: \"3c634709-46fb-46fd-a43b-7c6c0fc8e844\") " pod="openstack/keystone-db-sync-m5jld" Oct 08 19:40:50 crc kubenswrapper[4988]: I1008 19:40:50.999300 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f962k\" (UniqueName: \"kubernetes.io/projected/3c634709-46fb-46fd-a43b-7c6c0fc8e844-kube-api-access-f962k\") pod \"keystone-db-sync-m5jld\" (UID: \"3c634709-46fb-46fd-a43b-7c6c0fc8e844\") " pod="openstack/keystone-db-sync-m5jld" Oct 08 19:40:51 crc kubenswrapper[4988]: I1008 19:40:51.005209 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3c634709-46fb-46fd-a43b-7c6c0fc8e844-config-data\") pod \"keystone-db-sync-m5jld\" (UID: \"3c634709-46fb-46fd-a43b-7c6c0fc8e844\") " pod="openstack/keystone-db-sync-m5jld" Oct 08 19:40:51 crc kubenswrapper[4988]: I1008 19:40:51.017497 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3c634709-46fb-46fd-a43b-7c6c0fc8e844-combined-ca-bundle\") pod \"keystone-db-sync-m5jld\" (UID: \"3c634709-46fb-46fd-a43b-7c6c0fc8e844\") " pod="openstack/keystone-db-sync-m5jld" Oct 08 19:40:51 crc kubenswrapper[4988]: I1008 19:40:51.024170 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f962k\" (UniqueName: \"kubernetes.io/projected/3c634709-46fb-46fd-a43b-7c6c0fc8e844-kube-api-access-f962k\") pod \"keystone-db-sync-m5jld\" (UID: \"3c634709-46fb-46fd-a43b-7c6c0fc8e844\") " pod="openstack/keystone-db-sync-m5jld" Oct 08 19:40:51 crc kubenswrapper[4988]: I1008 19:40:51.157962 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-m5jld" Oct 08 19:40:51 crc kubenswrapper[4988]: I1008 19:40:51.628007 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-m5jld"] Oct 08 19:40:51 crc kubenswrapper[4988]: I1008 19:40:51.740971 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-m5jld" event={"ID":"3c634709-46fb-46fd-a43b-7c6c0fc8e844","Type":"ContainerStarted","Data":"fec798da10be2c79e35df1337b2637163e95981346e0c913133cc2dc463f6d5c"} Oct 08 19:40:52 crc kubenswrapper[4988]: I1008 19:40:52.754709 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-m5jld" event={"ID":"3c634709-46fb-46fd-a43b-7c6c0fc8e844","Type":"ContainerStarted","Data":"66762b138eba71245f007232f9cd0ef561419f725d3fe4ae225edaee9c59cd76"} Oct 08 19:40:52 crc kubenswrapper[4988]: I1008 19:40:52.777065 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-db-sync-m5jld" podStartSLOduration=2.777030865 podStartE2EDuration="2.777030865s" podCreationTimestamp="2025-10-08 19:40:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 19:40:52.775135444 +0000 UTC m=+5398.224978294" watchObservedRunningTime="2025-10-08 19:40:52.777030865 +0000 UTC m=+5398.226873685" Oct 08 19:40:53 crc kubenswrapper[4988]: I1008 19:40:53.768257 4988 generic.go:334] "Generic (PLEG): container finished" podID="3c634709-46fb-46fd-a43b-7c6c0fc8e844" containerID="66762b138eba71245f007232f9cd0ef561419f725d3fe4ae225edaee9c59cd76" exitCode=0 Oct 08 19:40:53 crc kubenswrapper[4988]: I1008 19:40:53.768338 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-m5jld" event={"ID":"3c634709-46fb-46fd-a43b-7c6c0fc8e844","Type":"ContainerDied","Data":"66762b138eba71245f007232f9cd0ef561419f725d3fe4ae225edaee9c59cd76"} Oct 08 19:40:55 crc kubenswrapper[4988]: I1008 19:40:55.141529 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-m5jld" Oct 08 19:40:55 crc kubenswrapper[4988]: I1008 19:40:55.284988 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-f962k\" (UniqueName: \"kubernetes.io/projected/3c634709-46fb-46fd-a43b-7c6c0fc8e844-kube-api-access-f962k\") pod \"3c634709-46fb-46fd-a43b-7c6c0fc8e844\" (UID: \"3c634709-46fb-46fd-a43b-7c6c0fc8e844\") " Oct 08 19:40:55 crc kubenswrapper[4988]: I1008 19:40:55.285470 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3c634709-46fb-46fd-a43b-7c6c0fc8e844-combined-ca-bundle\") pod \"3c634709-46fb-46fd-a43b-7c6c0fc8e844\" (UID: \"3c634709-46fb-46fd-a43b-7c6c0fc8e844\") " Oct 08 19:40:55 crc kubenswrapper[4988]: I1008 19:40:55.285664 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3c634709-46fb-46fd-a43b-7c6c0fc8e844-config-data\") pod \"3c634709-46fb-46fd-a43b-7c6c0fc8e844\" (UID: \"3c634709-46fb-46fd-a43b-7c6c0fc8e844\") " Oct 08 19:40:55 crc kubenswrapper[4988]: I1008 19:40:55.323962 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3c634709-46fb-46fd-a43b-7c6c0fc8e844-kube-api-access-f962k" (OuterVolumeSpecName: "kube-api-access-f962k") pod "3c634709-46fb-46fd-a43b-7c6c0fc8e844" (UID: "3c634709-46fb-46fd-a43b-7c6c0fc8e844"). InnerVolumeSpecName "kube-api-access-f962k". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 19:40:55 crc kubenswrapper[4988]: I1008 19:40:55.335901 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3c634709-46fb-46fd-a43b-7c6c0fc8e844-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "3c634709-46fb-46fd-a43b-7c6c0fc8e844" (UID: "3c634709-46fb-46fd-a43b-7c6c0fc8e844"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 19:40:55 crc kubenswrapper[4988]: I1008 19:40:55.357117 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3c634709-46fb-46fd-a43b-7c6c0fc8e844-config-data" (OuterVolumeSpecName: "config-data") pod "3c634709-46fb-46fd-a43b-7c6c0fc8e844" (UID: "3c634709-46fb-46fd-a43b-7c6c0fc8e844"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 19:40:55 crc kubenswrapper[4988]: I1008 19:40:55.387510 4988 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-f962k\" (UniqueName: \"kubernetes.io/projected/3c634709-46fb-46fd-a43b-7c6c0fc8e844-kube-api-access-f962k\") on node \"crc\" DevicePath \"\"" Oct 08 19:40:55 crc kubenswrapper[4988]: I1008 19:40:55.387543 4988 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3c634709-46fb-46fd-a43b-7c6c0fc8e844-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 19:40:55 crc kubenswrapper[4988]: I1008 19:40:55.387552 4988 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3c634709-46fb-46fd-a43b-7c6c0fc8e844-config-data\") on node \"crc\" DevicePath \"\"" Oct 08 19:40:55 crc kubenswrapper[4988]: I1008 19:40:55.790294 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-m5jld" event={"ID":"3c634709-46fb-46fd-a43b-7c6c0fc8e844","Type":"ContainerDied","Data":"fec798da10be2c79e35df1337b2637163e95981346e0c913133cc2dc463f6d5c"} Oct 08 19:40:55 crc kubenswrapper[4988]: I1008 19:40:55.790337 4988 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fec798da10be2c79e35df1337b2637163e95981346e0c913133cc2dc463f6d5c" Oct 08 19:40:55 crc kubenswrapper[4988]: I1008 19:40:55.790374 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-m5jld" Oct 08 19:40:56 crc kubenswrapper[4988]: I1008 19:40:56.041597 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-6c7c55f8b9-wlltj"] Oct 08 19:40:56 crc kubenswrapper[4988]: E1008 19:40:56.041927 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3c634709-46fb-46fd-a43b-7c6c0fc8e844" containerName="keystone-db-sync" Oct 08 19:40:56 crc kubenswrapper[4988]: I1008 19:40:56.041941 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="3c634709-46fb-46fd-a43b-7c6c0fc8e844" containerName="keystone-db-sync" Oct 08 19:40:56 crc kubenswrapper[4988]: I1008 19:40:56.042117 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="3c634709-46fb-46fd-a43b-7c6c0fc8e844" containerName="keystone-db-sync" Oct 08 19:40:56 crc kubenswrapper[4988]: I1008 19:40:56.042928 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6c7c55f8b9-wlltj" Oct 08 19:40:56 crc kubenswrapper[4988]: I1008 19:40:56.077060 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6c7c55f8b9-wlltj"] Oct 08 19:40:56 crc kubenswrapper[4988]: I1008 19:40:56.119457 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-5wnwz"] Oct 08 19:40:56 crc kubenswrapper[4988]: I1008 19:40:56.120472 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-5wnwz" Oct 08 19:40:56 crc kubenswrapper[4988]: I1008 19:40:56.126013 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-5wnwz"] Oct 08 19:40:56 crc kubenswrapper[4988]: I1008 19:40:56.133958 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Oct 08 19:40:56 crc kubenswrapper[4988]: I1008 19:40:56.134008 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-992fc" Oct 08 19:40:56 crc kubenswrapper[4988]: I1008 19:40:56.134218 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Oct 08 19:40:56 crc kubenswrapper[4988]: I1008 19:40:56.146051 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Oct 08 19:40:56 crc kubenswrapper[4988]: I1008 19:40:56.214543 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d45cfd02-b2f1-41ec-8751-6378224432ec-config-data\") pod \"keystone-bootstrap-5wnwz\" (UID: \"d45cfd02-b2f1-41ec-8751-6378224432ec\") " pod="openstack/keystone-bootstrap-5wnwz" Oct 08 19:40:56 crc kubenswrapper[4988]: I1008 19:40:56.214600 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/d45cfd02-b2f1-41ec-8751-6378224432ec-fernet-keys\") pod \"keystone-bootstrap-5wnwz\" (UID: \"d45cfd02-b2f1-41ec-8751-6378224432ec\") " pod="openstack/keystone-bootstrap-5wnwz" Oct 08 19:40:56 crc kubenswrapper[4988]: I1008 19:40:56.214625 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7lv46\" (UniqueName: \"kubernetes.io/projected/15716656-ce43-447b-9910-11b4813b5e52-kube-api-access-7lv46\") pod \"dnsmasq-dns-6c7c55f8b9-wlltj\" (UID: \"15716656-ce43-447b-9910-11b4813b5e52\") " pod="openstack/dnsmasq-dns-6c7c55f8b9-wlltj" Oct 08 19:40:56 crc kubenswrapper[4988]: I1008 19:40:56.214640 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-792zh\" (UniqueName: \"kubernetes.io/projected/d45cfd02-b2f1-41ec-8751-6378224432ec-kube-api-access-792zh\") pod \"keystone-bootstrap-5wnwz\" (UID: \"d45cfd02-b2f1-41ec-8751-6378224432ec\") " pod="openstack/keystone-bootstrap-5wnwz" Oct 08 19:40:56 crc kubenswrapper[4988]: I1008 19:40:56.214657 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/15716656-ce43-447b-9910-11b4813b5e52-ovsdbserver-nb\") pod \"dnsmasq-dns-6c7c55f8b9-wlltj\" (UID: \"15716656-ce43-447b-9910-11b4813b5e52\") " pod="openstack/dnsmasq-dns-6c7c55f8b9-wlltj" Oct 08 19:40:56 crc kubenswrapper[4988]: I1008 19:40:56.214670 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d45cfd02-b2f1-41ec-8751-6378224432ec-combined-ca-bundle\") pod \"keystone-bootstrap-5wnwz\" (UID: \"d45cfd02-b2f1-41ec-8751-6378224432ec\") " pod="openstack/keystone-bootstrap-5wnwz" Oct 08 19:40:56 crc kubenswrapper[4988]: I1008 19:40:56.214698 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/15716656-ce43-447b-9910-11b4813b5e52-dns-svc\") pod \"dnsmasq-dns-6c7c55f8b9-wlltj\" (UID: \"15716656-ce43-447b-9910-11b4813b5e52\") " pod="openstack/dnsmasq-dns-6c7c55f8b9-wlltj" Oct 08 19:40:56 crc kubenswrapper[4988]: I1008 19:40:56.214719 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/15716656-ce43-447b-9910-11b4813b5e52-ovsdbserver-sb\") pod \"dnsmasq-dns-6c7c55f8b9-wlltj\" (UID: \"15716656-ce43-447b-9910-11b4813b5e52\") " pod="openstack/dnsmasq-dns-6c7c55f8b9-wlltj" Oct 08 19:40:56 crc kubenswrapper[4988]: I1008 19:40:56.214737 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d45cfd02-b2f1-41ec-8751-6378224432ec-scripts\") pod \"keystone-bootstrap-5wnwz\" (UID: \"d45cfd02-b2f1-41ec-8751-6378224432ec\") " pod="openstack/keystone-bootstrap-5wnwz" Oct 08 19:40:56 crc kubenswrapper[4988]: I1008 19:40:56.214772 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/15716656-ce43-447b-9910-11b4813b5e52-config\") pod \"dnsmasq-dns-6c7c55f8b9-wlltj\" (UID: \"15716656-ce43-447b-9910-11b4813b5e52\") " pod="openstack/dnsmasq-dns-6c7c55f8b9-wlltj" Oct 08 19:40:56 crc kubenswrapper[4988]: I1008 19:40:56.214790 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/d45cfd02-b2f1-41ec-8751-6378224432ec-credential-keys\") pod \"keystone-bootstrap-5wnwz\" (UID: \"d45cfd02-b2f1-41ec-8751-6378224432ec\") " pod="openstack/keystone-bootstrap-5wnwz" Oct 08 19:40:56 crc kubenswrapper[4988]: I1008 19:40:56.315531 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d45cfd02-b2f1-41ec-8751-6378224432ec-scripts\") pod \"keystone-bootstrap-5wnwz\" (UID: \"d45cfd02-b2f1-41ec-8751-6378224432ec\") " pod="openstack/keystone-bootstrap-5wnwz" Oct 08 19:40:56 crc kubenswrapper[4988]: I1008 19:40:56.315624 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/15716656-ce43-447b-9910-11b4813b5e52-config\") pod \"dnsmasq-dns-6c7c55f8b9-wlltj\" (UID: \"15716656-ce43-447b-9910-11b4813b5e52\") " pod="openstack/dnsmasq-dns-6c7c55f8b9-wlltj" Oct 08 19:40:56 crc kubenswrapper[4988]: I1008 19:40:56.315644 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/d45cfd02-b2f1-41ec-8751-6378224432ec-credential-keys\") pod \"keystone-bootstrap-5wnwz\" (UID: \"d45cfd02-b2f1-41ec-8751-6378224432ec\") " pod="openstack/keystone-bootstrap-5wnwz" Oct 08 19:40:56 crc kubenswrapper[4988]: I1008 19:40:56.315691 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d45cfd02-b2f1-41ec-8751-6378224432ec-config-data\") pod \"keystone-bootstrap-5wnwz\" (UID: \"d45cfd02-b2f1-41ec-8751-6378224432ec\") " pod="openstack/keystone-bootstrap-5wnwz" Oct 08 19:40:56 crc kubenswrapper[4988]: I1008 19:40:56.315724 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/d45cfd02-b2f1-41ec-8751-6378224432ec-fernet-keys\") pod \"keystone-bootstrap-5wnwz\" (UID: \"d45cfd02-b2f1-41ec-8751-6378224432ec\") " pod="openstack/keystone-bootstrap-5wnwz" Oct 08 19:40:56 crc kubenswrapper[4988]: I1008 19:40:56.315746 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7lv46\" (UniqueName: \"kubernetes.io/projected/15716656-ce43-447b-9910-11b4813b5e52-kube-api-access-7lv46\") pod \"dnsmasq-dns-6c7c55f8b9-wlltj\" (UID: \"15716656-ce43-447b-9910-11b4813b5e52\") " pod="openstack/dnsmasq-dns-6c7c55f8b9-wlltj" Oct 08 19:40:56 crc kubenswrapper[4988]: I1008 19:40:56.315763 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-792zh\" (UniqueName: \"kubernetes.io/projected/d45cfd02-b2f1-41ec-8751-6378224432ec-kube-api-access-792zh\") pod \"keystone-bootstrap-5wnwz\" (UID: \"d45cfd02-b2f1-41ec-8751-6378224432ec\") " pod="openstack/keystone-bootstrap-5wnwz" Oct 08 19:40:56 crc kubenswrapper[4988]: I1008 19:40:56.315779 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/15716656-ce43-447b-9910-11b4813b5e52-ovsdbserver-nb\") pod \"dnsmasq-dns-6c7c55f8b9-wlltj\" (UID: \"15716656-ce43-447b-9910-11b4813b5e52\") " pod="openstack/dnsmasq-dns-6c7c55f8b9-wlltj" Oct 08 19:40:56 crc kubenswrapper[4988]: I1008 19:40:56.315794 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d45cfd02-b2f1-41ec-8751-6378224432ec-combined-ca-bundle\") pod \"keystone-bootstrap-5wnwz\" (UID: \"d45cfd02-b2f1-41ec-8751-6378224432ec\") " pod="openstack/keystone-bootstrap-5wnwz" Oct 08 19:40:56 crc kubenswrapper[4988]: I1008 19:40:56.315826 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/15716656-ce43-447b-9910-11b4813b5e52-dns-svc\") pod \"dnsmasq-dns-6c7c55f8b9-wlltj\" (UID: \"15716656-ce43-447b-9910-11b4813b5e52\") " pod="openstack/dnsmasq-dns-6c7c55f8b9-wlltj" Oct 08 19:40:56 crc kubenswrapper[4988]: I1008 19:40:56.315851 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/15716656-ce43-447b-9910-11b4813b5e52-ovsdbserver-sb\") pod \"dnsmasq-dns-6c7c55f8b9-wlltj\" (UID: \"15716656-ce43-447b-9910-11b4813b5e52\") " pod="openstack/dnsmasq-dns-6c7c55f8b9-wlltj" Oct 08 19:40:56 crc kubenswrapper[4988]: I1008 19:40:56.316594 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/15716656-ce43-447b-9910-11b4813b5e52-ovsdbserver-sb\") pod \"dnsmasq-dns-6c7c55f8b9-wlltj\" (UID: \"15716656-ce43-447b-9910-11b4813b5e52\") " pod="openstack/dnsmasq-dns-6c7c55f8b9-wlltj" Oct 08 19:40:56 crc kubenswrapper[4988]: I1008 19:40:56.317086 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/15716656-ce43-447b-9910-11b4813b5e52-ovsdbserver-nb\") pod \"dnsmasq-dns-6c7c55f8b9-wlltj\" (UID: \"15716656-ce43-447b-9910-11b4813b5e52\") " pod="openstack/dnsmasq-dns-6c7c55f8b9-wlltj" Oct 08 19:40:56 crc kubenswrapper[4988]: I1008 19:40:56.318138 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/15716656-ce43-447b-9910-11b4813b5e52-dns-svc\") pod \"dnsmasq-dns-6c7c55f8b9-wlltj\" (UID: \"15716656-ce43-447b-9910-11b4813b5e52\") " pod="openstack/dnsmasq-dns-6c7c55f8b9-wlltj" Oct 08 19:40:56 crc kubenswrapper[4988]: I1008 19:40:56.318157 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/15716656-ce43-447b-9910-11b4813b5e52-config\") pod \"dnsmasq-dns-6c7c55f8b9-wlltj\" (UID: \"15716656-ce43-447b-9910-11b4813b5e52\") " pod="openstack/dnsmasq-dns-6c7c55f8b9-wlltj" Oct 08 19:40:56 crc kubenswrapper[4988]: I1008 19:40:56.331901 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d45cfd02-b2f1-41ec-8751-6378224432ec-combined-ca-bundle\") pod \"keystone-bootstrap-5wnwz\" (UID: \"d45cfd02-b2f1-41ec-8751-6378224432ec\") " pod="openstack/keystone-bootstrap-5wnwz" Oct 08 19:40:56 crc kubenswrapper[4988]: I1008 19:40:56.333125 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d45cfd02-b2f1-41ec-8751-6378224432ec-scripts\") pod \"keystone-bootstrap-5wnwz\" (UID: \"d45cfd02-b2f1-41ec-8751-6378224432ec\") " pod="openstack/keystone-bootstrap-5wnwz" Oct 08 19:40:56 crc kubenswrapper[4988]: I1008 19:40:56.333939 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/d45cfd02-b2f1-41ec-8751-6378224432ec-credential-keys\") pod \"keystone-bootstrap-5wnwz\" (UID: \"d45cfd02-b2f1-41ec-8751-6378224432ec\") " pod="openstack/keystone-bootstrap-5wnwz" Oct 08 19:40:56 crc kubenswrapper[4988]: I1008 19:40:56.335902 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/d45cfd02-b2f1-41ec-8751-6378224432ec-fernet-keys\") pod \"keystone-bootstrap-5wnwz\" (UID: \"d45cfd02-b2f1-41ec-8751-6378224432ec\") " pod="openstack/keystone-bootstrap-5wnwz" Oct 08 19:40:56 crc kubenswrapper[4988]: I1008 19:40:56.336169 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d45cfd02-b2f1-41ec-8751-6378224432ec-config-data\") pod \"keystone-bootstrap-5wnwz\" (UID: \"d45cfd02-b2f1-41ec-8751-6378224432ec\") " pod="openstack/keystone-bootstrap-5wnwz" Oct 08 19:40:56 crc kubenswrapper[4988]: I1008 19:40:56.343528 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-792zh\" (UniqueName: \"kubernetes.io/projected/d45cfd02-b2f1-41ec-8751-6378224432ec-kube-api-access-792zh\") pod \"keystone-bootstrap-5wnwz\" (UID: \"d45cfd02-b2f1-41ec-8751-6378224432ec\") " pod="openstack/keystone-bootstrap-5wnwz" Oct 08 19:40:56 crc kubenswrapper[4988]: I1008 19:40:56.344187 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7lv46\" (UniqueName: \"kubernetes.io/projected/15716656-ce43-447b-9910-11b4813b5e52-kube-api-access-7lv46\") pod \"dnsmasq-dns-6c7c55f8b9-wlltj\" (UID: \"15716656-ce43-447b-9910-11b4813b5e52\") " pod="openstack/dnsmasq-dns-6c7c55f8b9-wlltj" Oct 08 19:40:56 crc kubenswrapper[4988]: I1008 19:40:56.362569 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6c7c55f8b9-wlltj" Oct 08 19:40:56 crc kubenswrapper[4988]: I1008 19:40:56.455559 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-5wnwz" Oct 08 19:40:56 crc kubenswrapper[4988]: I1008 19:40:56.854205 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6c7c55f8b9-wlltj"] Oct 08 19:40:56 crc kubenswrapper[4988]: W1008 19:40:56.860244 4988 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod15716656_ce43_447b_9910_11b4813b5e52.slice/crio-8ff82740ae1ec65178317bdd21e3481d29d174cffb920f3c9288c976b7d41fae WatchSource:0}: Error finding container 8ff82740ae1ec65178317bdd21e3481d29d174cffb920f3c9288c976b7d41fae: Status 404 returned error can't find the container with id 8ff82740ae1ec65178317bdd21e3481d29d174cffb920f3c9288c976b7d41fae Oct 08 19:40:56 crc kubenswrapper[4988]: I1008 19:40:56.929197 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-5wnwz"] Oct 08 19:40:57 crc kubenswrapper[4988]: I1008 19:40:57.822536 4988 generic.go:334] "Generic (PLEG): container finished" podID="15716656-ce43-447b-9910-11b4813b5e52" containerID="cff42b69f6d35ca4c5669a6d8580efd4430e29c89b59ff6444247a5a7a89664c" exitCode=0 Oct 08 19:40:57 crc kubenswrapper[4988]: I1008 19:40:57.822672 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6c7c55f8b9-wlltj" event={"ID":"15716656-ce43-447b-9910-11b4813b5e52","Type":"ContainerDied","Data":"cff42b69f6d35ca4c5669a6d8580efd4430e29c89b59ff6444247a5a7a89664c"} Oct 08 19:40:57 crc kubenswrapper[4988]: I1008 19:40:57.823259 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6c7c55f8b9-wlltj" event={"ID":"15716656-ce43-447b-9910-11b4813b5e52","Type":"ContainerStarted","Data":"8ff82740ae1ec65178317bdd21e3481d29d174cffb920f3c9288c976b7d41fae"} Oct 08 19:40:57 crc kubenswrapper[4988]: I1008 19:40:57.825921 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-5wnwz" event={"ID":"d45cfd02-b2f1-41ec-8751-6378224432ec","Type":"ContainerStarted","Data":"b72558f2ab7cfdd09ca75b37174b9784d823a4620a4720c222b2857202ad4232"} Oct 08 19:40:57 crc kubenswrapper[4988]: I1008 19:40:57.825965 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-5wnwz" event={"ID":"d45cfd02-b2f1-41ec-8751-6378224432ec","Type":"ContainerStarted","Data":"388e35f7173d1324bcce854d5ec284f43ced2c99e8342e82b23d0bc093e762be"} Oct 08 19:40:57 crc kubenswrapper[4988]: I1008 19:40:57.886627 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-5wnwz" podStartSLOduration=1.886605356 podStartE2EDuration="1.886605356s" podCreationTimestamp="2025-10-08 19:40:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 19:40:57.886119972 +0000 UTC m=+5403.335962772" watchObservedRunningTime="2025-10-08 19:40:57.886605356 +0000 UTC m=+5403.336448136" Oct 08 19:40:58 crc kubenswrapper[4988]: I1008 19:40:58.003732 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-wlkr2"] Oct 08 19:40:58 crc kubenswrapper[4988]: I1008 19:40:58.007070 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-wlkr2" Oct 08 19:40:58 crc kubenswrapper[4988]: I1008 19:40:58.019968 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-wlkr2"] Oct 08 19:40:58 crc kubenswrapper[4988]: I1008 19:40:58.047555 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jv24t\" (UniqueName: \"kubernetes.io/projected/ca92153f-9b18-45b1-9ba7-6d34cf132c1b-kube-api-access-jv24t\") pod \"certified-operators-wlkr2\" (UID: \"ca92153f-9b18-45b1-9ba7-6d34cf132c1b\") " pod="openshift-marketplace/certified-operators-wlkr2" Oct 08 19:40:58 crc kubenswrapper[4988]: I1008 19:40:58.047636 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ca92153f-9b18-45b1-9ba7-6d34cf132c1b-utilities\") pod \"certified-operators-wlkr2\" (UID: \"ca92153f-9b18-45b1-9ba7-6d34cf132c1b\") " pod="openshift-marketplace/certified-operators-wlkr2" Oct 08 19:40:58 crc kubenswrapper[4988]: I1008 19:40:58.047709 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ca92153f-9b18-45b1-9ba7-6d34cf132c1b-catalog-content\") pod \"certified-operators-wlkr2\" (UID: \"ca92153f-9b18-45b1-9ba7-6d34cf132c1b\") " pod="openshift-marketplace/certified-operators-wlkr2" Oct 08 19:40:58 crc kubenswrapper[4988]: I1008 19:40:58.148873 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ca92153f-9b18-45b1-9ba7-6d34cf132c1b-catalog-content\") pod \"certified-operators-wlkr2\" (UID: \"ca92153f-9b18-45b1-9ba7-6d34cf132c1b\") " pod="openshift-marketplace/certified-operators-wlkr2" Oct 08 19:40:58 crc kubenswrapper[4988]: I1008 19:40:58.148947 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jv24t\" (UniqueName: \"kubernetes.io/projected/ca92153f-9b18-45b1-9ba7-6d34cf132c1b-kube-api-access-jv24t\") pod \"certified-operators-wlkr2\" (UID: \"ca92153f-9b18-45b1-9ba7-6d34cf132c1b\") " pod="openshift-marketplace/certified-operators-wlkr2" Oct 08 19:40:58 crc kubenswrapper[4988]: I1008 19:40:58.148993 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ca92153f-9b18-45b1-9ba7-6d34cf132c1b-utilities\") pod \"certified-operators-wlkr2\" (UID: \"ca92153f-9b18-45b1-9ba7-6d34cf132c1b\") " pod="openshift-marketplace/certified-operators-wlkr2" Oct 08 19:40:58 crc kubenswrapper[4988]: I1008 19:40:58.149364 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ca92153f-9b18-45b1-9ba7-6d34cf132c1b-utilities\") pod \"certified-operators-wlkr2\" (UID: \"ca92153f-9b18-45b1-9ba7-6d34cf132c1b\") " pod="openshift-marketplace/certified-operators-wlkr2" Oct 08 19:40:58 crc kubenswrapper[4988]: I1008 19:40:58.151338 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ca92153f-9b18-45b1-9ba7-6d34cf132c1b-catalog-content\") pod \"certified-operators-wlkr2\" (UID: \"ca92153f-9b18-45b1-9ba7-6d34cf132c1b\") " pod="openshift-marketplace/certified-operators-wlkr2" Oct 08 19:40:58 crc kubenswrapper[4988]: I1008 19:40:58.168822 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jv24t\" (UniqueName: \"kubernetes.io/projected/ca92153f-9b18-45b1-9ba7-6d34cf132c1b-kube-api-access-jv24t\") pod \"certified-operators-wlkr2\" (UID: \"ca92153f-9b18-45b1-9ba7-6d34cf132c1b\") " pod="openshift-marketplace/certified-operators-wlkr2" Oct 08 19:40:58 crc kubenswrapper[4988]: I1008 19:40:58.365612 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-wlkr2" Oct 08 19:40:58 crc kubenswrapper[4988]: I1008 19:40:58.834363 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6c7c55f8b9-wlltj" event={"ID":"15716656-ce43-447b-9910-11b4813b5e52","Type":"ContainerStarted","Data":"df85a0680fc1f4db9b155d8405d389aa23264924aa46e3f4a41982e3c5d7966b"} Oct 08 19:40:58 crc kubenswrapper[4988]: I1008 19:40:58.860055 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-6c7c55f8b9-wlltj" podStartSLOduration=2.86003624 podStartE2EDuration="2.86003624s" podCreationTimestamp="2025-10-08 19:40:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 19:40:58.859941877 +0000 UTC m=+5404.309784647" watchObservedRunningTime="2025-10-08 19:40:58.86003624 +0000 UTC m=+5404.309879010" Oct 08 19:40:58 crc kubenswrapper[4988]: I1008 19:40:58.922143 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-wlkr2"] Oct 08 19:40:59 crc kubenswrapper[4988]: I1008 19:40:59.845839 4988 generic.go:334] "Generic (PLEG): container finished" podID="ca92153f-9b18-45b1-9ba7-6d34cf132c1b" containerID="ace2be587539f22a2afd389b7f1cd09c4cb32307275ac7cf11a1baa374045ff9" exitCode=0 Oct 08 19:40:59 crc kubenswrapper[4988]: I1008 19:40:59.845898 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wlkr2" event={"ID":"ca92153f-9b18-45b1-9ba7-6d34cf132c1b","Type":"ContainerDied","Data":"ace2be587539f22a2afd389b7f1cd09c4cb32307275ac7cf11a1baa374045ff9"} Oct 08 19:40:59 crc kubenswrapper[4988]: I1008 19:40:59.846171 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wlkr2" event={"ID":"ca92153f-9b18-45b1-9ba7-6d34cf132c1b","Type":"ContainerStarted","Data":"dfb53089fc70ac16c9d40be0ce64d1acf8b410e5e125327ca4c544f2be405a5c"} Oct 08 19:40:59 crc kubenswrapper[4988]: I1008 19:40:59.846543 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-6c7c55f8b9-wlltj" Oct 08 19:41:00 crc kubenswrapper[4988]: I1008 19:41:00.862475 4988 generic.go:334] "Generic (PLEG): container finished" podID="d45cfd02-b2f1-41ec-8751-6378224432ec" containerID="b72558f2ab7cfdd09ca75b37174b9784d823a4620a4720c222b2857202ad4232" exitCode=0 Oct 08 19:41:00 crc kubenswrapper[4988]: I1008 19:41:00.862571 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-5wnwz" event={"ID":"d45cfd02-b2f1-41ec-8751-6378224432ec","Type":"ContainerDied","Data":"b72558f2ab7cfdd09ca75b37174b9784d823a4620a4720c222b2857202ad4232"} Oct 08 19:41:01 crc kubenswrapper[4988]: I1008 19:41:01.881107 4988 generic.go:334] "Generic (PLEG): container finished" podID="ca92153f-9b18-45b1-9ba7-6d34cf132c1b" containerID="c28033bf02d45d29e61ff71fa479f26ef008a1b25ac620b3f020b9ec443204e2" exitCode=0 Oct 08 19:41:01 crc kubenswrapper[4988]: I1008 19:41:01.881187 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wlkr2" event={"ID":"ca92153f-9b18-45b1-9ba7-6d34cf132c1b","Type":"ContainerDied","Data":"c28033bf02d45d29e61ff71fa479f26ef008a1b25ac620b3f020b9ec443204e2"} Oct 08 19:41:02 crc kubenswrapper[4988]: I1008 19:41:02.295646 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-5wnwz" Oct 08 19:41:02 crc kubenswrapper[4988]: I1008 19:41:02.448883 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d45cfd02-b2f1-41ec-8751-6378224432ec-combined-ca-bundle\") pod \"d45cfd02-b2f1-41ec-8751-6378224432ec\" (UID: \"d45cfd02-b2f1-41ec-8751-6378224432ec\") " Oct 08 19:41:02 crc kubenswrapper[4988]: I1008 19:41:02.449181 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d45cfd02-b2f1-41ec-8751-6378224432ec-scripts\") pod \"d45cfd02-b2f1-41ec-8751-6378224432ec\" (UID: \"d45cfd02-b2f1-41ec-8751-6378224432ec\") " Oct 08 19:41:02 crc kubenswrapper[4988]: I1008 19:41:02.449271 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d45cfd02-b2f1-41ec-8751-6378224432ec-config-data\") pod \"d45cfd02-b2f1-41ec-8751-6378224432ec\" (UID: \"d45cfd02-b2f1-41ec-8751-6378224432ec\") " Oct 08 19:41:02 crc kubenswrapper[4988]: I1008 19:41:02.449347 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/d45cfd02-b2f1-41ec-8751-6378224432ec-fernet-keys\") pod \"d45cfd02-b2f1-41ec-8751-6378224432ec\" (UID: \"d45cfd02-b2f1-41ec-8751-6378224432ec\") " Oct 08 19:41:02 crc kubenswrapper[4988]: I1008 19:41:02.449462 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/d45cfd02-b2f1-41ec-8751-6378224432ec-credential-keys\") pod \"d45cfd02-b2f1-41ec-8751-6378224432ec\" (UID: \"d45cfd02-b2f1-41ec-8751-6378224432ec\") " Oct 08 19:41:02 crc kubenswrapper[4988]: I1008 19:41:02.449567 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-792zh\" (UniqueName: \"kubernetes.io/projected/d45cfd02-b2f1-41ec-8751-6378224432ec-kube-api-access-792zh\") pod \"d45cfd02-b2f1-41ec-8751-6378224432ec\" (UID: \"d45cfd02-b2f1-41ec-8751-6378224432ec\") " Oct 08 19:41:02 crc kubenswrapper[4988]: I1008 19:41:02.455296 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d45cfd02-b2f1-41ec-8751-6378224432ec-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "d45cfd02-b2f1-41ec-8751-6378224432ec" (UID: "d45cfd02-b2f1-41ec-8751-6378224432ec"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 19:41:02 crc kubenswrapper[4988]: I1008 19:41:02.455851 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d45cfd02-b2f1-41ec-8751-6378224432ec-scripts" (OuterVolumeSpecName: "scripts") pod "d45cfd02-b2f1-41ec-8751-6378224432ec" (UID: "d45cfd02-b2f1-41ec-8751-6378224432ec"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 19:41:02 crc kubenswrapper[4988]: I1008 19:41:02.457224 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d45cfd02-b2f1-41ec-8751-6378224432ec-kube-api-access-792zh" (OuterVolumeSpecName: "kube-api-access-792zh") pod "d45cfd02-b2f1-41ec-8751-6378224432ec" (UID: "d45cfd02-b2f1-41ec-8751-6378224432ec"). InnerVolumeSpecName "kube-api-access-792zh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 19:41:02 crc kubenswrapper[4988]: I1008 19:41:02.460693 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d45cfd02-b2f1-41ec-8751-6378224432ec-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "d45cfd02-b2f1-41ec-8751-6378224432ec" (UID: "d45cfd02-b2f1-41ec-8751-6378224432ec"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 19:41:02 crc kubenswrapper[4988]: I1008 19:41:02.495544 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d45cfd02-b2f1-41ec-8751-6378224432ec-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d45cfd02-b2f1-41ec-8751-6378224432ec" (UID: "d45cfd02-b2f1-41ec-8751-6378224432ec"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 19:41:02 crc kubenswrapper[4988]: I1008 19:41:02.496447 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d45cfd02-b2f1-41ec-8751-6378224432ec-config-data" (OuterVolumeSpecName: "config-data") pod "d45cfd02-b2f1-41ec-8751-6378224432ec" (UID: "d45cfd02-b2f1-41ec-8751-6378224432ec"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 19:41:02 crc kubenswrapper[4988]: I1008 19:41:02.551750 4988 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d45cfd02-b2f1-41ec-8751-6378224432ec-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 19:41:02 crc kubenswrapper[4988]: I1008 19:41:02.551791 4988 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d45cfd02-b2f1-41ec-8751-6378224432ec-scripts\") on node \"crc\" DevicePath \"\"" Oct 08 19:41:02 crc kubenswrapper[4988]: I1008 19:41:02.551805 4988 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d45cfd02-b2f1-41ec-8751-6378224432ec-config-data\") on node \"crc\" DevicePath \"\"" Oct 08 19:41:02 crc kubenswrapper[4988]: I1008 19:41:02.551817 4988 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/d45cfd02-b2f1-41ec-8751-6378224432ec-fernet-keys\") on node \"crc\" DevicePath \"\"" Oct 08 19:41:02 crc kubenswrapper[4988]: I1008 19:41:02.551828 4988 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/d45cfd02-b2f1-41ec-8751-6378224432ec-credential-keys\") on node \"crc\" DevicePath \"\"" Oct 08 19:41:02 crc kubenswrapper[4988]: I1008 19:41:02.551840 4988 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-792zh\" (UniqueName: \"kubernetes.io/projected/d45cfd02-b2f1-41ec-8751-6378224432ec-kube-api-access-792zh\") on node \"crc\" DevicePath \"\"" Oct 08 19:41:02 crc kubenswrapper[4988]: I1008 19:41:02.894877 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wlkr2" event={"ID":"ca92153f-9b18-45b1-9ba7-6d34cf132c1b","Type":"ContainerStarted","Data":"fb3c53030852e901fd1fca7bc8e154094e90f5254682590d512fccaae3a08221"} Oct 08 19:41:02 crc kubenswrapper[4988]: I1008 19:41:02.899160 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-5wnwz" event={"ID":"d45cfd02-b2f1-41ec-8751-6378224432ec","Type":"ContainerDied","Data":"388e35f7173d1324bcce854d5ec284f43ced2c99e8342e82b23d0bc093e762be"} Oct 08 19:41:02 crc kubenswrapper[4988]: I1008 19:41:02.899202 4988 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="388e35f7173d1324bcce854d5ec284f43ced2c99e8342e82b23d0bc093e762be" Oct 08 19:41:02 crc kubenswrapper[4988]: I1008 19:41:02.899256 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-5wnwz" Oct 08 19:41:02 crc kubenswrapper[4988]: I1008 19:41:02.928330 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-wlkr2" podStartSLOduration=3.12970616 podStartE2EDuration="5.928309942s" podCreationTimestamp="2025-10-08 19:40:57 +0000 UTC" firstStartedPulling="2025-10-08 19:40:59.850199645 +0000 UTC m=+5405.300042455" lastFinishedPulling="2025-10-08 19:41:02.648803457 +0000 UTC m=+5408.098646237" observedRunningTime="2025-10-08 19:41:02.923159118 +0000 UTC m=+5408.373001978" watchObservedRunningTime="2025-10-08 19:41:02.928309942 +0000 UTC m=+5408.378152722" Oct 08 19:41:03 crc kubenswrapper[4988]: I1008 19:41:03.078201 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-5wnwz"] Oct 08 19:41:03 crc kubenswrapper[4988]: I1008 19:41:03.090808 4988 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-5wnwz"] Oct 08 19:41:03 crc kubenswrapper[4988]: I1008 19:41:03.161409 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-kb7bl"] Oct 08 19:41:03 crc kubenswrapper[4988]: E1008 19:41:03.161850 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d45cfd02-b2f1-41ec-8751-6378224432ec" containerName="keystone-bootstrap" Oct 08 19:41:03 crc kubenswrapper[4988]: I1008 19:41:03.161870 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="d45cfd02-b2f1-41ec-8751-6378224432ec" containerName="keystone-bootstrap" Oct 08 19:41:03 crc kubenswrapper[4988]: I1008 19:41:03.162114 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="d45cfd02-b2f1-41ec-8751-6378224432ec" containerName="keystone-bootstrap" Oct 08 19:41:03 crc kubenswrapper[4988]: I1008 19:41:03.162889 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-kb7bl" Oct 08 19:41:03 crc kubenswrapper[4988]: I1008 19:41:03.163821 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/94108d41-9804-48e5-982a-6e0df6ba3d9f-combined-ca-bundle\") pod \"keystone-bootstrap-kb7bl\" (UID: \"94108d41-9804-48e5-982a-6e0df6ba3d9f\") " pod="openstack/keystone-bootstrap-kb7bl" Oct 08 19:41:03 crc kubenswrapper[4988]: I1008 19:41:03.163904 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/94108d41-9804-48e5-982a-6e0df6ba3d9f-credential-keys\") pod \"keystone-bootstrap-kb7bl\" (UID: \"94108d41-9804-48e5-982a-6e0df6ba3d9f\") " pod="openstack/keystone-bootstrap-kb7bl" Oct 08 19:41:03 crc kubenswrapper[4988]: I1008 19:41:03.163929 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c6tx9\" (UniqueName: \"kubernetes.io/projected/94108d41-9804-48e5-982a-6e0df6ba3d9f-kube-api-access-c6tx9\") pod \"keystone-bootstrap-kb7bl\" (UID: \"94108d41-9804-48e5-982a-6e0df6ba3d9f\") " pod="openstack/keystone-bootstrap-kb7bl" Oct 08 19:41:03 crc kubenswrapper[4988]: I1008 19:41:03.164113 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/94108d41-9804-48e5-982a-6e0df6ba3d9f-fernet-keys\") pod \"keystone-bootstrap-kb7bl\" (UID: \"94108d41-9804-48e5-982a-6e0df6ba3d9f\") " pod="openstack/keystone-bootstrap-kb7bl" Oct 08 19:41:03 crc kubenswrapper[4988]: I1008 19:41:03.164162 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/94108d41-9804-48e5-982a-6e0df6ba3d9f-config-data\") pod \"keystone-bootstrap-kb7bl\" (UID: \"94108d41-9804-48e5-982a-6e0df6ba3d9f\") " pod="openstack/keystone-bootstrap-kb7bl" Oct 08 19:41:03 crc kubenswrapper[4988]: I1008 19:41:03.164278 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/94108d41-9804-48e5-982a-6e0df6ba3d9f-scripts\") pod \"keystone-bootstrap-kb7bl\" (UID: \"94108d41-9804-48e5-982a-6e0df6ba3d9f\") " pod="openstack/keystone-bootstrap-kb7bl" Oct 08 19:41:03 crc kubenswrapper[4988]: I1008 19:41:03.212664 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Oct 08 19:41:03 crc kubenswrapper[4988]: I1008 19:41:03.212790 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Oct 08 19:41:03 crc kubenswrapper[4988]: I1008 19:41:03.213016 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-992fc" Oct 08 19:41:03 crc kubenswrapper[4988]: I1008 19:41:03.213090 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Oct 08 19:41:03 crc kubenswrapper[4988]: I1008 19:41:03.220751 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-kb7bl"] Oct 08 19:41:03 crc kubenswrapper[4988]: I1008 19:41:03.265422 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/94108d41-9804-48e5-982a-6e0df6ba3d9f-scripts\") pod \"keystone-bootstrap-kb7bl\" (UID: \"94108d41-9804-48e5-982a-6e0df6ba3d9f\") " pod="openstack/keystone-bootstrap-kb7bl" Oct 08 19:41:03 crc kubenswrapper[4988]: I1008 19:41:03.265484 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/94108d41-9804-48e5-982a-6e0df6ba3d9f-combined-ca-bundle\") pod \"keystone-bootstrap-kb7bl\" (UID: \"94108d41-9804-48e5-982a-6e0df6ba3d9f\") " pod="openstack/keystone-bootstrap-kb7bl" Oct 08 19:41:03 crc kubenswrapper[4988]: I1008 19:41:03.265546 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/94108d41-9804-48e5-982a-6e0df6ba3d9f-credential-keys\") pod \"keystone-bootstrap-kb7bl\" (UID: \"94108d41-9804-48e5-982a-6e0df6ba3d9f\") " pod="openstack/keystone-bootstrap-kb7bl" Oct 08 19:41:03 crc kubenswrapper[4988]: I1008 19:41:03.265564 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c6tx9\" (UniqueName: \"kubernetes.io/projected/94108d41-9804-48e5-982a-6e0df6ba3d9f-kube-api-access-c6tx9\") pod \"keystone-bootstrap-kb7bl\" (UID: \"94108d41-9804-48e5-982a-6e0df6ba3d9f\") " pod="openstack/keystone-bootstrap-kb7bl" Oct 08 19:41:03 crc kubenswrapper[4988]: I1008 19:41:03.265631 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/94108d41-9804-48e5-982a-6e0df6ba3d9f-fernet-keys\") pod \"keystone-bootstrap-kb7bl\" (UID: \"94108d41-9804-48e5-982a-6e0df6ba3d9f\") " pod="openstack/keystone-bootstrap-kb7bl" Oct 08 19:41:03 crc kubenswrapper[4988]: I1008 19:41:03.265655 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/94108d41-9804-48e5-982a-6e0df6ba3d9f-config-data\") pod \"keystone-bootstrap-kb7bl\" (UID: \"94108d41-9804-48e5-982a-6e0df6ba3d9f\") " pod="openstack/keystone-bootstrap-kb7bl" Oct 08 19:41:03 crc kubenswrapper[4988]: I1008 19:41:03.266192 4988 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d45cfd02-b2f1-41ec-8751-6378224432ec" path="/var/lib/kubelet/pods/d45cfd02-b2f1-41ec-8751-6378224432ec/volumes" Oct 08 19:41:03 crc kubenswrapper[4988]: I1008 19:41:03.271965 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/94108d41-9804-48e5-982a-6e0df6ba3d9f-credential-keys\") pod \"keystone-bootstrap-kb7bl\" (UID: \"94108d41-9804-48e5-982a-6e0df6ba3d9f\") " pod="openstack/keystone-bootstrap-kb7bl" Oct 08 19:41:03 crc kubenswrapper[4988]: I1008 19:41:03.272095 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/94108d41-9804-48e5-982a-6e0df6ba3d9f-combined-ca-bundle\") pod \"keystone-bootstrap-kb7bl\" (UID: \"94108d41-9804-48e5-982a-6e0df6ba3d9f\") " pod="openstack/keystone-bootstrap-kb7bl" Oct 08 19:41:03 crc kubenswrapper[4988]: I1008 19:41:03.272147 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/94108d41-9804-48e5-982a-6e0df6ba3d9f-scripts\") pod \"keystone-bootstrap-kb7bl\" (UID: \"94108d41-9804-48e5-982a-6e0df6ba3d9f\") " pod="openstack/keystone-bootstrap-kb7bl" Oct 08 19:41:03 crc kubenswrapper[4988]: I1008 19:41:03.272661 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/94108d41-9804-48e5-982a-6e0df6ba3d9f-config-data\") pod \"keystone-bootstrap-kb7bl\" (UID: \"94108d41-9804-48e5-982a-6e0df6ba3d9f\") " pod="openstack/keystone-bootstrap-kb7bl" Oct 08 19:41:03 crc kubenswrapper[4988]: I1008 19:41:03.273235 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/94108d41-9804-48e5-982a-6e0df6ba3d9f-fernet-keys\") pod \"keystone-bootstrap-kb7bl\" (UID: \"94108d41-9804-48e5-982a-6e0df6ba3d9f\") " pod="openstack/keystone-bootstrap-kb7bl" Oct 08 19:41:03 crc kubenswrapper[4988]: I1008 19:41:03.283268 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c6tx9\" (UniqueName: \"kubernetes.io/projected/94108d41-9804-48e5-982a-6e0df6ba3d9f-kube-api-access-c6tx9\") pod \"keystone-bootstrap-kb7bl\" (UID: \"94108d41-9804-48e5-982a-6e0df6ba3d9f\") " pod="openstack/keystone-bootstrap-kb7bl" Oct 08 19:41:03 crc kubenswrapper[4988]: I1008 19:41:03.359970 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-kb7bl" Oct 08 19:41:03 crc kubenswrapper[4988]: W1008 19:41:03.844791 4988 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod94108d41_9804_48e5_982a_6e0df6ba3d9f.slice/crio-41d2eb526bde6796a1f6851c98dfc7efe1bb19dc75aeaa2b9017bc4965e93d95 WatchSource:0}: Error finding container 41d2eb526bde6796a1f6851c98dfc7efe1bb19dc75aeaa2b9017bc4965e93d95: Status 404 returned error can't find the container with id 41d2eb526bde6796a1f6851c98dfc7efe1bb19dc75aeaa2b9017bc4965e93d95 Oct 08 19:41:03 crc kubenswrapper[4988]: I1008 19:41:03.849328 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-kb7bl"] Oct 08 19:41:03 crc kubenswrapper[4988]: I1008 19:41:03.912805 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-kb7bl" event={"ID":"94108d41-9804-48e5-982a-6e0df6ba3d9f","Type":"ContainerStarted","Data":"41d2eb526bde6796a1f6851c98dfc7efe1bb19dc75aeaa2b9017bc4965e93d95"} Oct 08 19:41:04 crc kubenswrapper[4988]: I1008 19:41:04.928302 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-kb7bl" event={"ID":"94108d41-9804-48e5-982a-6e0df6ba3d9f","Type":"ContainerStarted","Data":"36fd8e868f35d0f857726dd30cd07286a2b84b178bd372dde963221cb2f09f60"} Oct 08 19:41:04 crc kubenswrapper[4988]: I1008 19:41:04.950140 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-kb7bl" podStartSLOduration=1.950125141 podStartE2EDuration="1.950125141s" podCreationTimestamp="2025-10-08 19:41:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 19:41:04.946506736 +0000 UTC m=+5410.396349516" watchObservedRunningTime="2025-10-08 19:41:04.950125141 +0000 UTC m=+5410.399967931" Oct 08 19:41:06 crc kubenswrapper[4988]: I1008 19:41:06.364352 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-6c7c55f8b9-wlltj" Oct 08 19:41:06 crc kubenswrapper[4988]: I1008 19:41:06.465172 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-69864b896c-pw7dl"] Oct 08 19:41:06 crc kubenswrapper[4988]: I1008 19:41:06.465422 4988 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-69864b896c-pw7dl" podUID="4babecb5-3754-4372-8322-6bdddbc06931" containerName="dnsmasq-dns" containerID="cri-o://d1805129113610bb5b101c52737138b7d3b595ac13a200fb458d8b1748808654" gracePeriod=10 Oct 08 19:41:06 crc kubenswrapper[4988]: I1008 19:41:06.925673 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-69864b896c-pw7dl" Oct 08 19:41:06 crc kubenswrapper[4988]: I1008 19:41:06.947899 4988 generic.go:334] "Generic (PLEG): container finished" podID="4babecb5-3754-4372-8322-6bdddbc06931" containerID="d1805129113610bb5b101c52737138b7d3b595ac13a200fb458d8b1748808654" exitCode=0 Oct 08 19:41:06 crc kubenswrapper[4988]: I1008 19:41:06.947963 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-69864b896c-pw7dl" Oct 08 19:41:06 crc kubenswrapper[4988]: I1008 19:41:06.947961 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-69864b896c-pw7dl" event={"ID":"4babecb5-3754-4372-8322-6bdddbc06931","Type":"ContainerDied","Data":"d1805129113610bb5b101c52737138b7d3b595ac13a200fb458d8b1748808654"} Oct 08 19:41:06 crc kubenswrapper[4988]: I1008 19:41:06.948184 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-69864b896c-pw7dl" event={"ID":"4babecb5-3754-4372-8322-6bdddbc06931","Type":"ContainerDied","Data":"5d181cce6c5543682377478f69f2fdd9f09e4180f5e47455d0f0fd28384e7b0e"} Oct 08 19:41:06 crc kubenswrapper[4988]: I1008 19:41:06.948207 4988 scope.go:117] "RemoveContainer" containerID="d1805129113610bb5b101c52737138b7d3b595ac13a200fb458d8b1748808654" Oct 08 19:41:06 crc kubenswrapper[4988]: I1008 19:41:06.952800 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/4babecb5-3754-4372-8322-6bdddbc06931-ovsdbserver-sb\") pod \"4babecb5-3754-4372-8322-6bdddbc06931\" (UID: \"4babecb5-3754-4372-8322-6bdddbc06931\") " Oct 08 19:41:06 crc kubenswrapper[4988]: I1008 19:41:06.952872 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4babecb5-3754-4372-8322-6bdddbc06931-config\") pod \"4babecb5-3754-4372-8322-6bdddbc06931\" (UID: \"4babecb5-3754-4372-8322-6bdddbc06931\") " Oct 08 19:41:06 crc kubenswrapper[4988]: I1008 19:41:06.973667 4988 scope.go:117] "RemoveContainer" containerID="238814049da025e35cf5de0e82820d2f2b2e7db3b9e15fe4e489117316b60cdc" Oct 08 19:41:07 crc kubenswrapper[4988]: I1008 19:41:07.006520 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4babecb5-3754-4372-8322-6bdddbc06931-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "4babecb5-3754-4372-8322-6bdddbc06931" (UID: "4babecb5-3754-4372-8322-6bdddbc06931"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 19:41:07 crc kubenswrapper[4988]: I1008 19:41:07.009679 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4babecb5-3754-4372-8322-6bdddbc06931-config" (OuterVolumeSpecName: "config") pod "4babecb5-3754-4372-8322-6bdddbc06931" (UID: "4babecb5-3754-4372-8322-6bdddbc06931"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 19:41:07 crc kubenswrapper[4988]: I1008 19:41:07.053845 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4babecb5-3754-4372-8322-6bdddbc06931-dns-svc\") pod \"4babecb5-3754-4372-8322-6bdddbc06931\" (UID: \"4babecb5-3754-4372-8322-6bdddbc06931\") " Oct 08 19:41:07 crc kubenswrapper[4988]: I1008 19:41:07.053906 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4babecb5-3754-4372-8322-6bdddbc06931-ovsdbserver-nb\") pod \"4babecb5-3754-4372-8322-6bdddbc06931\" (UID: \"4babecb5-3754-4372-8322-6bdddbc06931\") " Oct 08 19:41:07 crc kubenswrapper[4988]: I1008 19:41:07.054001 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s7zwt\" (UniqueName: \"kubernetes.io/projected/4babecb5-3754-4372-8322-6bdddbc06931-kube-api-access-s7zwt\") pod \"4babecb5-3754-4372-8322-6bdddbc06931\" (UID: \"4babecb5-3754-4372-8322-6bdddbc06931\") " Oct 08 19:41:07 crc kubenswrapper[4988]: I1008 19:41:07.054353 4988 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/4babecb5-3754-4372-8322-6bdddbc06931-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 08 19:41:07 crc kubenswrapper[4988]: I1008 19:41:07.054372 4988 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4babecb5-3754-4372-8322-6bdddbc06931-config\") on node \"crc\" DevicePath \"\"" Oct 08 19:41:07 crc kubenswrapper[4988]: I1008 19:41:07.056560 4988 scope.go:117] "RemoveContainer" containerID="d1805129113610bb5b101c52737138b7d3b595ac13a200fb458d8b1748808654" Oct 08 19:41:07 crc kubenswrapper[4988]: I1008 19:41:07.057210 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4babecb5-3754-4372-8322-6bdddbc06931-kube-api-access-s7zwt" (OuterVolumeSpecName: "kube-api-access-s7zwt") pod "4babecb5-3754-4372-8322-6bdddbc06931" (UID: "4babecb5-3754-4372-8322-6bdddbc06931"). InnerVolumeSpecName "kube-api-access-s7zwt". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 19:41:07 crc kubenswrapper[4988]: E1008 19:41:07.061702 4988 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d1805129113610bb5b101c52737138b7d3b595ac13a200fb458d8b1748808654\": container with ID starting with d1805129113610bb5b101c52737138b7d3b595ac13a200fb458d8b1748808654 not found: ID does not exist" containerID="d1805129113610bb5b101c52737138b7d3b595ac13a200fb458d8b1748808654" Oct 08 19:41:07 crc kubenswrapper[4988]: I1008 19:41:07.061751 4988 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d1805129113610bb5b101c52737138b7d3b595ac13a200fb458d8b1748808654"} err="failed to get container status \"d1805129113610bb5b101c52737138b7d3b595ac13a200fb458d8b1748808654\": rpc error: code = NotFound desc = could not find container \"d1805129113610bb5b101c52737138b7d3b595ac13a200fb458d8b1748808654\": container with ID starting with d1805129113610bb5b101c52737138b7d3b595ac13a200fb458d8b1748808654 not found: ID does not exist" Oct 08 19:41:07 crc kubenswrapper[4988]: I1008 19:41:07.061778 4988 scope.go:117] "RemoveContainer" containerID="238814049da025e35cf5de0e82820d2f2b2e7db3b9e15fe4e489117316b60cdc" Oct 08 19:41:07 crc kubenswrapper[4988]: E1008 19:41:07.062207 4988 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"238814049da025e35cf5de0e82820d2f2b2e7db3b9e15fe4e489117316b60cdc\": container with ID starting with 238814049da025e35cf5de0e82820d2f2b2e7db3b9e15fe4e489117316b60cdc not found: ID does not exist" containerID="238814049da025e35cf5de0e82820d2f2b2e7db3b9e15fe4e489117316b60cdc" Oct 08 19:41:07 crc kubenswrapper[4988]: I1008 19:41:07.062241 4988 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"238814049da025e35cf5de0e82820d2f2b2e7db3b9e15fe4e489117316b60cdc"} err="failed to get container status \"238814049da025e35cf5de0e82820d2f2b2e7db3b9e15fe4e489117316b60cdc\": rpc error: code = NotFound desc = could not find container \"238814049da025e35cf5de0e82820d2f2b2e7db3b9e15fe4e489117316b60cdc\": container with ID starting with 238814049da025e35cf5de0e82820d2f2b2e7db3b9e15fe4e489117316b60cdc not found: ID does not exist" Oct 08 19:41:07 crc kubenswrapper[4988]: I1008 19:41:07.086657 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4babecb5-3754-4372-8322-6bdddbc06931-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "4babecb5-3754-4372-8322-6bdddbc06931" (UID: "4babecb5-3754-4372-8322-6bdddbc06931"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 19:41:07 crc kubenswrapper[4988]: I1008 19:41:07.090441 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4babecb5-3754-4372-8322-6bdddbc06931-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "4babecb5-3754-4372-8322-6bdddbc06931" (UID: "4babecb5-3754-4372-8322-6bdddbc06931"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 19:41:07 crc kubenswrapper[4988]: I1008 19:41:07.156156 4988 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4babecb5-3754-4372-8322-6bdddbc06931-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 08 19:41:07 crc kubenswrapper[4988]: I1008 19:41:07.156205 4988 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4babecb5-3754-4372-8322-6bdddbc06931-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 08 19:41:07 crc kubenswrapper[4988]: I1008 19:41:07.156221 4988 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s7zwt\" (UniqueName: \"kubernetes.io/projected/4babecb5-3754-4372-8322-6bdddbc06931-kube-api-access-s7zwt\") on node \"crc\" DevicePath \"\"" Oct 08 19:41:07 crc kubenswrapper[4988]: I1008 19:41:07.288830 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-69864b896c-pw7dl"] Oct 08 19:41:07 crc kubenswrapper[4988]: I1008 19:41:07.300923 4988 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-69864b896c-pw7dl"] Oct 08 19:41:07 crc kubenswrapper[4988]: I1008 19:41:07.966641 4988 generic.go:334] "Generic (PLEG): container finished" podID="94108d41-9804-48e5-982a-6e0df6ba3d9f" containerID="36fd8e868f35d0f857726dd30cd07286a2b84b178bd372dde963221cb2f09f60" exitCode=0 Oct 08 19:41:07 crc kubenswrapper[4988]: I1008 19:41:07.966718 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-kb7bl" event={"ID":"94108d41-9804-48e5-982a-6e0df6ba3d9f","Type":"ContainerDied","Data":"36fd8e868f35d0f857726dd30cd07286a2b84b178bd372dde963221cb2f09f60"} Oct 08 19:41:08 crc kubenswrapper[4988]: I1008 19:41:08.366739 4988 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-wlkr2" Oct 08 19:41:08 crc kubenswrapper[4988]: I1008 19:41:08.367265 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-wlkr2" Oct 08 19:41:08 crc kubenswrapper[4988]: I1008 19:41:08.451015 4988 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-wlkr2" Oct 08 19:41:09 crc kubenswrapper[4988]: I1008 19:41:09.035471 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-wlkr2" Oct 08 19:41:09 crc kubenswrapper[4988]: I1008 19:41:09.103501 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-wlkr2"] Oct 08 19:41:09 crc kubenswrapper[4988]: I1008 19:41:09.270386 4988 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4babecb5-3754-4372-8322-6bdddbc06931" path="/var/lib/kubelet/pods/4babecb5-3754-4372-8322-6bdddbc06931/volumes" Oct 08 19:41:09 crc kubenswrapper[4988]: I1008 19:41:09.412606 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-kb7bl" Oct 08 19:41:09 crc kubenswrapper[4988]: I1008 19:41:09.538459 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/94108d41-9804-48e5-982a-6e0df6ba3d9f-combined-ca-bundle\") pod \"94108d41-9804-48e5-982a-6e0df6ba3d9f\" (UID: \"94108d41-9804-48e5-982a-6e0df6ba3d9f\") " Oct 08 19:41:09 crc kubenswrapper[4988]: I1008 19:41:09.538565 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c6tx9\" (UniqueName: \"kubernetes.io/projected/94108d41-9804-48e5-982a-6e0df6ba3d9f-kube-api-access-c6tx9\") pod \"94108d41-9804-48e5-982a-6e0df6ba3d9f\" (UID: \"94108d41-9804-48e5-982a-6e0df6ba3d9f\") " Oct 08 19:41:09 crc kubenswrapper[4988]: I1008 19:41:09.538652 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/94108d41-9804-48e5-982a-6e0df6ba3d9f-config-data\") pod \"94108d41-9804-48e5-982a-6e0df6ba3d9f\" (UID: \"94108d41-9804-48e5-982a-6e0df6ba3d9f\") " Oct 08 19:41:09 crc kubenswrapper[4988]: I1008 19:41:09.538719 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/94108d41-9804-48e5-982a-6e0df6ba3d9f-fernet-keys\") pod \"94108d41-9804-48e5-982a-6e0df6ba3d9f\" (UID: \"94108d41-9804-48e5-982a-6e0df6ba3d9f\") " Oct 08 19:41:09 crc kubenswrapper[4988]: I1008 19:41:09.538758 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/94108d41-9804-48e5-982a-6e0df6ba3d9f-credential-keys\") pod \"94108d41-9804-48e5-982a-6e0df6ba3d9f\" (UID: \"94108d41-9804-48e5-982a-6e0df6ba3d9f\") " Oct 08 19:41:09 crc kubenswrapper[4988]: I1008 19:41:09.538797 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/94108d41-9804-48e5-982a-6e0df6ba3d9f-scripts\") pod \"94108d41-9804-48e5-982a-6e0df6ba3d9f\" (UID: \"94108d41-9804-48e5-982a-6e0df6ba3d9f\") " Oct 08 19:41:09 crc kubenswrapper[4988]: I1008 19:41:09.544559 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/94108d41-9804-48e5-982a-6e0df6ba3d9f-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "94108d41-9804-48e5-982a-6e0df6ba3d9f" (UID: "94108d41-9804-48e5-982a-6e0df6ba3d9f"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 19:41:09 crc kubenswrapper[4988]: I1008 19:41:09.544763 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/94108d41-9804-48e5-982a-6e0df6ba3d9f-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "94108d41-9804-48e5-982a-6e0df6ba3d9f" (UID: "94108d41-9804-48e5-982a-6e0df6ba3d9f"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 19:41:09 crc kubenswrapper[4988]: I1008 19:41:09.546807 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/94108d41-9804-48e5-982a-6e0df6ba3d9f-kube-api-access-c6tx9" (OuterVolumeSpecName: "kube-api-access-c6tx9") pod "94108d41-9804-48e5-982a-6e0df6ba3d9f" (UID: "94108d41-9804-48e5-982a-6e0df6ba3d9f"). InnerVolumeSpecName "kube-api-access-c6tx9". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 19:41:09 crc kubenswrapper[4988]: I1008 19:41:09.546906 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/94108d41-9804-48e5-982a-6e0df6ba3d9f-scripts" (OuterVolumeSpecName: "scripts") pod "94108d41-9804-48e5-982a-6e0df6ba3d9f" (UID: "94108d41-9804-48e5-982a-6e0df6ba3d9f"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 19:41:09 crc kubenswrapper[4988]: I1008 19:41:09.563039 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/94108d41-9804-48e5-982a-6e0df6ba3d9f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "94108d41-9804-48e5-982a-6e0df6ba3d9f" (UID: "94108d41-9804-48e5-982a-6e0df6ba3d9f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 19:41:09 crc kubenswrapper[4988]: I1008 19:41:09.569254 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/94108d41-9804-48e5-982a-6e0df6ba3d9f-config-data" (OuterVolumeSpecName: "config-data") pod "94108d41-9804-48e5-982a-6e0df6ba3d9f" (UID: "94108d41-9804-48e5-982a-6e0df6ba3d9f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 19:41:09 crc kubenswrapper[4988]: I1008 19:41:09.642389 4988 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c6tx9\" (UniqueName: \"kubernetes.io/projected/94108d41-9804-48e5-982a-6e0df6ba3d9f-kube-api-access-c6tx9\") on node \"crc\" DevicePath \"\"" Oct 08 19:41:09 crc kubenswrapper[4988]: I1008 19:41:09.642437 4988 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/94108d41-9804-48e5-982a-6e0df6ba3d9f-config-data\") on node \"crc\" DevicePath \"\"" Oct 08 19:41:09 crc kubenswrapper[4988]: I1008 19:41:09.642462 4988 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/94108d41-9804-48e5-982a-6e0df6ba3d9f-fernet-keys\") on node \"crc\" DevicePath \"\"" Oct 08 19:41:09 crc kubenswrapper[4988]: I1008 19:41:09.642471 4988 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/94108d41-9804-48e5-982a-6e0df6ba3d9f-credential-keys\") on node \"crc\" DevicePath \"\"" Oct 08 19:41:09 crc kubenswrapper[4988]: I1008 19:41:09.642497 4988 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/94108d41-9804-48e5-982a-6e0df6ba3d9f-scripts\") on node \"crc\" DevicePath \"\"" Oct 08 19:41:09 crc kubenswrapper[4988]: I1008 19:41:09.642506 4988 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/94108d41-9804-48e5-982a-6e0df6ba3d9f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 19:41:09 crc kubenswrapper[4988]: I1008 19:41:09.994172 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-kb7bl" event={"ID":"94108d41-9804-48e5-982a-6e0df6ba3d9f","Type":"ContainerDied","Data":"41d2eb526bde6796a1f6851c98dfc7efe1bb19dc75aeaa2b9017bc4965e93d95"} Oct 08 19:41:09 crc kubenswrapper[4988]: I1008 19:41:09.994214 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-kb7bl" Oct 08 19:41:09 crc kubenswrapper[4988]: I1008 19:41:09.994224 4988 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="41d2eb526bde6796a1f6851c98dfc7efe1bb19dc75aeaa2b9017bc4965e93d95" Oct 08 19:41:10 crc kubenswrapper[4988]: I1008 19:41:10.122963 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-7c6d678f8f-dxth8"] Oct 08 19:41:10 crc kubenswrapper[4988]: E1008 19:41:10.123371 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="94108d41-9804-48e5-982a-6e0df6ba3d9f" containerName="keystone-bootstrap" Oct 08 19:41:10 crc kubenswrapper[4988]: I1008 19:41:10.123388 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="94108d41-9804-48e5-982a-6e0df6ba3d9f" containerName="keystone-bootstrap" Oct 08 19:41:10 crc kubenswrapper[4988]: E1008 19:41:10.123424 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4babecb5-3754-4372-8322-6bdddbc06931" containerName="init" Oct 08 19:41:10 crc kubenswrapper[4988]: I1008 19:41:10.123432 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="4babecb5-3754-4372-8322-6bdddbc06931" containerName="init" Oct 08 19:41:10 crc kubenswrapper[4988]: E1008 19:41:10.123469 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4babecb5-3754-4372-8322-6bdddbc06931" containerName="dnsmasq-dns" Oct 08 19:41:10 crc kubenswrapper[4988]: I1008 19:41:10.123477 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="4babecb5-3754-4372-8322-6bdddbc06931" containerName="dnsmasq-dns" Oct 08 19:41:10 crc kubenswrapper[4988]: I1008 19:41:10.123656 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="4babecb5-3754-4372-8322-6bdddbc06931" containerName="dnsmasq-dns" Oct 08 19:41:10 crc kubenswrapper[4988]: I1008 19:41:10.123710 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="94108d41-9804-48e5-982a-6e0df6ba3d9f" containerName="keystone-bootstrap" Oct 08 19:41:10 crc kubenswrapper[4988]: I1008 19:41:10.124445 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-7c6d678f8f-dxth8" Oct 08 19:41:10 crc kubenswrapper[4988]: I1008 19:41:10.127371 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Oct 08 19:41:10 crc kubenswrapper[4988]: I1008 19:41:10.127580 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Oct 08 19:41:10 crc kubenswrapper[4988]: I1008 19:41:10.127643 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Oct 08 19:41:10 crc kubenswrapper[4988]: I1008 19:41:10.127372 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-992fc" Oct 08 19:41:10 crc kubenswrapper[4988]: I1008 19:41:10.127863 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-keystone-public-svc" Oct 08 19:41:10 crc kubenswrapper[4988]: I1008 19:41:10.127986 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-keystone-internal-svc" Oct 08 19:41:10 crc kubenswrapper[4988]: I1008 19:41:10.135930 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-7c6d678f8f-dxth8"] Oct 08 19:41:10 crc kubenswrapper[4988]: I1008 19:41:10.252773 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a92eb88b-08c7-4f05-abdf-87777d7d7170-config-data\") pod \"keystone-7c6d678f8f-dxth8\" (UID: \"a92eb88b-08c7-4f05-abdf-87777d7d7170\") " pod="openstack/keystone-7c6d678f8f-dxth8" Oct 08 19:41:10 crc kubenswrapper[4988]: I1008 19:41:10.252828 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/a92eb88b-08c7-4f05-abdf-87777d7d7170-credential-keys\") pod \"keystone-7c6d678f8f-dxth8\" (UID: \"a92eb88b-08c7-4f05-abdf-87777d7d7170\") " pod="openstack/keystone-7c6d678f8f-dxth8" Oct 08 19:41:10 crc kubenswrapper[4988]: I1008 19:41:10.252847 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/a92eb88b-08c7-4f05-abdf-87777d7d7170-internal-tls-certs\") pod \"keystone-7c6d678f8f-dxth8\" (UID: \"a92eb88b-08c7-4f05-abdf-87777d7d7170\") " pod="openstack/keystone-7c6d678f8f-dxth8" Oct 08 19:41:10 crc kubenswrapper[4988]: I1008 19:41:10.252939 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/a92eb88b-08c7-4f05-abdf-87777d7d7170-public-tls-certs\") pod \"keystone-7c6d678f8f-dxth8\" (UID: \"a92eb88b-08c7-4f05-abdf-87777d7d7170\") " pod="openstack/keystone-7c6d678f8f-dxth8" Oct 08 19:41:10 crc kubenswrapper[4988]: I1008 19:41:10.252975 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2hjzf\" (UniqueName: \"kubernetes.io/projected/a92eb88b-08c7-4f05-abdf-87777d7d7170-kube-api-access-2hjzf\") pod \"keystone-7c6d678f8f-dxth8\" (UID: \"a92eb88b-08c7-4f05-abdf-87777d7d7170\") " pod="openstack/keystone-7c6d678f8f-dxth8" Oct 08 19:41:10 crc kubenswrapper[4988]: I1008 19:41:10.253052 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a92eb88b-08c7-4f05-abdf-87777d7d7170-combined-ca-bundle\") pod \"keystone-7c6d678f8f-dxth8\" (UID: \"a92eb88b-08c7-4f05-abdf-87777d7d7170\") " pod="openstack/keystone-7c6d678f8f-dxth8" Oct 08 19:41:10 crc kubenswrapper[4988]: I1008 19:41:10.253128 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/a92eb88b-08c7-4f05-abdf-87777d7d7170-fernet-keys\") pod \"keystone-7c6d678f8f-dxth8\" (UID: \"a92eb88b-08c7-4f05-abdf-87777d7d7170\") " pod="openstack/keystone-7c6d678f8f-dxth8" Oct 08 19:41:10 crc kubenswrapper[4988]: I1008 19:41:10.253181 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a92eb88b-08c7-4f05-abdf-87777d7d7170-scripts\") pod \"keystone-7c6d678f8f-dxth8\" (UID: \"a92eb88b-08c7-4f05-abdf-87777d7d7170\") " pod="openstack/keystone-7c6d678f8f-dxth8" Oct 08 19:41:10 crc kubenswrapper[4988]: I1008 19:41:10.355303 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a92eb88b-08c7-4f05-abdf-87777d7d7170-combined-ca-bundle\") pod \"keystone-7c6d678f8f-dxth8\" (UID: \"a92eb88b-08c7-4f05-abdf-87777d7d7170\") " pod="openstack/keystone-7c6d678f8f-dxth8" Oct 08 19:41:10 crc kubenswrapper[4988]: I1008 19:41:10.355362 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/a92eb88b-08c7-4f05-abdf-87777d7d7170-fernet-keys\") pod \"keystone-7c6d678f8f-dxth8\" (UID: \"a92eb88b-08c7-4f05-abdf-87777d7d7170\") " pod="openstack/keystone-7c6d678f8f-dxth8" Oct 08 19:41:10 crc kubenswrapper[4988]: I1008 19:41:10.355436 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a92eb88b-08c7-4f05-abdf-87777d7d7170-scripts\") pod \"keystone-7c6d678f8f-dxth8\" (UID: \"a92eb88b-08c7-4f05-abdf-87777d7d7170\") " pod="openstack/keystone-7c6d678f8f-dxth8" Oct 08 19:41:10 crc kubenswrapper[4988]: I1008 19:41:10.355602 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a92eb88b-08c7-4f05-abdf-87777d7d7170-config-data\") pod \"keystone-7c6d678f8f-dxth8\" (UID: \"a92eb88b-08c7-4f05-abdf-87777d7d7170\") " pod="openstack/keystone-7c6d678f8f-dxth8" Oct 08 19:41:10 crc kubenswrapper[4988]: I1008 19:41:10.355646 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/a92eb88b-08c7-4f05-abdf-87777d7d7170-credential-keys\") pod \"keystone-7c6d678f8f-dxth8\" (UID: \"a92eb88b-08c7-4f05-abdf-87777d7d7170\") " pod="openstack/keystone-7c6d678f8f-dxth8" Oct 08 19:41:10 crc kubenswrapper[4988]: I1008 19:41:10.355669 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/a92eb88b-08c7-4f05-abdf-87777d7d7170-internal-tls-certs\") pod \"keystone-7c6d678f8f-dxth8\" (UID: \"a92eb88b-08c7-4f05-abdf-87777d7d7170\") " pod="openstack/keystone-7c6d678f8f-dxth8" Oct 08 19:41:10 crc kubenswrapper[4988]: I1008 19:41:10.355718 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/a92eb88b-08c7-4f05-abdf-87777d7d7170-public-tls-certs\") pod \"keystone-7c6d678f8f-dxth8\" (UID: \"a92eb88b-08c7-4f05-abdf-87777d7d7170\") " pod="openstack/keystone-7c6d678f8f-dxth8" Oct 08 19:41:10 crc kubenswrapper[4988]: I1008 19:41:10.355744 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2hjzf\" (UniqueName: \"kubernetes.io/projected/a92eb88b-08c7-4f05-abdf-87777d7d7170-kube-api-access-2hjzf\") pod \"keystone-7c6d678f8f-dxth8\" (UID: \"a92eb88b-08c7-4f05-abdf-87777d7d7170\") " pod="openstack/keystone-7c6d678f8f-dxth8" Oct 08 19:41:10 crc kubenswrapper[4988]: I1008 19:41:10.358948 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a92eb88b-08c7-4f05-abdf-87777d7d7170-scripts\") pod \"keystone-7c6d678f8f-dxth8\" (UID: \"a92eb88b-08c7-4f05-abdf-87777d7d7170\") " pod="openstack/keystone-7c6d678f8f-dxth8" Oct 08 19:41:10 crc kubenswrapper[4988]: I1008 19:41:10.359711 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/a92eb88b-08c7-4f05-abdf-87777d7d7170-internal-tls-certs\") pod \"keystone-7c6d678f8f-dxth8\" (UID: \"a92eb88b-08c7-4f05-abdf-87777d7d7170\") " pod="openstack/keystone-7c6d678f8f-dxth8" Oct 08 19:41:10 crc kubenswrapper[4988]: I1008 19:41:10.359744 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/a92eb88b-08c7-4f05-abdf-87777d7d7170-public-tls-certs\") pod \"keystone-7c6d678f8f-dxth8\" (UID: \"a92eb88b-08c7-4f05-abdf-87777d7d7170\") " pod="openstack/keystone-7c6d678f8f-dxth8" Oct 08 19:41:10 crc kubenswrapper[4988]: I1008 19:41:10.359796 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/a92eb88b-08c7-4f05-abdf-87777d7d7170-fernet-keys\") pod \"keystone-7c6d678f8f-dxth8\" (UID: \"a92eb88b-08c7-4f05-abdf-87777d7d7170\") " pod="openstack/keystone-7c6d678f8f-dxth8" Oct 08 19:41:10 crc kubenswrapper[4988]: I1008 19:41:10.360265 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a92eb88b-08c7-4f05-abdf-87777d7d7170-combined-ca-bundle\") pod \"keystone-7c6d678f8f-dxth8\" (UID: \"a92eb88b-08c7-4f05-abdf-87777d7d7170\") " pod="openstack/keystone-7c6d678f8f-dxth8" Oct 08 19:41:10 crc kubenswrapper[4988]: I1008 19:41:10.360535 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/a92eb88b-08c7-4f05-abdf-87777d7d7170-credential-keys\") pod \"keystone-7c6d678f8f-dxth8\" (UID: \"a92eb88b-08c7-4f05-abdf-87777d7d7170\") " pod="openstack/keystone-7c6d678f8f-dxth8" Oct 08 19:41:10 crc kubenswrapper[4988]: I1008 19:41:10.372209 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a92eb88b-08c7-4f05-abdf-87777d7d7170-config-data\") pod \"keystone-7c6d678f8f-dxth8\" (UID: \"a92eb88b-08c7-4f05-abdf-87777d7d7170\") " pod="openstack/keystone-7c6d678f8f-dxth8" Oct 08 19:41:10 crc kubenswrapper[4988]: I1008 19:41:10.374682 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2hjzf\" (UniqueName: \"kubernetes.io/projected/a92eb88b-08c7-4f05-abdf-87777d7d7170-kube-api-access-2hjzf\") pod \"keystone-7c6d678f8f-dxth8\" (UID: \"a92eb88b-08c7-4f05-abdf-87777d7d7170\") " pod="openstack/keystone-7c6d678f8f-dxth8" Oct 08 19:41:10 crc kubenswrapper[4988]: I1008 19:41:10.452167 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-7c6d678f8f-dxth8" Oct 08 19:41:10 crc kubenswrapper[4988]: I1008 19:41:10.987833 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-7c6d678f8f-dxth8"] Oct 08 19:41:11 crc kubenswrapper[4988]: I1008 19:41:11.010031 4988 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-wlkr2" podUID="ca92153f-9b18-45b1-9ba7-6d34cf132c1b" containerName="registry-server" containerID="cri-o://fb3c53030852e901fd1fca7bc8e154094e90f5254682590d512fccaae3a08221" gracePeriod=2 Oct 08 19:41:11 crc kubenswrapper[4988]: I1008 19:41:11.010140 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-7c6d678f8f-dxth8" event={"ID":"a92eb88b-08c7-4f05-abdf-87777d7d7170","Type":"ContainerStarted","Data":"f706aef4fbbf5bb952f8e554870bea9b3fa808420a4c816ef667d2c2359eae7e"} Oct 08 19:41:11 crc kubenswrapper[4988]: I1008 19:41:11.454379 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-wlkr2" Oct 08 19:41:11 crc kubenswrapper[4988]: I1008 19:41:11.479224 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ca92153f-9b18-45b1-9ba7-6d34cf132c1b-catalog-content\") pod \"ca92153f-9b18-45b1-9ba7-6d34cf132c1b\" (UID: \"ca92153f-9b18-45b1-9ba7-6d34cf132c1b\") " Oct 08 19:41:11 crc kubenswrapper[4988]: I1008 19:41:11.479283 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jv24t\" (UniqueName: \"kubernetes.io/projected/ca92153f-9b18-45b1-9ba7-6d34cf132c1b-kube-api-access-jv24t\") pod \"ca92153f-9b18-45b1-9ba7-6d34cf132c1b\" (UID: \"ca92153f-9b18-45b1-9ba7-6d34cf132c1b\") " Oct 08 19:41:11 crc kubenswrapper[4988]: I1008 19:41:11.484608 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ca92153f-9b18-45b1-9ba7-6d34cf132c1b-kube-api-access-jv24t" (OuterVolumeSpecName: "kube-api-access-jv24t") pod "ca92153f-9b18-45b1-9ba7-6d34cf132c1b" (UID: "ca92153f-9b18-45b1-9ba7-6d34cf132c1b"). InnerVolumeSpecName "kube-api-access-jv24t". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 19:41:11 crc kubenswrapper[4988]: I1008 19:41:11.560832 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ca92153f-9b18-45b1-9ba7-6d34cf132c1b-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "ca92153f-9b18-45b1-9ba7-6d34cf132c1b" (UID: "ca92153f-9b18-45b1-9ba7-6d34cf132c1b"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 19:41:11 crc kubenswrapper[4988]: I1008 19:41:11.581279 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ca92153f-9b18-45b1-9ba7-6d34cf132c1b-utilities\") pod \"ca92153f-9b18-45b1-9ba7-6d34cf132c1b\" (UID: \"ca92153f-9b18-45b1-9ba7-6d34cf132c1b\") " Oct 08 19:41:11 crc kubenswrapper[4988]: I1008 19:41:11.581812 4988 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ca92153f-9b18-45b1-9ba7-6d34cf132c1b-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 08 19:41:11 crc kubenswrapper[4988]: I1008 19:41:11.581832 4988 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jv24t\" (UniqueName: \"kubernetes.io/projected/ca92153f-9b18-45b1-9ba7-6d34cf132c1b-kube-api-access-jv24t\") on node \"crc\" DevicePath \"\"" Oct 08 19:41:11 crc kubenswrapper[4988]: I1008 19:41:11.582442 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ca92153f-9b18-45b1-9ba7-6d34cf132c1b-utilities" (OuterVolumeSpecName: "utilities") pod "ca92153f-9b18-45b1-9ba7-6d34cf132c1b" (UID: "ca92153f-9b18-45b1-9ba7-6d34cf132c1b"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 19:41:11 crc kubenswrapper[4988]: I1008 19:41:11.683528 4988 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ca92153f-9b18-45b1-9ba7-6d34cf132c1b-utilities\") on node \"crc\" DevicePath \"\"" Oct 08 19:41:12 crc kubenswrapper[4988]: I1008 19:41:12.026159 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-7c6d678f8f-dxth8" event={"ID":"a92eb88b-08c7-4f05-abdf-87777d7d7170","Type":"ContainerStarted","Data":"5c9d633cd71f76f8800bfcad583911814246d850911bf2abae91920c56aaf019"} Oct 08 19:41:12 crc kubenswrapper[4988]: I1008 19:41:12.026337 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/keystone-7c6d678f8f-dxth8" Oct 08 19:41:12 crc kubenswrapper[4988]: I1008 19:41:12.030213 4988 generic.go:334] "Generic (PLEG): container finished" podID="ca92153f-9b18-45b1-9ba7-6d34cf132c1b" containerID="fb3c53030852e901fd1fca7bc8e154094e90f5254682590d512fccaae3a08221" exitCode=0 Oct 08 19:41:12 crc kubenswrapper[4988]: I1008 19:41:12.030302 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wlkr2" event={"ID":"ca92153f-9b18-45b1-9ba7-6d34cf132c1b","Type":"ContainerDied","Data":"fb3c53030852e901fd1fca7bc8e154094e90f5254682590d512fccaae3a08221"} Oct 08 19:41:12 crc kubenswrapper[4988]: I1008 19:41:12.030346 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wlkr2" event={"ID":"ca92153f-9b18-45b1-9ba7-6d34cf132c1b","Type":"ContainerDied","Data":"dfb53089fc70ac16c9d40be0ce64d1acf8b410e5e125327ca4c544f2be405a5c"} Oct 08 19:41:12 crc kubenswrapper[4988]: I1008 19:41:12.030379 4988 scope.go:117] "RemoveContainer" containerID="fb3c53030852e901fd1fca7bc8e154094e90f5254682590d512fccaae3a08221" Oct 08 19:41:12 crc kubenswrapper[4988]: I1008 19:41:12.030713 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-wlkr2" Oct 08 19:41:12 crc kubenswrapper[4988]: I1008 19:41:12.061569 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-7c6d678f8f-dxth8" podStartSLOduration=2.061551409 podStartE2EDuration="2.061551409s" podCreationTimestamp="2025-10-08 19:41:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 19:41:12.05748328 +0000 UTC m=+5417.507326060" watchObservedRunningTime="2025-10-08 19:41:12.061551409 +0000 UTC m=+5417.511394189" Oct 08 19:41:12 crc kubenswrapper[4988]: I1008 19:41:12.077144 4988 scope.go:117] "RemoveContainer" containerID="c28033bf02d45d29e61ff71fa479f26ef008a1b25ac620b3f020b9ec443204e2" Oct 08 19:41:12 crc kubenswrapper[4988]: I1008 19:41:12.089264 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-wlkr2"] Oct 08 19:41:12 crc kubenswrapper[4988]: I1008 19:41:12.096672 4988 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-wlkr2"] Oct 08 19:41:12 crc kubenswrapper[4988]: I1008 19:41:12.107985 4988 scope.go:117] "RemoveContainer" containerID="ace2be587539f22a2afd389b7f1cd09c4cb32307275ac7cf11a1baa374045ff9" Oct 08 19:41:12 crc kubenswrapper[4988]: I1008 19:41:12.154530 4988 scope.go:117] "RemoveContainer" containerID="fb3c53030852e901fd1fca7bc8e154094e90f5254682590d512fccaae3a08221" Oct 08 19:41:12 crc kubenswrapper[4988]: E1008 19:41:12.154999 4988 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fb3c53030852e901fd1fca7bc8e154094e90f5254682590d512fccaae3a08221\": container with ID starting with fb3c53030852e901fd1fca7bc8e154094e90f5254682590d512fccaae3a08221 not found: ID does not exist" containerID="fb3c53030852e901fd1fca7bc8e154094e90f5254682590d512fccaae3a08221" Oct 08 19:41:12 crc kubenswrapper[4988]: I1008 19:41:12.155057 4988 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fb3c53030852e901fd1fca7bc8e154094e90f5254682590d512fccaae3a08221"} err="failed to get container status \"fb3c53030852e901fd1fca7bc8e154094e90f5254682590d512fccaae3a08221\": rpc error: code = NotFound desc = could not find container \"fb3c53030852e901fd1fca7bc8e154094e90f5254682590d512fccaae3a08221\": container with ID starting with fb3c53030852e901fd1fca7bc8e154094e90f5254682590d512fccaae3a08221 not found: ID does not exist" Oct 08 19:41:12 crc kubenswrapper[4988]: I1008 19:41:12.155122 4988 scope.go:117] "RemoveContainer" containerID="c28033bf02d45d29e61ff71fa479f26ef008a1b25ac620b3f020b9ec443204e2" Oct 08 19:41:12 crc kubenswrapper[4988]: E1008 19:41:12.155530 4988 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c28033bf02d45d29e61ff71fa479f26ef008a1b25ac620b3f020b9ec443204e2\": container with ID starting with c28033bf02d45d29e61ff71fa479f26ef008a1b25ac620b3f020b9ec443204e2 not found: ID does not exist" containerID="c28033bf02d45d29e61ff71fa479f26ef008a1b25ac620b3f020b9ec443204e2" Oct 08 19:41:12 crc kubenswrapper[4988]: I1008 19:41:12.155554 4988 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c28033bf02d45d29e61ff71fa479f26ef008a1b25ac620b3f020b9ec443204e2"} err="failed to get container status \"c28033bf02d45d29e61ff71fa479f26ef008a1b25ac620b3f020b9ec443204e2\": rpc error: code = NotFound desc = could not find container \"c28033bf02d45d29e61ff71fa479f26ef008a1b25ac620b3f020b9ec443204e2\": container with ID starting with c28033bf02d45d29e61ff71fa479f26ef008a1b25ac620b3f020b9ec443204e2 not found: ID does not exist" Oct 08 19:41:12 crc kubenswrapper[4988]: I1008 19:41:12.155570 4988 scope.go:117] "RemoveContainer" containerID="ace2be587539f22a2afd389b7f1cd09c4cb32307275ac7cf11a1baa374045ff9" Oct 08 19:41:12 crc kubenswrapper[4988]: E1008 19:41:12.155784 4988 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ace2be587539f22a2afd389b7f1cd09c4cb32307275ac7cf11a1baa374045ff9\": container with ID starting with ace2be587539f22a2afd389b7f1cd09c4cb32307275ac7cf11a1baa374045ff9 not found: ID does not exist" containerID="ace2be587539f22a2afd389b7f1cd09c4cb32307275ac7cf11a1baa374045ff9" Oct 08 19:41:12 crc kubenswrapper[4988]: I1008 19:41:12.155817 4988 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ace2be587539f22a2afd389b7f1cd09c4cb32307275ac7cf11a1baa374045ff9"} err="failed to get container status \"ace2be587539f22a2afd389b7f1cd09c4cb32307275ac7cf11a1baa374045ff9\": rpc error: code = NotFound desc = could not find container \"ace2be587539f22a2afd389b7f1cd09c4cb32307275ac7cf11a1baa374045ff9\": container with ID starting with ace2be587539f22a2afd389b7f1cd09c4cb32307275ac7cf11a1baa374045ff9 not found: ID does not exist" Oct 08 19:41:13 crc kubenswrapper[4988]: I1008 19:41:13.248523 4988 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ca92153f-9b18-45b1-9ba7-6d34cf132c1b" path="/var/lib/kubelet/pods/ca92153f-9b18-45b1-9ba7-6d34cf132c1b/volumes" Oct 08 19:41:42 crc kubenswrapper[4988]: I1008 19:41:42.056553 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/keystone-7c6d678f8f-dxth8" Oct 08 19:41:45 crc kubenswrapper[4988]: I1008 19:41:45.543120 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstackclient"] Oct 08 19:41:45 crc kubenswrapper[4988]: E1008 19:41:45.553268 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ca92153f-9b18-45b1-9ba7-6d34cf132c1b" containerName="extract-content" Oct 08 19:41:45 crc kubenswrapper[4988]: I1008 19:41:45.553313 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="ca92153f-9b18-45b1-9ba7-6d34cf132c1b" containerName="extract-content" Oct 08 19:41:45 crc kubenswrapper[4988]: E1008 19:41:45.553328 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ca92153f-9b18-45b1-9ba7-6d34cf132c1b" containerName="registry-server" Oct 08 19:41:45 crc kubenswrapper[4988]: I1008 19:41:45.553337 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="ca92153f-9b18-45b1-9ba7-6d34cf132c1b" containerName="registry-server" Oct 08 19:41:45 crc kubenswrapper[4988]: E1008 19:41:45.553354 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ca92153f-9b18-45b1-9ba7-6d34cf132c1b" containerName="extract-utilities" Oct 08 19:41:45 crc kubenswrapper[4988]: I1008 19:41:45.553364 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="ca92153f-9b18-45b1-9ba7-6d34cf132c1b" containerName="extract-utilities" Oct 08 19:41:45 crc kubenswrapper[4988]: I1008 19:41:45.553690 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="ca92153f-9b18-45b1-9ba7-6d34cf132c1b" containerName="registry-server" Oct 08 19:41:45 crc kubenswrapper[4988]: I1008 19:41:45.555180 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Oct 08 19:41:45 crc kubenswrapper[4988]: W1008 19:41:45.558228 4988 reflector.go:561] object-"openstack"/"openstackclient-openstackclient-dockercfg-xk79f": failed to list *v1.Secret: secrets "openstackclient-openstackclient-dockercfg-xk79f" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "openstack": no relationship found between node 'crc' and this object Oct 08 19:41:45 crc kubenswrapper[4988]: E1008 19:41:45.558294 4988 reflector.go:158] "Unhandled Error" err="object-\"openstack\"/\"openstackclient-openstackclient-dockercfg-xk79f\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"openstackclient-openstackclient-dockercfg-xk79f\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"openstack\": no relationship found between node 'crc' and this object" logger="UnhandledError" Oct 08 19:41:45 crc kubenswrapper[4988]: W1008 19:41:45.558621 4988 reflector.go:561] object-"openstack"/"openstack-config-secret": failed to list *v1.Secret: secrets "openstack-config-secret" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "openstack": no relationship found between node 'crc' and this object Oct 08 19:41:45 crc kubenswrapper[4988]: E1008 19:41:45.558660 4988 reflector.go:158] "Unhandled Error" err="object-\"openstack\"/\"openstack-config-secret\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"openstack-config-secret\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"openstack\": no relationship found between node 'crc' and this object" logger="UnhandledError" Oct 08 19:41:45 crc kubenswrapper[4988]: W1008 19:41:45.559056 4988 reflector.go:561] object-"openstack"/"openstack-config": failed to list *v1.ConfigMap: configmaps "openstack-config" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "openstack": no relationship found between node 'crc' and this object Oct 08 19:41:45 crc kubenswrapper[4988]: E1008 19:41:45.559095 4988 reflector.go:158] "Unhandled Error" err="object-\"openstack\"/\"openstack-config\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"openstack-config\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"openstack\": no relationship found between node 'crc' and this object" logger="UnhandledError" Oct 08 19:41:45 crc kubenswrapper[4988]: I1008 19:41:45.569496 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Oct 08 19:41:45 crc kubenswrapper[4988]: I1008 19:41:45.590627 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/openstackclient"] Oct 08 19:41:45 crc kubenswrapper[4988]: E1008 19:41:45.591810 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="unmounted volumes=[combined-ca-bundle kube-api-access-grk45 openstack-config openstack-config-secret], unattached volumes=[], failed to process volumes=[combined-ca-bundle kube-api-access-grk45 openstack-config openstack-config-secret]: context canceled" pod="openstack/openstackclient" podUID="fff483f2-c239-413a-8f63-873488837537" Oct 08 19:41:45 crc kubenswrapper[4988]: I1008 19:41:45.608047 4988 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/openstackclient"] Oct 08 19:41:45 crc kubenswrapper[4988]: I1008 19:41:45.616969 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstackclient"] Oct 08 19:41:45 crc kubenswrapper[4988]: I1008 19:41:45.619253 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Oct 08 19:41:45 crc kubenswrapper[4988]: I1008 19:41:45.624333 4988 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openstack/openstackclient" oldPodUID="fff483f2-c239-413a-8f63-873488837537" podUID="cd32ee99-e811-4f6a-8672-3836b7527cc9" Oct 08 19:41:45 crc kubenswrapper[4988]: I1008 19:41:45.632177 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Oct 08 19:41:45 crc kubenswrapper[4988]: I1008 19:41:45.672224 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/fff483f2-c239-413a-8f63-873488837537-openstack-config\") pod \"openstackclient\" (UID: \"fff483f2-c239-413a-8f63-873488837537\") " pod="openstack/openstackclient" Oct 08 19:41:45 crc kubenswrapper[4988]: I1008 19:41:45.672297 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fff483f2-c239-413a-8f63-873488837537-combined-ca-bundle\") pod \"openstackclient\" (UID: \"fff483f2-c239-413a-8f63-873488837537\") " pod="openstack/openstackclient" Oct 08 19:41:45 crc kubenswrapper[4988]: I1008 19:41:45.672344 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/fff483f2-c239-413a-8f63-873488837537-openstack-config-secret\") pod \"openstackclient\" (UID: \"fff483f2-c239-413a-8f63-873488837537\") " pod="openstack/openstackclient" Oct 08 19:41:45 crc kubenswrapper[4988]: I1008 19:41:45.672436 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-grk45\" (UniqueName: \"kubernetes.io/projected/fff483f2-c239-413a-8f63-873488837537-kube-api-access-grk45\") pod \"openstackclient\" (UID: \"fff483f2-c239-413a-8f63-873488837537\") " pod="openstack/openstackclient" Oct 08 19:41:45 crc kubenswrapper[4988]: I1008 19:41:45.773661 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/fff483f2-c239-413a-8f63-873488837537-openstack-config-secret\") pod \"openstackclient\" (UID: \"fff483f2-c239-413a-8f63-873488837537\") " pod="openstack/openstackclient" Oct 08 19:41:45 crc kubenswrapper[4988]: I1008 19:41:45.773706 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/cd32ee99-e811-4f6a-8672-3836b7527cc9-openstack-config-secret\") pod \"openstackclient\" (UID: \"cd32ee99-e811-4f6a-8672-3836b7527cc9\") " pod="openstack/openstackclient" Oct 08 19:41:45 crc kubenswrapper[4988]: I1008 19:41:45.773726 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cd32ee99-e811-4f6a-8672-3836b7527cc9-combined-ca-bundle\") pod \"openstackclient\" (UID: \"cd32ee99-e811-4f6a-8672-3836b7527cc9\") " pod="openstack/openstackclient" Oct 08 19:41:45 crc kubenswrapper[4988]: I1008 19:41:45.773766 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-grk45\" (UniqueName: \"kubernetes.io/projected/fff483f2-c239-413a-8f63-873488837537-kube-api-access-grk45\") pod \"openstackclient\" (UID: \"fff483f2-c239-413a-8f63-873488837537\") " pod="openstack/openstackclient" Oct 08 19:41:45 crc kubenswrapper[4988]: I1008 19:41:45.773819 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/cd32ee99-e811-4f6a-8672-3836b7527cc9-openstack-config\") pod \"openstackclient\" (UID: \"cd32ee99-e811-4f6a-8672-3836b7527cc9\") " pod="openstack/openstackclient" Oct 08 19:41:45 crc kubenswrapper[4988]: I1008 19:41:45.773847 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5lw74\" (UniqueName: \"kubernetes.io/projected/cd32ee99-e811-4f6a-8672-3836b7527cc9-kube-api-access-5lw74\") pod \"openstackclient\" (UID: \"cd32ee99-e811-4f6a-8672-3836b7527cc9\") " pod="openstack/openstackclient" Oct 08 19:41:45 crc kubenswrapper[4988]: I1008 19:41:45.773921 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/fff483f2-c239-413a-8f63-873488837537-openstack-config\") pod \"openstackclient\" (UID: \"fff483f2-c239-413a-8f63-873488837537\") " pod="openstack/openstackclient" Oct 08 19:41:45 crc kubenswrapper[4988]: I1008 19:41:45.774043 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fff483f2-c239-413a-8f63-873488837537-combined-ca-bundle\") pod \"openstackclient\" (UID: \"fff483f2-c239-413a-8f63-873488837537\") " pod="openstack/openstackclient" Oct 08 19:41:45 crc kubenswrapper[4988]: E1008 19:41:45.775812 4988 projected.go:194] Error preparing data for projected volume kube-api-access-grk45 for pod openstack/openstackclient: failed to fetch token: serviceaccounts "openstackclient-openstackclient" is forbidden: the UID in the bound object reference (fff483f2-c239-413a-8f63-873488837537) does not match the UID in record. The object might have been deleted and then recreated Oct 08 19:41:45 crc kubenswrapper[4988]: E1008 19:41:45.775925 4988 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/fff483f2-c239-413a-8f63-873488837537-kube-api-access-grk45 podName:fff483f2-c239-413a-8f63-873488837537 nodeName:}" failed. No retries permitted until 2025-10-08 19:41:46.275895766 +0000 UTC m=+5451.725738626 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-grk45" (UniqueName: "kubernetes.io/projected/fff483f2-c239-413a-8f63-873488837537-kube-api-access-grk45") pod "openstackclient" (UID: "fff483f2-c239-413a-8f63-873488837537") : failed to fetch token: serviceaccounts "openstackclient-openstackclient" is forbidden: the UID in the bound object reference (fff483f2-c239-413a-8f63-873488837537) does not match the UID in record. The object might have been deleted and then recreated Oct 08 19:41:45 crc kubenswrapper[4988]: I1008 19:41:45.783134 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fff483f2-c239-413a-8f63-873488837537-combined-ca-bundle\") pod \"openstackclient\" (UID: \"fff483f2-c239-413a-8f63-873488837537\") " pod="openstack/openstackclient" Oct 08 19:41:45 crc kubenswrapper[4988]: I1008 19:41:45.875847 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/cd32ee99-e811-4f6a-8672-3836b7527cc9-openstack-config-secret\") pod \"openstackclient\" (UID: \"cd32ee99-e811-4f6a-8672-3836b7527cc9\") " pod="openstack/openstackclient" Oct 08 19:41:45 crc kubenswrapper[4988]: I1008 19:41:45.875907 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cd32ee99-e811-4f6a-8672-3836b7527cc9-combined-ca-bundle\") pod \"openstackclient\" (UID: \"cd32ee99-e811-4f6a-8672-3836b7527cc9\") " pod="openstack/openstackclient" Oct 08 19:41:45 crc kubenswrapper[4988]: I1008 19:41:45.875995 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/cd32ee99-e811-4f6a-8672-3836b7527cc9-openstack-config\") pod \"openstackclient\" (UID: \"cd32ee99-e811-4f6a-8672-3836b7527cc9\") " pod="openstack/openstackclient" Oct 08 19:41:45 crc kubenswrapper[4988]: I1008 19:41:45.876030 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5lw74\" (UniqueName: \"kubernetes.io/projected/cd32ee99-e811-4f6a-8672-3836b7527cc9-kube-api-access-5lw74\") pod \"openstackclient\" (UID: \"cd32ee99-e811-4f6a-8672-3836b7527cc9\") " pod="openstack/openstackclient" Oct 08 19:41:45 crc kubenswrapper[4988]: I1008 19:41:45.879559 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cd32ee99-e811-4f6a-8672-3836b7527cc9-combined-ca-bundle\") pod \"openstackclient\" (UID: \"cd32ee99-e811-4f6a-8672-3836b7527cc9\") " pod="openstack/openstackclient" Oct 08 19:41:45 crc kubenswrapper[4988]: I1008 19:41:45.890905 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5lw74\" (UniqueName: \"kubernetes.io/projected/cd32ee99-e811-4f6a-8672-3836b7527cc9-kube-api-access-5lw74\") pod \"openstackclient\" (UID: \"cd32ee99-e811-4f6a-8672-3836b7527cc9\") " pod="openstack/openstackclient" Oct 08 19:41:46 crc kubenswrapper[4988]: I1008 19:41:46.285085 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-grk45\" (UniqueName: \"kubernetes.io/projected/fff483f2-c239-413a-8f63-873488837537-kube-api-access-grk45\") pod \"openstackclient\" (UID: \"fff483f2-c239-413a-8f63-873488837537\") " pod="openstack/openstackclient" Oct 08 19:41:46 crc kubenswrapper[4988]: E1008 19:41:46.288002 4988 projected.go:194] Error preparing data for projected volume kube-api-access-grk45 for pod openstack/openstackclient: failed to fetch token: serviceaccounts "openstackclient-openstackclient" is forbidden: the UID in the bound object reference (fff483f2-c239-413a-8f63-873488837537) does not match the UID in record. The object might have been deleted and then recreated Oct 08 19:41:46 crc kubenswrapper[4988]: E1008 19:41:46.288064 4988 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/fff483f2-c239-413a-8f63-873488837537-kube-api-access-grk45 podName:fff483f2-c239-413a-8f63-873488837537 nodeName:}" failed. No retries permitted until 2025-10-08 19:41:47.288047276 +0000 UTC m=+5452.737890036 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-grk45" (UniqueName: "kubernetes.io/projected/fff483f2-c239-413a-8f63-873488837537-kube-api-access-grk45") pod "openstackclient" (UID: "fff483f2-c239-413a-8f63-873488837537") : failed to fetch token: serviceaccounts "openstackclient-openstackclient" is forbidden: the UID in the bound object reference (fff483f2-c239-413a-8f63-873488837537) does not match the UID in record. The object might have been deleted and then recreated Oct 08 19:41:46 crc kubenswrapper[4988]: I1008 19:41:46.354924 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Oct 08 19:41:46 crc kubenswrapper[4988]: I1008 19:41:46.358256 4988 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openstack/openstackclient" oldPodUID="fff483f2-c239-413a-8f63-873488837537" podUID="cd32ee99-e811-4f6a-8672-3836b7527cc9" Oct 08 19:41:46 crc kubenswrapper[4988]: I1008 19:41:46.367461 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Oct 08 19:41:46 crc kubenswrapper[4988]: I1008 19:41:46.370453 4988 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openstack/openstackclient" oldPodUID="fff483f2-c239-413a-8f63-873488837537" podUID="cd32ee99-e811-4f6a-8672-3836b7527cc9" Oct 08 19:41:46 crc kubenswrapper[4988]: I1008 19:41:46.488714 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fff483f2-c239-413a-8f63-873488837537-combined-ca-bundle\") pod \"fff483f2-c239-413a-8f63-873488837537\" (UID: \"fff483f2-c239-413a-8f63-873488837537\") " Oct 08 19:41:46 crc kubenswrapper[4988]: I1008 19:41:46.489489 4988 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-grk45\" (UniqueName: \"kubernetes.io/projected/fff483f2-c239-413a-8f63-873488837537-kube-api-access-grk45\") on node \"crc\" DevicePath \"\"" Oct 08 19:41:46 crc kubenswrapper[4988]: I1008 19:41:46.494692 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fff483f2-c239-413a-8f63-873488837537-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "fff483f2-c239-413a-8f63-873488837537" (UID: "fff483f2-c239-413a-8f63-873488837537"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 19:41:46 crc kubenswrapper[4988]: I1008 19:41:46.591339 4988 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fff483f2-c239-413a-8f63-873488837537-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 19:41:46 crc kubenswrapper[4988]: E1008 19:41:46.774978 4988 secret.go:188] Couldn't get secret openstack/openstack-config-secret: failed to sync secret cache: timed out waiting for the condition Oct 08 19:41:46 crc kubenswrapper[4988]: E1008 19:41:46.775066 4988 configmap.go:193] Couldn't get configMap openstack/openstack-config: failed to sync configmap cache: timed out waiting for the condition Oct 08 19:41:46 crc kubenswrapper[4988]: E1008 19:41:46.775073 4988 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/fff483f2-c239-413a-8f63-873488837537-openstack-config-secret podName:fff483f2-c239-413a-8f63-873488837537 nodeName:}" failed. No retries permitted until 2025-10-08 19:41:47.275046586 +0000 UTC m=+5452.724889386 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "openstack-config-secret" (UniqueName: "kubernetes.io/secret/fff483f2-c239-413a-8f63-873488837537-openstack-config-secret") pod "openstackclient" (UID: "fff483f2-c239-413a-8f63-873488837537") : failed to sync secret cache: timed out waiting for the condition Oct 08 19:41:46 crc kubenswrapper[4988]: E1008 19:41:46.775149 4988 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/fff483f2-c239-413a-8f63-873488837537-openstack-config podName:fff483f2-c239-413a-8f63-873488837537 nodeName:}" failed. No retries permitted until 2025-10-08 19:41:47.275128949 +0000 UTC m=+5452.724971749 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "openstack-config" (UniqueName: "kubernetes.io/configmap/fff483f2-c239-413a-8f63-873488837537-openstack-config") pod "openstackclient" (UID: "fff483f2-c239-413a-8f63-873488837537") : failed to sync configmap cache: timed out waiting for the condition Oct 08 19:41:46 crc kubenswrapper[4988]: I1008 19:41:46.796035 4988 reconciler_common.go:293] "Volume detached for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/fff483f2-c239-413a-8f63-873488837537-openstack-config\") on node \"crc\" DevicePath \"\"" Oct 08 19:41:46 crc kubenswrapper[4988]: I1008 19:41:46.796097 4988 reconciler_common.go:293] "Volume detached for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/fff483f2-c239-413a-8f63-873488837537-openstack-config-secret\") on node \"crc\" DevicePath \"\"" Oct 08 19:41:46 crc kubenswrapper[4988]: E1008 19:41:46.876351 4988 secret.go:188] Couldn't get secret openstack/openstack-config-secret: failed to sync secret cache: timed out waiting for the condition Oct 08 19:41:46 crc kubenswrapper[4988]: E1008 19:41:46.876895 4988 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/cd32ee99-e811-4f6a-8672-3836b7527cc9-openstack-config-secret podName:cd32ee99-e811-4f6a-8672-3836b7527cc9 nodeName:}" failed. No retries permitted until 2025-10-08 19:41:47.376861292 +0000 UTC m=+5452.826704102 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "openstack-config-secret" (UniqueName: "kubernetes.io/secret/cd32ee99-e811-4f6a-8672-3836b7527cc9-openstack-config-secret") pod "openstackclient" (UID: "cd32ee99-e811-4f6a-8672-3836b7527cc9") : failed to sync secret cache: timed out waiting for the condition Oct 08 19:41:46 crc kubenswrapper[4988]: E1008 19:41:46.876509 4988 configmap.go:193] Couldn't get configMap openstack/openstack-config: failed to sync configmap cache: timed out waiting for the condition Oct 08 19:41:46 crc kubenswrapper[4988]: E1008 19:41:46.877749 4988 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/cd32ee99-e811-4f6a-8672-3836b7527cc9-openstack-config podName:cd32ee99-e811-4f6a-8672-3836b7527cc9 nodeName:}" failed. No retries permitted until 2025-10-08 19:41:47.37772057 +0000 UTC m=+5452.827563370 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "openstack-config" (UniqueName: "kubernetes.io/configmap/cd32ee99-e811-4f6a-8672-3836b7527cc9-openstack-config") pod "openstackclient" (UID: "cd32ee99-e811-4f6a-8672-3836b7527cc9") : failed to sync configmap cache: timed out waiting for the condition Oct 08 19:41:46 crc kubenswrapper[4988]: I1008 19:41:46.960769 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-config-secret" Oct 08 19:41:46 crc kubenswrapper[4988]: I1008 19:41:46.969104 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstackclient-openstackclient-dockercfg-xk79f" Oct 08 19:41:47 crc kubenswrapper[4988]: I1008 19:41:47.102424 4988 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config" Oct 08 19:41:47 crc kubenswrapper[4988]: I1008 19:41:47.260150 4988 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fff483f2-c239-413a-8f63-873488837537" path="/var/lib/kubelet/pods/fff483f2-c239-413a-8f63-873488837537/volumes" Oct 08 19:41:47 crc kubenswrapper[4988]: I1008 19:41:47.361387 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Oct 08 19:41:47 crc kubenswrapper[4988]: I1008 19:41:47.367328 4988 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openstack/openstackclient" oldPodUID="fff483f2-c239-413a-8f63-873488837537" podUID="cd32ee99-e811-4f6a-8672-3836b7527cc9" Oct 08 19:41:47 crc kubenswrapper[4988]: I1008 19:41:47.407410 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/cd32ee99-e811-4f6a-8672-3836b7527cc9-openstack-config\") pod \"openstackclient\" (UID: \"cd32ee99-e811-4f6a-8672-3836b7527cc9\") " pod="openstack/openstackclient" Oct 08 19:41:47 crc kubenswrapper[4988]: I1008 19:41:47.407571 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/cd32ee99-e811-4f6a-8672-3836b7527cc9-openstack-config-secret\") pod \"openstackclient\" (UID: \"cd32ee99-e811-4f6a-8672-3836b7527cc9\") " pod="openstack/openstackclient" Oct 08 19:41:47 crc kubenswrapper[4988]: I1008 19:41:47.408316 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/cd32ee99-e811-4f6a-8672-3836b7527cc9-openstack-config\") pod \"openstackclient\" (UID: \"cd32ee99-e811-4f6a-8672-3836b7527cc9\") " pod="openstack/openstackclient" Oct 08 19:41:47 crc kubenswrapper[4988]: I1008 19:41:47.411800 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/cd32ee99-e811-4f6a-8672-3836b7527cc9-openstack-config-secret\") pod \"openstackclient\" (UID: \"cd32ee99-e811-4f6a-8672-3836b7527cc9\") " pod="openstack/openstackclient" Oct 08 19:41:47 crc kubenswrapper[4988]: I1008 19:41:47.435978 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Oct 08 19:41:47 crc kubenswrapper[4988]: I1008 19:41:47.974099 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Oct 08 19:41:48 crc kubenswrapper[4988]: I1008 19:41:48.385155 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"cd32ee99-e811-4f6a-8672-3836b7527cc9","Type":"ContainerStarted","Data":"f20ab6daa874b15266bd84f68d85b3b1bb9566ce7795057403cb7229b3424127"} Oct 08 19:41:48 crc kubenswrapper[4988]: I1008 19:41:48.385559 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"cd32ee99-e811-4f6a-8672-3836b7527cc9","Type":"ContainerStarted","Data":"7c54db38bc96595337aa90db14197a7161f369da351862058d26b553854792e8"} Oct 08 19:41:48 crc kubenswrapper[4988]: I1008 19:41:48.410246 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstackclient" podStartSLOduration=3.410226545 podStartE2EDuration="3.410226545s" podCreationTimestamp="2025-10-08 19:41:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 19:41:48.40507414 +0000 UTC m=+5453.854916950" watchObservedRunningTime="2025-10-08 19:41:48.410226545 +0000 UTC m=+5453.860069325" Oct 08 19:42:03 crc kubenswrapper[4988]: I1008 19:42:03.096337 4988 scope.go:117] "RemoveContainer" containerID="a9930599f243f306a3e5c3973da6b1f9e04bcffe18286169854acaa0a07ef6f1" Oct 08 19:42:03 crc kubenswrapper[4988]: I1008 19:42:03.126469 4988 scope.go:117] "RemoveContainer" containerID="dcf87a67647e61ae96ca317046aa60dbfe871da46ec2397732c9a83ed077e54c" Oct 08 19:42:03 crc kubenswrapper[4988]: I1008 19:42:03.170073 4988 scope.go:117] "RemoveContainer" containerID="147b1b278899f1d5b3702571f4c2b413ed486c9aa9a015474beba8f457183f9a" Oct 08 19:42:03 crc kubenswrapper[4988]: I1008 19:42:03.211703 4988 scope.go:117] "RemoveContainer" containerID="bd8e78de51a3113df5ef108a4b3654ce67ad603f688627031a2267254b61c993" Oct 08 19:42:03 crc kubenswrapper[4988]: I1008 19:42:03.243519 4988 scope.go:117] "RemoveContainer" containerID="e9f20141a2a09213dbeecf8fa028433b50c47f2435e762d362eae8fa7625d987" Oct 08 19:42:03 crc kubenswrapper[4988]: I1008 19:42:03.279737 4988 scope.go:117] "RemoveContainer" containerID="9214b02e8010153a4c65cd88a96915ebf2fe9de3d5325d6c59491c347b8fbfb4" Oct 08 19:42:03 crc kubenswrapper[4988]: I1008 19:42:03.316154 4988 scope.go:117] "RemoveContainer" containerID="509d7603080340ee6579c4843868b1b26c3851e047d18904944d7e4c09fd834d" Oct 08 19:42:23 crc kubenswrapper[4988]: I1008 19:42:23.338153 4988 patch_prober.go:28] interesting pod/machine-config-daemon-nm54f container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 19:42:23 crc kubenswrapper[4988]: I1008 19:42:23.338824 4988 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 19:42:53 crc kubenswrapper[4988]: I1008 19:42:53.338092 4988 patch_prober.go:28] interesting pod/machine-config-daemon-nm54f container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 19:42:53 crc kubenswrapper[4988]: I1008 19:42:53.338867 4988 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 19:43:17 crc kubenswrapper[4988]: I1008 19:43:17.688298 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-8l6cs"] Oct 08 19:43:17 crc kubenswrapper[4988]: I1008 19:43:17.715648 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-8l6cs" Oct 08 19:43:17 crc kubenswrapper[4988]: I1008 19:43:17.739372 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-8l6cs"] Oct 08 19:43:17 crc kubenswrapper[4988]: I1008 19:43:17.835420 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hljzr\" (UniqueName: \"kubernetes.io/projected/10142596-954c-4381-a6bd-c1eccc39cb4b-kube-api-access-hljzr\") pod \"community-operators-8l6cs\" (UID: \"10142596-954c-4381-a6bd-c1eccc39cb4b\") " pod="openshift-marketplace/community-operators-8l6cs" Oct 08 19:43:17 crc kubenswrapper[4988]: I1008 19:43:17.835503 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/10142596-954c-4381-a6bd-c1eccc39cb4b-catalog-content\") pod \"community-operators-8l6cs\" (UID: \"10142596-954c-4381-a6bd-c1eccc39cb4b\") " pod="openshift-marketplace/community-operators-8l6cs" Oct 08 19:43:17 crc kubenswrapper[4988]: I1008 19:43:17.835546 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/10142596-954c-4381-a6bd-c1eccc39cb4b-utilities\") pod \"community-operators-8l6cs\" (UID: \"10142596-954c-4381-a6bd-c1eccc39cb4b\") " pod="openshift-marketplace/community-operators-8l6cs" Oct 08 19:43:17 crc kubenswrapper[4988]: I1008 19:43:17.937094 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hljzr\" (UniqueName: \"kubernetes.io/projected/10142596-954c-4381-a6bd-c1eccc39cb4b-kube-api-access-hljzr\") pod \"community-operators-8l6cs\" (UID: \"10142596-954c-4381-a6bd-c1eccc39cb4b\") " pod="openshift-marketplace/community-operators-8l6cs" Oct 08 19:43:17 crc kubenswrapper[4988]: I1008 19:43:17.937189 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/10142596-954c-4381-a6bd-c1eccc39cb4b-catalog-content\") pod \"community-operators-8l6cs\" (UID: \"10142596-954c-4381-a6bd-c1eccc39cb4b\") " pod="openshift-marketplace/community-operators-8l6cs" Oct 08 19:43:17 crc kubenswrapper[4988]: I1008 19:43:17.937251 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/10142596-954c-4381-a6bd-c1eccc39cb4b-utilities\") pod \"community-operators-8l6cs\" (UID: \"10142596-954c-4381-a6bd-c1eccc39cb4b\") " pod="openshift-marketplace/community-operators-8l6cs" Oct 08 19:43:17 crc kubenswrapper[4988]: I1008 19:43:17.937871 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/10142596-954c-4381-a6bd-c1eccc39cb4b-utilities\") pod \"community-operators-8l6cs\" (UID: \"10142596-954c-4381-a6bd-c1eccc39cb4b\") " pod="openshift-marketplace/community-operators-8l6cs" Oct 08 19:43:17 crc kubenswrapper[4988]: I1008 19:43:17.937872 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/10142596-954c-4381-a6bd-c1eccc39cb4b-catalog-content\") pod \"community-operators-8l6cs\" (UID: \"10142596-954c-4381-a6bd-c1eccc39cb4b\") " pod="openshift-marketplace/community-operators-8l6cs" Oct 08 19:43:17 crc kubenswrapper[4988]: I1008 19:43:17.960654 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hljzr\" (UniqueName: \"kubernetes.io/projected/10142596-954c-4381-a6bd-c1eccc39cb4b-kube-api-access-hljzr\") pod \"community-operators-8l6cs\" (UID: \"10142596-954c-4381-a6bd-c1eccc39cb4b\") " pod="openshift-marketplace/community-operators-8l6cs" Oct 08 19:43:18 crc kubenswrapper[4988]: I1008 19:43:18.050813 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-8l6cs" Oct 08 19:43:18 crc kubenswrapper[4988]: I1008 19:43:18.671524 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-8l6cs"] Oct 08 19:43:19 crc kubenswrapper[4988]: I1008 19:43:19.350946 4988 generic.go:334] "Generic (PLEG): container finished" podID="10142596-954c-4381-a6bd-c1eccc39cb4b" containerID="32e3052d1d5da059a366605ccf91aa2fa9443f5ac82b48c199729888a68bc90c" exitCode=0 Oct 08 19:43:19 crc kubenswrapper[4988]: I1008 19:43:19.351010 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-8l6cs" event={"ID":"10142596-954c-4381-a6bd-c1eccc39cb4b","Type":"ContainerDied","Data":"32e3052d1d5da059a366605ccf91aa2fa9443f5ac82b48c199729888a68bc90c"} Oct 08 19:43:19 crc kubenswrapper[4988]: I1008 19:43:19.351277 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-8l6cs" event={"ID":"10142596-954c-4381-a6bd-c1eccc39cb4b","Type":"ContainerStarted","Data":"789f3a42ac83ae670f22c311656330dbbec1e61de27e300dd3f86ba9965707cc"} Oct 08 19:43:20 crc kubenswrapper[4988]: I1008 19:43:20.362001 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-8l6cs" event={"ID":"10142596-954c-4381-a6bd-c1eccc39cb4b","Type":"ContainerStarted","Data":"fa4c9b6f07e7bc9bd80ed187a9d0420c485d126672ae50eb7c092a0d6fec581f"} Oct 08 19:43:21 crc kubenswrapper[4988]: I1008 19:43:21.373865 4988 generic.go:334] "Generic (PLEG): container finished" podID="10142596-954c-4381-a6bd-c1eccc39cb4b" containerID="fa4c9b6f07e7bc9bd80ed187a9d0420c485d126672ae50eb7c092a0d6fec581f" exitCode=0 Oct 08 19:43:21 crc kubenswrapper[4988]: I1008 19:43:21.373920 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-8l6cs" event={"ID":"10142596-954c-4381-a6bd-c1eccc39cb4b","Type":"ContainerDied","Data":"fa4c9b6f07e7bc9bd80ed187a9d0420c485d126672ae50eb7c092a0d6fec581f"} Oct 08 19:43:21 crc kubenswrapper[4988]: I1008 19:43:21.845645 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-create-htq7q"] Oct 08 19:43:21 crc kubenswrapper[4988]: I1008 19:43:21.846960 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-htq7q" Oct 08 19:43:21 crc kubenswrapper[4988]: I1008 19:43:21.861781 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-htq7q"] Oct 08 19:43:21 crc kubenswrapper[4988]: I1008 19:43:21.909255 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-prcnx\" (UniqueName: \"kubernetes.io/projected/45a3d8ec-0896-49d3-ac9b-f294c0f66f62-kube-api-access-prcnx\") pod \"barbican-db-create-htq7q\" (UID: \"45a3d8ec-0896-49d3-ac9b-f294c0f66f62\") " pod="openstack/barbican-db-create-htq7q" Oct 08 19:43:22 crc kubenswrapper[4988]: I1008 19:43:22.010864 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-prcnx\" (UniqueName: \"kubernetes.io/projected/45a3d8ec-0896-49d3-ac9b-f294c0f66f62-kube-api-access-prcnx\") pod \"barbican-db-create-htq7q\" (UID: \"45a3d8ec-0896-49d3-ac9b-f294c0f66f62\") " pod="openstack/barbican-db-create-htq7q" Oct 08 19:43:22 crc kubenswrapper[4988]: I1008 19:43:22.033600 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-prcnx\" (UniqueName: \"kubernetes.io/projected/45a3d8ec-0896-49d3-ac9b-f294c0f66f62-kube-api-access-prcnx\") pod \"barbican-db-create-htq7q\" (UID: \"45a3d8ec-0896-49d3-ac9b-f294c0f66f62\") " pod="openstack/barbican-db-create-htq7q" Oct 08 19:43:22 crc kubenswrapper[4988]: I1008 19:43:22.163286 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-htq7q" Oct 08 19:43:22 crc kubenswrapper[4988]: I1008 19:43:22.389324 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-8l6cs" event={"ID":"10142596-954c-4381-a6bd-c1eccc39cb4b","Type":"ContainerStarted","Data":"ed99f93fd6c53923af21f7f0f5d35c55e558f66c569c6921330b906ab9ae46a2"} Oct 08 19:43:22 crc kubenswrapper[4988]: I1008 19:43:22.410116 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-8l6cs" podStartSLOduration=2.842774622 podStartE2EDuration="5.410095531s" podCreationTimestamp="2025-10-08 19:43:17 +0000 UTC" firstStartedPulling="2025-10-08 19:43:19.354773399 +0000 UTC m=+5544.804616169" lastFinishedPulling="2025-10-08 19:43:21.922094308 +0000 UTC m=+5547.371937078" observedRunningTime="2025-10-08 19:43:22.403934595 +0000 UTC m=+5547.853777365" watchObservedRunningTime="2025-10-08 19:43:22.410095531 +0000 UTC m=+5547.859938301" Oct 08 19:43:22 crc kubenswrapper[4988]: I1008 19:43:22.672271 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-htq7q"] Oct 08 19:43:22 crc kubenswrapper[4988]: W1008 19:43:22.672682 4988 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod45a3d8ec_0896_49d3_ac9b_f294c0f66f62.slice/crio-853e1befb4fa7abb31723f741d33a5e5cff27e4a703258914d3f4b53a79a9dac WatchSource:0}: Error finding container 853e1befb4fa7abb31723f741d33a5e5cff27e4a703258914d3f4b53a79a9dac: Status 404 returned error can't find the container with id 853e1befb4fa7abb31723f741d33a5e5cff27e4a703258914d3f4b53a79a9dac Oct 08 19:43:23 crc kubenswrapper[4988]: I1008 19:43:23.338428 4988 patch_prober.go:28] interesting pod/machine-config-daemon-nm54f container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 19:43:23 crc kubenswrapper[4988]: I1008 19:43:23.338678 4988 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 19:43:23 crc kubenswrapper[4988]: I1008 19:43:23.338716 4988 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" Oct 08 19:43:23 crc kubenswrapper[4988]: I1008 19:43:23.339209 4988 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"bfa76a99ab92f7c2c936649e27734f81c2af70cdaf33874dac1d856a4b92f5f7"} pod="openshift-machine-config-operator/machine-config-daemon-nm54f" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 08 19:43:23 crc kubenswrapper[4988]: I1008 19:43:23.339265 4988 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" containerName="machine-config-daemon" containerID="cri-o://bfa76a99ab92f7c2c936649e27734f81c2af70cdaf33874dac1d856a4b92f5f7" gracePeriod=600 Oct 08 19:43:23 crc kubenswrapper[4988]: I1008 19:43:23.397606 4988 generic.go:334] "Generic (PLEG): container finished" podID="45a3d8ec-0896-49d3-ac9b-f294c0f66f62" containerID="d2e25e9a3e3089bde074997b30b906f7253eac35bf9c155ddd64bbb92eac324e" exitCode=0 Oct 08 19:43:23 crc kubenswrapper[4988]: I1008 19:43:23.397704 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-htq7q" event={"ID":"45a3d8ec-0896-49d3-ac9b-f294c0f66f62","Type":"ContainerDied","Data":"d2e25e9a3e3089bde074997b30b906f7253eac35bf9c155ddd64bbb92eac324e"} Oct 08 19:43:23 crc kubenswrapper[4988]: I1008 19:43:23.397744 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-htq7q" event={"ID":"45a3d8ec-0896-49d3-ac9b-f294c0f66f62","Type":"ContainerStarted","Data":"853e1befb4fa7abb31723f741d33a5e5cff27e4a703258914d3f4b53a79a9dac"} Oct 08 19:43:24 crc kubenswrapper[4988]: I1008 19:43:24.412360 4988 generic.go:334] "Generic (PLEG): container finished" podID="f328a20b-5283-407e-8a93-2ab2da599f18" containerID="bfa76a99ab92f7c2c936649e27734f81c2af70cdaf33874dac1d856a4b92f5f7" exitCode=0 Oct 08 19:43:24 crc kubenswrapper[4988]: I1008 19:43:24.412421 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" event={"ID":"f328a20b-5283-407e-8a93-2ab2da599f18","Type":"ContainerDied","Data":"bfa76a99ab92f7c2c936649e27734f81c2af70cdaf33874dac1d856a4b92f5f7"} Oct 08 19:43:24 crc kubenswrapper[4988]: I1008 19:43:24.413140 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" event={"ID":"f328a20b-5283-407e-8a93-2ab2da599f18","Type":"ContainerStarted","Data":"f2a29ed22dd7a48bbc929e65fbeceed969a669e78f65fbfb724b03dc0db6e947"} Oct 08 19:43:24 crc kubenswrapper[4988]: I1008 19:43:24.413182 4988 scope.go:117] "RemoveContainer" containerID="3cd52d2ff674f037f4b111ff08caa426d4d7977052a05e385f05362baabb8da9" Oct 08 19:43:24 crc kubenswrapper[4988]: I1008 19:43:24.795943 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-htq7q" Oct 08 19:43:24 crc kubenswrapper[4988]: I1008 19:43:24.959269 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-prcnx\" (UniqueName: \"kubernetes.io/projected/45a3d8ec-0896-49d3-ac9b-f294c0f66f62-kube-api-access-prcnx\") pod \"45a3d8ec-0896-49d3-ac9b-f294c0f66f62\" (UID: \"45a3d8ec-0896-49d3-ac9b-f294c0f66f62\") " Oct 08 19:43:24 crc kubenswrapper[4988]: I1008 19:43:24.979674 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/45a3d8ec-0896-49d3-ac9b-f294c0f66f62-kube-api-access-prcnx" (OuterVolumeSpecName: "kube-api-access-prcnx") pod "45a3d8ec-0896-49d3-ac9b-f294c0f66f62" (UID: "45a3d8ec-0896-49d3-ac9b-f294c0f66f62"). InnerVolumeSpecName "kube-api-access-prcnx". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 19:43:25 crc kubenswrapper[4988]: I1008 19:43:25.061229 4988 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-prcnx\" (UniqueName: \"kubernetes.io/projected/45a3d8ec-0896-49d3-ac9b-f294c0f66f62-kube-api-access-prcnx\") on node \"crc\" DevicePath \"\"" Oct 08 19:43:25 crc kubenswrapper[4988]: I1008 19:43:25.445490 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-htq7q" Oct 08 19:43:25 crc kubenswrapper[4988]: I1008 19:43:25.445484 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-htq7q" event={"ID":"45a3d8ec-0896-49d3-ac9b-f294c0f66f62","Type":"ContainerDied","Data":"853e1befb4fa7abb31723f741d33a5e5cff27e4a703258914d3f4b53a79a9dac"} Oct 08 19:43:25 crc kubenswrapper[4988]: I1008 19:43:25.445628 4988 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="853e1befb4fa7abb31723f741d33a5e5cff27e4a703258914d3f4b53a79a9dac" Oct 08 19:43:28 crc kubenswrapper[4988]: I1008 19:43:28.051330 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-8l6cs" Oct 08 19:43:28 crc kubenswrapper[4988]: I1008 19:43:28.052335 4988 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-8l6cs" Oct 08 19:43:28 crc kubenswrapper[4988]: I1008 19:43:28.125857 4988 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-8l6cs" Oct 08 19:43:28 crc kubenswrapper[4988]: I1008 19:43:28.533086 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-8l6cs" Oct 08 19:43:28 crc kubenswrapper[4988]: I1008 19:43:28.587419 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-8l6cs"] Oct 08 19:43:30 crc kubenswrapper[4988]: I1008 19:43:30.501668 4988 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-8l6cs" podUID="10142596-954c-4381-a6bd-c1eccc39cb4b" containerName="registry-server" containerID="cri-o://ed99f93fd6c53923af21f7f0f5d35c55e558f66c569c6921330b906ab9ae46a2" gracePeriod=2 Oct 08 19:43:31 crc kubenswrapper[4988]: I1008 19:43:31.051411 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-8l6cs" Oct 08 19:43:31 crc kubenswrapper[4988]: I1008 19:43:31.182927 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/10142596-954c-4381-a6bd-c1eccc39cb4b-catalog-content\") pod \"10142596-954c-4381-a6bd-c1eccc39cb4b\" (UID: \"10142596-954c-4381-a6bd-c1eccc39cb4b\") " Oct 08 19:43:31 crc kubenswrapper[4988]: I1008 19:43:31.183485 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hljzr\" (UniqueName: \"kubernetes.io/projected/10142596-954c-4381-a6bd-c1eccc39cb4b-kube-api-access-hljzr\") pod \"10142596-954c-4381-a6bd-c1eccc39cb4b\" (UID: \"10142596-954c-4381-a6bd-c1eccc39cb4b\") " Oct 08 19:43:31 crc kubenswrapper[4988]: I1008 19:43:31.183648 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/10142596-954c-4381-a6bd-c1eccc39cb4b-utilities\") pod \"10142596-954c-4381-a6bd-c1eccc39cb4b\" (UID: \"10142596-954c-4381-a6bd-c1eccc39cb4b\") " Oct 08 19:43:31 crc kubenswrapper[4988]: I1008 19:43:31.185043 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/10142596-954c-4381-a6bd-c1eccc39cb4b-utilities" (OuterVolumeSpecName: "utilities") pod "10142596-954c-4381-a6bd-c1eccc39cb4b" (UID: "10142596-954c-4381-a6bd-c1eccc39cb4b"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 19:43:31 crc kubenswrapper[4988]: I1008 19:43:31.193998 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/10142596-954c-4381-a6bd-c1eccc39cb4b-kube-api-access-hljzr" (OuterVolumeSpecName: "kube-api-access-hljzr") pod "10142596-954c-4381-a6bd-c1eccc39cb4b" (UID: "10142596-954c-4381-a6bd-c1eccc39cb4b"). InnerVolumeSpecName "kube-api-access-hljzr". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 19:43:31 crc kubenswrapper[4988]: I1008 19:43:31.240673 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/10142596-954c-4381-a6bd-c1eccc39cb4b-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "10142596-954c-4381-a6bd-c1eccc39cb4b" (UID: "10142596-954c-4381-a6bd-c1eccc39cb4b"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 19:43:31 crc kubenswrapper[4988]: I1008 19:43:31.286493 4988 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/10142596-954c-4381-a6bd-c1eccc39cb4b-utilities\") on node \"crc\" DevicePath \"\"" Oct 08 19:43:31 crc kubenswrapper[4988]: I1008 19:43:31.286544 4988 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/10142596-954c-4381-a6bd-c1eccc39cb4b-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 08 19:43:31 crc kubenswrapper[4988]: I1008 19:43:31.286565 4988 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hljzr\" (UniqueName: \"kubernetes.io/projected/10142596-954c-4381-a6bd-c1eccc39cb4b-kube-api-access-hljzr\") on node \"crc\" DevicePath \"\"" Oct 08 19:43:31 crc kubenswrapper[4988]: I1008 19:43:31.514516 4988 generic.go:334] "Generic (PLEG): container finished" podID="10142596-954c-4381-a6bd-c1eccc39cb4b" containerID="ed99f93fd6c53923af21f7f0f5d35c55e558f66c569c6921330b906ab9ae46a2" exitCode=0 Oct 08 19:43:31 crc kubenswrapper[4988]: I1008 19:43:31.514560 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-8l6cs" event={"ID":"10142596-954c-4381-a6bd-c1eccc39cb4b","Type":"ContainerDied","Data":"ed99f93fd6c53923af21f7f0f5d35c55e558f66c569c6921330b906ab9ae46a2"} Oct 08 19:43:31 crc kubenswrapper[4988]: I1008 19:43:31.514602 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-8l6cs" Oct 08 19:43:31 crc kubenswrapper[4988]: I1008 19:43:31.515325 4988 scope.go:117] "RemoveContainer" containerID="ed99f93fd6c53923af21f7f0f5d35c55e558f66c569c6921330b906ab9ae46a2" Oct 08 19:43:31 crc kubenswrapper[4988]: I1008 19:43:31.515246 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-8l6cs" event={"ID":"10142596-954c-4381-a6bd-c1eccc39cb4b","Type":"ContainerDied","Data":"789f3a42ac83ae670f22c311656330dbbec1e61de27e300dd3f86ba9965707cc"} Oct 08 19:43:31 crc kubenswrapper[4988]: I1008 19:43:31.544693 4988 scope.go:117] "RemoveContainer" containerID="fa4c9b6f07e7bc9bd80ed187a9d0420c485d126672ae50eb7c092a0d6fec581f" Oct 08 19:43:31 crc kubenswrapper[4988]: I1008 19:43:31.544711 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-8l6cs"] Oct 08 19:43:31 crc kubenswrapper[4988]: I1008 19:43:31.553403 4988 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-8l6cs"] Oct 08 19:43:31 crc kubenswrapper[4988]: I1008 19:43:31.564243 4988 scope.go:117] "RemoveContainer" containerID="32e3052d1d5da059a366605ccf91aa2fa9443f5ac82b48c199729888a68bc90c" Oct 08 19:43:31 crc kubenswrapper[4988]: I1008 19:43:31.603006 4988 scope.go:117] "RemoveContainer" containerID="ed99f93fd6c53923af21f7f0f5d35c55e558f66c569c6921330b906ab9ae46a2" Oct 08 19:43:31 crc kubenswrapper[4988]: E1008 19:43:31.604085 4988 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ed99f93fd6c53923af21f7f0f5d35c55e558f66c569c6921330b906ab9ae46a2\": container with ID starting with ed99f93fd6c53923af21f7f0f5d35c55e558f66c569c6921330b906ab9ae46a2 not found: ID does not exist" containerID="ed99f93fd6c53923af21f7f0f5d35c55e558f66c569c6921330b906ab9ae46a2" Oct 08 19:43:31 crc kubenswrapper[4988]: I1008 19:43:31.604113 4988 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ed99f93fd6c53923af21f7f0f5d35c55e558f66c569c6921330b906ab9ae46a2"} err="failed to get container status \"ed99f93fd6c53923af21f7f0f5d35c55e558f66c569c6921330b906ab9ae46a2\": rpc error: code = NotFound desc = could not find container \"ed99f93fd6c53923af21f7f0f5d35c55e558f66c569c6921330b906ab9ae46a2\": container with ID starting with ed99f93fd6c53923af21f7f0f5d35c55e558f66c569c6921330b906ab9ae46a2 not found: ID does not exist" Oct 08 19:43:31 crc kubenswrapper[4988]: I1008 19:43:31.604139 4988 scope.go:117] "RemoveContainer" containerID="fa4c9b6f07e7bc9bd80ed187a9d0420c485d126672ae50eb7c092a0d6fec581f" Oct 08 19:43:31 crc kubenswrapper[4988]: E1008 19:43:31.604425 4988 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fa4c9b6f07e7bc9bd80ed187a9d0420c485d126672ae50eb7c092a0d6fec581f\": container with ID starting with fa4c9b6f07e7bc9bd80ed187a9d0420c485d126672ae50eb7c092a0d6fec581f not found: ID does not exist" containerID="fa4c9b6f07e7bc9bd80ed187a9d0420c485d126672ae50eb7c092a0d6fec581f" Oct 08 19:43:31 crc kubenswrapper[4988]: I1008 19:43:31.604466 4988 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fa4c9b6f07e7bc9bd80ed187a9d0420c485d126672ae50eb7c092a0d6fec581f"} err="failed to get container status \"fa4c9b6f07e7bc9bd80ed187a9d0420c485d126672ae50eb7c092a0d6fec581f\": rpc error: code = NotFound desc = could not find container \"fa4c9b6f07e7bc9bd80ed187a9d0420c485d126672ae50eb7c092a0d6fec581f\": container with ID starting with fa4c9b6f07e7bc9bd80ed187a9d0420c485d126672ae50eb7c092a0d6fec581f not found: ID does not exist" Oct 08 19:43:31 crc kubenswrapper[4988]: I1008 19:43:31.604493 4988 scope.go:117] "RemoveContainer" containerID="32e3052d1d5da059a366605ccf91aa2fa9443f5ac82b48c199729888a68bc90c" Oct 08 19:43:31 crc kubenswrapper[4988]: E1008 19:43:31.604961 4988 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"32e3052d1d5da059a366605ccf91aa2fa9443f5ac82b48c199729888a68bc90c\": container with ID starting with 32e3052d1d5da059a366605ccf91aa2fa9443f5ac82b48c199729888a68bc90c not found: ID does not exist" containerID="32e3052d1d5da059a366605ccf91aa2fa9443f5ac82b48c199729888a68bc90c" Oct 08 19:43:31 crc kubenswrapper[4988]: I1008 19:43:31.604983 4988 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"32e3052d1d5da059a366605ccf91aa2fa9443f5ac82b48c199729888a68bc90c"} err="failed to get container status \"32e3052d1d5da059a366605ccf91aa2fa9443f5ac82b48c199729888a68bc90c\": rpc error: code = NotFound desc = could not find container \"32e3052d1d5da059a366605ccf91aa2fa9443f5ac82b48c199729888a68bc90c\": container with ID starting with 32e3052d1d5da059a366605ccf91aa2fa9443f5ac82b48c199729888a68bc90c not found: ID does not exist" Oct 08 19:43:31 crc kubenswrapper[4988]: I1008 19:43:31.863922 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-1c50-account-create-28g6x"] Oct 08 19:43:31 crc kubenswrapper[4988]: E1008 19:43:31.864435 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="10142596-954c-4381-a6bd-c1eccc39cb4b" containerName="registry-server" Oct 08 19:43:31 crc kubenswrapper[4988]: I1008 19:43:31.864469 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="10142596-954c-4381-a6bd-c1eccc39cb4b" containerName="registry-server" Oct 08 19:43:31 crc kubenswrapper[4988]: E1008 19:43:31.864493 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="10142596-954c-4381-a6bd-c1eccc39cb4b" containerName="extract-utilities" Oct 08 19:43:31 crc kubenswrapper[4988]: I1008 19:43:31.864506 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="10142596-954c-4381-a6bd-c1eccc39cb4b" containerName="extract-utilities" Oct 08 19:43:31 crc kubenswrapper[4988]: E1008 19:43:31.864539 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="10142596-954c-4381-a6bd-c1eccc39cb4b" containerName="extract-content" Oct 08 19:43:31 crc kubenswrapper[4988]: I1008 19:43:31.864550 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="10142596-954c-4381-a6bd-c1eccc39cb4b" containerName="extract-content" Oct 08 19:43:31 crc kubenswrapper[4988]: E1008 19:43:31.864570 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="45a3d8ec-0896-49d3-ac9b-f294c0f66f62" containerName="mariadb-database-create" Oct 08 19:43:31 crc kubenswrapper[4988]: I1008 19:43:31.864582 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="45a3d8ec-0896-49d3-ac9b-f294c0f66f62" containerName="mariadb-database-create" Oct 08 19:43:31 crc kubenswrapper[4988]: I1008 19:43:31.864820 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="10142596-954c-4381-a6bd-c1eccc39cb4b" containerName="registry-server" Oct 08 19:43:31 crc kubenswrapper[4988]: I1008 19:43:31.864865 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="45a3d8ec-0896-49d3-ac9b-f294c0f66f62" containerName="mariadb-database-create" Oct 08 19:43:31 crc kubenswrapper[4988]: I1008 19:43:31.865698 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-1c50-account-create-28g6x" Oct 08 19:43:31 crc kubenswrapper[4988]: I1008 19:43:31.869269 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-db-secret" Oct 08 19:43:31 crc kubenswrapper[4988]: I1008 19:43:31.873510 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-1c50-account-create-28g6x"] Oct 08 19:43:32 crc kubenswrapper[4988]: I1008 19:43:31.999814 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5lkl5\" (UniqueName: \"kubernetes.io/projected/3448b939-1d0f-4d1e-89f5-4a0786c4c82e-kube-api-access-5lkl5\") pod \"barbican-1c50-account-create-28g6x\" (UID: \"3448b939-1d0f-4d1e-89f5-4a0786c4c82e\") " pod="openstack/barbican-1c50-account-create-28g6x" Oct 08 19:43:32 crc kubenswrapper[4988]: I1008 19:43:32.102775 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5lkl5\" (UniqueName: \"kubernetes.io/projected/3448b939-1d0f-4d1e-89f5-4a0786c4c82e-kube-api-access-5lkl5\") pod \"barbican-1c50-account-create-28g6x\" (UID: \"3448b939-1d0f-4d1e-89f5-4a0786c4c82e\") " pod="openstack/barbican-1c50-account-create-28g6x" Oct 08 19:43:32 crc kubenswrapper[4988]: I1008 19:43:32.135750 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5lkl5\" (UniqueName: \"kubernetes.io/projected/3448b939-1d0f-4d1e-89f5-4a0786c4c82e-kube-api-access-5lkl5\") pod \"barbican-1c50-account-create-28g6x\" (UID: \"3448b939-1d0f-4d1e-89f5-4a0786c4c82e\") " pod="openstack/barbican-1c50-account-create-28g6x" Oct 08 19:43:32 crc kubenswrapper[4988]: I1008 19:43:32.186120 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-1c50-account-create-28g6x" Oct 08 19:43:32 crc kubenswrapper[4988]: I1008 19:43:32.713275 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-1c50-account-create-28g6x"] Oct 08 19:43:33 crc kubenswrapper[4988]: I1008 19:43:33.254254 4988 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="10142596-954c-4381-a6bd-c1eccc39cb4b" path="/var/lib/kubelet/pods/10142596-954c-4381-a6bd-c1eccc39cb4b/volumes" Oct 08 19:43:33 crc kubenswrapper[4988]: I1008 19:43:33.546238 4988 generic.go:334] "Generic (PLEG): container finished" podID="3448b939-1d0f-4d1e-89f5-4a0786c4c82e" containerID="0a5eba2345e58863ef2eebcec4b06b7eb809e54141dd50b136e9c8f229fd7938" exitCode=0 Oct 08 19:43:33 crc kubenswrapper[4988]: I1008 19:43:33.546366 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-1c50-account-create-28g6x" event={"ID":"3448b939-1d0f-4d1e-89f5-4a0786c4c82e","Type":"ContainerDied","Data":"0a5eba2345e58863ef2eebcec4b06b7eb809e54141dd50b136e9c8f229fd7938"} Oct 08 19:43:33 crc kubenswrapper[4988]: I1008 19:43:33.546466 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-1c50-account-create-28g6x" event={"ID":"3448b939-1d0f-4d1e-89f5-4a0786c4c82e","Type":"ContainerStarted","Data":"acab6960b783a8711433fcb0c07ccf2f2f7262ae71efe97f911ed5e33602314c"} Oct 08 19:43:34 crc kubenswrapper[4988]: I1008 19:43:34.990860 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-1c50-account-create-28g6x" Oct 08 19:43:35 crc kubenswrapper[4988]: I1008 19:43:35.157537 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5lkl5\" (UniqueName: \"kubernetes.io/projected/3448b939-1d0f-4d1e-89f5-4a0786c4c82e-kube-api-access-5lkl5\") pod \"3448b939-1d0f-4d1e-89f5-4a0786c4c82e\" (UID: \"3448b939-1d0f-4d1e-89f5-4a0786c4c82e\") " Oct 08 19:43:35 crc kubenswrapper[4988]: I1008 19:43:35.174367 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3448b939-1d0f-4d1e-89f5-4a0786c4c82e-kube-api-access-5lkl5" (OuterVolumeSpecName: "kube-api-access-5lkl5") pod "3448b939-1d0f-4d1e-89f5-4a0786c4c82e" (UID: "3448b939-1d0f-4d1e-89f5-4a0786c4c82e"). InnerVolumeSpecName "kube-api-access-5lkl5". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 19:43:35 crc kubenswrapper[4988]: I1008 19:43:35.260197 4988 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5lkl5\" (UniqueName: \"kubernetes.io/projected/3448b939-1d0f-4d1e-89f5-4a0786c4c82e-kube-api-access-5lkl5\") on node \"crc\" DevicePath \"\"" Oct 08 19:43:35 crc kubenswrapper[4988]: I1008 19:43:35.573019 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-1c50-account-create-28g6x" event={"ID":"3448b939-1d0f-4d1e-89f5-4a0786c4c82e","Type":"ContainerDied","Data":"acab6960b783a8711433fcb0c07ccf2f2f7262ae71efe97f911ed5e33602314c"} Oct 08 19:43:35 crc kubenswrapper[4988]: I1008 19:43:35.573080 4988 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="acab6960b783a8711433fcb0c07ccf2f2f7262ae71efe97f911ed5e33602314c" Oct 08 19:43:35 crc kubenswrapper[4988]: I1008 19:43:35.573686 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-1c50-account-create-28g6x" Oct 08 19:43:37 crc kubenswrapper[4988]: I1008 19:43:37.101765 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-sync-7z69t"] Oct 08 19:43:37 crc kubenswrapper[4988]: E1008 19:43:37.103498 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3448b939-1d0f-4d1e-89f5-4a0786c4c82e" containerName="mariadb-account-create" Oct 08 19:43:37 crc kubenswrapper[4988]: I1008 19:43:37.103660 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="3448b939-1d0f-4d1e-89f5-4a0786c4c82e" containerName="mariadb-account-create" Oct 08 19:43:37 crc kubenswrapper[4988]: I1008 19:43:37.104048 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="3448b939-1d0f-4d1e-89f5-4a0786c4c82e" containerName="mariadb-account-create" Oct 08 19:43:37 crc kubenswrapper[4988]: I1008 19:43:37.104896 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-7z69t" Oct 08 19:43:37 crc kubenswrapper[4988]: I1008 19:43:37.110085 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-7z69t"] Oct 08 19:43:37 crc kubenswrapper[4988]: I1008 19:43:37.147808 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-f5t2p" Oct 08 19:43:37 crc kubenswrapper[4988]: I1008 19:43:37.147808 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Oct 08 19:43:37 crc kubenswrapper[4988]: I1008 19:43:37.197742 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/02e9ed2c-31bf-4775-ab12-a9fafc13e68e-combined-ca-bundle\") pod \"barbican-db-sync-7z69t\" (UID: \"02e9ed2c-31bf-4775-ab12-a9fafc13e68e\") " pod="openstack/barbican-db-sync-7z69t" Oct 08 19:43:37 crc kubenswrapper[4988]: I1008 19:43:37.198379 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4m4hm\" (UniqueName: \"kubernetes.io/projected/02e9ed2c-31bf-4775-ab12-a9fafc13e68e-kube-api-access-4m4hm\") pod \"barbican-db-sync-7z69t\" (UID: \"02e9ed2c-31bf-4775-ab12-a9fafc13e68e\") " pod="openstack/barbican-db-sync-7z69t" Oct 08 19:43:37 crc kubenswrapper[4988]: I1008 19:43:37.198595 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/02e9ed2c-31bf-4775-ab12-a9fafc13e68e-db-sync-config-data\") pod \"barbican-db-sync-7z69t\" (UID: \"02e9ed2c-31bf-4775-ab12-a9fafc13e68e\") " pod="openstack/barbican-db-sync-7z69t" Oct 08 19:43:37 crc kubenswrapper[4988]: I1008 19:43:37.300099 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4m4hm\" (UniqueName: \"kubernetes.io/projected/02e9ed2c-31bf-4775-ab12-a9fafc13e68e-kube-api-access-4m4hm\") pod \"barbican-db-sync-7z69t\" (UID: \"02e9ed2c-31bf-4775-ab12-a9fafc13e68e\") " pod="openstack/barbican-db-sync-7z69t" Oct 08 19:43:37 crc kubenswrapper[4988]: I1008 19:43:37.300161 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/02e9ed2c-31bf-4775-ab12-a9fafc13e68e-db-sync-config-data\") pod \"barbican-db-sync-7z69t\" (UID: \"02e9ed2c-31bf-4775-ab12-a9fafc13e68e\") " pod="openstack/barbican-db-sync-7z69t" Oct 08 19:43:37 crc kubenswrapper[4988]: I1008 19:43:37.300230 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/02e9ed2c-31bf-4775-ab12-a9fafc13e68e-combined-ca-bundle\") pod \"barbican-db-sync-7z69t\" (UID: \"02e9ed2c-31bf-4775-ab12-a9fafc13e68e\") " pod="openstack/barbican-db-sync-7z69t" Oct 08 19:43:37 crc kubenswrapper[4988]: I1008 19:43:37.304516 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/02e9ed2c-31bf-4775-ab12-a9fafc13e68e-db-sync-config-data\") pod \"barbican-db-sync-7z69t\" (UID: \"02e9ed2c-31bf-4775-ab12-a9fafc13e68e\") " pod="openstack/barbican-db-sync-7z69t" Oct 08 19:43:37 crc kubenswrapper[4988]: I1008 19:43:37.304999 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/02e9ed2c-31bf-4775-ab12-a9fafc13e68e-combined-ca-bundle\") pod \"barbican-db-sync-7z69t\" (UID: \"02e9ed2c-31bf-4775-ab12-a9fafc13e68e\") " pod="openstack/barbican-db-sync-7z69t" Oct 08 19:43:37 crc kubenswrapper[4988]: I1008 19:43:37.317891 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4m4hm\" (UniqueName: \"kubernetes.io/projected/02e9ed2c-31bf-4775-ab12-a9fafc13e68e-kube-api-access-4m4hm\") pod \"barbican-db-sync-7z69t\" (UID: \"02e9ed2c-31bf-4775-ab12-a9fafc13e68e\") " pod="openstack/barbican-db-sync-7z69t" Oct 08 19:43:37 crc kubenswrapper[4988]: I1008 19:43:37.469731 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-7z69t" Oct 08 19:43:37 crc kubenswrapper[4988]: I1008 19:43:37.980114 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-7z69t"] Oct 08 19:43:38 crc kubenswrapper[4988]: I1008 19:43:38.614912 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-7z69t" event={"ID":"02e9ed2c-31bf-4775-ab12-a9fafc13e68e","Type":"ContainerStarted","Data":"08a3fd1a9c76b2dbcf7c8ee9804f4b7b48f3c54e14df3fcddc2085315cb58fa9"} Oct 08 19:43:38 crc kubenswrapper[4988]: I1008 19:43:38.614959 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-7z69t" event={"ID":"02e9ed2c-31bf-4775-ab12-a9fafc13e68e","Type":"ContainerStarted","Data":"f7f6445695c3eb7b3926701c6ad437698a56a976fc88c4e165ff5a00a6a5ddc1"} Oct 08 19:43:38 crc kubenswrapper[4988]: I1008 19:43:38.641674 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-db-sync-7z69t" podStartSLOduration=1.641639718 podStartE2EDuration="1.641639718s" podCreationTimestamp="2025-10-08 19:43:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 19:43:38.634720019 +0000 UTC m=+5564.084562789" watchObservedRunningTime="2025-10-08 19:43:38.641639718 +0000 UTC m=+5564.091482538" Oct 08 19:43:40 crc kubenswrapper[4988]: I1008 19:43:40.640663 4988 generic.go:334] "Generic (PLEG): container finished" podID="02e9ed2c-31bf-4775-ab12-a9fafc13e68e" containerID="08a3fd1a9c76b2dbcf7c8ee9804f4b7b48f3c54e14df3fcddc2085315cb58fa9" exitCode=0 Oct 08 19:43:40 crc kubenswrapper[4988]: I1008 19:43:40.640734 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-7z69t" event={"ID":"02e9ed2c-31bf-4775-ab12-a9fafc13e68e","Type":"ContainerDied","Data":"08a3fd1a9c76b2dbcf7c8ee9804f4b7b48f3c54e14df3fcddc2085315cb58fa9"} Oct 08 19:43:41 crc kubenswrapper[4988]: I1008 19:43:41.987423 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-7z69t" Oct 08 19:43:42 crc kubenswrapper[4988]: I1008 19:43:42.098602 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/02e9ed2c-31bf-4775-ab12-a9fafc13e68e-combined-ca-bundle\") pod \"02e9ed2c-31bf-4775-ab12-a9fafc13e68e\" (UID: \"02e9ed2c-31bf-4775-ab12-a9fafc13e68e\") " Oct 08 19:43:42 crc kubenswrapper[4988]: I1008 19:43:42.098790 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/02e9ed2c-31bf-4775-ab12-a9fafc13e68e-db-sync-config-data\") pod \"02e9ed2c-31bf-4775-ab12-a9fafc13e68e\" (UID: \"02e9ed2c-31bf-4775-ab12-a9fafc13e68e\") " Oct 08 19:43:42 crc kubenswrapper[4988]: I1008 19:43:42.098871 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4m4hm\" (UniqueName: \"kubernetes.io/projected/02e9ed2c-31bf-4775-ab12-a9fafc13e68e-kube-api-access-4m4hm\") pod \"02e9ed2c-31bf-4775-ab12-a9fafc13e68e\" (UID: \"02e9ed2c-31bf-4775-ab12-a9fafc13e68e\") " Oct 08 19:43:42 crc kubenswrapper[4988]: I1008 19:43:42.107642 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/02e9ed2c-31bf-4775-ab12-a9fafc13e68e-kube-api-access-4m4hm" (OuterVolumeSpecName: "kube-api-access-4m4hm") pod "02e9ed2c-31bf-4775-ab12-a9fafc13e68e" (UID: "02e9ed2c-31bf-4775-ab12-a9fafc13e68e"). InnerVolumeSpecName "kube-api-access-4m4hm". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 19:43:42 crc kubenswrapper[4988]: I1008 19:43:42.108047 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/02e9ed2c-31bf-4775-ab12-a9fafc13e68e-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "02e9ed2c-31bf-4775-ab12-a9fafc13e68e" (UID: "02e9ed2c-31bf-4775-ab12-a9fafc13e68e"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 19:43:42 crc kubenswrapper[4988]: I1008 19:43:42.145418 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/02e9ed2c-31bf-4775-ab12-a9fafc13e68e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "02e9ed2c-31bf-4775-ab12-a9fafc13e68e" (UID: "02e9ed2c-31bf-4775-ab12-a9fafc13e68e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 19:43:42 crc kubenswrapper[4988]: I1008 19:43:42.201731 4988 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/02e9ed2c-31bf-4775-ab12-a9fafc13e68e-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Oct 08 19:43:42 crc kubenswrapper[4988]: I1008 19:43:42.201780 4988 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4m4hm\" (UniqueName: \"kubernetes.io/projected/02e9ed2c-31bf-4775-ab12-a9fafc13e68e-kube-api-access-4m4hm\") on node \"crc\" DevicePath \"\"" Oct 08 19:43:42 crc kubenswrapper[4988]: I1008 19:43:42.201800 4988 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/02e9ed2c-31bf-4775-ab12-a9fafc13e68e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 19:43:42 crc kubenswrapper[4988]: I1008 19:43:42.663206 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-7z69t" event={"ID":"02e9ed2c-31bf-4775-ab12-a9fafc13e68e","Type":"ContainerDied","Data":"f7f6445695c3eb7b3926701c6ad437698a56a976fc88c4e165ff5a00a6a5ddc1"} Oct 08 19:43:42 crc kubenswrapper[4988]: I1008 19:43:42.663270 4988 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f7f6445695c3eb7b3926701c6ad437698a56a976fc88c4e165ff5a00a6a5ddc1" Oct 08 19:43:42 crc kubenswrapper[4988]: I1008 19:43:42.663321 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-7z69t" Oct 08 19:43:42 crc kubenswrapper[4988]: I1008 19:43:42.948467 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-worker-6d89d4fcc5-rcwl2"] Oct 08 19:43:42 crc kubenswrapper[4988]: E1008 19:43:42.948792 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="02e9ed2c-31bf-4775-ab12-a9fafc13e68e" containerName="barbican-db-sync" Oct 08 19:43:42 crc kubenswrapper[4988]: I1008 19:43:42.948804 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="02e9ed2c-31bf-4775-ab12-a9fafc13e68e" containerName="barbican-db-sync" Oct 08 19:43:42 crc kubenswrapper[4988]: I1008 19:43:42.948951 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="02e9ed2c-31bf-4775-ab12-a9fafc13e68e" containerName="barbican-db-sync" Oct 08 19:43:42 crc kubenswrapper[4988]: I1008 19:43:42.950047 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-6d89d4fcc5-rcwl2" Oct 08 19:43:42 crc kubenswrapper[4988]: I1008 19:43:42.951766 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Oct 08 19:43:42 crc kubenswrapper[4988]: I1008 19:43:42.952201 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-worker-config-data" Oct 08 19:43:42 crc kubenswrapper[4988]: I1008 19:43:42.952422 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-f5t2p" Oct 08 19:43:42 crc kubenswrapper[4988]: I1008 19:43:42.976835 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-6d89d4fcc5-rcwl2"] Oct 08 19:43:43 crc kubenswrapper[4988]: I1008 19:43:43.017266 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6f247168-21ff-44e9-8e52-f49698e93a65-config-data\") pod \"barbican-worker-6d89d4fcc5-rcwl2\" (UID: \"6f247168-21ff-44e9-8e52-f49698e93a65\") " pod="openstack/barbican-worker-6d89d4fcc5-rcwl2" Oct 08 19:43:43 crc kubenswrapper[4988]: I1008 19:43:43.017322 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6f247168-21ff-44e9-8e52-f49698e93a65-combined-ca-bundle\") pod \"barbican-worker-6d89d4fcc5-rcwl2\" (UID: \"6f247168-21ff-44e9-8e52-f49698e93a65\") " pod="openstack/barbican-worker-6d89d4fcc5-rcwl2" Oct 08 19:43:43 crc kubenswrapper[4988]: I1008 19:43:43.017382 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vhzfq\" (UniqueName: \"kubernetes.io/projected/6f247168-21ff-44e9-8e52-f49698e93a65-kube-api-access-vhzfq\") pod \"barbican-worker-6d89d4fcc5-rcwl2\" (UID: \"6f247168-21ff-44e9-8e52-f49698e93a65\") " pod="openstack/barbican-worker-6d89d4fcc5-rcwl2" Oct 08 19:43:43 crc kubenswrapper[4988]: I1008 19:43:43.017425 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/6f247168-21ff-44e9-8e52-f49698e93a65-config-data-custom\") pod \"barbican-worker-6d89d4fcc5-rcwl2\" (UID: \"6f247168-21ff-44e9-8e52-f49698e93a65\") " pod="openstack/barbican-worker-6d89d4fcc5-rcwl2" Oct 08 19:43:43 crc kubenswrapper[4988]: I1008 19:43:43.017481 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6f247168-21ff-44e9-8e52-f49698e93a65-logs\") pod \"barbican-worker-6d89d4fcc5-rcwl2\" (UID: \"6f247168-21ff-44e9-8e52-f49698e93a65\") " pod="openstack/barbican-worker-6d89d4fcc5-rcwl2" Oct 08 19:43:43 crc kubenswrapper[4988]: I1008 19:43:43.027978 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-keystone-listener-75d59957f4-c7sp8"] Oct 08 19:43:43 crc kubenswrapper[4988]: I1008 19:43:43.032359 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-75d59957f4-c7sp8" Oct 08 19:43:43 crc kubenswrapper[4988]: I1008 19:43:43.036002 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-keystone-listener-config-data" Oct 08 19:43:43 crc kubenswrapper[4988]: I1008 19:43:43.052216 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-75d59957f4-c7sp8"] Oct 08 19:43:43 crc kubenswrapper[4988]: I1008 19:43:43.072375 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-c5b46b9d5-npkpp"] Oct 08 19:43:43 crc kubenswrapper[4988]: I1008 19:43:43.073802 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-c5b46b9d5-npkpp" Oct 08 19:43:43 crc kubenswrapper[4988]: I1008 19:43:43.088034 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-c5b46b9d5-npkpp"] Oct 08 19:43:43 crc kubenswrapper[4988]: I1008 19:43:43.119462 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/04727a82-9855-47c2-9df2-97b9e7a636c1-config-data\") pod \"barbican-keystone-listener-75d59957f4-c7sp8\" (UID: \"04727a82-9855-47c2-9df2-97b9e7a636c1\") " pod="openstack/barbican-keystone-listener-75d59957f4-c7sp8" Oct 08 19:43:43 crc kubenswrapper[4988]: I1008 19:43:43.119520 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/04727a82-9855-47c2-9df2-97b9e7a636c1-config-data-custom\") pod \"barbican-keystone-listener-75d59957f4-c7sp8\" (UID: \"04727a82-9855-47c2-9df2-97b9e7a636c1\") " pod="openstack/barbican-keystone-listener-75d59957f4-c7sp8" Oct 08 19:43:43 crc kubenswrapper[4988]: I1008 19:43:43.119552 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/04727a82-9855-47c2-9df2-97b9e7a636c1-combined-ca-bundle\") pod \"barbican-keystone-listener-75d59957f4-c7sp8\" (UID: \"04727a82-9855-47c2-9df2-97b9e7a636c1\") " pod="openstack/barbican-keystone-listener-75d59957f4-c7sp8" Oct 08 19:43:43 crc kubenswrapper[4988]: I1008 19:43:43.119569 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/04727a82-9855-47c2-9df2-97b9e7a636c1-logs\") pod \"barbican-keystone-listener-75d59957f4-c7sp8\" (UID: \"04727a82-9855-47c2-9df2-97b9e7a636c1\") " pod="openstack/barbican-keystone-listener-75d59957f4-c7sp8" Oct 08 19:43:43 crc kubenswrapper[4988]: I1008 19:43:43.119601 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6f247168-21ff-44e9-8e52-f49698e93a65-logs\") pod \"barbican-worker-6d89d4fcc5-rcwl2\" (UID: \"6f247168-21ff-44e9-8e52-f49698e93a65\") " pod="openstack/barbican-worker-6d89d4fcc5-rcwl2" Oct 08 19:43:43 crc kubenswrapper[4988]: I1008 19:43:43.119636 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6f247168-21ff-44e9-8e52-f49698e93a65-config-data\") pod \"barbican-worker-6d89d4fcc5-rcwl2\" (UID: \"6f247168-21ff-44e9-8e52-f49698e93a65\") " pod="openstack/barbican-worker-6d89d4fcc5-rcwl2" Oct 08 19:43:43 crc kubenswrapper[4988]: I1008 19:43:43.119661 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6f247168-21ff-44e9-8e52-f49698e93a65-combined-ca-bundle\") pod \"barbican-worker-6d89d4fcc5-rcwl2\" (UID: \"6f247168-21ff-44e9-8e52-f49698e93a65\") " pod="openstack/barbican-worker-6d89d4fcc5-rcwl2" Oct 08 19:43:43 crc kubenswrapper[4988]: I1008 19:43:43.119694 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r4jpr\" (UniqueName: \"kubernetes.io/projected/04727a82-9855-47c2-9df2-97b9e7a636c1-kube-api-access-r4jpr\") pod \"barbican-keystone-listener-75d59957f4-c7sp8\" (UID: \"04727a82-9855-47c2-9df2-97b9e7a636c1\") " pod="openstack/barbican-keystone-listener-75d59957f4-c7sp8" Oct 08 19:43:43 crc kubenswrapper[4988]: I1008 19:43:43.119729 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vhzfq\" (UniqueName: \"kubernetes.io/projected/6f247168-21ff-44e9-8e52-f49698e93a65-kube-api-access-vhzfq\") pod \"barbican-worker-6d89d4fcc5-rcwl2\" (UID: \"6f247168-21ff-44e9-8e52-f49698e93a65\") " pod="openstack/barbican-worker-6d89d4fcc5-rcwl2" Oct 08 19:43:43 crc kubenswrapper[4988]: I1008 19:43:43.119759 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/6f247168-21ff-44e9-8e52-f49698e93a65-config-data-custom\") pod \"barbican-worker-6d89d4fcc5-rcwl2\" (UID: \"6f247168-21ff-44e9-8e52-f49698e93a65\") " pod="openstack/barbican-worker-6d89d4fcc5-rcwl2" Oct 08 19:43:43 crc kubenswrapper[4988]: I1008 19:43:43.121246 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6f247168-21ff-44e9-8e52-f49698e93a65-logs\") pod \"barbican-worker-6d89d4fcc5-rcwl2\" (UID: \"6f247168-21ff-44e9-8e52-f49698e93a65\") " pod="openstack/barbican-worker-6d89d4fcc5-rcwl2" Oct 08 19:43:43 crc kubenswrapper[4988]: I1008 19:43:43.124527 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6f247168-21ff-44e9-8e52-f49698e93a65-combined-ca-bundle\") pod \"barbican-worker-6d89d4fcc5-rcwl2\" (UID: \"6f247168-21ff-44e9-8e52-f49698e93a65\") " pod="openstack/barbican-worker-6d89d4fcc5-rcwl2" Oct 08 19:43:43 crc kubenswrapper[4988]: I1008 19:43:43.126376 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6f247168-21ff-44e9-8e52-f49698e93a65-config-data\") pod \"barbican-worker-6d89d4fcc5-rcwl2\" (UID: \"6f247168-21ff-44e9-8e52-f49698e93a65\") " pod="openstack/barbican-worker-6d89d4fcc5-rcwl2" Oct 08 19:43:43 crc kubenswrapper[4988]: I1008 19:43:43.147359 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/6f247168-21ff-44e9-8e52-f49698e93a65-config-data-custom\") pod \"barbican-worker-6d89d4fcc5-rcwl2\" (UID: \"6f247168-21ff-44e9-8e52-f49698e93a65\") " pod="openstack/barbican-worker-6d89d4fcc5-rcwl2" Oct 08 19:43:43 crc kubenswrapper[4988]: I1008 19:43:43.168046 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vhzfq\" (UniqueName: \"kubernetes.io/projected/6f247168-21ff-44e9-8e52-f49698e93a65-kube-api-access-vhzfq\") pod \"barbican-worker-6d89d4fcc5-rcwl2\" (UID: \"6f247168-21ff-44e9-8e52-f49698e93a65\") " pod="openstack/barbican-worker-6d89d4fcc5-rcwl2" Oct 08 19:43:43 crc kubenswrapper[4988]: I1008 19:43:43.200454 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-5d4dcf4888-mf8q8"] Oct 08 19:43:43 crc kubenswrapper[4988]: I1008 19:43:43.202262 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-5d4dcf4888-mf8q8" Oct 08 19:43:43 crc kubenswrapper[4988]: I1008 19:43:43.207047 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-api-config-data" Oct 08 19:43:43 crc kubenswrapper[4988]: I1008 19:43:43.215415 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-5d4dcf4888-mf8q8"] Oct 08 19:43:43 crc kubenswrapper[4988]: I1008 19:43:43.227254 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dce2f14f-a039-4d24-b861-af2e48f06d2d-config\") pod \"dnsmasq-dns-c5b46b9d5-npkpp\" (UID: \"dce2f14f-a039-4d24-b861-af2e48f06d2d\") " pod="openstack/dnsmasq-dns-c5b46b9d5-npkpp" Oct 08 19:43:43 crc kubenswrapper[4988]: I1008 19:43:43.227304 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/dce2f14f-a039-4d24-b861-af2e48f06d2d-ovsdbserver-nb\") pod \"dnsmasq-dns-c5b46b9d5-npkpp\" (UID: \"dce2f14f-a039-4d24-b861-af2e48f06d2d\") " pod="openstack/dnsmasq-dns-c5b46b9d5-npkpp" Oct 08 19:43:43 crc kubenswrapper[4988]: I1008 19:43:43.227372 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r4jpr\" (UniqueName: \"kubernetes.io/projected/04727a82-9855-47c2-9df2-97b9e7a636c1-kube-api-access-r4jpr\") pod \"barbican-keystone-listener-75d59957f4-c7sp8\" (UID: \"04727a82-9855-47c2-9df2-97b9e7a636c1\") " pod="openstack/barbican-keystone-listener-75d59957f4-c7sp8" Oct 08 19:43:43 crc kubenswrapper[4988]: I1008 19:43:43.227440 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-spnd2\" (UniqueName: \"kubernetes.io/projected/dce2f14f-a039-4d24-b861-af2e48f06d2d-kube-api-access-spnd2\") pod \"dnsmasq-dns-c5b46b9d5-npkpp\" (UID: \"dce2f14f-a039-4d24-b861-af2e48f06d2d\") " pod="openstack/dnsmasq-dns-c5b46b9d5-npkpp" Oct 08 19:43:43 crc kubenswrapper[4988]: I1008 19:43:43.227750 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/dce2f14f-a039-4d24-b861-af2e48f06d2d-ovsdbserver-sb\") pod \"dnsmasq-dns-c5b46b9d5-npkpp\" (UID: \"dce2f14f-a039-4d24-b861-af2e48f06d2d\") " pod="openstack/dnsmasq-dns-c5b46b9d5-npkpp" Oct 08 19:43:43 crc kubenswrapper[4988]: I1008 19:43:43.227780 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/04727a82-9855-47c2-9df2-97b9e7a636c1-config-data\") pod \"barbican-keystone-listener-75d59957f4-c7sp8\" (UID: \"04727a82-9855-47c2-9df2-97b9e7a636c1\") " pod="openstack/barbican-keystone-listener-75d59957f4-c7sp8" Oct 08 19:43:43 crc kubenswrapper[4988]: I1008 19:43:43.227836 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/dce2f14f-a039-4d24-b861-af2e48f06d2d-dns-svc\") pod \"dnsmasq-dns-c5b46b9d5-npkpp\" (UID: \"dce2f14f-a039-4d24-b861-af2e48f06d2d\") " pod="openstack/dnsmasq-dns-c5b46b9d5-npkpp" Oct 08 19:43:43 crc kubenswrapper[4988]: I1008 19:43:43.228249 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/04727a82-9855-47c2-9df2-97b9e7a636c1-config-data-custom\") pod \"barbican-keystone-listener-75d59957f4-c7sp8\" (UID: \"04727a82-9855-47c2-9df2-97b9e7a636c1\") " pod="openstack/barbican-keystone-listener-75d59957f4-c7sp8" Oct 08 19:43:43 crc kubenswrapper[4988]: I1008 19:43:43.228301 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/04727a82-9855-47c2-9df2-97b9e7a636c1-combined-ca-bundle\") pod \"barbican-keystone-listener-75d59957f4-c7sp8\" (UID: \"04727a82-9855-47c2-9df2-97b9e7a636c1\") " pod="openstack/barbican-keystone-listener-75d59957f4-c7sp8" Oct 08 19:43:43 crc kubenswrapper[4988]: I1008 19:43:43.228325 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/04727a82-9855-47c2-9df2-97b9e7a636c1-logs\") pod \"barbican-keystone-listener-75d59957f4-c7sp8\" (UID: \"04727a82-9855-47c2-9df2-97b9e7a636c1\") " pod="openstack/barbican-keystone-listener-75d59957f4-c7sp8" Oct 08 19:43:43 crc kubenswrapper[4988]: I1008 19:43:43.228758 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/04727a82-9855-47c2-9df2-97b9e7a636c1-logs\") pod \"barbican-keystone-listener-75d59957f4-c7sp8\" (UID: \"04727a82-9855-47c2-9df2-97b9e7a636c1\") " pod="openstack/barbican-keystone-listener-75d59957f4-c7sp8" Oct 08 19:43:43 crc kubenswrapper[4988]: I1008 19:43:43.232411 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/04727a82-9855-47c2-9df2-97b9e7a636c1-config-data\") pod \"barbican-keystone-listener-75d59957f4-c7sp8\" (UID: \"04727a82-9855-47c2-9df2-97b9e7a636c1\") " pod="openstack/barbican-keystone-listener-75d59957f4-c7sp8" Oct 08 19:43:43 crc kubenswrapper[4988]: I1008 19:43:43.237995 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/04727a82-9855-47c2-9df2-97b9e7a636c1-config-data-custom\") pod \"barbican-keystone-listener-75d59957f4-c7sp8\" (UID: \"04727a82-9855-47c2-9df2-97b9e7a636c1\") " pod="openstack/barbican-keystone-listener-75d59957f4-c7sp8" Oct 08 19:43:43 crc kubenswrapper[4988]: I1008 19:43:43.244815 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r4jpr\" (UniqueName: \"kubernetes.io/projected/04727a82-9855-47c2-9df2-97b9e7a636c1-kube-api-access-r4jpr\") pod \"barbican-keystone-listener-75d59957f4-c7sp8\" (UID: \"04727a82-9855-47c2-9df2-97b9e7a636c1\") " pod="openstack/barbican-keystone-listener-75d59957f4-c7sp8" Oct 08 19:43:43 crc kubenswrapper[4988]: I1008 19:43:43.244832 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/04727a82-9855-47c2-9df2-97b9e7a636c1-combined-ca-bundle\") pod \"barbican-keystone-listener-75d59957f4-c7sp8\" (UID: \"04727a82-9855-47c2-9df2-97b9e7a636c1\") " pod="openstack/barbican-keystone-listener-75d59957f4-c7sp8" Oct 08 19:43:43 crc kubenswrapper[4988]: I1008 19:43:43.275371 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-6d89d4fcc5-rcwl2" Oct 08 19:43:43 crc kubenswrapper[4988]: I1008 19:43:43.330012 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-spnd2\" (UniqueName: \"kubernetes.io/projected/dce2f14f-a039-4d24-b861-af2e48f06d2d-kube-api-access-spnd2\") pod \"dnsmasq-dns-c5b46b9d5-npkpp\" (UID: \"dce2f14f-a039-4d24-b861-af2e48f06d2d\") " pod="openstack/dnsmasq-dns-c5b46b9d5-npkpp" Oct 08 19:43:43 crc kubenswrapper[4988]: I1008 19:43:43.330059 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7e7725a2-0ed5-4f96-8be0-c359c1c40b26-logs\") pod \"barbican-api-5d4dcf4888-mf8q8\" (UID: \"7e7725a2-0ed5-4f96-8be0-c359c1c40b26\") " pod="openstack/barbican-api-5d4dcf4888-mf8q8" Oct 08 19:43:43 crc kubenswrapper[4988]: I1008 19:43:43.330087 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/dce2f14f-a039-4d24-b861-af2e48f06d2d-ovsdbserver-sb\") pod \"dnsmasq-dns-c5b46b9d5-npkpp\" (UID: \"dce2f14f-a039-4d24-b861-af2e48f06d2d\") " pod="openstack/dnsmasq-dns-c5b46b9d5-npkpp" Oct 08 19:43:43 crc kubenswrapper[4988]: I1008 19:43:43.330110 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n64x6\" (UniqueName: \"kubernetes.io/projected/7e7725a2-0ed5-4f96-8be0-c359c1c40b26-kube-api-access-n64x6\") pod \"barbican-api-5d4dcf4888-mf8q8\" (UID: \"7e7725a2-0ed5-4f96-8be0-c359c1c40b26\") " pod="openstack/barbican-api-5d4dcf4888-mf8q8" Oct 08 19:43:43 crc kubenswrapper[4988]: I1008 19:43:43.330135 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/dce2f14f-a039-4d24-b861-af2e48f06d2d-dns-svc\") pod \"dnsmasq-dns-c5b46b9d5-npkpp\" (UID: \"dce2f14f-a039-4d24-b861-af2e48f06d2d\") " pod="openstack/dnsmasq-dns-c5b46b9d5-npkpp" Oct 08 19:43:43 crc kubenswrapper[4988]: I1008 19:43:43.330162 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7e7725a2-0ed5-4f96-8be0-c359c1c40b26-combined-ca-bundle\") pod \"barbican-api-5d4dcf4888-mf8q8\" (UID: \"7e7725a2-0ed5-4f96-8be0-c359c1c40b26\") " pod="openstack/barbican-api-5d4dcf4888-mf8q8" Oct 08 19:43:43 crc kubenswrapper[4988]: I1008 19:43:43.330181 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/7e7725a2-0ed5-4f96-8be0-c359c1c40b26-config-data-custom\") pod \"barbican-api-5d4dcf4888-mf8q8\" (UID: \"7e7725a2-0ed5-4f96-8be0-c359c1c40b26\") " pod="openstack/barbican-api-5d4dcf4888-mf8q8" Oct 08 19:43:43 crc kubenswrapper[4988]: I1008 19:43:43.330220 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dce2f14f-a039-4d24-b861-af2e48f06d2d-config\") pod \"dnsmasq-dns-c5b46b9d5-npkpp\" (UID: \"dce2f14f-a039-4d24-b861-af2e48f06d2d\") " pod="openstack/dnsmasq-dns-c5b46b9d5-npkpp" Oct 08 19:43:43 crc kubenswrapper[4988]: I1008 19:43:43.330238 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/dce2f14f-a039-4d24-b861-af2e48f06d2d-ovsdbserver-nb\") pod \"dnsmasq-dns-c5b46b9d5-npkpp\" (UID: \"dce2f14f-a039-4d24-b861-af2e48f06d2d\") " pod="openstack/dnsmasq-dns-c5b46b9d5-npkpp" Oct 08 19:43:43 crc kubenswrapper[4988]: I1008 19:43:43.330257 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7e7725a2-0ed5-4f96-8be0-c359c1c40b26-config-data\") pod \"barbican-api-5d4dcf4888-mf8q8\" (UID: \"7e7725a2-0ed5-4f96-8be0-c359c1c40b26\") " pod="openstack/barbican-api-5d4dcf4888-mf8q8" Oct 08 19:43:43 crc kubenswrapper[4988]: I1008 19:43:43.331453 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/dce2f14f-a039-4d24-b861-af2e48f06d2d-ovsdbserver-sb\") pod \"dnsmasq-dns-c5b46b9d5-npkpp\" (UID: \"dce2f14f-a039-4d24-b861-af2e48f06d2d\") " pod="openstack/dnsmasq-dns-c5b46b9d5-npkpp" Oct 08 19:43:43 crc kubenswrapper[4988]: I1008 19:43:43.335199 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/dce2f14f-a039-4d24-b861-af2e48f06d2d-dns-svc\") pod \"dnsmasq-dns-c5b46b9d5-npkpp\" (UID: \"dce2f14f-a039-4d24-b861-af2e48f06d2d\") " pod="openstack/dnsmasq-dns-c5b46b9d5-npkpp" Oct 08 19:43:43 crc kubenswrapper[4988]: I1008 19:43:43.335287 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dce2f14f-a039-4d24-b861-af2e48f06d2d-config\") pod \"dnsmasq-dns-c5b46b9d5-npkpp\" (UID: \"dce2f14f-a039-4d24-b861-af2e48f06d2d\") " pod="openstack/dnsmasq-dns-c5b46b9d5-npkpp" Oct 08 19:43:43 crc kubenswrapper[4988]: I1008 19:43:43.335408 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/dce2f14f-a039-4d24-b861-af2e48f06d2d-ovsdbserver-nb\") pod \"dnsmasq-dns-c5b46b9d5-npkpp\" (UID: \"dce2f14f-a039-4d24-b861-af2e48f06d2d\") " pod="openstack/dnsmasq-dns-c5b46b9d5-npkpp" Oct 08 19:43:43 crc kubenswrapper[4988]: I1008 19:43:43.352263 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-spnd2\" (UniqueName: \"kubernetes.io/projected/dce2f14f-a039-4d24-b861-af2e48f06d2d-kube-api-access-spnd2\") pod \"dnsmasq-dns-c5b46b9d5-npkpp\" (UID: \"dce2f14f-a039-4d24-b861-af2e48f06d2d\") " pod="openstack/dnsmasq-dns-c5b46b9d5-npkpp" Oct 08 19:43:43 crc kubenswrapper[4988]: I1008 19:43:43.361998 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-75d59957f4-c7sp8" Oct 08 19:43:43 crc kubenswrapper[4988]: I1008 19:43:43.413324 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-c5b46b9d5-npkpp" Oct 08 19:43:43 crc kubenswrapper[4988]: I1008 19:43:43.432005 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7e7725a2-0ed5-4f96-8be0-c359c1c40b26-logs\") pod \"barbican-api-5d4dcf4888-mf8q8\" (UID: \"7e7725a2-0ed5-4f96-8be0-c359c1c40b26\") " pod="openstack/barbican-api-5d4dcf4888-mf8q8" Oct 08 19:43:43 crc kubenswrapper[4988]: I1008 19:43:43.432082 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n64x6\" (UniqueName: \"kubernetes.io/projected/7e7725a2-0ed5-4f96-8be0-c359c1c40b26-kube-api-access-n64x6\") pod \"barbican-api-5d4dcf4888-mf8q8\" (UID: \"7e7725a2-0ed5-4f96-8be0-c359c1c40b26\") " pod="openstack/barbican-api-5d4dcf4888-mf8q8" Oct 08 19:43:43 crc kubenswrapper[4988]: I1008 19:43:43.432133 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7e7725a2-0ed5-4f96-8be0-c359c1c40b26-combined-ca-bundle\") pod \"barbican-api-5d4dcf4888-mf8q8\" (UID: \"7e7725a2-0ed5-4f96-8be0-c359c1c40b26\") " pod="openstack/barbican-api-5d4dcf4888-mf8q8" Oct 08 19:43:43 crc kubenswrapper[4988]: I1008 19:43:43.432160 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/7e7725a2-0ed5-4f96-8be0-c359c1c40b26-config-data-custom\") pod \"barbican-api-5d4dcf4888-mf8q8\" (UID: \"7e7725a2-0ed5-4f96-8be0-c359c1c40b26\") " pod="openstack/barbican-api-5d4dcf4888-mf8q8" Oct 08 19:43:43 crc kubenswrapper[4988]: I1008 19:43:43.432206 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7e7725a2-0ed5-4f96-8be0-c359c1c40b26-config-data\") pod \"barbican-api-5d4dcf4888-mf8q8\" (UID: \"7e7725a2-0ed5-4f96-8be0-c359c1c40b26\") " pod="openstack/barbican-api-5d4dcf4888-mf8q8" Oct 08 19:43:43 crc kubenswrapper[4988]: I1008 19:43:43.432486 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7e7725a2-0ed5-4f96-8be0-c359c1c40b26-logs\") pod \"barbican-api-5d4dcf4888-mf8q8\" (UID: \"7e7725a2-0ed5-4f96-8be0-c359c1c40b26\") " pod="openstack/barbican-api-5d4dcf4888-mf8q8" Oct 08 19:43:43 crc kubenswrapper[4988]: I1008 19:43:43.437308 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7e7725a2-0ed5-4f96-8be0-c359c1c40b26-combined-ca-bundle\") pod \"barbican-api-5d4dcf4888-mf8q8\" (UID: \"7e7725a2-0ed5-4f96-8be0-c359c1c40b26\") " pod="openstack/barbican-api-5d4dcf4888-mf8q8" Oct 08 19:43:43 crc kubenswrapper[4988]: I1008 19:43:43.437998 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7e7725a2-0ed5-4f96-8be0-c359c1c40b26-config-data\") pod \"barbican-api-5d4dcf4888-mf8q8\" (UID: \"7e7725a2-0ed5-4f96-8be0-c359c1c40b26\") " pod="openstack/barbican-api-5d4dcf4888-mf8q8" Oct 08 19:43:43 crc kubenswrapper[4988]: I1008 19:43:43.442149 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/7e7725a2-0ed5-4f96-8be0-c359c1c40b26-config-data-custom\") pod \"barbican-api-5d4dcf4888-mf8q8\" (UID: \"7e7725a2-0ed5-4f96-8be0-c359c1c40b26\") " pod="openstack/barbican-api-5d4dcf4888-mf8q8" Oct 08 19:43:43 crc kubenswrapper[4988]: I1008 19:43:43.447606 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n64x6\" (UniqueName: \"kubernetes.io/projected/7e7725a2-0ed5-4f96-8be0-c359c1c40b26-kube-api-access-n64x6\") pod \"barbican-api-5d4dcf4888-mf8q8\" (UID: \"7e7725a2-0ed5-4f96-8be0-c359c1c40b26\") " pod="openstack/barbican-api-5d4dcf4888-mf8q8" Oct 08 19:43:43 crc kubenswrapper[4988]: I1008 19:43:43.515821 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-5d4dcf4888-mf8q8" Oct 08 19:43:43 crc kubenswrapper[4988]: I1008 19:43:43.756487 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-6d89d4fcc5-rcwl2"] Oct 08 19:43:43 crc kubenswrapper[4988]: W1008 19:43:43.763462 4988 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6f247168_21ff_44e9_8e52_f49698e93a65.slice/crio-171102d0de18f9a4fc524855f0de930a4695d7de96c6854f317169fa343a5220 WatchSource:0}: Error finding container 171102d0de18f9a4fc524855f0de930a4695d7de96c6854f317169fa343a5220: Status 404 returned error can't find the container with id 171102d0de18f9a4fc524855f0de930a4695d7de96c6854f317169fa343a5220 Oct 08 19:43:43 crc kubenswrapper[4988]: I1008 19:43:43.928949 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-75d59957f4-c7sp8"] Oct 08 19:43:43 crc kubenswrapper[4988]: I1008 19:43:43.937174 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-c5b46b9d5-npkpp"] Oct 08 19:43:44 crc kubenswrapper[4988]: I1008 19:43:44.067454 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-5d4dcf4888-mf8q8"] Oct 08 19:43:44 crc kubenswrapper[4988]: W1008 19:43:44.072161 4988 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7e7725a2_0ed5_4f96_8be0_c359c1c40b26.slice/crio-67cb970d59cbfb1cce9a565c2e70587bbbc3e3c39ae193a63e50252a2d25b230 WatchSource:0}: Error finding container 67cb970d59cbfb1cce9a565c2e70587bbbc3e3c39ae193a63e50252a2d25b230: Status 404 returned error can't find the container with id 67cb970d59cbfb1cce9a565c2e70587bbbc3e3c39ae193a63e50252a2d25b230 Oct 08 19:43:44 crc kubenswrapper[4988]: I1008 19:43:44.689733 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-6d89d4fcc5-rcwl2" event={"ID":"6f247168-21ff-44e9-8e52-f49698e93a65","Type":"ContainerStarted","Data":"18804ea6180e182e87c73fea1e5faa4848470733986f61f9da78a3cf9e0c035f"} Oct 08 19:43:44 crc kubenswrapper[4988]: I1008 19:43:44.690169 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-6d89d4fcc5-rcwl2" event={"ID":"6f247168-21ff-44e9-8e52-f49698e93a65","Type":"ContainerStarted","Data":"fdb1f9f66c2bdc3e76967d8e4a7ddad60eacada50c3cbe4a0dec9e3903a78b82"} Oct 08 19:43:44 crc kubenswrapper[4988]: I1008 19:43:44.690197 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-6d89d4fcc5-rcwl2" event={"ID":"6f247168-21ff-44e9-8e52-f49698e93a65","Type":"ContainerStarted","Data":"171102d0de18f9a4fc524855f0de930a4695d7de96c6854f317169fa343a5220"} Oct 08 19:43:44 crc kubenswrapper[4988]: I1008 19:43:44.692764 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-75d59957f4-c7sp8" event={"ID":"04727a82-9855-47c2-9df2-97b9e7a636c1","Type":"ContainerStarted","Data":"1d38b3356a7d2ab11b1528f1baa6b2417a8b34019f5aee98a2bc2ecae383f96e"} Oct 08 19:43:44 crc kubenswrapper[4988]: I1008 19:43:44.692820 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-75d59957f4-c7sp8" event={"ID":"04727a82-9855-47c2-9df2-97b9e7a636c1","Type":"ContainerStarted","Data":"8210e68aa7c353f74ba541a2f76b78761313a3ef1ffe07f1cc0d29b0835e3bcd"} Oct 08 19:43:44 crc kubenswrapper[4988]: I1008 19:43:44.692837 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-75d59957f4-c7sp8" event={"ID":"04727a82-9855-47c2-9df2-97b9e7a636c1","Type":"ContainerStarted","Data":"e94e633dc14d5d8301757f6e1525861db0811fc2d25d2767a3f01289d85683c9"} Oct 08 19:43:44 crc kubenswrapper[4988]: I1008 19:43:44.694735 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-5d4dcf4888-mf8q8" event={"ID":"7e7725a2-0ed5-4f96-8be0-c359c1c40b26","Type":"ContainerStarted","Data":"0bf7b18e26afc0536ed3ddb0599a0bfbd3db18373d1fdcaae401936156b6f838"} Oct 08 19:43:44 crc kubenswrapper[4988]: I1008 19:43:44.694783 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-5d4dcf4888-mf8q8" event={"ID":"7e7725a2-0ed5-4f96-8be0-c359c1c40b26","Type":"ContainerStarted","Data":"8a98ef35eb13a2c7d88a063b9880379ed72341e30de93d115da92883aa9a3c83"} Oct 08 19:43:44 crc kubenswrapper[4988]: I1008 19:43:44.694799 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-5d4dcf4888-mf8q8" event={"ID":"7e7725a2-0ed5-4f96-8be0-c359c1c40b26","Type":"ContainerStarted","Data":"67cb970d59cbfb1cce9a565c2e70587bbbc3e3c39ae193a63e50252a2d25b230"} Oct 08 19:43:44 crc kubenswrapper[4988]: I1008 19:43:44.694886 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-5d4dcf4888-mf8q8" Oct 08 19:43:44 crc kubenswrapper[4988]: I1008 19:43:44.694908 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-5d4dcf4888-mf8q8" Oct 08 19:43:44 crc kubenswrapper[4988]: I1008 19:43:44.696059 4988 generic.go:334] "Generic (PLEG): container finished" podID="dce2f14f-a039-4d24-b861-af2e48f06d2d" containerID="6edb06c5df4c60b50bb54de502320e580ab5ebc2c6963187cdda1b95895a15b3" exitCode=0 Oct 08 19:43:44 crc kubenswrapper[4988]: I1008 19:43:44.696096 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-c5b46b9d5-npkpp" event={"ID":"dce2f14f-a039-4d24-b861-af2e48f06d2d","Type":"ContainerDied","Data":"6edb06c5df4c60b50bb54de502320e580ab5ebc2c6963187cdda1b95895a15b3"} Oct 08 19:43:44 crc kubenswrapper[4988]: I1008 19:43:44.696116 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-c5b46b9d5-npkpp" event={"ID":"dce2f14f-a039-4d24-b861-af2e48f06d2d","Type":"ContainerStarted","Data":"8822a79eee32a009bb537dc232a185dc32a97b675bb803558b3b70c6a46db931"} Oct 08 19:43:44 crc kubenswrapper[4988]: I1008 19:43:44.749273 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-5d4dcf4888-mf8q8" podStartSLOduration=1.7492520360000001 podStartE2EDuration="1.749252036s" podCreationTimestamp="2025-10-08 19:43:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 19:43:44.741424856 +0000 UTC m=+5570.191267646" watchObservedRunningTime="2025-10-08 19:43:44.749252036 +0000 UTC m=+5570.199094806" Oct 08 19:43:44 crc kubenswrapper[4988]: I1008 19:43:44.751602 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-worker-6d89d4fcc5-rcwl2" podStartSLOduration=2.75158914 podStartE2EDuration="2.75158914s" podCreationTimestamp="2025-10-08 19:43:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 19:43:44.721529494 +0000 UTC m=+5570.171372284" watchObservedRunningTime="2025-10-08 19:43:44.75158914 +0000 UTC m=+5570.201431910" Oct 08 19:43:44 crc kubenswrapper[4988]: I1008 19:43:44.817789 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-keystone-listener-75d59957f4-c7sp8" podStartSLOduration=2.817738813 podStartE2EDuration="2.817738813s" podCreationTimestamp="2025-10-08 19:43:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 19:43:44.799774622 +0000 UTC m=+5570.249617392" watchObservedRunningTime="2025-10-08 19:43:44.817738813 +0000 UTC m=+5570.267581583" Oct 08 19:43:45 crc kubenswrapper[4988]: I1008 19:43:45.665338 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-5bb9747ff6-msjmt"] Oct 08 19:43:45 crc kubenswrapper[4988]: I1008 19:43:45.667810 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-5bb9747ff6-msjmt" Oct 08 19:43:45 crc kubenswrapper[4988]: I1008 19:43:45.670268 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-barbican-public-svc" Oct 08 19:43:45 crc kubenswrapper[4988]: I1008 19:43:45.675326 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-barbican-internal-svc" Oct 08 19:43:45 crc kubenswrapper[4988]: I1008 19:43:45.676373 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-5bb9747ff6-msjmt"] Oct 08 19:43:45 crc kubenswrapper[4988]: I1008 19:43:45.707553 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-c5b46b9d5-npkpp" event={"ID":"dce2f14f-a039-4d24-b861-af2e48f06d2d","Type":"ContainerStarted","Data":"7e3cc924ecc9eb8641f9e92f699dd1e29dd612ae90bccba932dfffe406eccc65"} Oct 08 19:43:45 crc kubenswrapper[4988]: I1008 19:43:45.707877 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-c5b46b9d5-npkpp" Oct 08 19:43:45 crc kubenswrapper[4988]: I1008 19:43:45.734530 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-c5b46b9d5-npkpp" podStartSLOduration=2.734499375 podStartE2EDuration="2.734499375s" podCreationTimestamp="2025-10-08 19:43:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 19:43:45.724377963 +0000 UTC m=+5571.174220733" watchObservedRunningTime="2025-10-08 19:43:45.734499375 +0000 UTC m=+5571.184342165" Oct 08 19:43:45 crc kubenswrapper[4988]: I1008 19:43:45.801788 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ab612ec9-381c-4cb0-b608-10a1bb5768bd-logs\") pod \"barbican-api-5bb9747ff6-msjmt\" (UID: \"ab612ec9-381c-4cb0-b608-10a1bb5768bd\") " pod="openstack/barbican-api-5bb9747ff6-msjmt" Oct 08 19:43:45 crc kubenswrapper[4988]: I1008 19:43:45.801884 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/ab612ec9-381c-4cb0-b608-10a1bb5768bd-public-tls-certs\") pod \"barbican-api-5bb9747ff6-msjmt\" (UID: \"ab612ec9-381c-4cb0-b608-10a1bb5768bd\") " pod="openstack/barbican-api-5bb9747ff6-msjmt" Oct 08 19:43:45 crc kubenswrapper[4988]: I1008 19:43:45.802055 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bg2tb\" (UniqueName: \"kubernetes.io/projected/ab612ec9-381c-4cb0-b608-10a1bb5768bd-kube-api-access-bg2tb\") pod \"barbican-api-5bb9747ff6-msjmt\" (UID: \"ab612ec9-381c-4cb0-b608-10a1bb5768bd\") " pod="openstack/barbican-api-5bb9747ff6-msjmt" Oct 08 19:43:45 crc kubenswrapper[4988]: I1008 19:43:45.802094 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/ab612ec9-381c-4cb0-b608-10a1bb5768bd-internal-tls-certs\") pod \"barbican-api-5bb9747ff6-msjmt\" (UID: \"ab612ec9-381c-4cb0-b608-10a1bb5768bd\") " pod="openstack/barbican-api-5bb9747ff6-msjmt" Oct 08 19:43:45 crc kubenswrapper[4988]: I1008 19:43:45.802437 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ab612ec9-381c-4cb0-b608-10a1bb5768bd-combined-ca-bundle\") pod \"barbican-api-5bb9747ff6-msjmt\" (UID: \"ab612ec9-381c-4cb0-b608-10a1bb5768bd\") " pod="openstack/barbican-api-5bb9747ff6-msjmt" Oct 08 19:43:45 crc kubenswrapper[4988]: I1008 19:43:45.802557 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/ab612ec9-381c-4cb0-b608-10a1bb5768bd-config-data-custom\") pod \"barbican-api-5bb9747ff6-msjmt\" (UID: \"ab612ec9-381c-4cb0-b608-10a1bb5768bd\") " pod="openstack/barbican-api-5bb9747ff6-msjmt" Oct 08 19:43:45 crc kubenswrapper[4988]: I1008 19:43:45.802619 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ab612ec9-381c-4cb0-b608-10a1bb5768bd-config-data\") pod \"barbican-api-5bb9747ff6-msjmt\" (UID: \"ab612ec9-381c-4cb0-b608-10a1bb5768bd\") " pod="openstack/barbican-api-5bb9747ff6-msjmt" Oct 08 19:43:45 crc kubenswrapper[4988]: I1008 19:43:45.904501 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ab612ec9-381c-4cb0-b608-10a1bb5768bd-logs\") pod \"barbican-api-5bb9747ff6-msjmt\" (UID: \"ab612ec9-381c-4cb0-b608-10a1bb5768bd\") " pod="openstack/barbican-api-5bb9747ff6-msjmt" Oct 08 19:43:45 crc kubenswrapper[4988]: I1008 19:43:45.905091 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ab612ec9-381c-4cb0-b608-10a1bb5768bd-logs\") pod \"barbican-api-5bb9747ff6-msjmt\" (UID: \"ab612ec9-381c-4cb0-b608-10a1bb5768bd\") " pod="openstack/barbican-api-5bb9747ff6-msjmt" Oct 08 19:43:45 crc kubenswrapper[4988]: I1008 19:43:45.905317 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/ab612ec9-381c-4cb0-b608-10a1bb5768bd-public-tls-certs\") pod \"barbican-api-5bb9747ff6-msjmt\" (UID: \"ab612ec9-381c-4cb0-b608-10a1bb5768bd\") " pod="openstack/barbican-api-5bb9747ff6-msjmt" Oct 08 19:43:45 crc kubenswrapper[4988]: I1008 19:43:45.905489 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bg2tb\" (UniqueName: \"kubernetes.io/projected/ab612ec9-381c-4cb0-b608-10a1bb5768bd-kube-api-access-bg2tb\") pod \"barbican-api-5bb9747ff6-msjmt\" (UID: \"ab612ec9-381c-4cb0-b608-10a1bb5768bd\") " pod="openstack/barbican-api-5bb9747ff6-msjmt" Oct 08 19:43:45 crc kubenswrapper[4988]: I1008 19:43:45.905941 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/ab612ec9-381c-4cb0-b608-10a1bb5768bd-internal-tls-certs\") pod \"barbican-api-5bb9747ff6-msjmt\" (UID: \"ab612ec9-381c-4cb0-b608-10a1bb5768bd\") " pod="openstack/barbican-api-5bb9747ff6-msjmt" Oct 08 19:43:45 crc kubenswrapper[4988]: I1008 19:43:45.906553 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ab612ec9-381c-4cb0-b608-10a1bb5768bd-combined-ca-bundle\") pod \"barbican-api-5bb9747ff6-msjmt\" (UID: \"ab612ec9-381c-4cb0-b608-10a1bb5768bd\") " pod="openstack/barbican-api-5bb9747ff6-msjmt" Oct 08 19:43:45 crc kubenswrapper[4988]: I1008 19:43:45.906623 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/ab612ec9-381c-4cb0-b608-10a1bb5768bd-config-data-custom\") pod \"barbican-api-5bb9747ff6-msjmt\" (UID: \"ab612ec9-381c-4cb0-b608-10a1bb5768bd\") " pod="openstack/barbican-api-5bb9747ff6-msjmt" Oct 08 19:43:45 crc kubenswrapper[4988]: I1008 19:43:45.906661 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ab612ec9-381c-4cb0-b608-10a1bb5768bd-config-data\") pod \"barbican-api-5bb9747ff6-msjmt\" (UID: \"ab612ec9-381c-4cb0-b608-10a1bb5768bd\") " pod="openstack/barbican-api-5bb9747ff6-msjmt" Oct 08 19:43:45 crc kubenswrapper[4988]: I1008 19:43:45.910664 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ab612ec9-381c-4cb0-b608-10a1bb5768bd-combined-ca-bundle\") pod \"barbican-api-5bb9747ff6-msjmt\" (UID: \"ab612ec9-381c-4cb0-b608-10a1bb5768bd\") " pod="openstack/barbican-api-5bb9747ff6-msjmt" Oct 08 19:43:45 crc kubenswrapper[4988]: I1008 19:43:45.911092 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/ab612ec9-381c-4cb0-b608-10a1bb5768bd-public-tls-certs\") pod \"barbican-api-5bb9747ff6-msjmt\" (UID: \"ab612ec9-381c-4cb0-b608-10a1bb5768bd\") " pod="openstack/barbican-api-5bb9747ff6-msjmt" Oct 08 19:43:45 crc kubenswrapper[4988]: I1008 19:43:45.911896 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/ab612ec9-381c-4cb0-b608-10a1bb5768bd-config-data-custom\") pod \"barbican-api-5bb9747ff6-msjmt\" (UID: \"ab612ec9-381c-4cb0-b608-10a1bb5768bd\") " pod="openstack/barbican-api-5bb9747ff6-msjmt" Oct 08 19:43:45 crc kubenswrapper[4988]: I1008 19:43:45.917928 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/ab612ec9-381c-4cb0-b608-10a1bb5768bd-internal-tls-certs\") pod \"barbican-api-5bb9747ff6-msjmt\" (UID: \"ab612ec9-381c-4cb0-b608-10a1bb5768bd\") " pod="openstack/barbican-api-5bb9747ff6-msjmt" Oct 08 19:43:45 crc kubenswrapper[4988]: I1008 19:43:45.919889 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ab612ec9-381c-4cb0-b608-10a1bb5768bd-config-data\") pod \"barbican-api-5bb9747ff6-msjmt\" (UID: \"ab612ec9-381c-4cb0-b608-10a1bb5768bd\") " pod="openstack/barbican-api-5bb9747ff6-msjmt" Oct 08 19:43:45 crc kubenswrapper[4988]: I1008 19:43:45.923531 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bg2tb\" (UniqueName: \"kubernetes.io/projected/ab612ec9-381c-4cb0-b608-10a1bb5768bd-kube-api-access-bg2tb\") pod \"barbican-api-5bb9747ff6-msjmt\" (UID: \"ab612ec9-381c-4cb0-b608-10a1bb5768bd\") " pod="openstack/barbican-api-5bb9747ff6-msjmt" Oct 08 19:43:45 crc kubenswrapper[4988]: I1008 19:43:45.985406 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-5bb9747ff6-msjmt" Oct 08 19:43:46 crc kubenswrapper[4988]: I1008 19:43:46.284047 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-5bb9747ff6-msjmt"] Oct 08 19:43:46 crc kubenswrapper[4988]: W1008 19:43:46.288226 4988 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podab612ec9_381c_4cb0_b608_10a1bb5768bd.slice/crio-2e90b3ea66e0ea62bf4303d318505d1d40e659340f6ab72be8cfc055781a6efd WatchSource:0}: Error finding container 2e90b3ea66e0ea62bf4303d318505d1d40e659340f6ab72be8cfc055781a6efd: Status 404 returned error can't find the container with id 2e90b3ea66e0ea62bf4303d318505d1d40e659340f6ab72be8cfc055781a6efd Oct 08 19:43:46 crc kubenswrapper[4988]: I1008 19:43:46.716043 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-5bb9747ff6-msjmt" event={"ID":"ab612ec9-381c-4cb0-b608-10a1bb5768bd","Type":"ContainerStarted","Data":"7be9d2917a24148c37430a761cb6e49f328947a397270ccdb79f443c01351cdd"} Oct 08 19:43:46 crc kubenswrapper[4988]: I1008 19:43:46.716421 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-5bb9747ff6-msjmt" event={"ID":"ab612ec9-381c-4cb0-b608-10a1bb5768bd","Type":"ContainerStarted","Data":"2e90b3ea66e0ea62bf4303d318505d1d40e659340f6ab72be8cfc055781a6efd"} Oct 08 19:43:47 crc kubenswrapper[4988]: I1008 19:43:47.737954 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-5bb9747ff6-msjmt" event={"ID":"ab612ec9-381c-4cb0-b608-10a1bb5768bd","Type":"ContainerStarted","Data":"a091d545b8a42c97a4189565dbc3ccfe3c5e48f71e72facb6125af03c3d6a41b"} Oct 08 19:43:47 crc kubenswrapper[4988]: I1008 19:43:47.738485 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-5bb9747ff6-msjmt" Oct 08 19:43:47 crc kubenswrapper[4988]: I1008 19:43:47.773789 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-5bb9747ff6-msjmt" podStartSLOduration=2.773759817 podStartE2EDuration="2.773759817s" podCreationTimestamp="2025-10-08 19:43:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 19:43:47.766246499 +0000 UTC m=+5573.216089309" watchObservedRunningTime="2025-10-08 19:43:47.773759817 +0000 UTC m=+5573.223602627" Oct 08 19:43:48 crc kubenswrapper[4988]: I1008 19:43:48.757844 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-5bb9747ff6-msjmt" Oct 08 19:43:53 crc kubenswrapper[4988]: I1008 19:43:53.415763 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-c5b46b9d5-npkpp" Oct 08 19:43:53 crc kubenswrapper[4988]: I1008 19:43:53.534752 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6c7c55f8b9-wlltj"] Oct 08 19:43:53 crc kubenswrapper[4988]: I1008 19:43:53.534972 4988 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-6c7c55f8b9-wlltj" podUID="15716656-ce43-447b-9910-11b4813b5e52" containerName="dnsmasq-dns" containerID="cri-o://df85a0680fc1f4db9b155d8405d389aa23264924aa46e3f4a41982e3c5d7966b" gracePeriod=10 Oct 08 19:43:53 crc kubenswrapper[4988]: I1008 19:43:53.832207 4988 generic.go:334] "Generic (PLEG): container finished" podID="15716656-ce43-447b-9910-11b4813b5e52" containerID="df85a0680fc1f4db9b155d8405d389aa23264924aa46e3f4a41982e3c5d7966b" exitCode=0 Oct 08 19:43:53 crc kubenswrapper[4988]: I1008 19:43:53.832359 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6c7c55f8b9-wlltj" event={"ID":"15716656-ce43-447b-9910-11b4813b5e52","Type":"ContainerDied","Data":"df85a0680fc1f4db9b155d8405d389aa23264924aa46e3f4a41982e3c5d7966b"} Oct 08 19:43:53 crc kubenswrapper[4988]: I1008 19:43:53.997080 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6c7c55f8b9-wlltj" Oct 08 19:43:54 crc kubenswrapper[4988]: I1008 19:43:54.075821 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7lv46\" (UniqueName: \"kubernetes.io/projected/15716656-ce43-447b-9910-11b4813b5e52-kube-api-access-7lv46\") pod \"15716656-ce43-447b-9910-11b4813b5e52\" (UID: \"15716656-ce43-447b-9910-11b4813b5e52\") " Oct 08 19:43:54 crc kubenswrapper[4988]: I1008 19:43:54.075897 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/15716656-ce43-447b-9910-11b4813b5e52-ovsdbserver-sb\") pod \"15716656-ce43-447b-9910-11b4813b5e52\" (UID: \"15716656-ce43-447b-9910-11b4813b5e52\") " Oct 08 19:43:54 crc kubenswrapper[4988]: I1008 19:43:54.076014 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/15716656-ce43-447b-9910-11b4813b5e52-config\") pod \"15716656-ce43-447b-9910-11b4813b5e52\" (UID: \"15716656-ce43-447b-9910-11b4813b5e52\") " Oct 08 19:43:54 crc kubenswrapper[4988]: I1008 19:43:54.076080 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/15716656-ce43-447b-9910-11b4813b5e52-dns-svc\") pod \"15716656-ce43-447b-9910-11b4813b5e52\" (UID: \"15716656-ce43-447b-9910-11b4813b5e52\") " Oct 08 19:43:54 crc kubenswrapper[4988]: I1008 19:43:54.076114 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/15716656-ce43-447b-9910-11b4813b5e52-ovsdbserver-nb\") pod \"15716656-ce43-447b-9910-11b4813b5e52\" (UID: \"15716656-ce43-447b-9910-11b4813b5e52\") " Oct 08 19:43:54 crc kubenswrapper[4988]: I1008 19:43:54.084508 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/15716656-ce43-447b-9910-11b4813b5e52-kube-api-access-7lv46" (OuterVolumeSpecName: "kube-api-access-7lv46") pod "15716656-ce43-447b-9910-11b4813b5e52" (UID: "15716656-ce43-447b-9910-11b4813b5e52"). InnerVolumeSpecName "kube-api-access-7lv46". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 19:43:54 crc kubenswrapper[4988]: I1008 19:43:54.126358 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/15716656-ce43-447b-9910-11b4813b5e52-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "15716656-ce43-447b-9910-11b4813b5e52" (UID: "15716656-ce43-447b-9910-11b4813b5e52"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 19:43:54 crc kubenswrapper[4988]: I1008 19:43:54.132550 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/15716656-ce43-447b-9910-11b4813b5e52-config" (OuterVolumeSpecName: "config") pod "15716656-ce43-447b-9910-11b4813b5e52" (UID: "15716656-ce43-447b-9910-11b4813b5e52"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 19:43:54 crc kubenswrapper[4988]: I1008 19:43:54.139149 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/15716656-ce43-447b-9910-11b4813b5e52-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "15716656-ce43-447b-9910-11b4813b5e52" (UID: "15716656-ce43-447b-9910-11b4813b5e52"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 19:43:54 crc kubenswrapper[4988]: I1008 19:43:54.148954 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/15716656-ce43-447b-9910-11b4813b5e52-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "15716656-ce43-447b-9910-11b4813b5e52" (UID: "15716656-ce43-447b-9910-11b4813b5e52"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 19:43:54 crc kubenswrapper[4988]: I1008 19:43:54.178251 4988 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/15716656-ce43-447b-9910-11b4813b5e52-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 08 19:43:54 crc kubenswrapper[4988]: I1008 19:43:54.178299 4988 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/15716656-ce43-447b-9910-11b4813b5e52-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 08 19:43:54 crc kubenswrapper[4988]: I1008 19:43:54.178322 4988 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7lv46\" (UniqueName: \"kubernetes.io/projected/15716656-ce43-447b-9910-11b4813b5e52-kube-api-access-7lv46\") on node \"crc\" DevicePath \"\"" Oct 08 19:43:54 crc kubenswrapper[4988]: I1008 19:43:54.178340 4988 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/15716656-ce43-447b-9910-11b4813b5e52-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 08 19:43:54 crc kubenswrapper[4988]: I1008 19:43:54.178355 4988 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/15716656-ce43-447b-9910-11b4813b5e52-config\") on node \"crc\" DevicePath \"\"" Oct 08 19:43:54 crc kubenswrapper[4988]: I1008 19:43:54.853080 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6c7c55f8b9-wlltj" event={"ID":"15716656-ce43-447b-9910-11b4813b5e52","Type":"ContainerDied","Data":"8ff82740ae1ec65178317bdd21e3481d29d174cffb920f3c9288c976b7d41fae"} Oct 08 19:43:54 crc kubenswrapper[4988]: I1008 19:43:54.853156 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6c7c55f8b9-wlltj" Oct 08 19:43:54 crc kubenswrapper[4988]: I1008 19:43:54.853467 4988 scope.go:117] "RemoveContainer" containerID="df85a0680fc1f4db9b155d8405d389aa23264924aa46e3f4a41982e3c5d7966b" Oct 08 19:43:54 crc kubenswrapper[4988]: I1008 19:43:54.885913 4988 scope.go:117] "RemoveContainer" containerID="cff42b69f6d35ca4c5669a6d8580efd4430e29c89b59ff6444247a5a7a89664c" Oct 08 19:43:54 crc kubenswrapper[4988]: I1008 19:43:54.908521 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-5d4dcf4888-mf8q8" Oct 08 19:43:54 crc kubenswrapper[4988]: I1008 19:43:54.918014 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6c7c55f8b9-wlltj"] Oct 08 19:43:54 crc kubenswrapper[4988]: I1008 19:43:54.926778 4988 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-6c7c55f8b9-wlltj"] Oct 08 19:43:54 crc kubenswrapper[4988]: I1008 19:43:54.953256 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-5d4dcf4888-mf8q8" Oct 08 19:43:55 crc kubenswrapper[4988]: I1008 19:43:55.255765 4988 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="15716656-ce43-447b-9910-11b4813b5e52" path="/var/lib/kubelet/pods/15716656-ce43-447b-9910-11b4813b5e52/volumes" Oct 08 19:43:57 crc kubenswrapper[4988]: I1008 19:43:57.301711 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-5bb9747ff6-msjmt" Oct 08 19:43:57 crc kubenswrapper[4988]: I1008 19:43:57.303824 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-5bb9747ff6-msjmt" Oct 08 19:43:57 crc kubenswrapper[4988]: I1008 19:43:57.432873 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-5d4dcf4888-mf8q8"] Oct 08 19:43:57 crc kubenswrapper[4988]: I1008 19:43:57.433160 4988 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-5d4dcf4888-mf8q8" podUID="7e7725a2-0ed5-4f96-8be0-c359c1c40b26" containerName="barbican-api-log" containerID="cri-o://8a98ef35eb13a2c7d88a063b9880379ed72341e30de93d115da92883aa9a3c83" gracePeriod=30 Oct 08 19:43:57 crc kubenswrapper[4988]: I1008 19:43:57.433604 4988 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-5d4dcf4888-mf8q8" podUID="7e7725a2-0ed5-4f96-8be0-c359c1c40b26" containerName="barbican-api" containerID="cri-o://0bf7b18e26afc0536ed3ddb0599a0bfbd3db18373d1fdcaae401936156b6f838" gracePeriod=30 Oct 08 19:43:57 crc kubenswrapper[4988]: I1008 19:43:57.884094 4988 generic.go:334] "Generic (PLEG): container finished" podID="7e7725a2-0ed5-4f96-8be0-c359c1c40b26" containerID="8a98ef35eb13a2c7d88a063b9880379ed72341e30de93d115da92883aa9a3c83" exitCode=143 Oct 08 19:43:57 crc kubenswrapper[4988]: I1008 19:43:57.884207 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-5d4dcf4888-mf8q8" event={"ID":"7e7725a2-0ed5-4f96-8be0-c359c1c40b26","Type":"ContainerDied","Data":"8a98ef35eb13a2c7d88a063b9880379ed72341e30de93d115da92883aa9a3c83"} Oct 08 19:44:00 crc kubenswrapper[4988]: I1008 19:44:00.612215 4988 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-5d4dcf4888-mf8q8" podUID="7e7725a2-0ed5-4f96-8be0-c359c1c40b26" containerName="barbican-api-log" probeResult="failure" output="Get \"http://10.217.1.39:9311/healthcheck\": read tcp 10.217.0.2:40898->10.217.1.39:9311: read: connection reset by peer" Oct 08 19:44:00 crc kubenswrapper[4988]: I1008 19:44:00.612543 4988 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-5d4dcf4888-mf8q8" podUID="7e7725a2-0ed5-4f96-8be0-c359c1c40b26" containerName="barbican-api" probeResult="failure" output="Get \"http://10.217.1.39:9311/healthcheck\": read tcp 10.217.0.2:40908->10.217.1.39:9311: read: connection reset by peer" Oct 08 19:44:00 crc kubenswrapper[4988]: I1008 19:44:00.922343 4988 generic.go:334] "Generic (PLEG): container finished" podID="7e7725a2-0ed5-4f96-8be0-c359c1c40b26" containerID="0bf7b18e26afc0536ed3ddb0599a0bfbd3db18373d1fdcaae401936156b6f838" exitCode=0 Oct 08 19:44:00 crc kubenswrapper[4988]: I1008 19:44:00.922410 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-5d4dcf4888-mf8q8" event={"ID":"7e7725a2-0ed5-4f96-8be0-c359c1c40b26","Type":"ContainerDied","Data":"0bf7b18e26afc0536ed3ddb0599a0bfbd3db18373d1fdcaae401936156b6f838"} Oct 08 19:44:01 crc kubenswrapper[4988]: I1008 19:44:01.079641 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-5d4dcf4888-mf8q8" Oct 08 19:44:01 crc kubenswrapper[4988]: I1008 19:44:01.150508 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/7e7725a2-0ed5-4f96-8be0-c359c1c40b26-config-data-custom\") pod \"7e7725a2-0ed5-4f96-8be0-c359c1c40b26\" (UID: \"7e7725a2-0ed5-4f96-8be0-c359c1c40b26\") " Oct 08 19:44:01 crc kubenswrapper[4988]: I1008 19:44:01.150584 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n64x6\" (UniqueName: \"kubernetes.io/projected/7e7725a2-0ed5-4f96-8be0-c359c1c40b26-kube-api-access-n64x6\") pod \"7e7725a2-0ed5-4f96-8be0-c359c1c40b26\" (UID: \"7e7725a2-0ed5-4f96-8be0-c359c1c40b26\") " Oct 08 19:44:01 crc kubenswrapper[4988]: I1008 19:44:01.150613 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7e7725a2-0ed5-4f96-8be0-c359c1c40b26-logs\") pod \"7e7725a2-0ed5-4f96-8be0-c359c1c40b26\" (UID: \"7e7725a2-0ed5-4f96-8be0-c359c1c40b26\") " Oct 08 19:44:01 crc kubenswrapper[4988]: I1008 19:44:01.150654 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7e7725a2-0ed5-4f96-8be0-c359c1c40b26-config-data\") pod \"7e7725a2-0ed5-4f96-8be0-c359c1c40b26\" (UID: \"7e7725a2-0ed5-4f96-8be0-c359c1c40b26\") " Oct 08 19:44:01 crc kubenswrapper[4988]: I1008 19:44:01.150782 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7e7725a2-0ed5-4f96-8be0-c359c1c40b26-combined-ca-bundle\") pod \"7e7725a2-0ed5-4f96-8be0-c359c1c40b26\" (UID: \"7e7725a2-0ed5-4f96-8be0-c359c1c40b26\") " Oct 08 19:44:01 crc kubenswrapper[4988]: I1008 19:44:01.151986 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7e7725a2-0ed5-4f96-8be0-c359c1c40b26-logs" (OuterVolumeSpecName: "logs") pod "7e7725a2-0ed5-4f96-8be0-c359c1c40b26" (UID: "7e7725a2-0ed5-4f96-8be0-c359c1c40b26"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 19:44:01 crc kubenswrapper[4988]: I1008 19:44:01.156122 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7e7725a2-0ed5-4f96-8be0-c359c1c40b26-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "7e7725a2-0ed5-4f96-8be0-c359c1c40b26" (UID: "7e7725a2-0ed5-4f96-8be0-c359c1c40b26"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 19:44:01 crc kubenswrapper[4988]: I1008 19:44:01.156360 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7e7725a2-0ed5-4f96-8be0-c359c1c40b26-kube-api-access-n64x6" (OuterVolumeSpecName: "kube-api-access-n64x6") pod "7e7725a2-0ed5-4f96-8be0-c359c1c40b26" (UID: "7e7725a2-0ed5-4f96-8be0-c359c1c40b26"). InnerVolumeSpecName "kube-api-access-n64x6". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 19:44:01 crc kubenswrapper[4988]: I1008 19:44:01.180298 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7e7725a2-0ed5-4f96-8be0-c359c1c40b26-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "7e7725a2-0ed5-4f96-8be0-c359c1c40b26" (UID: "7e7725a2-0ed5-4f96-8be0-c359c1c40b26"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 19:44:01 crc kubenswrapper[4988]: I1008 19:44:01.204901 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7e7725a2-0ed5-4f96-8be0-c359c1c40b26-config-data" (OuterVolumeSpecName: "config-data") pod "7e7725a2-0ed5-4f96-8be0-c359c1c40b26" (UID: "7e7725a2-0ed5-4f96-8be0-c359c1c40b26"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 19:44:01 crc kubenswrapper[4988]: I1008 19:44:01.251944 4988 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n64x6\" (UniqueName: \"kubernetes.io/projected/7e7725a2-0ed5-4f96-8be0-c359c1c40b26-kube-api-access-n64x6\") on node \"crc\" DevicePath \"\"" Oct 08 19:44:01 crc kubenswrapper[4988]: I1008 19:44:01.251971 4988 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7e7725a2-0ed5-4f96-8be0-c359c1c40b26-logs\") on node \"crc\" DevicePath \"\"" Oct 08 19:44:01 crc kubenswrapper[4988]: I1008 19:44:01.251982 4988 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7e7725a2-0ed5-4f96-8be0-c359c1c40b26-config-data\") on node \"crc\" DevicePath \"\"" Oct 08 19:44:01 crc kubenswrapper[4988]: I1008 19:44:01.251994 4988 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7e7725a2-0ed5-4f96-8be0-c359c1c40b26-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 19:44:01 crc kubenswrapper[4988]: I1008 19:44:01.252003 4988 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/7e7725a2-0ed5-4f96-8be0-c359c1c40b26-config-data-custom\") on node \"crc\" DevicePath \"\"" Oct 08 19:44:01 crc kubenswrapper[4988]: I1008 19:44:01.935005 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-5d4dcf4888-mf8q8" event={"ID":"7e7725a2-0ed5-4f96-8be0-c359c1c40b26","Type":"ContainerDied","Data":"67cb970d59cbfb1cce9a565c2e70587bbbc3e3c39ae193a63e50252a2d25b230"} Oct 08 19:44:01 crc kubenswrapper[4988]: I1008 19:44:01.935072 4988 scope.go:117] "RemoveContainer" containerID="0bf7b18e26afc0536ed3ddb0599a0bfbd3db18373d1fdcaae401936156b6f838" Oct 08 19:44:01 crc kubenswrapper[4988]: I1008 19:44:01.935190 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-5d4dcf4888-mf8q8" Oct 08 19:44:01 crc kubenswrapper[4988]: I1008 19:44:01.962057 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-5d4dcf4888-mf8q8"] Oct 08 19:44:01 crc kubenswrapper[4988]: I1008 19:44:01.968149 4988 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-api-5d4dcf4888-mf8q8"] Oct 08 19:44:01 crc kubenswrapper[4988]: I1008 19:44:01.970626 4988 scope.go:117] "RemoveContainer" containerID="8a98ef35eb13a2c7d88a063b9880379ed72341e30de93d115da92883aa9a3c83" Oct 08 19:44:03 crc kubenswrapper[4988]: I1008 19:44:03.254144 4988 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7e7725a2-0ed5-4f96-8be0-c359c1c40b26" path="/var/lib/kubelet/pods/7e7725a2-0ed5-4f96-8be0-c359c1c40b26/volumes" Oct 08 19:44:19 crc kubenswrapper[4988]: I1008 19:44:19.267056 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-create-t4zxj"] Oct 08 19:44:19 crc kubenswrapper[4988]: E1008 19:44:19.269580 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="15716656-ce43-447b-9910-11b4813b5e52" containerName="dnsmasq-dns" Oct 08 19:44:19 crc kubenswrapper[4988]: I1008 19:44:19.269693 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="15716656-ce43-447b-9910-11b4813b5e52" containerName="dnsmasq-dns" Oct 08 19:44:19 crc kubenswrapper[4988]: E1008 19:44:19.269785 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7e7725a2-0ed5-4f96-8be0-c359c1c40b26" containerName="barbican-api" Oct 08 19:44:19 crc kubenswrapper[4988]: I1008 19:44:19.269857 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="7e7725a2-0ed5-4f96-8be0-c359c1c40b26" containerName="barbican-api" Oct 08 19:44:19 crc kubenswrapper[4988]: E1008 19:44:19.269942 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7e7725a2-0ed5-4f96-8be0-c359c1c40b26" containerName="barbican-api-log" Oct 08 19:44:19 crc kubenswrapper[4988]: I1008 19:44:19.270013 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="7e7725a2-0ed5-4f96-8be0-c359c1c40b26" containerName="barbican-api-log" Oct 08 19:44:19 crc kubenswrapper[4988]: E1008 19:44:19.270102 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="15716656-ce43-447b-9910-11b4813b5e52" containerName="init" Oct 08 19:44:19 crc kubenswrapper[4988]: I1008 19:44:19.270182 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="15716656-ce43-447b-9910-11b4813b5e52" containerName="init" Oct 08 19:44:19 crc kubenswrapper[4988]: I1008 19:44:19.270470 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="7e7725a2-0ed5-4f96-8be0-c359c1c40b26" containerName="barbican-api" Oct 08 19:44:19 crc kubenswrapper[4988]: I1008 19:44:19.270565 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="7e7725a2-0ed5-4f96-8be0-c359c1c40b26" containerName="barbican-api-log" Oct 08 19:44:19 crc kubenswrapper[4988]: I1008 19:44:19.270656 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="15716656-ce43-447b-9910-11b4813b5e52" containerName="dnsmasq-dns" Oct 08 19:44:19 crc kubenswrapper[4988]: I1008 19:44:19.271425 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-t4zxj" Oct 08 19:44:19 crc kubenswrapper[4988]: I1008 19:44:19.274856 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-t4zxj"] Oct 08 19:44:19 crc kubenswrapper[4988]: I1008 19:44:19.313834 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-khz9f\" (UniqueName: \"kubernetes.io/projected/9dd78259-618d-4698-8f22-fa4ff6a75b75-kube-api-access-khz9f\") pod \"neutron-db-create-t4zxj\" (UID: \"9dd78259-618d-4698-8f22-fa4ff6a75b75\") " pod="openstack/neutron-db-create-t4zxj" Oct 08 19:44:19 crc kubenswrapper[4988]: I1008 19:44:19.416283 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-khz9f\" (UniqueName: \"kubernetes.io/projected/9dd78259-618d-4698-8f22-fa4ff6a75b75-kube-api-access-khz9f\") pod \"neutron-db-create-t4zxj\" (UID: \"9dd78259-618d-4698-8f22-fa4ff6a75b75\") " pod="openstack/neutron-db-create-t4zxj" Oct 08 19:44:19 crc kubenswrapper[4988]: I1008 19:44:19.440499 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-khz9f\" (UniqueName: \"kubernetes.io/projected/9dd78259-618d-4698-8f22-fa4ff6a75b75-kube-api-access-khz9f\") pod \"neutron-db-create-t4zxj\" (UID: \"9dd78259-618d-4698-8f22-fa4ff6a75b75\") " pod="openstack/neutron-db-create-t4zxj" Oct 08 19:44:19 crc kubenswrapper[4988]: I1008 19:44:19.588545 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-t4zxj" Oct 08 19:44:20 crc kubenswrapper[4988]: W1008 19:44:20.085219 4988 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9dd78259_618d_4698_8f22_fa4ff6a75b75.slice/crio-d7197cbed1f466b0ae027dd8871b1bacd3d0f00b387953a3fe5569899c7e1cf5 WatchSource:0}: Error finding container d7197cbed1f466b0ae027dd8871b1bacd3d0f00b387953a3fe5569899c7e1cf5: Status 404 returned error can't find the container with id d7197cbed1f466b0ae027dd8871b1bacd3d0f00b387953a3fe5569899c7e1cf5 Oct 08 19:44:20 crc kubenswrapper[4988]: I1008 19:44:20.086945 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-t4zxj"] Oct 08 19:44:20 crc kubenswrapper[4988]: I1008 19:44:20.139724 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-t4zxj" event={"ID":"9dd78259-618d-4698-8f22-fa4ff6a75b75","Type":"ContainerStarted","Data":"d7197cbed1f466b0ae027dd8871b1bacd3d0f00b387953a3fe5569899c7e1cf5"} Oct 08 19:44:21 crc kubenswrapper[4988]: I1008 19:44:21.173531 4988 generic.go:334] "Generic (PLEG): container finished" podID="9dd78259-618d-4698-8f22-fa4ff6a75b75" containerID="49f54ff31e8d408344336acfa6472b80772a333be76bcac7d65e889556f469c7" exitCode=0 Oct 08 19:44:21 crc kubenswrapper[4988]: I1008 19:44:21.173701 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-t4zxj" event={"ID":"9dd78259-618d-4698-8f22-fa4ff6a75b75","Type":"ContainerDied","Data":"49f54ff31e8d408344336acfa6472b80772a333be76bcac7d65e889556f469c7"} Oct 08 19:44:22 crc kubenswrapper[4988]: I1008 19:44:22.677862 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-t4zxj" Oct 08 19:44:22 crc kubenswrapper[4988]: I1008 19:44:22.798163 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-khz9f\" (UniqueName: \"kubernetes.io/projected/9dd78259-618d-4698-8f22-fa4ff6a75b75-kube-api-access-khz9f\") pod \"9dd78259-618d-4698-8f22-fa4ff6a75b75\" (UID: \"9dd78259-618d-4698-8f22-fa4ff6a75b75\") " Oct 08 19:44:22 crc kubenswrapper[4988]: I1008 19:44:22.810595 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9dd78259-618d-4698-8f22-fa4ff6a75b75-kube-api-access-khz9f" (OuterVolumeSpecName: "kube-api-access-khz9f") pod "9dd78259-618d-4698-8f22-fa4ff6a75b75" (UID: "9dd78259-618d-4698-8f22-fa4ff6a75b75"). InnerVolumeSpecName "kube-api-access-khz9f". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 19:44:22 crc kubenswrapper[4988]: I1008 19:44:22.900346 4988 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-khz9f\" (UniqueName: \"kubernetes.io/projected/9dd78259-618d-4698-8f22-fa4ff6a75b75-kube-api-access-khz9f\") on node \"crc\" DevicePath \"\"" Oct 08 19:44:23 crc kubenswrapper[4988]: I1008 19:44:23.201074 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-t4zxj" event={"ID":"9dd78259-618d-4698-8f22-fa4ff6a75b75","Type":"ContainerDied","Data":"d7197cbed1f466b0ae027dd8871b1bacd3d0f00b387953a3fe5569899c7e1cf5"} Oct 08 19:44:23 crc kubenswrapper[4988]: I1008 19:44:23.201402 4988 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d7197cbed1f466b0ae027dd8871b1bacd3d0f00b387953a3fe5569899c7e1cf5" Oct 08 19:44:23 crc kubenswrapper[4988]: I1008 19:44:23.201139 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-t4zxj" Oct 08 19:44:29 crc kubenswrapper[4988]: I1008 19:44:29.357079 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-8807-account-create-2rpmf"] Oct 08 19:44:29 crc kubenswrapper[4988]: E1008 19:44:29.358200 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9dd78259-618d-4698-8f22-fa4ff6a75b75" containerName="mariadb-database-create" Oct 08 19:44:29 crc kubenswrapper[4988]: I1008 19:44:29.358222 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="9dd78259-618d-4698-8f22-fa4ff6a75b75" containerName="mariadb-database-create" Oct 08 19:44:29 crc kubenswrapper[4988]: I1008 19:44:29.358600 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="9dd78259-618d-4698-8f22-fa4ff6a75b75" containerName="mariadb-database-create" Oct 08 19:44:29 crc kubenswrapper[4988]: I1008 19:44:29.359730 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-8807-account-create-2rpmf" Oct 08 19:44:29 crc kubenswrapper[4988]: I1008 19:44:29.362831 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-db-secret" Oct 08 19:44:29 crc kubenswrapper[4988]: I1008 19:44:29.370262 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-8807-account-create-2rpmf"] Oct 08 19:44:29 crc kubenswrapper[4988]: I1008 19:44:29.437242 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hjltx\" (UniqueName: \"kubernetes.io/projected/70fc1b25-cb88-4345-b70f-e72c74df4a0f-kube-api-access-hjltx\") pod \"neutron-8807-account-create-2rpmf\" (UID: \"70fc1b25-cb88-4345-b70f-e72c74df4a0f\") " pod="openstack/neutron-8807-account-create-2rpmf" Oct 08 19:44:29 crc kubenswrapper[4988]: I1008 19:44:29.540287 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hjltx\" (UniqueName: \"kubernetes.io/projected/70fc1b25-cb88-4345-b70f-e72c74df4a0f-kube-api-access-hjltx\") pod \"neutron-8807-account-create-2rpmf\" (UID: \"70fc1b25-cb88-4345-b70f-e72c74df4a0f\") " pod="openstack/neutron-8807-account-create-2rpmf" Oct 08 19:44:29 crc kubenswrapper[4988]: I1008 19:44:29.573682 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hjltx\" (UniqueName: \"kubernetes.io/projected/70fc1b25-cb88-4345-b70f-e72c74df4a0f-kube-api-access-hjltx\") pod \"neutron-8807-account-create-2rpmf\" (UID: \"70fc1b25-cb88-4345-b70f-e72c74df4a0f\") " pod="openstack/neutron-8807-account-create-2rpmf" Oct 08 19:44:29 crc kubenswrapper[4988]: I1008 19:44:29.685643 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-8807-account-create-2rpmf" Oct 08 19:44:29 crc kubenswrapper[4988]: I1008 19:44:29.987567 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-8807-account-create-2rpmf"] Oct 08 19:44:30 crc kubenswrapper[4988]: I1008 19:44:30.289374 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-8807-account-create-2rpmf" event={"ID":"70fc1b25-cb88-4345-b70f-e72c74df4a0f","Type":"ContainerStarted","Data":"6a6df4cf78aeb6f22196fcbff02c8005b18ea265c81fd0b5af42b23ab4537dff"} Oct 08 19:44:30 crc kubenswrapper[4988]: I1008 19:44:30.289678 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-8807-account-create-2rpmf" event={"ID":"70fc1b25-cb88-4345-b70f-e72c74df4a0f","Type":"ContainerStarted","Data":"2e98a6f1c65ef204fc59ea9cd3b5da4538b0cceb6056fec7b36c6a79ac8df292"} Oct 08 19:44:30 crc kubenswrapper[4988]: I1008 19:44:30.304815 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-8807-account-create-2rpmf" podStartSLOduration=1.304801189 podStartE2EDuration="1.304801189s" podCreationTimestamp="2025-10-08 19:44:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 19:44:30.304587732 +0000 UTC m=+5615.754430502" watchObservedRunningTime="2025-10-08 19:44:30.304801189 +0000 UTC m=+5615.754643959" Oct 08 19:44:31 crc kubenswrapper[4988]: I1008 19:44:31.300029 4988 generic.go:334] "Generic (PLEG): container finished" podID="70fc1b25-cb88-4345-b70f-e72c74df4a0f" containerID="6a6df4cf78aeb6f22196fcbff02c8005b18ea265c81fd0b5af42b23ab4537dff" exitCode=0 Oct 08 19:44:31 crc kubenswrapper[4988]: I1008 19:44:31.300136 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-8807-account-create-2rpmf" event={"ID":"70fc1b25-cb88-4345-b70f-e72c74df4a0f","Type":"ContainerDied","Data":"6a6df4cf78aeb6f22196fcbff02c8005b18ea265c81fd0b5af42b23ab4537dff"} Oct 08 19:44:32 crc kubenswrapper[4988]: I1008 19:44:32.756683 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-8807-account-create-2rpmf" Oct 08 19:44:32 crc kubenswrapper[4988]: I1008 19:44:32.799507 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hjltx\" (UniqueName: \"kubernetes.io/projected/70fc1b25-cb88-4345-b70f-e72c74df4a0f-kube-api-access-hjltx\") pod \"70fc1b25-cb88-4345-b70f-e72c74df4a0f\" (UID: \"70fc1b25-cb88-4345-b70f-e72c74df4a0f\") " Oct 08 19:44:32 crc kubenswrapper[4988]: I1008 19:44:32.806080 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/70fc1b25-cb88-4345-b70f-e72c74df4a0f-kube-api-access-hjltx" (OuterVolumeSpecName: "kube-api-access-hjltx") pod "70fc1b25-cb88-4345-b70f-e72c74df4a0f" (UID: "70fc1b25-cb88-4345-b70f-e72c74df4a0f"). InnerVolumeSpecName "kube-api-access-hjltx". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 19:44:32 crc kubenswrapper[4988]: I1008 19:44:32.901425 4988 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hjltx\" (UniqueName: \"kubernetes.io/projected/70fc1b25-cb88-4345-b70f-e72c74df4a0f-kube-api-access-hjltx\") on node \"crc\" DevicePath \"\"" Oct 08 19:44:33 crc kubenswrapper[4988]: I1008 19:44:33.326657 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-8807-account-create-2rpmf" event={"ID":"70fc1b25-cb88-4345-b70f-e72c74df4a0f","Type":"ContainerDied","Data":"2e98a6f1c65ef204fc59ea9cd3b5da4538b0cceb6056fec7b36c6a79ac8df292"} Oct 08 19:44:33 crc kubenswrapper[4988]: I1008 19:44:33.326722 4988 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2e98a6f1c65ef204fc59ea9cd3b5da4538b0cceb6056fec7b36c6a79ac8df292" Oct 08 19:44:33 crc kubenswrapper[4988]: I1008 19:44:33.326745 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-8807-account-create-2rpmf" Oct 08 19:44:34 crc kubenswrapper[4988]: I1008 19:44:34.525785 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-sync-r4nvb"] Oct 08 19:44:34 crc kubenswrapper[4988]: E1008 19:44:34.526179 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="70fc1b25-cb88-4345-b70f-e72c74df4a0f" containerName="mariadb-account-create" Oct 08 19:44:34 crc kubenswrapper[4988]: I1008 19:44:34.526194 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="70fc1b25-cb88-4345-b70f-e72c74df4a0f" containerName="mariadb-account-create" Oct 08 19:44:34 crc kubenswrapper[4988]: I1008 19:44:34.526437 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="70fc1b25-cb88-4345-b70f-e72c74df4a0f" containerName="mariadb-account-create" Oct 08 19:44:34 crc kubenswrapper[4988]: I1008 19:44:34.527112 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-r4nvb" Oct 08 19:44:34 crc kubenswrapper[4988]: I1008 19:44:34.529717 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-4tfrg" Oct 08 19:44:34 crc kubenswrapper[4988]: I1008 19:44:34.529726 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Oct 08 19:44:34 crc kubenswrapper[4988]: I1008 19:44:34.530727 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Oct 08 19:44:34 crc kubenswrapper[4988]: I1008 19:44:34.546672 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-r4nvb"] Oct 08 19:44:34 crc kubenswrapper[4988]: I1008 19:44:34.635956 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-42gk6\" (UniqueName: \"kubernetes.io/projected/74701b1e-3a3a-47e6-9794-46b76ec341e5-kube-api-access-42gk6\") pod \"neutron-db-sync-r4nvb\" (UID: \"74701b1e-3a3a-47e6-9794-46b76ec341e5\") " pod="openstack/neutron-db-sync-r4nvb" Oct 08 19:44:34 crc kubenswrapper[4988]: I1008 19:44:34.636215 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/74701b1e-3a3a-47e6-9794-46b76ec341e5-combined-ca-bundle\") pod \"neutron-db-sync-r4nvb\" (UID: \"74701b1e-3a3a-47e6-9794-46b76ec341e5\") " pod="openstack/neutron-db-sync-r4nvb" Oct 08 19:44:34 crc kubenswrapper[4988]: I1008 19:44:34.636274 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/74701b1e-3a3a-47e6-9794-46b76ec341e5-config\") pod \"neutron-db-sync-r4nvb\" (UID: \"74701b1e-3a3a-47e6-9794-46b76ec341e5\") " pod="openstack/neutron-db-sync-r4nvb" Oct 08 19:44:34 crc kubenswrapper[4988]: I1008 19:44:34.738021 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/74701b1e-3a3a-47e6-9794-46b76ec341e5-combined-ca-bundle\") pod \"neutron-db-sync-r4nvb\" (UID: \"74701b1e-3a3a-47e6-9794-46b76ec341e5\") " pod="openstack/neutron-db-sync-r4nvb" Oct 08 19:44:34 crc kubenswrapper[4988]: I1008 19:44:34.738511 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/74701b1e-3a3a-47e6-9794-46b76ec341e5-config\") pod \"neutron-db-sync-r4nvb\" (UID: \"74701b1e-3a3a-47e6-9794-46b76ec341e5\") " pod="openstack/neutron-db-sync-r4nvb" Oct 08 19:44:34 crc kubenswrapper[4988]: I1008 19:44:34.738625 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-42gk6\" (UniqueName: \"kubernetes.io/projected/74701b1e-3a3a-47e6-9794-46b76ec341e5-kube-api-access-42gk6\") pod \"neutron-db-sync-r4nvb\" (UID: \"74701b1e-3a3a-47e6-9794-46b76ec341e5\") " pod="openstack/neutron-db-sync-r4nvb" Oct 08 19:44:34 crc kubenswrapper[4988]: I1008 19:44:34.743627 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/74701b1e-3a3a-47e6-9794-46b76ec341e5-config\") pod \"neutron-db-sync-r4nvb\" (UID: \"74701b1e-3a3a-47e6-9794-46b76ec341e5\") " pod="openstack/neutron-db-sync-r4nvb" Oct 08 19:44:34 crc kubenswrapper[4988]: I1008 19:44:34.746077 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/74701b1e-3a3a-47e6-9794-46b76ec341e5-combined-ca-bundle\") pod \"neutron-db-sync-r4nvb\" (UID: \"74701b1e-3a3a-47e6-9794-46b76ec341e5\") " pod="openstack/neutron-db-sync-r4nvb" Oct 08 19:44:34 crc kubenswrapper[4988]: I1008 19:44:34.761276 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-42gk6\" (UniqueName: \"kubernetes.io/projected/74701b1e-3a3a-47e6-9794-46b76ec341e5-kube-api-access-42gk6\") pod \"neutron-db-sync-r4nvb\" (UID: \"74701b1e-3a3a-47e6-9794-46b76ec341e5\") " pod="openstack/neutron-db-sync-r4nvb" Oct 08 19:44:34 crc kubenswrapper[4988]: I1008 19:44:34.846776 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-r4nvb" Oct 08 19:44:35 crc kubenswrapper[4988]: I1008 19:44:35.321447 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-r4nvb"] Oct 08 19:44:35 crc kubenswrapper[4988]: I1008 19:44:35.350415 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-r4nvb" event={"ID":"74701b1e-3a3a-47e6-9794-46b76ec341e5","Type":"ContainerStarted","Data":"01294078efe2dce8617364a471429e131d46b8ca10a1fef7b2a18a3c0c86bc27"} Oct 08 19:44:36 crc kubenswrapper[4988]: I1008 19:44:36.365479 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-r4nvb" event={"ID":"74701b1e-3a3a-47e6-9794-46b76ec341e5","Type":"ContainerStarted","Data":"dff17961ab3ec0388707e7e082251560ce5645182a16215c1a0cc4d8267e7db6"} Oct 08 19:44:36 crc kubenswrapper[4988]: I1008 19:44:36.393120 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-db-sync-r4nvb" podStartSLOduration=2.393090733 podStartE2EDuration="2.393090733s" podCreationTimestamp="2025-10-08 19:44:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 19:44:36.3864021 +0000 UTC m=+5621.836244880" watchObservedRunningTime="2025-10-08 19:44:36.393090733 +0000 UTC m=+5621.842933533" Oct 08 19:44:39 crc kubenswrapper[4988]: I1008 19:44:39.403879 4988 generic.go:334] "Generic (PLEG): container finished" podID="74701b1e-3a3a-47e6-9794-46b76ec341e5" containerID="dff17961ab3ec0388707e7e082251560ce5645182a16215c1a0cc4d8267e7db6" exitCode=0 Oct 08 19:44:39 crc kubenswrapper[4988]: I1008 19:44:39.403989 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-r4nvb" event={"ID":"74701b1e-3a3a-47e6-9794-46b76ec341e5","Type":"ContainerDied","Data":"dff17961ab3ec0388707e7e082251560ce5645182a16215c1a0cc4d8267e7db6"} Oct 08 19:44:40 crc kubenswrapper[4988]: I1008 19:44:40.885609 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-r4nvb" Oct 08 19:44:41 crc kubenswrapper[4988]: I1008 19:44:41.062023 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-42gk6\" (UniqueName: \"kubernetes.io/projected/74701b1e-3a3a-47e6-9794-46b76ec341e5-kube-api-access-42gk6\") pod \"74701b1e-3a3a-47e6-9794-46b76ec341e5\" (UID: \"74701b1e-3a3a-47e6-9794-46b76ec341e5\") " Oct 08 19:44:41 crc kubenswrapper[4988]: I1008 19:44:41.062706 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/74701b1e-3a3a-47e6-9794-46b76ec341e5-config\") pod \"74701b1e-3a3a-47e6-9794-46b76ec341e5\" (UID: \"74701b1e-3a3a-47e6-9794-46b76ec341e5\") " Oct 08 19:44:41 crc kubenswrapper[4988]: I1008 19:44:41.062787 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/74701b1e-3a3a-47e6-9794-46b76ec341e5-combined-ca-bundle\") pod \"74701b1e-3a3a-47e6-9794-46b76ec341e5\" (UID: \"74701b1e-3a3a-47e6-9794-46b76ec341e5\") " Oct 08 19:44:41 crc kubenswrapper[4988]: I1008 19:44:41.068581 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/74701b1e-3a3a-47e6-9794-46b76ec341e5-kube-api-access-42gk6" (OuterVolumeSpecName: "kube-api-access-42gk6") pod "74701b1e-3a3a-47e6-9794-46b76ec341e5" (UID: "74701b1e-3a3a-47e6-9794-46b76ec341e5"). InnerVolumeSpecName "kube-api-access-42gk6". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 19:44:41 crc kubenswrapper[4988]: I1008 19:44:41.092287 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/74701b1e-3a3a-47e6-9794-46b76ec341e5-config" (OuterVolumeSpecName: "config") pod "74701b1e-3a3a-47e6-9794-46b76ec341e5" (UID: "74701b1e-3a3a-47e6-9794-46b76ec341e5"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 19:44:41 crc kubenswrapper[4988]: I1008 19:44:41.096142 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/74701b1e-3a3a-47e6-9794-46b76ec341e5-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "74701b1e-3a3a-47e6-9794-46b76ec341e5" (UID: "74701b1e-3a3a-47e6-9794-46b76ec341e5"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 19:44:41 crc kubenswrapper[4988]: I1008 19:44:41.165496 4988 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-42gk6\" (UniqueName: \"kubernetes.io/projected/74701b1e-3a3a-47e6-9794-46b76ec341e5-kube-api-access-42gk6\") on node \"crc\" DevicePath \"\"" Oct 08 19:44:41 crc kubenswrapper[4988]: I1008 19:44:41.166010 4988 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/74701b1e-3a3a-47e6-9794-46b76ec341e5-config\") on node \"crc\" DevicePath \"\"" Oct 08 19:44:41 crc kubenswrapper[4988]: I1008 19:44:41.166103 4988 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/74701b1e-3a3a-47e6-9794-46b76ec341e5-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 19:44:41 crc kubenswrapper[4988]: I1008 19:44:41.429212 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-r4nvb" event={"ID":"74701b1e-3a3a-47e6-9794-46b76ec341e5","Type":"ContainerDied","Data":"01294078efe2dce8617364a471429e131d46b8ca10a1fef7b2a18a3c0c86bc27"} Oct 08 19:44:41 crc kubenswrapper[4988]: I1008 19:44:41.429258 4988 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="01294078efe2dce8617364a471429e131d46b8ca10a1fef7b2a18a3c0c86bc27" Oct 08 19:44:41 crc kubenswrapper[4988]: I1008 19:44:41.429969 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-r4nvb" Oct 08 19:44:41 crc kubenswrapper[4988]: I1008 19:44:41.682276 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-fbc86f6b9-gwjfv"] Oct 08 19:44:41 crc kubenswrapper[4988]: E1008 19:44:41.688007 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="74701b1e-3a3a-47e6-9794-46b76ec341e5" containerName="neutron-db-sync" Oct 08 19:44:41 crc kubenswrapper[4988]: I1008 19:44:41.688040 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="74701b1e-3a3a-47e6-9794-46b76ec341e5" containerName="neutron-db-sync" Oct 08 19:44:41 crc kubenswrapper[4988]: I1008 19:44:41.688203 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="74701b1e-3a3a-47e6-9794-46b76ec341e5" containerName="neutron-db-sync" Oct 08 19:44:41 crc kubenswrapper[4988]: I1008 19:44:41.689109 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-fbc86f6b9-gwjfv" Oct 08 19:44:41 crc kubenswrapper[4988]: I1008 19:44:41.698630 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-fbc86f6b9-gwjfv"] Oct 08 19:44:41 crc kubenswrapper[4988]: I1008 19:44:41.784178 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c9742f58-7745-4b60-b61d-e864ceecf087-dns-svc\") pod \"dnsmasq-dns-fbc86f6b9-gwjfv\" (UID: \"c9742f58-7745-4b60-b61d-e864ceecf087\") " pod="openstack/dnsmasq-dns-fbc86f6b9-gwjfv" Oct 08 19:44:41 crc kubenswrapper[4988]: I1008 19:44:41.784251 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jdh86\" (UniqueName: \"kubernetes.io/projected/c9742f58-7745-4b60-b61d-e864ceecf087-kube-api-access-jdh86\") pod \"dnsmasq-dns-fbc86f6b9-gwjfv\" (UID: \"c9742f58-7745-4b60-b61d-e864ceecf087\") " pod="openstack/dnsmasq-dns-fbc86f6b9-gwjfv" Oct 08 19:44:41 crc kubenswrapper[4988]: I1008 19:44:41.784465 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c9742f58-7745-4b60-b61d-e864ceecf087-ovsdbserver-sb\") pod \"dnsmasq-dns-fbc86f6b9-gwjfv\" (UID: \"c9742f58-7745-4b60-b61d-e864ceecf087\") " pod="openstack/dnsmasq-dns-fbc86f6b9-gwjfv" Oct 08 19:44:41 crc kubenswrapper[4988]: I1008 19:44:41.784514 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c9742f58-7745-4b60-b61d-e864ceecf087-config\") pod \"dnsmasq-dns-fbc86f6b9-gwjfv\" (UID: \"c9742f58-7745-4b60-b61d-e864ceecf087\") " pod="openstack/dnsmasq-dns-fbc86f6b9-gwjfv" Oct 08 19:44:41 crc kubenswrapper[4988]: I1008 19:44:41.784548 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c9742f58-7745-4b60-b61d-e864ceecf087-ovsdbserver-nb\") pod \"dnsmasq-dns-fbc86f6b9-gwjfv\" (UID: \"c9742f58-7745-4b60-b61d-e864ceecf087\") " pod="openstack/dnsmasq-dns-fbc86f6b9-gwjfv" Oct 08 19:44:41 crc kubenswrapper[4988]: I1008 19:44:41.885803 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c9742f58-7745-4b60-b61d-e864ceecf087-config\") pod \"dnsmasq-dns-fbc86f6b9-gwjfv\" (UID: \"c9742f58-7745-4b60-b61d-e864ceecf087\") " pod="openstack/dnsmasq-dns-fbc86f6b9-gwjfv" Oct 08 19:44:41 crc kubenswrapper[4988]: I1008 19:44:41.885852 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c9742f58-7745-4b60-b61d-e864ceecf087-ovsdbserver-nb\") pod \"dnsmasq-dns-fbc86f6b9-gwjfv\" (UID: \"c9742f58-7745-4b60-b61d-e864ceecf087\") " pod="openstack/dnsmasq-dns-fbc86f6b9-gwjfv" Oct 08 19:44:41 crc kubenswrapper[4988]: I1008 19:44:41.885882 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c9742f58-7745-4b60-b61d-e864ceecf087-dns-svc\") pod \"dnsmasq-dns-fbc86f6b9-gwjfv\" (UID: \"c9742f58-7745-4b60-b61d-e864ceecf087\") " pod="openstack/dnsmasq-dns-fbc86f6b9-gwjfv" Oct 08 19:44:41 crc kubenswrapper[4988]: I1008 19:44:41.885911 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jdh86\" (UniqueName: \"kubernetes.io/projected/c9742f58-7745-4b60-b61d-e864ceecf087-kube-api-access-jdh86\") pod \"dnsmasq-dns-fbc86f6b9-gwjfv\" (UID: \"c9742f58-7745-4b60-b61d-e864ceecf087\") " pod="openstack/dnsmasq-dns-fbc86f6b9-gwjfv" Oct 08 19:44:41 crc kubenswrapper[4988]: I1008 19:44:41.886014 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c9742f58-7745-4b60-b61d-e864ceecf087-ovsdbserver-sb\") pod \"dnsmasq-dns-fbc86f6b9-gwjfv\" (UID: \"c9742f58-7745-4b60-b61d-e864ceecf087\") " pod="openstack/dnsmasq-dns-fbc86f6b9-gwjfv" Oct 08 19:44:41 crc kubenswrapper[4988]: I1008 19:44:41.887010 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c9742f58-7745-4b60-b61d-e864ceecf087-ovsdbserver-sb\") pod \"dnsmasq-dns-fbc86f6b9-gwjfv\" (UID: \"c9742f58-7745-4b60-b61d-e864ceecf087\") " pod="openstack/dnsmasq-dns-fbc86f6b9-gwjfv" Oct 08 19:44:41 crc kubenswrapper[4988]: I1008 19:44:41.887056 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c9742f58-7745-4b60-b61d-e864ceecf087-config\") pod \"dnsmasq-dns-fbc86f6b9-gwjfv\" (UID: \"c9742f58-7745-4b60-b61d-e864ceecf087\") " pod="openstack/dnsmasq-dns-fbc86f6b9-gwjfv" Oct 08 19:44:41 crc kubenswrapper[4988]: I1008 19:44:41.887056 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c9742f58-7745-4b60-b61d-e864ceecf087-ovsdbserver-nb\") pod \"dnsmasq-dns-fbc86f6b9-gwjfv\" (UID: \"c9742f58-7745-4b60-b61d-e864ceecf087\") " pod="openstack/dnsmasq-dns-fbc86f6b9-gwjfv" Oct 08 19:44:41 crc kubenswrapper[4988]: I1008 19:44:41.887536 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c9742f58-7745-4b60-b61d-e864ceecf087-dns-svc\") pod \"dnsmasq-dns-fbc86f6b9-gwjfv\" (UID: \"c9742f58-7745-4b60-b61d-e864ceecf087\") " pod="openstack/dnsmasq-dns-fbc86f6b9-gwjfv" Oct 08 19:44:41 crc kubenswrapper[4988]: I1008 19:44:41.898419 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-c458d797b-jlpvk"] Oct 08 19:44:41 crc kubenswrapper[4988]: I1008 19:44:41.900157 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-c458d797b-jlpvk" Oct 08 19:44:41 crc kubenswrapper[4988]: I1008 19:44:41.902637 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-ovndbs" Oct 08 19:44:41 crc kubenswrapper[4988]: I1008 19:44:41.903076 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Oct 08 19:44:41 crc kubenswrapper[4988]: I1008 19:44:41.903494 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-4tfrg" Oct 08 19:44:41 crc kubenswrapper[4988]: I1008 19:44:41.903467 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Oct 08 19:44:41 crc kubenswrapper[4988]: I1008 19:44:41.905659 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-c458d797b-jlpvk"] Oct 08 19:44:41 crc kubenswrapper[4988]: I1008 19:44:41.913863 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jdh86\" (UniqueName: \"kubernetes.io/projected/c9742f58-7745-4b60-b61d-e864ceecf087-kube-api-access-jdh86\") pod \"dnsmasq-dns-fbc86f6b9-gwjfv\" (UID: \"c9742f58-7745-4b60-b61d-e864ceecf087\") " pod="openstack/dnsmasq-dns-fbc86f6b9-gwjfv" Oct 08 19:44:41 crc kubenswrapper[4988]: I1008 19:44:41.987777 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8635e509-34ea-43b5-955d-db4677f28ad9-combined-ca-bundle\") pod \"neutron-c458d797b-jlpvk\" (UID: \"8635e509-34ea-43b5-955d-db4677f28ad9\") " pod="openstack/neutron-c458d797b-jlpvk" Oct 08 19:44:41 crc kubenswrapper[4988]: I1008 19:44:41.987852 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/8635e509-34ea-43b5-955d-db4677f28ad9-ovndb-tls-certs\") pod \"neutron-c458d797b-jlpvk\" (UID: \"8635e509-34ea-43b5-955d-db4677f28ad9\") " pod="openstack/neutron-c458d797b-jlpvk" Oct 08 19:44:41 crc kubenswrapper[4988]: I1008 19:44:41.987904 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/8635e509-34ea-43b5-955d-db4677f28ad9-httpd-config\") pod \"neutron-c458d797b-jlpvk\" (UID: \"8635e509-34ea-43b5-955d-db4677f28ad9\") " pod="openstack/neutron-c458d797b-jlpvk" Oct 08 19:44:41 crc kubenswrapper[4988]: I1008 19:44:41.987960 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/8635e509-34ea-43b5-955d-db4677f28ad9-config\") pod \"neutron-c458d797b-jlpvk\" (UID: \"8635e509-34ea-43b5-955d-db4677f28ad9\") " pod="openstack/neutron-c458d797b-jlpvk" Oct 08 19:44:41 crc kubenswrapper[4988]: I1008 19:44:41.987984 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dsj24\" (UniqueName: \"kubernetes.io/projected/8635e509-34ea-43b5-955d-db4677f28ad9-kube-api-access-dsj24\") pod \"neutron-c458d797b-jlpvk\" (UID: \"8635e509-34ea-43b5-955d-db4677f28ad9\") " pod="openstack/neutron-c458d797b-jlpvk" Oct 08 19:44:42 crc kubenswrapper[4988]: I1008 19:44:42.007311 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-fbc86f6b9-gwjfv" Oct 08 19:44:42 crc kubenswrapper[4988]: I1008 19:44:42.089642 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/8635e509-34ea-43b5-955d-db4677f28ad9-httpd-config\") pod \"neutron-c458d797b-jlpvk\" (UID: \"8635e509-34ea-43b5-955d-db4677f28ad9\") " pod="openstack/neutron-c458d797b-jlpvk" Oct 08 19:44:42 crc kubenswrapper[4988]: I1008 19:44:42.090282 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/8635e509-34ea-43b5-955d-db4677f28ad9-config\") pod \"neutron-c458d797b-jlpvk\" (UID: \"8635e509-34ea-43b5-955d-db4677f28ad9\") " pod="openstack/neutron-c458d797b-jlpvk" Oct 08 19:44:42 crc kubenswrapper[4988]: I1008 19:44:42.090465 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dsj24\" (UniqueName: \"kubernetes.io/projected/8635e509-34ea-43b5-955d-db4677f28ad9-kube-api-access-dsj24\") pod \"neutron-c458d797b-jlpvk\" (UID: \"8635e509-34ea-43b5-955d-db4677f28ad9\") " pod="openstack/neutron-c458d797b-jlpvk" Oct 08 19:44:42 crc kubenswrapper[4988]: I1008 19:44:42.090597 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8635e509-34ea-43b5-955d-db4677f28ad9-combined-ca-bundle\") pod \"neutron-c458d797b-jlpvk\" (UID: \"8635e509-34ea-43b5-955d-db4677f28ad9\") " pod="openstack/neutron-c458d797b-jlpvk" Oct 08 19:44:42 crc kubenswrapper[4988]: I1008 19:44:42.090757 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/8635e509-34ea-43b5-955d-db4677f28ad9-ovndb-tls-certs\") pod \"neutron-c458d797b-jlpvk\" (UID: \"8635e509-34ea-43b5-955d-db4677f28ad9\") " pod="openstack/neutron-c458d797b-jlpvk" Oct 08 19:44:42 crc kubenswrapper[4988]: I1008 19:44:42.093282 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/8635e509-34ea-43b5-955d-db4677f28ad9-httpd-config\") pod \"neutron-c458d797b-jlpvk\" (UID: \"8635e509-34ea-43b5-955d-db4677f28ad9\") " pod="openstack/neutron-c458d797b-jlpvk" Oct 08 19:44:42 crc kubenswrapper[4988]: I1008 19:44:42.093908 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/8635e509-34ea-43b5-955d-db4677f28ad9-config\") pod \"neutron-c458d797b-jlpvk\" (UID: \"8635e509-34ea-43b5-955d-db4677f28ad9\") " pod="openstack/neutron-c458d797b-jlpvk" Oct 08 19:44:42 crc kubenswrapper[4988]: I1008 19:44:42.095552 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/8635e509-34ea-43b5-955d-db4677f28ad9-ovndb-tls-certs\") pod \"neutron-c458d797b-jlpvk\" (UID: \"8635e509-34ea-43b5-955d-db4677f28ad9\") " pod="openstack/neutron-c458d797b-jlpvk" Oct 08 19:44:42 crc kubenswrapper[4988]: I1008 19:44:42.102238 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8635e509-34ea-43b5-955d-db4677f28ad9-combined-ca-bundle\") pod \"neutron-c458d797b-jlpvk\" (UID: \"8635e509-34ea-43b5-955d-db4677f28ad9\") " pod="openstack/neutron-c458d797b-jlpvk" Oct 08 19:44:42 crc kubenswrapper[4988]: I1008 19:44:42.116098 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dsj24\" (UniqueName: \"kubernetes.io/projected/8635e509-34ea-43b5-955d-db4677f28ad9-kube-api-access-dsj24\") pod \"neutron-c458d797b-jlpvk\" (UID: \"8635e509-34ea-43b5-955d-db4677f28ad9\") " pod="openstack/neutron-c458d797b-jlpvk" Oct 08 19:44:42 crc kubenswrapper[4988]: I1008 19:44:42.256465 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-c458d797b-jlpvk" Oct 08 19:44:42 crc kubenswrapper[4988]: I1008 19:44:42.470623 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-fbc86f6b9-gwjfv"] Oct 08 19:44:42 crc kubenswrapper[4988]: I1008 19:44:42.762483 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-c458d797b-jlpvk"] Oct 08 19:44:42 crc kubenswrapper[4988]: E1008 19:44:42.831753 4988 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc9742f58_7745_4b60_b61d_e864ceecf087.slice/crio-conmon-3cd75657be77ae5c79df8af727ac7439f43f1564944848bdd1528f1b162ff4a1.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc9742f58_7745_4b60_b61d_e864ceecf087.slice/crio-3cd75657be77ae5c79df8af727ac7439f43f1564944848bdd1528f1b162ff4a1.scope\": RecentStats: unable to find data in memory cache]" Oct 08 19:44:43 crc kubenswrapper[4988]: I1008 19:44:43.450726 4988 generic.go:334] "Generic (PLEG): container finished" podID="c9742f58-7745-4b60-b61d-e864ceecf087" containerID="3cd75657be77ae5c79df8af727ac7439f43f1564944848bdd1528f1b162ff4a1" exitCode=0 Oct 08 19:44:43 crc kubenswrapper[4988]: I1008 19:44:43.451152 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-fbc86f6b9-gwjfv" event={"ID":"c9742f58-7745-4b60-b61d-e864ceecf087","Type":"ContainerDied","Data":"3cd75657be77ae5c79df8af727ac7439f43f1564944848bdd1528f1b162ff4a1"} Oct 08 19:44:43 crc kubenswrapper[4988]: I1008 19:44:43.451179 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-fbc86f6b9-gwjfv" event={"ID":"c9742f58-7745-4b60-b61d-e864ceecf087","Type":"ContainerStarted","Data":"d87660c7e9ab7242d4e5b27d3809ac4c2356a06887540b0b5f2ae3d78655d5f3"} Oct 08 19:44:43 crc kubenswrapper[4988]: I1008 19:44:43.454950 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-c458d797b-jlpvk" event={"ID":"8635e509-34ea-43b5-955d-db4677f28ad9","Type":"ContainerStarted","Data":"d7d1cd0efaa71a26868202a003120cbac6962702516001eb5a915193a7148311"} Oct 08 19:44:43 crc kubenswrapper[4988]: I1008 19:44:43.455030 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-c458d797b-jlpvk" event={"ID":"8635e509-34ea-43b5-955d-db4677f28ad9","Type":"ContainerStarted","Data":"f9f6c9c6d01553f2e40f5d7f951b43c8679e350a27241c60aef27549dd34bab8"} Oct 08 19:44:43 crc kubenswrapper[4988]: I1008 19:44:43.455043 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-c458d797b-jlpvk" event={"ID":"8635e509-34ea-43b5-955d-db4677f28ad9","Type":"ContainerStarted","Data":"9a7bbfdd04db51f43746cacf553ca575ba49de3375bf5621c0d7e77afbab0d2b"} Oct 08 19:44:43 crc kubenswrapper[4988]: I1008 19:44:43.455119 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-c458d797b-jlpvk" Oct 08 19:44:43 crc kubenswrapper[4988]: I1008 19:44:43.501000 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-c458d797b-jlpvk" podStartSLOduration=2.500986228 podStartE2EDuration="2.500986228s" podCreationTimestamp="2025-10-08 19:44:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 19:44:43.497543588 +0000 UTC m=+5628.947386358" watchObservedRunningTime="2025-10-08 19:44:43.500986228 +0000 UTC m=+5628.950828998" Oct 08 19:44:44 crc kubenswrapper[4988]: I1008 19:44:44.427886 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-77f8f8964c-wclf6"] Oct 08 19:44:44 crc kubenswrapper[4988]: I1008 19:44:44.430921 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-77f8f8964c-wclf6" Oct 08 19:44:44 crc kubenswrapper[4988]: I1008 19:44:44.433960 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-internal-svc" Oct 08 19:44:44 crc kubenswrapper[4988]: I1008 19:44:44.434185 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-public-svc" Oct 08 19:44:44 crc kubenswrapper[4988]: I1008 19:44:44.442260 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-77f8f8964c-wclf6"] Oct 08 19:44:44 crc kubenswrapper[4988]: I1008 19:44:44.472144 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-fbc86f6b9-gwjfv" event={"ID":"c9742f58-7745-4b60-b61d-e864ceecf087","Type":"ContainerStarted","Data":"b491f91185e11de75a9e806ba7683fef2192b04aeda5c624f5f816ddf28e9426"} Oct 08 19:44:44 crc kubenswrapper[4988]: I1008 19:44:44.502563 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-fbc86f6b9-gwjfv" podStartSLOduration=3.502540675 podStartE2EDuration="3.502540675s" podCreationTimestamp="2025-10-08 19:44:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 19:44:44.495836001 +0000 UTC m=+5629.945678771" watchObservedRunningTime="2025-10-08 19:44:44.502540675 +0000 UTC m=+5629.952383445" Oct 08 19:44:44 crc kubenswrapper[4988]: I1008 19:44:44.532325 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/bc6e7082-3102-4dfb-af23-396d055e06a0-public-tls-certs\") pod \"neutron-77f8f8964c-wclf6\" (UID: \"bc6e7082-3102-4dfb-af23-396d055e06a0\") " pod="openstack/neutron-77f8f8964c-wclf6" Oct 08 19:44:44 crc kubenswrapper[4988]: I1008 19:44:44.532403 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hvnsr\" (UniqueName: \"kubernetes.io/projected/bc6e7082-3102-4dfb-af23-396d055e06a0-kube-api-access-hvnsr\") pod \"neutron-77f8f8964c-wclf6\" (UID: \"bc6e7082-3102-4dfb-af23-396d055e06a0\") " pod="openstack/neutron-77f8f8964c-wclf6" Oct 08 19:44:44 crc kubenswrapper[4988]: I1008 19:44:44.532464 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/bc6e7082-3102-4dfb-af23-396d055e06a0-config\") pod \"neutron-77f8f8964c-wclf6\" (UID: \"bc6e7082-3102-4dfb-af23-396d055e06a0\") " pod="openstack/neutron-77f8f8964c-wclf6" Oct 08 19:44:44 crc kubenswrapper[4988]: I1008 19:44:44.532480 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/bc6e7082-3102-4dfb-af23-396d055e06a0-httpd-config\") pod \"neutron-77f8f8964c-wclf6\" (UID: \"bc6e7082-3102-4dfb-af23-396d055e06a0\") " pod="openstack/neutron-77f8f8964c-wclf6" Oct 08 19:44:44 crc kubenswrapper[4988]: I1008 19:44:44.532621 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bc6e7082-3102-4dfb-af23-396d055e06a0-combined-ca-bundle\") pod \"neutron-77f8f8964c-wclf6\" (UID: \"bc6e7082-3102-4dfb-af23-396d055e06a0\") " pod="openstack/neutron-77f8f8964c-wclf6" Oct 08 19:44:44 crc kubenswrapper[4988]: I1008 19:44:44.532648 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/bc6e7082-3102-4dfb-af23-396d055e06a0-ovndb-tls-certs\") pod \"neutron-77f8f8964c-wclf6\" (UID: \"bc6e7082-3102-4dfb-af23-396d055e06a0\") " pod="openstack/neutron-77f8f8964c-wclf6" Oct 08 19:44:44 crc kubenswrapper[4988]: I1008 19:44:44.532671 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/bc6e7082-3102-4dfb-af23-396d055e06a0-internal-tls-certs\") pod \"neutron-77f8f8964c-wclf6\" (UID: \"bc6e7082-3102-4dfb-af23-396d055e06a0\") " pod="openstack/neutron-77f8f8964c-wclf6" Oct 08 19:44:44 crc kubenswrapper[4988]: I1008 19:44:44.634923 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bc6e7082-3102-4dfb-af23-396d055e06a0-combined-ca-bundle\") pod \"neutron-77f8f8964c-wclf6\" (UID: \"bc6e7082-3102-4dfb-af23-396d055e06a0\") " pod="openstack/neutron-77f8f8964c-wclf6" Oct 08 19:44:44 crc kubenswrapper[4988]: I1008 19:44:44.634992 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/bc6e7082-3102-4dfb-af23-396d055e06a0-ovndb-tls-certs\") pod \"neutron-77f8f8964c-wclf6\" (UID: \"bc6e7082-3102-4dfb-af23-396d055e06a0\") " pod="openstack/neutron-77f8f8964c-wclf6" Oct 08 19:44:44 crc kubenswrapper[4988]: I1008 19:44:44.635026 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/bc6e7082-3102-4dfb-af23-396d055e06a0-internal-tls-certs\") pod \"neutron-77f8f8964c-wclf6\" (UID: \"bc6e7082-3102-4dfb-af23-396d055e06a0\") " pod="openstack/neutron-77f8f8964c-wclf6" Oct 08 19:44:44 crc kubenswrapper[4988]: I1008 19:44:44.635062 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/bc6e7082-3102-4dfb-af23-396d055e06a0-public-tls-certs\") pod \"neutron-77f8f8964c-wclf6\" (UID: \"bc6e7082-3102-4dfb-af23-396d055e06a0\") " pod="openstack/neutron-77f8f8964c-wclf6" Oct 08 19:44:44 crc kubenswrapper[4988]: I1008 19:44:44.635114 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hvnsr\" (UniqueName: \"kubernetes.io/projected/bc6e7082-3102-4dfb-af23-396d055e06a0-kube-api-access-hvnsr\") pod \"neutron-77f8f8964c-wclf6\" (UID: \"bc6e7082-3102-4dfb-af23-396d055e06a0\") " pod="openstack/neutron-77f8f8964c-wclf6" Oct 08 19:44:44 crc kubenswrapper[4988]: I1008 19:44:44.635183 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/bc6e7082-3102-4dfb-af23-396d055e06a0-config\") pod \"neutron-77f8f8964c-wclf6\" (UID: \"bc6e7082-3102-4dfb-af23-396d055e06a0\") " pod="openstack/neutron-77f8f8964c-wclf6" Oct 08 19:44:44 crc kubenswrapper[4988]: I1008 19:44:44.635211 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/bc6e7082-3102-4dfb-af23-396d055e06a0-httpd-config\") pod \"neutron-77f8f8964c-wclf6\" (UID: \"bc6e7082-3102-4dfb-af23-396d055e06a0\") " pod="openstack/neutron-77f8f8964c-wclf6" Oct 08 19:44:44 crc kubenswrapper[4988]: I1008 19:44:44.639861 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/bc6e7082-3102-4dfb-af23-396d055e06a0-internal-tls-certs\") pod \"neutron-77f8f8964c-wclf6\" (UID: \"bc6e7082-3102-4dfb-af23-396d055e06a0\") " pod="openstack/neutron-77f8f8964c-wclf6" Oct 08 19:44:44 crc kubenswrapper[4988]: I1008 19:44:44.640169 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bc6e7082-3102-4dfb-af23-396d055e06a0-combined-ca-bundle\") pod \"neutron-77f8f8964c-wclf6\" (UID: \"bc6e7082-3102-4dfb-af23-396d055e06a0\") " pod="openstack/neutron-77f8f8964c-wclf6" Oct 08 19:44:44 crc kubenswrapper[4988]: I1008 19:44:44.642480 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/bc6e7082-3102-4dfb-af23-396d055e06a0-ovndb-tls-certs\") pod \"neutron-77f8f8964c-wclf6\" (UID: \"bc6e7082-3102-4dfb-af23-396d055e06a0\") " pod="openstack/neutron-77f8f8964c-wclf6" Oct 08 19:44:44 crc kubenswrapper[4988]: I1008 19:44:44.643271 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/bc6e7082-3102-4dfb-af23-396d055e06a0-config\") pod \"neutron-77f8f8964c-wclf6\" (UID: \"bc6e7082-3102-4dfb-af23-396d055e06a0\") " pod="openstack/neutron-77f8f8964c-wclf6" Oct 08 19:44:44 crc kubenswrapper[4988]: I1008 19:44:44.649170 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/bc6e7082-3102-4dfb-af23-396d055e06a0-httpd-config\") pod \"neutron-77f8f8964c-wclf6\" (UID: \"bc6e7082-3102-4dfb-af23-396d055e06a0\") " pod="openstack/neutron-77f8f8964c-wclf6" Oct 08 19:44:44 crc kubenswrapper[4988]: I1008 19:44:44.658835 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/bc6e7082-3102-4dfb-af23-396d055e06a0-public-tls-certs\") pod \"neutron-77f8f8964c-wclf6\" (UID: \"bc6e7082-3102-4dfb-af23-396d055e06a0\") " pod="openstack/neutron-77f8f8964c-wclf6" Oct 08 19:44:44 crc kubenswrapper[4988]: I1008 19:44:44.659941 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hvnsr\" (UniqueName: \"kubernetes.io/projected/bc6e7082-3102-4dfb-af23-396d055e06a0-kube-api-access-hvnsr\") pod \"neutron-77f8f8964c-wclf6\" (UID: \"bc6e7082-3102-4dfb-af23-396d055e06a0\") " pod="openstack/neutron-77f8f8964c-wclf6" Oct 08 19:44:44 crc kubenswrapper[4988]: I1008 19:44:44.755279 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-77f8f8964c-wclf6" Oct 08 19:44:45 crc kubenswrapper[4988]: I1008 19:44:45.285847 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-77f8f8964c-wclf6"] Oct 08 19:44:45 crc kubenswrapper[4988]: W1008 19:44:45.296539 4988 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podbc6e7082_3102_4dfb_af23_396d055e06a0.slice/crio-ab66a6446b4805a245fd3698d5c1306c758895089a3367adb66279dd8073fcf8 WatchSource:0}: Error finding container ab66a6446b4805a245fd3698d5c1306c758895089a3367adb66279dd8073fcf8: Status 404 returned error can't find the container with id ab66a6446b4805a245fd3698d5c1306c758895089a3367adb66279dd8073fcf8 Oct 08 19:44:45 crc kubenswrapper[4988]: I1008 19:44:45.481690 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-77f8f8964c-wclf6" event={"ID":"bc6e7082-3102-4dfb-af23-396d055e06a0","Type":"ContainerStarted","Data":"ab66a6446b4805a245fd3698d5c1306c758895089a3367adb66279dd8073fcf8"} Oct 08 19:44:45 crc kubenswrapper[4988]: I1008 19:44:45.482097 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-fbc86f6b9-gwjfv" Oct 08 19:44:46 crc kubenswrapper[4988]: I1008 19:44:46.495915 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-77f8f8964c-wclf6" event={"ID":"bc6e7082-3102-4dfb-af23-396d055e06a0","Type":"ContainerStarted","Data":"1aee535e464f78029bdd65eeab310e41138ac60105b557f3a16d56c6b7e8e4cd"} Oct 08 19:44:46 crc kubenswrapper[4988]: I1008 19:44:46.496364 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-77f8f8964c-wclf6" event={"ID":"bc6e7082-3102-4dfb-af23-396d055e06a0","Type":"ContainerStarted","Data":"6dbed9f818172047f5d03a62fe5cca5c201b0b87b6b48c3a773ee4a54ea8d969"} Oct 08 19:44:46 crc kubenswrapper[4988]: I1008 19:44:46.527468 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-77f8f8964c-wclf6" podStartSLOduration=2.527448242 podStartE2EDuration="2.527448242s" podCreationTimestamp="2025-10-08 19:44:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 19:44:46.52047949 +0000 UTC m=+5631.970322260" watchObservedRunningTime="2025-10-08 19:44:46.527448242 +0000 UTC m=+5631.977291012" Oct 08 19:44:47 crc kubenswrapper[4988]: I1008 19:44:47.508797 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-77f8f8964c-wclf6" Oct 08 19:44:52 crc kubenswrapper[4988]: I1008 19:44:52.009706 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-fbc86f6b9-gwjfv" Oct 08 19:44:52 crc kubenswrapper[4988]: I1008 19:44:52.104802 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-c5b46b9d5-npkpp"] Oct 08 19:44:52 crc kubenswrapper[4988]: I1008 19:44:52.105011 4988 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-c5b46b9d5-npkpp" podUID="dce2f14f-a039-4d24-b861-af2e48f06d2d" containerName="dnsmasq-dns" containerID="cri-o://7e3cc924ecc9eb8641f9e92f699dd1e29dd612ae90bccba932dfffe406eccc65" gracePeriod=10 Oct 08 19:44:52 crc kubenswrapper[4988]: I1008 19:44:52.557521 4988 generic.go:334] "Generic (PLEG): container finished" podID="dce2f14f-a039-4d24-b861-af2e48f06d2d" containerID="7e3cc924ecc9eb8641f9e92f699dd1e29dd612ae90bccba932dfffe406eccc65" exitCode=0 Oct 08 19:44:52 crc kubenswrapper[4988]: I1008 19:44:52.557582 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-c5b46b9d5-npkpp" event={"ID":"dce2f14f-a039-4d24-b861-af2e48f06d2d","Type":"ContainerDied","Data":"7e3cc924ecc9eb8641f9e92f699dd1e29dd612ae90bccba932dfffe406eccc65"} Oct 08 19:44:52 crc kubenswrapper[4988]: I1008 19:44:52.557885 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-c5b46b9d5-npkpp" event={"ID":"dce2f14f-a039-4d24-b861-af2e48f06d2d","Type":"ContainerDied","Data":"8822a79eee32a009bb537dc232a185dc32a97b675bb803558b3b70c6a46db931"} Oct 08 19:44:52 crc kubenswrapper[4988]: I1008 19:44:52.557910 4988 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8822a79eee32a009bb537dc232a185dc32a97b675bb803558b3b70c6a46db931" Oct 08 19:44:52 crc kubenswrapper[4988]: I1008 19:44:52.573960 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-c5b46b9d5-npkpp" Oct 08 19:44:52 crc kubenswrapper[4988]: I1008 19:44:52.608032 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/dce2f14f-a039-4d24-b861-af2e48f06d2d-dns-svc\") pod \"dce2f14f-a039-4d24-b861-af2e48f06d2d\" (UID: \"dce2f14f-a039-4d24-b861-af2e48f06d2d\") " Oct 08 19:44:52 crc kubenswrapper[4988]: I1008 19:44:52.608088 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/dce2f14f-a039-4d24-b861-af2e48f06d2d-ovsdbserver-sb\") pod \"dce2f14f-a039-4d24-b861-af2e48f06d2d\" (UID: \"dce2f14f-a039-4d24-b861-af2e48f06d2d\") " Oct 08 19:44:52 crc kubenswrapper[4988]: I1008 19:44:52.608271 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/dce2f14f-a039-4d24-b861-af2e48f06d2d-ovsdbserver-nb\") pod \"dce2f14f-a039-4d24-b861-af2e48f06d2d\" (UID: \"dce2f14f-a039-4d24-b861-af2e48f06d2d\") " Oct 08 19:44:52 crc kubenswrapper[4988]: I1008 19:44:52.608326 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-spnd2\" (UniqueName: \"kubernetes.io/projected/dce2f14f-a039-4d24-b861-af2e48f06d2d-kube-api-access-spnd2\") pod \"dce2f14f-a039-4d24-b861-af2e48f06d2d\" (UID: \"dce2f14f-a039-4d24-b861-af2e48f06d2d\") " Oct 08 19:44:52 crc kubenswrapper[4988]: I1008 19:44:52.608397 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dce2f14f-a039-4d24-b861-af2e48f06d2d-config\") pod \"dce2f14f-a039-4d24-b861-af2e48f06d2d\" (UID: \"dce2f14f-a039-4d24-b861-af2e48f06d2d\") " Oct 08 19:44:52 crc kubenswrapper[4988]: I1008 19:44:52.642958 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dce2f14f-a039-4d24-b861-af2e48f06d2d-kube-api-access-spnd2" (OuterVolumeSpecName: "kube-api-access-spnd2") pod "dce2f14f-a039-4d24-b861-af2e48f06d2d" (UID: "dce2f14f-a039-4d24-b861-af2e48f06d2d"). InnerVolumeSpecName "kube-api-access-spnd2". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 19:44:52 crc kubenswrapper[4988]: I1008 19:44:52.673760 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/dce2f14f-a039-4d24-b861-af2e48f06d2d-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "dce2f14f-a039-4d24-b861-af2e48f06d2d" (UID: "dce2f14f-a039-4d24-b861-af2e48f06d2d"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 19:44:52 crc kubenswrapper[4988]: I1008 19:44:52.676912 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/dce2f14f-a039-4d24-b861-af2e48f06d2d-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "dce2f14f-a039-4d24-b861-af2e48f06d2d" (UID: "dce2f14f-a039-4d24-b861-af2e48f06d2d"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 19:44:52 crc kubenswrapper[4988]: I1008 19:44:52.681852 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/dce2f14f-a039-4d24-b861-af2e48f06d2d-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "dce2f14f-a039-4d24-b861-af2e48f06d2d" (UID: "dce2f14f-a039-4d24-b861-af2e48f06d2d"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 19:44:52 crc kubenswrapper[4988]: I1008 19:44:52.682335 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/dce2f14f-a039-4d24-b861-af2e48f06d2d-config" (OuterVolumeSpecName: "config") pod "dce2f14f-a039-4d24-b861-af2e48f06d2d" (UID: "dce2f14f-a039-4d24-b861-af2e48f06d2d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 19:44:52 crc kubenswrapper[4988]: I1008 19:44:52.710553 4988 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-spnd2\" (UniqueName: \"kubernetes.io/projected/dce2f14f-a039-4d24-b861-af2e48f06d2d-kube-api-access-spnd2\") on node \"crc\" DevicePath \"\"" Oct 08 19:44:52 crc kubenswrapper[4988]: I1008 19:44:52.710585 4988 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dce2f14f-a039-4d24-b861-af2e48f06d2d-config\") on node \"crc\" DevicePath \"\"" Oct 08 19:44:52 crc kubenswrapper[4988]: I1008 19:44:52.710594 4988 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/dce2f14f-a039-4d24-b861-af2e48f06d2d-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 08 19:44:52 crc kubenswrapper[4988]: I1008 19:44:52.710603 4988 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/dce2f14f-a039-4d24-b861-af2e48f06d2d-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 08 19:44:52 crc kubenswrapper[4988]: I1008 19:44:52.710611 4988 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/dce2f14f-a039-4d24-b861-af2e48f06d2d-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 08 19:44:53 crc kubenswrapper[4988]: I1008 19:44:53.569299 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-c5b46b9d5-npkpp" Oct 08 19:44:53 crc kubenswrapper[4988]: I1008 19:44:53.620170 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-c5b46b9d5-npkpp"] Oct 08 19:44:53 crc kubenswrapper[4988]: I1008 19:44:53.631682 4988 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-c5b46b9d5-npkpp"] Oct 08 19:44:55 crc kubenswrapper[4988]: I1008 19:44:55.249737 4988 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="dce2f14f-a039-4d24-b861-af2e48f06d2d" path="/var/lib/kubelet/pods/dce2f14f-a039-4d24-b861-af2e48f06d2d/volumes" Oct 08 19:45:00 crc kubenswrapper[4988]: I1008 19:45:00.162937 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29332545-d245p"] Oct 08 19:45:00 crc kubenswrapper[4988]: E1008 19:45:00.163980 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dce2f14f-a039-4d24-b861-af2e48f06d2d" containerName="dnsmasq-dns" Oct 08 19:45:00 crc kubenswrapper[4988]: I1008 19:45:00.164000 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="dce2f14f-a039-4d24-b861-af2e48f06d2d" containerName="dnsmasq-dns" Oct 08 19:45:00 crc kubenswrapper[4988]: E1008 19:45:00.164032 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dce2f14f-a039-4d24-b861-af2e48f06d2d" containerName="init" Oct 08 19:45:00 crc kubenswrapper[4988]: I1008 19:45:00.164043 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="dce2f14f-a039-4d24-b861-af2e48f06d2d" containerName="init" Oct 08 19:45:00 crc kubenswrapper[4988]: I1008 19:45:00.164349 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="dce2f14f-a039-4d24-b861-af2e48f06d2d" containerName="dnsmasq-dns" Oct 08 19:45:00 crc kubenswrapper[4988]: I1008 19:45:00.165225 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29332545-d245p" Oct 08 19:45:00 crc kubenswrapper[4988]: I1008 19:45:00.170896 4988 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 08 19:45:00 crc kubenswrapper[4988]: I1008 19:45:00.172095 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 08 19:45:00 crc kubenswrapper[4988]: I1008 19:45:00.178859 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29332545-d245p"] Oct 08 19:45:00 crc kubenswrapper[4988]: I1008 19:45:00.353864 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c88f1935-f363-4596-9b74-fe8b3ee2946a-config-volume\") pod \"collect-profiles-29332545-d245p\" (UID: \"c88f1935-f363-4596-9b74-fe8b3ee2946a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29332545-d245p" Oct 08 19:45:00 crc kubenswrapper[4988]: I1008 19:45:00.354031 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/c88f1935-f363-4596-9b74-fe8b3ee2946a-secret-volume\") pod \"collect-profiles-29332545-d245p\" (UID: \"c88f1935-f363-4596-9b74-fe8b3ee2946a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29332545-d245p" Oct 08 19:45:00 crc kubenswrapper[4988]: I1008 19:45:00.354077 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ffkld\" (UniqueName: \"kubernetes.io/projected/c88f1935-f363-4596-9b74-fe8b3ee2946a-kube-api-access-ffkld\") pod \"collect-profiles-29332545-d245p\" (UID: \"c88f1935-f363-4596-9b74-fe8b3ee2946a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29332545-d245p" Oct 08 19:45:00 crc kubenswrapper[4988]: I1008 19:45:00.455526 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c88f1935-f363-4596-9b74-fe8b3ee2946a-config-volume\") pod \"collect-profiles-29332545-d245p\" (UID: \"c88f1935-f363-4596-9b74-fe8b3ee2946a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29332545-d245p" Oct 08 19:45:00 crc kubenswrapper[4988]: I1008 19:45:00.455705 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/c88f1935-f363-4596-9b74-fe8b3ee2946a-secret-volume\") pod \"collect-profiles-29332545-d245p\" (UID: \"c88f1935-f363-4596-9b74-fe8b3ee2946a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29332545-d245p" Oct 08 19:45:00 crc kubenswrapper[4988]: I1008 19:45:00.455774 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ffkld\" (UniqueName: \"kubernetes.io/projected/c88f1935-f363-4596-9b74-fe8b3ee2946a-kube-api-access-ffkld\") pod \"collect-profiles-29332545-d245p\" (UID: \"c88f1935-f363-4596-9b74-fe8b3ee2946a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29332545-d245p" Oct 08 19:45:00 crc kubenswrapper[4988]: I1008 19:45:00.456713 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c88f1935-f363-4596-9b74-fe8b3ee2946a-config-volume\") pod \"collect-profiles-29332545-d245p\" (UID: \"c88f1935-f363-4596-9b74-fe8b3ee2946a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29332545-d245p" Oct 08 19:45:00 crc kubenswrapper[4988]: I1008 19:45:00.465370 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/c88f1935-f363-4596-9b74-fe8b3ee2946a-secret-volume\") pod \"collect-profiles-29332545-d245p\" (UID: \"c88f1935-f363-4596-9b74-fe8b3ee2946a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29332545-d245p" Oct 08 19:45:00 crc kubenswrapper[4988]: I1008 19:45:00.482814 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ffkld\" (UniqueName: \"kubernetes.io/projected/c88f1935-f363-4596-9b74-fe8b3ee2946a-kube-api-access-ffkld\") pod \"collect-profiles-29332545-d245p\" (UID: \"c88f1935-f363-4596-9b74-fe8b3ee2946a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29332545-d245p" Oct 08 19:45:00 crc kubenswrapper[4988]: I1008 19:45:00.496322 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29332545-d245p" Oct 08 19:45:00 crc kubenswrapper[4988]: I1008 19:45:00.964431 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29332545-d245p"] Oct 08 19:45:00 crc kubenswrapper[4988]: W1008 19:45:00.964880 4988 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc88f1935_f363_4596_9b74_fe8b3ee2946a.slice/crio-84f377c88691aa615d302f5a458aa283128b78c6ff243adf829ae0263005e62e WatchSource:0}: Error finding container 84f377c88691aa615d302f5a458aa283128b78c6ff243adf829ae0263005e62e: Status 404 returned error can't find the container with id 84f377c88691aa615d302f5a458aa283128b78c6ff243adf829ae0263005e62e Oct 08 19:45:01 crc kubenswrapper[4988]: I1008 19:45:01.649403 4988 generic.go:334] "Generic (PLEG): container finished" podID="c88f1935-f363-4596-9b74-fe8b3ee2946a" containerID="0e7dce3df7eb9938242996e6b4d550500a33eb43c94bed56498f93c0669e77f1" exitCode=0 Oct 08 19:45:01 crc kubenswrapper[4988]: I1008 19:45:01.649471 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29332545-d245p" event={"ID":"c88f1935-f363-4596-9b74-fe8b3ee2946a","Type":"ContainerDied","Data":"0e7dce3df7eb9938242996e6b4d550500a33eb43c94bed56498f93c0669e77f1"} Oct 08 19:45:01 crc kubenswrapper[4988]: I1008 19:45:01.649762 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29332545-d245p" event={"ID":"c88f1935-f363-4596-9b74-fe8b3ee2946a","Type":"ContainerStarted","Data":"84f377c88691aa615d302f5a458aa283128b78c6ff243adf829ae0263005e62e"} Oct 08 19:45:03 crc kubenswrapper[4988]: I1008 19:45:03.064923 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29332545-d245p" Oct 08 19:45:03 crc kubenswrapper[4988]: I1008 19:45:03.214051 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/c88f1935-f363-4596-9b74-fe8b3ee2946a-secret-volume\") pod \"c88f1935-f363-4596-9b74-fe8b3ee2946a\" (UID: \"c88f1935-f363-4596-9b74-fe8b3ee2946a\") " Oct 08 19:45:03 crc kubenswrapper[4988]: I1008 19:45:03.214222 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ffkld\" (UniqueName: \"kubernetes.io/projected/c88f1935-f363-4596-9b74-fe8b3ee2946a-kube-api-access-ffkld\") pod \"c88f1935-f363-4596-9b74-fe8b3ee2946a\" (UID: \"c88f1935-f363-4596-9b74-fe8b3ee2946a\") " Oct 08 19:45:03 crc kubenswrapper[4988]: I1008 19:45:03.214378 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c88f1935-f363-4596-9b74-fe8b3ee2946a-config-volume\") pod \"c88f1935-f363-4596-9b74-fe8b3ee2946a\" (UID: \"c88f1935-f363-4596-9b74-fe8b3ee2946a\") " Oct 08 19:45:03 crc kubenswrapper[4988]: I1008 19:45:03.216003 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c88f1935-f363-4596-9b74-fe8b3ee2946a-config-volume" (OuterVolumeSpecName: "config-volume") pod "c88f1935-f363-4596-9b74-fe8b3ee2946a" (UID: "c88f1935-f363-4596-9b74-fe8b3ee2946a"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 19:45:03 crc kubenswrapper[4988]: I1008 19:45:03.224462 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c88f1935-f363-4596-9b74-fe8b3ee2946a-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "c88f1935-f363-4596-9b74-fe8b3ee2946a" (UID: "c88f1935-f363-4596-9b74-fe8b3ee2946a"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 19:45:03 crc kubenswrapper[4988]: I1008 19:45:03.230684 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c88f1935-f363-4596-9b74-fe8b3ee2946a-kube-api-access-ffkld" (OuterVolumeSpecName: "kube-api-access-ffkld") pod "c88f1935-f363-4596-9b74-fe8b3ee2946a" (UID: "c88f1935-f363-4596-9b74-fe8b3ee2946a"). InnerVolumeSpecName "kube-api-access-ffkld". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 19:45:03 crc kubenswrapper[4988]: I1008 19:45:03.316679 4988 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ffkld\" (UniqueName: \"kubernetes.io/projected/c88f1935-f363-4596-9b74-fe8b3ee2946a-kube-api-access-ffkld\") on node \"crc\" DevicePath \"\"" Oct 08 19:45:03 crc kubenswrapper[4988]: I1008 19:45:03.316714 4988 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c88f1935-f363-4596-9b74-fe8b3ee2946a-config-volume\") on node \"crc\" DevicePath \"\"" Oct 08 19:45:03 crc kubenswrapper[4988]: I1008 19:45:03.316725 4988 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/c88f1935-f363-4596-9b74-fe8b3ee2946a-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 08 19:45:03 crc kubenswrapper[4988]: I1008 19:45:03.669867 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29332545-d245p" event={"ID":"c88f1935-f363-4596-9b74-fe8b3ee2946a","Type":"ContainerDied","Data":"84f377c88691aa615d302f5a458aa283128b78c6ff243adf829ae0263005e62e"} Oct 08 19:45:03 crc kubenswrapper[4988]: I1008 19:45:03.670205 4988 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="84f377c88691aa615d302f5a458aa283128b78c6ff243adf829ae0263005e62e" Oct 08 19:45:03 crc kubenswrapper[4988]: I1008 19:45:03.669920 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29332545-d245p" Oct 08 19:45:04 crc kubenswrapper[4988]: I1008 19:45:04.200586 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29332500-85wnq"] Oct 08 19:45:04 crc kubenswrapper[4988]: I1008 19:45:04.208343 4988 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29332500-85wnq"] Oct 08 19:45:05 crc kubenswrapper[4988]: I1008 19:45:05.256703 4988 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8a70d02b-f876-4063-9dbc-a7159e12d8d8" path="/var/lib/kubelet/pods/8a70d02b-f876-4063-9dbc-a7159e12d8d8/volumes" Oct 08 19:45:12 crc kubenswrapper[4988]: I1008 19:45:12.269707 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-c458d797b-jlpvk" Oct 08 19:45:14 crc kubenswrapper[4988]: I1008 19:45:14.785806 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-77f8f8964c-wclf6" Oct 08 19:45:14 crc kubenswrapper[4988]: I1008 19:45:14.883008 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-c458d797b-jlpvk"] Oct 08 19:45:14 crc kubenswrapper[4988]: I1008 19:45:14.883263 4988 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-c458d797b-jlpvk" podUID="8635e509-34ea-43b5-955d-db4677f28ad9" containerName="neutron-api" containerID="cri-o://f9f6c9c6d01553f2e40f5d7f951b43c8679e350a27241c60aef27549dd34bab8" gracePeriod=30 Oct 08 19:45:14 crc kubenswrapper[4988]: I1008 19:45:14.883656 4988 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-c458d797b-jlpvk" podUID="8635e509-34ea-43b5-955d-db4677f28ad9" containerName="neutron-httpd" containerID="cri-o://d7d1cd0efaa71a26868202a003120cbac6962702516001eb5a915193a7148311" gracePeriod=30 Oct 08 19:45:15 crc kubenswrapper[4988]: I1008 19:45:15.796962 4988 generic.go:334] "Generic (PLEG): container finished" podID="8635e509-34ea-43b5-955d-db4677f28ad9" containerID="d7d1cd0efaa71a26868202a003120cbac6962702516001eb5a915193a7148311" exitCode=0 Oct 08 19:45:15 crc kubenswrapper[4988]: I1008 19:45:15.797297 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-c458d797b-jlpvk" event={"ID":"8635e509-34ea-43b5-955d-db4677f28ad9","Type":"ContainerDied","Data":"d7d1cd0efaa71a26868202a003120cbac6962702516001eb5a915193a7148311"} Oct 08 19:45:18 crc kubenswrapper[4988]: I1008 19:45:18.816790 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-c458d797b-jlpvk" Oct 08 19:45:18 crc kubenswrapper[4988]: I1008 19:45:18.824236 4988 generic.go:334] "Generic (PLEG): container finished" podID="8635e509-34ea-43b5-955d-db4677f28ad9" containerID="f9f6c9c6d01553f2e40f5d7f951b43c8679e350a27241c60aef27549dd34bab8" exitCode=0 Oct 08 19:45:18 crc kubenswrapper[4988]: I1008 19:45:18.824282 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-c458d797b-jlpvk" event={"ID":"8635e509-34ea-43b5-955d-db4677f28ad9","Type":"ContainerDied","Data":"f9f6c9c6d01553f2e40f5d7f951b43c8679e350a27241c60aef27549dd34bab8"} Oct 08 19:45:18 crc kubenswrapper[4988]: I1008 19:45:18.824310 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-c458d797b-jlpvk" event={"ID":"8635e509-34ea-43b5-955d-db4677f28ad9","Type":"ContainerDied","Data":"9a7bbfdd04db51f43746cacf553ca575ba49de3375bf5621c0d7e77afbab0d2b"} Oct 08 19:45:18 crc kubenswrapper[4988]: I1008 19:45:18.824326 4988 scope.go:117] "RemoveContainer" containerID="d7d1cd0efaa71a26868202a003120cbac6962702516001eb5a915193a7148311" Oct 08 19:45:18 crc kubenswrapper[4988]: I1008 19:45:18.824468 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-c458d797b-jlpvk" Oct 08 19:45:18 crc kubenswrapper[4988]: I1008 19:45:18.935095 4988 scope.go:117] "RemoveContainer" containerID="f9f6c9c6d01553f2e40f5d7f951b43c8679e350a27241c60aef27549dd34bab8" Oct 08 19:45:18 crc kubenswrapper[4988]: I1008 19:45:18.947934 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dsj24\" (UniqueName: \"kubernetes.io/projected/8635e509-34ea-43b5-955d-db4677f28ad9-kube-api-access-dsj24\") pod \"8635e509-34ea-43b5-955d-db4677f28ad9\" (UID: \"8635e509-34ea-43b5-955d-db4677f28ad9\") " Oct 08 19:45:18 crc kubenswrapper[4988]: I1008 19:45:18.947974 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/8635e509-34ea-43b5-955d-db4677f28ad9-httpd-config\") pod \"8635e509-34ea-43b5-955d-db4677f28ad9\" (UID: \"8635e509-34ea-43b5-955d-db4677f28ad9\") " Oct 08 19:45:18 crc kubenswrapper[4988]: I1008 19:45:18.948113 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8635e509-34ea-43b5-955d-db4677f28ad9-combined-ca-bundle\") pod \"8635e509-34ea-43b5-955d-db4677f28ad9\" (UID: \"8635e509-34ea-43b5-955d-db4677f28ad9\") " Oct 08 19:45:18 crc kubenswrapper[4988]: I1008 19:45:18.948149 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/8635e509-34ea-43b5-955d-db4677f28ad9-ovndb-tls-certs\") pod \"8635e509-34ea-43b5-955d-db4677f28ad9\" (UID: \"8635e509-34ea-43b5-955d-db4677f28ad9\") " Oct 08 19:45:18 crc kubenswrapper[4988]: I1008 19:45:18.948173 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/8635e509-34ea-43b5-955d-db4677f28ad9-config\") pod \"8635e509-34ea-43b5-955d-db4677f28ad9\" (UID: \"8635e509-34ea-43b5-955d-db4677f28ad9\") " Oct 08 19:45:18 crc kubenswrapper[4988]: I1008 19:45:18.964297 4988 scope.go:117] "RemoveContainer" containerID="d7d1cd0efaa71a26868202a003120cbac6962702516001eb5a915193a7148311" Oct 08 19:45:18 crc kubenswrapper[4988]: I1008 19:45:18.969018 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8635e509-34ea-43b5-955d-db4677f28ad9-httpd-config" (OuterVolumeSpecName: "httpd-config") pod "8635e509-34ea-43b5-955d-db4677f28ad9" (UID: "8635e509-34ea-43b5-955d-db4677f28ad9"). InnerVolumeSpecName "httpd-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 19:45:18 crc kubenswrapper[4988]: I1008 19:45:18.969114 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8635e509-34ea-43b5-955d-db4677f28ad9-kube-api-access-dsj24" (OuterVolumeSpecName: "kube-api-access-dsj24") pod "8635e509-34ea-43b5-955d-db4677f28ad9" (UID: "8635e509-34ea-43b5-955d-db4677f28ad9"). InnerVolumeSpecName "kube-api-access-dsj24". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 19:45:18 crc kubenswrapper[4988]: E1008 19:45:18.969813 4988 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d7d1cd0efaa71a26868202a003120cbac6962702516001eb5a915193a7148311\": container with ID starting with d7d1cd0efaa71a26868202a003120cbac6962702516001eb5a915193a7148311 not found: ID does not exist" containerID="d7d1cd0efaa71a26868202a003120cbac6962702516001eb5a915193a7148311" Oct 08 19:45:18 crc kubenswrapper[4988]: I1008 19:45:18.969875 4988 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d7d1cd0efaa71a26868202a003120cbac6962702516001eb5a915193a7148311"} err="failed to get container status \"d7d1cd0efaa71a26868202a003120cbac6962702516001eb5a915193a7148311\": rpc error: code = NotFound desc = could not find container \"d7d1cd0efaa71a26868202a003120cbac6962702516001eb5a915193a7148311\": container with ID starting with d7d1cd0efaa71a26868202a003120cbac6962702516001eb5a915193a7148311 not found: ID does not exist" Oct 08 19:45:18 crc kubenswrapper[4988]: I1008 19:45:18.969911 4988 scope.go:117] "RemoveContainer" containerID="f9f6c9c6d01553f2e40f5d7f951b43c8679e350a27241c60aef27549dd34bab8" Oct 08 19:45:18 crc kubenswrapper[4988]: E1008 19:45:18.970365 4988 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f9f6c9c6d01553f2e40f5d7f951b43c8679e350a27241c60aef27549dd34bab8\": container with ID starting with f9f6c9c6d01553f2e40f5d7f951b43c8679e350a27241c60aef27549dd34bab8 not found: ID does not exist" containerID="f9f6c9c6d01553f2e40f5d7f951b43c8679e350a27241c60aef27549dd34bab8" Oct 08 19:45:18 crc kubenswrapper[4988]: I1008 19:45:18.970408 4988 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f9f6c9c6d01553f2e40f5d7f951b43c8679e350a27241c60aef27549dd34bab8"} err="failed to get container status \"f9f6c9c6d01553f2e40f5d7f951b43c8679e350a27241c60aef27549dd34bab8\": rpc error: code = NotFound desc = could not find container \"f9f6c9c6d01553f2e40f5d7f951b43c8679e350a27241c60aef27549dd34bab8\": container with ID starting with f9f6c9c6d01553f2e40f5d7f951b43c8679e350a27241c60aef27549dd34bab8 not found: ID does not exist" Oct 08 19:45:19 crc kubenswrapper[4988]: I1008 19:45:19.001456 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8635e509-34ea-43b5-955d-db4677f28ad9-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "8635e509-34ea-43b5-955d-db4677f28ad9" (UID: "8635e509-34ea-43b5-955d-db4677f28ad9"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 19:45:19 crc kubenswrapper[4988]: I1008 19:45:19.028120 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8635e509-34ea-43b5-955d-db4677f28ad9-ovndb-tls-certs" (OuterVolumeSpecName: "ovndb-tls-certs") pod "8635e509-34ea-43b5-955d-db4677f28ad9" (UID: "8635e509-34ea-43b5-955d-db4677f28ad9"). InnerVolumeSpecName "ovndb-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 19:45:19 crc kubenswrapper[4988]: I1008 19:45:19.032932 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8635e509-34ea-43b5-955d-db4677f28ad9-config" (OuterVolumeSpecName: "config") pod "8635e509-34ea-43b5-955d-db4677f28ad9" (UID: "8635e509-34ea-43b5-955d-db4677f28ad9"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 19:45:19 crc kubenswrapper[4988]: I1008 19:45:19.051691 4988 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8635e509-34ea-43b5-955d-db4677f28ad9-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 19:45:19 crc kubenswrapper[4988]: I1008 19:45:19.051728 4988 reconciler_common.go:293] "Volume detached for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/8635e509-34ea-43b5-955d-db4677f28ad9-ovndb-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 08 19:45:19 crc kubenswrapper[4988]: I1008 19:45:19.051739 4988 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/8635e509-34ea-43b5-955d-db4677f28ad9-config\") on node \"crc\" DevicePath \"\"" Oct 08 19:45:19 crc kubenswrapper[4988]: I1008 19:45:19.051750 4988 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dsj24\" (UniqueName: \"kubernetes.io/projected/8635e509-34ea-43b5-955d-db4677f28ad9-kube-api-access-dsj24\") on node \"crc\" DevicePath \"\"" Oct 08 19:45:19 crc kubenswrapper[4988]: I1008 19:45:19.051762 4988 reconciler_common.go:293] "Volume detached for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/8635e509-34ea-43b5-955d-db4677f28ad9-httpd-config\") on node \"crc\" DevicePath \"\"" Oct 08 19:45:19 crc kubenswrapper[4988]: I1008 19:45:19.156749 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-c458d797b-jlpvk"] Oct 08 19:45:19 crc kubenswrapper[4988]: I1008 19:45:19.165640 4988 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-c458d797b-jlpvk"] Oct 08 19:45:19 crc kubenswrapper[4988]: I1008 19:45:19.255147 4988 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8635e509-34ea-43b5-955d-db4677f28ad9" path="/var/lib/kubelet/pods/8635e509-34ea-43b5-955d-db4677f28ad9/volumes" Oct 08 19:45:23 crc kubenswrapper[4988]: I1008 19:45:23.337775 4988 patch_prober.go:28] interesting pod/machine-config-daemon-nm54f container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 19:45:23 crc kubenswrapper[4988]: I1008 19:45:23.338537 4988 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 19:45:25 crc kubenswrapper[4988]: I1008 19:45:25.009310 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-ring-rebalance-t9tqb"] Oct 08 19:45:25 crc kubenswrapper[4988]: E1008 19:45:25.010169 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8635e509-34ea-43b5-955d-db4677f28ad9" containerName="neutron-httpd" Oct 08 19:45:25 crc kubenswrapper[4988]: I1008 19:45:25.010185 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="8635e509-34ea-43b5-955d-db4677f28ad9" containerName="neutron-httpd" Oct 08 19:45:25 crc kubenswrapper[4988]: E1008 19:45:25.010213 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c88f1935-f363-4596-9b74-fe8b3ee2946a" containerName="collect-profiles" Oct 08 19:45:25 crc kubenswrapper[4988]: I1008 19:45:25.010223 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="c88f1935-f363-4596-9b74-fe8b3ee2946a" containerName="collect-profiles" Oct 08 19:45:25 crc kubenswrapper[4988]: E1008 19:45:25.010241 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8635e509-34ea-43b5-955d-db4677f28ad9" containerName="neutron-api" Oct 08 19:45:25 crc kubenswrapper[4988]: I1008 19:45:25.010249 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="8635e509-34ea-43b5-955d-db4677f28ad9" containerName="neutron-api" Oct 08 19:45:25 crc kubenswrapper[4988]: I1008 19:45:25.010493 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="8635e509-34ea-43b5-955d-db4677f28ad9" containerName="neutron-httpd" Oct 08 19:45:25 crc kubenswrapper[4988]: I1008 19:45:25.010539 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="c88f1935-f363-4596-9b74-fe8b3ee2946a" containerName="collect-profiles" Oct 08 19:45:25 crc kubenswrapper[4988]: I1008 19:45:25.010551 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="8635e509-34ea-43b5-955d-db4677f28ad9" containerName="neutron-api" Oct 08 19:45:25 crc kubenswrapper[4988]: I1008 19:45:25.011261 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-t9tqb" Oct 08 19:45:25 crc kubenswrapper[4988]: I1008 19:45:25.018161 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-proxy-config-data" Oct 08 19:45:25 crc kubenswrapper[4988]: I1008 19:45:25.021985 4988 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-scripts" Oct 08 19:45:25 crc kubenswrapper[4988]: I1008 19:45:25.022165 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-conf" Oct 08 19:45:25 crc kubenswrapper[4988]: I1008 19:45:25.022275 4988 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-config-data" Oct 08 19:45:25 crc kubenswrapper[4988]: I1008 19:45:25.022426 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-swift-dockercfg-xhcps" Oct 08 19:45:25 crc kubenswrapper[4988]: I1008 19:45:25.041532 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-t9tqb"] Oct 08 19:45:25 crc kubenswrapper[4988]: I1008 19:45:25.067208 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-6d64fc8cdc-jb7cp"] Oct 08 19:45:25 crc kubenswrapper[4988]: I1008 19:45:25.074779 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6d64fc8cdc-jb7cp" Oct 08 19:45:25 crc kubenswrapper[4988]: I1008 19:45:25.093252 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6d64fc8cdc-jb7cp"] Oct 08 19:45:25 crc kubenswrapper[4988]: I1008 19:45:25.175718 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5bd22f4a-f899-4db4-88a1-10344841593b-ovsdbserver-sb\") pod \"dnsmasq-dns-6d64fc8cdc-jb7cp\" (UID: \"5bd22f4a-f899-4db4-88a1-10344841593b\") " pod="openstack/dnsmasq-dns-6d64fc8cdc-jb7cp" Oct 08 19:45:25 crc kubenswrapper[4988]: I1008 19:45:25.175773 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5bd22f4a-f899-4db4-88a1-10344841593b-config\") pod \"dnsmasq-dns-6d64fc8cdc-jb7cp\" (UID: \"5bd22f4a-f899-4db4-88a1-10344841593b\") " pod="openstack/dnsmasq-dns-6d64fc8cdc-jb7cp" Oct 08 19:45:25 crc kubenswrapper[4988]: I1008 19:45:25.175804 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/de3434fc-c4fa-4495-b3f3-7f232decfed9-dispersionconf\") pod \"swift-ring-rebalance-t9tqb\" (UID: \"de3434fc-c4fa-4495-b3f3-7f232decfed9\") " pod="openstack/swift-ring-rebalance-t9tqb" Oct 08 19:45:25 crc kubenswrapper[4988]: I1008 19:45:25.175875 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/de3434fc-c4fa-4495-b3f3-7f232decfed9-combined-ca-bundle\") pod \"swift-ring-rebalance-t9tqb\" (UID: \"de3434fc-c4fa-4495-b3f3-7f232decfed9\") " pod="openstack/swift-ring-rebalance-t9tqb" Oct 08 19:45:25 crc kubenswrapper[4988]: I1008 19:45:25.175901 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/de3434fc-c4fa-4495-b3f3-7f232decfed9-ring-data-devices\") pod \"swift-ring-rebalance-t9tqb\" (UID: \"de3434fc-c4fa-4495-b3f3-7f232decfed9\") " pod="openstack/swift-ring-rebalance-t9tqb" Oct 08 19:45:25 crc kubenswrapper[4988]: I1008 19:45:25.175925 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/de3434fc-c4fa-4495-b3f3-7f232decfed9-swiftconf\") pod \"swift-ring-rebalance-t9tqb\" (UID: \"de3434fc-c4fa-4495-b3f3-7f232decfed9\") " pod="openstack/swift-ring-rebalance-t9tqb" Oct 08 19:45:25 crc kubenswrapper[4988]: I1008 19:45:25.175956 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qwvjx\" (UniqueName: \"kubernetes.io/projected/5bd22f4a-f899-4db4-88a1-10344841593b-kube-api-access-qwvjx\") pod \"dnsmasq-dns-6d64fc8cdc-jb7cp\" (UID: \"5bd22f4a-f899-4db4-88a1-10344841593b\") " pod="openstack/dnsmasq-dns-6d64fc8cdc-jb7cp" Oct 08 19:45:25 crc kubenswrapper[4988]: I1008 19:45:25.175982 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k5f6q\" (UniqueName: \"kubernetes.io/projected/de3434fc-c4fa-4495-b3f3-7f232decfed9-kube-api-access-k5f6q\") pod \"swift-ring-rebalance-t9tqb\" (UID: \"de3434fc-c4fa-4495-b3f3-7f232decfed9\") " pod="openstack/swift-ring-rebalance-t9tqb" Oct 08 19:45:25 crc kubenswrapper[4988]: I1008 19:45:25.176020 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5bd22f4a-f899-4db4-88a1-10344841593b-dns-svc\") pod \"dnsmasq-dns-6d64fc8cdc-jb7cp\" (UID: \"5bd22f4a-f899-4db4-88a1-10344841593b\") " pod="openstack/dnsmasq-dns-6d64fc8cdc-jb7cp" Oct 08 19:45:25 crc kubenswrapper[4988]: I1008 19:45:25.176039 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5bd22f4a-f899-4db4-88a1-10344841593b-ovsdbserver-nb\") pod \"dnsmasq-dns-6d64fc8cdc-jb7cp\" (UID: \"5bd22f4a-f899-4db4-88a1-10344841593b\") " pod="openstack/dnsmasq-dns-6d64fc8cdc-jb7cp" Oct 08 19:45:25 crc kubenswrapper[4988]: I1008 19:45:25.176096 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/de3434fc-c4fa-4495-b3f3-7f232decfed9-scripts\") pod \"swift-ring-rebalance-t9tqb\" (UID: \"de3434fc-c4fa-4495-b3f3-7f232decfed9\") " pod="openstack/swift-ring-rebalance-t9tqb" Oct 08 19:45:25 crc kubenswrapper[4988]: I1008 19:45:25.176136 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/de3434fc-c4fa-4495-b3f3-7f232decfed9-etc-swift\") pod \"swift-ring-rebalance-t9tqb\" (UID: \"de3434fc-c4fa-4495-b3f3-7f232decfed9\") " pod="openstack/swift-ring-rebalance-t9tqb" Oct 08 19:45:25 crc kubenswrapper[4988]: I1008 19:45:25.277854 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/de3434fc-c4fa-4495-b3f3-7f232decfed9-combined-ca-bundle\") pod \"swift-ring-rebalance-t9tqb\" (UID: \"de3434fc-c4fa-4495-b3f3-7f232decfed9\") " pod="openstack/swift-ring-rebalance-t9tqb" Oct 08 19:45:25 crc kubenswrapper[4988]: I1008 19:45:25.277907 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/de3434fc-c4fa-4495-b3f3-7f232decfed9-ring-data-devices\") pod \"swift-ring-rebalance-t9tqb\" (UID: \"de3434fc-c4fa-4495-b3f3-7f232decfed9\") " pod="openstack/swift-ring-rebalance-t9tqb" Oct 08 19:45:25 crc kubenswrapper[4988]: I1008 19:45:25.277934 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/de3434fc-c4fa-4495-b3f3-7f232decfed9-swiftconf\") pod \"swift-ring-rebalance-t9tqb\" (UID: \"de3434fc-c4fa-4495-b3f3-7f232decfed9\") " pod="openstack/swift-ring-rebalance-t9tqb" Oct 08 19:45:25 crc kubenswrapper[4988]: I1008 19:45:25.277969 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qwvjx\" (UniqueName: \"kubernetes.io/projected/5bd22f4a-f899-4db4-88a1-10344841593b-kube-api-access-qwvjx\") pod \"dnsmasq-dns-6d64fc8cdc-jb7cp\" (UID: \"5bd22f4a-f899-4db4-88a1-10344841593b\") " pod="openstack/dnsmasq-dns-6d64fc8cdc-jb7cp" Oct 08 19:45:25 crc kubenswrapper[4988]: I1008 19:45:25.277993 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k5f6q\" (UniqueName: \"kubernetes.io/projected/de3434fc-c4fa-4495-b3f3-7f232decfed9-kube-api-access-k5f6q\") pod \"swift-ring-rebalance-t9tqb\" (UID: \"de3434fc-c4fa-4495-b3f3-7f232decfed9\") " pod="openstack/swift-ring-rebalance-t9tqb" Oct 08 19:45:25 crc kubenswrapper[4988]: I1008 19:45:25.278029 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5bd22f4a-f899-4db4-88a1-10344841593b-dns-svc\") pod \"dnsmasq-dns-6d64fc8cdc-jb7cp\" (UID: \"5bd22f4a-f899-4db4-88a1-10344841593b\") " pod="openstack/dnsmasq-dns-6d64fc8cdc-jb7cp" Oct 08 19:45:25 crc kubenswrapper[4988]: I1008 19:45:25.278053 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5bd22f4a-f899-4db4-88a1-10344841593b-ovsdbserver-nb\") pod \"dnsmasq-dns-6d64fc8cdc-jb7cp\" (UID: \"5bd22f4a-f899-4db4-88a1-10344841593b\") " pod="openstack/dnsmasq-dns-6d64fc8cdc-jb7cp" Oct 08 19:45:25 crc kubenswrapper[4988]: I1008 19:45:25.278106 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/de3434fc-c4fa-4495-b3f3-7f232decfed9-scripts\") pod \"swift-ring-rebalance-t9tqb\" (UID: \"de3434fc-c4fa-4495-b3f3-7f232decfed9\") " pod="openstack/swift-ring-rebalance-t9tqb" Oct 08 19:45:25 crc kubenswrapper[4988]: I1008 19:45:25.278143 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/de3434fc-c4fa-4495-b3f3-7f232decfed9-etc-swift\") pod \"swift-ring-rebalance-t9tqb\" (UID: \"de3434fc-c4fa-4495-b3f3-7f232decfed9\") " pod="openstack/swift-ring-rebalance-t9tqb" Oct 08 19:45:25 crc kubenswrapper[4988]: I1008 19:45:25.278177 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5bd22f4a-f899-4db4-88a1-10344841593b-ovsdbserver-sb\") pod \"dnsmasq-dns-6d64fc8cdc-jb7cp\" (UID: \"5bd22f4a-f899-4db4-88a1-10344841593b\") " pod="openstack/dnsmasq-dns-6d64fc8cdc-jb7cp" Oct 08 19:45:25 crc kubenswrapper[4988]: I1008 19:45:25.278207 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5bd22f4a-f899-4db4-88a1-10344841593b-config\") pod \"dnsmasq-dns-6d64fc8cdc-jb7cp\" (UID: \"5bd22f4a-f899-4db4-88a1-10344841593b\") " pod="openstack/dnsmasq-dns-6d64fc8cdc-jb7cp" Oct 08 19:45:25 crc kubenswrapper[4988]: I1008 19:45:25.278229 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/de3434fc-c4fa-4495-b3f3-7f232decfed9-dispersionconf\") pod \"swift-ring-rebalance-t9tqb\" (UID: \"de3434fc-c4fa-4495-b3f3-7f232decfed9\") " pod="openstack/swift-ring-rebalance-t9tqb" Oct 08 19:45:25 crc kubenswrapper[4988]: I1008 19:45:25.279608 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/de3434fc-c4fa-4495-b3f3-7f232decfed9-etc-swift\") pod \"swift-ring-rebalance-t9tqb\" (UID: \"de3434fc-c4fa-4495-b3f3-7f232decfed9\") " pod="openstack/swift-ring-rebalance-t9tqb" Oct 08 19:45:25 crc kubenswrapper[4988]: I1008 19:45:25.280308 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/de3434fc-c4fa-4495-b3f3-7f232decfed9-ring-data-devices\") pod \"swift-ring-rebalance-t9tqb\" (UID: \"de3434fc-c4fa-4495-b3f3-7f232decfed9\") " pod="openstack/swift-ring-rebalance-t9tqb" Oct 08 19:45:25 crc kubenswrapper[4988]: I1008 19:45:25.280365 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/de3434fc-c4fa-4495-b3f3-7f232decfed9-scripts\") pod \"swift-ring-rebalance-t9tqb\" (UID: \"de3434fc-c4fa-4495-b3f3-7f232decfed9\") " pod="openstack/swift-ring-rebalance-t9tqb" Oct 08 19:45:25 crc kubenswrapper[4988]: I1008 19:45:25.280629 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5bd22f4a-f899-4db4-88a1-10344841593b-dns-svc\") pod \"dnsmasq-dns-6d64fc8cdc-jb7cp\" (UID: \"5bd22f4a-f899-4db4-88a1-10344841593b\") " pod="openstack/dnsmasq-dns-6d64fc8cdc-jb7cp" Oct 08 19:45:25 crc kubenswrapper[4988]: I1008 19:45:25.280734 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5bd22f4a-f899-4db4-88a1-10344841593b-ovsdbserver-sb\") pod \"dnsmasq-dns-6d64fc8cdc-jb7cp\" (UID: \"5bd22f4a-f899-4db4-88a1-10344841593b\") " pod="openstack/dnsmasq-dns-6d64fc8cdc-jb7cp" Oct 08 19:45:25 crc kubenswrapper[4988]: I1008 19:45:25.280765 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5bd22f4a-f899-4db4-88a1-10344841593b-ovsdbserver-nb\") pod \"dnsmasq-dns-6d64fc8cdc-jb7cp\" (UID: \"5bd22f4a-f899-4db4-88a1-10344841593b\") " pod="openstack/dnsmasq-dns-6d64fc8cdc-jb7cp" Oct 08 19:45:25 crc kubenswrapper[4988]: I1008 19:45:25.280878 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5bd22f4a-f899-4db4-88a1-10344841593b-config\") pod \"dnsmasq-dns-6d64fc8cdc-jb7cp\" (UID: \"5bd22f4a-f899-4db4-88a1-10344841593b\") " pod="openstack/dnsmasq-dns-6d64fc8cdc-jb7cp" Oct 08 19:45:25 crc kubenswrapper[4988]: I1008 19:45:25.297294 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/de3434fc-c4fa-4495-b3f3-7f232decfed9-dispersionconf\") pod \"swift-ring-rebalance-t9tqb\" (UID: \"de3434fc-c4fa-4495-b3f3-7f232decfed9\") " pod="openstack/swift-ring-rebalance-t9tqb" Oct 08 19:45:25 crc kubenswrapper[4988]: I1008 19:45:25.297712 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/de3434fc-c4fa-4495-b3f3-7f232decfed9-combined-ca-bundle\") pod \"swift-ring-rebalance-t9tqb\" (UID: \"de3434fc-c4fa-4495-b3f3-7f232decfed9\") " pod="openstack/swift-ring-rebalance-t9tqb" Oct 08 19:45:25 crc kubenswrapper[4988]: I1008 19:45:25.301972 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qwvjx\" (UniqueName: \"kubernetes.io/projected/5bd22f4a-f899-4db4-88a1-10344841593b-kube-api-access-qwvjx\") pod \"dnsmasq-dns-6d64fc8cdc-jb7cp\" (UID: \"5bd22f4a-f899-4db4-88a1-10344841593b\") " pod="openstack/dnsmasq-dns-6d64fc8cdc-jb7cp" Oct 08 19:45:25 crc kubenswrapper[4988]: I1008 19:45:25.306086 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k5f6q\" (UniqueName: \"kubernetes.io/projected/de3434fc-c4fa-4495-b3f3-7f232decfed9-kube-api-access-k5f6q\") pod \"swift-ring-rebalance-t9tqb\" (UID: \"de3434fc-c4fa-4495-b3f3-7f232decfed9\") " pod="openstack/swift-ring-rebalance-t9tqb" Oct 08 19:45:25 crc kubenswrapper[4988]: I1008 19:45:25.306520 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/de3434fc-c4fa-4495-b3f3-7f232decfed9-swiftconf\") pod \"swift-ring-rebalance-t9tqb\" (UID: \"de3434fc-c4fa-4495-b3f3-7f232decfed9\") " pod="openstack/swift-ring-rebalance-t9tqb" Oct 08 19:45:25 crc kubenswrapper[4988]: I1008 19:45:25.336080 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-t9tqb" Oct 08 19:45:25 crc kubenswrapper[4988]: I1008 19:45:25.411380 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6d64fc8cdc-jb7cp" Oct 08 19:45:25 crc kubenswrapper[4988]: I1008 19:45:25.781166 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-t9tqb"] Oct 08 19:45:25 crc kubenswrapper[4988]: W1008 19:45:25.915992 4988 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5bd22f4a_f899_4db4_88a1_10344841593b.slice/crio-70ceb3dbac6d506652ffc91df3138eb6f9b4185c457725e9cddfba4a514c3e2a WatchSource:0}: Error finding container 70ceb3dbac6d506652ffc91df3138eb6f9b4185c457725e9cddfba4a514c3e2a: Status 404 returned error can't find the container with id 70ceb3dbac6d506652ffc91df3138eb6f9b4185c457725e9cddfba4a514c3e2a Oct 08 19:45:25 crc kubenswrapper[4988]: I1008 19:45:25.917966 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-t9tqb" event={"ID":"de3434fc-c4fa-4495-b3f3-7f232decfed9","Type":"ContainerStarted","Data":"ce0527c232d6cb4d95475048b0ea1fd748afdfa0f4a5ff109654451872209643"} Oct 08 19:45:25 crc kubenswrapper[4988]: I1008 19:45:25.919620 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6d64fc8cdc-jb7cp"] Oct 08 19:45:26 crc kubenswrapper[4988]: I1008 19:45:26.880349 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-proxy-5dfffdbbd6-gzqmd"] Oct 08 19:45:26 crc kubenswrapper[4988]: I1008 19:45:26.881867 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-5dfffdbbd6-gzqmd" Oct 08 19:45:26 crc kubenswrapper[4988]: I1008 19:45:26.883809 4988 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-files" Oct 08 19:45:26 crc kubenswrapper[4988]: I1008 19:45:26.896699 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-proxy-5dfffdbbd6-gzqmd"] Oct 08 19:45:26 crc kubenswrapper[4988]: I1008 19:45:26.940459 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-t9tqb" event={"ID":"de3434fc-c4fa-4495-b3f3-7f232decfed9","Type":"ContainerStarted","Data":"c886858da5e42884b9e8ec85277700eab6cb46f9bd3d7f965f102a724c3e7609"} Oct 08 19:45:26 crc kubenswrapper[4988]: I1008 19:45:26.942205 4988 generic.go:334] "Generic (PLEG): container finished" podID="5bd22f4a-f899-4db4-88a1-10344841593b" containerID="2570a1e23d022c1995fec65e70d8fc7ff21575a1cd4aaea4504a2d61ab9440c1" exitCode=0 Oct 08 19:45:26 crc kubenswrapper[4988]: I1008 19:45:26.942262 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6d64fc8cdc-jb7cp" event={"ID":"5bd22f4a-f899-4db4-88a1-10344841593b","Type":"ContainerDied","Data":"2570a1e23d022c1995fec65e70d8fc7ff21575a1cd4aaea4504a2d61ab9440c1"} Oct 08 19:45:26 crc kubenswrapper[4988]: I1008 19:45:26.942298 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6d64fc8cdc-jb7cp" event={"ID":"5bd22f4a-f899-4db4-88a1-10344841593b","Type":"ContainerStarted","Data":"70ceb3dbac6d506652ffc91df3138eb6f9b4185c457725e9cddfba4a514c3e2a"} Oct 08 19:45:26 crc kubenswrapper[4988]: I1008 19:45:26.968849 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-ring-rebalance-t9tqb" podStartSLOduration=2.968831825 podStartE2EDuration="2.968831825s" podCreationTimestamp="2025-10-08 19:45:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 19:45:26.95765794 +0000 UTC m=+5672.407500720" watchObservedRunningTime="2025-10-08 19:45:26.968831825 +0000 UTC m=+5672.418674595" Oct 08 19:45:27 crc kubenswrapper[4988]: I1008 19:45:27.008672 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7f4d7998-e850-47be-bc8b-b9c2c5f354d0-log-httpd\") pod \"swift-proxy-5dfffdbbd6-gzqmd\" (UID: \"7f4d7998-e850-47be-bc8b-b9c2c5f354d0\") " pod="openstack/swift-proxy-5dfffdbbd6-gzqmd" Oct 08 19:45:27 crc kubenswrapper[4988]: I1008 19:45:27.008816 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7f4d7998-e850-47be-bc8b-b9c2c5f354d0-config-data\") pod \"swift-proxy-5dfffdbbd6-gzqmd\" (UID: \"7f4d7998-e850-47be-bc8b-b9c2c5f354d0\") " pod="openstack/swift-proxy-5dfffdbbd6-gzqmd" Oct 08 19:45:27 crc kubenswrapper[4988]: I1008 19:45:27.008859 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pcngc\" (UniqueName: \"kubernetes.io/projected/7f4d7998-e850-47be-bc8b-b9c2c5f354d0-kube-api-access-pcngc\") pod \"swift-proxy-5dfffdbbd6-gzqmd\" (UID: \"7f4d7998-e850-47be-bc8b-b9c2c5f354d0\") " pod="openstack/swift-proxy-5dfffdbbd6-gzqmd" Oct 08 19:45:27 crc kubenswrapper[4988]: I1008 19:45:27.008901 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7f4d7998-e850-47be-bc8b-b9c2c5f354d0-run-httpd\") pod \"swift-proxy-5dfffdbbd6-gzqmd\" (UID: \"7f4d7998-e850-47be-bc8b-b9c2c5f354d0\") " pod="openstack/swift-proxy-5dfffdbbd6-gzqmd" Oct 08 19:45:27 crc kubenswrapper[4988]: I1008 19:45:27.009053 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/7f4d7998-e850-47be-bc8b-b9c2c5f354d0-etc-swift\") pod \"swift-proxy-5dfffdbbd6-gzqmd\" (UID: \"7f4d7998-e850-47be-bc8b-b9c2c5f354d0\") " pod="openstack/swift-proxy-5dfffdbbd6-gzqmd" Oct 08 19:45:27 crc kubenswrapper[4988]: I1008 19:45:27.009127 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7f4d7998-e850-47be-bc8b-b9c2c5f354d0-combined-ca-bundle\") pod \"swift-proxy-5dfffdbbd6-gzqmd\" (UID: \"7f4d7998-e850-47be-bc8b-b9c2c5f354d0\") " pod="openstack/swift-proxy-5dfffdbbd6-gzqmd" Oct 08 19:45:27 crc kubenswrapper[4988]: I1008 19:45:27.118799 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/7f4d7998-e850-47be-bc8b-b9c2c5f354d0-etc-swift\") pod \"swift-proxy-5dfffdbbd6-gzqmd\" (UID: \"7f4d7998-e850-47be-bc8b-b9c2c5f354d0\") " pod="openstack/swift-proxy-5dfffdbbd6-gzqmd" Oct 08 19:45:27 crc kubenswrapper[4988]: I1008 19:45:27.119087 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7f4d7998-e850-47be-bc8b-b9c2c5f354d0-combined-ca-bundle\") pod \"swift-proxy-5dfffdbbd6-gzqmd\" (UID: \"7f4d7998-e850-47be-bc8b-b9c2c5f354d0\") " pod="openstack/swift-proxy-5dfffdbbd6-gzqmd" Oct 08 19:45:27 crc kubenswrapper[4988]: I1008 19:45:27.119270 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7f4d7998-e850-47be-bc8b-b9c2c5f354d0-log-httpd\") pod \"swift-proxy-5dfffdbbd6-gzqmd\" (UID: \"7f4d7998-e850-47be-bc8b-b9c2c5f354d0\") " pod="openstack/swift-proxy-5dfffdbbd6-gzqmd" Oct 08 19:45:27 crc kubenswrapper[4988]: I1008 19:45:27.119405 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7f4d7998-e850-47be-bc8b-b9c2c5f354d0-config-data\") pod \"swift-proxy-5dfffdbbd6-gzqmd\" (UID: \"7f4d7998-e850-47be-bc8b-b9c2c5f354d0\") " pod="openstack/swift-proxy-5dfffdbbd6-gzqmd" Oct 08 19:45:27 crc kubenswrapper[4988]: I1008 19:45:27.119513 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pcngc\" (UniqueName: \"kubernetes.io/projected/7f4d7998-e850-47be-bc8b-b9c2c5f354d0-kube-api-access-pcngc\") pod \"swift-proxy-5dfffdbbd6-gzqmd\" (UID: \"7f4d7998-e850-47be-bc8b-b9c2c5f354d0\") " pod="openstack/swift-proxy-5dfffdbbd6-gzqmd" Oct 08 19:45:27 crc kubenswrapper[4988]: I1008 19:45:27.119605 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7f4d7998-e850-47be-bc8b-b9c2c5f354d0-run-httpd\") pod \"swift-proxy-5dfffdbbd6-gzqmd\" (UID: \"7f4d7998-e850-47be-bc8b-b9c2c5f354d0\") " pod="openstack/swift-proxy-5dfffdbbd6-gzqmd" Oct 08 19:45:27 crc kubenswrapper[4988]: I1008 19:45:27.120435 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7f4d7998-e850-47be-bc8b-b9c2c5f354d0-log-httpd\") pod \"swift-proxy-5dfffdbbd6-gzqmd\" (UID: \"7f4d7998-e850-47be-bc8b-b9c2c5f354d0\") " pod="openstack/swift-proxy-5dfffdbbd6-gzqmd" Oct 08 19:45:27 crc kubenswrapper[4988]: I1008 19:45:27.120595 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7f4d7998-e850-47be-bc8b-b9c2c5f354d0-run-httpd\") pod \"swift-proxy-5dfffdbbd6-gzqmd\" (UID: \"7f4d7998-e850-47be-bc8b-b9c2c5f354d0\") " pod="openstack/swift-proxy-5dfffdbbd6-gzqmd" Oct 08 19:45:27 crc kubenswrapper[4988]: I1008 19:45:27.127219 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/7f4d7998-e850-47be-bc8b-b9c2c5f354d0-etc-swift\") pod \"swift-proxy-5dfffdbbd6-gzqmd\" (UID: \"7f4d7998-e850-47be-bc8b-b9c2c5f354d0\") " pod="openstack/swift-proxy-5dfffdbbd6-gzqmd" Oct 08 19:45:27 crc kubenswrapper[4988]: I1008 19:45:27.127837 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7f4d7998-e850-47be-bc8b-b9c2c5f354d0-combined-ca-bundle\") pod \"swift-proxy-5dfffdbbd6-gzqmd\" (UID: \"7f4d7998-e850-47be-bc8b-b9c2c5f354d0\") " pod="openstack/swift-proxy-5dfffdbbd6-gzqmd" Oct 08 19:45:27 crc kubenswrapper[4988]: I1008 19:45:27.134290 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7f4d7998-e850-47be-bc8b-b9c2c5f354d0-config-data\") pod \"swift-proxy-5dfffdbbd6-gzqmd\" (UID: \"7f4d7998-e850-47be-bc8b-b9c2c5f354d0\") " pod="openstack/swift-proxy-5dfffdbbd6-gzqmd" Oct 08 19:45:27 crc kubenswrapper[4988]: I1008 19:45:27.138524 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pcngc\" (UniqueName: \"kubernetes.io/projected/7f4d7998-e850-47be-bc8b-b9c2c5f354d0-kube-api-access-pcngc\") pod \"swift-proxy-5dfffdbbd6-gzqmd\" (UID: \"7f4d7998-e850-47be-bc8b-b9c2c5f354d0\") " pod="openstack/swift-proxy-5dfffdbbd6-gzqmd" Oct 08 19:45:27 crc kubenswrapper[4988]: I1008 19:45:27.199247 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-5dfffdbbd6-gzqmd" Oct 08 19:45:27 crc kubenswrapper[4988]: I1008 19:45:27.844758 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-proxy-5dfffdbbd6-gzqmd"] Oct 08 19:45:27 crc kubenswrapper[4988]: I1008 19:45:27.950650 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-5dfffdbbd6-gzqmd" event={"ID":"7f4d7998-e850-47be-bc8b-b9c2c5f354d0","Type":"ContainerStarted","Data":"e483d4e1a94aaa04403fa6b98cc6a94d3eb0b133fd884d227b0c35bf8b0b7ca6"} Oct 08 19:45:27 crc kubenswrapper[4988]: I1008 19:45:27.952976 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6d64fc8cdc-jb7cp" event={"ID":"5bd22f4a-f899-4db4-88a1-10344841593b","Type":"ContainerStarted","Data":"ac7105c5d8e7c847fbb985bf98c080eaf4caf929bda6fd9f77fa4b83a42e3f8c"} Oct 08 19:45:27 crc kubenswrapper[4988]: I1008 19:45:27.980174 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-6d64fc8cdc-jb7cp" podStartSLOduration=2.980152463 podStartE2EDuration="2.980152463s" podCreationTimestamp="2025-10-08 19:45:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 19:45:27.978452879 +0000 UTC m=+5673.428295669" watchObservedRunningTime="2025-10-08 19:45:27.980152463 +0000 UTC m=+5673.429995233" Oct 08 19:45:28 crc kubenswrapper[4988]: I1008 19:45:28.966902 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-5dfffdbbd6-gzqmd" event={"ID":"7f4d7998-e850-47be-bc8b-b9c2c5f354d0","Type":"ContainerStarted","Data":"cd0754537c4a649ec0ef88b0c87d2d7aaf840e345d3451b5b177067ddbba6fb4"} Oct 08 19:45:28 crc kubenswrapper[4988]: I1008 19:45:28.967254 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-5dfffdbbd6-gzqmd" event={"ID":"7f4d7998-e850-47be-bc8b-b9c2c5f354d0","Type":"ContainerStarted","Data":"fc15ca42b349f5f45532359cc4b75f8b060255fe035d5a3e7aa4c423e79de0ea"} Oct 08 19:45:28 crc kubenswrapper[4988]: I1008 19:45:28.967412 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-6d64fc8cdc-jb7cp" Oct 08 19:45:28 crc kubenswrapper[4988]: I1008 19:45:28.967433 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/swift-proxy-5dfffdbbd6-gzqmd" Oct 08 19:45:28 crc kubenswrapper[4988]: I1008 19:45:28.967444 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/swift-proxy-5dfffdbbd6-gzqmd" Oct 08 19:45:28 crc kubenswrapper[4988]: I1008 19:45:28.993662 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-proxy-5dfffdbbd6-gzqmd" podStartSLOduration=2.99363321 podStartE2EDuration="2.99363321s" podCreationTimestamp="2025-10-08 19:45:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 19:45:28.992967218 +0000 UTC m=+5674.442809998" watchObservedRunningTime="2025-10-08 19:45:28.99363321 +0000 UTC m=+5674.443475980" Oct 08 19:45:29 crc kubenswrapper[4988]: I1008 19:45:29.022993 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-proxy-b8548dfd6-qblss"] Oct 08 19:45:29 crc kubenswrapper[4988]: I1008 19:45:29.024606 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-b8548dfd6-qblss" Oct 08 19:45:29 crc kubenswrapper[4988]: I1008 19:45:29.026911 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-swift-internal-svc" Oct 08 19:45:29 crc kubenswrapper[4988]: I1008 19:45:29.027131 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-swift-public-svc" Oct 08 19:45:29 crc kubenswrapper[4988]: I1008 19:45:29.044409 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-proxy-b8548dfd6-qblss"] Oct 08 19:45:29 crc kubenswrapper[4988]: I1008 19:45:29.151717 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/679a7d0a-3ad1-45c8-b1ff-6772a862379a-internal-tls-certs\") pod \"swift-proxy-b8548dfd6-qblss\" (UID: \"679a7d0a-3ad1-45c8-b1ff-6772a862379a\") " pod="openstack/swift-proxy-b8548dfd6-qblss" Oct 08 19:45:29 crc kubenswrapper[4988]: I1008 19:45:29.151752 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/679a7d0a-3ad1-45c8-b1ff-6772a862379a-combined-ca-bundle\") pod \"swift-proxy-b8548dfd6-qblss\" (UID: \"679a7d0a-3ad1-45c8-b1ff-6772a862379a\") " pod="openstack/swift-proxy-b8548dfd6-qblss" Oct 08 19:45:29 crc kubenswrapper[4988]: I1008 19:45:29.151782 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/679a7d0a-3ad1-45c8-b1ff-6772a862379a-log-httpd\") pod \"swift-proxy-b8548dfd6-qblss\" (UID: \"679a7d0a-3ad1-45c8-b1ff-6772a862379a\") " pod="openstack/swift-proxy-b8548dfd6-qblss" Oct 08 19:45:29 crc kubenswrapper[4988]: I1008 19:45:29.151812 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/679a7d0a-3ad1-45c8-b1ff-6772a862379a-public-tls-certs\") pod \"swift-proxy-b8548dfd6-qblss\" (UID: \"679a7d0a-3ad1-45c8-b1ff-6772a862379a\") " pod="openstack/swift-proxy-b8548dfd6-qblss" Oct 08 19:45:29 crc kubenswrapper[4988]: I1008 19:45:29.152066 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7mkld\" (UniqueName: \"kubernetes.io/projected/679a7d0a-3ad1-45c8-b1ff-6772a862379a-kube-api-access-7mkld\") pod \"swift-proxy-b8548dfd6-qblss\" (UID: \"679a7d0a-3ad1-45c8-b1ff-6772a862379a\") " pod="openstack/swift-proxy-b8548dfd6-qblss" Oct 08 19:45:29 crc kubenswrapper[4988]: I1008 19:45:29.152238 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/679a7d0a-3ad1-45c8-b1ff-6772a862379a-run-httpd\") pod \"swift-proxy-b8548dfd6-qblss\" (UID: \"679a7d0a-3ad1-45c8-b1ff-6772a862379a\") " pod="openstack/swift-proxy-b8548dfd6-qblss" Oct 08 19:45:29 crc kubenswrapper[4988]: I1008 19:45:29.152321 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/679a7d0a-3ad1-45c8-b1ff-6772a862379a-etc-swift\") pod \"swift-proxy-b8548dfd6-qblss\" (UID: \"679a7d0a-3ad1-45c8-b1ff-6772a862379a\") " pod="openstack/swift-proxy-b8548dfd6-qblss" Oct 08 19:45:29 crc kubenswrapper[4988]: I1008 19:45:29.152410 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/679a7d0a-3ad1-45c8-b1ff-6772a862379a-config-data\") pod \"swift-proxy-b8548dfd6-qblss\" (UID: \"679a7d0a-3ad1-45c8-b1ff-6772a862379a\") " pod="openstack/swift-proxy-b8548dfd6-qblss" Oct 08 19:45:29 crc kubenswrapper[4988]: I1008 19:45:29.254537 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/679a7d0a-3ad1-45c8-b1ff-6772a862379a-public-tls-certs\") pod \"swift-proxy-b8548dfd6-qblss\" (UID: \"679a7d0a-3ad1-45c8-b1ff-6772a862379a\") " pod="openstack/swift-proxy-b8548dfd6-qblss" Oct 08 19:45:29 crc kubenswrapper[4988]: I1008 19:45:29.254638 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7mkld\" (UniqueName: \"kubernetes.io/projected/679a7d0a-3ad1-45c8-b1ff-6772a862379a-kube-api-access-7mkld\") pod \"swift-proxy-b8548dfd6-qblss\" (UID: \"679a7d0a-3ad1-45c8-b1ff-6772a862379a\") " pod="openstack/swift-proxy-b8548dfd6-qblss" Oct 08 19:45:29 crc kubenswrapper[4988]: I1008 19:45:29.254676 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/679a7d0a-3ad1-45c8-b1ff-6772a862379a-run-httpd\") pod \"swift-proxy-b8548dfd6-qblss\" (UID: \"679a7d0a-3ad1-45c8-b1ff-6772a862379a\") " pod="openstack/swift-proxy-b8548dfd6-qblss" Oct 08 19:45:29 crc kubenswrapper[4988]: I1008 19:45:29.254698 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/679a7d0a-3ad1-45c8-b1ff-6772a862379a-etc-swift\") pod \"swift-proxy-b8548dfd6-qblss\" (UID: \"679a7d0a-3ad1-45c8-b1ff-6772a862379a\") " pod="openstack/swift-proxy-b8548dfd6-qblss" Oct 08 19:45:29 crc kubenswrapper[4988]: I1008 19:45:29.254723 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/679a7d0a-3ad1-45c8-b1ff-6772a862379a-config-data\") pod \"swift-proxy-b8548dfd6-qblss\" (UID: \"679a7d0a-3ad1-45c8-b1ff-6772a862379a\") " pod="openstack/swift-proxy-b8548dfd6-qblss" Oct 08 19:45:29 crc kubenswrapper[4988]: I1008 19:45:29.254787 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/679a7d0a-3ad1-45c8-b1ff-6772a862379a-combined-ca-bundle\") pod \"swift-proxy-b8548dfd6-qblss\" (UID: \"679a7d0a-3ad1-45c8-b1ff-6772a862379a\") " pod="openstack/swift-proxy-b8548dfd6-qblss" Oct 08 19:45:29 crc kubenswrapper[4988]: I1008 19:45:29.254806 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/679a7d0a-3ad1-45c8-b1ff-6772a862379a-internal-tls-certs\") pod \"swift-proxy-b8548dfd6-qblss\" (UID: \"679a7d0a-3ad1-45c8-b1ff-6772a862379a\") " pod="openstack/swift-proxy-b8548dfd6-qblss" Oct 08 19:45:29 crc kubenswrapper[4988]: I1008 19:45:29.254837 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/679a7d0a-3ad1-45c8-b1ff-6772a862379a-log-httpd\") pod \"swift-proxy-b8548dfd6-qblss\" (UID: \"679a7d0a-3ad1-45c8-b1ff-6772a862379a\") " pod="openstack/swift-proxy-b8548dfd6-qblss" Oct 08 19:45:29 crc kubenswrapper[4988]: I1008 19:45:29.255286 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/679a7d0a-3ad1-45c8-b1ff-6772a862379a-log-httpd\") pod \"swift-proxy-b8548dfd6-qblss\" (UID: \"679a7d0a-3ad1-45c8-b1ff-6772a862379a\") " pod="openstack/swift-proxy-b8548dfd6-qblss" Oct 08 19:45:29 crc kubenswrapper[4988]: I1008 19:45:29.260301 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/679a7d0a-3ad1-45c8-b1ff-6772a862379a-run-httpd\") pod \"swift-proxy-b8548dfd6-qblss\" (UID: \"679a7d0a-3ad1-45c8-b1ff-6772a862379a\") " pod="openstack/swift-proxy-b8548dfd6-qblss" Oct 08 19:45:29 crc kubenswrapper[4988]: I1008 19:45:29.263321 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/679a7d0a-3ad1-45c8-b1ff-6772a862379a-internal-tls-certs\") pod \"swift-proxy-b8548dfd6-qblss\" (UID: \"679a7d0a-3ad1-45c8-b1ff-6772a862379a\") " pod="openstack/swift-proxy-b8548dfd6-qblss" Oct 08 19:45:29 crc kubenswrapper[4988]: I1008 19:45:29.263531 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/679a7d0a-3ad1-45c8-b1ff-6772a862379a-public-tls-certs\") pod \"swift-proxy-b8548dfd6-qblss\" (UID: \"679a7d0a-3ad1-45c8-b1ff-6772a862379a\") " pod="openstack/swift-proxy-b8548dfd6-qblss" Oct 08 19:45:29 crc kubenswrapper[4988]: I1008 19:45:29.264752 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/679a7d0a-3ad1-45c8-b1ff-6772a862379a-etc-swift\") pod \"swift-proxy-b8548dfd6-qblss\" (UID: \"679a7d0a-3ad1-45c8-b1ff-6772a862379a\") " pod="openstack/swift-proxy-b8548dfd6-qblss" Oct 08 19:45:29 crc kubenswrapper[4988]: I1008 19:45:29.269638 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/679a7d0a-3ad1-45c8-b1ff-6772a862379a-config-data\") pod \"swift-proxy-b8548dfd6-qblss\" (UID: \"679a7d0a-3ad1-45c8-b1ff-6772a862379a\") " pod="openstack/swift-proxy-b8548dfd6-qblss" Oct 08 19:45:29 crc kubenswrapper[4988]: I1008 19:45:29.271206 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/679a7d0a-3ad1-45c8-b1ff-6772a862379a-combined-ca-bundle\") pod \"swift-proxy-b8548dfd6-qblss\" (UID: \"679a7d0a-3ad1-45c8-b1ff-6772a862379a\") " pod="openstack/swift-proxy-b8548dfd6-qblss" Oct 08 19:45:29 crc kubenswrapper[4988]: I1008 19:45:29.274065 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7mkld\" (UniqueName: \"kubernetes.io/projected/679a7d0a-3ad1-45c8-b1ff-6772a862379a-kube-api-access-7mkld\") pod \"swift-proxy-b8548dfd6-qblss\" (UID: \"679a7d0a-3ad1-45c8-b1ff-6772a862379a\") " pod="openstack/swift-proxy-b8548dfd6-qblss" Oct 08 19:45:29 crc kubenswrapper[4988]: I1008 19:45:29.392796 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-b8548dfd6-qblss" Oct 08 19:45:29 crc kubenswrapper[4988]: I1008 19:45:29.980259 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-t9tqb" event={"ID":"de3434fc-c4fa-4495-b3f3-7f232decfed9","Type":"ContainerDied","Data":"c886858da5e42884b9e8ec85277700eab6cb46f9bd3d7f965f102a724c3e7609"} Oct 08 19:45:29 crc kubenswrapper[4988]: I1008 19:45:29.980184 4988 generic.go:334] "Generic (PLEG): container finished" podID="de3434fc-c4fa-4495-b3f3-7f232decfed9" containerID="c886858da5e42884b9e8ec85277700eab6cb46f9bd3d7f965f102a724c3e7609" exitCode=0 Oct 08 19:45:30 crc kubenswrapper[4988]: I1008 19:45:30.026820 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-proxy-b8548dfd6-qblss"] Oct 08 19:45:30 crc kubenswrapper[4988]: W1008 19:45:30.029769 4988 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod679a7d0a_3ad1_45c8_b1ff_6772a862379a.slice/crio-8e4a089b518c62efb60d31abf7f1d6698ec3447b5ccaa5967f0bce61c1ab0cae WatchSource:0}: Error finding container 8e4a089b518c62efb60d31abf7f1d6698ec3447b5ccaa5967f0bce61c1ab0cae: Status 404 returned error can't find the container with id 8e4a089b518c62efb60d31abf7f1d6698ec3447b5ccaa5967f0bce61c1ab0cae Oct 08 19:45:31 crc kubenswrapper[4988]: I1008 19:45:31.000740 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-b8548dfd6-qblss" event={"ID":"679a7d0a-3ad1-45c8-b1ff-6772a862379a","Type":"ContainerStarted","Data":"93bc7bc357838737ead637c857a0d0512cd04671339ae7a4065f79e298a2fba3"} Oct 08 19:45:31 crc kubenswrapper[4988]: I1008 19:45:31.002657 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/swift-proxy-b8548dfd6-qblss" Oct 08 19:45:31 crc kubenswrapper[4988]: I1008 19:45:31.002687 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/swift-proxy-b8548dfd6-qblss" Oct 08 19:45:31 crc kubenswrapper[4988]: I1008 19:45:31.002706 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-b8548dfd6-qblss" event={"ID":"679a7d0a-3ad1-45c8-b1ff-6772a862379a","Type":"ContainerStarted","Data":"6306d8b0ef6a85a7a2244b7f14e74fd6bb9a55acd6e230d181b52f2fdb7d462a"} Oct 08 19:45:31 crc kubenswrapper[4988]: I1008 19:45:31.002730 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-b8548dfd6-qblss" event={"ID":"679a7d0a-3ad1-45c8-b1ff-6772a862379a","Type":"ContainerStarted","Data":"8e4a089b518c62efb60d31abf7f1d6698ec3447b5ccaa5967f0bce61c1ab0cae"} Oct 08 19:45:31 crc kubenswrapper[4988]: I1008 19:45:31.061424 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-proxy-b8548dfd6-qblss" podStartSLOduration=2.061370548 podStartE2EDuration="2.061370548s" podCreationTimestamp="2025-10-08 19:45:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 19:45:31.037847751 +0000 UTC m=+5676.487690571" watchObservedRunningTime="2025-10-08 19:45:31.061370548 +0000 UTC m=+5676.511213358" Oct 08 19:45:31 crc kubenswrapper[4988]: I1008 19:45:31.416186 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-t9tqb" Oct 08 19:45:31 crc kubenswrapper[4988]: I1008 19:45:31.506470 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/de3434fc-c4fa-4495-b3f3-7f232decfed9-combined-ca-bundle\") pod \"de3434fc-c4fa-4495-b3f3-7f232decfed9\" (UID: \"de3434fc-c4fa-4495-b3f3-7f232decfed9\") " Oct 08 19:45:31 crc kubenswrapper[4988]: I1008 19:45:31.506559 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/de3434fc-c4fa-4495-b3f3-7f232decfed9-dispersionconf\") pod \"de3434fc-c4fa-4495-b3f3-7f232decfed9\" (UID: \"de3434fc-c4fa-4495-b3f3-7f232decfed9\") " Oct 08 19:45:31 crc kubenswrapper[4988]: I1008 19:45:31.506585 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/de3434fc-c4fa-4495-b3f3-7f232decfed9-ring-data-devices\") pod \"de3434fc-c4fa-4495-b3f3-7f232decfed9\" (UID: \"de3434fc-c4fa-4495-b3f3-7f232decfed9\") " Oct 08 19:45:31 crc kubenswrapper[4988]: I1008 19:45:31.506633 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/de3434fc-c4fa-4495-b3f3-7f232decfed9-scripts\") pod \"de3434fc-c4fa-4495-b3f3-7f232decfed9\" (UID: \"de3434fc-c4fa-4495-b3f3-7f232decfed9\") " Oct 08 19:45:31 crc kubenswrapper[4988]: I1008 19:45:31.506694 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-k5f6q\" (UniqueName: \"kubernetes.io/projected/de3434fc-c4fa-4495-b3f3-7f232decfed9-kube-api-access-k5f6q\") pod \"de3434fc-c4fa-4495-b3f3-7f232decfed9\" (UID: \"de3434fc-c4fa-4495-b3f3-7f232decfed9\") " Oct 08 19:45:31 crc kubenswrapper[4988]: I1008 19:45:31.506734 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/de3434fc-c4fa-4495-b3f3-7f232decfed9-swiftconf\") pod \"de3434fc-c4fa-4495-b3f3-7f232decfed9\" (UID: \"de3434fc-c4fa-4495-b3f3-7f232decfed9\") " Oct 08 19:45:31 crc kubenswrapper[4988]: I1008 19:45:31.506823 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/de3434fc-c4fa-4495-b3f3-7f232decfed9-etc-swift\") pod \"de3434fc-c4fa-4495-b3f3-7f232decfed9\" (UID: \"de3434fc-c4fa-4495-b3f3-7f232decfed9\") " Oct 08 19:45:31 crc kubenswrapper[4988]: I1008 19:45:31.507778 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/de3434fc-c4fa-4495-b3f3-7f232decfed9-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "de3434fc-c4fa-4495-b3f3-7f232decfed9" (UID: "de3434fc-c4fa-4495-b3f3-7f232decfed9"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 19:45:31 crc kubenswrapper[4988]: I1008 19:45:31.508646 4988 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/de3434fc-c4fa-4495-b3f3-7f232decfed9-etc-swift\") on node \"crc\" DevicePath \"\"" Oct 08 19:45:31 crc kubenswrapper[4988]: I1008 19:45:31.514582 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/de3434fc-c4fa-4495-b3f3-7f232decfed9-kube-api-access-k5f6q" (OuterVolumeSpecName: "kube-api-access-k5f6q") pod "de3434fc-c4fa-4495-b3f3-7f232decfed9" (UID: "de3434fc-c4fa-4495-b3f3-7f232decfed9"). InnerVolumeSpecName "kube-api-access-k5f6q". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 19:45:31 crc kubenswrapper[4988]: I1008 19:45:31.514954 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/de3434fc-c4fa-4495-b3f3-7f232decfed9-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "de3434fc-c4fa-4495-b3f3-7f232decfed9" (UID: "de3434fc-c4fa-4495-b3f3-7f232decfed9"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 19:45:31 crc kubenswrapper[4988]: I1008 19:45:31.532289 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/de3434fc-c4fa-4495-b3f3-7f232decfed9-scripts" (OuterVolumeSpecName: "scripts") pod "de3434fc-c4fa-4495-b3f3-7f232decfed9" (UID: "de3434fc-c4fa-4495-b3f3-7f232decfed9"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 19:45:31 crc kubenswrapper[4988]: I1008 19:45:31.539479 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/de3434fc-c4fa-4495-b3f3-7f232decfed9-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "de3434fc-c4fa-4495-b3f3-7f232decfed9" (UID: "de3434fc-c4fa-4495-b3f3-7f232decfed9"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 19:45:31 crc kubenswrapper[4988]: I1008 19:45:31.564239 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/de3434fc-c4fa-4495-b3f3-7f232decfed9-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "de3434fc-c4fa-4495-b3f3-7f232decfed9" (UID: "de3434fc-c4fa-4495-b3f3-7f232decfed9"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 19:45:31 crc kubenswrapper[4988]: I1008 19:45:31.568959 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/de3434fc-c4fa-4495-b3f3-7f232decfed9-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "de3434fc-c4fa-4495-b3f3-7f232decfed9" (UID: "de3434fc-c4fa-4495-b3f3-7f232decfed9"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 19:45:31 crc kubenswrapper[4988]: I1008 19:45:31.610325 4988 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/de3434fc-c4fa-4495-b3f3-7f232decfed9-scripts\") on node \"crc\" DevicePath \"\"" Oct 08 19:45:31 crc kubenswrapper[4988]: I1008 19:45:31.610363 4988 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-k5f6q\" (UniqueName: \"kubernetes.io/projected/de3434fc-c4fa-4495-b3f3-7f232decfed9-kube-api-access-k5f6q\") on node \"crc\" DevicePath \"\"" Oct 08 19:45:31 crc kubenswrapper[4988]: I1008 19:45:31.610377 4988 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/de3434fc-c4fa-4495-b3f3-7f232decfed9-swiftconf\") on node \"crc\" DevicePath \"\"" Oct 08 19:45:31 crc kubenswrapper[4988]: I1008 19:45:31.610412 4988 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/de3434fc-c4fa-4495-b3f3-7f232decfed9-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 19:45:31 crc kubenswrapper[4988]: I1008 19:45:31.610427 4988 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/de3434fc-c4fa-4495-b3f3-7f232decfed9-dispersionconf\") on node \"crc\" DevicePath \"\"" Oct 08 19:45:31 crc kubenswrapper[4988]: I1008 19:45:31.610458 4988 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/de3434fc-c4fa-4495-b3f3-7f232decfed9-ring-data-devices\") on node \"crc\" DevicePath \"\"" Oct 08 19:45:32 crc kubenswrapper[4988]: I1008 19:45:32.015744 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-t9tqb" Oct 08 19:45:32 crc kubenswrapper[4988]: I1008 19:45:32.015772 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-t9tqb" event={"ID":"de3434fc-c4fa-4495-b3f3-7f232decfed9","Type":"ContainerDied","Data":"ce0527c232d6cb4d95475048b0ea1fd748afdfa0f4a5ff109654451872209643"} Oct 08 19:45:32 crc kubenswrapper[4988]: I1008 19:45:32.015890 4988 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ce0527c232d6cb4d95475048b0ea1fd748afdfa0f4a5ff109654451872209643" Oct 08 19:45:35 crc kubenswrapper[4988]: I1008 19:45:35.414356 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-6d64fc8cdc-jb7cp" Oct 08 19:45:35 crc kubenswrapper[4988]: I1008 19:45:35.508727 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-fbc86f6b9-gwjfv"] Oct 08 19:45:35 crc kubenswrapper[4988]: I1008 19:45:35.509006 4988 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-fbc86f6b9-gwjfv" podUID="c9742f58-7745-4b60-b61d-e864ceecf087" containerName="dnsmasq-dns" containerID="cri-o://b491f91185e11de75a9e806ba7683fef2192b04aeda5c624f5f816ddf28e9426" gracePeriod=10 Oct 08 19:45:36 crc kubenswrapper[4988]: I1008 19:45:36.004285 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-fbc86f6b9-gwjfv" Oct 08 19:45:36 crc kubenswrapper[4988]: I1008 19:45:36.058939 4988 generic.go:334] "Generic (PLEG): container finished" podID="c9742f58-7745-4b60-b61d-e864ceecf087" containerID="b491f91185e11de75a9e806ba7683fef2192b04aeda5c624f5f816ddf28e9426" exitCode=0 Oct 08 19:45:36 crc kubenswrapper[4988]: I1008 19:45:36.059007 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-fbc86f6b9-gwjfv" Oct 08 19:45:36 crc kubenswrapper[4988]: I1008 19:45:36.059023 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-fbc86f6b9-gwjfv" event={"ID":"c9742f58-7745-4b60-b61d-e864ceecf087","Type":"ContainerDied","Data":"b491f91185e11de75a9e806ba7683fef2192b04aeda5c624f5f816ddf28e9426"} Oct 08 19:45:36 crc kubenswrapper[4988]: I1008 19:45:36.060187 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-fbc86f6b9-gwjfv" event={"ID":"c9742f58-7745-4b60-b61d-e864ceecf087","Type":"ContainerDied","Data":"d87660c7e9ab7242d4e5b27d3809ac4c2356a06887540b0b5f2ae3d78655d5f3"} Oct 08 19:45:36 crc kubenswrapper[4988]: I1008 19:45:36.060220 4988 scope.go:117] "RemoveContainer" containerID="b491f91185e11de75a9e806ba7683fef2192b04aeda5c624f5f816ddf28e9426" Oct 08 19:45:36 crc kubenswrapper[4988]: I1008 19:45:36.096696 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c9742f58-7745-4b60-b61d-e864ceecf087-config\") pod \"c9742f58-7745-4b60-b61d-e864ceecf087\" (UID: \"c9742f58-7745-4b60-b61d-e864ceecf087\") " Oct 08 19:45:36 crc kubenswrapper[4988]: I1008 19:45:36.096841 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c9742f58-7745-4b60-b61d-e864ceecf087-ovsdbserver-nb\") pod \"c9742f58-7745-4b60-b61d-e864ceecf087\" (UID: \"c9742f58-7745-4b60-b61d-e864ceecf087\") " Oct 08 19:45:36 crc kubenswrapper[4988]: I1008 19:45:36.096925 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jdh86\" (UniqueName: \"kubernetes.io/projected/c9742f58-7745-4b60-b61d-e864ceecf087-kube-api-access-jdh86\") pod \"c9742f58-7745-4b60-b61d-e864ceecf087\" (UID: \"c9742f58-7745-4b60-b61d-e864ceecf087\") " Oct 08 19:45:36 crc kubenswrapper[4988]: I1008 19:45:36.097033 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c9742f58-7745-4b60-b61d-e864ceecf087-ovsdbserver-sb\") pod \"c9742f58-7745-4b60-b61d-e864ceecf087\" (UID: \"c9742f58-7745-4b60-b61d-e864ceecf087\") " Oct 08 19:45:36 crc kubenswrapper[4988]: I1008 19:45:36.097308 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c9742f58-7745-4b60-b61d-e864ceecf087-dns-svc\") pod \"c9742f58-7745-4b60-b61d-e864ceecf087\" (UID: \"c9742f58-7745-4b60-b61d-e864ceecf087\") " Oct 08 19:45:36 crc kubenswrapper[4988]: I1008 19:45:36.101838 4988 scope.go:117] "RemoveContainer" containerID="3cd75657be77ae5c79df8af727ac7439f43f1564944848bdd1528f1b162ff4a1" Oct 08 19:45:36 crc kubenswrapper[4988]: I1008 19:45:36.106900 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c9742f58-7745-4b60-b61d-e864ceecf087-kube-api-access-jdh86" (OuterVolumeSpecName: "kube-api-access-jdh86") pod "c9742f58-7745-4b60-b61d-e864ceecf087" (UID: "c9742f58-7745-4b60-b61d-e864ceecf087"). InnerVolumeSpecName "kube-api-access-jdh86". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 19:45:36 crc kubenswrapper[4988]: I1008 19:45:36.143962 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c9742f58-7745-4b60-b61d-e864ceecf087-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "c9742f58-7745-4b60-b61d-e864ceecf087" (UID: "c9742f58-7745-4b60-b61d-e864ceecf087"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 19:45:36 crc kubenswrapper[4988]: I1008 19:45:36.147351 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c9742f58-7745-4b60-b61d-e864ceecf087-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "c9742f58-7745-4b60-b61d-e864ceecf087" (UID: "c9742f58-7745-4b60-b61d-e864ceecf087"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 19:45:36 crc kubenswrapper[4988]: I1008 19:45:36.162059 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c9742f58-7745-4b60-b61d-e864ceecf087-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "c9742f58-7745-4b60-b61d-e864ceecf087" (UID: "c9742f58-7745-4b60-b61d-e864ceecf087"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 19:45:36 crc kubenswrapper[4988]: I1008 19:45:36.169287 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c9742f58-7745-4b60-b61d-e864ceecf087-config" (OuterVolumeSpecName: "config") pod "c9742f58-7745-4b60-b61d-e864ceecf087" (UID: "c9742f58-7745-4b60-b61d-e864ceecf087"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 19:45:36 crc kubenswrapper[4988]: I1008 19:45:36.201636 4988 scope.go:117] "RemoveContainer" containerID="b491f91185e11de75a9e806ba7683fef2192b04aeda5c624f5f816ddf28e9426" Oct 08 19:45:36 crc kubenswrapper[4988]: E1008 19:45:36.203135 4988 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b491f91185e11de75a9e806ba7683fef2192b04aeda5c624f5f816ddf28e9426\": container with ID starting with b491f91185e11de75a9e806ba7683fef2192b04aeda5c624f5f816ddf28e9426 not found: ID does not exist" containerID="b491f91185e11de75a9e806ba7683fef2192b04aeda5c624f5f816ddf28e9426" Oct 08 19:45:36 crc kubenswrapper[4988]: I1008 19:45:36.203173 4988 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b491f91185e11de75a9e806ba7683fef2192b04aeda5c624f5f816ddf28e9426"} err="failed to get container status \"b491f91185e11de75a9e806ba7683fef2192b04aeda5c624f5f816ddf28e9426\": rpc error: code = NotFound desc = could not find container \"b491f91185e11de75a9e806ba7683fef2192b04aeda5c624f5f816ddf28e9426\": container with ID starting with b491f91185e11de75a9e806ba7683fef2192b04aeda5c624f5f816ddf28e9426 not found: ID does not exist" Oct 08 19:45:36 crc kubenswrapper[4988]: I1008 19:45:36.203199 4988 scope.go:117] "RemoveContainer" containerID="3cd75657be77ae5c79df8af727ac7439f43f1564944848bdd1528f1b162ff4a1" Oct 08 19:45:36 crc kubenswrapper[4988]: I1008 19:45:36.203593 4988 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c9742f58-7745-4b60-b61d-e864ceecf087-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 08 19:45:36 crc kubenswrapper[4988]: I1008 19:45:36.203620 4988 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c9742f58-7745-4b60-b61d-e864ceecf087-config\") on node \"crc\" DevicePath \"\"" Oct 08 19:45:36 crc kubenswrapper[4988]: I1008 19:45:36.203638 4988 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c9742f58-7745-4b60-b61d-e864ceecf087-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 08 19:45:36 crc kubenswrapper[4988]: I1008 19:45:36.203655 4988 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jdh86\" (UniqueName: \"kubernetes.io/projected/c9742f58-7745-4b60-b61d-e864ceecf087-kube-api-access-jdh86\") on node \"crc\" DevicePath \"\"" Oct 08 19:45:36 crc kubenswrapper[4988]: I1008 19:45:36.203684 4988 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c9742f58-7745-4b60-b61d-e864ceecf087-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 08 19:45:36 crc kubenswrapper[4988]: E1008 19:45:36.203703 4988 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3cd75657be77ae5c79df8af727ac7439f43f1564944848bdd1528f1b162ff4a1\": container with ID starting with 3cd75657be77ae5c79df8af727ac7439f43f1564944848bdd1528f1b162ff4a1 not found: ID does not exist" containerID="3cd75657be77ae5c79df8af727ac7439f43f1564944848bdd1528f1b162ff4a1" Oct 08 19:45:36 crc kubenswrapper[4988]: I1008 19:45:36.203746 4988 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3cd75657be77ae5c79df8af727ac7439f43f1564944848bdd1528f1b162ff4a1"} err="failed to get container status \"3cd75657be77ae5c79df8af727ac7439f43f1564944848bdd1528f1b162ff4a1\": rpc error: code = NotFound desc = could not find container \"3cd75657be77ae5c79df8af727ac7439f43f1564944848bdd1528f1b162ff4a1\": container with ID starting with 3cd75657be77ae5c79df8af727ac7439f43f1564944848bdd1528f1b162ff4a1 not found: ID does not exist" Oct 08 19:45:36 crc kubenswrapper[4988]: I1008 19:45:36.385642 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-fbc86f6b9-gwjfv"] Oct 08 19:45:36 crc kubenswrapper[4988]: I1008 19:45:36.397879 4988 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-fbc86f6b9-gwjfv"] Oct 08 19:45:37 crc kubenswrapper[4988]: I1008 19:45:37.202876 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/swift-proxy-5dfffdbbd6-gzqmd" Oct 08 19:45:37 crc kubenswrapper[4988]: I1008 19:45:37.205723 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/swift-proxy-5dfffdbbd6-gzqmd" Oct 08 19:45:37 crc kubenswrapper[4988]: I1008 19:45:37.260436 4988 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c9742f58-7745-4b60-b61d-e864ceecf087" path="/var/lib/kubelet/pods/c9742f58-7745-4b60-b61d-e864ceecf087/volumes" Oct 08 19:45:39 crc kubenswrapper[4988]: I1008 19:45:39.399994 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/swift-proxy-b8548dfd6-qblss" Oct 08 19:45:39 crc kubenswrapper[4988]: I1008 19:45:39.400730 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/swift-proxy-b8548dfd6-qblss" Oct 08 19:45:39 crc kubenswrapper[4988]: I1008 19:45:39.532154 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/swift-proxy-5dfffdbbd6-gzqmd"] Oct 08 19:45:39 crc kubenswrapper[4988]: I1008 19:45:39.533027 4988 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-proxy-5dfffdbbd6-gzqmd" podUID="7f4d7998-e850-47be-bc8b-b9c2c5f354d0" containerName="proxy-httpd" containerID="cri-o://fc15ca42b349f5f45532359cc4b75f8b060255fe035d5a3e7aa4c423e79de0ea" gracePeriod=30 Oct 08 19:45:39 crc kubenswrapper[4988]: I1008 19:45:39.533173 4988 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-proxy-5dfffdbbd6-gzqmd" podUID="7f4d7998-e850-47be-bc8b-b9c2c5f354d0" containerName="proxy-server" containerID="cri-o://cd0754537c4a649ec0ef88b0c87d2d7aaf840e345d3451b5b177067ddbba6fb4" gracePeriod=30 Oct 08 19:45:40 crc kubenswrapper[4988]: I1008 19:45:40.105133 4988 generic.go:334] "Generic (PLEG): container finished" podID="7f4d7998-e850-47be-bc8b-b9c2c5f354d0" containerID="cd0754537c4a649ec0ef88b0c87d2d7aaf840e345d3451b5b177067ddbba6fb4" exitCode=0 Oct 08 19:45:40 crc kubenswrapper[4988]: I1008 19:45:40.105176 4988 generic.go:334] "Generic (PLEG): container finished" podID="7f4d7998-e850-47be-bc8b-b9c2c5f354d0" containerID="fc15ca42b349f5f45532359cc4b75f8b060255fe035d5a3e7aa4c423e79de0ea" exitCode=0 Oct 08 19:45:40 crc kubenswrapper[4988]: I1008 19:45:40.105200 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-5dfffdbbd6-gzqmd" event={"ID":"7f4d7998-e850-47be-bc8b-b9c2c5f354d0","Type":"ContainerDied","Data":"cd0754537c4a649ec0ef88b0c87d2d7aaf840e345d3451b5b177067ddbba6fb4"} Oct 08 19:45:40 crc kubenswrapper[4988]: I1008 19:45:40.105244 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-5dfffdbbd6-gzqmd" event={"ID":"7f4d7998-e850-47be-bc8b-b9c2c5f354d0","Type":"ContainerDied","Data":"fc15ca42b349f5f45532359cc4b75f8b060255fe035d5a3e7aa4c423e79de0ea"} Oct 08 19:45:40 crc kubenswrapper[4988]: I1008 19:45:40.352709 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-5dfffdbbd6-gzqmd" Oct 08 19:45:40 crc kubenswrapper[4988]: I1008 19:45:40.421614 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7f4d7998-e850-47be-bc8b-b9c2c5f354d0-log-httpd\") pod \"7f4d7998-e850-47be-bc8b-b9c2c5f354d0\" (UID: \"7f4d7998-e850-47be-bc8b-b9c2c5f354d0\") " Oct 08 19:45:40 crc kubenswrapper[4988]: I1008 19:45:40.421669 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/7f4d7998-e850-47be-bc8b-b9c2c5f354d0-etc-swift\") pod \"7f4d7998-e850-47be-bc8b-b9c2c5f354d0\" (UID: \"7f4d7998-e850-47be-bc8b-b9c2c5f354d0\") " Oct 08 19:45:40 crc kubenswrapper[4988]: I1008 19:45:40.421774 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7f4d7998-e850-47be-bc8b-b9c2c5f354d0-config-data\") pod \"7f4d7998-e850-47be-bc8b-b9c2c5f354d0\" (UID: \"7f4d7998-e850-47be-bc8b-b9c2c5f354d0\") " Oct 08 19:45:40 crc kubenswrapper[4988]: I1008 19:45:40.421796 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7f4d7998-e850-47be-bc8b-b9c2c5f354d0-run-httpd\") pod \"7f4d7998-e850-47be-bc8b-b9c2c5f354d0\" (UID: \"7f4d7998-e850-47be-bc8b-b9c2c5f354d0\") " Oct 08 19:45:40 crc kubenswrapper[4988]: I1008 19:45:40.422142 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7f4d7998-e850-47be-bc8b-b9c2c5f354d0-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "7f4d7998-e850-47be-bc8b-b9c2c5f354d0" (UID: "7f4d7998-e850-47be-bc8b-b9c2c5f354d0"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 19:45:40 crc kubenswrapper[4988]: I1008 19:45:40.422534 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7f4d7998-e850-47be-bc8b-b9c2c5f354d0-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "7f4d7998-e850-47be-bc8b-b9c2c5f354d0" (UID: "7f4d7998-e850-47be-bc8b-b9c2c5f354d0"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 19:45:40 crc kubenswrapper[4988]: I1008 19:45:40.422628 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pcngc\" (UniqueName: \"kubernetes.io/projected/7f4d7998-e850-47be-bc8b-b9c2c5f354d0-kube-api-access-pcngc\") pod \"7f4d7998-e850-47be-bc8b-b9c2c5f354d0\" (UID: \"7f4d7998-e850-47be-bc8b-b9c2c5f354d0\") " Oct 08 19:45:40 crc kubenswrapper[4988]: I1008 19:45:40.422736 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7f4d7998-e850-47be-bc8b-b9c2c5f354d0-combined-ca-bundle\") pod \"7f4d7998-e850-47be-bc8b-b9c2c5f354d0\" (UID: \"7f4d7998-e850-47be-bc8b-b9c2c5f354d0\") " Oct 08 19:45:40 crc kubenswrapper[4988]: I1008 19:45:40.424026 4988 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7f4d7998-e850-47be-bc8b-b9c2c5f354d0-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 08 19:45:40 crc kubenswrapper[4988]: I1008 19:45:40.424059 4988 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7f4d7998-e850-47be-bc8b-b9c2c5f354d0-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 08 19:45:40 crc kubenswrapper[4988]: I1008 19:45:40.427667 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7f4d7998-e850-47be-bc8b-b9c2c5f354d0-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "7f4d7998-e850-47be-bc8b-b9c2c5f354d0" (UID: "7f4d7998-e850-47be-bc8b-b9c2c5f354d0"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 19:45:40 crc kubenswrapper[4988]: I1008 19:45:40.436880 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7f4d7998-e850-47be-bc8b-b9c2c5f354d0-kube-api-access-pcngc" (OuterVolumeSpecName: "kube-api-access-pcngc") pod "7f4d7998-e850-47be-bc8b-b9c2c5f354d0" (UID: "7f4d7998-e850-47be-bc8b-b9c2c5f354d0"). InnerVolumeSpecName "kube-api-access-pcngc". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 19:45:40 crc kubenswrapper[4988]: I1008 19:45:40.478502 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7f4d7998-e850-47be-bc8b-b9c2c5f354d0-config-data" (OuterVolumeSpecName: "config-data") pod "7f4d7998-e850-47be-bc8b-b9c2c5f354d0" (UID: "7f4d7998-e850-47be-bc8b-b9c2c5f354d0"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 19:45:40 crc kubenswrapper[4988]: I1008 19:45:40.492374 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7f4d7998-e850-47be-bc8b-b9c2c5f354d0-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "7f4d7998-e850-47be-bc8b-b9c2c5f354d0" (UID: "7f4d7998-e850-47be-bc8b-b9c2c5f354d0"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 19:45:40 crc kubenswrapper[4988]: I1008 19:45:40.527801 4988 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7f4d7998-e850-47be-bc8b-b9c2c5f354d0-config-data\") on node \"crc\" DevicePath \"\"" Oct 08 19:45:40 crc kubenswrapper[4988]: I1008 19:45:40.527874 4988 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pcngc\" (UniqueName: \"kubernetes.io/projected/7f4d7998-e850-47be-bc8b-b9c2c5f354d0-kube-api-access-pcngc\") on node \"crc\" DevicePath \"\"" Oct 08 19:45:40 crc kubenswrapper[4988]: I1008 19:45:40.527902 4988 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7f4d7998-e850-47be-bc8b-b9c2c5f354d0-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 19:45:40 crc kubenswrapper[4988]: I1008 19:45:40.527930 4988 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/7f4d7998-e850-47be-bc8b-b9c2c5f354d0-etc-swift\") on node \"crc\" DevicePath \"\"" Oct 08 19:45:41 crc kubenswrapper[4988]: I1008 19:45:41.119807 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-5dfffdbbd6-gzqmd" event={"ID":"7f4d7998-e850-47be-bc8b-b9c2c5f354d0","Type":"ContainerDied","Data":"e483d4e1a94aaa04403fa6b98cc6a94d3eb0b133fd884d227b0c35bf8b0b7ca6"} Oct 08 19:45:41 crc kubenswrapper[4988]: I1008 19:45:41.119878 4988 scope.go:117] "RemoveContainer" containerID="cd0754537c4a649ec0ef88b0c87d2d7aaf840e345d3451b5b177067ddbba6fb4" Oct 08 19:45:41 crc kubenswrapper[4988]: I1008 19:45:41.119899 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-5dfffdbbd6-gzqmd" Oct 08 19:45:41 crc kubenswrapper[4988]: I1008 19:45:41.139338 4988 scope.go:117] "RemoveContainer" containerID="fc15ca42b349f5f45532359cc4b75f8b060255fe035d5a3e7aa4c423e79de0ea" Oct 08 19:45:41 crc kubenswrapper[4988]: I1008 19:45:41.167361 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/swift-proxy-5dfffdbbd6-gzqmd"] Oct 08 19:45:41 crc kubenswrapper[4988]: I1008 19:45:41.191078 4988 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/swift-proxy-5dfffdbbd6-gzqmd"] Oct 08 19:45:41 crc kubenswrapper[4988]: I1008 19:45:41.252926 4988 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7f4d7998-e850-47be-bc8b-b9c2c5f354d0" path="/var/lib/kubelet/pods/7f4d7998-e850-47be-bc8b-b9c2c5f354d0/volumes" Oct 08 19:45:45 crc kubenswrapper[4988]: I1008 19:45:45.642557 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-create-s2d5m"] Oct 08 19:45:45 crc kubenswrapper[4988]: E1008 19:45:45.643308 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="de3434fc-c4fa-4495-b3f3-7f232decfed9" containerName="swift-ring-rebalance" Oct 08 19:45:45 crc kubenswrapper[4988]: I1008 19:45:45.643320 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="de3434fc-c4fa-4495-b3f3-7f232decfed9" containerName="swift-ring-rebalance" Oct 08 19:45:45 crc kubenswrapper[4988]: E1008 19:45:45.643341 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7f4d7998-e850-47be-bc8b-b9c2c5f354d0" containerName="proxy-httpd" Oct 08 19:45:45 crc kubenswrapper[4988]: I1008 19:45:45.643346 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="7f4d7998-e850-47be-bc8b-b9c2c5f354d0" containerName="proxy-httpd" Oct 08 19:45:45 crc kubenswrapper[4988]: E1008 19:45:45.643359 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7f4d7998-e850-47be-bc8b-b9c2c5f354d0" containerName="proxy-server" Oct 08 19:45:45 crc kubenswrapper[4988]: I1008 19:45:45.643365 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="7f4d7998-e850-47be-bc8b-b9c2c5f354d0" containerName="proxy-server" Oct 08 19:45:45 crc kubenswrapper[4988]: E1008 19:45:45.643396 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c9742f58-7745-4b60-b61d-e864ceecf087" containerName="dnsmasq-dns" Oct 08 19:45:45 crc kubenswrapper[4988]: I1008 19:45:45.643402 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="c9742f58-7745-4b60-b61d-e864ceecf087" containerName="dnsmasq-dns" Oct 08 19:45:45 crc kubenswrapper[4988]: E1008 19:45:45.643415 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c9742f58-7745-4b60-b61d-e864ceecf087" containerName="init" Oct 08 19:45:45 crc kubenswrapper[4988]: I1008 19:45:45.643420 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="c9742f58-7745-4b60-b61d-e864ceecf087" containerName="init" Oct 08 19:45:45 crc kubenswrapper[4988]: I1008 19:45:45.643570 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="c9742f58-7745-4b60-b61d-e864ceecf087" containerName="dnsmasq-dns" Oct 08 19:45:45 crc kubenswrapper[4988]: I1008 19:45:45.643579 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="de3434fc-c4fa-4495-b3f3-7f232decfed9" containerName="swift-ring-rebalance" Oct 08 19:45:45 crc kubenswrapper[4988]: I1008 19:45:45.643590 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="7f4d7998-e850-47be-bc8b-b9c2c5f354d0" containerName="proxy-server" Oct 08 19:45:45 crc kubenswrapper[4988]: I1008 19:45:45.643603 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="7f4d7998-e850-47be-bc8b-b9c2c5f354d0" containerName="proxy-httpd" Oct 08 19:45:45 crc kubenswrapper[4988]: I1008 19:45:45.644130 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-s2d5m" Oct 08 19:45:45 crc kubenswrapper[4988]: I1008 19:45:45.655367 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-s2d5m"] Oct 08 19:45:45 crc kubenswrapper[4988]: I1008 19:45:45.728201 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-69zjz\" (UniqueName: \"kubernetes.io/projected/5681236e-e0dc-46cf-aa72-68e4b26454ec-kube-api-access-69zjz\") pod \"cinder-db-create-s2d5m\" (UID: \"5681236e-e0dc-46cf-aa72-68e4b26454ec\") " pod="openstack/cinder-db-create-s2d5m" Oct 08 19:45:45 crc kubenswrapper[4988]: I1008 19:45:45.829993 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-69zjz\" (UniqueName: \"kubernetes.io/projected/5681236e-e0dc-46cf-aa72-68e4b26454ec-kube-api-access-69zjz\") pod \"cinder-db-create-s2d5m\" (UID: \"5681236e-e0dc-46cf-aa72-68e4b26454ec\") " pod="openstack/cinder-db-create-s2d5m" Oct 08 19:45:45 crc kubenswrapper[4988]: I1008 19:45:45.850232 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-69zjz\" (UniqueName: \"kubernetes.io/projected/5681236e-e0dc-46cf-aa72-68e4b26454ec-kube-api-access-69zjz\") pod \"cinder-db-create-s2d5m\" (UID: \"5681236e-e0dc-46cf-aa72-68e4b26454ec\") " pod="openstack/cinder-db-create-s2d5m" Oct 08 19:45:45 crc kubenswrapper[4988]: I1008 19:45:45.973182 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-s2d5m" Oct 08 19:45:46 crc kubenswrapper[4988]: I1008 19:45:46.493306 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-s2d5m"] Oct 08 19:45:47 crc kubenswrapper[4988]: I1008 19:45:47.198252 4988 generic.go:334] "Generic (PLEG): container finished" podID="5681236e-e0dc-46cf-aa72-68e4b26454ec" containerID="2626533fad751ded722c45fcae26aab38c2d0861daefd298e02888a2a0a7b066" exitCode=0 Oct 08 19:45:47 crc kubenswrapper[4988]: I1008 19:45:47.198326 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-s2d5m" event={"ID":"5681236e-e0dc-46cf-aa72-68e4b26454ec","Type":"ContainerDied","Data":"2626533fad751ded722c45fcae26aab38c2d0861daefd298e02888a2a0a7b066"} Oct 08 19:45:47 crc kubenswrapper[4988]: I1008 19:45:47.198656 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-s2d5m" event={"ID":"5681236e-e0dc-46cf-aa72-68e4b26454ec","Type":"ContainerStarted","Data":"1f639460e84586dfc10a4252c1c9f1dcee83ade68ee1b7b282ef92133bfebf20"} Oct 08 19:45:48 crc kubenswrapper[4988]: I1008 19:45:48.603334 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-s2d5m" Oct 08 19:45:48 crc kubenswrapper[4988]: I1008 19:45:48.698549 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-69zjz\" (UniqueName: \"kubernetes.io/projected/5681236e-e0dc-46cf-aa72-68e4b26454ec-kube-api-access-69zjz\") pod \"5681236e-e0dc-46cf-aa72-68e4b26454ec\" (UID: \"5681236e-e0dc-46cf-aa72-68e4b26454ec\") " Oct 08 19:45:48 crc kubenswrapper[4988]: I1008 19:45:48.703677 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5681236e-e0dc-46cf-aa72-68e4b26454ec-kube-api-access-69zjz" (OuterVolumeSpecName: "kube-api-access-69zjz") pod "5681236e-e0dc-46cf-aa72-68e4b26454ec" (UID: "5681236e-e0dc-46cf-aa72-68e4b26454ec"). InnerVolumeSpecName "kube-api-access-69zjz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 19:45:48 crc kubenswrapper[4988]: I1008 19:45:48.800737 4988 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-69zjz\" (UniqueName: \"kubernetes.io/projected/5681236e-e0dc-46cf-aa72-68e4b26454ec-kube-api-access-69zjz\") on node \"crc\" DevicePath \"\"" Oct 08 19:45:49 crc kubenswrapper[4988]: I1008 19:45:49.223530 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-s2d5m" event={"ID":"5681236e-e0dc-46cf-aa72-68e4b26454ec","Type":"ContainerDied","Data":"1f639460e84586dfc10a4252c1c9f1dcee83ade68ee1b7b282ef92133bfebf20"} Oct 08 19:45:49 crc kubenswrapper[4988]: I1008 19:45:49.223886 4988 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1f639460e84586dfc10a4252c1c9f1dcee83ade68ee1b7b282ef92133bfebf20" Oct 08 19:45:49 crc kubenswrapper[4988]: I1008 19:45:49.223619 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-s2d5m" Oct 08 19:45:53 crc kubenswrapper[4988]: I1008 19:45:53.337970 4988 patch_prober.go:28] interesting pod/machine-config-daemon-nm54f container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 19:45:53 crc kubenswrapper[4988]: I1008 19:45:53.338369 4988 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 19:45:55 crc kubenswrapper[4988]: I1008 19:45:55.744008 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-17fa-account-create-rpcv2"] Oct 08 19:45:55 crc kubenswrapper[4988]: E1008 19:45:55.744905 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5681236e-e0dc-46cf-aa72-68e4b26454ec" containerName="mariadb-database-create" Oct 08 19:45:55 crc kubenswrapper[4988]: I1008 19:45:55.744944 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="5681236e-e0dc-46cf-aa72-68e4b26454ec" containerName="mariadb-database-create" Oct 08 19:45:55 crc kubenswrapper[4988]: I1008 19:45:55.745325 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="5681236e-e0dc-46cf-aa72-68e4b26454ec" containerName="mariadb-database-create" Oct 08 19:45:55 crc kubenswrapper[4988]: I1008 19:45:55.746432 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-17fa-account-create-rpcv2" Oct 08 19:45:55 crc kubenswrapper[4988]: I1008 19:45:55.748893 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-db-secret" Oct 08 19:45:55 crc kubenswrapper[4988]: I1008 19:45:55.756490 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-17fa-account-create-rpcv2"] Oct 08 19:45:55 crc kubenswrapper[4988]: I1008 19:45:55.840454 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x4b8j\" (UniqueName: \"kubernetes.io/projected/8f2a9de4-4576-416f-99cd-2fe38b4700db-kube-api-access-x4b8j\") pod \"cinder-17fa-account-create-rpcv2\" (UID: \"8f2a9de4-4576-416f-99cd-2fe38b4700db\") " pod="openstack/cinder-17fa-account-create-rpcv2" Oct 08 19:45:55 crc kubenswrapper[4988]: I1008 19:45:55.942716 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x4b8j\" (UniqueName: \"kubernetes.io/projected/8f2a9de4-4576-416f-99cd-2fe38b4700db-kube-api-access-x4b8j\") pod \"cinder-17fa-account-create-rpcv2\" (UID: \"8f2a9de4-4576-416f-99cd-2fe38b4700db\") " pod="openstack/cinder-17fa-account-create-rpcv2" Oct 08 19:45:55 crc kubenswrapper[4988]: I1008 19:45:55.967791 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x4b8j\" (UniqueName: \"kubernetes.io/projected/8f2a9de4-4576-416f-99cd-2fe38b4700db-kube-api-access-x4b8j\") pod \"cinder-17fa-account-create-rpcv2\" (UID: \"8f2a9de4-4576-416f-99cd-2fe38b4700db\") " pod="openstack/cinder-17fa-account-create-rpcv2" Oct 08 19:45:56 crc kubenswrapper[4988]: I1008 19:45:56.080326 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-17fa-account-create-rpcv2" Oct 08 19:45:56 crc kubenswrapper[4988]: I1008 19:45:56.529325 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-17fa-account-create-rpcv2"] Oct 08 19:45:57 crc kubenswrapper[4988]: I1008 19:45:57.324638 4988 generic.go:334] "Generic (PLEG): container finished" podID="8f2a9de4-4576-416f-99cd-2fe38b4700db" containerID="bf77b9c29c9c464a36804f797ad4a23d3ee5ddecf4d550cbf71fc2fbeac45e98" exitCode=0 Oct 08 19:45:57 crc kubenswrapper[4988]: I1008 19:45:57.324734 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-17fa-account-create-rpcv2" event={"ID":"8f2a9de4-4576-416f-99cd-2fe38b4700db","Type":"ContainerDied","Data":"bf77b9c29c9c464a36804f797ad4a23d3ee5ddecf4d550cbf71fc2fbeac45e98"} Oct 08 19:45:57 crc kubenswrapper[4988]: I1008 19:45:57.325065 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-17fa-account-create-rpcv2" event={"ID":"8f2a9de4-4576-416f-99cd-2fe38b4700db","Type":"ContainerStarted","Data":"084c4be38bf9ef0317da17998c6f677f6d8956138cb28d36776fb086701f4d30"} Oct 08 19:45:58 crc kubenswrapper[4988]: I1008 19:45:58.702817 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-17fa-account-create-rpcv2" Oct 08 19:45:58 crc kubenswrapper[4988]: I1008 19:45:58.801833 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x4b8j\" (UniqueName: \"kubernetes.io/projected/8f2a9de4-4576-416f-99cd-2fe38b4700db-kube-api-access-x4b8j\") pod \"8f2a9de4-4576-416f-99cd-2fe38b4700db\" (UID: \"8f2a9de4-4576-416f-99cd-2fe38b4700db\") " Oct 08 19:45:58 crc kubenswrapper[4988]: I1008 19:45:58.808183 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f2a9de4-4576-416f-99cd-2fe38b4700db-kube-api-access-x4b8j" (OuterVolumeSpecName: "kube-api-access-x4b8j") pod "8f2a9de4-4576-416f-99cd-2fe38b4700db" (UID: "8f2a9de4-4576-416f-99cd-2fe38b4700db"). InnerVolumeSpecName "kube-api-access-x4b8j". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 19:45:58 crc kubenswrapper[4988]: I1008 19:45:58.911780 4988 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x4b8j\" (UniqueName: \"kubernetes.io/projected/8f2a9de4-4576-416f-99cd-2fe38b4700db-kube-api-access-x4b8j\") on node \"crc\" DevicePath \"\"" Oct 08 19:45:59 crc kubenswrapper[4988]: I1008 19:45:59.361113 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-17fa-account-create-rpcv2" event={"ID":"8f2a9de4-4576-416f-99cd-2fe38b4700db","Type":"ContainerDied","Data":"084c4be38bf9ef0317da17998c6f677f6d8956138cb28d36776fb086701f4d30"} Oct 08 19:45:59 crc kubenswrapper[4988]: I1008 19:45:59.361176 4988 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="084c4be38bf9ef0317da17998c6f677f6d8956138cb28d36776fb086701f4d30" Oct 08 19:45:59 crc kubenswrapper[4988]: I1008 19:45:59.361177 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-17fa-account-create-rpcv2" Oct 08 19:46:01 crc kubenswrapper[4988]: I1008 19:46:01.100134 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-sync-sfpjb"] Oct 08 19:46:01 crc kubenswrapper[4988]: E1008 19:46:01.100864 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8f2a9de4-4576-416f-99cd-2fe38b4700db" containerName="mariadb-account-create" Oct 08 19:46:01 crc kubenswrapper[4988]: I1008 19:46:01.100879 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="8f2a9de4-4576-416f-99cd-2fe38b4700db" containerName="mariadb-account-create" Oct 08 19:46:01 crc kubenswrapper[4988]: I1008 19:46:01.101058 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="8f2a9de4-4576-416f-99cd-2fe38b4700db" containerName="mariadb-account-create" Oct 08 19:46:01 crc kubenswrapper[4988]: I1008 19:46:01.101657 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-sfpjb" Oct 08 19:46:01 crc kubenswrapper[4988]: I1008 19:46:01.110350 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-wzqmq" Oct 08 19:46:01 crc kubenswrapper[4988]: I1008 19:46:01.110586 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Oct 08 19:46:01 crc kubenswrapper[4988]: I1008 19:46:01.110683 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Oct 08 19:46:01 crc kubenswrapper[4988]: I1008 19:46:01.114945 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-sfpjb"] Oct 08 19:46:01 crc kubenswrapper[4988]: I1008 19:46:01.157212 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/ec8f6c64-9860-4a50-bd16-972351608049-etc-machine-id\") pod \"cinder-db-sync-sfpjb\" (UID: \"ec8f6c64-9860-4a50-bd16-972351608049\") " pod="openstack/cinder-db-sync-sfpjb" Oct 08 19:46:01 crc kubenswrapper[4988]: I1008 19:46:01.157288 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ec8f6c64-9860-4a50-bd16-972351608049-scripts\") pod \"cinder-db-sync-sfpjb\" (UID: \"ec8f6c64-9860-4a50-bd16-972351608049\") " pod="openstack/cinder-db-sync-sfpjb" Oct 08 19:46:01 crc kubenswrapper[4988]: I1008 19:46:01.157346 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ec8f6c64-9860-4a50-bd16-972351608049-combined-ca-bundle\") pod \"cinder-db-sync-sfpjb\" (UID: \"ec8f6c64-9860-4a50-bd16-972351608049\") " pod="openstack/cinder-db-sync-sfpjb" Oct 08 19:46:01 crc kubenswrapper[4988]: I1008 19:46:01.157377 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ec8f6c64-9860-4a50-bd16-972351608049-config-data\") pod \"cinder-db-sync-sfpjb\" (UID: \"ec8f6c64-9860-4a50-bd16-972351608049\") " pod="openstack/cinder-db-sync-sfpjb" Oct 08 19:46:01 crc kubenswrapper[4988]: I1008 19:46:01.157426 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/ec8f6c64-9860-4a50-bd16-972351608049-db-sync-config-data\") pod \"cinder-db-sync-sfpjb\" (UID: \"ec8f6c64-9860-4a50-bd16-972351608049\") " pod="openstack/cinder-db-sync-sfpjb" Oct 08 19:46:01 crc kubenswrapper[4988]: I1008 19:46:01.157463 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bbwwt\" (UniqueName: \"kubernetes.io/projected/ec8f6c64-9860-4a50-bd16-972351608049-kube-api-access-bbwwt\") pod \"cinder-db-sync-sfpjb\" (UID: \"ec8f6c64-9860-4a50-bd16-972351608049\") " pod="openstack/cinder-db-sync-sfpjb" Oct 08 19:46:01 crc kubenswrapper[4988]: I1008 19:46:01.259342 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ec8f6c64-9860-4a50-bd16-972351608049-config-data\") pod \"cinder-db-sync-sfpjb\" (UID: \"ec8f6c64-9860-4a50-bd16-972351608049\") " pod="openstack/cinder-db-sync-sfpjb" Oct 08 19:46:01 crc kubenswrapper[4988]: I1008 19:46:01.259424 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/ec8f6c64-9860-4a50-bd16-972351608049-db-sync-config-data\") pod \"cinder-db-sync-sfpjb\" (UID: \"ec8f6c64-9860-4a50-bd16-972351608049\") " pod="openstack/cinder-db-sync-sfpjb" Oct 08 19:46:01 crc kubenswrapper[4988]: I1008 19:46:01.259470 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bbwwt\" (UniqueName: \"kubernetes.io/projected/ec8f6c64-9860-4a50-bd16-972351608049-kube-api-access-bbwwt\") pod \"cinder-db-sync-sfpjb\" (UID: \"ec8f6c64-9860-4a50-bd16-972351608049\") " pod="openstack/cinder-db-sync-sfpjb" Oct 08 19:46:01 crc kubenswrapper[4988]: I1008 19:46:01.259506 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/ec8f6c64-9860-4a50-bd16-972351608049-etc-machine-id\") pod \"cinder-db-sync-sfpjb\" (UID: \"ec8f6c64-9860-4a50-bd16-972351608049\") " pod="openstack/cinder-db-sync-sfpjb" Oct 08 19:46:01 crc kubenswrapper[4988]: I1008 19:46:01.259574 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ec8f6c64-9860-4a50-bd16-972351608049-scripts\") pod \"cinder-db-sync-sfpjb\" (UID: \"ec8f6c64-9860-4a50-bd16-972351608049\") " pod="openstack/cinder-db-sync-sfpjb" Oct 08 19:46:01 crc kubenswrapper[4988]: I1008 19:46:01.259631 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ec8f6c64-9860-4a50-bd16-972351608049-combined-ca-bundle\") pod \"cinder-db-sync-sfpjb\" (UID: \"ec8f6c64-9860-4a50-bd16-972351608049\") " pod="openstack/cinder-db-sync-sfpjb" Oct 08 19:46:01 crc kubenswrapper[4988]: I1008 19:46:01.259738 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/ec8f6c64-9860-4a50-bd16-972351608049-etc-machine-id\") pod \"cinder-db-sync-sfpjb\" (UID: \"ec8f6c64-9860-4a50-bd16-972351608049\") " pod="openstack/cinder-db-sync-sfpjb" Oct 08 19:46:01 crc kubenswrapper[4988]: I1008 19:46:01.265826 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ec8f6c64-9860-4a50-bd16-972351608049-combined-ca-bundle\") pod \"cinder-db-sync-sfpjb\" (UID: \"ec8f6c64-9860-4a50-bd16-972351608049\") " pod="openstack/cinder-db-sync-sfpjb" Oct 08 19:46:01 crc kubenswrapper[4988]: I1008 19:46:01.265922 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/ec8f6c64-9860-4a50-bd16-972351608049-db-sync-config-data\") pod \"cinder-db-sync-sfpjb\" (UID: \"ec8f6c64-9860-4a50-bd16-972351608049\") " pod="openstack/cinder-db-sync-sfpjb" Oct 08 19:46:01 crc kubenswrapper[4988]: I1008 19:46:01.268279 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ec8f6c64-9860-4a50-bd16-972351608049-config-data\") pod \"cinder-db-sync-sfpjb\" (UID: \"ec8f6c64-9860-4a50-bd16-972351608049\") " pod="openstack/cinder-db-sync-sfpjb" Oct 08 19:46:01 crc kubenswrapper[4988]: I1008 19:46:01.269090 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ec8f6c64-9860-4a50-bd16-972351608049-scripts\") pod \"cinder-db-sync-sfpjb\" (UID: \"ec8f6c64-9860-4a50-bd16-972351608049\") " pod="openstack/cinder-db-sync-sfpjb" Oct 08 19:46:01 crc kubenswrapper[4988]: I1008 19:46:01.281943 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bbwwt\" (UniqueName: \"kubernetes.io/projected/ec8f6c64-9860-4a50-bd16-972351608049-kube-api-access-bbwwt\") pod \"cinder-db-sync-sfpjb\" (UID: \"ec8f6c64-9860-4a50-bd16-972351608049\") " pod="openstack/cinder-db-sync-sfpjb" Oct 08 19:46:01 crc kubenswrapper[4988]: I1008 19:46:01.430615 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-sfpjb" Oct 08 19:46:01 crc kubenswrapper[4988]: I1008 19:46:01.943919 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-sfpjb"] Oct 08 19:46:02 crc kubenswrapper[4988]: I1008 19:46:02.393708 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-sfpjb" event={"ID":"ec8f6c64-9860-4a50-bd16-972351608049","Type":"ContainerStarted","Data":"f4066f8ecd066e6220efc6da9946529cf807eff1cc738cb852733eac94ec8224"} Oct 08 19:46:03 crc kubenswrapper[4988]: I1008 19:46:03.412637 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-sfpjb" event={"ID":"ec8f6c64-9860-4a50-bd16-972351608049","Type":"ContainerStarted","Data":"e539312cc13f1978d1df1e563814b231de949749a70469f6361aa41906cadd6c"} Oct 08 19:46:03 crc kubenswrapper[4988]: I1008 19:46:03.454614 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-db-sync-sfpjb" podStartSLOduration=2.4545786290000002 podStartE2EDuration="2.454578629s" podCreationTimestamp="2025-10-08 19:46:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 19:46:03.447144454 +0000 UTC m=+5708.896987264" watchObservedRunningTime="2025-10-08 19:46:03.454578629 +0000 UTC m=+5708.904421439" Oct 08 19:46:03 crc kubenswrapper[4988]: I1008 19:46:03.649504 4988 scope.go:117] "RemoveContainer" containerID="9ddccb0f49cf424fcbc1f7bf2729999a66d10acdc7d77d5760f1c39fa7f1fd0e" Oct 08 19:46:03 crc kubenswrapper[4988]: I1008 19:46:03.678756 4988 scope.go:117] "RemoveContainer" containerID="d3f24e4ba303a1481781dd4658f349d18349f68f11a5f157594e62045a1fcf90" Oct 08 19:46:03 crc kubenswrapper[4988]: I1008 19:46:03.749093 4988 scope.go:117] "RemoveContainer" containerID="0a1aef9772ec4c79f7c96f4d80291f1758dec92998f6d480f86c70553420632e" Oct 08 19:46:05 crc kubenswrapper[4988]: I1008 19:46:05.439815 4988 generic.go:334] "Generic (PLEG): container finished" podID="ec8f6c64-9860-4a50-bd16-972351608049" containerID="e539312cc13f1978d1df1e563814b231de949749a70469f6361aa41906cadd6c" exitCode=0 Oct 08 19:46:05 crc kubenswrapper[4988]: I1008 19:46:05.439890 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-sfpjb" event={"ID":"ec8f6c64-9860-4a50-bd16-972351608049","Type":"ContainerDied","Data":"e539312cc13f1978d1df1e563814b231de949749a70469f6361aa41906cadd6c"} Oct 08 19:46:06 crc kubenswrapper[4988]: I1008 19:46:06.822090 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-sfpjb" Oct 08 19:46:06 crc kubenswrapper[4988]: I1008 19:46:06.872882 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/ec8f6c64-9860-4a50-bd16-972351608049-etc-machine-id\") pod \"ec8f6c64-9860-4a50-bd16-972351608049\" (UID: \"ec8f6c64-9860-4a50-bd16-972351608049\") " Oct 08 19:46:06 crc kubenswrapper[4988]: I1008 19:46:06.872984 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/ec8f6c64-9860-4a50-bd16-972351608049-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "ec8f6c64-9860-4a50-bd16-972351608049" (UID: "ec8f6c64-9860-4a50-bd16-972351608049"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 08 19:46:06 crc kubenswrapper[4988]: I1008 19:46:06.873052 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ec8f6c64-9860-4a50-bd16-972351608049-combined-ca-bundle\") pod \"ec8f6c64-9860-4a50-bd16-972351608049\" (UID: \"ec8f6c64-9860-4a50-bd16-972351608049\") " Oct 08 19:46:06 crc kubenswrapper[4988]: I1008 19:46:06.873106 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/ec8f6c64-9860-4a50-bd16-972351608049-db-sync-config-data\") pod \"ec8f6c64-9860-4a50-bd16-972351608049\" (UID: \"ec8f6c64-9860-4a50-bd16-972351608049\") " Oct 08 19:46:06 crc kubenswrapper[4988]: I1008 19:46:06.873151 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ec8f6c64-9860-4a50-bd16-972351608049-scripts\") pod \"ec8f6c64-9860-4a50-bd16-972351608049\" (UID: \"ec8f6c64-9860-4a50-bd16-972351608049\") " Oct 08 19:46:06 crc kubenswrapper[4988]: I1008 19:46:06.873199 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ec8f6c64-9860-4a50-bd16-972351608049-config-data\") pod \"ec8f6c64-9860-4a50-bd16-972351608049\" (UID: \"ec8f6c64-9860-4a50-bd16-972351608049\") " Oct 08 19:46:06 crc kubenswrapper[4988]: I1008 19:46:06.873242 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bbwwt\" (UniqueName: \"kubernetes.io/projected/ec8f6c64-9860-4a50-bd16-972351608049-kube-api-access-bbwwt\") pod \"ec8f6c64-9860-4a50-bd16-972351608049\" (UID: \"ec8f6c64-9860-4a50-bd16-972351608049\") " Oct 08 19:46:06 crc kubenswrapper[4988]: I1008 19:46:06.873633 4988 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/ec8f6c64-9860-4a50-bd16-972351608049-etc-machine-id\") on node \"crc\" DevicePath \"\"" Oct 08 19:46:06 crc kubenswrapper[4988]: I1008 19:46:06.879019 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ec8f6c64-9860-4a50-bd16-972351608049-scripts" (OuterVolumeSpecName: "scripts") pod "ec8f6c64-9860-4a50-bd16-972351608049" (UID: "ec8f6c64-9860-4a50-bd16-972351608049"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 19:46:06 crc kubenswrapper[4988]: I1008 19:46:06.879323 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ec8f6c64-9860-4a50-bd16-972351608049-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "ec8f6c64-9860-4a50-bd16-972351608049" (UID: "ec8f6c64-9860-4a50-bd16-972351608049"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 19:46:06 crc kubenswrapper[4988]: I1008 19:46:06.879766 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ec8f6c64-9860-4a50-bd16-972351608049-kube-api-access-bbwwt" (OuterVolumeSpecName: "kube-api-access-bbwwt") pod "ec8f6c64-9860-4a50-bd16-972351608049" (UID: "ec8f6c64-9860-4a50-bd16-972351608049"). InnerVolumeSpecName "kube-api-access-bbwwt". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 19:46:06 crc kubenswrapper[4988]: I1008 19:46:06.901712 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ec8f6c64-9860-4a50-bd16-972351608049-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ec8f6c64-9860-4a50-bd16-972351608049" (UID: "ec8f6c64-9860-4a50-bd16-972351608049"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 19:46:06 crc kubenswrapper[4988]: I1008 19:46:06.926560 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ec8f6c64-9860-4a50-bd16-972351608049-config-data" (OuterVolumeSpecName: "config-data") pod "ec8f6c64-9860-4a50-bd16-972351608049" (UID: "ec8f6c64-9860-4a50-bd16-972351608049"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 19:46:06 crc kubenswrapper[4988]: I1008 19:46:06.975446 4988 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ec8f6c64-9860-4a50-bd16-972351608049-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 19:46:06 crc kubenswrapper[4988]: I1008 19:46:06.975477 4988 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/ec8f6c64-9860-4a50-bd16-972351608049-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Oct 08 19:46:06 crc kubenswrapper[4988]: I1008 19:46:06.975488 4988 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ec8f6c64-9860-4a50-bd16-972351608049-scripts\") on node \"crc\" DevicePath \"\"" Oct 08 19:46:06 crc kubenswrapper[4988]: I1008 19:46:06.975497 4988 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ec8f6c64-9860-4a50-bd16-972351608049-config-data\") on node \"crc\" DevicePath \"\"" Oct 08 19:46:06 crc kubenswrapper[4988]: I1008 19:46:06.975506 4988 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bbwwt\" (UniqueName: \"kubernetes.io/projected/ec8f6c64-9860-4a50-bd16-972351608049-kube-api-access-bbwwt\") on node \"crc\" DevicePath \"\"" Oct 08 19:46:07 crc kubenswrapper[4988]: I1008 19:46:07.468841 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-sfpjb" event={"ID":"ec8f6c64-9860-4a50-bd16-972351608049","Type":"ContainerDied","Data":"f4066f8ecd066e6220efc6da9946529cf807eff1cc738cb852733eac94ec8224"} Oct 08 19:46:07 crc kubenswrapper[4988]: I1008 19:46:07.468902 4988 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f4066f8ecd066e6220efc6da9946529cf807eff1cc738cb852733eac94ec8224" Oct 08 19:46:07 crc kubenswrapper[4988]: I1008 19:46:07.468965 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-sfpjb" Oct 08 19:46:07 crc kubenswrapper[4988]: I1008 19:46:07.886309 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-56f9cf6987-4865b"] Oct 08 19:46:07 crc kubenswrapper[4988]: E1008 19:46:07.886741 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ec8f6c64-9860-4a50-bd16-972351608049" containerName="cinder-db-sync" Oct 08 19:46:07 crc kubenswrapper[4988]: I1008 19:46:07.886759 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="ec8f6c64-9860-4a50-bd16-972351608049" containerName="cinder-db-sync" Oct 08 19:46:07 crc kubenswrapper[4988]: I1008 19:46:07.886984 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="ec8f6c64-9860-4a50-bd16-972351608049" containerName="cinder-db-sync" Oct 08 19:46:07 crc kubenswrapper[4988]: I1008 19:46:07.887929 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-56f9cf6987-4865b" Oct 08 19:46:07 crc kubenswrapper[4988]: I1008 19:46:07.941250 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-56f9cf6987-4865b"] Oct 08 19:46:07 crc kubenswrapper[4988]: I1008 19:46:07.999357 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d4873fea-35df-48df-856f-9d4c61f4627a-config\") pod \"dnsmasq-dns-56f9cf6987-4865b\" (UID: \"d4873fea-35df-48df-856f-9d4c61f4627a\") " pod="openstack/dnsmasq-dns-56f9cf6987-4865b" Oct 08 19:46:07 crc kubenswrapper[4988]: I1008 19:46:07.999850 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d4873fea-35df-48df-856f-9d4c61f4627a-ovsdbserver-sb\") pod \"dnsmasq-dns-56f9cf6987-4865b\" (UID: \"d4873fea-35df-48df-856f-9d4c61f4627a\") " pod="openstack/dnsmasq-dns-56f9cf6987-4865b" Oct 08 19:46:07 crc kubenswrapper[4988]: I1008 19:46:07.999900 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d4873fea-35df-48df-856f-9d4c61f4627a-dns-svc\") pod \"dnsmasq-dns-56f9cf6987-4865b\" (UID: \"d4873fea-35df-48df-856f-9d4c61f4627a\") " pod="openstack/dnsmasq-dns-56f9cf6987-4865b" Oct 08 19:46:07 crc kubenswrapper[4988]: I1008 19:46:07.999919 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d4873fea-35df-48df-856f-9d4c61f4627a-ovsdbserver-nb\") pod \"dnsmasq-dns-56f9cf6987-4865b\" (UID: \"d4873fea-35df-48df-856f-9d4c61f4627a\") " pod="openstack/dnsmasq-dns-56f9cf6987-4865b" Oct 08 19:46:08 crc kubenswrapper[4988]: I1008 19:46:07.999949 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m2czm\" (UniqueName: \"kubernetes.io/projected/d4873fea-35df-48df-856f-9d4c61f4627a-kube-api-access-m2czm\") pod \"dnsmasq-dns-56f9cf6987-4865b\" (UID: \"d4873fea-35df-48df-856f-9d4c61f4627a\") " pod="openstack/dnsmasq-dns-56f9cf6987-4865b" Oct 08 19:46:08 crc kubenswrapper[4988]: I1008 19:46:08.039454 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Oct 08 19:46:08 crc kubenswrapper[4988]: I1008 19:46:08.045478 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 08 19:46:08 crc kubenswrapper[4988]: I1008 19:46:08.047847 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Oct 08 19:46:08 crc kubenswrapper[4988]: I1008 19:46:08.047959 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-wzqmq" Oct 08 19:46:08 crc kubenswrapper[4988]: I1008 19:46:08.048076 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Oct 08 19:46:08 crc kubenswrapper[4988]: I1008 19:46:08.048206 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Oct 08 19:46:08 crc kubenswrapper[4988]: I1008 19:46:08.050336 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Oct 08 19:46:08 crc kubenswrapper[4988]: I1008 19:46:08.102208 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d4873fea-35df-48df-856f-9d4c61f4627a-ovsdbserver-sb\") pod \"dnsmasq-dns-56f9cf6987-4865b\" (UID: \"d4873fea-35df-48df-856f-9d4c61f4627a\") " pod="openstack/dnsmasq-dns-56f9cf6987-4865b" Oct 08 19:46:08 crc kubenswrapper[4988]: I1008 19:46:08.102254 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ff18525a-5969-4ae8-be54-ae79a7c72821-scripts\") pod \"cinder-api-0\" (UID: \"ff18525a-5969-4ae8-be54-ae79a7c72821\") " pod="openstack/cinder-api-0" Oct 08 19:46:08 crc kubenswrapper[4988]: I1008 19:46:08.102320 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d4873fea-35df-48df-856f-9d4c61f4627a-dns-svc\") pod \"dnsmasq-dns-56f9cf6987-4865b\" (UID: \"d4873fea-35df-48df-856f-9d4c61f4627a\") " pod="openstack/dnsmasq-dns-56f9cf6987-4865b" Oct 08 19:46:08 crc kubenswrapper[4988]: I1008 19:46:08.102336 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d4873fea-35df-48df-856f-9d4c61f4627a-ovsdbserver-nb\") pod \"dnsmasq-dns-56f9cf6987-4865b\" (UID: \"d4873fea-35df-48df-856f-9d4c61f4627a\") " pod="openstack/dnsmasq-dns-56f9cf6987-4865b" Oct 08 19:46:08 crc kubenswrapper[4988]: I1008 19:46:08.102441 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m2czm\" (UniqueName: \"kubernetes.io/projected/d4873fea-35df-48df-856f-9d4c61f4627a-kube-api-access-m2czm\") pod \"dnsmasq-dns-56f9cf6987-4865b\" (UID: \"d4873fea-35df-48df-856f-9d4c61f4627a\") " pod="openstack/dnsmasq-dns-56f9cf6987-4865b" Oct 08 19:46:08 crc kubenswrapper[4988]: I1008 19:46:08.102745 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d4873fea-35df-48df-856f-9d4c61f4627a-config\") pod \"dnsmasq-dns-56f9cf6987-4865b\" (UID: \"d4873fea-35df-48df-856f-9d4c61f4627a\") " pod="openstack/dnsmasq-dns-56f9cf6987-4865b" Oct 08 19:46:08 crc kubenswrapper[4988]: I1008 19:46:08.103291 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d4873fea-35df-48df-856f-9d4c61f4627a-dns-svc\") pod \"dnsmasq-dns-56f9cf6987-4865b\" (UID: \"d4873fea-35df-48df-856f-9d4c61f4627a\") " pod="openstack/dnsmasq-dns-56f9cf6987-4865b" Oct 08 19:46:08 crc kubenswrapper[4988]: I1008 19:46:08.103312 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ff18525a-5969-4ae8-be54-ae79a7c72821-logs\") pod \"cinder-api-0\" (UID: \"ff18525a-5969-4ae8-be54-ae79a7c72821\") " pod="openstack/cinder-api-0" Oct 08 19:46:08 crc kubenswrapper[4988]: I1008 19:46:08.103263 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d4873fea-35df-48df-856f-9d4c61f4627a-ovsdbserver-nb\") pod \"dnsmasq-dns-56f9cf6987-4865b\" (UID: \"d4873fea-35df-48df-856f-9d4c61f4627a\") " pod="openstack/dnsmasq-dns-56f9cf6987-4865b" Oct 08 19:46:08 crc kubenswrapper[4988]: I1008 19:46:08.103336 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d4873fea-35df-48df-856f-9d4c61f4627a-ovsdbserver-sb\") pod \"dnsmasq-dns-56f9cf6987-4865b\" (UID: \"d4873fea-35df-48df-856f-9d4c61f4627a\") " pod="openstack/dnsmasq-dns-56f9cf6987-4865b" Oct 08 19:46:08 crc kubenswrapper[4988]: I1008 19:46:08.103549 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x7bgb\" (UniqueName: \"kubernetes.io/projected/ff18525a-5969-4ae8-be54-ae79a7c72821-kube-api-access-x7bgb\") pod \"cinder-api-0\" (UID: \"ff18525a-5969-4ae8-be54-ae79a7c72821\") " pod="openstack/cinder-api-0" Oct 08 19:46:08 crc kubenswrapper[4988]: I1008 19:46:08.103611 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/ff18525a-5969-4ae8-be54-ae79a7c72821-config-data-custom\") pod \"cinder-api-0\" (UID: \"ff18525a-5969-4ae8-be54-ae79a7c72821\") " pod="openstack/cinder-api-0" Oct 08 19:46:08 crc kubenswrapper[4988]: I1008 19:46:08.103664 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ff18525a-5969-4ae8-be54-ae79a7c72821-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"ff18525a-5969-4ae8-be54-ae79a7c72821\") " pod="openstack/cinder-api-0" Oct 08 19:46:08 crc kubenswrapper[4988]: I1008 19:46:08.103688 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ff18525a-5969-4ae8-be54-ae79a7c72821-config-data\") pod \"cinder-api-0\" (UID: \"ff18525a-5969-4ae8-be54-ae79a7c72821\") " pod="openstack/cinder-api-0" Oct 08 19:46:08 crc kubenswrapper[4988]: I1008 19:46:08.103772 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/ff18525a-5969-4ae8-be54-ae79a7c72821-etc-machine-id\") pod \"cinder-api-0\" (UID: \"ff18525a-5969-4ae8-be54-ae79a7c72821\") " pod="openstack/cinder-api-0" Oct 08 19:46:08 crc kubenswrapper[4988]: I1008 19:46:08.103883 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d4873fea-35df-48df-856f-9d4c61f4627a-config\") pod \"dnsmasq-dns-56f9cf6987-4865b\" (UID: \"d4873fea-35df-48df-856f-9d4c61f4627a\") " pod="openstack/dnsmasq-dns-56f9cf6987-4865b" Oct 08 19:46:08 crc kubenswrapper[4988]: I1008 19:46:08.127478 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m2czm\" (UniqueName: \"kubernetes.io/projected/d4873fea-35df-48df-856f-9d4c61f4627a-kube-api-access-m2czm\") pod \"dnsmasq-dns-56f9cf6987-4865b\" (UID: \"d4873fea-35df-48df-856f-9d4c61f4627a\") " pod="openstack/dnsmasq-dns-56f9cf6987-4865b" Oct 08 19:46:08 crc kubenswrapper[4988]: I1008 19:46:08.202360 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-56f9cf6987-4865b" Oct 08 19:46:08 crc kubenswrapper[4988]: I1008 19:46:08.206314 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x7bgb\" (UniqueName: \"kubernetes.io/projected/ff18525a-5969-4ae8-be54-ae79a7c72821-kube-api-access-x7bgb\") pod \"cinder-api-0\" (UID: \"ff18525a-5969-4ae8-be54-ae79a7c72821\") " pod="openstack/cinder-api-0" Oct 08 19:46:08 crc kubenswrapper[4988]: I1008 19:46:08.206357 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/ff18525a-5969-4ae8-be54-ae79a7c72821-config-data-custom\") pod \"cinder-api-0\" (UID: \"ff18525a-5969-4ae8-be54-ae79a7c72821\") " pod="openstack/cinder-api-0" Oct 08 19:46:08 crc kubenswrapper[4988]: I1008 19:46:08.206408 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ff18525a-5969-4ae8-be54-ae79a7c72821-config-data\") pod \"cinder-api-0\" (UID: \"ff18525a-5969-4ae8-be54-ae79a7c72821\") " pod="openstack/cinder-api-0" Oct 08 19:46:08 crc kubenswrapper[4988]: I1008 19:46:08.206426 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ff18525a-5969-4ae8-be54-ae79a7c72821-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"ff18525a-5969-4ae8-be54-ae79a7c72821\") " pod="openstack/cinder-api-0" Oct 08 19:46:08 crc kubenswrapper[4988]: I1008 19:46:08.206454 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/ff18525a-5969-4ae8-be54-ae79a7c72821-etc-machine-id\") pod \"cinder-api-0\" (UID: \"ff18525a-5969-4ae8-be54-ae79a7c72821\") " pod="openstack/cinder-api-0" Oct 08 19:46:08 crc kubenswrapper[4988]: I1008 19:46:08.206475 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ff18525a-5969-4ae8-be54-ae79a7c72821-scripts\") pod \"cinder-api-0\" (UID: \"ff18525a-5969-4ae8-be54-ae79a7c72821\") " pod="openstack/cinder-api-0" Oct 08 19:46:08 crc kubenswrapper[4988]: I1008 19:46:08.206559 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ff18525a-5969-4ae8-be54-ae79a7c72821-logs\") pod \"cinder-api-0\" (UID: \"ff18525a-5969-4ae8-be54-ae79a7c72821\") " pod="openstack/cinder-api-0" Oct 08 19:46:08 crc kubenswrapper[4988]: I1008 19:46:08.206866 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ff18525a-5969-4ae8-be54-ae79a7c72821-logs\") pod \"cinder-api-0\" (UID: \"ff18525a-5969-4ae8-be54-ae79a7c72821\") " pod="openstack/cinder-api-0" Oct 08 19:46:08 crc kubenswrapper[4988]: I1008 19:46:08.207428 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/ff18525a-5969-4ae8-be54-ae79a7c72821-etc-machine-id\") pod \"cinder-api-0\" (UID: \"ff18525a-5969-4ae8-be54-ae79a7c72821\") " pod="openstack/cinder-api-0" Oct 08 19:46:08 crc kubenswrapper[4988]: I1008 19:46:08.209697 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ff18525a-5969-4ae8-be54-ae79a7c72821-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"ff18525a-5969-4ae8-be54-ae79a7c72821\") " pod="openstack/cinder-api-0" Oct 08 19:46:08 crc kubenswrapper[4988]: I1008 19:46:08.210940 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ff18525a-5969-4ae8-be54-ae79a7c72821-config-data\") pod \"cinder-api-0\" (UID: \"ff18525a-5969-4ae8-be54-ae79a7c72821\") " pod="openstack/cinder-api-0" Oct 08 19:46:08 crc kubenswrapper[4988]: I1008 19:46:08.214895 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ff18525a-5969-4ae8-be54-ae79a7c72821-scripts\") pod \"cinder-api-0\" (UID: \"ff18525a-5969-4ae8-be54-ae79a7c72821\") " pod="openstack/cinder-api-0" Oct 08 19:46:08 crc kubenswrapper[4988]: I1008 19:46:08.216311 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/ff18525a-5969-4ae8-be54-ae79a7c72821-config-data-custom\") pod \"cinder-api-0\" (UID: \"ff18525a-5969-4ae8-be54-ae79a7c72821\") " pod="openstack/cinder-api-0" Oct 08 19:46:08 crc kubenswrapper[4988]: I1008 19:46:08.225188 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x7bgb\" (UniqueName: \"kubernetes.io/projected/ff18525a-5969-4ae8-be54-ae79a7c72821-kube-api-access-x7bgb\") pod \"cinder-api-0\" (UID: \"ff18525a-5969-4ae8-be54-ae79a7c72821\") " pod="openstack/cinder-api-0" Oct 08 19:46:08 crc kubenswrapper[4988]: I1008 19:46:08.366277 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 08 19:46:08 crc kubenswrapper[4988]: I1008 19:46:08.670548 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-56f9cf6987-4865b"] Oct 08 19:46:08 crc kubenswrapper[4988]: I1008 19:46:08.880264 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Oct 08 19:46:08 crc kubenswrapper[4988]: W1008 19:46:08.883941 4988 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podff18525a_5969_4ae8_be54_ae79a7c72821.slice/crio-2f636166ee133b4521f8d1d9ef56dd4fce73c09c353bade9e9998da5658051d8 WatchSource:0}: Error finding container 2f636166ee133b4521f8d1d9ef56dd4fce73c09c353bade9e9998da5658051d8: Status 404 returned error can't find the container with id 2f636166ee133b4521f8d1d9ef56dd4fce73c09c353bade9e9998da5658051d8 Oct 08 19:46:09 crc kubenswrapper[4988]: I1008 19:46:09.498798 4988 generic.go:334] "Generic (PLEG): container finished" podID="d4873fea-35df-48df-856f-9d4c61f4627a" containerID="4e91f30b65cc41952b68ff94905eb1417638da4d6d83f2d09bd6bd7e81c75141" exitCode=0 Oct 08 19:46:09 crc kubenswrapper[4988]: I1008 19:46:09.499138 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-56f9cf6987-4865b" event={"ID":"d4873fea-35df-48df-856f-9d4c61f4627a","Type":"ContainerDied","Data":"4e91f30b65cc41952b68ff94905eb1417638da4d6d83f2d09bd6bd7e81c75141"} Oct 08 19:46:09 crc kubenswrapper[4988]: I1008 19:46:09.499167 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-56f9cf6987-4865b" event={"ID":"d4873fea-35df-48df-856f-9d4c61f4627a","Type":"ContainerStarted","Data":"97a87a79dce8c2afc2e8864146bb2d9a670de19e5448babdd3e688a002944922"} Oct 08 19:46:09 crc kubenswrapper[4988]: I1008 19:46:09.507871 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"ff18525a-5969-4ae8-be54-ae79a7c72821","Type":"ContainerStarted","Data":"f884e699b534c487ba5f2f1c0ea87a02a01354f5f60f2517d3830f78d8d1897e"} Oct 08 19:46:09 crc kubenswrapper[4988]: I1008 19:46:09.507916 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"ff18525a-5969-4ae8-be54-ae79a7c72821","Type":"ContainerStarted","Data":"2f636166ee133b4521f8d1d9ef56dd4fce73c09c353bade9e9998da5658051d8"} Oct 08 19:46:10 crc kubenswrapper[4988]: I1008 19:46:10.110015 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Oct 08 19:46:10 crc kubenswrapper[4988]: I1008 19:46:10.517453 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"ff18525a-5969-4ae8-be54-ae79a7c72821","Type":"ContainerStarted","Data":"8e39c175ec7e6c0c93ececd2cbef4ce5fcd8eee1cd40b90491ca2981423c1b31"} Oct 08 19:46:10 crc kubenswrapper[4988]: I1008 19:46:10.517569 4988 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="ff18525a-5969-4ae8-be54-ae79a7c72821" containerName="cinder-api-log" containerID="cri-o://f884e699b534c487ba5f2f1c0ea87a02a01354f5f60f2517d3830f78d8d1897e" gracePeriod=30 Oct 08 19:46:10 crc kubenswrapper[4988]: I1008 19:46:10.517709 4988 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="ff18525a-5969-4ae8-be54-ae79a7c72821" containerName="cinder-api" containerID="cri-o://8e39c175ec7e6c0c93ececd2cbef4ce5fcd8eee1cd40b90491ca2981423c1b31" gracePeriod=30 Oct 08 19:46:10 crc kubenswrapper[4988]: I1008 19:46:10.517881 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Oct 08 19:46:10 crc kubenswrapper[4988]: I1008 19:46:10.522159 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-56f9cf6987-4865b" event={"ID":"d4873fea-35df-48df-856f-9d4c61f4627a","Type":"ContainerStarted","Data":"c866806d68920fec97681921584bced4ca900dab3bd184769e22ed92d82c8598"} Oct 08 19:46:10 crc kubenswrapper[4988]: I1008 19:46:10.522872 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-56f9cf6987-4865b" Oct 08 19:46:10 crc kubenswrapper[4988]: I1008 19:46:10.572964 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-56f9cf6987-4865b" podStartSLOduration=3.572947268 podStartE2EDuration="3.572947268s" podCreationTimestamp="2025-10-08 19:46:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 19:46:10.568879638 +0000 UTC m=+5716.018722408" watchObservedRunningTime="2025-10-08 19:46:10.572947268 +0000 UTC m=+5716.022790028" Oct 08 19:46:10 crc kubenswrapper[4988]: I1008 19:46:10.575850 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=2.57583962 podStartE2EDuration="2.57583962s" podCreationTimestamp="2025-10-08 19:46:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 19:46:10.551607949 +0000 UTC m=+5716.001450719" watchObservedRunningTime="2025-10-08 19:46:10.57583962 +0000 UTC m=+5716.025682390" Oct 08 19:46:11 crc kubenswrapper[4988]: I1008 19:46:11.102153 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 08 19:46:11 crc kubenswrapper[4988]: I1008 19:46:11.265311 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ff18525a-5969-4ae8-be54-ae79a7c72821-scripts\") pod \"ff18525a-5969-4ae8-be54-ae79a7c72821\" (UID: \"ff18525a-5969-4ae8-be54-ae79a7c72821\") " Oct 08 19:46:11 crc kubenswrapper[4988]: I1008 19:46:11.265368 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ff18525a-5969-4ae8-be54-ae79a7c72821-logs\") pod \"ff18525a-5969-4ae8-be54-ae79a7c72821\" (UID: \"ff18525a-5969-4ae8-be54-ae79a7c72821\") " Oct 08 19:46:11 crc kubenswrapper[4988]: I1008 19:46:11.265976 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ff18525a-5969-4ae8-be54-ae79a7c72821-combined-ca-bundle\") pod \"ff18525a-5969-4ae8-be54-ae79a7c72821\" (UID: \"ff18525a-5969-4ae8-be54-ae79a7c72821\") " Oct 08 19:46:11 crc kubenswrapper[4988]: I1008 19:46:11.266034 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/ff18525a-5969-4ae8-be54-ae79a7c72821-etc-machine-id\") pod \"ff18525a-5969-4ae8-be54-ae79a7c72821\" (UID: \"ff18525a-5969-4ae8-be54-ae79a7c72821\") " Oct 08 19:46:11 crc kubenswrapper[4988]: I1008 19:46:11.266060 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x7bgb\" (UniqueName: \"kubernetes.io/projected/ff18525a-5969-4ae8-be54-ae79a7c72821-kube-api-access-x7bgb\") pod \"ff18525a-5969-4ae8-be54-ae79a7c72821\" (UID: \"ff18525a-5969-4ae8-be54-ae79a7c72821\") " Oct 08 19:46:11 crc kubenswrapper[4988]: I1008 19:46:11.266093 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/ff18525a-5969-4ae8-be54-ae79a7c72821-config-data-custom\") pod \"ff18525a-5969-4ae8-be54-ae79a7c72821\" (UID: \"ff18525a-5969-4ae8-be54-ae79a7c72821\") " Oct 08 19:46:11 crc kubenswrapper[4988]: I1008 19:46:11.266114 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ff18525a-5969-4ae8-be54-ae79a7c72821-config-data\") pod \"ff18525a-5969-4ae8-be54-ae79a7c72821\" (UID: \"ff18525a-5969-4ae8-be54-ae79a7c72821\") " Oct 08 19:46:11 crc kubenswrapper[4988]: I1008 19:46:11.267525 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ff18525a-5969-4ae8-be54-ae79a7c72821-logs" (OuterVolumeSpecName: "logs") pod "ff18525a-5969-4ae8-be54-ae79a7c72821" (UID: "ff18525a-5969-4ae8-be54-ae79a7c72821"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 19:46:11 crc kubenswrapper[4988]: I1008 19:46:11.268032 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/ff18525a-5969-4ae8-be54-ae79a7c72821-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "ff18525a-5969-4ae8-be54-ae79a7c72821" (UID: "ff18525a-5969-4ae8-be54-ae79a7c72821"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 08 19:46:11 crc kubenswrapper[4988]: I1008 19:46:11.271330 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ff18525a-5969-4ae8-be54-ae79a7c72821-scripts" (OuterVolumeSpecName: "scripts") pod "ff18525a-5969-4ae8-be54-ae79a7c72821" (UID: "ff18525a-5969-4ae8-be54-ae79a7c72821"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 19:46:11 crc kubenswrapper[4988]: I1008 19:46:11.271958 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ff18525a-5969-4ae8-be54-ae79a7c72821-kube-api-access-x7bgb" (OuterVolumeSpecName: "kube-api-access-x7bgb") pod "ff18525a-5969-4ae8-be54-ae79a7c72821" (UID: "ff18525a-5969-4ae8-be54-ae79a7c72821"). InnerVolumeSpecName "kube-api-access-x7bgb". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 19:46:11 crc kubenswrapper[4988]: I1008 19:46:11.275487 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ff18525a-5969-4ae8-be54-ae79a7c72821-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "ff18525a-5969-4ae8-be54-ae79a7c72821" (UID: "ff18525a-5969-4ae8-be54-ae79a7c72821"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 19:46:11 crc kubenswrapper[4988]: I1008 19:46:11.308449 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ff18525a-5969-4ae8-be54-ae79a7c72821-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ff18525a-5969-4ae8-be54-ae79a7c72821" (UID: "ff18525a-5969-4ae8-be54-ae79a7c72821"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 19:46:11 crc kubenswrapper[4988]: I1008 19:46:11.314242 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ff18525a-5969-4ae8-be54-ae79a7c72821-config-data" (OuterVolumeSpecName: "config-data") pod "ff18525a-5969-4ae8-be54-ae79a7c72821" (UID: "ff18525a-5969-4ae8-be54-ae79a7c72821"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 19:46:11 crc kubenswrapper[4988]: I1008 19:46:11.368319 4988 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ff18525a-5969-4ae8-be54-ae79a7c72821-config-data\") on node \"crc\" DevicePath \"\"" Oct 08 19:46:11 crc kubenswrapper[4988]: I1008 19:46:11.368358 4988 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ff18525a-5969-4ae8-be54-ae79a7c72821-scripts\") on node \"crc\" DevicePath \"\"" Oct 08 19:46:11 crc kubenswrapper[4988]: I1008 19:46:11.368370 4988 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ff18525a-5969-4ae8-be54-ae79a7c72821-logs\") on node \"crc\" DevicePath \"\"" Oct 08 19:46:11 crc kubenswrapper[4988]: I1008 19:46:11.368397 4988 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ff18525a-5969-4ae8-be54-ae79a7c72821-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 19:46:11 crc kubenswrapper[4988]: I1008 19:46:11.368410 4988 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/ff18525a-5969-4ae8-be54-ae79a7c72821-etc-machine-id\") on node \"crc\" DevicePath \"\"" Oct 08 19:46:11 crc kubenswrapper[4988]: I1008 19:46:11.368427 4988 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x7bgb\" (UniqueName: \"kubernetes.io/projected/ff18525a-5969-4ae8-be54-ae79a7c72821-kube-api-access-x7bgb\") on node \"crc\" DevicePath \"\"" Oct 08 19:46:11 crc kubenswrapper[4988]: I1008 19:46:11.368441 4988 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/ff18525a-5969-4ae8-be54-ae79a7c72821-config-data-custom\") on node \"crc\" DevicePath \"\"" Oct 08 19:46:11 crc kubenswrapper[4988]: I1008 19:46:11.534924 4988 generic.go:334] "Generic (PLEG): container finished" podID="ff18525a-5969-4ae8-be54-ae79a7c72821" containerID="8e39c175ec7e6c0c93ececd2cbef4ce5fcd8eee1cd40b90491ca2981423c1b31" exitCode=0 Oct 08 19:46:11 crc kubenswrapper[4988]: I1008 19:46:11.534961 4988 generic.go:334] "Generic (PLEG): container finished" podID="ff18525a-5969-4ae8-be54-ae79a7c72821" containerID="f884e699b534c487ba5f2f1c0ea87a02a01354f5f60f2517d3830f78d8d1897e" exitCode=143 Oct 08 19:46:11 crc kubenswrapper[4988]: I1008 19:46:11.534980 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"ff18525a-5969-4ae8-be54-ae79a7c72821","Type":"ContainerDied","Data":"8e39c175ec7e6c0c93ececd2cbef4ce5fcd8eee1cd40b90491ca2981423c1b31"} Oct 08 19:46:11 crc kubenswrapper[4988]: I1008 19:46:11.535056 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"ff18525a-5969-4ae8-be54-ae79a7c72821","Type":"ContainerDied","Data":"f884e699b534c487ba5f2f1c0ea87a02a01354f5f60f2517d3830f78d8d1897e"} Oct 08 19:46:11 crc kubenswrapper[4988]: I1008 19:46:11.535072 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"ff18525a-5969-4ae8-be54-ae79a7c72821","Type":"ContainerDied","Data":"2f636166ee133b4521f8d1d9ef56dd4fce73c09c353bade9e9998da5658051d8"} Oct 08 19:46:11 crc kubenswrapper[4988]: I1008 19:46:11.535014 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 08 19:46:11 crc kubenswrapper[4988]: I1008 19:46:11.535112 4988 scope.go:117] "RemoveContainer" containerID="8e39c175ec7e6c0c93ececd2cbef4ce5fcd8eee1cd40b90491ca2981423c1b31" Oct 08 19:46:11 crc kubenswrapper[4988]: I1008 19:46:11.559576 4988 scope.go:117] "RemoveContainer" containerID="f884e699b534c487ba5f2f1c0ea87a02a01354f5f60f2517d3830f78d8d1897e" Oct 08 19:46:11 crc kubenswrapper[4988]: I1008 19:46:11.568142 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Oct 08 19:46:11 crc kubenswrapper[4988]: I1008 19:46:11.585551 4988 scope.go:117] "RemoveContainer" containerID="8e39c175ec7e6c0c93ececd2cbef4ce5fcd8eee1cd40b90491ca2981423c1b31" Oct 08 19:46:11 crc kubenswrapper[4988]: E1008 19:46:11.588776 4988 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8e39c175ec7e6c0c93ececd2cbef4ce5fcd8eee1cd40b90491ca2981423c1b31\": container with ID starting with 8e39c175ec7e6c0c93ececd2cbef4ce5fcd8eee1cd40b90491ca2981423c1b31 not found: ID does not exist" containerID="8e39c175ec7e6c0c93ececd2cbef4ce5fcd8eee1cd40b90491ca2981423c1b31" Oct 08 19:46:11 crc kubenswrapper[4988]: I1008 19:46:11.588819 4988 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8e39c175ec7e6c0c93ececd2cbef4ce5fcd8eee1cd40b90491ca2981423c1b31"} err="failed to get container status \"8e39c175ec7e6c0c93ececd2cbef4ce5fcd8eee1cd40b90491ca2981423c1b31\": rpc error: code = NotFound desc = could not find container \"8e39c175ec7e6c0c93ececd2cbef4ce5fcd8eee1cd40b90491ca2981423c1b31\": container with ID starting with 8e39c175ec7e6c0c93ececd2cbef4ce5fcd8eee1cd40b90491ca2981423c1b31 not found: ID does not exist" Oct 08 19:46:11 crc kubenswrapper[4988]: I1008 19:46:11.588846 4988 scope.go:117] "RemoveContainer" containerID="f884e699b534c487ba5f2f1c0ea87a02a01354f5f60f2517d3830f78d8d1897e" Oct 08 19:46:11 crc kubenswrapper[4988]: I1008 19:46:11.589088 4988 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-api-0"] Oct 08 19:46:11 crc kubenswrapper[4988]: E1008 19:46:11.589782 4988 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f884e699b534c487ba5f2f1c0ea87a02a01354f5f60f2517d3830f78d8d1897e\": container with ID starting with f884e699b534c487ba5f2f1c0ea87a02a01354f5f60f2517d3830f78d8d1897e not found: ID does not exist" containerID="f884e699b534c487ba5f2f1c0ea87a02a01354f5f60f2517d3830f78d8d1897e" Oct 08 19:46:11 crc kubenswrapper[4988]: I1008 19:46:11.589821 4988 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f884e699b534c487ba5f2f1c0ea87a02a01354f5f60f2517d3830f78d8d1897e"} err="failed to get container status \"f884e699b534c487ba5f2f1c0ea87a02a01354f5f60f2517d3830f78d8d1897e\": rpc error: code = NotFound desc = could not find container \"f884e699b534c487ba5f2f1c0ea87a02a01354f5f60f2517d3830f78d8d1897e\": container with ID starting with f884e699b534c487ba5f2f1c0ea87a02a01354f5f60f2517d3830f78d8d1897e not found: ID does not exist" Oct 08 19:46:11 crc kubenswrapper[4988]: I1008 19:46:11.589852 4988 scope.go:117] "RemoveContainer" containerID="8e39c175ec7e6c0c93ececd2cbef4ce5fcd8eee1cd40b90491ca2981423c1b31" Oct 08 19:46:11 crc kubenswrapper[4988]: I1008 19:46:11.590406 4988 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8e39c175ec7e6c0c93ececd2cbef4ce5fcd8eee1cd40b90491ca2981423c1b31"} err="failed to get container status \"8e39c175ec7e6c0c93ececd2cbef4ce5fcd8eee1cd40b90491ca2981423c1b31\": rpc error: code = NotFound desc = could not find container \"8e39c175ec7e6c0c93ececd2cbef4ce5fcd8eee1cd40b90491ca2981423c1b31\": container with ID starting with 8e39c175ec7e6c0c93ececd2cbef4ce5fcd8eee1cd40b90491ca2981423c1b31 not found: ID does not exist" Oct 08 19:46:11 crc kubenswrapper[4988]: I1008 19:46:11.590456 4988 scope.go:117] "RemoveContainer" containerID="f884e699b534c487ba5f2f1c0ea87a02a01354f5f60f2517d3830f78d8d1897e" Oct 08 19:46:11 crc kubenswrapper[4988]: I1008 19:46:11.590926 4988 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f884e699b534c487ba5f2f1c0ea87a02a01354f5f60f2517d3830f78d8d1897e"} err="failed to get container status \"f884e699b534c487ba5f2f1c0ea87a02a01354f5f60f2517d3830f78d8d1897e\": rpc error: code = NotFound desc = could not find container \"f884e699b534c487ba5f2f1c0ea87a02a01354f5f60f2517d3830f78d8d1897e\": container with ID starting with f884e699b534c487ba5f2f1c0ea87a02a01354f5f60f2517d3830f78d8d1897e not found: ID does not exist" Oct 08 19:46:11 crc kubenswrapper[4988]: I1008 19:46:11.604711 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Oct 08 19:46:11 crc kubenswrapper[4988]: E1008 19:46:11.605058 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ff18525a-5969-4ae8-be54-ae79a7c72821" containerName="cinder-api-log" Oct 08 19:46:11 crc kubenswrapper[4988]: I1008 19:46:11.605070 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="ff18525a-5969-4ae8-be54-ae79a7c72821" containerName="cinder-api-log" Oct 08 19:46:11 crc kubenswrapper[4988]: E1008 19:46:11.605099 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ff18525a-5969-4ae8-be54-ae79a7c72821" containerName="cinder-api" Oct 08 19:46:11 crc kubenswrapper[4988]: I1008 19:46:11.605104 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="ff18525a-5969-4ae8-be54-ae79a7c72821" containerName="cinder-api" Oct 08 19:46:11 crc kubenswrapper[4988]: I1008 19:46:11.605270 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="ff18525a-5969-4ae8-be54-ae79a7c72821" containerName="cinder-api" Oct 08 19:46:11 crc kubenswrapper[4988]: I1008 19:46:11.605289 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="ff18525a-5969-4ae8-be54-ae79a7c72821" containerName="cinder-api-log" Oct 08 19:46:11 crc kubenswrapper[4988]: I1008 19:46:11.606165 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 08 19:46:11 crc kubenswrapper[4988]: I1008 19:46:11.611353 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Oct 08 19:46:11 crc kubenswrapper[4988]: I1008 19:46:11.614327 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-public-svc" Oct 08 19:46:11 crc kubenswrapper[4988]: I1008 19:46:11.615429 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-internal-svc" Oct 08 19:46:11 crc kubenswrapper[4988]: I1008 19:46:11.616377 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-wzqmq" Oct 08 19:46:11 crc kubenswrapper[4988]: I1008 19:46:11.616521 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Oct 08 19:46:11 crc kubenswrapper[4988]: I1008 19:46:11.617056 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Oct 08 19:46:11 crc kubenswrapper[4988]: I1008 19:46:11.623640 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Oct 08 19:46:11 crc kubenswrapper[4988]: I1008 19:46:11.678903 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0c3843ef-c63c-41cb-b3ff-cf264790ca78-logs\") pod \"cinder-api-0\" (UID: \"0c3843ef-c63c-41cb-b3ff-cf264790ca78\") " pod="openstack/cinder-api-0" Oct 08 19:46:11 crc kubenswrapper[4988]: I1008 19:46:11.678961 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0c3843ef-c63c-41cb-b3ff-cf264790ca78-config-data\") pod \"cinder-api-0\" (UID: \"0c3843ef-c63c-41cb-b3ff-cf264790ca78\") " pod="openstack/cinder-api-0" Oct 08 19:46:11 crc kubenswrapper[4988]: I1008 19:46:11.679012 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/0c3843ef-c63c-41cb-b3ff-cf264790ca78-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"0c3843ef-c63c-41cb-b3ff-cf264790ca78\") " pod="openstack/cinder-api-0" Oct 08 19:46:11 crc kubenswrapper[4988]: I1008 19:46:11.679037 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/0c3843ef-c63c-41cb-b3ff-cf264790ca78-public-tls-certs\") pod \"cinder-api-0\" (UID: \"0c3843ef-c63c-41cb-b3ff-cf264790ca78\") " pod="openstack/cinder-api-0" Oct 08 19:46:11 crc kubenswrapper[4988]: I1008 19:46:11.679055 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x57nt\" (UniqueName: \"kubernetes.io/projected/0c3843ef-c63c-41cb-b3ff-cf264790ca78-kube-api-access-x57nt\") pod \"cinder-api-0\" (UID: \"0c3843ef-c63c-41cb-b3ff-cf264790ca78\") " pod="openstack/cinder-api-0" Oct 08 19:46:11 crc kubenswrapper[4988]: I1008 19:46:11.679091 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/0c3843ef-c63c-41cb-b3ff-cf264790ca78-config-data-custom\") pod \"cinder-api-0\" (UID: \"0c3843ef-c63c-41cb-b3ff-cf264790ca78\") " pod="openstack/cinder-api-0" Oct 08 19:46:11 crc kubenswrapper[4988]: I1008 19:46:11.679155 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0c3843ef-c63c-41cb-b3ff-cf264790ca78-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"0c3843ef-c63c-41cb-b3ff-cf264790ca78\") " pod="openstack/cinder-api-0" Oct 08 19:46:11 crc kubenswrapper[4988]: I1008 19:46:11.679173 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0c3843ef-c63c-41cb-b3ff-cf264790ca78-scripts\") pod \"cinder-api-0\" (UID: \"0c3843ef-c63c-41cb-b3ff-cf264790ca78\") " pod="openstack/cinder-api-0" Oct 08 19:46:11 crc kubenswrapper[4988]: I1008 19:46:11.679200 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/0c3843ef-c63c-41cb-b3ff-cf264790ca78-etc-machine-id\") pod \"cinder-api-0\" (UID: \"0c3843ef-c63c-41cb-b3ff-cf264790ca78\") " pod="openstack/cinder-api-0" Oct 08 19:46:11 crc kubenswrapper[4988]: I1008 19:46:11.781234 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0c3843ef-c63c-41cb-b3ff-cf264790ca78-logs\") pod \"cinder-api-0\" (UID: \"0c3843ef-c63c-41cb-b3ff-cf264790ca78\") " pod="openstack/cinder-api-0" Oct 08 19:46:11 crc kubenswrapper[4988]: I1008 19:46:11.781291 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0c3843ef-c63c-41cb-b3ff-cf264790ca78-config-data\") pod \"cinder-api-0\" (UID: \"0c3843ef-c63c-41cb-b3ff-cf264790ca78\") " pod="openstack/cinder-api-0" Oct 08 19:46:11 crc kubenswrapper[4988]: I1008 19:46:11.781338 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/0c3843ef-c63c-41cb-b3ff-cf264790ca78-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"0c3843ef-c63c-41cb-b3ff-cf264790ca78\") " pod="openstack/cinder-api-0" Oct 08 19:46:11 crc kubenswrapper[4988]: I1008 19:46:11.781362 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/0c3843ef-c63c-41cb-b3ff-cf264790ca78-public-tls-certs\") pod \"cinder-api-0\" (UID: \"0c3843ef-c63c-41cb-b3ff-cf264790ca78\") " pod="openstack/cinder-api-0" Oct 08 19:46:11 crc kubenswrapper[4988]: I1008 19:46:11.781391 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x57nt\" (UniqueName: \"kubernetes.io/projected/0c3843ef-c63c-41cb-b3ff-cf264790ca78-kube-api-access-x57nt\") pod \"cinder-api-0\" (UID: \"0c3843ef-c63c-41cb-b3ff-cf264790ca78\") " pod="openstack/cinder-api-0" Oct 08 19:46:11 crc kubenswrapper[4988]: I1008 19:46:11.781426 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/0c3843ef-c63c-41cb-b3ff-cf264790ca78-config-data-custom\") pod \"cinder-api-0\" (UID: \"0c3843ef-c63c-41cb-b3ff-cf264790ca78\") " pod="openstack/cinder-api-0" Oct 08 19:46:11 crc kubenswrapper[4988]: I1008 19:46:11.781460 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0c3843ef-c63c-41cb-b3ff-cf264790ca78-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"0c3843ef-c63c-41cb-b3ff-cf264790ca78\") " pod="openstack/cinder-api-0" Oct 08 19:46:11 crc kubenswrapper[4988]: I1008 19:46:11.781483 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0c3843ef-c63c-41cb-b3ff-cf264790ca78-scripts\") pod \"cinder-api-0\" (UID: \"0c3843ef-c63c-41cb-b3ff-cf264790ca78\") " pod="openstack/cinder-api-0" Oct 08 19:46:11 crc kubenswrapper[4988]: I1008 19:46:11.781509 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/0c3843ef-c63c-41cb-b3ff-cf264790ca78-etc-machine-id\") pod \"cinder-api-0\" (UID: \"0c3843ef-c63c-41cb-b3ff-cf264790ca78\") " pod="openstack/cinder-api-0" Oct 08 19:46:11 crc kubenswrapper[4988]: I1008 19:46:11.781569 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/0c3843ef-c63c-41cb-b3ff-cf264790ca78-etc-machine-id\") pod \"cinder-api-0\" (UID: \"0c3843ef-c63c-41cb-b3ff-cf264790ca78\") " pod="openstack/cinder-api-0" Oct 08 19:46:11 crc kubenswrapper[4988]: I1008 19:46:11.781704 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0c3843ef-c63c-41cb-b3ff-cf264790ca78-logs\") pod \"cinder-api-0\" (UID: \"0c3843ef-c63c-41cb-b3ff-cf264790ca78\") " pod="openstack/cinder-api-0" Oct 08 19:46:11 crc kubenswrapper[4988]: I1008 19:46:11.785107 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/0c3843ef-c63c-41cb-b3ff-cf264790ca78-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"0c3843ef-c63c-41cb-b3ff-cf264790ca78\") " pod="openstack/cinder-api-0" Oct 08 19:46:11 crc kubenswrapper[4988]: I1008 19:46:11.785140 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/0c3843ef-c63c-41cb-b3ff-cf264790ca78-config-data-custom\") pod \"cinder-api-0\" (UID: \"0c3843ef-c63c-41cb-b3ff-cf264790ca78\") " pod="openstack/cinder-api-0" Oct 08 19:46:11 crc kubenswrapper[4988]: I1008 19:46:11.785341 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0c3843ef-c63c-41cb-b3ff-cf264790ca78-scripts\") pod \"cinder-api-0\" (UID: \"0c3843ef-c63c-41cb-b3ff-cf264790ca78\") " pod="openstack/cinder-api-0" Oct 08 19:46:11 crc kubenswrapper[4988]: I1008 19:46:11.786889 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/0c3843ef-c63c-41cb-b3ff-cf264790ca78-public-tls-certs\") pod \"cinder-api-0\" (UID: \"0c3843ef-c63c-41cb-b3ff-cf264790ca78\") " pod="openstack/cinder-api-0" Oct 08 19:46:11 crc kubenswrapper[4988]: I1008 19:46:11.787825 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0c3843ef-c63c-41cb-b3ff-cf264790ca78-config-data\") pod \"cinder-api-0\" (UID: \"0c3843ef-c63c-41cb-b3ff-cf264790ca78\") " pod="openstack/cinder-api-0" Oct 08 19:46:11 crc kubenswrapper[4988]: I1008 19:46:11.789732 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0c3843ef-c63c-41cb-b3ff-cf264790ca78-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"0c3843ef-c63c-41cb-b3ff-cf264790ca78\") " pod="openstack/cinder-api-0" Oct 08 19:46:11 crc kubenswrapper[4988]: I1008 19:46:11.798747 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x57nt\" (UniqueName: \"kubernetes.io/projected/0c3843ef-c63c-41cb-b3ff-cf264790ca78-kube-api-access-x57nt\") pod \"cinder-api-0\" (UID: \"0c3843ef-c63c-41cb-b3ff-cf264790ca78\") " pod="openstack/cinder-api-0" Oct 08 19:46:11 crc kubenswrapper[4988]: I1008 19:46:11.932866 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 08 19:46:12 crc kubenswrapper[4988]: I1008 19:46:12.229163 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Oct 08 19:46:12 crc kubenswrapper[4988]: I1008 19:46:12.547041 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"0c3843ef-c63c-41cb-b3ff-cf264790ca78","Type":"ContainerStarted","Data":"ec34761646d3b54d0a6d731828ba8362f0d430caca1c60af73652e214ebd55fa"} Oct 08 19:46:13 crc kubenswrapper[4988]: I1008 19:46:13.251072 4988 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ff18525a-5969-4ae8-be54-ae79a7c72821" path="/var/lib/kubelet/pods/ff18525a-5969-4ae8-be54-ae79a7c72821/volumes" Oct 08 19:46:13 crc kubenswrapper[4988]: I1008 19:46:13.563154 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"0c3843ef-c63c-41cb-b3ff-cf264790ca78","Type":"ContainerStarted","Data":"3b86f1b55753a6ccd8391660045573be16c6a587576932bab614bff4f255ffef"} Oct 08 19:46:13 crc kubenswrapper[4988]: I1008 19:46:13.563207 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"0c3843ef-c63c-41cb-b3ff-cf264790ca78","Type":"ContainerStarted","Data":"ec266c8bc4c08ec14d15ce126d7820ac8681601b18c629e063ece19e3755a2ec"} Oct 08 19:46:13 crc kubenswrapper[4988]: I1008 19:46:13.563431 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Oct 08 19:46:13 crc kubenswrapper[4988]: I1008 19:46:13.597167 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=2.59713912 podStartE2EDuration="2.59713912s" podCreationTimestamp="2025-10-08 19:46:11 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 19:46:13.590852131 +0000 UTC m=+5719.040694931" watchObservedRunningTime="2025-10-08 19:46:13.59713912 +0000 UTC m=+5719.046981930" Oct 08 19:46:18 crc kubenswrapper[4988]: I1008 19:46:18.207506 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-56f9cf6987-4865b" Oct 08 19:46:18 crc kubenswrapper[4988]: I1008 19:46:18.299237 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6d64fc8cdc-jb7cp"] Oct 08 19:46:18 crc kubenswrapper[4988]: I1008 19:46:18.299649 4988 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-6d64fc8cdc-jb7cp" podUID="5bd22f4a-f899-4db4-88a1-10344841593b" containerName="dnsmasq-dns" containerID="cri-o://ac7105c5d8e7c847fbb985bf98c080eaf4caf929bda6fd9f77fa4b83a42e3f8c" gracePeriod=10 Oct 08 19:46:18 crc kubenswrapper[4988]: I1008 19:46:18.637274 4988 generic.go:334] "Generic (PLEG): container finished" podID="5bd22f4a-f899-4db4-88a1-10344841593b" containerID="ac7105c5d8e7c847fbb985bf98c080eaf4caf929bda6fd9f77fa4b83a42e3f8c" exitCode=0 Oct 08 19:46:18 crc kubenswrapper[4988]: I1008 19:46:18.637528 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6d64fc8cdc-jb7cp" event={"ID":"5bd22f4a-f899-4db4-88a1-10344841593b","Type":"ContainerDied","Data":"ac7105c5d8e7c847fbb985bf98c080eaf4caf929bda6fd9f77fa4b83a42e3f8c"} Oct 08 19:46:18 crc kubenswrapper[4988]: I1008 19:46:18.773989 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6d64fc8cdc-jb7cp" Oct 08 19:46:18 crc kubenswrapper[4988]: I1008 19:46:18.833601 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5bd22f4a-f899-4db4-88a1-10344841593b-config\") pod \"5bd22f4a-f899-4db4-88a1-10344841593b\" (UID: \"5bd22f4a-f899-4db4-88a1-10344841593b\") " Oct 08 19:46:18 crc kubenswrapper[4988]: I1008 19:46:18.833718 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5bd22f4a-f899-4db4-88a1-10344841593b-dns-svc\") pod \"5bd22f4a-f899-4db4-88a1-10344841593b\" (UID: \"5bd22f4a-f899-4db4-88a1-10344841593b\") " Oct 08 19:46:18 crc kubenswrapper[4988]: I1008 19:46:18.833852 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qwvjx\" (UniqueName: \"kubernetes.io/projected/5bd22f4a-f899-4db4-88a1-10344841593b-kube-api-access-qwvjx\") pod \"5bd22f4a-f899-4db4-88a1-10344841593b\" (UID: \"5bd22f4a-f899-4db4-88a1-10344841593b\") " Oct 08 19:46:18 crc kubenswrapper[4988]: I1008 19:46:18.833898 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5bd22f4a-f899-4db4-88a1-10344841593b-ovsdbserver-sb\") pod \"5bd22f4a-f899-4db4-88a1-10344841593b\" (UID: \"5bd22f4a-f899-4db4-88a1-10344841593b\") " Oct 08 19:46:18 crc kubenswrapper[4988]: I1008 19:46:18.833969 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5bd22f4a-f899-4db4-88a1-10344841593b-ovsdbserver-nb\") pod \"5bd22f4a-f899-4db4-88a1-10344841593b\" (UID: \"5bd22f4a-f899-4db4-88a1-10344841593b\") " Oct 08 19:46:18 crc kubenswrapper[4988]: I1008 19:46:18.854602 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5bd22f4a-f899-4db4-88a1-10344841593b-kube-api-access-qwvjx" (OuterVolumeSpecName: "kube-api-access-qwvjx") pod "5bd22f4a-f899-4db4-88a1-10344841593b" (UID: "5bd22f4a-f899-4db4-88a1-10344841593b"). InnerVolumeSpecName "kube-api-access-qwvjx". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 19:46:18 crc kubenswrapper[4988]: I1008 19:46:18.881343 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5bd22f4a-f899-4db4-88a1-10344841593b-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "5bd22f4a-f899-4db4-88a1-10344841593b" (UID: "5bd22f4a-f899-4db4-88a1-10344841593b"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 19:46:18 crc kubenswrapper[4988]: I1008 19:46:18.891432 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5bd22f4a-f899-4db4-88a1-10344841593b-config" (OuterVolumeSpecName: "config") pod "5bd22f4a-f899-4db4-88a1-10344841593b" (UID: "5bd22f4a-f899-4db4-88a1-10344841593b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 19:46:18 crc kubenswrapper[4988]: I1008 19:46:18.897908 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5bd22f4a-f899-4db4-88a1-10344841593b-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "5bd22f4a-f899-4db4-88a1-10344841593b" (UID: "5bd22f4a-f899-4db4-88a1-10344841593b"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 19:46:18 crc kubenswrapper[4988]: I1008 19:46:18.912262 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5bd22f4a-f899-4db4-88a1-10344841593b-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "5bd22f4a-f899-4db4-88a1-10344841593b" (UID: "5bd22f4a-f899-4db4-88a1-10344841593b"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 19:46:18 crc kubenswrapper[4988]: I1008 19:46:18.940171 4988 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qwvjx\" (UniqueName: \"kubernetes.io/projected/5bd22f4a-f899-4db4-88a1-10344841593b-kube-api-access-qwvjx\") on node \"crc\" DevicePath \"\"" Oct 08 19:46:18 crc kubenswrapper[4988]: I1008 19:46:18.940216 4988 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5bd22f4a-f899-4db4-88a1-10344841593b-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 08 19:46:18 crc kubenswrapper[4988]: I1008 19:46:18.940231 4988 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5bd22f4a-f899-4db4-88a1-10344841593b-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 08 19:46:18 crc kubenswrapper[4988]: I1008 19:46:18.940245 4988 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5bd22f4a-f899-4db4-88a1-10344841593b-config\") on node \"crc\" DevicePath \"\"" Oct 08 19:46:18 crc kubenswrapper[4988]: I1008 19:46:18.940257 4988 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5bd22f4a-f899-4db4-88a1-10344841593b-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 08 19:46:19 crc kubenswrapper[4988]: I1008 19:46:19.656614 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6d64fc8cdc-jb7cp" event={"ID":"5bd22f4a-f899-4db4-88a1-10344841593b","Type":"ContainerDied","Data":"70ceb3dbac6d506652ffc91df3138eb6f9b4185c457725e9cddfba4a514c3e2a"} Oct 08 19:46:19 crc kubenswrapper[4988]: I1008 19:46:19.656686 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6d64fc8cdc-jb7cp" Oct 08 19:46:19 crc kubenswrapper[4988]: I1008 19:46:19.657015 4988 scope.go:117] "RemoveContainer" containerID="ac7105c5d8e7c847fbb985bf98c080eaf4caf929bda6fd9f77fa4b83a42e3f8c" Oct 08 19:46:19 crc kubenswrapper[4988]: I1008 19:46:19.686109 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6d64fc8cdc-jb7cp"] Oct 08 19:46:19 crc kubenswrapper[4988]: I1008 19:46:19.695211 4988 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-6d64fc8cdc-jb7cp"] Oct 08 19:46:19 crc kubenswrapper[4988]: I1008 19:46:19.698778 4988 scope.go:117] "RemoveContainer" containerID="2570a1e23d022c1995fec65e70d8fc7ff21575a1cd4aaea4504a2d61ab9440c1" Oct 08 19:46:21 crc kubenswrapper[4988]: I1008 19:46:21.252556 4988 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5bd22f4a-f899-4db4-88a1-10344841593b" path="/var/lib/kubelet/pods/5bd22f4a-f899-4db4-88a1-10344841593b/volumes" Oct 08 19:46:23 crc kubenswrapper[4988]: I1008 19:46:23.337735 4988 patch_prober.go:28] interesting pod/machine-config-daemon-nm54f container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 19:46:23 crc kubenswrapper[4988]: I1008 19:46:23.338076 4988 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 19:46:23 crc kubenswrapper[4988]: I1008 19:46:23.338129 4988 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" Oct 08 19:46:23 crc kubenswrapper[4988]: I1008 19:46:23.338888 4988 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"f2a29ed22dd7a48bbc929e65fbeceed969a669e78f65fbfb724b03dc0db6e947"} pod="openshift-machine-config-operator/machine-config-daemon-nm54f" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 08 19:46:23 crc kubenswrapper[4988]: I1008 19:46:23.338962 4988 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" containerName="machine-config-daemon" containerID="cri-o://f2a29ed22dd7a48bbc929e65fbeceed969a669e78f65fbfb724b03dc0db6e947" gracePeriod=600 Oct 08 19:46:23 crc kubenswrapper[4988]: E1008 19:46:23.463487 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nm54f_openshift-machine-config-operator(f328a20b-5283-407e-8a93-2ab2da599f18)\"" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" Oct 08 19:46:23 crc kubenswrapper[4988]: I1008 19:46:23.702583 4988 generic.go:334] "Generic (PLEG): container finished" podID="f328a20b-5283-407e-8a93-2ab2da599f18" containerID="f2a29ed22dd7a48bbc929e65fbeceed969a669e78f65fbfb724b03dc0db6e947" exitCode=0 Oct 08 19:46:23 crc kubenswrapper[4988]: I1008 19:46:23.702644 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" event={"ID":"f328a20b-5283-407e-8a93-2ab2da599f18","Type":"ContainerDied","Data":"f2a29ed22dd7a48bbc929e65fbeceed969a669e78f65fbfb724b03dc0db6e947"} Oct 08 19:46:23 crc kubenswrapper[4988]: I1008 19:46:23.702700 4988 scope.go:117] "RemoveContainer" containerID="bfa76a99ab92f7c2c936649e27734f81c2af70cdaf33874dac1d856a4b92f5f7" Oct 08 19:46:23 crc kubenswrapper[4988]: I1008 19:46:23.703348 4988 scope.go:117] "RemoveContainer" containerID="f2a29ed22dd7a48bbc929e65fbeceed969a669e78f65fbfb724b03dc0db6e947" Oct 08 19:46:23 crc kubenswrapper[4988]: E1008 19:46:23.703741 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nm54f_openshift-machine-config-operator(f328a20b-5283-407e-8a93-2ab2da599f18)\"" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" Oct 08 19:46:23 crc kubenswrapper[4988]: I1008 19:46:23.839777 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/cinder-api-0" Oct 08 19:46:35 crc kubenswrapper[4988]: I1008 19:46:35.256707 4988 scope.go:117] "RemoveContainer" containerID="f2a29ed22dd7a48bbc929e65fbeceed969a669e78f65fbfb724b03dc0db6e947" Oct 08 19:46:35 crc kubenswrapper[4988]: E1008 19:46:35.258812 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nm54f_openshift-machine-config-operator(f328a20b-5283-407e-8a93-2ab2da599f18)\"" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" Oct 08 19:46:40 crc kubenswrapper[4988]: I1008 19:46:40.453994 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Oct 08 19:46:40 crc kubenswrapper[4988]: E1008 19:46:40.454742 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5bd22f4a-f899-4db4-88a1-10344841593b" containerName="init" Oct 08 19:46:40 crc kubenswrapper[4988]: I1008 19:46:40.454756 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="5bd22f4a-f899-4db4-88a1-10344841593b" containerName="init" Oct 08 19:46:40 crc kubenswrapper[4988]: E1008 19:46:40.454773 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5bd22f4a-f899-4db4-88a1-10344841593b" containerName="dnsmasq-dns" Oct 08 19:46:40 crc kubenswrapper[4988]: I1008 19:46:40.454779 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="5bd22f4a-f899-4db4-88a1-10344841593b" containerName="dnsmasq-dns" Oct 08 19:46:40 crc kubenswrapper[4988]: I1008 19:46:40.454965 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="5bd22f4a-f899-4db4-88a1-10344841593b" containerName="dnsmasq-dns" Oct 08 19:46:40 crc kubenswrapper[4988]: I1008 19:46:40.455899 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Oct 08 19:46:40 crc kubenswrapper[4988]: I1008 19:46:40.457953 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Oct 08 19:46:40 crc kubenswrapper[4988]: I1008 19:46:40.475857 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 08 19:46:40 crc kubenswrapper[4988]: I1008 19:46:40.582673 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/54a89dee-6134-4487-a44f-0be9ca5e8d12-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"54a89dee-6134-4487-a44f-0be9ca5e8d12\") " pod="openstack/cinder-scheduler-0" Oct 08 19:46:40 crc kubenswrapper[4988]: I1008 19:46:40.582755 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d9gr6\" (UniqueName: \"kubernetes.io/projected/54a89dee-6134-4487-a44f-0be9ca5e8d12-kube-api-access-d9gr6\") pod \"cinder-scheduler-0\" (UID: \"54a89dee-6134-4487-a44f-0be9ca5e8d12\") " pod="openstack/cinder-scheduler-0" Oct 08 19:46:40 crc kubenswrapper[4988]: I1008 19:46:40.582830 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/54a89dee-6134-4487-a44f-0be9ca5e8d12-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"54a89dee-6134-4487-a44f-0be9ca5e8d12\") " pod="openstack/cinder-scheduler-0" Oct 08 19:46:40 crc kubenswrapper[4988]: I1008 19:46:40.582878 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/54a89dee-6134-4487-a44f-0be9ca5e8d12-config-data\") pod \"cinder-scheduler-0\" (UID: \"54a89dee-6134-4487-a44f-0be9ca5e8d12\") " pod="openstack/cinder-scheduler-0" Oct 08 19:46:40 crc kubenswrapper[4988]: I1008 19:46:40.582912 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/54a89dee-6134-4487-a44f-0be9ca5e8d12-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"54a89dee-6134-4487-a44f-0be9ca5e8d12\") " pod="openstack/cinder-scheduler-0" Oct 08 19:46:40 crc kubenswrapper[4988]: I1008 19:46:40.582955 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/54a89dee-6134-4487-a44f-0be9ca5e8d12-scripts\") pod \"cinder-scheduler-0\" (UID: \"54a89dee-6134-4487-a44f-0be9ca5e8d12\") " pod="openstack/cinder-scheduler-0" Oct 08 19:46:40 crc kubenswrapper[4988]: I1008 19:46:40.685049 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d9gr6\" (UniqueName: \"kubernetes.io/projected/54a89dee-6134-4487-a44f-0be9ca5e8d12-kube-api-access-d9gr6\") pod \"cinder-scheduler-0\" (UID: \"54a89dee-6134-4487-a44f-0be9ca5e8d12\") " pod="openstack/cinder-scheduler-0" Oct 08 19:46:40 crc kubenswrapper[4988]: I1008 19:46:40.685153 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/54a89dee-6134-4487-a44f-0be9ca5e8d12-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"54a89dee-6134-4487-a44f-0be9ca5e8d12\") " pod="openstack/cinder-scheduler-0" Oct 08 19:46:40 crc kubenswrapper[4988]: I1008 19:46:40.685199 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/54a89dee-6134-4487-a44f-0be9ca5e8d12-config-data\") pod \"cinder-scheduler-0\" (UID: \"54a89dee-6134-4487-a44f-0be9ca5e8d12\") " pod="openstack/cinder-scheduler-0" Oct 08 19:46:40 crc kubenswrapper[4988]: I1008 19:46:40.685236 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/54a89dee-6134-4487-a44f-0be9ca5e8d12-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"54a89dee-6134-4487-a44f-0be9ca5e8d12\") " pod="openstack/cinder-scheduler-0" Oct 08 19:46:40 crc kubenswrapper[4988]: I1008 19:46:40.685279 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/54a89dee-6134-4487-a44f-0be9ca5e8d12-scripts\") pod \"cinder-scheduler-0\" (UID: \"54a89dee-6134-4487-a44f-0be9ca5e8d12\") " pod="openstack/cinder-scheduler-0" Oct 08 19:46:40 crc kubenswrapper[4988]: I1008 19:46:40.685320 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/54a89dee-6134-4487-a44f-0be9ca5e8d12-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"54a89dee-6134-4487-a44f-0be9ca5e8d12\") " pod="openstack/cinder-scheduler-0" Oct 08 19:46:40 crc kubenswrapper[4988]: I1008 19:46:40.685321 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/54a89dee-6134-4487-a44f-0be9ca5e8d12-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"54a89dee-6134-4487-a44f-0be9ca5e8d12\") " pod="openstack/cinder-scheduler-0" Oct 08 19:46:40 crc kubenswrapper[4988]: I1008 19:46:40.691984 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/54a89dee-6134-4487-a44f-0be9ca5e8d12-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"54a89dee-6134-4487-a44f-0be9ca5e8d12\") " pod="openstack/cinder-scheduler-0" Oct 08 19:46:40 crc kubenswrapper[4988]: I1008 19:46:40.692960 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/54a89dee-6134-4487-a44f-0be9ca5e8d12-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"54a89dee-6134-4487-a44f-0be9ca5e8d12\") " pod="openstack/cinder-scheduler-0" Oct 08 19:46:40 crc kubenswrapper[4988]: I1008 19:46:40.693574 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/54a89dee-6134-4487-a44f-0be9ca5e8d12-config-data\") pod \"cinder-scheduler-0\" (UID: \"54a89dee-6134-4487-a44f-0be9ca5e8d12\") " pod="openstack/cinder-scheduler-0" Oct 08 19:46:40 crc kubenswrapper[4988]: I1008 19:46:40.700236 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/54a89dee-6134-4487-a44f-0be9ca5e8d12-scripts\") pod \"cinder-scheduler-0\" (UID: \"54a89dee-6134-4487-a44f-0be9ca5e8d12\") " pod="openstack/cinder-scheduler-0" Oct 08 19:46:40 crc kubenswrapper[4988]: I1008 19:46:40.703128 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d9gr6\" (UniqueName: \"kubernetes.io/projected/54a89dee-6134-4487-a44f-0be9ca5e8d12-kube-api-access-d9gr6\") pod \"cinder-scheduler-0\" (UID: \"54a89dee-6134-4487-a44f-0be9ca5e8d12\") " pod="openstack/cinder-scheduler-0" Oct 08 19:46:40 crc kubenswrapper[4988]: I1008 19:46:40.775649 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Oct 08 19:46:41 crc kubenswrapper[4988]: I1008 19:46:41.284883 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 08 19:46:41 crc kubenswrapper[4988]: I1008 19:46:41.935199 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"54a89dee-6134-4487-a44f-0be9ca5e8d12","Type":"ContainerStarted","Data":"1c6dfaab8de2e1a91ba36fbc93c81beba38e40a30262ebb196364fd3d79d1f41"} Oct 08 19:46:42 crc kubenswrapper[4988]: I1008 19:46:42.052501 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Oct 08 19:46:42 crc kubenswrapper[4988]: I1008 19:46:42.053462 4988 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="0c3843ef-c63c-41cb-b3ff-cf264790ca78" containerName="cinder-api-log" containerID="cri-o://ec266c8bc4c08ec14d15ce126d7820ac8681601b18c629e063ece19e3755a2ec" gracePeriod=30 Oct 08 19:46:42 crc kubenswrapper[4988]: I1008 19:46:42.053687 4988 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="0c3843ef-c63c-41cb-b3ff-cf264790ca78" containerName="cinder-api" containerID="cri-o://3b86f1b55753a6ccd8391660045573be16c6a587576932bab614bff4f255ffef" gracePeriod=30 Oct 08 19:46:42 crc kubenswrapper[4988]: I1008 19:46:42.949559 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"54a89dee-6134-4487-a44f-0be9ca5e8d12","Type":"ContainerStarted","Data":"11df53bd3cfed3a4beff71d5b79fe236b4f77a15db2f72fa4786ef8495cdf597"} Oct 08 19:46:42 crc kubenswrapper[4988]: I1008 19:46:42.951208 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"54a89dee-6134-4487-a44f-0be9ca5e8d12","Type":"ContainerStarted","Data":"ca1666f1b723eb9cb65294642b7f95e0417ae640f7a2a716072076d392a8eb72"} Oct 08 19:46:42 crc kubenswrapper[4988]: I1008 19:46:42.952372 4988 generic.go:334] "Generic (PLEG): container finished" podID="0c3843ef-c63c-41cb-b3ff-cf264790ca78" containerID="ec266c8bc4c08ec14d15ce126d7820ac8681601b18c629e063ece19e3755a2ec" exitCode=143 Oct 08 19:46:42 crc kubenswrapper[4988]: I1008 19:46:42.952437 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"0c3843ef-c63c-41cb-b3ff-cf264790ca78","Type":"ContainerDied","Data":"ec266c8bc4c08ec14d15ce126d7820ac8681601b18c629e063ece19e3755a2ec"} Oct 08 19:46:42 crc kubenswrapper[4988]: I1008 19:46:42.978814 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=2.978791031 podStartE2EDuration="2.978791031s" podCreationTimestamp="2025-10-08 19:46:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 19:46:42.974375841 +0000 UTC m=+5748.424218631" watchObservedRunningTime="2025-10-08 19:46:42.978791031 +0000 UTC m=+5748.428633811" Oct 08 19:46:45 crc kubenswrapper[4988]: I1008 19:46:45.731304 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 08 19:46:45 crc kubenswrapper[4988]: I1008 19:46:45.775821 4988 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Oct 08 19:46:45 crc kubenswrapper[4988]: I1008 19:46:45.791428 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0c3843ef-c63c-41cb-b3ff-cf264790ca78-config-data\") pod \"0c3843ef-c63c-41cb-b3ff-cf264790ca78\" (UID: \"0c3843ef-c63c-41cb-b3ff-cf264790ca78\") " Oct 08 19:46:45 crc kubenswrapper[4988]: I1008 19:46:45.791502 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/0c3843ef-c63c-41cb-b3ff-cf264790ca78-config-data-custom\") pod \"0c3843ef-c63c-41cb-b3ff-cf264790ca78\" (UID: \"0c3843ef-c63c-41cb-b3ff-cf264790ca78\") " Oct 08 19:46:45 crc kubenswrapper[4988]: I1008 19:46:45.791537 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0c3843ef-c63c-41cb-b3ff-cf264790ca78-combined-ca-bundle\") pod \"0c3843ef-c63c-41cb-b3ff-cf264790ca78\" (UID: \"0c3843ef-c63c-41cb-b3ff-cf264790ca78\") " Oct 08 19:46:45 crc kubenswrapper[4988]: I1008 19:46:45.791591 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/0c3843ef-c63c-41cb-b3ff-cf264790ca78-etc-machine-id\") pod \"0c3843ef-c63c-41cb-b3ff-cf264790ca78\" (UID: \"0c3843ef-c63c-41cb-b3ff-cf264790ca78\") " Oct 08 19:46:45 crc kubenswrapper[4988]: I1008 19:46:45.791699 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/0c3843ef-c63c-41cb-b3ff-cf264790ca78-public-tls-certs\") pod \"0c3843ef-c63c-41cb-b3ff-cf264790ca78\" (UID: \"0c3843ef-c63c-41cb-b3ff-cf264790ca78\") " Oct 08 19:46:45 crc kubenswrapper[4988]: I1008 19:46:45.791750 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0c3843ef-c63c-41cb-b3ff-cf264790ca78-logs\") pod \"0c3843ef-c63c-41cb-b3ff-cf264790ca78\" (UID: \"0c3843ef-c63c-41cb-b3ff-cf264790ca78\") " Oct 08 19:46:45 crc kubenswrapper[4988]: I1008 19:46:45.791779 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/0c3843ef-c63c-41cb-b3ff-cf264790ca78-internal-tls-certs\") pod \"0c3843ef-c63c-41cb-b3ff-cf264790ca78\" (UID: \"0c3843ef-c63c-41cb-b3ff-cf264790ca78\") " Oct 08 19:46:45 crc kubenswrapper[4988]: I1008 19:46:45.791816 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0c3843ef-c63c-41cb-b3ff-cf264790ca78-scripts\") pod \"0c3843ef-c63c-41cb-b3ff-cf264790ca78\" (UID: \"0c3843ef-c63c-41cb-b3ff-cf264790ca78\") " Oct 08 19:46:45 crc kubenswrapper[4988]: I1008 19:46:45.791854 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x57nt\" (UniqueName: \"kubernetes.io/projected/0c3843ef-c63c-41cb-b3ff-cf264790ca78-kube-api-access-x57nt\") pod \"0c3843ef-c63c-41cb-b3ff-cf264790ca78\" (UID: \"0c3843ef-c63c-41cb-b3ff-cf264790ca78\") " Oct 08 19:46:45 crc kubenswrapper[4988]: I1008 19:46:45.793897 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/0c3843ef-c63c-41cb-b3ff-cf264790ca78-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "0c3843ef-c63c-41cb-b3ff-cf264790ca78" (UID: "0c3843ef-c63c-41cb-b3ff-cf264790ca78"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 08 19:46:45 crc kubenswrapper[4988]: I1008 19:46:45.796424 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0c3843ef-c63c-41cb-b3ff-cf264790ca78-logs" (OuterVolumeSpecName: "logs") pod "0c3843ef-c63c-41cb-b3ff-cf264790ca78" (UID: "0c3843ef-c63c-41cb-b3ff-cf264790ca78"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 19:46:45 crc kubenswrapper[4988]: I1008 19:46:45.802981 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0c3843ef-c63c-41cb-b3ff-cf264790ca78-kube-api-access-x57nt" (OuterVolumeSpecName: "kube-api-access-x57nt") pod "0c3843ef-c63c-41cb-b3ff-cf264790ca78" (UID: "0c3843ef-c63c-41cb-b3ff-cf264790ca78"). InnerVolumeSpecName "kube-api-access-x57nt". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 19:46:45 crc kubenswrapper[4988]: I1008 19:46:45.803505 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0c3843ef-c63c-41cb-b3ff-cf264790ca78-scripts" (OuterVolumeSpecName: "scripts") pod "0c3843ef-c63c-41cb-b3ff-cf264790ca78" (UID: "0c3843ef-c63c-41cb-b3ff-cf264790ca78"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 19:46:45 crc kubenswrapper[4988]: I1008 19:46:45.810103 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0c3843ef-c63c-41cb-b3ff-cf264790ca78-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "0c3843ef-c63c-41cb-b3ff-cf264790ca78" (UID: "0c3843ef-c63c-41cb-b3ff-cf264790ca78"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 19:46:45 crc kubenswrapper[4988]: I1008 19:46:45.854653 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0c3843ef-c63c-41cb-b3ff-cf264790ca78-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "0c3843ef-c63c-41cb-b3ff-cf264790ca78" (UID: "0c3843ef-c63c-41cb-b3ff-cf264790ca78"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 19:46:45 crc kubenswrapper[4988]: I1008 19:46:45.861665 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0c3843ef-c63c-41cb-b3ff-cf264790ca78-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "0c3843ef-c63c-41cb-b3ff-cf264790ca78" (UID: "0c3843ef-c63c-41cb-b3ff-cf264790ca78"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 19:46:45 crc kubenswrapper[4988]: I1008 19:46:45.871006 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0c3843ef-c63c-41cb-b3ff-cf264790ca78-config-data" (OuterVolumeSpecName: "config-data") pod "0c3843ef-c63c-41cb-b3ff-cf264790ca78" (UID: "0c3843ef-c63c-41cb-b3ff-cf264790ca78"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 19:46:45 crc kubenswrapper[4988]: I1008 19:46:45.886774 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0c3843ef-c63c-41cb-b3ff-cf264790ca78-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "0c3843ef-c63c-41cb-b3ff-cf264790ca78" (UID: "0c3843ef-c63c-41cb-b3ff-cf264790ca78"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 19:46:45 crc kubenswrapper[4988]: I1008 19:46:45.894204 4988 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/0c3843ef-c63c-41cb-b3ff-cf264790ca78-config-data-custom\") on node \"crc\" DevicePath \"\"" Oct 08 19:46:45 crc kubenswrapper[4988]: I1008 19:46:45.894921 4988 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0c3843ef-c63c-41cb-b3ff-cf264790ca78-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 19:46:45 crc kubenswrapper[4988]: I1008 19:46:45.894956 4988 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/0c3843ef-c63c-41cb-b3ff-cf264790ca78-etc-machine-id\") on node \"crc\" DevicePath \"\"" Oct 08 19:46:45 crc kubenswrapper[4988]: I1008 19:46:45.894969 4988 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/0c3843ef-c63c-41cb-b3ff-cf264790ca78-public-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 08 19:46:45 crc kubenswrapper[4988]: I1008 19:46:45.894981 4988 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0c3843ef-c63c-41cb-b3ff-cf264790ca78-logs\") on node \"crc\" DevicePath \"\"" Oct 08 19:46:45 crc kubenswrapper[4988]: I1008 19:46:45.894995 4988 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/0c3843ef-c63c-41cb-b3ff-cf264790ca78-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 08 19:46:45 crc kubenswrapper[4988]: I1008 19:46:45.895006 4988 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0c3843ef-c63c-41cb-b3ff-cf264790ca78-scripts\") on node \"crc\" DevicePath \"\"" Oct 08 19:46:45 crc kubenswrapper[4988]: I1008 19:46:45.895017 4988 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x57nt\" (UniqueName: \"kubernetes.io/projected/0c3843ef-c63c-41cb-b3ff-cf264790ca78-kube-api-access-x57nt\") on node \"crc\" DevicePath \"\"" Oct 08 19:46:45 crc kubenswrapper[4988]: I1008 19:46:45.895029 4988 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0c3843ef-c63c-41cb-b3ff-cf264790ca78-config-data\") on node \"crc\" DevicePath \"\"" Oct 08 19:46:46 crc kubenswrapper[4988]: I1008 19:46:46.000365 4988 generic.go:334] "Generic (PLEG): container finished" podID="0c3843ef-c63c-41cb-b3ff-cf264790ca78" containerID="3b86f1b55753a6ccd8391660045573be16c6a587576932bab614bff4f255ffef" exitCode=0 Oct 08 19:46:46 crc kubenswrapper[4988]: I1008 19:46:46.000421 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"0c3843ef-c63c-41cb-b3ff-cf264790ca78","Type":"ContainerDied","Data":"3b86f1b55753a6ccd8391660045573be16c6a587576932bab614bff4f255ffef"} Oct 08 19:46:46 crc kubenswrapper[4988]: I1008 19:46:46.000478 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"0c3843ef-c63c-41cb-b3ff-cf264790ca78","Type":"ContainerDied","Data":"ec34761646d3b54d0a6d731828ba8362f0d430caca1c60af73652e214ebd55fa"} Oct 08 19:46:46 crc kubenswrapper[4988]: I1008 19:46:46.000497 4988 scope.go:117] "RemoveContainer" containerID="3b86f1b55753a6ccd8391660045573be16c6a587576932bab614bff4f255ffef" Oct 08 19:46:46 crc kubenswrapper[4988]: I1008 19:46:46.000724 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 08 19:46:46 crc kubenswrapper[4988]: I1008 19:46:46.020363 4988 scope.go:117] "RemoveContainer" containerID="ec266c8bc4c08ec14d15ce126d7820ac8681601b18c629e063ece19e3755a2ec" Oct 08 19:46:46 crc kubenswrapper[4988]: I1008 19:46:46.030288 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Oct 08 19:46:46 crc kubenswrapper[4988]: I1008 19:46:46.037734 4988 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-api-0"] Oct 08 19:46:46 crc kubenswrapper[4988]: I1008 19:46:46.049411 4988 scope.go:117] "RemoveContainer" containerID="3b86f1b55753a6ccd8391660045573be16c6a587576932bab614bff4f255ffef" Oct 08 19:46:46 crc kubenswrapper[4988]: E1008 19:46:46.049780 4988 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3b86f1b55753a6ccd8391660045573be16c6a587576932bab614bff4f255ffef\": container with ID starting with 3b86f1b55753a6ccd8391660045573be16c6a587576932bab614bff4f255ffef not found: ID does not exist" containerID="3b86f1b55753a6ccd8391660045573be16c6a587576932bab614bff4f255ffef" Oct 08 19:46:46 crc kubenswrapper[4988]: I1008 19:46:46.049834 4988 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3b86f1b55753a6ccd8391660045573be16c6a587576932bab614bff4f255ffef"} err="failed to get container status \"3b86f1b55753a6ccd8391660045573be16c6a587576932bab614bff4f255ffef\": rpc error: code = NotFound desc = could not find container \"3b86f1b55753a6ccd8391660045573be16c6a587576932bab614bff4f255ffef\": container with ID starting with 3b86f1b55753a6ccd8391660045573be16c6a587576932bab614bff4f255ffef not found: ID does not exist" Oct 08 19:46:46 crc kubenswrapper[4988]: I1008 19:46:46.049867 4988 scope.go:117] "RemoveContainer" containerID="ec266c8bc4c08ec14d15ce126d7820ac8681601b18c629e063ece19e3755a2ec" Oct 08 19:46:46 crc kubenswrapper[4988]: E1008 19:46:46.050321 4988 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ec266c8bc4c08ec14d15ce126d7820ac8681601b18c629e063ece19e3755a2ec\": container with ID starting with ec266c8bc4c08ec14d15ce126d7820ac8681601b18c629e063ece19e3755a2ec not found: ID does not exist" containerID="ec266c8bc4c08ec14d15ce126d7820ac8681601b18c629e063ece19e3755a2ec" Oct 08 19:46:46 crc kubenswrapper[4988]: I1008 19:46:46.050550 4988 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ec266c8bc4c08ec14d15ce126d7820ac8681601b18c629e063ece19e3755a2ec"} err="failed to get container status \"ec266c8bc4c08ec14d15ce126d7820ac8681601b18c629e063ece19e3755a2ec\": rpc error: code = NotFound desc = could not find container \"ec266c8bc4c08ec14d15ce126d7820ac8681601b18c629e063ece19e3755a2ec\": container with ID starting with ec266c8bc4c08ec14d15ce126d7820ac8681601b18c629e063ece19e3755a2ec not found: ID does not exist" Oct 08 19:46:46 crc kubenswrapper[4988]: I1008 19:46:46.051731 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Oct 08 19:46:46 crc kubenswrapper[4988]: E1008 19:46:46.052146 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0c3843ef-c63c-41cb-b3ff-cf264790ca78" containerName="cinder-api" Oct 08 19:46:46 crc kubenswrapper[4988]: I1008 19:46:46.052163 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="0c3843ef-c63c-41cb-b3ff-cf264790ca78" containerName="cinder-api" Oct 08 19:46:46 crc kubenswrapper[4988]: E1008 19:46:46.052183 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0c3843ef-c63c-41cb-b3ff-cf264790ca78" containerName="cinder-api-log" Oct 08 19:46:46 crc kubenswrapper[4988]: I1008 19:46:46.052189 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="0c3843ef-c63c-41cb-b3ff-cf264790ca78" containerName="cinder-api-log" Oct 08 19:46:46 crc kubenswrapper[4988]: I1008 19:46:46.052334 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="0c3843ef-c63c-41cb-b3ff-cf264790ca78" containerName="cinder-api" Oct 08 19:46:46 crc kubenswrapper[4988]: I1008 19:46:46.052361 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="0c3843ef-c63c-41cb-b3ff-cf264790ca78" containerName="cinder-api-log" Oct 08 19:46:46 crc kubenswrapper[4988]: I1008 19:46:46.053264 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 08 19:46:46 crc kubenswrapper[4988]: I1008 19:46:46.057070 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-public-svc" Oct 08 19:46:46 crc kubenswrapper[4988]: I1008 19:46:46.057293 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Oct 08 19:46:46 crc kubenswrapper[4988]: I1008 19:46:46.057407 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-internal-svc" Oct 08 19:46:46 crc kubenswrapper[4988]: I1008 19:46:46.098807 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/15e8111d-836a-428e-b58f-b0df3f96c251-scripts\") pod \"cinder-api-0\" (UID: \"15e8111d-836a-428e-b58f-b0df3f96c251\") " pod="openstack/cinder-api-0" Oct 08 19:46:46 crc kubenswrapper[4988]: I1008 19:46:46.098861 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7gprg\" (UniqueName: \"kubernetes.io/projected/15e8111d-836a-428e-b58f-b0df3f96c251-kube-api-access-7gprg\") pod \"cinder-api-0\" (UID: \"15e8111d-836a-428e-b58f-b0df3f96c251\") " pod="openstack/cinder-api-0" Oct 08 19:46:46 crc kubenswrapper[4988]: I1008 19:46:46.098883 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/15e8111d-836a-428e-b58f-b0df3f96c251-config-data-custom\") pod \"cinder-api-0\" (UID: \"15e8111d-836a-428e-b58f-b0df3f96c251\") " pod="openstack/cinder-api-0" Oct 08 19:46:46 crc kubenswrapper[4988]: I1008 19:46:46.098931 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/15e8111d-836a-428e-b58f-b0df3f96c251-logs\") pod \"cinder-api-0\" (UID: \"15e8111d-836a-428e-b58f-b0df3f96c251\") " pod="openstack/cinder-api-0" Oct 08 19:46:46 crc kubenswrapper[4988]: I1008 19:46:46.098951 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/15e8111d-836a-428e-b58f-b0df3f96c251-public-tls-certs\") pod \"cinder-api-0\" (UID: \"15e8111d-836a-428e-b58f-b0df3f96c251\") " pod="openstack/cinder-api-0" Oct 08 19:46:46 crc kubenswrapper[4988]: I1008 19:46:46.098985 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/15e8111d-836a-428e-b58f-b0df3f96c251-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"15e8111d-836a-428e-b58f-b0df3f96c251\") " pod="openstack/cinder-api-0" Oct 08 19:46:46 crc kubenswrapper[4988]: I1008 19:46:46.099007 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/15e8111d-836a-428e-b58f-b0df3f96c251-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"15e8111d-836a-428e-b58f-b0df3f96c251\") " pod="openstack/cinder-api-0" Oct 08 19:46:46 crc kubenswrapper[4988]: I1008 19:46:46.099034 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/15e8111d-836a-428e-b58f-b0df3f96c251-etc-machine-id\") pod \"cinder-api-0\" (UID: \"15e8111d-836a-428e-b58f-b0df3f96c251\") " pod="openstack/cinder-api-0" Oct 08 19:46:46 crc kubenswrapper[4988]: I1008 19:46:46.099064 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/15e8111d-836a-428e-b58f-b0df3f96c251-config-data\") pod \"cinder-api-0\" (UID: \"15e8111d-836a-428e-b58f-b0df3f96c251\") " pod="openstack/cinder-api-0" Oct 08 19:46:46 crc kubenswrapper[4988]: I1008 19:46:46.104506 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Oct 08 19:46:46 crc kubenswrapper[4988]: I1008 19:46:46.200324 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/15e8111d-836a-428e-b58f-b0df3f96c251-scripts\") pod \"cinder-api-0\" (UID: \"15e8111d-836a-428e-b58f-b0df3f96c251\") " pod="openstack/cinder-api-0" Oct 08 19:46:46 crc kubenswrapper[4988]: I1008 19:46:46.200586 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7gprg\" (UniqueName: \"kubernetes.io/projected/15e8111d-836a-428e-b58f-b0df3f96c251-kube-api-access-7gprg\") pod \"cinder-api-0\" (UID: \"15e8111d-836a-428e-b58f-b0df3f96c251\") " pod="openstack/cinder-api-0" Oct 08 19:46:46 crc kubenswrapper[4988]: I1008 19:46:46.200661 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/15e8111d-836a-428e-b58f-b0df3f96c251-config-data-custom\") pod \"cinder-api-0\" (UID: \"15e8111d-836a-428e-b58f-b0df3f96c251\") " pod="openstack/cinder-api-0" Oct 08 19:46:46 crc kubenswrapper[4988]: I1008 19:46:46.200779 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/15e8111d-836a-428e-b58f-b0df3f96c251-logs\") pod \"cinder-api-0\" (UID: \"15e8111d-836a-428e-b58f-b0df3f96c251\") " pod="openstack/cinder-api-0" Oct 08 19:46:46 crc kubenswrapper[4988]: I1008 19:46:46.200888 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/15e8111d-836a-428e-b58f-b0df3f96c251-public-tls-certs\") pod \"cinder-api-0\" (UID: \"15e8111d-836a-428e-b58f-b0df3f96c251\") " pod="openstack/cinder-api-0" Oct 08 19:46:46 crc kubenswrapper[4988]: I1008 19:46:46.201051 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/15e8111d-836a-428e-b58f-b0df3f96c251-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"15e8111d-836a-428e-b58f-b0df3f96c251\") " pod="openstack/cinder-api-0" Oct 08 19:46:46 crc kubenswrapper[4988]: I1008 19:46:46.201163 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/15e8111d-836a-428e-b58f-b0df3f96c251-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"15e8111d-836a-428e-b58f-b0df3f96c251\") " pod="openstack/cinder-api-0" Oct 08 19:46:46 crc kubenswrapper[4988]: I1008 19:46:46.201279 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/15e8111d-836a-428e-b58f-b0df3f96c251-etc-machine-id\") pod \"cinder-api-0\" (UID: \"15e8111d-836a-428e-b58f-b0df3f96c251\") " pod="openstack/cinder-api-0" Oct 08 19:46:46 crc kubenswrapper[4988]: I1008 19:46:46.201403 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/15e8111d-836a-428e-b58f-b0df3f96c251-config-data\") pod \"cinder-api-0\" (UID: \"15e8111d-836a-428e-b58f-b0df3f96c251\") " pod="openstack/cinder-api-0" Oct 08 19:46:46 crc kubenswrapper[4988]: I1008 19:46:46.201404 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/15e8111d-836a-428e-b58f-b0df3f96c251-etc-machine-id\") pod \"cinder-api-0\" (UID: \"15e8111d-836a-428e-b58f-b0df3f96c251\") " pod="openstack/cinder-api-0" Oct 08 19:46:46 crc kubenswrapper[4988]: I1008 19:46:46.201315 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/15e8111d-836a-428e-b58f-b0df3f96c251-logs\") pod \"cinder-api-0\" (UID: \"15e8111d-836a-428e-b58f-b0df3f96c251\") " pod="openstack/cinder-api-0" Oct 08 19:46:46 crc kubenswrapper[4988]: I1008 19:46:46.206520 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/15e8111d-836a-428e-b58f-b0df3f96c251-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"15e8111d-836a-428e-b58f-b0df3f96c251\") " pod="openstack/cinder-api-0" Oct 08 19:46:46 crc kubenswrapper[4988]: I1008 19:46:46.206546 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/15e8111d-836a-428e-b58f-b0df3f96c251-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"15e8111d-836a-428e-b58f-b0df3f96c251\") " pod="openstack/cinder-api-0" Oct 08 19:46:46 crc kubenswrapper[4988]: I1008 19:46:46.207135 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/15e8111d-836a-428e-b58f-b0df3f96c251-config-data-custom\") pod \"cinder-api-0\" (UID: \"15e8111d-836a-428e-b58f-b0df3f96c251\") " pod="openstack/cinder-api-0" Oct 08 19:46:46 crc kubenswrapper[4988]: I1008 19:46:46.207740 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/15e8111d-836a-428e-b58f-b0df3f96c251-scripts\") pod \"cinder-api-0\" (UID: \"15e8111d-836a-428e-b58f-b0df3f96c251\") " pod="openstack/cinder-api-0" Oct 08 19:46:46 crc kubenswrapper[4988]: I1008 19:46:46.208064 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/15e8111d-836a-428e-b58f-b0df3f96c251-public-tls-certs\") pod \"cinder-api-0\" (UID: \"15e8111d-836a-428e-b58f-b0df3f96c251\") " pod="openstack/cinder-api-0" Oct 08 19:46:46 crc kubenswrapper[4988]: I1008 19:46:46.209023 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/15e8111d-836a-428e-b58f-b0df3f96c251-config-data\") pod \"cinder-api-0\" (UID: \"15e8111d-836a-428e-b58f-b0df3f96c251\") " pod="openstack/cinder-api-0" Oct 08 19:46:46 crc kubenswrapper[4988]: I1008 19:46:46.217715 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7gprg\" (UniqueName: \"kubernetes.io/projected/15e8111d-836a-428e-b58f-b0df3f96c251-kube-api-access-7gprg\") pod \"cinder-api-0\" (UID: \"15e8111d-836a-428e-b58f-b0df3f96c251\") " pod="openstack/cinder-api-0" Oct 08 19:46:46 crc kubenswrapper[4988]: I1008 19:46:46.377284 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 08 19:46:46 crc kubenswrapper[4988]: I1008 19:46:46.728256 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Oct 08 19:46:46 crc kubenswrapper[4988]: W1008 19:46:46.732673 4988 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod15e8111d_836a_428e_b58f_b0df3f96c251.slice/crio-394d1865f7fb57bd93db808bf417915d0a55510384454da9863c03cf1ab7dc3d WatchSource:0}: Error finding container 394d1865f7fb57bd93db808bf417915d0a55510384454da9863c03cf1ab7dc3d: Status 404 returned error can't find the container with id 394d1865f7fb57bd93db808bf417915d0a55510384454da9863c03cf1ab7dc3d Oct 08 19:46:47 crc kubenswrapper[4988]: I1008 19:46:47.015294 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"15e8111d-836a-428e-b58f-b0df3f96c251","Type":"ContainerStarted","Data":"394d1865f7fb57bd93db808bf417915d0a55510384454da9863c03cf1ab7dc3d"} Oct 08 19:46:47 crc kubenswrapper[4988]: I1008 19:46:47.257979 4988 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0c3843ef-c63c-41cb-b3ff-cf264790ca78" path="/var/lib/kubelet/pods/0c3843ef-c63c-41cb-b3ff-cf264790ca78/volumes" Oct 08 19:46:48 crc kubenswrapper[4988]: I1008 19:46:48.028688 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"15e8111d-836a-428e-b58f-b0df3f96c251","Type":"ContainerStarted","Data":"c850697811c4aba1f1bfc5debf97758ee0e604732426a237fe6d0bd365952a78"} Oct 08 19:46:49 crc kubenswrapper[4988]: I1008 19:46:49.047468 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"15e8111d-836a-428e-b58f-b0df3f96c251","Type":"ContainerStarted","Data":"1b34ae0006dd3e90f412e4772bf821cf0799decd35bdf8f5d0a0f8db9d633fc7"} Oct 08 19:46:49 crc kubenswrapper[4988]: I1008 19:46:49.047944 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Oct 08 19:46:49 crc kubenswrapper[4988]: I1008 19:46:49.097410 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=3.097368548 podStartE2EDuration="3.097368548s" podCreationTimestamp="2025-10-08 19:46:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 19:46:49.082139764 +0000 UTC m=+5754.531982614" watchObservedRunningTime="2025-10-08 19:46:49.097368548 +0000 UTC m=+5754.547211328" Oct 08 19:46:50 crc kubenswrapper[4988]: I1008 19:46:50.238195 4988 scope.go:117] "RemoveContainer" containerID="f2a29ed22dd7a48bbc929e65fbeceed969a669e78f65fbfb724b03dc0db6e947" Oct 08 19:46:50 crc kubenswrapper[4988]: E1008 19:46:50.238762 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nm54f_openshift-machine-config-operator(f328a20b-5283-407e-8a93-2ab2da599f18)\"" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" Oct 08 19:46:51 crc kubenswrapper[4988]: I1008 19:46:51.034010 4988 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Oct 08 19:46:51 crc kubenswrapper[4988]: I1008 19:46:51.071212 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 08 19:46:51 crc kubenswrapper[4988]: I1008 19:46:51.073190 4988 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="54a89dee-6134-4487-a44f-0be9ca5e8d12" containerName="cinder-scheduler" containerID="cri-o://ca1666f1b723eb9cb65294642b7f95e0417ae640f7a2a716072076d392a8eb72" gracePeriod=30 Oct 08 19:46:51 crc kubenswrapper[4988]: I1008 19:46:51.073679 4988 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="54a89dee-6134-4487-a44f-0be9ca5e8d12" containerName="probe" containerID="cri-o://11df53bd3cfed3a4beff71d5b79fe236b4f77a15db2f72fa4786ef8495cdf597" gracePeriod=30 Oct 08 19:46:52 crc kubenswrapper[4988]: I1008 19:46:52.088595 4988 generic.go:334] "Generic (PLEG): container finished" podID="54a89dee-6134-4487-a44f-0be9ca5e8d12" containerID="11df53bd3cfed3a4beff71d5b79fe236b4f77a15db2f72fa4786ef8495cdf597" exitCode=0 Oct 08 19:46:52 crc kubenswrapper[4988]: I1008 19:46:52.088683 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"54a89dee-6134-4487-a44f-0be9ca5e8d12","Type":"ContainerDied","Data":"11df53bd3cfed3a4beff71d5b79fe236b4f77a15db2f72fa4786ef8495cdf597"} Oct 08 19:46:52 crc kubenswrapper[4988]: I1008 19:46:52.980169 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Oct 08 19:46:53 crc kubenswrapper[4988]: I1008 19:46:53.050007 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/54a89dee-6134-4487-a44f-0be9ca5e8d12-config-data-custom\") pod \"54a89dee-6134-4487-a44f-0be9ca5e8d12\" (UID: \"54a89dee-6134-4487-a44f-0be9ca5e8d12\") " Oct 08 19:46:53 crc kubenswrapper[4988]: I1008 19:46:53.050081 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d9gr6\" (UniqueName: \"kubernetes.io/projected/54a89dee-6134-4487-a44f-0be9ca5e8d12-kube-api-access-d9gr6\") pod \"54a89dee-6134-4487-a44f-0be9ca5e8d12\" (UID: \"54a89dee-6134-4487-a44f-0be9ca5e8d12\") " Oct 08 19:46:53 crc kubenswrapper[4988]: I1008 19:46:53.050163 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/54a89dee-6134-4487-a44f-0be9ca5e8d12-etc-machine-id\") pod \"54a89dee-6134-4487-a44f-0be9ca5e8d12\" (UID: \"54a89dee-6134-4487-a44f-0be9ca5e8d12\") " Oct 08 19:46:53 crc kubenswrapper[4988]: I1008 19:46:53.050281 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/54a89dee-6134-4487-a44f-0be9ca5e8d12-scripts\") pod \"54a89dee-6134-4487-a44f-0be9ca5e8d12\" (UID: \"54a89dee-6134-4487-a44f-0be9ca5e8d12\") " Oct 08 19:46:53 crc kubenswrapper[4988]: I1008 19:46:53.050363 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/54a89dee-6134-4487-a44f-0be9ca5e8d12-config-data\") pod \"54a89dee-6134-4487-a44f-0be9ca5e8d12\" (UID: \"54a89dee-6134-4487-a44f-0be9ca5e8d12\") " Oct 08 19:46:53 crc kubenswrapper[4988]: I1008 19:46:53.050464 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/54a89dee-6134-4487-a44f-0be9ca5e8d12-combined-ca-bundle\") pod \"54a89dee-6134-4487-a44f-0be9ca5e8d12\" (UID: \"54a89dee-6134-4487-a44f-0be9ca5e8d12\") " Oct 08 19:46:53 crc kubenswrapper[4988]: I1008 19:46:53.051813 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/54a89dee-6134-4487-a44f-0be9ca5e8d12-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "54a89dee-6134-4487-a44f-0be9ca5e8d12" (UID: "54a89dee-6134-4487-a44f-0be9ca5e8d12"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 08 19:46:53 crc kubenswrapper[4988]: I1008 19:46:53.068073 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/54a89dee-6134-4487-a44f-0be9ca5e8d12-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "54a89dee-6134-4487-a44f-0be9ca5e8d12" (UID: "54a89dee-6134-4487-a44f-0be9ca5e8d12"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 19:46:53 crc kubenswrapper[4988]: I1008 19:46:53.068121 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/54a89dee-6134-4487-a44f-0be9ca5e8d12-scripts" (OuterVolumeSpecName: "scripts") pod "54a89dee-6134-4487-a44f-0be9ca5e8d12" (UID: "54a89dee-6134-4487-a44f-0be9ca5e8d12"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 19:46:53 crc kubenswrapper[4988]: I1008 19:46:53.068131 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/54a89dee-6134-4487-a44f-0be9ca5e8d12-kube-api-access-d9gr6" (OuterVolumeSpecName: "kube-api-access-d9gr6") pod "54a89dee-6134-4487-a44f-0be9ca5e8d12" (UID: "54a89dee-6134-4487-a44f-0be9ca5e8d12"). InnerVolumeSpecName "kube-api-access-d9gr6". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 19:46:53 crc kubenswrapper[4988]: I1008 19:46:53.112169 4988 generic.go:334] "Generic (PLEG): container finished" podID="54a89dee-6134-4487-a44f-0be9ca5e8d12" containerID="ca1666f1b723eb9cb65294642b7f95e0417ae640f7a2a716072076d392a8eb72" exitCode=0 Oct 08 19:46:53 crc kubenswrapper[4988]: I1008 19:46:53.112211 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"54a89dee-6134-4487-a44f-0be9ca5e8d12","Type":"ContainerDied","Data":"ca1666f1b723eb9cb65294642b7f95e0417ae640f7a2a716072076d392a8eb72"} Oct 08 19:46:53 crc kubenswrapper[4988]: I1008 19:46:53.112243 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"54a89dee-6134-4487-a44f-0be9ca5e8d12","Type":"ContainerDied","Data":"1c6dfaab8de2e1a91ba36fbc93c81beba38e40a30262ebb196364fd3d79d1f41"} Oct 08 19:46:53 crc kubenswrapper[4988]: I1008 19:46:53.112264 4988 scope.go:117] "RemoveContainer" containerID="11df53bd3cfed3a4beff71d5b79fe236b4f77a15db2f72fa4786ef8495cdf597" Oct 08 19:46:53 crc kubenswrapper[4988]: I1008 19:46:53.112285 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Oct 08 19:46:53 crc kubenswrapper[4988]: I1008 19:46:53.147548 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/54a89dee-6134-4487-a44f-0be9ca5e8d12-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "54a89dee-6134-4487-a44f-0be9ca5e8d12" (UID: "54a89dee-6134-4487-a44f-0be9ca5e8d12"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 19:46:53 crc kubenswrapper[4988]: I1008 19:46:53.153139 4988 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/54a89dee-6134-4487-a44f-0be9ca5e8d12-config-data-custom\") on node \"crc\" DevicePath \"\"" Oct 08 19:46:53 crc kubenswrapper[4988]: I1008 19:46:53.153170 4988 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d9gr6\" (UniqueName: \"kubernetes.io/projected/54a89dee-6134-4487-a44f-0be9ca5e8d12-kube-api-access-d9gr6\") on node \"crc\" DevicePath \"\"" Oct 08 19:46:53 crc kubenswrapper[4988]: I1008 19:46:53.153181 4988 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/54a89dee-6134-4487-a44f-0be9ca5e8d12-etc-machine-id\") on node \"crc\" DevicePath \"\"" Oct 08 19:46:53 crc kubenswrapper[4988]: I1008 19:46:53.153190 4988 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/54a89dee-6134-4487-a44f-0be9ca5e8d12-scripts\") on node \"crc\" DevicePath \"\"" Oct 08 19:46:53 crc kubenswrapper[4988]: I1008 19:46:53.153202 4988 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/54a89dee-6134-4487-a44f-0be9ca5e8d12-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 19:46:53 crc kubenswrapper[4988]: I1008 19:46:53.173690 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/54a89dee-6134-4487-a44f-0be9ca5e8d12-config-data" (OuterVolumeSpecName: "config-data") pod "54a89dee-6134-4487-a44f-0be9ca5e8d12" (UID: "54a89dee-6134-4487-a44f-0be9ca5e8d12"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 19:46:53 crc kubenswrapper[4988]: I1008 19:46:53.254363 4988 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/54a89dee-6134-4487-a44f-0be9ca5e8d12-config-data\") on node \"crc\" DevicePath \"\"" Oct 08 19:46:53 crc kubenswrapper[4988]: I1008 19:46:53.256473 4988 scope.go:117] "RemoveContainer" containerID="ca1666f1b723eb9cb65294642b7f95e0417ae640f7a2a716072076d392a8eb72" Oct 08 19:46:53 crc kubenswrapper[4988]: I1008 19:46:53.272710 4988 scope.go:117] "RemoveContainer" containerID="11df53bd3cfed3a4beff71d5b79fe236b4f77a15db2f72fa4786ef8495cdf597" Oct 08 19:46:53 crc kubenswrapper[4988]: E1008 19:46:53.273126 4988 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"11df53bd3cfed3a4beff71d5b79fe236b4f77a15db2f72fa4786ef8495cdf597\": container with ID starting with 11df53bd3cfed3a4beff71d5b79fe236b4f77a15db2f72fa4786ef8495cdf597 not found: ID does not exist" containerID="11df53bd3cfed3a4beff71d5b79fe236b4f77a15db2f72fa4786ef8495cdf597" Oct 08 19:46:53 crc kubenswrapper[4988]: I1008 19:46:53.273165 4988 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"11df53bd3cfed3a4beff71d5b79fe236b4f77a15db2f72fa4786ef8495cdf597"} err="failed to get container status \"11df53bd3cfed3a4beff71d5b79fe236b4f77a15db2f72fa4786ef8495cdf597\": rpc error: code = NotFound desc = could not find container \"11df53bd3cfed3a4beff71d5b79fe236b4f77a15db2f72fa4786ef8495cdf597\": container with ID starting with 11df53bd3cfed3a4beff71d5b79fe236b4f77a15db2f72fa4786ef8495cdf597 not found: ID does not exist" Oct 08 19:46:53 crc kubenswrapper[4988]: I1008 19:46:53.273190 4988 scope.go:117] "RemoveContainer" containerID="ca1666f1b723eb9cb65294642b7f95e0417ae640f7a2a716072076d392a8eb72" Oct 08 19:46:53 crc kubenswrapper[4988]: E1008 19:46:53.273471 4988 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ca1666f1b723eb9cb65294642b7f95e0417ae640f7a2a716072076d392a8eb72\": container with ID starting with ca1666f1b723eb9cb65294642b7f95e0417ae640f7a2a716072076d392a8eb72 not found: ID does not exist" containerID="ca1666f1b723eb9cb65294642b7f95e0417ae640f7a2a716072076d392a8eb72" Oct 08 19:46:53 crc kubenswrapper[4988]: I1008 19:46:53.273508 4988 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ca1666f1b723eb9cb65294642b7f95e0417ae640f7a2a716072076d392a8eb72"} err="failed to get container status \"ca1666f1b723eb9cb65294642b7f95e0417ae640f7a2a716072076d392a8eb72\": rpc error: code = NotFound desc = could not find container \"ca1666f1b723eb9cb65294642b7f95e0417ae640f7a2a716072076d392a8eb72\": container with ID starting with ca1666f1b723eb9cb65294642b7f95e0417ae640f7a2a716072076d392a8eb72 not found: ID does not exist" Oct 08 19:46:53 crc kubenswrapper[4988]: I1008 19:46:53.453419 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 08 19:46:53 crc kubenswrapper[4988]: I1008 19:46:53.467856 4988 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 08 19:46:53 crc kubenswrapper[4988]: I1008 19:46:53.508989 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Oct 08 19:46:53 crc kubenswrapper[4988]: E1008 19:46:53.511288 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="54a89dee-6134-4487-a44f-0be9ca5e8d12" containerName="cinder-scheduler" Oct 08 19:46:53 crc kubenswrapper[4988]: I1008 19:46:53.511341 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="54a89dee-6134-4487-a44f-0be9ca5e8d12" containerName="cinder-scheduler" Oct 08 19:46:53 crc kubenswrapper[4988]: E1008 19:46:53.516698 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="54a89dee-6134-4487-a44f-0be9ca5e8d12" containerName="probe" Oct 08 19:46:53 crc kubenswrapper[4988]: I1008 19:46:53.516745 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="54a89dee-6134-4487-a44f-0be9ca5e8d12" containerName="probe" Oct 08 19:46:53 crc kubenswrapper[4988]: I1008 19:46:53.517243 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="54a89dee-6134-4487-a44f-0be9ca5e8d12" containerName="cinder-scheduler" Oct 08 19:46:53 crc kubenswrapper[4988]: I1008 19:46:53.517310 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="54a89dee-6134-4487-a44f-0be9ca5e8d12" containerName="probe" Oct 08 19:46:53 crc kubenswrapper[4988]: I1008 19:46:53.519457 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Oct 08 19:46:53 crc kubenswrapper[4988]: I1008 19:46:53.524134 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Oct 08 19:46:53 crc kubenswrapper[4988]: I1008 19:46:53.534926 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 08 19:46:53 crc kubenswrapper[4988]: I1008 19:46:53.664001 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/95310a64-9176-420b-aca3-e97e6c821eed-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"95310a64-9176-420b-aca3-e97e6c821eed\") " pod="openstack/cinder-scheduler-0" Oct 08 19:46:53 crc kubenswrapper[4988]: I1008 19:46:53.664515 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/95310a64-9176-420b-aca3-e97e6c821eed-config-data\") pod \"cinder-scheduler-0\" (UID: \"95310a64-9176-420b-aca3-e97e6c821eed\") " pod="openstack/cinder-scheduler-0" Oct 08 19:46:53 crc kubenswrapper[4988]: I1008 19:46:53.664631 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/95310a64-9176-420b-aca3-e97e6c821eed-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"95310a64-9176-420b-aca3-e97e6c821eed\") " pod="openstack/cinder-scheduler-0" Oct 08 19:46:53 crc kubenswrapper[4988]: I1008 19:46:53.664739 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-swhms\" (UniqueName: \"kubernetes.io/projected/95310a64-9176-420b-aca3-e97e6c821eed-kube-api-access-swhms\") pod \"cinder-scheduler-0\" (UID: \"95310a64-9176-420b-aca3-e97e6c821eed\") " pod="openstack/cinder-scheduler-0" Oct 08 19:46:53 crc kubenswrapper[4988]: I1008 19:46:53.664826 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/95310a64-9176-420b-aca3-e97e6c821eed-scripts\") pod \"cinder-scheduler-0\" (UID: \"95310a64-9176-420b-aca3-e97e6c821eed\") " pod="openstack/cinder-scheduler-0" Oct 08 19:46:53 crc kubenswrapper[4988]: I1008 19:46:53.664879 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/95310a64-9176-420b-aca3-e97e6c821eed-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"95310a64-9176-420b-aca3-e97e6c821eed\") " pod="openstack/cinder-scheduler-0" Oct 08 19:46:53 crc kubenswrapper[4988]: I1008 19:46:53.766430 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/95310a64-9176-420b-aca3-e97e6c821eed-config-data\") pod \"cinder-scheduler-0\" (UID: \"95310a64-9176-420b-aca3-e97e6c821eed\") " pod="openstack/cinder-scheduler-0" Oct 08 19:46:53 crc kubenswrapper[4988]: I1008 19:46:53.766556 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/95310a64-9176-420b-aca3-e97e6c821eed-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"95310a64-9176-420b-aca3-e97e6c821eed\") " pod="openstack/cinder-scheduler-0" Oct 08 19:46:53 crc kubenswrapper[4988]: I1008 19:46:53.766639 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-swhms\" (UniqueName: \"kubernetes.io/projected/95310a64-9176-420b-aca3-e97e6c821eed-kube-api-access-swhms\") pod \"cinder-scheduler-0\" (UID: \"95310a64-9176-420b-aca3-e97e6c821eed\") " pod="openstack/cinder-scheduler-0" Oct 08 19:46:53 crc kubenswrapper[4988]: I1008 19:46:53.766739 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/95310a64-9176-420b-aca3-e97e6c821eed-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"95310a64-9176-420b-aca3-e97e6c821eed\") " pod="openstack/cinder-scheduler-0" Oct 08 19:46:53 crc kubenswrapper[4988]: I1008 19:46:53.766748 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/95310a64-9176-420b-aca3-e97e6c821eed-scripts\") pod \"cinder-scheduler-0\" (UID: \"95310a64-9176-420b-aca3-e97e6c821eed\") " pod="openstack/cinder-scheduler-0" Oct 08 19:46:53 crc kubenswrapper[4988]: I1008 19:46:53.767092 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/95310a64-9176-420b-aca3-e97e6c821eed-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"95310a64-9176-420b-aca3-e97e6c821eed\") " pod="openstack/cinder-scheduler-0" Oct 08 19:46:53 crc kubenswrapper[4988]: I1008 19:46:53.767161 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/95310a64-9176-420b-aca3-e97e6c821eed-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"95310a64-9176-420b-aca3-e97e6c821eed\") " pod="openstack/cinder-scheduler-0" Oct 08 19:46:53 crc kubenswrapper[4988]: I1008 19:46:53.770800 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/95310a64-9176-420b-aca3-e97e6c821eed-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"95310a64-9176-420b-aca3-e97e6c821eed\") " pod="openstack/cinder-scheduler-0" Oct 08 19:46:53 crc kubenswrapper[4988]: I1008 19:46:53.771474 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/95310a64-9176-420b-aca3-e97e6c821eed-config-data\") pod \"cinder-scheduler-0\" (UID: \"95310a64-9176-420b-aca3-e97e6c821eed\") " pod="openstack/cinder-scheduler-0" Oct 08 19:46:53 crc kubenswrapper[4988]: I1008 19:46:53.773438 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/95310a64-9176-420b-aca3-e97e6c821eed-scripts\") pod \"cinder-scheduler-0\" (UID: \"95310a64-9176-420b-aca3-e97e6c821eed\") " pod="openstack/cinder-scheduler-0" Oct 08 19:46:53 crc kubenswrapper[4988]: I1008 19:46:53.775151 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/95310a64-9176-420b-aca3-e97e6c821eed-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"95310a64-9176-420b-aca3-e97e6c821eed\") " pod="openstack/cinder-scheduler-0" Oct 08 19:46:53 crc kubenswrapper[4988]: I1008 19:46:53.787745 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-swhms\" (UniqueName: \"kubernetes.io/projected/95310a64-9176-420b-aca3-e97e6c821eed-kube-api-access-swhms\") pod \"cinder-scheduler-0\" (UID: \"95310a64-9176-420b-aca3-e97e6c821eed\") " pod="openstack/cinder-scheduler-0" Oct 08 19:46:53 crc kubenswrapper[4988]: I1008 19:46:53.854331 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Oct 08 19:46:54 crc kubenswrapper[4988]: W1008 19:46:54.189503 4988 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod95310a64_9176_420b_aca3_e97e6c821eed.slice/crio-cbd40e78d6e60b2c50fd31605d1ddd2473795d0586b92d11561a0b5459e3153d WatchSource:0}: Error finding container cbd40e78d6e60b2c50fd31605d1ddd2473795d0586b92d11561a0b5459e3153d: Status 404 returned error can't find the container with id cbd40e78d6e60b2c50fd31605d1ddd2473795d0586b92d11561a0b5459e3153d Oct 08 19:46:54 crc kubenswrapper[4988]: I1008 19:46:54.191358 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 08 19:46:55 crc kubenswrapper[4988]: I1008 19:46:55.164140 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"95310a64-9176-420b-aca3-e97e6c821eed","Type":"ContainerStarted","Data":"e758a12788d2844f773491a6314bd4f8b28368d1a3e23f1a0921094d7cc36f41"} Oct 08 19:46:55 crc kubenswrapper[4988]: I1008 19:46:55.165091 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"95310a64-9176-420b-aca3-e97e6c821eed","Type":"ContainerStarted","Data":"cbd40e78d6e60b2c50fd31605d1ddd2473795d0586b92d11561a0b5459e3153d"} Oct 08 19:46:55 crc kubenswrapper[4988]: I1008 19:46:55.250617 4988 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="54a89dee-6134-4487-a44f-0be9ca5e8d12" path="/var/lib/kubelet/pods/54a89dee-6134-4487-a44f-0be9ca5e8d12/volumes" Oct 08 19:46:56 crc kubenswrapper[4988]: I1008 19:46:56.178281 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"95310a64-9176-420b-aca3-e97e6c821eed","Type":"ContainerStarted","Data":"32aa9ca844058098b30990b941af9907fa4d36522a0884f0716ca047d88ede17"} Oct 08 19:46:56 crc kubenswrapper[4988]: I1008 19:46:56.204255 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=3.204237649 podStartE2EDuration="3.204237649s" podCreationTimestamp="2025-10-08 19:46:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 19:46:56.199763928 +0000 UTC m=+5761.649606688" watchObservedRunningTime="2025-10-08 19:46:56.204237649 +0000 UTC m=+5761.654080419" Oct 08 19:46:58 crc kubenswrapper[4988]: I1008 19:46:58.199087 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/cinder-api-0" Oct 08 19:46:58 crc kubenswrapper[4988]: I1008 19:46:58.855665 4988 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Oct 08 19:47:03 crc kubenswrapper[4988]: I1008 19:47:03.934815 4988 scope.go:117] "RemoveContainer" containerID="b72558f2ab7cfdd09ca75b37174b9784d823a4620a4720c222b2857202ad4232" Oct 08 19:47:04 crc kubenswrapper[4988]: I1008 19:47:04.073349 4988 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Oct 08 19:47:04 crc kubenswrapper[4988]: I1008 19:47:04.237375 4988 scope.go:117] "RemoveContainer" containerID="f2a29ed22dd7a48bbc929e65fbeceed969a669e78f65fbfb724b03dc0db6e947" Oct 08 19:47:04 crc kubenswrapper[4988]: E1008 19:47:04.237733 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nm54f_openshift-machine-config-operator(f328a20b-5283-407e-8a93-2ab2da599f18)\"" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" Oct 08 19:47:07 crc kubenswrapper[4988]: I1008 19:47:07.170150 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-create-dgvjr"] Oct 08 19:47:07 crc kubenswrapper[4988]: I1008 19:47:07.172460 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-dgvjr" Oct 08 19:47:07 crc kubenswrapper[4988]: I1008 19:47:07.181347 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-dgvjr"] Oct 08 19:47:07 crc kubenswrapper[4988]: I1008 19:47:07.274175 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nzbs6\" (UniqueName: \"kubernetes.io/projected/6effefe6-da26-4fbd-b3d2-099f884b60b6-kube-api-access-nzbs6\") pod \"glance-db-create-dgvjr\" (UID: \"6effefe6-da26-4fbd-b3d2-099f884b60b6\") " pod="openstack/glance-db-create-dgvjr" Oct 08 19:47:07 crc kubenswrapper[4988]: I1008 19:47:07.376337 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nzbs6\" (UniqueName: \"kubernetes.io/projected/6effefe6-da26-4fbd-b3d2-099f884b60b6-kube-api-access-nzbs6\") pod \"glance-db-create-dgvjr\" (UID: \"6effefe6-da26-4fbd-b3d2-099f884b60b6\") " pod="openstack/glance-db-create-dgvjr" Oct 08 19:47:07 crc kubenswrapper[4988]: I1008 19:47:07.401110 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nzbs6\" (UniqueName: \"kubernetes.io/projected/6effefe6-da26-4fbd-b3d2-099f884b60b6-kube-api-access-nzbs6\") pod \"glance-db-create-dgvjr\" (UID: \"6effefe6-da26-4fbd-b3d2-099f884b60b6\") " pod="openstack/glance-db-create-dgvjr" Oct 08 19:47:07 crc kubenswrapper[4988]: I1008 19:47:07.501590 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-dgvjr" Oct 08 19:47:08 crc kubenswrapper[4988]: I1008 19:47:08.006074 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-dgvjr"] Oct 08 19:47:08 crc kubenswrapper[4988]: I1008 19:47:08.305197 4988 generic.go:334] "Generic (PLEG): container finished" podID="6effefe6-da26-4fbd-b3d2-099f884b60b6" containerID="8940dc864fbeee9eb4739fbf376240780589859f1101f8442e27e18082c4f5fa" exitCode=0 Oct 08 19:47:08 crc kubenswrapper[4988]: I1008 19:47:08.305266 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-dgvjr" event={"ID":"6effefe6-da26-4fbd-b3d2-099f884b60b6","Type":"ContainerDied","Data":"8940dc864fbeee9eb4739fbf376240780589859f1101f8442e27e18082c4f5fa"} Oct 08 19:47:08 crc kubenswrapper[4988]: I1008 19:47:08.305310 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-dgvjr" event={"ID":"6effefe6-da26-4fbd-b3d2-099f884b60b6","Type":"ContainerStarted","Data":"f2b4ae27ceb67f0f1a8d7c709da7d2e40bcbcddb0841a8886a6d5052bdec732e"} Oct 08 19:47:09 crc kubenswrapper[4988]: I1008 19:47:09.709801 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-dgvjr" Oct 08 19:47:09 crc kubenswrapper[4988]: I1008 19:47:09.828263 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nzbs6\" (UniqueName: \"kubernetes.io/projected/6effefe6-da26-4fbd-b3d2-099f884b60b6-kube-api-access-nzbs6\") pod \"6effefe6-da26-4fbd-b3d2-099f884b60b6\" (UID: \"6effefe6-da26-4fbd-b3d2-099f884b60b6\") " Oct 08 19:47:09 crc kubenswrapper[4988]: I1008 19:47:09.840601 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6effefe6-da26-4fbd-b3d2-099f884b60b6-kube-api-access-nzbs6" (OuterVolumeSpecName: "kube-api-access-nzbs6") pod "6effefe6-da26-4fbd-b3d2-099f884b60b6" (UID: "6effefe6-da26-4fbd-b3d2-099f884b60b6"). InnerVolumeSpecName "kube-api-access-nzbs6". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 19:47:09 crc kubenswrapper[4988]: I1008 19:47:09.931679 4988 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nzbs6\" (UniqueName: \"kubernetes.io/projected/6effefe6-da26-4fbd-b3d2-099f884b60b6-kube-api-access-nzbs6\") on node \"crc\" DevicePath \"\"" Oct 08 19:47:10 crc kubenswrapper[4988]: I1008 19:47:10.328561 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-dgvjr" event={"ID":"6effefe6-da26-4fbd-b3d2-099f884b60b6","Type":"ContainerDied","Data":"f2b4ae27ceb67f0f1a8d7c709da7d2e40bcbcddb0841a8886a6d5052bdec732e"} Oct 08 19:47:10 crc kubenswrapper[4988]: I1008 19:47:10.328611 4988 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f2b4ae27ceb67f0f1a8d7c709da7d2e40bcbcddb0841a8886a6d5052bdec732e" Oct 08 19:47:10 crc kubenswrapper[4988]: I1008 19:47:10.328643 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-dgvjr" Oct 08 19:47:17 crc kubenswrapper[4988]: I1008 19:47:17.238013 4988 scope.go:117] "RemoveContainer" containerID="f2a29ed22dd7a48bbc929e65fbeceed969a669e78f65fbfb724b03dc0db6e947" Oct 08 19:47:17 crc kubenswrapper[4988]: E1008 19:47:17.239050 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nm54f_openshift-machine-config-operator(f328a20b-5283-407e-8a93-2ab2da599f18)\"" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" Oct 08 19:47:17 crc kubenswrapper[4988]: I1008 19:47:17.347022 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-a1b3-account-create-kzgsr"] Oct 08 19:47:17 crc kubenswrapper[4988]: E1008 19:47:17.347666 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6effefe6-da26-4fbd-b3d2-099f884b60b6" containerName="mariadb-database-create" Oct 08 19:47:17 crc kubenswrapper[4988]: I1008 19:47:17.347698 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="6effefe6-da26-4fbd-b3d2-099f884b60b6" containerName="mariadb-database-create" Oct 08 19:47:17 crc kubenswrapper[4988]: I1008 19:47:17.347970 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="6effefe6-da26-4fbd-b3d2-099f884b60b6" containerName="mariadb-database-create" Oct 08 19:47:17 crc kubenswrapper[4988]: I1008 19:47:17.348896 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-a1b3-account-create-kzgsr" Oct 08 19:47:17 crc kubenswrapper[4988]: I1008 19:47:17.353025 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-db-secret" Oct 08 19:47:17 crc kubenswrapper[4988]: I1008 19:47:17.360708 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-a1b3-account-create-kzgsr"] Oct 08 19:47:17 crc kubenswrapper[4988]: I1008 19:47:17.479933 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-stv5z\" (UniqueName: \"kubernetes.io/projected/869f964f-941e-4127-95cd-628c2e610300-kube-api-access-stv5z\") pod \"glance-a1b3-account-create-kzgsr\" (UID: \"869f964f-941e-4127-95cd-628c2e610300\") " pod="openstack/glance-a1b3-account-create-kzgsr" Oct 08 19:47:17 crc kubenswrapper[4988]: I1008 19:47:17.582490 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-stv5z\" (UniqueName: \"kubernetes.io/projected/869f964f-941e-4127-95cd-628c2e610300-kube-api-access-stv5z\") pod \"glance-a1b3-account-create-kzgsr\" (UID: \"869f964f-941e-4127-95cd-628c2e610300\") " pod="openstack/glance-a1b3-account-create-kzgsr" Oct 08 19:47:17 crc kubenswrapper[4988]: I1008 19:47:17.617417 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-stv5z\" (UniqueName: \"kubernetes.io/projected/869f964f-941e-4127-95cd-628c2e610300-kube-api-access-stv5z\") pod \"glance-a1b3-account-create-kzgsr\" (UID: \"869f964f-941e-4127-95cd-628c2e610300\") " pod="openstack/glance-a1b3-account-create-kzgsr" Oct 08 19:47:17 crc kubenswrapper[4988]: I1008 19:47:17.689091 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-a1b3-account-create-kzgsr" Oct 08 19:47:18 crc kubenswrapper[4988]: I1008 19:47:18.225890 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-a1b3-account-create-kzgsr"] Oct 08 19:47:18 crc kubenswrapper[4988]: W1008 19:47:18.228302 4988 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod869f964f_941e_4127_95cd_628c2e610300.slice/crio-17d5dc54b55ec49b8e109de7aa4825edaacc1b62ca172130e65f87b0f4aba1cc WatchSource:0}: Error finding container 17d5dc54b55ec49b8e109de7aa4825edaacc1b62ca172130e65f87b0f4aba1cc: Status 404 returned error can't find the container with id 17d5dc54b55ec49b8e109de7aa4825edaacc1b62ca172130e65f87b0f4aba1cc Oct 08 19:47:18 crc kubenswrapper[4988]: I1008 19:47:18.421916 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-a1b3-account-create-kzgsr" event={"ID":"869f964f-941e-4127-95cd-628c2e610300","Type":"ContainerStarted","Data":"17d5dc54b55ec49b8e109de7aa4825edaacc1b62ca172130e65f87b0f4aba1cc"} Oct 08 19:47:19 crc kubenswrapper[4988]: I1008 19:47:19.435511 4988 generic.go:334] "Generic (PLEG): container finished" podID="869f964f-941e-4127-95cd-628c2e610300" containerID="4fafc9509e524fbea5bed90e036340c69e0b654aeab3cdda3204b4864cd69ba7" exitCode=0 Oct 08 19:47:19 crc kubenswrapper[4988]: I1008 19:47:19.435597 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-a1b3-account-create-kzgsr" event={"ID":"869f964f-941e-4127-95cd-628c2e610300","Type":"ContainerDied","Data":"4fafc9509e524fbea5bed90e036340c69e0b654aeab3cdda3204b4864cd69ba7"} Oct 08 19:47:20 crc kubenswrapper[4988]: I1008 19:47:20.956053 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-a1b3-account-create-kzgsr" Oct 08 19:47:21 crc kubenswrapper[4988]: I1008 19:47:21.047698 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-stv5z\" (UniqueName: \"kubernetes.io/projected/869f964f-941e-4127-95cd-628c2e610300-kube-api-access-stv5z\") pod \"869f964f-941e-4127-95cd-628c2e610300\" (UID: \"869f964f-941e-4127-95cd-628c2e610300\") " Oct 08 19:47:21 crc kubenswrapper[4988]: I1008 19:47:21.053834 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/869f964f-941e-4127-95cd-628c2e610300-kube-api-access-stv5z" (OuterVolumeSpecName: "kube-api-access-stv5z") pod "869f964f-941e-4127-95cd-628c2e610300" (UID: "869f964f-941e-4127-95cd-628c2e610300"). InnerVolumeSpecName "kube-api-access-stv5z". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 19:47:21 crc kubenswrapper[4988]: I1008 19:47:21.149818 4988 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-stv5z\" (UniqueName: \"kubernetes.io/projected/869f964f-941e-4127-95cd-628c2e610300-kube-api-access-stv5z\") on node \"crc\" DevicePath \"\"" Oct 08 19:47:21 crc kubenswrapper[4988]: I1008 19:47:21.465328 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-a1b3-account-create-kzgsr" event={"ID":"869f964f-941e-4127-95cd-628c2e610300","Type":"ContainerDied","Data":"17d5dc54b55ec49b8e109de7aa4825edaacc1b62ca172130e65f87b0f4aba1cc"} Oct 08 19:47:21 crc kubenswrapper[4988]: I1008 19:47:21.465822 4988 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="17d5dc54b55ec49b8e109de7aa4825edaacc1b62ca172130e65f87b0f4aba1cc" Oct 08 19:47:21 crc kubenswrapper[4988]: I1008 19:47:21.465695 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-a1b3-account-create-kzgsr" Oct 08 19:47:22 crc kubenswrapper[4988]: I1008 19:47:22.510343 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-sync-pb78n"] Oct 08 19:47:22 crc kubenswrapper[4988]: E1008 19:47:22.512092 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="869f964f-941e-4127-95cd-628c2e610300" containerName="mariadb-account-create" Oct 08 19:47:22 crc kubenswrapper[4988]: I1008 19:47:22.512251 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="869f964f-941e-4127-95cd-628c2e610300" containerName="mariadb-account-create" Oct 08 19:47:22 crc kubenswrapper[4988]: I1008 19:47:22.512732 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="869f964f-941e-4127-95cd-628c2e610300" containerName="mariadb-account-create" Oct 08 19:47:22 crc kubenswrapper[4988]: I1008 19:47:22.513895 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-pb78n" Oct 08 19:47:22 crc kubenswrapper[4988]: I1008 19:47:22.516473 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-config-data" Oct 08 19:47:22 crc kubenswrapper[4988]: I1008 19:47:22.516974 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-b6m4j" Oct 08 19:47:22 crc kubenswrapper[4988]: I1008 19:47:22.521958 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-pb78n"] Oct 08 19:47:22 crc kubenswrapper[4988]: I1008 19:47:22.577956 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lgmcg\" (UniqueName: \"kubernetes.io/projected/3636ec06-cf9f-4d7f-8bbe-e67f97662074-kube-api-access-lgmcg\") pod \"glance-db-sync-pb78n\" (UID: \"3636ec06-cf9f-4d7f-8bbe-e67f97662074\") " pod="openstack/glance-db-sync-pb78n" Oct 08 19:47:22 crc kubenswrapper[4988]: I1008 19:47:22.578045 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/3636ec06-cf9f-4d7f-8bbe-e67f97662074-db-sync-config-data\") pod \"glance-db-sync-pb78n\" (UID: \"3636ec06-cf9f-4d7f-8bbe-e67f97662074\") " pod="openstack/glance-db-sync-pb78n" Oct 08 19:47:22 crc kubenswrapper[4988]: I1008 19:47:22.578077 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3636ec06-cf9f-4d7f-8bbe-e67f97662074-config-data\") pod \"glance-db-sync-pb78n\" (UID: \"3636ec06-cf9f-4d7f-8bbe-e67f97662074\") " pod="openstack/glance-db-sync-pb78n" Oct 08 19:47:22 crc kubenswrapper[4988]: I1008 19:47:22.578103 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3636ec06-cf9f-4d7f-8bbe-e67f97662074-combined-ca-bundle\") pod \"glance-db-sync-pb78n\" (UID: \"3636ec06-cf9f-4d7f-8bbe-e67f97662074\") " pod="openstack/glance-db-sync-pb78n" Oct 08 19:47:22 crc kubenswrapper[4988]: I1008 19:47:22.679682 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lgmcg\" (UniqueName: \"kubernetes.io/projected/3636ec06-cf9f-4d7f-8bbe-e67f97662074-kube-api-access-lgmcg\") pod \"glance-db-sync-pb78n\" (UID: \"3636ec06-cf9f-4d7f-8bbe-e67f97662074\") " pod="openstack/glance-db-sync-pb78n" Oct 08 19:47:22 crc kubenswrapper[4988]: I1008 19:47:22.679774 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/3636ec06-cf9f-4d7f-8bbe-e67f97662074-db-sync-config-data\") pod \"glance-db-sync-pb78n\" (UID: \"3636ec06-cf9f-4d7f-8bbe-e67f97662074\") " pod="openstack/glance-db-sync-pb78n" Oct 08 19:47:22 crc kubenswrapper[4988]: I1008 19:47:22.679817 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3636ec06-cf9f-4d7f-8bbe-e67f97662074-config-data\") pod \"glance-db-sync-pb78n\" (UID: \"3636ec06-cf9f-4d7f-8bbe-e67f97662074\") " pod="openstack/glance-db-sync-pb78n" Oct 08 19:47:22 crc kubenswrapper[4988]: I1008 19:47:22.679855 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3636ec06-cf9f-4d7f-8bbe-e67f97662074-combined-ca-bundle\") pod \"glance-db-sync-pb78n\" (UID: \"3636ec06-cf9f-4d7f-8bbe-e67f97662074\") " pod="openstack/glance-db-sync-pb78n" Oct 08 19:47:22 crc kubenswrapper[4988]: I1008 19:47:22.685866 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3636ec06-cf9f-4d7f-8bbe-e67f97662074-combined-ca-bundle\") pod \"glance-db-sync-pb78n\" (UID: \"3636ec06-cf9f-4d7f-8bbe-e67f97662074\") " pod="openstack/glance-db-sync-pb78n" Oct 08 19:47:22 crc kubenswrapper[4988]: I1008 19:47:22.686109 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/3636ec06-cf9f-4d7f-8bbe-e67f97662074-db-sync-config-data\") pod \"glance-db-sync-pb78n\" (UID: \"3636ec06-cf9f-4d7f-8bbe-e67f97662074\") " pod="openstack/glance-db-sync-pb78n" Oct 08 19:47:22 crc kubenswrapper[4988]: I1008 19:47:22.695636 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3636ec06-cf9f-4d7f-8bbe-e67f97662074-config-data\") pod \"glance-db-sync-pb78n\" (UID: \"3636ec06-cf9f-4d7f-8bbe-e67f97662074\") " pod="openstack/glance-db-sync-pb78n" Oct 08 19:47:22 crc kubenswrapper[4988]: I1008 19:47:22.705177 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lgmcg\" (UniqueName: \"kubernetes.io/projected/3636ec06-cf9f-4d7f-8bbe-e67f97662074-kube-api-access-lgmcg\") pod \"glance-db-sync-pb78n\" (UID: \"3636ec06-cf9f-4d7f-8bbe-e67f97662074\") " pod="openstack/glance-db-sync-pb78n" Oct 08 19:47:22 crc kubenswrapper[4988]: I1008 19:47:22.842735 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-pb78n" Oct 08 19:47:23 crc kubenswrapper[4988]: I1008 19:47:23.411130 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-pb78n"] Oct 08 19:47:23 crc kubenswrapper[4988]: I1008 19:47:23.485475 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-pb78n" event={"ID":"3636ec06-cf9f-4d7f-8bbe-e67f97662074","Type":"ContainerStarted","Data":"7115651c26a12e03c75ccf68c8a6971faf6e1c9729fcd4946112874e0e388408"} Oct 08 19:47:24 crc kubenswrapper[4988]: I1008 19:47:24.500066 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-pb78n" event={"ID":"3636ec06-cf9f-4d7f-8bbe-e67f97662074","Type":"ContainerStarted","Data":"f9fadee6bd841947a3cdc2d9f5693e3a99900cf16e24e4936647bdb1faa035ba"} Oct 08 19:47:24 crc kubenswrapper[4988]: I1008 19:47:24.536867 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-db-sync-pb78n" podStartSLOduration=2.536845254 podStartE2EDuration="2.536845254s" podCreationTimestamp="2025-10-08 19:47:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 19:47:24.534470799 +0000 UTC m=+5789.984313589" watchObservedRunningTime="2025-10-08 19:47:24.536845254 +0000 UTC m=+5789.986688034" Oct 08 19:47:27 crc kubenswrapper[4988]: I1008 19:47:27.538138 4988 generic.go:334] "Generic (PLEG): container finished" podID="3636ec06-cf9f-4d7f-8bbe-e67f97662074" containerID="f9fadee6bd841947a3cdc2d9f5693e3a99900cf16e24e4936647bdb1faa035ba" exitCode=0 Oct 08 19:47:27 crc kubenswrapper[4988]: I1008 19:47:27.538239 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-pb78n" event={"ID":"3636ec06-cf9f-4d7f-8bbe-e67f97662074","Type":"ContainerDied","Data":"f9fadee6bd841947a3cdc2d9f5693e3a99900cf16e24e4936647bdb1faa035ba"} Oct 08 19:47:29 crc kubenswrapper[4988]: I1008 19:47:29.012277 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-pb78n" Oct 08 19:47:29 crc kubenswrapper[4988]: I1008 19:47:29.207952 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3636ec06-cf9f-4d7f-8bbe-e67f97662074-config-data\") pod \"3636ec06-cf9f-4d7f-8bbe-e67f97662074\" (UID: \"3636ec06-cf9f-4d7f-8bbe-e67f97662074\") " Oct 08 19:47:29 crc kubenswrapper[4988]: I1008 19:47:29.208109 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3636ec06-cf9f-4d7f-8bbe-e67f97662074-combined-ca-bundle\") pod \"3636ec06-cf9f-4d7f-8bbe-e67f97662074\" (UID: \"3636ec06-cf9f-4d7f-8bbe-e67f97662074\") " Oct 08 19:47:29 crc kubenswrapper[4988]: I1008 19:47:29.208168 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lgmcg\" (UniqueName: \"kubernetes.io/projected/3636ec06-cf9f-4d7f-8bbe-e67f97662074-kube-api-access-lgmcg\") pod \"3636ec06-cf9f-4d7f-8bbe-e67f97662074\" (UID: \"3636ec06-cf9f-4d7f-8bbe-e67f97662074\") " Oct 08 19:47:29 crc kubenswrapper[4988]: I1008 19:47:29.208278 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/3636ec06-cf9f-4d7f-8bbe-e67f97662074-db-sync-config-data\") pod \"3636ec06-cf9f-4d7f-8bbe-e67f97662074\" (UID: \"3636ec06-cf9f-4d7f-8bbe-e67f97662074\") " Oct 08 19:47:29 crc kubenswrapper[4988]: I1008 19:47:29.216456 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3636ec06-cf9f-4d7f-8bbe-e67f97662074-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "3636ec06-cf9f-4d7f-8bbe-e67f97662074" (UID: "3636ec06-cf9f-4d7f-8bbe-e67f97662074"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 19:47:29 crc kubenswrapper[4988]: I1008 19:47:29.217168 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3636ec06-cf9f-4d7f-8bbe-e67f97662074-kube-api-access-lgmcg" (OuterVolumeSpecName: "kube-api-access-lgmcg") pod "3636ec06-cf9f-4d7f-8bbe-e67f97662074" (UID: "3636ec06-cf9f-4d7f-8bbe-e67f97662074"). InnerVolumeSpecName "kube-api-access-lgmcg". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 19:47:29 crc kubenswrapper[4988]: I1008 19:47:29.234553 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3636ec06-cf9f-4d7f-8bbe-e67f97662074-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "3636ec06-cf9f-4d7f-8bbe-e67f97662074" (UID: "3636ec06-cf9f-4d7f-8bbe-e67f97662074"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 19:47:29 crc kubenswrapper[4988]: I1008 19:47:29.285485 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3636ec06-cf9f-4d7f-8bbe-e67f97662074-config-data" (OuterVolumeSpecName: "config-data") pod "3636ec06-cf9f-4d7f-8bbe-e67f97662074" (UID: "3636ec06-cf9f-4d7f-8bbe-e67f97662074"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 19:47:29 crc kubenswrapper[4988]: I1008 19:47:29.311875 4988 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/3636ec06-cf9f-4d7f-8bbe-e67f97662074-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Oct 08 19:47:29 crc kubenswrapper[4988]: I1008 19:47:29.311926 4988 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3636ec06-cf9f-4d7f-8bbe-e67f97662074-config-data\") on node \"crc\" DevicePath \"\"" Oct 08 19:47:29 crc kubenswrapper[4988]: I1008 19:47:29.311946 4988 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3636ec06-cf9f-4d7f-8bbe-e67f97662074-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 19:47:29 crc kubenswrapper[4988]: I1008 19:47:29.311965 4988 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lgmcg\" (UniqueName: \"kubernetes.io/projected/3636ec06-cf9f-4d7f-8bbe-e67f97662074-kube-api-access-lgmcg\") on node \"crc\" DevicePath \"\"" Oct 08 19:47:29 crc kubenswrapper[4988]: I1008 19:47:29.569210 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-pb78n" event={"ID":"3636ec06-cf9f-4d7f-8bbe-e67f97662074","Type":"ContainerDied","Data":"7115651c26a12e03c75ccf68c8a6971faf6e1c9729fcd4946112874e0e388408"} Oct 08 19:47:29 crc kubenswrapper[4988]: I1008 19:47:29.569273 4988 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7115651c26a12e03c75ccf68c8a6971faf6e1c9729fcd4946112874e0e388408" Oct 08 19:47:29 crc kubenswrapper[4988]: I1008 19:47:29.569314 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-pb78n" Oct 08 19:47:29 crc kubenswrapper[4988]: I1008 19:47:29.874745 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Oct 08 19:47:29 crc kubenswrapper[4988]: E1008 19:47:29.875106 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3636ec06-cf9f-4d7f-8bbe-e67f97662074" containerName="glance-db-sync" Oct 08 19:47:29 crc kubenswrapper[4988]: I1008 19:47:29.875118 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="3636ec06-cf9f-4d7f-8bbe-e67f97662074" containerName="glance-db-sync" Oct 08 19:47:29 crc kubenswrapper[4988]: I1008 19:47:29.875266 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="3636ec06-cf9f-4d7f-8bbe-e67f97662074" containerName="glance-db-sync" Oct 08 19:47:29 crc kubenswrapper[4988]: I1008 19:47:29.876259 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 08 19:47:29 crc kubenswrapper[4988]: I1008 19:47:29.879116 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-scripts" Oct 08 19:47:29 crc kubenswrapper[4988]: I1008 19:47:29.879256 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-b6m4j" Oct 08 19:47:29 crc kubenswrapper[4988]: I1008 19:47:29.879708 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Oct 08 19:47:29 crc kubenswrapper[4988]: I1008 19:47:29.896155 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 08 19:47:30 crc kubenswrapper[4988]: I1008 19:47:30.020860 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5589f75ff5-qj6cv"] Oct 08 19:47:30 crc kubenswrapper[4988]: I1008 19:47:30.022599 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5589f75ff5-qj6cv" Oct 08 19:47:30 crc kubenswrapper[4988]: I1008 19:47:30.027110 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/089988dc-da74-4d2b-a9d3-c3ad66a17dac-logs\") pod \"glance-default-external-api-0\" (UID: \"089988dc-da74-4d2b-a9d3-c3ad66a17dac\") " pod="openstack/glance-default-external-api-0" Oct 08 19:47:30 crc kubenswrapper[4988]: I1008 19:47:30.027172 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xrrrt\" (UniqueName: \"kubernetes.io/projected/5943bf5d-dc4e-4699-af97-bbb845b9db43-kube-api-access-xrrrt\") pod \"dnsmasq-dns-5589f75ff5-qj6cv\" (UID: \"5943bf5d-dc4e-4699-af97-bbb845b9db43\") " pod="openstack/dnsmasq-dns-5589f75ff5-qj6cv" Oct 08 19:47:30 crc kubenswrapper[4988]: I1008 19:47:30.027218 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5943bf5d-dc4e-4699-af97-bbb845b9db43-config\") pod \"dnsmasq-dns-5589f75ff5-qj6cv\" (UID: \"5943bf5d-dc4e-4699-af97-bbb845b9db43\") " pod="openstack/dnsmasq-dns-5589f75ff5-qj6cv" Oct 08 19:47:30 crc kubenswrapper[4988]: I1008 19:47:30.027240 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5943bf5d-dc4e-4699-af97-bbb845b9db43-ovsdbserver-nb\") pod \"dnsmasq-dns-5589f75ff5-qj6cv\" (UID: \"5943bf5d-dc4e-4699-af97-bbb845b9db43\") " pod="openstack/dnsmasq-dns-5589f75ff5-qj6cv" Oct 08 19:47:30 crc kubenswrapper[4988]: I1008 19:47:30.027261 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/089988dc-da74-4d2b-a9d3-c3ad66a17dac-scripts\") pod \"glance-default-external-api-0\" (UID: \"089988dc-da74-4d2b-a9d3-c3ad66a17dac\") " pod="openstack/glance-default-external-api-0" Oct 08 19:47:30 crc kubenswrapper[4988]: I1008 19:47:30.027350 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5943bf5d-dc4e-4699-af97-bbb845b9db43-dns-svc\") pod \"dnsmasq-dns-5589f75ff5-qj6cv\" (UID: \"5943bf5d-dc4e-4699-af97-bbb845b9db43\") " pod="openstack/dnsmasq-dns-5589f75ff5-qj6cv" Oct 08 19:47:30 crc kubenswrapper[4988]: I1008 19:47:30.027408 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/089988dc-da74-4d2b-a9d3-c3ad66a17dac-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"089988dc-da74-4d2b-a9d3-c3ad66a17dac\") " pod="openstack/glance-default-external-api-0" Oct 08 19:47:30 crc kubenswrapper[4988]: I1008 19:47:30.027436 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5943bf5d-dc4e-4699-af97-bbb845b9db43-ovsdbserver-sb\") pod \"dnsmasq-dns-5589f75ff5-qj6cv\" (UID: \"5943bf5d-dc4e-4699-af97-bbb845b9db43\") " pod="openstack/dnsmasq-dns-5589f75ff5-qj6cv" Oct 08 19:47:30 crc kubenswrapper[4988]: I1008 19:47:30.027460 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/089988dc-da74-4d2b-a9d3-c3ad66a17dac-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"089988dc-da74-4d2b-a9d3-c3ad66a17dac\") " pod="openstack/glance-default-external-api-0" Oct 08 19:47:30 crc kubenswrapper[4988]: I1008 19:47:30.027499 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/089988dc-da74-4d2b-a9d3-c3ad66a17dac-config-data\") pod \"glance-default-external-api-0\" (UID: \"089988dc-da74-4d2b-a9d3-c3ad66a17dac\") " pod="openstack/glance-default-external-api-0" Oct 08 19:47:30 crc kubenswrapper[4988]: I1008 19:47:30.027522 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q2866\" (UniqueName: \"kubernetes.io/projected/089988dc-da74-4d2b-a9d3-c3ad66a17dac-kube-api-access-q2866\") pod \"glance-default-external-api-0\" (UID: \"089988dc-da74-4d2b-a9d3-c3ad66a17dac\") " pod="openstack/glance-default-external-api-0" Oct 08 19:47:30 crc kubenswrapper[4988]: I1008 19:47:30.039913 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5589f75ff5-qj6cv"] Oct 08 19:47:30 crc kubenswrapper[4988]: I1008 19:47:30.130584 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5943bf5d-dc4e-4699-af97-bbb845b9db43-dns-svc\") pod \"dnsmasq-dns-5589f75ff5-qj6cv\" (UID: \"5943bf5d-dc4e-4699-af97-bbb845b9db43\") " pod="openstack/dnsmasq-dns-5589f75ff5-qj6cv" Oct 08 19:47:30 crc kubenswrapper[4988]: I1008 19:47:30.130647 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/089988dc-da74-4d2b-a9d3-c3ad66a17dac-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"089988dc-da74-4d2b-a9d3-c3ad66a17dac\") " pod="openstack/glance-default-external-api-0" Oct 08 19:47:30 crc kubenswrapper[4988]: I1008 19:47:30.130671 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5943bf5d-dc4e-4699-af97-bbb845b9db43-ovsdbserver-sb\") pod \"dnsmasq-dns-5589f75ff5-qj6cv\" (UID: \"5943bf5d-dc4e-4699-af97-bbb845b9db43\") " pod="openstack/dnsmasq-dns-5589f75ff5-qj6cv" Oct 08 19:47:30 crc kubenswrapper[4988]: I1008 19:47:30.130690 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/089988dc-da74-4d2b-a9d3-c3ad66a17dac-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"089988dc-da74-4d2b-a9d3-c3ad66a17dac\") " pod="openstack/glance-default-external-api-0" Oct 08 19:47:30 crc kubenswrapper[4988]: I1008 19:47:30.130722 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/089988dc-da74-4d2b-a9d3-c3ad66a17dac-config-data\") pod \"glance-default-external-api-0\" (UID: \"089988dc-da74-4d2b-a9d3-c3ad66a17dac\") " pod="openstack/glance-default-external-api-0" Oct 08 19:47:30 crc kubenswrapper[4988]: I1008 19:47:30.130738 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q2866\" (UniqueName: \"kubernetes.io/projected/089988dc-da74-4d2b-a9d3-c3ad66a17dac-kube-api-access-q2866\") pod \"glance-default-external-api-0\" (UID: \"089988dc-da74-4d2b-a9d3-c3ad66a17dac\") " pod="openstack/glance-default-external-api-0" Oct 08 19:47:30 crc kubenswrapper[4988]: I1008 19:47:30.130760 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/089988dc-da74-4d2b-a9d3-c3ad66a17dac-logs\") pod \"glance-default-external-api-0\" (UID: \"089988dc-da74-4d2b-a9d3-c3ad66a17dac\") " pod="openstack/glance-default-external-api-0" Oct 08 19:47:30 crc kubenswrapper[4988]: I1008 19:47:30.130781 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xrrrt\" (UniqueName: \"kubernetes.io/projected/5943bf5d-dc4e-4699-af97-bbb845b9db43-kube-api-access-xrrrt\") pod \"dnsmasq-dns-5589f75ff5-qj6cv\" (UID: \"5943bf5d-dc4e-4699-af97-bbb845b9db43\") " pod="openstack/dnsmasq-dns-5589f75ff5-qj6cv" Oct 08 19:47:30 crc kubenswrapper[4988]: I1008 19:47:30.130813 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5943bf5d-dc4e-4699-af97-bbb845b9db43-config\") pod \"dnsmasq-dns-5589f75ff5-qj6cv\" (UID: \"5943bf5d-dc4e-4699-af97-bbb845b9db43\") " pod="openstack/dnsmasq-dns-5589f75ff5-qj6cv" Oct 08 19:47:30 crc kubenswrapper[4988]: I1008 19:47:30.130828 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5943bf5d-dc4e-4699-af97-bbb845b9db43-ovsdbserver-nb\") pod \"dnsmasq-dns-5589f75ff5-qj6cv\" (UID: \"5943bf5d-dc4e-4699-af97-bbb845b9db43\") " pod="openstack/dnsmasq-dns-5589f75ff5-qj6cv" Oct 08 19:47:30 crc kubenswrapper[4988]: I1008 19:47:30.130846 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/089988dc-da74-4d2b-a9d3-c3ad66a17dac-scripts\") pod \"glance-default-external-api-0\" (UID: \"089988dc-da74-4d2b-a9d3-c3ad66a17dac\") " pod="openstack/glance-default-external-api-0" Oct 08 19:47:30 crc kubenswrapper[4988]: I1008 19:47:30.139344 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/089988dc-da74-4d2b-a9d3-c3ad66a17dac-scripts\") pod \"glance-default-external-api-0\" (UID: \"089988dc-da74-4d2b-a9d3-c3ad66a17dac\") " pod="openstack/glance-default-external-api-0" Oct 08 19:47:30 crc kubenswrapper[4988]: I1008 19:47:30.139416 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/089988dc-da74-4d2b-a9d3-c3ad66a17dac-logs\") pod \"glance-default-external-api-0\" (UID: \"089988dc-da74-4d2b-a9d3-c3ad66a17dac\") " pod="openstack/glance-default-external-api-0" Oct 08 19:47:30 crc kubenswrapper[4988]: I1008 19:47:30.140143 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5943bf5d-dc4e-4699-af97-bbb845b9db43-config\") pod \"dnsmasq-dns-5589f75ff5-qj6cv\" (UID: \"5943bf5d-dc4e-4699-af97-bbb845b9db43\") " pod="openstack/dnsmasq-dns-5589f75ff5-qj6cv" Oct 08 19:47:30 crc kubenswrapper[4988]: I1008 19:47:30.140144 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5943bf5d-dc4e-4699-af97-bbb845b9db43-ovsdbserver-nb\") pod \"dnsmasq-dns-5589f75ff5-qj6cv\" (UID: \"5943bf5d-dc4e-4699-af97-bbb845b9db43\") " pod="openstack/dnsmasq-dns-5589f75ff5-qj6cv" Oct 08 19:47:30 crc kubenswrapper[4988]: I1008 19:47:30.141323 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5943bf5d-dc4e-4699-af97-bbb845b9db43-ovsdbserver-sb\") pod \"dnsmasq-dns-5589f75ff5-qj6cv\" (UID: \"5943bf5d-dc4e-4699-af97-bbb845b9db43\") " pod="openstack/dnsmasq-dns-5589f75ff5-qj6cv" Oct 08 19:47:30 crc kubenswrapper[4988]: I1008 19:47:30.141437 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/089988dc-da74-4d2b-a9d3-c3ad66a17dac-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"089988dc-da74-4d2b-a9d3-c3ad66a17dac\") " pod="openstack/glance-default-external-api-0" Oct 08 19:47:30 crc kubenswrapper[4988]: I1008 19:47:30.141999 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5943bf5d-dc4e-4699-af97-bbb845b9db43-dns-svc\") pod \"dnsmasq-dns-5589f75ff5-qj6cv\" (UID: \"5943bf5d-dc4e-4699-af97-bbb845b9db43\") " pod="openstack/dnsmasq-dns-5589f75ff5-qj6cv" Oct 08 19:47:30 crc kubenswrapper[4988]: I1008 19:47:30.146678 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/089988dc-da74-4d2b-a9d3-c3ad66a17dac-config-data\") pod \"glance-default-external-api-0\" (UID: \"089988dc-da74-4d2b-a9d3-c3ad66a17dac\") " pod="openstack/glance-default-external-api-0" Oct 08 19:47:30 crc kubenswrapper[4988]: I1008 19:47:30.152205 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/089988dc-da74-4d2b-a9d3-c3ad66a17dac-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"089988dc-da74-4d2b-a9d3-c3ad66a17dac\") " pod="openstack/glance-default-external-api-0" Oct 08 19:47:30 crc kubenswrapper[4988]: I1008 19:47:30.185054 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q2866\" (UniqueName: \"kubernetes.io/projected/089988dc-da74-4d2b-a9d3-c3ad66a17dac-kube-api-access-q2866\") pod \"glance-default-external-api-0\" (UID: \"089988dc-da74-4d2b-a9d3-c3ad66a17dac\") " pod="openstack/glance-default-external-api-0" Oct 08 19:47:30 crc kubenswrapper[4988]: I1008 19:47:30.201819 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 08 19:47:30 crc kubenswrapper[4988]: I1008 19:47:30.202199 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xrrrt\" (UniqueName: \"kubernetes.io/projected/5943bf5d-dc4e-4699-af97-bbb845b9db43-kube-api-access-xrrrt\") pod \"dnsmasq-dns-5589f75ff5-qj6cv\" (UID: \"5943bf5d-dc4e-4699-af97-bbb845b9db43\") " pod="openstack/dnsmasq-dns-5589f75ff5-qj6cv" Oct 08 19:47:30 crc kubenswrapper[4988]: I1008 19:47:30.238732 4988 scope.go:117] "RemoveContainer" containerID="f2a29ed22dd7a48bbc929e65fbeceed969a669e78f65fbfb724b03dc0db6e947" Oct 08 19:47:30 crc kubenswrapper[4988]: E1008 19:47:30.239173 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nm54f_openshift-machine-config-operator(f328a20b-5283-407e-8a93-2ab2da599f18)\"" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" Oct 08 19:47:30 crc kubenswrapper[4988]: I1008 19:47:30.269597 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 08 19:47:30 crc kubenswrapper[4988]: I1008 19:47:30.270995 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 08 19:47:30 crc kubenswrapper[4988]: I1008 19:47:30.275766 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Oct 08 19:47:30 crc kubenswrapper[4988]: I1008 19:47:30.308502 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 08 19:47:30 crc kubenswrapper[4988]: I1008 19:47:30.348743 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5589f75ff5-qj6cv" Oct 08 19:47:30 crc kubenswrapper[4988]: I1008 19:47:30.439217 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7198c3ac-8210-43f0-bd60-e18bb3da7de3-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"7198c3ac-8210-43f0-bd60-e18bb3da7de3\") " pod="openstack/glance-default-internal-api-0" Oct 08 19:47:30 crc kubenswrapper[4988]: I1008 19:47:30.439319 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-87v69\" (UniqueName: \"kubernetes.io/projected/7198c3ac-8210-43f0-bd60-e18bb3da7de3-kube-api-access-87v69\") pod \"glance-default-internal-api-0\" (UID: \"7198c3ac-8210-43f0-bd60-e18bb3da7de3\") " pod="openstack/glance-default-internal-api-0" Oct 08 19:47:30 crc kubenswrapper[4988]: I1008 19:47:30.439356 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/7198c3ac-8210-43f0-bd60-e18bb3da7de3-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"7198c3ac-8210-43f0-bd60-e18bb3da7de3\") " pod="openstack/glance-default-internal-api-0" Oct 08 19:47:30 crc kubenswrapper[4988]: I1008 19:47:30.439444 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7198c3ac-8210-43f0-bd60-e18bb3da7de3-logs\") pod \"glance-default-internal-api-0\" (UID: \"7198c3ac-8210-43f0-bd60-e18bb3da7de3\") " pod="openstack/glance-default-internal-api-0" Oct 08 19:47:30 crc kubenswrapper[4988]: I1008 19:47:30.439472 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7198c3ac-8210-43f0-bd60-e18bb3da7de3-scripts\") pod \"glance-default-internal-api-0\" (UID: \"7198c3ac-8210-43f0-bd60-e18bb3da7de3\") " pod="openstack/glance-default-internal-api-0" Oct 08 19:47:30 crc kubenswrapper[4988]: I1008 19:47:30.439568 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7198c3ac-8210-43f0-bd60-e18bb3da7de3-config-data\") pod \"glance-default-internal-api-0\" (UID: \"7198c3ac-8210-43f0-bd60-e18bb3da7de3\") " pod="openstack/glance-default-internal-api-0" Oct 08 19:47:30 crc kubenswrapper[4988]: I1008 19:47:30.541376 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7198c3ac-8210-43f0-bd60-e18bb3da7de3-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"7198c3ac-8210-43f0-bd60-e18bb3da7de3\") " pod="openstack/glance-default-internal-api-0" Oct 08 19:47:30 crc kubenswrapper[4988]: I1008 19:47:30.541497 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-87v69\" (UniqueName: \"kubernetes.io/projected/7198c3ac-8210-43f0-bd60-e18bb3da7de3-kube-api-access-87v69\") pod \"glance-default-internal-api-0\" (UID: \"7198c3ac-8210-43f0-bd60-e18bb3da7de3\") " pod="openstack/glance-default-internal-api-0" Oct 08 19:47:30 crc kubenswrapper[4988]: I1008 19:47:30.541544 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/7198c3ac-8210-43f0-bd60-e18bb3da7de3-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"7198c3ac-8210-43f0-bd60-e18bb3da7de3\") " pod="openstack/glance-default-internal-api-0" Oct 08 19:47:30 crc kubenswrapper[4988]: I1008 19:47:30.541583 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7198c3ac-8210-43f0-bd60-e18bb3da7de3-logs\") pod \"glance-default-internal-api-0\" (UID: \"7198c3ac-8210-43f0-bd60-e18bb3da7de3\") " pod="openstack/glance-default-internal-api-0" Oct 08 19:47:30 crc kubenswrapper[4988]: I1008 19:47:30.541610 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7198c3ac-8210-43f0-bd60-e18bb3da7de3-scripts\") pod \"glance-default-internal-api-0\" (UID: \"7198c3ac-8210-43f0-bd60-e18bb3da7de3\") " pod="openstack/glance-default-internal-api-0" Oct 08 19:47:30 crc kubenswrapper[4988]: I1008 19:47:30.541633 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7198c3ac-8210-43f0-bd60-e18bb3da7de3-config-data\") pod \"glance-default-internal-api-0\" (UID: \"7198c3ac-8210-43f0-bd60-e18bb3da7de3\") " pod="openstack/glance-default-internal-api-0" Oct 08 19:47:30 crc kubenswrapper[4988]: I1008 19:47:30.542512 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/7198c3ac-8210-43f0-bd60-e18bb3da7de3-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"7198c3ac-8210-43f0-bd60-e18bb3da7de3\") " pod="openstack/glance-default-internal-api-0" Oct 08 19:47:30 crc kubenswrapper[4988]: I1008 19:47:30.542684 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7198c3ac-8210-43f0-bd60-e18bb3da7de3-logs\") pod \"glance-default-internal-api-0\" (UID: \"7198c3ac-8210-43f0-bd60-e18bb3da7de3\") " pod="openstack/glance-default-internal-api-0" Oct 08 19:47:30 crc kubenswrapper[4988]: I1008 19:47:30.546491 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7198c3ac-8210-43f0-bd60-e18bb3da7de3-scripts\") pod \"glance-default-internal-api-0\" (UID: \"7198c3ac-8210-43f0-bd60-e18bb3da7de3\") " pod="openstack/glance-default-internal-api-0" Oct 08 19:47:30 crc kubenswrapper[4988]: I1008 19:47:30.550035 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7198c3ac-8210-43f0-bd60-e18bb3da7de3-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"7198c3ac-8210-43f0-bd60-e18bb3da7de3\") " pod="openstack/glance-default-internal-api-0" Oct 08 19:47:30 crc kubenswrapper[4988]: I1008 19:47:30.550508 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7198c3ac-8210-43f0-bd60-e18bb3da7de3-config-data\") pod \"glance-default-internal-api-0\" (UID: \"7198c3ac-8210-43f0-bd60-e18bb3da7de3\") " pod="openstack/glance-default-internal-api-0" Oct 08 19:47:30 crc kubenswrapper[4988]: I1008 19:47:30.558922 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-87v69\" (UniqueName: \"kubernetes.io/projected/7198c3ac-8210-43f0-bd60-e18bb3da7de3-kube-api-access-87v69\") pod \"glance-default-internal-api-0\" (UID: \"7198c3ac-8210-43f0-bd60-e18bb3da7de3\") " pod="openstack/glance-default-internal-api-0" Oct 08 19:47:30 crc kubenswrapper[4988]: I1008 19:47:30.612778 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 08 19:47:30 crc kubenswrapper[4988]: I1008 19:47:30.680183 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5589f75ff5-qj6cv"] Oct 08 19:47:30 crc kubenswrapper[4988]: I1008 19:47:30.876872 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 08 19:47:30 crc kubenswrapper[4988]: W1008 19:47:30.887432 4988 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod089988dc_da74_4d2b_a9d3_c3ad66a17dac.slice/crio-d4f4c0dff727040260bbd428edd53882fe244baa5f60b8d2a6c99e0f94902c72 WatchSource:0}: Error finding container d4f4c0dff727040260bbd428edd53882fe244baa5f60b8d2a6c99e0f94902c72: Status 404 returned error can't find the container with id d4f4c0dff727040260bbd428edd53882fe244baa5f60b8d2a6c99e0f94902c72 Oct 08 19:47:31 crc kubenswrapper[4988]: I1008 19:47:31.161368 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 08 19:47:31 crc kubenswrapper[4988]: I1008 19:47:31.266474 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 08 19:47:31 crc kubenswrapper[4988]: W1008 19:47:31.280645 4988 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7198c3ac_8210_43f0_bd60_e18bb3da7de3.slice/crio-905c84e528dbfdd9d49eaf89926f2273413eb50e311aa40f2860be6bc5c59f44 WatchSource:0}: Error finding container 905c84e528dbfdd9d49eaf89926f2273413eb50e311aa40f2860be6bc5c59f44: Status 404 returned error can't find the container with id 905c84e528dbfdd9d49eaf89926f2273413eb50e311aa40f2860be6bc5c59f44 Oct 08 19:47:31 crc kubenswrapper[4988]: I1008 19:47:31.598484 4988 generic.go:334] "Generic (PLEG): container finished" podID="5943bf5d-dc4e-4699-af97-bbb845b9db43" containerID="bfb03ed93c36eefdac388693d766b3ad6aa39d28b42fd1cc6881caf8e730d6a2" exitCode=0 Oct 08 19:47:31 crc kubenswrapper[4988]: I1008 19:47:31.598550 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5589f75ff5-qj6cv" event={"ID":"5943bf5d-dc4e-4699-af97-bbb845b9db43","Type":"ContainerDied","Data":"bfb03ed93c36eefdac388693d766b3ad6aa39d28b42fd1cc6881caf8e730d6a2"} Oct 08 19:47:31 crc kubenswrapper[4988]: I1008 19:47:31.598578 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5589f75ff5-qj6cv" event={"ID":"5943bf5d-dc4e-4699-af97-bbb845b9db43","Type":"ContainerStarted","Data":"0531fc115a562394dd0e6a1270979d2c8438dae8813dc865710678635fc7be41"} Oct 08 19:47:31 crc kubenswrapper[4988]: I1008 19:47:31.610230 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"089988dc-da74-4d2b-a9d3-c3ad66a17dac","Type":"ContainerStarted","Data":"eae9a4f1f2c8cb0bd59489b363c518b91d7ba118e959aaa791e2b4295597f7fc"} Oct 08 19:47:31 crc kubenswrapper[4988]: I1008 19:47:31.610282 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"089988dc-da74-4d2b-a9d3-c3ad66a17dac","Type":"ContainerStarted","Data":"d4f4c0dff727040260bbd428edd53882fe244baa5f60b8d2a6c99e0f94902c72"} Oct 08 19:47:31 crc kubenswrapper[4988]: I1008 19:47:31.612348 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"7198c3ac-8210-43f0-bd60-e18bb3da7de3","Type":"ContainerStarted","Data":"905c84e528dbfdd9d49eaf89926f2273413eb50e311aa40f2860be6bc5c59f44"} Oct 08 19:47:32 crc kubenswrapper[4988]: I1008 19:47:32.370076 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 08 19:47:32 crc kubenswrapper[4988]: I1008 19:47:32.623121 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5589f75ff5-qj6cv" event={"ID":"5943bf5d-dc4e-4699-af97-bbb845b9db43","Type":"ContainerStarted","Data":"38fe2bfa99b61db13654ec48a0d8f5ede91e22ae272e97c10552c183357640d0"} Oct 08 19:47:32 crc kubenswrapper[4988]: I1008 19:47:32.623262 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-5589f75ff5-qj6cv" Oct 08 19:47:32 crc kubenswrapper[4988]: I1008 19:47:32.624984 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"089988dc-da74-4d2b-a9d3-c3ad66a17dac","Type":"ContainerStarted","Data":"df1652a004c93e3aeab64779b2128aa987ae2c43bc9da3e84a0bdf38d1393a10"} Oct 08 19:47:32 crc kubenswrapper[4988]: I1008 19:47:32.625036 4988 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="089988dc-da74-4d2b-a9d3-c3ad66a17dac" containerName="glance-log" containerID="cri-o://eae9a4f1f2c8cb0bd59489b363c518b91d7ba118e959aaa791e2b4295597f7fc" gracePeriod=30 Oct 08 19:47:32 crc kubenswrapper[4988]: I1008 19:47:32.625056 4988 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="089988dc-da74-4d2b-a9d3-c3ad66a17dac" containerName="glance-httpd" containerID="cri-o://df1652a004c93e3aeab64779b2128aa987ae2c43bc9da3e84a0bdf38d1393a10" gracePeriod=30 Oct 08 19:47:32 crc kubenswrapper[4988]: I1008 19:47:32.626667 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"7198c3ac-8210-43f0-bd60-e18bb3da7de3","Type":"ContainerStarted","Data":"774051dbff0e9e9a4ecb1dd15c593ea0c50811d8ef6d13ab94101e4189441776"} Oct 08 19:47:32 crc kubenswrapper[4988]: I1008 19:47:32.626703 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"7198c3ac-8210-43f0-bd60-e18bb3da7de3","Type":"ContainerStarted","Data":"946855f6f4945fba05575de7262453fb75bd7867dea5aaea069f8748e2d681c9"} Oct 08 19:47:32 crc kubenswrapper[4988]: I1008 19:47:32.626763 4988 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="7198c3ac-8210-43f0-bd60-e18bb3da7de3" containerName="glance-log" containerID="cri-o://946855f6f4945fba05575de7262453fb75bd7867dea5aaea069f8748e2d681c9" gracePeriod=30 Oct 08 19:47:32 crc kubenswrapper[4988]: I1008 19:47:32.626798 4988 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="7198c3ac-8210-43f0-bd60-e18bb3da7de3" containerName="glance-httpd" containerID="cri-o://774051dbff0e9e9a4ecb1dd15c593ea0c50811d8ef6d13ab94101e4189441776" gracePeriod=30 Oct 08 19:47:32 crc kubenswrapper[4988]: I1008 19:47:32.647984 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-5589f75ff5-qj6cv" podStartSLOduration=3.64796787 podStartE2EDuration="3.64796787s" podCreationTimestamp="2025-10-08 19:47:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 19:47:32.641103311 +0000 UTC m=+5798.090946081" watchObservedRunningTime="2025-10-08 19:47:32.64796787 +0000 UTC m=+5798.097810640" Oct 08 19:47:32 crc kubenswrapper[4988]: I1008 19:47:32.671594 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=3.67157153 podStartE2EDuration="3.67157153s" podCreationTimestamp="2025-10-08 19:47:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 19:47:32.664989641 +0000 UTC m=+5798.114832411" watchObservedRunningTime="2025-10-08 19:47:32.67157153 +0000 UTC m=+5798.121414300" Oct 08 19:47:32 crc kubenswrapper[4988]: I1008 19:47:32.688195 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=2.688175858 podStartE2EDuration="2.688175858s" podCreationTimestamp="2025-10-08 19:47:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 19:47:32.682852629 +0000 UTC m=+5798.132695399" watchObservedRunningTime="2025-10-08 19:47:32.688175858 +0000 UTC m=+5798.138018628" Oct 08 19:47:33 crc kubenswrapper[4988]: I1008 19:47:33.380457 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 08 19:47:33 crc kubenswrapper[4988]: I1008 19:47:33.386789 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 08 19:47:33 crc kubenswrapper[4988]: I1008 19:47:33.490733 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7198c3ac-8210-43f0-bd60-e18bb3da7de3-combined-ca-bundle\") pod \"7198c3ac-8210-43f0-bd60-e18bb3da7de3\" (UID: \"7198c3ac-8210-43f0-bd60-e18bb3da7de3\") " Oct 08 19:47:33 crc kubenswrapper[4988]: I1008 19:47:33.490793 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/089988dc-da74-4d2b-a9d3-c3ad66a17dac-logs\") pod \"089988dc-da74-4d2b-a9d3-c3ad66a17dac\" (UID: \"089988dc-da74-4d2b-a9d3-c3ad66a17dac\") " Oct 08 19:47:33 crc kubenswrapper[4988]: I1008 19:47:33.490812 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/7198c3ac-8210-43f0-bd60-e18bb3da7de3-httpd-run\") pod \"7198c3ac-8210-43f0-bd60-e18bb3da7de3\" (UID: \"7198c3ac-8210-43f0-bd60-e18bb3da7de3\") " Oct 08 19:47:33 crc kubenswrapper[4988]: I1008 19:47:33.490832 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/089988dc-da74-4d2b-a9d3-c3ad66a17dac-scripts\") pod \"089988dc-da74-4d2b-a9d3-c3ad66a17dac\" (UID: \"089988dc-da74-4d2b-a9d3-c3ad66a17dac\") " Oct 08 19:47:33 crc kubenswrapper[4988]: I1008 19:47:33.490847 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7198c3ac-8210-43f0-bd60-e18bb3da7de3-scripts\") pod \"7198c3ac-8210-43f0-bd60-e18bb3da7de3\" (UID: \"7198c3ac-8210-43f0-bd60-e18bb3da7de3\") " Oct 08 19:47:33 crc kubenswrapper[4988]: I1008 19:47:33.490874 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q2866\" (UniqueName: \"kubernetes.io/projected/089988dc-da74-4d2b-a9d3-c3ad66a17dac-kube-api-access-q2866\") pod \"089988dc-da74-4d2b-a9d3-c3ad66a17dac\" (UID: \"089988dc-da74-4d2b-a9d3-c3ad66a17dac\") " Oct 08 19:47:33 crc kubenswrapper[4988]: I1008 19:47:33.490942 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/089988dc-da74-4d2b-a9d3-c3ad66a17dac-httpd-run\") pod \"089988dc-da74-4d2b-a9d3-c3ad66a17dac\" (UID: \"089988dc-da74-4d2b-a9d3-c3ad66a17dac\") " Oct 08 19:47:33 crc kubenswrapper[4988]: I1008 19:47:33.490958 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7198c3ac-8210-43f0-bd60-e18bb3da7de3-config-data\") pod \"7198c3ac-8210-43f0-bd60-e18bb3da7de3\" (UID: \"7198c3ac-8210-43f0-bd60-e18bb3da7de3\") " Oct 08 19:47:33 crc kubenswrapper[4988]: I1008 19:47:33.490977 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/089988dc-da74-4d2b-a9d3-c3ad66a17dac-combined-ca-bundle\") pod \"089988dc-da74-4d2b-a9d3-c3ad66a17dac\" (UID: \"089988dc-da74-4d2b-a9d3-c3ad66a17dac\") " Oct 08 19:47:33 crc kubenswrapper[4988]: I1008 19:47:33.490994 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7198c3ac-8210-43f0-bd60-e18bb3da7de3-logs\") pod \"7198c3ac-8210-43f0-bd60-e18bb3da7de3\" (UID: \"7198c3ac-8210-43f0-bd60-e18bb3da7de3\") " Oct 08 19:47:33 crc kubenswrapper[4988]: I1008 19:47:33.491058 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-87v69\" (UniqueName: \"kubernetes.io/projected/7198c3ac-8210-43f0-bd60-e18bb3da7de3-kube-api-access-87v69\") pod \"7198c3ac-8210-43f0-bd60-e18bb3da7de3\" (UID: \"7198c3ac-8210-43f0-bd60-e18bb3da7de3\") " Oct 08 19:47:33 crc kubenswrapper[4988]: I1008 19:47:33.491092 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/089988dc-da74-4d2b-a9d3-c3ad66a17dac-config-data\") pod \"089988dc-da74-4d2b-a9d3-c3ad66a17dac\" (UID: \"089988dc-da74-4d2b-a9d3-c3ad66a17dac\") " Oct 08 19:47:33 crc kubenswrapper[4988]: I1008 19:47:33.491244 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/089988dc-da74-4d2b-a9d3-c3ad66a17dac-logs" (OuterVolumeSpecName: "logs") pod "089988dc-da74-4d2b-a9d3-c3ad66a17dac" (UID: "089988dc-da74-4d2b-a9d3-c3ad66a17dac"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 19:47:33 crc kubenswrapper[4988]: I1008 19:47:33.491302 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7198c3ac-8210-43f0-bd60-e18bb3da7de3-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "7198c3ac-8210-43f0-bd60-e18bb3da7de3" (UID: "7198c3ac-8210-43f0-bd60-e18bb3da7de3"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 19:47:33 crc kubenswrapper[4988]: I1008 19:47:33.491651 4988 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/089988dc-da74-4d2b-a9d3-c3ad66a17dac-logs\") on node \"crc\" DevicePath \"\"" Oct 08 19:47:33 crc kubenswrapper[4988]: I1008 19:47:33.491669 4988 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/7198c3ac-8210-43f0-bd60-e18bb3da7de3-httpd-run\") on node \"crc\" DevicePath \"\"" Oct 08 19:47:33 crc kubenswrapper[4988]: I1008 19:47:33.492253 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7198c3ac-8210-43f0-bd60-e18bb3da7de3-logs" (OuterVolumeSpecName: "logs") pod "7198c3ac-8210-43f0-bd60-e18bb3da7de3" (UID: "7198c3ac-8210-43f0-bd60-e18bb3da7de3"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 19:47:33 crc kubenswrapper[4988]: I1008 19:47:33.492363 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/089988dc-da74-4d2b-a9d3-c3ad66a17dac-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "089988dc-da74-4d2b-a9d3-c3ad66a17dac" (UID: "089988dc-da74-4d2b-a9d3-c3ad66a17dac"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 19:47:33 crc kubenswrapper[4988]: I1008 19:47:33.496693 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7198c3ac-8210-43f0-bd60-e18bb3da7de3-scripts" (OuterVolumeSpecName: "scripts") pod "7198c3ac-8210-43f0-bd60-e18bb3da7de3" (UID: "7198c3ac-8210-43f0-bd60-e18bb3da7de3"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 19:47:33 crc kubenswrapper[4988]: I1008 19:47:33.497072 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/089988dc-da74-4d2b-a9d3-c3ad66a17dac-scripts" (OuterVolumeSpecName: "scripts") pod "089988dc-da74-4d2b-a9d3-c3ad66a17dac" (UID: "089988dc-da74-4d2b-a9d3-c3ad66a17dac"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 19:47:33 crc kubenswrapper[4988]: I1008 19:47:33.497080 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7198c3ac-8210-43f0-bd60-e18bb3da7de3-kube-api-access-87v69" (OuterVolumeSpecName: "kube-api-access-87v69") pod "7198c3ac-8210-43f0-bd60-e18bb3da7de3" (UID: "7198c3ac-8210-43f0-bd60-e18bb3da7de3"). InnerVolumeSpecName "kube-api-access-87v69". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 19:47:33 crc kubenswrapper[4988]: I1008 19:47:33.497210 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/089988dc-da74-4d2b-a9d3-c3ad66a17dac-kube-api-access-q2866" (OuterVolumeSpecName: "kube-api-access-q2866") pod "089988dc-da74-4d2b-a9d3-c3ad66a17dac" (UID: "089988dc-da74-4d2b-a9d3-c3ad66a17dac"). InnerVolumeSpecName "kube-api-access-q2866". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 19:47:33 crc kubenswrapper[4988]: I1008 19:47:33.524483 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/089988dc-da74-4d2b-a9d3-c3ad66a17dac-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "089988dc-da74-4d2b-a9d3-c3ad66a17dac" (UID: "089988dc-da74-4d2b-a9d3-c3ad66a17dac"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 19:47:33 crc kubenswrapper[4988]: I1008 19:47:33.525406 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7198c3ac-8210-43f0-bd60-e18bb3da7de3-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "7198c3ac-8210-43f0-bd60-e18bb3da7de3" (UID: "7198c3ac-8210-43f0-bd60-e18bb3da7de3"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 19:47:33 crc kubenswrapper[4988]: I1008 19:47:33.553760 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/089988dc-da74-4d2b-a9d3-c3ad66a17dac-config-data" (OuterVolumeSpecName: "config-data") pod "089988dc-da74-4d2b-a9d3-c3ad66a17dac" (UID: "089988dc-da74-4d2b-a9d3-c3ad66a17dac"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 19:47:33 crc kubenswrapper[4988]: I1008 19:47:33.554258 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7198c3ac-8210-43f0-bd60-e18bb3da7de3-config-data" (OuterVolumeSpecName: "config-data") pod "7198c3ac-8210-43f0-bd60-e18bb3da7de3" (UID: "7198c3ac-8210-43f0-bd60-e18bb3da7de3"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 19:47:33 crc kubenswrapper[4988]: I1008 19:47:33.592937 4988 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/089988dc-da74-4d2b-a9d3-c3ad66a17dac-httpd-run\") on node \"crc\" DevicePath \"\"" Oct 08 19:47:33 crc kubenswrapper[4988]: I1008 19:47:33.592968 4988 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7198c3ac-8210-43f0-bd60-e18bb3da7de3-config-data\") on node \"crc\" DevicePath \"\"" Oct 08 19:47:33 crc kubenswrapper[4988]: I1008 19:47:33.592977 4988 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/089988dc-da74-4d2b-a9d3-c3ad66a17dac-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 19:47:33 crc kubenswrapper[4988]: I1008 19:47:33.592991 4988 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7198c3ac-8210-43f0-bd60-e18bb3da7de3-logs\") on node \"crc\" DevicePath \"\"" Oct 08 19:47:33 crc kubenswrapper[4988]: I1008 19:47:33.593000 4988 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-87v69\" (UniqueName: \"kubernetes.io/projected/7198c3ac-8210-43f0-bd60-e18bb3da7de3-kube-api-access-87v69\") on node \"crc\" DevicePath \"\"" Oct 08 19:47:33 crc kubenswrapper[4988]: I1008 19:47:33.593008 4988 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/089988dc-da74-4d2b-a9d3-c3ad66a17dac-config-data\") on node \"crc\" DevicePath \"\"" Oct 08 19:47:33 crc kubenswrapper[4988]: I1008 19:47:33.593018 4988 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7198c3ac-8210-43f0-bd60-e18bb3da7de3-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 19:47:33 crc kubenswrapper[4988]: I1008 19:47:33.593026 4988 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/089988dc-da74-4d2b-a9d3-c3ad66a17dac-scripts\") on node \"crc\" DevicePath \"\"" Oct 08 19:47:33 crc kubenswrapper[4988]: I1008 19:47:33.593033 4988 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7198c3ac-8210-43f0-bd60-e18bb3da7de3-scripts\") on node \"crc\" DevicePath \"\"" Oct 08 19:47:33 crc kubenswrapper[4988]: I1008 19:47:33.593040 4988 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q2866\" (UniqueName: \"kubernetes.io/projected/089988dc-da74-4d2b-a9d3-c3ad66a17dac-kube-api-access-q2866\") on node \"crc\" DevicePath \"\"" Oct 08 19:47:33 crc kubenswrapper[4988]: I1008 19:47:33.635904 4988 generic.go:334] "Generic (PLEG): container finished" podID="7198c3ac-8210-43f0-bd60-e18bb3da7de3" containerID="774051dbff0e9e9a4ecb1dd15c593ea0c50811d8ef6d13ab94101e4189441776" exitCode=0 Oct 08 19:47:33 crc kubenswrapper[4988]: I1008 19:47:33.635938 4988 generic.go:334] "Generic (PLEG): container finished" podID="7198c3ac-8210-43f0-bd60-e18bb3da7de3" containerID="946855f6f4945fba05575de7262453fb75bd7867dea5aaea069f8748e2d681c9" exitCode=143 Oct 08 19:47:33 crc kubenswrapper[4988]: I1008 19:47:33.635992 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"7198c3ac-8210-43f0-bd60-e18bb3da7de3","Type":"ContainerDied","Data":"774051dbff0e9e9a4ecb1dd15c593ea0c50811d8ef6d13ab94101e4189441776"} Oct 08 19:47:33 crc kubenswrapper[4988]: I1008 19:47:33.636045 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"7198c3ac-8210-43f0-bd60-e18bb3da7de3","Type":"ContainerDied","Data":"946855f6f4945fba05575de7262453fb75bd7867dea5aaea069f8748e2d681c9"} Oct 08 19:47:33 crc kubenswrapper[4988]: I1008 19:47:33.636056 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"7198c3ac-8210-43f0-bd60-e18bb3da7de3","Type":"ContainerDied","Data":"905c84e528dbfdd9d49eaf89926f2273413eb50e311aa40f2860be6bc5c59f44"} Oct 08 19:47:33 crc kubenswrapper[4988]: I1008 19:47:33.636073 4988 scope.go:117] "RemoveContainer" containerID="774051dbff0e9e9a4ecb1dd15c593ea0c50811d8ef6d13ab94101e4189441776" Oct 08 19:47:33 crc kubenswrapper[4988]: I1008 19:47:33.636012 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 08 19:47:33 crc kubenswrapper[4988]: I1008 19:47:33.640727 4988 generic.go:334] "Generic (PLEG): container finished" podID="089988dc-da74-4d2b-a9d3-c3ad66a17dac" containerID="df1652a004c93e3aeab64779b2128aa987ae2c43bc9da3e84a0bdf38d1393a10" exitCode=0 Oct 08 19:47:33 crc kubenswrapper[4988]: I1008 19:47:33.640844 4988 generic.go:334] "Generic (PLEG): container finished" podID="089988dc-da74-4d2b-a9d3-c3ad66a17dac" containerID="eae9a4f1f2c8cb0bd59489b363c518b91d7ba118e959aaa791e2b4295597f7fc" exitCode=143 Oct 08 19:47:33 crc kubenswrapper[4988]: I1008 19:47:33.641806 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 08 19:47:33 crc kubenswrapper[4988]: I1008 19:47:33.645564 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"089988dc-da74-4d2b-a9d3-c3ad66a17dac","Type":"ContainerDied","Data":"df1652a004c93e3aeab64779b2128aa987ae2c43bc9da3e84a0bdf38d1393a10"} Oct 08 19:47:33 crc kubenswrapper[4988]: I1008 19:47:33.645609 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"089988dc-da74-4d2b-a9d3-c3ad66a17dac","Type":"ContainerDied","Data":"eae9a4f1f2c8cb0bd59489b363c518b91d7ba118e959aaa791e2b4295597f7fc"} Oct 08 19:47:33 crc kubenswrapper[4988]: I1008 19:47:33.645620 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"089988dc-da74-4d2b-a9d3-c3ad66a17dac","Type":"ContainerDied","Data":"d4f4c0dff727040260bbd428edd53882fe244baa5f60b8d2a6c99e0f94902c72"} Oct 08 19:47:33 crc kubenswrapper[4988]: I1008 19:47:33.669444 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 08 19:47:33 crc kubenswrapper[4988]: I1008 19:47:33.691237 4988 scope.go:117] "RemoveContainer" containerID="946855f6f4945fba05575de7262453fb75bd7867dea5aaea069f8748e2d681c9" Oct 08 19:47:33 crc kubenswrapper[4988]: I1008 19:47:33.742691 4988 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 08 19:47:33 crc kubenswrapper[4988]: I1008 19:47:33.746455 4988 scope.go:117] "RemoveContainer" containerID="774051dbff0e9e9a4ecb1dd15c593ea0c50811d8ef6d13ab94101e4189441776" Oct 08 19:47:33 crc kubenswrapper[4988]: E1008 19:47:33.747371 4988 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"774051dbff0e9e9a4ecb1dd15c593ea0c50811d8ef6d13ab94101e4189441776\": container with ID starting with 774051dbff0e9e9a4ecb1dd15c593ea0c50811d8ef6d13ab94101e4189441776 not found: ID does not exist" containerID="774051dbff0e9e9a4ecb1dd15c593ea0c50811d8ef6d13ab94101e4189441776" Oct 08 19:47:33 crc kubenswrapper[4988]: I1008 19:47:33.747440 4988 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"774051dbff0e9e9a4ecb1dd15c593ea0c50811d8ef6d13ab94101e4189441776"} err="failed to get container status \"774051dbff0e9e9a4ecb1dd15c593ea0c50811d8ef6d13ab94101e4189441776\": rpc error: code = NotFound desc = could not find container \"774051dbff0e9e9a4ecb1dd15c593ea0c50811d8ef6d13ab94101e4189441776\": container with ID starting with 774051dbff0e9e9a4ecb1dd15c593ea0c50811d8ef6d13ab94101e4189441776 not found: ID does not exist" Oct 08 19:47:33 crc kubenswrapper[4988]: I1008 19:47:33.747466 4988 scope.go:117] "RemoveContainer" containerID="946855f6f4945fba05575de7262453fb75bd7867dea5aaea069f8748e2d681c9" Oct 08 19:47:33 crc kubenswrapper[4988]: E1008 19:47:33.747836 4988 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"946855f6f4945fba05575de7262453fb75bd7867dea5aaea069f8748e2d681c9\": container with ID starting with 946855f6f4945fba05575de7262453fb75bd7867dea5aaea069f8748e2d681c9 not found: ID does not exist" containerID="946855f6f4945fba05575de7262453fb75bd7867dea5aaea069f8748e2d681c9" Oct 08 19:47:33 crc kubenswrapper[4988]: I1008 19:47:33.747851 4988 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"946855f6f4945fba05575de7262453fb75bd7867dea5aaea069f8748e2d681c9"} err="failed to get container status \"946855f6f4945fba05575de7262453fb75bd7867dea5aaea069f8748e2d681c9\": rpc error: code = NotFound desc = could not find container \"946855f6f4945fba05575de7262453fb75bd7867dea5aaea069f8748e2d681c9\": container with ID starting with 946855f6f4945fba05575de7262453fb75bd7867dea5aaea069f8748e2d681c9 not found: ID does not exist" Oct 08 19:47:33 crc kubenswrapper[4988]: I1008 19:47:33.747864 4988 scope.go:117] "RemoveContainer" containerID="774051dbff0e9e9a4ecb1dd15c593ea0c50811d8ef6d13ab94101e4189441776" Oct 08 19:47:33 crc kubenswrapper[4988]: I1008 19:47:33.748025 4988 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"774051dbff0e9e9a4ecb1dd15c593ea0c50811d8ef6d13ab94101e4189441776"} err="failed to get container status \"774051dbff0e9e9a4ecb1dd15c593ea0c50811d8ef6d13ab94101e4189441776\": rpc error: code = NotFound desc = could not find container \"774051dbff0e9e9a4ecb1dd15c593ea0c50811d8ef6d13ab94101e4189441776\": container with ID starting with 774051dbff0e9e9a4ecb1dd15c593ea0c50811d8ef6d13ab94101e4189441776 not found: ID does not exist" Oct 08 19:47:33 crc kubenswrapper[4988]: I1008 19:47:33.748038 4988 scope.go:117] "RemoveContainer" containerID="946855f6f4945fba05575de7262453fb75bd7867dea5aaea069f8748e2d681c9" Oct 08 19:47:33 crc kubenswrapper[4988]: I1008 19:47:33.748202 4988 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"946855f6f4945fba05575de7262453fb75bd7867dea5aaea069f8748e2d681c9"} err="failed to get container status \"946855f6f4945fba05575de7262453fb75bd7867dea5aaea069f8748e2d681c9\": rpc error: code = NotFound desc = could not find container \"946855f6f4945fba05575de7262453fb75bd7867dea5aaea069f8748e2d681c9\": container with ID starting with 946855f6f4945fba05575de7262453fb75bd7867dea5aaea069f8748e2d681c9 not found: ID does not exist" Oct 08 19:47:33 crc kubenswrapper[4988]: I1008 19:47:33.748214 4988 scope.go:117] "RemoveContainer" containerID="df1652a004c93e3aeab64779b2128aa987ae2c43bc9da3e84a0bdf38d1393a10" Oct 08 19:47:33 crc kubenswrapper[4988]: I1008 19:47:33.759509 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 08 19:47:33 crc kubenswrapper[4988]: E1008 19:47:33.759910 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="089988dc-da74-4d2b-a9d3-c3ad66a17dac" containerName="glance-log" Oct 08 19:47:33 crc kubenswrapper[4988]: I1008 19:47:33.759924 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="089988dc-da74-4d2b-a9d3-c3ad66a17dac" containerName="glance-log" Oct 08 19:47:33 crc kubenswrapper[4988]: E1008 19:47:33.759942 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7198c3ac-8210-43f0-bd60-e18bb3da7de3" containerName="glance-httpd" Oct 08 19:47:33 crc kubenswrapper[4988]: I1008 19:47:33.759947 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="7198c3ac-8210-43f0-bd60-e18bb3da7de3" containerName="glance-httpd" Oct 08 19:47:33 crc kubenswrapper[4988]: E1008 19:47:33.759970 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7198c3ac-8210-43f0-bd60-e18bb3da7de3" containerName="glance-log" Oct 08 19:47:33 crc kubenswrapper[4988]: I1008 19:47:33.759977 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="7198c3ac-8210-43f0-bd60-e18bb3da7de3" containerName="glance-log" Oct 08 19:47:33 crc kubenswrapper[4988]: E1008 19:47:33.760001 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="089988dc-da74-4d2b-a9d3-c3ad66a17dac" containerName="glance-httpd" Oct 08 19:47:33 crc kubenswrapper[4988]: I1008 19:47:33.760007 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="089988dc-da74-4d2b-a9d3-c3ad66a17dac" containerName="glance-httpd" Oct 08 19:47:33 crc kubenswrapper[4988]: I1008 19:47:33.760182 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="7198c3ac-8210-43f0-bd60-e18bb3da7de3" containerName="glance-log" Oct 08 19:47:33 crc kubenswrapper[4988]: I1008 19:47:33.760210 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="7198c3ac-8210-43f0-bd60-e18bb3da7de3" containerName="glance-httpd" Oct 08 19:47:33 crc kubenswrapper[4988]: I1008 19:47:33.760222 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="089988dc-da74-4d2b-a9d3-c3ad66a17dac" containerName="glance-log" Oct 08 19:47:33 crc kubenswrapper[4988]: I1008 19:47:33.760231 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="089988dc-da74-4d2b-a9d3-c3ad66a17dac" containerName="glance-httpd" Oct 08 19:47:33 crc kubenswrapper[4988]: I1008 19:47:33.762170 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 08 19:47:33 crc kubenswrapper[4988]: I1008 19:47:33.767184 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Oct 08 19:47:33 crc kubenswrapper[4988]: I1008 19:47:33.767694 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-scripts" Oct 08 19:47:33 crc kubenswrapper[4988]: I1008 19:47:33.767944 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Oct 08 19:47:33 crc kubenswrapper[4988]: I1008 19:47:33.768151 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-b6m4j" Oct 08 19:47:33 crc kubenswrapper[4988]: I1008 19:47:33.774577 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 08 19:47:33 crc kubenswrapper[4988]: I1008 19:47:33.774758 4988 scope.go:117] "RemoveContainer" containerID="eae9a4f1f2c8cb0bd59489b363c518b91d7ba118e959aaa791e2b4295597f7fc" Oct 08 19:47:33 crc kubenswrapper[4988]: I1008 19:47:33.785421 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 08 19:47:33 crc kubenswrapper[4988]: I1008 19:47:33.794611 4988 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 08 19:47:33 crc kubenswrapper[4988]: I1008 19:47:33.803779 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Oct 08 19:47:33 crc kubenswrapper[4988]: I1008 19:47:33.804061 4988 scope.go:117] "RemoveContainer" containerID="df1652a004c93e3aeab64779b2128aa987ae2c43bc9da3e84a0bdf38d1393a10" Oct 08 19:47:33 crc kubenswrapper[4988]: I1008 19:47:33.806447 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 08 19:47:33 crc kubenswrapper[4988]: E1008 19:47:33.807927 4988 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"df1652a004c93e3aeab64779b2128aa987ae2c43bc9da3e84a0bdf38d1393a10\": container with ID starting with df1652a004c93e3aeab64779b2128aa987ae2c43bc9da3e84a0bdf38d1393a10 not found: ID does not exist" containerID="df1652a004c93e3aeab64779b2128aa987ae2c43bc9da3e84a0bdf38d1393a10" Oct 08 19:47:33 crc kubenswrapper[4988]: I1008 19:47:33.808042 4988 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"df1652a004c93e3aeab64779b2128aa987ae2c43bc9da3e84a0bdf38d1393a10"} err="failed to get container status \"df1652a004c93e3aeab64779b2128aa987ae2c43bc9da3e84a0bdf38d1393a10\": rpc error: code = NotFound desc = could not find container \"df1652a004c93e3aeab64779b2128aa987ae2c43bc9da3e84a0bdf38d1393a10\": container with ID starting with df1652a004c93e3aeab64779b2128aa987ae2c43bc9da3e84a0bdf38d1393a10 not found: ID does not exist" Oct 08 19:47:33 crc kubenswrapper[4988]: I1008 19:47:33.808138 4988 scope.go:117] "RemoveContainer" containerID="eae9a4f1f2c8cb0bd59489b363c518b91d7ba118e959aaa791e2b4295597f7fc" Oct 08 19:47:33 crc kubenswrapper[4988]: E1008 19:47:33.809835 4988 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"eae9a4f1f2c8cb0bd59489b363c518b91d7ba118e959aaa791e2b4295597f7fc\": container with ID starting with eae9a4f1f2c8cb0bd59489b363c518b91d7ba118e959aaa791e2b4295597f7fc not found: ID does not exist" containerID="eae9a4f1f2c8cb0bd59489b363c518b91d7ba118e959aaa791e2b4295597f7fc" Oct 08 19:47:33 crc kubenswrapper[4988]: I1008 19:47:33.809949 4988 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"eae9a4f1f2c8cb0bd59489b363c518b91d7ba118e959aaa791e2b4295597f7fc"} err="failed to get container status \"eae9a4f1f2c8cb0bd59489b363c518b91d7ba118e959aaa791e2b4295597f7fc\": rpc error: code = NotFound desc = could not find container \"eae9a4f1f2c8cb0bd59489b363c518b91d7ba118e959aaa791e2b4295597f7fc\": container with ID starting with eae9a4f1f2c8cb0bd59489b363c518b91d7ba118e959aaa791e2b4295597f7fc not found: ID does not exist" Oct 08 19:47:33 crc kubenswrapper[4988]: I1008 19:47:33.810835 4988 scope.go:117] "RemoveContainer" containerID="df1652a004c93e3aeab64779b2128aa987ae2c43bc9da3e84a0bdf38d1393a10" Oct 08 19:47:33 crc kubenswrapper[4988]: I1008 19:47:33.809834 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Oct 08 19:47:33 crc kubenswrapper[4988]: I1008 19:47:33.809905 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Oct 08 19:47:33 crc kubenswrapper[4988]: I1008 19:47:33.814950 4988 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"df1652a004c93e3aeab64779b2128aa987ae2c43bc9da3e84a0bdf38d1393a10"} err="failed to get container status \"df1652a004c93e3aeab64779b2128aa987ae2c43bc9da3e84a0bdf38d1393a10\": rpc error: code = NotFound desc = could not find container \"df1652a004c93e3aeab64779b2128aa987ae2c43bc9da3e84a0bdf38d1393a10\": container with ID starting with df1652a004c93e3aeab64779b2128aa987ae2c43bc9da3e84a0bdf38d1393a10 not found: ID does not exist" Oct 08 19:47:33 crc kubenswrapper[4988]: I1008 19:47:33.815007 4988 scope.go:117] "RemoveContainer" containerID="eae9a4f1f2c8cb0bd59489b363c518b91d7ba118e959aaa791e2b4295597f7fc" Oct 08 19:47:33 crc kubenswrapper[4988]: I1008 19:47:33.815269 4988 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"eae9a4f1f2c8cb0bd59489b363c518b91d7ba118e959aaa791e2b4295597f7fc"} err="failed to get container status \"eae9a4f1f2c8cb0bd59489b363c518b91d7ba118e959aaa791e2b4295597f7fc\": rpc error: code = NotFound desc = could not find container \"eae9a4f1f2c8cb0bd59489b363c518b91d7ba118e959aaa791e2b4295597f7fc\": container with ID starting with eae9a4f1f2c8cb0bd59489b363c518b91d7ba118e959aaa791e2b4295597f7fc not found: ID does not exist" Oct 08 19:47:33 crc kubenswrapper[4988]: I1008 19:47:33.816526 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 08 19:47:33 crc kubenswrapper[4988]: I1008 19:47:33.900886 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/185123ed-b356-4268-af3d-e44ea04bbf3f-logs\") pod \"glance-default-external-api-0\" (UID: \"185123ed-b356-4268-af3d-e44ea04bbf3f\") " pod="openstack/glance-default-external-api-0" Oct 08 19:47:33 crc kubenswrapper[4988]: I1008 19:47:33.900937 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d29618a6-2810-4d35-b5b3-2369d4a1c096-config-data\") pod \"glance-default-internal-api-0\" (UID: \"d29618a6-2810-4d35-b5b3-2369d4a1c096\") " pod="openstack/glance-default-internal-api-0" Oct 08 19:47:33 crc kubenswrapper[4988]: I1008 19:47:33.900976 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d29618a6-2810-4d35-b5b3-2369d4a1c096-scripts\") pod \"glance-default-internal-api-0\" (UID: \"d29618a6-2810-4d35-b5b3-2369d4a1c096\") " pod="openstack/glance-default-internal-api-0" Oct 08 19:47:33 crc kubenswrapper[4988]: I1008 19:47:33.900995 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/d29618a6-2810-4d35-b5b3-2369d4a1c096-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"d29618a6-2810-4d35-b5b3-2369d4a1c096\") " pod="openstack/glance-default-internal-api-0" Oct 08 19:47:33 crc kubenswrapper[4988]: I1008 19:47:33.901019 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d29618a6-2810-4d35-b5b3-2369d4a1c096-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"d29618a6-2810-4d35-b5b3-2369d4a1c096\") " pod="openstack/glance-default-internal-api-0" Oct 08 19:47:33 crc kubenswrapper[4988]: I1008 19:47:33.901040 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/185123ed-b356-4268-af3d-e44ea04bbf3f-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"185123ed-b356-4268-af3d-e44ea04bbf3f\") " pod="openstack/glance-default-external-api-0" Oct 08 19:47:33 crc kubenswrapper[4988]: I1008 19:47:33.901061 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/185123ed-b356-4268-af3d-e44ea04bbf3f-config-data\") pod \"glance-default-external-api-0\" (UID: \"185123ed-b356-4268-af3d-e44ea04bbf3f\") " pod="openstack/glance-default-external-api-0" Oct 08 19:47:33 crc kubenswrapper[4988]: I1008 19:47:33.901079 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gtpmt\" (UniqueName: \"kubernetes.io/projected/d29618a6-2810-4d35-b5b3-2369d4a1c096-kube-api-access-gtpmt\") pod \"glance-default-internal-api-0\" (UID: \"d29618a6-2810-4d35-b5b3-2369d4a1c096\") " pod="openstack/glance-default-internal-api-0" Oct 08 19:47:33 crc kubenswrapper[4988]: I1008 19:47:33.901099 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5fcjj\" (UniqueName: \"kubernetes.io/projected/185123ed-b356-4268-af3d-e44ea04bbf3f-kube-api-access-5fcjj\") pod \"glance-default-external-api-0\" (UID: \"185123ed-b356-4268-af3d-e44ea04bbf3f\") " pod="openstack/glance-default-external-api-0" Oct 08 19:47:33 crc kubenswrapper[4988]: I1008 19:47:33.901164 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d29618a6-2810-4d35-b5b3-2369d4a1c096-logs\") pod \"glance-default-internal-api-0\" (UID: \"d29618a6-2810-4d35-b5b3-2369d4a1c096\") " pod="openstack/glance-default-internal-api-0" Oct 08 19:47:33 crc kubenswrapper[4988]: I1008 19:47:33.901182 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/185123ed-b356-4268-af3d-e44ea04bbf3f-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"185123ed-b356-4268-af3d-e44ea04bbf3f\") " pod="openstack/glance-default-external-api-0" Oct 08 19:47:33 crc kubenswrapper[4988]: I1008 19:47:33.901197 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/185123ed-b356-4268-af3d-e44ea04bbf3f-scripts\") pod \"glance-default-external-api-0\" (UID: \"185123ed-b356-4268-af3d-e44ea04bbf3f\") " pod="openstack/glance-default-external-api-0" Oct 08 19:47:33 crc kubenswrapper[4988]: I1008 19:47:33.901231 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d29618a6-2810-4d35-b5b3-2369d4a1c096-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"d29618a6-2810-4d35-b5b3-2369d4a1c096\") " pod="openstack/glance-default-internal-api-0" Oct 08 19:47:33 crc kubenswrapper[4988]: I1008 19:47:33.901254 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/185123ed-b356-4268-af3d-e44ea04bbf3f-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"185123ed-b356-4268-af3d-e44ea04bbf3f\") " pod="openstack/glance-default-external-api-0" Oct 08 19:47:34 crc kubenswrapper[4988]: I1008 19:47:34.003048 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d29618a6-2810-4d35-b5b3-2369d4a1c096-logs\") pod \"glance-default-internal-api-0\" (UID: \"d29618a6-2810-4d35-b5b3-2369d4a1c096\") " pod="openstack/glance-default-internal-api-0" Oct 08 19:47:34 crc kubenswrapper[4988]: I1008 19:47:34.003091 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/185123ed-b356-4268-af3d-e44ea04bbf3f-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"185123ed-b356-4268-af3d-e44ea04bbf3f\") " pod="openstack/glance-default-external-api-0" Oct 08 19:47:34 crc kubenswrapper[4988]: I1008 19:47:34.003108 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/185123ed-b356-4268-af3d-e44ea04bbf3f-scripts\") pod \"glance-default-external-api-0\" (UID: \"185123ed-b356-4268-af3d-e44ea04bbf3f\") " pod="openstack/glance-default-external-api-0" Oct 08 19:47:34 crc kubenswrapper[4988]: I1008 19:47:34.003146 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d29618a6-2810-4d35-b5b3-2369d4a1c096-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"d29618a6-2810-4d35-b5b3-2369d4a1c096\") " pod="openstack/glance-default-internal-api-0" Oct 08 19:47:34 crc kubenswrapper[4988]: I1008 19:47:34.003163 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/185123ed-b356-4268-af3d-e44ea04bbf3f-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"185123ed-b356-4268-af3d-e44ea04bbf3f\") " pod="openstack/glance-default-external-api-0" Oct 08 19:47:34 crc kubenswrapper[4988]: I1008 19:47:34.003186 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/185123ed-b356-4268-af3d-e44ea04bbf3f-logs\") pod \"glance-default-external-api-0\" (UID: \"185123ed-b356-4268-af3d-e44ea04bbf3f\") " pod="openstack/glance-default-external-api-0" Oct 08 19:47:34 crc kubenswrapper[4988]: I1008 19:47:34.003213 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d29618a6-2810-4d35-b5b3-2369d4a1c096-config-data\") pod \"glance-default-internal-api-0\" (UID: \"d29618a6-2810-4d35-b5b3-2369d4a1c096\") " pod="openstack/glance-default-internal-api-0" Oct 08 19:47:34 crc kubenswrapper[4988]: I1008 19:47:34.003245 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d29618a6-2810-4d35-b5b3-2369d4a1c096-scripts\") pod \"glance-default-internal-api-0\" (UID: \"d29618a6-2810-4d35-b5b3-2369d4a1c096\") " pod="openstack/glance-default-internal-api-0" Oct 08 19:47:34 crc kubenswrapper[4988]: I1008 19:47:34.003265 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/d29618a6-2810-4d35-b5b3-2369d4a1c096-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"d29618a6-2810-4d35-b5b3-2369d4a1c096\") " pod="openstack/glance-default-internal-api-0" Oct 08 19:47:34 crc kubenswrapper[4988]: I1008 19:47:34.003288 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d29618a6-2810-4d35-b5b3-2369d4a1c096-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"d29618a6-2810-4d35-b5b3-2369d4a1c096\") " pod="openstack/glance-default-internal-api-0" Oct 08 19:47:34 crc kubenswrapper[4988]: I1008 19:47:34.003307 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/185123ed-b356-4268-af3d-e44ea04bbf3f-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"185123ed-b356-4268-af3d-e44ea04bbf3f\") " pod="openstack/glance-default-external-api-0" Oct 08 19:47:34 crc kubenswrapper[4988]: I1008 19:47:34.003328 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/185123ed-b356-4268-af3d-e44ea04bbf3f-config-data\") pod \"glance-default-external-api-0\" (UID: \"185123ed-b356-4268-af3d-e44ea04bbf3f\") " pod="openstack/glance-default-external-api-0" Oct 08 19:47:34 crc kubenswrapper[4988]: I1008 19:47:34.003347 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gtpmt\" (UniqueName: \"kubernetes.io/projected/d29618a6-2810-4d35-b5b3-2369d4a1c096-kube-api-access-gtpmt\") pod \"glance-default-internal-api-0\" (UID: \"d29618a6-2810-4d35-b5b3-2369d4a1c096\") " pod="openstack/glance-default-internal-api-0" Oct 08 19:47:34 crc kubenswrapper[4988]: I1008 19:47:34.003368 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5fcjj\" (UniqueName: \"kubernetes.io/projected/185123ed-b356-4268-af3d-e44ea04bbf3f-kube-api-access-5fcjj\") pod \"glance-default-external-api-0\" (UID: \"185123ed-b356-4268-af3d-e44ea04bbf3f\") " pod="openstack/glance-default-external-api-0" Oct 08 19:47:34 crc kubenswrapper[4988]: I1008 19:47:34.003709 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/185123ed-b356-4268-af3d-e44ea04bbf3f-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"185123ed-b356-4268-af3d-e44ea04bbf3f\") " pod="openstack/glance-default-external-api-0" Oct 08 19:47:34 crc kubenswrapper[4988]: I1008 19:47:34.003769 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d29618a6-2810-4d35-b5b3-2369d4a1c096-logs\") pod \"glance-default-internal-api-0\" (UID: \"d29618a6-2810-4d35-b5b3-2369d4a1c096\") " pod="openstack/glance-default-internal-api-0" Oct 08 19:47:34 crc kubenswrapper[4988]: I1008 19:47:34.004031 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/d29618a6-2810-4d35-b5b3-2369d4a1c096-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"d29618a6-2810-4d35-b5b3-2369d4a1c096\") " pod="openstack/glance-default-internal-api-0" Oct 08 19:47:34 crc kubenswrapper[4988]: I1008 19:47:34.004506 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/185123ed-b356-4268-af3d-e44ea04bbf3f-logs\") pod \"glance-default-external-api-0\" (UID: \"185123ed-b356-4268-af3d-e44ea04bbf3f\") " pod="openstack/glance-default-external-api-0" Oct 08 19:47:34 crc kubenswrapper[4988]: I1008 19:47:34.007516 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/185123ed-b356-4268-af3d-e44ea04bbf3f-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"185123ed-b356-4268-af3d-e44ea04bbf3f\") " pod="openstack/glance-default-external-api-0" Oct 08 19:47:34 crc kubenswrapper[4988]: I1008 19:47:34.007716 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/185123ed-b356-4268-af3d-e44ea04bbf3f-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"185123ed-b356-4268-af3d-e44ea04bbf3f\") " pod="openstack/glance-default-external-api-0" Oct 08 19:47:34 crc kubenswrapper[4988]: I1008 19:47:34.008820 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/185123ed-b356-4268-af3d-e44ea04bbf3f-scripts\") pod \"glance-default-external-api-0\" (UID: \"185123ed-b356-4268-af3d-e44ea04bbf3f\") " pod="openstack/glance-default-external-api-0" Oct 08 19:47:34 crc kubenswrapper[4988]: I1008 19:47:34.009013 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d29618a6-2810-4d35-b5b3-2369d4a1c096-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"d29618a6-2810-4d35-b5b3-2369d4a1c096\") " pod="openstack/glance-default-internal-api-0" Oct 08 19:47:34 crc kubenswrapper[4988]: I1008 19:47:34.009058 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d29618a6-2810-4d35-b5b3-2369d4a1c096-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"d29618a6-2810-4d35-b5b3-2369d4a1c096\") " pod="openstack/glance-default-internal-api-0" Oct 08 19:47:34 crc kubenswrapper[4988]: I1008 19:47:34.011349 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d29618a6-2810-4d35-b5b3-2369d4a1c096-config-data\") pod \"glance-default-internal-api-0\" (UID: \"d29618a6-2810-4d35-b5b3-2369d4a1c096\") " pod="openstack/glance-default-internal-api-0" Oct 08 19:47:34 crc kubenswrapper[4988]: I1008 19:47:34.012694 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/185123ed-b356-4268-af3d-e44ea04bbf3f-config-data\") pod \"glance-default-external-api-0\" (UID: \"185123ed-b356-4268-af3d-e44ea04bbf3f\") " pod="openstack/glance-default-external-api-0" Oct 08 19:47:34 crc kubenswrapper[4988]: I1008 19:47:34.012877 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d29618a6-2810-4d35-b5b3-2369d4a1c096-scripts\") pod \"glance-default-internal-api-0\" (UID: \"d29618a6-2810-4d35-b5b3-2369d4a1c096\") " pod="openstack/glance-default-internal-api-0" Oct 08 19:47:34 crc kubenswrapper[4988]: I1008 19:47:34.020885 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5fcjj\" (UniqueName: \"kubernetes.io/projected/185123ed-b356-4268-af3d-e44ea04bbf3f-kube-api-access-5fcjj\") pod \"glance-default-external-api-0\" (UID: \"185123ed-b356-4268-af3d-e44ea04bbf3f\") " pod="openstack/glance-default-external-api-0" Oct 08 19:47:34 crc kubenswrapper[4988]: I1008 19:47:34.024689 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gtpmt\" (UniqueName: \"kubernetes.io/projected/d29618a6-2810-4d35-b5b3-2369d4a1c096-kube-api-access-gtpmt\") pod \"glance-default-internal-api-0\" (UID: \"d29618a6-2810-4d35-b5b3-2369d4a1c096\") " pod="openstack/glance-default-internal-api-0" Oct 08 19:47:34 crc kubenswrapper[4988]: I1008 19:47:34.091848 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 08 19:47:34 crc kubenswrapper[4988]: I1008 19:47:34.124035 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 08 19:47:34 crc kubenswrapper[4988]: I1008 19:47:34.662603 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 08 19:47:34 crc kubenswrapper[4988]: W1008 19:47:34.671580 4988 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd29618a6_2810_4d35_b5b3_2369d4a1c096.slice/crio-5048f4c53afd16b33e344d6fd27661ebed068a0c390359e8edb0f0df863bbe36 WatchSource:0}: Error finding container 5048f4c53afd16b33e344d6fd27661ebed068a0c390359e8edb0f0df863bbe36: Status 404 returned error can't find the container with id 5048f4c53afd16b33e344d6fd27661ebed068a0c390359e8edb0f0df863bbe36 Oct 08 19:47:34 crc kubenswrapper[4988]: I1008 19:47:34.769153 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 08 19:47:34 crc kubenswrapper[4988]: W1008 19:47:34.772242 4988 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod185123ed_b356_4268_af3d_e44ea04bbf3f.slice/crio-61685588cdeb820835a56ec6a4dbaee1105b8876ff91f5008a50025350a6f30d WatchSource:0}: Error finding container 61685588cdeb820835a56ec6a4dbaee1105b8876ff91f5008a50025350a6f30d: Status 404 returned error can't find the container with id 61685588cdeb820835a56ec6a4dbaee1105b8876ff91f5008a50025350a6f30d Oct 08 19:47:35 crc kubenswrapper[4988]: I1008 19:47:35.249673 4988 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="089988dc-da74-4d2b-a9d3-c3ad66a17dac" path="/var/lib/kubelet/pods/089988dc-da74-4d2b-a9d3-c3ad66a17dac/volumes" Oct 08 19:47:35 crc kubenswrapper[4988]: I1008 19:47:35.250660 4988 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7198c3ac-8210-43f0-bd60-e18bb3da7de3" path="/var/lib/kubelet/pods/7198c3ac-8210-43f0-bd60-e18bb3da7de3/volumes" Oct 08 19:47:35 crc kubenswrapper[4988]: I1008 19:47:35.678808 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"185123ed-b356-4268-af3d-e44ea04bbf3f","Type":"ContainerStarted","Data":"3e47ea1ab135b55cab053319b8c923376c4cda044c92fb262fdd29d259c417da"} Oct 08 19:47:35 crc kubenswrapper[4988]: I1008 19:47:35.678855 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"185123ed-b356-4268-af3d-e44ea04bbf3f","Type":"ContainerStarted","Data":"61685588cdeb820835a56ec6a4dbaee1105b8876ff91f5008a50025350a6f30d"} Oct 08 19:47:35 crc kubenswrapper[4988]: I1008 19:47:35.680540 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"d29618a6-2810-4d35-b5b3-2369d4a1c096","Type":"ContainerStarted","Data":"77b13c99dddd5b6f4b11423db1a866862933fa4da7f96cd4e3f36292c533bb98"} Oct 08 19:47:35 crc kubenswrapper[4988]: I1008 19:47:35.680560 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"d29618a6-2810-4d35-b5b3-2369d4a1c096","Type":"ContainerStarted","Data":"5048f4c53afd16b33e344d6fd27661ebed068a0c390359e8edb0f0df863bbe36"} Oct 08 19:47:36 crc kubenswrapper[4988]: I1008 19:47:36.698747 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"d29618a6-2810-4d35-b5b3-2369d4a1c096","Type":"ContainerStarted","Data":"63520e1979dd97c54c047ccdf382e36d66fc5644ef9ac1bde7c697c98007cbdd"} Oct 08 19:47:36 crc kubenswrapper[4988]: I1008 19:47:36.702254 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"185123ed-b356-4268-af3d-e44ea04bbf3f","Type":"ContainerStarted","Data":"1decc70dd2302ff2a788701c560d989e84796d90fcfb7cf0324d8c48e524d42e"} Oct 08 19:47:36 crc kubenswrapper[4988]: I1008 19:47:36.738022 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=3.737980873 podStartE2EDuration="3.737980873s" podCreationTimestamp="2025-10-08 19:47:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 19:47:36.731568849 +0000 UTC m=+5802.181411699" watchObservedRunningTime="2025-10-08 19:47:36.737980873 +0000 UTC m=+5802.187823653" Oct 08 19:47:36 crc kubenswrapper[4988]: I1008 19:47:36.773715 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=3.773694168 podStartE2EDuration="3.773694168s" podCreationTimestamp="2025-10-08 19:47:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 19:47:36.766648215 +0000 UTC m=+5802.216491055" watchObservedRunningTime="2025-10-08 19:47:36.773694168 +0000 UTC m=+5802.223536948" Oct 08 19:47:37 crc kubenswrapper[4988]: I1008 19:47:37.124348 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-8zqnt"] Oct 08 19:47:37 crc kubenswrapper[4988]: I1008 19:47:37.146056 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-8zqnt" Oct 08 19:47:37 crc kubenswrapper[4988]: I1008 19:47:37.146910 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-8zqnt"] Oct 08 19:47:37 crc kubenswrapper[4988]: I1008 19:47:37.263308 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8wzr9\" (UniqueName: \"kubernetes.io/projected/55444eeb-4abd-4b8f-840a-2cd7e508feae-kube-api-access-8wzr9\") pod \"redhat-marketplace-8zqnt\" (UID: \"55444eeb-4abd-4b8f-840a-2cd7e508feae\") " pod="openshift-marketplace/redhat-marketplace-8zqnt" Oct 08 19:47:37 crc kubenswrapper[4988]: I1008 19:47:37.263419 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/55444eeb-4abd-4b8f-840a-2cd7e508feae-catalog-content\") pod \"redhat-marketplace-8zqnt\" (UID: \"55444eeb-4abd-4b8f-840a-2cd7e508feae\") " pod="openshift-marketplace/redhat-marketplace-8zqnt" Oct 08 19:47:37 crc kubenswrapper[4988]: I1008 19:47:37.263795 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/55444eeb-4abd-4b8f-840a-2cd7e508feae-utilities\") pod \"redhat-marketplace-8zqnt\" (UID: \"55444eeb-4abd-4b8f-840a-2cd7e508feae\") " pod="openshift-marketplace/redhat-marketplace-8zqnt" Oct 08 19:47:37 crc kubenswrapper[4988]: I1008 19:47:37.365651 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/55444eeb-4abd-4b8f-840a-2cd7e508feae-catalog-content\") pod \"redhat-marketplace-8zqnt\" (UID: \"55444eeb-4abd-4b8f-840a-2cd7e508feae\") " pod="openshift-marketplace/redhat-marketplace-8zqnt" Oct 08 19:47:37 crc kubenswrapper[4988]: I1008 19:47:37.365777 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/55444eeb-4abd-4b8f-840a-2cd7e508feae-utilities\") pod \"redhat-marketplace-8zqnt\" (UID: \"55444eeb-4abd-4b8f-840a-2cd7e508feae\") " pod="openshift-marketplace/redhat-marketplace-8zqnt" Oct 08 19:47:37 crc kubenswrapper[4988]: I1008 19:47:37.365832 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8wzr9\" (UniqueName: \"kubernetes.io/projected/55444eeb-4abd-4b8f-840a-2cd7e508feae-kube-api-access-8wzr9\") pod \"redhat-marketplace-8zqnt\" (UID: \"55444eeb-4abd-4b8f-840a-2cd7e508feae\") " pod="openshift-marketplace/redhat-marketplace-8zqnt" Oct 08 19:47:37 crc kubenswrapper[4988]: I1008 19:47:37.366506 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/55444eeb-4abd-4b8f-840a-2cd7e508feae-catalog-content\") pod \"redhat-marketplace-8zqnt\" (UID: \"55444eeb-4abd-4b8f-840a-2cd7e508feae\") " pod="openshift-marketplace/redhat-marketplace-8zqnt" Oct 08 19:47:37 crc kubenswrapper[4988]: I1008 19:47:37.366711 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/55444eeb-4abd-4b8f-840a-2cd7e508feae-utilities\") pod \"redhat-marketplace-8zqnt\" (UID: \"55444eeb-4abd-4b8f-840a-2cd7e508feae\") " pod="openshift-marketplace/redhat-marketplace-8zqnt" Oct 08 19:47:37 crc kubenswrapper[4988]: I1008 19:47:37.399313 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8wzr9\" (UniqueName: \"kubernetes.io/projected/55444eeb-4abd-4b8f-840a-2cd7e508feae-kube-api-access-8wzr9\") pod \"redhat-marketplace-8zqnt\" (UID: \"55444eeb-4abd-4b8f-840a-2cd7e508feae\") " pod="openshift-marketplace/redhat-marketplace-8zqnt" Oct 08 19:47:37 crc kubenswrapper[4988]: I1008 19:47:37.481802 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-8zqnt" Oct 08 19:47:37 crc kubenswrapper[4988]: I1008 19:47:37.965039 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-8zqnt"] Oct 08 19:47:37 crc kubenswrapper[4988]: W1008 19:47:37.965119 4988 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod55444eeb_4abd_4b8f_840a_2cd7e508feae.slice/crio-c81d0b80ae5383cf3950beec554447311405accab41f53616b387d8439c6c950 WatchSource:0}: Error finding container c81d0b80ae5383cf3950beec554447311405accab41f53616b387d8439c6c950: Status 404 returned error can't find the container with id c81d0b80ae5383cf3950beec554447311405accab41f53616b387d8439c6c950 Oct 08 19:47:38 crc kubenswrapper[4988]: I1008 19:47:38.727926 4988 generic.go:334] "Generic (PLEG): container finished" podID="55444eeb-4abd-4b8f-840a-2cd7e508feae" containerID="b2ae7687b60cc018eb56d49d99e8db2f83fd91a0473c485d24e8182b518313d2" exitCode=0 Oct 08 19:47:38 crc kubenswrapper[4988]: I1008 19:47:38.728037 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8zqnt" event={"ID":"55444eeb-4abd-4b8f-840a-2cd7e508feae","Type":"ContainerDied","Data":"b2ae7687b60cc018eb56d49d99e8db2f83fd91a0473c485d24e8182b518313d2"} Oct 08 19:47:38 crc kubenswrapper[4988]: I1008 19:47:38.728550 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8zqnt" event={"ID":"55444eeb-4abd-4b8f-840a-2cd7e508feae","Type":"ContainerStarted","Data":"c81d0b80ae5383cf3950beec554447311405accab41f53616b387d8439c6c950"} Oct 08 19:47:38 crc kubenswrapper[4988]: I1008 19:47:38.730531 4988 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 08 19:47:39 crc kubenswrapper[4988]: I1008 19:47:39.743827 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8zqnt" event={"ID":"55444eeb-4abd-4b8f-840a-2cd7e508feae","Type":"ContainerStarted","Data":"a7cc809429fada31476c348ee9a9bc12595d75bacbfad867aa41712ca1ac92a9"} Oct 08 19:47:40 crc kubenswrapper[4988]: I1008 19:47:40.350574 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-5589f75ff5-qj6cv" Oct 08 19:47:40 crc kubenswrapper[4988]: I1008 19:47:40.436799 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-56f9cf6987-4865b"] Oct 08 19:47:40 crc kubenswrapper[4988]: I1008 19:47:40.437254 4988 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-56f9cf6987-4865b" podUID="d4873fea-35df-48df-856f-9d4c61f4627a" containerName="dnsmasq-dns" containerID="cri-o://c866806d68920fec97681921584bced4ca900dab3bd184769e22ed92d82c8598" gracePeriod=10 Oct 08 19:47:40 crc kubenswrapper[4988]: I1008 19:47:40.761154 4988 generic.go:334] "Generic (PLEG): container finished" podID="d4873fea-35df-48df-856f-9d4c61f4627a" containerID="c866806d68920fec97681921584bced4ca900dab3bd184769e22ed92d82c8598" exitCode=0 Oct 08 19:47:40 crc kubenswrapper[4988]: I1008 19:47:40.761256 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-56f9cf6987-4865b" event={"ID":"d4873fea-35df-48df-856f-9d4c61f4627a","Type":"ContainerDied","Data":"c866806d68920fec97681921584bced4ca900dab3bd184769e22ed92d82c8598"} Oct 08 19:47:40 crc kubenswrapper[4988]: I1008 19:47:40.763278 4988 generic.go:334] "Generic (PLEG): container finished" podID="55444eeb-4abd-4b8f-840a-2cd7e508feae" containerID="a7cc809429fada31476c348ee9a9bc12595d75bacbfad867aa41712ca1ac92a9" exitCode=0 Oct 08 19:47:40 crc kubenswrapper[4988]: I1008 19:47:40.763307 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8zqnt" event={"ID":"55444eeb-4abd-4b8f-840a-2cd7e508feae","Type":"ContainerDied","Data":"a7cc809429fada31476c348ee9a9bc12595d75bacbfad867aa41712ca1ac92a9"} Oct 08 19:47:40 crc kubenswrapper[4988]: I1008 19:47:40.945977 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-56f9cf6987-4865b" Oct 08 19:47:41 crc kubenswrapper[4988]: I1008 19:47:41.047486 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d4873fea-35df-48df-856f-9d4c61f4627a-dns-svc\") pod \"d4873fea-35df-48df-856f-9d4c61f4627a\" (UID: \"d4873fea-35df-48df-856f-9d4c61f4627a\") " Oct 08 19:47:41 crc kubenswrapper[4988]: I1008 19:47:41.047637 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m2czm\" (UniqueName: \"kubernetes.io/projected/d4873fea-35df-48df-856f-9d4c61f4627a-kube-api-access-m2czm\") pod \"d4873fea-35df-48df-856f-9d4c61f4627a\" (UID: \"d4873fea-35df-48df-856f-9d4c61f4627a\") " Oct 08 19:47:41 crc kubenswrapper[4988]: I1008 19:47:41.047664 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d4873fea-35df-48df-856f-9d4c61f4627a-ovsdbserver-sb\") pod \"d4873fea-35df-48df-856f-9d4c61f4627a\" (UID: \"d4873fea-35df-48df-856f-9d4c61f4627a\") " Oct 08 19:47:41 crc kubenswrapper[4988]: I1008 19:47:41.047692 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d4873fea-35df-48df-856f-9d4c61f4627a-ovsdbserver-nb\") pod \"d4873fea-35df-48df-856f-9d4c61f4627a\" (UID: \"d4873fea-35df-48df-856f-9d4c61f4627a\") " Oct 08 19:47:41 crc kubenswrapper[4988]: I1008 19:47:41.047735 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d4873fea-35df-48df-856f-9d4c61f4627a-config\") pod \"d4873fea-35df-48df-856f-9d4c61f4627a\" (UID: \"d4873fea-35df-48df-856f-9d4c61f4627a\") " Oct 08 19:47:41 crc kubenswrapper[4988]: I1008 19:47:41.052814 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d4873fea-35df-48df-856f-9d4c61f4627a-kube-api-access-m2czm" (OuterVolumeSpecName: "kube-api-access-m2czm") pod "d4873fea-35df-48df-856f-9d4c61f4627a" (UID: "d4873fea-35df-48df-856f-9d4c61f4627a"). InnerVolumeSpecName "kube-api-access-m2czm". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 19:47:41 crc kubenswrapper[4988]: I1008 19:47:41.097199 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d4873fea-35df-48df-856f-9d4c61f4627a-config" (OuterVolumeSpecName: "config") pod "d4873fea-35df-48df-856f-9d4c61f4627a" (UID: "d4873fea-35df-48df-856f-9d4c61f4627a"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 19:47:41 crc kubenswrapper[4988]: I1008 19:47:41.097243 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d4873fea-35df-48df-856f-9d4c61f4627a-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "d4873fea-35df-48df-856f-9d4c61f4627a" (UID: "d4873fea-35df-48df-856f-9d4c61f4627a"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 19:47:41 crc kubenswrapper[4988]: I1008 19:47:41.100860 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d4873fea-35df-48df-856f-9d4c61f4627a-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "d4873fea-35df-48df-856f-9d4c61f4627a" (UID: "d4873fea-35df-48df-856f-9d4c61f4627a"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 19:47:41 crc kubenswrapper[4988]: I1008 19:47:41.127980 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d4873fea-35df-48df-856f-9d4c61f4627a-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "d4873fea-35df-48df-856f-9d4c61f4627a" (UID: "d4873fea-35df-48df-856f-9d4c61f4627a"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 19:47:41 crc kubenswrapper[4988]: I1008 19:47:41.161262 4988 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d4873fea-35df-48df-856f-9d4c61f4627a-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 08 19:47:41 crc kubenswrapper[4988]: I1008 19:47:41.161512 4988 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d4873fea-35df-48df-856f-9d4c61f4627a-config\") on node \"crc\" DevicePath \"\"" Oct 08 19:47:41 crc kubenswrapper[4988]: I1008 19:47:41.161589 4988 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d4873fea-35df-48df-856f-9d4c61f4627a-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 08 19:47:41 crc kubenswrapper[4988]: I1008 19:47:41.161684 4988 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-m2czm\" (UniqueName: \"kubernetes.io/projected/d4873fea-35df-48df-856f-9d4c61f4627a-kube-api-access-m2czm\") on node \"crc\" DevicePath \"\"" Oct 08 19:47:41 crc kubenswrapper[4988]: I1008 19:47:41.161756 4988 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d4873fea-35df-48df-856f-9d4c61f4627a-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 08 19:47:41 crc kubenswrapper[4988]: I1008 19:47:41.779582 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8zqnt" event={"ID":"55444eeb-4abd-4b8f-840a-2cd7e508feae","Type":"ContainerStarted","Data":"aca74f33c18255fffe143078a94b5568d01b1050a6b1d04a44fed5c893c8f205"} Oct 08 19:47:41 crc kubenswrapper[4988]: I1008 19:47:41.784005 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-56f9cf6987-4865b" event={"ID":"d4873fea-35df-48df-856f-9d4c61f4627a","Type":"ContainerDied","Data":"97a87a79dce8c2afc2e8864146bb2d9a670de19e5448babdd3e688a002944922"} Oct 08 19:47:41 crc kubenswrapper[4988]: I1008 19:47:41.784057 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-56f9cf6987-4865b" Oct 08 19:47:41 crc kubenswrapper[4988]: I1008 19:47:41.784103 4988 scope.go:117] "RemoveContainer" containerID="c866806d68920fec97681921584bced4ca900dab3bd184769e22ed92d82c8598" Oct 08 19:47:41 crc kubenswrapper[4988]: I1008 19:47:41.816012 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-8zqnt" podStartSLOduration=2.195187103 podStartE2EDuration="4.815983923s" podCreationTimestamp="2025-10-08 19:47:37 +0000 UTC" firstStartedPulling="2025-10-08 19:47:38.730084528 +0000 UTC m=+5804.179927308" lastFinishedPulling="2025-10-08 19:47:41.350881348 +0000 UTC m=+5806.800724128" observedRunningTime="2025-10-08 19:47:41.811674125 +0000 UTC m=+5807.261516985" watchObservedRunningTime="2025-10-08 19:47:41.815983923 +0000 UTC m=+5807.265826733" Oct 08 19:47:41 crc kubenswrapper[4988]: I1008 19:47:41.816346 4988 scope.go:117] "RemoveContainer" containerID="4e91f30b65cc41952b68ff94905eb1417638da4d6d83f2d09bd6bd7e81c75141" Oct 08 19:47:41 crc kubenswrapper[4988]: I1008 19:47:41.839502 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-56f9cf6987-4865b"] Oct 08 19:47:41 crc kubenswrapper[4988]: I1008 19:47:41.845936 4988 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-56f9cf6987-4865b"] Oct 08 19:47:43 crc kubenswrapper[4988]: I1008 19:47:43.253614 4988 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d4873fea-35df-48df-856f-9d4c61f4627a" path="/var/lib/kubelet/pods/d4873fea-35df-48df-856f-9d4c61f4627a/volumes" Oct 08 19:47:44 crc kubenswrapper[4988]: I1008 19:47:44.092475 4988 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Oct 08 19:47:44 crc kubenswrapper[4988]: I1008 19:47:44.092551 4988 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Oct 08 19:47:44 crc kubenswrapper[4988]: I1008 19:47:44.124999 4988 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Oct 08 19:47:44 crc kubenswrapper[4988]: I1008 19:47:44.125076 4988 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Oct 08 19:47:44 crc kubenswrapper[4988]: I1008 19:47:44.153002 4988 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Oct 08 19:47:44 crc kubenswrapper[4988]: I1008 19:47:44.170517 4988 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Oct 08 19:47:44 crc kubenswrapper[4988]: I1008 19:47:44.195013 4988 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Oct 08 19:47:44 crc kubenswrapper[4988]: I1008 19:47:44.218110 4988 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Oct 08 19:47:44 crc kubenswrapper[4988]: I1008 19:47:44.245025 4988 scope.go:117] "RemoveContainer" containerID="f2a29ed22dd7a48bbc929e65fbeceed969a669e78f65fbfb724b03dc0db6e947" Oct 08 19:47:44 crc kubenswrapper[4988]: E1008 19:47:44.245740 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nm54f_openshift-machine-config-operator(f328a20b-5283-407e-8a93-2ab2da599f18)\"" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" Oct 08 19:47:44 crc kubenswrapper[4988]: I1008 19:47:44.823544 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Oct 08 19:47:44 crc kubenswrapper[4988]: I1008 19:47:44.823622 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Oct 08 19:47:44 crc kubenswrapper[4988]: I1008 19:47:44.823646 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Oct 08 19:47:44 crc kubenswrapper[4988]: I1008 19:47:44.823663 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Oct 08 19:47:46 crc kubenswrapper[4988]: I1008 19:47:46.583456 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Oct 08 19:47:46 crc kubenswrapper[4988]: I1008 19:47:46.589822 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Oct 08 19:47:46 crc kubenswrapper[4988]: I1008 19:47:46.698763 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Oct 08 19:47:46 crc kubenswrapper[4988]: I1008 19:47:46.818481 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Oct 08 19:47:47 crc kubenswrapper[4988]: I1008 19:47:47.482077 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-8zqnt" Oct 08 19:47:47 crc kubenswrapper[4988]: I1008 19:47:47.482913 4988 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-8zqnt" Oct 08 19:47:47 crc kubenswrapper[4988]: I1008 19:47:47.557759 4988 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-8zqnt" Oct 08 19:47:47 crc kubenswrapper[4988]: I1008 19:47:47.917085 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-8zqnt" Oct 08 19:47:47 crc kubenswrapper[4988]: I1008 19:47:47.984853 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-8zqnt"] Oct 08 19:47:49 crc kubenswrapper[4988]: I1008 19:47:49.881020 4988 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-8zqnt" podUID="55444eeb-4abd-4b8f-840a-2cd7e508feae" containerName="registry-server" containerID="cri-o://aca74f33c18255fffe143078a94b5568d01b1050a6b1d04a44fed5c893c8f205" gracePeriod=2 Oct 08 19:47:50 crc kubenswrapper[4988]: I1008 19:47:50.429119 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-8zqnt" Oct 08 19:47:50 crc kubenswrapper[4988]: I1008 19:47:50.548774 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/55444eeb-4abd-4b8f-840a-2cd7e508feae-utilities\") pod \"55444eeb-4abd-4b8f-840a-2cd7e508feae\" (UID: \"55444eeb-4abd-4b8f-840a-2cd7e508feae\") " Oct 08 19:47:50 crc kubenswrapper[4988]: I1008 19:47:50.548902 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8wzr9\" (UniqueName: \"kubernetes.io/projected/55444eeb-4abd-4b8f-840a-2cd7e508feae-kube-api-access-8wzr9\") pod \"55444eeb-4abd-4b8f-840a-2cd7e508feae\" (UID: \"55444eeb-4abd-4b8f-840a-2cd7e508feae\") " Oct 08 19:47:50 crc kubenswrapper[4988]: I1008 19:47:50.549183 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/55444eeb-4abd-4b8f-840a-2cd7e508feae-catalog-content\") pod \"55444eeb-4abd-4b8f-840a-2cd7e508feae\" (UID: \"55444eeb-4abd-4b8f-840a-2cd7e508feae\") " Oct 08 19:47:50 crc kubenswrapper[4988]: I1008 19:47:50.550583 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/55444eeb-4abd-4b8f-840a-2cd7e508feae-utilities" (OuterVolumeSpecName: "utilities") pod "55444eeb-4abd-4b8f-840a-2cd7e508feae" (UID: "55444eeb-4abd-4b8f-840a-2cd7e508feae"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 19:47:50 crc kubenswrapper[4988]: I1008 19:47:50.555131 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/55444eeb-4abd-4b8f-840a-2cd7e508feae-kube-api-access-8wzr9" (OuterVolumeSpecName: "kube-api-access-8wzr9") pod "55444eeb-4abd-4b8f-840a-2cd7e508feae" (UID: "55444eeb-4abd-4b8f-840a-2cd7e508feae"). InnerVolumeSpecName "kube-api-access-8wzr9". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 19:47:50 crc kubenswrapper[4988]: I1008 19:47:50.567424 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/55444eeb-4abd-4b8f-840a-2cd7e508feae-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "55444eeb-4abd-4b8f-840a-2cd7e508feae" (UID: "55444eeb-4abd-4b8f-840a-2cd7e508feae"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 19:47:50 crc kubenswrapper[4988]: I1008 19:47:50.652774 4988 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/55444eeb-4abd-4b8f-840a-2cd7e508feae-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 08 19:47:50 crc kubenswrapper[4988]: I1008 19:47:50.653213 4988 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/55444eeb-4abd-4b8f-840a-2cd7e508feae-utilities\") on node \"crc\" DevicePath \"\"" Oct 08 19:47:50 crc kubenswrapper[4988]: I1008 19:47:50.653239 4988 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8wzr9\" (UniqueName: \"kubernetes.io/projected/55444eeb-4abd-4b8f-840a-2cd7e508feae-kube-api-access-8wzr9\") on node \"crc\" DevicePath \"\"" Oct 08 19:47:50 crc kubenswrapper[4988]: I1008 19:47:50.906850 4988 generic.go:334] "Generic (PLEG): container finished" podID="55444eeb-4abd-4b8f-840a-2cd7e508feae" containerID="aca74f33c18255fffe143078a94b5568d01b1050a6b1d04a44fed5c893c8f205" exitCode=0 Oct 08 19:47:50 crc kubenswrapper[4988]: I1008 19:47:50.906908 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8zqnt" event={"ID":"55444eeb-4abd-4b8f-840a-2cd7e508feae","Type":"ContainerDied","Data":"aca74f33c18255fffe143078a94b5568d01b1050a6b1d04a44fed5c893c8f205"} Oct 08 19:47:50 crc kubenswrapper[4988]: I1008 19:47:50.906940 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8zqnt" event={"ID":"55444eeb-4abd-4b8f-840a-2cd7e508feae","Type":"ContainerDied","Data":"c81d0b80ae5383cf3950beec554447311405accab41f53616b387d8439c6c950"} Oct 08 19:47:50 crc kubenswrapper[4988]: I1008 19:47:50.906961 4988 scope.go:117] "RemoveContainer" containerID="aca74f33c18255fffe143078a94b5568d01b1050a6b1d04a44fed5c893c8f205" Oct 08 19:47:50 crc kubenswrapper[4988]: I1008 19:47:50.907144 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-8zqnt" Oct 08 19:47:50 crc kubenswrapper[4988]: I1008 19:47:50.954803 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-8zqnt"] Oct 08 19:47:50 crc kubenswrapper[4988]: I1008 19:47:50.956351 4988 scope.go:117] "RemoveContainer" containerID="a7cc809429fada31476c348ee9a9bc12595d75bacbfad867aa41712ca1ac92a9" Oct 08 19:47:50 crc kubenswrapper[4988]: I1008 19:47:50.963372 4988 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-8zqnt"] Oct 08 19:47:50 crc kubenswrapper[4988]: I1008 19:47:50.974680 4988 scope.go:117] "RemoveContainer" containerID="b2ae7687b60cc018eb56d49d99e8db2f83fd91a0473c485d24e8182b518313d2" Oct 08 19:47:51 crc kubenswrapper[4988]: I1008 19:47:51.010272 4988 scope.go:117] "RemoveContainer" containerID="aca74f33c18255fffe143078a94b5568d01b1050a6b1d04a44fed5c893c8f205" Oct 08 19:47:51 crc kubenswrapper[4988]: E1008 19:47:51.010758 4988 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"aca74f33c18255fffe143078a94b5568d01b1050a6b1d04a44fed5c893c8f205\": container with ID starting with aca74f33c18255fffe143078a94b5568d01b1050a6b1d04a44fed5c893c8f205 not found: ID does not exist" containerID="aca74f33c18255fffe143078a94b5568d01b1050a6b1d04a44fed5c893c8f205" Oct 08 19:47:51 crc kubenswrapper[4988]: I1008 19:47:51.010805 4988 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"aca74f33c18255fffe143078a94b5568d01b1050a6b1d04a44fed5c893c8f205"} err="failed to get container status \"aca74f33c18255fffe143078a94b5568d01b1050a6b1d04a44fed5c893c8f205\": rpc error: code = NotFound desc = could not find container \"aca74f33c18255fffe143078a94b5568d01b1050a6b1d04a44fed5c893c8f205\": container with ID starting with aca74f33c18255fffe143078a94b5568d01b1050a6b1d04a44fed5c893c8f205 not found: ID does not exist" Oct 08 19:47:51 crc kubenswrapper[4988]: I1008 19:47:51.010835 4988 scope.go:117] "RemoveContainer" containerID="a7cc809429fada31476c348ee9a9bc12595d75bacbfad867aa41712ca1ac92a9" Oct 08 19:47:51 crc kubenswrapper[4988]: E1008 19:47:51.011201 4988 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a7cc809429fada31476c348ee9a9bc12595d75bacbfad867aa41712ca1ac92a9\": container with ID starting with a7cc809429fada31476c348ee9a9bc12595d75bacbfad867aa41712ca1ac92a9 not found: ID does not exist" containerID="a7cc809429fada31476c348ee9a9bc12595d75bacbfad867aa41712ca1ac92a9" Oct 08 19:47:51 crc kubenswrapper[4988]: I1008 19:47:51.011224 4988 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a7cc809429fada31476c348ee9a9bc12595d75bacbfad867aa41712ca1ac92a9"} err="failed to get container status \"a7cc809429fada31476c348ee9a9bc12595d75bacbfad867aa41712ca1ac92a9\": rpc error: code = NotFound desc = could not find container \"a7cc809429fada31476c348ee9a9bc12595d75bacbfad867aa41712ca1ac92a9\": container with ID starting with a7cc809429fada31476c348ee9a9bc12595d75bacbfad867aa41712ca1ac92a9 not found: ID does not exist" Oct 08 19:47:51 crc kubenswrapper[4988]: I1008 19:47:51.011241 4988 scope.go:117] "RemoveContainer" containerID="b2ae7687b60cc018eb56d49d99e8db2f83fd91a0473c485d24e8182b518313d2" Oct 08 19:47:51 crc kubenswrapper[4988]: E1008 19:47:51.011595 4988 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b2ae7687b60cc018eb56d49d99e8db2f83fd91a0473c485d24e8182b518313d2\": container with ID starting with b2ae7687b60cc018eb56d49d99e8db2f83fd91a0473c485d24e8182b518313d2 not found: ID does not exist" containerID="b2ae7687b60cc018eb56d49d99e8db2f83fd91a0473c485d24e8182b518313d2" Oct 08 19:47:51 crc kubenswrapper[4988]: I1008 19:47:51.011627 4988 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b2ae7687b60cc018eb56d49d99e8db2f83fd91a0473c485d24e8182b518313d2"} err="failed to get container status \"b2ae7687b60cc018eb56d49d99e8db2f83fd91a0473c485d24e8182b518313d2\": rpc error: code = NotFound desc = could not find container \"b2ae7687b60cc018eb56d49d99e8db2f83fd91a0473c485d24e8182b518313d2\": container with ID starting with b2ae7687b60cc018eb56d49d99e8db2f83fd91a0473c485d24e8182b518313d2 not found: ID does not exist" Oct 08 19:47:51 crc kubenswrapper[4988]: I1008 19:47:51.254959 4988 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="55444eeb-4abd-4b8f-840a-2cd7e508feae" path="/var/lib/kubelet/pods/55444eeb-4abd-4b8f-840a-2cd7e508feae/volumes" Oct 08 19:47:53 crc kubenswrapper[4988]: I1008 19:47:53.107911 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-create-smwmq"] Oct 08 19:47:53 crc kubenswrapper[4988]: E1008 19:47:53.108733 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="55444eeb-4abd-4b8f-840a-2cd7e508feae" containerName="extract-utilities" Oct 08 19:47:53 crc kubenswrapper[4988]: I1008 19:47:53.108753 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="55444eeb-4abd-4b8f-840a-2cd7e508feae" containerName="extract-utilities" Oct 08 19:47:53 crc kubenswrapper[4988]: E1008 19:47:53.108769 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="55444eeb-4abd-4b8f-840a-2cd7e508feae" containerName="registry-server" Oct 08 19:47:53 crc kubenswrapper[4988]: I1008 19:47:53.108777 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="55444eeb-4abd-4b8f-840a-2cd7e508feae" containerName="registry-server" Oct 08 19:47:53 crc kubenswrapper[4988]: E1008 19:47:53.108809 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="55444eeb-4abd-4b8f-840a-2cd7e508feae" containerName="extract-content" Oct 08 19:47:53 crc kubenswrapper[4988]: I1008 19:47:53.108817 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="55444eeb-4abd-4b8f-840a-2cd7e508feae" containerName="extract-content" Oct 08 19:47:53 crc kubenswrapper[4988]: E1008 19:47:53.108830 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d4873fea-35df-48df-856f-9d4c61f4627a" containerName="init" Oct 08 19:47:53 crc kubenswrapper[4988]: I1008 19:47:53.108837 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="d4873fea-35df-48df-856f-9d4c61f4627a" containerName="init" Oct 08 19:47:53 crc kubenswrapper[4988]: E1008 19:47:53.108857 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d4873fea-35df-48df-856f-9d4c61f4627a" containerName="dnsmasq-dns" Oct 08 19:47:53 crc kubenswrapper[4988]: I1008 19:47:53.108864 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="d4873fea-35df-48df-856f-9d4c61f4627a" containerName="dnsmasq-dns" Oct 08 19:47:53 crc kubenswrapper[4988]: I1008 19:47:53.109080 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="55444eeb-4abd-4b8f-840a-2cd7e508feae" containerName="registry-server" Oct 08 19:47:53 crc kubenswrapper[4988]: I1008 19:47:53.109100 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="d4873fea-35df-48df-856f-9d4c61f4627a" containerName="dnsmasq-dns" Oct 08 19:47:53 crc kubenswrapper[4988]: I1008 19:47:53.109824 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-smwmq" Oct 08 19:47:53 crc kubenswrapper[4988]: I1008 19:47:53.118548 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-smwmq"] Oct 08 19:47:53 crc kubenswrapper[4988]: I1008 19:47:53.202338 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8fjzs\" (UniqueName: \"kubernetes.io/projected/5965ce16-2d47-4658-8429-eb4bcff30a21-kube-api-access-8fjzs\") pod \"placement-db-create-smwmq\" (UID: \"5965ce16-2d47-4658-8429-eb4bcff30a21\") " pod="openstack/placement-db-create-smwmq" Oct 08 19:47:53 crc kubenswrapper[4988]: I1008 19:47:53.304753 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8fjzs\" (UniqueName: \"kubernetes.io/projected/5965ce16-2d47-4658-8429-eb4bcff30a21-kube-api-access-8fjzs\") pod \"placement-db-create-smwmq\" (UID: \"5965ce16-2d47-4658-8429-eb4bcff30a21\") " pod="openstack/placement-db-create-smwmq" Oct 08 19:47:53 crc kubenswrapper[4988]: I1008 19:47:53.329440 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8fjzs\" (UniqueName: \"kubernetes.io/projected/5965ce16-2d47-4658-8429-eb4bcff30a21-kube-api-access-8fjzs\") pod \"placement-db-create-smwmq\" (UID: \"5965ce16-2d47-4658-8429-eb4bcff30a21\") " pod="openstack/placement-db-create-smwmq" Oct 08 19:47:53 crc kubenswrapper[4988]: I1008 19:47:53.436189 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-smwmq" Oct 08 19:47:53 crc kubenswrapper[4988]: I1008 19:47:53.922759 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-smwmq"] Oct 08 19:47:53 crc kubenswrapper[4988]: W1008 19:47:53.937906 4988 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5965ce16_2d47_4658_8429_eb4bcff30a21.slice/crio-f3dccabe715b5a7386b8708a5fd780b9c4ce65779f2cb5cf81ddd72a9fd0fb79 WatchSource:0}: Error finding container f3dccabe715b5a7386b8708a5fd780b9c4ce65779f2cb5cf81ddd72a9fd0fb79: Status 404 returned error can't find the container with id f3dccabe715b5a7386b8708a5fd780b9c4ce65779f2cb5cf81ddd72a9fd0fb79 Oct 08 19:47:54 crc kubenswrapper[4988]: I1008 19:47:54.953758 4988 generic.go:334] "Generic (PLEG): container finished" podID="5965ce16-2d47-4658-8429-eb4bcff30a21" containerID="2d47400cbad5a6a81cdb1dd2bc01f4bcff2cf43f37c997940b5f61fdee846285" exitCode=0 Oct 08 19:47:54 crc kubenswrapper[4988]: I1008 19:47:54.953830 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-smwmq" event={"ID":"5965ce16-2d47-4658-8429-eb4bcff30a21","Type":"ContainerDied","Data":"2d47400cbad5a6a81cdb1dd2bc01f4bcff2cf43f37c997940b5f61fdee846285"} Oct 08 19:47:54 crc kubenswrapper[4988]: I1008 19:47:54.954153 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-smwmq" event={"ID":"5965ce16-2d47-4658-8429-eb4bcff30a21","Type":"ContainerStarted","Data":"f3dccabe715b5a7386b8708a5fd780b9c4ce65779f2cb5cf81ddd72a9fd0fb79"} Oct 08 19:47:56 crc kubenswrapper[4988]: I1008 19:47:56.397210 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-smwmq" Oct 08 19:47:56 crc kubenswrapper[4988]: I1008 19:47:56.570935 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8fjzs\" (UniqueName: \"kubernetes.io/projected/5965ce16-2d47-4658-8429-eb4bcff30a21-kube-api-access-8fjzs\") pod \"5965ce16-2d47-4658-8429-eb4bcff30a21\" (UID: \"5965ce16-2d47-4658-8429-eb4bcff30a21\") " Oct 08 19:47:56 crc kubenswrapper[4988]: I1008 19:47:56.579774 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5965ce16-2d47-4658-8429-eb4bcff30a21-kube-api-access-8fjzs" (OuterVolumeSpecName: "kube-api-access-8fjzs") pod "5965ce16-2d47-4658-8429-eb4bcff30a21" (UID: "5965ce16-2d47-4658-8429-eb4bcff30a21"). InnerVolumeSpecName "kube-api-access-8fjzs". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 19:47:56 crc kubenswrapper[4988]: I1008 19:47:56.673512 4988 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8fjzs\" (UniqueName: \"kubernetes.io/projected/5965ce16-2d47-4658-8429-eb4bcff30a21-kube-api-access-8fjzs\") on node \"crc\" DevicePath \"\"" Oct 08 19:47:56 crc kubenswrapper[4988]: I1008 19:47:56.985303 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-smwmq" event={"ID":"5965ce16-2d47-4658-8429-eb4bcff30a21","Type":"ContainerDied","Data":"f3dccabe715b5a7386b8708a5fd780b9c4ce65779f2cb5cf81ddd72a9fd0fb79"} Oct 08 19:47:56 crc kubenswrapper[4988]: I1008 19:47:56.985344 4988 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f3dccabe715b5a7386b8708a5fd780b9c4ce65779f2cb5cf81ddd72a9fd0fb79" Oct 08 19:47:56 crc kubenswrapper[4988]: I1008 19:47:56.985432 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-smwmq" Oct 08 19:47:57 crc kubenswrapper[4988]: I1008 19:47:57.238337 4988 scope.go:117] "RemoveContainer" containerID="f2a29ed22dd7a48bbc929e65fbeceed969a669e78f65fbfb724b03dc0db6e947" Oct 08 19:47:57 crc kubenswrapper[4988]: E1008 19:47:57.238747 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nm54f_openshift-machine-config-operator(f328a20b-5283-407e-8a93-2ab2da599f18)\"" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" Oct 08 19:48:03 crc kubenswrapper[4988]: I1008 19:48:03.200867 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-16fa-account-create-vxrv2"] Oct 08 19:48:03 crc kubenswrapper[4988]: E1008 19:48:03.202132 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5965ce16-2d47-4658-8429-eb4bcff30a21" containerName="mariadb-database-create" Oct 08 19:48:03 crc kubenswrapper[4988]: I1008 19:48:03.202156 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="5965ce16-2d47-4658-8429-eb4bcff30a21" containerName="mariadb-database-create" Oct 08 19:48:03 crc kubenswrapper[4988]: I1008 19:48:03.202572 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="5965ce16-2d47-4658-8429-eb4bcff30a21" containerName="mariadb-database-create" Oct 08 19:48:03 crc kubenswrapper[4988]: I1008 19:48:03.203640 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-16fa-account-create-vxrv2" Oct 08 19:48:03 crc kubenswrapper[4988]: I1008 19:48:03.206852 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-db-secret" Oct 08 19:48:03 crc kubenswrapper[4988]: I1008 19:48:03.219164 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-16fa-account-create-vxrv2"] Oct 08 19:48:03 crc kubenswrapper[4988]: I1008 19:48:03.229231 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l6q2t\" (UniqueName: \"kubernetes.io/projected/5388dded-ae47-41e5-934d-7b196fb9ce79-kube-api-access-l6q2t\") pod \"placement-16fa-account-create-vxrv2\" (UID: \"5388dded-ae47-41e5-934d-7b196fb9ce79\") " pod="openstack/placement-16fa-account-create-vxrv2" Oct 08 19:48:03 crc kubenswrapper[4988]: I1008 19:48:03.331285 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l6q2t\" (UniqueName: \"kubernetes.io/projected/5388dded-ae47-41e5-934d-7b196fb9ce79-kube-api-access-l6q2t\") pod \"placement-16fa-account-create-vxrv2\" (UID: \"5388dded-ae47-41e5-934d-7b196fb9ce79\") " pod="openstack/placement-16fa-account-create-vxrv2" Oct 08 19:48:03 crc kubenswrapper[4988]: I1008 19:48:03.365939 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l6q2t\" (UniqueName: \"kubernetes.io/projected/5388dded-ae47-41e5-934d-7b196fb9ce79-kube-api-access-l6q2t\") pod \"placement-16fa-account-create-vxrv2\" (UID: \"5388dded-ae47-41e5-934d-7b196fb9ce79\") " pod="openstack/placement-16fa-account-create-vxrv2" Oct 08 19:48:03 crc kubenswrapper[4988]: I1008 19:48:03.536186 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-16fa-account-create-vxrv2" Oct 08 19:48:04 crc kubenswrapper[4988]: I1008 19:48:04.058378 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-16fa-account-create-vxrv2"] Oct 08 19:48:04 crc kubenswrapper[4988]: W1008 19:48:04.062218 4988 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5388dded_ae47_41e5_934d_7b196fb9ce79.slice/crio-8185264dea6a00f1eac83fe490d907f55cf51da577adc7a89b585d336ad0080d WatchSource:0}: Error finding container 8185264dea6a00f1eac83fe490d907f55cf51da577adc7a89b585d336ad0080d: Status 404 returned error can't find the container with id 8185264dea6a00f1eac83fe490d907f55cf51da577adc7a89b585d336ad0080d Oct 08 19:48:05 crc kubenswrapper[4988]: I1008 19:48:05.080142 4988 generic.go:334] "Generic (PLEG): container finished" podID="5388dded-ae47-41e5-934d-7b196fb9ce79" containerID="d7106d9f01fda802281545842274b315267e733b064e24d743fcae62e7dd0bde" exitCode=0 Oct 08 19:48:05 crc kubenswrapper[4988]: I1008 19:48:05.080221 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-16fa-account-create-vxrv2" event={"ID":"5388dded-ae47-41e5-934d-7b196fb9ce79","Type":"ContainerDied","Data":"d7106d9f01fda802281545842274b315267e733b064e24d743fcae62e7dd0bde"} Oct 08 19:48:05 crc kubenswrapper[4988]: I1008 19:48:05.080528 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-16fa-account-create-vxrv2" event={"ID":"5388dded-ae47-41e5-934d-7b196fb9ce79","Type":"ContainerStarted","Data":"8185264dea6a00f1eac83fe490d907f55cf51da577adc7a89b585d336ad0080d"} Oct 08 19:48:06 crc kubenswrapper[4988]: I1008 19:48:06.479854 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-16fa-account-create-vxrv2" Oct 08 19:48:06 crc kubenswrapper[4988]: I1008 19:48:06.499776 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-l6q2t\" (UniqueName: \"kubernetes.io/projected/5388dded-ae47-41e5-934d-7b196fb9ce79-kube-api-access-l6q2t\") pod \"5388dded-ae47-41e5-934d-7b196fb9ce79\" (UID: \"5388dded-ae47-41e5-934d-7b196fb9ce79\") " Oct 08 19:48:06 crc kubenswrapper[4988]: I1008 19:48:06.508118 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5388dded-ae47-41e5-934d-7b196fb9ce79-kube-api-access-l6q2t" (OuterVolumeSpecName: "kube-api-access-l6q2t") pod "5388dded-ae47-41e5-934d-7b196fb9ce79" (UID: "5388dded-ae47-41e5-934d-7b196fb9ce79"). InnerVolumeSpecName "kube-api-access-l6q2t". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 19:48:06 crc kubenswrapper[4988]: I1008 19:48:06.602710 4988 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-l6q2t\" (UniqueName: \"kubernetes.io/projected/5388dded-ae47-41e5-934d-7b196fb9ce79-kube-api-access-l6q2t\") on node \"crc\" DevicePath \"\"" Oct 08 19:48:07 crc kubenswrapper[4988]: I1008 19:48:07.109598 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-16fa-account-create-vxrv2" event={"ID":"5388dded-ae47-41e5-934d-7b196fb9ce79","Type":"ContainerDied","Data":"8185264dea6a00f1eac83fe490d907f55cf51da577adc7a89b585d336ad0080d"} Oct 08 19:48:07 crc kubenswrapper[4988]: I1008 19:48:07.109668 4988 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8185264dea6a00f1eac83fe490d907f55cf51da577adc7a89b585d336ad0080d" Oct 08 19:48:07 crc kubenswrapper[4988]: I1008 19:48:07.109750 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-16fa-account-create-vxrv2" Oct 08 19:48:08 crc kubenswrapper[4988]: I1008 19:48:08.519070 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-sync-hsz6g"] Oct 08 19:48:08 crc kubenswrapper[4988]: E1008 19:48:08.519778 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5388dded-ae47-41e5-934d-7b196fb9ce79" containerName="mariadb-account-create" Oct 08 19:48:08 crc kubenswrapper[4988]: I1008 19:48:08.519808 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="5388dded-ae47-41e5-934d-7b196fb9ce79" containerName="mariadb-account-create" Oct 08 19:48:08 crc kubenswrapper[4988]: I1008 19:48:08.520073 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="5388dded-ae47-41e5-934d-7b196fb9ce79" containerName="mariadb-account-create" Oct 08 19:48:08 crc kubenswrapper[4988]: I1008 19:48:08.520822 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-hsz6g" Oct 08 19:48:08 crc kubenswrapper[4988]: I1008 19:48:08.525430 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-pgrsr" Oct 08 19:48:08 crc kubenswrapper[4988]: I1008 19:48:08.525558 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Oct 08 19:48:08 crc kubenswrapper[4988]: I1008 19:48:08.525698 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Oct 08 19:48:08 crc kubenswrapper[4988]: I1008 19:48:08.525743 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-hsz6g"] Oct 08 19:48:08 crc kubenswrapper[4988]: I1008 19:48:08.536996 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-7595cd99bc-sq75r"] Oct 08 19:48:08 crc kubenswrapper[4988]: I1008 19:48:08.538569 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7595cd99bc-sq75r" Oct 08 19:48:08 crc kubenswrapper[4988]: I1008 19:48:08.572953 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7595cd99bc-sq75r"] Oct 08 19:48:08 crc kubenswrapper[4988]: I1008 19:48:08.644372 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cc2cdc3d-5598-4c70-b47a-790abb8ecd2d-logs\") pod \"placement-db-sync-hsz6g\" (UID: \"cc2cdc3d-5598-4c70-b47a-790abb8ecd2d\") " pod="openstack/placement-db-sync-hsz6g" Oct 08 19:48:08 crc kubenswrapper[4988]: I1008 19:48:08.644452 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/98ad3c06-9746-432f-a4e8-8540a58404cb-dns-svc\") pod \"dnsmasq-dns-7595cd99bc-sq75r\" (UID: \"98ad3c06-9746-432f-a4e8-8540a58404cb\") " pod="openstack/dnsmasq-dns-7595cd99bc-sq75r" Oct 08 19:48:08 crc kubenswrapper[4988]: I1008 19:48:08.644486 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/98ad3c06-9746-432f-a4e8-8540a58404cb-config\") pod \"dnsmasq-dns-7595cd99bc-sq75r\" (UID: \"98ad3c06-9746-432f-a4e8-8540a58404cb\") " pod="openstack/dnsmasq-dns-7595cd99bc-sq75r" Oct 08 19:48:08 crc kubenswrapper[4988]: I1008 19:48:08.644508 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cc2cdc3d-5598-4c70-b47a-790abb8ecd2d-combined-ca-bundle\") pod \"placement-db-sync-hsz6g\" (UID: \"cc2cdc3d-5598-4c70-b47a-790abb8ecd2d\") " pod="openstack/placement-db-sync-hsz6g" Oct 08 19:48:08 crc kubenswrapper[4988]: I1008 19:48:08.644529 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xkmrh\" (UniqueName: \"kubernetes.io/projected/cc2cdc3d-5598-4c70-b47a-790abb8ecd2d-kube-api-access-xkmrh\") pod \"placement-db-sync-hsz6g\" (UID: \"cc2cdc3d-5598-4c70-b47a-790abb8ecd2d\") " pod="openstack/placement-db-sync-hsz6g" Oct 08 19:48:08 crc kubenswrapper[4988]: I1008 19:48:08.644618 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/98ad3c06-9746-432f-a4e8-8540a58404cb-ovsdbserver-sb\") pod \"dnsmasq-dns-7595cd99bc-sq75r\" (UID: \"98ad3c06-9746-432f-a4e8-8540a58404cb\") " pod="openstack/dnsmasq-dns-7595cd99bc-sq75r" Oct 08 19:48:08 crc kubenswrapper[4988]: I1008 19:48:08.644673 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cc2cdc3d-5598-4c70-b47a-790abb8ecd2d-config-data\") pod \"placement-db-sync-hsz6g\" (UID: \"cc2cdc3d-5598-4c70-b47a-790abb8ecd2d\") " pod="openstack/placement-db-sync-hsz6g" Oct 08 19:48:08 crc kubenswrapper[4988]: I1008 19:48:08.644770 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cc2cdc3d-5598-4c70-b47a-790abb8ecd2d-scripts\") pod \"placement-db-sync-hsz6g\" (UID: \"cc2cdc3d-5598-4c70-b47a-790abb8ecd2d\") " pod="openstack/placement-db-sync-hsz6g" Oct 08 19:48:08 crc kubenswrapper[4988]: I1008 19:48:08.644886 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-85v9g\" (UniqueName: \"kubernetes.io/projected/98ad3c06-9746-432f-a4e8-8540a58404cb-kube-api-access-85v9g\") pod \"dnsmasq-dns-7595cd99bc-sq75r\" (UID: \"98ad3c06-9746-432f-a4e8-8540a58404cb\") " pod="openstack/dnsmasq-dns-7595cd99bc-sq75r" Oct 08 19:48:08 crc kubenswrapper[4988]: I1008 19:48:08.644982 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/98ad3c06-9746-432f-a4e8-8540a58404cb-ovsdbserver-nb\") pod \"dnsmasq-dns-7595cd99bc-sq75r\" (UID: \"98ad3c06-9746-432f-a4e8-8540a58404cb\") " pod="openstack/dnsmasq-dns-7595cd99bc-sq75r" Oct 08 19:48:08 crc kubenswrapper[4988]: I1008 19:48:08.746870 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cc2cdc3d-5598-4c70-b47a-790abb8ecd2d-config-data\") pod \"placement-db-sync-hsz6g\" (UID: \"cc2cdc3d-5598-4c70-b47a-790abb8ecd2d\") " pod="openstack/placement-db-sync-hsz6g" Oct 08 19:48:08 crc kubenswrapper[4988]: I1008 19:48:08.746925 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cc2cdc3d-5598-4c70-b47a-790abb8ecd2d-scripts\") pod \"placement-db-sync-hsz6g\" (UID: \"cc2cdc3d-5598-4c70-b47a-790abb8ecd2d\") " pod="openstack/placement-db-sync-hsz6g" Oct 08 19:48:08 crc kubenswrapper[4988]: I1008 19:48:08.746961 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-85v9g\" (UniqueName: \"kubernetes.io/projected/98ad3c06-9746-432f-a4e8-8540a58404cb-kube-api-access-85v9g\") pod \"dnsmasq-dns-7595cd99bc-sq75r\" (UID: \"98ad3c06-9746-432f-a4e8-8540a58404cb\") " pod="openstack/dnsmasq-dns-7595cd99bc-sq75r" Oct 08 19:48:08 crc kubenswrapper[4988]: I1008 19:48:08.746997 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/98ad3c06-9746-432f-a4e8-8540a58404cb-ovsdbserver-nb\") pod \"dnsmasq-dns-7595cd99bc-sq75r\" (UID: \"98ad3c06-9746-432f-a4e8-8540a58404cb\") " pod="openstack/dnsmasq-dns-7595cd99bc-sq75r" Oct 08 19:48:08 crc kubenswrapper[4988]: I1008 19:48:08.747015 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cc2cdc3d-5598-4c70-b47a-790abb8ecd2d-logs\") pod \"placement-db-sync-hsz6g\" (UID: \"cc2cdc3d-5598-4c70-b47a-790abb8ecd2d\") " pod="openstack/placement-db-sync-hsz6g" Oct 08 19:48:08 crc kubenswrapper[4988]: I1008 19:48:08.747040 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/98ad3c06-9746-432f-a4e8-8540a58404cb-dns-svc\") pod \"dnsmasq-dns-7595cd99bc-sq75r\" (UID: \"98ad3c06-9746-432f-a4e8-8540a58404cb\") " pod="openstack/dnsmasq-dns-7595cd99bc-sq75r" Oct 08 19:48:08 crc kubenswrapper[4988]: I1008 19:48:08.747070 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/98ad3c06-9746-432f-a4e8-8540a58404cb-config\") pod \"dnsmasq-dns-7595cd99bc-sq75r\" (UID: \"98ad3c06-9746-432f-a4e8-8540a58404cb\") " pod="openstack/dnsmasq-dns-7595cd99bc-sq75r" Oct 08 19:48:08 crc kubenswrapper[4988]: I1008 19:48:08.747091 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cc2cdc3d-5598-4c70-b47a-790abb8ecd2d-combined-ca-bundle\") pod \"placement-db-sync-hsz6g\" (UID: \"cc2cdc3d-5598-4c70-b47a-790abb8ecd2d\") " pod="openstack/placement-db-sync-hsz6g" Oct 08 19:48:08 crc kubenswrapper[4988]: I1008 19:48:08.747110 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xkmrh\" (UniqueName: \"kubernetes.io/projected/cc2cdc3d-5598-4c70-b47a-790abb8ecd2d-kube-api-access-xkmrh\") pod \"placement-db-sync-hsz6g\" (UID: \"cc2cdc3d-5598-4c70-b47a-790abb8ecd2d\") " pod="openstack/placement-db-sync-hsz6g" Oct 08 19:48:08 crc kubenswrapper[4988]: I1008 19:48:08.747164 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/98ad3c06-9746-432f-a4e8-8540a58404cb-ovsdbserver-sb\") pod \"dnsmasq-dns-7595cd99bc-sq75r\" (UID: \"98ad3c06-9746-432f-a4e8-8540a58404cb\") " pod="openstack/dnsmasq-dns-7595cd99bc-sq75r" Oct 08 19:48:08 crc kubenswrapper[4988]: I1008 19:48:08.748009 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/98ad3c06-9746-432f-a4e8-8540a58404cb-config\") pod \"dnsmasq-dns-7595cd99bc-sq75r\" (UID: \"98ad3c06-9746-432f-a4e8-8540a58404cb\") " pod="openstack/dnsmasq-dns-7595cd99bc-sq75r" Oct 08 19:48:08 crc kubenswrapper[4988]: I1008 19:48:08.748053 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cc2cdc3d-5598-4c70-b47a-790abb8ecd2d-logs\") pod \"placement-db-sync-hsz6g\" (UID: \"cc2cdc3d-5598-4c70-b47a-790abb8ecd2d\") " pod="openstack/placement-db-sync-hsz6g" Oct 08 19:48:08 crc kubenswrapper[4988]: I1008 19:48:08.748151 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/98ad3c06-9746-432f-a4e8-8540a58404cb-dns-svc\") pod \"dnsmasq-dns-7595cd99bc-sq75r\" (UID: \"98ad3c06-9746-432f-a4e8-8540a58404cb\") " pod="openstack/dnsmasq-dns-7595cd99bc-sq75r" Oct 08 19:48:08 crc kubenswrapper[4988]: I1008 19:48:08.748294 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/98ad3c06-9746-432f-a4e8-8540a58404cb-ovsdbserver-nb\") pod \"dnsmasq-dns-7595cd99bc-sq75r\" (UID: \"98ad3c06-9746-432f-a4e8-8540a58404cb\") " pod="openstack/dnsmasq-dns-7595cd99bc-sq75r" Oct 08 19:48:08 crc kubenswrapper[4988]: I1008 19:48:08.748521 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/98ad3c06-9746-432f-a4e8-8540a58404cb-ovsdbserver-sb\") pod \"dnsmasq-dns-7595cd99bc-sq75r\" (UID: \"98ad3c06-9746-432f-a4e8-8540a58404cb\") " pod="openstack/dnsmasq-dns-7595cd99bc-sq75r" Oct 08 19:48:08 crc kubenswrapper[4988]: I1008 19:48:08.752753 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cc2cdc3d-5598-4c70-b47a-790abb8ecd2d-config-data\") pod \"placement-db-sync-hsz6g\" (UID: \"cc2cdc3d-5598-4c70-b47a-790abb8ecd2d\") " pod="openstack/placement-db-sync-hsz6g" Oct 08 19:48:08 crc kubenswrapper[4988]: I1008 19:48:08.753972 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cc2cdc3d-5598-4c70-b47a-790abb8ecd2d-combined-ca-bundle\") pod \"placement-db-sync-hsz6g\" (UID: \"cc2cdc3d-5598-4c70-b47a-790abb8ecd2d\") " pod="openstack/placement-db-sync-hsz6g" Oct 08 19:48:08 crc kubenswrapper[4988]: I1008 19:48:08.762691 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cc2cdc3d-5598-4c70-b47a-790abb8ecd2d-scripts\") pod \"placement-db-sync-hsz6g\" (UID: \"cc2cdc3d-5598-4c70-b47a-790abb8ecd2d\") " pod="openstack/placement-db-sync-hsz6g" Oct 08 19:48:08 crc kubenswrapper[4988]: I1008 19:48:08.766573 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-85v9g\" (UniqueName: \"kubernetes.io/projected/98ad3c06-9746-432f-a4e8-8540a58404cb-kube-api-access-85v9g\") pod \"dnsmasq-dns-7595cd99bc-sq75r\" (UID: \"98ad3c06-9746-432f-a4e8-8540a58404cb\") " pod="openstack/dnsmasq-dns-7595cd99bc-sq75r" Oct 08 19:48:08 crc kubenswrapper[4988]: I1008 19:48:08.766761 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xkmrh\" (UniqueName: \"kubernetes.io/projected/cc2cdc3d-5598-4c70-b47a-790abb8ecd2d-kube-api-access-xkmrh\") pod \"placement-db-sync-hsz6g\" (UID: \"cc2cdc3d-5598-4c70-b47a-790abb8ecd2d\") " pod="openstack/placement-db-sync-hsz6g" Oct 08 19:48:08 crc kubenswrapper[4988]: I1008 19:48:08.851047 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-hsz6g" Oct 08 19:48:08 crc kubenswrapper[4988]: I1008 19:48:08.859954 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7595cd99bc-sq75r" Oct 08 19:48:09 crc kubenswrapper[4988]: I1008 19:48:09.355558 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-hsz6g"] Oct 08 19:48:09 crc kubenswrapper[4988]: I1008 19:48:09.422029 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7595cd99bc-sq75r"] Oct 08 19:48:09 crc kubenswrapper[4988]: W1008 19:48:09.427828 4988 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod98ad3c06_9746_432f_a4e8_8540a58404cb.slice/crio-775e237db11993d7f40d25781b28a945efb9ee387d3c88f7ff69bccbcfd039f4 WatchSource:0}: Error finding container 775e237db11993d7f40d25781b28a945efb9ee387d3c88f7ff69bccbcfd039f4: Status 404 returned error can't find the container with id 775e237db11993d7f40d25781b28a945efb9ee387d3c88f7ff69bccbcfd039f4 Oct 08 19:48:10 crc kubenswrapper[4988]: I1008 19:48:10.158309 4988 generic.go:334] "Generic (PLEG): container finished" podID="98ad3c06-9746-432f-a4e8-8540a58404cb" containerID="42d1e21314688dabe77bbf2187b2f2ed13a25ff0f0a692a5589c46386a06fa7b" exitCode=0 Oct 08 19:48:10 crc kubenswrapper[4988]: I1008 19:48:10.158568 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7595cd99bc-sq75r" event={"ID":"98ad3c06-9746-432f-a4e8-8540a58404cb","Type":"ContainerDied","Data":"42d1e21314688dabe77bbf2187b2f2ed13a25ff0f0a692a5589c46386a06fa7b"} Oct 08 19:48:10 crc kubenswrapper[4988]: I1008 19:48:10.158964 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7595cd99bc-sq75r" event={"ID":"98ad3c06-9746-432f-a4e8-8540a58404cb","Type":"ContainerStarted","Data":"775e237db11993d7f40d25781b28a945efb9ee387d3c88f7ff69bccbcfd039f4"} Oct 08 19:48:10 crc kubenswrapper[4988]: I1008 19:48:10.165737 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-hsz6g" event={"ID":"cc2cdc3d-5598-4c70-b47a-790abb8ecd2d","Type":"ContainerStarted","Data":"0c0821c37b7f2e7b751b73b3e7164c71c0e4721c7cc79b3d9bf529bc9019a275"} Oct 08 19:48:10 crc kubenswrapper[4988]: I1008 19:48:10.165801 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-hsz6g" event={"ID":"cc2cdc3d-5598-4c70-b47a-790abb8ecd2d","Type":"ContainerStarted","Data":"43cf111f473291c1b3224132d206b0dba39a9e06f9adf6872efdaab704a3a0b3"} Oct 08 19:48:10 crc kubenswrapper[4988]: I1008 19:48:10.238781 4988 scope.go:117] "RemoveContainer" containerID="f2a29ed22dd7a48bbc929e65fbeceed969a669e78f65fbfb724b03dc0db6e947" Oct 08 19:48:10 crc kubenswrapper[4988]: E1008 19:48:10.239110 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nm54f_openshift-machine-config-operator(f328a20b-5283-407e-8a93-2ab2da599f18)\"" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" Oct 08 19:48:11 crc kubenswrapper[4988]: I1008 19:48:11.175116 4988 generic.go:334] "Generic (PLEG): container finished" podID="cc2cdc3d-5598-4c70-b47a-790abb8ecd2d" containerID="0c0821c37b7f2e7b751b73b3e7164c71c0e4721c7cc79b3d9bf529bc9019a275" exitCode=0 Oct 08 19:48:11 crc kubenswrapper[4988]: I1008 19:48:11.175195 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-hsz6g" event={"ID":"cc2cdc3d-5598-4c70-b47a-790abb8ecd2d","Type":"ContainerDied","Data":"0c0821c37b7f2e7b751b73b3e7164c71c0e4721c7cc79b3d9bf529bc9019a275"} Oct 08 19:48:11 crc kubenswrapper[4988]: I1008 19:48:11.179775 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7595cd99bc-sq75r" event={"ID":"98ad3c06-9746-432f-a4e8-8540a58404cb","Type":"ContainerStarted","Data":"f8e480b1825776a01bf18e448caa9ac4c2322e267eff0611a05881ef60ff97d4"} Oct 08 19:48:11 crc kubenswrapper[4988]: I1008 19:48:11.179965 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-7595cd99bc-sq75r" Oct 08 19:48:11 crc kubenswrapper[4988]: I1008 19:48:11.223167 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-7595cd99bc-sq75r" podStartSLOduration=3.2231461120000002 podStartE2EDuration="3.223146112s" podCreationTimestamp="2025-10-08 19:48:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 19:48:11.213201085 +0000 UTC m=+5836.663043855" watchObservedRunningTime="2025-10-08 19:48:11.223146112 +0000 UTC m=+5836.672988902" Oct 08 19:48:12 crc kubenswrapper[4988]: I1008 19:48:12.606285 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-hsz6g" Oct 08 19:48:12 crc kubenswrapper[4988]: I1008 19:48:12.643146 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cc2cdc3d-5598-4c70-b47a-790abb8ecd2d-config-data\") pod \"cc2cdc3d-5598-4c70-b47a-790abb8ecd2d\" (UID: \"cc2cdc3d-5598-4c70-b47a-790abb8ecd2d\") " Oct 08 19:48:12 crc kubenswrapper[4988]: I1008 19:48:12.643288 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xkmrh\" (UniqueName: \"kubernetes.io/projected/cc2cdc3d-5598-4c70-b47a-790abb8ecd2d-kube-api-access-xkmrh\") pod \"cc2cdc3d-5598-4c70-b47a-790abb8ecd2d\" (UID: \"cc2cdc3d-5598-4c70-b47a-790abb8ecd2d\") " Oct 08 19:48:12 crc kubenswrapper[4988]: I1008 19:48:12.643367 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cc2cdc3d-5598-4c70-b47a-790abb8ecd2d-combined-ca-bundle\") pod \"cc2cdc3d-5598-4c70-b47a-790abb8ecd2d\" (UID: \"cc2cdc3d-5598-4c70-b47a-790abb8ecd2d\") " Oct 08 19:48:12 crc kubenswrapper[4988]: I1008 19:48:12.643414 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cc2cdc3d-5598-4c70-b47a-790abb8ecd2d-scripts\") pod \"cc2cdc3d-5598-4c70-b47a-790abb8ecd2d\" (UID: \"cc2cdc3d-5598-4c70-b47a-790abb8ecd2d\") " Oct 08 19:48:12 crc kubenswrapper[4988]: I1008 19:48:12.643455 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cc2cdc3d-5598-4c70-b47a-790abb8ecd2d-logs\") pod \"cc2cdc3d-5598-4c70-b47a-790abb8ecd2d\" (UID: \"cc2cdc3d-5598-4c70-b47a-790abb8ecd2d\") " Oct 08 19:48:12 crc kubenswrapper[4988]: I1008 19:48:12.644486 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cc2cdc3d-5598-4c70-b47a-790abb8ecd2d-logs" (OuterVolumeSpecName: "logs") pod "cc2cdc3d-5598-4c70-b47a-790abb8ecd2d" (UID: "cc2cdc3d-5598-4c70-b47a-790abb8ecd2d"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 19:48:12 crc kubenswrapper[4988]: I1008 19:48:12.648710 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cc2cdc3d-5598-4c70-b47a-790abb8ecd2d-scripts" (OuterVolumeSpecName: "scripts") pod "cc2cdc3d-5598-4c70-b47a-790abb8ecd2d" (UID: "cc2cdc3d-5598-4c70-b47a-790abb8ecd2d"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 19:48:12 crc kubenswrapper[4988]: I1008 19:48:12.652291 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cc2cdc3d-5598-4c70-b47a-790abb8ecd2d-kube-api-access-xkmrh" (OuterVolumeSpecName: "kube-api-access-xkmrh") pod "cc2cdc3d-5598-4c70-b47a-790abb8ecd2d" (UID: "cc2cdc3d-5598-4c70-b47a-790abb8ecd2d"). InnerVolumeSpecName "kube-api-access-xkmrh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 19:48:12 crc kubenswrapper[4988]: I1008 19:48:12.675887 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cc2cdc3d-5598-4c70-b47a-790abb8ecd2d-config-data" (OuterVolumeSpecName: "config-data") pod "cc2cdc3d-5598-4c70-b47a-790abb8ecd2d" (UID: "cc2cdc3d-5598-4c70-b47a-790abb8ecd2d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 19:48:12 crc kubenswrapper[4988]: I1008 19:48:12.695688 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cc2cdc3d-5598-4c70-b47a-790abb8ecd2d-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "cc2cdc3d-5598-4c70-b47a-790abb8ecd2d" (UID: "cc2cdc3d-5598-4c70-b47a-790abb8ecd2d"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 19:48:12 crc kubenswrapper[4988]: I1008 19:48:12.746200 4988 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cc2cdc3d-5598-4c70-b47a-790abb8ecd2d-config-data\") on node \"crc\" DevicePath \"\"" Oct 08 19:48:12 crc kubenswrapper[4988]: I1008 19:48:12.746493 4988 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xkmrh\" (UniqueName: \"kubernetes.io/projected/cc2cdc3d-5598-4c70-b47a-790abb8ecd2d-kube-api-access-xkmrh\") on node \"crc\" DevicePath \"\"" Oct 08 19:48:12 crc kubenswrapper[4988]: I1008 19:48:12.746657 4988 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cc2cdc3d-5598-4c70-b47a-790abb8ecd2d-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 19:48:12 crc kubenswrapper[4988]: I1008 19:48:12.746781 4988 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cc2cdc3d-5598-4c70-b47a-790abb8ecd2d-scripts\") on node \"crc\" DevicePath \"\"" Oct 08 19:48:12 crc kubenswrapper[4988]: I1008 19:48:12.746897 4988 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cc2cdc3d-5598-4c70-b47a-790abb8ecd2d-logs\") on node \"crc\" DevicePath \"\"" Oct 08 19:48:13 crc kubenswrapper[4988]: I1008 19:48:13.200034 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-hsz6g" event={"ID":"cc2cdc3d-5598-4c70-b47a-790abb8ecd2d","Type":"ContainerDied","Data":"43cf111f473291c1b3224132d206b0dba39a9e06f9adf6872efdaab704a3a0b3"} Oct 08 19:48:13 crc kubenswrapper[4988]: I1008 19:48:13.200092 4988 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="43cf111f473291c1b3224132d206b0dba39a9e06f9adf6872efdaab704a3a0b3" Oct 08 19:48:13 crc kubenswrapper[4988]: I1008 19:48:13.200092 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-hsz6g" Oct 08 19:48:13 crc kubenswrapper[4988]: I1008 19:48:13.739110 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-766746c8cd-vsm49"] Oct 08 19:48:13 crc kubenswrapper[4988]: E1008 19:48:13.740576 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cc2cdc3d-5598-4c70-b47a-790abb8ecd2d" containerName="placement-db-sync" Oct 08 19:48:13 crc kubenswrapper[4988]: I1008 19:48:13.740700 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="cc2cdc3d-5598-4c70-b47a-790abb8ecd2d" containerName="placement-db-sync" Oct 08 19:48:13 crc kubenswrapper[4988]: I1008 19:48:13.741014 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="cc2cdc3d-5598-4c70-b47a-790abb8ecd2d" containerName="placement-db-sync" Oct 08 19:48:13 crc kubenswrapper[4988]: I1008 19:48:13.742244 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-766746c8cd-vsm49" Oct 08 19:48:13 crc kubenswrapper[4988]: I1008 19:48:13.746011 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Oct 08 19:48:13 crc kubenswrapper[4988]: I1008 19:48:13.746199 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Oct 08 19:48:13 crc kubenswrapper[4988]: I1008 19:48:13.746267 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-pgrsr" Oct 08 19:48:13 crc kubenswrapper[4988]: I1008 19:48:13.748003 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-placement-internal-svc" Oct 08 19:48:13 crc kubenswrapper[4988]: I1008 19:48:13.748189 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-placement-public-svc" Oct 08 19:48:13 crc kubenswrapper[4988]: I1008 19:48:13.764671 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-766746c8cd-vsm49"] Oct 08 19:48:13 crc kubenswrapper[4988]: I1008 19:48:13.867604 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/dbb701b2-bfa0-4a83-b0db-91bd2d047e68-internal-tls-certs\") pod \"placement-766746c8cd-vsm49\" (UID: \"dbb701b2-bfa0-4a83-b0db-91bd2d047e68\") " pod="openstack/placement-766746c8cd-vsm49" Oct 08 19:48:13 crc kubenswrapper[4988]: I1008 19:48:13.867923 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/dbb701b2-bfa0-4a83-b0db-91bd2d047e68-logs\") pod \"placement-766746c8cd-vsm49\" (UID: \"dbb701b2-bfa0-4a83-b0db-91bd2d047e68\") " pod="openstack/placement-766746c8cd-vsm49" Oct 08 19:48:13 crc kubenswrapper[4988]: I1008 19:48:13.868161 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9sm6v\" (UniqueName: \"kubernetes.io/projected/dbb701b2-bfa0-4a83-b0db-91bd2d047e68-kube-api-access-9sm6v\") pod \"placement-766746c8cd-vsm49\" (UID: \"dbb701b2-bfa0-4a83-b0db-91bd2d047e68\") " pod="openstack/placement-766746c8cd-vsm49" Oct 08 19:48:13 crc kubenswrapper[4988]: I1008 19:48:13.868288 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dbb701b2-bfa0-4a83-b0db-91bd2d047e68-combined-ca-bundle\") pod \"placement-766746c8cd-vsm49\" (UID: \"dbb701b2-bfa0-4a83-b0db-91bd2d047e68\") " pod="openstack/placement-766746c8cd-vsm49" Oct 08 19:48:13 crc kubenswrapper[4988]: I1008 19:48:13.868529 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/dbb701b2-bfa0-4a83-b0db-91bd2d047e68-public-tls-certs\") pod \"placement-766746c8cd-vsm49\" (UID: \"dbb701b2-bfa0-4a83-b0db-91bd2d047e68\") " pod="openstack/placement-766746c8cd-vsm49" Oct 08 19:48:13 crc kubenswrapper[4988]: I1008 19:48:13.868636 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/dbb701b2-bfa0-4a83-b0db-91bd2d047e68-scripts\") pod \"placement-766746c8cd-vsm49\" (UID: \"dbb701b2-bfa0-4a83-b0db-91bd2d047e68\") " pod="openstack/placement-766746c8cd-vsm49" Oct 08 19:48:13 crc kubenswrapper[4988]: I1008 19:48:13.868905 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dbb701b2-bfa0-4a83-b0db-91bd2d047e68-config-data\") pod \"placement-766746c8cd-vsm49\" (UID: \"dbb701b2-bfa0-4a83-b0db-91bd2d047e68\") " pod="openstack/placement-766746c8cd-vsm49" Oct 08 19:48:13 crc kubenswrapper[4988]: I1008 19:48:13.971044 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dbb701b2-bfa0-4a83-b0db-91bd2d047e68-config-data\") pod \"placement-766746c8cd-vsm49\" (UID: \"dbb701b2-bfa0-4a83-b0db-91bd2d047e68\") " pod="openstack/placement-766746c8cd-vsm49" Oct 08 19:48:13 crc kubenswrapper[4988]: I1008 19:48:13.971320 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/dbb701b2-bfa0-4a83-b0db-91bd2d047e68-internal-tls-certs\") pod \"placement-766746c8cd-vsm49\" (UID: \"dbb701b2-bfa0-4a83-b0db-91bd2d047e68\") " pod="openstack/placement-766746c8cd-vsm49" Oct 08 19:48:13 crc kubenswrapper[4988]: I1008 19:48:13.971352 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/dbb701b2-bfa0-4a83-b0db-91bd2d047e68-logs\") pod \"placement-766746c8cd-vsm49\" (UID: \"dbb701b2-bfa0-4a83-b0db-91bd2d047e68\") " pod="openstack/placement-766746c8cd-vsm49" Oct 08 19:48:13 crc kubenswrapper[4988]: I1008 19:48:13.971378 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9sm6v\" (UniqueName: \"kubernetes.io/projected/dbb701b2-bfa0-4a83-b0db-91bd2d047e68-kube-api-access-9sm6v\") pod \"placement-766746c8cd-vsm49\" (UID: \"dbb701b2-bfa0-4a83-b0db-91bd2d047e68\") " pod="openstack/placement-766746c8cd-vsm49" Oct 08 19:48:13 crc kubenswrapper[4988]: I1008 19:48:13.971424 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dbb701b2-bfa0-4a83-b0db-91bd2d047e68-combined-ca-bundle\") pod \"placement-766746c8cd-vsm49\" (UID: \"dbb701b2-bfa0-4a83-b0db-91bd2d047e68\") " pod="openstack/placement-766746c8cd-vsm49" Oct 08 19:48:13 crc kubenswrapper[4988]: I1008 19:48:13.971454 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/dbb701b2-bfa0-4a83-b0db-91bd2d047e68-public-tls-certs\") pod \"placement-766746c8cd-vsm49\" (UID: \"dbb701b2-bfa0-4a83-b0db-91bd2d047e68\") " pod="openstack/placement-766746c8cd-vsm49" Oct 08 19:48:13 crc kubenswrapper[4988]: I1008 19:48:13.971478 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/dbb701b2-bfa0-4a83-b0db-91bd2d047e68-scripts\") pod \"placement-766746c8cd-vsm49\" (UID: \"dbb701b2-bfa0-4a83-b0db-91bd2d047e68\") " pod="openstack/placement-766746c8cd-vsm49" Oct 08 19:48:13 crc kubenswrapper[4988]: I1008 19:48:13.972315 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/dbb701b2-bfa0-4a83-b0db-91bd2d047e68-logs\") pod \"placement-766746c8cd-vsm49\" (UID: \"dbb701b2-bfa0-4a83-b0db-91bd2d047e68\") " pod="openstack/placement-766746c8cd-vsm49" Oct 08 19:48:13 crc kubenswrapper[4988]: I1008 19:48:13.975952 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dbb701b2-bfa0-4a83-b0db-91bd2d047e68-combined-ca-bundle\") pod \"placement-766746c8cd-vsm49\" (UID: \"dbb701b2-bfa0-4a83-b0db-91bd2d047e68\") " pod="openstack/placement-766746c8cd-vsm49" Oct 08 19:48:13 crc kubenswrapper[4988]: I1008 19:48:13.976616 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/dbb701b2-bfa0-4a83-b0db-91bd2d047e68-scripts\") pod \"placement-766746c8cd-vsm49\" (UID: \"dbb701b2-bfa0-4a83-b0db-91bd2d047e68\") " pod="openstack/placement-766746c8cd-vsm49" Oct 08 19:48:13 crc kubenswrapper[4988]: I1008 19:48:13.976788 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/dbb701b2-bfa0-4a83-b0db-91bd2d047e68-public-tls-certs\") pod \"placement-766746c8cd-vsm49\" (UID: \"dbb701b2-bfa0-4a83-b0db-91bd2d047e68\") " pod="openstack/placement-766746c8cd-vsm49" Oct 08 19:48:13 crc kubenswrapper[4988]: I1008 19:48:13.977251 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/dbb701b2-bfa0-4a83-b0db-91bd2d047e68-internal-tls-certs\") pod \"placement-766746c8cd-vsm49\" (UID: \"dbb701b2-bfa0-4a83-b0db-91bd2d047e68\") " pod="openstack/placement-766746c8cd-vsm49" Oct 08 19:48:13 crc kubenswrapper[4988]: I1008 19:48:13.977682 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dbb701b2-bfa0-4a83-b0db-91bd2d047e68-config-data\") pod \"placement-766746c8cd-vsm49\" (UID: \"dbb701b2-bfa0-4a83-b0db-91bd2d047e68\") " pod="openstack/placement-766746c8cd-vsm49" Oct 08 19:48:13 crc kubenswrapper[4988]: I1008 19:48:13.997712 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9sm6v\" (UniqueName: \"kubernetes.io/projected/dbb701b2-bfa0-4a83-b0db-91bd2d047e68-kube-api-access-9sm6v\") pod \"placement-766746c8cd-vsm49\" (UID: \"dbb701b2-bfa0-4a83-b0db-91bd2d047e68\") " pod="openstack/placement-766746c8cd-vsm49" Oct 08 19:48:14 crc kubenswrapper[4988]: I1008 19:48:14.067830 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-766746c8cd-vsm49" Oct 08 19:48:14 crc kubenswrapper[4988]: I1008 19:48:14.704597 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-766746c8cd-vsm49"] Oct 08 19:48:15 crc kubenswrapper[4988]: I1008 19:48:15.222768 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-766746c8cd-vsm49" event={"ID":"dbb701b2-bfa0-4a83-b0db-91bd2d047e68","Type":"ContainerStarted","Data":"caa0b0670bb9478b63551c9dcb4921f3ee1c18b1fe947f5698e8019eae926009"} Oct 08 19:48:15 crc kubenswrapper[4988]: I1008 19:48:15.223064 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-766746c8cd-vsm49" event={"ID":"dbb701b2-bfa0-4a83-b0db-91bd2d047e68","Type":"ContainerStarted","Data":"56e4fa5ab1a12e6f41614b3a2dcd9b4abbfd4418802c15cf9dbb222ef3cbb872"} Oct 08 19:48:16 crc kubenswrapper[4988]: I1008 19:48:16.238568 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-766746c8cd-vsm49" event={"ID":"dbb701b2-bfa0-4a83-b0db-91bd2d047e68","Type":"ContainerStarted","Data":"4ae0096d218d4bbaf0f9539a92c6140f1143e1bb586fd14cac6c3d8aa4ecb2d6"} Oct 08 19:48:16 crc kubenswrapper[4988]: I1008 19:48:16.240555 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-766746c8cd-vsm49" Oct 08 19:48:16 crc kubenswrapper[4988]: I1008 19:48:16.240679 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-766746c8cd-vsm49" Oct 08 19:48:16 crc kubenswrapper[4988]: I1008 19:48:16.275493 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-766746c8cd-vsm49" podStartSLOduration=3.275458414 podStartE2EDuration="3.275458414s" podCreationTimestamp="2025-10-08 19:48:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 19:48:16.268381459 +0000 UTC m=+5841.718224289" watchObservedRunningTime="2025-10-08 19:48:16.275458414 +0000 UTC m=+5841.725301264" Oct 08 19:48:18 crc kubenswrapper[4988]: I1008 19:48:18.861681 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-7595cd99bc-sq75r" Oct 08 19:48:18 crc kubenswrapper[4988]: I1008 19:48:18.924497 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5589f75ff5-qj6cv"] Oct 08 19:48:18 crc kubenswrapper[4988]: I1008 19:48:18.924834 4988 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-5589f75ff5-qj6cv" podUID="5943bf5d-dc4e-4699-af97-bbb845b9db43" containerName="dnsmasq-dns" containerID="cri-o://38fe2bfa99b61db13654ec48a0d8f5ede91e22ae272e97c10552c183357640d0" gracePeriod=10 Oct 08 19:48:19 crc kubenswrapper[4988]: I1008 19:48:19.313035 4988 generic.go:334] "Generic (PLEG): container finished" podID="5943bf5d-dc4e-4699-af97-bbb845b9db43" containerID="38fe2bfa99b61db13654ec48a0d8f5ede91e22ae272e97c10552c183357640d0" exitCode=0 Oct 08 19:48:19 crc kubenswrapper[4988]: I1008 19:48:19.313094 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5589f75ff5-qj6cv" event={"ID":"5943bf5d-dc4e-4699-af97-bbb845b9db43","Type":"ContainerDied","Data":"38fe2bfa99b61db13654ec48a0d8f5ede91e22ae272e97c10552c183357640d0"} Oct 08 19:48:19 crc kubenswrapper[4988]: I1008 19:48:19.457031 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5589f75ff5-qj6cv" Oct 08 19:48:19 crc kubenswrapper[4988]: I1008 19:48:19.495004 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5943bf5d-dc4e-4699-af97-bbb845b9db43-ovsdbserver-nb\") pod \"5943bf5d-dc4e-4699-af97-bbb845b9db43\" (UID: \"5943bf5d-dc4e-4699-af97-bbb845b9db43\") " Oct 08 19:48:19 crc kubenswrapper[4988]: I1008 19:48:19.495054 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5943bf5d-dc4e-4699-af97-bbb845b9db43-dns-svc\") pod \"5943bf5d-dc4e-4699-af97-bbb845b9db43\" (UID: \"5943bf5d-dc4e-4699-af97-bbb845b9db43\") " Oct 08 19:48:19 crc kubenswrapper[4988]: I1008 19:48:19.495102 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xrrrt\" (UniqueName: \"kubernetes.io/projected/5943bf5d-dc4e-4699-af97-bbb845b9db43-kube-api-access-xrrrt\") pod \"5943bf5d-dc4e-4699-af97-bbb845b9db43\" (UID: \"5943bf5d-dc4e-4699-af97-bbb845b9db43\") " Oct 08 19:48:19 crc kubenswrapper[4988]: I1008 19:48:19.495142 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5943bf5d-dc4e-4699-af97-bbb845b9db43-config\") pod \"5943bf5d-dc4e-4699-af97-bbb845b9db43\" (UID: \"5943bf5d-dc4e-4699-af97-bbb845b9db43\") " Oct 08 19:48:19 crc kubenswrapper[4988]: I1008 19:48:19.495174 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5943bf5d-dc4e-4699-af97-bbb845b9db43-ovsdbserver-sb\") pod \"5943bf5d-dc4e-4699-af97-bbb845b9db43\" (UID: \"5943bf5d-dc4e-4699-af97-bbb845b9db43\") " Oct 08 19:48:19 crc kubenswrapper[4988]: I1008 19:48:19.502463 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5943bf5d-dc4e-4699-af97-bbb845b9db43-kube-api-access-xrrrt" (OuterVolumeSpecName: "kube-api-access-xrrrt") pod "5943bf5d-dc4e-4699-af97-bbb845b9db43" (UID: "5943bf5d-dc4e-4699-af97-bbb845b9db43"). InnerVolumeSpecName "kube-api-access-xrrrt". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 19:48:19 crc kubenswrapper[4988]: I1008 19:48:19.545324 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5943bf5d-dc4e-4699-af97-bbb845b9db43-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "5943bf5d-dc4e-4699-af97-bbb845b9db43" (UID: "5943bf5d-dc4e-4699-af97-bbb845b9db43"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 19:48:19 crc kubenswrapper[4988]: I1008 19:48:19.553501 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5943bf5d-dc4e-4699-af97-bbb845b9db43-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "5943bf5d-dc4e-4699-af97-bbb845b9db43" (UID: "5943bf5d-dc4e-4699-af97-bbb845b9db43"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 19:48:19 crc kubenswrapper[4988]: I1008 19:48:19.562359 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5943bf5d-dc4e-4699-af97-bbb845b9db43-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "5943bf5d-dc4e-4699-af97-bbb845b9db43" (UID: "5943bf5d-dc4e-4699-af97-bbb845b9db43"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 19:48:19 crc kubenswrapper[4988]: I1008 19:48:19.562445 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5943bf5d-dc4e-4699-af97-bbb845b9db43-config" (OuterVolumeSpecName: "config") pod "5943bf5d-dc4e-4699-af97-bbb845b9db43" (UID: "5943bf5d-dc4e-4699-af97-bbb845b9db43"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 19:48:19 crc kubenswrapper[4988]: I1008 19:48:19.597181 4988 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5943bf5d-dc4e-4699-af97-bbb845b9db43-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 08 19:48:19 crc kubenswrapper[4988]: I1008 19:48:19.597211 4988 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5943bf5d-dc4e-4699-af97-bbb845b9db43-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 08 19:48:19 crc kubenswrapper[4988]: I1008 19:48:19.597222 4988 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xrrrt\" (UniqueName: \"kubernetes.io/projected/5943bf5d-dc4e-4699-af97-bbb845b9db43-kube-api-access-xrrrt\") on node \"crc\" DevicePath \"\"" Oct 08 19:48:19 crc kubenswrapper[4988]: I1008 19:48:19.597233 4988 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5943bf5d-dc4e-4699-af97-bbb845b9db43-config\") on node \"crc\" DevicePath \"\"" Oct 08 19:48:19 crc kubenswrapper[4988]: I1008 19:48:19.597243 4988 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5943bf5d-dc4e-4699-af97-bbb845b9db43-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 08 19:48:20 crc kubenswrapper[4988]: I1008 19:48:20.330041 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5589f75ff5-qj6cv" event={"ID":"5943bf5d-dc4e-4699-af97-bbb845b9db43","Type":"ContainerDied","Data":"0531fc115a562394dd0e6a1270979d2c8438dae8813dc865710678635fc7be41"} Oct 08 19:48:20 crc kubenswrapper[4988]: I1008 19:48:20.330594 4988 scope.go:117] "RemoveContainer" containerID="38fe2bfa99b61db13654ec48a0d8f5ede91e22ae272e97c10552c183357640d0" Oct 08 19:48:20 crc kubenswrapper[4988]: I1008 19:48:20.330117 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5589f75ff5-qj6cv" Oct 08 19:48:20 crc kubenswrapper[4988]: I1008 19:48:20.390514 4988 scope.go:117] "RemoveContainer" containerID="bfb03ed93c36eefdac388693d766b3ad6aa39d28b42fd1cc6881caf8e730d6a2" Oct 08 19:48:20 crc kubenswrapper[4988]: I1008 19:48:20.395031 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5589f75ff5-qj6cv"] Oct 08 19:48:20 crc kubenswrapper[4988]: I1008 19:48:20.425972 4988 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5589f75ff5-qj6cv"] Oct 08 19:48:21 crc kubenswrapper[4988]: I1008 19:48:21.255346 4988 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5943bf5d-dc4e-4699-af97-bbb845b9db43" path="/var/lib/kubelet/pods/5943bf5d-dc4e-4699-af97-bbb845b9db43/volumes" Oct 08 19:48:25 crc kubenswrapper[4988]: I1008 19:48:25.245264 4988 scope.go:117] "RemoveContainer" containerID="f2a29ed22dd7a48bbc929e65fbeceed969a669e78f65fbfb724b03dc0db6e947" Oct 08 19:48:25 crc kubenswrapper[4988]: E1008 19:48:25.246240 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nm54f_openshift-machine-config-operator(f328a20b-5283-407e-8a93-2ab2da599f18)\"" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" Oct 08 19:48:39 crc kubenswrapper[4988]: I1008 19:48:39.238789 4988 scope.go:117] "RemoveContainer" containerID="f2a29ed22dd7a48bbc929e65fbeceed969a669e78f65fbfb724b03dc0db6e947" Oct 08 19:48:39 crc kubenswrapper[4988]: E1008 19:48:39.242116 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nm54f_openshift-machine-config-operator(f328a20b-5283-407e-8a93-2ab2da599f18)\"" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" Oct 08 19:48:45 crc kubenswrapper[4988]: I1008 19:48:45.086988 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-766746c8cd-vsm49" Oct 08 19:48:45 crc kubenswrapper[4988]: I1008 19:48:45.090065 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-766746c8cd-vsm49" Oct 08 19:48:53 crc kubenswrapper[4988]: I1008 19:48:53.239263 4988 scope.go:117] "RemoveContainer" containerID="f2a29ed22dd7a48bbc929e65fbeceed969a669e78f65fbfb724b03dc0db6e947" Oct 08 19:48:53 crc kubenswrapper[4988]: E1008 19:48:53.240476 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nm54f_openshift-machine-config-operator(f328a20b-5283-407e-8a93-2ab2da599f18)\"" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" Oct 08 19:49:06 crc kubenswrapper[4988]: I1008 19:49:06.904485 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-db-create-mdbqg"] Oct 08 19:49:06 crc kubenswrapper[4988]: E1008 19:49:06.905508 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5943bf5d-dc4e-4699-af97-bbb845b9db43" containerName="init" Oct 08 19:49:06 crc kubenswrapper[4988]: I1008 19:49:06.905528 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="5943bf5d-dc4e-4699-af97-bbb845b9db43" containerName="init" Oct 08 19:49:06 crc kubenswrapper[4988]: E1008 19:49:06.905539 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5943bf5d-dc4e-4699-af97-bbb845b9db43" containerName="dnsmasq-dns" Oct 08 19:49:06 crc kubenswrapper[4988]: I1008 19:49:06.905547 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="5943bf5d-dc4e-4699-af97-bbb845b9db43" containerName="dnsmasq-dns" Oct 08 19:49:06 crc kubenswrapper[4988]: I1008 19:49:06.905760 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="5943bf5d-dc4e-4699-af97-bbb845b9db43" containerName="dnsmasq-dns" Oct 08 19:49:06 crc kubenswrapper[4988]: I1008 19:49:06.906671 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-mdbqg" Oct 08 19:49:06 crc kubenswrapper[4988]: I1008 19:49:06.913592 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-mdbqg"] Oct 08 19:49:06 crc kubenswrapper[4988]: I1008 19:49:06.989033 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-db-create-g2pmg"] Oct 08 19:49:06 crc kubenswrapper[4988]: I1008 19:49:06.990693 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-g2pmg" Oct 08 19:49:06 crc kubenswrapper[4988]: I1008 19:49:06.997570 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-g2pmg"] Oct 08 19:49:07 crc kubenswrapper[4988]: I1008 19:49:07.083789 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-57kn8\" (UniqueName: \"kubernetes.io/projected/42622d8d-30ea-4b08-b87c-3c77397d9c0e-kube-api-access-57kn8\") pod \"nova-api-db-create-mdbqg\" (UID: \"42622d8d-30ea-4b08-b87c-3c77397d9c0e\") " pod="openstack/nova-api-db-create-mdbqg" Oct 08 19:49:07 crc kubenswrapper[4988]: I1008 19:49:07.111533 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-db-create-ctbvx"] Oct 08 19:49:07 crc kubenswrapper[4988]: I1008 19:49:07.113728 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-ctbvx" Oct 08 19:49:07 crc kubenswrapper[4988]: I1008 19:49:07.124279 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-ctbvx"] Oct 08 19:49:07 crc kubenswrapper[4988]: I1008 19:49:07.185215 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r5bsr\" (UniqueName: \"kubernetes.io/projected/fd9a23d7-0671-49d5-b63b-2d632743a81c-kube-api-access-r5bsr\") pod \"nova-cell0-db-create-g2pmg\" (UID: \"fd9a23d7-0671-49d5-b63b-2d632743a81c\") " pod="openstack/nova-cell0-db-create-g2pmg" Oct 08 19:49:07 crc kubenswrapper[4988]: I1008 19:49:07.185304 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-57kn8\" (UniqueName: \"kubernetes.io/projected/42622d8d-30ea-4b08-b87c-3c77397d9c0e-kube-api-access-57kn8\") pod \"nova-api-db-create-mdbqg\" (UID: \"42622d8d-30ea-4b08-b87c-3c77397d9c0e\") " pod="openstack/nova-api-db-create-mdbqg" Oct 08 19:49:07 crc kubenswrapper[4988]: I1008 19:49:07.208200 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-57kn8\" (UniqueName: \"kubernetes.io/projected/42622d8d-30ea-4b08-b87c-3c77397d9c0e-kube-api-access-57kn8\") pod \"nova-api-db-create-mdbqg\" (UID: \"42622d8d-30ea-4b08-b87c-3c77397d9c0e\") " pod="openstack/nova-api-db-create-mdbqg" Oct 08 19:49:07 crc kubenswrapper[4988]: I1008 19:49:07.225469 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-mdbqg" Oct 08 19:49:07 crc kubenswrapper[4988]: I1008 19:49:07.287327 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r5bsr\" (UniqueName: \"kubernetes.io/projected/fd9a23d7-0671-49d5-b63b-2d632743a81c-kube-api-access-r5bsr\") pod \"nova-cell0-db-create-g2pmg\" (UID: \"fd9a23d7-0671-49d5-b63b-2d632743a81c\") " pod="openstack/nova-cell0-db-create-g2pmg" Oct 08 19:49:07 crc kubenswrapper[4988]: I1008 19:49:07.287725 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ps4c8\" (UniqueName: \"kubernetes.io/projected/8a3a3ca1-d84a-45f0-8879-cdc313823841-kube-api-access-ps4c8\") pod \"nova-cell1-db-create-ctbvx\" (UID: \"8a3a3ca1-d84a-45f0-8879-cdc313823841\") " pod="openstack/nova-cell1-db-create-ctbvx" Oct 08 19:49:07 crc kubenswrapper[4988]: I1008 19:49:07.308189 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r5bsr\" (UniqueName: \"kubernetes.io/projected/fd9a23d7-0671-49d5-b63b-2d632743a81c-kube-api-access-r5bsr\") pod \"nova-cell0-db-create-g2pmg\" (UID: \"fd9a23d7-0671-49d5-b63b-2d632743a81c\") " pod="openstack/nova-cell0-db-create-g2pmg" Oct 08 19:49:07 crc kubenswrapper[4988]: I1008 19:49:07.314645 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-g2pmg" Oct 08 19:49:07 crc kubenswrapper[4988]: I1008 19:49:07.392512 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ps4c8\" (UniqueName: \"kubernetes.io/projected/8a3a3ca1-d84a-45f0-8879-cdc313823841-kube-api-access-ps4c8\") pod \"nova-cell1-db-create-ctbvx\" (UID: \"8a3a3ca1-d84a-45f0-8879-cdc313823841\") " pod="openstack/nova-cell1-db-create-ctbvx" Oct 08 19:49:07 crc kubenswrapper[4988]: I1008 19:49:07.421076 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ps4c8\" (UniqueName: \"kubernetes.io/projected/8a3a3ca1-d84a-45f0-8879-cdc313823841-kube-api-access-ps4c8\") pod \"nova-cell1-db-create-ctbvx\" (UID: \"8a3a3ca1-d84a-45f0-8879-cdc313823841\") " pod="openstack/nova-cell1-db-create-ctbvx" Oct 08 19:49:07 crc kubenswrapper[4988]: I1008 19:49:07.431522 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-ctbvx" Oct 08 19:49:07 crc kubenswrapper[4988]: I1008 19:49:07.737537 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-mdbqg"] Oct 08 19:49:07 crc kubenswrapper[4988]: I1008 19:49:07.785738 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-ctbvx"] Oct 08 19:49:07 crc kubenswrapper[4988]: W1008 19:49:07.796430 4988 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8a3a3ca1_d84a_45f0_8879_cdc313823841.slice/crio-65524b09a97b691294b2a94ef43514ad4bdddd30c5957dc40d8db49f0af8f53b WatchSource:0}: Error finding container 65524b09a97b691294b2a94ef43514ad4bdddd30c5957dc40d8db49f0af8f53b: Status 404 returned error can't find the container with id 65524b09a97b691294b2a94ef43514ad4bdddd30c5957dc40d8db49f0af8f53b Oct 08 19:49:07 crc kubenswrapper[4988]: I1008 19:49:07.829606 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-g2pmg"] Oct 08 19:49:07 crc kubenswrapper[4988]: W1008 19:49:07.832646 4988 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podfd9a23d7_0671_49d5_b63b_2d632743a81c.slice/crio-37b010bc673e05f45dcad10d7143c75247eb23e6e0b0d404d486b0c4f9151f1c WatchSource:0}: Error finding container 37b010bc673e05f45dcad10d7143c75247eb23e6e0b0d404d486b0c4f9151f1c: Status 404 returned error can't find the container with id 37b010bc673e05f45dcad10d7143c75247eb23e6e0b0d404d486b0c4f9151f1c Oct 08 19:49:07 crc kubenswrapper[4988]: I1008 19:49:07.864353 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-ctbvx" event={"ID":"8a3a3ca1-d84a-45f0-8879-cdc313823841","Type":"ContainerStarted","Data":"65524b09a97b691294b2a94ef43514ad4bdddd30c5957dc40d8db49f0af8f53b"} Oct 08 19:49:07 crc kubenswrapper[4988]: I1008 19:49:07.869081 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-g2pmg" event={"ID":"fd9a23d7-0671-49d5-b63b-2d632743a81c","Type":"ContainerStarted","Data":"37b010bc673e05f45dcad10d7143c75247eb23e6e0b0d404d486b0c4f9151f1c"} Oct 08 19:49:07 crc kubenswrapper[4988]: I1008 19:49:07.870344 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-mdbqg" event={"ID":"42622d8d-30ea-4b08-b87c-3c77397d9c0e","Type":"ContainerStarted","Data":"2df99cf78901965e40c25ce2620a6aa97fcafd389d94a43cac8640fec03709f4"} Oct 08 19:49:08 crc kubenswrapper[4988]: I1008 19:49:08.239441 4988 scope.go:117] "RemoveContainer" containerID="f2a29ed22dd7a48bbc929e65fbeceed969a669e78f65fbfb724b03dc0db6e947" Oct 08 19:49:08 crc kubenswrapper[4988]: E1008 19:49:08.239885 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nm54f_openshift-machine-config-operator(f328a20b-5283-407e-8a93-2ab2da599f18)\"" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" Oct 08 19:49:08 crc kubenswrapper[4988]: I1008 19:49:08.882213 4988 generic.go:334] "Generic (PLEG): container finished" podID="fd9a23d7-0671-49d5-b63b-2d632743a81c" containerID="95c7fc4b25b5bddeed68db71ab1b187e3958e71ccbdf4c0e924fb8a3fc37b446" exitCode=0 Oct 08 19:49:08 crc kubenswrapper[4988]: I1008 19:49:08.882302 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-g2pmg" event={"ID":"fd9a23d7-0671-49d5-b63b-2d632743a81c","Type":"ContainerDied","Data":"95c7fc4b25b5bddeed68db71ab1b187e3958e71ccbdf4c0e924fb8a3fc37b446"} Oct 08 19:49:08 crc kubenswrapper[4988]: I1008 19:49:08.884747 4988 generic.go:334] "Generic (PLEG): container finished" podID="42622d8d-30ea-4b08-b87c-3c77397d9c0e" containerID="97202d3b51d442b5a41c083e3250d51fc15b8e002a829755567476a33bfba747" exitCode=0 Oct 08 19:49:08 crc kubenswrapper[4988]: I1008 19:49:08.884940 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-mdbqg" event={"ID":"42622d8d-30ea-4b08-b87c-3c77397d9c0e","Type":"ContainerDied","Data":"97202d3b51d442b5a41c083e3250d51fc15b8e002a829755567476a33bfba747"} Oct 08 19:49:08 crc kubenswrapper[4988]: I1008 19:49:08.888604 4988 generic.go:334] "Generic (PLEG): container finished" podID="8a3a3ca1-d84a-45f0-8879-cdc313823841" containerID="89036c30ed9153f070b806de127d546654637116b052af9ceb17921797fabcc7" exitCode=0 Oct 08 19:49:08 crc kubenswrapper[4988]: I1008 19:49:08.888718 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-ctbvx" event={"ID":"8a3a3ca1-d84a-45f0-8879-cdc313823841","Type":"ContainerDied","Data":"89036c30ed9153f070b806de127d546654637116b052af9ceb17921797fabcc7"} Oct 08 19:49:10 crc kubenswrapper[4988]: I1008 19:49:10.366546 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-mdbqg" Oct 08 19:49:10 crc kubenswrapper[4988]: I1008 19:49:10.375239 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-g2pmg" Oct 08 19:49:10 crc kubenswrapper[4988]: I1008 19:49:10.386560 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-ctbvx" Oct 08 19:49:10 crc kubenswrapper[4988]: I1008 19:49:10.456926 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-57kn8\" (UniqueName: \"kubernetes.io/projected/42622d8d-30ea-4b08-b87c-3c77397d9c0e-kube-api-access-57kn8\") pod \"42622d8d-30ea-4b08-b87c-3c77397d9c0e\" (UID: \"42622d8d-30ea-4b08-b87c-3c77397d9c0e\") " Oct 08 19:49:10 crc kubenswrapper[4988]: I1008 19:49:10.457023 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r5bsr\" (UniqueName: \"kubernetes.io/projected/fd9a23d7-0671-49d5-b63b-2d632743a81c-kube-api-access-r5bsr\") pod \"fd9a23d7-0671-49d5-b63b-2d632743a81c\" (UID: \"fd9a23d7-0671-49d5-b63b-2d632743a81c\") " Oct 08 19:49:10 crc kubenswrapper[4988]: I1008 19:49:10.462228 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/42622d8d-30ea-4b08-b87c-3c77397d9c0e-kube-api-access-57kn8" (OuterVolumeSpecName: "kube-api-access-57kn8") pod "42622d8d-30ea-4b08-b87c-3c77397d9c0e" (UID: "42622d8d-30ea-4b08-b87c-3c77397d9c0e"). InnerVolumeSpecName "kube-api-access-57kn8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 19:49:10 crc kubenswrapper[4988]: I1008 19:49:10.463807 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fd9a23d7-0671-49d5-b63b-2d632743a81c-kube-api-access-r5bsr" (OuterVolumeSpecName: "kube-api-access-r5bsr") pod "fd9a23d7-0671-49d5-b63b-2d632743a81c" (UID: "fd9a23d7-0671-49d5-b63b-2d632743a81c"). InnerVolumeSpecName "kube-api-access-r5bsr". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 19:49:10 crc kubenswrapper[4988]: I1008 19:49:10.558700 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ps4c8\" (UniqueName: \"kubernetes.io/projected/8a3a3ca1-d84a-45f0-8879-cdc313823841-kube-api-access-ps4c8\") pod \"8a3a3ca1-d84a-45f0-8879-cdc313823841\" (UID: \"8a3a3ca1-d84a-45f0-8879-cdc313823841\") " Oct 08 19:49:10 crc kubenswrapper[4988]: I1008 19:49:10.559347 4988 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-57kn8\" (UniqueName: \"kubernetes.io/projected/42622d8d-30ea-4b08-b87c-3c77397d9c0e-kube-api-access-57kn8\") on node \"crc\" DevicePath \"\"" Oct 08 19:49:10 crc kubenswrapper[4988]: I1008 19:49:10.559361 4988 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r5bsr\" (UniqueName: \"kubernetes.io/projected/fd9a23d7-0671-49d5-b63b-2d632743a81c-kube-api-access-r5bsr\") on node \"crc\" DevicePath \"\"" Oct 08 19:49:10 crc kubenswrapper[4988]: I1008 19:49:10.562017 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8a3a3ca1-d84a-45f0-8879-cdc313823841-kube-api-access-ps4c8" (OuterVolumeSpecName: "kube-api-access-ps4c8") pod "8a3a3ca1-d84a-45f0-8879-cdc313823841" (UID: "8a3a3ca1-d84a-45f0-8879-cdc313823841"). InnerVolumeSpecName "kube-api-access-ps4c8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 19:49:10 crc kubenswrapper[4988]: I1008 19:49:10.661272 4988 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ps4c8\" (UniqueName: \"kubernetes.io/projected/8a3a3ca1-d84a-45f0-8879-cdc313823841-kube-api-access-ps4c8\") on node \"crc\" DevicePath \"\"" Oct 08 19:49:10 crc kubenswrapper[4988]: I1008 19:49:10.925145 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-ctbvx" event={"ID":"8a3a3ca1-d84a-45f0-8879-cdc313823841","Type":"ContainerDied","Data":"65524b09a97b691294b2a94ef43514ad4bdddd30c5957dc40d8db49f0af8f53b"} Oct 08 19:49:10 crc kubenswrapper[4988]: I1008 19:49:10.925211 4988 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="65524b09a97b691294b2a94ef43514ad4bdddd30c5957dc40d8db49f0af8f53b" Oct 08 19:49:10 crc kubenswrapper[4988]: I1008 19:49:10.925300 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-ctbvx" Oct 08 19:49:10 crc kubenswrapper[4988]: I1008 19:49:10.930737 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-g2pmg" Oct 08 19:49:10 crc kubenswrapper[4988]: I1008 19:49:10.930748 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-g2pmg" event={"ID":"fd9a23d7-0671-49d5-b63b-2d632743a81c","Type":"ContainerDied","Data":"37b010bc673e05f45dcad10d7143c75247eb23e6e0b0d404d486b0c4f9151f1c"} Oct 08 19:49:10 crc kubenswrapper[4988]: I1008 19:49:10.930796 4988 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="37b010bc673e05f45dcad10d7143c75247eb23e6e0b0d404d486b0c4f9151f1c" Oct 08 19:49:10 crc kubenswrapper[4988]: I1008 19:49:10.934816 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-mdbqg" event={"ID":"42622d8d-30ea-4b08-b87c-3c77397d9c0e","Type":"ContainerDied","Data":"2df99cf78901965e40c25ce2620a6aa97fcafd389d94a43cac8640fec03709f4"} Oct 08 19:49:10 crc kubenswrapper[4988]: I1008 19:49:10.934865 4988 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2df99cf78901965e40c25ce2620a6aa97fcafd389d94a43cac8640fec03709f4" Oct 08 19:49:10 crc kubenswrapper[4988]: I1008 19:49:10.934928 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-mdbqg" Oct 08 19:49:17 crc kubenswrapper[4988]: I1008 19:49:17.059252 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-015a-account-create-t8ks5"] Oct 08 19:49:17 crc kubenswrapper[4988]: E1008 19:49:17.060103 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="42622d8d-30ea-4b08-b87c-3c77397d9c0e" containerName="mariadb-database-create" Oct 08 19:49:17 crc kubenswrapper[4988]: I1008 19:49:17.060204 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="42622d8d-30ea-4b08-b87c-3c77397d9c0e" containerName="mariadb-database-create" Oct 08 19:49:17 crc kubenswrapper[4988]: E1008 19:49:17.060225 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8a3a3ca1-d84a-45f0-8879-cdc313823841" containerName="mariadb-database-create" Oct 08 19:49:17 crc kubenswrapper[4988]: I1008 19:49:17.060233 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="8a3a3ca1-d84a-45f0-8879-cdc313823841" containerName="mariadb-database-create" Oct 08 19:49:17 crc kubenswrapper[4988]: E1008 19:49:17.060256 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fd9a23d7-0671-49d5-b63b-2d632743a81c" containerName="mariadb-database-create" Oct 08 19:49:17 crc kubenswrapper[4988]: I1008 19:49:17.060264 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="fd9a23d7-0671-49d5-b63b-2d632743a81c" containerName="mariadb-database-create" Oct 08 19:49:17 crc kubenswrapper[4988]: I1008 19:49:17.060503 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="8a3a3ca1-d84a-45f0-8879-cdc313823841" containerName="mariadb-database-create" Oct 08 19:49:17 crc kubenswrapper[4988]: I1008 19:49:17.060523 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="42622d8d-30ea-4b08-b87c-3c77397d9c0e" containerName="mariadb-database-create" Oct 08 19:49:17 crc kubenswrapper[4988]: I1008 19:49:17.060548 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="fd9a23d7-0671-49d5-b63b-2d632743a81c" containerName="mariadb-database-create" Oct 08 19:49:17 crc kubenswrapper[4988]: I1008 19:49:17.061209 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-015a-account-create-t8ks5" Oct 08 19:49:17 crc kubenswrapper[4988]: I1008 19:49:17.064894 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-db-secret" Oct 08 19:49:17 crc kubenswrapper[4988]: I1008 19:49:17.093875 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-015a-account-create-t8ks5"] Oct 08 19:49:17 crc kubenswrapper[4988]: I1008 19:49:17.204451 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6mszz\" (UniqueName: \"kubernetes.io/projected/19e6cdb4-5c6f-4e69-a918-9f88dd304d9e-kube-api-access-6mszz\") pod \"nova-api-015a-account-create-t8ks5\" (UID: \"19e6cdb4-5c6f-4e69-a918-9f88dd304d9e\") " pod="openstack/nova-api-015a-account-create-t8ks5" Oct 08 19:49:17 crc kubenswrapper[4988]: I1008 19:49:17.256362 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-0c51-account-create-frwnd"] Oct 08 19:49:17 crc kubenswrapper[4988]: I1008 19:49:17.258088 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-0c51-account-create-frwnd"] Oct 08 19:49:17 crc kubenswrapper[4988]: I1008 19:49:17.258210 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-0c51-account-create-frwnd" Oct 08 19:49:17 crc kubenswrapper[4988]: I1008 19:49:17.262369 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-db-secret" Oct 08 19:49:17 crc kubenswrapper[4988]: I1008 19:49:17.306161 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6mszz\" (UniqueName: \"kubernetes.io/projected/19e6cdb4-5c6f-4e69-a918-9f88dd304d9e-kube-api-access-6mszz\") pod \"nova-api-015a-account-create-t8ks5\" (UID: \"19e6cdb4-5c6f-4e69-a918-9f88dd304d9e\") " pod="openstack/nova-api-015a-account-create-t8ks5" Oct 08 19:49:17 crc kubenswrapper[4988]: I1008 19:49:17.336901 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6mszz\" (UniqueName: \"kubernetes.io/projected/19e6cdb4-5c6f-4e69-a918-9f88dd304d9e-kube-api-access-6mszz\") pod \"nova-api-015a-account-create-t8ks5\" (UID: \"19e6cdb4-5c6f-4e69-a918-9f88dd304d9e\") " pod="openstack/nova-api-015a-account-create-t8ks5" Oct 08 19:49:17 crc kubenswrapper[4988]: I1008 19:49:17.396641 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-015a-account-create-t8ks5" Oct 08 19:49:17 crc kubenswrapper[4988]: I1008 19:49:17.408638 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4x5b5\" (UniqueName: \"kubernetes.io/projected/f3f270ce-b57a-4d31-a77a-5ba66dabab78-kube-api-access-4x5b5\") pod \"nova-cell0-0c51-account-create-frwnd\" (UID: \"f3f270ce-b57a-4d31-a77a-5ba66dabab78\") " pod="openstack/nova-cell0-0c51-account-create-frwnd" Oct 08 19:49:17 crc kubenswrapper[4988]: I1008 19:49:17.471071 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-ef5f-account-create-mn8dp"] Oct 08 19:49:17 crc kubenswrapper[4988]: I1008 19:49:17.473406 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-ef5f-account-create-mn8dp"] Oct 08 19:49:17 crc kubenswrapper[4988]: I1008 19:49:17.473535 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-ef5f-account-create-mn8dp" Oct 08 19:49:17 crc kubenswrapper[4988]: I1008 19:49:17.482176 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-db-secret" Oct 08 19:49:17 crc kubenswrapper[4988]: I1008 19:49:17.511038 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4x5b5\" (UniqueName: \"kubernetes.io/projected/f3f270ce-b57a-4d31-a77a-5ba66dabab78-kube-api-access-4x5b5\") pod \"nova-cell0-0c51-account-create-frwnd\" (UID: \"f3f270ce-b57a-4d31-a77a-5ba66dabab78\") " pod="openstack/nova-cell0-0c51-account-create-frwnd" Oct 08 19:49:17 crc kubenswrapper[4988]: I1008 19:49:17.527880 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4x5b5\" (UniqueName: \"kubernetes.io/projected/f3f270ce-b57a-4d31-a77a-5ba66dabab78-kube-api-access-4x5b5\") pod \"nova-cell0-0c51-account-create-frwnd\" (UID: \"f3f270ce-b57a-4d31-a77a-5ba66dabab78\") " pod="openstack/nova-cell0-0c51-account-create-frwnd" Oct 08 19:49:17 crc kubenswrapper[4988]: I1008 19:49:17.584631 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-0c51-account-create-frwnd" Oct 08 19:49:17 crc kubenswrapper[4988]: I1008 19:49:17.612550 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6xknz\" (UniqueName: \"kubernetes.io/projected/4cde7ab7-3097-45ad-a312-98ddc074114c-kube-api-access-6xknz\") pod \"nova-cell1-ef5f-account-create-mn8dp\" (UID: \"4cde7ab7-3097-45ad-a312-98ddc074114c\") " pod="openstack/nova-cell1-ef5f-account-create-mn8dp" Oct 08 19:49:17 crc kubenswrapper[4988]: I1008 19:49:17.714983 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6xknz\" (UniqueName: \"kubernetes.io/projected/4cde7ab7-3097-45ad-a312-98ddc074114c-kube-api-access-6xknz\") pod \"nova-cell1-ef5f-account-create-mn8dp\" (UID: \"4cde7ab7-3097-45ad-a312-98ddc074114c\") " pod="openstack/nova-cell1-ef5f-account-create-mn8dp" Oct 08 19:49:17 crc kubenswrapper[4988]: I1008 19:49:17.736594 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6xknz\" (UniqueName: \"kubernetes.io/projected/4cde7ab7-3097-45ad-a312-98ddc074114c-kube-api-access-6xknz\") pod \"nova-cell1-ef5f-account-create-mn8dp\" (UID: \"4cde7ab7-3097-45ad-a312-98ddc074114c\") " pod="openstack/nova-cell1-ef5f-account-create-mn8dp" Oct 08 19:49:17 crc kubenswrapper[4988]: I1008 19:49:17.891659 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-ef5f-account-create-mn8dp" Oct 08 19:49:17 crc kubenswrapper[4988]: I1008 19:49:17.893760 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-015a-account-create-t8ks5"] Oct 08 19:49:18 crc kubenswrapper[4988]: I1008 19:49:18.024218 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-015a-account-create-t8ks5" event={"ID":"19e6cdb4-5c6f-4e69-a918-9f88dd304d9e","Type":"ContainerStarted","Data":"242089320daea8d8c90a3a5138c7fa6439b2776cf036045f19f13e72715bb77c"} Oct 08 19:49:18 crc kubenswrapper[4988]: I1008 19:49:18.036312 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-0c51-account-create-frwnd"] Oct 08 19:49:18 crc kubenswrapper[4988]: W1008 19:49:18.058732 4988 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf3f270ce_b57a_4d31_a77a_5ba66dabab78.slice/crio-0503c0f80860aa9ec3c47b10af5cfceb49477f99653e8762a6264a0551e328cf WatchSource:0}: Error finding container 0503c0f80860aa9ec3c47b10af5cfceb49477f99653e8762a6264a0551e328cf: Status 404 returned error can't find the container with id 0503c0f80860aa9ec3c47b10af5cfceb49477f99653e8762a6264a0551e328cf Oct 08 19:49:18 crc kubenswrapper[4988]: I1008 19:49:18.340861 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-ef5f-account-create-mn8dp"] Oct 08 19:49:19 crc kubenswrapper[4988]: I1008 19:49:19.038084 4988 generic.go:334] "Generic (PLEG): container finished" podID="4cde7ab7-3097-45ad-a312-98ddc074114c" containerID="17bffefafaad07ac2d268dde4cf519ffef397e22b0bb24d9fc9fd17189af4020" exitCode=0 Oct 08 19:49:19 crc kubenswrapper[4988]: I1008 19:49:19.038176 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-ef5f-account-create-mn8dp" event={"ID":"4cde7ab7-3097-45ad-a312-98ddc074114c","Type":"ContainerDied","Data":"17bffefafaad07ac2d268dde4cf519ffef397e22b0bb24d9fc9fd17189af4020"} Oct 08 19:49:19 crc kubenswrapper[4988]: I1008 19:49:19.038570 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-ef5f-account-create-mn8dp" event={"ID":"4cde7ab7-3097-45ad-a312-98ddc074114c","Type":"ContainerStarted","Data":"1788e602c1378e827f737d4a1ddfcedbfdf0fff22c6d9941429fabb7ed6f2bbf"} Oct 08 19:49:19 crc kubenswrapper[4988]: I1008 19:49:19.041066 4988 generic.go:334] "Generic (PLEG): container finished" podID="19e6cdb4-5c6f-4e69-a918-9f88dd304d9e" containerID="4ff1fec466874c2f74cd25cc619db431c83691c072385c72f632aceaa00e88ec" exitCode=0 Oct 08 19:49:19 crc kubenswrapper[4988]: I1008 19:49:19.041197 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-015a-account-create-t8ks5" event={"ID":"19e6cdb4-5c6f-4e69-a918-9f88dd304d9e","Type":"ContainerDied","Data":"4ff1fec466874c2f74cd25cc619db431c83691c072385c72f632aceaa00e88ec"} Oct 08 19:49:19 crc kubenswrapper[4988]: I1008 19:49:19.043925 4988 generic.go:334] "Generic (PLEG): container finished" podID="f3f270ce-b57a-4d31-a77a-5ba66dabab78" containerID="30d556b219e1e75ea900a02038621eaa6233a071efd905b622e3fc38cb801be5" exitCode=0 Oct 08 19:49:19 crc kubenswrapper[4988]: I1008 19:49:19.043952 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-0c51-account-create-frwnd" event={"ID":"f3f270ce-b57a-4d31-a77a-5ba66dabab78","Type":"ContainerDied","Data":"30d556b219e1e75ea900a02038621eaa6233a071efd905b622e3fc38cb801be5"} Oct 08 19:49:19 crc kubenswrapper[4988]: I1008 19:49:19.043968 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-0c51-account-create-frwnd" event={"ID":"f3f270ce-b57a-4d31-a77a-5ba66dabab78","Type":"ContainerStarted","Data":"0503c0f80860aa9ec3c47b10af5cfceb49477f99653e8762a6264a0551e328cf"} Oct 08 19:49:19 crc kubenswrapper[4988]: I1008 19:49:19.240526 4988 scope.go:117] "RemoveContainer" containerID="f2a29ed22dd7a48bbc929e65fbeceed969a669e78f65fbfb724b03dc0db6e947" Oct 08 19:49:19 crc kubenswrapper[4988]: E1008 19:49:19.241709 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nm54f_openshift-machine-config-operator(f328a20b-5283-407e-8a93-2ab2da599f18)\"" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" Oct 08 19:49:20 crc kubenswrapper[4988]: I1008 19:49:20.546081 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-015a-account-create-t8ks5" Oct 08 19:49:20 crc kubenswrapper[4988]: I1008 19:49:20.553207 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-0c51-account-create-frwnd" Oct 08 19:49:20 crc kubenswrapper[4988]: I1008 19:49:20.576416 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-ef5f-account-create-mn8dp" Oct 08 19:49:20 crc kubenswrapper[4988]: I1008 19:49:20.673169 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4x5b5\" (UniqueName: \"kubernetes.io/projected/f3f270ce-b57a-4d31-a77a-5ba66dabab78-kube-api-access-4x5b5\") pod \"f3f270ce-b57a-4d31-a77a-5ba66dabab78\" (UID: \"f3f270ce-b57a-4d31-a77a-5ba66dabab78\") " Oct 08 19:49:20 crc kubenswrapper[4988]: I1008 19:49:20.673291 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6xknz\" (UniqueName: \"kubernetes.io/projected/4cde7ab7-3097-45ad-a312-98ddc074114c-kube-api-access-6xknz\") pod \"4cde7ab7-3097-45ad-a312-98ddc074114c\" (UID: \"4cde7ab7-3097-45ad-a312-98ddc074114c\") " Oct 08 19:49:20 crc kubenswrapper[4988]: I1008 19:49:20.673378 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6mszz\" (UniqueName: \"kubernetes.io/projected/19e6cdb4-5c6f-4e69-a918-9f88dd304d9e-kube-api-access-6mszz\") pod \"19e6cdb4-5c6f-4e69-a918-9f88dd304d9e\" (UID: \"19e6cdb4-5c6f-4e69-a918-9f88dd304d9e\") " Oct 08 19:49:20 crc kubenswrapper[4988]: I1008 19:49:20.681240 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/19e6cdb4-5c6f-4e69-a918-9f88dd304d9e-kube-api-access-6mszz" (OuterVolumeSpecName: "kube-api-access-6mszz") pod "19e6cdb4-5c6f-4e69-a918-9f88dd304d9e" (UID: "19e6cdb4-5c6f-4e69-a918-9f88dd304d9e"). InnerVolumeSpecName "kube-api-access-6mszz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 19:49:20 crc kubenswrapper[4988]: I1008 19:49:20.681825 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4cde7ab7-3097-45ad-a312-98ddc074114c-kube-api-access-6xknz" (OuterVolumeSpecName: "kube-api-access-6xknz") pod "4cde7ab7-3097-45ad-a312-98ddc074114c" (UID: "4cde7ab7-3097-45ad-a312-98ddc074114c"). InnerVolumeSpecName "kube-api-access-6xknz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 19:49:20 crc kubenswrapper[4988]: I1008 19:49:20.682241 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f3f270ce-b57a-4d31-a77a-5ba66dabab78-kube-api-access-4x5b5" (OuterVolumeSpecName: "kube-api-access-4x5b5") pod "f3f270ce-b57a-4d31-a77a-5ba66dabab78" (UID: "f3f270ce-b57a-4d31-a77a-5ba66dabab78"). InnerVolumeSpecName "kube-api-access-4x5b5". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 19:49:20 crc kubenswrapper[4988]: I1008 19:49:20.775538 4988 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6xknz\" (UniqueName: \"kubernetes.io/projected/4cde7ab7-3097-45ad-a312-98ddc074114c-kube-api-access-6xknz\") on node \"crc\" DevicePath \"\"" Oct 08 19:49:20 crc kubenswrapper[4988]: I1008 19:49:20.775596 4988 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6mszz\" (UniqueName: \"kubernetes.io/projected/19e6cdb4-5c6f-4e69-a918-9f88dd304d9e-kube-api-access-6mszz\") on node \"crc\" DevicePath \"\"" Oct 08 19:49:20 crc kubenswrapper[4988]: I1008 19:49:20.775615 4988 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4x5b5\" (UniqueName: \"kubernetes.io/projected/f3f270ce-b57a-4d31-a77a-5ba66dabab78-kube-api-access-4x5b5\") on node \"crc\" DevicePath \"\"" Oct 08 19:49:21 crc kubenswrapper[4988]: I1008 19:49:21.066445 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-015a-account-create-t8ks5" event={"ID":"19e6cdb4-5c6f-4e69-a918-9f88dd304d9e","Type":"ContainerDied","Data":"242089320daea8d8c90a3a5138c7fa6439b2776cf036045f19f13e72715bb77c"} Oct 08 19:49:21 crc kubenswrapper[4988]: I1008 19:49:21.066872 4988 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="242089320daea8d8c90a3a5138c7fa6439b2776cf036045f19f13e72715bb77c" Oct 08 19:49:21 crc kubenswrapper[4988]: I1008 19:49:21.066961 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-015a-account-create-t8ks5" Oct 08 19:49:21 crc kubenswrapper[4988]: I1008 19:49:21.074244 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-0c51-account-create-frwnd" event={"ID":"f3f270ce-b57a-4d31-a77a-5ba66dabab78","Type":"ContainerDied","Data":"0503c0f80860aa9ec3c47b10af5cfceb49477f99653e8762a6264a0551e328cf"} Oct 08 19:49:21 crc kubenswrapper[4988]: I1008 19:49:21.074281 4988 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0503c0f80860aa9ec3c47b10af5cfceb49477f99653e8762a6264a0551e328cf" Oct 08 19:49:21 crc kubenswrapper[4988]: I1008 19:49:21.074291 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-0c51-account-create-frwnd" Oct 08 19:49:21 crc kubenswrapper[4988]: I1008 19:49:21.076068 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-ef5f-account-create-mn8dp" event={"ID":"4cde7ab7-3097-45ad-a312-98ddc074114c","Type":"ContainerDied","Data":"1788e602c1378e827f737d4a1ddfcedbfdf0fff22c6d9941429fabb7ed6f2bbf"} Oct 08 19:49:21 crc kubenswrapper[4988]: I1008 19:49:21.076125 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-ef5f-account-create-mn8dp" Oct 08 19:49:21 crc kubenswrapper[4988]: I1008 19:49:21.076132 4988 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1788e602c1378e827f737d4a1ddfcedbfdf0fff22c6d9941429fabb7ed6f2bbf" Oct 08 19:49:22 crc kubenswrapper[4988]: I1008 19:49:22.505547 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-db-sync-fscm7"] Oct 08 19:49:22 crc kubenswrapper[4988]: E1008 19:49:22.506135 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="19e6cdb4-5c6f-4e69-a918-9f88dd304d9e" containerName="mariadb-account-create" Oct 08 19:49:22 crc kubenswrapper[4988]: I1008 19:49:22.506147 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="19e6cdb4-5c6f-4e69-a918-9f88dd304d9e" containerName="mariadb-account-create" Oct 08 19:49:22 crc kubenswrapper[4988]: E1008 19:49:22.506162 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f3f270ce-b57a-4d31-a77a-5ba66dabab78" containerName="mariadb-account-create" Oct 08 19:49:22 crc kubenswrapper[4988]: I1008 19:49:22.506167 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="f3f270ce-b57a-4d31-a77a-5ba66dabab78" containerName="mariadb-account-create" Oct 08 19:49:22 crc kubenswrapper[4988]: E1008 19:49:22.506183 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4cde7ab7-3097-45ad-a312-98ddc074114c" containerName="mariadb-account-create" Oct 08 19:49:22 crc kubenswrapper[4988]: I1008 19:49:22.506190 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="4cde7ab7-3097-45ad-a312-98ddc074114c" containerName="mariadb-account-create" Oct 08 19:49:22 crc kubenswrapper[4988]: I1008 19:49:22.506398 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="19e6cdb4-5c6f-4e69-a918-9f88dd304d9e" containerName="mariadb-account-create" Oct 08 19:49:22 crc kubenswrapper[4988]: I1008 19:49:22.506425 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="f3f270ce-b57a-4d31-a77a-5ba66dabab78" containerName="mariadb-account-create" Oct 08 19:49:22 crc kubenswrapper[4988]: I1008 19:49:22.506443 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="4cde7ab7-3097-45ad-a312-98ddc074114c" containerName="mariadb-account-create" Oct 08 19:49:22 crc kubenswrapper[4988]: I1008 19:49:22.507141 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-fscm7" Oct 08 19:49:22 crc kubenswrapper[4988]: I1008 19:49:22.517287 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-lcgs9" Oct 08 19:49:22 crc kubenswrapper[4988]: I1008 19:49:22.517431 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-scripts" Oct 08 19:49:22 crc kubenswrapper[4988]: I1008 19:49:22.517873 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Oct 08 19:49:22 crc kubenswrapper[4988]: I1008 19:49:22.528720 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-fscm7"] Oct 08 19:49:22 crc kubenswrapper[4988]: I1008 19:49:22.610550 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f95e96b8-7f14-47a8-9d7e-1dca47b4b468-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-fscm7\" (UID: \"f95e96b8-7f14-47a8-9d7e-1dca47b4b468\") " pod="openstack/nova-cell0-conductor-db-sync-fscm7" Oct 08 19:49:22 crc kubenswrapper[4988]: I1008 19:49:22.610937 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qcqmb\" (UniqueName: \"kubernetes.io/projected/f95e96b8-7f14-47a8-9d7e-1dca47b4b468-kube-api-access-qcqmb\") pod \"nova-cell0-conductor-db-sync-fscm7\" (UID: \"f95e96b8-7f14-47a8-9d7e-1dca47b4b468\") " pod="openstack/nova-cell0-conductor-db-sync-fscm7" Oct 08 19:49:22 crc kubenswrapper[4988]: I1008 19:49:22.610978 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f95e96b8-7f14-47a8-9d7e-1dca47b4b468-config-data\") pod \"nova-cell0-conductor-db-sync-fscm7\" (UID: \"f95e96b8-7f14-47a8-9d7e-1dca47b4b468\") " pod="openstack/nova-cell0-conductor-db-sync-fscm7" Oct 08 19:49:22 crc kubenswrapper[4988]: I1008 19:49:22.610999 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f95e96b8-7f14-47a8-9d7e-1dca47b4b468-scripts\") pod \"nova-cell0-conductor-db-sync-fscm7\" (UID: \"f95e96b8-7f14-47a8-9d7e-1dca47b4b468\") " pod="openstack/nova-cell0-conductor-db-sync-fscm7" Oct 08 19:49:22 crc kubenswrapper[4988]: I1008 19:49:22.712940 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qcqmb\" (UniqueName: \"kubernetes.io/projected/f95e96b8-7f14-47a8-9d7e-1dca47b4b468-kube-api-access-qcqmb\") pod \"nova-cell0-conductor-db-sync-fscm7\" (UID: \"f95e96b8-7f14-47a8-9d7e-1dca47b4b468\") " pod="openstack/nova-cell0-conductor-db-sync-fscm7" Oct 08 19:49:22 crc kubenswrapper[4988]: I1008 19:49:22.713225 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f95e96b8-7f14-47a8-9d7e-1dca47b4b468-config-data\") pod \"nova-cell0-conductor-db-sync-fscm7\" (UID: \"f95e96b8-7f14-47a8-9d7e-1dca47b4b468\") " pod="openstack/nova-cell0-conductor-db-sync-fscm7" Oct 08 19:49:22 crc kubenswrapper[4988]: I1008 19:49:22.713337 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f95e96b8-7f14-47a8-9d7e-1dca47b4b468-scripts\") pod \"nova-cell0-conductor-db-sync-fscm7\" (UID: \"f95e96b8-7f14-47a8-9d7e-1dca47b4b468\") " pod="openstack/nova-cell0-conductor-db-sync-fscm7" Oct 08 19:49:22 crc kubenswrapper[4988]: I1008 19:49:22.713537 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f95e96b8-7f14-47a8-9d7e-1dca47b4b468-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-fscm7\" (UID: \"f95e96b8-7f14-47a8-9d7e-1dca47b4b468\") " pod="openstack/nova-cell0-conductor-db-sync-fscm7" Oct 08 19:49:22 crc kubenswrapper[4988]: I1008 19:49:22.719156 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f95e96b8-7f14-47a8-9d7e-1dca47b4b468-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-fscm7\" (UID: \"f95e96b8-7f14-47a8-9d7e-1dca47b4b468\") " pod="openstack/nova-cell0-conductor-db-sync-fscm7" Oct 08 19:49:22 crc kubenswrapper[4988]: I1008 19:49:22.719851 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f95e96b8-7f14-47a8-9d7e-1dca47b4b468-config-data\") pod \"nova-cell0-conductor-db-sync-fscm7\" (UID: \"f95e96b8-7f14-47a8-9d7e-1dca47b4b468\") " pod="openstack/nova-cell0-conductor-db-sync-fscm7" Oct 08 19:49:22 crc kubenswrapper[4988]: I1008 19:49:22.727991 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f95e96b8-7f14-47a8-9d7e-1dca47b4b468-scripts\") pod \"nova-cell0-conductor-db-sync-fscm7\" (UID: \"f95e96b8-7f14-47a8-9d7e-1dca47b4b468\") " pod="openstack/nova-cell0-conductor-db-sync-fscm7" Oct 08 19:49:22 crc kubenswrapper[4988]: I1008 19:49:22.733015 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qcqmb\" (UniqueName: \"kubernetes.io/projected/f95e96b8-7f14-47a8-9d7e-1dca47b4b468-kube-api-access-qcqmb\") pod \"nova-cell0-conductor-db-sync-fscm7\" (UID: \"f95e96b8-7f14-47a8-9d7e-1dca47b4b468\") " pod="openstack/nova-cell0-conductor-db-sync-fscm7" Oct 08 19:49:22 crc kubenswrapper[4988]: I1008 19:49:22.826958 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-fscm7" Oct 08 19:49:23 crc kubenswrapper[4988]: I1008 19:49:23.282954 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-fscm7"] Oct 08 19:49:24 crc kubenswrapper[4988]: I1008 19:49:24.112612 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-fscm7" event={"ID":"f95e96b8-7f14-47a8-9d7e-1dca47b4b468","Type":"ContainerStarted","Data":"a156581bd3e8962fa43eb67139042f378b710e919ab897fb5ab73789cc40f5ad"} Oct 08 19:49:24 crc kubenswrapper[4988]: I1008 19:49:24.112939 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-fscm7" event={"ID":"f95e96b8-7f14-47a8-9d7e-1dca47b4b468","Type":"ContainerStarted","Data":"f35b30d4b6bba713851901a578e2e8c725b91b586952f07add26b8a91649b758"} Oct 08 19:49:24 crc kubenswrapper[4988]: I1008 19:49:24.140269 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-db-sync-fscm7" podStartSLOduration=2.14024417 podStartE2EDuration="2.14024417s" podCreationTimestamp="2025-10-08 19:49:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 19:49:24.12829634 +0000 UTC m=+5909.578139150" watchObservedRunningTime="2025-10-08 19:49:24.14024417 +0000 UTC m=+5909.590086950" Oct 08 19:49:29 crc kubenswrapper[4988]: I1008 19:49:29.166363 4988 generic.go:334] "Generic (PLEG): container finished" podID="f95e96b8-7f14-47a8-9d7e-1dca47b4b468" containerID="a156581bd3e8962fa43eb67139042f378b710e919ab897fb5ab73789cc40f5ad" exitCode=0 Oct 08 19:49:29 crc kubenswrapper[4988]: I1008 19:49:29.166493 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-fscm7" event={"ID":"f95e96b8-7f14-47a8-9d7e-1dca47b4b468","Type":"ContainerDied","Data":"a156581bd3e8962fa43eb67139042f378b710e919ab897fb5ab73789cc40f5ad"} Oct 08 19:49:30 crc kubenswrapper[4988]: I1008 19:49:30.555012 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-fscm7" Oct 08 19:49:30 crc kubenswrapper[4988]: I1008 19:49:30.689219 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f95e96b8-7f14-47a8-9d7e-1dca47b4b468-config-data\") pod \"f95e96b8-7f14-47a8-9d7e-1dca47b4b468\" (UID: \"f95e96b8-7f14-47a8-9d7e-1dca47b4b468\") " Oct 08 19:49:30 crc kubenswrapper[4988]: I1008 19:49:30.689409 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f95e96b8-7f14-47a8-9d7e-1dca47b4b468-scripts\") pod \"f95e96b8-7f14-47a8-9d7e-1dca47b4b468\" (UID: \"f95e96b8-7f14-47a8-9d7e-1dca47b4b468\") " Oct 08 19:49:30 crc kubenswrapper[4988]: I1008 19:49:30.689484 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qcqmb\" (UniqueName: \"kubernetes.io/projected/f95e96b8-7f14-47a8-9d7e-1dca47b4b468-kube-api-access-qcqmb\") pod \"f95e96b8-7f14-47a8-9d7e-1dca47b4b468\" (UID: \"f95e96b8-7f14-47a8-9d7e-1dca47b4b468\") " Oct 08 19:49:30 crc kubenswrapper[4988]: I1008 19:49:30.689560 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f95e96b8-7f14-47a8-9d7e-1dca47b4b468-combined-ca-bundle\") pod \"f95e96b8-7f14-47a8-9d7e-1dca47b4b468\" (UID: \"f95e96b8-7f14-47a8-9d7e-1dca47b4b468\") " Oct 08 19:49:30 crc kubenswrapper[4988]: I1008 19:49:30.697310 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f95e96b8-7f14-47a8-9d7e-1dca47b4b468-scripts" (OuterVolumeSpecName: "scripts") pod "f95e96b8-7f14-47a8-9d7e-1dca47b4b468" (UID: "f95e96b8-7f14-47a8-9d7e-1dca47b4b468"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 19:49:30 crc kubenswrapper[4988]: I1008 19:49:30.697763 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f95e96b8-7f14-47a8-9d7e-1dca47b4b468-kube-api-access-qcqmb" (OuterVolumeSpecName: "kube-api-access-qcqmb") pod "f95e96b8-7f14-47a8-9d7e-1dca47b4b468" (UID: "f95e96b8-7f14-47a8-9d7e-1dca47b4b468"). InnerVolumeSpecName "kube-api-access-qcqmb". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 19:49:30 crc kubenswrapper[4988]: I1008 19:49:30.728698 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f95e96b8-7f14-47a8-9d7e-1dca47b4b468-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f95e96b8-7f14-47a8-9d7e-1dca47b4b468" (UID: "f95e96b8-7f14-47a8-9d7e-1dca47b4b468"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 19:49:30 crc kubenswrapper[4988]: I1008 19:49:30.741683 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f95e96b8-7f14-47a8-9d7e-1dca47b4b468-config-data" (OuterVolumeSpecName: "config-data") pod "f95e96b8-7f14-47a8-9d7e-1dca47b4b468" (UID: "f95e96b8-7f14-47a8-9d7e-1dca47b4b468"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 19:49:30 crc kubenswrapper[4988]: I1008 19:49:30.791580 4988 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f95e96b8-7f14-47a8-9d7e-1dca47b4b468-config-data\") on node \"crc\" DevicePath \"\"" Oct 08 19:49:30 crc kubenswrapper[4988]: I1008 19:49:30.791612 4988 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f95e96b8-7f14-47a8-9d7e-1dca47b4b468-scripts\") on node \"crc\" DevicePath \"\"" Oct 08 19:49:30 crc kubenswrapper[4988]: I1008 19:49:30.791623 4988 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qcqmb\" (UniqueName: \"kubernetes.io/projected/f95e96b8-7f14-47a8-9d7e-1dca47b4b468-kube-api-access-qcqmb\") on node \"crc\" DevicePath \"\"" Oct 08 19:49:30 crc kubenswrapper[4988]: I1008 19:49:30.791632 4988 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f95e96b8-7f14-47a8-9d7e-1dca47b4b468-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 19:49:31 crc kubenswrapper[4988]: I1008 19:49:31.188403 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-fscm7" event={"ID":"f95e96b8-7f14-47a8-9d7e-1dca47b4b468","Type":"ContainerDied","Data":"f35b30d4b6bba713851901a578e2e8c725b91b586952f07add26b8a91649b758"} Oct 08 19:49:31 crc kubenswrapper[4988]: I1008 19:49:31.188451 4988 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f35b30d4b6bba713851901a578e2e8c725b91b586952f07add26b8a91649b758" Oct 08 19:49:31 crc kubenswrapper[4988]: I1008 19:49:31.188462 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-fscm7" Oct 08 19:49:31 crc kubenswrapper[4988]: I1008 19:49:31.277519 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-0"] Oct 08 19:49:31 crc kubenswrapper[4988]: E1008 19:49:31.278046 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f95e96b8-7f14-47a8-9d7e-1dca47b4b468" containerName="nova-cell0-conductor-db-sync" Oct 08 19:49:31 crc kubenswrapper[4988]: I1008 19:49:31.278075 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="f95e96b8-7f14-47a8-9d7e-1dca47b4b468" containerName="nova-cell0-conductor-db-sync" Oct 08 19:49:31 crc kubenswrapper[4988]: I1008 19:49:31.278432 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="f95e96b8-7f14-47a8-9d7e-1dca47b4b468" containerName="nova-cell0-conductor-db-sync" Oct 08 19:49:31 crc kubenswrapper[4988]: I1008 19:49:31.279278 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Oct 08 19:49:31 crc kubenswrapper[4988]: I1008 19:49:31.281660 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-lcgs9" Oct 08 19:49:31 crc kubenswrapper[4988]: I1008 19:49:31.282169 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Oct 08 19:49:31 crc kubenswrapper[4988]: I1008 19:49:31.304563 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Oct 08 19:49:31 crc kubenswrapper[4988]: I1008 19:49:31.401943 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/667c78af-4a60-4271-88c6-ef5ac33982a6-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"667c78af-4a60-4271-88c6-ef5ac33982a6\") " pod="openstack/nova-cell0-conductor-0" Oct 08 19:49:31 crc kubenswrapper[4988]: I1008 19:49:31.402124 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/667c78af-4a60-4271-88c6-ef5ac33982a6-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"667c78af-4a60-4271-88c6-ef5ac33982a6\") " pod="openstack/nova-cell0-conductor-0" Oct 08 19:49:31 crc kubenswrapper[4988]: I1008 19:49:31.402589 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wrwf7\" (UniqueName: \"kubernetes.io/projected/667c78af-4a60-4271-88c6-ef5ac33982a6-kube-api-access-wrwf7\") pod \"nova-cell0-conductor-0\" (UID: \"667c78af-4a60-4271-88c6-ef5ac33982a6\") " pod="openstack/nova-cell0-conductor-0" Oct 08 19:49:31 crc kubenswrapper[4988]: I1008 19:49:31.504738 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/667c78af-4a60-4271-88c6-ef5ac33982a6-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"667c78af-4a60-4271-88c6-ef5ac33982a6\") " pod="openstack/nova-cell0-conductor-0" Oct 08 19:49:31 crc kubenswrapper[4988]: I1008 19:49:31.504835 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/667c78af-4a60-4271-88c6-ef5ac33982a6-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"667c78af-4a60-4271-88c6-ef5ac33982a6\") " pod="openstack/nova-cell0-conductor-0" Oct 08 19:49:31 crc kubenswrapper[4988]: I1008 19:49:31.504924 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wrwf7\" (UniqueName: \"kubernetes.io/projected/667c78af-4a60-4271-88c6-ef5ac33982a6-kube-api-access-wrwf7\") pod \"nova-cell0-conductor-0\" (UID: \"667c78af-4a60-4271-88c6-ef5ac33982a6\") " pod="openstack/nova-cell0-conductor-0" Oct 08 19:49:31 crc kubenswrapper[4988]: I1008 19:49:31.511201 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/667c78af-4a60-4271-88c6-ef5ac33982a6-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"667c78af-4a60-4271-88c6-ef5ac33982a6\") " pod="openstack/nova-cell0-conductor-0" Oct 08 19:49:31 crc kubenswrapper[4988]: I1008 19:49:31.511855 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/667c78af-4a60-4271-88c6-ef5ac33982a6-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"667c78af-4a60-4271-88c6-ef5ac33982a6\") " pod="openstack/nova-cell0-conductor-0" Oct 08 19:49:31 crc kubenswrapper[4988]: I1008 19:49:31.536559 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wrwf7\" (UniqueName: \"kubernetes.io/projected/667c78af-4a60-4271-88c6-ef5ac33982a6-kube-api-access-wrwf7\") pod \"nova-cell0-conductor-0\" (UID: \"667c78af-4a60-4271-88c6-ef5ac33982a6\") " pod="openstack/nova-cell0-conductor-0" Oct 08 19:49:31 crc kubenswrapper[4988]: I1008 19:49:31.602987 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Oct 08 19:49:32 crc kubenswrapper[4988]: I1008 19:49:32.184907 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Oct 08 19:49:32 crc kubenswrapper[4988]: I1008 19:49:32.208985 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"667c78af-4a60-4271-88c6-ef5ac33982a6","Type":"ContainerStarted","Data":"cc8d4a55b18c04765f9700f188f9b9acc26b4e0b1269759d871dc3ad10d53413"} Oct 08 19:49:32 crc kubenswrapper[4988]: I1008 19:49:32.238516 4988 scope.go:117] "RemoveContainer" containerID="f2a29ed22dd7a48bbc929e65fbeceed969a669e78f65fbfb724b03dc0db6e947" Oct 08 19:49:32 crc kubenswrapper[4988]: E1008 19:49:32.238970 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nm54f_openshift-machine-config-operator(f328a20b-5283-407e-8a93-2ab2da599f18)\"" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" Oct 08 19:49:33 crc kubenswrapper[4988]: I1008 19:49:33.263819 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"667c78af-4a60-4271-88c6-ef5ac33982a6","Type":"ContainerStarted","Data":"c2884809446b412404e04b0517c34a161c686d7e35b437b35238d006fdbe1300"} Oct 08 19:49:33 crc kubenswrapper[4988]: I1008 19:49:33.264221 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell0-conductor-0" Oct 08 19:49:33 crc kubenswrapper[4988]: I1008 19:49:33.288049 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-0" podStartSLOduration=2.287992477 podStartE2EDuration="2.287992477s" podCreationTimestamp="2025-10-08 19:49:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 19:49:33.277498313 +0000 UTC m=+5918.727341154" watchObservedRunningTime="2025-10-08 19:49:33.287992477 +0000 UTC m=+5918.737835287" Oct 08 19:49:41 crc kubenswrapper[4988]: I1008 19:49:41.641780 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell0-conductor-0" Oct 08 19:49:42 crc kubenswrapper[4988]: I1008 19:49:42.141853 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-cell-mapping-jtj56"] Oct 08 19:49:42 crc kubenswrapper[4988]: I1008 19:49:42.142956 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-jtj56" Oct 08 19:49:42 crc kubenswrapper[4988]: I1008 19:49:42.145512 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-config-data" Oct 08 19:49:42 crc kubenswrapper[4988]: I1008 19:49:42.145916 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-scripts" Oct 08 19:49:42 crc kubenswrapper[4988]: I1008 19:49:42.158306 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-jtj56"] Oct 08 19:49:42 crc kubenswrapper[4988]: I1008 19:49:42.228992 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/586b0be5-17fd-4984-87ba-a8c2862ddbc0-scripts\") pod \"nova-cell0-cell-mapping-jtj56\" (UID: \"586b0be5-17fd-4984-87ba-a8c2862ddbc0\") " pod="openstack/nova-cell0-cell-mapping-jtj56" Oct 08 19:49:42 crc kubenswrapper[4988]: I1008 19:49:42.229375 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2kt7x\" (UniqueName: \"kubernetes.io/projected/586b0be5-17fd-4984-87ba-a8c2862ddbc0-kube-api-access-2kt7x\") pod \"nova-cell0-cell-mapping-jtj56\" (UID: \"586b0be5-17fd-4984-87ba-a8c2862ddbc0\") " pod="openstack/nova-cell0-cell-mapping-jtj56" Oct 08 19:49:42 crc kubenswrapper[4988]: I1008 19:49:42.229748 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/586b0be5-17fd-4984-87ba-a8c2862ddbc0-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-jtj56\" (UID: \"586b0be5-17fd-4984-87ba-a8c2862ddbc0\") " pod="openstack/nova-cell0-cell-mapping-jtj56" Oct 08 19:49:42 crc kubenswrapper[4988]: I1008 19:49:42.230350 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/586b0be5-17fd-4984-87ba-a8c2862ddbc0-config-data\") pod \"nova-cell0-cell-mapping-jtj56\" (UID: \"586b0be5-17fd-4984-87ba-a8c2862ddbc0\") " pod="openstack/nova-cell0-cell-mapping-jtj56" Oct 08 19:49:42 crc kubenswrapper[4988]: I1008 19:49:42.325068 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 08 19:49:42 crc kubenswrapper[4988]: I1008 19:49:42.326171 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Oct 08 19:49:42 crc kubenswrapper[4988]: I1008 19:49:42.329487 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Oct 08 19:49:42 crc kubenswrapper[4988]: I1008 19:49:42.331827 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2kt7x\" (UniqueName: \"kubernetes.io/projected/586b0be5-17fd-4984-87ba-a8c2862ddbc0-kube-api-access-2kt7x\") pod \"nova-cell0-cell-mapping-jtj56\" (UID: \"586b0be5-17fd-4984-87ba-a8c2862ddbc0\") " pod="openstack/nova-cell0-cell-mapping-jtj56" Oct 08 19:49:42 crc kubenswrapper[4988]: I1008 19:49:42.331909 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/586b0be5-17fd-4984-87ba-a8c2862ddbc0-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-jtj56\" (UID: \"586b0be5-17fd-4984-87ba-a8c2862ddbc0\") " pod="openstack/nova-cell0-cell-mapping-jtj56" Oct 08 19:49:42 crc kubenswrapper[4988]: I1008 19:49:42.331981 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/586b0be5-17fd-4984-87ba-a8c2862ddbc0-config-data\") pod \"nova-cell0-cell-mapping-jtj56\" (UID: \"586b0be5-17fd-4984-87ba-a8c2862ddbc0\") " pod="openstack/nova-cell0-cell-mapping-jtj56" Oct 08 19:49:42 crc kubenswrapper[4988]: I1008 19:49:42.332077 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/586b0be5-17fd-4984-87ba-a8c2862ddbc0-scripts\") pod \"nova-cell0-cell-mapping-jtj56\" (UID: \"586b0be5-17fd-4984-87ba-a8c2862ddbc0\") " pod="openstack/nova-cell0-cell-mapping-jtj56" Oct 08 19:49:42 crc kubenswrapper[4988]: I1008 19:49:42.340245 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/586b0be5-17fd-4984-87ba-a8c2862ddbc0-scripts\") pod \"nova-cell0-cell-mapping-jtj56\" (UID: \"586b0be5-17fd-4984-87ba-a8c2862ddbc0\") " pod="openstack/nova-cell0-cell-mapping-jtj56" Oct 08 19:49:42 crc kubenswrapper[4988]: I1008 19:49:42.343668 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/586b0be5-17fd-4984-87ba-a8c2862ddbc0-config-data\") pod \"nova-cell0-cell-mapping-jtj56\" (UID: \"586b0be5-17fd-4984-87ba-a8c2862ddbc0\") " pod="openstack/nova-cell0-cell-mapping-jtj56" Oct 08 19:49:42 crc kubenswrapper[4988]: I1008 19:49:42.354931 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/586b0be5-17fd-4984-87ba-a8c2862ddbc0-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-jtj56\" (UID: \"586b0be5-17fd-4984-87ba-a8c2862ddbc0\") " pod="openstack/nova-cell0-cell-mapping-jtj56" Oct 08 19:49:42 crc kubenswrapper[4988]: I1008 19:49:42.357005 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 08 19:49:42 crc kubenswrapper[4988]: I1008 19:49:42.362855 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2kt7x\" (UniqueName: \"kubernetes.io/projected/586b0be5-17fd-4984-87ba-a8c2862ddbc0-kube-api-access-2kt7x\") pod \"nova-cell0-cell-mapping-jtj56\" (UID: \"586b0be5-17fd-4984-87ba-a8c2862ddbc0\") " pod="openstack/nova-cell0-cell-mapping-jtj56" Oct 08 19:49:42 crc kubenswrapper[4988]: I1008 19:49:42.417705 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Oct 08 19:49:42 crc kubenswrapper[4988]: I1008 19:49:42.425337 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 08 19:49:42 crc kubenswrapper[4988]: I1008 19:49:42.432745 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Oct 08 19:49:42 crc kubenswrapper[4988]: I1008 19:49:42.434255 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6h5hk\" (UniqueName: \"kubernetes.io/projected/225d108a-28ee-4fea-9f01-c95cf08a0413-kube-api-access-6h5hk\") pod \"nova-cell1-novncproxy-0\" (UID: \"225d108a-28ee-4fea-9f01-c95cf08a0413\") " pod="openstack/nova-cell1-novncproxy-0" Oct 08 19:49:42 crc kubenswrapper[4988]: I1008 19:49:42.434307 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/225d108a-28ee-4fea-9f01-c95cf08a0413-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"225d108a-28ee-4fea-9f01-c95cf08a0413\") " pod="openstack/nova-cell1-novncproxy-0" Oct 08 19:49:42 crc kubenswrapper[4988]: I1008 19:49:42.434353 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/225d108a-28ee-4fea-9f01-c95cf08a0413-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"225d108a-28ee-4fea-9f01-c95cf08a0413\") " pod="openstack/nova-cell1-novncproxy-0" Oct 08 19:49:42 crc kubenswrapper[4988]: I1008 19:49:42.440788 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 08 19:49:42 crc kubenswrapper[4988]: I1008 19:49:42.526931 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Oct 08 19:49:42 crc kubenswrapper[4988]: I1008 19:49:42.528192 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 08 19:49:42 crc kubenswrapper[4988]: I1008 19:49:42.534992 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Oct 08 19:49:42 crc kubenswrapper[4988]: I1008 19:49:42.536133 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6h5hk\" (UniqueName: \"kubernetes.io/projected/225d108a-28ee-4fea-9f01-c95cf08a0413-kube-api-access-6h5hk\") pod \"nova-cell1-novncproxy-0\" (UID: \"225d108a-28ee-4fea-9f01-c95cf08a0413\") " pod="openstack/nova-cell1-novncproxy-0" Oct 08 19:49:42 crc kubenswrapper[4988]: I1008 19:49:42.536170 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g47kl\" (UniqueName: \"kubernetes.io/projected/b354976f-9d09-4023-a869-d373f9717585-kube-api-access-g47kl\") pod \"nova-metadata-0\" (UID: \"b354976f-9d09-4023-a869-d373f9717585\") " pod="openstack/nova-metadata-0" Oct 08 19:49:42 crc kubenswrapper[4988]: I1008 19:49:42.536220 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/225d108a-28ee-4fea-9f01-c95cf08a0413-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"225d108a-28ee-4fea-9f01-c95cf08a0413\") " pod="openstack/nova-cell1-novncproxy-0" Oct 08 19:49:42 crc kubenswrapper[4988]: I1008 19:49:42.536260 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b354976f-9d09-4023-a869-d373f9717585-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"b354976f-9d09-4023-a869-d373f9717585\") " pod="openstack/nova-metadata-0" Oct 08 19:49:42 crc kubenswrapper[4988]: I1008 19:49:42.536286 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/225d108a-28ee-4fea-9f01-c95cf08a0413-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"225d108a-28ee-4fea-9f01-c95cf08a0413\") " pod="openstack/nova-cell1-novncproxy-0" Oct 08 19:49:42 crc kubenswrapper[4988]: I1008 19:49:42.536333 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b354976f-9d09-4023-a869-d373f9717585-logs\") pod \"nova-metadata-0\" (UID: \"b354976f-9d09-4023-a869-d373f9717585\") " pod="openstack/nova-metadata-0" Oct 08 19:49:42 crc kubenswrapper[4988]: I1008 19:49:42.536363 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b354976f-9d09-4023-a869-d373f9717585-config-data\") pod \"nova-metadata-0\" (UID: \"b354976f-9d09-4023-a869-d373f9717585\") " pod="openstack/nova-metadata-0" Oct 08 19:49:42 crc kubenswrapper[4988]: I1008 19:49:42.542502 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-jtj56" Oct 08 19:49:42 crc kubenswrapper[4988]: I1008 19:49:42.542813 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/225d108a-28ee-4fea-9f01-c95cf08a0413-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"225d108a-28ee-4fea-9f01-c95cf08a0413\") " pod="openstack/nova-cell1-novncproxy-0" Oct 08 19:49:42 crc kubenswrapper[4988]: I1008 19:49:42.543821 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/225d108a-28ee-4fea-9f01-c95cf08a0413-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"225d108a-28ee-4fea-9f01-c95cf08a0413\") " pod="openstack/nova-cell1-novncproxy-0" Oct 08 19:49:42 crc kubenswrapper[4988]: I1008 19:49:42.549682 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Oct 08 19:49:42 crc kubenswrapper[4988]: I1008 19:49:42.557454 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5bf7c8d69c-j8v4t"] Oct 08 19:49:42 crc kubenswrapper[4988]: I1008 19:49:42.559046 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5bf7c8d69c-j8v4t" Oct 08 19:49:42 crc kubenswrapper[4988]: I1008 19:49:42.569051 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Oct 08 19:49:42 crc kubenswrapper[4988]: I1008 19:49:42.570416 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 08 19:49:42 crc kubenswrapper[4988]: I1008 19:49:42.578698 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Oct 08 19:49:42 crc kubenswrapper[4988]: I1008 19:49:42.581841 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6h5hk\" (UniqueName: \"kubernetes.io/projected/225d108a-28ee-4fea-9f01-c95cf08a0413-kube-api-access-6h5hk\") pod \"nova-cell1-novncproxy-0\" (UID: \"225d108a-28ee-4fea-9f01-c95cf08a0413\") " pod="openstack/nova-cell1-novncproxy-0" Oct 08 19:49:42 crc kubenswrapper[4988]: I1008 19:49:42.599487 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 08 19:49:42 crc kubenswrapper[4988]: I1008 19:49:42.637738 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/984822c7-f554-4c5b-bba1-3cd471eee0f3-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"984822c7-f554-4c5b-bba1-3cd471eee0f3\") " pod="openstack/nova-api-0" Oct 08 19:49:42 crc kubenswrapper[4988]: I1008 19:49:42.637795 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ee5a3e37-5a2f-4af4-9d64-70444901144a-dns-svc\") pod \"dnsmasq-dns-5bf7c8d69c-j8v4t\" (UID: \"ee5a3e37-5a2f-4af4-9d64-70444901144a\") " pod="openstack/dnsmasq-dns-5bf7c8d69c-j8v4t" Oct 08 19:49:42 crc kubenswrapper[4988]: I1008 19:49:42.637824 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g47kl\" (UniqueName: \"kubernetes.io/projected/b354976f-9d09-4023-a869-d373f9717585-kube-api-access-g47kl\") pod \"nova-metadata-0\" (UID: \"b354976f-9d09-4023-a869-d373f9717585\") " pod="openstack/nova-metadata-0" Oct 08 19:49:42 crc kubenswrapper[4988]: I1008 19:49:42.637849 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ee5a3e37-5a2f-4af4-9d64-70444901144a-config\") pod \"dnsmasq-dns-5bf7c8d69c-j8v4t\" (UID: \"ee5a3e37-5a2f-4af4-9d64-70444901144a\") " pod="openstack/dnsmasq-dns-5bf7c8d69c-j8v4t" Oct 08 19:49:42 crc kubenswrapper[4988]: I1008 19:49:42.637884 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kl9pp\" (UniqueName: \"kubernetes.io/projected/984822c7-f554-4c5b-bba1-3cd471eee0f3-kube-api-access-kl9pp\") pod \"nova-api-0\" (UID: \"984822c7-f554-4c5b-bba1-3cd471eee0f3\") " pod="openstack/nova-api-0" Oct 08 19:49:42 crc kubenswrapper[4988]: I1008 19:49:42.637921 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/984822c7-f554-4c5b-bba1-3cd471eee0f3-config-data\") pod \"nova-api-0\" (UID: \"984822c7-f554-4c5b-bba1-3cd471eee0f3\") " pod="openstack/nova-api-0" Oct 08 19:49:42 crc kubenswrapper[4988]: I1008 19:49:42.637939 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b354976f-9d09-4023-a869-d373f9717585-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"b354976f-9d09-4023-a869-d373f9717585\") " pod="openstack/nova-metadata-0" Oct 08 19:49:42 crc kubenswrapper[4988]: I1008 19:49:42.637959 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ee5a3e37-5a2f-4af4-9d64-70444901144a-ovsdbserver-nb\") pod \"dnsmasq-dns-5bf7c8d69c-j8v4t\" (UID: \"ee5a3e37-5a2f-4af4-9d64-70444901144a\") " pod="openstack/dnsmasq-dns-5bf7c8d69c-j8v4t" Oct 08 19:49:42 crc kubenswrapper[4988]: I1008 19:49:42.637979 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2ceb61b1-8b93-4d44-b27f-ac96c4566be6-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"2ceb61b1-8b93-4d44-b27f-ac96c4566be6\") " pod="openstack/nova-scheduler-0" Oct 08 19:49:42 crc kubenswrapper[4988]: I1008 19:49:42.638000 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2ceb61b1-8b93-4d44-b27f-ac96c4566be6-config-data\") pod \"nova-scheduler-0\" (UID: \"2ceb61b1-8b93-4d44-b27f-ac96c4566be6\") " pod="openstack/nova-scheduler-0" Oct 08 19:49:42 crc kubenswrapper[4988]: I1008 19:49:42.638036 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ee5a3e37-5a2f-4af4-9d64-70444901144a-ovsdbserver-sb\") pod \"dnsmasq-dns-5bf7c8d69c-j8v4t\" (UID: \"ee5a3e37-5a2f-4af4-9d64-70444901144a\") " pod="openstack/dnsmasq-dns-5bf7c8d69c-j8v4t" Oct 08 19:49:42 crc kubenswrapper[4988]: I1008 19:49:42.638562 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b354976f-9d09-4023-a869-d373f9717585-logs\") pod \"nova-metadata-0\" (UID: \"b354976f-9d09-4023-a869-d373f9717585\") " pod="openstack/nova-metadata-0" Oct 08 19:49:42 crc kubenswrapper[4988]: I1008 19:49:42.638633 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/984822c7-f554-4c5b-bba1-3cd471eee0f3-logs\") pod \"nova-api-0\" (UID: \"984822c7-f554-4c5b-bba1-3cd471eee0f3\") " pod="openstack/nova-api-0" Oct 08 19:49:42 crc kubenswrapper[4988]: I1008 19:49:42.638682 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b354976f-9d09-4023-a869-d373f9717585-config-data\") pod \"nova-metadata-0\" (UID: \"b354976f-9d09-4023-a869-d373f9717585\") " pod="openstack/nova-metadata-0" Oct 08 19:49:42 crc kubenswrapper[4988]: I1008 19:49:42.638744 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rrmn2\" (UniqueName: \"kubernetes.io/projected/2ceb61b1-8b93-4d44-b27f-ac96c4566be6-kube-api-access-rrmn2\") pod \"nova-scheduler-0\" (UID: \"2ceb61b1-8b93-4d44-b27f-ac96c4566be6\") " pod="openstack/nova-scheduler-0" Oct 08 19:49:42 crc kubenswrapper[4988]: I1008 19:49:42.638863 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hl89n\" (UniqueName: \"kubernetes.io/projected/ee5a3e37-5a2f-4af4-9d64-70444901144a-kube-api-access-hl89n\") pod \"dnsmasq-dns-5bf7c8d69c-j8v4t\" (UID: \"ee5a3e37-5a2f-4af4-9d64-70444901144a\") " pod="openstack/dnsmasq-dns-5bf7c8d69c-j8v4t" Oct 08 19:49:42 crc kubenswrapper[4988]: I1008 19:49:42.639419 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b354976f-9d09-4023-a869-d373f9717585-logs\") pod \"nova-metadata-0\" (UID: \"b354976f-9d09-4023-a869-d373f9717585\") " pod="openstack/nova-metadata-0" Oct 08 19:49:42 crc kubenswrapper[4988]: I1008 19:49:42.642214 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b354976f-9d09-4023-a869-d373f9717585-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"b354976f-9d09-4023-a869-d373f9717585\") " pod="openstack/nova-metadata-0" Oct 08 19:49:42 crc kubenswrapper[4988]: I1008 19:49:42.646700 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b354976f-9d09-4023-a869-d373f9717585-config-data\") pod \"nova-metadata-0\" (UID: \"b354976f-9d09-4023-a869-d373f9717585\") " pod="openstack/nova-metadata-0" Oct 08 19:49:42 crc kubenswrapper[4988]: I1008 19:49:42.651463 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5bf7c8d69c-j8v4t"] Oct 08 19:49:42 crc kubenswrapper[4988]: I1008 19:49:42.661002 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g47kl\" (UniqueName: \"kubernetes.io/projected/b354976f-9d09-4023-a869-d373f9717585-kube-api-access-g47kl\") pod \"nova-metadata-0\" (UID: \"b354976f-9d09-4023-a869-d373f9717585\") " pod="openstack/nova-metadata-0" Oct 08 19:49:42 crc kubenswrapper[4988]: I1008 19:49:42.736810 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Oct 08 19:49:42 crc kubenswrapper[4988]: I1008 19:49:42.751988 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kl9pp\" (UniqueName: \"kubernetes.io/projected/984822c7-f554-4c5b-bba1-3cd471eee0f3-kube-api-access-kl9pp\") pod \"nova-api-0\" (UID: \"984822c7-f554-4c5b-bba1-3cd471eee0f3\") " pod="openstack/nova-api-0" Oct 08 19:49:42 crc kubenswrapper[4988]: I1008 19:49:42.752067 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/984822c7-f554-4c5b-bba1-3cd471eee0f3-config-data\") pod \"nova-api-0\" (UID: \"984822c7-f554-4c5b-bba1-3cd471eee0f3\") " pod="openstack/nova-api-0" Oct 08 19:49:42 crc kubenswrapper[4988]: I1008 19:49:42.752098 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ee5a3e37-5a2f-4af4-9d64-70444901144a-ovsdbserver-nb\") pod \"dnsmasq-dns-5bf7c8d69c-j8v4t\" (UID: \"ee5a3e37-5a2f-4af4-9d64-70444901144a\") " pod="openstack/dnsmasq-dns-5bf7c8d69c-j8v4t" Oct 08 19:49:42 crc kubenswrapper[4988]: I1008 19:49:42.752119 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2ceb61b1-8b93-4d44-b27f-ac96c4566be6-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"2ceb61b1-8b93-4d44-b27f-ac96c4566be6\") " pod="openstack/nova-scheduler-0" Oct 08 19:49:42 crc kubenswrapper[4988]: I1008 19:49:42.752140 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2ceb61b1-8b93-4d44-b27f-ac96c4566be6-config-data\") pod \"nova-scheduler-0\" (UID: \"2ceb61b1-8b93-4d44-b27f-ac96c4566be6\") " pod="openstack/nova-scheduler-0" Oct 08 19:49:42 crc kubenswrapper[4988]: I1008 19:49:42.752190 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ee5a3e37-5a2f-4af4-9d64-70444901144a-ovsdbserver-sb\") pod \"dnsmasq-dns-5bf7c8d69c-j8v4t\" (UID: \"ee5a3e37-5a2f-4af4-9d64-70444901144a\") " pod="openstack/dnsmasq-dns-5bf7c8d69c-j8v4t" Oct 08 19:49:42 crc kubenswrapper[4988]: I1008 19:49:42.752230 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/984822c7-f554-4c5b-bba1-3cd471eee0f3-logs\") pod \"nova-api-0\" (UID: \"984822c7-f554-4c5b-bba1-3cd471eee0f3\") " pod="openstack/nova-api-0" Oct 08 19:49:42 crc kubenswrapper[4988]: I1008 19:49:42.752264 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rrmn2\" (UniqueName: \"kubernetes.io/projected/2ceb61b1-8b93-4d44-b27f-ac96c4566be6-kube-api-access-rrmn2\") pod \"nova-scheduler-0\" (UID: \"2ceb61b1-8b93-4d44-b27f-ac96c4566be6\") " pod="openstack/nova-scheduler-0" Oct 08 19:49:42 crc kubenswrapper[4988]: I1008 19:49:42.752307 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hl89n\" (UniqueName: \"kubernetes.io/projected/ee5a3e37-5a2f-4af4-9d64-70444901144a-kube-api-access-hl89n\") pod \"dnsmasq-dns-5bf7c8d69c-j8v4t\" (UID: \"ee5a3e37-5a2f-4af4-9d64-70444901144a\") " pod="openstack/dnsmasq-dns-5bf7c8d69c-j8v4t" Oct 08 19:49:42 crc kubenswrapper[4988]: I1008 19:49:42.752337 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/984822c7-f554-4c5b-bba1-3cd471eee0f3-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"984822c7-f554-4c5b-bba1-3cd471eee0f3\") " pod="openstack/nova-api-0" Oct 08 19:49:42 crc kubenswrapper[4988]: I1008 19:49:42.752362 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ee5a3e37-5a2f-4af4-9d64-70444901144a-dns-svc\") pod \"dnsmasq-dns-5bf7c8d69c-j8v4t\" (UID: \"ee5a3e37-5a2f-4af4-9d64-70444901144a\") " pod="openstack/dnsmasq-dns-5bf7c8d69c-j8v4t" Oct 08 19:49:42 crc kubenswrapper[4988]: I1008 19:49:42.752415 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ee5a3e37-5a2f-4af4-9d64-70444901144a-config\") pod \"dnsmasq-dns-5bf7c8d69c-j8v4t\" (UID: \"ee5a3e37-5a2f-4af4-9d64-70444901144a\") " pod="openstack/dnsmasq-dns-5bf7c8d69c-j8v4t" Oct 08 19:49:42 crc kubenswrapper[4988]: I1008 19:49:42.752953 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/984822c7-f554-4c5b-bba1-3cd471eee0f3-logs\") pod \"nova-api-0\" (UID: \"984822c7-f554-4c5b-bba1-3cd471eee0f3\") " pod="openstack/nova-api-0" Oct 08 19:49:42 crc kubenswrapper[4988]: I1008 19:49:42.752980 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ee5a3e37-5a2f-4af4-9d64-70444901144a-ovsdbserver-nb\") pod \"dnsmasq-dns-5bf7c8d69c-j8v4t\" (UID: \"ee5a3e37-5a2f-4af4-9d64-70444901144a\") " pod="openstack/dnsmasq-dns-5bf7c8d69c-j8v4t" Oct 08 19:49:42 crc kubenswrapper[4988]: I1008 19:49:42.753119 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ee5a3e37-5a2f-4af4-9d64-70444901144a-config\") pod \"dnsmasq-dns-5bf7c8d69c-j8v4t\" (UID: \"ee5a3e37-5a2f-4af4-9d64-70444901144a\") " pod="openstack/dnsmasq-dns-5bf7c8d69c-j8v4t" Oct 08 19:49:42 crc kubenswrapper[4988]: I1008 19:49:42.756279 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ee5a3e37-5a2f-4af4-9d64-70444901144a-dns-svc\") pod \"dnsmasq-dns-5bf7c8d69c-j8v4t\" (UID: \"ee5a3e37-5a2f-4af4-9d64-70444901144a\") " pod="openstack/dnsmasq-dns-5bf7c8d69c-j8v4t" Oct 08 19:49:42 crc kubenswrapper[4988]: I1008 19:49:42.757167 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ee5a3e37-5a2f-4af4-9d64-70444901144a-ovsdbserver-sb\") pod \"dnsmasq-dns-5bf7c8d69c-j8v4t\" (UID: \"ee5a3e37-5a2f-4af4-9d64-70444901144a\") " pod="openstack/dnsmasq-dns-5bf7c8d69c-j8v4t" Oct 08 19:49:42 crc kubenswrapper[4988]: I1008 19:49:42.757613 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 08 19:49:42 crc kubenswrapper[4988]: I1008 19:49:42.765218 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2ceb61b1-8b93-4d44-b27f-ac96c4566be6-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"2ceb61b1-8b93-4d44-b27f-ac96c4566be6\") " pod="openstack/nova-scheduler-0" Oct 08 19:49:42 crc kubenswrapper[4988]: I1008 19:49:42.765275 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2ceb61b1-8b93-4d44-b27f-ac96c4566be6-config-data\") pod \"nova-scheduler-0\" (UID: \"2ceb61b1-8b93-4d44-b27f-ac96c4566be6\") " pod="openstack/nova-scheduler-0" Oct 08 19:49:42 crc kubenswrapper[4988]: I1008 19:49:42.766592 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/984822c7-f554-4c5b-bba1-3cd471eee0f3-config-data\") pod \"nova-api-0\" (UID: \"984822c7-f554-4c5b-bba1-3cd471eee0f3\") " pod="openstack/nova-api-0" Oct 08 19:49:42 crc kubenswrapper[4988]: I1008 19:49:42.769570 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rrmn2\" (UniqueName: \"kubernetes.io/projected/2ceb61b1-8b93-4d44-b27f-ac96c4566be6-kube-api-access-rrmn2\") pod \"nova-scheduler-0\" (UID: \"2ceb61b1-8b93-4d44-b27f-ac96c4566be6\") " pod="openstack/nova-scheduler-0" Oct 08 19:49:42 crc kubenswrapper[4988]: I1008 19:49:42.771246 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kl9pp\" (UniqueName: \"kubernetes.io/projected/984822c7-f554-4c5b-bba1-3cd471eee0f3-kube-api-access-kl9pp\") pod \"nova-api-0\" (UID: \"984822c7-f554-4c5b-bba1-3cd471eee0f3\") " pod="openstack/nova-api-0" Oct 08 19:49:42 crc kubenswrapper[4988]: I1008 19:49:42.773196 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/984822c7-f554-4c5b-bba1-3cd471eee0f3-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"984822c7-f554-4c5b-bba1-3cd471eee0f3\") " pod="openstack/nova-api-0" Oct 08 19:49:42 crc kubenswrapper[4988]: I1008 19:49:42.779711 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hl89n\" (UniqueName: \"kubernetes.io/projected/ee5a3e37-5a2f-4af4-9d64-70444901144a-kube-api-access-hl89n\") pod \"dnsmasq-dns-5bf7c8d69c-j8v4t\" (UID: \"ee5a3e37-5a2f-4af4-9d64-70444901144a\") " pod="openstack/dnsmasq-dns-5bf7c8d69c-j8v4t" Oct 08 19:49:42 crc kubenswrapper[4988]: I1008 19:49:42.996693 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 08 19:49:43 crc kubenswrapper[4988]: I1008 19:49:43.007472 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5bf7c8d69c-j8v4t" Oct 08 19:49:43 crc kubenswrapper[4988]: I1008 19:49:43.030108 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 08 19:49:43 crc kubenswrapper[4988]: I1008 19:49:43.095960 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-jtj56"] Oct 08 19:49:43 crc kubenswrapper[4988]: I1008 19:49:43.145741 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 08 19:49:43 crc kubenswrapper[4988]: W1008 19:49:43.166769 4988 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb354976f_9d09_4023_a869_d373f9717585.slice/crio-407616015a59726cfadc2992428cbb8518590b7fffd80151d8be99d907b7b16e WatchSource:0}: Error finding container 407616015a59726cfadc2992428cbb8518590b7fffd80151d8be99d907b7b16e: Status 404 returned error can't find the container with id 407616015a59726cfadc2992428cbb8518590b7fffd80151d8be99d907b7b16e Oct 08 19:49:43 crc kubenswrapper[4988]: I1008 19:49:43.271315 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-db-sync-tmx75"] Oct 08 19:49:43 crc kubenswrapper[4988]: I1008 19:49:43.272583 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-tmx75"] Oct 08 19:49:43 crc kubenswrapper[4988]: I1008 19:49:43.272660 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-tmx75" Oct 08 19:49:43 crc kubenswrapper[4988]: I1008 19:49:43.272976 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 08 19:49:43 crc kubenswrapper[4988]: I1008 19:49:43.275674 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-scripts" Oct 08 19:49:43 crc kubenswrapper[4988]: I1008 19:49:43.275846 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Oct 08 19:49:43 crc kubenswrapper[4988]: I1008 19:49:43.351841 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-jtj56" event={"ID":"586b0be5-17fd-4984-87ba-a8c2862ddbc0","Type":"ContainerStarted","Data":"2b6ba334e548c9ad6b05c7e939f8f084e178ea5d7ab3d7a4a27bc3c0d7c5b3ef"} Oct 08 19:49:43 crc kubenswrapper[4988]: I1008 19:49:43.352722 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"b354976f-9d09-4023-a869-d373f9717585","Type":"ContainerStarted","Data":"407616015a59726cfadc2992428cbb8518590b7fffd80151d8be99d907b7b16e"} Oct 08 19:49:43 crc kubenswrapper[4988]: I1008 19:49:43.354344 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"225d108a-28ee-4fea-9f01-c95cf08a0413","Type":"ContainerStarted","Data":"3eb885a68ffa63b5ac9af98ef09790cdc37c41f9b37458743cfd06ab0daa82ef"} Oct 08 19:49:43 crc kubenswrapper[4988]: I1008 19:49:43.369382 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f6f69dea-85ca-4d7e-9f1c-05f11cadd3b9-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-tmx75\" (UID: \"f6f69dea-85ca-4d7e-9f1c-05f11cadd3b9\") " pod="openstack/nova-cell1-conductor-db-sync-tmx75" Oct 08 19:49:43 crc kubenswrapper[4988]: I1008 19:49:43.369432 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9fgsh\" (UniqueName: \"kubernetes.io/projected/f6f69dea-85ca-4d7e-9f1c-05f11cadd3b9-kube-api-access-9fgsh\") pod \"nova-cell1-conductor-db-sync-tmx75\" (UID: \"f6f69dea-85ca-4d7e-9f1c-05f11cadd3b9\") " pod="openstack/nova-cell1-conductor-db-sync-tmx75" Oct 08 19:49:43 crc kubenswrapper[4988]: I1008 19:49:43.369459 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f6f69dea-85ca-4d7e-9f1c-05f11cadd3b9-scripts\") pod \"nova-cell1-conductor-db-sync-tmx75\" (UID: \"f6f69dea-85ca-4d7e-9f1c-05f11cadd3b9\") " pod="openstack/nova-cell1-conductor-db-sync-tmx75" Oct 08 19:49:43 crc kubenswrapper[4988]: I1008 19:49:43.369493 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f6f69dea-85ca-4d7e-9f1c-05f11cadd3b9-config-data\") pod \"nova-cell1-conductor-db-sync-tmx75\" (UID: \"f6f69dea-85ca-4d7e-9f1c-05f11cadd3b9\") " pod="openstack/nova-cell1-conductor-db-sync-tmx75" Oct 08 19:49:43 crc kubenswrapper[4988]: I1008 19:49:43.470663 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f6f69dea-85ca-4d7e-9f1c-05f11cadd3b9-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-tmx75\" (UID: \"f6f69dea-85ca-4d7e-9f1c-05f11cadd3b9\") " pod="openstack/nova-cell1-conductor-db-sync-tmx75" Oct 08 19:49:43 crc kubenswrapper[4988]: I1008 19:49:43.470710 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9fgsh\" (UniqueName: \"kubernetes.io/projected/f6f69dea-85ca-4d7e-9f1c-05f11cadd3b9-kube-api-access-9fgsh\") pod \"nova-cell1-conductor-db-sync-tmx75\" (UID: \"f6f69dea-85ca-4d7e-9f1c-05f11cadd3b9\") " pod="openstack/nova-cell1-conductor-db-sync-tmx75" Oct 08 19:49:43 crc kubenswrapper[4988]: I1008 19:49:43.470738 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f6f69dea-85ca-4d7e-9f1c-05f11cadd3b9-scripts\") pod \"nova-cell1-conductor-db-sync-tmx75\" (UID: \"f6f69dea-85ca-4d7e-9f1c-05f11cadd3b9\") " pod="openstack/nova-cell1-conductor-db-sync-tmx75" Oct 08 19:49:43 crc kubenswrapper[4988]: I1008 19:49:43.470772 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f6f69dea-85ca-4d7e-9f1c-05f11cadd3b9-config-data\") pod \"nova-cell1-conductor-db-sync-tmx75\" (UID: \"f6f69dea-85ca-4d7e-9f1c-05f11cadd3b9\") " pod="openstack/nova-cell1-conductor-db-sync-tmx75" Oct 08 19:49:43 crc kubenswrapper[4988]: I1008 19:49:43.477753 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f6f69dea-85ca-4d7e-9f1c-05f11cadd3b9-scripts\") pod \"nova-cell1-conductor-db-sync-tmx75\" (UID: \"f6f69dea-85ca-4d7e-9f1c-05f11cadd3b9\") " pod="openstack/nova-cell1-conductor-db-sync-tmx75" Oct 08 19:49:43 crc kubenswrapper[4988]: I1008 19:49:43.478208 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f6f69dea-85ca-4d7e-9f1c-05f11cadd3b9-config-data\") pod \"nova-cell1-conductor-db-sync-tmx75\" (UID: \"f6f69dea-85ca-4d7e-9f1c-05f11cadd3b9\") " pod="openstack/nova-cell1-conductor-db-sync-tmx75" Oct 08 19:49:43 crc kubenswrapper[4988]: I1008 19:49:43.488003 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f6f69dea-85ca-4d7e-9f1c-05f11cadd3b9-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-tmx75\" (UID: \"f6f69dea-85ca-4d7e-9f1c-05f11cadd3b9\") " pod="openstack/nova-cell1-conductor-db-sync-tmx75" Oct 08 19:49:43 crc kubenswrapper[4988]: I1008 19:49:43.492229 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9fgsh\" (UniqueName: \"kubernetes.io/projected/f6f69dea-85ca-4d7e-9f1c-05f11cadd3b9-kube-api-access-9fgsh\") pod \"nova-cell1-conductor-db-sync-tmx75\" (UID: \"f6f69dea-85ca-4d7e-9f1c-05f11cadd3b9\") " pod="openstack/nova-cell1-conductor-db-sync-tmx75" Oct 08 19:49:43 crc kubenswrapper[4988]: I1008 19:49:43.602184 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5bf7c8d69c-j8v4t"] Oct 08 19:49:43 crc kubenswrapper[4988]: I1008 19:49:43.612150 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-tmx75" Oct 08 19:49:43 crc kubenswrapper[4988]: I1008 19:49:43.616736 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Oct 08 19:49:43 crc kubenswrapper[4988]: I1008 19:49:43.770679 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 08 19:49:44 crc kubenswrapper[4988]: I1008 19:49:44.237374 4988 scope.go:117] "RemoveContainer" containerID="f2a29ed22dd7a48bbc929e65fbeceed969a669e78f65fbfb724b03dc0db6e947" Oct 08 19:49:44 crc kubenswrapper[4988]: E1008 19:49:44.238470 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nm54f_openshift-machine-config-operator(f328a20b-5283-407e-8a93-2ab2da599f18)\"" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" Oct 08 19:49:44 crc kubenswrapper[4988]: I1008 19:49:44.266117 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-tmx75"] Oct 08 19:49:44 crc kubenswrapper[4988]: I1008 19:49:44.397606 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"225d108a-28ee-4fea-9f01-c95cf08a0413","Type":"ContainerStarted","Data":"d477d542c96ea814ef6296fa18119feacedc946338168f73d0dff3014091cc0a"} Oct 08 19:49:44 crc kubenswrapper[4988]: I1008 19:49:44.403970 4988 generic.go:334] "Generic (PLEG): container finished" podID="ee5a3e37-5a2f-4af4-9d64-70444901144a" containerID="adc6bc94734471245a6c79f42ab92cdc1b52cf63f4e217321e573b19c62af828" exitCode=0 Oct 08 19:49:44 crc kubenswrapper[4988]: I1008 19:49:44.404137 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5bf7c8d69c-j8v4t" event={"ID":"ee5a3e37-5a2f-4af4-9d64-70444901144a","Type":"ContainerDied","Data":"adc6bc94734471245a6c79f42ab92cdc1b52cf63f4e217321e573b19c62af828"} Oct 08 19:49:44 crc kubenswrapper[4988]: I1008 19:49:44.404193 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5bf7c8d69c-j8v4t" event={"ID":"ee5a3e37-5a2f-4af4-9d64-70444901144a","Type":"ContainerStarted","Data":"55e3f22e8294f3486ffcd7709ca7dd047dbf01ac2601c28f50c80dda0da31c59"} Oct 08 19:49:44 crc kubenswrapper[4988]: I1008 19:49:44.417364 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-tmx75" event={"ID":"f6f69dea-85ca-4d7e-9f1c-05f11cadd3b9","Type":"ContainerStarted","Data":"9ecf571ea874f003e8866a4166050422888f259d06b800e7586518fedb847cf1"} Oct 08 19:49:44 crc kubenswrapper[4988]: I1008 19:49:44.424156 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=2.424138182 podStartE2EDuration="2.424138182s" podCreationTimestamp="2025-10-08 19:49:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 19:49:44.411919463 +0000 UTC m=+5929.861762254" watchObservedRunningTime="2025-10-08 19:49:44.424138182 +0000 UTC m=+5929.873980952" Oct 08 19:49:44 crc kubenswrapper[4988]: I1008 19:49:44.426172 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"2ceb61b1-8b93-4d44-b27f-ac96c4566be6","Type":"ContainerStarted","Data":"95a43697a44a1358aec0023227996593f0f41fa671f8322b52c7d43b9e0dbc7b"} Oct 08 19:49:44 crc kubenswrapper[4988]: I1008 19:49:44.426224 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"2ceb61b1-8b93-4d44-b27f-ac96c4566be6","Type":"ContainerStarted","Data":"36e84c38650d1dfc479d96ff045e9d44a4924ddd01761ecdd740b8a34e3952b0"} Oct 08 19:49:44 crc kubenswrapper[4988]: I1008 19:49:44.435285 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-jtj56" event={"ID":"586b0be5-17fd-4984-87ba-a8c2862ddbc0","Type":"ContainerStarted","Data":"521bb2202c7abb16ee78ce44796f153e3a190098a143f6b2bcd30634e128b462"} Oct 08 19:49:44 crc kubenswrapper[4988]: I1008 19:49:44.464991 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.4649737 podStartE2EDuration="2.4649737s" podCreationTimestamp="2025-10-08 19:49:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 19:49:44.464124633 +0000 UTC m=+5929.913967403" watchObservedRunningTime="2025-10-08 19:49:44.4649737 +0000 UTC m=+5929.914816470" Oct 08 19:49:44 crc kubenswrapper[4988]: I1008 19:49:44.465777 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"b354976f-9d09-4023-a869-d373f9717585","Type":"ContainerStarted","Data":"540181e64243b98c638416ffa7acbeac7561c2b19121b4f26f23f9bbd19800b4"} Oct 08 19:49:44 crc kubenswrapper[4988]: I1008 19:49:44.465820 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"b354976f-9d09-4023-a869-d373f9717585","Type":"ContainerStarted","Data":"6aceccbda0d681f4b250f907f606350db065a93a521c2a5c166be26d461622bf"} Oct 08 19:49:44 crc kubenswrapper[4988]: I1008 19:49:44.488597 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"984822c7-f554-4c5b-bba1-3cd471eee0f3","Type":"ContainerStarted","Data":"cad4dfedacebaed69682037b44f45eb1f5e008bde623318e3394ee03a8575c82"} Oct 08 19:49:44 crc kubenswrapper[4988]: I1008 19:49:44.488647 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"984822c7-f554-4c5b-bba1-3cd471eee0f3","Type":"ContainerStarted","Data":"657bf93b8bae9ec7afe75fbc48ccfc45ccf9af760218e29cfbce56c270f7347f"} Oct 08 19:49:44 crc kubenswrapper[4988]: I1008 19:49:44.488656 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"984822c7-f554-4c5b-bba1-3cd471eee0f3","Type":"ContainerStarted","Data":"5fd78401f11670490a43fd869a607ae304f9e60a4b2db9d48a8a14541ccbce91"} Oct 08 19:49:44 crc kubenswrapper[4988]: I1008 19:49:44.512624 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-cell-mapping-jtj56" podStartSLOduration=2.5126054939999998 podStartE2EDuration="2.512605494s" podCreationTimestamp="2025-10-08 19:49:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 19:49:44.481721622 +0000 UTC m=+5929.931564392" watchObservedRunningTime="2025-10-08 19:49:44.512605494 +0000 UTC m=+5929.962448264" Oct 08 19:49:44 crc kubenswrapper[4988]: I1008 19:49:44.523802 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.523784059 podStartE2EDuration="2.523784059s" podCreationTimestamp="2025-10-08 19:49:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 19:49:44.501858933 +0000 UTC m=+5929.951701713" watchObservedRunningTime="2025-10-08 19:49:44.523784059 +0000 UTC m=+5929.973626829" Oct 08 19:49:44 crc kubenswrapper[4988]: I1008 19:49:44.551930 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.551912434 podStartE2EDuration="2.551912434s" podCreationTimestamp="2025-10-08 19:49:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 19:49:44.530233765 +0000 UTC m=+5929.980076535" watchObservedRunningTime="2025-10-08 19:49:44.551912434 +0000 UTC m=+5930.001755204" Oct 08 19:49:45 crc kubenswrapper[4988]: I1008 19:49:45.506561 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5bf7c8d69c-j8v4t" event={"ID":"ee5a3e37-5a2f-4af4-9d64-70444901144a","Type":"ContainerStarted","Data":"f29503fc288795cbc96059d6a35444db4b2e4c9402d74d1e3a005c7dc00086af"} Oct 08 19:49:45 crc kubenswrapper[4988]: I1008 19:49:45.506933 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-5bf7c8d69c-j8v4t" Oct 08 19:49:45 crc kubenswrapper[4988]: I1008 19:49:45.511201 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-tmx75" event={"ID":"f6f69dea-85ca-4d7e-9f1c-05f11cadd3b9","Type":"ContainerStarted","Data":"9b8bbc55b526532b8819a7ac1a54a4c1aa8fc31d5cca19ae4663664bf7137162"} Oct 08 19:49:45 crc kubenswrapper[4988]: I1008 19:49:45.528575 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-5bf7c8d69c-j8v4t" podStartSLOduration=3.528559439 podStartE2EDuration="3.528559439s" podCreationTimestamp="2025-10-08 19:49:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 19:49:45.524987385 +0000 UTC m=+5930.974830155" watchObservedRunningTime="2025-10-08 19:49:45.528559439 +0000 UTC m=+5930.978402209" Oct 08 19:49:45 crc kubenswrapper[4988]: I1008 19:49:45.546375 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-db-sync-tmx75" podStartSLOduration=2.546357225 podStartE2EDuration="2.546357225s" podCreationTimestamp="2025-10-08 19:49:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 19:49:45.54494364 +0000 UTC m=+5930.994786430" watchObservedRunningTime="2025-10-08 19:49:45.546357225 +0000 UTC m=+5930.996199995" Oct 08 19:49:46 crc kubenswrapper[4988]: I1008 19:49:46.372314 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Oct 08 19:49:46 crc kubenswrapper[4988]: I1008 19:49:46.378343 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 08 19:49:46 crc kubenswrapper[4988]: I1008 19:49:46.516724 4988 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="b354976f-9d09-4023-a869-d373f9717585" containerName="nova-metadata-log" containerID="cri-o://6aceccbda0d681f4b250f907f606350db065a93a521c2a5c166be26d461622bf" gracePeriod=30 Oct 08 19:49:46 crc kubenswrapper[4988]: I1008 19:49:46.517078 4988 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell1-novncproxy-0" podUID="225d108a-28ee-4fea-9f01-c95cf08a0413" containerName="nova-cell1-novncproxy-novncproxy" containerID="cri-o://d477d542c96ea814ef6296fa18119feacedc946338168f73d0dff3014091cc0a" gracePeriod=30 Oct 08 19:49:46 crc kubenswrapper[4988]: I1008 19:49:46.517078 4988 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="b354976f-9d09-4023-a869-d373f9717585" containerName="nova-metadata-metadata" containerID="cri-o://540181e64243b98c638416ffa7acbeac7561c2b19121b4f26f23f9bbd19800b4" gracePeriod=30 Oct 08 19:49:47 crc kubenswrapper[4988]: I1008 19:49:47.271330 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 08 19:49:47 crc kubenswrapper[4988]: I1008 19:49:47.360572 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b354976f-9d09-4023-a869-d373f9717585-config-data\") pod \"b354976f-9d09-4023-a869-d373f9717585\" (UID: \"b354976f-9d09-4023-a869-d373f9717585\") " Oct 08 19:49:47 crc kubenswrapper[4988]: I1008 19:49:47.360670 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b354976f-9d09-4023-a869-d373f9717585-combined-ca-bundle\") pod \"b354976f-9d09-4023-a869-d373f9717585\" (UID: \"b354976f-9d09-4023-a869-d373f9717585\") " Oct 08 19:49:47 crc kubenswrapper[4988]: I1008 19:49:47.360824 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b354976f-9d09-4023-a869-d373f9717585-logs\") pod \"b354976f-9d09-4023-a869-d373f9717585\" (UID: \"b354976f-9d09-4023-a869-d373f9717585\") " Oct 08 19:49:47 crc kubenswrapper[4988]: I1008 19:49:47.360861 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-g47kl\" (UniqueName: \"kubernetes.io/projected/b354976f-9d09-4023-a869-d373f9717585-kube-api-access-g47kl\") pod \"b354976f-9d09-4023-a869-d373f9717585\" (UID: \"b354976f-9d09-4023-a869-d373f9717585\") " Oct 08 19:49:47 crc kubenswrapper[4988]: I1008 19:49:47.361943 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b354976f-9d09-4023-a869-d373f9717585-logs" (OuterVolumeSpecName: "logs") pod "b354976f-9d09-4023-a869-d373f9717585" (UID: "b354976f-9d09-4023-a869-d373f9717585"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 19:49:47 crc kubenswrapper[4988]: I1008 19:49:47.380902 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b354976f-9d09-4023-a869-d373f9717585-kube-api-access-g47kl" (OuterVolumeSpecName: "kube-api-access-g47kl") pod "b354976f-9d09-4023-a869-d373f9717585" (UID: "b354976f-9d09-4023-a869-d373f9717585"). InnerVolumeSpecName "kube-api-access-g47kl". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 19:49:47 crc kubenswrapper[4988]: I1008 19:49:47.395484 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b354976f-9d09-4023-a869-d373f9717585-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "b354976f-9d09-4023-a869-d373f9717585" (UID: "b354976f-9d09-4023-a869-d373f9717585"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 19:49:47 crc kubenswrapper[4988]: I1008 19:49:47.426352 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b354976f-9d09-4023-a869-d373f9717585-config-data" (OuterVolumeSpecName: "config-data") pod "b354976f-9d09-4023-a869-d373f9717585" (UID: "b354976f-9d09-4023-a869-d373f9717585"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 19:49:47 crc kubenswrapper[4988]: I1008 19:49:47.464582 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Oct 08 19:49:47 crc kubenswrapper[4988]: I1008 19:49:47.467458 4988 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-g47kl\" (UniqueName: \"kubernetes.io/projected/b354976f-9d09-4023-a869-d373f9717585-kube-api-access-g47kl\") on node \"crc\" DevicePath \"\"" Oct 08 19:49:47 crc kubenswrapper[4988]: I1008 19:49:47.467495 4988 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b354976f-9d09-4023-a869-d373f9717585-config-data\") on node \"crc\" DevicePath \"\"" Oct 08 19:49:47 crc kubenswrapper[4988]: I1008 19:49:47.467506 4988 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b354976f-9d09-4023-a869-d373f9717585-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 19:49:47 crc kubenswrapper[4988]: I1008 19:49:47.467514 4988 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b354976f-9d09-4023-a869-d373f9717585-logs\") on node \"crc\" DevicePath \"\"" Oct 08 19:49:47 crc kubenswrapper[4988]: I1008 19:49:47.536742 4988 generic.go:334] "Generic (PLEG): container finished" podID="225d108a-28ee-4fea-9f01-c95cf08a0413" containerID="d477d542c96ea814ef6296fa18119feacedc946338168f73d0dff3014091cc0a" exitCode=0 Oct 08 19:49:47 crc kubenswrapper[4988]: I1008 19:49:47.536828 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"225d108a-28ee-4fea-9f01-c95cf08a0413","Type":"ContainerDied","Data":"d477d542c96ea814ef6296fa18119feacedc946338168f73d0dff3014091cc0a"} Oct 08 19:49:47 crc kubenswrapper[4988]: I1008 19:49:47.536868 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"225d108a-28ee-4fea-9f01-c95cf08a0413","Type":"ContainerDied","Data":"3eb885a68ffa63b5ac9af98ef09790cdc37c41f9b37458743cfd06ab0daa82ef"} Oct 08 19:49:47 crc kubenswrapper[4988]: I1008 19:49:47.536895 4988 scope.go:117] "RemoveContainer" containerID="d477d542c96ea814ef6296fa18119feacedc946338168f73d0dff3014091cc0a" Oct 08 19:49:47 crc kubenswrapper[4988]: I1008 19:49:47.537027 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Oct 08 19:49:47 crc kubenswrapper[4988]: I1008 19:49:47.542565 4988 generic.go:334] "Generic (PLEG): container finished" podID="b354976f-9d09-4023-a869-d373f9717585" containerID="540181e64243b98c638416ffa7acbeac7561c2b19121b4f26f23f9bbd19800b4" exitCode=0 Oct 08 19:49:47 crc kubenswrapper[4988]: I1008 19:49:47.542774 4988 generic.go:334] "Generic (PLEG): container finished" podID="b354976f-9d09-4023-a869-d373f9717585" containerID="6aceccbda0d681f4b250f907f606350db065a93a521c2a5c166be26d461622bf" exitCode=143 Oct 08 19:49:47 crc kubenswrapper[4988]: I1008 19:49:47.542861 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"b354976f-9d09-4023-a869-d373f9717585","Type":"ContainerDied","Data":"540181e64243b98c638416ffa7acbeac7561c2b19121b4f26f23f9bbd19800b4"} Oct 08 19:49:47 crc kubenswrapper[4988]: I1008 19:49:47.542973 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"b354976f-9d09-4023-a869-d373f9717585","Type":"ContainerDied","Data":"6aceccbda0d681f4b250f907f606350db065a93a521c2a5c166be26d461622bf"} Oct 08 19:49:47 crc kubenswrapper[4988]: I1008 19:49:47.543048 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"b354976f-9d09-4023-a869-d373f9717585","Type":"ContainerDied","Data":"407616015a59726cfadc2992428cbb8518590b7fffd80151d8be99d907b7b16e"} Oct 08 19:49:47 crc kubenswrapper[4988]: I1008 19:49:47.543164 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 08 19:49:47 crc kubenswrapper[4988]: I1008 19:49:47.558640 4988 scope.go:117] "RemoveContainer" containerID="d477d542c96ea814ef6296fa18119feacedc946338168f73d0dff3014091cc0a" Oct 08 19:49:47 crc kubenswrapper[4988]: E1008 19:49:47.559123 4988 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d477d542c96ea814ef6296fa18119feacedc946338168f73d0dff3014091cc0a\": container with ID starting with d477d542c96ea814ef6296fa18119feacedc946338168f73d0dff3014091cc0a not found: ID does not exist" containerID="d477d542c96ea814ef6296fa18119feacedc946338168f73d0dff3014091cc0a" Oct 08 19:49:47 crc kubenswrapper[4988]: I1008 19:49:47.559153 4988 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d477d542c96ea814ef6296fa18119feacedc946338168f73d0dff3014091cc0a"} err="failed to get container status \"d477d542c96ea814ef6296fa18119feacedc946338168f73d0dff3014091cc0a\": rpc error: code = NotFound desc = could not find container \"d477d542c96ea814ef6296fa18119feacedc946338168f73d0dff3014091cc0a\": container with ID starting with d477d542c96ea814ef6296fa18119feacedc946338168f73d0dff3014091cc0a not found: ID does not exist" Oct 08 19:49:47 crc kubenswrapper[4988]: I1008 19:49:47.559173 4988 scope.go:117] "RemoveContainer" containerID="540181e64243b98c638416ffa7acbeac7561c2b19121b4f26f23f9bbd19800b4" Oct 08 19:49:47 crc kubenswrapper[4988]: I1008 19:49:47.568260 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6h5hk\" (UniqueName: \"kubernetes.io/projected/225d108a-28ee-4fea-9f01-c95cf08a0413-kube-api-access-6h5hk\") pod \"225d108a-28ee-4fea-9f01-c95cf08a0413\" (UID: \"225d108a-28ee-4fea-9f01-c95cf08a0413\") " Oct 08 19:49:47 crc kubenswrapper[4988]: I1008 19:49:47.568299 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/225d108a-28ee-4fea-9f01-c95cf08a0413-config-data\") pod \"225d108a-28ee-4fea-9f01-c95cf08a0413\" (UID: \"225d108a-28ee-4fea-9f01-c95cf08a0413\") " Oct 08 19:49:47 crc kubenswrapper[4988]: I1008 19:49:47.568337 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/225d108a-28ee-4fea-9f01-c95cf08a0413-combined-ca-bundle\") pod \"225d108a-28ee-4fea-9f01-c95cf08a0413\" (UID: \"225d108a-28ee-4fea-9f01-c95cf08a0413\") " Oct 08 19:49:47 crc kubenswrapper[4988]: I1008 19:49:47.571846 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/225d108a-28ee-4fea-9f01-c95cf08a0413-kube-api-access-6h5hk" (OuterVolumeSpecName: "kube-api-access-6h5hk") pod "225d108a-28ee-4fea-9f01-c95cf08a0413" (UID: "225d108a-28ee-4fea-9f01-c95cf08a0413"). InnerVolumeSpecName "kube-api-access-6h5hk". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 19:49:47 crc kubenswrapper[4988]: I1008 19:49:47.587113 4988 scope.go:117] "RemoveContainer" containerID="6aceccbda0d681f4b250f907f606350db065a93a521c2a5c166be26d461622bf" Oct 08 19:49:47 crc kubenswrapper[4988]: I1008 19:49:47.602819 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/225d108a-28ee-4fea-9f01-c95cf08a0413-config-data" (OuterVolumeSpecName: "config-data") pod "225d108a-28ee-4fea-9f01-c95cf08a0413" (UID: "225d108a-28ee-4fea-9f01-c95cf08a0413"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 19:49:47 crc kubenswrapper[4988]: I1008 19:49:47.607527 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Oct 08 19:49:47 crc kubenswrapper[4988]: I1008 19:49:47.615564 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/225d108a-28ee-4fea-9f01-c95cf08a0413-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "225d108a-28ee-4fea-9f01-c95cf08a0413" (UID: "225d108a-28ee-4fea-9f01-c95cf08a0413"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 19:49:47 crc kubenswrapper[4988]: I1008 19:49:47.615686 4988 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Oct 08 19:49:47 crc kubenswrapper[4988]: I1008 19:49:47.625969 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Oct 08 19:49:47 crc kubenswrapper[4988]: E1008 19:49:47.626325 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="225d108a-28ee-4fea-9f01-c95cf08a0413" containerName="nova-cell1-novncproxy-novncproxy" Oct 08 19:49:47 crc kubenswrapper[4988]: I1008 19:49:47.626341 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="225d108a-28ee-4fea-9f01-c95cf08a0413" containerName="nova-cell1-novncproxy-novncproxy" Oct 08 19:49:47 crc kubenswrapper[4988]: E1008 19:49:47.626363 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b354976f-9d09-4023-a869-d373f9717585" containerName="nova-metadata-metadata" Oct 08 19:49:47 crc kubenswrapper[4988]: I1008 19:49:47.626369 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="b354976f-9d09-4023-a869-d373f9717585" containerName="nova-metadata-metadata" Oct 08 19:49:47 crc kubenswrapper[4988]: E1008 19:49:47.626403 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b354976f-9d09-4023-a869-d373f9717585" containerName="nova-metadata-log" Oct 08 19:49:47 crc kubenswrapper[4988]: I1008 19:49:47.626410 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="b354976f-9d09-4023-a869-d373f9717585" containerName="nova-metadata-log" Oct 08 19:49:47 crc kubenswrapper[4988]: I1008 19:49:47.626579 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="b354976f-9d09-4023-a869-d373f9717585" containerName="nova-metadata-metadata" Oct 08 19:49:47 crc kubenswrapper[4988]: I1008 19:49:47.626600 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="b354976f-9d09-4023-a869-d373f9717585" containerName="nova-metadata-log" Oct 08 19:49:47 crc kubenswrapper[4988]: I1008 19:49:47.626613 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="225d108a-28ee-4fea-9f01-c95cf08a0413" containerName="nova-cell1-novncproxy-novncproxy" Oct 08 19:49:47 crc kubenswrapper[4988]: I1008 19:49:47.628249 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 08 19:49:47 crc kubenswrapper[4988]: I1008 19:49:47.631019 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 08 19:49:47 crc kubenswrapper[4988]: I1008 19:49:47.631647 4988 scope.go:117] "RemoveContainer" containerID="540181e64243b98c638416ffa7acbeac7561c2b19121b4f26f23f9bbd19800b4" Oct 08 19:49:47 crc kubenswrapper[4988]: E1008 19:49:47.632038 4988 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"540181e64243b98c638416ffa7acbeac7561c2b19121b4f26f23f9bbd19800b4\": container with ID starting with 540181e64243b98c638416ffa7acbeac7561c2b19121b4f26f23f9bbd19800b4 not found: ID does not exist" containerID="540181e64243b98c638416ffa7acbeac7561c2b19121b4f26f23f9bbd19800b4" Oct 08 19:49:47 crc kubenswrapper[4988]: I1008 19:49:47.632064 4988 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"540181e64243b98c638416ffa7acbeac7561c2b19121b4f26f23f9bbd19800b4"} err="failed to get container status \"540181e64243b98c638416ffa7acbeac7561c2b19121b4f26f23f9bbd19800b4\": rpc error: code = NotFound desc = could not find container \"540181e64243b98c638416ffa7acbeac7561c2b19121b4f26f23f9bbd19800b4\": container with ID starting with 540181e64243b98c638416ffa7acbeac7561c2b19121b4f26f23f9bbd19800b4 not found: ID does not exist" Oct 08 19:49:47 crc kubenswrapper[4988]: I1008 19:49:47.632208 4988 scope.go:117] "RemoveContainer" containerID="6aceccbda0d681f4b250f907f606350db065a93a521c2a5c166be26d461622bf" Oct 08 19:49:47 crc kubenswrapper[4988]: I1008 19:49:47.632261 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Oct 08 19:49:47 crc kubenswrapper[4988]: I1008 19:49:47.632288 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Oct 08 19:49:47 crc kubenswrapper[4988]: E1008 19:49:47.632492 4988 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6aceccbda0d681f4b250f907f606350db065a93a521c2a5c166be26d461622bf\": container with ID starting with 6aceccbda0d681f4b250f907f606350db065a93a521c2a5c166be26d461622bf not found: ID does not exist" containerID="6aceccbda0d681f4b250f907f606350db065a93a521c2a5c166be26d461622bf" Oct 08 19:49:47 crc kubenswrapper[4988]: I1008 19:49:47.632535 4988 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6aceccbda0d681f4b250f907f606350db065a93a521c2a5c166be26d461622bf"} err="failed to get container status \"6aceccbda0d681f4b250f907f606350db065a93a521c2a5c166be26d461622bf\": rpc error: code = NotFound desc = could not find container \"6aceccbda0d681f4b250f907f606350db065a93a521c2a5c166be26d461622bf\": container with ID starting with 6aceccbda0d681f4b250f907f606350db065a93a521c2a5c166be26d461622bf not found: ID does not exist" Oct 08 19:49:47 crc kubenswrapper[4988]: I1008 19:49:47.632566 4988 scope.go:117] "RemoveContainer" containerID="540181e64243b98c638416ffa7acbeac7561c2b19121b4f26f23f9bbd19800b4" Oct 08 19:49:47 crc kubenswrapper[4988]: I1008 19:49:47.632924 4988 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"540181e64243b98c638416ffa7acbeac7561c2b19121b4f26f23f9bbd19800b4"} err="failed to get container status \"540181e64243b98c638416ffa7acbeac7561c2b19121b4f26f23f9bbd19800b4\": rpc error: code = NotFound desc = could not find container \"540181e64243b98c638416ffa7acbeac7561c2b19121b4f26f23f9bbd19800b4\": container with ID starting with 540181e64243b98c638416ffa7acbeac7561c2b19121b4f26f23f9bbd19800b4 not found: ID does not exist" Oct 08 19:49:47 crc kubenswrapper[4988]: I1008 19:49:47.632962 4988 scope.go:117] "RemoveContainer" containerID="6aceccbda0d681f4b250f907f606350db065a93a521c2a5c166be26d461622bf" Oct 08 19:49:47 crc kubenswrapper[4988]: I1008 19:49:47.633172 4988 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6aceccbda0d681f4b250f907f606350db065a93a521c2a5c166be26d461622bf"} err="failed to get container status \"6aceccbda0d681f4b250f907f606350db065a93a521c2a5c166be26d461622bf\": rpc error: code = NotFound desc = could not find container \"6aceccbda0d681f4b250f907f606350db065a93a521c2a5c166be26d461622bf\": container with ID starting with 6aceccbda0d681f4b250f907f606350db065a93a521c2a5c166be26d461622bf not found: ID does not exist" Oct 08 19:49:47 crc kubenswrapper[4988]: I1008 19:49:47.670244 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/62ad0fcb-bc01-493c-8fc6-4a186e4e197f-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"62ad0fcb-bc01-493c-8fc6-4a186e4e197f\") " pod="openstack/nova-metadata-0" Oct 08 19:49:47 crc kubenswrapper[4988]: I1008 19:49:47.670294 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/62ad0fcb-bc01-493c-8fc6-4a186e4e197f-logs\") pod \"nova-metadata-0\" (UID: \"62ad0fcb-bc01-493c-8fc6-4a186e4e197f\") " pod="openstack/nova-metadata-0" Oct 08 19:49:47 crc kubenswrapper[4988]: I1008 19:49:47.670338 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/62ad0fcb-bc01-493c-8fc6-4a186e4e197f-config-data\") pod \"nova-metadata-0\" (UID: \"62ad0fcb-bc01-493c-8fc6-4a186e4e197f\") " pod="openstack/nova-metadata-0" Oct 08 19:49:47 crc kubenswrapper[4988]: I1008 19:49:47.670360 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/62ad0fcb-bc01-493c-8fc6-4a186e4e197f-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"62ad0fcb-bc01-493c-8fc6-4a186e4e197f\") " pod="openstack/nova-metadata-0" Oct 08 19:49:47 crc kubenswrapper[4988]: I1008 19:49:47.670436 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q9nkw\" (UniqueName: \"kubernetes.io/projected/62ad0fcb-bc01-493c-8fc6-4a186e4e197f-kube-api-access-q9nkw\") pod \"nova-metadata-0\" (UID: \"62ad0fcb-bc01-493c-8fc6-4a186e4e197f\") " pod="openstack/nova-metadata-0" Oct 08 19:49:47 crc kubenswrapper[4988]: I1008 19:49:47.670576 4988 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6h5hk\" (UniqueName: \"kubernetes.io/projected/225d108a-28ee-4fea-9f01-c95cf08a0413-kube-api-access-6h5hk\") on node \"crc\" DevicePath \"\"" Oct 08 19:49:47 crc kubenswrapper[4988]: I1008 19:49:47.670610 4988 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/225d108a-28ee-4fea-9f01-c95cf08a0413-config-data\") on node \"crc\" DevicePath \"\"" Oct 08 19:49:47 crc kubenswrapper[4988]: I1008 19:49:47.670623 4988 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/225d108a-28ee-4fea-9f01-c95cf08a0413-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 19:49:47 crc kubenswrapper[4988]: I1008 19:49:47.772557 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/62ad0fcb-bc01-493c-8fc6-4a186e4e197f-config-data\") pod \"nova-metadata-0\" (UID: \"62ad0fcb-bc01-493c-8fc6-4a186e4e197f\") " pod="openstack/nova-metadata-0" Oct 08 19:49:47 crc kubenswrapper[4988]: I1008 19:49:47.772604 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/62ad0fcb-bc01-493c-8fc6-4a186e4e197f-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"62ad0fcb-bc01-493c-8fc6-4a186e4e197f\") " pod="openstack/nova-metadata-0" Oct 08 19:49:47 crc kubenswrapper[4988]: I1008 19:49:47.772865 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q9nkw\" (UniqueName: \"kubernetes.io/projected/62ad0fcb-bc01-493c-8fc6-4a186e4e197f-kube-api-access-q9nkw\") pod \"nova-metadata-0\" (UID: \"62ad0fcb-bc01-493c-8fc6-4a186e4e197f\") " pod="openstack/nova-metadata-0" Oct 08 19:49:47 crc kubenswrapper[4988]: I1008 19:49:47.772936 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/62ad0fcb-bc01-493c-8fc6-4a186e4e197f-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"62ad0fcb-bc01-493c-8fc6-4a186e4e197f\") " pod="openstack/nova-metadata-0" Oct 08 19:49:47 crc kubenswrapper[4988]: I1008 19:49:47.772964 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/62ad0fcb-bc01-493c-8fc6-4a186e4e197f-logs\") pod \"nova-metadata-0\" (UID: \"62ad0fcb-bc01-493c-8fc6-4a186e4e197f\") " pod="openstack/nova-metadata-0" Oct 08 19:49:47 crc kubenswrapper[4988]: I1008 19:49:47.773312 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/62ad0fcb-bc01-493c-8fc6-4a186e4e197f-logs\") pod \"nova-metadata-0\" (UID: \"62ad0fcb-bc01-493c-8fc6-4a186e4e197f\") " pod="openstack/nova-metadata-0" Oct 08 19:49:47 crc kubenswrapper[4988]: I1008 19:49:47.777059 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/62ad0fcb-bc01-493c-8fc6-4a186e4e197f-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"62ad0fcb-bc01-493c-8fc6-4a186e4e197f\") " pod="openstack/nova-metadata-0" Oct 08 19:49:47 crc kubenswrapper[4988]: I1008 19:49:47.777490 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/62ad0fcb-bc01-493c-8fc6-4a186e4e197f-config-data\") pod \"nova-metadata-0\" (UID: \"62ad0fcb-bc01-493c-8fc6-4a186e4e197f\") " pod="openstack/nova-metadata-0" Oct 08 19:49:47 crc kubenswrapper[4988]: I1008 19:49:47.777502 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/62ad0fcb-bc01-493c-8fc6-4a186e4e197f-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"62ad0fcb-bc01-493c-8fc6-4a186e4e197f\") " pod="openstack/nova-metadata-0" Oct 08 19:49:47 crc kubenswrapper[4988]: I1008 19:49:47.793472 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q9nkw\" (UniqueName: \"kubernetes.io/projected/62ad0fcb-bc01-493c-8fc6-4a186e4e197f-kube-api-access-q9nkw\") pod \"nova-metadata-0\" (UID: \"62ad0fcb-bc01-493c-8fc6-4a186e4e197f\") " pod="openstack/nova-metadata-0" Oct 08 19:49:47 crc kubenswrapper[4988]: I1008 19:49:47.961069 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 08 19:49:47 crc kubenswrapper[4988]: I1008 19:49:47.967459 4988 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 08 19:49:47 crc kubenswrapper[4988]: I1008 19:49:47.979760 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 08 19:49:47 crc kubenswrapper[4988]: I1008 19:49:47.985966 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Oct 08 19:49:47 crc kubenswrapper[4988]: I1008 19:49:47.987661 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Oct 08 19:49:47 crc kubenswrapper[4988]: I1008 19:49:47.987999 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-novncproxy-cell1-vencrypt" Oct 08 19:49:47 crc kubenswrapper[4988]: I1008 19:49:47.988115 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-novncproxy-cell1-public-svc" Oct 08 19:49:47 crc kubenswrapper[4988]: I1008 19:49:47.994497 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 08 19:49:47 crc kubenswrapper[4988]: I1008 19:49:47.997763 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Oct 08 19:49:48 crc kubenswrapper[4988]: I1008 19:49:48.007344 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 08 19:49:48 crc kubenswrapper[4988]: I1008 19:49:48.078183 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wrnj2\" (UniqueName: \"kubernetes.io/projected/48846c2d-55ec-4747-b3ab-4ed54cebc731-kube-api-access-wrnj2\") pod \"nova-cell1-novncproxy-0\" (UID: \"48846c2d-55ec-4747-b3ab-4ed54cebc731\") " pod="openstack/nova-cell1-novncproxy-0" Oct 08 19:49:48 crc kubenswrapper[4988]: I1008 19:49:48.078235 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/48846c2d-55ec-4747-b3ab-4ed54cebc731-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"48846c2d-55ec-4747-b3ab-4ed54cebc731\") " pod="openstack/nova-cell1-novncproxy-0" Oct 08 19:49:48 crc kubenswrapper[4988]: I1008 19:49:48.078302 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/48846c2d-55ec-4747-b3ab-4ed54cebc731-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"48846c2d-55ec-4747-b3ab-4ed54cebc731\") " pod="openstack/nova-cell1-novncproxy-0" Oct 08 19:49:48 crc kubenswrapper[4988]: I1008 19:49:48.078375 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/48846c2d-55ec-4747-b3ab-4ed54cebc731-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"48846c2d-55ec-4747-b3ab-4ed54cebc731\") " pod="openstack/nova-cell1-novncproxy-0" Oct 08 19:49:48 crc kubenswrapper[4988]: I1008 19:49:48.078478 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/48846c2d-55ec-4747-b3ab-4ed54cebc731-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"48846c2d-55ec-4747-b3ab-4ed54cebc731\") " pod="openstack/nova-cell1-novncproxy-0" Oct 08 19:49:48 crc kubenswrapper[4988]: I1008 19:49:48.179691 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/48846c2d-55ec-4747-b3ab-4ed54cebc731-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"48846c2d-55ec-4747-b3ab-4ed54cebc731\") " pod="openstack/nova-cell1-novncproxy-0" Oct 08 19:49:48 crc kubenswrapper[4988]: I1008 19:49:48.181313 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/48846c2d-55ec-4747-b3ab-4ed54cebc731-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"48846c2d-55ec-4747-b3ab-4ed54cebc731\") " pod="openstack/nova-cell1-novncproxy-0" Oct 08 19:49:48 crc kubenswrapper[4988]: I1008 19:49:48.181425 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/48846c2d-55ec-4747-b3ab-4ed54cebc731-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"48846c2d-55ec-4747-b3ab-4ed54cebc731\") " pod="openstack/nova-cell1-novncproxy-0" Oct 08 19:49:48 crc kubenswrapper[4988]: I1008 19:49:48.183076 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wrnj2\" (UniqueName: \"kubernetes.io/projected/48846c2d-55ec-4747-b3ab-4ed54cebc731-kube-api-access-wrnj2\") pod \"nova-cell1-novncproxy-0\" (UID: \"48846c2d-55ec-4747-b3ab-4ed54cebc731\") " pod="openstack/nova-cell1-novncproxy-0" Oct 08 19:49:48 crc kubenswrapper[4988]: I1008 19:49:48.183112 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/48846c2d-55ec-4747-b3ab-4ed54cebc731-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"48846c2d-55ec-4747-b3ab-4ed54cebc731\") " pod="openstack/nova-cell1-novncproxy-0" Oct 08 19:49:48 crc kubenswrapper[4988]: I1008 19:49:48.185677 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/48846c2d-55ec-4747-b3ab-4ed54cebc731-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"48846c2d-55ec-4747-b3ab-4ed54cebc731\") " pod="openstack/nova-cell1-novncproxy-0" Oct 08 19:49:48 crc kubenswrapper[4988]: I1008 19:49:48.185683 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/48846c2d-55ec-4747-b3ab-4ed54cebc731-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"48846c2d-55ec-4747-b3ab-4ed54cebc731\") " pod="openstack/nova-cell1-novncproxy-0" Oct 08 19:49:48 crc kubenswrapper[4988]: I1008 19:49:48.185973 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/48846c2d-55ec-4747-b3ab-4ed54cebc731-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"48846c2d-55ec-4747-b3ab-4ed54cebc731\") " pod="openstack/nova-cell1-novncproxy-0" Oct 08 19:49:48 crc kubenswrapper[4988]: I1008 19:49:48.189006 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/48846c2d-55ec-4747-b3ab-4ed54cebc731-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"48846c2d-55ec-4747-b3ab-4ed54cebc731\") " pod="openstack/nova-cell1-novncproxy-0" Oct 08 19:49:48 crc kubenswrapper[4988]: I1008 19:49:48.205125 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wrnj2\" (UniqueName: \"kubernetes.io/projected/48846c2d-55ec-4747-b3ab-4ed54cebc731-kube-api-access-wrnj2\") pod \"nova-cell1-novncproxy-0\" (UID: \"48846c2d-55ec-4747-b3ab-4ed54cebc731\") " pod="openstack/nova-cell1-novncproxy-0" Oct 08 19:49:48 crc kubenswrapper[4988]: I1008 19:49:48.440605 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Oct 08 19:49:48 crc kubenswrapper[4988]: I1008 19:49:48.502460 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 08 19:49:48 crc kubenswrapper[4988]: I1008 19:49:48.556454 4988 generic.go:334] "Generic (PLEG): container finished" podID="f6f69dea-85ca-4d7e-9f1c-05f11cadd3b9" containerID="9b8bbc55b526532b8819a7ac1a54a4c1aa8fc31d5cca19ae4663664bf7137162" exitCode=0 Oct 08 19:49:48 crc kubenswrapper[4988]: I1008 19:49:48.556539 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-tmx75" event={"ID":"f6f69dea-85ca-4d7e-9f1c-05f11cadd3b9","Type":"ContainerDied","Data":"9b8bbc55b526532b8819a7ac1a54a4c1aa8fc31d5cca19ae4663664bf7137162"} Oct 08 19:49:48 crc kubenswrapper[4988]: I1008 19:49:48.568748 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"62ad0fcb-bc01-493c-8fc6-4a186e4e197f","Type":"ContainerStarted","Data":"4c2f5b5a3c0bb5f6cf9f0b4dbf57712ba1b5c160bfa4803f733182db9279073a"} Oct 08 19:49:48 crc kubenswrapper[4988]: I1008 19:49:48.934228 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 08 19:49:48 crc kubenswrapper[4988]: W1008 19:49:48.936616 4988 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod48846c2d_55ec_4747_b3ab_4ed54cebc731.slice/crio-08c892371d8ace34de30baabe995dc9ec45e95ddf67c67cd4f4aa30459404d12 WatchSource:0}: Error finding container 08c892371d8ace34de30baabe995dc9ec45e95ddf67c67cd4f4aa30459404d12: Status 404 returned error can't find the container with id 08c892371d8ace34de30baabe995dc9ec45e95ddf67c67cd4f4aa30459404d12 Oct 08 19:49:49 crc kubenswrapper[4988]: I1008 19:49:49.258551 4988 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="225d108a-28ee-4fea-9f01-c95cf08a0413" path="/var/lib/kubelet/pods/225d108a-28ee-4fea-9f01-c95cf08a0413/volumes" Oct 08 19:49:49 crc kubenswrapper[4988]: I1008 19:49:49.259904 4988 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b354976f-9d09-4023-a869-d373f9717585" path="/var/lib/kubelet/pods/b354976f-9d09-4023-a869-d373f9717585/volumes" Oct 08 19:49:49 crc kubenswrapper[4988]: I1008 19:49:49.599007 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"62ad0fcb-bc01-493c-8fc6-4a186e4e197f","Type":"ContainerStarted","Data":"9717aa641b94b3af71e62103a9ba8c139f01ae406cf769d77ea28bcd2a2a482e"} Oct 08 19:49:49 crc kubenswrapper[4988]: I1008 19:49:49.599075 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"62ad0fcb-bc01-493c-8fc6-4a186e4e197f","Type":"ContainerStarted","Data":"c14335abef6035f10ff1740b717b228c578b085bebb38659384f1fc7518b10f1"} Oct 08 19:49:49 crc kubenswrapper[4988]: I1008 19:49:49.611103 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"48846c2d-55ec-4747-b3ab-4ed54cebc731","Type":"ContainerStarted","Data":"c6f5a1d62b6dc2b6df9dd5ce93877ba9673d582c02e540577c75b8f1e543a789"} Oct 08 19:49:49 crc kubenswrapper[4988]: I1008 19:49:49.611189 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"48846c2d-55ec-4747-b3ab-4ed54cebc731","Type":"ContainerStarted","Data":"08c892371d8ace34de30baabe995dc9ec45e95ddf67c67cd4f4aa30459404d12"} Oct 08 19:49:49 crc kubenswrapper[4988]: I1008 19:49:49.617606 4988 generic.go:334] "Generic (PLEG): container finished" podID="586b0be5-17fd-4984-87ba-a8c2862ddbc0" containerID="521bb2202c7abb16ee78ce44796f153e3a190098a143f6b2bcd30634e128b462" exitCode=0 Oct 08 19:49:49 crc kubenswrapper[4988]: I1008 19:49:49.617825 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-jtj56" event={"ID":"586b0be5-17fd-4984-87ba-a8c2862ddbc0","Type":"ContainerDied","Data":"521bb2202c7abb16ee78ce44796f153e3a190098a143f6b2bcd30634e128b462"} Oct 08 19:49:49 crc kubenswrapper[4988]: I1008 19:49:49.634961 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.634943752 podStartE2EDuration="2.634943752s" podCreationTimestamp="2025-10-08 19:49:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 19:49:49.621188235 +0000 UTC m=+5935.071031045" watchObservedRunningTime="2025-10-08 19:49:49.634943752 +0000 UTC m=+5935.084786532" Oct 08 19:49:49 crc kubenswrapper[4988]: I1008 19:49:49.657925 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=2.657907212 podStartE2EDuration="2.657907212s" podCreationTimestamp="2025-10-08 19:49:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 19:49:49.647515512 +0000 UTC m=+5935.097358312" watchObservedRunningTime="2025-10-08 19:49:49.657907212 +0000 UTC m=+5935.107749992" Oct 08 19:49:50 crc kubenswrapper[4988]: I1008 19:49:50.079029 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-tmx75" Oct 08 19:49:50 crc kubenswrapper[4988]: I1008 19:49:50.138317 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f6f69dea-85ca-4d7e-9f1c-05f11cadd3b9-combined-ca-bundle\") pod \"f6f69dea-85ca-4d7e-9f1c-05f11cadd3b9\" (UID: \"f6f69dea-85ca-4d7e-9f1c-05f11cadd3b9\") " Oct 08 19:49:50 crc kubenswrapper[4988]: I1008 19:49:50.138471 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f6f69dea-85ca-4d7e-9f1c-05f11cadd3b9-config-data\") pod \"f6f69dea-85ca-4d7e-9f1c-05f11cadd3b9\" (UID: \"f6f69dea-85ca-4d7e-9f1c-05f11cadd3b9\") " Oct 08 19:49:50 crc kubenswrapper[4988]: I1008 19:49:50.138568 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9fgsh\" (UniqueName: \"kubernetes.io/projected/f6f69dea-85ca-4d7e-9f1c-05f11cadd3b9-kube-api-access-9fgsh\") pod \"f6f69dea-85ca-4d7e-9f1c-05f11cadd3b9\" (UID: \"f6f69dea-85ca-4d7e-9f1c-05f11cadd3b9\") " Oct 08 19:49:50 crc kubenswrapper[4988]: I1008 19:49:50.138611 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f6f69dea-85ca-4d7e-9f1c-05f11cadd3b9-scripts\") pod \"f6f69dea-85ca-4d7e-9f1c-05f11cadd3b9\" (UID: \"f6f69dea-85ca-4d7e-9f1c-05f11cadd3b9\") " Oct 08 19:49:50 crc kubenswrapper[4988]: I1008 19:49:50.157659 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f6f69dea-85ca-4d7e-9f1c-05f11cadd3b9-scripts" (OuterVolumeSpecName: "scripts") pod "f6f69dea-85ca-4d7e-9f1c-05f11cadd3b9" (UID: "f6f69dea-85ca-4d7e-9f1c-05f11cadd3b9"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 19:49:50 crc kubenswrapper[4988]: I1008 19:49:50.157755 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f6f69dea-85ca-4d7e-9f1c-05f11cadd3b9-kube-api-access-9fgsh" (OuterVolumeSpecName: "kube-api-access-9fgsh") pod "f6f69dea-85ca-4d7e-9f1c-05f11cadd3b9" (UID: "f6f69dea-85ca-4d7e-9f1c-05f11cadd3b9"). InnerVolumeSpecName "kube-api-access-9fgsh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 19:49:50 crc kubenswrapper[4988]: I1008 19:49:50.187750 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f6f69dea-85ca-4d7e-9f1c-05f11cadd3b9-config-data" (OuterVolumeSpecName: "config-data") pod "f6f69dea-85ca-4d7e-9f1c-05f11cadd3b9" (UID: "f6f69dea-85ca-4d7e-9f1c-05f11cadd3b9"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 19:49:50 crc kubenswrapper[4988]: I1008 19:49:50.196730 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f6f69dea-85ca-4d7e-9f1c-05f11cadd3b9-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f6f69dea-85ca-4d7e-9f1c-05f11cadd3b9" (UID: "f6f69dea-85ca-4d7e-9f1c-05f11cadd3b9"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 19:49:50 crc kubenswrapper[4988]: I1008 19:49:50.241760 4988 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f6f69dea-85ca-4d7e-9f1c-05f11cadd3b9-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 19:49:50 crc kubenswrapper[4988]: I1008 19:49:50.241791 4988 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f6f69dea-85ca-4d7e-9f1c-05f11cadd3b9-config-data\") on node \"crc\" DevicePath \"\"" Oct 08 19:49:50 crc kubenswrapper[4988]: I1008 19:49:50.241805 4988 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9fgsh\" (UniqueName: \"kubernetes.io/projected/f6f69dea-85ca-4d7e-9f1c-05f11cadd3b9-kube-api-access-9fgsh\") on node \"crc\" DevicePath \"\"" Oct 08 19:49:50 crc kubenswrapper[4988]: I1008 19:49:50.241817 4988 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f6f69dea-85ca-4d7e-9f1c-05f11cadd3b9-scripts\") on node \"crc\" DevicePath \"\"" Oct 08 19:49:50 crc kubenswrapper[4988]: I1008 19:49:50.661984 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-tmx75" Oct 08 19:49:50 crc kubenswrapper[4988]: I1008 19:49:50.665774 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-tmx75" event={"ID":"f6f69dea-85ca-4d7e-9f1c-05f11cadd3b9","Type":"ContainerDied","Data":"9ecf571ea874f003e8866a4166050422888f259d06b800e7586518fedb847cf1"} Oct 08 19:49:50 crc kubenswrapper[4988]: I1008 19:49:50.665894 4988 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9ecf571ea874f003e8866a4166050422888f259d06b800e7586518fedb847cf1" Oct 08 19:49:50 crc kubenswrapper[4988]: I1008 19:49:50.701492 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-0"] Oct 08 19:49:50 crc kubenswrapper[4988]: E1008 19:49:50.702038 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f6f69dea-85ca-4d7e-9f1c-05f11cadd3b9" containerName="nova-cell1-conductor-db-sync" Oct 08 19:49:50 crc kubenswrapper[4988]: I1008 19:49:50.702058 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="f6f69dea-85ca-4d7e-9f1c-05f11cadd3b9" containerName="nova-cell1-conductor-db-sync" Oct 08 19:49:50 crc kubenswrapper[4988]: I1008 19:49:50.702310 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="f6f69dea-85ca-4d7e-9f1c-05f11cadd3b9" containerName="nova-cell1-conductor-db-sync" Oct 08 19:49:50 crc kubenswrapper[4988]: I1008 19:49:50.703160 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Oct 08 19:49:50 crc kubenswrapper[4988]: I1008 19:49:50.708974 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Oct 08 19:49:50 crc kubenswrapper[4988]: I1008 19:49:50.713649 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Oct 08 19:49:50 crc kubenswrapper[4988]: I1008 19:49:50.751189 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9htsf\" (UniqueName: \"kubernetes.io/projected/730ab813-7314-4ff9-bd44-1eeeff96e912-kube-api-access-9htsf\") pod \"nova-cell1-conductor-0\" (UID: \"730ab813-7314-4ff9-bd44-1eeeff96e912\") " pod="openstack/nova-cell1-conductor-0" Oct 08 19:49:50 crc kubenswrapper[4988]: I1008 19:49:50.751301 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/730ab813-7314-4ff9-bd44-1eeeff96e912-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"730ab813-7314-4ff9-bd44-1eeeff96e912\") " pod="openstack/nova-cell1-conductor-0" Oct 08 19:49:50 crc kubenswrapper[4988]: I1008 19:49:50.751349 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/730ab813-7314-4ff9-bd44-1eeeff96e912-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"730ab813-7314-4ff9-bd44-1eeeff96e912\") " pod="openstack/nova-cell1-conductor-0" Oct 08 19:49:50 crc kubenswrapper[4988]: E1008 19:49:50.810789 4988 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf6f69dea_85ca_4d7e_9f1c_05f11cadd3b9.slice\": RecentStats: unable to find data in memory cache]" Oct 08 19:49:50 crc kubenswrapper[4988]: I1008 19:49:50.853076 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9htsf\" (UniqueName: \"kubernetes.io/projected/730ab813-7314-4ff9-bd44-1eeeff96e912-kube-api-access-9htsf\") pod \"nova-cell1-conductor-0\" (UID: \"730ab813-7314-4ff9-bd44-1eeeff96e912\") " pod="openstack/nova-cell1-conductor-0" Oct 08 19:49:50 crc kubenswrapper[4988]: I1008 19:49:50.853174 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/730ab813-7314-4ff9-bd44-1eeeff96e912-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"730ab813-7314-4ff9-bd44-1eeeff96e912\") " pod="openstack/nova-cell1-conductor-0" Oct 08 19:49:50 crc kubenswrapper[4988]: I1008 19:49:50.853204 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/730ab813-7314-4ff9-bd44-1eeeff96e912-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"730ab813-7314-4ff9-bd44-1eeeff96e912\") " pod="openstack/nova-cell1-conductor-0" Oct 08 19:49:50 crc kubenswrapper[4988]: I1008 19:49:50.869867 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/730ab813-7314-4ff9-bd44-1eeeff96e912-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"730ab813-7314-4ff9-bd44-1eeeff96e912\") " pod="openstack/nova-cell1-conductor-0" Oct 08 19:49:50 crc kubenswrapper[4988]: I1008 19:49:50.870627 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/730ab813-7314-4ff9-bd44-1eeeff96e912-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"730ab813-7314-4ff9-bd44-1eeeff96e912\") " pod="openstack/nova-cell1-conductor-0" Oct 08 19:49:50 crc kubenswrapper[4988]: I1008 19:49:50.885969 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9htsf\" (UniqueName: \"kubernetes.io/projected/730ab813-7314-4ff9-bd44-1eeeff96e912-kube-api-access-9htsf\") pod \"nova-cell1-conductor-0\" (UID: \"730ab813-7314-4ff9-bd44-1eeeff96e912\") " pod="openstack/nova-cell1-conductor-0" Oct 08 19:49:51 crc kubenswrapper[4988]: I1008 19:49:51.027875 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Oct 08 19:49:51 crc kubenswrapper[4988]: I1008 19:49:51.055898 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-jtj56" Oct 08 19:49:51 crc kubenswrapper[4988]: I1008 19:49:51.157959 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/586b0be5-17fd-4984-87ba-a8c2862ddbc0-combined-ca-bundle\") pod \"586b0be5-17fd-4984-87ba-a8c2862ddbc0\" (UID: \"586b0be5-17fd-4984-87ba-a8c2862ddbc0\") " Oct 08 19:49:51 crc kubenswrapper[4988]: I1008 19:49:51.158040 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2kt7x\" (UniqueName: \"kubernetes.io/projected/586b0be5-17fd-4984-87ba-a8c2862ddbc0-kube-api-access-2kt7x\") pod \"586b0be5-17fd-4984-87ba-a8c2862ddbc0\" (UID: \"586b0be5-17fd-4984-87ba-a8c2862ddbc0\") " Oct 08 19:49:51 crc kubenswrapper[4988]: I1008 19:49:51.158110 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/586b0be5-17fd-4984-87ba-a8c2862ddbc0-config-data\") pod \"586b0be5-17fd-4984-87ba-a8c2862ddbc0\" (UID: \"586b0be5-17fd-4984-87ba-a8c2862ddbc0\") " Oct 08 19:49:51 crc kubenswrapper[4988]: I1008 19:49:51.158135 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/586b0be5-17fd-4984-87ba-a8c2862ddbc0-scripts\") pod \"586b0be5-17fd-4984-87ba-a8c2862ddbc0\" (UID: \"586b0be5-17fd-4984-87ba-a8c2862ddbc0\") " Oct 08 19:49:51 crc kubenswrapper[4988]: I1008 19:49:51.162512 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/586b0be5-17fd-4984-87ba-a8c2862ddbc0-scripts" (OuterVolumeSpecName: "scripts") pod "586b0be5-17fd-4984-87ba-a8c2862ddbc0" (UID: "586b0be5-17fd-4984-87ba-a8c2862ddbc0"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 19:49:51 crc kubenswrapper[4988]: I1008 19:49:51.162937 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/586b0be5-17fd-4984-87ba-a8c2862ddbc0-kube-api-access-2kt7x" (OuterVolumeSpecName: "kube-api-access-2kt7x") pod "586b0be5-17fd-4984-87ba-a8c2862ddbc0" (UID: "586b0be5-17fd-4984-87ba-a8c2862ddbc0"). InnerVolumeSpecName "kube-api-access-2kt7x". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 19:49:51 crc kubenswrapper[4988]: I1008 19:49:51.189523 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/586b0be5-17fd-4984-87ba-a8c2862ddbc0-config-data" (OuterVolumeSpecName: "config-data") pod "586b0be5-17fd-4984-87ba-a8c2862ddbc0" (UID: "586b0be5-17fd-4984-87ba-a8c2862ddbc0"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 19:49:51 crc kubenswrapper[4988]: I1008 19:49:51.194086 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/586b0be5-17fd-4984-87ba-a8c2862ddbc0-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "586b0be5-17fd-4984-87ba-a8c2862ddbc0" (UID: "586b0be5-17fd-4984-87ba-a8c2862ddbc0"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 19:49:51 crc kubenswrapper[4988]: I1008 19:49:51.259726 4988 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/586b0be5-17fd-4984-87ba-a8c2862ddbc0-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 19:49:51 crc kubenswrapper[4988]: I1008 19:49:51.259760 4988 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2kt7x\" (UniqueName: \"kubernetes.io/projected/586b0be5-17fd-4984-87ba-a8c2862ddbc0-kube-api-access-2kt7x\") on node \"crc\" DevicePath \"\"" Oct 08 19:49:51 crc kubenswrapper[4988]: I1008 19:49:51.259770 4988 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/586b0be5-17fd-4984-87ba-a8c2862ddbc0-config-data\") on node \"crc\" DevicePath \"\"" Oct 08 19:49:51 crc kubenswrapper[4988]: I1008 19:49:51.259779 4988 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/586b0be5-17fd-4984-87ba-a8c2862ddbc0-scripts\") on node \"crc\" DevicePath \"\"" Oct 08 19:49:51 crc kubenswrapper[4988]: I1008 19:49:51.531917 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Oct 08 19:49:51 crc kubenswrapper[4988]: W1008 19:49:51.535470 4988 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod730ab813_7314_4ff9_bd44_1eeeff96e912.slice/crio-53af335385340a15cf262e6df1d2842c11d4629fab2cc1d9081351bdd45830dc WatchSource:0}: Error finding container 53af335385340a15cf262e6df1d2842c11d4629fab2cc1d9081351bdd45830dc: Status 404 returned error can't find the container with id 53af335385340a15cf262e6df1d2842c11d4629fab2cc1d9081351bdd45830dc Oct 08 19:49:51 crc kubenswrapper[4988]: I1008 19:49:51.676151 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-jtj56" event={"ID":"586b0be5-17fd-4984-87ba-a8c2862ddbc0","Type":"ContainerDied","Data":"2b6ba334e548c9ad6b05c7e939f8f084e178ea5d7ab3d7a4a27bc3c0d7c5b3ef"} Oct 08 19:49:51 crc kubenswrapper[4988]: I1008 19:49:51.677903 4988 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2b6ba334e548c9ad6b05c7e939f8f084e178ea5d7ab3d7a4a27bc3c0d7c5b3ef" Oct 08 19:49:51 crc kubenswrapper[4988]: I1008 19:49:51.678109 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-jtj56" Oct 08 19:49:51 crc kubenswrapper[4988]: I1008 19:49:51.680803 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"730ab813-7314-4ff9-bd44-1eeeff96e912","Type":"ContainerStarted","Data":"53af335385340a15cf262e6df1d2842c11d4629fab2cc1d9081351bdd45830dc"} Oct 08 19:49:51 crc kubenswrapper[4988]: I1008 19:49:51.867650 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 08 19:49:51 crc kubenswrapper[4988]: I1008 19:49:51.867916 4988 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="984822c7-f554-4c5b-bba1-3cd471eee0f3" containerName="nova-api-log" containerID="cri-o://657bf93b8bae9ec7afe75fbc48ccfc45ccf9af760218e29cfbce56c270f7347f" gracePeriod=30 Oct 08 19:49:51 crc kubenswrapper[4988]: I1008 19:49:51.868017 4988 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="984822c7-f554-4c5b-bba1-3cd471eee0f3" containerName="nova-api-api" containerID="cri-o://cad4dfedacebaed69682037b44f45eb1f5e008bde623318e3394ee03a8575c82" gracePeriod=30 Oct 08 19:49:51 crc kubenswrapper[4988]: I1008 19:49:51.890254 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Oct 08 19:49:51 crc kubenswrapper[4988]: I1008 19:49:51.890566 4988 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="2ceb61b1-8b93-4d44-b27f-ac96c4566be6" containerName="nova-scheduler-scheduler" containerID="cri-o://95a43697a44a1358aec0023227996593f0f41fa671f8322b52c7d43b9e0dbc7b" gracePeriod=30 Oct 08 19:49:51 crc kubenswrapper[4988]: I1008 19:49:51.906534 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Oct 08 19:49:51 crc kubenswrapper[4988]: I1008 19:49:51.907130 4988 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="62ad0fcb-bc01-493c-8fc6-4a186e4e197f" containerName="nova-metadata-log" containerID="cri-o://c14335abef6035f10ff1740b717b228c578b085bebb38659384f1fc7518b10f1" gracePeriod=30 Oct 08 19:49:51 crc kubenswrapper[4988]: I1008 19:49:51.907277 4988 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="62ad0fcb-bc01-493c-8fc6-4a186e4e197f" containerName="nova-metadata-metadata" containerID="cri-o://9717aa641b94b3af71e62103a9ba8c139f01ae406cf769d77ea28bcd2a2a482e" gracePeriod=30 Oct 08 19:49:52 crc kubenswrapper[4988]: I1008 19:49:52.351049 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 08 19:49:52 crc kubenswrapper[4988]: I1008 19:49:52.444702 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 08 19:49:52 crc kubenswrapper[4988]: I1008 19:49:52.487133 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/62ad0fcb-bc01-493c-8fc6-4a186e4e197f-combined-ca-bundle\") pod \"62ad0fcb-bc01-493c-8fc6-4a186e4e197f\" (UID: \"62ad0fcb-bc01-493c-8fc6-4a186e4e197f\") " Oct 08 19:49:52 crc kubenswrapper[4988]: I1008 19:49:52.487310 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/984822c7-f554-4c5b-bba1-3cd471eee0f3-logs\") pod \"984822c7-f554-4c5b-bba1-3cd471eee0f3\" (UID: \"984822c7-f554-4c5b-bba1-3cd471eee0f3\") " Oct 08 19:49:52 crc kubenswrapper[4988]: I1008 19:49:52.487346 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/62ad0fcb-bc01-493c-8fc6-4a186e4e197f-config-data\") pod \"62ad0fcb-bc01-493c-8fc6-4a186e4e197f\" (UID: \"62ad0fcb-bc01-493c-8fc6-4a186e4e197f\") " Oct 08 19:49:52 crc kubenswrapper[4988]: I1008 19:49:52.487366 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/62ad0fcb-bc01-493c-8fc6-4a186e4e197f-nova-metadata-tls-certs\") pod \"62ad0fcb-bc01-493c-8fc6-4a186e4e197f\" (UID: \"62ad0fcb-bc01-493c-8fc6-4a186e4e197f\") " Oct 08 19:49:52 crc kubenswrapper[4988]: I1008 19:49:52.487439 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q9nkw\" (UniqueName: \"kubernetes.io/projected/62ad0fcb-bc01-493c-8fc6-4a186e4e197f-kube-api-access-q9nkw\") pod \"62ad0fcb-bc01-493c-8fc6-4a186e4e197f\" (UID: \"62ad0fcb-bc01-493c-8fc6-4a186e4e197f\") " Oct 08 19:49:52 crc kubenswrapper[4988]: I1008 19:49:52.487468 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kl9pp\" (UniqueName: \"kubernetes.io/projected/984822c7-f554-4c5b-bba1-3cd471eee0f3-kube-api-access-kl9pp\") pod \"984822c7-f554-4c5b-bba1-3cd471eee0f3\" (UID: \"984822c7-f554-4c5b-bba1-3cd471eee0f3\") " Oct 08 19:49:52 crc kubenswrapper[4988]: I1008 19:49:52.487504 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/62ad0fcb-bc01-493c-8fc6-4a186e4e197f-logs\") pod \"62ad0fcb-bc01-493c-8fc6-4a186e4e197f\" (UID: \"62ad0fcb-bc01-493c-8fc6-4a186e4e197f\") " Oct 08 19:49:52 crc kubenswrapper[4988]: I1008 19:49:52.487536 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/984822c7-f554-4c5b-bba1-3cd471eee0f3-combined-ca-bundle\") pod \"984822c7-f554-4c5b-bba1-3cd471eee0f3\" (UID: \"984822c7-f554-4c5b-bba1-3cd471eee0f3\") " Oct 08 19:49:52 crc kubenswrapper[4988]: I1008 19:49:52.487570 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/984822c7-f554-4c5b-bba1-3cd471eee0f3-config-data\") pod \"984822c7-f554-4c5b-bba1-3cd471eee0f3\" (UID: \"984822c7-f554-4c5b-bba1-3cd471eee0f3\") " Oct 08 19:49:52 crc kubenswrapper[4988]: I1008 19:49:52.488112 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/984822c7-f554-4c5b-bba1-3cd471eee0f3-logs" (OuterVolumeSpecName: "logs") pod "984822c7-f554-4c5b-bba1-3cd471eee0f3" (UID: "984822c7-f554-4c5b-bba1-3cd471eee0f3"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 19:49:52 crc kubenswrapper[4988]: I1008 19:49:52.488269 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/62ad0fcb-bc01-493c-8fc6-4a186e4e197f-logs" (OuterVolumeSpecName: "logs") pod "62ad0fcb-bc01-493c-8fc6-4a186e4e197f" (UID: "62ad0fcb-bc01-493c-8fc6-4a186e4e197f"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 19:49:52 crc kubenswrapper[4988]: I1008 19:49:52.488837 4988 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/62ad0fcb-bc01-493c-8fc6-4a186e4e197f-logs\") on node \"crc\" DevicePath \"\"" Oct 08 19:49:52 crc kubenswrapper[4988]: I1008 19:49:52.488865 4988 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/984822c7-f554-4c5b-bba1-3cd471eee0f3-logs\") on node \"crc\" DevicePath \"\"" Oct 08 19:49:52 crc kubenswrapper[4988]: I1008 19:49:52.494449 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/984822c7-f554-4c5b-bba1-3cd471eee0f3-kube-api-access-kl9pp" (OuterVolumeSpecName: "kube-api-access-kl9pp") pod "984822c7-f554-4c5b-bba1-3cd471eee0f3" (UID: "984822c7-f554-4c5b-bba1-3cd471eee0f3"). InnerVolumeSpecName "kube-api-access-kl9pp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 19:49:52 crc kubenswrapper[4988]: I1008 19:49:52.495519 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/62ad0fcb-bc01-493c-8fc6-4a186e4e197f-kube-api-access-q9nkw" (OuterVolumeSpecName: "kube-api-access-q9nkw") pod "62ad0fcb-bc01-493c-8fc6-4a186e4e197f" (UID: "62ad0fcb-bc01-493c-8fc6-4a186e4e197f"). InnerVolumeSpecName "kube-api-access-q9nkw". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 19:49:52 crc kubenswrapper[4988]: I1008 19:49:52.514676 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/62ad0fcb-bc01-493c-8fc6-4a186e4e197f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "62ad0fcb-bc01-493c-8fc6-4a186e4e197f" (UID: "62ad0fcb-bc01-493c-8fc6-4a186e4e197f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 19:49:52 crc kubenswrapper[4988]: I1008 19:49:52.515135 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/984822c7-f554-4c5b-bba1-3cd471eee0f3-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "984822c7-f554-4c5b-bba1-3cd471eee0f3" (UID: "984822c7-f554-4c5b-bba1-3cd471eee0f3"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 19:49:52 crc kubenswrapper[4988]: I1008 19:49:52.519764 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/62ad0fcb-bc01-493c-8fc6-4a186e4e197f-config-data" (OuterVolumeSpecName: "config-data") pod "62ad0fcb-bc01-493c-8fc6-4a186e4e197f" (UID: "62ad0fcb-bc01-493c-8fc6-4a186e4e197f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 19:49:52 crc kubenswrapper[4988]: I1008 19:49:52.521249 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/984822c7-f554-4c5b-bba1-3cd471eee0f3-config-data" (OuterVolumeSpecName: "config-data") pod "984822c7-f554-4c5b-bba1-3cd471eee0f3" (UID: "984822c7-f554-4c5b-bba1-3cd471eee0f3"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 19:49:52 crc kubenswrapper[4988]: I1008 19:49:52.548606 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/62ad0fcb-bc01-493c-8fc6-4a186e4e197f-nova-metadata-tls-certs" (OuterVolumeSpecName: "nova-metadata-tls-certs") pod "62ad0fcb-bc01-493c-8fc6-4a186e4e197f" (UID: "62ad0fcb-bc01-493c-8fc6-4a186e4e197f"). InnerVolumeSpecName "nova-metadata-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 19:49:52 crc kubenswrapper[4988]: I1008 19:49:52.597010 4988 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/984822c7-f554-4c5b-bba1-3cd471eee0f3-config-data\") on node \"crc\" DevicePath \"\"" Oct 08 19:49:52 crc kubenswrapper[4988]: I1008 19:49:52.597041 4988 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/62ad0fcb-bc01-493c-8fc6-4a186e4e197f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 19:49:52 crc kubenswrapper[4988]: I1008 19:49:52.597053 4988 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/62ad0fcb-bc01-493c-8fc6-4a186e4e197f-config-data\") on node \"crc\" DevicePath \"\"" Oct 08 19:49:52 crc kubenswrapper[4988]: I1008 19:49:52.597061 4988 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/62ad0fcb-bc01-493c-8fc6-4a186e4e197f-nova-metadata-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 08 19:49:52 crc kubenswrapper[4988]: I1008 19:49:52.597071 4988 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q9nkw\" (UniqueName: \"kubernetes.io/projected/62ad0fcb-bc01-493c-8fc6-4a186e4e197f-kube-api-access-q9nkw\") on node \"crc\" DevicePath \"\"" Oct 08 19:49:52 crc kubenswrapper[4988]: I1008 19:49:52.597079 4988 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kl9pp\" (UniqueName: \"kubernetes.io/projected/984822c7-f554-4c5b-bba1-3cd471eee0f3-kube-api-access-kl9pp\") on node \"crc\" DevicePath \"\"" Oct 08 19:49:52 crc kubenswrapper[4988]: I1008 19:49:52.597087 4988 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/984822c7-f554-4c5b-bba1-3cd471eee0f3-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 19:49:52 crc kubenswrapper[4988]: I1008 19:49:52.698341 4988 generic.go:334] "Generic (PLEG): container finished" podID="984822c7-f554-4c5b-bba1-3cd471eee0f3" containerID="cad4dfedacebaed69682037b44f45eb1f5e008bde623318e3394ee03a8575c82" exitCode=0 Oct 08 19:49:52 crc kubenswrapper[4988]: I1008 19:49:52.698678 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 08 19:49:52 crc kubenswrapper[4988]: I1008 19:49:52.698691 4988 generic.go:334] "Generic (PLEG): container finished" podID="984822c7-f554-4c5b-bba1-3cd471eee0f3" containerID="657bf93b8bae9ec7afe75fbc48ccfc45ccf9af760218e29cfbce56c270f7347f" exitCode=143 Oct 08 19:49:52 crc kubenswrapper[4988]: I1008 19:49:52.698579 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"984822c7-f554-4c5b-bba1-3cd471eee0f3","Type":"ContainerDied","Data":"cad4dfedacebaed69682037b44f45eb1f5e008bde623318e3394ee03a8575c82"} Oct 08 19:49:52 crc kubenswrapper[4988]: I1008 19:49:52.698770 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"984822c7-f554-4c5b-bba1-3cd471eee0f3","Type":"ContainerDied","Data":"657bf93b8bae9ec7afe75fbc48ccfc45ccf9af760218e29cfbce56c270f7347f"} Oct 08 19:49:52 crc kubenswrapper[4988]: I1008 19:49:52.698788 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"984822c7-f554-4c5b-bba1-3cd471eee0f3","Type":"ContainerDied","Data":"5fd78401f11670490a43fd869a607ae304f9e60a4b2db9d48a8a14541ccbce91"} Oct 08 19:49:52 crc kubenswrapper[4988]: I1008 19:49:52.698807 4988 scope.go:117] "RemoveContainer" containerID="cad4dfedacebaed69682037b44f45eb1f5e008bde623318e3394ee03a8575c82" Oct 08 19:49:52 crc kubenswrapper[4988]: I1008 19:49:52.713830 4988 generic.go:334] "Generic (PLEG): container finished" podID="62ad0fcb-bc01-493c-8fc6-4a186e4e197f" containerID="9717aa641b94b3af71e62103a9ba8c139f01ae406cf769d77ea28bcd2a2a482e" exitCode=0 Oct 08 19:49:52 crc kubenswrapper[4988]: I1008 19:49:52.713858 4988 generic.go:334] "Generic (PLEG): container finished" podID="62ad0fcb-bc01-493c-8fc6-4a186e4e197f" containerID="c14335abef6035f10ff1740b717b228c578b085bebb38659384f1fc7518b10f1" exitCode=143 Oct 08 19:49:52 crc kubenswrapper[4988]: I1008 19:49:52.713936 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 08 19:49:52 crc kubenswrapper[4988]: I1008 19:49:52.713932 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"62ad0fcb-bc01-493c-8fc6-4a186e4e197f","Type":"ContainerDied","Data":"9717aa641b94b3af71e62103a9ba8c139f01ae406cf769d77ea28bcd2a2a482e"} Oct 08 19:49:52 crc kubenswrapper[4988]: I1008 19:49:52.714214 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"62ad0fcb-bc01-493c-8fc6-4a186e4e197f","Type":"ContainerDied","Data":"c14335abef6035f10ff1740b717b228c578b085bebb38659384f1fc7518b10f1"} Oct 08 19:49:52 crc kubenswrapper[4988]: I1008 19:49:52.714282 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"62ad0fcb-bc01-493c-8fc6-4a186e4e197f","Type":"ContainerDied","Data":"4c2f5b5a3c0bb5f6cf9f0b4dbf57712ba1b5c160bfa4803f733182db9279073a"} Oct 08 19:49:52 crc kubenswrapper[4988]: I1008 19:49:52.722950 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"730ab813-7314-4ff9-bd44-1eeeff96e912","Type":"ContainerStarted","Data":"f14ec0486f35029bb67d558c8eb8d27090175a66965eb8a80d480f8632247426"} Oct 08 19:49:52 crc kubenswrapper[4988]: I1008 19:49:52.723285 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-conductor-0" Oct 08 19:49:52 crc kubenswrapper[4988]: I1008 19:49:52.745148 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-0" podStartSLOduration=2.745130058 podStartE2EDuration="2.745130058s" podCreationTimestamp="2025-10-08 19:49:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 19:49:52.737722803 +0000 UTC m=+5938.187565633" watchObservedRunningTime="2025-10-08 19:49:52.745130058 +0000 UTC m=+5938.194972828" Oct 08 19:49:52 crc kubenswrapper[4988]: I1008 19:49:52.747941 4988 scope.go:117] "RemoveContainer" containerID="657bf93b8bae9ec7afe75fbc48ccfc45ccf9af760218e29cfbce56c270f7347f" Oct 08 19:49:52 crc kubenswrapper[4988]: I1008 19:49:52.771866 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 08 19:49:52 crc kubenswrapper[4988]: I1008 19:49:52.784318 4988 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Oct 08 19:49:52 crc kubenswrapper[4988]: I1008 19:49:52.808295 4988 scope.go:117] "RemoveContainer" containerID="cad4dfedacebaed69682037b44f45eb1f5e008bde623318e3394ee03a8575c82" Oct 08 19:49:52 crc kubenswrapper[4988]: I1008 19:49:52.808427 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Oct 08 19:49:52 crc kubenswrapper[4988]: E1008 19:49:52.810088 4988 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cad4dfedacebaed69682037b44f45eb1f5e008bde623318e3394ee03a8575c82\": container with ID starting with cad4dfedacebaed69682037b44f45eb1f5e008bde623318e3394ee03a8575c82 not found: ID does not exist" containerID="cad4dfedacebaed69682037b44f45eb1f5e008bde623318e3394ee03a8575c82" Oct 08 19:49:52 crc kubenswrapper[4988]: I1008 19:49:52.810129 4988 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cad4dfedacebaed69682037b44f45eb1f5e008bde623318e3394ee03a8575c82"} err="failed to get container status \"cad4dfedacebaed69682037b44f45eb1f5e008bde623318e3394ee03a8575c82\": rpc error: code = NotFound desc = could not find container \"cad4dfedacebaed69682037b44f45eb1f5e008bde623318e3394ee03a8575c82\": container with ID starting with cad4dfedacebaed69682037b44f45eb1f5e008bde623318e3394ee03a8575c82 not found: ID does not exist" Oct 08 19:49:52 crc kubenswrapper[4988]: I1008 19:49:52.810157 4988 scope.go:117] "RemoveContainer" containerID="657bf93b8bae9ec7afe75fbc48ccfc45ccf9af760218e29cfbce56c270f7347f" Oct 08 19:49:52 crc kubenswrapper[4988]: I1008 19:49:52.817608 4988 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Oct 08 19:49:52 crc kubenswrapper[4988]: E1008 19:49:52.826583 4988 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"657bf93b8bae9ec7afe75fbc48ccfc45ccf9af760218e29cfbce56c270f7347f\": container with ID starting with 657bf93b8bae9ec7afe75fbc48ccfc45ccf9af760218e29cfbce56c270f7347f not found: ID does not exist" containerID="657bf93b8bae9ec7afe75fbc48ccfc45ccf9af760218e29cfbce56c270f7347f" Oct 08 19:49:52 crc kubenswrapper[4988]: I1008 19:49:52.826642 4988 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"657bf93b8bae9ec7afe75fbc48ccfc45ccf9af760218e29cfbce56c270f7347f"} err="failed to get container status \"657bf93b8bae9ec7afe75fbc48ccfc45ccf9af760218e29cfbce56c270f7347f\": rpc error: code = NotFound desc = could not find container \"657bf93b8bae9ec7afe75fbc48ccfc45ccf9af760218e29cfbce56c270f7347f\": container with ID starting with 657bf93b8bae9ec7afe75fbc48ccfc45ccf9af760218e29cfbce56c270f7347f not found: ID does not exist" Oct 08 19:49:52 crc kubenswrapper[4988]: I1008 19:49:52.826682 4988 scope.go:117] "RemoveContainer" containerID="cad4dfedacebaed69682037b44f45eb1f5e008bde623318e3394ee03a8575c82" Oct 08 19:49:52 crc kubenswrapper[4988]: I1008 19:49:52.831138 4988 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cad4dfedacebaed69682037b44f45eb1f5e008bde623318e3394ee03a8575c82"} err="failed to get container status \"cad4dfedacebaed69682037b44f45eb1f5e008bde623318e3394ee03a8575c82\": rpc error: code = NotFound desc = could not find container \"cad4dfedacebaed69682037b44f45eb1f5e008bde623318e3394ee03a8575c82\": container with ID starting with cad4dfedacebaed69682037b44f45eb1f5e008bde623318e3394ee03a8575c82 not found: ID does not exist" Oct 08 19:49:52 crc kubenswrapper[4988]: I1008 19:49:52.831182 4988 scope.go:117] "RemoveContainer" containerID="657bf93b8bae9ec7afe75fbc48ccfc45ccf9af760218e29cfbce56c270f7347f" Oct 08 19:49:52 crc kubenswrapper[4988]: I1008 19:49:52.831287 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Oct 08 19:49:52 crc kubenswrapper[4988]: E1008 19:49:52.831682 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="62ad0fcb-bc01-493c-8fc6-4a186e4e197f" containerName="nova-metadata-metadata" Oct 08 19:49:52 crc kubenswrapper[4988]: I1008 19:49:52.831699 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="62ad0fcb-bc01-493c-8fc6-4a186e4e197f" containerName="nova-metadata-metadata" Oct 08 19:49:52 crc kubenswrapper[4988]: E1008 19:49:52.831716 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="984822c7-f554-4c5b-bba1-3cd471eee0f3" containerName="nova-api-api" Oct 08 19:49:52 crc kubenswrapper[4988]: I1008 19:49:52.831725 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="984822c7-f554-4c5b-bba1-3cd471eee0f3" containerName="nova-api-api" Oct 08 19:49:52 crc kubenswrapper[4988]: E1008 19:49:52.831743 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="62ad0fcb-bc01-493c-8fc6-4a186e4e197f" containerName="nova-metadata-log" Oct 08 19:49:52 crc kubenswrapper[4988]: I1008 19:49:52.831752 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="62ad0fcb-bc01-493c-8fc6-4a186e4e197f" containerName="nova-metadata-log" Oct 08 19:49:52 crc kubenswrapper[4988]: E1008 19:49:52.831770 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="586b0be5-17fd-4984-87ba-a8c2862ddbc0" containerName="nova-manage" Oct 08 19:49:52 crc kubenswrapper[4988]: I1008 19:49:52.831778 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="586b0be5-17fd-4984-87ba-a8c2862ddbc0" containerName="nova-manage" Oct 08 19:49:52 crc kubenswrapper[4988]: E1008 19:49:52.831789 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="984822c7-f554-4c5b-bba1-3cd471eee0f3" containerName="nova-api-log" Oct 08 19:49:52 crc kubenswrapper[4988]: I1008 19:49:52.831795 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="984822c7-f554-4c5b-bba1-3cd471eee0f3" containerName="nova-api-log" Oct 08 19:49:52 crc kubenswrapper[4988]: I1008 19:49:52.831957 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="62ad0fcb-bc01-493c-8fc6-4a186e4e197f" containerName="nova-metadata-metadata" Oct 08 19:49:52 crc kubenswrapper[4988]: I1008 19:49:52.831982 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="984822c7-f554-4c5b-bba1-3cd471eee0f3" containerName="nova-api-log" Oct 08 19:49:52 crc kubenswrapper[4988]: I1008 19:49:52.832000 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="586b0be5-17fd-4984-87ba-a8c2862ddbc0" containerName="nova-manage" Oct 08 19:49:52 crc kubenswrapper[4988]: I1008 19:49:52.832013 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="62ad0fcb-bc01-493c-8fc6-4a186e4e197f" containerName="nova-metadata-log" Oct 08 19:49:52 crc kubenswrapper[4988]: I1008 19:49:52.832025 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="984822c7-f554-4c5b-bba1-3cd471eee0f3" containerName="nova-api-api" Oct 08 19:49:52 crc kubenswrapper[4988]: I1008 19:49:52.833068 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 08 19:49:52 crc kubenswrapper[4988]: I1008 19:49:52.835075 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 08 19:49:52 crc kubenswrapper[4988]: I1008 19:49:52.836336 4988 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"657bf93b8bae9ec7afe75fbc48ccfc45ccf9af760218e29cfbce56c270f7347f"} err="failed to get container status \"657bf93b8bae9ec7afe75fbc48ccfc45ccf9af760218e29cfbce56c270f7347f\": rpc error: code = NotFound desc = could not find container \"657bf93b8bae9ec7afe75fbc48ccfc45ccf9af760218e29cfbce56c270f7347f\": container with ID starting with 657bf93b8bae9ec7afe75fbc48ccfc45ccf9af760218e29cfbce56c270f7347f not found: ID does not exist" Oct 08 19:49:52 crc kubenswrapper[4988]: I1008 19:49:52.836368 4988 scope.go:117] "RemoveContainer" containerID="9717aa641b94b3af71e62103a9ba8c139f01ae406cf769d77ea28bcd2a2a482e" Oct 08 19:49:52 crc kubenswrapper[4988]: I1008 19:49:52.836881 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Oct 08 19:49:52 crc kubenswrapper[4988]: I1008 19:49:52.842796 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Oct 08 19:49:52 crc kubenswrapper[4988]: I1008 19:49:52.844709 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 08 19:49:52 crc kubenswrapper[4988]: I1008 19:49:52.846659 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Oct 08 19:49:52 crc kubenswrapper[4988]: I1008 19:49:52.846944 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Oct 08 19:49:52 crc kubenswrapper[4988]: I1008 19:49:52.848782 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 08 19:49:52 crc kubenswrapper[4988]: I1008 19:49:52.880852 4988 scope.go:117] "RemoveContainer" containerID="c14335abef6035f10ff1740b717b228c578b085bebb38659384f1fc7518b10f1" Oct 08 19:49:52 crc kubenswrapper[4988]: I1008 19:49:52.899508 4988 scope.go:117] "RemoveContainer" containerID="9717aa641b94b3af71e62103a9ba8c139f01ae406cf769d77ea28bcd2a2a482e" Oct 08 19:49:52 crc kubenswrapper[4988]: E1008 19:49:52.899941 4988 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9717aa641b94b3af71e62103a9ba8c139f01ae406cf769d77ea28bcd2a2a482e\": container with ID starting with 9717aa641b94b3af71e62103a9ba8c139f01ae406cf769d77ea28bcd2a2a482e not found: ID does not exist" containerID="9717aa641b94b3af71e62103a9ba8c139f01ae406cf769d77ea28bcd2a2a482e" Oct 08 19:49:52 crc kubenswrapper[4988]: I1008 19:49:52.900148 4988 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9717aa641b94b3af71e62103a9ba8c139f01ae406cf769d77ea28bcd2a2a482e"} err="failed to get container status \"9717aa641b94b3af71e62103a9ba8c139f01ae406cf769d77ea28bcd2a2a482e\": rpc error: code = NotFound desc = could not find container \"9717aa641b94b3af71e62103a9ba8c139f01ae406cf769d77ea28bcd2a2a482e\": container with ID starting with 9717aa641b94b3af71e62103a9ba8c139f01ae406cf769d77ea28bcd2a2a482e not found: ID does not exist" Oct 08 19:49:52 crc kubenswrapper[4988]: I1008 19:49:52.900239 4988 scope.go:117] "RemoveContainer" containerID="c14335abef6035f10ff1740b717b228c578b085bebb38659384f1fc7518b10f1" Oct 08 19:49:52 crc kubenswrapper[4988]: E1008 19:49:52.900590 4988 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c14335abef6035f10ff1740b717b228c578b085bebb38659384f1fc7518b10f1\": container with ID starting with c14335abef6035f10ff1740b717b228c578b085bebb38659384f1fc7518b10f1 not found: ID does not exist" containerID="c14335abef6035f10ff1740b717b228c578b085bebb38659384f1fc7518b10f1" Oct 08 19:49:52 crc kubenswrapper[4988]: I1008 19:49:52.900620 4988 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c14335abef6035f10ff1740b717b228c578b085bebb38659384f1fc7518b10f1"} err="failed to get container status \"c14335abef6035f10ff1740b717b228c578b085bebb38659384f1fc7518b10f1\": rpc error: code = NotFound desc = could not find container \"c14335abef6035f10ff1740b717b228c578b085bebb38659384f1fc7518b10f1\": container with ID starting with c14335abef6035f10ff1740b717b228c578b085bebb38659384f1fc7518b10f1 not found: ID does not exist" Oct 08 19:49:52 crc kubenswrapper[4988]: I1008 19:49:52.900644 4988 scope.go:117] "RemoveContainer" containerID="9717aa641b94b3af71e62103a9ba8c139f01ae406cf769d77ea28bcd2a2a482e" Oct 08 19:49:52 crc kubenswrapper[4988]: I1008 19:49:52.900991 4988 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9717aa641b94b3af71e62103a9ba8c139f01ae406cf769d77ea28bcd2a2a482e"} err="failed to get container status \"9717aa641b94b3af71e62103a9ba8c139f01ae406cf769d77ea28bcd2a2a482e\": rpc error: code = NotFound desc = could not find container \"9717aa641b94b3af71e62103a9ba8c139f01ae406cf769d77ea28bcd2a2a482e\": container with ID starting with 9717aa641b94b3af71e62103a9ba8c139f01ae406cf769d77ea28bcd2a2a482e not found: ID does not exist" Oct 08 19:49:52 crc kubenswrapper[4988]: I1008 19:49:52.901021 4988 scope.go:117] "RemoveContainer" containerID="c14335abef6035f10ff1740b717b228c578b085bebb38659384f1fc7518b10f1" Oct 08 19:49:52 crc kubenswrapper[4988]: I1008 19:49:52.901204 4988 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c14335abef6035f10ff1740b717b228c578b085bebb38659384f1fc7518b10f1"} err="failed to get container status \"c14335abef6035f10ff1740b717b228c578b085bebb38659384f1fc7518b10f1\": rpc error: code = NotFound desc = could not find container \"c14335abef6035f10ff1740b717b228c578b085bebb38659384f1fc7518b10f1\": container with ID starting with c14335abef6035f10ff1740b717b228c578b085bebb38659384f1fc7518b10f1 not found: ID does not exist" Oct 08 19:49:52 crc kubenswrapper[4988]: I1008 19:49:52.906470 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/efa711f6-0608-4910-95c6-ca068c902dc1-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"efa711f6-0608-4910-95c6-ca068c902dc1\") " pod="openstack/nova-api-0" Oct 08 19:49:52 crc kubenswrapper[4988]: I1008 19:49:52.906518 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5f14d61e-922a-4af3-8970-3978205b32f6-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"5f14d61e-922a-4af3-8970-3978205b32f6\") " pod="openstack/nova-metadata-0" Oct 08 19:49:52 crc kubenswrapper[4988]: I1008 19:49:52.906553 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kt2k6\" (UniqueName: \"kubernetes.io/projected/efa711f6-0608-4910-95c6-ca068c902dc1-kube-api-access-kt2k6\") pod \"nova-api-0\" (UID: \"efa711f6-0608-4910-95c6-ca068c902dc1\") " pod="openstack/nova-api-0" Oct 08 19:49:52 crc kubenswrapper[4988]: I1008 19:49:52.906819 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5f14d61e-922a-4af3-8970-3978205b32f6-config-data\") pod \"nova-metadata-0\" (UID: \"5f14d61e-922a-4af3-8970-3978205b32f6\") " pod="openstack/nova-metadata-0" Oct 08 19:49:52 crc kubenswrapper[4988]: I1008 19:49:52.906870 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/efa711f6-0608-4910-95c6-ca068c902dc1-logs\") pod \"nova-api-0\" (UID: \"efa711f6-0608-4910-95c6-ca068c902dc1\") " pod="openstack/nova-api-0" Oct 08 19:49:52 crc kubenswrapper[4988]: I1008 19:49:52.906952 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/efa711f6-0608-4910-95c6-ca068c902dc1-config-data\") pod \"nova-api-0\" (UID: \"efa711f6-0608-4910-95c6-ca068c902dc1\") " pod="openstack/nova-api-0" Oct 08 19:49:52 crc kubenswrapper[4988]: I1008 19:49:52.906981 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5f14d61e-922a-4af3-8970-3978205b32f6-logs\") pod \"nova-metadata-0\" (UID: \"5f14d61e-922a-4af3-8970-3978205b32f6\") " pod="openstack/nova-metadata-0" Oct 08 19:49:52 crc kubenswrapper[4988]: I1008 19:49:52.907014 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/5f14d61e-922a-4af3-8970-3978205b32f6-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"5f14d61e-922a-4af3-8970-3978205b32f6\") " pod="openstack/nova-metadata-0" Oct 08 19:49:52 crc kubenswrapper[4988]: I1008 19:49:52.907044 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zxq4b\" (UniqueName: \"kubernetes.io/projected/5f14d61e-922a-4af3-8970-3978205b32f6-kube-api-access-zxq4b\") pod \"nova-metadata-0\" (UID: \"5f14d61e-922a-4af3-8970-3978205b32f6\") " pod="openstack/nova-metadata-0" Oct 08 19:49:53 crc kubenswrapper[4988]: I1008 19:49:53.008465 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zxq4b\" (UniqueName: \"kubernetes.io/projected/5f14d61e-922a-4af3-8970-3978205b32f6-kube-api-access-zxq4b\") pod \"nova-metadata-0\" (UID: \"5f14d61e-922a-4af3-8970-3978205b32f6\") " pod="openstack/nova-metadata-0" Oct 08 19:49:53 crc kubenswrapper[4988]: I1008 19:49:53.008551 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/efa711f6-0608-4910-95c6-ca068c902dc1-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"efa711f6-0608-4910-95c6-ca068c902dc1\") " pod="openstack/nova-api-0" Oct 08 19:49:53 crc kubenswrapper[4988]: I1008 19:49:53.008582 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5f14d61e-922a-4af3-8970-3978205b32f6-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"5f14d61e-922a-4af3-8970-3978205b32f6\") " pod="openstack/nova-metadata-0" Oct 08 19:49:53 crc kubenswrapper[4988]: I1008 19:49:53.008621 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kt2k6\" (UniqueName: \"kubernetes.io/projected/efa711f6-0608-4910-95c6-ca068c902dc1-kube-api-access-kt2k6\") pod \"nova-api-0\" (UID: \"efa711f6-0608-4910-95c6-ca068c902dc1\") " pod="openstack/nova-api-0" Oct 08 19:49:53 crc kubenswrapper[4988]: I1008 19:49:53.008698 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5f14d61e-922a-4af3-8970-3978205b32f6-config-data\") pod \"nova-metadata-0\" (UID: \"5f14d61e-922a-4af3-8970-3978205b32f6\") " pod="openstack/nova-metadata-0" Oct 08 19:49:53 crc kubenswrapper[4988]: I1008 19:49:53.008728 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/efa711f6-0608-4910-95c6-ca068c902dc1-logs\") pod \"nova-api-0\" (UID: \"efa711f6-0608-4910-95c6-ca068c902dc1\") " pod="openstack/nova-api-0" Oct 08 19:49:53 crc kubenswrapper[4988]: I1008 19:49:53.008761 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/efa711f6-0608-4910-95c6-ca068c902dc1-config-data\") pod \"nova-api-0\" (UID: \"efa711f6-0608-4910-95c6-ca068c902dc1\") " pod="openstack/nova-api-0" Oct 08 19:49:53 crc kubenswrapper[4988]: I1008 19:49:53.008784 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5f14d61e-922a-4af3-8970-3978205b32f6-logs\") pod \"nova-metadata-0\" (UID: \"5f14d61e-922a-4af3-8970-3978205b32f6\") " pod="openstack/nova-metadata-0" Oct 08 19:49:53 crc kubenswrapper[4988]: I1008 19:49:53.008826 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/5f14d61e-922a-4af3-8970-3978205b32f6-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"5f14d61e-922a-4af3-8970-3978205b32f6\") " pod="openstack/nova-metadata-0" Oct 08 19:49:53 crc kubenswrapper[4988]: I1008 19:49:53.009550 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/efa711f6-0608-4910-95c6-ca068c902dc1-logs\") pod \"nova-api-0\" (UID: \"efa711f6-0608-4910-95c6-ca068c902dc1\") " pod="openstack/nova-api-0" Oct 08 19:49:53 crc kubenswrapper[4988]: I1008 19:49:53.009754 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-5bf7c8d69c-j8v4t" Oct 08 19:49:53 crc kubenswrapper[4988]: I1008 19:49:53.010685 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5f14d61e-922a-4af3-8970-3978205b32f6-logs\") pod \"nova-metadata-0\" (UID: \"5f14d61e-922a-4af3-8970-3978205b32f6\") " pod="openstack/nova-metadata-0" Oct 08 19:49:53 crc kubenswrapper[4988]: I1008 19:49:53.014242 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/efa711f6-0608-4910-95c6-ca068c902dc1-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"efa711f6-0608-4910-95c6-ca068c902dc1\") " pod="openstack/nova-api-0" Oct 08 19:49:53 crc kubenswrapper[4988]: I1008 19:49:53.014513 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/efa711f6-0608-4910-95c6-ca068c902dc1-config-data\") pod \"nova-api-0\" (UID: \"efa711f6-0608-4910-95c6-ca068c902dc1\") " pod="openstack/nova-api-0" Oct 08 19:49:53 crc kubenswrapper[4988]: I1008 19:49:53.018812 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5f14d61e-922a-4af3-8970-3978205b32f6-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"5f14d61e-922a-4af3-8970-3978205b32f6\") " pod="openstack/nova-metadata-0" Oct 08 19:49:53 crc kubenswrapper[4988]: I1008 19:49:53.025337 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5f14d61e-922a-4af3-8970-3978205b32f6-config-data\") pod \"nova-metadata-0\" (UID: \"5f14d61e-922a-4af3-8970-3978205b32f6\") " pod="openstack/nova-metadata-0" Oct 08 19:49:53 crc kubenswrapper[4988]: I1008 19:49:53.028501 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kt2k6\" (UniqueName: \"kubernetes.io/projected/efa711f6-0608-4910-95c6-ca068c902dc1-kube-api-access-kt2k6\") pod \"nova-api-0\" (UID: \"efa711f6-0608-4910-95c6-ca068c902dc1\") " pod="openstack/nova-api-0" Oct 08 19:49:53 crc kubenswrapper[4988]: I1008 19:49:53.032676 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zxq4b\" (UniqueName: \"kubernetes.io/projected/5f14d61e-922a-4af3-8970-3978205b32f6-kube-api-access-zxq4b\") pod \"nova-metadata-0\" (UID: \"5f14d61e-922a-4af3-8970-3978205b32f6\") " pod="openstack/nova-metadata-0" Oct 08 19:49:53 crc kubenswrapper[4988]: I1008 19:49:53.043114 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/5f14d61e-922a-4af3-8970-3978205b32f6-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"5f14d61e-922a-4af3-8970-3978205b32f6\") " pod="openstack/nova-metadata-0" Oct 08 19:49:53 crc kubenswrapper[4988]: I1008 19:49:53.092437 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7595cd99bc-sq75r"] Oct 08 19:49:53 crc kubenswrapper[4988]: I1008 19:49:53.092729 4988 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-7595cd99bc-sq75r" podUID="98ad3c06-9746-432f-a4e8-8540a58404cb" containerName="dnsmasq-dns" containerID="cri-o://f8e480b1825776a01bf18e448caa9ac4c2322e267eff0611a05881ef60ff97d4" gracePeriod=10 Oct 08 19:49:53 crc kubenswrapper[4988]: I1008 19:49:53.178862 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 08 19:49:53 crc kubenswrapper[4988]: I1008 19:49:53.187344 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 08 19:49:53 crc kubenswrapper[4988]: I1008 19:49:53.269595 4988 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="62ad0fcb-bc01-493c-8fc6-4a186e4e197f" path="/var/lib/kubelet/pods/62ad0fcb-bc01-493c-8fc6-4a186e4e197f/volumes" Oct 08 19:49:53 crc kubenswrapper[4988]: I1008 19:49:53.270740 4988 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="984822c7-f554-4c5b-bba1-3cd471eee0f3" path="/var/lib/kubelet/pods/984822c7-f554-4c5b-bba1-3cd471eee0f3/volumes" Oct 08 19:49:53 crc kubenswrapper[4988]: I1008 19:49:53.442072 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Oct 08 19:49:53 crc kubenswrapper[4988]: I1008 19:49:53.545589 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7595cd99bc-sq75r" Oct 08 19:49:53 crc kubenswrapper[4988]: I1008 19:49:53.620681 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-85v9g\" (UniqueName: \"kubernetes.io/projected/98ad3c06-9746-432f-a4e8-8540a58404cb-kube-api-access-85v9g\") pod \"98ad3c06-9746-432f-a4e8-8540a58404cb\" (UID: \"98ad3c06-9746-432f-a4e8-8540a58404cb\") " Oct 08 19:49:53 crc kubenswrapper[4988]: I1008 19:49:53.620792 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/98ad3c06-9746-432f-a4e8-8540a58404cb-ovsdbserver-nb\") pod \"98ad3c06-9746-432f-a4e8-8540a58404cb\" (UID: \"98ad3c06-9746-432f-a4e8-8540a58404cb\") " Oct 08 19:49:53 crc kubenswrapper[4988]: I1008 19:49:53.620840 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/98ad3c06-9746-432f-a4e8-8540a58404cb-dns-svc\") pod \"98ad3c06-9746-432f-a4e8-8540a58404cb\" (UID: \"98ad3c06-9746-432f-a4e8-8540a58404cb\") " Oct 08 19:49:53 crc kubenswrapper[4988]: I1008 19:49:53.620871 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/98ad3c06-9746-432f-a4e8-8540a58404cb-config\") pod \"98ad3c06-9746-432f-a4e8-8540a58404cb\" (UID: \"98ad3c06-9746-432f-a4e8-8540a58404cb\") " Oct 08 19:49:53 crc kubenswrapper[4988]: I1008 19:49:53.620917 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/98ad3c06-9746-432f-a4e8-8540a58404cb-ovsdbserver-sb\") pod \"98ad3c06-9746-432f-a4e8-8540a58404cb\" (UID: \"98ad3c06-9746-432f-a4e8-8540a58404cb\") " Oct 08 19:49:53 crc kubenswrapper[4988]: I1008 19:49:53.626034 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/98ad3c06-9746-432f-a4e8-8540a58404cb-kube-api-access-85v9g" (OuterVolumeSpecName: "kube-api-access-85v9g") pod "98ad3c06-9746-432f-a4e8-8540a58404cb" (UID: "98ad3c06-9746-432f-a4e8-8540a58404cb"). InnerVolumeSpecName "kube-api-access-85v9g". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 19:49:53 crc kubenswrapper[4988]: I1008 19:49:53.674933 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/98ad3c06-9746-432f-a4e8-8540a58404cb-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "98ad3c06-9746-432f-a4e8-8540a58404cb" (UID: "98ad3c06-9746-432f-a4e8-8540a58404cb"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 19:49:53 crc kubenswrapper[4988]: I1008 19:49:53.678996 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/98ad3c06-9746-432f-a4e8-8540a58404cb-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "98ad3c06-9746-432f-a4e8-8540a58404cb" (UID: "98ad3c06-9746-432f-a4e8-8540a58404cb"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 19:49:53 crc kubenswrapper[4988]: I1008 19:49:53.679024 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/98ad3c06-9746-432f-a4e8-8540a58404cb-config" (OuterVolumeSpecName: "config") pod "98ad3c06-9746-432f-a4e8-8540a58404cb" (UID: "98ad3c06-9746-432f-a4e8-8540a58404cb"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 19:49:53 crc kubenswrapper[4988]: I1008 19:49:53.682818 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/98ad3c06-9746-432f-a4e8-8540a58404cb-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "98ad3c06-9746-432f-a4e8-8540a58404cb" (UID: "98ad3c06-9746-432f-a4e8-8540a58404cb"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 19:49:53 crc kubenswrapper[4988]: I1008 19:49:53.696782 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 08 19:49:53 crc kubenswrapper[4988]: I1008 19:49:53.723963 4988 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/98ad3c06-9746-432f-a4e8-8540a58404cb-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 08 19:49:53 crc kubenswrapper[4988]: I1008 19:49:53.724000 4988 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/98ad3c06-9746-432f-a4e8-8540a58404cb-config\") on node \"crc\" DevicePath \"\"" Oct 08 19:49:53 crc kubenswrapper[4988]: I1008 19:49:53.724026 4988 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/98ad3c06-9746-432f-a4e8-8540a58404cb-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 08 19:49:53 crc kubenswrapper[4988]: I1008 19:49:53.724041 4988 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-85v9g\" (UniqueName: \"kubernetes.io/projected/98ad3c06-9746-432f-a4e8-8540a58404cb-kube-api-access-85v9g\") on node \"crc\" DevicePath \"\"" Oct 08 19:49:53 crc kubenswrapper[4988]: I1008 19:49:53.724056 4988 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/98ad3c06-9746-432f-a4e8-8540a58404cb-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 08 19:49:53 crc kubenswrapper[4988]: I1008 19:49:53.738646 4988 generic.go:334] "Generic (PLEG): container finished" podID="98ad3c06-9746-432f-a4e8-8540a58404cb" containerID="f8e480b1825776a01bf18e448caa9ac4c2322e267eff0611a05881ef60ff97d4" exitCode=0 Oct 08 19:49:53 crc kubenswrapper[4988]: I1008 19:49:53.738704 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7595cd99bc-sq75r" event={"ID":"98ad3c06-9746-432f-a4e8-8540a58404cb","Type":"ContainerDied","Data":"f8e480b1825776a01bf18e448caa9ac4c2322e267eff0611a05881ef60ff97d4"} Oct 08 19:49:53 crc kubenswrapper[4988]: I1008 19:49:53.738730 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7595cd99bc-sq75r" event={"ID":"98ad3c06-9746-432f-a4e8-8540a58404cb","Type":"ContainerDied","Data":"775e237db11993d7f40d25781b28a945efb9ee387d3c88f7ff69bccbcfd039f4"} Oct 08 19:49:53 crc kubenswrapper[4988]: I1008 19:49:53.738746 4988 scope.go:117] "RemoveContainer" containerID="f8e480b1825776a01bf18e448caa9ac4c2322e267eff0611a05881ef60ff97d4" Oct 08 19:49:53 crc kubenswrapper[4988]: I1008 19:49:53.738785 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7595cd99bc-sq75r" Oct 08 19:49:53 crc kubenswrapper[4988]: I1008 19:49:53.740521 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"5f14d61e-922a-4af3-8970-3978205b32f6","Type":"ContainerStarted","Data":"077eda34e459005c1accac5d31da7759702a984edd6e74e264c42a777203051b"} Oct 08 19:49:53 crc kubenswrapper[4988]: I1008 19:49:53.774300 4988 scope.go:117] "RemoveContainer" containerID="42d1e21314688dabe77bbf2187b2f2ed13a25ff0f0a692a5589c46386a06fa7b" Oct 08 19:49:53 crc kubenswrapper[4988]: I1008 19:49:53.777620 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7595cd99bc-sq75r"] Oct 08 19:49:53 crc kubenswrapper[4988]: I1008 19:49:53.784332 4988 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-7595cd99bc-sq75r"] Oct 08 19:49:53 crc kubenswrapper[4988]: W1008 19:49:53.801747 4988 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podefa711f6_0608_4910_95c6_ca068c902dc1.slice/crio-05de85528b67ae1cb67fbedf33d49e52a6d5843733639d930233c975908e8530 WatchSource:0}: Error finding container 05de85528b67ae1cb67fbedf33d49e52a6d5843733639d930233c975908e8530: Status 404 returned error can't find the container with id 05de85528b67ae1cb67fbedf33d49e52a6d5843733639d930233c975908e8530 Oct 08 19:49:53 crc kubenswrapper[4988]: I1008 19:49:53.803720 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 08 19:49:53 crc kubenswrapper[4988]: I1008 19:49:53.914799 4988 scope.go:117] "RemoveContainer" containerID="f8e480b1825776a01bf18e448caa9ac4c2322e267eff0611a05881ef60ff97d4" Oct 08 19:49:53 crc kubenswrapper[4988]: E1008 19:49:53.915276 4988 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f8e480b1825776a01bf18e448caa9ac4c2322e267eff0611a05881ef60ff97d4\": container with ID starting with f8e480b1825776a01bf18e448caa9ac4c2322e267eff0611a05881ef60ff97d4 not found: ID does not exist" containerID="f8e480b1825776a01bf18e448caa9ac4c2322e267eff0611a05881ef60ff97d4" Oct 08 19:49:53 crc kubenswrapper[4988]: I1008 19:49:53.915306 4988 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f8e480b1825776a01bf18e448caa9ac4c2322e267eff0611a05881ef60ff97d4"} err="failed to get container status \"f8e480b1825776a01bf18e448caa9ac4c2322e267eff0611a05881ef60ff97d4\": rpc error: code = NotFound desc = could not find container \"f8e480b1825776a01bf18e448caa9ac4c2322e267eff0611a05881ef60ff97d4\": container with ID starting with f8e480b1825776a01bf18e448caa9ac4c2322e267eff0611a05881ef60ff97d4 not found: ID does not exist" Oct 08 19:49:53 crc kubenswrapper[4988]: I1008 19:49:53.915331 4988 scope.go:117] "RemoveContainer" containerID="42d1e21314688dabe77bbf2187b2f2ed13a25ff0f0a692a5589c46386a06fa7b" Oct 08 19:49:53 crc kubenswrapper[4988]: E1008 19:49:53.915758 4988 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"42d1e21314688dabe77bbf2187b2f2ed13a25ff0f0a692a5589c46386a06fa7b\": container with ID starting with 42d1e21314688dabe77bbf2187b2f2ed13a25ff0f0a692a5589c46386a06fa7b not found: ID does not exist" containerID="42d1e21314688dabe77bbf2187b2f2ed13a25ff0f0a692a5589c46386a06fa7b" Oct 08 19:49:53 crc kubenswrapper[4988]: I1008 19:49:53.915784 4988 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"42d1e21314688dabe77bbf2187b2f2ed13a25ff0f0a692a5589c46386a06fa7b"} err="failed to get container status \"42d1e21314688dabe77bbf2187b2f2ed13a25ff0f0a692a5589c46386a06fa7b\": rpc error: code = NotFound desc = could not find container \"42d1e21314688dabe77bbf2187b2f2ed13a25ff0f0a692a5589c46386a06fa7b\": container with ID starting with 42d1e21314688dabe77bbf2187b2f2ed13a25ff0f0a692a5589c46386a06fa7b not found: ID does not exist" Oct 08 19:49:54 crc kubenswrapper[4988]: I1008 19:49:54.753637 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"5f14d61e-922a-4af3-8970-3978205b32f6","Type":"ContainerStarted","Data":"e7720181df256d9211fee9d5c764d0cb9e14a49d34bf5a1cb91970c7ea88885a"} Oct 08 19:49:54 crc kubenswrapper[4988]: I1008 19:49:54.753942 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"5f14d61e-922a-4af3-8970-3978205b32f6","Type":"ContainerStarted","Data":"5ffd7570d857b0f82c9a1ebbfe4188c6574c9bb793a0e3c222ec6c6ecaeac7ec"} Oct 08 19:49:54 crc kubenswrapper[4988]: I1008 19:49:54.756318 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"efa711f6-0608-4910-95c6-ca068c902dc1","Type":"ContainerStarted","Data":"bf3c3be108df9c98c2da09b4764b05c791456a337938789fa421accdb32ad76b"} Oct 08 19:49:54 crc kubenswrapper[4988]: I1008 19:49:54.756364 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"efa711f6-0608-4910-95c6-ca068c902dc1","Type":"ContainerStarted","Data":"06b9b7565da09481b02fa8b303f324c8d0ca3b7bdd1787eed0d9426af1559000"} Oct 08 19:49:54 crc kubenswrapper[4988]: I1008 19:49:54.756377 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"efa711f6-0608-4910-95c6-ca068c902dc1","Type":"ContainerStarted","Data":"05de85528b67ae1cb67fbedf33d49e52a6d5843733639d930233c975908e8530"} Oct 08 19:49:54 crc kubenswrapper[4988]: I1008 19:49:54.788851 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.788822443 podStartE2EDuration="2.788822443s" podCreationTimestamp="2025-10-08 19:49:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 19:49:54.781320095 +0000 UTC m=+5940.231162905" watchObservedRunningTime="2025-10-08 19:49:54.788822443 +0000 UTC m=+5940.238665253" Oct 08 19:49:54 crc kubenswrapper[4988]: I1008 19:49:54.806710 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.806678591 podStartE2EDuration="2.806678591s" podCreationTimestamp="2025-10-08 19:49:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 19:49:54.802414526 +0000 UTC m=+5940.252257346" watchObservedRunningTime="2025-10-08 19:49:54.806678591 +0000 UTC m=+5940.256521401" Oct 08 19:49:55 crc kubenswrapper[4988]: I1008 19:49:55.248066 4988 scope.go:117] "RemoveContainer" containerID="f2a29ed22dd7a48bbc929e65fbeceed969a669e78f65fbfb724b03dc0db6e947" Oct 08 19:49:55 crc kubenswrapper[4988]: E1008 19:49:55.248623 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nm54f_openshift-machine-config-operator(f328a20b-5283-407e-8a93-2ab2da599f18)\"" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" Oct 08 19:49:55 crc kubenswrapper[4988]: I1008 19:49:55.258840 4988 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="98ad3c06-9746-432f-a4e8-8540a58404cb" path="/var/lib/kubelet/pods/98ad3c06-9746-432f-a4e8-8540a58404cb/volumes" Oct 08 19:49:56 crc kubenswrapper[4988]: I1008 19:49:56.078616 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-conductor-0" Oct 08 19:49:58 crc kubenswrapper[4988]: I1008 19:49:58.188378 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Oct 08 19:49:58 crc kubenswrapper[4988]: I1008 19:49:58.188454 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Oct 08 19:49:58 crc kubenswrapper[4988]: I1008 19:49:58.441268 4988 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-cell1-novncproxy-0" Oct 08 19:49:58 crc kubenswrapper[4988]: I1008 19:49:58.470159 4988 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-cell1-novncproxy-0" Oct 08 19:49:58 crc kubenswrapper[4988]: I1008 19:49:58.845191 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-novncproxy-0" Oct 08 19:49:59 crc kubenswrapper[4988]: I1008 19:49:59.035710 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-cell-mapping-lmgq7"] Oct 08 19:49:59 crc kubenswrapper[4988]: E1008 19:49:59.036081 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="98ad3c06-9746-432f-a4e8-8540a58404cb" containerName="dnsmasq-dns" Oct 08 19:49:59 crc kubenswrapper[4988]: I1008 19:49:59.036094 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="98ad3c06-9746-432f-a4e8-8540a58404cb" containerName="dnsmasq-dns" Oct 08 19:49:59 crc kubenswrapper[4988]: E1008 19:49:59.036113 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="98ad3c06-9746-432f-a4e8-8540a58404cb" containerName="init" Oct 08 19:49:59 crc kubenswrapper[4988]: I1008 19:49:59.036119 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="98ad3c06-9746-432f-a4e8-8540a58404cb" containerName="init" Oct 08 19:49:59 crc kubenswrapper[4988]: I1008 19:49:59.036303 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="98ad3c06-9746-432f-a4e8-8540a58404cb" containerName="dnsmasq-dns" Oct 08 19:49:59 crc kubenswrapper[4988]: I1008 19:49:59.036935 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-lmgq7" Oct 08 19:49:59 crc kubenswrapper[4988]: I1008 19:49:59.040250 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-config-data" Oct 08 19:49:59 crc kubenswrapper[4988]: I1008 19:49:59.043191 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-scripts" Oct 08 19:49:59 crc kubenswrapper[4988]: I1008 19:49:59.061818 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-lmgq7"] Oct 08 19:49:59 crc kubenswrapper[4988]: I1008 19:49:59.142976 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fa6bc882-bc50-4842-91a2-626c0e0329f4-scripts\") pod \"nova-cell1-cell-mapping-lmgq7\" (UID: \"fa6bc882-bc50-4842-91a2-626c0e0329f4\") " pod="openstack/nova-cell1-cell-mapping-lmgq7" Oct 08 19:49:59 crc kubenswrapper[4988]: I1008 19:49:59.143043 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5wf8b\" (UniqueName: \"kubernetes.io/projected/fa6bc882-bc50-4842-91a2-626c0e0329f4-kube-api-access-5wf8b\") pod \"nova-cell1-cell-mapping-lmgq7\" (UID: \"fa6bc882-bc50-4842-91a2-626c0e0329f4\") " pod="openstack/nova-cell1-cell-mapping-lmgq7" Oct 08 19:49:59 crc kubenswrapper[4988]: I1008 19:49:59.143128 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fa6bc882-bc50-4842-91a2-626c0e0329f4-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-lmgq7\" (UID: \"fa6bc882-bc50-4842-91a2-626c0e0329f4\") " pod="openstack/nova-cell1-cell-mapping-lmgq7" Oct 08 19:49:59 crc kubenswrapper[4988]: I1008 19:49:59.143214 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fa6bc882-bc50-4842-91a2-626c0e0329f4-config-data\") pod \"nova-cell1-cell-mapping-lmgq7\" (UID: \"fa6bc882-bc50-4842-91a2-626c0e0329f4\") " pod="openstack/nova-cell1-cell-mapping-lmgq7" Oct 08 19:49:59 crc kubenswrapper[4988]: I1008 19:49:59.245148 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fa6bc882-bc50-4842-91a2-626c0e0329f4-config-data\") pod \"nova-cell1-cell-mapping-lmgq7\" (UID: \"fa6bc882-bc50-4842-91a2-626c0e0329f4\") " pod="openstack/nova-cell1-cell-mapping-lmgq7" Oct 08 19:49:59 crc kubenswrapper[4988]: I1008 19:49:59.245268 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fa6bc882-bc50-4842-91a2-626c0e0329f4-scripts\") pod \"nova-cell1-cell-mapping-lmgq7\" (UID: \"fa6bc882-bc50-4842-91a2-626c0e0329f4\") " pod="openstack/nova-cell1-cell-mapping-lmgq7" Oct 08 19:49:59 crc kubenswrapper[4988]: I1008 19:49:59.245312 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5wf8b\" (UniqueName: \"kubernetes.io/projected/fa6bc882-bc50-4842-91a2-626c0e0329f4-kube-api-access-5wf8b\") pod \"nova-cell1-cell-mapping-lmgq7\" (UID: \"fa6bc882-bc50-4842-91a2-626c0e0329f4\") " pod="openstack/nova-cell1-cell-mapping-lmgq7" Oct 08 19:49:59 crc kubenswrapper[4988]: I1008 19:49:59.245482 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fa6bc882-bc50-4842-91a2-626c0e0329f4-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-lmgq7\" (UID: \"fa6bc882-bc50-4842-91a2-626c0e0329f4\") " pod="openstack/nova-cell1-cell-mapping-lmgq7" Oct 08 19:49:59 crc kubenswrapper[4988]: I1008 19:49:59.255130 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fa6bc882-bc50-4842-91a2-626c0e0329f4-config-data\") pod \"nova-cell1-cell-mapping-lmgq7\" (UID: \"fa6bc882-bc50-4842-91a2-626c0e0329f4\") " pod="openstack/nova-cell1-cell-mapping-lmgq7" Oct 08 19:49:59 crc kubenswrapper[4988]: I1008 19:49:59.255248 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fa6bc882-bc50-4842-91a2-626c0e0329f4-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-lmgq7\" (UID: \"fa6bc882-bc50-4842-91a2-626c0e0329f4\") " pod="openstack/nova-cell1-cell-mapping-lmgq7" Oct 08 19:49:59 crc kubenswrapper[4988]: I1008 19:49:59.256819 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fa6bc882-bc50-4842-91a2-626c0e0329f4-scripts\") pod \"nova-cell1-cell-mapping-lmgq7\" (UID: \"fa6bc882-bc50-4842-91a2-626c0e0329f4\") " pod="openstack/nova-cell1-cell-mapping-lmgq7" Oct 08 19:49:59 crc kubenswrapper[4988]: I1008 19:49:59.269926 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5wf8b\" (UniqueName: \"kubernetes.io/projected/fa6bc882-bc50-4842-91a2-626c0e0329f4-kube-api-access-5wf8b\") pod \"nova-cell1-cell-mapping-lmgq7\" (UID: \"fa6bc882-bc50-4842-91a2-626c0e0329f4\") " pod="openstack/nova-cell1-cell-mapping-lmgq7" Oct 08 19:49:59 crc kubenswrapper[4988]: I1008 19:49:59.355710 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-lmgq7" Oct 08 19:49:59 crc kubenswrapper[4988]: I1008 19:49:59.844456 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-lmgq7"] Oct 08 19:50:00 crc kubenswrapper[4988]: I1008 19:50:00.847009 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-lmgq7" event={"ID":"fa6bc882-bc50-4842-91a2-626c0e0329f4","Type":"ContainerStarted","Data":"2a2987792e85f659046e9d2bfb64d4bb2f1b90ded6eb9bc60cc5b8fe63002c6f"} Oct 08 19:50:00 crc kubenswrapper[4988]: I1008 19:50:00.847341 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-lmgq7" event={"ID":"fa6bc882-bc50-4842-91a2-626c0e0329f4","Type":"ContainerStarted","Data":"6a12aba568fdd944be7ed55a4c6272b623ea086a1cf604dffd55025a9968e29b"} Oct 08 19:50:00 crc kubenswrapper[4988]: I1008 19:50:00.872326 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-cell-mapping-lmgq7" podStartSLOduration=1.872311533 podStartE2EDuration="1.872311533s" podCreationTimestamp="2025-10-08 19:49:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 19:50:00.868013177 +0000 UTC m=+5946.317855957" watchObservedRunningTime="2025-10-08 19:50:00.872311533 +0000 UTC m=+5946.322154303" Oct 08 19:50:03 crc kubenswrapper[4988]: I1008 19:50:03.179892 4988 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 08 19:50:03 crc kubenswrapper[4988]: I1008 19:50:03.180476 4988 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 08 19:50:03 crc kubenswrapper[4988]: I1008 19:50:03.188078 4988 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Oct 08 19:50:03 crc kubenswrapper[4988]: I1008 19:50:03.188147 4988 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Oct 08 19:50:04 crc kubenswrapper[4988]: I1008 19:50:04.278616 4988 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="5f14d61e-922a-4af3-8970-3978205b32f6" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.1.94:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 08 19:50:04 crc kubenswrapper[4988]: I1008 19:50:04.278748 4988 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="efa711f6-0608-4910-95c6-ca068c902dc1" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.1.93:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 08 19:50:04 crc kubenswrapper[4988]: I1008 19:50:04.278923 4988 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="5f14d61e-922a-4af3-8970-3978205b32f6" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.1.94:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 08 19:50:04 crc kubenswrapper[4988]: I1008 19:50:04.278601 4988 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="efa711f6-0608-4910-95c6-ca068c902dc1" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.1.93:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 08 19:50:04 crc kubenswrapper[4988]: I1008 19:50:04.280592 4988 scope.go:117] "RemoveContainer" containerID="7e3cc924ecc9eb8641f9e92f699dd1e29dd612ae90bccba932dfffe406eccc65" Oct 08 19:50:04 crc kubenswrapper[4988]: I1008 19:50:04.323439 4988 scope.go:117] "RemoveContainer" containerID="6edb06c5df4c60b50bb54de502320e580ab5ebc2c6963187cdda1b95895a15b3" Oct 08 19:50:04 crc kubenswrapper[4988]: I1008 19:50:04.898190 4988 generic.go:334] "Generic (PLEG): container finished" podID="fa6bc882-bc50-4842-91a2-626c0e0329f4" containerID="2a2987792e85f659046e9d2bfb64d4bb2f1b90ded6eb9bc60cc5b8fe63002c6f" exitCode=0 Oct 08 19:50:04 crc kubenswrapper[4988]: I1008 19:50:04.898300 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-lmgq7" event={"ID":"fa6bc882-bc50-4842-91a2-626c0e0329f4","Type":"ContainerDied","Data":"2a2987792e85f659046e9d2bfb64d4bb2f1b90ded6eb9bc60cc5b8fe63002c6f"} Oct 08 19:50:06 crc kubenswrapper[4988]: I1008 19:50:06.239001 4988 scope.go:117] "RemoveContainer" containerID="f2a29ed22dd7a48bbc929e65fbeceed969a669e78f65fbfb724b03dc0db6e947" Oct 08 19:50:06 crc kubenswrapper[4988]: E1008 19:50:06.239972 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nm54f_openshift-machine-config-operator(f328a20b-5283-407e-8a93-2ab2da599f18)\"" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" Oct 08 19:50:06 crc kubenswrapper[4988]: I1008 19:50:06.361275 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-lmgq7" Oct 08 19:50:06 crc kubenswrapper[4988]: I1008 19:50:06.429127 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fa6bc882-bc50-4842-91a2-626c0e0329f4-combined-ca-bundle\") pod \"fa6bc882-bc50-4842-91a2-626c0e0329f4\" (UID: \"fa6bc882-bc50-4842-91a2-626c0e0329f4\") " Oct 08 19:50:06 crc kubenswrapper[4988]: I1008 19:50:06.429203 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5wf8b\" (UniqueName: \"kubernetes.io/projected/fa6bc882-bc50-4842-91a2-626c0e0329f4-kube-api-access-5wf8b\") pod \"fa6bc882-bc50-4842-91a2-626c0e0329f4\" (UID: \"fa6bc882-bc50-4842-91a2-626c0e0329f4\") " Oct 08 19:50:06 crc kubenswrapper[4988]: I1008 19:50:06.429402 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fa6bc882-bc50-4842-91a2-626c0e0329f4-scripts\") pod \"fa6bc882-bc50-4842-91a2-626c0e0329f4\" (UID: \"fa6bc882-bc50-4842-91a2-626c0e0329f4\") " Oct 08 19:50:06 crc kubenswrapper[4988]: I1008 19:50:06.429498 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fa6bc882-bc50-4842-91a2-626c0e0329f4-config-data\") pod \"fa6bc882-bc50-4842-91a2-626c0e0329f4\" (UID: \"fa6bc882-bc50-4842-91a2-626c0e0329f4\") " Oct 08 19:50:06 crc kubenswrapper[4988]: I1008 19:50:06.435871 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fa6bc882-bc50-4842-91a2-626c0e0329f4-scripts" (OuterVolumeSpecName: "scripts") pod "fa6bc882-bc50-4842-91a2-626c0e0329f4" (UID: "fa6bc882-bc50-4842-91a2-626c0e0329f4"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 19:50:06 crc kubenswrapper[4988]: I1008 19:50:06.444427 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fa6bc882-bc50-4842-91a2-626c0e0329f4-kube-api-access-5wf8b" (OuterVolumeSpecName: "kube-api-access-5wf8b") pod "fa6bc882-bc50-4842-91a2-626c0e0329f4" (UID: "fa6bc882-bc50-4842-91a2-626c0e0329f4"). InnerVolumeSpecName "kube-api-access-5wf8b". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 19:50:06 crc kubenswrapper[4988]: I1008 19:50:06.475912 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fa6bc882-bc50-4842-91a2-626c0e0329f4-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "fa6bc882-bc50-4842-91a2-626c0e0329f4" (UID: "fa6bc882-bc50-4842-91a2-626c0e0329f4"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 19:50:06 crc kubenswrapper[4988]: I1008 19:50:06.479698 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fa6bc882-bc50-4842-91a2-626c0e0329f4-config-data" (OuterVolumeSpecName: "config-data") pod "fa6bc882-bc50-4842-91a2-626c0e0329f4" (UID: "fa6bc882-bc50-4842-91a2-626c0e0329f4"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 19:50:06 crc kubenswrapper[4988]: I1008 19:50:06.532081 4988 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fa6bc882-bc50-4842-91a2-626c0e0329f4-scripts\") on node \"crc\" DevicePath \"\"" Oct 08 19:50:06 crc kubenswrapper[4988]: I1008 19:50:06.532128 4988 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fa6bc882-bc50-4842-91a2-626c0e0329f4-config-data\") on node \"crc\" DevicePath \"\"" Oct 08 19:50:06 crc kubenswrapper[4988]: I1008 19:50:06.532146 4988 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fa6bc882-bc50-4842-91a2-626c0e0329f4-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 19:50:06 crc kubenswrapper[4988]: I1008 19:50:06.532161 4988 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5wf8b\" (UniqueName: \"kubernetes.io/projected/fa6bc882-bc50-4842-91a2-626c0e0329f4-kube-api-access-5wf8b\") on node \"crc\" DevicePath \"\"" Oct 08 19:50:06 crc kubenswrapper[4988]: I1008 19:50:06.926522 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-lmgq7" event={"ID":"fa6bc882-bc50-4842-91a2-626c0e0329f4","Type":"ContainerDied","Data":"6a12aba568fdd944be7ed55a4c6272b623ea086a1cf604dffd55025a9968e29b"} Oct 08 19:50:06 crc kubenswrapper[4988]: I1008 19:50:06.926588 4988 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6a12aba568fdd944be7ed55a4c6272b623ea086a1cf604dffd55025a9968e29b" Oct 08 19:50:06 crc kubenswrapper[4988]: I1008 19:50:06.926596 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-lmgq7" Oct 08 19:50:07 crc kubenswrapper[4988]: I1008 19:50:07.135168 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 08 19:50:07 crc kubenswrapper[4988]: I1008 19:50:07.135777 4988 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="efa711f6-0608-4910-95c6-ca068c902dc1" containerName="nova-api-log" containerID="cri-o://06b9b7565da09481b02fa8b303f324c8d0ca3b7bdd1787eed0d9426af1559000" gracePeriod=30 Oct 08 19:50:07 crc kubenswrapper[4988]: I1008 19:50:07.135897 4988 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="efa711f6-0608-4910-95c6-ca068c902dc1" containerName="nova-api-api" containerID="cri-o://bf3c3be108df9c98c2da09b4764b05c791456a337938789fa421accdb32ad76b" gracePeriod=30 Oct 08 19:50:07 crc kubenswrapper[4988]: I1008 19:50:07.210079 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Oct 08 19:50:07 crc kubenswrapper[4988]: I1008 19:50:07.210322 4988 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="5f14d61e-922a-4af3-8970-3978205b32f6" containerName="nova-metadata-log" containerID="cri-o://5ffd7570d857b0f82c9a1ebbfe4188c6574c9bb793a0e3c222ec6c6ecaeac7ec" gracePeriod=30 Oct 08 19:50:07 crc kubenswrapper[4988]: I1008 19:50:07.210470 4988 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="5f14d61e-922a-4af3-8970-3978205b32f6" containerName="nova-metadata-metadata" containerID="cri-o://e7720181df256d9211fee9d5c764d0cb9e14a49d34bf5a1cb91970c7ea88885a" gracePeriod=30 Oct 08 19:50:07 crc kubenswrapper[4988]: I1008 19:50:07.940565 4988 generic.go:334] "Generic (PLEG): container finished" podID="5f14d61e-922a-4af3-8970-3978205b32f6" containerID="5ffd7570d857b0f82c9a1ebbfe4188c6574c9bb793a0e3c222ec6c6ecaeac7ec" exitCode=143 Oct 08 19:50:07 crc kubenswrapper[4988]: I1008 19:50:07.940673 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"5f14d61e-922a-4af3-8970-3978205b32f6","Type":"ContainerDied","Data":"5ffd7570d857b0f82c9a1ebbfe4188c6574c9bb793a0e3c222ec6c6ecaeac7ec"} Oct 08 19:50:07 crc kubenswrapper[4988]: I1008 19:50:07.943713 4988 generic.go:334] "Generic (PLEG): container finished" podID="efa711f6-0608-4910-95c6-ca068c902dc1" containerID="06b9b7565da09481b02fa8b303f324c8d0ca3b7bdd1787eed0d9426af1559000" exitCode=143 Oct 08 19:50:07 crc kubenswrapper[4988]: I1008 19:50:07.943751 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"efa711f6-0608-4910-95c6-ca068c902dc1","Type":"ContainerDied","Data":"06b9b7565da09481b02fa8b303f324c8d0ca3b7bdd1787eed0d9426af1559000"} Oct 08 19:50:10 crc kubenswrapper[4988]: I1008 19:50:10.802660 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 08 19:50:10 crc kubenswrapper[4988]: I1008 19:50:10.808773 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 08 19:50:10 crc kubenswrapper[4988]: I1008 19:50:10.928786 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5f14d61e-922a-4af3-8970-3978205b32f6-logs\") pod \"5f14d61e-922a-4af3-8970-3978205b32f6\" (UID: \"5f14d61e-922a-4af3-8970-3978205b32f6\") " Oct 08 19:50:10 crc kubenswrapper[4988]: I1008 19:50:10.929053 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/efa711f6-0608-4910-95c6-ca068c902dc1-logs\") pod \"efa711f6-0608-4910-95c6-ca068c902dc1\" (UID: \"efa711f6-0608-4910-95c6-ca068c902dc1\") " Oct 08 19:50:10 crc kubenswrapper[4988]: I1008 19:50:10.929119 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5f14d61e-922a-4af3-8970-3978205b32f6-combined-ca-bundle\") pod \"5f14d61e-922a-4af3-8970-3978205b32f6\" (UID: \"5f14d61e-922a-4af3-8970-3978205b32f6\") " Oct 08 19:50:10 crc kubenswrapper[4988]: I1008 19:50:10.929180 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/5f14d61e-922a-4af3-8970-3978205b32f6-nova-metadata-tls-certs\") pod \"5f14d61e-922a-4af3-8970-3978205b32f6\" (UID: \"5f14d61e-922a-4af3-8970-3978205b32f6\") " Oct 08 19:50:10 crc kubenswrapper[4988]: I1008 19:50:10.929285 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kt2k6\" (UniqueName: \"kubernetes.io/projected/efa711f6-0608-4910-95c6-ca068c902dc1-kube-api-access-kt2k6\") pod \"efa711f6-0608-4910-95c6-ca068c902dc1\" (UID: \"efa711f6-0608-4910-95c6-ca068c902dc1\") " Oct 08 19:50:10 crc kubenswrapper[4988]: I1008 19:50:10.929336 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zxq4b\" (UniqueName: \"kubernetes.io/projected/5f14d61e-922a-4af3-8970-3978205b32f6-kube-api-access-zxq4b\") pod \"5f14d61e-922a-4af3-8970-3978205b32f6\" (UID: \"5f14d61e-922a-4af3-8970-3978205b32f6\") " Oct 08 19:50:10 crc kubenswrapper[4988]: I1008 19:50:10.929459 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5f14d61e-922a-4af3-8970-3978205b32f6-config-data\") pod \"5f14d61e-922a-4af3-8970-3978205b32f6\" (UID: \"5f14d61e-922a-4af3-8970-3978205b32f6\") " Oct 08 19:50:10 crc kubenswrapper[4988]: I1008 19:50:10.929713 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/efa711f6-0608-4910-95c6-ca068c902dc1-config-data\") pod \"efa711f6-0608-4910-95c6-ca068c902dc1\" (UID: \"efa711f6-0608-4910-95c6-ca068c902dc1\") " Oct 08 19:50:10 crc kubenswrapper[4988]: I1008 19:50:10.929820 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/efa711f6-0608-4910-95c6-ca068c902dc1-combined-ca-bundle\") pod \"efa711f6-0608-4910-95c6-ca068c902dc1\" (UID: \"efa711f6-0608-4910-95c6-ca068c902dc1\") " Oct 08 19:50:10 crc kubenswrapper[4988]: I1008 19:50:10.930041 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5f14d61e-922a-4af3-8970-3978205b32f6-logs" (OuterVolumeSpecName: "logs") pod "5f14d61e-922a-4af3-8970-3978205b32f6" (UID: "5f14d61e-922a-4af3-8970-3978205b32f6"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 19:50:10 crc kubenswrapper[4988]: I1008 19:50:10.931171 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/efa711f6-0608-4910-95c6-ca068c902dc1-logs" (OuterVolumeSpecName: "logs") pod "efa711f6-0608-4910-95c6-ca068c902dc1" (UID: "efa711f6-0608-4910-95c6-ca068c902dc1"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 19:50:10 crc kubenswrapper[4988]: I1008 19:50:10.931171 4988 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5f14d61e-922a-4af3-8970-3978205b32f6-logs\") on node \"crc\" DevicePath \"\"" Oct 08 19:50:10 crc kubenswrapper[4988]: I1008 19:50:10.935603 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/efa711f6-0608-4910-95c6-ca068c902dc1-kube-api-access-kt2k6" (OuterVolumeSpecName: "kube-api-access-kt2k6") pod "efa711f6-0608-4910-95c6-ca068c902dc1" (UID: "efa711f6-0608-4910-95c6-ca068c902dc1"). InnerVolumeSpecName "kube-api-access-kt2k6". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 19:50:10 crc kubenswrapper[4988]: I1008 19:50:10.935949 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5f14d61e-922a-4af3-8970-3978205b32f6-kube-api-access-zxq4b" (OuterVolumeSpecName: "kube-api-access-zxq4b") pod "5f14d61e-922a-4af3-8970-3978205b32f6" (UID: "5f14d61e-922a-4af3-8970-3978205b32f6"). InnerVolumeSpecName "kube-api-access-zxq4b". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 19:50:10 crc kubenswrapper[4988]: I1008 19:50:10.975976 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/efa711f6-0608-4910-95c6-ca068c902dc1-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "efa711f6-0608-4910-95c6-ca068c902dc1" (UID: "efa711f6-0608-4910-95c6-ca068c902dc1"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 19:50:10 crc kubenswrapper[4988]: I1008 19:50:10.980651 4988 generic.go:334] "Generic (PLEG): container finished" podID="5f14d61e-922a-4af3-8970-3978205b32f6" containerID="e7720181df256d9211fee9d5c764d0cb9e14a49d34bf5a1cb91970c7ea88885a" exitCode=0 Oct 08 19:50:10 crc kubenswrapper[4988]: I1008 19:50:10.980733 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"5f14d61e-922a-4af3-8970-3978205b32f6","Type":"ContainerDied","Data":"e7720181df256d9211fee9d5c764d0cb9e14a49d34bf5a1cb91970c7ea88885a"} Oct 08 19:50:10 crc kubenswrapper[4988]: I1008 19:50:10.980769 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5f14d61e-922a-4af3-8970-3978205b32f6-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "5f14d61e-922a-4af3-8970-3978205b32f6" (UID: "5f14d61e-922a-4af3-8970-3978205b32f6"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 19:50:10 crc kubenswrapper[4988]: I1008 19:50:10.980796 4988 scope.go:117] "RemoveContainer" containerID="e7720181df256d9211fee9d5c764d0cb9e14a49d34bf5a1cb91970c7ea88885a" Oct 08 19:50:10 crc kubenswrapper[4988]: I1008 19:50:10.980784 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"5f14d61e-922a-4af3-8970-3978205b32f6","Type":"ContainerDied","Data":"077eda34e459005c1accac5d31da7759702a984edd6e74e264c42a777203051b"} Oct 08 19:50:10 crc kubenswrapper[4988]: I1008 19:50:10.981238 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 08 19:50:10 crc kubenswrapper[4988]: I1008 19:50:10.981306 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5f14d61e-922a-4af3-8970-3978205b32f6-config-data" (OuterVolumeSpecName: "config-data") pod "5f14d61e-922a-4af3-8970-3978205b32f6" (UID: "5f14d61e-922a-4af3-8970-3978205b32f6"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 19:50:10 crc kubenswrapper[4988]: I1008 19:50:10.985740 4988 generic.go:334] "Generic (PLEG): container finished" podID="efa711f6-0608-4910-95c6-ca068c902dc1" containerID="bf3c3be108df9c98c2da09b4764b05c791456a337938789fa421accdb32ad76b" exitCode=0 Oct 08 19:50:10 crc kubenswrapper[4988]: I1008 19:50:10.985798 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"efa711f6-0608-4910-95c6-ca068c902dc1","Type":"ContainerDied","Data":"bf3c3be108df9c98c2da09b4764b05c791456a337938789fa421accdb32ad76b"} Oct 08 19:50:10 crc kubenswrapper[4988]: I1008 19:50:10.985833 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"efa711f6-0608-4910-95c6-ca068c902dc1","Type":"ContainerDied","Data":"05de85528b67ae1cb67fbedf33d49e52a6d5843733639d930233c975908e8530"} Oct 08 19:50:10 crc kubenswrapper[4988]: I1008 19:50:10.986018 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 08 19:50:10 crc kubenswrapper[4988]: I1008 19:50:10.991449 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/efa711f6-0608-4910-95c6-ca068c902dc1-config-data" (OuterVolumeSpecName: "config-data") pod "efa711f6-0608-4910-95c6-ca068c902dc1" (UID: "efa711f6-0608-4910-95c6-ca068c902dc1"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 19:50:11 crc kubenswrapper[4988]: I1008 19:50:11.008595 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5f14d61e-922a-4af3-8970-3978205b32f6-nova-metadata-tls-certs" (OuterVolumeSpecName: "nova-metadata-tls-certs") pod "5f14d61e-922a-4af3-8970-3978205b32f6" (UID: "5f14d61e-922a-4af3-8970-3978205b32f6"). InnerVolumeSpecName "nova-metadata-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 19:50:11 crc kubenswrapper[4988]: I1008 19:50:11.022155 4988 scope.go:117] "RemoveContainer" containerID="5ffd7570d857b0f82c9a1ebbfe4188c6574c9bb793a0e3c222ec6c6ecaeac7ec" Oct 08 19:50:11 crc kubenswrapper[4988]: I1008 19:50:11.035844 4988 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/efa711f6-0608-4910-95c6-ca068c902dc1-logs\") on node \"crc\" DevicePath \"\"" Oct 08 19:50:11 crc kubenswrapper[4988]: I1008 19:50:11.035895 4988 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5f14d61e-922a-4af3-8970-3978205b32f6-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 19:50:11 crc kubenswrapper[4988]: I1008 19:50:11.035916 4988 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/5f14d61e-922a-4af3-8970-3978205b32f6-nova-metadata-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 08 19:50:11 crc kubenswrapper[4988]: I1008 19:50:11.035938 4988 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kt2k6\" (UniqueName: \"kubernetes.io/projected/efa711f6-0608-4910-95c6-ca068c902dc1-kube-api-access-kt2k6\") on node \"crc\" DevicePath \"\"" Oct 08 19:50:11 crc kubenswrapper[4988]: I1008 19:50:11.035955 4988 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zxq4b\" (UniqueName: \"kubernetes.io/projected/5f14d61e-922a-4af3-8970-3978205b32f6-kube-api-access-zxq4b\") on node \"crc\" DevicePath \"\"" Oct 08 19:50:11 crc kubenswrapper[4988]: I1008 19:50:11.035971 4988 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5f14d61e-922a-4af3-8970-3978205b32f6-config-data\") on node \"crc\" DevicePath \"\"" Oct 08 19:50:11 crc kubenswrapper[4988]: I1008 19:50:11.035988 4988 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/efa711f6-0608-4910-95c6-ca068c902dc1-config-data\") on node \"crc\" DevicePath \"\"" Oct 08 19:50:11 crc kubenswrapper[4988]: I1008 19:50:11.036004 4988 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/efa711f6-0608-4910-95c6-ca068c902dc1-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 19:50:11 crc kubenswrapper[4988]: I1008 19:50:11.050514 4988 scope.go:117] "RemoveContainer" containerID="e7720181df256d9211fee9d5c764d0cb9e14a49d34bf5a1cb91970c7ea88885a" Oct 08 19:50:11 crc kubenswrapper[4988]: E1008 19:50:11.051221 4988 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e7720181df256d9211fee9d5c764d0cb9e14a49d34bf5a1cb91970c7ea88885a\": container with ID starting with e7720181df256d9211fee9d5c764d0cb9e14a49d34bf5a1cb91970c7ea88885a not found: ID does not exist" containerID="e7720181df256d9211fee9d5c764d0cb9e14a49d34bf5a1cb91970c7ea88885a" Oct 08 19:50:11 crc kubenswrapper[4988]: I1008 19:50:11.051274 4988 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e7720181df256d9211fee9d5c764d0cb9e14a49d34bf5a1cb91970c7ea88885a"} err="failed to get container status \"e7720181df256d9211fee9d5c764d0cb9e14a49d34bf5a1cb91970c7ea88885a\": rpc error: code = NotFound desc = could not find container \"e7720181df256d9211fee9d5c764d0cb9e14a49d34bf5a1cb91970c7ea88885a\": container with ID starting with e7720181df256d9211fee9d5c764d0cb9e14a49d34bf5a1cb91970c7ea88885a not found: ID does not exist" Oct 08 19:50:11 crc kubenswrapper[4988]: I1008 19:50:11.051314 4988 scope.go:117] "RemoveContainer" containerID="5ffd7570d857b0f82c9a1ebbfe4188c6574c9bb793a0e3c222ec6c6ecaeac7ec" Oct 08 19:50:11 crc kubenswrapper[4988]: E1008 19:50:11.051769 4988 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5ffd7570d857b0f82c9a1ebbfe4188c6574c9bb793a0e3c222ec6c6ecaeac7ec\": container with ID starting with 5ffd7570d857b0f82c9a1ebbfe4188c6574c9bb793a0e3c222ec6c6ecaeac7ec not found: ID does not exist" containerID="5ffd7570d857b0f82c9a1ebbfe4188c6574c9bb793a0e3c222ec6c6ecaeac7ec" Oct 08 19:50:11 crc kubenswrapper[4988]: I1008 19:50:11.051820 4988 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5ffd7570d857b0f82c9a1ebbfe4188c6574c9bb793a0e3c222ec6c6ecaeac7ec"} err="failed to get container status \"5ffd7570d857b0f82c9a1ebbfe4188c6574c9bb793a0e3c222ec6c6ecaeac7ec\": rpc error: code = NotFound desc = could not find container \"5ffd7570d857b0f82c9a1ebbfe4188c6574c9bb793a0e3c222ec6c6ecaeac7ec\": container with ID starting with 5ffd7570d857b0f82c9a1ebbfe4188c6574c9bb793a0e3c222ec6c6ecaeac7ec not found: ID does not exist" Oct 08 19:50:11 crc kubenswrapper[4988]: I1008 19:50:11.051847 4988 scope.go:117] "RemoveContainer" containerID="bf3c3be108df9c98c2da09b4764b05c791456a337938789fa421accdb32ad76b" Oct 08 19:50:11 crc kubenswrapper[4988]: I1008 19:50:11.073260 4988 scope.go:117] "RemoveContainer" containerID="06b9b7565da09481b02fa8b303f324c8d0ca3b7bdd1787eed0d9426af1559000" Oct 08 19:50:11 crc kubenswrapper[4988]: I1008 19:50:11.100607 4988 scope.go:117] "RemoveContainer" containerID="bf3c3be108df9c98c2da09b4764b05c791456a337938789fa421accdb32ad76b" Oct 08 19:50:11 crc kubenswrapper[4988]: E1008 19:50:11.101010 4988 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bf3c3be108df9c98c2da09b4764b05c791456a337938789fa421accdb32ad76b\": container with ID starting with bf3c3be108df9c98c2da09b4764b05c791456a337938789fa421accdb32ad76b not found: ID does not exist" containerID="bf3c3be108df9c98c2da09b4764b05c791456a337938789fa421accdb32ad76b" Oct 08 19:50:11 crc kubenswrapper[4988]: I1008 19:50:11.101048 4988 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bf3c3be108df9c98c2da09b4764b05c791456a337938789fa421accdb32ad76b"} err="failed to get container status \"bf3c3be108df9c98c2da09b4764b05c791456a337938789fa421accdb32ad76b\": rpc error: code = NotFound desc = could not find container \"bf3c3be108df9c98c2da09b4764b05c791456a337938789fa421accdb32ad76b\": container with ID starting with bf3c3be108df9c98c2da09b4764b05c791456a337938789fa421accdb32ad76b not found: ID does not exist" Oct 08 19:50:11 crc kubenswrapper[4988]: I1008 19:50:11.101074 4988 scope.go:117] "RemoveContainer" containerID="06b9b7565da09481b02fa8b303f324c8d0ca3b7bdd1787eed0d9426af1559000" Oct 08 19:50:11 crc kubenswrapper[4988]: E1008 19:50:11.101907 4988 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"06b9b7565da09481b02fa8b303f324c8d0ca3b7bdd1787eed0d9426af1559000\": container with ID starting with 06b9b7565da09481b02fa8b303f324c8d0ca3b7bdd1787eed0d9426af1559000 not found: ID does not exist" containerID="06b9b7565da09481b02fa8b303f324c8d0ca3b7bdd1787eed0d9426af1559000" Oct 08 19:50:11 crc kubenswrapper[4988]: I1008 19:50:11.101933 4988 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"06b9b7565da09481b02fa8b303f324c8d0ca3b7bdd1787eed0d9426af1559000"} err="failed to get container status \"06b9b7565da09481b02fa8b303f324c8d0ca3b7bdd1787eed0d9426af1559000\": rpc error: code = NotFound desc = could not find container \"06b9b7565da09481b02fa8b303f324c8d0ca3b7bdd1787eed0d9426af1559000\": container with ID starting with 06b9b7565da09481b02fa8b303f324c8d0ca3b7bdd1787eed0d9426af1559000 not found: ID does not exist" Oct 08 19:50:11 crc kubenswrapper[4988]: I1008 19:50:11.331190 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Oct 08 19:50:11 crc kubenswrapper[4988]: I1008 19:50:11.360700 4988 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Oct 08 19:50:11 crc kubenswrapper[4988]: E1008 19:50:11.369703 4988 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podefa711f6_0608_4910_95c6_ca068c902dc1.slice/crio-05de85528b67ae1cb67fbedf33d49e52a6d5843733639d930233c975908e8530\": RecentStats: unable to find data in memory cache]" Oct 08 19:50:11 crc kubenswrapper[4988]: I1008 19:50:11.378672 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 08 19:50:11 crc kubenswrapper[4988]: I1008 19:50:11.381212 4988 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Oct 08 19:50:11 crc kubenswrapper[4988]: I1008 19:50:11.388449 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Oct 08 19:50:11 crc kubenswrapper[4988]: E1008 19:50:11.388964 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="efa711f6-0608-4910-95c6-ca068c902dc1" containerName="nova-api-log" Oct 08 19:50:11 crc kubenswrapper[4988]: I1008 19:50:11.388988 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="efa711f6-0608-4910-95c6-ca068c902dc1" containerName="nova-api-log" Oct 08 19:50:11 crc kubenswrapper[4988]: E1008 19:50:11.389018 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5f14d61e-922a-4af3-8970-3978205b32f6" containerName="nova-metadata-log" Oct 08 19:50:11 crc kubenswrapper[4988]: I1008 19:50:11.389027 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="5f14d61e-922a-4af3-8970-3978205b32f6" containerName="nova-metadata-log" Oct 08 19:50:11 crc kubenswrapper[4988]: E1008 19:50:11.389045 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fa6bc882-bc50-4842-91a2-626c0e0329f4" containerName="nova-manage" Oct 08 19:50:11 crc kubenswrapper[4988]: I1008 19:50:11.389055 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="fa6bc882-bc50-4842-91a2-626c0e0329f4" containerName="nova-manage" Oct 08 19:50:11 crc kubenswrapper[4988]: E1008 19:50:11.389084 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5f14d61e-922a-4af3-8970-3978205b32f6" containerName="nova-metadata-metadata" Oct 08 19:50:11 crc kubenswrapper[4988]: I1008 19:50:11.389092 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="5f14d61e-922a-4af3-8970-3978205b32f6" containerName="nova-metadata-metadata" Oct 08 19:50:11 crc kubenswrapper[4988]: E1008 19:50:11.389104 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="efa711f6-0608-4910-95c6-ca068c902dc1" containerName="nova-api-api" Oct 08 19:50:11 crc kubenswrapper[4988]: I1008 19:50:11.389113 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="efa711f6-0608-4910-95c6-ca068c902dc1" containerName="nova-api-api" Oct 08 19:50:11 crc kubenswrapper[4988]: I1008 19:50:11.389337 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="5f14d61e-922a-4af3-8970-3978205b32f6" containerName="nova-metadata-log" Oct 08 19:50:11 crc kubenswrapper[4988]: I1008 19:50:11.389356 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="fa6bc882-bc50-4842-91a2-626c0e0329f4" containerName="nova-manage" Oct 08 19:50:11 crc kubenswrapper[4988]: I1008 19:50:11.389368 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="5f14d61e-922a-4af3-8970-3978205b32f6" containerName="nova-metadata-metadata" Oct 08 19:50:11 crc kubenswrapper[4988]: I1008 19:50:11.389401 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="efa711f6-0608-4910-95c6-ca068c902dc1" containerName="nova-api-log" Oct 08 19:50:11 crc kubenswrapper[4988]: I1008 19:50:11.389415 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="efa711f6-0608-4910-95c6-ca068c902dc1" containerName="nova-api-api" Oct 08 19:50:11 crc kubenswrapper[4988]: I1008 19:50:11.390604 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 08 19:50:11 crc kubenswrapper[4988]: I1008 19:50:11.392769 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Oct 08 19:50:11 crc kubenswrapper[4988]: I1008 19:50:11.393211 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Oct 08 19:50:11 crc kubenswrapper[4988]: I1008 19:50:11.396504 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 08 19:50:11 crc kubenswrapper[4988]: I1008 19:50:11.404422 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Oct 08 19:50:11 crc kubenswrapper[4988]: I1008 19:50:11.406501 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 08 19:50:11 crc kubenswrapper[4988]: I1008 19:50:11.408002 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Oct 08 19:50:11 crc kubenswrapper[4988]: I1008 19:50:11.412608 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 08 19:50:11 crc kubenswrapper[4988]: I1008 19:50:11.443621 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/96d0365d-bfac-45d3-85e6-a89e540e433f-config-data\") pod \"nova-metadata-0\" (UID: \"96d0365d-bfac-45d3-85e6-a89e540e433f\") " pod="openstack/nova-metadata-0" Oct 08 19:50:11 crc kubenswrapper[4988]: I1008 19:50:11.443675 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qwzhn\" (UniqueName: \"kubernetes.io/projected/96d0365d-bfac-45d3-85e6-a89e540e433f-kube-api-access-qwzhn\") pod \"nova-metadata-0\" (UID: \"96d0365d-bfac-45d3-85e6-a89e540e433f\") " pod="openstack/nova-metadata-0" Oct 08 19:50:11 crc kubenswrapper[4988]: I1008 19:50:11.443714 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/96d0365d-bfac-45d3-85e6-a89e540e433f-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"96d0365d-bfac-45d3-85e6-a89e540e433f\") " pod="openstack/nova-metadata-0" Oct 08 19:50:11 crc kubenswrapper[4988]: I1008 19:50:11.443886 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/96d0365d-bfac-45d3-85e6-a89e540e433f-logs\") pod \"nova-metadata-0\" (UID: \"96d0365d-bfac-45d3-85e6-a89e540e433f\") " pod="openstack/nova-metadata-0" Oct 08 19:50:11 crc kubenswrapper[4988]: I1008 19:50:11.443983 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/96d0365d-bfac-45d3-85e6-a89e540e433f-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"96d0365d-bfac-45d3-85e6-a89e540e433f\") " pod="openstack/nova-metadata-0" Oct 08 19:50:11 crc kubenswrapper[4988]: I1008 19:50:11.545779 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/96d0365d-bfac-45d3-85e6-a89e540e433f-logs\") pod \"nova-metadata-0\" (UID: \"96d0365d-bfac-45d3-85e6-a89e540e433f\") " pod="openstack/nova-metadata-0" Oct 08 19:50:11 crc kubenswrapper[4988]: I1008 19:50:11.545856 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/96d0365d-bfac-45d3-85e6-a89e540e433f-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"96d0365d-bfac-45d3-85e6-a89e540e433f\") " pod="openstack/nova-metadata-0" Oct 08 19:50:11 crc kubenswrapper[4988]: I1008 19:50:11.545954 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f6d28178-c627-4be2-b5c1-d660f8f2dbb4-logs\") pod \"nova-api-0\" (UID: \"f6d28178-c627-4be2-b5c1-d660f8f2dbb4\") " pod="openstack/nova-api-0" Oct 08 19:50:11 crc kubenswrapper[4988]: I1008 19:50:11.546021 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f6d28178-c627-4be2-b5c1-d660f8f2dbb4-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"f6d28178-c627-4be2-b5c1-d660f8f2dbb4\") " pod="openstack/nova-api-0" Oct 08 19:50:11 crc kubenswrapper[4988]: I1008 19:50:11.546056 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/96d0365d-bfac-45d3-85e6-a89e540e433f-config-data\") pod \"nova-metadata-0\" (UID: \"96d0365d-bfac-45d3-85e6-a89e540e433f\") " pod="openstack/nova-metadata-0" Oct 08 19:50:11 crc kubenswrapper[4988]: I1008 19:50:11.546084 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f6d28178-c627-4be2-b5c1-d660f8f2dbb4-config-data\") pod \"nova-api-0\" (UID: \"f6d28178-c627-4be2-b5c1-d660f8f2dbb4\") " pod="openstack/nova-api-0" Oct 08 19:50:11 crc kubenswrapper[4988]: I1008 19:50:11.546108 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qwzhn\" (UniqueName: \"kubernetes.io/projected/96d0365d-bfac-45d3-85e6-a89e540e433f-kube-api-access-qwzhn\") pod \"nova-metadata-0\" (UID: \"96d0365d-bfac-45d3-85e6-a89e540e433f\") " pod="openstack/nova-metadata-0" Oct 08 19:50:11 crc kubenswrapper[4988]: I1008 19:50:11.546143 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rtndv\" (UniqueName: \"kubernetes.io/projected/f6d28178-c627-4be2-b5c1-d660f8f2dbb4-kube-api-access-rtndv\") pod \"nova-api-0\" (UID: \"f6d28178-c627-4be2-b5c1-d660f8f2dbb4\") " pod="openstack/nova-api-0" Oct 08 19:50:11 crc kubenswrapper[4988]: I1008 19:50:11.546167 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/96d0365d-bfac-45d3-85e6-a89e540e433f-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"96d0365d-bfac-45d3-85e6-a89e540e433f\") " pod="openstack/nova-metadata-0" Oct 08 19:50:11 crc kubenswrapper[4988]: I1008 19:50:11.546409 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/96d0365d-bfac-45d3-85e6-a89e540e433f-logs\") pod \"nova-metadata-0\" (UID: \"96d0365d-bfac-45d3-85e6-a89e540e433f\") " pod="openstack/nova-metadata-0" Oct 08 19:50:11 crc kubenswrapper[4988]: I1008 19:50:11.551284 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/96d0365d-bfac-45d3-85e6-a89e540e433f-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"96d0365d-bfac-45d3-85e6-a89e540e433f\") " pod="openstack/nova-metadata-0" Oct 08 19:50:11 crc kubenswrapper[4988]: I1008 19:50:11.551880 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/96d0365d-bfac-45d3-85e6-a89e540e433f-config-data\") pod \"nova-metadata-0\" (UID: \"96d0365d-bfac-45d3-85e6-a89e540e433f\") " pod="openstack/nova-metadata-0" Oct 08 19:50:11 crc kubenswrapper[4988]: I1008 19:50:11.559042 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/96d0365d-bfac-45d3-85e6-a89e540e433f-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"96d0365d-bfac-45d3-85e6-a89e540e433f\") " pod="openstack/nova-metadata-0" Oct 08 19:50:11 crc kubenswrapper[4988]: I1008 19:50:11.566440 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qwzhn\" (UniqueName: \"kubernetes.io/projected/96d0365d-bfac-45d3-85e6-a89e540e433f-kube-api-access-qwzhn\") pod \"nova-metadata-0\" (UID: \"96d0365d-bfac-45d3-85e6-a89e540e433f\") " pod="openstack/nova-metadata-0" Oct 08 19:50:11 crc kubenswrapper[4988]: I1008 19:50:11.647277 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f6d28178-c627-4be2-b5c1-d660f8f2dbb4-logs\") pod \"nova-api-0\" (UID: \"f6d28178-c627-4be2-b5c1-d660f8f2dbb4\") " pod="openstack/nova-api-0" Oct 08 19:50:11 crc kubenswrapper[4988]: I1008 19:50:11.647717 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f6d28178-c627-4be2-b5c1-d660f8f2dbb4-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"f6d28178-c627-4be2-b5c1-d660f8f2dbb4\") " pod="openstack/nova-api-0" Oct 08 19:50:11 crc kubenswrapper[4988]: I1008 19:50:11.647765 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f6d28178-c627-4be2-b5c1-d660f8f2dbb4-config-data\") pod \"nova-api-0\" (UID: \"f6d28178-c627-4be2-b5c1-d660f8f2dbb4\") " pod="openstack/nova-api-0" Oct 08 19:50:11 crc kubenswrapper[4988]: I1008 19:50:11.647807 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rtndv\" (UniqueName: \"kubernetes.io/projected/f6d28178-c627-4be2-b5c1-d660f8f2dbb4-kube-api-access-rtndv\") pod \"nova-api-0\" (UID: \"f6d28178-c627-4be2-b5c1-d660f8f2dbb4\") " pod="openstack/nova-api-0" Oct 08 19:50:11 crc kubenswrapper[4988]: I1008 19:50:11.648539 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f6d28178-c627-4be2-b5c1-d660f8f2dbb4-logs\") pod \"nova-api-0\" (UID: \"f6d28178-c627-4be2-b5c1-d660f8f2dbb4\") " pod="openstack/nova-api-0" Oct 08 19:50:11 crc kubenswrapper[4988]: I1008 19:50:11.651232 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f6d28178-c627-4be2-b5c1-d660f8f2dbb4-config-data\") pod \"nova-api-0\" (UID: \"f6d28178-c627-4be2-b5c1-d660f8f2dbb4\") " pod="openstack/nova-api-0" Oct 08 19:50:11 crc kubenswrapper[4988]: I1008 19:50:11.651691 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f6d28178-c627-4be2-b5c1-d660f8f2dbb4-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"f6d28178-c627-4be2-b5c1-d660f8f2dbb4\") " pod="openstack/nova-api-0" Oct 08 19:50:11 crc kubenswrapper[4988]: I1008 19:50:11.665730 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rtndv\" (UniqueName: \"kubernetes.io/projected/f6d28178-c627-4be2-b5c1-d660f8f2dbb4-kube-api-access-rtndv\") pod \"nova-api-0\" (UID: \"f6d28178-c627-4be2-b5c1-d660f8f2dbb4\") " pod="openstack/nova-api-0" Oct 08 19:50:11 crc kubenswrapper[4988]: I1008 19:50:11.750333 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 08 19:50:11 crc kubenswrapper[4988]: I1008 19:50:11.757829 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 08 19:50:12 crc kubenswrapper[4988]: I1008 19:50:12.267583 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 08 19:50:12 crc kubenswrapper[4988]: W1008 19:50:12.271710 4988 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod96d0365d_bfac_45d3_85e6_a89e540e433f.slice/crio-d7128b4e27ffb9ac708e3296be5bc84babf586304faad2e0d0444570805873b1 WatchSource:0}: Error finding container d7128b4e27ffb9ac708e3296be5bc84babf586304faad2e0d0444570805873b1: Status 404 returned error can't find the container with id d7128b4e27ffb9ac708e3296be5bc84babf586304faad2e0d0444570805873b1 Oct 08 19:50:12 crc kubenswrapper[4988]: W1008 19:50:12.278198 4988 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf6d28178_c627_4be2_b5c1_d660f8f2dbb4.slice/crio-f18d9cd976e668b8113026acae7688e30b7fa9600f3b9425bae52a7ebdecf937 WatchSource:0}: Error finding container f18d9cd976e668b8113026acae7688e30b7fa9600f3b9425bae52a7ebdecf937: Status 404 returned error can't find the container with id f18d9cd976e668b8113026acae7688e30b7fa9600f3b9425bae52a7ebdecf937 Oct 08 19:50:12 crc kubenswrapper[4988]: I1008 19:50:12.278313 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 08 19:50:13 crc kubenswrapper[4988]: I1008 19:50:13.010232 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"96d0365d-bfac-45d3-85e6-a89e540e433f","Type":"ContainerStarted","Data":"f3c33b39236a5e6ebb4d7f65e628002c11c1202f89454a370f1f038359f92254"} Oct 08 19:50:13 crc kubenswrapper[4988]: I1008 19:50:13.010633 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"96d0365d-bfac-45d3-85e6-a89e540e433f","Type":"ContainerStarted","Data":"183c4a8fb4867be027b4eaa6827471b499f33dddf4c08c32851630fbe21ef74c"} Oct 08 19:50:13 crc kubenswrapper[4988]: I1008 19:50:13.010655 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"96d0365d-bfac-45d3-85e6-a89e540e433f","Type":"ContainerStarted","Data":"d7128b4e27ffb9ac708e3296be5bc84babf586304faad2e0d0444570805873b1"} Oct 08 19:50:13 crc kubenswrapper[4988]: I1008 19:50:13.014744 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"f6d28178-c627-4be2-b5c1-d660f8f2dbb4","Type":"ContainerStarted","Data":"a291e0fd40de41d8c18058f44c29dcc41afba6d653817a42371a2fa8ae0d896c"} Oct 08 19:50:13 crc kubenswrapper[4988]: I1008 19:50:13.014812 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"f6d28178-c627-4be2-b5c1-d660f8f2dbb4","Type":"ContainerStarted","Data":"87d4a0bd1cbfe9bee62db4b538afe1dd7810577476cb342c2fe3a8fe7e4d4ae8"} Oct 08 19:50:13 crc kubenswrapper[4988]: I1008 19:50:13.014833 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"f6d28178-c627-4be2-b5c1-d660f8f2dbb4","Type":"ContainerStarted","Data":"f18d9cd976e668b8113026acae7688e30b7fa9600f3b9425bae52a7ebdecf937"} Oct 08 19:50:13 crc kubenswrapper[4988]: I1008 19:50:13.066327 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.066309126 podStartE2EDuration="2.066309126s" podCreationTimestamp="2025-10-08 19:50:11 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 19:50:13.038618026 +0000 UTC m=+5958.488460796" watchObservedRunningTime="2025-10-08 19:50:13.066309126 +0000 UTC m=+5958.516151906" Oct 08 19:50:13 crc kubenswrapper[4988]: I1008 19:50:13.073911 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.073896257 podStartE2EDuration="2.073896257s" podCreationTimestamp="2025-10-08 19:50:11 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 19:50:13.064325012 +0000 UTC m=+5958.514167792" watchObservedRunningTime="2025-10-08 19:50:13.073896257 +0000 UTC m=+5958.523739037" Oct 08 19:50:13 crc kubenswrapper[4988]: I1008 19:50:13.260373 4988 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5f14d61e-922a-4af3-8970-3978205b32f6" path="/var/lib/kubelet/pods/5f14d61e-922a-4af3-8970-3978205b32f6/volumes" Oct 08 19:50:13 crc kubenswrapper[4988]: I1008 19:50:13.261820 4988 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="efa711f6-0608-4910-95c6-ca068c902dc1" path="/var/lib/kubelet/pods/efa711f6-0608-4910-95c6-ca068c902dc1/volumes" Oct 08 19:50:16 crc kubenswrapper[4988]: I1008 19:50:16.751290 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Oct 08 19:50:16 crc kubenswrapper[4988]: I1008 19:50:16.751538 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Oct 08 19:50:21 crc kubenswrapper[4988]: I1008 19:50:21.237652 4988 scope.go:117] "RemoveContainer" containerID="f2a29ed22dd7a48bbc929e65fbeceed969a669e78f65fbfb724b03dc0db6e947" Oct 08 19:50:21 crc kubenswrapper[4988]: E1008 19:50:21.238285 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nm54f_openshift-machine-config-operator(f328a20b-5283-407e-8a93-2ab2da599f18)\"" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" Oct 08 19:50:21 crc kubenswrapper[4988]: I1008 19:50:21.750683 4988 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Oct 08 19:50:21 crc kubenswrapper[4988]: I1008 19:50:21.750772 4988 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Oct 08 19:50:21 crc kubenswrapper[4988]: I1008 19:50:21.758443 4988 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 08 19:50:21 crc kubenswrapper[4988]: I1008 19:50:21.758501 4988 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 08 19:50:22 crc kubenswrapper[4988]: I1008 19:50:22.117074 4988 generic.go:334] "Generic (PLEG): container finished" podID="2ceb61b1-8b93-4d44-b27f-ac96c4566be6" containerID="95a43697a44a1358aec0023227996593f0f41fa671f8322b52c7d43b9e0dbc7b" exitCode=137 Oct 08 19:50:22 crc kubenswrapper[4988]: I1008 19:50:22.117409 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"2ceb61b1-8b93-4d44-b27f-ac96c4566be6","Type":"ContainerDied","Data":"95a43697a44a1358aec0023227996593f0f41fa671f8322b52c7d43b9e0dbc7b"} Oct 08 19:50:22 crc kubenswrapper[4988]: I1008 19:50:22.317068 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 08 19:50:22 crc kubenswrapper[4988]: I1008 19:50:22.391651 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2ceb61b1-8b93-4d44-b27f-ac96c4566be6-combined-ca-bundle\") pod \"2ceb61b1-8b93-4d44-b27f-ac96c4566be6\" (UID: \"2ceb61b1-8b93-4d44-b27f-ac96c4566be6\") " Oct 08 19:50:22 crc kubenswrapper[4988]: I1008 19:50:22.391811 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2ceb61b1-8b93-4d44-b27f-ac96c4566be6-config-data\") pod \"2ceb61b1-8b93-4d44-b27f-ac96c4566be6\" (UID: \"2ceb61b1-8b93-4d44-b27f-ac96c4566be6\") " Oct 08 19:50:22 crc kubenswrapper[4988]: I1008 19:50:22.391869 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rrmn2\" (UniqueName: \"kubernetes.io/projected/2ceb61b1-8b93-4d44-b27f-ac96c4566be6-kube-api-access-rrmn2\") pod \"2ceb61b1-8b93-4d44-b27f-ac96c4566be6\" (UID: \"2ceb61b1-8b93-4d44-b27f-ac96c4566be6\") " Oct 08 19:50:22 crc kubenswrapper[4988]: I1008 19:50:22.396502 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2ceb61b1-8b93-4d44-b27f-ac96c4566be6-kube-api-access-rrmn2" (OuterVolumeSpecName: "kube-api-access-rrmn2") pod "2ceb61b1-8b93-4d44-b27f-ac96c4566be6" (UID: "2ceb61b1-8b93-4d44-b27f-ac96c4566be6"). InnerVolumeSpecName "kube-api-access-rrmn2". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 19:50:22 crc kubenswrapper[4988]: I1008 19:50:22.424110 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2ceb61b1-8b93-4d44-b27f-ac96c4566be6-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "2ceb61b1-8b93-4d44-b27f-ac96c4566be6" (UID: "2ceb61b1-8b93-4d44-b27f-ac96c4566be6"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 19:50:22 crc kubenswrapper[4988]: I1008 19:50:22.441756 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2ceb61b1-8b93-4d44-b27f-ac96c4566be6-config-data" (OuterVolumeSpecName: "config-data") pod "2ceb61b1-8b93-4d44-b27f-ac96c4566be6" (UID: "2ceb61b1-8b93-4d44-b27f-ac96c4566be6"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 19:50:22 crc kubenswrapper[4988]: I1008 19:50:22.495837 4988 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2ceb61b1-8b93-4d44-b27f-ac96c4566be6-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 19:50:22 crc kubenswrapper[4988]: I1008 19:50:22.495993 4988 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2ceb61b1-8b93-4d44-b27f-ac96c4566be6-config-data\") on node \"crc\" DevicePath \"\"" Oct 08 19:50:22 crc kubenswrapper[4988]: I1008 19:50:22.496022 4988 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rrmn2\" (UniqueName: \"kubernetes.io/projected/2ceb61b1-8b93-4d44-b27f-ac96c4566be6-kube-api-access-rrmn2\") on node \"crc\" DevicePath \"\"" Oct 08 19:50:22 crc kubenswrapper[4988]: I1008 19:50:22.762671 4988 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="96d0365d-bfac-45d3-85e6-a89e540e433f" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.1.96:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 08 19:50:22 crc kubenswrapper[4988]: I1008 19:50:22.762713 4988 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="96d0365d-bfac-45d3-85e6-a89e540e433f" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.1.96:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 08 19:50:22 crc kubenswrapper[4988]: I1008 19:50:22.845532 4988 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="f6d28178-c627-4be2-b5c1-d660f8f2dbb4" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.1.97:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 08 19:50:22 crc kubenswrapper[4988]: I1008 19:50:22.845820 4988 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="f6d28178-c627-4be2-b5c1-d660f8f2dbb4" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.1.97:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 08 19:50:23 crc kubenswrapper[4988]: I1008 19:50:23.127258 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"2ceb61b1-8b93-4d44-b27f-ac96c4566be6","Type":"ContainerDied","Data":"36e84c38650d1dfc479d96ff045e9d44a4924ddd01761ecdd740b8a34e3952b0"} Oct 08 19:50:23 crc kubenswrapper[4988]: I1008 19:50:23.127315 4988 scope.go:117] "RemoveContainer" containerID="95a43697a44a1358aec0023227996593f0f41fa671f8322b52c7d43b9e0dbc7b" Oct 08 19:50:23 crc kubenswrapper[4988]: I1008 19:50:23.127461 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 08 19:50:23 crc kubenswrapper[4988]: I1008 19:50:23.175578 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Oct 08 19:50:23 crc kubenswrapper[4988]: I1008 19:50:23.185104 4988 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Oct 08 19:50:23 crc kubenswrapper[4988]: I1008 19:50:23.200111 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Oct 08 19:50:23 crc kubenswrapper[4988]: E1008 19:50:23.200513 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2ceb61b1-8b93-4d44-b27f-ac96c4566be6" containerName="nova-scheduler-scheduler" Oct 08 19:50:23 crc kubenswrapper[4988]: I1008 19:50:23.200536 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="2ceb61b1-8b93-4d44-b27f-ac96c4566be6" containerName="nova-scheduler-scheduler" Oct 08 19:50:23 crc kubenswrapper[4988]: I1008 19:50:23.200778 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="2ceb61b1-8b93-4d44-b27f-ac96c4566be6" containerName="nova-scheduler-scheduler" Oct 08 19:50:23 crc kubenswrapper[4988]: I1008 19:50:23.201560 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 08 19:50:23 crc kubenswrapper[4988]: I1008 19:50:23.204020 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Oct 08 19:50:23 crc kubenswrapper[4988]: I1008 19:50:23.227494 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Oct 08 19:50:23 crc kubenswrapper[4988]: I1008 19:50:23.259085 4988 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2ceb61b1-8b93-4d44-b27f-ac96c4566be6" path="/var/lib/kubelet/pods/2ceb61b1-8b93-4d44-b27f-ac96c4566be6/volumes" Oct 08 19:50:23 crc kubenswrapper[4988]: I1008 19:50:23.313978 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ca7e4fbf-efe7-4ae8-b85c-104d24f83f95-config-data\") pod \"nova-scheduler-0\" (UID: \"ca7e4fbf-efe7-4ae8-b85c-104d24f83f95\") " pod="openstack/nova-scheduler-0" Oct 08 19:50:23 crc kubenswrapper[4988]: I1008 19:50:23.314041 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gmtwd\" (UniqueName: \"kubernetes.io/projected/ca7e4fbf-efe7-4ae8-b85c-104d24f83f95-kube-api-access-gmtwd\") pod \"nova-scheduler-0\" (UID: \"ca7e4fbf-efe7-4ae8-b85c-104d24f83f95\") " pod="openstack/nova-scheduler-0" Oct 08 19:50:23 crc kubenswrapper[4988]: I1008 19:50:23.314225 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ca7e4fbf-efe7-4ae8-b85c-104d24f83f95-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"ca7e4fbf-efe7-4ae8-b85c-104d24f83f95\") " pod="openstack/nova-scheduler-0" Oct 08 19:50:23 crc kubenswrapper[4988]: I1008 19:50:23.417320 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ca7e4fbf-efe7-4ae8-b85c-104d24f83f95-config-data\") pod \"nova-scheduler-0\" (UID: \"ca7e4fbf-efe7-4ae8-b85c-104d24f83f95\") " pod="openstack/nova-scheduler-0" Oct 08 19:50:23 crc kubenswrapper[4988]: I1008 19:50:23.418514 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gmtwd\" (UniqueName: \"kubernetes.io/projected/ca7e4fbf-efe7-4ae8-b85c-104d24f83f95-kube-api-access-gmtwd\") pod \"nova-scheduler-0\" (UID: \"ca7e4fbf-efe7-4ae8-b85c-104d24f83f95\") " pod="openstack/nova-scheduler-0" Oct 08 19:50:23 crc kubenswrapper[4988]: I1008 19:50:23.418904 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ca7e4fbf-efe7-4ae8-b85c-104d24f83f95-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"ca7e4fbf-efe7-4ae8-b85c-104d24f83f95\") " pod="openstack/nova-scheduler-0" Oct 08 19:50:23 crc kubenswrapper[4988]: I1008 19:50:23.422478 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ca7e4fbf-efe7-4ae8-b85c-104d24f83f95-config-data\") pod \"nova-scheduler-0\" (UID: \"ca7e4fbf-efe7-4ae8-b85c-104d24f83f95\") " pod="openstack/nova-scheduler-0" Oct 08 19:50:23 crc kubenswrapper[4988]: I1008 19:50:23.425834 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ca7e4fbf-efe7-4ae8-b85c-104d24f83f95-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"ca7e4fbf-efe7-4ae8-b85c-104d24f83f95\") " pod="openstack/nova-scheduler-0" Oct 08 19:50:23 crc kubenswrapper[4988]: I1008 19:50:23.439342 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gmtwd\" (UniqueName: \"kubernetes.io/projected/ca7e4fbf-efe7-4ae8-b85c-104d24f83f95-kube-api-access-gmtwd\") pod \"nova-scheduler-0\" (UID: \"ca7e4fbf-efe7-4ae8-b85c-104d24f83f95\") " pod="openstack/nova-scheduler-0" Oct 08 19:50:23 crc kubenswrapper[4988]: I1008 19:50:23.520619 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 08 19:50:24 crc kubenswrapper[4988]: I1008 19:50:24.040936 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Oct 08 19:50:24 crc kubenswrapper[4988]: W1008 19:50:24.041983 4988 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podca7e4fbf_efe7_4ae8_b85c_104d24f83f95.slice/crio-b681f0306d45f7110985bb8e895a96f594a212828bde833223fca8c5337ef990 WatchSource:0}: Error finding container b681f0306d45f7110985bb8e895a96f594a212828bde833223fca8c5337ef990: Status 404 returned error can't find the container with id b681f0306d45f7110985bb8e895a96f594a212828bde833223fca8c5337ef990 Oct 08 19:50:24 crc kubenswrapper[4988]: I1008 19:50:24.150914 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"ca7e4fbf-efe7-4ae8-b85c-104d24f83f95","Type":"ContainerStarted","Data":"b681f0306d45f7110985bb8e895a96f594a212828bde833223fca8c5337ef990"} Oct 08 19:50:25 crc kubenswrapper[4988]: I1008 19:50:25.170928 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"ca7e4fbf-efe7-4ae8-b85c-104d24f83f95","Type":"ContainerStarted","Data":"5c2dfbd0d5dd3874131abf5da252d865ed64608cd12ff2aad33f7474101e0229"} Oct 08 19:50:25 crc kubenswrapper[4988]: I1008 19:50:25.195209 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.195179806 podStartE2EDuration="2.195179806s" podCreationTimestamp="2025-10-08 19:50:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 19:50:25.190814188 +0000 UTC m=+5970.640656998" watchObservedRunningTime="2025-10-08 19:50:25.195179806 +0000 UTC m=+5970.645022616" Oct 08 19:50:28 crc kubenswrapper[4988]: I1008 19:50:28.521425 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Oct 08 19:50:31 crc kubenswrapper[4988]: I1008 19:50:31.758730 4988 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Oct 08 19:50:31 crc kubenswrapper[4988]: I1008 19:50:31.763356 4988 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Oct 08 19:50:31 crc kubenswrapper[4988]: I1008 19:50:31.763812 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Oct 08 19:50:31 crc kubenswrapper[4988]: I1008 19:50:31.768897 4988 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Oct 08 19:50:31 crc kubenswrapper[4988]: I1008 19:50:31.771771 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Oct 08 19:50:31 crc kubenswrapper[4988]: I1008 19:50:31.771956 4988 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Oct 08 19:50:31 crc kubenswrapper[4988]: I1008 19:50:31.772020 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Oct 08 19:50:32 crc kubenswrapper[4988]: I1008 19:50:32.249818 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Oct 08 19:50:32 crc kubenswrapper[4988]: I1008 19:50:32.255293 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Oct 08 19:50:32 crc kubenswrapper[4988]: I1008 19:50:32.258154 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Oct 08 19:50:32 crc kubenswrapper[4988]: I1008 19:50:32.480322 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-65d478dd7c-lc957"] Oct 08 19:50:32 crc kubenswrapper[4988]: I1008 19:50:32.481832 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-65d478dd7c-lc957" Oct 08 19:50:32 crc kubenswrapper[4988]: I1008 19:50:32.491561 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-65d478dd7c-lc957"] Oct 08 19:50:32 crc kubenswrapper[4988]: I1008 19:50:32.523226 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5c03ce9a-8cfe-4830-a131-e5b6353bef7c-ovsdbserver-nb\") pod \"dnsmasq-dns-65d478dd7c-lc957\" (UID: \"5c03ce9a-8cfe-4830-a131-e5b6353bef7c\") " pod="openstack/dnsmasq-dns-65d478dd7c-lc957" Oct 08 19:50:32 crc kubenswrapper[4988]: I1008 19:50:32.523506 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lcfxp\" (UniqueName: \"kubernetes.io/projected/5c03ce9a-8cfe-4830-a131-e5b6353bef7c-kube-api-access-lcfxp\") pod \"dnsmasq-dns-65d478dd7c-lc957\" (UID: \"5c03ce9a-8cfe-4830-a131-e5b6353bef7c\") " pod="openstack/dnsmasq-dns-65d478dd7c-lc957" Oct 08 19:50:32 crc kubenswrapper[4988]: I1008 19:50:32.523548 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5c03ce9a-8cfe-4830-a131-e5b6353bef7c-dns-svc\") pod \"dnsmasq-dns-65d478dd7c-lc957\" (UID: \"5c03ce9a-8cfe-4830-a131-e5b6353bef7c\") " pod="openstack/dnsmasq-dns-65d478dd7c-lc957" Oct 08 19:50:32 crc kubenswrapper[4988]: I1008 19:50:32.523674 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5c03ce9a-8cfe-4830-a131-e5b6353bef7c-config\") pod \"dnsmasq-dns-65d478dd7c-lc957\" (UID: \"5c03ce9a-8cfe-4830-a131-e5b6353bef7c\") " pod="openstack/dnsmasq-dns-65d478dd7c-lc957" Oct 08 19:50:32 crc kubenswrapper[4988]: I1008 19:50:32.523810 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5c03ce9a-8cfe-4830-a131-e5b6353bef7c-ovsdbserver-sb\") pod \"dnsmasq-dns-65d478dd7c-lc957\" (UID: \"5c03ce9a-8cfe-4830-a131-e5b6353bef7c\") " pod="openstack/dnsmasq-dns-65d478dd7c-lc957" Oct 08 19:50:32 crc kubenswrapper[4988]: I1008 19:50:32.625792 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5c03ce9a-8cfe-4830-a131-e5b6353bef7c-config\") pod \"dnsmasq-dns-65d478dd7c-lc957\" (UID: \"5c03ce9a-8cfe-4830-a131-e5b6353bef7c\") " pod="openstack/dnsmasq-dns-65d478dd7c-lc957" Oct 08 19:50:32 crc kubenswrapper[4988]: I1008 19:50:32.625860 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5c03ce9a-8cfe-4830-a131-e5b6353bef7c-ovsdbserver-sb\") pod \"dnsmasq-dns-65d478dd7c-lc957\" (UID: \"5c03ce9a-8cfe-4830-a131-e5b6353bef7c\") " pod="openstack/dnsmasq-dns-65d478dd7c-lc957" Oct 08 19:50:32 crc kubenswrapper[4988]: I1008 19:50:32.625936 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5c03ce9a-8cfe-4830-a131-e5b6353bef7c-ovsdbserver-nb\") pod \"dnsmasq-dns-65d478dd7c-lc957\" (UID: \"5c03ce9a-8cfe-4830-a131-e5b6353bef7c\") " pod="openstack/dnsmasq-dns-65d478dd7c-lc957" Oct 08 19:50:32 crc kubenswrapper[4988]: I1008 19:50:32.625991 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lcfxp\" (UniqueName: \"kubernetes.io/projected/5c03ce9a-8cfe-4830-a131-e5b6353bef7c-kube-api-access-lcfxp\") pod \"dnsmasq-dns-65d478dd7c-lc957\" (UID: \"5c03ce9a-8cfe-4830-a131-e5b6353bef7c\") " pod="openstack/dnsmasq-dns-65d478dd7c-lc957" Oct 08 19:50:32 crc kubenswrapper[4988]: I1008 19:50:32.626010 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5c03ce9a-8cfe-4830-a131-e5b6353bef7c-dns-svc\") pod \"dnsmasq-dns-65d478dd7c-lc957\" (UID: \"5c03ce9a-8cfe-4830-a131-e5b6353bef7c\") " pod="openstack/dnsmasq-dns-65d478dd7c-lc957" Oct 08 19:50:32 crc kubenswrapper[4988]: I1008 19:50:32.626872 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5c03ce9a-8cfe-4830-a131-e5b6353bef7c-dns-svc\") pod \"dnsmasq-dns-65d478dd7c-lc957\" (UID: \"5c03ce9a-8cfe-4830-a131-e5b6353bef7c\") " pod="openstack/dnsmasq-dns-65d478dd7c-lc957" Oct 08 19:50:32 crc kubenswrapper[4988]: I1008 19:50:32.627007 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5c03ce9a-8cfe-4830-a131-e5b6353bef7c-config\") pod \"dnsmasq-dns-65d478dd7c-lc957\" (UID: \"5c03ce9a-8cfe-4830-a131-e5b6353bef7c\") " pod="openstack/dnsmasq-dns-65d478dd7c-lc957" Oct 08 19:50:32 crc kubenswrapper[4988]: I1008 19:50:32.627426 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5c03ce9a-8cfe-4830-a131-e5b6353bef7c-ovsdbserver-nb\") pod \"dnsmasq-dns-65d478dd7c-lc957\" (UID: \"5c03ce9a-8cfe-4830-a131-e5b6353bef7c\") " pod="openstack/dnsmasq-dns-65d478dd7c-lc957" Oct 08 19:50:32 crc kubenswrapper[4988]: I1008 19:50:32.627959 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5c03ce9a-8cfe-4830-a131-e5b6353bef7c-ovsdbserver-sb\") pod \"dnsmasq-dns-65d478dd7c-lc957\" (UID: \"5c03ce9a-8cfe-4830-a131-e5b6353bef7c\") " pod="openstack/dnsmasq-dns-65d478dd7c-lc957" Oct 08 19:50:32 crc kubenswrapper[4988]: I1008 19:50:32.645868 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lcfxp\" (UniqueName: \"kubernetes.io/projected/5c03ce9a-8cfe-4830-a131-e5b6353bef7c-kube-api-access-lcfxp\") pod \"dnsmasq-dns-65d478dd7c-lc957\" (UID: \"5c03ce9a-8cfe-4830-a131-e5b6353bef7c\") " pod="openstack/dnsmasq-dns-65d478dd7c-lc957" Oct 08 19:50:32 crc kubenswrapper[4988]: I1008 19:50:32.803180 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-65d478dd7c-lc957" Oct 08 19:50:33 crc kubenswrapper[4988]: I1008 19:50:33.264726 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-65d478dd7c-lc957"] Oct 08 19:50:33 crc kubenswrapper[4988]: W1008 19:50:33.269456 4988 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5c03ce9a_8cfe_4830_a131_e5b6353bef7c.slice/crio-8e2da6caaa60825f6ac114028a3bde9ad140aad58d71713ccf7cff4d35497bdf WatchSource:0}: Error finding container 8e2da6caaa60825f6ac114028a3bde9ad140aad58d71713ccf7cff4d35497bdf: Status 404 returned error can't find the container with id 8e2da6caaa60825f6ac114028a3bde9ad140aad58d71713ccf7cff4d35497bdf Oct 08 19:50:33 crc kubenswrapper[4988]: I1008 19:50:33.521809 4988 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Oct 08 19:50:33 crc kubenswrapper[4988]: I1008 19:50:33.548983 4988 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Oct 08 19:50:34 crc kubenswrapper[4988]: I1008 19:50:34.270988 4988 generic.go:334] "Generic (PLEG): container finished" podID="5c03ce9a-8cfe-4830-a131-e5b6353bef7c" containerID="de44ce794ae6a6705f6329d861517cd9df6c73b5e3feec35b602075e9037d78e" exitCode=0 Oct 08 19:50:34 crc kubenswrapper[4988]: I1008 19:50:34.271088 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-65d478dd7c-lc957" event={"ID":"5c03ce9a-8cfe-4830-a131-e5b6353bef7c","Type":"ContainerDied","Data":"de44ce794ae6a6705f6329d861517cd9df6c73b5e3feec35b602075e9037d78e"} Oct 08 19:50:34 crc kubenswrapper[4988]: I1008 19:50:34.271441 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-65d478dd7c-lc957" event={"ID":"5c03ce9a-8cfe-4830-a131-e5b6353bef7c","Type":"ContainerStarted","Data":"8e2da6caaa60825f6ac114028a3bde9ad140aad58d71713ccf7cff4d35497bdf"} Oct 08 19:50:34 crc kubenswrapper[4988]: I1008 19:50:34.320618 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Oct 08 19:50:34 crc kubenswrapper[4988]: I1008 19:50:34.895565 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 08 19:50:35 crc kubenswrapper[4988]: I1008 19:50:35.245148 4988 scope.go:117] "RemoveContainer" containerID="f2a29ed22dd7a48bbc929e65fbeceed969a669e78f65fbfb724b03dc0db6e947" Oct 08 19:50:35 crc kubenswrapper[4988]: E1008 19:50:35.245503 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nm54f_openshift-machine-config-operator(f328a20b-5283-407e-8a93-2ab2da599f18)\"" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" Oct 08 19:50:35 crc kubenswrapper[4988]: I1008 19:50:35.281832 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-65d478dd7c-lc957" event={"ID":"5c03ce9a-8cfe-4830-a131-e5b6353bef7c","Type":"ContainerStarted","Data":"a2613649e8a5b46817e3a16690ad3f809127cf8b260464db7c2c75c1cf186c5a"} Oct 08 19:50:35 crc kubenswrapper[4988]: I1008 19:50:35.282051 4988 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="f6d28178-c627-4be2-b5c1-d660f8f2dbb4" containerName="nova-api-log" containerID="cri-o://87d4a0bd1cbfe9bee62db4b538afe1dd7810577476cb342c2fe3a8fe7e4d4ae8" gracePeriod=30 Oct 08 19:50:35 crc kubenswrapper[4988]: I1008 19:50:35.282698 4988 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="f6d28178-c627-4be2-b5c1-d660f8f2dbb4" containerName="nova-api-api" containerID="cri-o://a291e0fd40de41d8c18058f44c29dcc41afba6d653817a42371a2fa8ae0d896c" gracePeriod=30 Oct 08 19:50:35 crc kubenswrapper[4988]: I1008 19:50:35.309196 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-65d478dd7c-lc957" podStartSLOduration=3.309179998 podStartE2EDuration="3.309179998s" podCreationTimestamp="2025-10-08 19:50:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 19:50:35.30801231 +0000 UTC m=+5980.757855100" watchObservedRunningTime="2025-10-08 19:50:35.309179998 +0000 UTC m=+5980.759022758" Oct 08 19:50:36 crc kubenswrapper[4988]: I1008 19:50:36.295229 4988 generic.go:334] "Generic (PLEG): container finished" podID="f6d28178-c627-4be2-b5c1-d660f8f2dbb4" containerID="87d4a0bd1cbfe9bee62db4b538afe1dd7810577476cb342c2fe3a8fe7e4d4ae8" exitCode=143 Oct 08 19:50:36 crc kubenswrapper[4988]: I1008 19:50:36.295349 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"f6d28178-c627-4be2-b5c1-d660f8f2dbb4","Type":"ContainerDied","Data":"87d4a0bd1cbfe9bee62db4b538afe1dd7810577476cb342c2fe3a8fe7e4d4ae8"} Oct 08 19:50:36 crc kubenswrapper[4988]: I1008 19:50:36.296044 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-65d478dd7c-lc957" Oct 08 19:50:38 crc kubenswrapper[4988]: I1008 19:50:38.074452 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-create-zk7zh"] Oct 08 19:50:38 crc kubenswrapper[4988]: I1008 19:50:38.082906 4988 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-create-zk7zh"] Oct 08 19:50:39 crc kubenswrapper[4988]: I1008 19:50:39.198427 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 08 19:50:39 crc kubenswrapper[4988]: I1008 19:50:39.247508 4988 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e6a11d4a-8b71-47bf-a169-f814d7c5455c" path="/var/lib/kubelet/pods/e6a11d4a-8b71-47bf-a169-f814d7c5455c/volumes" Oct 08 19:50:39 crc kubenswrapper[4988]: I1008 19:50:39.255545 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rtndv\" (UniqueName: \"kubernetes.io/projected/f6d28178-c627-4be2-b5c1-d660f8f2dbb4-kube-api-access-rtndv\") pod \"f6d28178-c627-4be2-b5c1-d660f8f2dbb4\" (UID: \"f6d28178-c627-4be2-b5c1-d660f8f2dbb4\") " Oct 08 19:50:39 crc kubenswrapper[4988]: I1008 19:50:39.255696 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f6d28178-c627-4be2-b5c1-d660f8f2dbb4-combined-ca-bundle\") pod \"f6d28178-c627-4be2-b5c1-d660f8f2dbb4\" (UID: \"f6d28178-c627-4be2-b5c1-d660f8f2dbb4\") " Oct 08 19:50:39 crc kubenswrapper[4988]: I1008 19:50:39.255813 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f6d28178-c627-4be2-b5c1-d660f8f2dbb4-config-data\") pod \"f6d28178-c627-4be2-b5c1-d660f8f2dbb4\" (UID: \"f6d28178-c627-4be2-b5c1-d660f8f2dbb4\") " Oct 08 19:50:39 crc kubenswrapper[4988]: I1008 19:50:39.255848 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f6d28178-c627-4be2-b5c1-d660f8f2dbb4-logs\") pod \"f6d28178-c627-4be2-b5c1-d660f8f2dbb4\" (UID: \"f6d28178-c627-4be2-b5c1-d660f8f2dbb4\") " Oct 08 19:50:39 crc kubenswrapper[4988]: I1008 19:50:39.256853 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f6d28178-c627-4be2-b5c1-d660f8f2dbb4-logs" (OuterVolumeSpecName: "logs") pod "f6d28178-c627-4be2-b5c1-d660f8f2dbb4" (UID: "f6d28178-c627-4be2-b5c1-d660f8f2dbb4"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 19:50:39 crc kubenswrapper[4988]: I1008 19:50:39.262178 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f6d28178-c627-4be2-b5c1-d660f8f2dbb4-kube-api-access-rtndv" (OuterVolumeSpecName: "kube-api-access-rtndv") pod "f6d28178-c627-4be2-b5c1-d660f8f2dbb4" (UID: "f6d28178-c627-4be2-b5c1-d660f8f2dbb4"). InnerVolumeSpecName "kube-api-access-rtndv". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 19:50:39 crc kubenswrapper[4988]: I1008 19:50:39.294468 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f6d28178-c627-4be2-b5c1-d660f8f2dbb4-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f6d28178-c627-4be2-b5c1-d660f8f2dbb4" (UID: "f6d28178-c627-4be2-b5c1-d660f8f2dbb4"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 19:50:39 crc kubenswrapper[4988]: I1008 19:50:39.331112 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f6d28178-c627-4be2-b5c1-d660f8f2dbb4-config-data" (OuterVolumeSpecName: "config-data") pod "f6d28178-c627-4be2-b5c1-d660f8f2dbb4" (UID: "f6d28178-c627-4be2-b5c1-d660f8f2dbb4"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 19:50:39 crc kubenswrapper[4988]: I1008 19:50:39.342682 4988 generic.go:334] "Generic (PLEG): container finished" podID="f6d28178-c627-4be2-b5c1-d660f8f2dbb4" containerID="a291e0fd40de41d8c18058f44c29dcc41afba6d653817a42371a2fa8ae0d896c" exitCode=0 Oct 08 19:50:39 crc kubenswrapper[4988]: I1008 19:50:39.342720 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"f6d28178-c627-4be2-b5c1-d660f8f2dbb4","Type":"ContainerDied","Data":"a291e0fd40de41d8c18058f44c29dcc41afba6d653817a42371a2fa8ae0d896c"} Oct 08 19:50:39 crc kubenswrapper[4988]: I1008 19:50:39.342745 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"f6d28178-c627-4be2-b5c1-d660f8f2dbb4","Type":"ContainerDied","Data":"f18d9cd976e668b8113026acae7688e30b7fa9600f3b9425bae52a7ebdecf937"} Oct 08 19:50:39 crc kubenswrapper[4988]: I1008 19:50:39.342761 4988 scope.go:117] "RemoveContainer" containerID="a291e0fd40de41d8c18058f44c29dcc41afba6d653817a42371a2fa8ae0d896c" Oct 08 19:50:39 crc kubenswrapper[4988]: I1008 19:50:39.342863 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 08 19:50:39 crc kubenswrapper[4988]: I1008 19:50:39.357749 4988 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f6d28178-c627-4be2-b5c1-d660f8f2dbb4-config-data\") on node \"crc\" DevicePath \"\"" Oct 08 19:50:39 crc kubenswrapper[4988]: I1008 19:50:39.357782 4988 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f6d28178-c627-4be2-b5c1-d660f8f2dbb4-logs\") on node \"crc\" DevicePath \"\"" Oct 08 19:50:39 crc kubenswrapper[4988]: I1008 19:50:39.357793 4988 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rtndv\" (UniqueName: \"kubernetes.io/projected/f6d28178-c627-4be2-b5c1-d660f8f2dbb4-kube-api-access-rtndv\") on node \"crc\" DevicePath \"\"" Oct 08 19:50:39 crc kubenswrapper[4988]: I1008 19:50:39.357804 4988 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f6d28178-c627-4be2-b5c1-d660f8f2dbb4-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 19:50:39 crc kubenswrapper[4988]: I1008 19:50:39.399013 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 08 19:50:39 crc kubenswrapper[4988]: I1008 19:50:39.404278 4988 scope.go:117] "RemoveContainer" containerID="87d4a0bd1cbfe9bee62db4b538afe1dd7810577476cb342c2fe3a8fe7e4d4ae8" Oct 08 19:50:39 crc kubenswrapper[4988]: I1008 19:50:39.410953 4988 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Oct 08 19:50:39 crc kubenswrapper[4988]: I1008 19:50:39.423901 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Oct 08 19:50:39 crc kubenswrapper[4988]: E1008 19:50:39.424294 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f6d28178-c627-4be2-b5c1-d660f8f2dbb4" containerName="nova-api-log" Oct 08 19:50:39 crc kubenswrapper[4988]: I1008 19:50:39.424313 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="f6d28178-c627-4be2-b5c1-d660f8f2dbb4" containerName="nova-api-log" Oct 08 19:50:39 crc kubenswrapper[4988]: E1008 19:50:39.424335 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f6d28178-c627-4be2-b5c1-d660f8f2dbb4" containerName="nova-api-api" Oct 08 19:50:39 crc kubenswrapper[4988]: I1008 19:50:39.424341 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="f6d28178-c627-4be2-b5c1-d660f8f2dbb4" containerName="nova-api-api" Oct 08 19:50:39 crc kubenswrapper[4988]: I1008 19:50:39.425466 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="f6d28178-c627-4be2-b5c1-d660f8f2dbb4" containerName="nova-api-log" Oct 08 19:50:39 crc kubenswrapper[4988]: I1008 19:50:39.425490 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="f6d28178-c627-4be2-b5c1-d660f8f2dbb4" containerName="nova-api-api" Oct 08 19:50:39 crc kubenswrapper[4988]: I1008 19:50:39.426433 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 08 19:50:39 crc kubenswrapper[4988]: I1008 19:50:39.429379 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-internal-svc" Oct 08 19:50:39 crc kubenswrapper[4988]: I1008 19:50:39.429468 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Oct 08 19:50:39 crc kubenswrapper[4988]: I1008 19:50:39.429565 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-public-svc" Oct 08 19:50:39 crc kubenswrapper[4988]: I1008 19:50:39.438217 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 08 19:50:39 crc kubenswrapper[4988]: I1008 19:50:39.446015 4988 scope.go:117] "RemoveContainer" containerID="a291e0fd40de41d8c18058f44c29dcc41afba6d653817a42371a2fa8ae0d896c" Oct 08 19:50:39 crc kubenswrapper[4988]: E1008 19:50:39.446374 4988 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a291e0fd40de41d8c18058f44c29dcc41afba6d653817a42371a2fa8ae0d896c\": container with ID starting with a291e0fd40de41d8c18058f44c29dcc41afba6d653817a42371a2fa8ae0d896c not found: ID does not exist" containerID="a291e0fd40de41d8c18058f44c29dcc41afba6d653817a42371a2fa8ae0d896c" Oct 08 19:50:39 crc kubenswrapper[4988]: I1008 19:50:39.446476 4988 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a291e0fd40de41d8c18058f44c29dcc41afba6d653817a42371a2fa8ae0d896c"} err="failed to get container status \"a291e0fd40de41d8c18058f44c29dcc41afba6d653817a42371a2fa8ae0d896c\": rpc error: code = NotFound desc = could not find container \"a291e0fd40de41d8c18058f44c29dcc41afba6d653817a42371a2fa8ae0d896c\": container with ID starting with a291e0fd40de41d8c18058f44c29dcc41afba6d653817a42371a2fa8ae0d896c not found: ID does not exist" Oct 08 19:50:39 crc kubenswrapper[4988]: I1008 19:50:39.446503 4988 scope.go:117] "RemoveContainer" containerID="87d4a0bd1cbfe9bee62db4b538afe1dd7810577476cb342c2fe3a8fe7e4d4ae8" Oct 08 19:50:39 crc kubenswrapper[4988]: E1008 19:50:39.446774 4988 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"87d4a0bd1cbfe9bee62db4b538afe1dd7810577476cb342c2fe3a8fe7e4d4ae8\": container with ID starting with 87d4a0bd1cbfe9bee62db4b538afe1dd7810577476cb342c2fe3a8fe7e4d4ae8 not found: ID does not exist" containerID="87d4a0bd1cbfe9bee62db4b538afe1dd7810577476cb342c2fe3a8fe7e4d4ae8" Oct 08 19:50:39 crc kubenswrapper[4988]: I1008 19:50:39.446808 4988 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"87d4a0bd1cbfe9bee62db4b538afe1dd7810577476cb342c2fe3a8fe7e4d4ae8"} err="failed to get container status \"87d4a0bd1cbfe9bee62db4b538afe1dd7810577476cb342c2fe3a8fe7e4d4ae8\": rpc error: code = NotFound desc = could not find container \"87d4a0bd1cbfe9bee62db4b538afe1dd7810577476cb342c2fe3a8fe7e4d4ae8\": container with ID starting with 87d4a0bd1cbfe9bee62db4b538afe1dd7810577476cb342c2fe3a8fe7e4d4ae8 not found: ID does not exist" Oct 08 19:50:39 crc kubenswrapper[4988]: I1008 19:50:39.474036 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/60e4b83d-3db4-4f5a-aca8-06ee1ee85ea7-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"60e4b83d-3db4-4f5a-aca8-06ee1ee85ea7\") " pod="openstack/nova-api-0" Oct 08 19:50:39 crc kubenswrapper[4988]: I1008 19:50:39.474126 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/60e4b83d-3db4-4f5a-aca8-06ee1ee85ea7-public-tls-certs\") pod \"nova-api-0\" (UID: \"60e4b83d-3db4-4f5a-aca8-06ee1ee85ea7\") " pod="openstack/nova-api-0" Oct 08 19:50:39 crc kubenswrapper[4988]: I1008 19:50:39.474192 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/60e4b83d-3db4-4f5a-aca8-06ee1ee85ea7-internal-tls-certs\") pod \"nova-api-0\" (UID: \"60e4b83d-3db4-4f5a-aca8-06ee1ee85ea7\") " pod="openstack/nova-api-0" Oct 08 19:50:39 crc kubenswrapper[4988]: I1008 19:50:39.474229 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/60e4b83d-3db4-4f5a-aca8-06ee1ee85ea7-config-data\") pod \"nova-api-0\" (UID: \"60e4b83d-3db4-4f5a-aca8-06ee1ee85ea7\") " pod="openstack/nova-api-0" Oct 08 19:50:39 crc kubenswrapper[4988]: I1008 19:50:39.474317 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/60e4b83d-3db4-4f5a-aca8-06ee1ee85ea7-logs\") pod \"nova-api-0\" (UID: \"60e4b83d-3db4-4f5a-aca8-06ee1ee85ea7\") " pod="openstack/nova-api-0" Oct 08 19:50:39 crc kubenswrapper[4988]: I1008 19:50:39.474351 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7fm97\" (UniqueName: \"kubernetes.io/projected/60e4b83d-3db4-4f5a-aca8-06ee1ee85ea7-kube-api-access-7fm97\") pod \"nova-api-0\" (UID: \"60e4b83d-3db4-4f5a-aca8-06ee1ee85ea7\") " pod="openstack/nova-api-0" Oct 08 19:50:39 crc kubenswrapper[4988]: I1008 19:50:39.576412 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/60e4b83d-3db4-4f5a-aca8-06ee1ee85ea7-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"60e4b83d-3db4-4f5a-aca8-06ee1ee85ea7\") " pod="openstack/nova-api-0" Oct 08 19:50:39 crc kubenswrapper[4988]: I1008 19:50:39.576515 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/60e4b83d-3db4-4f5a-aca8-06ee1ee85ea7-public-tls-certs\") pod \"nova-api-0\" (UID: \"60e4b83d-3db4-4f5a-aca8-06ee1ee85ea7\") " pod="openstack/nova-api-0" Oct 08 19:50:39 crc kubenswrapper[4988]: I1008 19:50:39.576571 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/60e4b83d-3db4-4f5a-aca8-06ee1ee85ea7-internal-tls-certs\") pod \"nova-api-0\" (UID: \"60e4b83d-3db4-4f5a-aca8-06ee1ee85ea7\") " pod="openstack/nova-api-0" Oct 08 19:50:39 crc kubenswrapper[4988]: I1008 19:50:39.576595 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/60e4b83d-3db4-4f5a-aca8-06ee1ee85ea7-config-data\") pod \"nova-api-0\" (UID: \"60e4b83d-3db4-4f5a-aca8-06ee1ee85ea7\") " pod="openstack/nova-api-0" Oct 08 19:50:39 crc kubenswrapper[4988]: I1008 19:50:39.576668 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/60e4b83d-3db4-4f5a-aca8-06ee1ee85ea7-logs\") pod \"nova-api-0\" (UID: \"60e4b83d-3db4-4f5a-aca8-06ee1ee85ea7\") " pod="openstack/nova-api-0" Oct 08 19:50:39 crc kubenswrapper[4988]: I1008 19:50:39.576693 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7fm97\" (UniqueName: \"kubernetes.io/projected/60e4b83d-3db4-4f5a-aca8-06ee1ee85ea7-kube-api-access-7fm97\") pod \"nova-api-0\" (UID: \"60e4b83d-3db4-4f5a-aca8-06ee1ee85ea7\") " pod="openstack/nova-api-0" Oct 08 19:50:39 crc kubenswrapper[4988]: I1008 19:50:39.577546 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/60e4b83d-3db4-4f5a-aca8-06ee1ee85ea7-logs\") pod \"nova-api-0\" (UID: \"60e4b83d-3db4-4f5a-aca8-06ee1ee85ea7\") " pod="openstack/nova-api-0" Oct 08 19:50:39 crc kubenswrapper[4988]: I1008 19:50:39.580649 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/60e4b83d-3db4-4f5a-aca8-06ee1ee85ea7-internal-tls-certs\") pod \"nova-api-0\" (UID: \"60e4b83d-3db4-4f5a-aca8-06ee1ee85ea7\") " pod="openstack/nova-api-0" Oct 08 19:50:39 crc kubenswrapper[4988]: I1008 19:50:39.580689 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/60e4b83d-3db4-4f5a-aca8-06ee1ee85ea7-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"60e4b83d-3db4-4f5a-aca8-06ee1ee85ea7\") " pod="openstack/nova-api-0" Oct 08 19:50:39 crc kubenswrapper[4988]: I1008 19:50:39.580740 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/60e4b83d-3db4-4f5a-aca8-06ee1ee85ea7-config-data\") pod \"nova-api-0\" (UID: \"60e4b83d-3db4-4f5a-aca8-06ee1ee85ea7\") " pod="openstack/nova-api-0" Oct 08 19:50:39 crc kubenswrapper[4988]: I1008 19:50:39.581211 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/60e4b83d-3db4-4f5a-aca8-06ee1ee85ea7-public-tls-certs\") pod \"nova-api-0\" (UID: \"60e4b83d-3db4-4f5a-aca8-06ee1ee85ea7\") " pod="openstack/nova-api-0" Oct 08 19:50:39 crc kubenswrapper[4988]: I1008 19:50:39.596131 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7fm97\" (UniqueName: \"kubernetes.io/projected/60e4b83d-3db4-4f5a-aca8-06ee1ee85ea7-kube-api-access-7fm97\") pod \"nova-api-0\" (UID: \"60e4b83d-3db4-4f5a-aca8-06ee1ee85ea7\") " pod="openstack/nova-api-0" Oct 08 19:50:39 crc kubenswrapper[4988]: I1008 19:50:39.773834 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 08 19:50:40 crc kubenswrapper[4988]: I1008 19:50:40.282056 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 08 19:50:40 crc kubenswrapper[4988]: I1008 19:50:40.357462 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"60e4b83d-3db4-4f5a-aca8-06ee1ee85ea7","Type":"ContainerStarted","Data":"a47b5cd9b728e721f1efe6ff52e9a59abff1e7b84636b4af0a4e3b42718087f9"} Oct 08 19:50:41 crc kubenswrapper[4988]: I1008 19:50:41.248875 4988 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f6d28178-c627-4be2-b5c1-d660f8f2dbb4" path="/var/lib/kubelet/pods/f6d28178-c627-4be2-b5c1-d660f8f2dbb4/volumes" Oct 08 19:50:41 crc kubenswrapper[4988]: I1008 19:50:41.374183 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"60e4b83d-3db4-4f5a-aca8-06ee1ee85ea7","Type":"ContainerStarted","Data":"607110804b1f15f95b3a567b056ee8c4c1b1f154bbc886dc1d8716bf75980d29"} Oct 08 19:50:41 crc kubenswrapper[4988]: I1008 19:50:41.374236 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"60e4b83d-3db4-4f5a-aca8-06ee1ee85ea7","Type":"ContainerStarted","Data":"cb80a17c24a9f12755e932ef41d6a5bb75e4e04a3d0ba21999c59874b0080b6a"} Oct 08 19:50:41 crc kubenswrapper[4988]: I1008 19:50:41.402156 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.402131119 podStartE2EDuration="2.402131119s" podCreationTimestamp="2025-10-08 19:50:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 19:50:41.393834286 +0000 UTC m=+5986.843677096" watchObservedRunningTime="2025-10-08 19:50:41.402131119 +0000 UTC m=+5986.851973929" Oct 08 19:50:42 crc kubenswrapper[4988]: I1008 19:50:42.804919 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-65d478dd7c-lc957" Oct 08 19:50:42 crc kubenswrapper[4988]: I1008 19:50:42.910348 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5bf7c8d69c-j8v4t"] Oct 08 19:50:42 crc kubenswrapper[4988]: I1008 19:50:42.910670 4988 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-5bf7c8d69c-j8v4t" podUID="ee5a3e37-5a2f-4af4-9d64-70444901144a" containerName="dnsmasq-dns" containerID="cri-o://f29503fc288795cbc96059d6a35444db4b2e4c9402d74d1e3a005c7dc00086af" gracePeriod=10 Oct 08 19:50:43 crc kubenswrapper[4988]: I1008 19:50:43.008673 4988 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-5bf7c8d69c-j8v4t" podUID="ee5a3e37-5a2f-4af4-9d64-70444901144a" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.1.88:5353: connect: connection refused" Oct 08 19:50:43 crc kubenswrapper[4988]: I1008 19:50:43.377336 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5bf7c8d69c-j8v4t" Oct 08 19:50:43 crc kubenswrapper[4988]: I1008 19:50:43.393640 4988 generic.go:334] "Generic (PLEG): container finished" podID="ee5a3e37-5a2f-4af4-9d64-70444901144a" containerID="f29503fc288795cbc96059d6a35444db4b2e4c9402d74d1e3a005c7dc00086af" exitCode=0 Oct 08 19:50:43 crc kubenswrapper[4988]: I1008 19:50:43.393697 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5bf7c8d69c-j8v4t" event={"ID":"ee5a3e37-5a2f-4af4-9d64-70444901144a","Type":"ContainerDied","Data":"f29503fc288795cbc96059d6a35444db4b2e4c9402d74d1e3a005c7dc00086af"} Oct 08 19:50:43 crc kubenswrapper[4988]: I1008 19:50:43.393732 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5bf7c8d69c-j8v4t" event={"ID":"ee5a3e37-5a2f-4af4-9d64-70444901144a","Type":"ContainerDied","Data":"55e3f22e8294f3486ffcd7709ca7dd047dbf01ac2601c28f50c80dda0da31c59"} Oct 08 19:50:43 crc kubenswrapper[4988]: I1008 19:50:43.393732 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5bf7c8d69c-j8v4t" Oct 08 19:50:43 crc kubenswrapper[4988]: I1008 19:50:43.393752 4988 scope.go:117] "RemoveContainer" containerID="f29503fc288795cbc96059d6a35444db4b2e4c9402d74d1e3a005c7dc00086af" Oct 08 19:50:43 crc kubenswrapper[4988]: I1008 19:50:43.417803 4988 scope.go:117] "RemoveContainer" containerID="adc6bc94734471245a6c79f42ab92cdc1b52cf63f4e217321e573b19c62af828" Oct 08 19:50:43 crc kubenswrapper[4988]: I1008 19:50:43.442255 4988 scope.go:117] "RemoveContainer" containerID="f29503fc288795cbc96059d6a35444db4b2e4c9402d74d1e3a005c7dc00086af" Oct 08 19:50:43 crc kubenswrapper[4988]: E1008 19:50:43.442909 4988 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f29503fc288795cbc96059d6a35444db4b2e4c9402d74d1e3a005c7dc00086af\": container with ID starting with f29503fc288795cbc96059d6a35444db4b2e4c9402d74d1e3a005c7dc00086af not found: ID does not exist" containerID="f29503fc288795cbc96059d6a35444db4b2e4c9402d74d1e3a005c7dc00086af" Oct 08 19:50:43 crc kubenswrapper[4988]: I1008 19:50:43.442966 4988 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f29503fc288795cbc96059d6a35444db4b2e4c9402d74d1e3a005c7dc00086af"} err="failed to get container status \"f29503fc288795cbc96059d6a35444db4b2e4c9402d74d1e3a005c7dc00086af\": rpc error: code = NotFound desc = could not find container \"f29503fc288795cbc96059d6a35444db4b2e4c9402d74d1e3a005c7dc00086af\": container with ID starting with f29503fc288795cbc96059d6a35444db4b2e4c9402d74d1e3a005c7dc00086af not found: ID does not exist" Oct 08 19:50:43 crc kubenswrapper[4988]: I1008 19:50:43.443009 4988 scope.go:117] "RemoveContainer" containerID="adc6bc94734471245a6c79f42ab92cdc1b52cf63f4e217321e573b19c62af828" Oct 08 19:50:43 crc kubenswrapper[4988]: E1008 19:50:43.443597 4988 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"adc6bc94734471245a6c79f42ab92cdc1b52cf63f4e217321e573b19c62af828\": container with ID starting with adc6bc94734471245a6c79f42ab92cdc1b52cf63f4e217321e573b19c62af828 not found: ID does not exist" containerID="adc6bc94734471245a6c79f42ab92cdc1b52cf63f4e217321e573b19c62af828" Oct 08 19:50:43 crc kubenswrapper[4988]: I1008 19:50:43.443647 4988 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"adc6bc94734471245a6c79f42ab92cdc1b52cf63f4e217321e573b19c62af828"} err="failed to get container status \"adc6bc94734471245a6c79f42ab92cdc1b52cf63f4e217321e573b19c62af828\": rpc error: code = NotFound desc = could not find container \"adc6bc94734471245a6c79f42ab92cdc1b52cf63f4e217321e573b19c62af828\": container with ID starting with adc6bc94734471245a6c79f42ab92cdc1b52cf63f4e217321e573b19c62af828 not found: ID does not exist" Oct 08 19:50:43 crc kubenswrapper[4988]: I1008 19:50:43.458315 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ee5a3e37-5a2f-4af4-9d64-70444901144a-ovsdbserver-nb\") pod \"ee5a3e37-5a2f-4af4-9d64-70444901144a\" (UID: \"ee5a3e37-5a2f-4af4-9d64-70444901144a\") " Oct 08 19:50:43 crc kubenswrapper[4988]: I1008 19:50:43.458376 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ee5a3e37-5a2f-4af4-9d64-70444901144a-dns-svc\") pod \"ee5a3e37-5a2f-4af4-9d64-70444901144a\" (UID: \"ee5a3e37-5a2f-4af4-9d64-70444901144a\") " Oct 08 19:50:43 crc kubenswrapper[4988]: I1008 19:50:43.458502 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hl89n\" (UniqueName: \"kubernetes.io/projected/ee5a3e37-5a2f-4af4-9d64-70444901144a-kube-api-access-hl89n\") pod \"ee5a3e37-5a2f-4af4-9d64-70444901144a\" (UID: \"ee5a3e37-5a2f-4af4-9d64-70444901144a\") " Oct 08 19:50:43 crc kubenswrapper[4988]: I1008 19:50:43.458695 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ee5a3e37-5a2f-4af4-9d64-70444901144a-ovsdbserver-sb\") pod \"ee5a3e37-5a2f-4af4-9d64-70444901144a\" (UID: \"ee5a3e37-5a2f-4af4-9d64-70444901144a\") " Oct 08 19:50:43 crc kubenswrapper[4988]: I1008 19:50:43.458733 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ee5a3e37-5a2f-4af4-9d64-70444901144a-config\") pod \"ee5a3e37-5a2f-4af4-9d64-70444901144a\" (UID: \"ee5a3e37-5a2f-4af4-9d64-70444901144a\") " Oct 08 19:50:43 crc kubenswrapper[4988]: I1008 19:50:43.464321 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ee5a3e37-5a2f-4af4-9d64-70444901144a-kube-api-access-hl89n" (OuterVolumeSpecName: "kube-api-access-hl89n") pod "ee5a3e37-5a2f-4af4-9d64-70444901144a" (UID: "ee5a3e37-5a2f-4af4-9d64-70444901144a"). InnerVolumeSpecName "kube-api-access-hl89n". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 19:50:43 crc kubenswrapper[4988]: I1008 19:50:43.510493 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ee5a3e37-5a2f-4af4-9d64-70444901144a-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "ee5a3e37-5a2f-4af4-9d64-70444901144a" (UID: "ee5a3e37-5a2f-4af4-9d64-70444901144a"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 19:50:43 crc kubenswrapper[4988]: I1008 19:50:43.511286 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ee5a3e37-5a2f-4af4-9d64-70444901144a-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "ee5a3e37-5a2f-4af4-9d64-70444901144a" (UID: "ee5a3e37-5a2f-4af4-9d64-70444901144a"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 19:50:43 crc kubenswrapper[4988]: I1008 19:50:43.523197 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ee5a3e37-5a2f-4af4-9d64-70444901144a-config" (OuterVolumeSpecName: "config") pod "ee5a3e37-5a2f-4af4-9d64-70444901144a" (UID: "ee5a3e37-5a2f-4af4-9d64-70444901144a"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 19:50:43 crc kubenswrapper[4988]: I1008 19:50:43.524956 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ee5a3e37-5a2f-4af4-9d64-70444901144a-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "ee5a3e37-5a2f-4af4-9d64-70444901144a" (UID: "ee5a3e37-5a2f-4af4-9d64-70444901144a"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 19:50:43 crc kubenswrapper[4988]: I1008 19:50:43.561607 4988 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ee5a3e37-5a2f-4af4-9d64-70444901144a-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 08 19:50:43 crc kubenswrapper[4988]: I1008 19:50:43.561656 4988 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ee5a3e37-5a2f-4af4-9d64-70444901144a-config\") on node \"crc\" DevicePath \"\"" Oct 08 19:50:43 crc kubenswrapper[4988]: I1008 19:50:43.561675 4988 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ee5a3e37-5a2f-4af4-9d64-70444901144a-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 08 19:50:43 crc kubenswrapper[4988]: I1008 19:50:43.561691 4988 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ee5a3e37-5a2f-4af4-9d64-70444901144a-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 08 19:50:43 crc kubenswrapper[4988]: I1008 19:50:43.561711 4988 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hl89n\" (UniqueName: \"kubernetes.io/projected/ee5a3e37-5a2f-4af4-9d64-70444901144a-kube-api-access-hl89n\") on node \"crc\" DevicePath \"\"" Oct 08 19:50:43 crc kubenswrapper[4988]: I1008 19:50:43.726273 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5bf7c8d69c-j8v4t"] Oct 08 19:50:43 crc kubenswrapper[4988]: I1008 19:50:43.734049 4988 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5bf7c8d69c-j8v4t"] Oct 08 19:50:45 crc kubenswrapper[4988]: I1008 19:50:45.252088 4988 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ee5a3e37-5a2f-4af4-9d64-70444901144a" path="/var/lib/kubelet/pods/ee5a3e37-5a2f-4af4-9d64-70444901144a/volumes" Oct 08 19:50:48 crc kubenswrapper[4988]: I1008 19:50:48.239312 4988 scope.go:117] "RemoveContainer" containerID="f2a29ed22dd7a48bbc929e65fbeceed969a669e78f65fbfb724b03dc0db6e947" Oct 08 19:50:48 crc kubenswrapper[4988]: E1008 19:50:48.240382 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nm54f_openshift-machine-config-operator(f328a20b-5283-407e-8a93-2ab2da599f18)\"" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" Oct 08 19:50:49 crc kubenswrapper[4988]: I1008 19:50:49.060330 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-70ac-account-create-wmgnd"] Oct 08 19:50:49 crc kubenswrapper[4988]: I1008 19:50:49.075592 4988 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-70ac-account-create-wmgnd"] Oct 08 19:50:49 crc kubenswrapper[4988]: I1008 19:50:49.252455 4988 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="175bafdf-d19e-4c77-9296-6a150c92a009" path="/var/lib/kubelet/pods/175bafdf-d19e-4c77-9296-6a150c92a009/volumes" Oct 08 19:50:49 crc kubenswrapper[4988]: I1008 19:50:49.774693 4988 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 08 19:50:49 crc kubenswrapper[4988]: I1008 19:50:49.774728 4988 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 08 19:50:50 crc kubenswrapper[4988]: I1008 19:50:50.790607 4988 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="60e4b83d-3db4-4f5a-aca8-06ee1ee85ea7" containerName="nova-api-api" probeResult="failure" output="Get \"https://10.217.1.100:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 08 19:50:50 crc kubenswrapper[4988]: I1008 19:50:50.790613 4988 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="60e4b83d-3db4-4f5a-aca8-06ee1ee85ea7" containerName="nova-api-log" probeResult="failure" output="Get \"https://10.217.1.100:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 08 19:50:55 crc kubenswrapper[4988]: I1008 19:50:55.045753 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-sync-m5jld"] Oct 08 19:50:55 crc kubenswrapper[4988]: I1008 19:50:55.060830 4988 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-sync-m5jld"] Oct 08 19:50:55 crc kubenswrapper[4988]: I1008 19:50:55.249033 4988 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3c634709-46fb-46fd-a43b-7c6c0fc8e844" path="/var/lib/kubelet/pods/3c634709-46fb-46fd-a43b-7c6c0fc8e844/volumes" Oct 08 19:50:59 crc kubenswrapper[4988]: I1008 19:50:59.240858 4988 scope.go:117] "RemoveContainer" containerID="f2a29ed22dd7a48bbc929e65fbeceed969a669e78f65fbfb724b03dc0db6e947" Oct 08 19:50:59 crc kubenswrapper[4988]: E1008 19:50:59.242112 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nm54f_openshift-machine-config-operator(f328a20b-5283-407e-8a93-2ab2da599f18)\"" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" Oct 08 19:50:59 crc kubenswrapper[4988]: I1008 19:50:59.784928 4988 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Oct 08 19:50:59 crc kubenswrapper[4988]: I1008 19:50:59.785824 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Oct 08 19:50:59 crc kubenswrapper[4988]: I1008 19:50:59.786045 4988 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Oct 08 19:50:59 crc kubenswrapper[4988]: I1008 19:50:59.799965 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Oct 08 19:51:00 crc kubenswrapper[4988]: I1008 19:51:00.597910 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Oct 08 19:51:00 crc kubenswrapper[4988]: I1008 19:51:00.609192 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Oct 08 19:51:04 crc kubenswrapper[4988]: I1008 19:51:04.526321 4988 scope.go:117] "RemoveContainer" containerID="24b1cee3957790a1ad65dc1e2719eba944956a68854fff075295d74d45fe7f8d" Oct 08 19:51:04 crc kubenswrapper[4988]: I1008 19:51:04.567651 4988 scope.go:117] "RemoveContainer" containerID="16f993453e17b5da7222fd493e51ced51bce9012568fdd229415c647988b5f60" Oct 08 19:51:04 crc kubenswrapper[4988]: I1008 19:51:04.653102 4988 scope.go:117] "RemoveContainer" containerID="66762b138eba71245f007232f9cd0ef561419f725d3fe4ae225edaee9c59cd76" Oct 08 19:51:10 crc kubenswrapper[4988]: I1008 19:51:10.050662 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-kb7bl"] Oct 08 19:51:10 crc kubenswrapper[4988]: I1008 19:51:10.072948 4988 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-kb7bl"] Oct 08 19:51:11 crc kubenswrapper[4988]: I1008 19:51:11.250155 4988 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="94108d41-9804-48e5-982a-6e0df6ba3d9f" path="/var/lib/kubelet/pods/94108d41-9804-48e5-982a-6e0df6ba3d9f/volumes" Oct 08 19:51:14 crc kubenswrapper[4988]: I1008 19:51:14.239015 4988 scope.go:117] "RemoveContainer" containerID="f2a29ed22dd7a48bbc929e65fbeceed969a669e78f65fbfb724b03dc0db6e947" Oct 08 19:51:14 crc kubenswrapper[4988]: E1008 19:51:14.239927 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nm54f_openshift-machine-config-operator(f328a20b-5283-407e-8a93-2ab2da599f18)\"" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" Oct 08 19:51:25 crc kubenswrapper[4988]: I1008 19:51:25.369612 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-2xddw"] Oct 08 19:51:25 crc kubenswrapper[4988]: E1008 19:51:25.370698 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ee5a3e37-5a2f-4af4-9d64-70444901144a" containerName="init" Oct 08 19:51:25 crc kubenswrapper[4988]: I1008 19:51:25.370722 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="ee5a3e37-5a2f-4af4-9d64-70444901144a" containerName="init" Oct 08 19:51:25 crc kubenswrapper[4988]: E1008 19:51:25.370792 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ee5a3e37-5a2f-4af4-9d64-70444901144a" containerName="dnsmasq-dns" Oct 08 19:51:25 crc kubenswrapper[4988]: I1008 19:51:25.370806 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="ee5a3e37-5a2f-4af4-9d64-70444901144a" containerName="dnsmasq-dns" Oct 08 19:51:25 crc kubenswrapper[4988]: I1008 19:51:25.371140 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="ee5a3e37-5a2f-4af4-9d64-70444901144a" containerName="dnsmasq-dns" Oct 08 19:51:25 crc kubenswrapper[4988]: I1008 19:51:25.372061 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-2xddw" Oct 08 19:51:25 crc kubenswrapper[4988]: I1008 19:51:25.373549 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncontroller-ovncontroller-dockercfg-7bhzq" Oct 08 19:51:25 crc kubenswrapper[4988]: I1008 19:51:25.373908 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovncontroller-ovndbs" Oct 08 19:51:25 crc kubenswrapper[4988]: I1008 19:51:25.383925 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-ovs-4667z"] Oct 08 19:51:25 crc kubenswrapper[4988]: I1008 19:51:25.385970 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-4667z" Oct 08 19:51:25 crc kubenswrapper[4988]: I1008 19:51:25.390818 4988 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-scripts" Oct 08 19:51:25 crc kubenswrapper[4988]: I1008 19:51:25.415085 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-2xddw"] Oct 08 19:51:25 crc kubenswrapper[4988]: I1008 19:51:25.422206 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-4667z"] Oct 08 19:51:25 crc kubenswrapper[4988]: I1008 19:51:25.520729 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a1187bea-e139-42e7-bd42-8086ed52bceb-scripts\") pod \"ovn-controller-2xddw\" (UID: \"a1187bea-e139-42e7-bd42-8086ed52bceb\") " pod="openstack/ovn-controller-2xddw" Oct 08 19:51:25 crc kubenswrapper[4988]: I1008 19:51:25.520798 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/a1187bea-e139-42e7-bd42-8086ed52bceb-var-log-ovn\") pod \"ovn-controller-2xddw\" (UID: \"a1187bea-e139-42e7-bd42-8086ed52bceb\") " pod="openstack/ovn-controller-2xddw" Oct 08 19:51:25 crc kubenswrapper[4988]: I1008 19:51:25.520833 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/a1187bea-e139-42e7-bd42-8086ed52bceb-var-run\") pod \"ovn-controller-2xddw\" (UID: \"a1187bea-e139-42e7-bd42-8086ed52bceb\") " pod="openstack/ovn-controller-2xddw" Oct 08 19:51:25 crc kubenswrapper[4988]: I1008 19:51:25.520863 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/eb6a17a8-576e-4d61-8467-cca342654b2c-etc-ovs\") pod \"ovn-controller-ovs-4667z\" (UID: \"eb6a17a8-576e-4d61-8467-cca342654b2c\") " pod="openstack/ovn-controller-ovs-4667z" Oct 08 19:51:25 crc kubenswrapper[4988]: I1008 19:51:25.520993 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/eb6a17a8-576e-4d61-8467-cca342654b2c-var-run\") pod \"ovn-controller-ovs-4667z\" (UID: \"eb6a17a8-576e-4d61-8467-cca342654b2c\") " pod="openstack/ovn-controller-ovs-4667z" Oct 08 19:51:25 crc kubenswrapper[4988]: I1008 19:51:25.521038 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/a1187bea-e139-42e7-bd42-8086ed52bceb-var-run-ovn\") pod \"ovn-controller-2xddw\" (UID: \"a1187bea-e139-42e7-bd42-8086ed52bceb\") " pod="openstack/ovn-controller-2xddw" Oct 08 19:51:25 crc kubenswrapper[4988]: I1008 19:51:25.521067 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/a1187bea-e139-42e7-bd42-8086ed52bceb-ovn-controller-tls-certs\") pod \"ovn-controller-2xddw\" (UID: \"a1187bea-e139-42e7-bd42-8086ed52bceb\") " pod="openstack/ovn-controller-2xddw" Oct 08 19:51:25 crc kubenswrapper[4988]: I1008 19:51:25.521097 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6wnxm\" (UniqueName: \"kubernetes.io/projected/eb6a17a8-576e-4d61-8467-cca342654b2c-kube-api-access-6wnxm\") pod \"ovn-controller-ovs-4667z\" (UID: \"eb6a17a8-576e-4d61-8467-cca342654b2c\") " pod="openstack/ovn-controller-ovs-4667z" Oct 08 19:51:25 crc kubenswrapper[4988]: I1008 19:51:25.521119 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qmdlf\" (UniqueName: \"kubernetes.io/projected/a1187bea-e139-42e7-bd42-8086ed52bceb-kube-api-access-qmdlf\") pod \"ovn-controller-2xddw\" (UID: \"a1187bea-e139-42e7-bd42-8086ed52bceb\") " pod="openstack/ovn-controller-2xddw" Oct 08 19:51:25 crc kubenswrapper[4988]: I1008 19:51:25.521152 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/eb6a17a8-576e-4d61-8467-cca342654b2c-var-log\") pod \"ovn-controller-ovs-4667z\" (UID: \"eb6a17a8-576e-4d61-8467-cca342654b2c\") " pod="openstack/ovn-controller-ovs-4667z" Oct 08 19:51:25 crc kubenswrapper[4988]: I1008 19:51:25.521182 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/eb6a17a8-576e-4d61-8467-cca342654b2c-var-lib\") pod \"ovn-controller-ovs-4667z\" (UID: \"eb6a17a8-576e-4d61-8467-cca342654b2c\") " pod="openstack/ovn-controller-ovs-4667z" Oct 08 19:51:25 crc kubenswrapper[4988]: I1008 19:51:25.521232 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/eb6a17a8-576e-4d61-8467-cca342654b2c-scripts\") pod \"ovn-controller-ovs-4667z\" (UID: \"eb6a17a8-576e-4d61-8467-cca342654b2c\") " pod="openstack/ovn-controller-ovs-4667z" Oct 08 19:51:25 crc kubenswrapper[4988]: I1008 19:51:25.521256 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a1187bea-e139-42e7-bd42-8086ed52bceb-combined-ca-bundle\") pod \"ovn-controller-2xddw\" (UID: \"a1187bea-e139-42e7-bd42-8086ed52bceb\") " pod="openstack/ovn-controller-2xddw" Oct 08 19:51:25 crc kubenswrapper[4988]: I1008 19:51:25.623097 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/a1187bea-e139-42e7-bd42-8086ed52bceb-var-run-ovn\") pod \"ovn-controller-2xddw\" (UID: \"a1187bea-e139-42e7-bd42-8086ed52bceb\") " pod="openstack/ovn-controller-2xddw" Oct 08 19:51:25 crc kubenswrapper[4988]: I1008 19:51:25.623148 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/a1187bea-e139-42e7-bd42-8086ed52bceb-ovn-controller-tls-certs\") pod \"ovn-controller-2xddw\" (UID: \"a1187bea-e139-42e7-bd42-8086ed52bceb\") " pod="openstack/ovn-controller-2xddw" Oct 08 19:51:25 crc kubenswrapper[4988]: I1008 19:51:25.623174 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6wnxm\" (UniqueName: \"kubernetes.io/projected/eb6a17a8-576e-4d61-8467-cca342654b2c-kube-api-access-6wnxm\") pod \"ovn-controller-ovs-4667z\" (UID: \"eb6a17a8-576e-4d61-8467-cca342654b2c\") " pod="openstack/ovn-controller-ovs-4667z" Oct 08 19:51:25 crc kubenswrapper[4988]: I1008 19:51:25.623191 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qmdlf\" (UniqueName: \"kubernetes.io/projected/a1187bea-e139-42e7-bd42-8086ed52bceb-kube-api-access-qmdlf\") pod \"ovn-controller-2xddw\" (UID: \"a1187bea-e139-42e7-bd42-8086ed52bceb\") " pod="openstack/ovn-controller-2xddw" Oct 08 19:51:25 crc kubenswrapper[4988]: I1008 19:51:25.623224 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/eb6a17a8-576e-4d61-8467-cca342654b2c-var-log\") pod \"ovn-controller-ovs-4667z\" (UID: \"eb6a17a8-576e-4d61-8467-cca342654b2c\") " pod="openstack/ovn-controller-ovs-4667z" Oct 08 19:51:25 crc kubenswrapper[4988]: I1008 19:51:25.623252 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/eb6a17a8-576e-4d61-8467-cca342654b2c-scripts\") pod \"ovn-controller-ovs-4667z\" (UID: \"eb6a17a8-576e-4d61-8467-cca342654b2c\") " pod="openstack/ovn-controller-ovs-4667z" Oct 08 19:51:25 crc kubenswrapper[4988]: I1008 19:51:25.623265 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/eb6a17a8-576e-4d61-8467-cca342654b2c-var-lib\") pod \"ovn-controller-ovs-4667z\" (UID: \"eb6a17a8-576e-4d61-8467-cca342654b2c\") " pod="openstack/ovn-controller-ovs-4667z" Oct 08 19:51:25 crc kubenswrapper[4988]: I1008 19:51:25.623280 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a1187bea-e139-42e7-bd42-8086ed52bceb-combined-ca-bundle\") pod \"ovn-controller-2xddw\" (UID: \"a1187bea-e139-42e7-bd42-8086ed52bceb\") " pod="openstack/ovn-controller-2xddw" Oct 08 19:51:25 crc kubenswrapper[4988]: I1008 19:51:25.623318 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a1187bea-e139-42e7-bd42-8086ed52bceb-scripts\") pod \"ovn-controller-2xddw\" (UID: \"a1187bea-e139-42e7-bd42-8086ed52bceb\") " pod="openstack/ovn-controller-2xddw" Oct 08 19:51:25 crc kubenswrapper[4988]: I1008 19:51:25.623345 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/a1187bea-e139-42e7-bd42-8086ed52bceb-var-log-ovn\") pod \"ovn-controller-2xddw\" (UID: \"a1187bea-e139-42e7-bd42-8086ed52bceb\") " pod="openstack/ovn-controller-2xddw" Oct 08 19:51:25 crc kubenswrapper[4988]: I1008 19:51:25.623360 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/a1187bea-e139-42e7-bd42-8086ed52bceb-var-run\") pod \"ovn-controller-2xddw\" (UID: \"a1187bea-e139-42e7-bd42-8086ed52bceb\") " pod="openstack/ovn-controller-2xddw" Oct 08 19:51:25 crc kubenswrapper[4988]: I1008 19:51:25.623378 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/eb6a17a8-576e-4d61-8467-cca342654b2c-etc-ovs\") pod \"ovn-controller-ovs-4667z\" (UID: \"eb6a17a8-576e-4d61-8467-cca342654b2c\") " pod="openstack/ovn-controller-ovs-4667z" Oct 08 19:51:25 crc kubenswrapper[4988]: I1008 19:51:25.623478 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/eb6a17a8-576e-4d61-8467-cca342654b2c-var-run\") pod \"ovn-controller-ovs-4667z\" (UID: \"eb6a17a8-576e-4d61-8467-cca342654b2c\") " pod="openstack/ovn-controller-ovs-4667z" Oct 08 19:51:25 crc kubenswrapper[4988]: I1008 19:51:25.623796 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/eb6a17a8-576e-4d61-8467-cca342654b2c-var-run\") pod \"ovn-controller-ovs-4667z\" (UID: \"eb6a17a8-576e-4d61-8467-cca342654b2c\") " pod="openstack/ovn-controller-ovs-4667z" Oct 08 19:51:25 crc kubenswrapper[4988]: I1008 19:51:25.624509 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/a1187bea-e139-42e7-bd42-8086ed52bceb-var-run\") pod \"ovn-controller-2xddw\" (UID: \"a1187bea-e139-42e7-bd42-8086ed52bceb\") " pod="openstack/ovn-controller-2xddw" Oct 08 19:51:25 crc kubenswrapper[4988]: I1008 19:51:25.624572 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/eb6a17a8-576e-4d61-8467-cca342654b2c-var-lib\") pod \"ovn-controller-ovs-4667z\" (UID: \"eb6a17a8-576e-4d61-8467-cca342654b2c\") " pod="openstack/ovn-controller-ovs-4667z" Oct 08 19:51:25 crc kubenswrapper[4988]: I1008 19:51:25.624561 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/a1187bea-e139-42e7-bd42-8086ed52bceb-var-log-ovn\") pod \"ovn-controller-2xddw\" (UID: \"a1187bea-e139-42e7-bd42-8086ed52bceb\") " pod="openstack/ovn-controller-2xddw" Oct 08 19:51:25 crc kubenswrapper[4988]: I1008 19:51:25.624523 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/eb6a17a8-576e-4d61-8467-cca342654b2c-etc-ovs\") pod \"ovn-controller-ovs-4667z\" (UID: \"eb6a17a8-576e-4d61-8467-cca342654b2c\") " pod="openstack/ovn-controller-ovs-4667z" Oct 08 19:51:25 crc kubenswrapper[4988]: I1008 19:51:25.624856 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/a1187bea-e139-42e7-bd42-8086ed52bceb-var-run-ovn\") pod \"ovn-controller-2xddw\" (UID: \"a1187bea-e139-42e7-bd42-8086ed52bceb\") " pod="openstack/ovn-controller-2xddw" Oct 08 19:51:25 crc kubenswrapper[4988]: I1008 19:51:25.624939 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/eb6a17a8-576e-4d61-8467-cca342654b2c-var-log\") pod \"ovn-controller-ovs-4667z\" (UID: \"eb6a17a8-576e-4d61-8467-cca342654b2c\") " pod="openstack/ovn-controller-ovs-4667z" Oct 08 19:51:25 crc kubenswrapper[4988]: I1008 19:51:25.636226 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a1187bea-e139-42e7-bd42-8086ed52bceb-combined-ca-bundle\") pod \"ovn-controller-2xddw\" (UID: \"a1187bea-e139-42e7-bd42-8086ed52bceb\") " pod="openstack/ovn-controller-2xddw" Oct 08 19:51:25 crc kubenswrapper[4988]: I1008 19:51:25.636832 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/a1187bea-e139-42e7-bd42-8086ed52bceb-ovn-controller-tls-certs\") pod \"ovn-controller-2xddw\" (UID: \"a1187bea-e139-42e7-bd42-8086ed52bceb\") " pod="openstack/ovn-controller-2xddw" Oct 08 19:51:25 crc kubenswrapper[4988]: I1008 19:51:25.638462 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a1187bea-e139-42e7-bd42-8086ed52bceb-scripts\") pod \"ovn-controller-2xddw\" (UID: \"a1187bea-e139-42e7-bd42-8086ed52bceb\") " pod="openstack/ovn-controller-2xddw" Oct 08 19:51:25 crc kubenswrapper[4988]: I1008 19:51:25.639031 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/eb6a17a8-576e-4d61-8467-cca342654b2c-scripts\") pod \"ovn-controller-ovs-4667z\" (UID: \"eb6a17a8-576e-4d61-8467-cca342654b2c\") " pod="openstack/ovn-controller-ovs-4667z" Oct 08 19:51:25 crc kubenswrapper[4988]: I1008 19:51:25.643061 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qmdlf\" (UniqueName: \"kubernetes.io/projected/a1187bea-e139-42e7-bd42-8086ed52bceb-kube-api-access-qmdlf\") pod \"ovn-controller-2xddw\" (UID: \"a1187bea-e139-42e7-bd42-8086ed52bceb\") " pod="openstack/ovn-controller-2xddw" Oct 08 19:51:25 crc kubenswrapper[4988]: I1008 19:51:25.648152 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6wnxm\" (UniqueName: \"kubernetes.io/projected/eb6a17a8-576e-4d61-8467-cca342654b2c-kube-api-access-6wnxm\") pod \"ovn-controller-ovs-4667z\" (UID: \"eb6a17a8-576e-4d61-8467-cca342654b2c\") " pod="openstack/ovn-controller-ovs-4667z" Oct 08 19:51:25 crc kubenswrapper[4988]: I1008 19:51:25.689976 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-2xddw" Oct 08 19:51:25 crc kubenswrapper[4988]: I1008 19:51:25.704543 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-4667z" Oct 08 19:51:26 crc kubenswrapper[4988]: I1008 19:51:26.207435 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-2xddw"] Oct 08 19:51:26 crc kubenswrapper[4988]: I1008 19:51:26.237360 4988 scope.go:117] "RemoveContainer" containerID="f2a29ed22dd7a48bbc929e65fbeceed969a669e78f65fbfb724b03dc0db6e947" Oct 08 19:51:26 crc kubenswrapper[4988]: W1008 19:51:26.581638 4988 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podeb6a17a8_576e_4d61_8467_cca342654b2c.slice/crio-e439cdc73793ade2cce66d0e678cf1bbd41eeb0bea4c51e8913aa9fcf525ee3c WatchSource:0}: Error finding container e439cdc73793ade2cce66d0e678cf1bbd41eeb0bea4c51e8913aa9fcf525ee3c: Status 404 returned error can't find the container with id e439cdc73793ade2cce66d0e678cf1bbd41eeb0bea4c51e8913aa9fcf525ee3c Oct 08 19:51:26 crc kubenswrapper[4988]: I1008 19:51:26.583504 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-4667z"] Oct 08 19:51:26 crc kubenswrapper[4988]: I1008 19:51:26.614955 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-metrics-pdhvt"] Oct 08 19:51:26 crc kubenswrapper[4988]: I1008 19:51:26.616054 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-pdhvt" Oct 08 19:51:26 crc kubenswrapper[4988]: I1008 19:51:26.618972 4988 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-metrics-config" Oct 08 19:51:26 crc kubenswrapper[4988]: I1008 19:51:26.630929 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-pdhvt"] Oct 08 19:51:26 crc kubenswrapper[4988]: I1008 19:51:26.765456 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/8c2854b9-1c17-48e4-82c7-84a4eb44f407-ovn-rundir\") pod \"ovn-controller-metrics-pdhvt\" (UID: \"8c2854b9-1c17-48e4-82c7-84a4eb44f407\") " pod="openstack/ovn-controller-metrics-pdhvt" Oct 08 19:51:26 crc kubenswrapper[4988]: I1008 19:51:26.765516 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8c2854b9-1c17-48e4-82c7-84a4eb44f407-config\") pod \"ovn-controller-metrics-pdhvt\" (UID: \"8c2854b9-1c17-48e4-82c7-84a4eb44f407\") " pod="openstack/ovn-controller-metrics-pdhvt" Oct 08 19:51:26 crc kubenswrapper[4988]: I1008 19:51:26.765728 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8c2854b9-1c17-48e4-82c7-84a4eb44f407-combined-ca-bundle\") pod \"ovn-controller-metrics-pdhvt\" (UID: \"8c2854b9-1c17-48e4-82c7-84a4eb44f407\") " pod="openstack/ovn-controller-metrics-pdhvt" Oct 08 19:51:26 crc kubenswrapper[4988]: I1008 19:51:26.765762 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/8c2854b9-1c17-48e4-82c7-84a4eb44f407-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-pdhvt\" (UID: \"8c2854b9-1c17-48e4-82c7-84a4eb44f407\") " pod="openstack/ovn-controller-metrics-pdhvt" Oct 08 19:51:26 crc kubenswrapper[4988]: I1008 19:51:26.765866 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/8c2854b9-1c17-48e4-82c7-84a4eb44f407-ovs-rundir\") pod \"ovn-controller-metrics-pdhvt\" (UID: \"8c2854b9-1c17-48e4-82c7-84a4eb44f407\") " pod="openstack/ovn-controller-metrics-pdhvt" Oct 08 19:51:26 crc kubenswrapper[4988]: I1008 19:51:26.765915 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gnztw\" (UniqueName: \"kubernetes.io/projected/8c2854b9-1c17-48e4-82c7-84a4eb44f407-kube-api-access-gnztw\") pod \"ovn-controller-metrics-pdhvt\" (UID: \"8c2854b9-1c17-48e4-82c7-84a4eb44f407\") " pod="openstack/ovn-controller-metrics-pdhvt" Oct 08 19:51:26 crc kubenswrapper[4988]: I1008 19:51:26.867581 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8c2854b9-1c17-48e4-82c7-84a4eb44f407-combined-ca-bundle\") pod \"ovn-controller-metrics-pdhvt\" (UID: \"8c2854b9-1c17-48e4-82c7-84a4eb44f407\") " pod="openstack/ovn-controller-metrics-pdhvt" Oct 08 19:51:26 crc kubenswrapper[4988]: I1008 19:51:26.867801 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/8c2854b9-1c17-48e4-82c7-84a4eb44f407-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-pdhvt\" (UID: \"8c2854b9-1c17-48e4-82c7-84a4eb44f407\") " pod="openstack/ovn-controller-metrics-pdhvt" Oct 08 19:51:26 crc kubenswrapper[4988]: I1008 19:51:26.867841 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/8c2854b9-1c17-48e4-82c7-84a4eb44f407-ovs-rundir\") pod \"ovn-controller-metrics-pdhvt\" (UID: \"8c2854b9-1c17-48e4-82c7-84a4eb44f407\") " pod="openstack/ovn-controller-metrics-pdhvt" Oct 08 19:51:26 crc kubenswrapper[4988]: I1008 19:51:26.867871 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gnztw\" (UniqueName: \"kubernetes.io/projected/8c2854b9-1c17-48e4-82c7-84a4eb44f407-kube-api-access-gnztw\") pod \"ovn-controller-metrics-pdhvt\" (UID: \"8c2854b9-1c17-48e4-82c7-84a4eb44f407\") " pod="openstack/ovn-controller-metrics-pdhvt" Oct 08 19:51:26 crc kubenswrapper[4988]: I1008 19:51:26.867905 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/8c2854b9-1c17-48e4-82c7-84a4eb44f407-ovn-rundir\") pod \"ovn-controller-metrics-pdhvt\" (UID: \"8c2854b9-1c17-48e4-82c7-84a4eb44f407\") " pod="openstack/ovn-controller-metrics-pdhvt" Oct 08 19:51:26 crc kubenswrapper[4988]: I1008 19:51:26.867928 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8c2854b9-1c17-48e4-82c7-84a4eb44f407-config\") pod \"ovn-controller-metrics-pdhvt\" (UID: \"8c2854b9-1c17-48e4-82c7-84a4eb44f407\") " pod="openstack/ovn-controller-metrics-pdhvt" Oct 08 19:51:26 crc kubenswrapper[4988]: I1008 19:51:26.868206 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/8c2854b9-1c17-48e4-82c7-84a4eb44f407-ovs-rundir\") pod \"ovn-controller-metrics-pdhvt\" (UID: \"8c2854b9-1c17-48e4-82c7-84a4eb44f407\") " pod="openstack/ovn-controller-metrics-pdhvt" Oct 08 19:51:26 crc kubenswrapper[4988]: I1008 19:51:26.868310 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/8c2854b9-1c17-48e4-82c7-84a4eb44f407-ovn-rundir\") pod \"ovn-controller-metrics-pdhvt\" (UID: \"8c2854b9-1c17-48e4-82c7-84a4eb44f407\") " pod="openstack/ovn-controller-metrics-pdhvt" Oct 08 19:51:26 crc kubenswrapper[4988]: I1008 19:51:26.868690 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8c2854b9-1c17-48e4-82c7-84a4eb44f407-config\") pod \"ovn-controller-metrics-pdhvt\" (UID: \"8c2854b9-1c17-48e4-82c7-84a4eb44f407\") " pod="openstack/ovn-controller-metrics-pdhvt" Oct 08 19:51:26 crc kubenswrapper[4988]: I1008 19:51:26.876213 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/8c2854b9-1c17-48e4-82c7-84a4eb44f407-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-pdhvt\" (UID: \"8c2854b9-1c17-48e4-82c7-84a4eb44f407\") " pod="openstack/ovn-controller-metrics-pdhvt" Oct 08 19:51:26 crc kubenswrapper[4988]: I1008 19:51:26.876323 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8c2854b9-1c17-48e4-82c7-84a4eb44f407-combined-ca-bundle\") pod \"ovn-controller-metrics-pdhvt\" (UID: \"8c2854b9-1c17-48e4-82c7-84a4eb44f407\") " pod="openstack/ovn-controller-metrics-pdhvt" Oct 08 19:51:26 crc kubenswrapper[4988]: I1008 19:51:26.892234 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gnztw\" (UniqueName: \"kubernetes.io/projected/8c2854b9-1c17-48e4-82c7-84a4eb44f407-kube-api-access-gnztw\") pod \"ovn-controller-metrics-pdhvt\" (UID: \"8c2854b9-1c17-48e4-82c7-84a4eb44f407\") " pod="openstack/ovn-controller-metrics-pdhvt" Oct 08 19:51:26 crc kubenswrapper[4988]: I1008 19:51:26.915887 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-4667z" event={"ID":"eb6a17a8-576e-4d61-8467-cca342654b2c","Type":"ContainerStarted","Data":"e439cdc73793ade2cce66d0e678cf1bbd41eeb0bea4c51e8913aa9fcf525ee3c"} Oct 08 19:51:26 crc kubenswrapper[4988]: I1008 19:51:26.922795 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" event={"ID":"f328a20b-5283-407e-8a93-2ab2da599f18","Type":"ContainerStarted","Data":"07bebacafd1c8af09c722a5994e896b519bd87cf3a710207584b6b2c926371f1"} Oct 08 19:51:26 crc kubenswrapper[4988]: I1008 19:51:26.924831 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-2xddw" event={"ID":"a1187bea-e139-42e7-bd42-8086ed52bceb","Type":"ContainerStarted","Data":"f077b9780b9bb2be2fb911c96b742c6a71100b18f2a4dd1d19e31a0ed33a3846"} Oct 08 19:51:26 crc kubenswrapper[4988]: I1008 19:51:26.924859 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-2xddw" event={"ID":"a1187bea-e139-42e7-bd42-8086ed52bceb","Type":"ContainerStarted","Data":"dff0ee11f5832caf78baed1b702b4dad53fbeb8e05bfa6961b978eecbfa7b884"} Oct 08 19:51:26 crc kubenswrapper[4988]: I1008 19:51:26.925293 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-2xddw" Oct 08 19:51:26 crc kubenswrapper[4988]: I1008 19:51:26.990745 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-2xddw" podStartSLOduration=1.990717274 podStartE2EDuration="1.990717274s" podCreationTimestamp="2025-10-08 19:51:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 19:51:26.968823769 +0000 UTC m=+6032.418666549" watchObservedRunningTime="2025-10-08 19:51:26.990717274 +0000 UTC m=+6032.440560064" Oct 08 19:51:27 crc kubenswrapper[4988]: I1008 19:51:27.010916 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-pdhvt" Oct 08 19:51:27 crc kubenswrapper[4988]: I1008 19:51:27.017502 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/octavia-db-create-lfm67"] Oct 08 19:51:27 crc kubenswrapper[4988]: I1008 19:51:27.018747 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-db-create-lfm67" Oct 08 19:51:27 crc kubenswrapper[4988]: I1008 19:51:27.037484 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-db-create-lfm67"] Oct 08 19:51:27 crc kubenswrapper[4988]: I1008 19:51:27.177465 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zrlrv\" (UniqueName: \"kubernetes.io/projected/0dcba654-49b6-4faf-9013-cca23d157c74-kube-api-access-zrlrv\") pod \"octavia-db-create-lfm67\" (UID: \"0dcba654-49b6-4faf-9013-cca23d157c74\") " pod="openstack/octavia-db-create-lfm67" Oct 08 19:51:27 crc kubenswrapper[4988]: I1008 19:51:27.279852 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zrlrv\" (UniqueName: \"kubernetes.io/projected/0dcba654-49b6-4faf-9013-cca23d157c74-kube-api-access-zrlrv\") pod \"octavia-db-create-lfm67\" (UID: \"0dcba654-49b6-4faf-9013-cca23d157c74\") " pod="openstack/octavia-db-create-lfm67" Oct 08 19:51:27 crc kubenswrapper[4988]: I1008 19:51:27.300204 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zrlrv\" (UniqueName: \"kubernetes.io/projected/0dcba654-49b6-4faf-9013-cca23d157c74-kube-api-access-zrlrv\") pod \"octavia-db-create-lfm67\" (UID: \"0dcba654-49b6-4faf-9013-cca23d157c74\") " pod="openstack/octavia-db-create-lfm67" Oct 08 19:51:27 crc kubenswrapper[4988]: I1008 19:51:27.352559 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-db-create-lfm67" Oct 08 19:51:27 crc kubenswrapper[4988]: I1008 19:51:27.565544 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-pdhvt"] Oct 08 19:51:27 crc kubenswrapper[4988]: W1008 19:51:27.574140 4988 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8c2854b9_1c17_48e4_82c7_84a4eb44f407.slice/crio-a16fc7fe80a422e99b5ae8e94195daed8e0f8d77b0f7620a13f7b1bb01310286 WatchSource:0}: Error finding container a16fc7fe80a422e99b5ae8e94195daed8e0f8d77b0f7620a13f7b1bb01310286: Status 404 returned error can't find the container with id a16fc7fe80a422e99b5ae8e94195daed8e0f8d77b0f7620a13f7b1bb01310286 Oct 08 19:51:27 crc kubenswrapper[4988]: I1008 19:51:27.846449 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-db-create-lfm67"] Oct 08 19:51:27 crc kubenswrapper[4988]: W1008 19:51:27.852607 4988 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0dcba654_49b6_4faf_9013_cca23d157c74.slice/crio-767b7dff02eda4edc66a3b3f6caae28da06fb5ab209c93aa1ac02c6d609bddba WatchSource:0}: Error finding container 767b7dff02eda4edc66a3b3f6caae28da06fb5ab209c93aa1ac02c6d609bddba: Status 404 returned error can't find the container with id 767b7dff02eda4edc66a3b3f6caae28da06fb5ab209c93aa1ac02c6d609bddba Oct 08 19:51:27 crc kubenswrapper[4988]: I1008 19:51:27.994484 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-pdhvt" event={"ID":"8c2854b9-1c17-48e4-82c7-84a4eb44f407","Type":"ContainerStarted","Data":"3e1c560594e4d0c85c386f0a8892981a0e005b58e1e28a20993734fa14f5cc26"} Oct 08 19:51:27 crc kubenswrapper[4988]: I1008 19:51:27.994847 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-pdhvt" event={"ID":"8c2854b9-1c17-48e4-82c7-84a4eb44f407","Type":"ContainerStarted","Data":"a16fc7fe80a422e99b5ae8e94195daed8e0f8d77b0f7620a13f7b1bb01310286"} Oct 08 19:51:27 crc kubenswrapper[4988]: I1008 19:51:27.999203 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-db-create-lfm67" event={"ID":"0dcba654-49b6-4faf-9013-cca23d157c74","Type":"ContainerStarted","Data":"767b7dff02eda4edc66a3b3f6caae28da06fb5ab209c93aa1ac02c6d609bddba"} Oct 08 19:51:28 crc kubenswrapper[4988]: I1008 19:51:28.011485 4988 generic.go:334] "Generic (PLEG): container finished" podID="eb6a17a8-576e-4d61-8467-cca342654b2c" containerID="5d595ff06405c53bc1b7f428e4cb290f856af5c741f228bccd42bdd6b808c834" exitCode=0 Oct 08 19:51:28 crc kubenswrapper[4988]: I1008 19:51:28.014036 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-4667z" event={"ID":"eb6a17a8-576e-4d61-8467-cca342654b2c","Type":"ContainerDied","Data":"5d595ff06405c53bc1b7f428e4cb290f856af5c741f228bccd42bdd6b808c834"} Oct 08 19:51:28 crc kubenswrapper[4988]: I1008 19:51:28.056036 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-metrics-pdhvt" podStartSLOduration=2.056014258 podStartE2EDuration="2.056014258s" podCreationTimestamp="2025-10-08 19:51:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 19:51:28.04853764 +0000 UTC m=+6033.498380410" watchObservedRunningTime="2025-10-08 19:51:28.056014258 +0000 UTC m=+6033.505857028" Oct 08 19:51:29 crc kubenswrapper[4988]: I1008 19:51:29.045780 4988 generic.go:334] "Generic (PLEG): container finished" podID="0dcba654-49b6-4faf-9013-cca23d157c74" containerID="c154f8ad7eb6341238ad4b33fbbc84fb4d6309bbbd4d2aab84058e8f4371b9c8" exitCode=0 Oct 08 19:51:29 crc kubenswrapper[4988]: I1008 19:51:29.045837 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-db-create-lfm67" event={"ID":"0dcba654-49b6-4faf-9013-cca23d157c74","Type":"ContainerDied","Data":"c154f8ad7eb6341238ad4b33fbbc84fb4d6309bbbd4d2aab84058e8f4371b9c8"} Oct 08 19:51:29 crc kubenswrapper[4988]: I1008 19:51:29.050127 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-4667z" event={"ID":"eb6a17a8-576e-4d61-8467-cca342654b2c","Type":"ContainerStarted","Data":"5a51fd446ac2ec2e95e5ae82d022986572694cbc4e45c0c43c0ddd6709ac2171"} Oct 08 19:51:29 crc kubenswrapper[4988]: I1008 19:51:29.050167 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-4667z" event={"ID":"eb6a17a8-576e-4d61-8467-cca342654b2c","Type":"ContainerStarted","Data":"64a1d11a1d91d059da22f58dccdc56e532de6efeae75efd574add7f52513674c"} Oct 08 19:51:29 crc kubenswrapper[4988]: I1008 19:51:29.050952 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-4667z" Oct 08 19:51:29 crc kubenswrapper[4988]: I1008 19:51:29.051064 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-4667z" Oct 08 19:51:29 crc kubenswrapper[4988]: I1008 19:51:29.084898 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-ovs-4667z" podStartSLOduration=4.084882883 podStartE2EDuration="4.084882883s" podCreationTimestamp="2025-10-08 19:51:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 19:51:29.077160168 +0000 UTC m=+6034.527002938" watchObservedRunningTime="2025-10-08 19:51:29.084882883 +0000 UTC m=+6034.534725653" Oct 08 19:51:30 crc kubenswrapper[4988]: I1008 19:51:30.421313 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-db-create-lfm67" Oct 08 19:51:30 crc kubenswrapper[4988]: I1008 19:51:30.548632 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zrlrv\" (UniqueName: \"kubernetes.io/projected/0dcba654-49b6-4faf-9013-cca23d157c74-kube-api-access-zrlrv\") pod \"0dcba654-49b6-4faf-9013-cca23d157c74\" (UID: \"0dcba654-49b6-4faf-9013-cca23d157c74\") " Oct 08 19:51:30 crc kubenswrapper[4988]: I1008 19:51:30.558626 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0dcba654-49b6-4faf-9013-cca23d157c74-kube-api-access-zrlrv" (OuterVolumeSpecName: "kube-api-access-zrlrv") pod "0dcba654-49b6-4faf-9013-cca23d157c74" (UID: "0dcba654-49b6-4faf-9013-cca23d157c74"). InnerVolumeSpecName "kube-api-access-zrlrv". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 19:51:30 crc kubenswrapper[4988]: I1008 19:51:30.651368 4988 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zrlrv\" (UniqueName: \"kubernetes.io/projected/0dcba654-49b6-4faf-9013-cca23d157c74-kube-api-access-zrlrv\") on node \"crc\" DevicePath \"\"" Oct 08 19:51:31 crc kubenswrapper[4988]: I1008 19:51:31.073644 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-db-create-lfm67" event={"ID":"0dcba654-49b6-4faf-9013-cca23d157c74","Type":"ContainerDied","Data":"767b7dff02eda4edc66a3b3f6caae28da06fb5ab209c93aa1ac02c6d609bddba"} Oct 08 19:51:31 crc kubenswrapper[4988]: I1008 19:51:31.073702 4988 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="767b7dff02eda4edc66a3b3f6caae28da06fb5ab209c93aa1ac02c6d609bddba" Oct 08 19:51:31 crc kubenswrapper[4988]: I1008 19:51:31.073784 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-db-create-lfm67" Oct 08 19:51:37 crc kubenswrapper[4988]: I1008 19:51:37.731926 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/octavia-70cf-account-create-66n9b"] Oct 08 19:51:37 crc kubenswrapper[4988]: E1008 19:51:37.732702 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0dcba654-49b6-4faf-9013-cca23d157c74" containerName="mariadb-database-create" Oct 08 19:51:37 crc kubenswrapper[4988]: I1008 19:51:37.732715 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="0dcba654-49b6-4faf-9013-cca23d157c74" containerName="mariadb-database-create" Oct 08 19:51:37 crc kubenswrapper[4988]: I1008 19:51:37.732905 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="0dcba654-49b6-4faf-9013-cca23d157c74" containerName="mariadb-database-create" Oct 08 19:51:37 crc kubenswrapper[4988]: I1008 19:51:37.733547 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-70cf-account-create-66n9b" Oct 08 19:51:37 crc kubenswrapper[4988]: I1008 19:51:37.736639 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"octavia-db-secret" Oct 08 19:51:37 crc kubenswrapper[4988]: I1008 19:51:37.755358 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-70cf-account-create-66n9b"] Oct 08 19:51:37 crc kubenswrapper[4988]: I1008 19:51:37.816906 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gd9n6\" (UniqueName: \"kubernetes.io/projected/75cd79f3-6abb-4eb2-9e2a-02c6da1eb074-kube-api-access-gd9n6\") pod \"octavia-70cf-account-create-66n9b\" (UID: \"75cd79f3-6abb-4eb2-9e2a-02c6da1eb074\") " pod="openstack/octavia-70cf-account-create-66n9b" Oct 08 19:51:37 crc kubenswrapper[4988]: I1008 19:51:37.919256 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gd9n6\" (UniqueName: \"kubernetes.io/projected/75cd79f3-6abb-4eb2-9e2a-02c6da1eb074-kube-api-access-gd9n6\") pod \"octavia-70cf-account-create-66n9b\" (UID: \"75cd79f3-6abb-4eb2-9e2a-02c6da1eb074\") " pod="openstack/octavia-70cf-account-create-66n9b" Oct 08 19:51:37 crc kubenswrapper[4988]: I1008 19:51:37.941486 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gd9n6\" (UniqueName: \"kubernetes.io/projected/75cd79f3-6abb-4eb2-9e2a-02c6da1eb074-kube-api-access-gd9n6\") pod \"octavia-70cf-account-create-66n9b\" (UID: \"75cd79f3-6abb-4eb2-9e2a-02c6da1eb074\") " pod="openstack/octavia-70cf-account-create-66n9b" Oct 08 19:51:38 crc kubenswrapper[4988]: I1008 19:51:38.053825 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-70cf-account-create-66n9b" Oct 08 19:51:38 crc kubenswrapper[4988]: I1008 19:51:38.643580 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-70cf-account-create-66n9b"] Oct 08 19:51:39 crc kubenswrapper[4988]: I1008 19:51:39.174033 4988 generic.go:334] "Generic (PLEG): container finished" podID="75cd79f3-6abb-4eb2-9e2a-02c6da1eb074" containerID="8b3820b3ba038da4c218186f9a302029947854aaaaac288e9947ee6eb75efb33" exitCode=0 Oct 08 19:51:39 crc kubenswrapper[4988]: I1008 19:51:39.174106 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-70cf-account-create-66n9b" event={"ID":"75cd79f3-6abb-4eb2-9e2a-02c6da1eb074","Type":"ContainerDied","Data":"8b3820b3ba038da4c218186f9a302029947854aaaaac288e9947ee6eb75efb33"} Oct 08 19:51:39 crc kubenswrapper[4988]: I1008 19:51:39.174166 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-70cf-account-create-66n9b" event={"ID":"75cd79f3-6abb-4eb2-9e2a-02c6da1eb074","Type":"ContainerStarted","Data":"1fb8287728ff527d3dbd697eef5035773f28cab60a7f6423b1e1454773ca4d62"} Oct 08 19:51:40 crc kubenswrapper[4988]: I1008 19:51:40.596024 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-70cf-account-create-66n9b" Oct 08 19:51:40 crc kubenswrapper[4988]: I1008 19:51:40.677329 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gd9n6\" (UniqueName: \"kubernetes.io/projected/75cd79f3-6abb-4eb2-9e2a-02c6da1eb074-kube-api-access-gd9n6\") pod \"75cd79f3-6abb-4eb2-9e2a-02c6da1eb074\" (UID: \"75cd79f3-6abb-4eb2-9e2a-02c6da1eb074\") " Oct 08 19:51:40 crc kubenswrapper[4988]: I1008 19:51:40.682731 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/75cd79f3-6abb-4eb2-9e2a-02c6da1eb074-kube-api-access-gd9n6" (OuterVolumeSpecName: "kube-api-access-gd9n6") pod "75cd79f3-6abb-4eb2-9e2a-02c6da1eb074" (UID: "75cd79f3-6abb-4eb2-9e2a-02c6da1eb074"). InnerVolumeSpecName "kube-api-access-gd9n6". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 19:51:40 crc kubenswrapper[4988]: I1008 19:51:40.779420 4988 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gd9n6\" (UniqueName: \"kubernetes.io/projected/75cd79f3-6abb-4eb2-9e2a-02c6da1eb074-kube-api-access-gd9n6\") on node \"crc\" DevicePath \"\"" Oct 08 19:51:41 crc kubenswrapper[4988]: I1008 19:51:41.197592 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-70cf-account-create-66n9b" event={"ID":"75cd79f3-6abb-4eb2-9e2a-02c6da1eb074","Type":"ContainerDied","Data":"1fb8287728ff527d3dbd697eef5035773f28cab60a7f6423b1e1454773ca4d62"} Oct 08 19:51:41 crc kubenswrapper[4988]: I1008 19:51:41.197942 4988 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1fb8287728ff527d3dbd697eef5035773f28cab60a7f6423b1e1454773ca4d62" Oct 08 19:51:41 crc kubenswrapper[4988]: I1008 19:51:41.197721 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-70cf-account-create-66n9b" Oct 08 19:51:43 crc kubenswrapper[4988]: I1008 19:51:43.857280 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/octavia-persistence-db-create-7f7s2"] Oct 08 19:51:43 crc kubenswrapper[4988]: E1008 19:51:43.859089 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="75cd79f3-6abb-4eb2-9e2a-02c6da1eb074" containerName="mariadb-account-create" Oct 08 19:51:43 crc kubenswrapper[4988]: I1008 19:51:43.859198 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="75cd79f3-6abb-4eb2-9e2a-02c6da1eb074" containerName="mariadb-account-create" Oct 08 19:51:43 crc kubenswrapper[4988]: I1008 19:51:43.859555 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="75cd79f3-6abb-4eb2-9e2a-02c6da1eb074" containerName="mariadb-account-create" Oct 08 19:51:43 crc kubenswrapper[4988]: I1008 19:51:43.860440 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-persistence-db-create-7f7s2" Oct 08 19:51:43 crc kubenswrapper[4988]: I1008 19:51:43.906789 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-persistence-db-create-7f7s2"] Oct 08 19:51:43 crc kubenswrapper[4988]: I1008 19:51:43.946082 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4djh9\" (UniqueName: \"kubernetes.io/projected/039811dd-898b-4b0f-bf09-6530a670233e-kube-api-access-4djh9\") pod \"octavia-persistence-db-create-7f7s2\" (UID: \"039811dd-898b-4b0f-bf09-6530a670233e\") " pod="openstack/octavia-persistence-db-create-7f7s2" Oct 08 19:51:44 crc kubenswrapper[4988]: I1008 19:51:44.047342 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4djh9\" (UniqueName: \"kubernetes.io/projected/039811dd-898b-4b0f-bf09-6530a670233e-kube-api-access-4djh9\") pod \"octavia-persistence-db-create-7f7s2\" (UID: \"039811dd-898b-4b0f-bf09-6530a670233e\") " pod="openstack/octavia-persistence-db-create-7f7s2" Oct 08 19:51:44 crc kubenswrapper[4988]: I1008 19:51:44.072008 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4djh9\" (UniqueName: \"kubernetes.io/projected/039811dd-898b-4b0f-bf09-6530a670233e-kube-api-access-4djh9\") pod \"octavia-persistence-db-create-7f7s2\" (UID: \"039811dd-898b-4b0f-bf09-6530a670233e\") " pod="openstack/octavia-persistence-db-create-7f7s2" Oct 08 19:51:44 crc kubenswrapper[4988]: I1008 19:51:44.192656 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-persistence-db-create-7f7s2" Oct 08 19:51:44 crc kubenswrapper[4988]: I1008 19:51:44.649819 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-persistence-db-create-7f7s2"] Oct 08 19:51:45 crc kubenswrapper[4988]: I1008 19:51:45.234514 4988 generic.go:334] "Generic (PLEG): container finished" podID="039811dd-898b-4b0f-bf09-6530a670233e" containerID="fcb652f27029f5c4ec2dad99a9ccd16e58b1b8bc4b61fe19dd9f66678a1e730f" exitCode=0 Oct 08 19:51:45 crc kubenswrapper[4988]: I1008 19:51:45.234730 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-persistence-db-create-7f7s2" event={"ID":"039811dd-898b-4b0f-bf09-6530a670233e","Type":"ContainerDied","Data":"fcb652f27029f5c4ec2dad99a9ccd16e58b1b8bc4b61fe19dd9f66678a1e730f"} Oct 08 19:51:45 crc kubenswrapper[4988]: I1008 19:51:45.234827 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-persistence-db-create-7f7s2" event={"ID":"039811dd-898b-4b0f-bf09-6530a670233e","Type":"ContainerStarted","Data":"4a66813571480b76c38d9bae31d97876429415b1b9b6900eea7aad652073a925"} Oct 08 19:51:46 crc kubenswrapper[4988]: I1008 19:51:46.726473 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-persistence-db-create-7f7s2" Oct 08 19:51:46 crc kubenswrapper[4988]: I1008 19:51:46.802391 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4djh9\" (UniqueName: \"kubernetes.io/projected/039811dd-898b-4b0f-bf09-6530a670233e-kube-api-access-4djh9\") pod \"039811dd-898b-4b0f-bf09-6530a670233e\" (UID: \"039811dd-898b-4b0f-bf09-6530a670233e\") " Oct 08 19:51:46 crc kubenswrapper[4988]: I1008 19:51:46.809368 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/039811dd-898b-4b0f-bf09-6530a670233e-kube-api-access-4djh9" (OuterVolumeSpecName: "kube-api-access-4djh9") pod "039811dd-898b-4b0f-bf09-6530a670233e" (UID: "039811dd-898b-4b0f-bf09-6530a670233e"). InnerVolumeSpecName "kube-api-access-4djh9". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 19:51:46 crc kubenswrapper[4988]: I1008 19:51:46.905284 4988 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4djh9\" (UniqueName: \"kubernetes.io/projected/039811dd-898b-4b0f-bf09-6530a670233e-kube-api-access-4djh9\") on node \"crc\" DevicePath \"\"" Oct 08 19:51:47 crc kubenswrapper[4988]: I1008 19:51:47.255355 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-persistence-db-create-7f7s2" event={"ID":"039811dd-898b-4b0f-bf09-6530a670233e","Type":"ContainerDied","Data":"4a66813571480b76c38d9bae31d97876429415b1b9b6900eea7aad652073a925"} Oct 08 19:51:47 crc kubenswrapper[4988]: I1008 19:51:47.255398 4988 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4a66813571480b76c38d9bae31d97876429415b1b9b6900eea7aad652073a925" Oct 08 19:51:47 crc kubenswrapper[4988]: I1008 19:51:47.255467 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-persistence-db-create-7f7s2" Oct 08 19:51:54 crc kubenswrapper[4988]: I1008 19:51:54.823755 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/octavia-0b5f-account-create-mvdkn"] Oct 08 19:51:54 crc kubenswrapper[4988]: E1008 19:51:54.824964 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="039811dd-898b-4b0f-bf09-6530a670233e" containerName="mariadb-database-create" Oct 08 19:51:54 crc kubenswrapper[4988]: I1008 19:51:54.824987 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="039811dd-898b-4b0f-bf09-6530a670233e" containerName="mariadb-database-create" Oct 08 19:51:54 crc kubenswrapper[4988]: I1008 19:51:54.825291 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="039811dd-898b-4b0f-bf09-6530a670233e" containerName="mariadb-database-create" Oct 08 19:51:54 crc kubenswrapper[4988]: I1008 19:51:54.826294 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-0b5f-account-create-mvdkn" Oct 08 19:51:54 crc kubenswrapper[4988]: I1008 19:51:54.836587 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"octavia-persistence-db-secret" Oct 08 19:51:54 crc kubenswrapper[4988]: I1008 19:51:54.845283 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-0b5f-account-create-mvdkn"] Oct 08 19:51:55 crc kubenswrapper[4988]: I1008 19:51:55.018039 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6t6mv\" (UniqueName: \"kubernetes.io/projected/868eea53-6829-45cf-9ebd-c504223fe1ae-kube-api-access-6t6mv\") pod \"octavia-0b5f-account-create-mvdkn\" (UID: \"868eea53-6829-45cf-9ebd-c504223fe1ae\") " pod="openstack/octavia-0b5f-account-create-mvdkn" Oct 08 19:51:55 crc kubenswrapper[4988]: I1008 19:51:55.120432 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6t6mv\" (UniqueName: \"kubernetes.io/projected/868eea53-6829-45cf-9ebd-c504223fe1ae-kube-api-access-6t6mv\") pod \"octavia-0b5f-account-create-mvdkn\" (UID: \"868eea53-6829-45cf-9ebd-c504223fe1ae\") " pod="openstack/octavia-0b5f-account-create-mvdkn" Oct 08 19:51:55 crc kubenswrapper[4988]: I1008 19:51:55.151451 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6t6mv\" (UniqueName: \"kubernetes.io/projected/868eea53-6829-45cf-9ebd-c504223fe1ae-kube-api-access-6t6mv\") pod \"octavia-0b5f-account-create-mvdkn\" (UID: \"868eea53-6829-45cf-9ebd-c504223fe1ae\") " pod="openstack/octavia-0b5f-account-create-mvdkn" Oct 08 19:51:55 crc kubenswrapper[4988]: I1008 19:51:55.163265 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-0b5f-account-create-mvdkn" Oct 08 19:51:55 crc kubenswrapper[4988]: I1008 19:51:55.637357 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-0b5f-account-create-mvdkn"] Oct 08 19:51:55 crc kubenswrapper[4988]: W1008 19:51:55.642018 4988 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod868eea53_6829_45cf_9ebd_c504223fe1ae.slice/crio-51d7aa10535d69929b0d940d20a15c0e9d2a58b76e0610d3a5b023e597a0650d WatchSource:0}: Error finding container 51d7aa10535d69929b0d940d20a15c0e9d2a58b76e0610d3a5b023e597a0650d: Status 404 returned error can't find the container with id 51d7aa10535d69929b0d940d20a15c0e9d2a58b76e0610d3a5b023e597a0650d Oct 08 19:51:55 crc kubenswrapper[4988]: I1008 19:51:55.646895 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"octavia-persistence-db-secret" Oct 08 19:51:56 crc kubenswrapper[4988]: I1008 19:51:56.417718 4988 generic.go:334] "Generic (PLEG): container finished" podID="868eea53-6829-45cf-9ebd-c504223fe1ae" containerID="4aa0d55b4f3d154ace085d1fa54fbcd4ff8e3d440884648b943272a9e4954e63" exitCode=0 Oct 08 19:51:56 crc kubenswrapper[4988]: I1008 19:51:56.417871 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-0b5f-account-create-mvdkn" event={"ID":"868eea53-6829-45cf-9ebd-c504223fe1ae","Type":"ContainerDied","Data":"4aa0d55b4f3d154ace085d1fa54fbcd4ff8e3d440884648b943272a9e4954e63"} Oct 08 19:51:56 crc kubenswrapper[4988]: I1008 19:51:56.418109 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-0b5f-account-create-mvdkn" event={"ID":"868eea53-6829-45cf-9ebd-c504223fe1ae","Type":"ContainerStarted","Data":"51d7aa10535d69929b0d940d20a15c0e9d2a58b76e0610d3a5b023e597a0650d"} Oct 08 19:51:57 crc kubenswrapper[4988]: I1008 19:51:57.885486 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-0b5f-account-create-mvdkn" Oct 08 19:51:57 crc kubenswrapper[4988]: I1008 19:51:57.989206 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6t6mv\" (UniqueName: \"kubernetes.io/projected/868eea53-6829-45cf-9ebd-c504223fe1ae-kube-api-access-6t6mv\") pod \"868eea53-6829-45cf-9ebd-c504223fe1ae\" (UID: \"868eea53-6829-45cf-9ebd-c504223fe1ae\") " Oct 08 19:51:57 crc kubenswrapper[4988]: I1008 19:51:57.997785 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/868eea53-6829-45cf-9ebd-c504223fe1ae-kube-api-access-6t6mv" (OuterVolumeSpecName: "kube-api-access-6t6mv") pod "868eea53-6829-45cf-9ebd-c504223fe1ae" (UID: "868eea53-6829-45cf-9ebd-c504223fe1ae"). InnerVolumeSpecName "kube-api-access-6t6mv". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 19:51:58 crc kubenswrapper[4988]: I1008 19:51:58.092348 4988 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6t6mv\" (UniqueName: \"kubernetes.io/projected/868eea53-6829-45cf-9ebd-c504223fe1ae-kube-api-access-6t6mv\") on node \"crc\" DevicePath \"\"" Oct 08 19:51:58 crc kubenswrapper[4988]: I1008 19:51:58.446561 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-0b5f-account-create-mvdkn" event={"ID":"868eea53-6829-45cf-9ebd-c504223fe1ae","Type":"ContainerDied","Data":"51d7aa10535d69929b0d940d20a15c0e9d2a58b76e0610d3a5b023e597a0650d"} Oct 08 19:51:58 crc kubenswrapper[4988]: I1008 19:51:58.447043 4988 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="51d7aa10535d69929b0d940d20a15c0e9d2a58b76e0610d3a5b023e597a0650d" Oct 08 19:51:58 crc kubenswrapper[4988]: I1008 19:51:58.446679 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-0b5f-account-create-mvdkn" Oct 08 19:52:00 crc kubenswrapper[4988]: I1008 19:52:00.767712 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-2xddw" Oct 08 19:52:00 crc kubenswrapper[4988]: I1008 19:52:00.780352 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-4667z" Oct 08 19:52:00 crc kubenswrapper[4988]: I1008 19:52:00.787662 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-4667z" Oct 08 19:52:00 crc kubenswrapper[4988]: I1008 19:52:00.930168 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-2xddw-config-cmjlc"] Oct 08 19:52:00 crc kubenswrapper[4988]: E1008 19:52:00.930580 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="868eea53-6829-45cf-9ebd-c504223fe1ae" containerName="mariadb-account-create" Oct 08 19:52:00 crc kubenswrapper[4988]: I1008 19:52:00.930595 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="868eea53-6829-45cf-9ebd-c504223fe1ae" containerName="mariadb-account-create" Oct 08 19:52:00 crc kubenswrapper[4988]: I1008 19:52:00.930794 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="868eea53-6829-45cf-9ebd-c504223fe1ae" containerName="mariadb-account-create" Oct 08 19:52:00 crc kubenswrapper[4988]: I1008 19:52:00.931758 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-2xddw-config-cmjlc" Oct 08 19:52:00 crc kubenswrapper[4988]: I1008 19:52:00.933628 4988 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-extra-scripts" Oct 08 19:52:00 crc kubenswrapper[4988]: I1008 19:52:00.960826 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-2xddw-config-cmjlc"] Oct 08 19:52:00 crc kubenswrapper[4988]: I1008 19:52:00.975588 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/a1d7994b-dab0-4357-984a-93070a03f8d8-var-log-ovn\") pod \"ovn-controller-2xddw-config-cmjlc\" (UID: \"a1d7994b-dab0-4357-984a-93070a03f8d8\") " pod="openstack/ovn-controller-2xddw-config-cmjlc" Oct 08 19:52:00 crc kubenswrapper[4988]: I1008 19:52:00.975710 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wp7bl\" (UniqueName: \"kubernetes.io/projected/a1d7994b-dab0-4357-984a-93070a03f8d8-kube-api-access-wp7bl\") pod \"ovn-controller-2xddw-config-cmjlc\" (UID: \"a1d7994b-dab0-4357-984a-93070a03f8d8\") " pod="openstack/ovn-controller-2xddw-config-cmjlc" Oct 08 19:52:00 crc kubenswrapper[4988]: I1008 19:52:00.975760 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/a1d7994b-dab0-4357-984a-93070a03f8d8-var-run-ovn\") pod \"ovn-controller-2xddw-config-cmjlc\" (UID: \"a1d7994b-dab0-4357-984a-93070a03f8d8\") " pod="openstack/ovn-controller-2xddw-config-cmjlc" Oct 08 19:52:00 crc kubenswrapper[4988]: I1008 19:52:00.975830 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/a1d7994b-dab0-4357-984a-93070a03f8d8-additional-scripts\") pod \"ovn-controller-2xddw-config-cmjlc\" (UID: \"a1d7994b-dab0-4357-984a-93070a03f8d8\") " pod="openstack/ovn-controller-2xddw-config-cmjlc" Oct 08 19:52:00 crc kubenswrapper[4988]: I1008 19:52:00.975884 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a1d7994b-dab0-4357-984a-93070a03f8d8-scripts\") pod \"ovn-controller-2xddw-config-cmjlc\" (UID: \"a1d7994b-dab0-4357-984a-93070a03f8d8\") " pod="openstack/ovn-controller-2xddw-config-cmjlc" Oct 08 19:52:00 crc kubenswrapper[4988]: I1008 19:52:00.975930 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/a1d7994b-dab0-4357-984a-93070a03f8d8-var-run\") pod \"ovn-controller-2xddw-config-cmjlc\" (UID: \"a1d7994b-dab0-4357-984a-93070a03f8d8\") " pod="openstack/ovn-controller-2xddw-config-cmjlc" Oct 08 19:52:00 crc kubenswrapper[4988]: I1008 19:52:00.984106 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/octavia-api-5d77847989-vhxvn"] Oct 08 19:52:00 crc kubenswrapper[4988]: I1008 19:52:00.986264 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-api-5d77847989-vhxvn" Oct 08 19:52:00 crc kubenswrapper[4988]: I1008 19:52:00.990697 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"octavia-octavia-dockercfg-ttz46" Oct 08 19:52:01 crc kubenswrapper[4988]: I1008 19:52:01.005239 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"octavia-api-config-data" Oct 08 19:52:01 crc kubenswrapper[4988]: I1008 19:52:01.007286 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"octavia-api-scripts" Oct 08 19:52:01 crc kubenswrapper[4988]: I1008 19:52:01.009341 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-octavia-ovndbs" Oct 08 19:52:01 crc kubenswrapper[4988]: I1008 19:52:01.036766 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-api-5d77847989-vhxvn"] Oct 08 19:52:01 crc kubenswrapper[4988]: I1008 19:52:01.078417 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/c144edcb-2dfc-4f6d-85a3-77ae04fd9fb5-config-data-merged\") pod \"octavia-api-5d77847989-vhxvn\" (UID: \"c144edcb-2dfc-4f6d-85a3-77ae04fd9fb5\") " pod="openstack/octavia-api-5d77847989-vhxvn" Oct 08 19:52:01 crc kubenswrapper[4988]: I1008 19:52:01.078460 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c144edcb-2dfc-4f6d-85a3-77ae04fd9fb5-scripts\") pod \"octavia-api-5d77847989-vhxvn\" (UID: \"c144edcb-2dfc-4f6d-85a3-77ae04fd9fb5\") " pod="openstack/octavia-api-5d77847989-vhxvn" Oct 08 19:52:01 crc kubenswrapper[4988]: I1008 19:52:01.078485 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/a1d7994b-dab0-4357-984a-93070a03f8d8-var-log-ovn\") pod \"ovn-controller-2xddw-config-cmjlc\" (UID: \"a1d7994b-dab0-4357-984a-93070a03f8d8\") " pod="openstack/ovn-controller-2xddw-config-cmjlc" Oct 08 19:52:01 crc kubenswrapper[4988]: I1008 19:52:01.078506 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"octavia-run\" (UniqueName: \"kubernetes.io/empty-dir/c144edcb-2dfc-4f6d-85a3-77ae04fd9fb5-octavia-run\") pod \"octavia-api-5d77847989-vhxvn\" (UID: \"c144edcb-2dfc-4f6d-85a3-77ae04fd9fb5\") " pod="openstack/octavia-api-5d77847989-vhxvn" Oct 08 19:52:01 crc kubenswrapper[4988]: I1008 19:52:01.078522 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c144edcb-2dfc-4f6d-85a3-77ae04fd9fb5-config-data\") pod \"octavia-api-5d77847989-vhxvn\" (UID: \"c144edcb-2dfc-4f6d-85a3-77ae04fd9fb5\") " pod="openstack/octavia-api-5d77847989-vhxvn" Oct 08 19:52:01 crc kubenswrapper[4988]: I1008 19:52:01.078791 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/c144edcb-2dfc-4f6d-85a3-77ae04fd9fb5-ovndb-tls-certs\") pod \"octavia-api-5d77847989-vhxvn\" (UID: \"c144edcb-2dfc-4f6d-85a3-77ae04fd9fb5\") " pod="openstack/octavia-api-5d77847989-vhxvn" Oct 08 19:52:01 crc kubenswrapper[4988]: I1008 19:52:01.078881 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wp7bl\" (UniqueName: \"kubernetes.io/projected/a1d7994b-dab0-4357-984a-93070a03f8d8-kube-api-access-wp7bl\") pod \"ovn-controller-2xddw-config-cmjlc\" (UID: \"a1d7994b-dab0-4357-984a-93070a03f8d8\") " pod="openstack/ovn-controller-2xddw-config-cmjlc" Oct 08 19:52:01 crc kubenswrapper[4988]: I1008 19:52:01.078950 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/a1d7994b-dab0-4357-984a-93070a03f8d8-var-run-ovn\") pod \"ovn-controller-2xddw-config-cmjlc\" (UID: \"a1d7994b-dab0-4357-984a-93070a03f8d8\") " pod="openstack/ovn-controller-2xddw-config-cmjlc" Oct 08 19:52:01 crc kubenswrapper[4988]: I1008 19:52:01.079046 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/a1d7994b-dab0-4357-984a-93070a03f8d8-additional-scripts\") pod \"ovn-controller-2xddw-config-cmjlc\" (UID: \"a1d7994b-dab0-4357-984a-93070a03f8d8\") " pod="openstack/ovn-controller-2xddw-config-cmjlc" Oct 08 19:52:01 crc kubenswrapper[4988]: I1008 19:52:01.079082 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a1d7994b-dab0-4357-984a-93070a03f8d8-scripts\") pod \"ovn-controller-2xddw-config-cmjlc\" (UID: \"a1d7994b-dab0-4357-984a-93070a03f8d8\") " pod="openstack/ovn-controller-2xddw-config-cmjlc" Oct 08 19:52:01 crc kubenswrapper[4988]: I1008 19:52:01.079147 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c144edcb-2dfc-4f6d-85a3-77ae04fd9fb5-combined-ca-bundle\") pod \"octavia-api-5d77847989-vhxvn\" (UID: \"c144edcb-2dfc-4f6d-85a3-77ae04fd9fb5\") " pod="openstack/octavia-api-5d77847989-vhxvn" Oct 08 19:52:01 crc kubenswrapper[4988]: I1008 19:52:01.079190 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/a1d7994b-dab0-4357-984a-93070a03f8d8-var-run\") pod \"ovn-controller-2xddw-config-cmjlc\" (UID: \"a1d7994b-dab0-4357-984a-93070a03f8d8\") " pod="openstack/ovn-controller-2xddw-config-cmjlc" Oct 08 19:52:01 crc kubenswrapper[4988]: I1008 19:52:01.079076 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/a1d7994b-dab0-4357-984a-93070a03f8d8-var-log-ovn\") pod \"ovn-controller-2xddw-config-cmjlc\" (UID: \"a1d7994b-dab0-4357-984a-93070a03f8d8\") " pod="openstack/ovn-controller-2xddw-config-cmjlc" Oct 08 19:52:01 crc kubenswrapper[4988]: I1008 19:52:01.079427 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/a1d7994b-dab0-4357-984a-93070a03f8d8-var-run\") pod \"ovn-controller-2xddw-config-cmjlc\" (UID: \"a1d7994b-dab0-4357-984a-93070a03f8d8\") " pod="openstack/ovn-controller-2xddw-config-cmjlc" Oct 08 19:52:01 crc kubenswrapper[4988]: I1008 19:52:01.079499 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/a1d7994b-dab0-4357-984a-93070a03f8d8-var-run-ovn\") pod \"ovn-controller-2xddw-config-cmjlc\" (UID: \"a1d7994b-dab0-4357-984a-93070a03f8d8\") " pod="openstack/ovn-controller-2xddw-config-cmjlc" Oct 08 19:52:01 crc kubenswrapper[4988]: I1008 19:52:01.079878 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/a1d7994b-dab0-4357-984a-93070a03f8d8-additional-scripts\") pod \"ovn-controller-2xddw-config-cmjlc\" (UID: \"a1d7994b-dab0-4357-984a-93070a03f8d8\") " pod="openstack/ovn-controller-2xddw-config-cmjlc" Oct 08 19:52:01 crc kubenswrapper[4988]: I1008 19:52:01.081374 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a1d7994b-dab0-4357-984a-93070a03f8d8-scripts\") pod \"ovn-controller-2xddw-config-cmjlc\" (UID: \"a1d7994b-dab0-4357-984a-93070a03f8d8\") " pod="openstack/ovn-controller-2xddw-config-cmjlc" Oct 08 19:52:01 crc kubenswrapper[4988]: I1008 19:52:01.126494 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wp7bl\" (UniqueName: \"kubernetes.io/projected/a1d7994b-dab0-4357-984a-93070a03f8d8-kube-api-access-wp7bl\") pod \"ovn-controller-2xddw-config-cmjlc\" (UID: \"a1d7994b-dab0-4357-984a-93070a03f8d8\") " pod="openstack/ovn-controller-2xddw-config-cmjlc" Oct 08 19:52:01 crc kubenswrapper[4988]: I1008 19:52:01.180351 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c144edcb-2dfc-4f6d-85a3-77ae04fd9fb5-combined-ca-bundle\") pod \"octavia-api-5d77847989-vhxvn\" (UID: \"c144edcb-2dfc-4f6d-85a3-77ae04fd9fb5\") " pod="openstack/octavia-api-5d77847989-vhxvn" Oct 08 19:52:01 crc kubenswrapper[4988]: I1008 19:52:01.180488 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/c144edcb-2dfc-4f6d-85a3-77ae04fd9fb5-config-data-merged\") pod \"octavia-api-5d77847989-vhxvn\" (UID: \"c144edcb-2dfc-4f6d-85a3-77ae04fd9fb5\") " pod="openstack/octavia-api-5d77847989-vhxvn" Oct 08 19:52:01 crc kubenswrapper[4988]: I1008 19:52:01.180516 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c144edcb-2dfc-4f6d-85a3-77ae04fd9fb5-scripts\") pod \"octavia-api-5d77847989-vhxvn\" (UID: \"c144edcb-2dfc-4f6d-85a3-77ae04fd9fb5\") " pod="openstack/octavia-api-5d77847989-vhxvn" Oct 08 19:52:01 crc kubenswrapper[4988]: I1008 19:52:01.180548 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"octavia-run\" (UniqueName: \"kubernetes.io/empty-dir/c144edcb-2dfc-4f6d-85a3-77ae04fd9fb5-octavia-run\") pod \"octavia-api-5d77847989-vhxvn\" (UID: \"c144edcb-2dfc-4f6d-85a3-77ae04fd9fb5\") " pod="openstack/octavia-api-5d77847989-vhxvn" Oct 08 19:52:01 crc kubenswrapper[4988]: I1008 19:52:01.180591 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c144edcb-2dfc-4f6d-85a3-77ae04fd9fb5-config-data\") pod \"octavia-api-5d77847989-vhxvn\" (UID: \"c144edcb-2dfc-4f6d-85a3-77ae04fd9fb5\") " pod="openstack/octavia-api-5d77847989-vhxvn" Oct 08 19:52:01 crc kubenswrapper[4988]: I1008 19:52:01.180628 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/c144edcb-2dfc-4f6d-85a3-77ae04fd9fb5-ovndb-tls-certs\") pod \"octavia-api-5d77847989-vhxvn\" (UID: \"c144edcb-2dfc-4f6d-85a3-77ae04fd9fb5\") " pod="openstack/octavia-api-5d77847989-vhxvn" Oct 08 19:52:01 crc kubenswrapper[4988]: I1008 19:52:01.182067 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"octavia-run\" (UniqueName: \"kubernetes.io/empty-dir/c144edcb-2dfc-4f6d-85a3-77ae04fd9fb5-octavia-run\") pod \"octavia-api-5d77847989-vhxvn\" (UID: \"c144edcb-2dfc-4f6d-85a3-77ae04fd9fb5\") " pod="openstack/octavia-api-5d77847989-vhxvn" Oct 08 19:52:01 crc kubenswrapper[4988]: I1008 19:52:01.182846 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/c144edcb-2dfc-4f6d-85a3-77ae04fd9fb5-config-data-merged\") pod \"octavia-api-5d77847989-vhxvn\" (UID: \"c144edcb-2dfc-4f6d-85a3-77ae04fd9fb5\") " pod="openstack/octavia-api-5d77847989-vhxvn" Oct 08 19:52:01 crc kubenswrapper[4988]: I1008 19:52:01.189695 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c144edcb-2dfc-4f6d-85a3-77ae04fd9fb5-scripts\") pod \"octavia-api-5d77847989-vhxvn\" (UID: \"c144edcb-2dfc-4f6d-85a3-77ae04fd9fb5\") " pod="openstack/octavia-api-5d77847989-vhxvn" Oct 08 19:52:01 crc kubenswrapper[4988]: I1008 19:52:01.189747 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c144edcb-2dfc-4f6d-85a3-77ae04fd9fb5-combined-ca-bundle\") pod \"octavia-api-5d77847989-vhxvn\" (UID: \"c144edcb-2dfc-4f6d-85a3-77ae04fd9fb5\") " pod="openstack/octavia-api-5d77847989-vhxvn" Oct 08 19:52:01 crc kubenswrapper[4988]: I1008 19:52:01.190216 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/c144edcb-2dfc-4f6d-85a3-77ae04fd9fb5-ovndb-tls-certs\") pod \"octavia-api-5d77847989-vhxvn\" (UID: \"c144edcb-2dfc-4f6d-85a3-77ae04fd9fb5\") " pod="openstack/octavia-api-5d77847989-vhxvn" Oct 08 19:52:01 crc kubenswrapper[4988]: I1008 19:52:01.190625 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c144edcb-2dfc-4f6d-85a3-77ae04fd9fb5-config-data\") pod \"octavia-api-5d77847989-vhxvn\" (UID: \"c144edcb-2dfc-4f6d-85a3-77ae04fd9fb5\") " pod="openstack/octavia-api-5d77847989-vhxvn" Oct 08 19:52:01 crc kubenswrapper[4988]: I1008 19:52:01.256604 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-2xddw-config-cmjlc" Oct 08 19:52:01 crc kubenswrapper[4988]: I1008 19:52:01.328444 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-api-5d77847989-vhxvn" Oct 08 19:52:01 crc kubenswrapper[4988]: I1008 19:52:01.726041 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-2xddw-config-cmjlc"] Oct 08 19:52:01 crc kubenswrapper[4988]: I1008 19:52:01.841152 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-api-5d77847989-vhxvn"] Oct 08 19:52:01 crc kubenswrapper[4988]: W1008 19:52:01.847650 4988 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc144edcb_2dfc_4f6d_85a3_77ae04fd9fb5.slice/crio-cf1c3bc8156553244fb189779f0c5254dd623357704835e658e0fad467cb9724 WatchSource:0}: Error finding container cf1c3bc8156553244fb189779f0c5254dd623357704835e658e0fad467cb9724: Status 404 returned error can't find the container with id cf1c3bc8156553244fb189779f0c5254dd623357704835e658e0fad467cb9724 Oct 08 19:52:02 crc kubenswrapper[4988]: I1008 19:52:02.492779 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-api-5d77847989-vhxvn" event={"ID":"c144edcb-2dfc-4f6d-85a3-77ae04fd9fb5","Type":"ContainerStarted","Data":"cf1c3bc8156553244fb189779f0c5254dd623357704835e658e0fad467cb9724"} Oct 08 19:52:02 crc kubenswrapper[4988]: I1008 19:52:02.500297 4988 generic.go:334] "Generic (PLEG): container finished" podID="a1d7994b-dab0-4357-984a-93070a03f8d8" containerID="3c4b8767c0a6bc569abb1d4ca4d565a13e4a771dbf0b8eedc8117cf869c70d65" exitCode=0 Oct 08 19:52:02 crc kubenswrapper[4988]: I1008 19:52:02.500355 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-2xddw-config-cmjlc" event={"ID":"a1d7994b-dab0-4357-984a-93070a03f8d8","Type":"ContainerDied","Data":"3c4b8767c0a6bc569abb1d4ca4d565a13e4a771dbf0b8eedc8117cf869c70d65"} Oct 08 19:52:02 crc kubenswrapper[4988]: I1008 19:52:02.500386 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-2xddw-config-cmjlc" event={"ID":"a1d7994b-dab0-4357-984a-93070a03f8d8","Type":"ContainerStarted","Data":"bb5346ddfd41702a6c983e7db4106017c31d98620395210d375c13921b6abb31"} Oct 08 19:52:03 crc kubenswrapper[4988]: I1008 19:52:03.910789 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-2xddw-config-cmjlc" Oct 08 19:52:03 crc kubenswrapper[4988]: I1008 19:52:03.936966 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a1d7994b-dab0-4357-984a-93070a03f8d8-scripts\") pod \"a1d7994b-dab0-4357-984a-93070a03f8d8\" (UID: \"a1d7994b-dab0-4357-984a-93070a03f8d8\") " Oct 08 19:52:03 crc kubenswrapper[4988]: I1008 19:52:03.937017 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/a1d7994b-dab0-4357-984a-93070a03f8d8-var-log-ovn\") pod \"a1d7994b-dab0-4357-984a-93070a03f8d8\" (UID: \"a1d7994b-dab0-4357-984a-93070a03f8d8\") " Oct 08 19:52:03 crc kubenswrapper[4988]: I1008 19:52:03.937124 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a1d7994b-dab0-4357-984a-93070a03f8d8-var-log-ovn" (OuterVolumeSpecName: "var-log-ovn") pod "a1d7994b-dab0-4357-984a-93070a03f8d8" (UID: "a1d7994b-dab0-4357-984a-93070a03f8d8"). InnerVolumeSpecName "var-log-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 08 19:52:03 crc kubenswrapper[4988]: I1008 19:52:03.937249 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/a1d7994b-dab0-4357-984a-93070a03f8d8-var-run-ovn\") pod \"a1d7994b-dab0-4357-984a-93070a03f8d8\" (UID: \"a1d7994b-dab0-4357-984a-93070a03f8d8\") " Oct 08 19:52:03 crc kubenswrapper[4988]: I1008 19:52:03.937302 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/a1d7994b-dab0-4357-984a-93070a03f8d8-var-run\") pod \"a1d7994b-dab0-4357-984a-93070a03f8d8\" (UID: \"a1d7994b-dab0-4357-984a-93070a03f8d8\") " Oct 08 19:52:03 crc kubenswrapper[4988]: I1008 19:52:03.937377 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wp7bl\" (UniqueName: \"kubernetes.io/projected/a1d7994b-dab0-4357-984a-93070a03f8d8-kube-api-access-wp7bl\") pod \"a1d7994b-dab0-4357-984a-93070a03f8d8\" (UID: \"a1d7994b-dab0-4357-984a-93070a03f8d8\") " Oct 08 19:52:03 crc kubenswrapper[4988]: I1008 19:52:03.937540 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/a1d7994b-dab0-4357-984a-93070a03f8d8-additional-scripts\") pod \"a1d7994b-dab0-4357-984a-93070a03f8d8\" (UID: \"a1d7994b-dab0-4357-984a-93070a03f8d8\") " Oct 08 19:52:03 crc kubenswrapper[4988]: I1008 19:52:03.938030 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a1d7994b-dab0-4357-984a-93070a03f8d8-scripts" (OuterVolumeSpecName: "scripts") pod "a1d7994b-dab0-4357-984a-93070a03f8d8" (UID: "a1d7994b-dab0-4357-984a-93070a03f8d8"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 19:52:03 crc kubenswrapper[4988]: I1008 19:52:03.938230 4988 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a1d7994b-dab0-4357-984a-93070a03f8d8-scripts\") on node \"crc\" DevicePath \"\"" Oct 08 19:52:03 crc kubenswrapper[4988]: I1008 19:52:03.938248 4988 reconciler_common.go:293] "Volume detached for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/a1d7994b-dab0-4357-984a-93070a03f8d8-var-log-ovn\") on node \"crc\" DevicePath \"\"" Oct 08 19:52:03 crc kubenswrapper[4988]: I1008 19:52:03.938270 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a1d7994b-dab0-4357-984a-93070a03f8d8-var-run" (OuterVolumeSpecName: "var-run") pod "a1d7994b-dab0-4357-984a-93070a03f8d8" (UID: "a1d7994b-dab0-4357-984a-93070a03f8d8"). InnerVolumeSpecName "var-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 08 19:52:03 crc kubenswrapper[4988]: I1008 19:52:03.938421 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a1d7994b-dab0-4357-984a-93070a03f8d8-var-run-ovn" (OuterVolumeSpecName: "var-run-ovn") pod "a1d7994b-dab0-4357-984a-93070a03f8d8" (UID: "a1d7994b-dab0-4357-984a-93070a03f8d8"). InnerVolumeSpecName "var-run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 08 19:52:03 crc kubenswrapper[4988]: I1008 19:52:03.938516 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a1d7994b-dab0-4357-984a-93070a03f8d8-additional-scripts" (OuterVolumeSpecName: "additional-scripts") pod "a1d7994b-dab0-4357-984a-93070a03f8d8" (UID: "a1d7994b-dab0-4357-984a-93070a03f8d8"). InnerVolumeSpecName "additional-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 19:52:03 crc kubenswrapper[4988]: I1008 19:52:03.946601 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a1d7994b-dab0-4357-984a-93070a03f8d8-kube-api-access-wp7bl" (OuterVolumeSpecName: "kube-api-access-wp7bl") pod "a1d7994b-dab0-4357-984a-93070a03f8d8" (UID: "a1d7994b-dab0-4357-984a-93070a03f8d8"). InnerVolumeSpecName "kube-api-access-wp7bl". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 19:52:04 crc kubenswrapper[4988]: I1008 19:52:04.039556 4988 reconciler_common.go:293] "Volume detached for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/a1d7994b-dab0-4357-984a-93070a03f8d8-var-run-ovn\") on node \"crc\" DevicePath \"\"" Oct 08 19:52:04 crc kubenswrapper[4988]: I1008 19:52:04.039584 4988 reconciler_common.go:293] "Volume detached for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/a1d7994b-dab0-4357-984a-93070a03f8d8-var-run\") on node \"crc\" DevicePath \"\"" Oct 08 19:52:04 crc kubenswrapper[4988]: I1008 19:52:04.039594 4988 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wp7bl\" (UniqueName: \"kubernetes.io/projected/a1d7994b-dab0-4357-984a-93070a03f8d8-kube-api-access-wp7bl\") on node \"crc\" DevicePath \"\"" Oct 08 19:52:04 crc kubenswrapper[4988]: I1008 19:52:04.039604 4988 reconciler_common.go:293] "Volume detached for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/a1d7994b-dab0-4357-984a-93070a03f8d8-additional-scripts\") on node \"crc\" DevicePath \"\"" Oct 08 19:52:04 crc kubenswrapper[4988]: I1008 19:52:04.525609 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-2xddw-config-cmjlc" event={"ID":"a1d7994b-dab0-4357-984a-93070a03f8d8","Type":"ContainerDied","Data":"bb5346ddfd41702a6c983e7db4106017c31d98620395210d375c13921b6abb31"} Oct 08 19:52:04 crc kubenswrapper[4988]: I1008 19:52:04.525843 4988 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="bb5346ddfd41702a6c983e7db4106017c31d98620395210d375c13921b6abb31" Oct 08 19:52:04 crc kubenswrapper[4988]: I1008 19:52:04.525725 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-2xddw-config-cmjlc" Oct 08 19:52:04 crc kubenswrapper[4988]: I1008 19:52:04.820505 4988 scope.go:117] "RemoveContainer" containerID="36fd8e868f35d0f857726dd30cd07286a2b84b178bd372dde963221cb2f09f60" Oct 08 19:52:05 crc kubenswrapper[4988]: I1008 19:52:05.009959 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-2xddw-config-cmjlc"] Oct 08 19:52:05 crc kubenswrapper[4988]: I1008 19:52:05.020075 4988 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-2xddw-config-cmjlc"] Oct 08 19:52:05 crc kubenswrapper[4988]: I1008 19:52:05.250183 4988 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a1d7994b-dab0-4357-984a-93070a03f8d8" path="/var/lib/kubelet/pods/a1d7994b-dab0-4357-984a-93070a03f8d8/volumes" Oct 08 19:52:11 crc kubenswrapper[4988]: I1008 19:52:11.596370 4988 generic.go:334] "Generic (PLEG): container finished" podID="c144edcb-2dfc-4f6d-85a3-77ae04fd9fb5" containerID="2b1a5349c2854b7638f49bce3966bd7c10039d7a5ce4ab4106b399cb7368f25c" exitCode=0 Oct 08 19:52:11 crc kubenswrapper[4988]: I1008 19:52:11.596506 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-api-5d77847989-vhxvn" event={"ID":"c144edcb-2dfc-4f6d-85a3-77ae04fd9fb5","Type":"ContainerDied","Data":"2b1a5349c2854b7638f49bce3966bd7c10039d7a5ce4ab4106b399cb7368f25c"} Oct 08 19:52:12 crc kubenswrapper[4988]: I1008 19:52:12.607779 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-api-5d77847989-vhxvn" event={"ID":"c144edcb-2dfc-4f6d-85a3-77ae04fd9fb5","Type":"ContainerStarted","Data":"571330e5b42c1e9fd691583f34d35e8b73f643c5b7dc4b0f8b4e40e4b973a33e"} Oct 08 19:52:12 crc kubenswrapper[4988]: I1008 19:52:12.608151 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-api-5d77847989-vhxvn" event={"ID":"c144edcb-2dfc-4f6d-85a3-77ae04fd9fb5","Type":"ContainerStarted","Data":"28dbba51a65b539a92846427a3d45c9ac7d4d4f18b74b5d9d539aee4c3142293"} Oct 08 19:52:12 crc kubenswrapper[4988]: I1008 19:52:12.609483 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/octavia-api-5d77847989-vhxvn" Oct 08 19:52:12 crc kubenswrapper[4988]: I1008 19:52:12.609505 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/octavia-api-5d77847989-vhxvn" Oct 08 19:52:12 crc kubenswrapper[4988]: I1008 19:52:12.641406 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/octavia-api-5d77847989-vhxvn" podStartSLOduration=3.967874878 podStartE2EDuration="12.641375639s" podCreationTimestamp="2025-10-08 19:52:00 +0000 UTC" firstStartedPulling="2025-10-08 19:52:01.855436017 +0000 UTC m=+6067.305278787" lastFinishedPulling="2025-10-08 19:52:10.528936768 +0000 UTC m=+6075.978779548" observedRunningTime="2025-10-08 19:52:12.634861151 +0000 UTC m=+6078.084703921" watchObservedRunningTime="2025-10-08 19:52:12.641375639 +0000 UTC m=+6078.091218409" Oct 08 19:52:20 crc kubenswrapper[4988]: I1008 19:52:20.347732 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/octavia-api-5d77847989-vhxvn" Oct 08 19:52:21 crc kubenswrapper[4988]: I1008 19:52:21.362425 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/octavia-rsyslog-xdhtp"] Oct 08 19:52:21 crc kubenswrapper[4988]: E1008 19:52:21.363160 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a1d7994b-dab0-4357-984a-93070a03f8d8" containerName="ovn-config" Oct 08 19:52:21 crc kubenswrapper[4988]: I1008 19:52:21.363177 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="a1d7994b-dab0-4357-984a-93070a03f8d8" containerName="ovn-config" Oct 08 19:52:21 crc kubenswrapper[4988]: I1008 19:52:21.363427 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="a1d7994b-dab0-4357-984a-93070a03f8d8" containerName="ovn-config" Oct 08 19:52:21 crc kubenswrapper[4988]: I1008 19:52:21.364719 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-rsyslog-xdhtp" Oct 08 19:52:21 crc kubenswrapper[4988]: I1008 19:52:21.367406 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"octavia-rsyslog-config-data" Oct 08 19:52:21 crc kubenswrapper[4988]: I1008 19:52:21.367406 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"octavia-rsyslog-scripts" Oct 08 19:52:21 crc kubenswrapper[4988]: I1008 19:52:21.368441 4988 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"octavia-hmport-map" Oct 08 19:52:21 crc kubenswrapper[4988]: I1008 19:52:21.378786 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-rsyslog-xdhtp"] Oct 08 19:52:21 crc kubenswrapper[4988]: I1008 19:52:21.497083 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/2171d012-b738-48cf-a6e7-673d2996c676-config-data-merged\") pod \"octavia-rsyslog-xdhtp\" (UID: \"2171d012-b738-48cf-a6e7-673d2996c676\") " pod="openstack/octavia-rsyslog-xdhtp" Oct 08 19:52:21 crc kubenswrapper[4988]: I1008 19:52:21.497160 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hm-ports\" (UniqueName: \"kubernetes.io/configmap/2171d012-b738-48cf-a6e7-673d2996c676-hm-ports\") pod \"octavia-rsyslog-xdhtp\" (UID: \"2171d012-b738-48cf-a6e7-673d2996c676\") " pod="openstack/octavia-rsyslog-xdhtp" Oct 08 19:52:21 crc kubenswrapper[4988]: I1008 19:52:21.497240 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2171d012-b738-48cf-a6e7-673d2996c676-scripts\") pod \"octavia-rsyslog-xdhtp\" (UID: \"2171d012-b738-48cf-a6e7-673d2996c676\") " pod="openstack/octavia-rsyslog-xdhtp" Oct 08 19:52:21 crc kubenswrapper[4988]: I1008 19:52:21.497330 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2171d012-b738-48cf-a6e7-673d2996c676-config-data\") pod \"octavia-rsyslog-xdhtp\" (UID: \"2171d012-b738-48cf-a6e7-673d2996c676\") " pod="openstack/octavia-rsyslog-xdhtp" Oct 08 19:52:21 crc kubenswrapper[4988]: I1008 19:52:21.599174 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2171d012-b738-48cf-a6e7-673d2996c676-scripts\") pod \"octavia-rsyslog-xdhtp\" (UID: \"2171d012-b738-48cf-a6e7-673d2996c676\") " pod="openstack/octavia-rsyslog-xdhtp" Oct 08 19:52:21 crc kubenswrapper[4988]: I1008 19:52:21.599261 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2171d012-b738-48cf-a6e7-673d2996c676-config-data\") pod \"octavia-rsyslog-xdhtp\" (UID: \"2171d012-b738-48cf-a6e7-673d2996c676\") " pod="openstack/octavia-rsyslog-xdhtp" Oct 08 19:52:21 crc kubenswrapper[4988]: I1008 19:52:21.599304 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/2171d012-b738-48cf-a6e7-673d2996c676-config-data-merged\") pod \"octavia-rsyslog-xdhtp\" (UID: \"2171d012-b738-48cf-a6e7-673d2996c676\") " pod="openstack/octavia-rsyslog-xdhtp" Oct 08 19:52:21 crc kubenswrapper[4988]: I1008 19:52:21.599381 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hm-ports\" (UniqueName: \"kubernetes.io/configmap/2171d012-b738-48cf-a6e7-673d2996c676-hm-ports\") pod \"octavia-rsyslog-xdhtp\" (UID: \"2171d012-b738-48cf-a6e7-673d2996c676\") " pod="openstack/octavia-rsyslog-xdhtp" Oct 08 19:52:21 crc kubenswrapper[4988]: I1008 19:52:21.599971 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/2171d012-b738-48cf-a6e7-673d2996c676-config-data-merged\") pod \"octavia-rsyslog-xdhtp\" (UID: \"2171d012-b738-48cf-a6e7-673d2996c676\") " pod="openstack/octavia-rsyslog-xdhtp" Oct 08 19:52:21 crc kubenswrapper[4988]: I1008 19:52:21.600371 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hm-ports\" (UniqueName: \"kubernetes.io/configmap/2171d012-b738-48cf-a6e7-673d2996c676-hm-ports\") pod \"octavia-rsyslog-xdhtp\" (UID: \"2171d012-b738-48cf-a6e7-673d2996c676\") " pod="openstack/octavia-rsyslog-xdhtp" Oct 08 19:52:21 crc kubenswrapper[4988]: I1008 19:52:21.604534 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2171d012-b738-48cf-a6e7-673d2996c676-config-data\") pod \"octavia-rsyslog-xdhtp\" (UID: \"2171d012-b738-48cf-a6e7-673d2996c676\") " pod="openstack/octavia-rsyslog-xdhtp" Oct 08 19:52:21 crc kubenswrapper[4988]: I1008 19:52:21.606829 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2171d012-b738-48cf-a6e7-673d2996c676-scripts\") pod \"octavia-rsyslog-xdhtp\" (UID: \"2171d012-b738-48cf-a6e7-673d2996c676\") " pod="openstack/octavia-rsyslog-xdhtp" Oct 08 19:52:21 crc kubenswrapper[4988]: I1008 19:52:21.726129 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-rsyslog-xdhtp" Oct 08 19:52:22 crc kubenswrapper[4988]: I1008 19:52:22.102210 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/octavia-image-upload-678599687f-xxfmc"] Oct 08 19:52:22 crc kubenswrapper[4988]: I1008 19:52:22.106058 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-image-upload-678599687f-xxfmc" Oct 08 19:52:22 crc kubenswrapper[4988]: I1008 19:52:22.110011 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"octavia-config-data" Oct 08 19:52:22 crc kubenswrapper[4988]: I1008 19:52:22.114022 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-image-upload-678599687f-xxfmc"] Oct 08 19:52:22 crc kubenswrapper[4988]: I1008 19:52:22.212573 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"amphora-image\" (UniqueName: \"kubernetes.io/empty-dir/43ab7316-2b23-4b0c-9146-0d87054cc1e9-amphora-image\") pod \"octavia-image-upload-678599687f-xxfmc\" (UID: \"43ab7316-2b23-4b0c-9146-0d87054cc1e9\") " pod="openstack/octavia-image-upload-678599687f-xxfmc" Oct 08 19:52:22 crc kubenswrapper[4988]: I1008 19:52:22.212765 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/43ab7316-2b23-4b0c-9146-0d87054cc1e9-httpd-config\") pod \"octavia-image-upload-678599687f-xxfmc\" (UID: \"43ab7316-2b23-4b0c-9146-0d87054cc1e9\") " pod="openstack/octavia-image-upload-678599687f-xxfmc" Oct 08 19:52:22 crc kubenswrapper[4988]: I1008 19:52:22.291478 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-rsyslog-xdhtp"] Oct 08 19:52:22 crc kubenswrapper[4988]: I1008 19:52:22.315289 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/43ab7316-2b23-4b0c-9146-0d87054cc1e9-httpd-config\") pod \"octavia-image-upload-678599687f-xxfmc\" (UID: \"43ab7316-2b23-4b0c-9146-0d87054cc1e9\") " pod="openstack/octavia-image-upload-678599687f-xxfmc" Oct 08 19:52:22 crc kubenswrapper[4988]: I1008 19:52:22.315689 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"amphora-image\" (UniqueName: \"kubernetes.io/empty-dir/43ab7316-2b23-4b0c-9146-0d87054cc1e9-amphora-image\") pod \"octavia-image-upload-678599687f-xxfmc\" (UID: \"43ab7316-2b23-4b0c-9146-0d87054cc1e9\") " pod="openstack/octavia-image-upload-678599687f-xxfmc" Oct 08 19:52:22 crc kubenswrapper[4988]: I1008 19:52:22.316828 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"amphora-image\" (UniqueName: \"kubernetes.io/empty-dir/43ab7316-2b23-4b0c-9146-0d87054cc1e9-amphora-image\") pod \"octavia-image-upload-678599687f-xxfmc\" (UID: \"43ab7316-2b23-4b0c-9146-0d87054cc1e9\") " pod="openstack/octavia-image-upload-678599687f-xxfmc" Oct 08 19:52:22 crc kubenswrapper[4988]: I1008 19:52:22.323702 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/43ab7316-2b23-4b0c-9146-0d87054cc1e9-httpd-config\") pod \"octavia-image-upload-678599687f-xxfmc\" (UID: \"43ab7316-2b23-4b0c-9146-0d87054cc1e9\") " pod="openstack/octavia-image-upload-678599687f-xxfmc" Oct 08 19:52:22 crc kubenswrapper[4988]: I1008 19:52:22.439995 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-image-upload-678599687f-xxfmc" Oct 08 19:52:22 crc kubenswrapper[4988]: I1008 19:52:22.729615 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-rsyslog-xdhtp" event={"ID":"2171d012-b738-48cf-a6e7-673d2996c676","Type":"ContainerStarted","Data":"d289a824b92803e3d81342822df5d9e225d5f11f9a69eebdbdb612f5c04e318f"} Oct 08 19:52:22 crc kubenswrapper[4988]: I1008 19:52:22.849391 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/octavia-db-sync-r8g9v"] Oct 08 19:52:22 crc kubenswrapper[4988]: I1008 19:52:22.851337 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-db-sync-r8g9v" Oct 08 19:52:22 crc kubenswrapper[4988]: I1008 19:52:22.853206 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"octavia-scripts" Oct 08 19:52:22 crc kubenswrapper[4988]: I1008 19:52:22.891070 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-db-sync-r8g9v"] Oct 08 19:52:22 crc kubenswrapper[4988]: I1008 19:52:22.919386 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-image-upload-678599687f-xxfmc"] Oct 08 19:52:22 crc kubenswrapper[4988]: W1008 19:52:22.923318 4988 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod43ab7316_2b23_4b0c_9146_0d87054cc1e9.slice/crio-5bd2dd3f735acd07283f5231a7dc7019749b6b75a2cc8fa76cba82ccf691ec91 WatchSource:0}: Error finding container 5bd2dd3f735acd07283f5231a7dc7019749b6b75a2cc8fa76cba82ccf691ec91: Status 404 returned error can't find the container with id 5bd2dd3f735acd07283f5231a7dc7019749b6b75a2cc8fa76cba82ccf691ec91 Oct 08 19:52:22 crc kubenswrapper[4988]: I1008 19:52:22.929497 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f75923d6-45dc-42fa-9f2b-2a64fb7490b1-scripts\") pod \"octavia-db-sync-r8g9v\" (UID: \"f75923d6-45dc-42fa-9f2b-2a64fb7490b1\") " pod="openstack/octavia-db-sync-r8g9v" Oct 08 19:52:22 crc kubenswrapper[4988]: I1008 19:52:22.929532 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f75923d6-45dc-42fa-9f2b-2a64fb7490b1-config-data\") pod \"octavia-db-sync-r8g9v\" (UID: \"f75923d6-45dc-42fa-9f2b-2a64fb7490b1\") " pod="openstack/octavia-db-sync-r8g9v" Oct 08 19:52:22 crc kubenswrapper[4988]: I1008 19:52:22.929553 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/f75923d6-45dc-42fa-9f2b-2a64fb7490b1-config-data-merged\") pod \"octavia-db-sync-r8g9v\" (UID: \"f75923d6-45dc-42fa-9f2b-2a64fb7490b1\") " pod="openstack/octavia-db-sync-r8g9v" Oct 08 19:52:22 crc kubenswrapper[4988]: I1008 19:52:22.929905 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f75923d6-45dc-42fa-9f2b-2a64fb7490b1-combined-ca-bundle\") pod \"octavia-db-sync-r8g9v\" (UID: \"f75923d6-45dc-42fa-9f2b-2a64fb7490b1\") " pod="openstack/octavia-db-sync-r8g9v" Oct 08 19:52:23 crc kubenswrapper[4988]: I1008 19:52:23.032962 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f75923d6-45dc-42fa-9f2b-2a64fb7490b1-combined-ca-bundle\") pod \"octavia-db-sync-r8g9v\" (UID: \"f75923d6-45dc-42fa-9f2b-2a64fb7490b1\") " pod="openstack/octavia-db-sync-r8g9v" Oct 08 19:52:23 crc kubenswrapper[4988]: I1008 19:52:23.033576 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f75923d6-45dc-42fa-9f2b-2a64fb7490b1-scripts\") pod \"octavia-db-sync-r8g9v\" (UID: \"f75923d6-45dc-42fa-9f2b-2a64fb7490b1\") " pod="openstack/octavia-db-sync-r8g9v" Oct 08 19:52:23 crc kubenswrapper[4988]: I1008 19:52:23.033640 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f75923d6-45dc-42fa-9f2b-2a64fb7490b1-config-data\") pod \"octavia-db-sync-r8g9v\" (UID: \"f75923d6-45dc-42fa-9f2b-2a64fb7490b1\") " pod="openstack/octavia-db-sync-r8g9v" Oct 08 19:52:23 crc kubenswrapper[4988]: I1008 19:52:23.033859 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/f75923d6-45dc-42fa-9f2b-2a64fb7490b1-config-data-merged\") pod \"octavia-db-sync-r8g9v\" (UID: \"f75923d6-45dc-42fa-9f2b-2a64fb7490b1\") " pod="openstack/octavia-db-sync-r8g9v" Oct 08 19:52:23 crc kubenswrapper[4988]: I1008 19:52:23.034959 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/f75923d6-45dc-42fa-9f2b-2a64fb7490b1-config-data-merged\") pod \"octavia-db-sync-r8g9v\" (UID: \"f75923d6-45dc-42fa-9f2b-2a64fb7490b1\") " pod="openstack/octavia-db-sync-r8g9v" Oct 08 19:52:23 crc kubenswrapper[4988]: I1008 19:52:23.040535 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f75923d6-45dc-42fa-9f2b-2a64fb7490b1-combined-ca-bundle\") pod \"octavia-db-sync-r8g9v\" (UID: \"f75923d6-45dc-42fa-9f2b-2a64fb7490b1\") " pod="openstack/octavia-db-sync-r8g9v" Oct 08 19:52:23 crc kubenswrapper[4988]: I1008 19:52:23.045119 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f75923d6-45dc-42fa-9f2b-2a64fb7490b1-config-data\") pod \"octavia-db-sync-r8g9v\" (UID: \"f75923d6-45dc-42fa-9f2b-2a64fb7490b1\") " pod="openstack/octavia-db-sync-r8g9v" Oct 08 19:52:23 crc kubenswrapper[4988]: I1008 19:52:23.048344 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f75923d6-45dc-42fa-9f2b-2a64fb7490b1-scripts\") pod \"octavia-db-sync-r8g9v\" (UID: \"f75923d6-45dc-42fa-9f2b-2a64fb7490b1\") " pod="openstack/octavia-db-sync-r8g9v" Oct 08 19:52:23 crc kubenswrapper[4988]: I1008 19:52:23.185242 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-db-sync-r8g9v" Oct 08 19:52:23 crc kubenswrapper[4988]: I1008 19:52:23.742489 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-image-upload-678599687f-xxfmc" event={"ID":"43ab7316-2b23-4b0c-9146-0d87054cc1e9","Type":"ContainerStarted","Data":"5bd2dd3f735acd07283f5231a7dc7019749b6b75a2cc8fa76cba82ccf691ec91"} Oct 08 19:52:23 crc kubenswrapper[4988]: I1008 19:52:23.956448 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-db-sync-r8g9v"] Oct 08 19:52:24 crc kubenswrapper[4988]: I1008 19:52:24.382108 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/octavia-api-5d77847989-vhxvn" Oct 08 19:52:24 crc kubenswrapper[4988]: I1008 19:52:24.772464 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-rsyslog-xdhtp" event={"ID":"2171d012-b738-48cf-a6e7-673d2996c676","Type":"ContainerStarted","Data":"f94a75bf155acd6f5bc84e8f1281a699929d4fc32473b5ca5159faeefecbf2dc"} Oct 08 19:52:24 crc kubenswrapper[4988]: I1008 19:52:24.780209 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-db-sync-r8g9v" event={"ID":"f75923d6-45dc-42fa-9f2b-2a64fb7490b1","Type":"ContainerStarted","Data":"b2fec19bc320ef3a8b59ec6771f5db2c19ab3770f6a63c2166ed9b75fd26760a"} Oct 08 19:52:25 crc kubenswrapper[4988]: I1008 19:52:25.793482 4988 generic.go:334] "Generic (PLEG): container finished" podID="f75923d6-45dc-42fa-9f2b-2a64fb7490b1" containerID="0eb81c8aa96066cadf521105121ac612d65857e5ef8c6e6fe7e132225a6a4e9f" exitCode=0 Oct 08 19:52:25 crc kubenswrapper[4988]: I1008 19:52:25.793660 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-db-sync-r8g9v" event={"ID":"f75923d6-45dc-42fa-9f2b-2a64fb7490b1","Type":"ContainerDied","Data":"0eb81c8aa96066cadf521105121ac612d65857e5ef8c6e6fe7e132225a6a4e9f"} Oct 08 19:52:26 crc kubenswrapper[4988]: I1008 19:52:26.807670 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-db-sync-r8g9v" event={"ID":"f75923d6-45dc-42fa-9f2b-2a64fb7490b1","Type":"ContainerStarted","Data":"fd73f70995f5184740422e93c2520de613e32ba386667d673b23f5aaf4dba04b"} Oct 08 19:52:26 crc kubenswrapper[4988]: I1008 19:52:26.810084 4988 generic.go:334] "Generic (PLEG): container finished" podID="2171d012-b738-48cf-a6e7-673d2996c676" containerID="f94a75bf155acd6f5bc84e8f1281a699929d4fc32473b5ca5159faeefecbf2dc" exitCode=0 Oct 08 19:52:26 crc kubenswrapper[4988]: I1008 19:52:26.810158 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-rsyslog-xdhtp" event={"ID":"2171d012-b738-48cf-a6e7-673d2996c676","Type":"ContainerDied","Data":"f94a75bf155acd6f5bc84e8f1281a699929d4fc32473b5ca5159faeefecbf2dc"} Oct 08 19:52:26 crc kubenswrapper[4988]: I1008 19:52:26.822684 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/octavia-db-sync-r8g9v" podStartSLOduration=4.8226654920000005 podStartE2EDuration="4.822665492s" podCreationTimestamp="2025-10-08 19:52:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 19:52:26.820904816 +0000 UTC m=+6092.270747596" watchObservedRunningTime="2025-10-08 19:52:26.822665492 +0000 UTC m=+6092.272508262" Oct 08 19:52:29 crc kubenswrapper[4988]: I1008 19:52:29.852325 4988 generic.go:334] "Generic (PLEG): container finished" podID="f75923d6-45dc-42fa-9f2b-2a64fb7490b1" containerID="fd73f70995f5184740422e93c2520de613e32ba386667d673b23f5aaf4dba04b" exitCode=0 Oct 08 19:52:29 crc kubenswrapper[4988]: I1008 19:52:29.852443 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-db-sync-r8g9v" event={"ID":"f75923d6-45dc-42fa-9f2b-2a64fb7490b1","Type":"ContainerDied","Data":"fd73f70995f5184740422e93c2520de613e32ba386667d673b23f5aaf4dba04b"} Oct 08 19:52:32 crc kubenswrapper[4988]: I1008 19:52:32.885741 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-db-sync-r8g9v" event={"ID":"f75923d6-45dc-42fa-9f2b-2a64fb7490b1","Type":"ContainerDied","Data":"b2fec19bc320ef3a8b59ec6771f5db2c19ab3770f6a63c2166ed9b75fd26760a"} Oct 08 19:52:32 crc kubenswrapper[4988]: I1008 19:52:32.886461 4988 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b2fec19bc320ef3a8b59ec6771f5db2c19ab3770f6a63c2166ed9b75fd26760a" Oct 08 19:52:32 crc kubenswrapper[4988]: I1008 19:52:32.900561 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-db-sync-r8g9v" Oct 08 19:52:33 crc kubenswrapper[4988]: I1008 19:52:33.053153 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f75923d6-45dc-42fa-9f2b-2a64fb7490b1-config-data\") pod \"f75923d6-45dc-42fa-9f2b-2a64fb7490b1\" (UID: \"f75923d6-45dc-42fa-9f2b-2a64fb7490b1\") " Oct 08 19:52:33 crc kubenswrapper[4988]: I1008 19:52:33.053234 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f75923d6-45dc-42fa-9f2b-2a64fb7490b1-scripts\") pod \"f75923d6-45dc-42fa-9f2b-2a64fb7490b1\" (UID: \"f75923d6-45dc-42fa-9f2b-2a64fb7490b1\") " Oct 08 19:52:33 crc kubenswrapper[4988]: I1008 19:52:33.053349 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f75923d6-45dc-42fa-9f2b-2a64fb7490b1-combined-ca-bundle\") pod \"f75923d6-45dc-42fa-9f2b-2a64fb7490b1\" (UID: \"f75923d6-45dc-42fa-9f2b-2a64fb7490b1\") " Oct 08 19:52:33 crc kubenswrapper[4988]: I1008 19:52:33.053458 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/f75923d6-45dc-42fa-9f2b-2a64fb7490b1-config-data-merged\") pod \"f75923d6-45dc-42fa-9f2b-2a64fb7490b1\" (UID: \"f75923d6-45dc-42fa-9f2b-2a64fb7490b1\") " Oct 08 19:52:33 crc kubenswrapper[4988]: I1008 19:52:33.065639 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f75923d6-45dc-42fa-9f2b-2a64fb7490b1-config-data" (OuterVolumeSpecName: "config-data") pod "f75923d6-45dc-42fa-9f2b-2a64fb7490b1" (UID: "f75923d6-45dc-42fa-9f2b-2a64fb7490b1"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 19:52:33 crc kubenswrapper[4988]: I1008 19:52:33.073845 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f75923d6-45dc-42fa-9f2b-2a64fb7490b1-scripts" (OuterVolumeSpecName: "scripts") pod "f75923d6-45dc-42fa-9f2b-2a64fb7490b1" (UID: "f75923d6-45dc-42fa-9f2b-2a64fb7490b1"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 19:52:33 crc kubenswrapper[4988]: I1008 19:52:33.084645 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f75923d6-45dc-42fa-9f2b-2a64fb7490b1-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f75923d6-45dc-42fa-9f2b-2a64fb7490b1" (UID: "f75923d6-45dc-42fa-9f2b-2a64fb7490b1"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 19:52:33 crc kubenswrapper[4988]: I1008 19:52:33.092635 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f75923d6-45dc-42fa-9f2b-2a64fb7490b1-config-data-merged" (OuterVolumeSpecName: "config-data-merged") pod "f75923d6-45dc-42fa-9f2b-2a64fb7490b1" (UID: "f75923d6-45dc-42fa-9f2b-2a64fb7490b1"). InnerVolumeSpecName "config-data-merged". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 19:52:33 crc kubenswrapper[4988]: I1008 19:52:33.156070 4988 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f75923d6-45dc-42fa-9f2b-2a64fb7490b1-config-data\") on node \"crc\" DevicePath \"\"" Oct 08 19:52:33 crc kubenswrapper[4988]: I1008 19:52:33.156112 4988 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f75923d6-45dc-42fa-9f2b-2a64fb7490b1-scripts\") on node \"crc\" DevicePath \"\"" Oct 08 19:52:33 crc kubenswrapper[4988]: I1008 19:52:33.156124 4988 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f75923d6-45dc-42fa-9f2b-2a64fb7490b1-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 19:52:33 crc kubenswrapper[4988]: I1008 19:52:33.156164 4988 reconciler_common.go:293] "Volume detached for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/f75923d6-45dc-42fa-9f2b-2a64fb7490b1-config-data-merged\") on node \"crc\" DevicePath \"\"" Oct 08 19:52:33 crc kubenswrapper[4988]: I1008 19:52:33.903996 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-rsyslog-xdhtp" event={"ID":"2171d012-b738-48cf-a6e7-673d2996c676","Type":"ContainerStarted","Data":"8d8ca76d73313a24f2921e12ba93695553eb3f6f302c504a7c677a274d747159"} Oct 08 19:52:33 crc kubenswrapper[4988]: I1008 19:52:33.904748 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/octavia-rsyslog-xdhtp" Oct 08 19:52:33 crc kubenswrapper[4988]: I1008 19:52:33.909757 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-db-sync-r8g9v" Oct 08 19:52:33 crc kubenswrapper[4988]: I1008 19:52:33.909779 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-image-upload-678599687f-xxfmc" event={"ID":"43ab7316-2b23-4b0c-9146-0d87054cc1e9","Type":"ContainerStarted","Data":"bb8901ca13542d2d72987c27ec92ba393cc266bf8aa8ca904b9ac77634514a88"} Oct 08 19:52:33 crc kubenswrapper[4988]: I1008 19:52:33.941164 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/octavia-rsyslog-xdhtp" podStartSLOduration=1.864120748 podStartE2EDuration="12.941140563s" podCreationTimestamp="2025-10-08 19:52:21 +0000 UTC" firstStartedPulling="2025-10-08 19:52:22.292950262 +0000 UTC m=+6087.742793032" lastFinishedPulling="2025-10-08 19:52:33.369970067 +0000 UTC m=+6098.819812847" observedRunningTime="2025-10-08 19:52:33.929599596 +0000 UTC m=+6099.379442376" watchObservedRunningTime="2025-10-08 19:52:33.941140563 +0000 UTC m=+6099.390983343" Oct 08 19:52:34 crc kubenswrapper[4988]: I1008 19:52:34.971142 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/octavia-api-764db59bb9-p78lt"] Oct 08 19:52:34 crc kubenswrapper[4988]: E1008 19:52:34.972003 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f75923d6-45dc-42fa-9f2b-2a64fb7490b1" containerName="init" Oct 08 19:52:34 crc kubenswrapper[4988]: I1008 19:52:34.972020 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="f75923d6-45dc-42fa-9f2b-2a64fb7490b1" containerName="init" Oct 08 19:52:34 crc kubenswrapper[4988]: E1008 19:52:34.972049 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f75923d6-45dc-42fa-9f2b-2a64fb7490b1" containerName="octavia-db-sync" Oct 08 19:52:34 crc kubenswrapper[4988]: I1008 19:52:34.972059 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="f75923d6-45dc-42fa-9f2b-2a64fb7490b1" containerName="octavia-db-sync" Oct 08 19:52:34 crc kubenswrapper[4988]: I1008 19:52:34.972300 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="f75923d6-45dc-42fa-9f2b-2a64fb7490b1" containerName="octavia-db-sync" Oct 08 19:52:34 crc kubenswrapper[4988]: I1008 19:52:34.975367 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-api-764db59bb9-p78lt" Oct 08 19:52:34 crc kubenswrapper[4988]: I1008 19:52:34.977983 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-octavia-internal-svc" Oct 08 19:52:34 crc kubenswrapper[4988]: I1008 19:52:34.978013 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-octavia-public-svc" Oct 08 19:52:34 crc kubenswrapper[4988]: I1008 19:52:34.991800 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-api-764db59bb9-p78lt"] Oct 08 19:52:35 crc kubenswrapper[4988]: E1008 19:52:35.084949 4988 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod43ab7316_2b23_4b0c_9146_0d87054cc1e9.slice/crio-bb8901ca13542d2d72987c27ec92ba393cc266bf8aa8ca904b9ac77634514a88.scope\": RecentStats: unable to find data in memory cache]" Oct 08 19:52:35 crc kubenswrapper[4988]: I1008 19:52:35.131027 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a0c334ad-5ffa-41d9-8cdd-c5c50372137d-scripts\") pod \"octavia-api-764db59bb9-p78lt\" (UID: \"a0c334ad-5ffa-41d9-8cdd-c5c50372137d\") " pod="openstack/octavia-api-764db59bb9-p78lt" Oct 08 19:52:35 crc kubenswrapper[4988]: I1008 19:52:35.131113 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"octavia-run\" (UniqueName: \"kubernetes.io/empty-dir/a0c334ad-5ffa-41d9-8cdd-c5c50372137d-octavia-run\") pod \"octavia-api-764db59bb9-p78lt\" (UID: \"a0c334ad-5ffa-41d9-8cdd-c5c50372137d\") " pod="openstack/octavia-api-764db59bb9-p78lt" Oct 08 19:52:35 crc kubenswrapper[4988]: I1008 19:52:35.131266 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/a0c334ad-5ffa-41d9-8cdd-c5c50372137d-internal-tls-certs\") pod \"octavia-api-764db59bb9-p78lt\" (UID: \"a0c334ad-5ffa-41d9-8cdd-c5c50372137d\") " pod="openstack/octavia-api-764db59bb9-p78lt" Oct 08 19:52:35 crc kubenswrapper[4988]: I1008 19:52:35.131445 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/a0c334ad-5ffa-41d9-8cdd-c5c50372137d-ovndb-tls-certs\") pod \"octavia-api-764db59bb9-p78lt\" (UID: \"a0c334ad-5ffa-41d9-8cdd-c5c50372137d\") " pod="openstack/octavia-api-764db59bb9-p78lt" Oct 08 19:52:35 crc kubenswrapper[4988]: I1008 19:52:35.131578 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/a0c334ad-5ffa-41d9-8cdd-c5c50372137d-public-tls-certs\") pod \"octavia-api-764db59bb9-p78lt\" (UID: \"a0c334ad-5ffa-41d9-8cdd-c5c50372137d\") " pod="openstack/octavia-api-764db59bb9-p78lt" Oct 08 19:52:35 crc kubenswrapper[4988]: I1008 19:52:35.131626 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/a0c334ad-5ffa-41d9-8cdd-c5c50372137d-config-data-merged\") pod \"octavia-api-764db59bb9-p78lt\" (UID: \"a0c334ad-5ffa-41d9-8cdd-c5c50372137d\") " pod="openstack/octavia-api-764db59bb9-p78lt" Oct 08 19:52:35 crc kubenswrapper[4988]: I1008 19:52:35.131668 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a0c334ad-5ffa-41d9-8cdd-c5c50372137d-combined-ca-bundle\") pod \"octavia-api-764db59bb9-p78lt\" (UID: \"a0c334ad-5ffa-41d9-8cdd-c5c50372137d\") " pod="openstack/octavia-api-764db59bb9-p78lt" Oct 08 19:52:35 crc kubenswrapper[4988]: I1008 19:52:35.131798 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a0c334ad-5ffa-41d9-8cdd-c5c50372137d-config-data\") pod \"octavia-api-764db59bb9-p78lt\" (UID: \"a0c334ad-5ffa-41d9-8cdd-c5c50372137d\") " pod="openstack/octavia-api-764db59bb9-p78lt" Oct 08 19:52:35 crc kubenswrapper[4988]: I1008 19:52:35.234251 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/a0c334ad-5ffa-41d9-8cdd-c5c50372137d-ovndb-tls-certs\") pod \"octavia-api-764db59bb9-p78lt\" (UID: \"a0c334ad-5ffa-41d9-8cdd-c5c50372137d\") " pod="openstack/octavia-api-764db59bb9-p78lt" Oct 08 19:52:35 crc kubenswrapper[4988]: I1008 19:52:35.235685 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/a0c334ad-5ffa-41d9-8cdd-c5c50372137d-public-tls-certs\") pod \"octavia-api-764db59bb9-p78lt\" (UID: \"a0c334ad-5ffa-41d9-8cdd-c5c50372137d\") " pod="openstack/octavia-api-764db59bb9-p78lt" Oct 08 19:52:35 crc kubenswrapper[4988]: I1008 19:52:35.235739 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/a0c334ad-5ffa-41d9-8cdd-c5c50372137d-config-data-merged\") pod \"octavia-api-764db59bb9-p78lt\" (UID: \"a0c334ad-5ffa-41d9-8cdd-c5c50372137d\") " pod="openstack/octavia-api-764db59bb9-p78lt" Oct 08 19:52:35 crc kubenswrapper[4988]: I1008 19:52:35.235787 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a0c334ad-5ffa-41d9-8cdd-c5c50372137d-combined-ca-bundle\") pod \"octavia-api-764db59bb9-p78lt\" (UID: \"a0c334ad-5ffa-41d9-8cdd-c5c50372137d\") " pod="openstack/octavia-api-764db59bb9-p78lt" Oct 08 19:52:35 crc kubenswrapper[4988]: I1008 19:52:35.235882 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a0c334ad-5ffa-41d9-8cdd-c5c50372137d-config-data\") pod \"octavia-api-764db59bb9-p78lt\" (UID: \"a0c334ad-5ffa-41d9-8cdd-c5c50372137d\") " pod="openstack/octavia-api-764db59bb9-p78lt" Oct 08 19:52:35 crc kubenswrapper[4988]: I1008 19:52:35.235931 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a0c334ad-5ffa-41d9-8cdd-c5c50372137d-scripts\") pod \"octavia-api-764db59bb9-p78lt\" (UID: \"a0c334ad-5ffa-41d9-8cdd-c5c50372137d\") " pod="openstack/octavia-api-764db59bb9-p78lt" Oct 08 19:52:35 crc kubenswrapper[4988]: I1008 19:52:35.236003 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"octavia-run\" (UniqueName: \"kubernetes.io/empty-dir/a0c334ad-5ffa-41d9-8cdd-c5c50372137d-octavia-run\") pod \"octavia-api-764db59bb9-p78lt\" (UID: \"a0c334ad-5ffa-41d9-8cdd-c5c50372137d\") " pod="openstack/octavia-api-764db59bb9-p78lt" Oct 08 19:52:35 crc kubenswrapper[4988]: I1008 19:52:35.236076 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/a0c334ad-5ffa-41d9-8cdd-c5c50372137d-internal-tls-certs\") pod \"octavia-api-764db59bb9-p78lt\" (UID: \"a0c334ad-5ffa-41d9-8cdd-c5c50372137d\") " pod="openstack/octavia-api-764db59bb9-p78lt" Oct 08 19:52:35 crc kubenswrapper[4988]: I1008 19:52:35.236215 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/a0c334ad-5ffa-41d9-8cdd-c5c50372137d-config-data-merged\") pod \"octavia-api-764db59bb9-p78lt\" (UID: \"a0c334ad-5ffa-41d9-8cdd-c5c50372137d\") " pod="openstack/octavia-api-764db59bb9-p78lt" Oct 08 19:52:35 crc kubenswrapper[4988]: I1008 19:52:35.236480 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"octavia-run\" (UniqueName: \"kubernetes.io/empty-dir/a0c334ad-5ffa-41d9-8cdd-c5c50372137d-octavia-run\") pod \"octavia-api-764db59bb9-p78lt\" (UID: \"a0c334ad-5ffa-41d9-8cdd-c5c50372137d\") " pod="openstack/octavia-api-764db59bb9-p78lt" Oct 08 19:52:35 crc kubenswrapper[4988]: I1008 19:52:35.244013 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a0c334ad-5ffa-41d9-8cdd-c5c50372137d-combined-ca-bundle\") pod \"octavia-api-764db59bb9-p78lt\" (UID: \"a0c334ad-5ffa-41d9-8cdd-c5c50372137d\") " pod="openstack/octavia-api-764db59bb9-p78lt" Oct 08 19:52:35 crc kubenswrapper[4988]: I1008 19:52:35.244240 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/a0c334ad-5ffa-41d9-8cdd-c5c50372137d-internal-tls-certs\") pod \"octavia-api-764db59bb9-p78lt\" (UID: \"a0c334ad-5ffa-41d9-8cdd-c5c50372137d\") " pod="openstack/octavia-api-764db59bb9-p78lt" Oct 08 19:52:35 crc kubenswrapper[4988]: I1008 19:52:35.244285 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/a0c334ad-5ffa-41d9-8cdd-c5c50372137d-ovndb-tls-certs\") pod \"octavia-api-764db59bb9-p78lt\" (UID: \"a0c334ad-5ffa-41d9-8cdd-c5c50372137d\") " pod="openstack/octavia-api-764db59bb9-p78lt" Oct 08 19:52:35 crc kubenswrapper[4988]: I1008 19:52:35.245221 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a0c334ad-5ffa-41d9-8cdd-c5c50372137d-config-data\") pod \"octavia-api-764db59bb9-p78lt\" (UID: \"a0c334ad-5ffa-41d9-8cdd-c5c50372137d\") " pod="openstack/octavia-api-764db59bb9-p78lt" Oct 08 19:52:35 crc kubenswrapper[4988]: I1008 19:52:35.245380 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/a0c334ad-5ffa-41d9-8cdd-c5c50372137d-public-tls-certs\") pod \"octavia-api-764db59bb9-p78lt\" (UID: \"a0c334ad-5ffa-41d9-8cdd-c5c50372137d\") " pod="openstack/octavia-api-764db59bb9-p78lt" Oct 08 19:52:35 crc kubenswrapper[4988]: I1008 19:52:35.246973 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a0c334ad-5ffa-41d9-8cdd-c5c50372137d-scripts\") pod \"octavia-api-764db59bb9-p78lt\" (UID: \"a0c334ad-5ffa-41d9-8cdd-c5c50372137d\") " pod="openstack/octavia-api-764db59bb9-p78lt" Oct 08 19:52:35 crc kubenswrapper[4988]: I1008 19:52:35.313294 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-api-764db59bb9-p78lt" Oct 08 19:52:35 crc kubenswrapper[4988]: W1008 19:52:35.907292 4988 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda0c334ad_5ffa_41d9_8cdd_c5c50372137d.slice/crio-0be8218d526213c6c772e338de7972cc448e8132542540cb127888507e05e2bd WatchSource:0}: Error finding container 0be8218d526213c6c772e338de7972cc448e8132542540cb127888507e05e2bd: Status 404 returned error can't find the container with id 0be8218d526213c6c772e338de7972cc448e8132542540cb127888507e05e2bd Oct 08 19:52:35 crc kubenswrapper[4988]: I1008 19:52:35.913441 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-api-764db59bb9-p78lt"] Oct 08 19:52:35 crc kubenswrapper[4988]: I1008 19:52:35.935666 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-api-764db59bb9-p78lt" event={"ID":"a0c334ad-5ffa-41d9-8cdd-c5c50372137d","Type":"ContainerStarted","Data":"0be8218d526213c6c772e338de7972cc448e8132542540cb127888507e05e2bd"} Oct 08 19:52:36 crc kubenswrapper[4988]: I1008 19:52:36.954170 4988 generic.go:334] "Generic (PLEG): container finished" podID="43ab7316-2b23-4b0c-9146-0d87054cc1e9" containerID="bb8901ca13542d2d72987c27ec92ba393cc266bf8aa8ca904b9ac77634514a88" exitCode=0 Oct 08 19:52:36 crc kubenswrapper[4988]: I1008 19:52:36.954307 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-image-upload-678599687f-xxfmc" event={"ID":"43ab7316-2b23-4b0c-9146-0d87054cc1e9","Type":"ContainerDied","Data":"bb8901ca13542d2d72987c27ec92ba393cc266bf8aa8ca904b9ac77634514a88"} Oct 08 19:52:36 crc kubenswrapper[4988]: I1008 19:52:36.958843 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-api-764db59bb9-p78lt" event={"ID":"a0c334ad-5ffa-41d9-8cdd-c5c50372137d","Type":"ContainerStarted","Data":"466111f11ee8f1b8c81a4520008192c7dc3f210d3e9b472e0473e93326c6acc0"} Oct 08 19:52:37 crc kubenswrapper[4988]: I1008 19:52:37.968221 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-image-upload-678599687f-xxfmc" event={"ID":"43ab7316-2b23-4b0c-9146-0d87054cc1e9","Type":"ContainerStarted","Data":"41dc2fb836716fd813681aad2b2be8e323d7090abe1c69c1ad51df34c6c908f7"} Oct 08 19:52:37 crc kubenswrapper[4988]: I1008 19:52:37.969497 4988 generic.go:334] "Generic (PLEG): container finished" podID="a0c334ad-5ffa-41d9-8cdd-c5c50372137d" containerID="466111f11ee8f1b8c81a4520008192c7dc3f210d3e9b472e0473e93326c6acc0" exitCode=0 Oct 08 19:52:37 crc kubenswrapper[4988]: I1008 19:52:37.969529 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-api-764db59bb9-p78lt" event={"ID":"a0c334ad-5ffa-41d9-8cdd-c5c50372137d","Type":"ContainerDied","Data":"466111f11ee8f1b8c81a4520008192c7dc3f210d3e9b472e0473e93326c6acc0"} Oct 08 19:52:38 crc kubenswrapper[4988]: I1008 19:52:38.014368 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/octavia-image-upload-678599687f-xxfmc" podStartSLOduration=5.437134535 podStartE2EDuration="16.014314721s" podCreationTimestamp="2025-10-08 19:52:22 +0000 UTC" firstStartedPulling="2025-10-08 19:52:22.925629683 +0000 UTC m=+6088.375472453" lastFinishedPulling="2025-10-08 19:52:33.502809869 +0000 UTC m=+6098.952652639" observedRunningTime="2025-10-08 19:52:37.989068329 +0000 UTC m=+6103.438911099" watchObservedRunningTime="2025-10-08 19:52:38.014314721 +0000 UTC m=+6103.464157531" Oct 08 19:52:38 crc kubenswrapper[4988]: I1008 19:52:38.981588 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-api-764db59bb9-p78lt" event={"ID":"a0c334ad-5ffa-41d9-8cdd-c5c50372137d","Type":"ContainerStarted","Data":"974f4133a98a2ba84f5e5f690f49c7a0c0b652e81d03893b8d261e9c8b3249a0"} Oct 08 19:52:38 crc kubenswrapper[4988]: I1008 19:52:38.981810 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-api-764db59bb9-p78lt" event={"ID":"a0c334ad-5ffa-41d9-8cdd-c5c50372137d","Type":"ContainerStarted","Data":"1381734b810306ade7ad84e215da50332fd261a6c5f4d1f5365583b84ea08d72"} Oct 08 19:52:38 crc kubenswrapper[4988]: I1008 19:52:38.981824 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/octavia-api-764db59bb9-p78lt" Oct 08 19:52:38 crc kubenswrapper[4988]: I1008 19:52:38.981834 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/octavia-api-764db59bb9-p78lt" Oct 08 19:52:51 crc kubenswrapper[4988]: I1008 19:52:51.772720 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/octavia-rsyslog-xdhtp" Oct 08 19:52:51 crc kubenswrapper[4988]: I1008 19:52:51.793773 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/octavia-api-764db59bb9-p78lt" podStartSLOduration=17.793759241 podStartE2EDuration="17.793759241s" podCreationTimestamp="2025-10-08 19:52:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 19:52:39.028280512 +0000 UTC m=+6104.478123282" watchObservedRunningTime="2025-10-08 19:52:51.793759241 +0000 UTC m=+6117.243602011" Oct 08 19:52:54 crc kubenswrapper[4988]: I1008 19:52:54.182720 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/octavia-api-764db59bb9-p78lt" Oct 08 19:52:54 crc kubenswrapper[4988]: I1008 19:52:54.241986 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/octavia-api-764db59bb9-p78lt" Oct 08 19:52:54 crc kubenswrapper[4988]: I1008 19:52:54.315020 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/octavia-api-5d77847989-vhxvn"] Oct 08 19:52:54 crc kubenswrapper[4988]: I1008 19:52:54.315396 4988 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/octavia-api-5d77847989-vhxvn" podUID="c144edcb-2dfc-4f6d-85a3-77ae04fd9fb5" containerName="octavia-api" containerID="cri-o://28dbba51a65b539a92846427a3d45c9ac7d4d4f18b74b5d9d539aee4c3142293" gracePeriod=30 Oct 08 19:52:54 crc kubenswrapper[4988]: I1008 19:52:54.315556 4988 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/octavia-api-5d77847989-vhxvn" podUID="c144edcb-2dfc-4f6d-85a3-77ae04fd9fb5" containerName="octavia-api-provider-agent" containerID="cri-o://571330e5b42c1e9fd691583f34d35e8b73f643c5b7dc4b0f8b4e40e4b973a33e" gracePeriod=30 Oct 08 19:52:55 crc kubenswrapper[4988]: I1008 19:52:55.149842 4988 generic.go:334] "Generic (PLEG): container finished" podID="c144edcb-2dfc-4f6d-85a3-77ae04fd9fb5" containerID="571330e5b42c1e9fd691583f34d35e8b73f643c5b7dc4b0f8b4e40e4b973a33e" exitCode=0 Oct 08 19:52:55 crc kubenswrapper[4988]: I1008 19:52:55.149926 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-api-5d77847989-vhxvn" event={"ID":"c144edcb-2dfc-4f6d-85a3-77ae04fd9fb5","Type":"ContainerDied","Data":"571330e5b42c1e9fd691583f34d35e8b73f643c5b7dc4b0f8b4e40e4b973a33e"} Oct 08 19:52:55 crc kubenswrapper[4988]: I1008 19:52:55.866511 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/octavia-image-upload-678599687f-xxfmc"] Oct 08 19:52:55 crc kubenswrapper[4988]: I1008 19:52:55.866806 4988 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/octavia-image-upload-678599687f-xxfmc" podUID="43ab7316-2b23-4b0c-9146-0d87054cc1e9" containerName="octavia-amphora-httpd" containerID="cri-o://41dc2fb836716fd813681aad2b2be8e323d7090abe1c69c1ad51df34c6c908f7" gracePeriod=30 Oct 08 19:52:56 crc kubenswrapper[4988]: I1008 19:52:56.172234 4988 generic.go:334] "Generic (PLEG): container finished" podID="43ab7316-2b23-4b0c-9146-0d87054cc1e9" containerID="41dc2fb836716fd813681aad2b2be8e323d7090abe1c69c1ad51df34c6c908f7" exitCode=0 Oct 08 19:52:56 crc kubenswrapper[4988]: I1008 19:52:56.172637 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-image-upload-678599687f-xxfmc" event={"ID":"43ab7316-2b23-4b0c-9146-0d87054cc1e9","Type":"ContainerDied","Data":"41dc2fb836716fd813681aad2b2be8e323d7090abe1c69c1ad51df34c6c908f7"} Oct 08 19:52:56 crc kubenswrapper[4988]: I1008 19:52:56.372686 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-image-upload-678599687f-xxfmc" Oct 08 19:52:56 crc kubenswrapper[4988]: I1008 19:52:56.536618 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"amphora-image\" (UniqueName: \"kubernetes.io/empty-dir/43ab7316-2b23-4b0c-9146-0d87054cc1e9-amphora-image\") pod \"43ab7316-2b23-4b0c-9146-0d87054cc1e9\" (UID: \"43ab7316-2b23-4b0c-9146-0d87054cc1e9\") " Oct 08 19:52:56 crc kubenswrapper[4988]: I1008 19:52:56.536771 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/43ab7316-2b23-4b0c-9146-0d87054cc1e9-httpd-config\") pod \"43ab7316-2b23-4b0c-9146-0d87054cc1e9\" (UID: \"43ab7316-2b23-4b0c-9146-0d87054cc1e9\") " Oct 08 19:52:56 crc kubenswrapper[4988]: I1008 19:52:56.564303 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43ab7316-2b23-4b0c-9146-0d87054cc1e9-httpd-config" (OuterVolumeSpecName: "httpd-config") pod "43ab7316-2b23-4b0c-9146-0d87054cc1e9" (UID: "43ab7316-2b23-4b0c-9146-0d87054cc1e9"). InnerVolumeSpecName "httpd-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 19:52:56 crc kubenswrapper[4988]: I1008 19:52:56.599659 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/43ab7316-2b23-4b0c-9146-0d87054cc1e9-amphora-image" (OuterVolumeSpecName: "amphora-image") pod "43ab7316-2b23-4b0c-9146-0d87054cc1e9" (UID: "43ab7316-2b23-4b0c-9146-0d87054cc1e9"). InnerVolumeSpecName "amphora-image". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 19:52:56 crc kubenswrapper[4988]: I1008 19:52:56.640224 4988 reconciler_common.go:293] "Volume detached for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/43ab7316-2b23-4b0c-9146-0d87054cc1e9-httpd-config\") on node \"crc\" DevicePath \"\"" Oct 08 19:52:56 crc kubenswrapper[4988]: I1008 19:52:56.640311 4988 reconciler_common.go:293] "Volume detached for volume \"amphora-image\" (UniqueName: \"kubernetes.io/empty-dir/43ab7316-2b23-4b0c-9146-0d87054cc1e9-amphora-image\") on node \"crc\" DevicePath \"\"" Oct 08 19:52:57 crc kubenswrapper[4988]: I1008 19:52:57.187201 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-image-upload-678599687f-xxfmc" event={"ID":"43ab7316-2b23-4b0c-9146-0d87054cc1e9","Type":"ContainerDied","Data":"5bd2dd3f735acd07283f5231a7dc7019749b6b75a2cc8fa76cba82ccf691ec91"} Oct 08 19:52:57 crc kubenswrapper[4988]: I1008 19:52:57.187281 4988 scope.go:117] "RemoveContainer" containerID="41dc2fb836716fd813681aad2b2be8e323d7090abe1c69c1ad51df34c6c908f7" Oct 08 19:52:57 crc kubenswrapper[4988]: I1008 19:52:57.187447 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-image-upload-678599687f-xxfmc" Oct 08 19:52:57 crc kubenswrapper[4988]: I1008 19:52:57.208603 4988 scope.go:117] "RemoveContainer" containerID="bb8901ca13542d2d72987c27ec92ba393cc266bf8aa8ca904b9ac77634514a88" Oct 08 19:52:57 crc kubenswrapper[4988]: I1008 19:52:57.250446 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/octavia-image-upload-678599687f-xxfmc"] Oct 08 19:52:57 crc kubenswrapper[4988]: I1008 19:52:57.259606 4988 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/octavia-image-upload-678599687f-xxfmc"] Oct 08 19:52:57 crc kubenswrapper[4988]: I1008 19:52:57.972868 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-api-5d77847989-vhxvn" Oct 08 19:52:58 crc kubenswrapper[4988]: I1008 19:52:58.066078 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c144edcb-2dfc-4f6d-85a3-77ae04fd9fb5-scripts\") pod \"c144edcb-2dfc-4f6d-85a3-77ae04fd9fb5\" (UID: \"c144edcb-2dfc-4f6d-85a3-77ae04fd9fb5\") " Oct 08 19:52:58 crc kubenswrapper[4988]: I1008 19:52:58.066413 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c144edcb-2dfc-4f6d-85a3-77ae04fd9fb5-combined-ca-bundle\") pod \"c144edcb-2dfc-4f6d-85a3-77ae04fd9fb5\" (UID: \"c144edcb-2dfc-4f6d-85a3-77ae04fd9fb5\") " Oct 08 19:52:58 crc kubenswrapper[4988]: I1008 19:52:58.066491 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/c144edcb-2dfc-4f6d-85a3-77ae04fd9fb5-config-data-merged\") pod \"c144edcb-2dfc-4f6d-85a3-77ae04fd9fb5\" (UID: \"c144edcb-2dfc-4f6d-85a3-77ae04fd9fb5\") " Oct 08 19:52:58 crc kubenswrapper[4988]: I1008 19:52:58.066542 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c144edcb-2dfc-4f6d-85a3-77ae04fd9fb5-config-data\") pod \"c144edcb-2dfc-4f6d-85a3-77ae04fd9fb5\" (UID: \"c144edcb-2dfc-4f6d-85a3-77ae04fd9fb5\") " Oct 08 19:52:58 crc kubenswrapper[4988]: I1008 19:52:58.066618 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/c144edcb-2dfc-4f6d-85a3-77ae04fd9fb5-ovndb-tls-certs\") pod \"c144edcb-2dfc-4f6d-85a3-77ae04fd9fb5\" (UID: \"c144edcb-2dfc-4f6d-85a3-77ae04fd9fb5\") " Oct 08 19:52:58 crc kubenswrapper[4988]: I1008 19:52:58.066661 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"octavia-run\" (UniqueName: \"kubernetes.io/empty-dir/c144edcb-2dfc-4f6d-85a3-77ae04fd9fb5-octavia-run\") pod \"c144edcb-2dfc-4f6d-85a3-77ae04fd9fb5\" (UID: \"c144edcb-2dfc-4f6d-85a3-77ae04fd9fb5\") " Oct 08 19:52:58 crc kubenswrapper[4988]: I1008 19:52:58.067322 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c144edcb-2dfc-4f6d-85a3-77ae04fd9fb5-octavia-run" (OuterVolumeSpecName: "octavia-run") pod "c144edcb-2dfc-4f6d-85a3-77ae04fd9fb5" (UID: "c144edcb-2dfc-4f6d-85a3-77ae04fd9fb5"). InnerVolumeSpecName "octavia-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 19:52:58 crc kubenswrapper[4988]: I1008 19:52:58.088317 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c144edcb-2dfc-4f6d-85a3-77ae04fd9fb5-config-data" (OuterVolumeSpecName: "config-data") pod "c144edcb-2dfc-4f6d-85a3-77ae04fd9fb5" (UID: "c144edcb-2dfc-4f6d-85a3-77ae04fd9fb5"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 19:52:58 crc kubenswrapper[4988]: I1008 19:52:58.091069 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c144edcb-2dfc-4f6d-85a3-77ae04fd9fb5-scripts" (OuterVolumeSpecName: "scripts") pod "c144edcb-2dfc-4f6d-85a3-77ae04fd9fb5" (UID: "c144edcb-2dfc-4f6d-85a3-77ae04fd9fb5"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 19:52:58 crc kubenswrapper[4988]: I1008 19:52:58.119230 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c144edcb-2dfc-4f6d-85a3-77ae04fd9fb5-config-data-merged" (OuterVolumeSpecName: "config-data-merged") pod "c144edcb-2dfc-4f6d-85a3-77ae04fd9fb5" (UID: "c144edcb-2dfc-4f6d-85a3-77ae04fd9fb5"). InnerVolumeSpecName "config-data-merged". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 19:52:58 crc kubenswrapper[4988]: I1008 19:52:58.134653 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c144edcb-2dfc-4f6d-85a3-77ae04fd9fb5-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "c144edcb-2dfc-4f6d-85a3-77ae04fd9fb5" (UID: "c144edcb-2dfc-4f6d-85a3-77ae04fd9fb5"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 19:52:58 crc kubenswrapper[4988]: I1008 19:52:58.168443 4988 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c144edcb-2dfc-4f6d-85a3-77ae04fd9fb5-config-data\") on node \"crc\" DevicePath \"\"" Oct 08 19:52:58 crc kubenswrapper[4988]: I1008 19:52:58.168474 4988 reconciler_common.go:293] "Volume detached for volume \"octavia-run\" (UniqueName: \"kubernetes.io/empty-dir/c144edcb-2dfc-4f6d-85a3-77ae04fd9fb5-octavia-run\") on node \"crc\" DevicePath \"\"" Oct 08 19:52:58 crc kubenswrapper[4988]: I1008 19:52:58.168484 4988 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c144edcb-2dfc-4f6d-85a3-77ae04fd9fb5-scripts\") on node \"crc\" DevicePath \"\"" Oct 08 19:52:58 crc kubenswrapper[4988]: I1008 19:52:58.168492 4988 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c144edcb-2dfc-4f6d-85a3-77ae04fd9fb5-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 19:52:58 crc kubenswrapper[4988]: I1008 19:52:58.168501 4988 reconciler_common.go:293] "Volume detached for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/c144edcb-2dfc-4f6d-85a3-77ae04fd9fb5-config-data-merged\") on node \"crc\" DevicePath \"\"" Oct 08 19:52:58 crc kubenswrapper[4988]: I1008 19:52:58.207847 4988 generic.go:334] "Generic (PLEG): container finished" podID="c144edcb-2dfc-4f6d-85a3-77ae04fd9fb5" containerID="28dbba51a65b539a92846427a3d45c9ac7d4d4f18b74b5d9d539aee4c3142293" exitCode=0 Oct 08 19:52:58 crc kubenswrapper[4988]: I1008 19:52:58.207901 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-api-5d77847989-vhxvn" event={"ID":"c144edcb-2dfc-4f6d-85a3-77ae04fd9fb5","Type":"ContainerDied","Data":"28dbba51a65b539a92846427a3d45c9ac7d4d4f18b74b5d9d539aee4c3142293"} Oct 08 19:52:58 crc kubenswrapper[4988]: I1008 19:52:58.207951 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-api-5d77847989-vhxvn" Oct 08 19:52:58 crc kubenswrapper[4988]: I1008 19:52:58.207972 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-api-5d77847989-vhxvn" event={"ID":"c144edcb-2dfc-4f6d-85a3-77ae04fd9fb5","Type":"ContainerDied","Data":"cf1c3bc8156553244fb189779f0c5254dd623357704835e658e0fad467cb9724"} Oct 08 19:52:58 crc kubenswrapper[4988]: I1008 19:52:58.207990 4988 scope.go:117] "RemoveContainer" containerID="571330e5b42c1e9fd691583f34d35e8b73f643c5b7dc4b0f8b4e40e4b973a33e" Oct 08 19:52:58 crc kubenswrapper[4988]: I1008 19:52:58.234514 4988 scope.go:117] "RemoveContainer" containerID="28dbba51a65b539a92846427a3d45c9ac7d4d4f18b74b5d9d539aee4c3142293" Oct 08 19:52:58 crc kubenswrapper[4988]: I1008 19:52:58.269626 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c144edcb-2dfc-4f6d-85a3-77ae04fd9fb5-ovndb-tls-certs" (OuterVolumeSpecName: "ovndb-tls-certs") pod "c144edcb-2dfc-4f6d-85a3-77ae04fd9fb5" (UID: "c144edcb-2dfc-4f6d-85a3-77ae04fd9fb5"). InnerVolumeSpecName "ovndb-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 19:52:58 crc kubenswrapper[4988]: I1008 19:52:58.269992 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/c144edcb-2dfc-4f6d-85a3-77ae04fd9fb5-ovndb-tls-certs\") pod \"c144edcb-2dfc-4f6d-85a3-77ae04fd9fb5\" (UID: \"c144edcb-2dfc-4f6d-85a3-77ae04fd9fb5\") " Oct 08 19:52:58 crc kubenswrapper[4988]: W1008 19:52:58.270085 4988 empty_dir.go:500] Warning: Unmount skipped because path does not exist: /var/lib/kubelet/pods/c144edcb-2dfc-4f6d-85a3-77ae04fd9fb5/volumes/kubernetes.io~secret/ovndb-tls-certs Oct 08 19:52:58 crc kubenswrapper[4988]: I1008 19:52:58.270096 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c144edcb-2dfc-4f6d-85a3-77ae04fd9fb5-ovndb-tls-certs" (OuterVolumeSpecName: "ovndb-tls-certs") pod "c144edcb-2dfc-4f6d-85a3-77ae04fd9fb5" (UID: "c144edcb-2dfc-4f6d-85a3-77ae04fd9fb5"). InnerVolumeSpecName "ovndb-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 19:52:58 crc kubenswrapper[4988]: I1008 19:52:58.271079 4988 reconciler_common.go:293] "Volume detached for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/c144edcb-2dfc-4f6d-85a3-77ae04fd9fb5-ovndb-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 08 19:52:58 crc kubenswrapper[4988]: I1008 19:52:58.361352 4988 scope.go:117] "RemoveContainer" containerID="2b1a5349c2854b7638f49bce3966bd7c10039d7a5ce4ab4106b399cb7368f25c" Oct 08 19:52:58 crc kubenswrapper[4988]: I1008 19:52:58.400537 4988 scope.go:117] "RemoveContainer" containerID="571330e5b42c1e9fd691583f34d35e8b73f643c5b7dc4b0f8b4e40e4b973a33e" Oct 08 19:52:58 crc kubenswrapper[4988]: E1008 19:52:58.401050 4988 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"571330e5b42c1e9fd691583f34d35e8b73f643c5b7dc4b0f8b4e40e4b973a33e\": container with ID starting with 571330e5b42c1e9fd691583f34d35e8b73f643c5b7dc4b0f8b4e40e4b973a33e not found: ID does not exist" containerID="571330e5b42c1e9fd691583f34d35e8b73f643c5b7dc4b0f8b4e40e4b973a33e" Oct 08 19:52:58 crc kubenswrapper[4988]: I1008 19:52:58.401096 4988 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"571330e5b42c1e9fd691583f34d35e8b73f643c5b7dc4b0f8b4e40e4b973a33e"} err="failed to get container status \"571330e5b42c1e9fd691583f34d35e8b73f643c5b7dc4b0f8b4e40e4b973a33e\": rpc error: code = NotFound desc = could not find container \"571330e5b42c1e9fd691583f34d35e8b73f643c5b7dc4b0f8b4e40e4b973a33e\": container with ID starting with 571330e5b42c1e9fd691583f34d35e8b73f643c5b7dc4b0f8b4e40e4b973a33e not found: ID does not exist" Oct 08 19:52:58 crc kubenswrapper[4988]: I1008 19:52:58.401127 4988 scope.go:117] "RemoveContainer" containerID="28dbba51a65b539a92846427a3d45c9ac7d4d4f18b74b5d9d539aee4c3142293" Oct 08 19:52:58 crc kubenswrapper[4988]: E1008 19:52:58.401823 4988 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"28dbba51a65b539a92846427a3d45c9ac7d4d4f18b74b5d9d539aee4c3142293\": container with ID starting with 28dbba51a65b539a92846427a3d45c9ac7d4d4f18b74b5d9d539aee4c3142293 not found: ID does not exist" containerID="28dbba51a65b539a92846427a3d45c9ac7d4d4f18b74b5d9d539aee4c3142293" Oct 08 19:52:58 crc kubenswrapper[4988]: I1008 19:52:58.401853 4988 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"28dbba51a65b539a92846427a3d45c9ac7d4d4f18b74b5d9d539aee4c3142293"} err="failed to get container status \"28dbba51a65b539a92846427a3d45c9ac7d4d4f18b74b5d9d539aee4c3142293\": rpc error: code = NotFound desc = could not find container \"28dbba51a65b539a92846427a3d45c9ac7d4d4f18b74b5d9d539aee4c3142293\": container with ID starting with 28dbba51a65b539a92846427a3d45c9ac7d4d4f18b74b5d9d539aee4c3142293 not found: ID does not exist" Oct 08 19:52:58 crc kubenswrapper[4988]: I1008 19:52:58.401873 4988 scope.go:117] "RemoveContainer" containerID="2b1a5349c2854b7638f49bce3966bd7c10039d7a5ce4ab4106b399cb7368f25c" Oct 08 19:52:58 crc kubenswrapper[4988]: E1008 19:52:58.402634 4988 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2b1a5349c2854b7638f49bce3966bd7c10039d7a5ce4ab4106b399cb7368f25c\": container with ID starting with 2b1a5349c2854b7638f49bce3966bd7c10039d7a5ce4ab4106b399cb7368f25c not found: ID does not exist" containerID="2b1a5349c2854b7638f49bce3966bd7c10039d7a5ce4ab4106b399cb7368f25c" Oct 08 19:52:58 crc kubenswrapper[4988]: I1008 19:52:58.402684 4988 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2b1a5349c2854b7638f49bce3966bd7c10039d7a5ce4ab4106b399cb7368f25c"} err="failed to get container status \"2b1a5349c2854b7638f49bce3966bd7c10039d7a5ce4ab4106b399cb7368f25c\": rpc error: code = NotFound desc = could not find container \"2b1a5349c2854b7638f49bce3966bd7c10039d7a5ce4ab4106b399cb7368f25c\": container with ID starting with 2b1a5349c2854b7638f49bce3966bd7c10039d7a5ce4ab4106b399cb7368f25c not found: ID does not exist" Oct 08 19:52:58 crc kubenswrapper[4988]: I1008 19:52:58.560658 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/octavia-api-5d77847989-vhxvn"] Oct 08 19:52:58 crc kubenswrapper[4988]: I1008 19:52:58.577591 4988 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/octavia-api-5d77847989-vhxvn"] Oct 08 19:52:59 crc kubenswrapper[4988]: I1008 19:52:59.257199 4988 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="43ab7316-2b23-4b0c-9146-0d87054cc1e9" path="/var/lib/kubelet/pods/43ab7316-2b23-4b0c-9146-0d87054cc1e9/volumes" Oct 08 19:52:59 crc kubenswrapper[4988]: I1008 19:52:59.258197 4988 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c144edcb-2dfc-4f6d-85a3-77ae04fd9fb5" path="/var/lib/kubelet/pods/c144edcb-2dfc-4f6d-85a3-77ae04fd9fb5/volumes" Oct 08 19:53:01 crc kubenswrapper[4988]: I1008 19:53:01.423575 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/octavia-image-upload-678599687f-2f6h5"] Oct 08 19:53:01 crc kubenswrapper[4988]: E1008 19:53:01.424793 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c144edcb-2dfc-4f6d-85a3-77ae04fd9fb5" containerName="octavia-api" Oct 08 19:53:01 crc kubenswrapper[4988]: I1008 19:53:01.424822 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="c144edcb-2dfc-4f6d-85a3-77ae04fd9fb5" containerName="octavia-api" Oct 08 19:53:01 crc kubenswrapper[4988]: E1008 19:53:01.424869 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c144edcb-2dfc-4f6d-85a3-77ae04fd9fb5" containerName="octavia-api-provider-agent" Oct 08 19:53:01 crc kubenswrapper[4988]: I1008 19:53:01.424883 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="c144edcb-2dfc-4f6d-85a3-77ae04fd9fb5" containerName="octavia-api-provider-agent" Oct 08 19:53:01 crc kubenswrapper[4988]: E1008 19:53:01.424909 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c144edcb-2dfc-4f6d-85a3-77ae04fd9fb5" containerName="init" Oct 08 19:53:01 crc kubenswrapper[4988]: I1008 19:53:01.424923 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="c144edcb-2dfc-4f6d-85a3-77ae04fd9fb5" containerName="init" Oct 08 19:53:01 crc kubenswrapper[4988]: E1008 19:53:01.424941 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="43ab7316-2b23-4b0c-9146-0d87054cc1e9" containerName="octavia-amphora-httpd" Oct 08 19:53:01 crc kubenswrapper[4988]: I1008 19:53:01.424955 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="43ab7316-2b23-4b0c-9146-0d87054cc1e9" containerName="octavia-amphora-httpd" Oct 08 19:53:01 crc kubenswrapper[4988]: E1008 19:53:01.425003 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="43ab7316-2b23-4b0c-9146-0d87054cc1e9" containerName="init" Oct 08 19:53:01 crc kubenswrapper[4988]: I1008 19:53:01.425015 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="43ab7316-2b23-4b0c-9146-0d87054cc1e9" containerName="init" Oct 08 19:53:01 crc kubenswrapper[4988]: I1008 19:53:01.425350 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="43ab7316-2b23-4b0c-9146-0d87054cc1e9" containerName="octavia-amphora-httpd" Oct 08 19:53:01 crc kubenswrapper[4988]: I1008 19:53:01.425450 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="c144edcb-2dfc-4f6d-85a3-77ae04fd9fb5" containerName="octavia-api" Oct 08 19:53:01 crc kubenswrapper[4988]: I1008 19:53:01.425487 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="c144edcb-2dfc-4f6d-85a3-77ae04fd9fb5" containerName="octavia-api-provider-agent" Oct 08 19:53:01 crc kubenswrapper[4988]: I1008 19:53:01.427345 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-image-upload-678599687f-2f6h5" Oct 08 19:53:01 crc kubenswrapper[4988]: I1008 19:53:01.433685 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"octavia-config-data" Oct 08 19:53:01 crc kubenswrapper[4988]: I1008 19:53:01.450763 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-image-upload-678599687f-2f6h5"] Oct 08 19:53:01 crc kubenswrapper[4988]: I1008 19:53:01.540730 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"amphora-image\" (UniqueName: \"kubernetes.io/empty-dir/68ad708f-db19-445d-a991-1d1a011a4a9f-amphora-image\") pod \"octavia-image-upload-678599687f-2f6h5\" (UID: \"68ad708f-db19-445d-a991-1d1a011a4a9f\") " pod="openstack/octavia-image-upload-678599687f-2f6h5" Oct 08 19:53:01 crc kubenswrapper[4988]: I1008 19:53:01.541012 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/68ad708f-db19-445d-a991-1d1a011a4a9f-httpd-config\") pod \"octavia-image-upload-678599687f-2f6h5\" (UID: \"68ad708f-db19-445d-a991-1d1a011a4a9f\") " pod="openstack/octavia-image-upload-678599687f-2f6h5" Oct 08 19:53:01 crc kubenswrapper[4988]: I1008 19:53:01.643213 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"amphora-image\" (UniqueName: \"kubernetes.io/empty-dir/68ad708f-db19-445d-a991-1d1a011a4a9f-amphora-image\") pod \"octavia-image-upload-678599687f-2f6h5\" (UID: \"68ad708f-db19-445d-a991-1d1a011a4a9f\") " pod="openstack/octavia-image-upload-678599687f-2f6h5" Oct 08 19:53:01 crc kubenswrapper[4988]: I1008 19:53:01.643557 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/68ad708f-db19-445d-a991-1d1a011a4a9f-httpd-config\") pod \"octavia-image-upload-678599687f-2f6h5\" (UID: \"68ad708f-db19-445d-a991-1d1a011a4a9f\") " pod="openstack/octavia-image-upload-678599687f-2f6h5" Oct 08 19:53:01 crc kubenswrapper[4988]: I1008 19:53:01.644287 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"amphora-image\" (UniqueName: \"kubernetes.io/empty-dir/68ad708f-db19-445d-a991-1d1a011a4a9f-amphora-image\") pod \"octavia-image-upload-678599687f-2f6h5\" (UID: \"68ad708f-db19-445d-a991-1d1a011a4a9f\") " pod="openstack/octavia-image-upload-678599687f-2f6h5" Oct 08 19:53:01 crc kubenswrapper[4988]: I1008 19:53:01.654411 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/68ad708f-db19-445d-a991-1d1a011a4a9f-httpd-config\") pod \"octavia-image-upload-678599687f-2f6h5\" (UID: \"68ad708f-db19-445d-a991-1d1a011a4a9f\") " pod="openstack/octavia-image-upload-678599687f-2f6h5" Oct 08 19:53:01 crc kubenswrapper[4988]: I1008 19:53:01.748681 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-image-upload-678599687f-2f6h5" Oct 08 19:53:02 crc kubenswrapper[4988]: I1008 19:53:02.259983 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-image-upload-678599687f-2f6h5"] Oct 08 19:53:02 crc kubenswrapper[4988]: I1008 19:53:02.264539 4988 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 08 19:53:03 crc kubenswrapper[4988]: I1008 19:53:03.271811 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-image-upload-678599687f-2f6h5" event={"ID":"68ad708f-db19-445d-a991-1d1a011a4a9f","Type":"ContainerStarted","Data":"b185ba9019f11726a7fe38ff87218f7e18eb65c76012daeb7b90e0af47811031"} Oct 08 19:53:03 crc kubenswrapper[4988]: I1008 19:53:03.273731 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-image-upload-678599687f-2f6h5" event={"ID":"68ad708f-db19-445d-a991-1d1a011a4a9f","Type":"ContainerStarted","Data":"d32dc2320f9cc90530885ed06904bf686d1a25083693a7425b92f102198a4855"} Oct 08 19:53:06 crc kubenswrapper[4988]: I1008 19:53:06.312822 4988 generic.go:334] "Generic (PLEG): container finished" podID="68ad708f-db19-445d-a991-1d1a011a4a9f" containerID="b185ba9019f11726a7fe38ff87218f7e18eb65c76012daeb7b90e0af47811031" exitCode=0 Oct 08 19:53:06 crc kubenswrapper[4988]: I1008 19:53:06.312949 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-image-upload-678599687f-2f6h5" event={"ID":"68ad708f-db19-445d-a991-1d1a011a4a9f","Type":"ContainerDied","Data":"b185ba9019f11726a7fe38ff87218f7e18eb65c76012daeb7b90e0af47811031"} Oct 08 19:53:07 crc kubenswrapper[4988]: I1008 19:53:07.326960 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-image-upload-678599687f-2f6h5" event={"ID":"68ad708f-db19-445d-a991-1d1a011a4a9f","Type":"ContainerStarted","Data":"2747b19611502726b3530904286dde2d83a9611906ee0e938bf299d318b41e08"} Oct 08 19:53:07 crc kubenswrapper[4988]: I1008 19:53:07.353892 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/octavia-image-upload-678599687f-2f6h5" podStartSLOduration=5.948104026 podStartE2EDuration="6.353858364s" podCreationTimestamp="2025-10-08 19:53:01 +0000 UTC" firstStartedPulling="2025-10-08 19:53:02.264124562 +0000 UTC m=+6127.713967372" lastFinishedPulling="2025-10-08 19:53:02.66987893 +0000 UTC m=+6128.119721710" observedRunningTime="2025-10-08 19:53:07.347226953 +0000 UTC m=+6132.797069723" watchObservedRunningTime="2025-10-08 19:53:07.353858364 +0000 UTC m=+6132.803701134" Oct 08 19:53:25 crc kubenswrapper[4988]: I1008 19:53:25.082876 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-create-htq7q"] Oct 08 19:53:25 crc kubenswrapper[4988]: I1008 19:53:25.094768 4988 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-create-htq7q"] Oct 08 19:53:25 crc kubenswrapper[4988]: I1008 19:53:25.250259 4988 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="45a3d8ec-0896-49d3-ac9b-f294c0f66f62" path="/var/lib/kubelet/pods/45a3d8ec-0896-49d3-ac9b-f294c0f66f62/volumes" Oct 08 19:53:27 crc kubenswrapper[4988]: I1008 19:53:27.145565 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/octavia-healthmanager-hhkfh"] Oct 08 19:53:27 crc kubenswrapper[4988]: I1008 19:53:27.149220 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-healthmanager-hhkfh" Oct 08 19:53:27 crc kubenswrapper[4988]: I1008 19:53:27.159769 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"octavia-healthmanager-config-data" Oct 08 19:53:27 crc kubenswrapper[4988]: I1008 19:53:27.159813 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"octavia-certs-secret" Oct 08 19:53:27 crc kubenswrapper[4988]: I1008 19:53:27.159778 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"octavia-healthmanager-scripts" Oct 08 19:53:27 crc kubenswrapper[4988]: I1008 19:53:27.167621 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-healthmanager-hhkfh"] Oct 08 19:53:27 crc kubenswrapper[4988]: I1008 19:53:27.230005 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/d7a04c8d-1f4c-4dd8-98b6-7afdc47bece7-config-data-merged\") pod \"octavia-healthmanager-hhkfh\" (UID: \"d7a04c8d-1f4c-4dd8-98b6-7afdc47bece7\") " pod="openstack/octavia-healthmanager-hhkfh" Oct 08 19:53:27 crc kubenswrapper[4988]: I1008 19:53:27.230426 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d7a04c8d-1f4c-4dd8-98b6-7afdc47bece7-scripts\") pod \"octavia-healthmanager-hhkfh\" (UID: \"d7a04c8d-1f4c-4dd8-98b6-7afdc47bece7\") " pod="openstack/octavia-healthmanager-hhkfh" Oct 08 19:53:27 crc kubenswrapper[4988]: I1008 19:53:27.230767 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hm-ports\" (UniqueName: \"kubernetes.io/configmap/d7a04c8d-1f4c-4dd8-98b6-7afdc47bece7-hm-ports\") pod \"octavia-healthmanager-hhkfh\" (UID: \"d7a04c8d-1f4c-4dd8-98b6-7afdc47bece7\") " pod="openstack/octavia-healthmanager-hhkfh" Oct 08 19:53:27 crc kubenswrapper[4988]: I1008 19:53:27.230825 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d7a04c8d-1f4c-4dd8-98b6-7afdc47bece7-config-data\") pod \"octavia-healthmanager-hhkfh\" (UID: \"d7a04c8d-1f4c-4dd8-98b6-7afdc47bece7\") " pod="openstack/octavia-healthmanager-hhkfh" Oct 08 19:53:27 crc kubenswrapper[4988]: I1008 19:53:27.230943 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"amphora-certs\" (UniqueName: \"kubernetes.io/secret/d7a04c8d-1f4c-4dd8-98b6-7afdc47bece7-amphora-certs\") pod \"octavia-healthmanager-hhkfh\" (UID: \"d7a04c8d-1f4c-4dd8-98b6-7afdc47bece7\") " pod="openstack/octavia-healthmanager-hhkfh" Oct 08 19:53:27 crc kubenswrapper[4988]: I1008 19:53:27.231007 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d7a04c8d-1f4c-4dd8-98b6-7afdc47bece7-combined-ca-bundle\") pod \"octavia-healthmanager-hhkfh\" (UID: \"d7a04c8d-1f4c-4dd8-98b6-7afdc47bece7\") " pod="openstack/octavia-healthmanager-hhkfh" Oct 08 19:53:27 crc kubenswrapper[4988]: I1008 19:53:27.333208 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/d7a04c8d-1f4c-4dd8-98b6-7afdc47bece7-config-data-merged\") pod \"octavia-healthmanager-hhkfh\" (UID: \"d7a04c8d-1f4c-4dd8-98b6-7afdc47bece7\") " pod="openstack/octavia-healthmanager-hhkfh" Oct 08 19:53:27 crc kubenswrapper[4988]: I1008 19:53:27.333452 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d7a04c8d-1f4c-4dd8-98b6-7afdc47bece7-scripts\") pod \"octavia-healthmanager-hhkfh\" (UID: \"d7a04c8d-1f4c-4dd8-98b6-7afdc47bece7\") " pod="openstack/octavia-healthmanager-hhkfh" Oct 08 19:53:27 crc kubenswrapper[4988]: I1008 19:53:27.333750 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hm-ports\" (UniqueName: \"kubernetes.io/configmap/d7a04c8d-1f4c-4dd8-98b6-7afdc47bece7-hm-ports\") pod \"octavia-healthmanager-hhkfh\" (UID: \"d7a04c8d-1f4c-4dd8-98b6-7afdc47bece7\") " pod="openstack/octavia-healthmanager-hhkfh" Oct 08 19:53:27 crc kubenswrapper[4988]: I1008 19:53:27.333799 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d7a04c8d-1f4c-4dd8-98b6-7afdc47bece7-config-data\") pod \"octavia-healthmanager-hhkfh\" (UID: \"d7a04c8d-1f4c-4dd8-98b6-7afdc47bece7\") " pod="openstack/octavia-healthmanager-hhkfh" Oct 08 19:53:27 crc kubenswrapper[4988]: I1008 19:53:27.333872 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"amphora-certs\" (UniqueName: \"kubernetes.io/secret/d7a04c8d-1f4c-4dd8-98b6-7afdc47bece7-amphora-certs\") pod \"octavia-healthmanager-hhkfh\" (UID: \"d7a04c8d-1f4c-4dd8-98b6-7afdc47bece7\") " pod="openstack/octavia-healthmanager-hhkfh" Oct 08 19:53:27 crc kubenswrapper[4988]: I1008 19:53:27.333920 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d7a04c8d-1f4c-4dd8-98b6-7afdc47bece7-combined-ca-bundle\") pod \"octavia-healthmanager-hhkfh\" (UID: \"d7a04c8d-1f4c-4dd8-98b6-7afdc47bece7\") " pod="openstack/octavia-healthmanager-hhkfh" Oct 08 19:53:27 crc kubenswrapper[4988]: I1008 19:53:27.334676 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/d7a04c8d-1f4c-4dd8-98b6-7afdc47bece7-config-data-merged\") pod \"octavia-healthmanager-hhkfh\" (UID: \"d7a04c8d-1f4c-4dd8-98b6-7afdc47bece7\") " pod="openstack/octavia-healthmanager-hhkfh" Oct 08 19:53:27 crc kubenswrapper[4988]: I1008 19:53:27.339171 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hm-ports\" (UniqueName: \"kubernetes.io/configmap/d7a04c8d-1f4c-4dd8-98b6-7afdc47bece7-hm-ports\") pod \"octavia-healthmanager-hhkfh\" (UID: \"d7a04c8d-1f4c-4dd8-98b6-7afdc47bece7\") " pod="openstack/octavia-healthmanager-hhkfh" Oct 08 19:53:27 crc kubenswrapper[4988]: I1008 19:53:27.345021 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d7a04c8d-1f4c-4dd8-98b6-7afdc47bece7-combined-ca-bundle\") pod \"octavia-healthmanager-hhkfh\" (UID: \"d7a04c8d-1f4c-4dd8-98b6-7afdc47bece7\") " pod="openstack/octavia-healthmanager-hhkfh" Oct 08 19:53:27 crc kubenswrapper[4988]: I1008 19:53:27.345483 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d7a04c8d-1f4c-4dd8-98b6-7afdc47bece7-config-data\") pod \"octavia-healthmanager-hhkfh\" (UID: \"d7a04c8d-1f4c-4dd8-98b6-7afdc47bece7\") " pod="openstack/octavia-healthmanager-hhkfh" Oct 08 19:53:27 crc kubenswrapper[4988]: I1008 19:53:27.346159 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d7a04c8d-1f4c-4dd8-98b6-7afdc47bece7-scripts\") pod \"octavia-healthmanager-hhkfh\" (UID: \"d7a04c8d-1f4c-4dd8-98b6-7afdc47bece7\") " pod="openstack/octavia-healthmanager-hhkfh" Oct 08 19:53:27 crc kubenswrapper[4988]: I1008 19:53:27.346655 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"amphora-certs\" (UniqueName: \"kubernetes.io/secret/d7a04c8d-1f4c-4dd8-98b6-7afdc47bece7-amphora-certs\") pod \"octavia-healthmanager-hhkfh\" (UID: \"d7a04c8d-1f4c-4dd8-98b6-7afdc47bece7\") " pod="openstack/octavia-healthmanager-hhkfh" Oct 08 19:53:27 crc kubenswrapper[4988]: I1008 19:53:27.468032 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-healthmanager-hhkfh" Oct 08 19:53:28 crc kubenswrapper[4988]: I1008 19:53:28.095709 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-healthmanager-hhkfh"] Oct 08 19:53:28 crc kubenswrapper[4988]: I1008 19:53:28.248240 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/octavia-housekeeping-zwdsf"] Oct 08 19:53:28 crc kubenswrapper[4988]: I1008 19:53:28.252011 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-housekeeping-zwdsf" Oct 08 19:53:28 crc kubenswrapper[4988]: I1008 19:53:28.256298 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"octavia-housekeeping-scripts" Oct 08 19:53:28 crc kubenswrapper[4988]: I1008 19:53:28.260922 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"octavia-housekeeping-config-data" Oct 08 19:53:28 crc kubenswrapper[4988]: I1008 19:53:28.281007 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-housekeeping-zwdsf"] Oct 08 19:53:28 crc kubenswrapper[4988]: I1008 19:53:28.358295 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hm-ports\" (UniqueName: \"kubernetes.io/configmap/19ef63af-3cf2-40c9-9002-3a5536d518b9-hm-ports\") pod \"octavia-housekeeping-zwdsf\" (UID: \"19ef63af-3cf2-40c9-9002-3a5536d518b9\") " pod="openstack/octavia-housekeeping-zwdsf" Oct 08 19:53:28 crc kubenswrapper[4988]: I1008 19:53:28.358380 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"amphora-certs\" (UniqueName: \"kubernetes.io/secret/19ef63af-3cf2-40c9-9002-3a5536d518b9-amphora-certs\") pod \"octavia-housekeeping-zwdsf\" (UID: \"19ef63af-3cf2-40c9-9002-3a5536d518b9\") " pod="openstack/octavia-housekeeping-zwdsf" Oct 08 19:53:28 crc kubenswrapper[4988]: I1008 19:53:28.358513 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/19ef63af-3cf2-40c9-9002-3a5536d518b9-combined-ca-bundle\") pod \"octavia-housekeeping-zwdsf\" (UID: \"19ef63af-3cf2-40c9-9002-3a5536d518b9\") " pod="openstack/octavia-housekeeping-zwdsf" Oct 08 19:53:28 crc kubenswrapper[4988]: I1008 19:53:28.358561 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/19ef63af-3cf2-40c9-9002-3a5536d518b9-config-data-merged\") pod \"octavia-housekeeping-zwdsf\" (UID: \"19ef63af-3cf2-40c9-9002-3a5536d518b9\") " pod="openstack/octavia-housekeeping-zwdsf" Oct 08 19:53:28 crc kubenswrapper[4988]: I1008 19:53:28.358634 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/19ef63af-3cf2-40c9-9002-3a5536d518b9-scripts\") pod \"octavia-housekeeping-zwdsf\" (UID: \"19ef63af-3cf2-40c9-9002-3a5536d518b9\") " pod="openstack/octavia-housekeeping-zwdsf" Oct 08 19:53:28 crc kubenswrapper[4988]: I1008 19:53:28.358650 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/19ef63af-3cf2-40c9-9002-3a5536d518b9-config-data\") pod \"octavia-housekeeping-zwdsf\" (UID: \"19ef63af-3cf2-40c9-9002-3a5536d518b9\") " pod="openstack/octavia-housekeeping-zwdsf" Oct 08 19:53:28 crc kubenswrapper[4988]: I1008 19:53:28.460679 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hm-ports\" (UniqueName: \"kubernetes.io/configmap/19ef63af-3cf2-40c9-9002-3a5536d518b9-hm-ports\") pod \"octavia-housekeeping-zwdsf\" (UID: \"19ef63af-3cf2-40c9-9002-3a5536d518b9\") " pod="openstack/octavia-housekeeping-zwdsf" Oct 08 19:53:28 crc kubenswrapper[4988]: I1008 19:53:28.461289 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"amphora-certs\" (UniqueName: \"kubernetes.io/secret/19ef63af-3cf2-40c9-9002-3a5536d518b9-amphora-certs\") pod \"octavia-housekeeping-zwdsf\" (UID: \"19ef63af-3cf2-40c9-9002-3a5536d518b9\") " pod="openstack/octavia-housekeeping-zwdsf" Oct 08 19:53:28 crc kubenswrapper[4988]: I1008 19:53:28.461353 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/19ef63af-3cf2-40c9-9002-3a5536d518b9-combined-ca-bundle\") pod \"octavia-housekeeping-zwdsf\" (UID: \"19ef63af-3cf2-40c9-9002-3a5536d518b9\") " pod="openstack/octavia-housekeeping-zwdsf" Oct 08 19:53:28 crc kubenswrapper[4988]: I1008 19:53:28.461403 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/19ef63af-3cf2-40c9-9002-3a5536d518b9-config-data-merged\") pod \"octavia-housekeeping-zwdsf\" (UID: \"19ef63af-3cf2-40c9-9002-3a5536d518b9\") " pod="openstack/octavia-housekeeping-zwdsf" Oct 08 19:53:28 crc kubenswrapper[4988]: I1008 19:53:28.461476 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/19ef63af-3cf2-40c9-9002-3a5536d518b9-scripts\") pod \"octavia-housekeeping-zwdsf\" (UID: \"19ef63af-3cf2-40c9-9002-3a5536d518b9\") " pod="openstack/octavia-housekeeping-zwdsf" Oct 08 19:53:28 crc kubenswrapper[4988]: I1008 19:53:28.461502 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/19ef63af-3cf2-40c9-9002-3a5536d518b9-config-data\") pod \"octavia-housekeeping-zwdsf\" (UID: \"19ef63af-3cf2-40c9-9002-3a5536d518b9\") " pod="openstack/octavia-housekeeping-zwdsf" Oct 08 19:53:28 crc kubenswrapper[4988]: I1008 19:53:28.461877 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hm-ports\" (UniqueName: \"kubernetes.io/configmap/19ef63af-3cf2-40c9-9002-3a5536d518b9-hm-ports\") pod \"octavia-housekeeping-zwdsf\" (UID: \"19ef63af-3cf2-40c9-9002-3a5536d518b9\") " pod="openstack/octavia-housekeeping-zwdsf" Oct 08 19:53:28 crc kubenswrapper[4988]: I1008 19:53:28.462293 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/19ef63af-3cf2-40c9-9002-3a5536d518b9-config-data-merged\") pod \"octavia-housekeeping-zwdsf\" (UID: \"19ef63af-3cf2-40c9-9002-3a5536d518b9\") " pod="openstack/octavia-housekeeping-zwdsf" Oct 08 19:53:28 crc kubenswrapper[4988]: I1008 19:53:28.469786 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/19ef63af-3cf2-40c9-9002-3a5536d518b9-config-data\") pod \"octavia-housekeeping-zwdsf\" (UID: \"19ef63af-3cf2-40c9-9002-3a5536d518b9\") " pod="openstack/octavia-housekeeping-zwdsf" Oct 08 19:53:28 crc kubenswrapper[4988]: I1008 19:53:28.477115 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"amphora-certs\" (UniqueName: \"kubernetes.io/secret/19ef63af-3cf2-40c9-9002-3a5536d518b9-amphora-certs\") pod \"octavia-housekeeping-zwdsf\" (UID: \"19ef63af-3cf2-40c9-9002-3a5536d518b9\") " pod="openstack/octavia-housekeeping-zwdsf" Oct 08 19:53:28 crc kubenswrapper[4988]: I1008 19:53:28.478061 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/19ef63af-3cf2-40c9-9002-3a5536d518b9-scripts\") pod \"octavia-housekeeping-zwdsf\" (UID: \"19ef63af-3cf2-40c9-9002-3a5536d518b9\") " pod="openstack/octavia-housekeeping-zwdsf" Oct 08 19:53:28 crc kubenswrapper[4988]: I1008 19:53:28.481129 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/19ef63af-3cf2-40c9-9002-3a5536d518b9-combined-ca-bundle\") pod \"octavia-housekeeping-zwdsf\" (UID: \"19ef63af-3cf2-40c9-9002-3a5536d518b9\") " pod="openstack/octavia-housekeeping-zwdsf" Oct 08 19:53:28 crc kubenswrapper[4988]: I1008 19:53:28.572453 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-healthmanager-hhkfh" event={"ID":"d7a04c8d-1f4c-4dd8-98b6-7afdc47bece7","Type":"ContainerStarted","Data":"3cf2e29ba2b4a0dad0ac675b6bdd6cba61204fcfa948e959de172a107e2a3de2"} Oct 08 19:53:28 crc kubenswrapper[4988]: I1008 19:53:28.576470 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-housekeeping-zwdsf" Oct 08 19:53:29 crc kubenswrapper[4988]: I1008 19:53:29.182811 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-housekeeping-zwdsf"] Oct 08 19:53:29 crc kubenswrapper[4988]: I1008 19:53:29.583858 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-housekeeping-zwdsf" event={"ID":"19ef63af-3cf2-40c9-9002-3a5536d518b9","Type":"ContainerStarted","Data":"3cc0d98cd982a3d51e7689a40d11d845752b96c54018ee5af9b8040bacb25fe3"} Oct 08 19:53:29 crc kubenswrapper[4988]: I1008 19:53:29.585901 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-healthmanager-hhkfh" event={"ID":"d7a04c8d-1f4c-4dd8-98b6-7afdc47bece7","Type":"ContainerStarted","Data":"8daac4cef954f7bbdd3100c6cf19ed91b0e168bb85d8dc2c0db5dafdc5f3be31"} Oct 08 19:53:30 crc kubenswrapper[4988]: I1008 19:53:30.081090 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/octavia-worker-rtcdp"] Oct 08 19:53:30 crc kubenswrapper[4988]: I1008 19:53:30.084213 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-worker-rtcdp" Oct 08 19:53:30 crc kubenswrapper[4988]: I1008 19:53:30.086744 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"octavia-worker-config-data" Oct 08 19:53:30 crc kubenswrapper[4988]: I1008 19:53:30.087222 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"octavia-worker-scripts" Oct 08 19:53:30 crc kubenswrapper[4988]: I1008 19:53:30.090402 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-worker-rtcdp"] Oct 08 19:53:30 crc kubenswrapper[4988]: I1008 19:53:30.093159 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/baa5ab97-2a79-4125-b1f8-5f0e7af21c6e-config-data-merged\") pod \"octavia-worker-rtcdp\" (UID: \"baa5ab97-2a79-4125-b1f8-5f0e7af21c6e\") " pod="openstack/octavia-worker-rtcdp" Oct 08 19:53:30 crc kubenswrapper[4988]: I1008 19:53:30.093618 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"amphora-certs\" (UniqueName: \"kubernetes.io/secret/baa5ab97-2a79-4125-b1f8-5f0e7af21c6e-amphora-certs\") pod \"octavia-worker-rtcdp\" (UID: \"baa5ab97-2a79-4125-b1f8-5f0e7af21c6e\") " pod="openstack/octavia-worker-rtcdp" Oct 08 19:53:30 crc kubenswrapper[4988]: I1008 19:53:30.093695 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/baa5ab97-2a79-4125-b1f8-5f0e7af21c6e-scripts\") pod \"octavia-worker-rtcdp\" (UID: \"baa5ab97-2a79-4125-b1f8-5f0e7af21c6e\") " pod="openstack/octavia-worker-rtcdp" Oct 08 19:53:30 crc kubenswrapper[4988]: I1008 19:53:30.093891 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hm-ports\" (UniqueName: \"kubernetes.io/configmap/baa5ab97-2a79-4125-b1f8-5f0e7af21c6e-hm-ports\") pod \"octavia-worker-rtcdp\" (UID: \"baa5ab97-2a79-4125-b1f8-5f0e7af21c6e\") " pod="openstack/octavia-worker-rtcdp" Oct 08 19:53:30 crc kubenswrapper[4988]: I1008 19:53:30.094137 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/baa5ab97-2a79-4125-b1f8-5f0e7af21c6e-combined-ca-bundle\") pod \"octavia-worker-rtcdp\" (UID: \"baa5ab97-2a79-4125-b1f8-5f0e7af21c6e\") " pod="openstack/octavia-worker-rtcdp" Oct 08 19:53:30 crc kubenswrapper[4988]: I1008 19:53:30.094221 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/baa5ab97-2a79-4125-b1f8-5f0e7af21c6e-config-data\") pod \"octavia-worker-rtcdp\" (UID: \"baa5ab97-2a79-4125-b1f8-5f0e7af21c6e\") " pod="openstack/octavia-worker-rtcdp" Oct 08 19:53:30 crc kubenswrapper[4988]: I1008 19:53:30.196087 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/baa5ab97-2a79-4125-b1f8-5f0e7af21c6e-config-data\") pod \"octavia-worker-rtcdp\" (UID: \"baa5ab97-2a79-4125-b1f8-5f0e7af21c6e\") " pod="openstack/octavia-worker-rtcdp" Oct 08 19:53:30 crc kubenswrapper[4988]: I1008 19:53:30.196150 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/baa5ab97-2a79-4125-b1f8-5f0e7af21c6e-config-data-merged\") pod \"octavia-worker-rtcdp\" (UID: \"baa5ab97-2a79-4125-b1f8-5f0e7af21c6e\") " pod="openstack/octavia-worker-rtcdp" Oct 08 19:53:30 crc kubenswrapper[4988]: I1008 19:53:30.196187 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"amphora-certs\" (UniqueName: \"kubernetes.io/secret/baa5ab97-2a79-4125-b1f8-5f0e7af21c6e-amphora-certs\") pod \"octavia-worker-rtcdp\" (UID: \"baa5ab97-2a79-4125-b1f8-5f0e7af21c6e\") " pod="openstack/octavia-worker-rtcdp" Oct 08 19:53:30 crc kubenswrapper[4988]: I1008 19:53:30.196207 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/baa5ab97-2a79-4125-b1f8-5f0e7af21c6e-scripts\") pod \"octavia-worker-rtcdp\" (UID: \"baa5ab97-2a79-4125-b1f8-5f0e7af21c6e\") " pod="openstack/octavia-worker-rtcdp" Oct 08 19:53:30 crc kubenswrapper[4988]: I1008 19:53:30.196264 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hm-ports\" (UniqueName: \"kubernetes.io/configmap/baa5ab97-2a79-4125-b1f8-5f0e7af21c6e-hm-ports\") pod \"octavia-worker-rtcdp\" (UID: \"baa5ab97-2a79-4125-b1f8-5f0e7af21c6e\") " pod="openstack/octavia-worker-rtcdp" Oct 08 19:53:30 crc kubenswrapper[4988]: I1008 19:53:30.196335 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/baa5ab97-2a79-4125-b1f8-5f0e7af21c6e-combined-ca-bundle\") pod \"octavia-worker-rtcdp\" (UID: \"baa5ab97-2a79-4125-b1f8-5f0e7af21c6e\") " pod="openstack/octavia-worker-rtcdp" Oct 08 19:53:30 crc kubenswrapper[4988]: I1008 19:53:30.196640 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/baa5ab97-2a79-4125-b1f8-5f0e7af21c6e-config-data-merged\") pod \"octavia-worker-rtcdp\" (UID: \"baa5ab97-2a79-4125-b1f8-5f0e7af21c6e\") " pod="openstack/octavia-worker-rtcdp" Oct 08 19:53:30 crc kubenswrapper[4988]: I1008 19:53:30.197596 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hm-ports\" (UniqueName: \"kubernetes.io/configmap/baa5ab97-2a79-4125-b1f8-5f0e7af21c6e-hm-ports\") pod \"octavia-worker-rtcdp\" (UID: \"baa5ab97-2a79-4125-b1f8-5f0e7af21c6e\") " pod="openstack/octavia-worker-rtcdp" Oct 08 19:53:30 crc kubenswrapper[4988]: I1008 19:53:30.202768 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/baa5ab97-2a79-4125-b1f8-5f0e7af21c6e-scripts\") pod \"octavia-worker-rtcdp\" (UID: \"baa5ab97-2a79-4125-b1f8-5f0e7af21c6e\") " pod="openstack/octavia-worker-rtcdp" Oct 08 19:53:30 crc kubenswrapper[4988]: I1008 19:53:30.205280 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/baa5ab97-2a79-4125-b1f8-5f0e7af21c6e-config-data\") pod \"octavia-worker-rtcdp\" (UID: \"baa5ab97-2a79-4125-b1f8-5f0e7af21c6e\") " pod="openstack/octavia-worker-rtcdp" Oct 08 19:53:30 crc kubenswrapper[4988]: I1008 19:53:30.206128 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"amphora-certs\" (UniqueName: \"kubernetes.io/secret/baa5ab97-2a79-4125-b1f8-5f0e7af21c6e-amphora-certs\") pod \"octavia-worker-rtcdp\" (UID: \"baa5ab97-2a79-4125-b1f8-5f0e7af21c6e\") " pod="openstack/octavia-worker-rtcdp" Oct 08 19:53:30 crc kubenswrapper[4988]: I1008 19:53:30.221757 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/baa5ab97-2a79-4125-b1f8-5f0e7af21c6e-combined-ca-bundle\") pod \"octavia-worker-rtcdp\" (UID: \"baa5ab97-2a79-4125-b1f8-5f0e7af21c6e\") " pod="openstack/octavia-worker-rtcdp" Oct 08 19:53:30 crc kubenswrapper[4988]: I1008 19:53:30.420738 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-worker-rtcdp" Oct 08 19:53:30 crc kubenswrapper[4988]: I1008 19:53:30.606741 4988 generic.go:334] "Generic (PLEG): container finished" podID="d7a04c8d-1f4c-4dd8-98b6-7afdc47bece7" containerID="8daac4cef954f7bbdd3100c6cf19ed91b0e168bb85d8dc2c0db5dafdc5f3be31" exitCode=0 Oct 08 19:53:30 crc kubenswrapper[4988]: I1008 19:53:30.606795 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-healthmanager-hhkfh" event={"ID":"d7a04c8d-1f4c-4dd8-98b6-7afdc47bece7","Type":"ContainerDied","Data":"8daac4cef954f7bbdd3100c6cf19ed91b0e168bb85d8dc2c0db5dafdc5f3be31"} Oct 08 19:53:30 crc kubenswrapper[4988]: I1008 19:53:30.985691 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-worker-rtcdp"] Oct 08 19:53:31 crc kubenswrapper[4988]: I1008 19:53:31.621494 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-healthmanager-hhkfh" event={"ID":"d7a04c8d-1f4c-4dd8-98b6-7afdc47bece7","Type":"ContainerStarted","Data":"929b34f46a3eb1ee1ddb45686f01a4b613a69bc79777ec6379ecbcb7652b589c"} Oct 08 19:53:31 crc kubenswrapper[4988]: I1008 19:53:31.621848 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/octavia-healthmanager-hhkfh" Oct 08 19:53:31 crc kubenswrapper[4988]: I1008 19:53:31.624721 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-housekeeping-zwdsf" event={"ID":"19ef63af-3cf2-40c9-9002-3a5536d518b9","Type":"ContainerStarted","Data":"354b45463325568c831edede27eb114923a78effb27ed0906e3e03f4489336c5"} Oct 08 19:53:31 crc kubenswrapper[4988]: I1008 19:53:31.634963 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-worker-rtcdp" event={"ID":"baa5ab97-2a79-4125-b1f8-5f0e7af21c6e","Type":"ContainerStarted","Data":"e961e995f7ae4a082db4c5967f19cf7709851f6b9259935eef733d0460b09b22"} Oct 08 19:53:31 crc kubenswrapper[4988]: I1008 19:53:31.645845 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/octavia-healthmanager-hhkfh" podStartSLOduration=4.645826435 podStartE2EDuration="4.645826435s" podCreationTimestamp="2025-10-08 19:53:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 19:53:31.643586513 +0000 UTC m=+6157.093429293" watchObservedRunningTime="2025-10-08 19:53:31.645826435 +0000 UTC m=+6157.095669215" Oct 08 19:53:32 crc kubenswrapper[4988]: I1008 19:53:32.649932 4988 generic.go:334] "Generic (PLEG): container finished" podID="19ef63af-3cf2-40c9-9002-3a5536d518b9" containerID="354b45463325568c831edede27eb114923a78effb27ed0906e3e03f4489336c5" exitCode=0 Oct 08 19:53:32 crc kubenswrapper[4988]: I1008 19:53:32.650179 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-housekeeping-zwdsf" event={"ID":"19ef63af-3cf2-40c9-9002-3a5536d518b9","Type":"ContainerDied","Data":"354b45463325568c831edede27eb114923a78effb27ed0906e3e03f4489336c5"} Oct 08 19:53:32 crc kubenswrapper[4988]: I1008 19:53:32.651024 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-housekeeping-zwdsf" event={"ID":"19ef63af-3cf2-40c9-9002-3a5536d518b9","Type":"ContainerStarted","Data":"d74faa611a20327b51cfbc659613d3456b33f78de6ad17bf1647cd390d8d9c83"} Oct 08 19:53:32 crc kubenswrapper[4988]: I1008 19:53:32.651043 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/octavia-housekeeping-zwdsf" Oct 08 19:53:32 crc kubenswrapper[4988]: I1008 19:53:32.675033 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/octavia-housekeeping-zwdsf" podStartSLOduration=3.522630322 podStartE2EDuration="4.675015758s" podCreationTimestamp="2025-10-08 19:53:28 +0000 UTC" firstStartedPulling="2025-10-08 19:53:29.190128274 +0000 UTC m=+6154.639971074" lastFinishedPulling="2025-10-08 19:53:30.34251374 +0000 UTC m=+6155.792356510" observedRunningTime="2025-10-08 19:53:32.668697666 +0000 UTC m=+6158.118540456" watchObservedRunningTime="2025-10-08 19:53:32.675015758 +0000 UTC m=+6158.124858528" Oct 08 19:53:33 crc kubenswrapper[4988]: I1008 19:53:33.661305 4988 generic.go:334] "Generic (PLEG): container finished" podID="baa5ab97-2a79-4125-b1f8-5f0e7af21c6e" containerID="8ff5ae7daf06a2bf7740f0fca88387c6258fae8e2e8a04e1945acf63b685a7c5" exitCode=0 Oct 08 19:53:33 crc kubenswrapper[4988]: I1008 19:53:33.661415 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-worker-rtcdp" event={"ID":"baa5ab97-2a79-4125-b1f8-5f0e7af21c6e","Type":"ContainerDied","Data":"8ff5ae7daf06a2bf7740f0fca88387c6258fae8e2e8a04e1945acf63b685a7c5"} Oct 08 19:53:34 crc kubenswrapper[4988]: I1008 19:53:34.687330 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-worker-rtcdp" event={"ID":"baa5ab97-2a79-4125-b1f8-5f0e7af21c6e","Type":"ContainerStarted","Data":"e8b50722e6656ea0479689003677d6b0de75c50ac3048752072c7eae5a76c8c5"} Oct 08 19:53:34 crc kubenswrapper[4988]: I1008 19:53:34.688131 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/octavia-worker-rtcdp" Oct 08 19:53:34 crc kubenswrapper[4988]: I1008 19:53:34.724159 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/octavia-worker-rtcdp" podStartSLOduration=3.32996701 podStartE2EDuration="4.724137799s" podCreationTimestamp="2025-10-08 19:53:30 +0000 UTC" firstStartedPulling="2025-10-08 19:53:30.99360797 +0000 UTC m=+6156.443450740" lastFinishedPulling="2025-10-08 19:53:32.387778759 +0000 UTC m=+6157.837621529" observedRunningTime="2025-10-08 19:53:34.715272906 +0000 UTC m=+6160.165115676" watchObservedRunningTime="2025-10-08 19:53:34.724137799 +0000 UTC m=+6160.173980579" Oct 08 19:53:36 crc kubenswrapper[4988]: I1008 19:53:36.025835 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-1c50-account-create-28g6x"] Oct 08 19:53:36 crc kubenswrapper[4988]: I1008 19:53:36.033774 4988 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-1c50-account-create-28g6x"] Oct 08 19:53:37 crc kubenswrapper[4988]: I1008 19:53:37.252854 4988 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3448b939-1d0f-4d1e-89f5-4a0786c4c82e" path="/var/lib/kubelet/pods/3448b939-1d0f-4d1e-89f5-4a0786c4c82e/volumes" Oct 08 19:53:42 crc kubenswrapper[4988]: I1008 19:53:42.033453 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-sync-7z69t"] Oct 08 19:53:42 crc kubenswrapper[4988]: I1008 19:53:42.043187 4988 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-sync-7z69t"] Oct 08 19:53:42 crc kubenswrapper[4988]: I1008 19:53:42.501251 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/octavia-healthmanager-hhkfh" Oct 08 19:53:43 crc kubenswrapper[4988]: I1008 19:53:43.252259 4988 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="02e9ed2c-31bf-4775-ab12-a9fafc13e68e" path="/var/lib/kubelet/pods/02e9ed2c-31bf-4775-ab12-a9fafc13e68e/volumes" Oct 08 19:53:43 crc kubenswrapper[4988]: I1008 19:53:43.622737 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/octavia-housekeeping-zwdsf" Oct 08 19:53:45 crc kubenswrapper[4988]: I1008 19:53:45.477282 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/octavia-worker-rtcdp" Oct 08 19:53:53 crc kubenswrapper[4988]: I1008 19:53:53.337694 4988 patch_prober.go:28] interesting pod/machine-config-daemon-nm54f container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 19:53:53 crc kubenswrapper[4988]: I1008 19:53:53.338266 4988 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 19:54:03 crc kubenswrapper[4988]: E1008 19:54:03.368550 4988 upgradeaware.go:427] Error proxying data from client to backend: readfrom tcp 38.102.83.102:58814->38.102.83.102:38495: write tcp 38.102.83.102:58814->38.102.83.102:38495: write: broken pipe Oct 08 19:54:10 crc kubenswrapper[4988]: I1008 19:54:10.574148 4988 scope.go:117] "RemoveContainer" containerID="d2e25e9a3e3089bde074997b30b906f7253eac35bf9c155ddd64bbb92eac324e" Oct 08 19:54:10 crc kubenswrapper[4988]: I1008 19:54:10.610076 4988 scope.go:117] "RemoveContainer" containerID="08a3fd1a9c76b2dbcf7c8ee9804f4b7b48f3c54e14df3fcddc2085315cb58fa9" Oct 08 19:54:10 crc kubenswrapper[4988]: I1008 19:54:10.697449 4988 scope.go:117] "RemoveContainer" containerID="0a5eba2345e58863ef2eebcec4b06b7eb809e54141dd50b136e9c8f229fd7938" Oct 08 19:54:15 crc kubenswrapper[4988]: I1008 19:54:15.804402 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-2jm7s"] Oct 08 19:54:15 crc kubenswrapper[4988]: I1008 19:54:15.806915 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-2jm7s" Oct 08 19:54:15 crc kubenswrapper[4988]: I1008 19:54:15.838785 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-2jm7s"] Oct 08 19:54:15 crc kubenswrapper[4988]: I1008 19:54:15.882290 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dd00ff3f-0c19-43d8-8a05-80ad79711731-catalog-content\") pod \"community-operators-2jm7s\" (UID: \"dd00ff3f-0c19-43d8-8a05-80ad79711731\") " pod="openshift-marketplace/community-operators-2jm7s" Oct 08 19:54:15 crc kubenswrapper[4988]: I1008 19:54:15.882348 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dd00ff3f-0c19-43d8-8a05-80ad79711731-utilities\") pod \"community-operators-2jm7s\" (UID: \"dd00ff3f-0c19-43d8-8a05-80ad79711731\") " pod="openshift-marketplace/community-operators-2jm7s" Oct 08 19:54:15 crc kubenswrapper[4988]: I1008 19:54:15.882444 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fgmsl\" (UniqueName: \"kubernetes.io/projected/dd00ff3f-0c19-43d8-8a05-80ad79711731-kube-api-access-fgmsl\") pod \"community-operators-2jm7s\" (UID: \"dd00ff3f-0c19-43d8-8a05-80ad79711731\") " pod="openshift-marketplace/community-operators-2jm7s" Oct 08 19:54:15 crc kubenswrapper[4988]: I1008 19:54:15.983674 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dd00ff3f-0c19-43d8-8a05-80ad79711731-catalog-content\") pod \"community-operators-2jm7s\" (UID: \"dd00ff3f-0c19-43d8-8a05-80ad79711731\") " pod="openshift-marketplace/community-operators-2jm7s" Oct 08 19:54:15 crc kubenswrapper[4988]: I1008 19:54:15.983718 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dd00ff3f-0c19-43d8-8a05-80ad79711731-utilities\") pod \"community-operators-2jm7s\" (UID: \"dd00ff3f-0c19-43d8-8a05-80ad79711731\") " pod="openshift-marketplace/community-operators-2jm7s" Oct 08 19:54:15 crc kubenswrapper[4988]: I1008 19:54:15.983757 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fgmsl\" (UniqueName: \"kubernetes.io/projected/dd00ff3f-0c19-43d8-8a05-80ad79711731-kube-api-access-fgmsl\") pod \"community-operators-2jm7s\" (UID: \"dd00ff3f-0c19-43d8-8a05-80ad79711731\") " pod="openshift-marketplace/community-operators-2jm7s" Oct 08 19:54:15 crc kubenswrapper[4988]: I1008 19:54:15.984477 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dd00ff3f-0c19-43d8-8a05-80ad79711731-catalog-content\") pod \"community-operators-2jm7s\" (UID: \"dd00ff3f-0c19-43d8-8a05-80ad79711731\") " pod="openshift-marketplace/community-operators-2jm7s" Oct 08 19:54:15 crc kubenswrapper[4988]: I1008 19:54:15.984581 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dd00ff3f-0c19-43d8-8a05-80ad79711731-utilities\") pod \"community-operators-2jm7s\" (UID: \"dd00ff3f-0c19-43d8-8a05-80ad79711731\") " pod="openshift-marketplace/community-operators-2jm7s" Oct 08 19:54:16 crc kubenswrapper[4988]: I1008 19:54:16.020937 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fgmsl\" (UniqueName: \"kubernetes.io/projected/dd00ff3f-0c19-43d8-8a05-80ad79711731-kube-api-access-fgmsl\") pod \"community-operators-2jm7s\" (UID: \"dd00ff3f-0c19-43d8-8a05-80ad79711731\") " pod="openshift-marketplace/community-operators-2jm7s" Oct 08 19:54:16 crc kubenswrapper[4988]: I1008 19:54:16.126486 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-2jm7s" Oct 08 19:54:16 crc kubenswrapper[4988]: I1008 19:54:16.708414 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-2jm7s"] Oct 08 19:54:17 crc kubenswrapper[4988]: I1008 19:54:17.187603 4988 generic.go:334] "Generic (PLEG): container finished" podID="dd00ff3f-0c19-43d8-8a05-80ad79711731" containerID="7d091e741f2608b8c6e8449e8f0ea55a2424a323ff499a8a7ba86aa621893690" exitCode=0 Oct 08 19:54:17 crc kubenswrapper[4988]: I1008 19:54:17.187670 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-2jm7s" event={"ID":"dd00ff3f-0c19-43d8-8a05-80ad79711731","Type":"ContainerDied","Data":"7d091e741f2608b8c6e8449e8f0ea55a2424a323ff499a8a7ba86aa621893690"} Oct 08 19:54:17 crc kubenswrapper[4988]: I1008 19:54:17.187709 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-2jm7s" event={"ID":"dd00ff3f-0c19-43d8-8a05-80ad79711731","Type":"ContainerStarted","Data":"2d54bfe4bcb2083f17b73390fd3a2543df44f47bb6d809d582e7c70307862330"} Oct 08 19:54:19 crc kubenswrapper[4988]: I1008 19:54:19.214471 4988 generic.go:334] "Generic (PLEG): container finished" podID="dd00ff3f-0c19-43d8-8a05-80ad79711731" containerID="18a74433931c8691f496240afa0d5ef264cfe8235cb580225b48b2fa68150461" exitCode=0 Oct 08 19:54:19 crc kubenswrapper[4988]: I1008 19:54:19.214589 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-2jm7s" event={"ID":"dd00ff3f-0c19-43d8-8a05-80ad79711731","Type":"ContainerDied","Data":"18a74433931c8691f496240afa0d5ef264cfe8235cb580225b48b2fa68150461"} Oct 08 19:54:20 crc kubenswrapper[4988]: I1008 19:54:20.235604 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-2jm7s" event={"ID":"dd00ff3f-0c19-43d8-8a05-80ad79711731","Type":"ContainerStarted","Data":"da1342de8829768d18b20476c8cce90f2ec9a1c19cc5a7da3f94bd198a60c3c0"} Oct 08 19:54:20 crc kubenswrapper[4988]: I1008 19:54:20.264221 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-2jm7s" podStartSLOduration=2.791072653 podStartE2EDuration="5.264202801s" podCreationTimestamp="2025-10-08 19:54:15 +0000 UTC" firstStartedPulling="2025-10-08 19:54:17.191217427 +0000 UTC m=+6202.641060217" lastFinishedPulling="2025-10-08 19:54:19.664347595 +0000 UTC m=+6205.114190365" observedRunningTime="2025-10-08 19:54:20.260051929 +0000 UTC m=+6205.709894699" watchObservedRunningTime="2025-10-08 19:54:20.264202801 +0000 UTC m=+6205.714045571" Oct 08 19:54:23 crc kubenswrapper[4988]: I1008 19:54:23.061361 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-create-t4zxj"] Oct 08 19:54:23 crc kubenswrapper[4988]: I1008 19:54:23.076261 4988 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-create-t4zxj"] Oct 08 19:54:23 crc kubenswrapper[4988]: I1008 19:54:23.255492 4988 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9dd78259-618d-4698-8f22-fa4ff6a75b75" path="/var/lib/kubelet/pods/9dd78259-618d-4698-8f22-fa4ff6a75b75/volumes" Oct 08 19:54:23 crc kubenswrapper[4988]: I1008 19:54:23.338334 4988 patch_prober.go:28] interesting pod/machine-config-daemon-nm54f container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 19:54:23 crc kubenswrapper[4988]: I1008 19:54:23.338410 4988 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 19:54:26 crc kubenswrapper[4988]: I1008 19:54:26.127714 4988 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-2jm7s" Oct 08 19:54:26 crc kubenswrapper[4988]: I1008 19:54:26.128229 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-2jm7s" Oct 08 19:54:26 crc kubenswrapper[4988]: I1008 19:54:26.193692 4988 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-2jm7s" Oct 08 19:54:26 crc kubenswrapper[4988]: I1008 19:54:26.349122 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-2jm7s" Oct 08 19:54:26 crc kubenswrapper[4988]: I1008 19:54:26.449365 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-2jm7s"] Oct 08 19:54:28 crc kubenswrapper[4988]: I1008 19:54:28.316137 4988 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-2jm7s" podUID="dd00ff3f-0c19-43d8-8a05-80ad79711731" containerName="registry-server" containerID="cri-o://da1342de8829768d18b20476c8cce90f2ec9a1c19cc5a7da3f94bd198a60c3c0" gracePeriod=2 Oct 08 19:54:28 crc kubenswrapper[4988]: I1008 19:54:28.888181 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-2jm7s" Oct 08 19:54:28 crc kubenswrapper[4988]: I1008 19:54:28.979710 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fgmsl\" (UniqueName: \"kubernetes.io/projected/dd00ff3f-0c19-43d8-8a05-80ad79711731-kube-api-access-fgmsl\") pod \"dd00ff3f-0c19-43d8-8a05-80ad79711731\" (UID: \"dd00ff3f-0c19-43d8-8a05-80ad79711731\") " Oct 08 19:54:28 crc kubenswrapper[4988]: I1008 19:54:28.979794 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dd00ff3f-0c19-43d8-8a05-80ad79711731-catalog-content\") pod \"dd00ff3f-0c19-43d8-8a05-80ad79711731\" (UID: \"dd00ff3f-0c19-43d8-8a05-80ad79711731\") " Oct 08 19:54:28 crc kubenswrapper[4988]: I1008 19:54:28.980087 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dd00ff3f-0c19-43d8-8a05-80ad79711731-utilities\") pod \"dd00ff3f-0c19-43d8-8a05-80ad79711731\" (UID: \"dd00ff3f-0c19-43d8-8a05-80ad79711731\") " Oct 08 19:54:28 crc kubenswrapper[4988]: I1008 19:54:28.980928 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/dd00ff3f-0c19-43d8-8a05-80ad79711731-utilities" (OuterVolumeSpecName: "utilities") pod "dd00ff3f-0c19-43d8-8a05-80ad79711731" (UID: "dd00ff3f-0c19-43d8-8a05-80ad79711731"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 19:54:28 crc kubenswrapper[4988]: I1008 19:54:28.997017 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dd00ff3f-0c19-43d8-8a05-80ad79711731-kube-api-access-fgmsl" (OuterVolumeSpecName: "kube-api-access-fgmsl") pod "dd00ff3f-0c19-43d8-8a05-80ad79711731" (UID: "dd00ff3f-0c19-43d8-8a05-80ad79711731"). InnerVolumeSpecName "kube-api-access-fgmsl". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 19:54:29 crc kubenswrapper[4988]: I1008 19:54:29.049791 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/dd00ff3f-0c19-43d8-8a05-80ad79711731-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "dd00ff3f-0c19-43d8-8a05-80ad79711731" (UID: "dd00ff3f-0c19-43d8-8a05-80ad79711731"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 19:54:29 crc kubenswrapper[4988]: I1008 19:54:29.082690 4988 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fgmsl\" (UniqueName: \"kubernetes.io/projected/dd00ff3f-0c19-43d8-8a05-80ad79711731-kube-api-access-fgmsl\") on node \"crc\" DevicePath \"\"" Oct 08 19:54:29 crc kubenswrapper[4988]: I1008 19:54:29.082729 4988 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dd00ff3f-0c19-43d8-8a05-80ad79711731-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 08 19:54:29 crc kubenswrapper[4988]: I1008 19:54:29.082738 4988 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dd00ff3f-0c19-43d8-8a05-80ad79711731-utilities\") on node \"crc\" DevicePath \"\"" Oct 08 19:54:29 crc kubenswrapper[4988]: I1008 19:54:29.331135 4988 generic.go:334] "Generic (PLEG): container finished" podID="dd00ff3f-0c19-43d8-8a05-80ad79711731" containerID="da1342de8829768d18b20476c8cce90f2ec9a1c19cc5a7da3f94bd198a60c3c0" exitCode=0 Oct 08 19:54:29 crc kubenswrapper[4988]: I1008 19:54:29.331202 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-2jm7s" event={"ID":"dd00ff3f-0c19-43d8-8a05-80ad79711731","Type":"ContainerDied","Data":"da1342de8829768d18b20476c8cce90f2ec9a1c19cc5a7da3f94bd198a60c3c0"} Oct 08 19:54:29 crc kubenswrapper[4988]: I1008 19:54:29.331265 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-2jm7s" Oct 08 19:54:29 crc kubenswrapper[4988]: I1008 19:54:29.331313 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-2jm7s" event={"ID":"dd00ff3f-0c19-43d8-8a05-80ad79711731","Type":"ContainerDied","Data":"2d54bfe4bcb2083f17b73390fd3a2543df44f47bb6d809d582e7c70307862330"} Oct 08 19:54:29 crc kubenswrapper[4988]: I1008 19:54:29.331349 4988 scope.go:117] "RemoveContainer" containerID="da1342de8829768d18b20476c8cce90f2ec9a1c19cc5a7da3f94bd198a60c3c0" Oct 08 19:54:29 crc kubenswrapper[4988]: I1008 19:54:29.369185 4988 scope.go:117] "RemoveContainer" containerID="18a74433931c8691f496240afa0d5ef264cfe8235cb580225b48b2fa68150461" Oct 08 19:54:29 crc kubenswrapper[4988]: I1008 19:54:29.371624 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-2jm7s"] Oct 08 19:54:29 crc kubenswrapper[4988]: I1008 19:54:29.386145 4988 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-2jm7s"] Oct 08 19:54:29 crc kubenswrapper[4988]: I1008 19:54:29.414921 4988 scope.go:117] "RemoveContainer" containerID="7d091e741f2608b8c6e8449e8f0ea55a2424a323ff499a8a7ba86aa621893690" Oct 08 19:54:29 crc kubenswrapper[4988]: I1008 19:54:29.459859 4988 scope.go:117] "RemoveContainer" containerID="da1342de8829768d18b20476c8cce90f2ec9a1c19cc5a7da3f94bd198a60c3c0" Oct 08 19:54:29 crc kubenswrapper[4988]: E1008 19:54:29.460444 4988 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"da1342de8829768d18b20476c8cce90f2ec9a1c19cc5a7da3f94bd198a60c3c0\": container with ID starting with da1342de8829768d18b20476c8cce90f2ec9a1c19cc5a7da3f94bd198a60c3c0 not found: ID does not exist" containerID="da1342de8829768d18b20476c8cce90f2ec9a1c19cc5a7da3f94bd198a60c3c0" Oct 08 19:54:29 crc kubenswrapper[4988]: I1008 19:54:29.460473 4988 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"da1342de8829768d18b20476c8cce90f2ec9a1c19cc5a7da3f94bd198a60c3c0"} err="failed to get container status \"da1342de8829768d18b20476c8cce90f2ec9a1c19cc5a7da3f94bd198a60c3c0\": rpc error: code = NotFound desc = could not find container \"da1342de8829768d18b20476c8cce90f2ec9a1c19cc5a7da3f94bd198a60c3c0\": container with ID starting with da1342de8829768d18b20476c8cce90f2ec9a1c19cc5a7da3f94bd198a60c3c0 not found: ID does not exist" Oct 08 19:54:29 crc kubenswrapper[4988]: I1008 19:54:29.460497 4988 scope.go:117] "RemoveContainer" containerID="18a74433931c8691f496240afa0d5ef264cfe8235cb580225b48b2fa68150461" Oct 08 19:54:29 crc kubenswrapper[4988]: E1008 19:54:29.460760 4988 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"18a74433931c8691f496240afa0d5ef264cfe8235cb580225b48b2fa68150461\": container with ID starting with 18a74433931c8691f496240afa0d5ef264cfe8235cb580225b48b2fa68150461 not found: ID does not exist" containerID="18a74433931c8691f496240afa0d5ef264cfe8235cb580225b48b2fa68150461" Oct 08 19:54:29 crc kubenswrapper[4988]: I1008 19:54:29.460779 4988 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"18a74433931c8691f496240afa0d5ef264cfe8235cb580225b48b2fa68150461"} err="failed to get container status \"18a74433931c8691f496240afa0d5ef264cfe8235cb580225b48b2fa68150461\": rpc error: code = NotFound desc = could not find container \"18a74433931c8691f496240afa0d5ef264cfe8235cb580225b48b2fa68150461\": container with ID starting with 18a74433931c8691f496240afa0d5ef264cfe8235cb580225b48b2fa68150461 not found: ID does not exist" Oct 08 19:54:29 crc kubenswrapper[4988]: I1008 19:54:29.460791 4988 scope.go:117] "RemoveContainer" containerID="7d091e741f2608b8c6e8449e8f0ea55a2424a323ff499a8a7ba86aa621893690" Oct 08 19:54:29 crc kubenswrapper[4988]: E1008 19:54:29.461257 4988 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7d091e741f2608b8c6e8449e8f0ea55a2424a323ff499a8a7ba86aa621893690\": container with ID starting with 7d091e741f2608b8c6e8449e8f0ea55a2424a323ff499a8a7ba86aa621893690 not found: ID does not exist" containerID="7d091e741f2608b8c6e8449e8f0ea55a2424a323ff499a8a7ba86aa621893690" Oct 08 19:54:29 crc kubenswrapper[4988]: I1008 19:54:29.461277 4988 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7d091e741f2608b8c6e8449e8f0ea55a2424a323ff499a8a7ba86aa621893690"} err="failed to get container status \"7d091e741f2608b8c6e8449e8f0ea55a2424a323ff499a8a7ba86aa621893690\": rpc error: code = NotFound desc = could not find container \"7d091e741f2608b8c6e8449e8f0ea55a2424a323ff499a8a7ba86aa621893690\": container with ID starting with 7d091e741f2608b8c6e8449e8f0ea55a2424a323ff499a8a7ba86aa621893690 not found: ID does not exist" Oct 08 19:54:31 crc kubenswrapper[4988]: I1008 19:54:31.252805 4988 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="dd00ff3f-0c19-43d8-8a05-80ad79711731" path="/var/lib/kubelet/pods/dd00ff3f-0c19-43d8-8a05-80ad79711731/volumes" Oct 08 19:54:33 crc kubenswrapper[4988]: I1008 19:54:33.040462 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-8807-account-create-2rpmf"] Oct 08 19:54:33 crc kubenswrapper[4988]: I1008 19:54:33.058061 4988 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-8807-account-create-2rpmf"] Oct 08 19:54:33 crc kubenswrapper[4988]: I1008 19:54:33.259190 4988 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="70fc1b25-cb88-4345-b70f-e72c74df4a0f" path="/var/lib/kubelet/pods/70fc1b25-cb88-4345-b70f-e72c74df4a0f/volumes" Oct 08 19:54:37 crc kubenswrapper[4988]: I1008 19:54:37.616233 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-7b9d74c97-mthd5"] Oct 08 19:54:37 crc kubenswrapper[4988]: E1008 19:54:37.617315 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dd00ff3f-0c19-43d8-8a05-80ad79711731" containerName="extract-content" Oct 08 19:54:37 crc kubenswrapper[4988]: I1008 19:54:37.617333 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="dd00ff3f-0c19-43d8-8a05-80ad79711731" containerName="extract-content" Oct 08 19:54:37 crc kubenswrapper[4988]: E1008 19:54:37.617359 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dd00ff3f-0c19-43d8-8a05-80ad79711731" containerName="extract-utilities" Oct 08 19:54:37 crc kubenswrapper[4988]: I1008 19:54:37.617367 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="dd00ff3f-0c19-43d8-8a05-80ad79711731" containerName="extract-utilities" Oct 08 19:54:37 crc kubenswrapper[4988]: E1008 19:54:37.617412 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dd00ff3f-0c19-43d8-8a05-80ad79711731" containerName="registry-server" Oct 08 19:54:37 crc kubenswrapper[4988]: I1008 19:54:37.617420 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="dd00ff3f-0c19-43d8-8a05-80ad79711731" containerName="registry-server" Oct 08 19:54:37 crc kubenswrapper[4988]: I1008 19:54:37.617744 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="dd00ff3f-0c19-43d8-8a05-80ad79711731" containerName="registry-server" Oct 08 19:54:37 crc kubenswrapper[4988]: I1008 19:54:37.619125 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-7b9d74c97-mthd5" Oct 08 19:54:37 crc kubenswrapper[4988]: I1008 19:54:37.630402 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"horizon" Oct 08 19:54:37 crc kubenswrapper[4988]: I1008 19:54:37.630587 4988 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"horizon-scripts" Oct 08 19:54:37 crc kubenswrapper[4988]: I1008 19:54:37.631508 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"horizon-horizon-dockercfg-ndmlb" Oct 08 19:54:37 crc kubenswrapper[4988]: I1008 19:54:37.631726 4988 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"horizon-config-data" Oct 08 19:54:37 crc kubenswrapper[4988]: I1008 19:54:37.638604 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-7b9d74c97-mthd5"] Oct 08 19:54:37 crc kubenswrapper[4988]: I1008 19:54:37.643949 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 08 19:54:37 crc kubenswrapper[4988]: I1008 19:54:37.644180 4988 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="185123ed-b356-4268-af3d-e44ea04bbf3f" containerName="glance-log" containerID="cri-o://3e47ea1ab135b55cab053319b8c923376c4cda044c92fb262fdd29d259c417da" gracePeriod=30 Oct 08 19:54:37 crc kubenswrapper[4988]: I1008 19:54:37.644309 4988 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="185123ed-b356-4268-af3d-e44ea04bbf3f" containerName="glance-httpd" containerID="cri-o://1decc70dd2302ff2a788701c560d989e84796d90fcfb7cf0324d8c48e524d42e" gracePeriod=30 Oct 08 19:54:37 crc kubenswrapper[4988]: I1008 19:54:37.722877 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 08 19:54:37 crc kubenswrapper[4988]: I1008 19:54:37.723128 4988 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="d29618a6-2810-4d35-b5b3-2369d4a1c096" containerName="glance-log" containerID="cri-o://77b13c99dddd5b6f4b11423db1a866862933fa4da7f96cd4e3f36292c533bb98" gracePeriod=30 Oct 08 19:54:37 crc kubenswrapper[4988]: I1008 19:54:37.723573 4988 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="d29618a6-2810-4d35-b5b3-2369d4a1c096" containerName="glance-httpd" containerID="cri-o://63520e1979dd97c54c047ccdf382e36d66fc5644ef9ac1bde7c697c98007cbdd" gracePeriod=30 Oct 08 19:54:37 crc kubenswrapper[4988]: I1008 19:54:37.729442 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-54cd997f77-cs4qd"] Oct 08 19:54:37 crc kubenswrapper[4988]: I1008 19:54:37.733240 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-54cd997f77-cs4qd" Oct 08 19:54:37 crc kubenswrapper[4988]: I1008 19:54:37.737574 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-54cd997f77-cs4qd"] Oct 08 19:54:37 crc kubenswrapper[4988]: I1008 19:54:37.779643 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/322a3c03-38ac-4ef2-8980-c1976f855a81-config-data\") pod \"horizon-7b9d74c97-mthd5\" (UID: \"322a3c03-38ac-4ef2-8980-c1976f855a81\") " pod="openstack/horizon-7b9d74c97-mthd5" Oct 08 19:54:37 crc kubenswrapper[4988]: I1008 19:54:37.779740 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/322a3c03-38ac-4ef2-8980-c1976f855a81-scripts\") pod \"horizon-7b9d74c97-mthd5\" (UID: \"322a3c03-38ac-4ef2-8980-c1976f855a81\") " pod="openstack/horizon-7b9d74c97-mthd5" Oct 08 19:54:37 crc kubenswrapper[4988]: I1008 19:54:37.779805 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/322a3c03-38ac-4ef2-8980-c1976f855a81-logs\") pod \"horizon-7b9d74c97-mthd5\" (UID: \"322a3c03-38ac-4ef2-8980-c1976f855a81\") " pod="openstack/horizon-7b9d74c97-mthd5" Oct 08 19:54:37 crc kubenswrapper[4988]: I1008 19:54:37.779828 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9ppgg\" (UniqueName: \"kubernetes.io/projected/322a3c03-38ac-4ef2-8980-c1976f855a81-kube-api-access-9ppgg\") pod \"horizon-7b9d74c97-mthd5\" (UID: \"322a3c03-38ac-4ef2-8980-c1976f855a81\") " pod="openstack/horizon-7b9d74c97-mthd5" Oct 08 19:54:37 crc kubenswrapper[4988]: I1008 19:54:37.779871 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/322a3c03-38ac-4ef2-8980-c1976f855a81-horizon-secret-key\") pod \"horizon-7b9d74c97-mthd5\" (UID: \"322a3c03-38ac-4ef2-8980-c1976f855a81\") " pod="openstack/horizon-7b9d74c97-mthd5" Oct 08 19:54:37 crc kubenswrapper[4988]: I1008 19:54:37.881174 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/83fe9318-96ca-4f74-b838-9997e59c1ea2-scripts\") pod \"horizon-54cd997f77-cs4qd\" (UID: \"83fe9318-96ca-4f74-b838-9997e59c1ea2\") " pod="openstack/horizon-54cd997f77-cs4qd" Oct 08 19:54:37 crc kubenswrapper[4988]: I1008 19:54:37.881242 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/322a3c03-38ac-4ef2-8980-c1976f855a81-horizon-secret-key\") pod \"horizon-7b9d74c97-mthd5\" (UID: \"322a3c03-38ac-4ef2-8980-c1976f855a81\") " pod="openstack/horizon-7b9d74c97-mthd5" Oct 08 19:54:37 crc kubenswrapper[4988]: I1008 19:54:37.881299 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/322a3c03-38ac-4ef2-8980-c1976f855a81-config-data\") pod \"horizon-7b9d74c97-mthd5\" (UID: \"322a3c03-38ac-4ef2-8980-c1976f855a81\") " pod="openstack/horizon-7b9d74c97-mthd5" Oct 08 19:54:37 crc kubenswrapper[4988]: I1008 19:54:37.881373 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/322a3c03-38ac-4ef2-8980-c1976f855a81-scripts\") pod \"horizon-7b9d74c97-mthd5\" (UID: \"322a3c03-38ac-4ef2-8980-c1976f855a81\") " pod="openstack/horizon-7b9d74c97-mthd5" Oct 08 19:54:37 crc kubenswrapper[4988]: I1008 19:54:37.881424 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/83fe9318-96ca-4f74-b838-9997e59c1ea2-config-data\") pod \"horizon-54cd997f77-cs4qd\" (UID: \"83fe9318-96ca-4f74-b838-9997e59c1ea2\") " pod="openstack/horizon-54cd997f77-cs4qd" Oct 08 19:54:37 crc kubenswrapper[4988]: I1008 19:54:37.881458 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/83fe9318-96ca-4f74-b838-9997e59c1ea2-logs\") pod \"horizon-54cd997f77-cs4qd\" (UID: \"83fe9318-96ca-4f74-b838-9997e59c1ea2\") " pod="openstack/horizon-54cd997f77-cs4qd" Oct 08 19:54:37 crc kubenswrapper[4988]: I1008 19:54:37.881483 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/322a3c03-38ac-4ef2-8980-c1976f855a81-logs\") pod \"horizon-7b9d74c97-mthd5\" (UID: \"322a3c03-38ac-4ef2-8980-c1976f855a81\") " pod="openstack/horizon-7b9d74c97-mthd5" Oct 08 19:54:37 crc kubenswrapper[4988]: I1008 19:54:37.881504 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9ppgg\" (UniqueName: \"kubernetes.io/projected/322a3c03-38ac-4ef2-8980-c1976f855a81-kube-api-access-9ppgg\") pod \"horizon-7b9d74c97-mthd5\" (UID: \"322a3c03-38ac-4ef2-8980-c1976f855a81\") " pod="openstack/horizon-7b9d74c97-mthd5" Oct 08 19:54:37 crc kubenswrapper[4988]: I1008 19:54:37.881526 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8bm6z\" (UniqueName: \"kubernetes.io/projected/83fe9318-96ca-4f74-b838-9997e59c1ea2-kube-api-access-8bm6z\") pod \"horizon-54cd997f77-cs4qd\" (UID: \"83fe9318-96ca-4f74-b838-9997e59c1ea2\") " pod="openstack/horizon-54cd997f77-cs4qd" Oct 08 19:54:37 crc kubenswrapper[4988]: I1008 19:54:37.881663 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/83fe9318-96ca-4f74-b838-9997e59c1ea2-horizon-secret-key\") pod \"horizon-54cd997f77-cs4qd\" (UID: \"83fe9318-96ca-4f74-b838-9997e59c1ea2\") " pod="openstack/horizon-54cd997f77-cs4qd" Oct 08 19:54:37 crc kubenswrapper[4988]: I1008 19:54:37.882218 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/322a3c03-38ac-4ef2-8980-c1976f855a81-logs\") pod \"horizon-7b9d74c97-mthd5\" (UID: \"322a3c03-38ac-4ef2-8980-c1976f855a81\") " pod="openstack/horizon-7b9d74c97-mthd5" Oct 08 19:54:37 crc kubenswrapper[4988]: I1008 19:54:37.883038 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/322a3c03-38ac-4ef2-8980-c1976f855a81-config-data\") pod \"horizon-7b9d74c97-mthd5\" (UID: \"322a3c03-38ac-4ef2-8980-c1976f855a81\") " pod="openstack/horizon-7b9d74c97-mthd5" Oct 08 19:54:37 crc kubenswrapper[4988]: I1008 19:54:37.883536 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/322a3c03-38ac-4ef2-8980-c1976f855a81-scripts\") pod \"horizon-7b9d74c97-mthd5\" (UID: \"322a3c03-38ac-4ef2-8980-c1976f855a81\") " pod="openstack/horizon-7b9d74c97-mthd5" Oct 08 19:54:37 crc kubenswrapper[4988]: I1008 19:54:37.887163 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/322a3c03-38ac-4ef2-8980-c1976f855a81-horizon-secret-key\") pod \"horizon-7b9d74c97-mthd5\" (UID: \"322a3c03-38ac-4ef2-8980-c1976f855a81\") " pod="openstack/horizon-7b9d74c97-mthd5" Oct 08 19:54:37 crc kubenswrapper[4988]: I1008 19:54:37.895493 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9ppgg\" (UniqueName: \"kubernetes.io/projected/322a3c03-38ac-4ef2-8980-c1976f855a81-kube-api-access-9ppgg\") pod \"horizon-7b9d74c97-mthd5\" (UID: \"322a3c03-38ac-4ef2-8980-c1976f855a81\") " pod="openstack/horizon-7b9d74c97-mthd5" Oct 08 19:54:37 crc kubenswrapper[4988]: I1008 19:54:37.971295 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-7b9d74c97-mthd5" Oct 08 19:54:37 crc kubenswrapper[4988]: I1008 19:54:37.983789 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/83fe9318-96ca-4f74-b838-9997e59c1ea2-config-data\") pod \"horizon-54cd997f77-cs4qd\" (UID: \"83fe9318-96ca-4f74-b838-9997e59c1ea2\") " pod="openstack/horizon-54cd997f77-cs4qd" Oct 08 19:54:37 crc kubenswrapper[4988]: I1008 19:54:37.983872 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/83fe9318-96ca-4f74-b838-9997e59c1ea2-logs\") pod \"horizon-54cd997f77-cs4qd\" (UID: \"83fe9318-96ca-4f74-b838-9997e59c1ea2\") " pod="openstack/horizon-54cd997f77-cs4qd" Oct 08 19:54:37 crc kubenswrapper[4988]: I1008 19:54:37.983937 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8bm6z\" (UniqueName: \"kubernetes.io/projected/83fe9318-96ca-4f74-b838-9997e59c1ea2-kube-api-access-8bm6z\") pod \"horizon-54cd997f77-cs4qd\" (UID: \"83fe9318-96ca-4f74-b838-9997e59c1ea2\") " pod="openstack/horizon-54cd997f77-cs4qd" Oct 08 19:54:37 crc kubenswrapper[4988]: I1008 19:54:37.983962 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/83fe9318-96ca-4f74-b838-9997e59c1ea2-horizon-secret-key\") pod \"horizon-54cd997f77-cs4qd\" (UID: \"83fe9318-96ca-4f74-b838-9997e59c1ea2\") " pod="openstack/horizon-54cd997f77-cs4qd" Oct 08 19:54:37 crc kubenswrapper[4988]: I1008 19:54:37.983997 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/83fe9318-96ca-4f74-b838-9997e59c1ea2-scripts\") pod \"horizon-54cd997f77-cs4qd\" (UID: \"83fe9318-96ca-4f74-b838-9997e59c1ea2\") " pod="openstack/horizon-54cd997f77-cs4qd" Oct 08 19:54:37 crc kubenswrapper[4988]: I1008 19:54:37.984435 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/83fe9318-96ca-4f74-b838-9997e59c1ea2-logs\") pod \"horizon-54cd997f77-cs4qd\" (UID: \"83fe9318-96ca-4f74-b838-9997e59c1ea2\") " pod="openstack/horizon-54cd997f77-cs4qd" Oct 08 19:54:37 crc kubenswrapper[4988]: I1008 19:54:37.985997 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/83fe9318-96ca-4f74-b838-9997e59c1ea2-config-data\") pod \"horizon-54cd997f77-cs4qd\" (UID: \"83fe9318-96ca-4f74-b838-9997e59c1ea2\") " pod="openstack/horizon-54cd997f77-cs4qd" Oct 08 19:54:37 crc kubenswrapper[4988]: I1008 19:54:37.989008 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/83fe9318-96ca-4f74-b838-9997e59c1ea2-scripts\") pod \"horizon-54cd997f77-cs4qd\" (UID: \"83fe9318-96ca-4f74-b838-9997e59c1ea2\") " pod="openstack/horizon-54cd997f77-cs4qd" Oct 08 19:54:37 crc kubenswrapper[4988]: I1008 19:54:37.990250 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/83fe9318-96ca-4f74-b838-9997e59c1ea2-horizon-secret-key\") pod \"horizon-54cd997f77-cs4qd\" (UID: \"83fe9318-96ca-4f74-b838-9997e59c1ea2\") " pod="openstack/horizon-54cd997f77-cs4qd" Oct 08 19:54:38 crc kubenswrapper[4988]: I1008 19:54:37.999946 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8bm6z\" (UniqueName: \"kubernetes.io/projected/83fe9318-96ca-4f74-b838-9997e59c1ea2-kube-api-access-8bm6z\") pod \"horizon-54cd997f77-cs4qd\" (UID: \"83fe9318-96ca-4f74-b838-9997e59c1ea2\") " pod="openstack/horizon-54cd997f77-cs4qd" Oct 08 19:54:38 crc kubenswrapper[4988]: I1008 19:54:38.052521 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-54cd997f77-cs4qd" Oct 08 19:54:38 crc kubenswrapper[4988]: I1008 19:54:38.438455 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-7b9d74c97-mthd5"] Oct 08 19:54:38 crc kubenswrapper[4988]: I1008 19:54:38.449508 4988 generic.go:334] "Generic (PLEG): container finished" podID="185123ed-b356-4268-af3d-e44ea04bbf3f" containerID="3e47ea1ab135b55cab053319b8c923376c4cda044c92fb262fdd29d259c417da" exitCode=143 Oct 08 19:54:38 crc kubenswrapper[4988]: I1008 19:54:38.449583 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"185123ed-b356-4268-af3d-e44ea04bbf3f","Type":"ContainerDied","Data":"3e47ea1ab135b55cab053319b8c923376c4cda044c92fb262fdd29d259c417da"} Oct 08 19:54:38 crc kubenswrapper[4988]: W1008 19:54:38.451834 4988 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod322a3c03_38ac_4ef2_8980_c1976f855a81.slice/crio-ed1ba5812189b29ecd1ed422dc3e95f1ab70f71ba0dad04343513bf59fdf71a2 WatchSource:0}: Error finding container ed1ba5812189b29ecd1ed422dc3e95f1ab70f71ba0dad04343513bf59fdf71a2: Status 404 returned error can't find the container with id ed1ba5812189b29ecd1ed422dc3e95f1ab70f71ba0dad04343513bf59fdf71a2 Oct 08 19:54:38 crc kubenswrapper[4988]: I1008 19:54:38.455925 4988 generic.go:334] "Generic (PLEG): container finished" podID="d29618a6-2810-4d35-b5b3-2369d4a1c096" containerID="77b13c99dddd5b6f4b11423db1a866862933fa4da7f96cd4e3f36292c533bb98" exitCode=143 Oct 08 19:54:38 crc kubenswrapper[4988]: I1008 19:54:38.455959 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"d29618a6-2810-4d35-b5b3-2369d4a1c096","Type":"ContainerDied","Data":"77b13c99dddd5b6f4b11423db1a866862933fa4da7f96cd4e3f36292c533bb98"} Oct 08 19:54:38 crc kubenswrapper[4988]: I1008 19:54:38.549495 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-54cd997f77-cs4qd"] Oct 08 19:54:39 crc kubenswrapper[4988]: I1008 19:54:39.415072 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-54cd997f77-cs4qd"] Oct 08 19:54:39 crc kubenswrapper[4988]: I1008 19:54:39.439337 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-d95cf97bd-hmknh"] Oct 08 19:54:39 crc kubenswrapper[4988]: I1008 19:54:39.447326 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-d95cf97bd-hmknh" Oct 08 19:54:39 crc kubenswrapper[4988]: I1008 19:54:39.449598 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-horizon-svc" Oct 08 19:54:39 crc kubenswrapper[4988]: I1008 19:54:39.460193 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-d95cf97bd-hmknh"] Oct 08 19:54:39 crc kubenswrapper[4988]: I1008 19:54:39.485717 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-54cd997f77-cs4qd" event={"ID":"83fe9318-96ca-4f74-b838-9997e59c1ea2","Type":"ContainerStarted","Data":"8fef69f66e9971ee638657baf4eeae76feafb4bfec6e030bc00f580c5aa86a92"} Oct 08 19:54:39 crc kubenswrapper[4988]: I1008 19:54:39.490617 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7b9d74c97-mthd5" event={"ID":"322a3c03-38ac-4ef2-8980-c1976f855a81","Type":"ContainerStarted","Data":"ed1ba5812189b29ecd1ed422dc3e95f1ab70f71ba0dad04343513bf59fdf71a2"} Oct 08 19:54:39 crc kubenswrapper[4988]: I1008 19:54:39.516168 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-7b9d74c97-mthd5"] Oct 08 19:54:39 crc kubenswrapper[4988]: I1008 19:54:39.524503 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-7d85fc8646-gv4lm"] Oct 08 19:54:39 crc kubenswrapper[4988]: I1008 19:54:39.526184 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-7d85fc8646-gv4lm" Oct 08 19:54:39 crc kubenswrapper[4988]: I1008 19:54:39.533581 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-7d85fc8646-gv4lm"] Oct 08 19:54:39 crc kubenswrapper[4988]: I1008 19:54:39.535359 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/f45d9be7-457d-472d-8c9f-463f9d241eae-horizon-tls-certs\") pod \"horizon-d95cf97bd-hmknh\" (UID: \"f45d9be7-457d-472d-8c9f-463f9d241eae\") " pod="openstack/horizon-d95cf97bd-hmknh" Oct 08 19:54:39 crc kubenswrapper[4988]: I1008 19:54:39.535427 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/f45d9be7-457d-472d-8c9f-463f9d241eae-config-data\") pod \"horizon-d95cf97bd-hmknh\" (UID: \"f45d9be7-457d-472d-8c9f-463f9d241eae\") " pod="openstack/horizon-d95cf97bd-hmknh" Oct 08 19:54:39 crc kubenswrapper[4988]: I1008 19:54:39.535494 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f45d9be7-457d-472d-8c9f-463f9d241eae-scripts\") pod \"horizon-d95cf97bd-hmknh\" (UID: \"f45d9be7-457d-472d-8c9f-463f9d241eae\") " pod="openstack/horizon-d95cf97bd-hmknh" Oct 08 19:54:39 crc kubenswrapper[4988]: I1008 19:54:39.535543 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-klg9v\" (UniqueName: \"kubernetes.io/projected/f45d9be7-457d-472d-8c9f-463f9d241eae-kube-api-access-klg9v\") pod \"horizon-d95cf97bd-hmknh\" (UID: \"f45d9be7-457d-472d-8c9f-463f9d241eae\") " pod="openstack/horizon-d95cf97bd-hmknh" Oct 08 19:54:39 crc kubenswrapper[4988]: I1008 19:54:39.535643 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/f45d9be7-457d-472d-8c9f-463f9d241eae-horizon-secret-key\") pod \"horizon-d95cf97bd-hmknh\" (UID: \"f45d9be7-457d-472d-8c9f-463f9d241eae\") " pod="openstack/horizon-d95cf97bd-hmknh" Oct 08 19:54:39 crc kubenswrapper[4988]: I1008 19:54:39.535695 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f45d9be7-457d-472d-8c9f-463f9d241eae-combined-ca-bundle\") pod \"horizon-d95cf97bd-hmknh\" (UID: \"f45d9be7-457d-472d-8c9f-463f9d241eae\") " pod="openstack/horizon-d95cf97bd-hmknh" Oct 08 19:54:39 crc kubenswrapper[4988]: I1008 19:54:39.535726 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f45d9be7-457d-472d-8c9f-463f9d241eae-logs\") pod \"horizon-d95cf97bd-hmknh\" (UID: \"f45d9be7-457d-472d-8c9f-463f9d241eae\") " pod="openstack/horizon-d95cf97bd-hmknh" Oct 08 19:54:39 crc kubenswrapper[4988]: I1008 19:54:39.637990 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/f45d9be7-457d-472d-8c9f-463f9d241eae-config-data\") pod \"horizon-d95cf97bd-hmknh\" (UID: \"f45d9be7-457d-472d-8c9f-463f9d241eae\") " pod="openstack/horizon-d95cf97bd-hmknh" Oct 08 19:54:39 crc kubenswrapper[4988]: I1008 19:54:39.638068 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c533a2e1-7213-4410-b862-5eed5e1fe8e9-logs\") pod \"horizon-7d85fc8646-gv4lm\" (UID: \"c533a2e1-7213-4410-b862-5eed5e1fe8e9\") " pod="openstack/horizon-7d85fc8646-gv4lm" Oct 08 19:54:39 crc kubenswrapper[4988]: I1008 19:54:39.638092 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f45d9be7-457d-472d-8c9f-463f9d241eae-scripts\") pod \"horizon-d95cf97bd-hmknh\" (UID: \"f45d9be7-457d-472d-8c9f-463f9d241eae\") " pod="openstack/horizon-d95cf97bd-hmknh" Oct 08 19:54:39 crc kubenswrapper[4988]: I1008 19:54:39.638151 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-klg9v\" (UniqueName: \"kubernetes.io/projected/f45d9be7-457d-472d-8c9f-463f9d241eae-kube-api-access-klg9v\") pod \"horizon-d95cf97bd-hmknh\" (UID: \"f45d9be7-457d-472d-8c9f-463f9d241eae\") " pod="openstack/horizon-d95cf97bd-hmknh" Oct 08 19:54:39 crc kubenswrapper[4988]: I1008 19:54:39.638208 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/f45d9be7-457d-472d-8c9f-463f9d241eae-horizon-secret-key\") pod \"horizon-d95cf97bd-hmknh\" (UID: \"f45d9be7-457d-472d-8c9f-463f9d241eae\") " pod="openstack/horizon-d95cf97bd-hmknh" Oct 08 19:54:39 crc kubenswrapper[4988]: I1008 19:54:39.638228 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/c533a2e1-7213-4410-b862-5eed5e1fe8e9-horizon-tls-certs\") pod \"horizon-7d85fc8646-gv4lm\" (UID: \"c533a2e1-7213-4410-b862-5eed5e1fe8e9\") " pod="openstack/horizon-7d85fc8646-gv4lm" Oct 08 19:54:39 crc kubenswrapper[4988]: I1008 19:54:39.638252 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8hpsk\" (UniqueName: \"kubernetes.io/projected/c533a2e1-7213-4410-b862-5eed5e1fe8e9-kube-api-access-8hpsk\") pod \"horizon-7d85fc8646-gv4lm\" (UID: \"c533a2e1-7213-4410-b862-5eed5e1fe8e9\") " pod="openstack/horizon-7d85fc8646-gv4lm" Oct 08 19:54:39 crc kubenswrapper[4988]: I1008 19:54:39.638274 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f45d9be7-457d-472d-8c9f-463f9d241eae-combined-ca-bundle\") pod \"horizon-d95cf97bd-hmknh\" (UID: \"f45d9be7-457d-472d-8c9f-463f9d241eae\") " pod="openstack/horizon-d95cf97bd-hmknh" Oct 08 19:54:39 crc kubenswrapper[4988]: I1008 19:54:39.638288 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f45d9be7-457d-472d-8c9f-463f9d241eae-logs\") pod \"horizon-d95cf97bd-hmknh\" (UID: \"f45d9be7-457d-472d-8c9f-463f9d241eae\") " pod="openstack/horizon-d95cf97bd-hmknh" Oct 08 19:54:39 crc kubenswrapper[4988]: I1008 19:54:39.638336 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c533a2e1-7213-4410-b862-5eed5e1fe8e9-combined-ca-bundle\") pod \"horizon-7d85fc8646-gv4lm\" (UID: \"c533a2e1-7213-4410-b862-5eed5e1fe8e9\") " pod="openstack/horizon-7d85fc8646-gv4lm" Oct 08 19:54:39 crc kubenswrapper[4988]: I1008 19:54:39.638378 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/c533a2e1-7213-4410-b862-5eed5e1fe8e9-horizon-secret-key\") pod \"horizon-7d85fc8646-gv4lm\" (UID: \"c533a2e1-7213-4410-b862-5eed5e1fe8e9\") " pod="openstack/horizon-7d85fc8646-gv4lm" Oct 08 19:54:39 crc kubenswrapper[4988]: I1008 19:54:39.638432 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c533a2e1-7213-4410-b862-5eed5e1fe8e9-scripts\") pod \"horizon-7d85fc8646-gv4lm\" (UID: \"c533a2e1-7213-4410-b862-5eed5e1fe8e9\") " pod="openstack/horizon-7d85fc8646-gv4lm" Oct 08 19:54:39 crc kubenswrapper[4988]: I1008 19:54:39.638519 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/c533a2e1-7213-4410-b862-5eed5e1fe8e9-config-data\") pod \"horizon-7d85fc8646-gv4lm\" (UID: \"c533a2e1-7213-4410-b862-5eed5e1fe8e9\") " pod="openstack/horizon-7d85fc8646-gv4lm" Oct 08 19:54:39 crc kubenswrapper[4988]: I1008 19:54:39.638549 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/f45d9be7-457d-472d-8c9f-463f9d241eae-horizon-tls-certs\") pod \"horizon-d95cf97bd-hmknh\" (UID: \"f45d9be7-457d-472d-8c9f-463f9d241eae\") " pod="openstack/horizon-d95cf97bd-hmknh" Oct 08 19:54:39 crc kubenswrapper[4988]: I1008 19:54:39.638934 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f45d9be7-457d-472d-8c9f-463f9d241eae-scripts\") pod \"horizon-d95cf97bd-hmknh\" (UID: \"f45d9be7-457d-472d-8c9f-463f9d241eae\") " pod="openstack/horizon-d95cf97bd-hmknh" Oct 08 19:54:39 crc kubenswrapper[4988]: I1008 19:54:39.639516 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f45d9be7-457d-472d-8c9f-463f9d241eae-logs\") pod \"horizon-d95cf97bd-hmknh\" (UID: \"f45d9be7-457d-472d-8c9f-463f9d241eae\") " pod="openstack/horizon-d95cf97bd-hmknh" Oct 08 19:54:39 crc kubenswrapper[4988]: I1008 19:54:39.640061 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/f45d9be7-457d-472d-8c9f-463f9d241eae-config-data\") pod \"horizon-d95cf97bd-hmknh\" (UID: \"f45d9be7-457d-472d-8c9f-463f9d241eae\") " pod="openstack/horizon-d95cf97bd-hmknh" Oct 08 19:54:39 crc kubenswrapper[4988]: I1008 19:54:39.644236 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f45d9be7-457d-472d-8c9f-463f9d241eae-combined-ca-bundle\") pod \"horizon-d95cf97bd-hmknh\" (UID: \"f45d9be7-457d-472d-8c9f-463f9d241eae\") " pod="openstack/horizon-d95cf97bd-hmknh" Oct 08 19:54:39 crc kubenswrapper[4988]: I1008 19:54:39.646889 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/f45d9be7-457d-472d-8c9f-463f9d241eae-horizon-secret-key\") pod \"horizon-d95cf97bd-hmknh\" (UID: \"f45d9be7-457d-472d-8c9f-463f9d241eae\") " pod="openstack/horizon-d95cf97bd-hmknh" Oct 08 19:54:39 crc kubenswrapper[4988]: I1008 19:54:39.647861 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/f45d9be7-457d-472d-8c9f-463f9d241eae-horizon-tls-certs\") pod \"horizon-d95cf97bd-hmknh\" (UID: \"f45d9be7-457d-472d-8c9f-463f9d241eae\") " pod="openstack/horizon-d95cf97bd-hmknh" Oct 08 19:54:39 crc kubenswrapper[4988]: I1008 19:54:39.655871 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-klg9v\" (UniqueName: \"kubernetes.io/projected/f45d9be7-457d-472d-8c9f-463f9d241eae-kube-api-access-klg9v\") pod \"horizon-d95cf97bd-hmknh\" (UID: \"f45d9be7-457d-472d-8c9f-463f9d241eae\") " pod="openstack/horizon-d95cf97bd-hmknh" Oct 08 19:54:39 crc kubenswrapper[4988]: I1008 19:54:39.741644 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c533a2e1-7213-4410-b862-5eed5e1fe8e9-combined-ca-bundle\") pod \"horizon-7d85fc8646-gv4lm\" (UID: \"c533a2e1-7213-4410-b862-5eed5e1fe8e9\") " pod="openstack/horizon-7d85fc8646-gv4lm" Oct 08 19:54:39 crc kubenswrapper[4988]: I1008 19:54:39.741700 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/c533a2e1-7213-4410-b862-5eed5e1fe8e9-horizon-secret-key\") pod \"horizon-7d85fc8646-gv4lm\" (UID: \"c533a2e1-7213-4410-b862-5eed5e1fe8e9\") " pod="openstack/horizon-7d85fc8646-gv4lm" Oct 08 19:54:39 crc kubenswrapper[4988]: I1008 19:54:39.741723 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c533a2e1-7213-4410-b862-5eed5e1fe8e9-scripts\") pod \"horizon-7d85fc8646-gv4lm\" (UID: \"c533a2e1-7213-4410-b862-5eed5e1fe8e9\") " pod="openstack/horizon-7d85fc8646-gv4lm" Oct 08 19:54:39 crc kubenswrapper[4988]: I1008 19:54:39.741796 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/c533a2e1-7213-4410-b862-5eed5e1fe8e9-config-data\") pod \"horizon-7d85fc8646-gv4lm\" (UID: \"c533a2e1-7213-4410-b862-5eed5e1fe8e9\") " pod="openstack/horizon-7d85fc8646-gv4lm" Oct 08 19:54:39 crc kubenswrapper[4988]: I1008 19:54:39.741836 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c533a2e1-7213-4410-b862-5eed5e1fe8e9-logs\") pod \"horizon-7d85fc8646-gv4lm\" (UID: \"c533a2e1-7213-4410-b862-5eed5e1fe8e9\") " pod="openstack/horizon-7d85fc8646-gv4lm" Oct 08 19:54:39 crc kubenswrapper[4988]: I1008 19:54:39.741884 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/c533a2e1-7213-4410-b862-5eed5e1fe8e9-horizon-tls-certs\") pod \"horizon-7d85fc8646-gv4lm\" (UID: \"c533a2e1-7213-4410-b862-5eed5e1fe8e9\") " pod="openstack/horizon-7d85fc8646-gv4lm" Oct 08 19:54:39 crc kubenswrapper[4988]: I1008 19:54:39.741905 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8hpsk\" (UniqueName: \"kubernetes.io/projected/c533a2e1-7213-4410-b862-5eed5e1fe8e9-kube-api-access-8hpsk\") pod \"horizon-7d85fc8646-gv4lm\" (UID: \"c533a2e1-7213-4410-b862-5eed5e1fe8e9\") " pod="openstack/horizon-7d85fc8646-gv4lm" Oct 08 19:54:39 crc kubenswrapper[4988]: I1008 19:54:39.742618 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c533a2e1-7213-4410-b862-5eed5e1fe8e9-logs\") pod \"horizon-7d85fc8646-gv4lm\" (UID: \"c533a2e1-7213-4410-b862-5eed5e1fe8e9\") " pod="openstack/horizon-7d85fc8646-gv4lm" Oct 08 19:54:39 crc kubenswrapper[4988]: I1008 19:54:39.742984 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c533a2e1-7213-4410-b862-5eed5e1fe8e9-scripts\") pod \"horizon-7d85fc8646-gv4lm\" (UID: \"c533a2e1-7213-4410-b862-5eed5e1fe8e9\") " pod="openstack/horizon-7d85fc8646-gv4lm" Oct 08 19:54:39 crc kubenswrapper[4988]: I1008 19:54:39.743986 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/c533a2e1-7213-4410-b862-5eed5e1fe8e9-config-data\") pod \"horizon-7d85fc8646-gv4lm\" (UID: \"c533a2e1-7213-4410-b862-5eed5e1fe8e9\") " pod="openstack/horizon-7d85fc8646-gv4lm" Oct 08 19:54:39 crc kubenswrapper[4988]: I1008 19:54:39.745245 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c533a2e1-7213-4410-b862-5eed5e1fe8e9-combined-ca-bundle\") pod \"horizon-7d85fc8646-gv4lm\" (UID: \"c533a2e1-7213-4410-b862-5eed5e1fe8e9\") " pod="openstack/horizon-7d85fc8646-gv4lm" Oct 08 19:54:39 crc kubenswrapper[4988]: I1008 19:54:39.745831 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/c533a2e1-7213-4410-b862-5eed5e1fe8e9-horizon-tls-certs\") pod \"horizon-7d85fc8646-gv4lm\" (UID: \"c533a2e1-7213-4410-b862-5eed5e1fe8e9\") " pod="openstack/horizon-7d85fc8646-gv4lm" Oct 08 19:54:39 crc kubenswrapper[4988]: I1008 19:54:39.747884 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/c533a2e1-7213-4410-b862-5eed5e1fe8e9-horizon-secret-key\") pod \"horizon-7d85fc8646-gv4lm\" (UID: \"c533a2e1-7213-4410-b862-5eed5e1fe8e9\") " pod="openstack/horizon-7d85fc8646-gv4lm" Oct 08 19:54:39 crc kubenswrapper[4988]: I1008 19:54:39.757081 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8hpsk\" (UniqueName: \"kubernetes.io/projected/c533a2e1-7213-4410-b862-5eed5e1fe8e9-kube-api-access-8hpsk\") pod \"horizon-7d85fc8646-gv4lm\" (UID: \"c533a2e1-7213-4410-b862-5eed5e1fe8e9\") " pod="openstack/horizon-7d85fc8646-gv4lm" Oct 08 19:54:39 crc kubenswrapper[4988]: I1008 19:54:39.784666 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-d95cf97bd-hmknh" Oct 08 19:54:39 crc kubenswrapper[4988]: I1008 19:54:39.845081 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-7d85fc8646-gv4lm" Oct 08 19:54:40 crc kubenswrapper[4988]: I1008 19:54:40.275113 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-d95cf97bd-hmknh"] Oct 08 19:54:40 crc kubenswrapper[4988]: W1008 19:54:40.279053 4988 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf45d9be7_457d_472d_8c9f_463f9d241eae.slice/crio-b9b01b86503ad5dc6c6e49585c6699cfae1d882b6e9c9bfa15c09a673663b3cb WatchSource:0}: Error finding container b9b01b86503ad5dc6c6e49585c6699cfae1d882b6e9c9bfa15c09a673663b3cb: Status 404 returned error can't find the container with id b9b01b86503ad5dc6c6e49585c6699cfae1d882b6e9c9bfa15c09a673663b3cb Oct 08 19:54:40 crc kubenswrapper[4988]: I1008 19:54:40.369065 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-7d85fc8646-gv4lm"] Oct 08 19:54:40 crc kubenswrapper[4988]: I1008 19:54:40.506543 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7d85fc8646-gv4lm" event={"ID":"c533a2e1-7213-4410-b862-5eed5e1fe8e9","Type":"ContainerStarted","Data":"60876b72a1bcd0717dd169ac584f8f5bd73e37c53808d766bf8e74b7823bf6cc"} Oct 08 19:54:40 crc kubenswrapper[4988]: I1008 19:54:40.508521 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-d95cf97bd-hmknh" event={"ID":"f45d9be7-457d-472d-8c9f-463f9d241eae","Type":"ContainerStarted","Data":"b9b01b86503ad5dc6c6e49585c6699cfae1d882b6e9c9bfa15c09a673663b3cb"} Oct 08 19:54:41 crc kubenswrapper[4988]: I1008 19:54:41.036703 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-sync-r4nvb"] Oct 08 19:54:41 crc kubenswrapper[4988]: I1008 19:54:41.046097 4988 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-sync-r4nvb"] Oct 08 19:54:41 crc kubenswrapper[4988]: I1008 19:54:41.248499 4988 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="74701b1e-3a3a-47e6-9794-46b76ec341e5" path="/var/lib/kubelet/pods/74701b1e-3a3a-47e6-9794-46b76ec341e5/volumes" Oct 08 19:54:41 crc kubenswrapper[4988]: I1008 19:54:41.519353 4988 generic.go:334] "Generic (PLEG): container finished" podID="d29618a6-2810-4d35-b5b3-2369d4a1c096" containerID="63520e1979dd97c54c047ccdf382e36d66fc5644ef9ac1bde7c697c98007cbdd" exitCode=0 Oct 08 19:54:41 crc kubenswrapper[4988]: I1008 19:54:41.519449 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"d29618a6-2810-4d35-b5b3-2369d4a1c096","Type":"ContainerDied","Data":"63520e1979dd97c54c047ccdf382e36d66fc5644ef9ac1bde7c697c98007cbdd"} Oct 08 19:54:41 crc kubenswrapper[4988]: I1008 19:54:41.521449 4988 generic.go:334] "Generic (PLEG): container finished" podID="185123ed-b356-4268-af3d-e44ea04bbf3f" containerID="1decc70dd2302ff2a788701c560d989e84796d90fcfb7cf0324d8c48e524d42e" exitCode=0 Oct 08 19:54:41 crc kubenswrapper[4988]: I1008 19:54:41.521477 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"185123ed-b356-4268-af3d-e44ea04bbf3f","Type":"ContainerDied","Data":"1decc70dd2302ff2a788701c560d989e84796d90fcfb7cf0324d8c48e524d42e"} Oct 08 19:54:45 crc kubenswrapper[4988]: I1008 19:54:45.889986 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 08 19:54:45 crc kubenswrapper[4988]: I1008 19:54:45.989771 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5fcjj\" (UniqueName: \"kubernetes.io/projected/185123ed-b356-4268-af3d-e44ea04bbf3f-kube-api-access-5fcjj\") pod \"185123ed-b356-4268-af3d-e44ea04bbf3f\" (UID: \"185123ed-b356-4268-af3d-e44ea04bbf3f\") " Oct 08 19:54:45 crc kubenswrapper[4988]: I1008 19:54:45.990225 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/185123ed-b356-4268-af3d-e44ea04bbf3f-combined-ca-bundle\") pod \"185123ed-b356-4268-af3d-e44ea04bbf3f\" (UID: \"185123ed-b356-4268-af3d-e44ea04bbf3f\") " Oct 08 19:54:45 crc kubenswrapper[4988]: I1008 19:54:45.990264 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/185123ed-b356-4268-af3d-e44ea04bbf3f-logs\") pod \"185123ed-b356-4268-af3d-e44ea04bbf3f\" (UID: \"185123ed-b356-4268-af3d-e44ea04bbf3f\") " Oct 08 19:54:45 crc kubenswrapper[4988]: I1008 19:54:45.990281 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/185123ed-b356-4268-af3d-e44ea04bbf3f-scripts\") pod \"185123ed-b356-4268-af3d-e44ea04bbf3f\" (UID: \"185123ed-b356-4268-af3d-e44ea04bbf3f\") " Oct 08 19:54:45 crc kubenswrapper[4988]: I1008 19:54:45.990356 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/185123ed-b356-4268-af3d-e44ea04bbf3f-httpd-run\") pod \"185123ed-b356-4268-af3d-e44ea04bbf3f\" (UID: \"185123ed-b356-4268-af3d-e44ea04bbf3f\") " Oct 08 19:54:45 crc kubenswrapper[4988]: I1008 19:54:45.990610 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/185123ed-b356-4268-af3d-e44ea04bbf3f-config-data\") pod \"185123ed-b356-4268-af3d-e44ea04bbf3f\" (UID: \"185123ed-b356-4268-af3d-e44ea04bbf3f\") " Oct 08 19:54:45 crc kubenswrapper[4988]: I1008 19:54:45.990650 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/185123ed-b356-4268-af3d-e44ea04bbf3f-public-tls-certs\") pod \"185123ed-b356-4268-af3d-e44ea04bbf3f\" (UID: \"185123ed-b356-4268-af3d-e44ea04bbf3f\") " Oct 08 19:54:45 crc kubenswrapper[4988]: I1008 19:54:45.991427 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/185123ed-b356-4268-af3d-e44ea04bbf3f-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "185123ed-b356-4268-af3d-e44ea04bbf3f" (UID: "185123ed-b356-4268-af3d-e44ea04bbf3f"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 19:54:45 crc kubenswrapper[4988]: I1008 19:54:45.996726 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/185123ed-b356-4268-af3d-e44ea04bbf3f-logs" (OuterVolumeSpecName: "logs") pod "185123ed-b356-4268-af3d-e44ea04bbf3f" (UID: "185123ed-b356-4268-af3d-e44ea04bbf3f"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 19:54:46 crc kubenswrapper[4988]: I1008 19:54:46.002604 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/185123ed-b356-4268-af3d-e44ea04bbf3f-scripts" (OuterVolumeSpecName: "scripts") pod "185123ed-b356-4268-af3d-e44ea04bbf3f" (UID: "185123ed-b356-4268-af3d-e44ea04bbf3f"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 19:54:46 crc kubenswrapper[4988]: I1008 19:54:46.010744 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/185123ed-b356-4268-af3d-e44ea04bbf3f-kube-api-access-5fcjj" (OuterVolumeSpecName: "kube-api-access-5fcjj") pod "185123ed-b356-4268-af3d-e44ea04bbf3f" (UID: "185123ed-b356-4268-af3d-e44ea04bbf3f"). InnerVolumeSpecName "kube-api-access-5fcjj". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 19:54:46 crc kubenswrapper[4988]: I1008 19:54:46.046113 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 08 19:54:46 crc kubenswrapper[4988]: I1008 19:54:46.074048 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/185123ed-b356-4268-af3d-e44ea04bbf3f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "185123ed-b356-4268-af3d-e44ea04bbf3f" (UID: "185123ed-b356-4268-af3d-e44ea04bbf3f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 19:54:46 crc kubenswrapper[4988]: I1008 19:54:46.095278 4988 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5fcjj\" (UniqueName: \"kubernetes.io/projected/185123ed-b356-4268-af3d-e44ea04bbf3f-kube-api-access-5fcjj\") on node \"crc\" DevicePath \"\"" Oct 08 19:54:46 crc kubenswrapper[4988]: I1008 19:54:46.095313 4988 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/185123ed-b356-4268-af3d-e44ea04bbf3f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 19:54:46 crc kubenswrapper[4988]: I1008 19:54:46.095325 4988 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/185123ed-b356-4268-af3d-e44ea04bbf3f-logs\") on node \"crc\" DevicePath \"\"" Oct 08 19:54:46 crc kubenswrapper[4988]: I1008 19:54:46.095337 4988 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/185123ed-b356-4268-af3d-e44ea04bbf3f-scripts\") on node \"crc\" DevicePath \"\"" Oct 08 19:54:46 crc kubenswrapper[4988]: I1008 19:54:46.095347 4988 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/185123ed-b356-4268-af3d-e44ea04bbf3f-httpd-run\") on node \"crc\" DevicePath \"\"" Oct 08 19:54:46 crc kubenswrapper[4988]: I1008 19:54:46.101815 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/185123ed-b356-4268-af3d-e44ea04bbf3f-config-data" (OuterVolumeSpecName: "config-data") pod "185123ed-b356-4268-af3d-e44ea04bbf3f" (UID: "185123ed-b356-4268-af3d-e44ea04bbf3f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 19:54:46 crc kubenswrapper[4988]: I1008 19:54:46.109295 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/185123ed-b356-4268-af3d-e44ea04bbf3f-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "185123ed-b356-4268-af3d-e44ea04bbf3f" (UID: "185123ed-b356-4268-af3d-e44ea04bbf3f"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 19:54:46 crc kubenswrapper[4988]: I1008 19:54:46.196092 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d29618a6-2810-4d35-b5b3-2369d4a1c096-scripts\") pod \"d29618a6-2810-4d35-b5b3-2369d4a1c096\" (UID: \"d29618a6-2810-4d35-b5b3-2369d4a1c096\") " Oct 08 19:54:46 crc kubenswrapper[4988]: I1008 19:54:46.196171 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d29618a6-2810-4d35-b5b3-2369d4a1c096-combined-ca-bundle\") pod \"d29618a6-2810-4d35-b5b3-2369d4a1c096\" (UID: \"d29618a6-2810-4d35-b5b3-2369d4a1c096\") " Oct 08 19:54:46 crc kubenswrapper[4988]: I1008 19:54:46.196472 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d29618a6-2810-4d35-b5b3-2369d4a1c096-logs\") pod \"d29618a6-2810-4d35-b5b3-2369d4a1c096\" (UID: \"d29618a6-2810-4d35-b5b3-2369d4a1c096\") " Oct 08 19:54:46 crc kubenswrapper[4988]: I1008 19:54:46.196505 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d29618a6-2810-4d35-b5b3-2369d4a1c096-config-data\") pod \"d29618a6-2810-4d35-b5b3-2369d4a1c096\" (UID: \"d29618a6-2810-4d35-b5b3-2369d4a1c096\") " Oct 08 19:54:46 crc kubenswrapper[4988]: I1008 19:54:46.196583 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d29618a6-2810-4d35-b5b3-2369d4a1c096-internal-tls-certs\") pod \"d29618a6-2810-4d35-b5b3-2369d4a1c096\" (UID: \"d29618a6-2810-4d35-b5b3-2369d4a1c096\") " Oct 08 19:54:46 crc kubenswrapper[4988]: I1008 19:54:46.196656 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gtpmt\" (UniqueName: \"kubernetes.io/projected/d29618a6-2810-4d35-b5b3-2369d4a1c096-kube-api-access-gtpmt\") pod \"d29618a6-2810-4d35-b5b3-2369d4a1c096\" (UID: \"d29618a6-2810-4d35-b5b3-2369d4a1c096\") " Oct 08 19:54:46 crc kubenswrapper[4988]: I1008 19:54:46.196746 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/d29618a6-2810-4d35-b5b3-2369d4a1c096-httpd-run\") pod \"d29618a6-2810-4d35-b5b3-2369d4a1c096\" (UID: \"d29618a6-2810-4d35-b5b3-2369d4a1c096\") " Oct 08 19:54:46 crc kubenswrapper[4988]: I1008 19:54:46.197117 4988 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/185123ed-b356-4268-af3d-e44ea04bbf3f-config-data\") on node \"crc\" DevicePath \"\"" Oct 08 19:54:46 crc kubenswrapper[4988]: I1008 19:54:46.197132 4988 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/185123ed-b356-4268-af3d-e44ea04bbf3f-public-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 08 19:54:46 crc kubenswrapper[4988]: I1008 19:54:46.197770 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d29618a6-2810-4d35-b5b3-2369d4a1c096-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "d29618a6-2810-4d35-b5b3-2369d4a1c096" (UID: "d29618a6-2810-4d35-b5b3-2369d4a1c096"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 19:54:46 crc kubenswrapper[4988]: I1008 19:54:46.201498 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d29618a6-2810-4d35-b5b3-2369d4a1c096-kube-api-access-gtpmt" (OuterVolumeSpecName: "kube-api-access-gtpmt") pod "d29618a6-2810-4d35-b5b3-2369d4a1c096" (UID: "d29618a6-2810-4d35-b5b3-2369d4a1c096"). InnerVolumeSpecName "kube-api-access-gtpmt". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 19:54:46 crc kubenswrapper[4988]: I1008 19:54:46.202501 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d29618a6-2810-4d35-b5b3-2369d4a1c096-logs" (OuterVolumeSpecName: "logs") pod "d29618a6-2810-4d35-b5b3-2369d4a1c096" (UID: "d29618a6-2810-4d35-b5b3-2369d4a1c096"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 19:54:46 crc kubenswrapper[4988]: I1008 19:54:46.202513 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d29618a6-2810-4d35-b5b3-2369d4a1c096-scripts" (OuterVolumeSpecName: "scripts") pod "d29618a6-2810-4d35-b5b3-2369d4a1c096" (UID: "d29618a6-2810-4d35-b5b3-2369d4a1c096"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 19:54:46 crc kubenswrapper[4988]: I1008 19:54:46.275267 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d29618a6-2810-4d35-b5b3-2369d4a1c096-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d29618a6-2810-4d35-b5b3-2369d4a1c096" (UID: "d29618a6-2810-4d35-b5b3-2369d4a1c096"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 19:54:46 crc kubenswrapper[4988]: I1008 19:54:46.299541 4988 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d29618a6-2810-4d35-b5b3-2369d4a1c096-scripts\") on node \"crc\" DevicePath \"\"" Oct 08 19:54:46 crc kubenswrapper[4988]: I1008 19:54:46.299571 4988 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d29618a6-2810-4d35-b5b3-2369d4a1c096-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 19:54:46 crc kubenswrapper[4988]: I1008 19:54:46.299584 4988 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d29618a6-2810-4d35-b5b3-2369d4a1c096-logs\") on node \"crc\" DevicePath \"\"" Oct 08 19:54:46 crc kubenswrapper[4988]: I1008 19:54:46.299617 4988 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gtpmt\" (UniqueName: \"kubernetes.io/projected/d29618a6-2810-4d35-b5b3-2369d4a1c096-kube-api-access-gtpmt\") on node \"crc\" DevicePath \"\"" Oct 08 19:54:46 crc kubenswrapper[4988]: I1008 19:54:46.299629 4988 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/d29618a6-2810-4d35-b5b3-2369d4a1c096-httpd-run\") on node \"crc\" DevicePath \"\"" Oct 08 19:54:46 crc kubenswrapper[4988]: I1008 19:54:46.332676 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d29618a6-2810-4d35-b5b3-2369d4a1c096-config-data" (OuterVolumeSpecName: "config-data") pod "d29618a6-2810-4d35-b5b3-2369d4a1c096" (UID: "d29618a6-2810-4d35-b5b3-2369d4a1c096"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 19:54:46 crc kubenswrapper[4988]: I1008 19:54:46.336506 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d29618a6-2810-4d35-b5b3-2369d4a1c096-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "d29618a6-2810-4d35-b5b3-2369d4a1c096" (UID: "d29618a6-2810-4d35-b5b3-2369d4a1c096"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 19:54:46 crc kubenswrapper[4988]: I1008 19:54:46.401257 4988 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d29618a6-2810-4d35-b5b3-2369d4a1c096-config-data\") on node \"crc\" DevicePath \"\"" Oct 08 19:54:46 crc kubenswrapper[4988]: I1008 19:54:46.401291 4988 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d29618a6-2810-4d35-b5b3-2369d4a1c096-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 08 19:54:46 crc kubenswrapper[4988]: I1008 19:54:46.595093 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-d95cf97bd-hmknh" event={"ID":"f45d9be7-457d-472d-8c9f-463f9d241eae","Type":"ContainerStarted","Data":"35268501f379cc03d23dd17a36b891c830fa111aea8c3f4c0dac3ce3f4b66d27"} Oct 08 19:54:46 crc kubenswrapper[4988]: I1008 19:54:46.595363 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-d95cf97bd-hmknh" event={"ID":"f45d9be7-457d-472d-8c9f-463f9d241eae","Type":"ContainerStarted","Data":"3044dddba8423c50138cfe88cad812bb1f3b03df4ece17ff02aad32480deb876"} Oct 08 19:54:46 crc kubenswrapper[4988]: I1008 19:54:46.598593 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-54cd997f77-cs4qd" event={"ID":"83fe9318-96ca-4f74-b838-9997e59c1ea2","Type":"ContainerStarted","Data":"b4342b0a8a8029ce5a9743a91ea0040e09494640e7d44f66a617c680c51cb14f"} Oct 08 19:54:46 crc kubenswrapper[4988]: I1008 19:54:46.598621 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-54cd997f77-cs4qd" event={"ID":"83fe9318-96ca-4f74-b838-9997e59c1ea2","Type":"ContainerStarted","Data":"02475e19a95215d237081d184062efcf0c3a50790eac23452187746453d8d6b8"} Oct 08 19:54:46 crc kubenswrapper[4988]: I1008 19:54:46.598712 4988 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-54cd997f77-cs4qd" podUID="83fe9318-96ca-4f74-b838-9997e59c1ea2" containerName="horizon-log" containerID="cri-o://02475e19a95215d237081d184062efcf0c3a50790eac23452187746453d8d6b8" gracePeriod=30 Oct 08 19:54:46 crc kubenswrapper[4988]: I1008 19:54:46.598826 4988 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-54cd997f77-cs4qd" podUID="83fe9318-96ca-4f74-b838-9997e59c1ea2" containerName="horizon" containerID="cri-o://b4342b0a8a8029ce5a9743a91ea0040e09494640e7d44f66a617c680c51cb14f" gracePeriod=30 Oct 08 19:54:46 crc kubenswrapper[4988]: I1008 19:54:46.602730 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7b9d74c97-mthd5" event={"ID":"322a3c03-38ac-4ef2-8980-c1976f855a81","Type":"ContainerStarted","Data":"1d4c823b49d5e59e790c219ce79ca123a64288d8978ef4875e9ade7e28b36d25"} Oct 08 19:54:46 crc kubenswrapper[4988]: I1008 19:54:46.602766 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7b9d74c97-mthd5" event={"ID":"322a3c03-38ac-4ef2-8980-c1976f855a81","Type":"ContainerStarted","Data":"1ae46144f520666a647fe5640b334d415fb0e6714b2c54837415df1f1b0a681d"} Oct 08 19:54:46 crc kubenswrapper[4988]: I1008 19:54:46.602858 4988 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-7b9d74c97-mthd5" podUID="322a3c03-38ac-4ef2-8980-c1976f855a81" containerName="horizon-log" containerID="cri-o://1ae46144f520666a647fe5640b334d415fb0e6714b2c54837415df1f1b0a681d" gracePeriod=30 Oct 08 19:54:46 crc kubenswrapper[4988]: I1008 19:54:46.602941 4988 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-7b9d74c97-mthd5" podUID="322a3c03-38ac-4ef2-8980-c1976f855a81" containerName="horizon" containerID="cri-o://1d4c823b49d5e59e790c219ce79ca123a64288d8978ef4875e9ade7e28b36d25" gracePeriod=30 Oct 08 19:54:46 crc kubenswrapper[4988]: I1008 19:54:46.610214 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"185123ed-b356-4268-af3d-e44ea04bbf3f","Type":"ContainerDied","Data":"61685588cdeb820835a56ec6a4dbaee1105b8876ff91f5008a50025350a6f30d"} Oct 08 19:54:46 crc kubenswrapper[4988]: I1008 19:54:46.610269 4988 scope.go:117] "RemoveContainer" containerID="1decc70dd2302ff2a788701c560d989e84796d90fcfb7cf0324d8c48e524d42e" Oct 08 19:54:46 crc kubenswrapper[4988]: I1008 19:54:46.610473 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 08 19:54:46 crc kubenswrapper[4988]: I1008 19:54:46.616268 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-d95cf97bd-hmknh" podStartSLOduration=2.284017036 podStartE2EDuration="7.616253654s" podCreationTimestamp="2025-10-08 19:54:39 +0000 UTC" firstStartedPulling="2025-10-08 19:54:40.282088552 +0000 UTC m=+6225.731931322" lastFinishedPulling="2025-10-08 19:54:45.61432517 +0000 UTC m=+6231.064167940" observedRunningTime="2025-10-08 19:54:46.613414993 +0000 UTC m=+6232.063257753" watchObservedRunningTime="2025-10-08 19:54:46.616253654 +0000 UTC m=+6232.066096424" Oct 08 19:54:46 crc kubenswrapper[4988]: I1008 19:54:46.634450 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7d85fc8646-gv4lm" event={"ID":"c533a2e1-7213-4410-b862-5eed5e1fe8e9","Type":"ContainerStarted","Data":"24a201cb0f08c481ca0dc4c546b32eeb4e83b514268ccaffc0f84cb043a4ef21"} Oct 08 19:54:46 crc kubenswrapper[4988]: I1008 19:54:46.634538 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7d85fc8646-gv4lm" event={"ID":"c533a2e1-7213-4410-b862-5eed5e1fe8e9","Type":"ContainerStarted","Data":"bfd13f227cb2064f23761c7a562d35816e2cb13aba4850cb48a1f6228a4798cc"} Oct 08 19:54:46 crc kubenswrapper[4988]: I1008 19:54:46.643121 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"d29618a6-2810-4d35-b5b3-2369d4a1c096","Type":"ContainerDied","Data":"5048f4c53afd16b33e344d6fd27661ebed068a0c390359e8edb0f0df863bbe36"} Oct 08 19:54:46 crc kubenswrapper[4988]: I1008 19:54:46.643210 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 08 19:54:46 crc kubenswrapper[4988]: I1008 19:54:46.652510 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-54cd997f77-cs4qd" podStartSLOduration=2.592051652 podStartE2EDuration="9.652487708s" podCreationTimestamp="2025-10-08 19:54:37 +0000 UTC" firstStartedPulling="2025-10-08 19:54:38.550989111 +0000 UTC m=+6224.000831881" lastFinishedPulling="2025-10-08 19:54:45.611425167 +0000 UTC m=+6231.061267937" observedRunningTime="2025-10-08 19:54:46.636559641 +0000 UTC m=+6232.086402411" watchObservedRunningTime="2025-10-08 19:54:46.652487708 +0000 UTC m=+6232.102330478" Oct 08 19:54:46 crc kubenswrapper[4988]: I1008 19:54:46.657339 4988 scope.go:117] "RemoveContainer" containerID="3e47ea1ab135b55cab053319b8c923376c4cda044c92fb262fdd29d259c417da" Oct 08 19:54:46 crc kubenswrapper[4988]: I1008 19:54:46.664682 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-7b9d74c97-mthd5" podStartSLOduration=2.508082917 podStartE2EDuration="9.664664956s" podCreationTimestamp="2025-10-08 19:54:37 +0000 UTC" firstStartedPulling="2025-10-08 19:54:38.455773508 +0000 UTC m=+6223.905616278" lastFinishedPulling="2025-10-08 19:54:45.612355557 +0000 UTC m=+6231.062198317" observedRunningTime="2025-10-08 19:54:46.654688219 +0000 UTC m=+6232.104530989" watchObservedRunningTime="2025-10-08 19:54:46.664664956 +0000 UTC m=+6232.114507726" Oct 08 19:54:46 crc kubenswrapper[4988]: I1008 19:54:46.677585 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 08 19:54:46 crc kubenswrapper[4988]: I1008 19:54:46.694722 4988 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 08 19:54:46 crc kubenswrapper[4988]: I1008 19:54:46.698003 4988 scope.go:117] "RemoveContainer" containerID="63520e1979dd97c54c047ccdf382e36d66fc5644ef9ac1bde7c697c98007cbdd" Oct 08 19:54:46 crc kubenswrapper[4988]: I1008 19:54:46.712755 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Oct 08 19:54:46 crc kubenswrapper[4988]: E1008 19:54:46.717859 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d29618a6-2810-4d35-b5b3-2369d4a1c096" containerName="glance-httpd" Oct 08 19:54:46 crc kubenswrapper[4988]: I1008 19:54:46.717888 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="d29618a6-2810-4d35-b5b3-2369d4a1c096" containerName="glance-httpd" Oct 08 19:54:46 crc kubenswrapper[4988]: E1008 19:54:46.717909 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d29618a6-2810-4d35-b5b3-2369d4a1c096" containerName="glance-log" Oct 08 19:54:46 crc kubenswrapper[4988]: I1008 19:54:46.717919 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="d29618a6-2810-4d35-b5b3-2369d4a1c096" containerName="glance-log" Oct 08 19:54:46 crc kubenswrapper[4988]: E1008 19:54:46.717951 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="185123ed-b356-4268-af3d-e44ea04bbf3f" containerName="glance-httpd" Oct 08 19:54:46 crc kubenswrapper[4988]: I1008 19:54:46.717957 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="185123ed-b356-4268-af3d-e44ea04bbf3f" containerName="glance-httpd" Oct 08 19:54:46 crc kubenswrapper[4988]: E1008 19:54:46.717974 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="185123ed-b356-4268-af3d-e44ea04bbf3f" containerName="glance-log" Oct 08 19:54:46 crc kubenswrapper[4988]: I1008 19:54:46.717980 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="185123ed-b356-4268-af3d-e44ea04bbf3f" containerName="glance-log" Oct 08 19:54:46 crc kubenswrapper[4988]: I1008 19:54:46.718181 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="185123ed-b356-4268-af3d-e44ea04bbf3f" containerName="glance-httpd" Oct 08 19:54:46 crc kubenswrapper[4988]: I1008 19:54:46.718196 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="d29618a6-2810-4d35-b5b3-2369d4a1c096" containerName="glance-httpd" Oct 08 19:54:46 crc kubenswrapper[4988]: I1008 19:54:46.718214 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="d29618a6-2810-4d35-b5b3-2369d4a1c096" containerName="glance-log" Oct 08 19:54:46 crc kubenswrapper[4988]: I1008 19:54:46.718228 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="185123ed-b356-4268-af3d-e44ea04bbf3f" containerName="glance-log" Oct 08 19:54:46 crc kubenswrapper[4988]: I1008 19:54:46.718324 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-7d85fc8646-gv4lm" podStartSLOduration=2.495873404 podStartE2EDuration="7.718300595s" podCreationTimestamp="2025-10-08 19:54:39 +0000 UTC" firstStartedPulling="2025-10-08 19:54:40.387656184 +0000 UTC m=+6225.837498954" lastFinishedPulling="2025-10-08 19:54:45.610083375 +0000 UTC m=+6231.059926145" observedRunningTime="2025-10-08 19:54:46.696594624 +0000 UTC m=+6232.146437394" watchObservedRunningTime="2025-10-08 19:54:46.718300595 +0000 UTC m=+6232.168143375" Oct 08 19:54:46 crc kubenswrapper[4988]: I1008 19:54:46.719317 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 08 19:54:46 crc kubenswrapper[4988]: I1008 19:54:46.722674 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Oct 08 19:54:46 crc kubenswrapper[4988]: I1008 19:54:46.722860 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-b6m4j" Oct 08 19:54:46 crc kubenswrapper[4988]: I1008 19:54:46.722959 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-scripts" Oct 08 19:54:46 crc kubenswrapper[4988]: I1008 19:54:46.729344 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Oct 08 19:54:46 crc kubenswrapper[4988]: I1008 19:54:46.735122 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 08 19:54:46 crc kubenswrapper[4988]: I1008 19:54:46.746634 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 08 19:54:46 crc kubenswrapper[4988]: I1008 19:54:46.748491 4988 scope.go:117] "RemoveContainer" containerID="77b13c99dddd5b6f4b11423db1a866862933fa4da7f96cd4e3f36292c533bb98" Oct 08 19:54:46 crc kubenswrapper[4988]: I1008 19:54:46.756556 4988 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 08 19:54:46 crc kubenswrapper[4988]: I1008 19:54:46.768629 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 08 19:54:46 crc kubenswrapper[4988]: I1008 19:54:46.770562 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 08 19:54:46 crc kubenswrapper[4988]: I1008 19:54:46.786936 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Oct 08 19:54:46 crc kubenswrapper[4988]: I1008 19:54:46.787307 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Oct 08 19:54:46 crc kubenswrapper[4988]: I1008 19:54:46.800538 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 08 19:54:46 crc kubenswrapper[4988]: I1008 19:54:46.810481 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b6e8a219-5fcb-4640-8672-8278e9e628a7-logs\") pod \"glance-default-external-api-0\" (UID: \"b6e8a219-5fcb-4640-8672-8278e9e628a7\") " pod="openstack/glance-default-external-api-0" Oct 08 19:54:46 crc kubenswrapper[4988]: I1008 19:54:46.810659 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b6e8a219-5fcb-4640-8672-8278e9e628a7-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"b6e8a219-5fcb-4640-8672-8278e9e628a7\") " pod="openstack/glance-default-external-api-0" Oct 08 19:54:46 crc kubenswrapper[4988]: I1008 19:54:46.810967 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b6e8a219-5fcb-4640-8672-8278e9e628a7-config-data\") pod \"glance-default-external-api-0\" (UID: \"b6e8a219-5fcb-4640-8672-8278e9e628a7\") " pod="openstack/glance-default-external-api-0" Oct 08 19:54:46 crc kubenswrapper[4988]: I1008 19:54:46.811330 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/b6e8a219-5fcb-4640-8672-8278e9e628a7-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"b6e8a219-5fcb-4640-8672-8278e9e628a7\") " pod="openstack/glance-default-external-api-0" Oct 08 19:54:46 crc kubenswrapper[4988]: I1008 19:54:46.811430 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d57b2\" (UniqueName: \"kubernetes.io/projected/b6e8a219-5fcb-4640-8672-8278e9e628a7-kube-api-access-d57b2\") pod \"glance-default-external-api-0\" (UID: \"b6e8a219-5fcb-4640-8672-8278e9e628a7\") " pod="openstack/glance-default-external-api-0" Oct 08 19:54:46 crc kubenswrapper[4988]: I1008 19:54:46.811462 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b6e8a219-5fcb-4640-8672-8278e9e628a7-scripts\") pod \"glance-default-external-api-0\" (UID: \"b6e8a219-5fcb-4640-8672-8278e9e628a7\") " pod="openstack/glance-default-external-api-0" Oct 08 19:54:46 crc kubenswrapper[4988]: I1008 19:54:46.811690 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/b6e8a219-5fcb-4640-8672-8278e9e628a7-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"b6e8a219-5fcb-4640-8672-8278e9e628a7\") " pod="openstack/glance-default-external-api-0" Oct 08 19:54:46 crc kubenswrapper[4988]: I1008 19:54:46.914407 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b6e8a219-5fcb-4640-8672-8278e9e628a7-logs\") pod \"glance-default-external-api-0\" (UID: \"b6e8a219-5fcb-4640-8672-8278e9e628a7\") " pod="openstack/glance-default-external-api-0" Oct 08 19:54:46 crc kubenswrapper[4988]: I1008 19:54:46.914516 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b6e8a219-5fcb-4640-8672-8278e9e628a7-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"b6e8a219-5fcb-4640-8672-8278e9e628a7\") " pod="openstack/glance-default-external-api-0" Oct 08 19:54:46 crc kubenswrapper[4988]: I1008 19:54:46.914846 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b6e8a219-5fcb-4640-8672-8278e9e628a7-logs\") pod \"glance-default-external-api-0\" (UID: \"b6e8a219-5fcb-4640-8672-8278e9e628a7\") " pod="openstack/glance-default-external-api-0" Oct 08 19:54:46 crc kubenswrapper[4988]: I1008 19:54:46.915787 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/2b071964-bede-4593-bf58-b5deaa573d05-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"2b071964-bede-4593-bf58-b5deaa573d05\") " pod="openstack/glance-default-internal-api-0" Oct 08 19:54:46 crc kubenswrapper[4988]: I1008 19:54:46.915871 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b6e8a219-5fcb-4640-8672-8278e9e628a7-config-data\") pod \"glance-default-external-api-0\" (UID: \"b6e8a219-5fcb-4640-8672-8278e9e628a7\") " pod="openstack/glance-default-external-api-0" Oct 08 19:54:46 crc kubenswrapper[4988]: I1008 19:54:46.915915 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/2b071964-bede-4593-bf58-b5deaa573d05-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"2b071964-bede-4593-bf58-b5deaa573d05\") " pod="openstack/glance-default-internal-api-0" Oct 08 19:54:46 crc kubenswrapper[4988]: I1008 19:54:46.915964 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2b071964-bede-4593-bf58-b5deaa573d05-config-data\") pod \"glance-default-internal-api-0\" (UID: \"2b071964-bede-4593-bf58-b5deaa573d05\") " pod="openstack/glance-default-internal-api-0" Oct 08 19:54:46 crc kubenswrapper[4988]: I1008 19:54:46.915995 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2b071964-bede-4593-bf58-b5deaa573d05-scripts\") pod \"glance-default-internal-api-0\" (UID: \"2b071964-bede-4593-bf58-b5deaa573d05\") " pod="openstack/glance-default-internal-api-0" Oct 08 19:54:46 crc kubenswrapper[4988]: I1008 19:54:46.916310 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2b071964-bede-4593-bf58-b5deaa573d05-logs\") pod \"glance-default-internal-api-0\" (UID: \"2b071964-bede-4593-bf58-b5deaa573d05\") " pod="openstack/glance-default-internal-api-0" Oct 08 19:54:46 crc kubenswrapper[4988]: I1008 19:54:46.916399 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/b6e8a219-5fcb-4640-8672-8278e9e628a7-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"b6e8a219-5fcb-4640-8672-8278e9e628a7\") " pod="openstack/glance-default-external-api-0" Oct 08 19:54:46 crc kubenswrapper[4988]: I1008 19:54:46.916510 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d57b2\" (UniqueName: \"kubernetes.io/projected/b6e8a219-5fcb-4640-8672-8278e9e628a7-kube-api-access-d57b2\") pod \"glance-default-external-api-0\" (UID: \"b6e8a219-5fcb-4640-8672-8278e9e628a7\") " pod="openstack/glance-default-external-api-0" Oct 08 19:54:46 crc kubenswrapper[4988]: I1008 19:54:46.916552 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b6e8a219-5fcb-4640-8672-8278e9e628a7-scripts\") pod \"glance-default-external-api-0\" (UID: \"b6e8a219-5fcb-4640-8672-8278e9e628a7\") " pod="openstack/glance-default-external-api-0" Oct 08 19:54:46 crc kubenswrapper[4988]: I1008 19:54:46.916589 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2b071964-bede-4593-bf58-b5deaa573d05-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"2b071964-bede-4593-bf58-b5deaa573d05\") " pod="openstack/glance-default-internal-api-0" Oct 08 19:54:46 crc kubenswrapper[4988]: I1008 19:54:46.916646 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/b6e8a219-5fcb-4640-8672-8278e9e628a7-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"b6e8a219-5fcb-4640-8672-8278e9e628a7\") " pod="openstack/glance-default-external-api-0" Oct 08 19:54:46 crc kubenswrapper[4988]: I1008 19:54:46.916718 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hnmj5\" (UniqueName: \"kubernetes.io/projected/2b071964-bede-4593-bf58-b5deaa573d05-kube-api-access-hnmj5\") pod \"glance-default-internal-api-0\" (UID: \"2b071964-bede-4593-bf58-b5deaa573d05\") " pod="openstack/glance-default-internal-api-0" Oct 08 19:54:46 crc kubenswrapper[4988]: I1008 19:54:46.916735 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/b6e8a219-5fcb-4640-8672-8278e9e628a7-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"b6e8a219-5fcb-4640-8672-8278e9e628a7\") " pod="openstack/glance-default-external-api-0" Oct 08 19:54:46 crc kubenswrapper[4988]: I1008 19:54:46.919689 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b6e8a219-5fcb-4640-8672-8278e9e628a7-scripts\") pod \"glance-default-external-api-0\" (UID: \"b6e8a219-5fcb-4640-8672-8278e9e628a7\") " pod="openstack/glance-default-external-api-0" Oct 08 19:54:46 crc kubenswrapper[4988]: I1008 19:54:46.919757 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b6e8a219-5fcb-4640-8672-8278e9e628a7-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"b6e8a219-5fcb-4640-8672-8278e9e628a7\") " pod="openstack/glance-default-external-api-0" Oct 08 19:54:46 crc kubenswrapper[4988]: I1008 19:54:46.920885 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/b6e8a219-5fcb-4640-8672-8278e9e628a7-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"b6e8a219-5fcb-4640-8672-8278e9e628a7\") " pod="openstack/glance-default-external-api-0" Oct 08 19:54:46 crc kubenswrapper[4988]: I1008 19:54:46.921852 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b6e8a219-5fcb-4640-8672-8278e9e628a7-config-data\") pod \"glance-default-external-api-0\" (UID: \"b6e8a219-5fcb-4640-8672-8278e9e628a7\") " pod="openstack/glance-default-external-api-0" Oct 08 19:54:46 crc kubenswrapper[4988]: I1008 19:54:46.937724 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d57b2\" (UniqueName: \"kubernetes.io/projected/b6e8a219-5fcb-4640-8672-8278e9e628a7-kube-api-access-d57b2\") pod \"glance-default-external-api-0\" (UID: \"b6e8a219-5fcb-4640-8672-8278e9e628a7\") " pod="openstack/glance-default-external-api-0" Oct 08 19:54:47 crc kubenswrapper[4988]: I1008 19:54:47.018908 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2b071964-bede-4593-bf58-b5deaa573d05-logs\") pod \"glance-default-internal-api-0\" (UID: \"2b071964-bede-4593-bf58-b5deaa573d05\") " pod="openstack/glance-default-internal-api-0" Oct 08 19:54:47 crc kubenswrapper[4988]: I1008 19:54:47.019059 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2b071964-bede-4593-bf58-b5deaa573d05-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"2b071964-bede-4593-bf58-b5deaa573d05\") " pod="openstack/glance-default-internal-api-0" Oct 08 19:54:47 crc kubenswrapper[4988]: I1008 19:54:47.019106 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hnmj5\" (UniqueName: \"kubernetes.io/projected/2b071964-bede-4593-bf58-b5deaa573d05-kube-api-access-hnmj5\") pod \"glance-default-internal-api-0\" (UID: \"2b071964-bede-4593-bf58-b5deaa573d05\") " pod="openstack/glance-default-internal-api-0" Oct 08 19:54:47 crc kubenswrapper[4988]: I1008 19:54:47.019174 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/2b071964-bede-4593-bf58-b5deaa573d05-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"2b071964-bede-4593-bf58-b5deaa573d05\") " pod="openstack/glance-default-internal-api-0" Oct 08 19:54:47 crc kubenswrapper[4988]: I1008 19:54:47.019200 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/2b071964-bede-4593-bf58-b5deaa573d05-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"2b071964-bede-4593-bf58-b5deaa573d05\") " pod="openstack/glance-default-internal-api-0" Oct 08 19:54:47 crc kubenswrapper[4988]: I1008 19:54:47.019222 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2b071964-bede-4593-bf58-b5deaa573d05-config-data\") pod \"glance-default-internal-api-0\" (UID: \"2b071964-bede-4593-bf58-b5deaa573d05\") " pod="openstack/glance-default-internal-api-0" Oct 08 19:54:47 crc kubenswrapper[4988]: I1008 19:54:47.019238 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2b071964-bede-4593-bf58-b5deaa573d05-scripts\") pod \"glance-default-internal-api-0\" (UID: \"2b071964-bede-4593-bf58-b5deaa573d05\") " pod="openstack/glance-default-internal-api-0" Oct 08 19:54:47 crc kubenswrapper[4988]: I1008 19:54:47.019714 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/2b071964-bede-4593-bf58-b5deaa573d05-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"2b071964-bede-4593-bf58-b5deaa573d05\") " pod="openstack/glance-default-internal-api-0" Oct 08 19:54:47 crc kubenswrapper[4988]: I1008 19:54:47.019741 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2b071964-bede-4593-bf58-b5deaa573d05-logs\") pod \"glance-default-internal-api-0\" (UID: \"2b071964-bede-4593-bf58-b5deaa573d05\") " pod="openstack/glance-default-internal-api-0" Oct 08 19:54:47 crc kubenswrapper[4988]: I1008 19:54:47.023143 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2b071964-bede-4593-bf58-b5deaa573d05-scripts\") pod \"glance-default-internal-api-0\" (UID: \"2b071964-bede-4593-bf58-b5deaa573d05\") " pod="openstack/glance-default-internal-api-0" Oct 08 19:54:47 crc kubenswrapper[4988]: I1008 19:54:47.028024 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2b071964-bede-4593-bf58-b5deaa573d05-config-data\") pod \"glance-default-internal-api-0\" (UID: \"2b071964-bede-4593-bf58-b5deaa573d05\") " pod="openstack/glance-default-internal-api-0" Oct 08 19:54:47 crc kubenswrapper[4988]: I1008 19:54:47.028847 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/2b071964-bede-4593-bf58-b5deaa573d05-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"2b071964-bede-4593-bf58-b5deaa573d05\") " pod="openstack/glance-default-internal-api-0" Oct 08 19:54:47 crc kubenswrapper[4988]: I1008 19:54:47.030114 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2b071964-bede-4593-bf58-b5deaa573d05-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"2b071964-bede-4593-bf58-b5deaa573d05\") " pod="openstack/glance-default-internal-api-0" Oct 08 19:54:47 crc kubenswrapper[4988]: I1008 19:54:47.043023 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 08 19:54:47 crc kubenswrapper[4988]: I1008 19:54:47.043594 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hnmj5\" (UniqueName: \"kubernetes.io/projected/2b071964-bede-4593-bf58-b5deaa573d05-kube-api-access-hnmj5\") pod \"glance-default-internal-api-0\" (UID: \"2b071964-bede-4593-bf58-b5deaa573d05\") " pod="openstack/glance-default-internal-api-0" Oct 08 19:54:47 crc kubenswrapper[4988]: I1008 19:54:47.101868 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 08 19:54:47 crc kubenswrapper[4988]: I1008 19:54:47.251744 4988 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="185123ed-b356-4268-af3d-e44ea04bbf3f" path="/var/lib/kubelet/pods/185123ed-b356-4268-af3d-e44ea04bbf3f/volumes" Oct 08 19:54:47 crc kubenswrapper[4988]: I1008 19:54:47.253037 4988 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d29618a6-2810-4d35-b5b3-2369d4a1c096" path="/var/lib/kubelet/pods/d29618a6-2810-4d35-b5b3-2369d4a1c096/volumes" Oct 08 19:54:47 crc kubenswrapper[4988]: I1008 19:54:47.626371 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 08 19:54:47 crc kubenswrapper[4988]: I1008 19:54:47.670336 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"b6e8a219-5fcb-4640-8672-8278e9e628a7","Type":"ContainerStarted","Data":"8b697b7f99b272c328ab9f416a1668e5c07c974f0415d3f8c6342e7d3b309a0b"} Oct 08 19:54:47 crc kubenswrapper[4988]: W1008 19:54:47.755375 4988 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2b071964_bede_4593_bf58_b5deaa573d05.slice/crio-bb4ab6992206490d4869740c5bae59e84f959aca2a63535d5a85268649ff8541 WatchSource:0}: Error finding container bb4ab6992206490d4869740c5bae59e84f959aca2a63535d5a85268649ff8541: Status 404 returned error can't find the container with id bb4ab6992206490d4869740c5bae59e84f959aca2a63535d5a85268649ff8541 Oct 08 19:54:47 crc kubenswrapper[4988]: I1008 19:54:47.766069 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 08 19:54:47 crc kubenswrapper[4988]: I1008 19:54:47.972090 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-7b9d74c97-mthd5" Oct 08 19:54:48 crc kubenswrapper[4988]: I1008 19:54:48.053430 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-54cd997f77-cs4qd" Oct 08 19:54:48 crc kubenswrapper[4988]: I1008 19:54:48.688271 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"b6e8a219-5fcb-4640-8672-8278e9e628a7","Type":"ContainerStarted","Data":"bce058db7f79df17b8ee547bfaa7d9a4fc864720bf9a5243ec9ababc242c6a6e"} Oct 08 19:54:48 crc kubenswrapper[4988]: I1008 19:54:48.691257 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"2b071964-bede-4593-bf58-b5deaa573d05","Type":"ContainerStarted","Data":"df0be386cb7d376ab03d4de6fdf0940bd54cf51cccf18f0c5f087388e56da787"} Oct 08 19:54:48 crc kubenswrapper[4988]: I1008 19:54:48.691299 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"2b071964-bede-4593-bf58-b5deaa573d05","Type":"ContainerStarted","Data":"bb4ab6992206490d4869740c5bae59e84f959aca2a63535d5a85268649ff8541"} Oct 08 19:54:49 crc kubenswrapper[4988]: I1008 19:54:49.704582 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"2b071964-bede-4593-bf58-b5deaa573d05","Type":"ContainerStarted","Data":"d4c9bd3183059fdd811aa3496446831e7ccfab39230c37e8ce5556544ba137f2"} Oct 08 19:54:49 crc kubenswrapper[4988]: I1008 19:54:49.707229 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"b6e8a219-5fcb-4640-8672-8278e9e628a7","Type":"ContainerStarted","Data":"0926b1ae6b02f8a7bcf8ae8e083a105cf9b21bf9d2bd37630ce042af6e2679cd"} Oct 08 19:54:49 crc kubenswrapper[4988]: I1008 19:54:49.736872 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=3.7368415649999998 podStartE2EDuration="3.736841565s" podCreationTimestamp="2025-10-08 19:54:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 19:54:49.72600864 +0000 UTC m=+6235.175851450" watchObservedRunningTime="2025-10-08 19:54:49.736841565 +0000 UTC m=+6235.186684355" Oct 08 19:54:49 crc kubenswrapper[4988]: I1008 19:54:49.753208 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=3.753187225 podStartE2EDuration="3.753187225s" podCreationTimestamp="2025-10-08 19:54:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 19:54:49.7407839 +0000 UTC m=+6235.190626670" watchObservedRunningTime="2025-10-08 19:54:49.753187225 +0000 UTC m=+6235.203029995" Oct 08 19:54:49 crc kubenswrapper[4988]: I1008 19:54:49.785306 4988 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-d95cf97bd-hmknh" Oct 08 19:54:49 crc kubenswrapper[4988]: I1008 19:54:49.785357 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-d95cf97bd-hmknh" Oct 08 19:54:49 crc kubenswrapper[4988]: I1008 19:54:49.846053 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-7d85fc8646-gv4lm" Oct 08 19:54:49 crc kubenswrapper[4988]: I1008 19:54:49.846130 4988 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-7d85fc8646-gv4lm" Oct 08 19:54:53 crc kubenswrapper[4988]: I1008 19:54:53.338334 4988 patch_prober.go:28] interesting pod/machine-config-daemon-nm54f container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 19:54:53 crc kubenswrapper[4988]: I1008 19:54:53.338996 4988 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 19:54:53 crc kubenswrapper[4988]: I1008 19:54:53.339060 4988 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" Oct 08 19:54:53 crc kubenswrapper[4988]: I1008 19:54:53.340043 4988 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"07bebacafd1c8af09c722a5994e896b519bd87cf3a710207584b6b2c926371f1"} pod="openshift-machine-config-operator/machine-config-daemon-nm54f" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 08 19:54:53 crc kubenswrapper[4988]: I1008 19:54:53.340126 4988 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" containerName="machine-config-daemon" containerID="cri-o://07bebacafd1c8af09c722a5994e896b519bd87cf3a710207584b6b2c926371f1" gracePeriod=600 Oct 08 19:54:53 crc kubenswrapper[4988]: I1008 19:54:53.757847 4988 generic.go:334] "Generic (PLEG): container finished" podID="f328a20b-5283-407e-8a93-2ab2da599f18" containerID="07bebacafd1c8af09c722a5994e896b519bd87cf3a710207584b6b2c926371f1" exitCode=0 Oct 08 19:54:53 crc kubenswrapper[4988]: I1008 19:54:53.758224 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" event={"ID":"f328a20b-5283-407e-8a93-2ab2da599f18","Type":"ContainerDied","Data":"07bebacafd1c8af09c722a5994e896b519bd87cf3a710207584b6b2c926371f1"} Oct 08 19:54:53 crc kubenswrapper[4988]: I1008 19:54:53.758265 4988 scope.go:117] "RemoveContainer" containerID="f2a29ed22dd7a48bbc929e65fbeceed969a669e78f65fbfb724b03dc0db6e947" Oct 08 19:54:54 crc kubenswrapper[4988]: I1008 19:54:54.770073 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" event={"ID":"f328a20b-5283-407e-8a93-2ab2da599f18","Type":"ContainerStarted","Data":"439af713ff941f973b8758bd6af3c097808f76ff520b96ff10bd3268811393f7"} Oct 08 19:54:57 crc kubenswrapper[4988]: I1008 19:54:57.044672 4988 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Oct 08 19:54:57 crc kubenswrapper[4988]: I1008 19:54:57.045216 4988 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Oct 08 19:54:57 crc kubenswrapper[4988]: I1008 19:54:57.102266 4988 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Oct 08 19:54:57 crc kubenswrapper[4988]: I1008 19:54:57.102359 4988 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Oct 08 19:54:57 crc kubenswrapper[4988]: I1008 19:54:57.103718 4988 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Oct 08 19:54:57 crc kubenswrapper[4988]: I1008 19:54:57.119208 4988 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Oct 08 19:54:57 crc kubenswrapper[4988]: I1008 19:54:57.176344 4988 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Oct 08 19:54:57 crc kubenswrapper[4988]: I1008 19:54:57.178521 4988 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Oct 08 19:54:57 crc kubenswrapper[4988]: I1008 19:54:57.798100 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Oct 08 19:54:57 crc kubenswrapper[4988]: I1008 19:54:57.798145 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Oct 08 19:54:57 crc kubenswrapper[4988]: I1008 19:54:57.798160 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Oct 08 19:54:57 crc kubenswrapper[4988]: I1008 19:54:57.798172 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Oct 08 19:54:59 crc kubenswrapper[4988]: I1008 19:54:59.787359 4988 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-d95cf97bd-hmknh" podUID="f45d9be7-457d-472d-8c9f-463f9d241eae" containerName="horizon" probeResult="failure" output="Get \"https://10.217.1.121:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.1.121:8443: connect: connection refused" Oct 08 19:54:59 crc kubenswrapper[4988]: I1008 19:54:59.848215 4988 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-7d85fc8646-gv4lm" podUID="c533a2e1-7213-4410-b862-5eed5e1fe8e9" containerName="horizon" probeResult="failure" output="Get \"https://10.217.1.122:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.1.122:8443: connect: connection refused" Oct 08 19:54:59 crc kubenswrapper[4988]: I1008 19:54:59.893075 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Oct 08 19:54:59 crc kubenswrapper[4988]: I1008 19:54:59.893186 4988 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 08 19:54:59 crc kubenswrapper[4988]: I1008 19:54:59.899001 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Oct 08 19:55:00 crc kubenswrapper[4988]: I1008 19:55:00.117412 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Oct 08 19:55:00 crc kubenswrapper[4988]: I1008 19:55:00.117496 4988 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 08 19:55:00 crc kubenswrapper[4988]: I1008 19:55:00.622959 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Oct 08 19:55:10 crc kubenswrapper[4988]: I1008 19:55:10.854783 4988 scope.go:117] "RemoveContainer" containerID="6a6df4cf78aeb6f22196fcbff02c8005b18ea265c81fd0b5af42b23ab4537dff" Oct 08 19:55:10 crc kubenswrapper[4988]: I1008 19:55:10.884282 4988 scope.go:117] "RemoveContainer" containerID="49f54ff31e8d408344336acfa6472b80772a333be76bcac7d65e889556f469c7" Oct 08 19:55:10 crc kubenswrapper[4988]: I1008 19:55:10.945577 4988 scope.go:117] "RemoveContainer" containerID="dff17961ab3ec0388707e7e082251560ce5645182a16215c1a0cc4d8267e7db6" Oct 08 19:55:11 crc kubenswrapper[4988]: I1008 19:55:11.482069 4988 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/horizon-d95cf97bd-hmknh" Oct 08 19:55:11 crc kubenswrapper[4988]: I1008 19:55:11.734109 4988 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/horizon-7d85fc8646-gv4lm" Oct 08 19:55:13 crc kubenswrapper[4988]: I1008 19:55:13.136275 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/horizon-d95cf97bd-hmknh" Oct 08 19:55:13 crc kubenswrapper[4988]: I1008 19:55:13.458077 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/horizon-7d85fc8646-gv4lm" Oct 08 19:55:13 crc kubenswrapper[4988]: I1008 19:55:13.577528 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-d95cf97bd-hmknh"] Oct 08 19:55:14 crc kubenswrapper[4988]: I1008 19:55:14.037648 4988 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-d95cf97bd-hmknh" podUID="f45d9be7-457d-472d-8c9f-463f9d241eae" containerName="horizon-log" containerID="cri-o://3044dddba8423c50138cfe88cad812bb1f3b03df4ece17ff02aad32480deb876" gracePeriod=30 Oct 08 19:55:14 crc kubenswrapper[4988]: I1008 19:55:14.037747 4988 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-d95cf97bd-hmknh" podUID="f45d9be7-457d-472d-8c9f-463f9d241eae" containerName="horizon" containerID="cri-o://35268501f379cc03d23dd17a36b891c830fa111aea8c3f4c0dac3ce3f4b66d27" gracePeriod=30 Oct 08 19:55:16 crc kubenswrapper[4988]: E1008 19:55:16.966226 4988 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod83fe9318_96ca_4f74_b838_9997e59c1ea2.slice/crio-b4342b0a8a8029ce5a9743a91ea0040e09494640e7d44f66a617c680c51cb14f.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod322a3c03_38ac_4ef2_8980_c1976f855a81.slice/crio-1ae46144f520666a647fe5640b334d415fb0e6714b2c54837415df1f1b0a681d.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod83fe9318_96ca_4f74_b838_9997e59c1ea2.slice/crio-conmon-02475e19a95215d237081d184062efcf0c3a50790eac23452187746453d8d6b8.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod322a3c03_38ac_4ef2_8980_c1976f855a81.slice/crio-1d4c823b49d5e59e790c219ce79ca123a64288d8978ef4875e9ade7e28b36d25.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod322a3c03_38ac_4ef2_8980_c1976f855a81.slice/crio-conmon-1d4c823b49d5e59e790c219ce79ca123a64288d8978ef4875e9ade7e28b36d25.scope\": RecentStats: unable to find data in memory cache]" Oct 08 19:55:17 crc kubenswrapper[4988]: I1008 19:55:17.079809 4988 generic.go:334] "Generic (PLEG): container finished" podID="83fe9318-96ca-4f74-b838-9997e59c1ea2" containerID="b4342b0a8a8029ce5a9743a91ea0040e09494640e7d44f66a617c680c51cb14f" exitCode=137 Oct 08 19:55:17 crc kubenswrapper[4988]: I1008 19:55:17.079913 4988 generic.go:334] "Generic (PLEG): container finished" podID="83fe9318-96ca-4f74-b838-9997e59c1ea2" containerID="02475e19a95215d237081d184062efcf0c3a50790eac23452187746453d8d6b8" exitCode=137 Oct 08 19:55:17 crc kubenswrapper[4988]: I1008 19:55:17.079998 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-54cd997f77-cs4qd" event={"ID":"83fe9318-96ca-4f74-b838-9997e59c1ea2","Type":"ContainerDied","Data":"b4342b0a8a8029ce5a9743a91ea0040e09494640e7d44f66a617c680c51cb14f"} Oct 08 19:55:17 crc kubenswrapper[4988]: I1008 19:55:17.080108 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-54cd997f77-cs4qd" event={"ID":"83fe9318-96ca-4f74-b838-9997e59c1ea2","Type":"ContainerDied","Data":"02475e19a95215d237081d184062efcf0c3a50790eac23452187746453d8d6b8"} Oct 08 19:55:17 crc kubenswrapper[4988]: I1008 19:55:17.082845 4988 generic.go:334] "Generic (PLEG): container finished" podID="322a3c03-38ac-4ef2-8980-c1976f855a81" containerID="1d4c823b49d5e59e790c219ce79ca123a64288d8978ef4875e9ade7e28b36d25" exitCode=137 Oct 08 19:55:17 crc kubenswrapper[4988]: I1008 19:55:17.082935 4988 generic.go:334] "Generic (PLEG): container finished" podID="322a3c03-38ac-4ef2-8980-c1976f855a81" containerID="1ae46144f520666a647fe5640b334d415fb0e6714b2c54837415df1f1b0a681d" exitCode=137 Oct 08 19:55:17 crc kubenswrapper[4988]: I1008 19:55:17.083030 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7b9d74c97-mthd5" event={"ID":"322a3c03-38ac-4ef2-8980-c1976f855a81","Type":"ContainerDied","Data":"1d4c823b49d5e59e790c219ce79ca123a64288d8978ef4875e9ade7e28b36d25"} Oct 08 19:55:17 crc kubenswrapper[4988]: I1008 19:55:17.083090 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7b9d74c97-mthd5" event={"ID":"322a3c03-38ac-4ef2-8980-c1976f855a81","Type":"ContainerDied","Data":"1ae46144f520666a647fe5640b334d415fb0e6714b2c54837415df1f1b0a681d"} Oct 08 19:55:17 crc kubenswrapper[4988]: I1008 19:55:17.083145 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7b9d74c97-mthd5" event={"ID":"322a3c03-38ac-4ef2-8980-c1976f855a81","Type":"ContainerDied","Data":"ed1ba5812189b29ecd1ed422dc3e95f1ab70f71ba0dad04343513bf59fdf71a2"} Oct 08 19:55:17 crc kubenswrapper[4988]: I1008 19:55:17.083215 4988 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ed1ba5812189b29ecd1ed422dc3e95f1ab70f71ba0dad04343513bf59fdf71a2" Oct 08 19:55:17 crc kubenswrapper[4988]: I1008 19:55:17.105782 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-7b9d74c97-mthd5" Oct 08 19:55:17 crc kubenswrapper[4988]: I1008 19:55:17.118250 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-54cd997f77-cs4qd" Oct 08 19:55:17 crc kubenswrapper[4988]: I1008 19:55:17.212937 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/322a3c03-38ac-4ef2-8980-c1976f855a81-logs\") pod \"322a3c03-38ac-4ef2-8980-c1976f855a81\" (UID: \"322a3c03-38ac-4ef2-8980-c1976f855a81\") " Oct 08 19:55:17 crc kubenswrapper[4988]: I1008 19:55:17.212994 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8bm6z\" (UniqueName: \"kubernetes.io/projected/83fe9318-96ca-4f74-b838-9997e59c1ea2-kube-api-access-8bm6z\") pod \"83fe9318-96ca-4f74-b838-9997e59c1ea2\" (UID: \"83fe9318-96ca-4f74-b838-9997e59c1ea2\") " Oct 08 19:55:17 crc kubenswrapper[4988]: I1008 19:55:17.213030 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/83fe9318-96ca-4f74-b838-9997e59c1ea2-horizon-secret-key\") pod \"83fe9318-96ca-4f74-b838-9997e59c1ea2\" (UID: \"83fe9318-96ca-4f74-b838-9997e59c1ea2\") " Oct 08 19:55:17 crc kubenswrapper[4988]: I1008 19:55:17.213085 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9ppgg\" (UniqueName: \"kubernetes.io/projected/322a3c03-38ac-4ef2-8980-c1976f855a81-kube-api-access-9ppgg\") pod \"322a3c03-38ac-4ef2-8980-c1976f855a81\" (UID: \"322a3c03-38ac-4ef2-8980-c1976f855a81\") " Oct 08 19:55:17 crc kubenswrapper[4988]: I1008 19:55:17.213129 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/322a3c03-38ac-4ef2-8980-c1976f855a81-config-data\") pod \"322a3c03-38ac-4ef2-8980-c1976f855a81\" (UID: \"322a3c03-38ac-4ef2-8980-c1976f855a81\") " Oct 08 19:55:17 crc kubenswrapper[4988]: I1008 19:55:17.213191 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/83fe9318-96ca-4f74-b838-9997e59c1ea2-logs\") pod \"83fe9318-96ca-4f74-b838-9997e59c1ea2\" (UID: \"83fe9318-96ca-4f74-b838-9997e59c1ea2\") " Oct 08 19:55:17 crc kubenswrapper[4988]: I1008 19:55:17.213229 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/83fe9318-96ca-4f74-b838-9997e59c1ea2-scripts\") pod \"83fe9318-96ca-4f74-b838-9997e59c1ea2\" (UID: \"83fe9318-96ca-4f74-b838-9997e59c1ea2\") " Oct 08 19:55:17 crc kubenswrapper[4988]: I1008 19:55:17.213284 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/322a3c03-38ac-4ef2-8980-c1976f855a81-scripts\") pod \"322a3c03-38ac-4ef2-8980-c1976f855a81\" (UID: \"322a3c03-38ac-4ef2-8980-c1976f855a81\") " Oct 08 19:55:17 crc kubenswrapper[4988]: I1008 19:55:17.213318 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/83fe9318-96ca-4f74-b838-9997e59c1ea2-config-data\") pod \"83fe9318-96ca-4f74-b838-9997e59c1ea2\" (UID: \"83fe9318-96ca-4f74-b838-9997e59c1ea2\") " Oct 08 19:55:17 crc kubenswrapper[4988]: I1008 19:55:17.213414 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/322a3c03-38ac-4ef2-8980-c1976f855a81-horizon-secret-key\") pod \"322a3c03-38ac-4ef2-8980-c1976f855a81\" (UID: \"322a3c03-38ac-4ef2-8980-c1976f855a81\") " Oct 08 19:55:17 crc kubenswrapper[4988]: I1008 19:55:17.213550 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/322a3c03-38ac-4ef2-8980-c1976f855a81-logs" (OuterVolumeSpecName: "logs") pod "322a3c03-38ac-4ef2-8980-c1976f855a81" (UID: "322a3c03-38ac-4ef2-8980-c1976f855a81"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 19:55:17 crc kubenswrapper[4988]: I1008 19:55:17.214033 4988 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/322a3c03-38ac-4ef2-8980-c1976f855a81-logs\") on node \"crc\" DevicePath \"\"" Oct 08 19:55:17 crc kubenswrapper[4988]: I1008 19:55:17.214721 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/83fe9318-96ca-4f74-b838-9997e59c1ea2-logs" (OuterVolumeSpecName: "logs") pod "83fe9318-96ca-4f74-b838-9997e59c1ea2" (UID: "83fe9318-96ca-4f74-b838-9997e59c1ea2"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 19:55:17 crc kubenswrapper[4988]: I1008 19:55:17.219612 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/322a3c03-38ac-4ef2-8980-c1976f855a81-kube-api-access-9ppgg" (OuterVolumeSpecName: "kube-api-access-9ppgg") pod "322a3c03-38ac-4ef2-8980-c1976f855a81" (UID: "322a3c03-38ac-4ef2-8980-c1976f855a81"). InnerVolumeSpecName "kube-api-access-9ppgg". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 19:55:17 crc kubenswrapper[4988]: I1008 19:55:17.220004 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/83fe9318-96ca-4f74-b838-9997e59c1ea2-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "83fe9318-96ca-4f74-b838-9997e59c1ea2" (UID: "83fe9318-96ca-4f74-b838-9997e59c1ea2"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 19:55:17 crc kubenswrapper[4988]: I1008 19:55:17.222601 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/322a3c03-38ac-4ef2-8980-c1976f855a81-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "322a3c03-38ac-4ef2-8980-c1976f855a81" (UID: "322a3c03-38ac-4ef2-8980-c1976f855a81"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 19:55:17 crc kubenswrapper[4988]: I1008 19:55:17.225044 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/83fe9318-96ca-4f74-b838-9997e59c1ea2-kube-api-access-8bm6z" (OuterVolumeSpecName: "kube-api-access-8bm6z") pod "83fe9318-96ca-4f74-b838-9997e59c1ea2" (UID: "83fe9318-96ca-4f74-b838-9997e59c1ea2"). InnerVolumeSpecName "kube-api-access-8bm6z". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 19:55:17 crc kubenswrapper[4988]: I1008 19:55:17.241319 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/322a3c03-38ac-4ef2-8980-c1976f855a81-config-data" (OuterVolumeSpecName: "config-data") pod "322a3c03-38ac-4ef2-8980-c1976f855a81" (UID: "322a3c03-38ac-4ef2-8980-c1976f855a81"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 19:55:17 crc kubenswrapper[4988]: I1008 19:55:17.251484 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/83fe9318-96ca-4f74-b838-9997e59c1ea2-config-data" (OuterVolumeSpecName: "config-data") pod "83fe9318-96ca-4f74-b838-9997e59c1ea2" (UID: "83fe9318-96ca-4f74-b838-9997e59c1ea2"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 19:55:17 crc kubenswrapper[4988]: I1008 19:55:17.252983 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/83fe9318-96ca-4f74-b838-9997e59c1ea2-scripts" (OuterVolumeSpecName: "scripts") pod "83fe9318-96ca-4f74-b838-9997e59c1ea2" (UID: "83fe9318-96ca-4f74-b838-9997e59c1ea2"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 19:55:17 crc kubenswrapper[4988]: I1008 19:55:17.257295 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/322a3c03-38ac-4ef2-8980-c1976f855a81-scripts" (OuterVolumeSpecName: "scripts") pod "322a3c03-38ac-4ef2-8980-c1976f855a81" (UID: "322a3c03-38ac-4ef2-8980-c1976f855a81"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 19:55:17 crc kubenswrapper[4988]: I1008 19:55:17.316672 4988 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/322a3c03-38ac-4ef2-8980-c1976f855a81-scripts\") on node \"crc\" DevicePath \"\"" Oct 08 19:55:17 crc kubenswrapper[4988]: I1008 19:55:17.316719 4988 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/83fe9318-96ca-4f74-b838-9997e59c1ea2-config-data\") on node \"crc\" DevicePath \"\"" Oct 08 19:55:17 crc kubenswrapper[4988]: I1008 19:55:17.316733 4988 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/322a3c03-38ac-4ef2-8980-c1976f855a81-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Oct 08 19:55:17 crc kubenswrapper[4988]: I1008 19:55:17.316747 4988 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8bm6z\" (UniqueName: \"kubernetes.io/projected/83fe9318-96ca-4f74-b838-9997e59c1ea2-kube-api-access-8bm6z\") on node \"crc\" DevicePath \"\"" Oct 08 19:55:17 crc kubenswrapper[4988]: I1008 19:55:17.316759 4988 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/83fe9318-96ca-4f74-b838-9997e59c1ea2-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Oct 08 19:55:17 crc kubenswrapper[4988]: I1008 19:55:17.316770 4988 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9ppgg\" (UniqueName: \"kubernetes.io/projected/322a3c03-38ac-4ef2-8980-c1976f855a81-kube-api-access-9ppgg\") on node \"crc\" DevicePath \"\"" Oct 08 19:55:17 crc kubenswrapper[4988]: I1008 19:55:17.316780 4988 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/322a3c03-38ac-4ef2-8980-c1976f855a81-config-data\") on node \"crc\" DevicePath \"\"" Oct 08 19:55:17 crc kubenswrapper[4988]: I1008 19:55:17.316791 4988 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/83fe9318-96ca-4f74-b838-9997e59c1ea2-logs\") on node \"crc\" DevicePath \"\"" Oct 08 19:55:17 crc kubenswrapper[4988]: I1008 19:55:17.316803 4988 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/83fe9318-96ca-4f74-b838-9997e59c1ea2-scripts\") on node \"crc\" DevicePath \"\"" Oct 08 19:55:18 crc kubenswrapper[4988]: I1008 19:55:18.101345 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-54cd997f77-cs4qd" Oct 08 19:55:18 crc kubenswrapper[4988]: I1008 19:55:18.101337 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-54cd997f77-cs4qd" event={"ID":"83fe9318-96ca-4f74-b838-9997e59c1ea2","Type":"ContainerDied","Data":"8fef69f66e9971ee638657baf4eeae76feafb4bfec6e030bc00f580c5aa86a92"} Oct 08 19:55:18 crc kubenswrapper[4988]: I1008 19:55:18.102045 4988 scope.go:117] "RemoveContainer" containerID="b4342b0a8a8029ce5a9743a91ea0040e09494640e7d44f66a617c680c51cb14f" Oct 08 19:55:18 crc kubenswrapper[4988]: I1008 19:55:18.106091 4988 generic.go:334] "Generic (PLEG): container finished" podID="f45d9be7-457d-472d-8c9f-463f9d241eae" containerID="35268501f379cc03d23dd17a36b891c830fa111aea8c3f4c0dac3ce3f4b66d27" exitCode=0 Oct 08 19:55:18 crc kubenswrapper[4988]: I1008 19:55:18.106199 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-7b9d74c97-mthd5" Oct 08 19:55:18 crc kubenswrapper[4988]: I1008 19:55:18.106335 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-d95cf97bd-hmknh" event={"ID":"f45d9be7-457d-472d-8c9f-463f9d241eae","Type":"ContainerDied","Data":"35268501f379cc03d23dd17a36b891c830fa111aea8c3f4c0dac3ce3f4b66d27"} Oct 08 19:55:18 crc kubenswrapper[4988]: I1008 19:55:18.153475 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-54cd997f77-cs4qd"] Oct 08 19:55:18 crc kubenswrapper[4988]: I1008 19:55:18.164460 4988 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-54cd997f77-cs4qd"] Oct 08 19:55:18 crc kubenswrapper[4988]: I1008 19:55:18.176800 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-7b9d74c97-mthd5"] Oct 08 19:55:18 crc kubenswrapper[4988]: I1008 19:55:18.187053 4988 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-7b9d74c97-mthd5"] Oct 08 19:55:18 crc kubenswrapper[4988]: I1008 19:55:18.299744 4988 scope.go:117] "RemoveContainer" containerID="02475e19a95215d237081d184062efcf0c3a50790eac23452187746453d8d6b8" Oct 08 19:55:19 crc kubenswrapper[4988]: I1008 19:55:19.253871 4988 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="322a3c03-38ac-4ef2-8980-c1976f855a81" path="/var/lib/kubelet/pods/322a3c03-38ac-4ef2-8980-c1976f855a81/volumes" Oct 08 19:55:19 crc kubenswrapper[4988]: I1008 19:55:19.255374 4988 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="83fe9318-96ca-4f74-b838-9997e59c1ea2" path="/var/lib/kubelet/pods/83fe9318-96ca-4f74-b838-9997e59c1ea2/volumes" Oct 08 19:55:19 crc kubenswrapper[4988]: I1008 19:55:19.785644 4988 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-d95cf97bd-hmknh" podUID="f45d9be7-457d-472d-8c9f-463f9d241eae" containerName="horizon" probeResult="failure" output="Get \"https://10.217.1.121:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.1.121:8443: connect: connection refused" Oct 08 19:55:29 crc kubenswrapper[4988]: I1008 19:55:29.785372 4988 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-d95cf97bd-hmknh" podUID="f45d9be7-457d-472d-8c9f-463f9d241eae" containerName="horizon" probeResult="failure" output="Get \"https://10.217.1.121:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.1.121:8443: connect: connection refused" Oct 08 19:55:39 crc kubenswrapper[4988]: I1008 19:55:39.790161 4988 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-d95cf97bd-hmknh" podUID="f45d9be7-457d-472d-8c9f-463f9d241eae" containerName="horizon" probeResult="failure" output="Get \"https://10.217.1.121:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.1.121:8443: connect: connection refused" Oct 08 19:55:39 crc kubenswrapper[4988]: I1008 19:55:39.791046 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-d95cf97bd-hmknh" Oct 08 19:55:44 crc kubenswrapper[4988]: I1008 19:55:44.420579 4988 generic.go:334] "Generic (PLEG): container finished" podID="f45d9be7-457d-472d-8c9f-463f9d241eae" containerID="3044dddba8423c50138cfe88cad812bb1f3b03df4ece17ff02aad32480deb876" exitCode=137 Oct 08 19:55:44 crc kubenswrapper[4988]: I1008 19:55:44.420676 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-d95cf97bd-hmknh" event={"ID":"f45d9be7-457d-472d-8c9f-463f9d241eae","Type":"ContainerDied","Data":"3044dddba8423c50138cfe88cad812bb1f3b03df4ece17ff02aad32480deb876"} Oct 08 19:55:44 crc kubenswrapper[4988]: I1008 19:55:44.549963 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-d95cf97bd-hmknh" Oct 08 19:55:44 crc kubenswrapper[4988]: I1008 19:55:44.723306 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/f45d9be7-457d-472d-8c9f-463f9d241eae-horizon-tls-certs\") pod \"f45d9be7-457d-472d-8c9f-463f9d241eae\" (UID: \"f45d9be7-457d-472d-8c9f-463f9d241eae\") " Oct 08 19:55:44 crc kubenswrapper[4988]: I1008 19:55:44.723415 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/f45d9be7-457d-472d-8c9f-463f9d241eae-config-data\") pod \"f45d9be7-457d-472d-8c9f-463f9d241eae\" (UID: \"f45d9be7-457d-472d-8c9f-463f9d241eae\") " Oct 08 19:55:44 crc kubenswrapper[4988]: I1008 19:55:44.723463 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f45d9be7-457d-472d-8c9f-463f9d241eae-combined-ca-bundle\") pod \"f45d9be7-457d-472d-8c9f-463f9d241eae\" (UID: \"f45d9be7-457d-472d-8c9f-463f9d241eae\") " Oct 08 19:55:44 crc kubenswrapper[4988]: I1008 19:55:44.723580 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-klg9v\" (UniqueName: \"kubernetes.io/projected/f45d9be7-457d-472d-8c9f-463f9d241eae-kube-api-access-klg9v\") pod \"f45d9be7-457d-472d-8c9f-463f9d241eae\" (UID: \"f45d9be7-457d-472d-8c9f-463f9d241eae\") " Oct 08 19:55:44 crc kubenswrapper[4988]: I1008 19:55:44.723619 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f45d9be7-457d-472d-8c9f-463f9d241eae-scripts\") pod \"f45d9be7-457d-472d-8c9f-463f9d241eae\" (UID: \"f45d9be7-457d-472d-8c9f-463f9d241eae\") " Oct 08 19:55:44 crc kubenswrapper[4988]: I1008 19:55:44.723659 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/f45d9be7-457d-472d-8c9f-463f9d241eae-horizon-secret-key\") pod \"f45d9be7-457d-472d-8c9f-463f9d241eae\" (UID: \"f45d9be7-457d-472d-8c9f-463f9d241eae\") " Oct 08 19:55:44 crc kubenswrapper[4988]: I1008 19:55:44.723858 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f45d9be7-457d-472d-8c9f-463f9d241eae-logs\") pod \"f45d9be7-457d-472d-8c9f-463f9d241eae\" (UID: \"f45d9be7-457d-472d-8c9f-463f9d241eae\") " Oct 08 19:55:44 crc kubenswrapper[4988]: I1008 19:55:44.725798 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f45d9be7-457d-472d-8c9f-463f9d241eae-logs" (OuterVolumeSpecName: "logs") pod "f45d9be7-457d-472d-8c9f-463f9d241eae" (UID: "f45d9be7-457d-472d-8c9f-463f9d241eae"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 19:55:44 crc kubenswrapper[4988]: I1008 19:55:44.732327 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f45d9be7-457d-472d-8c9f-463f9d241eae-kube-api-access-klg9v" (OuterVolumeSpecName: "kube-api-access-klg9v") pod "f45d9be7-457d-472d-8c9f-463f9d241eae" (UID: "f45d9be7-457d-472d-8c9f-463f9d241eae"). InnerVolumeSpecName "kube-api-access-klg9v". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 19:55:44 crc kubenswrapper[4988]: I1008 19:55:44.732608 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f45d9be7-457d-472d-8c9f-463f9d241eae-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "f45d9be7-457d-472d-8c9f-463f9d241eae" (UID: "f45d9be7-457d-472d-8c9f-463f9d241eae"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 19:55:44 crc kubenswrapper[4988]: I1008 19:55:44.753133 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f45d9be7-457d-472d-8c9f-463f9d241eae-config-data" (OuterVolumeSpecName: "config-data") pod "f45d9be7-457d-472d-8c9f-463f9d241eae" (UID: "f45d9be7-457d-472d-8c9f-463f9d241eae"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 19:55:44 crc kubenswrapper[4988]: I1008 19:55:44.785491 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f45d9be7-457d-472d-8c9f-463f9d241eae-scripts" (OuterVolumeSpecName: "scripts") pod "f45d9be7-457d-472d-8c9f-463f9d241eae" (UID: "f45d9be7-457d-472d-8c9f-463f9d241eae"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 19:55:44 crc kubenswrapper[4988]: I1008 19:55:44.788191 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f45d9be7-457d-472d-8c9f-463f9d241eae-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f45d9be7-457d-472d-8c9f-463f9d241eae" (UID: "f45d9be7-457d-472d-8c9f-463f9d241eae"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 19:55:44 crc kubenswrapper[4988]: I1008 19:55:44.804566 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f45d9be7-457d-472d-8c9f-463f9d241eae-horizon-tls-certs" (OuterVolumeSpecName: "horizon-tls-certs") pod "f45d9be7-457d-472d-8c9f-463f9d241eae" (UID: "f45d9be7-457d-472d-8c9f-463f9d241eae"). InnerVolumeSpecName "horizon-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 19:55:44 crc kubenswrapper[4988]: I1008 19:55:44.827343 4988 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f45d9be7-457d-472d-8c9f-463f9d241eae-logs\") on node \"crc\" DevicePath \"\"" Oct 08 19:55:44 crc kubenswrapper[4988]: I1008 19:55:44.827402 4988 reconciler_common.go:293] "Volume detached for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/f45d9be7-457d-472d-8c9f-463f9d241eae-horizon-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 08 19:55:44 crc kubenswrapper[4988]: I1008 19:55:44.827414 4988 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/f45d9be7-457d-472d-8c9f-463f9d241eae-config-data\") on node \"crc\" DevicePath \"\"" Oct 08 19:55:44 crc kubenswrapper[4988]: I1008 19:55:44.827424 4988 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f45d9be7-457d-472d-8c9f-463f9d241eae-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 19:55:44 crc kubenswrapper[4988]: I1008 19:55:44.827434 4988 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-klg9v\" (UniqueName: \"kubernetes.io/projected/f45d9be7-457d-472d-8c9f-463f9d241eae-kube-api-access-klg9v\") on node \"crc\" DevicePath \"\"" Oct 08 19:55:44 crc kubenswrapper[4988]: I1008 19:55:44.827442 4988 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f45d9be7-457d-472d-8c9f-463f9d241eae-scripts\") on node \"crc\" DevicePath \"\"" Oct 08 19:55:44 crc kubenswrapper[4988]: I1008 19:55:44.827452 4988 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/f45d9be7-457d-472d-8c9f-463f9d241eae-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Oct 08 19:55:45 crc kubenswrapper[4988]: I1008 19:55:45.438223 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-d95cf97bd-hmknh" event={"ID":"f45d9be7-457d-472d-8c9f-463f9d241eae","Type":"ContainerDied","Data":"b9b01b86503ad5dc6c6e49585c6699cfae1d882b6e9c9bfa15c09a673663b3cb"} Oct 08 19:55:45 crc kubenswrapper[4988]: I1008 19:55:45.439860 4988 scope.go:117] "RemoveContainer" containerID="35268501f379cc03d23dd17a36b891c830fa111aea8c3f4c0dac3ce3f4b66d27" Oct 08 19:55:45 crc kubenswrapper[4988]: I1008 19:55:45.438309 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-d95cf97bd-hmknh" Oct 08 19:55:45 crc kubenswrapper[4988]: I1008 19:55:45.485486 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-d95cf97bd-hmknh"] Oct 08 19:55:45 crc kubenswrapper[4988]: I1008 19:55:45.499783 4988 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-d95cf97bd-hmknh"] Oct 08 19:55:45 crc kubenswrapper[4988]: I1008 19:55:45.681980 4988 scope.go:117] "RemoveContainer" containerID="3044dddba8423c50138cfe88cad812bb1f3b03df4ece17ff02aad32480deb876" Oct 08 19:55:47 crc kubenswrapper[4988]: I1008 19:55:47.257744 4988 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f45d9be7-457d-472d-8c9f-463f9d241eae" path="/var/lib/kubelet/pods/f45d9be7-457d-472d-8c9f-463f9d241eae/volumes" Oct 08 19:55:49 crc kubenswrapper[4988]: I1008 19:55:49.048727 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-create-s2d5m"] Oct 08 19:55:49 crc kubenswrapper[4988]: I1008 19:55:49.058765 4988 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-create-s2d5m"] Oct 08 19:55:49 crc kubenswrapper[4988]: I1008 19:55:49.258711 4988 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5681236e-e0dc-46cf-aa72-68e4b26454ec" path="/var/lib/kubelet/pods/5681236e-e0dc-46cf-aa72-68e4b26454ec/volumes" Oct 08 19:55:55 crc kubenswrapper[4988]: I1008 19:55:55.348656 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-64f8b4fb78-xcwgf"] Oct 08 19:55:55 crc kubenswrapper[4988]: E1008 19:55:55.349372 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="83fe9318-96ca-4f74-b838-9997e59c1ea2" containerName="horizon-log" Oct 08 19:55:55 crc kubenswrapper[4988]: I1008 19:55:55.349399 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="83fe9318-96ca-4f74-b838-9997e59c1ea2" containerName="horizon-log" Oct 08 19:55:55 crc kubenswrapper[4988]: E1008 19:55:55.349413 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f45d9be7-457d-472d-8c9f-463f9d241eae" containerName="horizon" Oct 08 19:55:55 crc kubenswrapper[4988]: I1008 19:55:55.349418 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="f45d9be7-457d-472d-8c9f-463f9d241eae" containerName="horizon" Oct 08 19:55:55 crc kubenswrapper[4988]: E1008 19:55:55.349444 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="322a3c03-38ac-4ef2-8980-c1976f855a81" containerName="horizon" Oct 08 19:55:55 crc kubenswrapper[4988]: I1008 19:55:55.349450 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="322a3c03-38ac-4ef2-8980-c1976f855a81" containerName="horizon" Oct 08 19:55:55 crc kubenswrapper[4988]: E1008 19:55:55.349465 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="83fe9318-96ca-4f74-b838-9997e59c1ea2" containerName="horizon" Oct 08 19:55:55 crc kubenswrapper[4988]: I1008 19:55:55.349473 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="83fe9318-96ca-4f74-b838-9997e59c1ea2" containerName="horizon" Oct 08 19:55:55 crc kubenswrapper[4988]: E1008 19:55:55.349484 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="322a3c03-38ac-4ef2-8980-c1976f855a81" containerName="horizon-log" Oct 08 19:55:55 crc kubenswrapper[4988]: I1008 19:55:55.349490 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="322a3c03-38ac-4ef2-8980-c1976f855a81" containerName="horizon-log" Oct 08 19:55:55 crc kubenswrapper[4988]: E1008 19:55:55.349506 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f45d9be7-457d-472d-8c9f-463f9d241eae" containerName="horizon-log" Oct 08 19:55:55 crc kubenswrapper[4988]: I1008 19:55:55.349512 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="f45d9be7-457d-472d-8c9f-463f9d241eae" containerName="horizon-log" Oct 08 19:55:55 crc kubenswrapper[4988]: I1008 19:55:55.349679 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="f45d9be7-457d-472d-8c9f-463f9d241eae" containerName="horizon-log" Oct 08 19:55:55 crc kubenswrapper[4988]: I1008 19:55:55.349690 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="83fe9318-96ca-4f74-b838-9997e59c1ea2" containerName="horizon-log" Oct 08 19:55:55 crc kubenswrapper[4988]: I1008 19:55:55.349702 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="f45d9be7-457d-472d-8c9f-463f9d241eae" containerName="horizon" Oct 08 19:55:55 crc kubenswrapper[4988]: I1008 19:55:55.349718 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="322a3c03-38ac-4ef2-8980-c1976f855a81" containerName="horizon-log" Oct 08 19:55:55 crc kubenswrapper[4988]: I1008 19:55:55.349729 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="322a3c03-38ac-4ef2-8980-c1976f855a81" containerName="horizon" Oct 08 19:55:55 crc kubenswrapper[4988]: I1008 19:55:55.349740 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="83fe9318-96ca-4f74-b838-9997e59c1ea2" containerName="horizon" Oct 08 19:55:55 crc kubenswrapper[4988]: I1008 19:55:55.350713 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-64f8b4fb78-xcwgf" Oct 08 19:55:55 crc kubenswrapper[4988]: I1008 19:55:55.382853 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-64f8b4fb78-xcwgf"] Oct 08 19:55:55 crc kubenswrapper[4988]: I1008 19:55:55.496149 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/f084c3ca-3754-4e7e-82dd-f0f3029d7102-horizon-secret-key\") pod \"horizon-64f8b4fb78-xcwgf\" (UID: \"f084c3ca-3754-4e7e-82dd-f0f3029d7102\") " pod="openstack/horizon-64f8b4fb78-xcwgf" Oct 08 19:55:55 crc kubenswrapper[4988]: I1008 19:55:55.496256 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f084c3ca-3754-4e7e-82dd-f0f3029d7102-scripts\") pod \"horizon-64f8b4fb78-xcwgf\" (UID: \"f084c3ca-3754-4e7e-82dd-f0f3029d7102\") " pod="openstack/horizon-64f8b4fb78-xcwgf" Oct 08 19:55:55 crc kubenswrapper[4988]: I1008 19:55:55.496295 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/f084c3ca-3754-4e7e-82dd-f0f3029d7102-config-data\") pod \"horizon-64f8b4fb78-xcwgf\" (UID: \"f084c3ca-3754-4e7e-82dd-f0f3029d7102\") " pod="openstack/horizon-64f8b4fb78-xcwgf" Oct 08 19:55:55 crc kubenswrapper[4988]: I1008 19:55:55.496314 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f084c3ca-3754-4e7e-82dd-f0f3029d7102-combined-ca-bundle\") pod \"horizon-64f8b4fb78-xcwgf\" (UID: \"f084c3ca-3754-4e7e-82dd-f0f3029d7102\") " pod="openstack/horizon-64f8b4fb78-xcwgf" Oct 08 19:55:55 crc kubenswrapper[4988]: I1008 19:55:55.496336 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/f084c3ca-3754-4e7e-82dd-f0f3029d7102-horizon-tls-certs\") pod \"horizon-64f8b4fb78-xcwgf\" (UID: \"f084c3ca-3754-4e7e-82dd-f0f3029d7102\") " pod="openstack/horizon-64f8b4fb78-xcwgf" Oct 08 19:55:55 crc kubenswrapper[4988]: I1008 19:55:55.496627 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tgk2h\" (UniqueName: \"kubernetes.io/projected/f084c3ca-3754-4e7e-82dd-f0f3029d7102-kube-api-access-tgk2h\") pod \"horizon-64f8b4fb78-xcwgf\" (UID: \"f084c3ca-3754-4e7e-82dd-f0f3029d7102\") " pod="openstack/horizon-64f8b4fb78-xcwgf" Oct 08 19:55:55 crc kubenswrapper[4988]: I1008 19:55:55.496772 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f084c3ca-3754-4e7e-82dd-f0f3029d7102-logs\") pod \"horizon-64f8b4fb78-xcwgf\" (UID: \"f084c3ca-3754-4e7e-82dd-f0f3029d7102\") " pod="openstack/horizon-64f8b4fb78-xcwgf" Oct 08 19:55:55 crc kubenswrapper[4988]: I1008 19:55:55.599278 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/f084c3ca-3754-4e7e-82dd-f0f3029d7102-horizon-secret-key\") pod \"horizon-64f8b4fb78-xcwgf\" (UID: \"f084c3ca-3754-4e7e-82dd-f0f3029d7102\") " pod="openstack/horizon-64f8b4fb78-xcwgf" Oct 08 19:55:55 crc kubenswrapper[4988]: I1008 19:55:55.599482 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f084c3ca-3754-4e7e-82dd-f0f3029d7102-scripts\") pod \"horizon-64f8b4fb78-xcwgf\" (UID: \"f084c3ca-3754-4e7e-82dd-f0f3029d7102\") " pod="openstack/horizon-64f8b4fb78-xcwgf" Oct 08 19:55:55 crc kubenswrapper[4988]: I1008 19:55:55.599559 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/f084c3ca-3754-4e7e-82dd-f0f3029d7102-config-data\") pod \"horizon-64f8b4fb78-xcwgf\" (UID: \"f084c3ca-3754-4e7e-82dd-f0f3029d7102\") " pod="openstack/horizon-64f8b4fb78-xcwgf" Oct 08 19:55:55 crc kubenswrapper[4988]: I1008 19:55:55.599637 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f084c3ca-3754-4e7e-82dd-f0f3029d7102-combined-ca-bundle\") pod \"horizon-64f8b4fb78-xcwgf\" (UID: \"f084c3ca-3754-4e7e-82dd-f0f3029d7102\") " pod="openstack/horizon-64f8b4fb78-xcwgf" Oct 08 19:55:55 crc kubenswrapper[4988]: I1008 19:55:55.599685 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/f084c3ca-3754-4e7e-82dd-f0f3029d7102-horizon-tls-certs\") pod \"horizon-64f8b4fb78-xcwgf\" (UID: \"f084c3ca-3754-4e7e-82dd-f0f3029d7102\") " pod="openstack/horizon-64f8b4fb78-xcwgf" Oct 08 19:55:55 crc kubenswrapper[4988]: I1008 19:55:55.599803 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tgk2h\" (UniqueName: \"kubernetes.io/projected/f084c3ca-3754-4e7e-82dd-f0f3029d7102-kube-api-access-tgk2h\") pod \"horizon-64f8b4fb78-xcwgf\" (UID: \"f084c3ca-3754-4e7e-82dd-f0f3029d7102\") " pod="openstack/horizon-64f8b4fb78-xcwgf" Oct 08 19:55:55 crc kubenswrapper[4988]: I1008 19:55:55.599902 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f084c3ca-3754-4e7e-82dd-f0f3029d7102-logs\") pod \"horizon-64f8b4fb78-xcwgf\" (UID: \"f084c3ca-3754-4e7e-82dd-f0f3029d7102\") " pod="openstack/horizon-64f8b4fb78-xcwgf" Oct 08 19:55:55 crc kubenswrapper[4988]: I1008 19:55:55.600680 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f084c3ca-3754-4e7e-82dd-f0f3029d7102-logs\") pod \"horizon-64f8b4fb78-xcwgf\" (UID: \"f084c3ca-3754-4e7e-82dd-f0f3029d7102\") " pod="openstack/horizon-64f8b4fb78-xcwgf" Oct 08 19:55:55 crc kubenswrapper[4988]: I1008 19:55:55.601504 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f084c3ca-3754-4e7e-82dd-f0f3029d7102-scripts\") pod \"horizon-64f8b4fb78-xcwgf\" (UID: \"f084c3ca-3754-4e7e-82dd-f0f3029d7102\") " pod="openstack/horizon-64f8b4fb78-xcwgf" Oct 08 19:55:55 crc kubenswrapper[4988]: I1008 19:55:55.603591 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/f084c3ca-3754-4e7e-82dd-f0f3029d7102-config-data\") pod \"horizon-64f8b4fb78-xcwgf\" (UID: \"f084c3ca-3754-4e7e-82dd-f0f3029d7102\") " pod="openstack/horizon-64f8b4fb78-xcwgf" Oct 08 19:55:55 crc kubenswrapper[4988]: I1008 19:55:55.610423 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f084c3ca-3754-4e7e-82dd-f0f3029d7102-combined-ca-bundle\") pod \"horizon-64f8b4fb78-xcwgf\" (UID: \"f084c3ca-3754-4e7e-82dd-f0f3029d7102\") " pod="openstack/horizon-64f8b4fb78-xcwgf" Oct 08 19:55:55 crc kubenswrapper[4988]: I1008 19:55:55.611010 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/f084c3ca-3754-4e7e-82dd-f0f3029d7102-horizon-tls-certs\") pod \"horizon-64f8b4fb78-xcwgf\" (UID: \"f084c3ca-3754-4e7e-82dd-f0f3029d7102\") " pod="openstack/horizon-64f8b4fb78-xcwgf" Oct 08 19:55:55 crc kubenswrapper[4988]: I1008 19:55:55.628929 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/f084c3ca-3754-4e7e-82dd-f0f3029d7102-horizon-secret-key\") pod \"horizon-64f8b4fb78-xcwgf\" (UID: \"f084c3ca-3754-4e7e-82dd-f0f3029d7102\") " pod="openstack/horizon-64f8b4fb78-xcwgf" Oct 08 19:55:55 crc kubenswrapper[4988]: I1008 19:55:55.633345 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tgk2h\" (UniqueName: \"kubernetes.io/projected/f084c3ca-3754-4e7e-82dd-f0f3029d7102-kube-api-access-tgk2h\") pod \"horizon-64f8b4fb78-xcwgf\" (UID: \"f084c3ca-3754-4e7e-82dd-f0f3029d7102\") " pod="openstack/horizon-64f8b4fb78-xcwgf" Oct 08 19:55:55 crc kubenswrapper[4988]: I1008 19:55:55.667416 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-64f8b4fb78-xcwgf" Oct 08 19:55:56 crc kubenswrapper[4988]: I1008 19:55:56.187915 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-64f8b4fb78-xcwgf"] Oct 08 19:55:56 crc kubenswrapper[4988]: I1008 19:55:56.604661 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-64f8b4fb78-xcwgf" event={"ID":"f084c3ca-3754-4e7e-82dd-f0f3029d7102","Type":"ContainerStarted","Data":"a12f1c01b2d2060a7595c312368509e6d6d924ef6a97beb5417039ddb3ebfee9"} Oct 08 19:55:56 crc kubenswrapper[4988]: I1008 19:55:56.604990 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-64f8b4fb78-xcwgf" event={"ID":"f084c3ca-3754-4e7e-82dd-f0f3029d7102","Type":"ContainerStarted","Data":"77c07534844f541eb1eec4d2b321c1bb3079918a991c078b373b1e191921b22c"} Oct 08 19:55:56 crc kubenswrapper[4988]: I1008 19:55:56.605014 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-64f8b4fb78-xcwgf" event={"ID":"f084c3ca-3754-4e7e-82dd-f0f3029d7102","Type":"ContainerStarted","Data":"f258e3d77a798f007c3a0debfeaffb018a1d14212c3d8e99365538f50e10d40f"} Oct 08 19:55:56 crc kubenswrapper[4988]: I1008 19:55:56.645298 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-64f8b4fb78-xcwgf" podStartSLOduration=1.6452696439999999 podStartE2EDuration="1.645269644s" podCreationTimestamp="2025-10-08 19:55:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 19:55:56.637508627 +0000 UTC m=+6302.087351427" watchObservedRunningTime="2025-10-08 19:55:56.645269644 +0000 UTC m=+6302.095112454" Oct 08 19:55:57 crc kubenswrapper[4988]: I1008 19:55:57.014183 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-db-create-rz6jz"] Oct 08 19:55:57 crc kubenswrapper[4988]: I1008 19:55:57.015450 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-create-rz6jz" Oct 08 19:55:57 crc kubenswrapper[4988]: I1008 19:55:57.024422 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-db-create-rz6jz"] Oct 08 19:55:57 crc kubenswrapper[4988]: I1008 19:55:57.047355 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cbflg\" (UniqueName: \"kubernetes.io/projected/23ccee36-ec99-4075-95f8-19cb2f402c3b-kube-api-access-cbflg\") pod \"heat-db-create-rz6jz\" (UID: \"23ccee36-ec99-4075-95f8-19cb2f402c3b\") " pod="openstack/heat-db-create-rz6jz" Oct 08 19:55:57 crc kubenswrapper[4988]: I1008 19:55:57.149559 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cbflg\" (UniqueName: \"kubernetes.io/projected/23ccee36-ec99-4075-95f8-19cb2f402c3b-kube-api-access-cbflg\") pod \"heat-db-create-rz6jz\" (UID: \"23ccee36-ec99-4075-95f8-19cb2f402c3b\") " pod="openstack/heat-db-create-rz6jz" Oct 08 19:55:57 crc kubenswrapper[4988]: I1008 19:55:57.171118 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cbflg\" (UniqueName: \"kubernetes.io/projected/23ccee36-ec99-4075-95f8-19cb2f402c3b-kube-api-access-cbflg\") pod \"heat-db-create-rz6jz\" (UID: \"23ccee36-ec99-4075-95f8-19cb2f402c3b\") " pod="openstack/heat-db-create-rz6jz" Oct 08 19:55:57 crc kubenswrapper[4988]: I1008 19:55:57.342050 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-create-rz6jz" Oct 08 19:55:57 crc kubenswrapper[4988]: I1008 19:55:57.848681 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-db-create-rz6jz"] Oct 08 19:55:57 crc kubenswrapper[4988]: W1008 19:55:57.850633 4988 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod23ccee36_ec99_4075_95f8_19cb2f402c3b.slice/crio-67e02b4087b7e45a8568b9522c199ca8bb3b35ab2723f82d5074b637431e0397 WatchSource:0}: Error finding container 67e02b4087b7e45a8568b9522c199ca8bb3b35ab2723f82d5074b637431e0397: Status 404 returned error can't find the container with id 67e02b4087b7e45a8568b9522c199ca8bb3b35ab2723f82d5074b637431e0397 Oct 08 19:55:58 crc kubenswrapper[4988]: I1008 19:55:58.632000 4988 generic.go:334] "Generic (PLEG): container finished" podID="23ccee36-ec99-4075-95f8-19cb2f402c3b" containerID="2dd58f0f20f70d265cd0868a92e6a83410886f5cfaa37d2fb006a899a6b89420" exitCode=0 Oct 08 19:55:58 crc kubenswrapper[4988]: I1008 19:55:58.632143 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-create-rz6jz" event={"ID":"23ccee36-ec99-4075-95f8-19cb2f402c3b","Type":"ContainerDied","Data":"2dd58f0f20f70d265cd0868a92e6a83410886f5cfaa37d2fb006a899a6b89420"} Oct 08 19:55:58 crc kubenswrapper[4988]: I1008 19:55:58.632667 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-create-rz6jz" event={"ID":"23ccee36-ec99-4075-95f8-19cb2f402c3b","Type":"ContainerStarted","Data":"67e02b4087b7e45a8568b9522c199ca8bb3b35ab2723f82d5074b637431e0397"} Oct 08 19:55:59 crc kubenswrapper[4988]: I1008 19:55:59.043784 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-17fa-account-create-rpcv2"] Oct 08 19:55:59 crc kubenswrapper[4988]: I1008 19:55:59.057791 4988 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-17fa-account-create-rpcv2"] Oct 08 19:55:59 crc kubenswrapper[4988]: I1008 19:55:59.262988 4988 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8f2a9de4-4576-416f-99cd-2fe38b4700db" path="/var/lib/kubelet/pods/8f2a9de4-4576-416f-99cd-2fe38b4700db/volumes" Oct 08 19:56:00 crc kubenswrapper[4988]: I1008 19:56:00.109149 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-create-rz6jz" Oct 08 19:56:00 crc kubenswrapper[4988]: I1008 19:56:00.242342 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cbflg\" (UniqueName: \"kubernetes.io/projected/23ccee36-ec99-4075-95f8-19cb2f402c3b-kube-api-access-cbflg\") pod \"23ccee36-ec99-4075-95f8-19cb2f402c3b\" (UID: \"23ccee36-ec99-4075-95f8-19cb2f402c3b\") " Oct 08 19:56:00 crc kubenswrapper[4988]: I1008 19:56:00.249107 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/23ccee36-ec99-4075-95f8-19cb2f402c3b-kube-api-access-cbflg" (OuterVolumeSpecName: "kube-api-access-cbflg") pod "23ccee36-ec99-4075-95f8-19cb2f402c3b" (UID: "23ccee36-ec99-4075-95f8-19cb2f402c3b"). InnerVolumeSpecName "kube-api-access-cbflg". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 19:56:00 crc kubenswrapper[4988]: I1008 19:56:00.345486 4988 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cbflg\" (UniqueName: \"kubernetes.io/projected/23ccee36-ec99-4075-95f8-19cb2f402c3b-kube-api-access-cbflg\") on node \"crc\" DevicePath \"\"" Oct 08 19:56:00 crc kubenswrapper[4988]: I1008 19:56:00.661031 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-create-rz6jz" event={"ID":"23ccee36-ec99-4075-95f8-19cb2f402c3b","Type":"ContainerDied","Data":"67e02b4087b7e45a8568b9522c199ca8bb3b35ab2723f82d5074b637431e0397"} Oct 08 19:56:00 crc kubenswrapper[4988]: I1008 19:56:00.661075 4988 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="67e02b4087b7e45a8568b9522c199ca8bb3b35ab2723f82d5074b637431e0397" Oct 08 19:56:00 crc kubenswrapper[4988]: I1008 19:56:00.661142 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-create-rz6jz" Oct 08 19:56:05 crc kubenswrapper[4988]: I1008 19:56:05.668331 4988 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-64f8b4fb78-xcwgf" Oct 08 19:56:05 crc kubenswrapper[4988]: I1008 19:56:05.668985 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-64f8b4fb78-xcwgf" Oct 08 19:56:07 crc kubenswrapper[4988]: I1008 19:56:07.053565 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-sync-sfpjb"] Oct 08 19:56:07 crc kubenswrapper[4988]: I1008 19:56:07.070000 4988 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-sync-sfpjb"] Oct 08 19:56:07 crc kubenswrapper[4988]: I1008 19:56:07.129791 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-6119-account-create-fcfhx"] Oct 08 19:56:07 crc kubenswrapper[4988]: E1008 19:56:07.130879 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="23ccee36-ec99-4075-95f8-19cb2f402c3b" containerName="mariadb-database-create" Oct 08 19:56:07 crc kubenswrapper[4988]: I1008 19:56:07.130925 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="23ccee36-ec99-4075-95f8-19cb2f402c3b" containerName="mariadb-database-create" Oct 08 19:56:07 crc kubenswrapper[4988]: I1008 19:56:07.131511 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="23ccee36-ec99-4075-95f8-19cb2f402c3b" containerName="mariadb-database-create" Oct 08 19:56:07 crc kubenswrapper[4988]: I1008 19:56:07.133211 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-6119-account-create-fcfhx" Oct 08 19:56:07 crc kubenswrapper[4988]: I1008 19:56:07.141502 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-db-secret" Oct 08 19:56:07 crc kubenswrapper[4988]: I1008 19:56:07.141527 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-6119-account-create-fcfhx"] Oct 08 19:56:07 crc kubenswrapper[4988]: I1008 19:56:07.228240 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fszxs\" (UniqueName: \"kubernetes.io/projected/d4b96fb0-1c91-4514-938b-5dbffa503fc1-kube-api-access-fszxs\") pod \"heat-6119-account-create-fcfhx\" (UID: \"d4b96fb0-1c91-4514-938b-5dbffa503fc1\") " pod="openstack/heat-6119-account-create-fcfhx" Oct 08 19:56:07 crc kubenswrapper[4988]: I1008 19:56:07.259842 4988 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ec8f6c64-9860-4a50-bd16-972351608049" path="/var/lib/kubelet/pods/ec8f6c64-9860-4a50-bd16-972351608049/volumes" Oct 08 19:56:07 crc kubenswrapper[4988]: I1008 19:56:07.330484 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fszxs\" (UniqueName: \"kubernetes.io/projected/d4b96fb0-1c91-4514-938b-5dbffa503fc1-kube-api-access-fszxs\") pod \"heat-6119-account-create-fcfhx\" (UID: \"d4b96fb0-1c91-4514-938b-5dbffa503fc1\") " pod="openstack/heat-6119-account-create-fcfhx" Oct 08 19:56:07 crc kubenswrapper[4988]: I1008 19:56:07.354734 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fszxs\" (UniqueName: \"kubernetes.io/projected/d4b96fb0-1c91-4514-938b-5dbffa503fc1-kube-api-access-fszxs\") pod \"heat-6119-account-create-fcfhx\" (UID: \"d4b96fb0-1c91-4514-938b-5dbffa503fc1\") " pod="openstack/heat-6119-account-create-fcfhx" Oct 08 19:56:07 crc kubenswrapper[4988]: I1008 19:56:07.475771 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-6119-account-create-fcfhx" Oct 08 19:56:08 crc kubenswrapper[4988]: I1008 19:56:08.054270 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-6119-account-create-fcfhx"] Oct 08 19:56:08 crc kubenswrapper[4988]: W1008 19:56:08.065731 4988 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd4b96fb0_1c91_4514_938b_5dbffa503fc1.slice/crio-b532a7aa72ee0ddb140d950fe0fdf3e82dcaa9fdc6da16911899dd0a8bd13cf2 WatchSource:0}: Error finding container b532a7aa72ee0ddb140d950fe0fdf3e82dcaa9fdc6da16911899dd0a8bd13cf2: Status 404 returned error can't find the container with id b532a7aa72ee0ddb140d950fe0fdf3e82dcaa9fdc6da16911899dd0a8bd13cf2 Oct 08 19:56:08 crc kubenswrapper[4988]: E1008 19:56:08.523189 4988 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd4b96fb0_1c91_4514_938b_5dbffa503fc1.slice/crio-conmon-18c2bea687f803c935f24db50662fe3c14af212c5ec3794b7cd3defdea79da15.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd4b96fb0_1c91_4514_938b_5dbffa503fc1.slice/crio-18c2bea687f803c935f24db50662fe3c14af212c5ec3794b7cd3defdea79da15.scope\": RecentStats: unable to find data in memory cache]" Oct 08 19:56:08 crc kubenswrapper[4988]: I1008 19:56:08.782115 4988 generic.go:334] "Generic (PLEG): container finished" podID="d4b96fb0-1c91-4514-938b-5dbffa503fc1" containerID="18c2bea687f803c935f24db50662fe3c14af212c5ec3794b7cd3defdea79da15" exitCode=0 Oct 08 19:56:08 crc kubenswrapper[4988]: I1008 19:56:08.782206 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-6119-account-create-fcfhx" event={"ID":"d4b96fb0-1c91-4514-938b-5dbffa503fc1","Type":"ContainerDied","Data":"18c2bea687f803c935f24db50662fe3c14af212c5ec3794b7cd3defdea79da15"} Oct 08 19:56:08 crc kubenswrapper[4988]: I1008 19:56:08.782513 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-6119-account-create-fcfhx" event={"ID":"d4b96fb0-1c91-4514-938b-5dbffa503fc1","Type":"ContainerStarted","Data":"b532a7aa72ee0ddb140d950fe0fdf3e82dcaa9fdc6da16911899dd0a8bd13cf2"} Oct 08 19:56:10 crc kubenswrapper[4988]: I1008 19:56:10.234829 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-6119-account-create-fcfhx" Oct 08 19:56:10 crc kubenswrapper[4988]: I1008 19:56:10.310719 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fszxs\" (UniqueName: \"kubernetes.io/projected/d4b96fb0-1c91-4514-938b-5dbffa503fc1-kube-api-access-fszxs\") pod \"d4b96fb0-1c91-4514-938b-5dbffa503fc1\" (UID: \"d4b96fb0-1c91-4514-938b-5dbffa503fc1\") " Oct 08 19:56:10 crc kubenswrapper[4988]: I1008 19:56:10.320167 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d4b96fb0-1c91-4514-938b-5dbffa503fc1-kube-api-access-fszxs" (OuterVolumeSpecName: "kube-api-access-fszxs") pod "d4b96fb0-1c91-4514-938b-5dbffa503fc1" (UID: "d4b96fb0-1c91-4514-938b-5dbffa503fc1"). InnerVolumeSpecName "kube-api-access-fszxs". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 19:56:10 crc kubenswrapper[4988]: I1008 19:56:10.413851 4988 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fszxs\" (UniqueName: \"kubernetes.io/projected/d4b96fb0-1c91-4514-938b-5dbffa503fc1-kube-api-access-fszxs\") on node \"crc\" DevicePath \"\"" Oct 08 19:56:10 crc kubenswrapper[4988]: I1008 19:56:10.829758 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-6119-account-create-fcfhx" event={"ID":"d4b96fb0-1c91-4514-938b-5dbffa503fc1","Type":"ContainerDied","Data":"b532a7aa72ee0ddb140d950fe0fdf3e82dcaa9fdc6da16911899dd0a8bd13cf2"} Oct 08 19:56:10 crc kubenswrapper[4988]: I1008 19:56:10.829803 4988 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b532a7aa72ee0ddb140d950fe0fdf3e82dcaa9fdc6da16911899dd0a8bd13cf2" Oct 08 19:56:10 crc kubenswrapper[4988]: I1008 19:56:10.829902 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-6119-account-create-fcfhx" Oct 08 19:56:11 crc kubenswrapper[4988]: I1008 19:56:11.180193 4988 scope.go:117] "RemoveContainer" containerID="e539312cc13f1978d1df1e563814b231de949749a70469f6361aa41906cadd6c" Oct 08 19:56:11 crc kubenswrapper[4988]: I1008 19:56:11.233428 4988 scope.go:117] "RemoveContainer" containerID="2626533fad751ded722c45fcae26aab38c2d0861daefd298e02888a2a0a7b066" Oct 08 19:56:11 crc kubenswrapper[4988]: I1008 19:56:11.268740 4988 scope.go:117] "RemoveContainer" containerID="bf77b9c29c9c464a36804f797ad4a23d3ee5ddecf4d550cbf71fc2fbeac45e98" Oct 08 19:56:12 crc kubenswrapper[4988]: I1008 19:56:12.300135 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-db-sync-xs9q8"] Oct 08 19:56:12 crc kubenswrapper[4988]: E1008 19:56:12.300610 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d4b96fb0-1c91-4514-938b-5dbffa503fc1" containerName="mariadb-account-create" Oct 08 19:56:12 crc kubenswrapper[4988]: I1008 19:56:12.300625 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="d4b96fb0-1c91-4514-938b-5dbffa503fc1" containerName="mariadb-account-create" Oct 08 19:56:12 crc kubenswrapper[4988]: I1008 19:56:12.300916 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="d4b96fb0-1c91-4514-938b-5dbffa503fc1" containerName="mariadb-account-create" Oct 08 19:56:12 crc kubenswrapper[4988]: I1008 19:56:12.301676 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-sync-xs9q8" Oct 08 19:56:12 crc kubenswrapper[4988]: I1008 19:56:12.307045 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-heat-dockercfg-cnlv2" Oct 08 19:56:12 crc kubenswrapper[4988]: I1008 19:56:12.312857 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-config-data" Oct 08 19:56:12 crc kubenswrapper[4988]: I1008 19:56:12.347619 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-db-sync-xs9q8"] Oct 08 19:56:12 crc kubenswrapper[4988]: I1008 19:56:12.467870 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/162b2195-0c02-4cfc-ae52-ac4cdc1e1a9c-config-data\") pod \"heat-db-sync-xs9q8\" (UID: \"162b2195-0c02-4cfc-ae52-ac4cdc1e1a9c\") " pod="openstack/heat-db-sync-xs9q8" Oct 08 19:56:12 crc kubenswrapper[4988]: I1008 19:56:12.467956 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-snnll\" (UniqueName: \"kubernetes.io/projected/162b2195-0c02-4cfc-ae52-ac4cdc1e1a9c-kube-api-access-snnll\") pod \"heat-db-sync-xs9q8\" (UID: \"162b2195-0c02-4cfc-ae52-ac4cdc1e1a9c\") " pod="openstack/heat-db-sync-xs9q8" Oct 08 19:56:12 crc kubenswrapper[4988]: I1008 19:56:12.468280 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/162b2195-0c02-4cfc-ae52-ac4cdc1e1a9c-combined-ca-bundle\") pod \"heat-db-sync-xs9q8\" (UID: \"162b2195-0c02-4cfc-ae52-ac4cdc1e1a9c\") " pod="openstack/heat-db-sync-xs9q8" Oct 08 19:56:12 crc kubenswrapper[4988]: I1008 19:56:12.570489 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/162b2195-0c02-4cfc-ae52-ac4cdc1e1a9c-config-data\") pod \"heat-db-sync-xs9q8\" (UID: \"162b2195-0c02-4cfc-ae52-ac4cdc1e1a9c\") " pod="openstack/heat-db-sync-xs9q8" Oct 08 19:56:12 crc kubenswrapper[4988]: I1008 19:56:12.570925 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-snnll\" (UniqueName: \"kubernetes.io/projected/162b2195-0c02-4cfc-ae52-ac4cdc1e1a9c-kube-api-access-snnll\") pod \"heat-db-sync-xs9q8\" (UID: \"162b2195-0c02-4cfc-ae52-ac4cdc1e1a9c\") " pod="openstack/heat-db-sync-xs9q8" Oct 08 19:56:12 crc kubenswrapper[4988]: I1008 19:56:12.571137 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/162b2195-0c02-4cfc-ae52-ac4cdc1e1a9c-combined-ca-bundle\") pod \"heat-db-sync-xs9q8\" (UID: \"162b2195-0c02-4cfc-ae52-ac4cdc1e1a9c\") " pod="openstack/heat-db-sync-xs9q8" Oct 08 19:56:12 crc kubenswrapper[4988]: I1008 19:56:12.576711 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/162b2195-0c02-4cfc-ae52-ac4cdc1e1a9c-combined-ca-bundle\") pod \"heat-db-sync-xs9q8\" (UID: \"162b2195-0c02-4cfc-ae52-ac4cdc1e1a9c\") " pod="openstack/heat-db-sync-xs9q8" Oct 08 19:56:12 crc kubenswrapper[4988]: I1008 19:56:12.586281 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-snnll\" (UniqueName: \"kubernetes.io/projected/162b2195-0c02-4cfc-ae52-ac4cdc1e1a9c-kube-api-access-snnll\") pod \"heat-db-sync-xs9q8\" (UID: \"162b2195-0c02-4cfc-ae52-ac4cdc1e1a9c\") " pod="openstack/heat-db-sync-xs9q8" Oct 08 19:56:12 crc kubenswrapper[4988]: I1008 19:56:12.587490 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/162b2195-0c02-4cfc-ae52-ac4cdc1e1a9c-config-data\") pod \"heat-db-sync-xs9q8\" (UID: \"162b2195-0c02-4cfc-ae52-ac4cdc1e1a9c\") " pod="openstack/heat-db-sync-xs9q8" Oct 08 19:56:12 crc kubenswrapper[4988]: I1008 19:56:12.651423 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-sync-xs9q8" Oct 08 19:56:13 crc kubenswrapper[4988]: I1008 19:56:13.159850 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-db-sync-xs9q8"] Oct 08 19:56:13 crc kubenswrapper[4988]: W1008 19:56:13.164667 4988 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod162b2195_0c02_4cfc_ae52_ac4cdc1e1a9c.slice/crio-4f2ba48827917045fd09ca2a73e5951c6866937321f72a027a8fb5ace14730c3 WatchSource:0}: Error finding container 4f2ba48827917045fd09ca2a73e5951c6866937321f72a027a8fb5ace14730c3: Status 404 returned error can't find the container with id 4f2ba48827917045fd09ca2a73e5951c6866937321f72a027a8fb5ace14730c3 Oct 08 19:56:13 crc kubenswrapper[4988]: I1008 19:56:13.885667 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-sync-xs9q8" event={"ID":"162b2195-0c02-4cfc-ae52-ac4cdc1e1a9c","Type":"ContainerStarted","Data":"4f2ba48827917045fd09ca2a73e5951c6866937321f72a027a8fb5ace14730c3"} Oct 08 19:56:17 crc kubenswrapper[4988]: I1008 19:56:17.386947 4988 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/horizon-64f8b4fb78-xcwgf" Oct 08 19:56:19 crc kubenswrapper[4988]: I1008 19:56:19.089690 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/horizon-64f8b4fb78-xcwgf" Oct 08 19:56:19 crc kubenswrapper[4988]: I1008 19:56:19.154305 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-7d85fc8646-gv4lm"] Oct 08 19:56:19 crc kubenswrapper[4988]: I1008 19:56:19.154619 4988 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-7d85fc8646-gv4lm" podUID="c533a2e1-7213-4410-b862-5eed5e1fe8e9" containerName="horizon-log" containerID="cri-o://bfd13f227cb2064f23761c7a562d35816e2cb13aba4850cb48a1f6228a4798cc" gracePeriod=30 Oct 08 19:56:19 crc kubenswrapper[4988]: I1008 19:56:19.154811 4988 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-7d85fc8646-gv4lm" podUID="c533a2e1-7213-4410-b862-5eed5e1fe8e9" containerName="horizon" containerID="cri-o://24a201cb0f08c481ca0dc4c546b32eeb4e83b514268ccaffc0f84cb043a4ef21" gracePeriod=30 Oct 08 19:56:20 crc kubenswrapper[4988]: I1008 19:56:20.953991 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-sync-xs9q8" event={"ID":"162b2195-0c02-4cfc-ae52-ac4cdc1e1a9c","Type":"ContainerStarted","Data":"2f3511be8d5426549a31710170af0dbe435492910a7fc827500070e46b3487ec"} Oct 08 19:56:20 crc kubenswrapper[4988]: I1008 19:56:20.972758 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-db-sync-xs9q8" podStartSLOduration=2.426722068 podStartE2EDuration="8.972736469s" podCreationTimestamp="2025-10-08 19:56:12 +0000 UTC" firstStartedPulling="2025-10-08 19:56:13.166265659 +0000 UTC m=+6318.616108429" lastFinishedPulling="2025-10-08 19:56:19.71228006 +0000 UTC m=+6325.162122830" observedRunningTime="2025-10-08 19:56:20.96744787 +0000 UTC m=+6326.417290660" watchObservedRunningTime="2025-10-08 19:56:20.972736469 +0000 UTC m=+6326.422579249" Oct 08 19:56:22 crc kubenswrapper[4988]: I1008 19:56:22.309617 4988 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-7d85fc8646-gv4lm" podUID="c533a2e1-7213-4410-b862-5eed5e1fe8e9" containerName="horizon" probeResult="failure" output="Get \"https://10.217.1.122:8443/dashboard/auth/login/?next=/dashboard/\": read tcp 10.217.0.2:59094->10.217.1.122:8443: read: connection reset by peer" Oct 08 19:56:22 crc kubenswrapper[4988]: I1008 19:56:22.980603 4988 generic.go:334] "Generic (PLEG): container finished" podID="c533a2e1-7213-4410-b862-5eed5e1fe8e9" containerID="24a201cb0f08c481ca0dc4c546b32eeb4e83b514268ccaffc0f84cb043a4ef21" exitCode=0 Oct 08 19:56:22 crc kubenswrapper[4988]: I1008 19:56:22.980730 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7d85fc8646-gv4lm" event={"ID":"c533a2e1-7213-4410-b862-5eed5e1fe8e9","Type":"ContainerDied","Data":"24a201cb0f08c481ca0dc4c546b32eeb4e83b514268ccaffc0f84cb043a4ef21"} Oct 08 19:56:22 crc kubenswrapper[4988]: I1008 19:56:22.984185 4988 generic.go:334] "Generic (PLEG): container finished" podID="162b2195-0c02-4cfc-ae52-ac4cdc1e1a9c" containerID="2f3511be8d5426549a31710170af0dbe435492910a7fc827500070e46b3487ec" exitCode=0 Oct 08 19:56:22 crc kubenswrapper[4988]: I1008 19:56:22.984258 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-sync-xs9q8" event={"ID":"162b2195-0c02-4cfc-ae52-ac4cdc1e1a9c","Type":"ContainerDied","Data":"2f3511be8d5426549a31710170af0dbe435492910a7fc827500070e46b3487ec"} Oct 08 19:56:24 crc kubenswrapper[4988]: I1008 19:56:24.406918 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-sync-xs9q8" Oct 08 19:56:24 crc kubenswrapper[4988]: I1008 19:56:24.429986 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/162b2195-0c02-4cfc-ae52-ac4cdc1e1a9c-combined-ca-bundle\") pod \"162b2195-0c02-4cfc-ae52-ac4cdc1e1a9c\" (UID: \"162b2195-0c02-4cfc-ae52-ac4cdc1e1a9c\") " Oct 08 19:56:24 crc kubenswrapper[4988]: I1008 19:56:24.430220 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-snnll\" (UniqueName: \"kubernetes.io/projected/162b2195-0c02-4cfc-ae52-ac4cdc1e1a9c-kube-api-access-snnll\") pod \"162b2195-0c02-4cfc-ae52-ac4cdc1e1a9c\" (UID: \"162b2195-0c02-4cfc-ae52-ac4cdc1e1a9c\") " Oct 08 19:56:24 crc kubenswrapper[4988]: I1008 19:56:24.430297 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/162b2195-0c02-4cfc-ae52-ac4cdc1e1a9c-config-data\") pod \"162b2195-0c02-4cfc-ae52-ac4cdc1e1a9c\" (UID: \"162b2195-0c02-4cfc-ae52-ac4cdc1e1a9c\") " Oct 08 19:56:24 crc kubenswrapper[4988]: I1008 19:56:24.436093 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/162b2195-0c02-4cfc-ae52-ac4cdc1e1a9c-kube-api-access-snnll" (OuterVolumeSpecName: "kube-api-access-snnll") pod "162b2195-0c02-4cfc-ae52-ac4cdc1e1a9c" (UID: "162b2195-0c02-4cfc-ae52-ac4cdc1e1a9c"). InnerVolumeSpecName "kube-api-access-snnll". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 19:56:24 crc kubenswrapper[4988]: I1008 19:56:24.462819 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/162b2195-0c02-4cfc-ae52-ac4cdc1e1a9c-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "162b2195-0c02-4cfc-ae52-ac4cdc1e1a9c" (UID: "162b2195-0c02-4cfc-ae52-ac4cdc1e1a9c"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 19:56:24 crc kubenswrapper[4988]: I1008 19:56:24.532282 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/162b2195-0c02-4cfc-ae52-ac4cdc1e1a9c-config-data" (OuterVolumeSpecName: "config-data") pod "162b2195-0c02-4cfc-ae52-ac4cdc1e1a9c" (UID: "162b2195-0c02-4cfc-ae52-ac4cdc1e1a9c"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 19:56:24 crc kubenswrapper[4988]: I1008 19:56:24.532785 4988 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-snnll\" (UniqueName: \"kubernetes.io/projected/162b2195-0c02-4cfc-ae52-ac4cdc1e1a9c-kube-api-access-snnll\") on node \"crc\" DevicePath \"\"" Oct 08 19:56:24 crc kubenswrapper[4988]: I1008 19:56:24.532819 4988 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/162b2195-0c02-4cfc-ae52-ac4cdc1e1a9c-config-data\") on node \"crc\" DevicePath \"\"" Oct 08 19:56:24 crc kubenswrapper[4988]: I1008 19:56:24.532832 4988 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/162b2195-0c02-4cfc-ae52-ac4cdc1e1a9c-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 19:56:25 crc kubenswrapper[4988]: I1008 19:56:25.013049 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-sync-xs9q8" event={"ID":"162b2195-0c02-4cfc-ae52-ac4cdc1e1a9c","Type":"ContainerDied","Data":"4f2ba48827917045fd09ca2a73e5951c6866937321f72a027a8fb5ace14730c3"} Oct 08 19:56:25 crc kubenswrapper[4988]: I1008 19:56:25.013456 4988 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4f2ba48827917045fd09ca2a73e5951c6866937321f72a027a8fb5ace14730c3" Oct 08 19:56:25 crc kubenswrapper[4988]: I1008 19:56:25.013125 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-sync-xs9q8" Oct 08 19:56:26 crc kubenswrapper[4988]: I1008 19:56:26.066562 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-engine-d897f4749-rlhsn"] Oct 08 19:56:26 crc kubenswrapper[4988]: E1008 19:56:26.067131 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="162b2195-0c02-4cfc-ae52-ac4cdc1e1a9c" containerName="heat-db-sync" Oct 08 19:56:26 crc kubenswrapper[4988]: I1008 19:56:26.067152 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="162b2195-0c02-4cfc-ae52-ac4cdc1e1a9c" containerName="heat-db-sync" Oct 08 19:56:26 crc kubenswrapper[4988]: I1008 19:56:26.067435 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="162b2195-0c02-4cfc-ae52-ac4cdc1e1a9c" containerName="heat-db-sync" Oct 08 19:56:26 crc kubenswrapper[4988]: I1008 19:56:26.068179 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-engine-d897f4749-rlhsn" Oct 08 19:56:26 crc kubenswrapper[4988]: I1008 19:56:26.071696 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-heat-dockercfg-cnlv2" Oct 08 19:56:26 crc kubenswrapper[4988]: I1008 19:56:26.071696 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-engine-config-data" Oct 08 19:56:26 crc kubenswrapper[4988]: I1008 19:56:26.071893 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-config-data" Oct 08 19:56:26 crc kubenswrapper[4988]: I1008 19:56:26.106475 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-engine-d897f4749-rlhsn"] Oct 08 19:56:26 crc kubenswrapper[4988]: I1008 19:56:26.167815 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fdbbb87a-7afb-402c-83fd-d4582e06c6ba-config-data\") pod \"heat-engine-d897f4749-rlhsn\" (UID: \"fdbbb87a-7afb-402c-83fd-d4582e06c6ba\") " pod="openstack/heat-engine-d897f4749-rlhsn" Oct 08 19:56:26 crc kubenswrapper[4988]: I1008 19:56:26.167885 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dp87j\" (UniqueName: \"kubernetes.io/projected/fdbbb87a-7afb-402c-83fd-d4582e06c6ba-kube-api-access-dp87j\") pod \"heat-engine-d897f4749-rlhsn\" (UID: \"fdbbb87a-7afb-402c-83fd-d4582e06c6ba\") " pod="openstack/heat-engine-d897f4749-rlhsn" Oct 08 19:56:26 crc kubenswrapper[4988]: I1008 19:56:26.167960 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fdbbb87a-7afb-402c-83fd-d4582e06c6ba-combined-ca-bundle\") pod \"heat-engine-d897f4749-rlhsn\" (UID: \"fdbbb87a-7afb-402c-83fd-d4582e06c6ba\") " pod="openstack/heat-engine-d897f4749-rlhsn" Oct 08 19:56:26 crc kubenswrapper[4988]: I1008 19:56:26.167999 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/fdbbb87a-7afb-402c-83fd-d4582e06c6ba-config-data-custom\") pod \"heat-engine-d897f4749-rlhsn\" (UID: \"fdbbb87a-7afb-402c-83fd-d4582e06c6ba\") " pod="openstack/heat-engine-d897f4749-rlhsn" Oct 08 19:56:26 crc kubenswrapper[4988]: I1008 19:56:26.192347 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-api-8664dcb6b9-jf2pf"] Oct 08 19:56:26 crc kubenswrapper[4988]: I1008 19:56:26.194001 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-8664dcb6b9-jf2pf" Oct 08 19:56:26 crc kubenswrapper[4988]: I1008 19:56:26.198669 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-api-config-data" Oct 08 19:56:26 crc kubenswrapper[4988]: I1008 19:56:26.200155 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-api-8664dcb6b9-jf2pf"] Oct 08 19:56:26 crc kubenswrapper[4988]: I1008 19:56:26.269685 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fdbbb87a-7afb-402c-83fd-d4582e06c6ba-config-data\") pod \"heat-engine-d897f4749-rlhsn\" (UID: \"fdbbb87a-7afb-402c-83fd-d4582e06c6ba\") " pod="openstack/heat-engine-d897f4749-rlhsn" Oct 08 19:56:26 crc kubenswrapper[4988]: I1008 19:56:26.269731 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8g292\" (UniqueName: \"kubernetes.io/projected/255ab071-02d7-46b6-af97-cd84f085b097-kube-api-access-8g292\") pod \"heat-api-8664dcb6b9-jf2pf\" (UID: \"255ab071-02d7-46b6-af97-cd84f085b097\") " pod="openstack/heat-api-8664dcb6b9-jf2pf" Oct 08 19:56:26 crc kubenswrapper[4988]: I1008 19:56:26.269768 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dp87j\" (UniqueName: \"kubernetes.io/projected/fdbbb87a-7afb-402c-83fd-d4582e06c6ba-kube-api-access-dp87j\") pod \"heat-engine-d897f4749-rlhsn\" (UID: \"fdbbb87a-7afb-402c-83fd-d4582e06c6ba\") " pod="openstack/heat-engine-d897f4749-rlhsn" Oct 08 19:56:26 crc kubenswrapper[4988]: I1008 19:56:26.269856 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fdbbb87a-7afb-402c-83fd-d4582e06c6ba-combined-ca-bundle\") pod \"heat-engine-d897f4749-rlhsn\" (UID: \"fdbbb87a-7afb-402c-83fd-d4582e06c6ba\") " pod="openstack/heat-engine-d897f4749-rlhsn" Oct 08 19:56:26 crc kubenswrapper[4988]: I1008 19:56:26.269905 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/fdbbb87a-7afb-402c-83fd-d4582e06c6ba-config-data-custom\") pod \"heat-engine-d897f4749-rlhsn\" (UID: \"fdbbb87a-7afb-402c-83fd-d4582e06c6ba\") " pod="openstack/heat-engine-d897f4749-rlhsn" Oct 08 19:56:26 crc kubenswrapper[4988]: I1008 19:56:26.269932 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/255ab071-02d7-46b6-af97-cd84f085b097-combined-ca-bundle\") pod \"heat-api-8664dcb6b9-jf2pf\" (UID: \"255ab071-02d7-46b6-af97-cd84f085b097\") " pod="openstack/heat-api-8664dcb6b9-jf2pf" Oct 08 19:56:26 crc kubenswrapper[4988]: I1008 19:56:26.269947 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/255ab071-02d7-46b6-af97-cd84f085b097-config-data-custom\") pod \"heat-api-8664dcb6b9-jf2pf\" (UID: \"255ab071-02d7-46b6-af97-cd84f085b097\") " pod="openstack/heat-api-8664dcb6b9-jf2pf" Oct 08 19:56:26 crc kubenswrapper[4988]: I1008 19:56:26.269972 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/255ab071-02d7-46b6-af97-cd84f085b097-config-data\") pod \"heat-api-8664dcb6b9-jf2pf\" (UID: \"255ab071-02d7-46b6-af97-cd84f085b097\") " pod="openstack/heat-api-8664dcb6b9-jf2pf" Oct 08 19:56:26 crc kubenswrapper[4988]: I1008 19:56:26.274185 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-cfnapi-56bdcfdf8b-dbkmh"] Oct 08 19:56:26 crc kubenswrapper[4988]: I1008 19:56:26.278456 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-56bdcfdf8b-dbkmh" Oct 08 19:56:26 crc kubenswrapper[4988]: I1008 19:56:26.278538 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/fdbbb87a-7afb-402c-83fd-d4582e06c6ba-config-data-custom\") pod \"heat-engine-d897f4749-rlhsn\" (UID: \"fdbbb87a-7afb-402c-83fd-d4582e06c6ba\") " pod="openstack/heat-engine-d897f4749-rlhsn" Oct 08 19:56:26 crc kubenswrapper[4988]: I1008 19:56:26.280284 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-cfnapi-config-data" Oct 08 19:56:26 crc kubenswrapper[4988]: I1008 19:56:26.283606 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fdbbb87a-7afb-402c-83fd-d4582e06c6ba-combined-ca-bundle\") pod \"heat-engine-d897f4749-rlhsn\" (UID: \"fdbbb87a-7afb-402c-83fd-d4582e06c6ba\") " pod="openstack/heat-engine-d897f4749-rlhsn" Oct 08 19:56:26 crc kubenswrapper[4988]: I1008 19:56:26.284774 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-cfnapi-56bdcfdf8b-dbkmh"] Oct 08 19:56:26 crc kubenswrapper[4988]: I1008 19:56:26.302133 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dp87j\" (UniqueName: \"kubernetes.io/projected/fdbbb87a-7afb-402c-83fd-d4582e06c6ba-kube-api-access-dp87j\") pod \"heat-engine-d897f4749-rlhsn\" (UID: \"fdbbb87a-7afb-402c-83fd-d4582e06c6ba\") " pod="openstack/heat-engine-d897f4749-rlhsn" Oct 08 19:56:26 crc kubenswrapper[4988]: I1008 19:56:26.305927 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fdbbb87a-7afb-402c-83fd-d4582e06c6ba-config-data\") pod \"heat-engine-d897f4749-rlhsn\" (UID: \"fdbbb87a-7afb-402c-83fd-d4582e06c6ba\") " pod="openstack/heat-engine-d897f4749-rlhsn" Oct 08 19:56:26 crc kubenswrapper[4988]: I1008 19:56:26.372077 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6aeded19-1e22-46b5-8e88-ce9ba127269f-combined-ca-bundle\") pod \"heat-cfnapi-56bdcfdf8b-dbkmh\" (UID: \"6aeded19-1e22-46b5-8e88-ce9ba127269f\") " pod="openstack/heat-cfnapi-56bdcfdf8b-dbkmh" Oct 08 19:56:26 crc kubenswrapper[4988]: I1008 19:56:26.372125 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8g292\" (UniqueName: \"kubernetes.io/projected/255ab071-02d7-46b6-af97-cd84f085b097-kube-api-access-8g292\") pod \"heat-api-8664dcb6b9-jf2pf\" (UID: \"255ab071-02d7-46b6-af97-cd84f085b097\") " pod="openstack/heat-api-8664dcb6b9-jf2pf" Oct 08 19:56:26 crc kubenswrapper[4988]: I1008 19:56:26.372163 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6aeded19-1e22-46b5-8e88-ce9ba127269f-config-data\") pod \"heat-cfnapi-56bdcfdf8b-dbkmh\" (UID: \"6aeded19-1e22-46b5-8e88-ce9ba127269f\") " pod="openstack/heat-cfnapi-56bdcfdf8b-dbkmh" Oct 08 19:56:26 crc kubenswrapper[4988]: I1008 19:56:26.372203 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zfdhk\" (UniqueName: \"kubernetes.io/projected/6aeded19-1e22-46b5-8e88-ce9ba127269f-kube-api-access-zfdhk\") pod \"heat-cfnapi-56bdcfdf8b-dbkmh\" (UID: \"6aeded19-1e22-46b5-8e88-ce9ba127269f\") " pod="openstack/heat-cfnapi-56bdcfdf8b-dbkmh" Oct 08 19:56:26 crc kubenswrapper[4988]: I1008 19:56:26.372246 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/6aeded19-1e22-46b5-8e88-ce9ba127269f-config-data-custom\") pod \"heat-cfnapi-56bdcfdf8b-dbkmh\" (UID: \"6aeded19-1e22-46b5-8e88-ce9ba127269f\") " pod="openstack/heat-cfnapi-56bdcfdf8b-dbkmh" Oct 08 19:56:26 crc kubenswrapper[4988]: I1008 19:56:26.372276 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/255ab071-02d7-46b6-af97-cd84f085b097-combined-ca-bundle\") pod \"heat-api-8664dcb6b9-jf2pf\" (UID: \"255ab071-02d7-46b6-af97-cd84f085b097\") " pod="openstack/heat-api-8664dcb6b9-jf2pf" Oct 08 19:56:26 crc kubenswrapper[4988]: I1008 19:56:26.372292 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/255ab071-02d7-46b6-af97-cd84f085b097-config-data-custom\") pod \"heat-api-8664dcb6b9-jf2pf\" (UID: \"255ab071-02d7-46b6-af97-cd84f085b097\") " pod="openstack/heat-api-8664dcb6b9-jf2pf" Oct 08 19:56:26 crc kubenswrapper[4988]: I1008 19:56:26.372315 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/255ab071-02d7-46b6-af97-cd84f085b097-config-data\") pod \"heat-api-8664dcb6b9-jf2pf\" (UID: \"255ab071-02d7-46b6-af97-cd84f085b097\") " pod="openstack/heat-api-8664dcb6b9-jf2pf" Oct 08 19:56:26 crc kubenswrapper[4988]: I1008 19:56:26.377111 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/255ab071-02d7-46b6-af97-cd84f085b097-config-data-custom\") pod \"heat-api-8664dcb6b9-jf2pf\" (UID: \"255ab071-02d7-46b6-af97-cd84f085b097\") " pod="openstack/heat-api-8664dcb6b9-jf2pf" Oct 08 19:56:26 crc kubenswrapper[4988]: I1008 19:56:26.378513 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/255ab071-02d7-46b6-af97-cd84f085b097-config-data\") pod \"heat-api-8664dcb6b9-jf2pf\" (UID: \"255ab071-02d7-46b6-af97-cd84f085b097\") " pod="openstack/heat-api-8664dcb6b9-jf2pf" Oct 08 19:56:26 crc kubenswrapper[4988]: I1008 19:56:26.383080 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/255ab071-02d7-46b6-af97-cd84f085b097-combined-ca-bundle\") pod \"heat-api-8664dcb6b9-jf2pf\" (UID: \"255ab071-02d7-46b6-af97-cd84f085b097\") " pod="openstack/heat-api-8664dcb6b9-jf2pf" Oct 08 19:56:26 crc kubenswrapper[4988]: I1008 19:56:26.389372 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8g292\" (UniqueName: \"kubernetes.io/projected/255ab071-02d7-46b6-af97-cd84f085b097-kube-api-access-8g292\") pod \"heat-api-8664dcb6b9-jf2pf\" (UID: \"255ab071-02d7-46b6-af97-cd84f085b097\") " pod="openstack/heat-api-8664dcb6b9-jf2pf" Oct 08 19:56:26 crc kubenswrapper[4988]: I1008 19:56:26.399579 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-engine-d897f4749-rlhsn" Oct 08 19:56:26 crc kubenswrapper[4988]: I1008 19:56:26.475573 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6aeded19-1e22-46b5-8e88-ce9ba127269f-config-data\") pod \"heat-cfnapi-56bdcfdf8b-dbkmh\" (UID: \"6aeded19-1e22-46b5-8e88-ce9ba127269f\") " pod="openstack/heat-cfnapi-56bdcfdf8b-dbkmh" Oct 08 19:56:26 crc kubenswrapper[4988]: I1008 19:56:26.475650 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zfdhk\" (UniqueName: \"kubernetes.io/projected/6aeded19-1e22-46b5-8e88-ce9ba127269f-kube-api-access-zfdhk\") pod \"heat-cfnapi-56bdcfdf8b-dbkmh\" (UID: \"6aeded19-1e22-46b5-8e88-ce9ba127269f\") " pod="openstack/heat-cfnapi-56bdcfdf8b-dbkmh" Oct 08 19:56:26 crc kubenswrapper[4988]: I1008 19:56:26.475706 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/6aeded19-1e22-46b5-8e88-ce9ba127269f-config-data-custom\") pod \"heat-cfnapi-56bdcfdf8b-dbkmh\" (UID: \"6aeded19-1e22-46b5-8e88-ce9ba127269f\") " pod="openstack/heat-cfnapi-56bdcfdf8b-dbkmh" Oct 08 19:56:26 crc kubenswrapper[4988]: I1008 19:56:26.475817 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6aeded19-1e22-46b5-8e88-ce9ba127269f-combined-ca-bundle\") pod \"heat-cfnapi-56bdcfdf8b-dbkmh\" (UID: \"6aeded19-1e22-46b5-8e88-ce9ba127269f\") " pod="openstack/heat-cfnapi-56bdcfdf8b-dbkmh" Oct 08 19:56:26 crc kubenswrapper[4988]: I1008 19:56:26.481130 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6aeded19-1e22-46b5-8e88-ce9ba127269f-combined-ca-bundle\") pod \"heat-cfnapi-56bdcfdf8b-dbkmh\" (UID: \"6aeded19-1e22-46b5-8e88-ce9ba127269f\") " pod="openstack/heat-cfnapi-56bdcfdf8b-dbkmh" Oct 08 19:56:26 crc kubenswrapper[4988]: I1008 19:56:26.492191 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/6aeded19-1e22-46b5-8e88-ce9ba127269f-config-data-custom\") pod \"heat-cfnapi-56bdcfdf8b-dbkmh\" (UID: \"6aeded19-1e22-46b5-8e88-ce9ba127269f\") " pod="openstack/heat-cfnapi-56bdcfdf8b-dbkmh" Oct 08 19:56:26 crc kubenswrapper[4988]: I1008 19:56:26.492462 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6aeded19-1e22-46b5-8e88-ce9ba127269f-config-data\") pod \"heat-cfnapi-56bdcfdf8b-dbkmh\" (UID: \"6aeded19-1e22-46b5-8e88-ce9ba127269f\") " pod="openstack/heat-cfnapi-56bdcfdf8b-dbkmh" Oct 08 19:56:26 crc kubenswrapper[4988]: I1008 19:56:26.494954 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zfdhk\" (UniqueName: \"kubernetes.io/projected/6aeded19-1e22-46b5-8e88-ce9ba127269f-kube-api-access-zfdhk\") pod \"heat-cfnapi-56bdcfdf8b-dbkmh\" (UID: \"6aeded19-1e22-46b5-8e88-ce9ba127269f\") " pod="openstack/heat-cfnapi-56bdcfdf8b-dbkmh" Oct 08 19:56:26 crc kubenswrapper[4988]: I1008 19:56:26.522868 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-8664dcb6b9-jf2pf" Oct 08 19:56:26 crc kubenswrapper[4988]: I1008 19:56:26.553476 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-56bdcfdf8b-dbkmh" Oct 08 19:56:26 crc kubenswrapper[4988]: I1008 19:56:26.908453 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-engine-d897f4749-rlhsn"] Oct 08 19:56:27 crc kubenswrapper[4988]: I1008 19:56:27.033908 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-engine-d897f4749-rlhsn" event={"ID":"fdbbb87a-7afb-402c-83fd-d4582e06c6ba","Type":"ContainerStarted","Data":"f6214419fd3decc57de42d8836027405c6c2bd3700fdcaaa99a9a4101da92b9d"} Oct 08 19:56:27 crc kubenswrapper[4988]: I1008 19:56:27.058108 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-api-8664dcb6b9-jf2pf"] Oct 08 19:56:27 crc kubenswrapper[4988]: I1008 19:56:27.157351 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-cfnapi-56bdcfdf8b-dbkmh"] Oct 08 19:56:27 crc kubenswrapper[4988]: W1008 19:56:27.166477 4988 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6aeded19_1e22_46b5_8e88_ce9ba127269f.slice/crio-ad0d72c598d9925765913496cce57e27b9a598d1b282086861baae2d03c8d989 WatchSource:0}: Error finding container ad0d72c598d9925765913496cce57e27b9a598d1b282086861baae2d03c8d989: Status 404 returned error can't find the container with id ad0d72c598d9925765913496cce57e27b9a598d1b282086861baae2d03c8d989 Oct 08 19:56:28 crc kubenswrapper[4988]: I1008 19:56:28.043918 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-engine-d897f4749-rlhsn" event={"ID":"fdbbb87a-7afb-402c-83fd-d4582e06c6ba","Type":"ContainerStarted","Data":"c29143467f18a0d9cb5e4136506cbf1b8520fd0496108cf77db65c962e6ea146"} Oct 08 19:56:28 crc kubenswrapper[4988]: I1008 19:56:28.044213 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-engine-d897f4749-rlhsn" Oct 08 19:56:28 crc kubenswrapper[4988]: I1008 19:56:28.045443 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-56bdcfdf8b-dbkmh" event={"ID":"6aeded19-1e22-46b5-8e88-ce9ba127269f","Type":"ContainerStarted","Data":"ad0d72c598d9925765913496cce57e27b9a598d1b282086861baae2d03c8d989"} Oct 08 19:56:28 crc kubenswrapper[4988]: I1008 19:56:28.046829 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-8664dcb6b9-jf2pf" event={"ID":"255ab071-02d7-46b6-af97-cd84f085b097","Type":"ContainerStarted","Data":"66a6de057679cd07810c71c9a7402c7a0ecc6b93c98a269421e26c301a24115f"} Oct 08 19:56:28 crc kubenswrapper[4988]: I1008 19:56:28.061587 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-engine-d897f4749-rlhsn" podStartSLOduration=2.06156754 podStartE2EDuration="2.06156754s" podCreationTimestamp="2025-10-08 19:56:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 19:56:28.057426179 +0000 UTC m=+6333.507268949" watchObservedRunningTime="2025-10-08 19:56:28.06156754 +0000 UTC m=+6333.511410310" Oct 08 19:56:29 crc kubenswrapper[4988]: I1008 19:56:29.846224 4988 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-7d85fc8646-gv4lm" podUID="c533a2e1-7213-4410-b862-5eed5e1fe8e9" containerName="horizon" probeResult="failure" output="Get \"https://10.217.1.122:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.1.122:8443: connect: connection refused" Oct 08 19:56:30 crc kubenswrapper[4988]: I1008 19:56:30.066563 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-56bdcfdf8b-dbkmh" event={"ID":"6aeded19-1e22-46b5-8e88-ce9ba127269f","Type":"ContainerStarted","Data":"33c6f5043911fe4478477ea2b8457d1abcff028ec754adec057bfb6a524ecb3b"} Oct 08 19:56:30 crc kubenswrapper[4988]: I1008 19:56:30.066645 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-cfnapi-56bdcfdf8b-dbkmh" Oct 08 19:56:30 crc kubenswrapper[4988]: I1008 19:56:30.069041 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-8664dcb6b9-jf2pf" event={"ID":"255ab071-02d7-46b6-af97-cd84f085b097","Type":"ContainerStarted","Data":"ea11f6f5c59b0839c5743a2f38ec67a1da94e3ef717c1e36f809ce296a109d4e"} Oct 08 19:56:30 crc kubenswrapper[4988]: I1008 19:56:30.070128 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-api-8664dcb6b9-jf2pf" Oct 08 19:56:30 crc kubenswrapper[4988]: I1008 19:56:30.109548 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-cfnapi-56bdcfdf8b-dbkmh" podStartSLOduration=1.998928665 podStartE2EDuration="4.109531254s" podCreationTimestamp="2025-10-08 19:56:26 +0000 UTC" firstStartedPulling="2025-10-08 19:56:27.168560075 +0000 UTC m=+6332.618402845" lastFinishedPulling="2025-10-08 19:56:29.279162644 +0000 UTC m=+6334.729005434" observedRunningTime="2025-10-08 19:56:30.085893011 +0000 UTC m=+6335.535735781" watchObservedRunningTime="2025-10-08 19:56:30.109531254 +0000 UTC m=+6335.559374024" Oct 08 19:56:30 crc kubenswrapper[4988]: I1008 19:56:30.109827 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-api-8664dcb6b9-jf2pf" podStartSLOduration=1.906864793 podStartE2EDuration="4.109824084s" podCreationTimestamp="2025-10-08 19:56:26 +0000 UTC" firstStartedPulling="2025-10-08 19:56:27.068306282 +0000 UTC m=+6332.518149052" lastFinishedPulling="2025-10-08 19:56:29.271265573 +0000 UTC m=+6334.721108343" observedRunningTime="2025-10-08 19:56:30.104204255 +0000 UTC m=+6335.554047035" watchObservedRunningTime="2025-10-08 19:56:30.109824084 +0000 UTC m=+6335.559666854" Oct 08 19:56:33 crc kubenswrapper[4988]: I1008 19:56:33.062431 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-engine-78cc4595db-p87kc"] Oct 08 19:56:33 crc kubenswrapper[4988]: I1008 19:56:33.064720 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-engine-78cc4595db-p87kc" Oct 08 19:56:33 crc kubenswrapper[4988]: I1008 19:56:33.079749 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-cfnapi-b969444b9-mhxbw"] Oct 08 19:56:33 crc kubenswrapper[4988]: I1008 19:56:33.081338 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-b969444b9-mhxbw" Oct 08 19:56:33 crc kubenswrapper[4988]: I1008 19:56:33.091530 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-engine-78cc4595db-p87kc"] Oct 08 19:56:33 crc kubenswrapper[4988]: I1008 19:56:33.102668 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-api-59756cd4b5-b6ghv"] Oct 08 19:56:33 crc kubenswrapper[4988]: I1008 19:56:33.104279 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-59756cd4b5-b6ghv" Oct 08 19:56:33 crc kubenswrapper[4988]: I1008 19:56:33.120233 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-cfnapi-b969444b9-mhxbw"] Oct 08 19:56:33 crc kubenswrapper[4988]: I1008 19:56:33.128520 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/807b4691-54e6-4ae9-abeb-8e95ab06578a-combined-ca-bundle\") pod \"heat-engine-78cc4595db-p87kc\" (UID: \"807b4691-54e6-4ae9-abeb-8e95ab06578a\") " pod="openstack/heat-engine-78cc4595db-p87kc" Oct 08 19:56:33 crc kubenswrapper[4988]: I1008 19:56:33.128610 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z27k8\" (UniqueName: \"kubernetes.io/projected/9fb22c64-8f41-496f-b894-d1c7d9b80125-kube-api-access-z27k8\") pod \"heat-cfnapi-b969444b9-mhxbw\" (UID: \"9fb22c64-8f41-496f-b894-d1c7d9b80125\") " pod="openstack/heat-cfnapi-b969444b9-mhxbw" Oct 08 19:56:33 crc kubenswrapper[4988]: I1008 19:56:33.128665 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/807b4691-54e6-4ae9-abeb-8e95ab06578a-config-data-custom\") pod \"heat-engine-78cc4595db-p87kc\" (UID: \"807b4691-54e6-4ae9-abeb-8e95ab06578a\") " pod="openstack/heat-engine-78cc4595db-p87kc" Oct 08 19:56:33 crc kubenswrapper[4988]: I1008 19:56:33.128736 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9fb22c64-8f41-496f-b894-d1c7d9b80125-config-data\") pod \"heat-cfnapi-b969444b9-mhxbw\" (UID: \"9fb22c64-8f41-496f-b894-d1c7d9b80125\") " pod="openstack/heat-cfnapi-b969444b9-mhxbw" Oct 08 19:56:33 crc kubenswrapper[4988]: I1008 19:56:33.128816 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b722t\" (UniqueName: \"kubernetes.io/projected/ae3bd686-c190-4513-814a-0e6352ce3fdb-kube-api-access-b722t\") pod \"heat-api-59756cd4b5-b6ghv\" (UID: \"ae3bd686-c190-4513-814a-0e6352ce3fdb\") " pod="openstack/heat-api-59756cd4b5-b6ghv" Oct 08 19:56:33 crc kubenswrapper[4988]: I1008 19:56:33.128868 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ae3bd686-c190-4513-814a-0e6352ce3fdb-config-data\") pod \"heat-api-59756cd4b5-b6ghv\" (UID: \"ae3bd686-c190-4513-814a-0e6352ce3fdb\") " pod="openstack/heat-api-59756cd4b5-b6ghv" Oct 08 19:56:33 crc kubenswrapper[4988]: I1008 19:56:33.128913 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ldr5r\" (UniqueName: \"kubernetes.io/projected/807b4691-54e6-4ae9-abeb-8e95ab06578a-kube-api-access-ldr5r\") pod \"heat-engine-78cc4595db-p87kc\" (UID: \"807b4691-54e6-4ae9-abeb-8e95ab06578a\") " pod="openstack/heat-engine-78cc4595db-p87kc" Oct 08 19:56:33 crc kubenswrapper[4988]: I1008 19:56:33.128976 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/9fb22c64-8f41-496f-b894-d1c7d9b80125-config-data-custom\") pod \"heat-cfnapi-b969444b9-mhxbw\" (UID: \"9fb22c64-8f41-496f-b894-d1c7d9b80125\") " pod="openstack/heat-cfnapi-b969444b9-mhxbw" Oct 08 19:56:33 crc kubenswrapper[4988]: I1008 19:56:33.129013 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/807b4691-54e6-4ae9-abeb-8e95ab06578a-config-data\") pod \"heat-engine-78cc4595db-p87kc\" (UID: \"807b4691-54e6-4ae9-abeb-8e95ab06578a\") " pod="openstack/heat-engine-78cc4595db-p87kc" Oct 08 19:56:33 crc kubenswrapper[4988]: I1008 19:56:33.129156 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ae3bd686-c190-4513-814a-0e6352ce3fdb-combined-ca-bundle\") pod \"heat-api-59756cd4b5-b6ghv\" (UID: \"ae3bd686-c190-4513-814a-0e6352ce3fdb\") " pod="openstack/heat-api-59756cd4b5-b6ghv" Oct 08 19:56:33 crc kubenswrapper[4988]: I1008 19:56:33.129274 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9fb22c64-8f41-496f-b894-d1c7d9b80125-combined-ca-bundle\") pod \"heat-cfnapi-b969444b9-mhxbw\" (UID: \"9fb22c64-8f41-496f-b894-d1c7d9b80125\") " pod="openstack/heat-cfnapi-b969444b9-mhxbw" Oct 08 19:56:33 crc kubenswrapper[4988]: I1008 19:56:33.129322 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/ae3bd686-c190-4513-814a-0e6352ce3fdb-config-data-custom\") pod \"heat-api-59756cd4b5-b6ghv\" (UID: \"ae3bd686-c190-4513-814a-0e6352ce3fdb\") " pod="openstack/heat-api-59756cd4b5-b6ghv" Oct 08 19:56:33 crc kubenswrapper[4988]: I1008 19:56:33.137145 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-api-59756cd4b5-b6ghv"] Oct 08 19:56:33 crc kubenswrapper[4988]: I1008 19:56:33.231168 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/9fb22c64-8f41-496f-b894-d1c7d9b80125-config-data-custom\") pod \"heat-cfnapi-b969444b9-mhxbw\" (UID: \"9fb22c64-8f41-496f-b894-d1c7d9b80125\") " pod="openstack/heat-cfnapi-b969444b9-mhxbw" Oct 08 19:56:33 crc kubenswrapper[4988]: I1008 19:56:33.231279 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/807b4691-54e6-4ae9-abeb-8e95ab06578a-config-data\") pod \"heat-engine-78cc4595db-p87kc\" (UID: \"807b4691-54e6-4ae9-abeb-8e95ab06578a\") " pod="openstack/heat-engine-78cc4595db-p87kc" Oct 08 19:56:33 crc kubenswrapper[4988]: I1008 19:56:33.231325 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ae3bd686-c190-4513-814a-0e6352ce3fdb-combined-ca-bundle\") pod \"heat-api-59756cd4b5-b6ghv\" (UID: \"ae3bd686-c190-4513-814a-0e6352ce3fdb\") " pod="openstack/heat-api-59756cd4b5-b6ghv" Oct 08 19:56:33 crc kubenswrapper[4988]: I1008 19:56:33.231425 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9fb22c64-8f41-496f-b894-d1c7d9b80125-combined-ca-bundle\") pod \"heat-cfnapi-b969444b9-mhxbw\" (UID: \"9fb22c64-8f41-496f-b894-d1c7d9b80125\") " pod="openstack/heat-cfnapi-b969444b9-mhxbw" Oct 08 19:56:33 crc kubenswrapper[4988]: I1008 19:56:33.231504 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/ae3bd686-c190-4513-814a-0e6352ce3fdb-config-data-custom\") pod \"heat-api-59756cd4b5-b6ghv\" (UID: \"ae3bd686-c190-4513-814a-0e6352ce3fdb\") " pod="openstack/heat-api-59756cd4b5-b6ghv" Oct 08 19:56:33 crc kubenswrapper[4988]: I1008 19:56:33.231585 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/807b4691-54e6-4ae9-abeb-8e95ab06578a-combined-ca-bundle\") pod \"heat-engine-78cc4595db-p87kc\" (UID: \"807b4691-54e6-4ae9-abeb-8e95ab06578a\") " pod="openstack/heat-engine-78cc4595db-p87kc" Oct 08 19:56:33 crc kubenswrapper[4988]: I1008 19:56:33.231647 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z27k8\" (UniqueName: \"kubernetes.io/projected/9fb22c64-8f41-496f-b894-d1c7d9b80125-kube-api-access-z27k8\") pod \"heat-cfnapi-b969444b9-mhxbw\" (UID: \"9fb22c64-8f41-496f-b894-d1c7d9b80125\") " pod="openstack/heat-cfnapi-b969444b9-mhxbw" Oct 08 19:56:33 crc kubenswrapper[4988]: I1008 19:56:33.231709 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/807b4691-54e6-4ae9-abeb-8e95ab06578a-config-data-custom\") pod \"heat-engine-78cc4595db-p87kc\" (UID: \"807b4691-54e6-4ae9-abeb-8e95ab06578a\") " pod="openstack/heat-engine-78cc4595db-p87kc" Oct 08 19:56:33 crc kubenswrapper[4988]: I1008 19:56:33.231780 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9fb22c64-8f41-496f-b894-d1c7d9b80125-config-data\") pod \"heat-cfnapi-b969444b9-mhxbw\" (UID: \"9fb22c64-8f41-496f-b894-d1c7d9b80125\") " pod="openstack/heat-cfnapi-b969444b9-mhxbw" Oct 08 19:56:33 crc kubenswrapper[4988]: I1008 19:56:33.231857 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b722t\" (UniqueName: \"kubernetes.io/projected/ae3bd686-c190-4513-814a-0e6352ce3fdb-kube-api-access-b722t\") pod \"heat-api-59756cd4b5-b6ghv\" (UID: \"ae3bd686-c190-4513-814a-0e6352ce3fdb\") " pod="openstack/heat-api-59756cd4b5-b6ghv" Oct 08 19:56:33 crc kubenswrapper[4988]: I1008 19:56:33.231907 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ae3bd686-c190-4513-814a-0e6352ce3fdb-config-data\") pod \"heat-api-59756cd4b5-b6ghv\" (UID: \"ae3bd686-c190-4513-814a-0e6352ce3fdb\") " pod="openstack/heat-api-59756cd4b5-b6ghv" Oct 08 19:56:33 crc kubenswrapper[4988]: I1008 19:56:33.231957 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ldr5r\" (UniqueName: \"kubernetes.io/projected/807b4691-54e6-4ae9-abeb-8e95ab06578a-kube-api-access-ldr5r\") pod \"heat-engine-78cc4595db-p87kc\" (UID: \"807b4691-54e6-4ae9-abeb-8e95ab06578a\") " pod="openstack/heat-engine-78cc4595db-p87kc" Oct 08 19:56:33 crc kubenswrapper[4988]: I1008 19:56:33.237993 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/807b4691-54e6-4ae9-abeb-8e95ab06578a-combined-ca-bundle\") pod \"heat-engine-78cc4595db-p87kc\" (UID: \"807b4691-54e6-4ae9-abeb-8e95ab06578a\") " pod="openstack/heat-engine-78cc4595db-p87kc" Oct 08 19:56:33 crc kubenswrapper[4988]: I1008 19:56:33.238775 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9fb22c64-8f41-496f-b894-d1c7d9b80125-config-data\") pod \"heat-cfnapi-b969444b9-mhxbw\" (UID: \"9fb22c64-8f41-496f-b894-d1c7d9b80125\") " pod="openstack/heat-cfnapi-b969444b9-mhxbw" Oct 08 19:56:33 crc kubenswrapper[4988]: I1008 19:56:33.239108 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/ae3bd686-c190-4513-814a-0e6352ce3fdb-config-data-custom\") pod \"heat-api-59756cd4b5-b6ghv\" (UID: \"ae3bd686-c190-4513-814a-0e6352ce3fdb\") " pod="openstack/heat-api-59756cd4b5-b6ghv" Oct 08 19:56:33 crc kubenswrapper[4988]: I1008 19:56:33.239539 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9fb22c64-8f41-496f-b894-d1c7d9b80125-combined-ca-bundle\") pod \"heat-cfnapi-b969444b9-mhxbw\" (UID: \"9fb22c64-8f41-496f-b894-d1c7d9b80125\") " pod="openstack/heat-cfnapi-b969444b9-mhxbw" Oct 08 19:56:33 crc kubenswrapper[4988]: I1008 19:56:33.245819 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/807b4691-54e6-4ae9-abeb-8e95ab06578a-config-data-custom\") pod \"heat-engine-78cc4595db-p87kc\" (UID: \"807b4691-54e6-4ae9-abeb-8e95ab06578a\") " pod="openstack/heat-engine-78cc4595db-p87kc" Oct 08 19:56:33 crc kubenswrapper[4988]: I1008 19:56:33.246404 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ae3bd686-c190-4513-814a-0e6352ce3fdb-config-data\") pod \"heat-api-59756cd4b5-b6ghv\" (UID: \"ae3bd686-c190-4513-814a-0e6352ce3fdb\") " pod="openstack/heat-api-59756cd4b5-b6ghv" Oct 08 19:56:33 crc kubenswrapper[4988]: I1008 19:56:33.250448 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ae3bd686-c190-4513-814a-0e6352ce3fdb-combined-ca-bundle\") pod \"heat-api-59756cd4b5-b6ghv\" (UID: \"ae3bd686-c190-4513-814a-0e6352ce3fdb\") " pod="openstack/heat-api-59756cd4b5-b6ghv" Oct 08 19:56:33 crc kubenswrapper[4988]: I1008 19:56:33.250608 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/807b4691-54e6-4ae9-abeb-8e95ab06578a-config-data\") pod \"heat-engine-78cc4595db-p87kc\" (UID: \"807b4691-54e6-4ae9-abeb-8e95ab06578a\") " pod="openstack/heat-engine-78cc4595db-p87kc" Oct 08 19:56:33 crc kubenswrapper[4988]: I1008 19:56:33.251884 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b722t\" (UniqueName: \"kubernetes.io/projected/ae3bd686-c190-4513-814a-0e6352ce3fdb-kube-api-access-b722t\") pod \"heat-api-59756cd4b5-b6ghv\" (UID: \"ae3bd686-c190-4513-814a-0e6352ce3fdb\") " pod="openstack/heat-api-59756cd4b5-b6ghv" Oct 08 19:56:33 crc kubenswrapper[4988]: I1008 19:56:33.255091 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/9fb22c64-8f41-496f-b894-d1c7d9b80125-config-data-custom\") pod \"heat-cfnapi-b969444b9-mhxbw\" (UID: \"9fb22c64-8f41-496f-b894-d1c7d9b80125\") " pod="openstack/heat-cfnapi-b969444b9-mhxbw" Oct 08 19:56:33 crc kubenswrapper[4988]: I1008 19:56:33.256558 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z27k8\" (UniqueName: \"kubernetes.io/projected/9fb22c64-8f41-496f-b894-d1c7d9b80125-kube-api-access-z27k8\") pod \"heat-cfnapi-b969444b9-mhxbw\" (UID: \"9fb22c64-8f41-496f-b894-d1c7d9b80125\") " pod="openstack/heat-cfnapi-b969444b9-mhxbw" Oct 08 19:56:33 crc kubenswrapper[4988]: I1008 19:56:33.256621 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ldr5r\" (UniqueName: \"kubernetes.io/projected/807b4691-54e6-4ae9-abeb-8e95ab06578a-kube-api-access-ldr5r\") pod \"heat-engine-78cc4595db-p87kc\" (UID: \"807b4691-54e6-4ae9-abeb-8e95ab06578a\") " pod="openstack/heat-engine-78cc4595db-p87kc" Oct 08 19:56:33 crc kubenswrapper[4988]: I1008 19:56:33.438327 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-engine-78cc4595db-p87kc" Oct 08 19:56:33 crc kubenswrapper[4988]: I1008 19:56:33.450241 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-b969444b9-mhxbw" Oct 08 19:56:33 crc kubenswrapper[4988]: I1008 19:56:33.471557 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-59756cd4b5-b6ghv" Oct 08 19:56:34 crc kubenswrapper[4988]: I1008 19:56:33.999435 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-engine-78cc4595db-p87kc"] Oct 08 19:56:34 crc kubenswrapper[4988]: W1008 19:56:34.005374 4988 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod807b4691_54e6_4ae9_abeb_8e95ab06578a.slice/crio-bf63f1e87409a7788fad4ffcd5ce4cbd7ff23e32f359a22c98cc1ad4e2724b6d WatchSource:0}: Error finding container bf63f1e87409a7788fad4ffcd5ce4cbd7ff23e32f359a22c98cc1ad4e2724b6d: Status 404 returned error can't find the container with id bf63f1e87409a7788fad4ffcd5ce4cbd7ff23e32f359a22c98cc1ad4e2724b6d Oct 08 19:56:34 crc kubenswrapper[4988]: W1008 19:56:34.009891 4988 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9fb22c64_8f41_496f_b894_d1c7d9b80125.slice/crio-2281e4fe7eb5a009b1302cb3881ca439ab4c236333f51bc20071f187cda6884a WatchSource:0}: Error finding container 2281e4fe7eb5a009b1302cb3881ca439ab4c236333f51bc20071f187cda6884a: Status 404 returned error can't find the container with id 2281e4fe7eb5a009b1302cb3881ca439ab4c236333f51bc20071f187cda6884a Oct 08 19:56:34 crc kubenswrapper[4988]: I1008 19:56:34.012236 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-cfnapi-b969444b9-mhxbw"] Oct 08 19:56:34 crc kubenswrapper[4988]: I1008 19:56:34.135972 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-api-59756cd4b5-b6ghv"] Oct 08 19:56:34 crc kubenswrapper[4988]: I1008 19:56:34.137651 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-engine-78cc4595db-p87kc" event={"ID":"807b4691-54e6-4ae9-abeb-8e95ab06578a","Type":"ContainerStarted","Data":"bf63f1e87409a7788fad4ffcd5ce4cbd7ff23e32f359a22c98cc1ad4e2724b6d"} Oct 08 19:56:34 crc kubenswrapper[4988]: I1008 19:56:34.145259 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-b969444b9-mhxbw" event={"ID":"9fb22c64-8f41-496f-b894-d1c7d9b80125","Type":"ContainerStarted","Data":"2281e4fe7eb5a009b1302cb3881ca439ab4c236333f51bc20071f187cda6884a"} Oct 08 19:56:34 crc kubenswrapper[4988]: I1008 19:56:34.202251 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-api-8664dcb6b9-jf2pf"] Oct 08 19:56:34 crc kubenswrapper[4988]: I1008 19:56:34.202466 4988 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/heat-api-8664dcb6b9-jf2pf" podUID="255ab071-02d7-46b6-af97-cd84f085b097" containerName="heat-api" containerID="cri-o://ea11f6f5c59b0839c5743a2f38ec67a1da94e3ef717c1e36f809ce296a109d4e" gracePeriod=60 Oct 08 19:56:34 crc kubenswrapper[4988]: I1008 19:56:34.216682 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-cfnapi-56bdcfdf8b-dbkmh"] Oct 08 19:56:34 crc kubenswrapper[4988]: I1008 19:56:34.216932 4988 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/heat-cfnapi-56bdcfdf8b-dbkmh" podUID="6aeded19-1e22-46b5-8e88-ce9ba127269f" containerName="heat-cfnapi" containerID="cri-o://33c6f5043911fe4478477ea2b8457d1abcff028ec754adec057bfb6a524ecb3b" gracePeriod=60 Oct 08 19:56:34 crc kubenswrapper[4988]: I1008 19:56:34.234599 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-api-594c995686-vt22w"] Oct 08 19:56:34 crc kubenswrapper[4988]: I1008 19:56:34.236307 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-594c995686-vt22w" Oct 08 19:56:34 crc kubenswrapper[4988]: I1008 19:56:34.243764 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-heat-api-public-svc" Oct 08 19:56:34 crc kubenswrapper[4988]: I1008 19:56:34.244153 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-heat-api-internal-svc" Oct 08 19:56:34 crc kubenswrapper[4988]: I1008 19:56:34.248558 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-cfnapi-9587dcd9d-cx57b"] Oct 08 19:56:34 crc kubenswrapper[4988]: I1008 19:56:34.250021 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-9587dcd9d-cx57b" Oct 08 19:56:34 crc kubenswrapper[4988]: I1008 19:56:34.253303 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-heat-cfnapi-public-svc" Oct 08 19:56:34 crc kubenswrapper[4988]: I1008 19:56:34.253627 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-heat-cfnapi-internal-svc" Oct 08 19:56:34 crc kubenswrapper[4988]: I1008 19:56:34.255987 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7f6fs\" (UniqueName: \"kubernetes.io/projected/e6dd3357-9d84-4c92-8e40-69f19dd64a83-kube-api-access-7f6fs\") pod \"heat-cfnapi-9587dcd9d-cx57b\" (UID: \"e6dd3357-9d84-4c92-8e40-69f19dd64a83\") " pod="openstack/heat-cfnapi-9587dcd9d-cx57b" Oct 08 19:56:34 crc kubenswrapper[4988]: I1008 19:56:34.256132 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e988e6d2-d2c3-46b6-a6e4-e8524f3fbe8a-combined-ca-bundle\") pod \"heat-api-594c995686-vt22w\" (UID: \"e988e6d2-d2c3-46b6-a6e4-e8524f3fbe8a\") " pod="openstack/heat-api-594c995686-vt22w" Oct 08 19:56:34 crc kubenswrapper[4988]: I1008 19:56:34.256242 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/e988e6d2-d2c3-46b6-a6e4-e8524f3fbe8a-internal-tls-certs\") pod \"heat-api-594c995686-vt22w\" (UID: \"e988e6d2-d2c3-46b6-a6e4-e8524f3fbe8a\") " pod="openstack/heat-api-594c995686-vt22w" Oct 08 19:56:34 crc kubenswrapper[4988]: I1008 19:56:34.256338 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e988e6d2-d2c3-46b6-a6e4-e8524f3fbe8a-config-data\") pod \"heat-api-594c995686-vt22w\" (UID: \"e988e6d2-d2c3-46b6-a6e4-e8524f3fbe8a\") " pod="openstack/heat-api-594c995686-vt22w" Oct 08 19:56:34 crc kubenswrapper[4988]: I1008 19:56:34.256457 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e6dd3357-9d84-4c92-8e40-69f19dd64a83-public-tls-certs\") pod \"heat-cfnapi-9587dcd9d-cx57b\" (UID: \"e6dd3357-9d84-4c92-8e40-69f19dd64a83\") " pod="openstack/heat-cfnapi-9587dcd9d-cx57b" Oct 08 19:56:34 crc kubenswrapper[4988]: I1008 19:56:34.256551 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e6dd3357-9d84-4c92-8e40-69f19dd64a83-combined-ca-bundle\") pod \"heat-cfnapi-9587dcd9d-cx57b\" (UID: \"e6dd3357-9d84-4c92-8e40-69f19dd64a83\") " pod="openstack/heat-cfnapi-9587dcd9d-cx57b" Oct 08 19:56:34 crc kubenswrapper[4988]: I1008 19:56:34.256667 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e988e6d2-d2c3-46b6-a6e4-e8524f3fbe8a-public-tls-certs\") pod \"heat-api-594c995686-vt22w\" (UID: \"e988e6d2-d2c3-46b6-a6e4-e8524f3fbe8a\") " pod="openstack/heat-api-594c995686-vt22w" Oct 08 19:56:34 crc kubenswrapper[4988]: I1008 19:56:34.256754 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e6dd3357-9d84-4c92-8e40-69f19dd64a83-config-data-custom\") pod \"heat-cfnapi-9587dcd9d-cx57b\" (UID: \"e6dd3357-9d84-4c92-8e40-69f19dd64a83\") " pod="openstack/heat-cfnapi-9587dcd9d-cx57b" Oct 08 19:56:34 crc kubenswrapper[4988]: I1008 19:56:34.256825 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e6dd3357-9d84-4c92-8e40-69f19dd64a83-config-data\") pod \"heat-cfnapi-9587dcd9d-cx57b\" (UID: \"e6dd3357-9d84-4c92-8e40-69f19dd64a83\") " pod="openstack/heat-cfnapi-9587dcd9d-cx57b" Oct 08 19:56:34 crc kubenswrapper[4988]: I1008 19:56:34.256925 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dvp9p\" (UniqueName: \"kubernetes.io/projected/e988e6d2-d2c3-46b6-a6e4-e8524f3fbe8a-kube-api-access-dvp9p\") pod \"heat-api-594c995686-vt22w\" (UID: \"e988e6d2-d2c3-46b6-a6e4-e8524f3fbe8a\") " pod="openstack/heat-api-594c995686-vt22w" Oct 08 19:56:34 crc kubenswrapper[4988]: I1008 19:56:34.257011 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e988e6d2-d2c3-46b6-a6e4-e8524f3fbe8a-config-data-custom\") pod \"heat-api-594c995686-vt22w\" (UID: \"e988e6d2-d2c3-46b6-a6e4-e8524f3fbe8a\") " pod="openstack/heat-api-594c995686-vt22w" Oct 08 19:56:34 crc kubenswrapper[4988]: I1008 19:56:34.257076 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/e6dd3357-9d84-4c92-8e40-69f19dd64a83-internal-tls-certs\") pod \"heat-cfnapi-9587dcd9d-cx57b\" (UID: \"e6dd3357-9d84-4c92-8e40-69f19dd64a83\") " pod="openstack/heat-cfnapi-9587dcd9d-cx57b" Oct 08 19:56:34 crc kubenswrapper[4988]: I1008 19:56:34.265512 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-api-594c995686-vt22w"] Oct 08 19:56:34 crc kubenswrapper[4988]: I1008 19:56:34.275111 4988 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/heat-api-8664dcb6b9-jf2pf" podUID="255ab071-02d7-46b6-af97-cd84f085b097" containerName="heat-api" probeResult="failure" output="Get \"http://10.217.1.130:8004/healthcheck\": EOF" Oct 08 19:56:34 crc kubenswrapper[4988]: I1008 19:56:34.278076 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-cfnapi-9587dcd9d-cx57b"] Oct 08 19:56:34 crc kubenswrapper[4988]: I1008 19:56:34.358640 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/e988e6d2-d2c3-46b6-a6e4-e8524f3fbe8a-internal-tls-certs\") pod \"heat-api-594c995686-vt22w\" (UID: \"e988e6d2-d2c3-46b6-a6e4-e8524f3fbe8a\") " pod="openstack/heat-api-594c995686-vt22w" Oct 08 19:56:34 crc kubenswrapper[4988]: I1008 19:56:34.358717 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e988e6d2-d2c3-46b6-a6e4-e8524f3fbe8a-config-data\") pod \"heat-api-594c995686-vt22w\" (UID: \"e988e6d2-d2c3-46b6-a6e4-e8524f3fbe8a\") " pod="openstack/heat-api-594c995686-vt22w" Oct 08 19:56:34 crc kubenswrapper[4988]: I1008 19:56:34.358775 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e6dd3357-9d84-4c92-8e40-69f19dd64a83-public-tls-certs\") pod \"heat-cfnapi-9587dcd9d-cx57b\" (UID: \"e6dd3357-9d84-4c92-8e40-69f19dd64a83\") " pod="openstack/heat-cfnapi-9587dcd9d-cx57b" Oct 08 19:56:34 crc kubenswrapper[4988]: I1008 19:56:34.358810 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e6dd3357-9d84-4c92-8e40-69f19dd64a83-combined-ca-bundle\") pod \"heat-cfnapi-9587dcd9d-cx57b\" (UID: \"e6dd3357-9d84-4c92-8e40-69f19dd64a83\") " pod="openstack/heat-cfnapi-9587dcd9d-cx57b" Oct 08 19:56:34 crc kubenswrapper[4988]: I1008 19:56:34.358867 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e988e6d2-d2c3-46b6-a6e4-e8524f3fbe8a-public-tls-certs\") pod \"heat-api-594c995686-vt22w\" (UID: \"e988e6d2-d2c3-46b6-a6e4-e8524f3fbe8a\") " pod="openstack/heat-api-594c995686-vt22w" Oct 08 19:56:34 crc kubenswrapper[4988]: I1008 19:56:34.358893 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e6dd3357-9d84-4c92-8e40-69f19dd64a83-config-data-custom\") pod \"heat-cfnapi-9587dcd9d-cx57b\" (UID: \"e6dd3357-9d84-4c92-8e40-69f19dd64a83\") " pod="openstack/heat-cfnapi-9587dcd9d-cx57b" Oct 08 19:56:34 crc kubenswrapper[4988]: I1008 19:56:34.358911 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e6dd3357-9d84-4c92-8e40-69f19dd64a83-config-data\") pod \"heat-cfnapi-9587dcd9d-cx57b\" (UID: \"e6dd3357-9d84-4c92-8e40-69f19dd64a83\") " pod="openstack/heat-cfnapi-9587dcd9d-cx57b" Oct 08 19:56:34 crc kubenswrapper[4988]: I1008 19:56:34.358945 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dvp9p\" (UniqueName: \"kubernetes.io/projected/e988e6d2-d2c3-46b6-a6e4-e8524f3fbe8a-kube-api-access-dvp9p\") pod \"heat-api-594c995686-vt22w\" (UID: \"e988e6d2-d2c3-46b6-a6e4-e8524f3fbe8a\") " pod="openstack/heat-api-594c995686-vt22w" Oct 08 19:56:34 crc kubenswrapper[4988]: I1008 19:56:34.358979 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e988e6d2-d2c3-46b6-a6e4-e8524f3fbe8a-config-data-custom\") pod \"heat-api-594c995686-vt22w\" (UID: \"e988e6d2-d2c3-46b6-a6e4-e8524f3fbe8a\") " pod="openstack/heat-api-594c995686-vt22w" Oct 08 19:56:34 crc kubenswrapper[4988]: I1008 19:56:34.359001 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/e6dd3357-9d84-4c92-8e40-69f19dd64a83-internal-tls-certs\") pod \"heat-cfnapi-9587dcd9d-cx57b\" (UID: \"e6dd3357-9d84-4c92-8e40-69f19dd64a83\") " pod="openstack/heat-cfnapi-9587dcd9d-cx57b" Oct 08 19:56:34 crc kubenswrapper[4988]: I1008 19:56:34.359028 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7f6fs\" (UniqueName: \"kubernetes.io/projected/e6dd3357-9d84-4c92-8e40-69f19dd64a83-kube-api-access-7f6fs\") pod \"heat-cfnapi-9587dcd9d-cx57b\" (UID: \"e6dd3357-9d84-4c92-8e40-69f19dd64a83\") " pod="openstack/heat-cfnapi-9587dcd9d-cx57b" Oct 08 19:56:34 crc kubenswrapper[4988]: I1008 19:56:34.359048 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e988e6d2-d2c3-46b6-a6e4-e8524f3fbe8a-combined-ca-bundle\") pod \"heat-api-594c995686-vt22w\" (UID: \"e988e6d2-d2c3-46b6-a6e4-e8524f3fbe8a\") " pod="openstack/heat-api-594c995686-vt22w" Oct 08 19:56:34 crc kubenswrapper[4988]: I1008 19:56:34.365101 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e988e6d2-d2c3-46b6-a6e4-e8524f3fbe8a-config-data-custom\") pod \"heat-api-594c995686-vt22w\" (UID: \"e988e6d2-d2c3-46b6-a6e4-e8524f3fbe8a\") " pod="openstack/heat-api-594c995686-vt22w" Oct 08 19:56:34 crc kubenswrapper[4988]: I1008 19:56:34.365429 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e988e6d2-d2c3-46b6-a6e4-e8524f3fbe8a-public-tls-certs\") pod \"heat-api-594c995686-vt22w\" (UID: \"e988e6d2-d2c3-46b6-a6e4-e8524f3fbe8a\") " pod="openstack/heat-api-594c995686-vt22w" Oct 08 19:56:34 crc kubenswrapper[4988]: I1008 19:56:34.366443 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e6dd3357-9d84-4c92-8e40-69f19dd64a83-config-data-custom\") pod \"heat-cfnapi-9587dcd9d-cx57b\" (UID: \"e6dd3357-9d84-4c92-8e40-69f19dd64a83\") " pod="openstack/heat-cfnapi-9587dcd9d-cx57b" Oct 08 19:56:34 crc kubenswrapper[4988]: I1008 19:56:34.366810 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/e6dd3357-9d84-4c92-8e40-69f19dd64a83-internal-tls-certs\") pod \"heat-cfnapi-9587dcd9d-cx57b\" (UID: \"e6dd3357-9d84-4c92-8e40-69f19dd64a83\") " pod="openstack/heat-cfnapi-9587dcd9d-cx57b" Oct 08 19:56:34 crc kubenswrapper[4988]: I1008 19:56:34.368053 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e988e6d2-d2c3-46b6-a6e4-e8524f3fbe8a-config-data\") pod \"heat-api-594c995686-vt22w\" (UID: \"e988e6d2-d2c3-46b6-a6e4-e8524f3fbe8a\") " pod="openstack/heat-api-594c995686-vt22w" Oct 08 19:56:34 crc kubenswrapper[4988]: I1008 19:56:34.368468 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e988e6d2-d2c3-46b6-a6e4-e8524f3fbe8a-combined-ca-bundle\") pod \"heat-api-594c995686-vt22w\" (UID: \"e988e6d2-d2c3-46b6-a6e4-e8524f3fbe8a\") " pod="openstack/heat-api-594c995686-vt22w" Oct 08 19:56:34 crc kubenswrapper[4988]: I1008 19:56:34.368739 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e6dd3357-9d84-4c92-8e40-69f19dd64a83-combined-ca-bundle\") pod \"heat-cfnapi-9587dcd9d-cx57b\" (UID: \"e6dd3357-9d84-4c92-8e40-69f19dd64a83\") " pod="openstack/heat-cfnapi-9587dcd9d-cx57b" Oct 08 19:56:34 crc kubenswrapper[4988]: I1008 19:56:34.368857 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e6dd3357-9d84-4c92-8e40-69f19dd64a83-public-tls-certs\") pod \"heat-cfnapi-9587dcd9d-cx57b\" (UID: \"e6dd3357-9d84-4c92-8e40-69f19dd64a83\") " pod="openstack/heat-cfnapi-9587dcd9d-cx57b" Oct 08 19:56:34 crc kubenswrapper[4988]: I1008 19:56:34.369244 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/e988e6d2-d2c3-46b6-a6e4-e8524f3fbe8a-internal-tls-certs\") pod \"heat-api-594c995686-vt22w\" (UID: \"e988e6d2-d2c3-46b6-a6e4-e8524f3fbe8a\") " pod="openstack/heat-api-594c995686-vt22w" Oct 08 19:56:34 crc kubenswrapper[4988]: I1008 19:56:34.370776 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e6dd3357-9d84-4c92-8e40-69f19dd64a83-config-data\") pod \"heat-cfnapi-9587dcd9d-cx57b\" (UID: \"e6dd3357-9d84-4c92-8e40-69f19dd64a83\") " pod="openstack/heat-cfnapi-9587dcd9d-cx57b" Oct 08 19:56:34 crc kubenswrapper[4988]: I1008 19:56:34.387409 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7f6fs\" (UniqueName: \"kubernetes.io/projected/e6dd3357-9d84-4c92-8e40-69f19dd64a83-kube-api-access-7f6fs\") pod \"heat-cfnapi-9587dcd9d-cx57b\" (UID: \"e6dd3357-9d84-4c92-8e40-69f19dd64a83\") " pod="openstack/heat-cfnapi-9587dcd9d-cx57b" Oct 08 19:56:34 crc kubenswrapper[4988]: I1008 19:56:34.389358 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dvp9p\" (UniqueName: \"kubernetes.io/projected/e988e6d2-d2c3-46b6-a6e4-e8524f3fbe8a-kube-api-access-dvp9p\") pod \"heat-api-594c995686-vt22w\" (UID: \"e988e6d2-d2c3-46b6-a6e4-e8524f3fbe8a\") " pod="openstack/heat-api-594c995686-vt22w" Oct 08 19:56:34 crc kubenswrapper[4988]: I1008 19:56:34.522933 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-594c995686-vt22w" Oct 08 19:56:34 crc kubenswrapper[4988]: I1008 19:56:34.537532 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-9587dcd9d-cx57b" Oct 08 19:56:35 crc kubenswrapper[4988]: I1008 19:56:35.068155 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-api-594c995686-vt22w"] Oct 08 19:56:35 crc kubenswrapper[4988]: I1008 19:56:35.178990 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-cfnapi-9587dcd9d-cx57b"] Oct 08 19:56:35 crc kubenswrapper[4988]: I1008 19:56:35.188704 4988 generic.go:334] "Generic (PLEG): container finished" podID="ae3bd686-c190-4513-814a-0e6352ce3fdb" containerID="12a7d85683e0e2845002b1d08926e124e6a8781e988492861698b1a45012beb5" exitCode=1 Oct 08 19:56:35 crc kubenswrapper[4988]: I1008 19:56:35.188984 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-59756cd4b5-b6ghv" event={"ID":"ae3bd686-c190-4513-814a-0e6352ce3fdb","Type":"ContainerDied","Data":"12a7d85683e0e2845002b1d08926e124e6a8781e988492861698b1a45012beb5"} Oct 08 19:56:35 crc kubenswrapper[4988]: I1008 19:56:35.189033 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-59756cd4b5-b6ghv" event={"ID":"ae3bd686-c190-4513-814a-0e6352ce3fdb","Type":"ContainerStarted","Data":"ed4eed2a902d11aa9b2f6aaaf97742fd1ee0b30ef18b1ad0e0efb2bcca74c036"} Oct 08 19:56:35 crc kubenswrapper[4988]: I1008 19:56:35.189215 4988 scope.go:117] "RemoveContainer" containerID="12a7d85683e0e2845002b1d08926e124e6a8781e988492861698b1a45012beb5" Oct 08 19:56:35 crc kubenswrapper[4988]: I1008 19:56:35.196130 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-engine-78cc4595db-p87kc" event={"ID":"807b4691-54e6-4ae9-abeb-8e95ab06578a","Type":"ContainerStarted","Data":"b4031768e12919fef49bf38e16d6108605a3a32afab7e3eade954f8c301801ca"} Oct 08 19:56:35 crc kubenswrapper[4988]: I1008 19:56:35.197345 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-engine-78cc4595db-p87kc" Oct 08 19:56:35 crc kubenswrapper[4988]: I1008 19:56:35.200913 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-594c995686-vt22w" event={"ID":"e988e6d2-d2c3-46b6-a6e4-e8524f3fbe8a","Type":"ContainerStarted","Data":"9b516b7e45f2acaebcf26faa1cfb5a386543bbc3fbee735497c0d5556ff05d7a"} Oct 08 19:56:35 crc kubenswrapper[4988]: I1008 19:56:35.216549 4988 generic.go:334] "Generic (PLEG): container finished" podID="9fb22c64-8f41-496f-b894-d1c7d9b80125" containerID="f5f3c14a7960ecbc10974b6dda14c478684d2a5ac4fd0ea933ee98b7cbbb3a73" exitCode=1 Oct 08 19:56:35 crc kubenswrapper[4988]: I1008 19:56:35.216637 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-b969444b9-mhxbw" event={"ID":"9fb22c64-8f41-496f-b894-d1c7d9b80125","Type":"ContainerDied","Data":"f5f3c14a7960ecbc10974b6dda14c478684d2a5ac4fd0ea933ee98b7cbbb3a73"} Oct 08 19:56:35 crc kubenswrapper[4988]: I1008 19:56:35.218794 4988 scope.go:117] "RemoveContainer" containerID="f5f3c14a7960ecbc10974b6dda14c478684d2a5ac4fd0ea933ee98b7cbbb3a73" Oct 08 19:56:35 crc kubenswrapper[4988]: I1008 19:56:35.260872 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-engine-78cc4595db-p87kc" podStartSLOduration=2.260802148 podStartE2EDuration="2.260802148s" podCreationTimestamp="2025-10-08 19:56:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 19:56:35.230901985 +0000 UTC m=+6340.680744755" watchObservedRunningTime="2025-10-08 19:56:35.260802148 +0000 UTC m=+6340.710644918" Oct 08 19:56:35 crc kubenswrapper[4988]: I1008 19:56:35.815734 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/heat-cfnapi-56bdcfdf8b-dbkmh" Oct 08 19:56:36 crc kubenswrapper[4988]: I1008 19:56:36.228003 4988 generic.go:334] "Generic (PLEG): container finished" podID="9fb22c64-8f41-496f-b894-d1c7d9b80125" containerID="35d074e20a1fa24b5d08bbd8fd08ca66a673a3e894293a2e998ac660f4dc7f21" exitCode=1 Oct 08 19:56:36 crc kubenswrapper[4988]: I1008 19:56:36.228073 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-b969444b9-mhxbw" event={"ID":"9fb22c64-8f41-496f-b894-d1c7d9b80125","Type":"ContainerDied","Data":"35d074e20a1fa24b5d08bbd8fd08ca66a673a3e894293a2e998ac660f4dc7f21"} Oct 08 19:56:36 crc kubenswrapper[4988]: I1008 19:56:36.228110 4988 scope.go:117] "RemoveContainer" containerID="f5f3c14a7960ecbc10974b6dda14c478684d2a5ac4fd0ea933ee98b7cbbb3a73" Oct 08 19:56:36 crc kubenswrapper[4988]: I1008 19:56:36.228915 4988 scope.go:117] "RemoveContainer" containerID="35d074e20a1fa24b5d08bbd8fd08ca66a673a3e894293a2e998ac660f4dc7f21" Oct 08 19:56:36 crc kubenswrapper[4988]: E1008 19:56:36.229247 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-cfnapi\" with CrashLoopBackOff: \"back-off 10s restarting failed container=heat-cfnapi pod=heat-cfnapi-b969444b9-mhxbw_openstack(9fb22c64-8f41-496f-b894-d1c7d9b80125)\"" pod="openstack/heat-cfnapi-b969444b9-mhxbw" podUID="9fb22c64-8f41-496f-b894-d1c7d9b80125" Oct 08 19:56:36 crc kubenswrapper[4988]: I1008 19:56:36.232344 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-9587dcd9d-cx57b" event={"ID":"e6dd3357-9d84-4c92-8e40-69f19dd64a83","Type":"ContainerStarted","Data":"528c0eb578eae1ce92021b34592a0bd963c6bb6afc7e353a1000c01d020d7019"} Oct 08 19:56:36 crc kubenswrapper[4988]: I1008 19:56:36.232427 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-9587dcd9d-cx57b" event={"ID":"e6dd3357-9d84-4c92-8e40-69f19dd64a83","Type":"ContainerStarted","Data":"b815f9e00db1db0a6aa715446ec7580ff7bc48a45775e1610d461caa22593ef4"} Oct 08 19:56:36 crc kubenswrapper[4988]: I1008 19:56:36.232449 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-cfnapi-9587dcd9d-cx57b" Oct 08 19:56:36 crc kubenswrapper[4988]: I1008 19:56:36.235229 4988 generic.go:334] "Generic (PLEG): container finished" podID="ae3bd686-c190-4513-814a-0e6352ce3fdb" containerID="bbac93eebb6b76224a38744fafb8a9d8eeaebaa8e928e59c1848e75dab667290" exitCode=1 Oct 08 19:56:36 crc kubenswrapper[4988]: I1008 19:56:36.235275 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-59756cd4b5-b6ghv" event={"ID":"ae3bd686-c190-4513-814a-0e6352ce3fdb","Type":"ContainerDied","Data":"bbac93eebb6b76224a38744fafb8a9d8eeaebaa8e928e59c1848e75dab667290"} Oct 08 19:56:36 crc kubenswrapper[4988]: I1008 19:56:36.235655 4988 scope.go:117] "RemoveContainer" containerID="bbac93eebb6b76224a38744fafb8a9d8eeaebaa8e928e59c1848e75dab667290" Oct 08 19:56:36 crc kubenswrapper[4988]: E1008 19:56:36.235861 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-api\" with CrashLoopBackOff: \"back-off 10s restarting failed container=heat-api pod=heat-api-59756cd4b5-b6ghv_openstack(ae3bd686-c190-4513-814a-0e6352ce3fdb)\"" pod="openstack/heat-api-59756cd4b5-b6ghv" podUID="ae3bd686-c190-4513-814a-0e6352ce3fdb" Oct 08 19:56:36 crc kubenswrapper[4988]: I1008 19:56:36.251348 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-594c995686-vt22w" event={"ID":"e988e6d2-d2c3-46b6-a6e4-e8524f3fbe8a","Type":"ContainerStarted","Data":"46290c0f2ed7c7512b6d0081250e4fe56ad4ba385422aa70d035381fab83892a"} Oct 08 19:56:36 crc kubenswrapper[4988]: I1008 19:56:36.269597 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-cfnapi-9587dcd9d-cx57b" podStartSLOduration=2.26957488 podStartE2EDuration="2.26957488s" podCreationTimestamp="2025-10-08 19:56:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 19:56:36.26172545 +0000 UTC m=+6341.711568240" watchObservedRunningTime="2025-10-08 19:56:36.26957488 +0000 UTC m=+6341.719417660" Oct 08 19:56:36 crc kubenswrapper[4988]: I1008 19:56:36.296087 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-api-594c995686-vt22w" podStartSLOduration=2.296066464 podStartE2EDuration="2.296066464s" podCreationTimestamp="2025-10-08 19:56:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 19:56:36.290615931 +0000 UTC m=+6341.740458701" watchObservedRunningTime="2025-10-08 19:56:36.296066464 +0000 UTC m=+6341.745909234" Oct 08 19:56:36 crc kubenswrapper[4988]: I1008 19:56:36.323827 4988 scope.go:117] "RemoveContainer" containerID="12a7d85683e0e2845002b1d08926e124e6a8781e988492861698b1a45012beb5" Oct 08 19:56:37 crc kubenswrapper[4988]: I1008 19:56:37.260202 4988 scope.go:117] "RemoveContainer" containerID="35d074e20a1fa24b5d08bbd8fd08ca66a673a3e894293a2e998ac660f4dc7f21" Oct 08 19:56:37 crc kubenswrapper[4988]: E1008 19:56:37.260849 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-cfnapi\" with CrashLoopBackOff: \"back-off 10s restarting failed container=heat-cfnapi pod=heat-cfnapi-b969444b9-mhxbw_openstack(9fb22c64-8f41-496f-b894-d1c7d9b80125)\"" pod="openstack/heat-cfnapi-b969444b9-mhxbw" podUID="9fb22c64-8f41-496f-b894-d1c7d9b80125" Oct 08 19:56:37 crc kubenswrapper[4988]: I1008 19:56:37.261780 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-api-594c995686-vt22w" Oct 08 19:56:37 crc kubenswrapper[4988]: I1008 19:56:37.262274 4988 scope.go:117] "RemoveContainer" containerID="bbac93eebb6b76224a38744fafb8a9d8eeaebaa8e928e59c1848e75dab667290" Oct 08 19:56:37 crc kubenswrapper[4988]: E1008 19:56:37.262514 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-api\" with CrashLoopBackOff: \"back-off 10s restarting failed container=heat-api pod=heat-api-59756cd4b5-b6ghv_openstack(ae3bd686-c190-4513-814a-0e6352ce3fdb)\"" pod="openstack/heat-api-59756cd4b5-b6ghv" podUID="ae3bd686-c190-4513-814a-0e6352ce3fdb" Oct 08 19:56:38 crc kubenswrapper[4988]: I1008 19:56:38.451483 4988 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openstack/heat-cfnapi-b969444b9-mhxbw" Oct 08 19:56:38 crc kubenswrapper[4988]: I1008 19:56:38.451754 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-cfnapi-b969444b9-mhxbw" Oct 08 19:56:38 crc kubenswrapper[4988]: I1008 19:56:38.452323 4988 scope.go:117] "RemoveContainer" containerID="35d074e20a1fa24b5d08bbd8fd08ca66a673a3e894293a2e998ac660f4dc7f21" Oct 08 19:56:38 crc kubenswrapper[4988]: E1008 19:56:38.452653 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-cfnapi\" with CrashLoopBackOff: \"back-off 10s restarting failed container=heat-cfnapi pod=heat-cfnapi-b969444b9-mhxbw_openstack(9fb22c64-8f41-496f-b894-d1c7d9b80125)\"" pod="openstack/heat-cfnapi-b969444b9-mhxbw" podUID="9fb22c64-8f41-496f-b894-d1c7d9b80125" Oct 08 19:56:38 crc kubenswrapper[4988]: I1008 19:56:38.472707 4988 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openstack/heat-api-59756cd4b5-b6ghv" Oct 08 19:56:38 crc kubenswrapper[4988]: I1008 19:56:38.472795 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-api-59756cd4b5-b6ghv" Oct 08 19:56:38 crc kubenswrapper[4988]: I1008 19:56:38.473676 4988 scope.go:117] "RemoveContainer" containerID="bbac93eebb6b76224a38744fafb8a9d8eeaebaa8e928e59c1848e75dab667290" Oct 08 19:56:38 crc kubenswrapper[4988]: E1008 19:56:38.474044 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-api\" with CrashLoopBackOff: \"back-off 10s restarting failed container=heat-api pod=heat-api-59756cd4b5-b6ghv_openstack(ae3bd686-c190-4513-814a-0e6352ce3fdb)\"" pod="openstack/heat-api-59756cd4b5-b6ghv" podUID="ae3bd686-c190-4513-814a-0e6352ce3fdb" Oct 08 19:56:39 crc kubenswrapper[4988]: I1008 19:56:39.634571 4988 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/heat-cfnapi-56bdcfdf8b-dbkmh" podUID="6aeded19-1e22-46b5-8e88-ce9ba127269f" containerName="heat-cfnapi" probeResult="failure" output="Get \"http://10.217.1.131:8000/healthcheck\": read tcp 10.217.0.2:53010->10.217.1.131:8000: read: connection reset by peer" Oct 08 19:56:39 crc kubenswrapper[4988]: I1008 19:56:39.651347 4988 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/heat-api-8664dcb6b9-jf2pf" podUID="255ab071-02d7-46b6-af97-cd84f085b097" containerName="heat-api" probeResult="failure" output="Get \"http://10.217.1.130:8004/healthcheck\": read tcp 10.217.0.2:54868->10.217.1.130:8004: read: connection reset by peer" Oct 08 19:56:39 crc kubenswrapper[4988]: I1008 19:56:39.845850 4988 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-7d85fc8646-gv4lm" podUID="c533a2e1-7213-4410-b862-5eed5e1fe8e9" containerName="horizon" probeResult="failure" output="Get \"https://10.217.1.122:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.1.122:8443: connect: connection refused" Oct 08 19:56:39 crc kubenswrapper[4988]: I1008 19:56:39.846294 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-7d85fc8646-gv4lm" Oct 08 19:56:40 crc kubenswrapper[4988]: I1008 19:56:40.295315 4988 generic.go:334] "Generic (PLEG): container finished" podID="255ab071-02d7-46b6-af97-cd84f085b097" containerID="ea11f6f5c59b0839c5743a2f38ec67a1da94e3ef717c1e36f809ce296a109d4e" exitCode=0 Oct 08 19:56:40 crc kubenswrapper[4988]: I1008 19:56:40.295438 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-8664dcb6b9-jf2pf" event={"ID":"255ab071-02d7-46b6-af97-cd84f085b097","Type":"ContainerDied","Data":"ea11f6f5c59b0839c5743a2f38ec67a1da94e3ef717c1e36f809ce296a109d4e"} Oct 08 19:56:40 crc kubenswrapper[4988]: I1008 19:56:40.295624 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-8664dcb6b9-jf2pf" event={"ID":"255ab071-02d7-46b6-af97-cd84f085b097","Type":"ContainerDied","Data":"66a6de057679cd07810c71c9a7402c7a0ecc6b93c98a269421e26c301a24115f"} Oct 08 19:56:40 crc kubenswrapper[4988]: I1008 19:56:40.295640 4988 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="66a6de057679cd07810c71c9a7402c7a0ecc6b93c98a269421e26c301a24115f" Oct 08 19:56:40 crc kubenswrapper[4988]: I1008 19:56:40.297543 4988 generic.go:334] "Generic (PLEG): container finished" podID="6aeded19-1e22-46b5-8e88-ce9ba127269f" containerID="33c6f5043911fe4478477ea2b8457d1abcff028ec754adec057bfb6a524ecb3b" exitCode=0 Oct 08 19:56:40 crc kubenswrapper[4988]: I1008 19:56:40.297570 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-56bdcfdf8b-dbkmh" event={"ID":"6aeded19-1e22-46b5-8e88-ce9ba127269f","Type":"ContainerDied","Data":"33c6f5043911fe4478477ea2b8457d1abcff028ec754adec057bfb6a524ecb3b"} Oct 08 19:56:40 crc kubenswrapper[4988]: I1008 19:56:40.297585 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-56bdcfdf8b-dbkmh" event={"ID":"6aeded19-1e22-46b5-8e88-ce9ba127269f","Type":"ContainerDied","Data":"ad0d72c598d9925765913496cce57e27b9a598d1b282086861baae2d03c8d989"} Oct 08 19:56:40 crc kubenswrapper[4988]: I1008 19:56:40.297593 4988 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ad0d72c598d9925765913496cce57e27b9a598d1b282086861baae2d03c8d989" Oct 08 19:56:40 crc kubenswrapper[4988]: I1008 19:56:40.341320 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-8664dcb6b9-jf2pf" Oct 08 19:56:40 crc kubenswrapper[4988]: I1008 19:56:40.345821 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-56bdcfdf8b-dbkmh" Oct 08 19:56:40 crc kubenswrapper[4988]: I1008 19:56:40.409697 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8g292\" (UniqueName: \"kubernetes.io/projected/255ab071-02d7-46b6-af97-cd84f085b097-kube-api-access-8g292\") pod \"255ab071-02d7-46b6-af97-cd84f085b097\" (UID: \"255ab071-02d7-46b6-af97-cd84f085b097\") " Oct 08 19:56:40 crc kubenswrapper[4988]: I1008 19:56:40.409730 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6aeded19-1e22-46b5-8e88-ce9ba127269f-config-data\") pod \"6aeded19-1e22-46b5-8e88-ce9ba127269f\" (UID: \"6aeded19-1e22-46b5-8e88-ce9ba127269f\") " Oct 08 19:56:40 crc kubenswrapper[4988]: I1008 19:56:40.409771 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/6aeded19-1e22-46b5-8e88-ce9ba127269f-config-data-custom\") pod \"6aeded19-1e22-46b5-8e88-ce9ba127269f\" (UID: \"6aeded19-1e22-46b5-8e88-ce9ba127269f\") " Oct 08 19:56:40 crc kubenswrapper[4988]: I1008 19:56:40.409821 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/255ab071-02d7-46b6-af97-cd84f085b097-combined-ca-bundle\") pod \"255ab071-02d7-46b6-af97-cd84f085b097\" (UID: \"255ab071-02d7-46b6-af97-cd84f085b097\") " Oct 08 19:56:40 crc kubenswrapper[4988]: I1008 19:56:40.409872 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/255ab071-02d7-46b6-af97-cd84f085b097-config-data-custom\") pod \"255ab071-02d7-46b6-af97-cd84f085b097\" (UID: \"255ab071-02d7-46b6-af97-cd84f085b097\") " Oct 08 19:56:40 crc kubenswrapper[4988]: I1008 19:56:40.409896 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/255ab071-02d7-46b6-af97-cd84f085b097-config-data\") pod \"255ab071-02d7-46b6-af97-cd84f085b097\" (UID: \"255ab071-02d7-46b6-af97-cd84f085b097\") " Oct 08 19:56:40 crc kubenswrapper[4988]: I1008 19:56:40.409923 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zfdhk\" (UniqueName: \"kubernetes.io/projected/6aeded19-1e22-46b5-8e88-ce9ba127269f-kube-api-access-zfdhk\") pod \"6aeded19-1e22-46b5-8e88-ce9ba127269f\" (UID: \"6aeded19-1e22-46b5-8e88-ce9ba127269f\") " Oct 08 19:56:40 crc kubenswrapper[4988]: I1008 19:56:40.409970 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6aeded19-1e22-46b5-8e88-ce9ba127269f-combined-ca-bundle\") pod \"6aeded19-1e22-46b5-8e88-ce9ba127269f\" (UID: \"6aeded19-1e22-46b5-8e88-ce9ba127269f\") " Oct 08 19:56:40 crc kubenswrapper[4988]: I1008 19:56:40.416085 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6aeded19-1e22-46b5-8e88-ce9ba127269f-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "6aeded19-1e22-46b5-8e88-ce9ba127269f" (UID: "6aeded19-1e22-46b5-8e88-ce9ba127269f"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 19:56:40 crc kubenswrapper[4988]: I1008 19:56:40.417275 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/255ab071-02d7-46b6-af97-cd84f085b097-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "255ab071-02d7-46b6-af97-cd84f085b097" (UID: "255ab071-02d7-46b6-af97-cd84f085b097"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 19:56:40 crc kubenswrapper[4988]: I1008 19:56:40.418603 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/255ab071-02d7-46b6-af97-cd84f085b097-kube-api-access-8g292" (OuterVolumeSpecName: "kube-api-access-8g292") pod "255ab071-02d7-46b6-af97-cd84f085b097" (UID: "255ab071-02d7-46b6-af97-cd84f085b097"). InnerVolumeSpecName "kube-api-access-8g292". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 19:56:40 crc kubenswrapper[4988]: I1008 19:56:40.421132 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6aeded19-1e22-46b5-8e88-ce9ba127269f-kube-api-access-zfdhk" (OuterVolumeSpecName: "kube-api-access-zfdhk") pod "6aeded19-1e22-46b5-8e88-ce9ba127269f" (UID: "6aeded19-1e22-46b5-8e88-ce9ba127269f"). InnerVolumeSpecName "kube-api-access-zfdhk". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 19:56:40 crc kubenswrapper[4988]: I1008 19:56:40.437874 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6aeded19-1e22-46b5-8e88-ce9ba127269f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "6aeded19-1e22-46b5-8e88-ce9ba127269f" (UID: "6aeded19-1e22-46b5-8e88-ce9ba127269f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 19:56:40 crc kubenswrapper[4988]: I1008 19:56:40.442978 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/255ab071-02d7-46b6-af97-cd84f085b097-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "255ab071-02d7-46b6-af97-cd84f085b097" (UID: "255ab071-02d7-46b6-af97-cd84f085b097"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 19:56:40 crc kubenswrapper[4988]: I1008 19:56:40.478535 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/255ab071-02d7-46b6-af97-cd84f085b097-config-data" (OuterVolumeSpecName: "config-data") pod "255ab071-02d7-46b6-af97-cd84f085b097" (UID: "255ab071-02d7-46b6-af97-cd84f085b097"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 19:56:40 crc kubenswrapper[4988]: I1008 19:56:40.495199 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6aeded19-1e22-46b5-8e88-ce9ba127269f-config-data" (OuterVolumeSpecName: "config-data") pod "6aeded19-1e22-46b5-8e88-ce9ba127269f" (UID: "6aeded19-1e22-46b5-8e88-ce9ba127269f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 19:56:40 crc kubenswrapper[4988]: I1008 19:56:40.511825 4988 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/6aeded19-1e22-46b5-8e88-ce9ba127269f-config-data-custom\") on node \"crc\" DevicePath \"\"" Oct 08 19:56:40 crc kubenswrapper[4988]: I1008 19:56:40.511859 4988 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/255ab071-02d7-46b6-af97-cd84f085b097-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 19:56:40 crc kubenswrapper[4988]: I1008 19:56:40.511872 4988 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/255ab071-02d7-46b6-af97-cd84f085b097-config-data-custom\") on node \"crc\" DevicePath \"\"" Oct 08 19:56:40 crc kubenswrapper[4988]: I1008 19:56:40.511886 4988 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/255ab071-02d7-46b6-af97-cd84f085b097-config-data\") on node \"crc\" DevicePath \"\"" Oct 08 19:56:40 crc kubenswrapper[4988]: I1008 19:56:40.511900 4988 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zfdhk\" (UniqueName: \"kubernetes.io/projected/6aeded19-1e22-46b5-8e88-ce9ba127269f-kube-api-access-zfdhk\") on node \"crc\" DevicePath \"\"" Oct 08 19:56:40 crc kubenswrapper[4988]: I1008 19:56:40.511913 4988 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6aeded19-1e22-46b5-8e88-ce9ba127269f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 19:56:40 crc kubenswrapper[4988]: I1008 19:56:40.511926 4988 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8g292\" (UniqueName: \"kubernetes.io/projected/255ab071-02d7-46b6-af97-cd84f085b097-kube-api-access-8g292\") on node \"crc\" DevicePath \"\"" Oct 08 19:56:40 crc kubenswrapper[4988]: I1008 19:56:40.511937 4988 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6aeded19-1e22-46b5-8e88-ce9ba127269f-config-data\") on node \"crc\" DevicePath \"\"" Oct 08 19:56:41 crc kubenswrapper[4988]: I1008 19:56:41.315293 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-8664dcb6b9-jf2pf" Oct 08 19:56:41 crc kubenswrapper[4988]: I1008 19:56:41.315485 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-56bdcfdf8b-dbkmh" Oct 08 19:56:41 crc kubenswrapper[4988]: I1008 19:56:41.365219 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-api-8664dcb6b9-jf2pf"] Oct 08 19:56:41 crc kubenswrapper[4988]: I1008 19:56:41.375887 4988 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-api-8664dcb6b9-jf2pf"] Oct 08 19:56:41 crc kubenswrapper[4988]: I1008 19:56:41.386247 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-cfnapi-56bdcfdf8b-dbkmh"] Oct 08 19:56:41 crc kubenswrapper[4988]: I1008 19:56:41.395702 4988 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-cfnapi-56bdcfdf8b-dbkmh"] Oct 08 19:56:43 crc kubenswrapper[4988]: I1008 19:56:43.252225 4988 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="255ab071-02d7-46b6-af97-cd84f085b097" path="/var/lib/kubelet/pods/255ab071-02d7-46b6-af97-cd84f085b097/volumes" Oct 08 19:56:43 crc kubenswrapper[4988]: I1008 19:56:43.255015 4988 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6aeded19-1e22-46b5-8e88-ce9ba127269f" path="/var/lib/kubelet/pods/6aeded19-1e22-46b5-8e88-ce9ba127269f/volumes" Oct 08 19:56:45 crc kubenswrapper[4988]: I1008 19:56:45.875427 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/heat-cfnapi-9587dcd9d-cx57b" Oct 08 19:56:45 crc kubenswrapper[4988]: I1008 19:56:45.921804 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/heat-api-594c995686-vt22w" Oct 08 19:56:45 crc kubenswrapper[4988]: I1008 19:56:45.953616 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-cfnapi-b969444b9-mhxbw"] Oct 08 19:56:45 crc kubenswrapper[4988]: I1008 19:56:45.998458 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-api-59756cd4b5-b6ghv"] Oct 08 19:56:46 crc kubenswrapper[4988]: I1008 19:56:46.391968 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-b969444b9-mhxbw" event={"ID":"9fb22c64-8f41-496f-b894-d1c7d9b80125","Type":"ContainerDied","Data":"2281e4fe7eb5a009b1302cb3881ca439ab4c236333f51bc20071f187cda6884a"} Oct 08 19:56:46 crc kubenswrapper[4988]: I1008 19:56:46.392336 4988 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2281e4fe7eb5a009b1302cb3881ca439ab4c236333f51bc20071f187cda6884a" Oct 08 19:56:46 crc kubenswrapper[4988]: I1008 19:56:46.432806 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/heat-engine-d897f4749-rlhsn" Oct 08 19:56:46 crc kubenswrapper[4988]: I1008 19:56:46.526853 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-b969444b9-mhxbw" Oct 08 19:56:46 crc kubenswrapper[4988]: I1008 19:56:46.532208 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-59756cd4b5-b6ghv" Oct 08 19:56:46 crc kubenswrapper[4988]: I1008 19:56:46.671123 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9fb22c64-8f41-496f-b894-d1c7d9b80125-combined-ca-bundle\") pod \"9fb22c64-8f41-496f-b894-d1c7d9b80125\" (UID: \"9fb22c64-8f41-496f-b894-d1c7d9b80125\") " Oct 08 19:56:46 crc kubenswrapper[4988]: I1008 19:56:46.671245 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/ae3bd686-c190-4513-814a-0e6352ce3fdb-config-data-custom\") pod \"ae3bd686-c190-4513-814a-0e6352ce3fdb\" (UID: \"ae3bd686-c190-4513-814a-0e6352ce3fdb\") " Oct 08 19:56:46 crc kubenswrapper[4988]: I1008 19:56:46.671302 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9fb22c64-8f41-496f-b894-d1c7d9b80125-config-data\") pod \"9fb22c64-8f41-496f-b894-d1c7d9b80125\" (UID: \"9fb22c64-8f41-496f-b894-d1c7d9b80125\") " Oct 08 19:56:46 crc kubenswrapper[4988]: I1008 19:56:46.671434 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z27k8\" (UniqueName: \"kubernetes.io/projected/9fb22c64-8f41-496f-b894-d1c7d9b80125-kube-api-access-z27k8\") pod \"9fb22c64-8f41-496f-b894-d1c7d9b80125\" (UID: \"9fb22c64-8f41-496f-b894-d1c7d9b80125\") " Oct 08 19:56:46 crc kubenswrapper[4988]: I1008 19:56:46.671458 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b722t\" (UniqueName: \"kubernetes.io/projected/ae3bd686-c190-4513-814a-0e6352ce3fdb-kube-api-access-b722t\") pod \"ae3bd686-c190-4513-814a-0e6352ce3fdb\" (UID: \"ae3bd686-c190-4513-814a-0e6352ce3fdb\") " Oct 08 19:56:46 crc kubenswrapper[4988]: I1008 19:56:46.671494 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ae3bd686-c190-4513-814a-0e6352ce3fdb-combined-ca-bundle\") pod \"ae3bd686-c190-4513-814a-0e6352ce3fdb\" (UID: \"ae3bd686-c190-4513-814a-0e6352ce3fdb\") " Oct 08 19:56:46 crc kubenswrapper[4988]: I1008 19:56:46.671523 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/9fb22c64-8f41-496f-b894-d1c7d9b80125-config-data-custom\") pod \"9fb22c64-8f41-496f-b894-d1c7d9b80125\" (UID: \"9fb22c64-8f41-496f-b894-d1c7d9b80125\") " Oct 08 19:56:46 crc kubenswrapper[4988]: I1008 19:56:46.671559 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ae3bd686-c190-4513-814a-0e6352ce3fdb-config-data\") pod \"ae3bd686-c190-4513-814a-0e6352ce3fdb\" (UID: \"ae3bd686-c190-4513-814a-0e6352ce3fdb\") " Oct 08 19:56:46 crc kubenswrapper[4988]: I1008 19:56:46.685651 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ae3bd686-c190-4513-814a-0e6352ce3fdb-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "ae3bd686-c190-4513-814a-0e6352ce3fdb" (UID: "ae3bd686-c190-4513-814a-0e6352ce3fdb"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 19:56:46 crc kubenswrapper[4988]: I1008 19:56:46.693681 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ae3bd686-c190-4513-814a-0e6352ce3fdb-kube-api-access-b722t" (OuterVolumeSpecName: "kube-api-access-b722t") pod "ae3bd686-c190-4513-814a-0e6352ce3fdb" (UID: "ae3bd686-c190-4513-814a-0e6352ce3fdb"). InnerVolumeSpecName "kube-api-access-b722t". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 19:56:46 crc kubenswrapper[4988]: I1008 19:56:46.698565 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9fb22c64-8f41-496f-b894-d1c7d9b80125-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "9fb22c64-8f41-496f-b894-d1c7d9b80125" (UID: "9fb22c64-8f41-496f-b894-d1c7d9b80125"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 19:56:46 crc kubenswrapper[4988]: I1008 19:56:46.715582 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9fb22c64-8f41-496f-b894-d1c7d9b80125-kube-api-access-z27k8" (OuterVolumeSpecName: "kube-api-access-z27k8") pod "9fb22c64-8f41-496f-b894-d1c7d9b80125" (UID: "9fb22c64-8f41-496f-b894-d1c7d9b80125"). InnerVolumeSpecName "kube-api-access-z27k8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 19:56:46 crc kubenswrapper[4988]: I1008 19:56:46.765604 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ae3bd686-c190-4513-814a-0e6352ce3fdb-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ae3bd686-c190-4513-814a-0e6352ce3fdb" (UID: "ae3bd686-c190-4513-814a-0e6352ce3fdb"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 19:56:46 crc kubenswrapper[4988]: I1008 19:56:46.774715 4988 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z27k8\" (UniqueName: \"kubernetes.io/projected/9fb22c64-8f41-496f-b894-d1c7d9b80125-kube-api-access-z27k8\") on node \"crc\" DevicePath \"\"" Oct 08 19:56:46 crc kubenswrapper[4988]: I1008 19:56:46.774743 4988 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b722t\" (UniqueName: \"kubernetes.io/projected/ae3bd686-c190-4513-814a-0e6352ce3fdb-kube-api-access-b722t\") on node \"crc\" DevicePath \"\"" Oct 08 19:56:46 crc kubenswrapper[4988]: I1008 19:56:46.774752 4988 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ae3bd686-c190-4513-814a-0e6352ce3fdb-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 19:56:46 crc kubenswrapper[4988]: I1008 19:56:46.774764 4988 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/9fb22c64-8f41-496f-b894-d1c7d9b80125-config-data-custom\") on node \"crc\" DevicePath \"\"" Oct 08 19:56:46 crc kubenswrapper[4988]: I1008 19:56:46.774779 4988 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/ae3bd686-c190-4513-814a-0e6352ce3fdb-config-data-custom\") on node \"crc\" DevicePath \"\"" Oct 08 19:56:46 crc kubenswrapper[4988]: I1008 19:56:46.805567 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9fb22c64-8f41-496f-b894-d1c7d9b80125-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "9fb22c64-8f41-496f-b894-d1c7d9b80125" (UID: "9fb22c64-8f41-496f-b894-d1c7d9b80125"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 19:56:46 crc kubenswrapper[4988]: I1008 19:56:46.812763 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9fb22c64-8f41-496f-b894-d1c7d9b80125-config-data" (OuterVolumeSpecName: "config-data") pod "9fb22c64-8f41-496f-b894-d1c7d9b80125" (UID: "9fb22c64-8f41-496f-b894-d1c7d9b80125"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 19:56:46 crc kubenswrapper[4988]: I1008 19:56:46.815503 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ae3bd686-c190-4513-814a-0e6352ce3fdb-config-data" (OuterVolumeSpecName: "config-data") pod "ae3bd686-c190-4513-814a-0e6352ce3fdb" (UID: "ae3bd686-c190-4513-814a-0e6352ce3fdb"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 19:56:46 crc kubenswrapper[4988]: I1008 19:56:46.876964 4988 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ae3bd686-c190-4513-814a-0e6352ce3fdb-config-data\") on node \"crc\" DevicePath \"\"" Oct 08 19:56:46 crc kubenswrapper[4988]: I1008 19:56:46.876995 4988 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9fb22c64-8f41-496f-b894-d1c7d9b80125-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 19:56:46 crc kubenswrapper[4988]: I1008 19:56:46.877008 4988 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9fb22c64-8f41-496f-b894-d1c7d9b80125-config-data\") on node \"crc\" DevicePath \"\"" Oct 08 19:56:47 crc kubenswrapper[4988]: I1008 19:56:47.406557 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-b969444b9-mhxbw" Oct 08 19:56:47 crc kubenswrapper[4988]: I1008 19:56:47.406555 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-59756cd4b5-b6ghv" event={"ID":"ae3bd686-c190-4513-814a-0e6352ce3fdb","Type":"ContainerDied","Data":"ed4eed2a902d11aa9b2f6aaaf97742fd1ee0b30ef18b1ad0e0efb2bcca74c036"} Oct 08 19:56:47 crc kubenswrapper[4988]: I1008 19:56:47.406586 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-59756cd4b5-b6ghv" Oct 08 19:56:47 crc kubenswrapper[4988]: I1008 19:56:47.406997 4988 scope.go:117] "RemoveContainer" containerID="bbac93eebb6b76224a38744fafb8a9d8eeaebaa8e928e59c1848e75dab667290" Oct 08 19:56:47 crc kubenswrapper[4988]: I1008 19:56:47.456921 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-api-59756cd4b5-b6ghv"] Oct 08 19:56:47 crc kubenswrapper[4988]: I1008 19:56:47.466960 4988 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-api-59756cd4b5-b6ghv"] Oct 08 19:56:47 crc kubenswrapper[4988]: I1008 19:56:47.491433 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-cfnapi-b969444b9-mhxbw"] Oct 08 19:56:47 crc kubenswrapper[4988]: I1008 19:56:47.503397 4988 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-cfnapi-b969444b9-mhxbw"] Oct 08 19:56:49 crc kubenswrapper[4988]: I1008 19:56:49.262196 4988 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9fb22c64-8f41-496f-b894-d1c7d9b80125" path="/var/lib/kubelet/pods/9fb22c64-8f41-496f-b894-d1c7d9b80125/volumes" Oct 08 19:56:49 crc kubenswrapper[4988]: I1008 19:56:49.263496 4988 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ae3bd686-c190-4513-814a-0e6352ce3fdb" path="/var/lib/kubelet/pods/ae3bd686-c190-4513-814a-0e6352ce3fdb/volumes" Oct 08 19:56:49 crc kubenswrapper[4988]: I1008 19:56:49.442248 4988 generic.go:334] "Generic (PLEG): container finished" podID="c533a2e1-7213-4410-b862-5eed5e1fe8e9" containerID="bfd13f227cb2064f23761c7a562d35816e2cb13aba4850cb48a1f6228a4798cc" exitCode=137 Oct 08 19:56:49 crc kubenswrapper[4988]: I1008 19:56:49.442538 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7d85fc8646-gv4lm" event={"ID":"c533a2e1-7213-4410-b862-5eed5e1fe8e9","Type":"ContainerDied","Data":"bfd13f227cb2064f23761c7a562d35816e2cb13aba4850cb48a1f6228a4798cc"} Oct 08 19:56:49 crc kubenswrapper[4988]: I1008 19:56:49.672087 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-7d85fc8646-gv4lm" Oct 08 19:56:49 crc kubenswrapper[4988]: I1008 19:56:49.840573 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c533a2e1-7213-4410-b862-5eed5e1fe8e9-combined-ca-bundle\") pod \"c533a2e1-7213-4410-b862-5eed5e1fe8e9\" (UID: \"c533a2e1-7213-4410-b862-5eed5e1fe8e9\") " Oct 08 19:56:49 crc kubenswrapper[4988]: I1008 19:56:49.840639 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8hpsk\" (UniqueName: \"kubernetes.io/projected/c533a2e1-7213-4410-b862-5eed5e1fe8e9-kube-api-access-8hpsk\") pod \"c533a2e1-7213-4410-b862-5eed5e1fe8e9\" (UID: \"c533a2e1-7213-4410-b862-5eed5e1fe8e9\") " Oct 08 19:56:49 crc kubenswrapper[4988]: I1008 19:56:49.840680 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c533a2e1-7213-4410-b862-5eed5e1fe8e9-logs\") pod \"c533a2e1-7213-4410-b862-5eed5e1fe8e9\" (UID: \"c533a2e1-7213-4410-b862-5eed5e1fe8e9\") " Oct 08 19:56:49 crc kubenswrapper[4988]: I1008 19:56:49.840751 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/c533a2e1-7213-4410-b862-5eed5e1fe8e9-horizon-secret-key\") pod \"c533a2e1-7213-4410-b862-5eed5e1fe8e9\" (UID: \"c533a2e1-7213-4410-b862-5eed5e1fe8e9\") " Oct 08 19:56:49 crc kubenswrapper[4988]: I1008 19:56:49.840796 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c533a2e1-7213-4410-b862-5eed5e1fe8e9-scripts\") pod \"c533a2e1-7213-4410-b862-5eed5e1fe8e9\" (UID: \"c533a2e1-7213-4410-b862-5eed5e1fe8e9\") " Oct 08 19:56:49 crc kubenswrapper[4988]: I1008 19:56:49.840913 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/c533a2e1-7213-4410-b862-5eed5e1fe8e9-config-data\") pod \"c533a2e1-7213-4410-b862-5eed5e1fe8e9\" (UID: \"c533a2e1-7213-4410-b862-5eed5e1fe8e9\") " Oct 08 19:56:49 crc kubenswrapper[4988]: I1008 19:56:49.840959 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/c533a2e1-7213-4410-b862-5eed5e1fe8e9-horizon-tls-certs\") pod \"c533a2e1-7213-4410-b862-5eed5e1fe8e9\" (UID: \"c533a2e1-7213-4410-b862-5eed5e1fe8e9\") " Oct 08 19:56:49 crc kubenswrapper[4988]: I1008 19:56:49.841996 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c533a2e1-7213-4410-b862-5eed5e1fe8e9-logs" (OuterVolumeSpecName: "logs") pod "c533a2e1-7213-4410-b862-5eed5e1fe8e9" (UID: "c533a2e1-7213-4410-b862-5eed5e1fe8e9"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 19:56:49 crc kubenswrapper[4988]: I1008 19:56:49.846757 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c533a2e1-7213-4410-b862-5eed5e1fe8e9-kube-api-access-8hpsk" (OuterVolumeSpecName: "kube-api-access-8hpsk") pod "c533a2e1-7213-4410-b862-5eed5e1fe8e9" (UID: "c533a2e1-7213-4410-b862-5eed5e1fe8e9"). InnerVolumeSpecName "kube-api-access-8hpsk". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 19:56:49 crc kubenswrapper[4988]: I1008 19:56:49.858775 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c533a2e1-7213-4410-b862-5eed5e1fe8e9-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "c533a2e1-7213-4410-b862-5eed5e1fe8e9" (UID: "c533a2e1-7213-4410-b862-5eed5e1fe8e9"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 19:56:49 crc kubenswrapper[4988]: I1008 19:56:49.874480 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c533a2e1-7213-4410-b862-5eed5e1fe8e9-scripts" (OuterVolumeSpecName: "scripts") pod "c533a2e1-7213-4410-b862-5eed5e1fe8e9" (UID: "c533a2e1-7213-4410-b862-5eed5e1fe8e9"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 19:56:49 crc kubenswrapper[4988]: I1008 19:56:49.876374 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c533a2e1-7213-4410-b862-5eed5e1fe8e9-config-data" (OuterVolumeSpecName: "config-data") pod "c533a2e1-7213-4410-b862-5eed5e1fe8e9" (UID: "c533a2e1-7213-4410-b862-5eed5e1fe8e9"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 19:56:49 crc kubenswrapper[4988]: I1008 19:56:49.892587 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c533a2e1-7213-4410-b862-5eed5e1fe8e9-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "c533a2e1-7213-4410-b862-5eed5e1fe8e9" (UID: "c533a2e1-7213-4410-b862-5eed5e1fe8e9"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 19:56:49 crc kubenswrapper[4988]: I1008 19:56:49.913231 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c533a2e1-7213-4410-b862-5eed5e1fe8e9-horizon-tls-certs" (OuterVolumeSpecName: "horizon-tls-certs") pod "c533a2e1-7213-4410-b862-5eed5e1fe8e9" (UID: "c533a2e1-7213-4410-b862-5eed5e1fe8e9"). InnerVolumeSpecName "horizon-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 19:56:49 crc kubenswrapper[4988]: I1008 19:56:49.943331 4988 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/c533a2e1-7213-4410-b862-5eed5e1fe8e9-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Oct 08 19:56:49 crc kubenswrapper[4988]: I1008 19:56:49.943704 4988 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c533a2e1-7213-4410-b862-5eed5e1fe8e9-scripts\") on node \"crc\" DevicePath \"\"" Oct 08 19:56:49 crc kubenswrapper[4988]: I1008 19:56:49.943716 4988 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/c533a2e1-7213-4410-b862-5eed5e1fe8e9-config-data\") on node \"crc\" DevicePath \"\"" Oct 08 19:56:49 crc kubenswrapper[4988]: I1008 19:56:49.943724 4988 reconciler_common.go:293] "Volume detached for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/c533a2e1-7213-4410-b862-5eed5e1fe8e9-horizon-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 08 19:56:49 crc kubenswrapper[4988]: I1008 19:56:49.943733 4988 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c533a2e1-7213-4410-b862-5eed5e1fe8e9-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 19:56:49 crc kubenswrapper[4988]: I1008 19:56:49.943742 4988 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8hpsk\" (UniqueName: \"kubernetes.io/projected/c533a2e1-7213-4410-b862-5eed5e1fe8e9-kube-api-access-8hpsk\") on node \"crc\" DevicePath \"\"" Oct 08 19:56:49 crc kubenswrapper[4988]: I1008 19:56:49.943754 4988 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c533a2e1-7213-4410-b862-5eed5e1fe8e9-logs\") on node \"crc\" DevicePath \"\"" Oct 08 19:56:50 crc kubenswrapper[4988]: I1008 19:56:50.457845 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7d85fc8646-gv4lm" event={"ID":"c533a2e1-7213-4410-b862-5eed5e1fe8e9","Type":"ContainerDied","Data":"60876b72a1bcd0717dd169ac584f8f5bd73e37c53808d766bf8e74b7823bf6cc"} Oct 08 19:56:50 crc kubenswrapper[4988]: I1008 19:56:50.457951 4988 scope.go:117] "RemoveContainer" containerID="24a201cb0f08c481ca0dc4c546b32eeb4e83b514268ccaffc0f84cb043a4ef21" Oct 08 19:56:50 crc kubenswrapper[4988]: I1008 19:56:50.458233 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-7d85fc8646-gv4lm" Oct 08 19:56:50 crc kubenswrapper[4988]: I1008 19:56:50.529434 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-7d85fc8646-gv4lm"] Oct 08 19:56:50 crc kubenswrapper[4988]: I1008 19:56:50.547167 4988 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-7d85fc8646-gv4lm"] Oct 08 19:56:50 crc kubenswrapper[4988]: I1008 19:56:50.656073 4988 scope.go:117] "RemoveContainer" containerID="bfd13f227cb2064f23761c7a562d35816e2cb13aba4850cb48a1f6228a4798cc" Oct 08 19:56:51 crc kubenswrapper[4988]: I1008 19:56:51.258651 4988 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c533a2e1-7213-4410-b862-5eed5e1fe8e9" path="/var/lib/kubelet/pods/c533a2e1-7213-4410-b862-5eed5e1fe8e9/volumes" Oct 08 19:56:53 crc kubenswrapper[4988]: I1008 19:56:53.338427 4988 patch_prober.go:28] interesting pod/machine-config-daemon-nm54f container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 19:56:53 crc kubenswrapper[4988]: I1008 19:56:53.338870 4988 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 19:56:53 crc kubenswrapper[4988]: I1008 19:56:53.472146 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/heat-engine-78cc4595db-p87kc" Oct 08 19:56:53 crc kubenswrapper[4988]: I1008 19:56:53.526881 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-engine-d897f4749-rlhsn"] Oct 08 19:56:53 crc kubenswrapper[4988]: I1008 19:56:53.527394 4988 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/heat-engine-d897f4749-rlhsn" podUID="fdbbb87a-7afb-402c-83fd-d4582e06c6ba" containerName="heat-engine" containerID="cri-o://c29143467f18a0d9cb5e4136506cbf1b8520fd0496108cf77db65c962e6ea146" gracePeriod=60 Oct 08 19:56:56 crc kubenswrapper[4988]: E1008 19:56:56.402288 4988 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="c29143467f18a0d9cb5e4136506cbf1b8520fd0496108cf77db65c962e6ea146" cmd=["/usr/bin/pgrep","-r","DRST","heat-engine"] Oct 08 19:56:56 crc kubenswrapper[4988]: E1008 19:56:56.404183 4988 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="c29143467f18a0d9cb5e4136506cbf1b8520fd0496108cf77db65c962e6ea146" cmd=["/usr/bin/pgrep","-r","DRST","heat-engine"] Oct 08 19:56:56 crc kubenswrapper[4988]: E1008 19:56:56.405996 4988 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="c29143467f18a0d9cb5e4136506cbf1b8520fd0496108cf77db65c962e6ea146" cmd=["/usr/bin/pgrep","-r","DRST","heat-engine"] Oct 08 19:56:56 crc kubenswrapper[4988]: E1008 19:56:56.406038 4988 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/heat-engine-d897f4749-rlhsn" podUID="fdbbb87a-7afb-402c-83fd-d4582e06c6ba" containerName="heat-engine" Oct 08 19:57:05 crc kubenswrapper[4988]: I1008 19:57:05.451453 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-engine-d897f4749-rlhsn" Oct 08 19:57:05 crc kubenswrapper[4988]: I1008 19:57:05.640509 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dp87j\" (UniqueName: \"kubernetes.io/projected/fdbbb87a-7afb-402c-83fd-d4582e06c6ba-kube-api-access-dp87j\") pod \"fdbbb87a-7afb-402c-83fd-d4582e06c6ba\" (UID: \"fdbbb87a-7afb-402c-83fd-d4582e06c6ba\") " Oct 08 19:57:05 crc kubenswrapper[4988]: I1008 19:57:05.640689 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/fdbbb87a-7afb-402c-83fd-d4582e06c6ba-config-data-custom\") pod \"fdbbb87a-7afb-402c-83fd-d4582e06c6ba\" (UID: \"fdbbb87a-7afb-402c-83fd-d4582e06c6ba\") " Oct 08 19:57:05 crc kubenswrapper[4988]: I1008 19:57:05.640974 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fdbbb87a-7afb-402c-83fd-d4582e06c6ba-config-data\") pod \"fdbbb87a-7afb-402c-83fd-d4582e06c6ba\" (UID: \"fdbbb87a-7afb-402c-83fd-d4582e06c6ba\") " Oct 08 19:57:05 crc kubenswrapper[4988]: I1008 19:57:05.641182 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fdbbb87a-7afb-402c-83fd-d4582e06c6ba-combined-ca-bundle\") pod \"fdbbb87a-7afb-402c-83fd-d4582e06c6ba\" (UID: \"fdbbb87a-7afb-402c-83fd-d4582e06c6ba\") " Oct 08 19:57:05 crc kubenswrapper[4988]: I1008 19:57:05.647092 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fdbbb87a-7afb-402c-83fd-d4582e06c6ba-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "fdbbb87a-7afb-402c-83fd-d4582e06c6ba" (UID: "fdbbb87a-7afb-402c-83fd-d4582e06c6ba"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 19:57:05 crc kubenswrapper[4988]: I1008 19:57:05.652425 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fdbbb87a-7afb-402c-83fd-d4582e06c6ba-kube-api-access-dp87j" (OuterVolumeSpecName: "kube-api-access-dp87j") pod "fdbbb87a-7afb-402c-83fd-d4582e06c6ba" (UID: "fdbbb87a-7afb-402c-83fd-d4582e06c6ba"). InnerVolumeSpecName "kube-api-access-dp87j". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 19:57:05 crc kubenswrapper[4988]: I1008 19:57:05.672008 4988 generic.go:334] "Generic (PLEG): container finished" podID="fdbbb87a-7afb-402c-83fd-d4582e06c6ba" containerID="c29143467f18a0d9cb5e4136506cbf1b8520fd0496108cf77db65c962e6ea146" exitCode=0 Oct 08 19:57:05 crc kubenswrapper[4988]: I1008 19:57:05.672083 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-engine-d897f4749-rlhsn" event={"ID":"fdbbb87a-7afb-402c-83fd-d4582e06c6ba","Type":"ContainerDied","Data":"c29143467f18a0d9cb5e4136506cbf1b8520fd0496108cf77db65c962e6ea146"} Oct 08 19:57:05 crc kubenswrapper[4988]: I1008 19:57:05.672125 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-engine-d897f4749-rlhsn" event={"ID":"fdbbb87a-7afb-402c-83fd-d4582e06c6ba","Type":"ContainerDied","Data":"f6214419fd3decc57de42d8836027405c6c2bd3700fdcaaa99a9a4101da92b9d"} Oct 08 19:57:05 crc kubenswrapper[4988]: I1008 19:57:05.672143 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-engine-d897f4749-rlhsn" Oct 08 19:57:05 crc kubenswrapper[4988]: I1008 19:57:05.672148 4988 scope.go:117] "RemoveContainer" containerID="c29143467f18a0d9cb5e4136506cbf1b8520fd0496108cf77db65c962e6ea146" Oct 08 19:57:05 crc kubenswrapper[4988]: I1008 19:57:05.695456 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fdbbb87a-7afb-402c-83fd-d4582e06c6ba-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "fdbbb87a-7afb-402c-83fd-d4582e06c6ba" (UID: "fdbbb87a-7afb-402c-83fd-d4582e06c6ba"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 19:57:05 crc kubenswrapper[4988]: I1008 19:57:05.725420 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fdbbb87a-7afb-402c-83fd-d4582e06c6ba-config-data" (OuterVolumeSpecName: "config-data") pod "fdbbb87a-7afb-402c-83fd-d4582e06c6ba" (UID: "fdbbb87a-7afb-402c-83fd-d4582e06c6ba"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 19:57:05 crc kubenswrapper[4988]: I1008 19:57:05.744759 4988 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dp87j\" (UniqueName: \"kubernetes.io/projected/fdbbb87a-7afb-402c-83fd-d4582e06c6ba-kube-api-access-dp87j\") on node \"crc\" DevicePath \"\"" Oct 08 19:57:05 crc kubenswrapper[4988]: I1008 19:57:05.744786 4988 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/fdbbb87a-7afb-402c-83fd-d4582e06c6ba-config-data-custom\") on node \"crc\" DevicePath \"\"" Oct 08 19:57:05 crc kubenswrapper[4988]: I1008 19:57:05.744795 4988 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fdbbb87a-7afb-402c-83fd-d4582e06c6ba-config-data\") on node \"crc\" DevicePath \"\"" Oct 08 19:57:05 crc kubenswrapper[4988]: I1008 19:57:05.744804 4988 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fdbbb87a-7afb-402c-83fd-d4582e06c6ba-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 19:57:05 crc kubenswrapper[4988]: I1008 19:57:05.800092 4988 scope.go:117] "RemoveContainer" containerID="c29143467f18a0d9cb5e4136506cbf1b8520fd0496108cf77db65c962e6ea146" Oct 08 19:57:05 crc kubenswrapper[4988]: E1008 19:57:05.800642 4988 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c29143467f18a0d9cb5e4136506cbf1b8520fd0496108cf77db65c962e6ea146\": container with ID starting with c29143467f18a0d9cb5e4136506cbf1b8520fd0496108cf77db65c962e6ea146 not found: ID does not exist" containerID="c29143467f18a0d9cb5e4136506cbf1b8520fd0496108cf77db65c962e6ea146" Oct 08 19:57:05 crc kubenswrapper[4988]: I1008 19:57:05.800737 4988 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c29143467f18a0d9cb5e4136506cbf1b8520fd0496108cf77db65c962e6ea146"} err="failed to get container status \"c29143467f18a0d9cb5e4136506cbf1b8520fd0496108cf77db65c962e6ea146\": rpc error: code = NotFound desc = could not find container \"c29143467f18a0d9cb5e4136506cbf1b8520fd0496108cf77db65c962e6ea146\": container with ID starting with c29143467f18a0d9cb5e4136506cbf1b8520fd0496108cf77db65c962e6ea146 not found: ID does not exist" Oct 08 19:57:06 crc kubenswrapper[4988]: I1008 19:57:06.026522 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-engine-d897f4749-rlhsn"] Oct 08 19:57:06 crc kubenswrapper[4988]: I1008 19:57:06.033024 4988 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-engine-d897f4749-rlhsn"] Oct 08 19:57:07 crc kubenswrapper[4988]: I1008 19:57:07.250752 4988 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fdbbb87a-7afb-402c-83fd-d4582e06c6ba" path="/var/lib/kubelet/pods/fdbbb87a-7afb-402c-83fd-d4582e06c6ba/volumes" Oct 08 19:57:10 crc kubenswrapper[4988]: I1008 19:57:10.059803 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-create-dgvjr"] Oct 08 19:57:10 crc kubenswrapper[4988]: I1008 19:57:10.071054 4988 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-create-dgvjr"] Oct 08 19:57:11 crc kubenswrapper[4988]: I1008 19:57:11.251909 4988 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6effefe6-da26-4fbd-b3d2-099f884b60b6" path="/var/lib/kubelet/pods/6effefe6-da26-4fbd-b3d2-099f884b60b6/volumes" Oct 08 19:57:11 crc kubenswrapper[4988]: I1008 19:57:11.419787 4988 scope.go:117] "RemoveContainer" containerID="8940dc864fbeee9eb4739fbf376240780589859f1101f8442e27e18082c4f5fa" Oct 08 19:57:21 crc kubenswrapper[4988]: I1008 19:57:21.045660 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-a1b3-account-create-kzgsr"] Oct 08 19:57:21 crc kubenswrapper[4988]: I1008 19:57:21.054871 4988 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-a1b3-account-create-kzgsr"] Oct 08 19:57:21 crc kubenswrapper[4988]: I1008 19:57:21.260464 4988 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="869f964f-941e-4127-95cd-628c2e610300" path="/var/lib/kubelet/pods/869f964f-941e-4127-95cd-628c2e610300/volumes" Oct 08 19:57:23 crc kubenswrapper[4988]: I1008 19:57:23.337807 4988 patch_prober.go:28] interesting pod/machine-config-daemon-nm54f container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 19:57:23 crc kubenswrapper[4988]: I1008 19:57:23.338460 4988 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 19:57:29 crc kubenswrapper[4988]: I1008 19:57:29.051969 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-sync-pb78n"] Oct 08 19:57:29 crc kubenswrapper[4988]: I1008 19:57:29.064803 4988 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-sync-pb78n"] Oct 08 19:57:29 crc kubenswrapper[4988]: I1008 19:57:29.271790 4988 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3636ec06-cf9f-4d7f-8bbe-e67f97662074" path="/var/lib/kubelet/pods/3636ec06-cf9f-4d7f-8bbe-e67f97662074/volumes" Oct 08 19:57:29 crc kubenswrapper[4988]: I1008 19:57:29.718574 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2djwvff"] Oct 08 19:57:29 crc kubenswrapper[4988]: E1008 19:57:29.719142 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9fb22c64-8f41-496f-b894-d1c7d9b80125" containerName="heat-cfnapi" Oct 08 19:57:29 crc kubenswrapper[4988]: I1008 19:57:29.719169 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="9fb22c64-8f41-496f-b894-d1c7d9b80125" containerName="heat-cfnapi" Oct 08 19:57:29 crc kubenswrapper[4988]: E1008 19:57:29.719190 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="255ab071-02d7-46b6-af97-cd84f085b097" containerName="heat-api" Oct 08 19:57:29 crc kubenswrapper[4988]: I1008 19:57:29.719200 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="255ab071-02d7-46b6-af97-cd84f085b097" containerName="heat-api" Oct 08 19:57:29 crc kubenswrapper[4988]: E1008 19:57:29.719223 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c533a2e1-7213-4410-b862-5eed5e1fe8e9" containerName="horizon-log" Oct 08 19:57:29 crc kubenswrapper[4988]: I1008 19:57:29.719233 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="c533a2e1-7213-4410-b862-5eed5e1fe8e9" containerName="horizon-log" Oct 08 19:57:29 crc kubenswrapper[4988]: E1008 19:57:29.719258 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ae3bd686-c190-4513-814a-0e6352ce3fdb" containerName="heat-api" Oct 08 19:57:29 crc kubenswrapper[4988]: I1008 19:57:29.719268 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="ae3bd686-c190-4513-814a-0e6352ce3fdb" containerName="heat-api" Oct 08 19:57:29 crc kubenswrapper[4988]: E1008 19:57:29.719298 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fdbbb87a-7afb-402c-83fd-d4582e06c6ba" containerName="heat-engine" Oct 08 19:57:29 crc kubenswrapper[4988]: I1008 19:57:29.719308 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="fdbbb87a-7afb-402c-83fd-d4582e06c6ba" containerName="heat-engine" Oct 08 19:57:29 crc kubenswrapper[4988]: E1008 19:57:29.719324 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c533a2e1-7213-4410-b862-5eed5e1fe8e9" containerName="horizon" Oct 08 19:57:29 crc kubenswrapper[4988]: I1008 19:57:29.719334 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="c533a2e1-7213-4410-b862-5eed5e1fe8e9" containerName="horizon" Oct 08 19:57:29 crc kubenswrapper[4988]: E1008 19:57:29.719356 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6aeded19-1e22-46b5-8e88-ce9ba127269f" containerName="heat-cfnapi" Oct 08 19:57:29 crc kubenswrapper[4988]: I1008 19:57:29.719365 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="6aeded19-1e22-46b5-8e88-ce9ba127269f" containerName="heat-cfnapi" Oct 08 19:57:29 crc kubenswrapper[4988]: I1008 19:57:29.719682 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="9fb22c64-8f41-496f-b894-d1c7d9b80125" containerName="heat-cfnapi" Oct 08 19:57:29 crc kubenswrapper[4988]: I1008 19:57:29.719709 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="c533a2e1-7213-4410-b862-5eed5e1fe8e9" containerName="horizon" Oct 08 19:57:29 crc kubenswrapper[4988]: I1008 19:57:29.719732 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="ae3bd686-c190-4513-814a-0e6352ce3fdb" containerName="heat-api" Oct 08 19:57:29 crc kubenswrapper[4988]: I1008 19:57:29.719751 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="255ab071-02d7-46b6-af97-cd84f085b097" containerName="heat-api" Oct 08 19:57:29 crc kubenswrapper[4988]: I1008 19:57:29.719776 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="9fb22c64-8f41-496f-b894-d1c7d9b80125" containerName="heat-cfnapi" Oct 08 19:57:29 crc kubenswrapper[4988]: I1008 19:57:29.719793 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="6aeded19-1e22-46b5-8e88-ce9ba127269f" containerName="heat-cfnapi" Oct 08 19:57:29 crc kubenswrapper[4988]: I1008 19:57:29.719810 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="fdbbb87a-7afb-402c-83fd-d4582e06c6ba" containerName="heat-engine" Oct 08 19:57:29 crc kubenswrapper[4988]: I1008 19:57:29.719829 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="c533a2e1-7213-4410-b862-5eed5e1fe8e9" containerName="horizon-log" Oct 08 19:57:29 crc kubenswrapper[4988]: E1008 19:57:29.720122 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9fb22c64-8f41-496f-b894-d1c7d9b80125" containerName="heat-cfnapi" Oct 08 19:57:29 crc kubenswrapper[4988]: I1008 19:57:29.720136 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="9fb22c64-8f41-496f-b894-d1c7d9b80125" containerName="heat-cfnapi" Oct 08 19:57:29 crc kubenswrapper[4988]: E1008 19:57:29.720165 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ae3bd686-c190-4513-814a-0e6352ce3fdb" containerName="heat-api" Oct 08 19:57:29 crc kubenswrapper[4988]: I1008 19:57:29.720175 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="ae3bd686-c190-4513-814a-0e6352ce3fdb" containerName="heat-api" Oct 08 19:57:29 crc kubenswrapper[4988]: I1008 19:57:29.720497 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="ae3bd686-c190-4513-814a-0e6352ce3fdb" containerName="heat-api" Oct 08 19:57:29 crc kubenswrapper[4988]: I1008 19:57:29.721893 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2djwvff" Oct 08 19:57:29 crc kubenswrapper[4988]: I1008 19:57:29.725776 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Oct 08 19:57:29 crc kubenswrapper[4988]: I1008 19:57:29.748332 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2djwvff"] Oct 08 19:57:29 crc kubenswrapper[4988]: I1008 19:57:29.815483 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/e7b36ee1-53f7-46fc-a92c-bdf893239e77-util\") pod \"a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2djwvff\" (UID: \"e7b36ee1-53f7-46fc-a92c-bdf893239e77\") " pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2djwvff" Oct 08 19:57:29 crc kubenswrapper[4988]: I1008 19:57:29.815769 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/e7b36ee1-53f7-46fc-a92c-bdf893239e77-bundle\") pod \"a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2djwvff\" (UID: \"e7b36ee1-53f7-46fc-a92c-bdf893239e77\") " pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2djwvff" Oct 08 19:57:29 crc kubenswrapper[4988]: I1008 19:57:29.815845 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m8sfw\" (UniqueName: \"kubernetes.io/projected/e7b36ee1-53f7-46fc-a92c-bdf893239e77-kube-api-access-m8sfw\") pod \"a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2djwvff\" (UID: \"e7b36ee1-53f7-46fc-a92c-bdf893239e77\") " pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2djwvff" Oct 08 19:57:29 crc kubenswrapper[4988]: I1008 19:57:29.918687 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/e7b36ee1-53f7-46fc-a92c-bdf893239e77-util\") pod \"a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2djwvff\" (UID: \"e7b36ee1-53f7-46fc-a92c-bdf893239e77\") " pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2djwvff" Oct 08 19:57:29 crc kubenswrapper[4988]: I1008 19:57:29.918924 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/e7b36ee1-53f7-46fc-a92c-bdf893239e77-bundle\") pod \"a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2djwvff\" (UID: \"e7b36ee1-53f7-46fc-a92c-bdf893239e77\") " pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2djwvff" Oct 08 19:57:29 crc kubenswrapper[4988]: I1008 19:57:29.919000 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m8sfw\" (UniqueName: \"kubernetes.io/projected/e7b36ee1-53f7-46fc-a92c-bdf893239e77-kube-api-access-m8sfw\") pod \"a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2djwvff\" (UID: \"e7b36ee1-53f7-46fc-a92c-bdf893239e77\") " pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2djwvff" Oct 08 19:57:29 crc kubenswrapper[4988]: I1008 19:57:29.919519 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/e7b36ee1-53f7-46fc-a92c-bdf893239e77-util\") pod \"a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2djwvff\" (UID: \"e7b36ee1-53f7-46fc-a92c-bdf893239e77\") " pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2djwvff" Oct 08 19:57:29 crc kubenswrapper[4988]: I1008 19:57:29.919768 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/e7b36ee1-53f7-46fc-a92c-bdf893239e77-bundle\") pod \"a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2djwvff\" (UID: \"e7b36ee1-53f7-46fc-a92c-bdf893239e77\") " pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2djwvff" Oct 08 19:57:29 crc kubenswrapper[4988]: I1008 19:57:29.955066 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m8sfw\" (UniqueName: \"kubernetes.io/projected/e7b36ee1-53f7-46fc-a92c-bdf893239e77-kube-api-access-m8sfw\") pod \"a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2djwvff\" (UID: \"e7b36ee1-53f7-46fc-a92c-bdf893239e77\") " pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2djwvff" Oct 08 19:57:30 crc kubenswrapper[4988]: I1008 19:57:30.060722 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2djwvff" Oct 08 19:57:30 crc kubenswrapper[4988]: I1008 19:57:30.417730 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2djwvff"] Oct 08 19:57:30 crc kubenswrapper[4988]: I1008 19:57:30.997851 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2djwvff" event={"ID":"e7b36ee1-53f7-46fc-a92c-bdf893239e77","Type":"ContainerStarted","Data":"c2061f83a823446ef45082c0d711d8c195cc02ed6788301a5bfc423d007b0862"} Oct 08 19:57:30 crc kubenswrapper[4988]: I1008 19:57:30.998162 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2djwvff" event={"ID":"e7b36ee1-53f7-46fc-a92c-bdf893239e77","Type":"ContainerStarted","Data":"51163fca74005ba0d95c4a59a8854767a82ca2df5fa9f187a155d8df4f4084ce"} Oct 08 19:57:32 crc kubenswrapper[4988]: I1008 19:57:32.013828 4988 generic.go:334] "Generic (PLEG): container finished" podID="e7b36ee1-53f7-46fc-a92c-bdf893239e77" containerID="c2061f83a823446ef45082c0d711d8c195cc02ed6788301a5bfc423d007b0862" exitCode=0 Oct 08 19:57:32 crc kubenswrapper[4988]: I1008 19:57:32.013881 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2djwvff" event={"ID":"e7b36ee1-53f7-46fc-a92c-bdf893239e77","Type":"ContainerDied","Data":"c2061f83a823446ef45082c0d711d8c195cc02ed6788301a5bfc423d007b0862"} Oct 08 19:57:32 crc kubenswrapper[4988]: I1008 19:57:32.069908 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-rrm85"] Oct 08 19:57:32 crc kubenswrapper[4988]: I1008 19:57:32.083044 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-rrm85" Oct 08 19:57:32 crc kubenswrapper[4988]: I1008 19:57:32.093137 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-rrm85"] Oct 08 19:57:32 crc kubenswrapper[4988]: I1008 19:57:32.270660 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6zzc4\" (UniqueName: \"kubernetes.io/projected/45843c4e-b4ba-4b4f-b1d9-ceeafb52e794-kube-api-access-6zzc4\") pod \"redhat-operators-rrm85\" (UID: \"45843c4e-b4ba-4b4f-b1d9-ceeafb52e794\") " pod="openshift-marketplace/redhat-operators-rrm85" Oct 08 19:57:32 crc kubenswrapper[4988]: I1008 19:57:32.270718 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/45843c4e-b4ba-4b4f-b1d9-ceeafb52e794-catalog-content\") pod \"redhat-operators-rrm85\" (UID: \"45843c4e-b4ba-4b4f-b1d9-ceeafb52e794\") " pod="openshift-marketplace/redhat-operators-rrm85" Oct 08 19:57:32 crc kubenswrapper[4988]: I1008 19:57:32.270954 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/45843c4e-b4ba-4b4f-b1d9-ceeafb52e794-utilities\") pod \"redhat-operators-rrm85\" (UID: \"45843c4e-b4ba-4b4f-b1d9-ceeafb52e794\") " pod="openshift-marketplace/redhat-operators-rrm85" Oct 08 19:57:32 crc kubenswrapper[4988]: I1008 19:57:32.372694 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6zzc4\" (UniqueName: \"kubernetes.io/projected/45843c4e-b4ba-4b4f-b1d9-ceeafb52e794-kube-api-access-6zzc4\") pod \"redhat-operators-rrm85\" (UID: \"45843c4e-b4ba-4b4f-b1d9-ceeafb52e794\") " pod="openshift-marketplace/redhat-operators-rrm85" Oct 08 19:57:32 crc kubenswrapper[4988]: I1008 19:57:32.372748 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/45843c4e-b4ba-4b4f-b1d9-ceeafb52e794-catalog-content\") pod \"redhat-operators-rrm85\" (UID: \"45843c4e-b4ba-4b4f-b1d9-ceeafb52e794\") " pod="openshift-marketplace/redhat-operators-rrm85" Oct 08 19:57:32 crc kubenswrapper[4988]: I1008 19:57:32.372834 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/45843c4e-b4ba-4b4f-b1d9-ceeafb52e794-utilities\") pod \"redhat-operators-rrm85\" (UID: \"45843c4e-b4ba-4b4f-b1d9-ceeafb52e794\") " pod="openshift-marketplace/redhat-operators-rrm85" Oct 08 19:57:32 crc kubenswrapper[4988]: I1008 19:57:32.373986 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/45843c4e-b4ba-4b4f-b1d9-ceeafb52e794-catalog-content\") pod \"redhat-operators-rrm85\" (UID: \"45843c4e-b4ba-4b4f-b1d9-ceeafb52e794\") " pod="openshift-marketplace/redhat-operators-rrm85" Oct 08 19:57:32 crc kubenswrapper[4988]: I1008 19:57:32.374018 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/45843c4e-b4ba-4b4f-b1d9-ceeafb52e794-utilities\") pod \"redhat-operators-rrm85\" (UID: \"45843c4e-b4ba-4b4f-b1d9-ceeafb52e794\") " pod="openshift-marketplace/redhat-operators-rrm85" Oct 08 19:57:32 crc kubenswrapper[4988]: I1008 19:57:32.390894 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6zzc4\" (UniqueName: \"kubernetes.io/projected/45843c4e-b4ba-4b4f-b1d9-ceeafb52e794-kube-api-access-6zzc4\") pod \"redhat-operators-rrm85\" (UID: \"45843c4e-b4ba-4b4f-b1d9-ceeafb52e794\") " pod="openshift-marketplace/redhat-operators-rrm85" Oct 08 19:57:32 crc kubenswrapper[4988]: I1008 19:57:32.430658 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-rrm85" Oct 08 19:57:32 crc kubenswrapper[4988]: I1008 19:57:32.909765 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-rrm85"] Oct 08 19:57:32 crc kubenswrapper[4988]: W1008 19:57:32.986751 4988 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod45843c4e_b4ba_4b4f_b1d9_ceeafb52e794.slice/crio-f5e3a7a148fd5096b9e2cf8261a6ba9764d6f911839c5ef7179d7bdba26a54e8 WatchSource:0}: Error finding container f5e3a7a148fd5096b9e2cf8261a6ba9764d6f911839c5ef7179d7bdba26a54e8: Status 404 returned error can't find the container with id f5e3a7a148fd5096b9e2cf8261a6ba9764d6f911839c5ef7179d7bdba26a54e8 Oct 08 19:57:33 crc kubenswrapper[4988]: I1008 19:57:33.041773 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rrm85" event={"ID":"45843c4e-b4ba-4b4f-b1d9-ceeafb52e794","Type":"ContainerStarted","Data":"f5e3a7a148fd5096b9e2cf8261a6ba9764d6f911839c5ef7179d7bdba26a54e8"} Oct 08 19:57:34 crc kubenswrapper[4988]: I1008 19:57:34.059643 4988 generic.go:334] "Generic (PLEG): container finished" podID="45843c4e-b4ba-4b4f-b1d9-ceeafb52e794" containerID="6881bff31c8fd6399e7581d8a3047ad71ff216cdc5694e04cd78641657858591" exitCode=0 Oct 08 19:57:34 crc kubenswrapper[4988]: I1008 19:57:34.059716 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rrm85" event={"ID":"45843c4e-b4ba-4b4f-b1d9-ceeafb52e794","Type":"ContainerDied","Data":"6881bff31c8fd6399e7581d8a3047ad71ff216cdc5694e04cd78641657858591"} Oct 08 19:57:34 crc kubenswrapper[4988]: I1008 19:57:34.068198 4988 generic.go:334] "Generic (PLEG): container finished" podID="e7b36ee1-53f7-46fc-a92c-bdf893239e77" containerID="fb177145a017ea1b7fbd29db885456fa7797d524432ab17d20545726423f3db7" exitCode=0 Oct 08 19:57:34 crc kubenswrapper[4988]: I1008 19:57:34.068245 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2djwvff" event={"ID":"e7b36ee1-53f7-46fc-a92c-bdf893239e77","Type":"ContainerDied","Data":"fb177145a017ea1b7fbd29db885456fa7797d524432ab17d20545726423f3db7"} Oct 08 19:57:35 crc kubenswrapper[4988]: I1008 19:57:35.079094 4988 generic.go:334] "Generic (PLEG): container finished" podID="e7b36ee1-53f7-46fc-a92c-bdf893239e77" containerID="79c5f305187f6502c9a34e047a379b9654eee4ac4b1912fb06f1a6d777aaa129" exitCode=0 Oct 08 19:57:35 crc kubenswrapper[4988]: I1008 19:57:35.079300 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2djwvff" event={"ID":"e7b36ee1-53f7-46fc-a92c-bdf893239e77","Type":"ContainerDied","Data":"79c5f305187f6502c9a34e047a379b9654eee4ac4b1912fb06f1a6d777aaa129"} Oct 08 19:57:36 crc kubenswrapper[4988]: I1008 19:57:36.098430 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rrm85" event={"ID":"45843c4e-b4ba-4b4f-b1d9-ceeafb52e794","Type":"ContainerStarted","Data":"f4b64dc676d09b73bd0e5e7c83afa900e5dc94774ba7d3446540e7650d58659b"} Oct 08 19:57:36 crc kubenswrapper[4988]: I1008 19:57:36.508489 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2djwvff" Oct 08 19:57:36 crc kubenswrapper[4988]: I1008 19:57:36.681841 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m8sfw\" (UniqueName: \"kubernetes.io/projected/e7b36ee1-53f7-46fc-a92c-bdf893239e77-kube-api-access-m8sfw\") pod \"e7b36ee1-53f7-46fc-a92c-bdf893239e77\" (UID: \"e7b36ee1-53f7-46fc-a92c-bdf893239e77\") " Oct 08 19:57:36 crc kubenswrapper[4988]: I1008 19:57:36.681894 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/e7b36ee1-53f7-46fc-a92c-bdf893239e77-bundle\") pod \"e7b36ee1-53f7-46fc-a92c-bdf893239e77\" (UID: \"e7b36ee1-53f7-46fc-a92c-bdf893239e77\") " Oct 08 19:57:36 crc kubenswrapper[4988]: I1008 19:57:36.681933 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/e7b36ee1-53f7-46fc-a92c-bdf893239e77-util\") pod \"e7b36ee1-53f7-46fc-a92c-bdf893239e77\" (UID: \"e7b36ee1-53f7-46fc-a92c-bdf893239e77\") " Oct 08 19:57:36 crc kubenswrapper[4988]: I1008 19:57:36.684979 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e7b36ee1-53f7-46fc-a92c-bdf893239e77-bundle" (OuterVolumeSpecName: "bundle") pod "e7b36ee1-53f7-46fc-a92c-bdf893239e77" (UID: "e7b36ee1-53f7-46fc-a92c-bdf893239e77"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 19:57:36 crc kubenswrapper[4988]: I1008 19:57:36.691349 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e7b36ee1-53f7-46fc-a92c-bdf893239e77-kube-api-access-m8sfw" (OuterVolumeSpecName: "kube-api-access-m8sfw") pod "e7b36ee1-53f7-46fc-a92c-bdf893239e77" (UID: "e7b36ee1-53f7-46fc-a92c-bdf893239e77"). InnerVolumeSpecName "kube-api-access-m8sfw". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 19:57:36 crc kubenswrapper[4988]: I1008 19:57:36.784223 4988 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-m8sfw\" (UniqueName: \"kubernetes.io/projected/e7b36ee1-53f7-46fc-a92c-bdf893239e77-kube-api-access-m8sfw\") on node \"crc\" DevicePath \"\"" Oct 08 19:57:36 crc kubenswrapper[4988]: I1008 19:57:36.784267 4988 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/e7b36ee1-53f7-46fc-a92c-bdf893239e77-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 19:57:37 crc kubenswrapper[4988]: I1008 19:57:37.118738 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2djwvff" Oct 08 19:57:37 crc kubenswrapper[4988]: I1008 19:57:37.118729 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2djwvff" event={"ID":"e7b36ee1-53f7-46fc-a92c-bdf893239e77","Type":"ContainerDied","Data":"51163fca74005ba0d95c4a59a8854767a82ca2df5fa9f187a155d8df4f4084ce"} Oct 08 19:57:37 crc kubenswrapper[4988]: I1008 19:57:37.120612 4988 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="51163fca74005ba0d95c4a59a8854767a82ca2df5fa9f187a155d8df4f4084ce" Oct 08 19:57:37 crc kubenswrapper[4988]: I1008 19:57:37.124638 4988 generic.go:334] "Generic (PLEG): container finished" podID="45843c4e-b4ba-4b4f-b1d9-ceeafb52e794" containerID="f4b64dc676d09b73bd0e5e7c83afa900e5dc94774ba7d3446540e7650d58659b" exitCode=0 Oct 08 19:57:37 crc kubenswrapper[4988]: I1008 19:57:37.124729 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rrm85" event={"ID":"45843c4e-b4ba-4b4f-b1d9-ceeafb52e794","Type":"ContainerDied","Data":"f4b64dc676d09b73bd0e5e7c83afa900e5dc94774ba7d3446540e7650d58659b"} Oct 08 19:57:37 crc kubenswrapper[4988]: I1008 19:57:37.134985 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e7b36ee1-53f7-46fc-a92c-bdf893239e77-util" (OuterVolumeSpecName: "util") pod "e7b36ee1-53f7-46fc-a92c-bdf893239e77" (UID: "e7b36ee1-53f7-46fc-a92c-bdf893239e77"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 19:57:37 crc kubenswrapper[4988]: I1008 19:57:37.191782 4988 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/e7b36ee1-53f7-46fc-a92c-bdf893239e77-util\") on node \"crc\" DevicePath \"\"" Oct 08 19:57:38 crc kubenswrapper[4988]: I1008 19:57:38.136707 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rrm85" event={"ID":"45843c4e-b4ba-4b4f-b1d9-ceeafb52e794","Type":"ContainerStarted","Data":"d620feb5bdb67de79dc31481e4abe24b19e6e7779e0417024d4655eb274bba06"} Oct 08 19:57:38 crc kubenswrapper[4988]: I1008 19:57:38.154638 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-rrm85" podStartSLOduration=2.610373175 podStartE2EDuration="6.15461613s" podCreationTimestamp="2025-10-08 19:57:32 +0000 UTC" firstStartedPulling="2025-10-08 19:57:34.063208307 +0000 UTC m=+6399.513051087" lastFinishedPulling="2025-10-08 19:57:37.607451232 +0000 UTC m=+6403.057294042" observedRunningTime="2025-10-08 19:57:38.153050301 +0000 UTC m=+6403.602893091" watchObservedRunningTime="2025-10-08 19:57:38.15461613 +0000 UTC m=+6403.604458940" Oct 08 19:57:42 crc kubenswrapper[4988]: I1008 19:57:42.431558 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-rrm85" Oct 08 19:57:42 crc kubenswrapper[4988]: I1008 19:57:42.432138 4988 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-rrm85" Oct 08 19:57:43 crc kubenswrapper[4988]: I1008 19:57:43.480271 4988 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-rrm85" podUID="45843c4e-b4ba-4b4f-b1d9-ceeafb52e794" containerName="registry-server" probeResult="failure" output=< Oct 08 19:57:43 crc kubenswrapper[4988]: timeout: failed to connect service ":50051" within 1s Oct 08 19:57:43 crc kubenswrapper[4988]: > Oct 08 19:57:45 crc kubenswrapper[4988]: I1008 19:57:45.040161 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/obo-prometheus-operator-7c8cf85677-pwbvx"] Oct 08 19:57:45 crc kubenswrapper[4988]: E1008 19:57:45.041439 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e7b36ee1-53f7-46fc-a92c-bdf893239e77" containerName="extract" Oct 08 19:57:45 crc kubenswrapper[4988]: I1008 19:57:45.041459 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="e7b36ee1-53f7-46fc-a92c-bdf893239e77" containerName="extract" Oct 08 19:57:45 crc kubenswrapper[4988]: E1008 19:57:45.041528 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e7b36ee1-53f7-46fc-a92c-bdf893239e77" containerName="pull" Oct 08 19:57:45 crc kubenswrapper[4988]: I1008 19:57:45.041535 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="e7b36ee1-53f7-46fc-a92c-bdf893239e77" containerName="pull" Oct 08 19:57:45 crc kubenswrapper[4988]: E1008 19:57:45.041594 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e7b36ee1-53f7-46fc-a92c-bdf893239e77" containerName="util" Oct 08 19:57:45 crc kubenswrapper[4988]: I1008 19:57:45.041602 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="e7b36ee1-53f7-46fc-a92c-bdf893239e77" containerName="util" Oct 08 19:57:45 crc kubenswrapper[4988]: I1008 19:57:45.042395 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="e7b36ee1-53f7-46fc-a92c-bdf893239e77" containerName="extract" Oct 08 19:57:45 crc kubenswrapper[4988]: I1008 19:57:45.043367 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-7c8cf85677-pwbvx" Oct 08 19:57:45 crc kubenswrapper[4988]: I1008 19:57:45.052034 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-7c8cf85677-pwbvx"] Oct 08 19:57:45 crc kubenswrapper[4988]: I1008 19:57:45.069968 4988 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operators"/"openshift-service-ca.crt" Oct 08 19:57:45 crc kubenswrapper[4988]: I1008 19:57:45.070249 4988 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operators"/"kube-root-ca.crt" Oct 08 19:57:45 crc kubenswrapper[4988]: I1008 19:57:45.070378 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"obo-prometheus-operator-dockercfg-sp7sp" Oct 08 19:57:45 crc kubenswrapper[4988]: I1008 19:57:45.146040 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-6b789f9dcf-d7jwn"] Oct 08 19:57:45 crc kubenswrapper[4988]: I1008 19:57:45.158006 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-6b789f9dcf-d7jwn" Oct 08 19:57:45 crc kubenswrapper[4988]: I1008 19:57:45.160705 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"obo-prometheus-operator-admission-webhook-service-cert" Oct 08 19:57:45 crc kubenswrapper[4988]: I1008 19:57:45.161425 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-6b789f9dcf-cs98t"] Oct 08 19:57:45 crc kubenswrapper[4988]: I1008 19:57:45.163070 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-6b789f9dcf-cs98t" Oct 08 19:57:45 crc kubenswrapper[4988]: I1008 19:57:45.165150 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"obo-prometheus-operator-admission-webhook-dockercfg-mh5sg" Oct 08 19:57:45 crc kubenswrapper[4988]: I1008 19:57:45.169778 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2mxm5\" (UniqueName: \"kubernetes.io/projected/38d89d99-8b9b-43e1-85ad-d928aad3d3d1-kube-api-access-2mxm5\") pod \"obo-prometheus-operator-7c8cf85677-pwbvx\" (UID: \"38d89d99-8b9b-43e1-85ad-d928aad3d3d1\") " pod="openshift-operators/obo-prometheus-operator-7c8cf85677-pwbvx" Oct 08 19:57:45 crc kubenswrapper[4988]: I1008 19:57:45.176176 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-6b789f9dcf-d7jwn"] Oct 08 19:57:45 crc kubenswrapper[4988]: I1008 19:57:45.189427 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-6b789f9dcf-cs98t"] Oct 08 19:57:45 crc kubenswrapper[4988]: I1008 19:57:45.271548 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/d625ac6d-4570-4a73-8e1f-14ea64510864-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-6b789f9dcf-cs98t\" (UID: \"d625ac6d-4570-4a73-8e1f-14ea64510864\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-6b789f9dcf-cs98t" Oct 08 19:57:45 crc kubenswrapper[4988]: I1008 19:57:45.271597 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/3a9085fc-8d56-486f-a107-3e03f9ae0e64-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-6b789f9dcf-d7jwn\" (UID: \"3a9085fc-8d56-486f-a107-3e03f9ae0e64\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-6b789f9dcf-d7jwn" Oct 08 19:57:45 crc kubenswrapper[4988]: I1008 19:57:45.271653 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/d625ac6d-4570-4a73-8e1f-14ea64510864-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-6b789f9dcf-cs98t\" (UID: \"d625ac6d-4570-4a73-8e1f-14ea64510864\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-6b789f9dcf-cs98t" Oct 08 19:57:45 crc kubenswrapper[4988]: I1008 19:57:45.271707 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2mxm5\" (UniqueName: \"kubernetes.io/projected/38d89d99-8b9b-43e1-85ad-d928aad3d3d1-kube-api-access-2mxm5\") pod \"obo-prometheus-operator-7c8cf85677-pwbvx\" (UID: \"38d89d99-8b9b-43e1-85ad-d928aad3d3d1\") " pod="openshift-operators/obo-prometheus-operator-7c8cf85677-pwbvx" Oct 08 19:57:45 crc kubenswrapper[4988]: I1008 19:57:45.271785 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/3a9085fc-8d56-486f-a107-3e03f9ae0e64-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-6b789f9dcf-d7jwn\" (UID: \"3a9085fc-8d56-486f-a107-3e03f9ae0e64\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-6b789f9dcf-d7jwn" Oct 08 19:57:45 crc kubenswrapper[4988]: I1008 19:57:45.293645 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2mxm5\" (UniqueName: \"kubernetes.io/projected/38d89d99-8b9b-43e1-85ad-d928aad3d3d1-kube-api-access-2mxm5\") pod \"obo-prometheus-operator-7c8cf85677-pwbvx\" (UID: \"38d89d99-8b9b-43e1-85ad-d928aad3d3d1\") " pod="openshift-operators/obo-prometheus-operator-7c8cf85677-pwbvx" Oct 08 19:57:45 crc kubenswrapper[4988]: I1008 19:57:45.343296 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/observability-operator-cc5f78dfc-7qjkj"] Oct 08 19:57:45 crc kubenswrapper[4988]: I1008 19:57:45.344586 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/observability-operator-cc5f78dfc-7qjkj" Oct 08 19:57:45 crc kubenswrapper[4988]: I1008 19:57:45.347085 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"observability-operator-tls" Oct 08 19:57:45 crc kubenswrapper[4988]: I1008 19:57:45.355767 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"observability-operator-sa-dockercfg-4xrrq" Oct 08 19:57:45 crc kubenswrapper[4988]: I1008 19:57:45.361501 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/observability-operator-cc5f78dfc-7qjkj"] Oct 08 19:57:45 crc kubenswrapper[4988]: I1008 19:57:45.383494 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/d625ac6d-4570-4a73-8e1f-14ea64510864-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-6b789f9dcf-cs98t\" (UID: \"d625ac6d-4570-4a73-8e1f-14ea64510864\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-6b789f9dcf-cs98t" Oct 08 19:57:45 crc kubenswrapper[4988]: I1008 19:57:45.383538 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/3a9085fc-8d56-486f-a107-3e03f9ae0e64-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-6b789f9dcf-d7jwn\" (UID: \"3a9085fc-8d56-486f-a107-3e03f9ae0e64\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-6b789f9dcf-d7jwn" Oct 08 19:57:45 crc kubenswrapper[4988]: I1008 19:57:45.383602 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/d625ac6d-4570-4a73-8e1f-14ea64510864-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-6b789f9dcf-cs98t\" (UID: \"d625ac6d-4570-4a73-8e1f-14ea64510864\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-6b789f9dcf-cs98t" Oct 08 19:57:45 crc kubenswrapper[4988]: I1008 19:57:45.387500 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/d625ac6d-4570-4a73-8e1f-14ea64510864-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-6b789f9dcf-cs98t\" (UID: \"d625ac6d-4570-4a73-8e1f-14ea64510864\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-6b789f9dcf-cs98t" Oct 08 19:57:45 crc kubenswrapper[4988]: I1008 19:57:45.389691 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-7c8cf85677-pwbvx" Oct 08 19:57:45 crc kubenswrapper[4988]: I1008 19:57:45.389879 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/3a9085fc-8d56-486f-a107-3e03f9ae0e64-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-6b789f9dcf-d7jwn\" (UID: \"3a9085fc-8d56-486f-a107-3e03f9ae0e64\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-6b789f9dcf-d7jwn" Oct 08 19:57:45 crc kubenswrapper[4988]: I1008 19:57:45.390624 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/3a9085fc-8d56-486f-a107-3e03f9ae0e64-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-6b789f9dcf-d7jwn\" (UID: \"3a9085fc-8d56-486f-a107-3e03f9ae0e64\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-6b789f9dcf-d7jwn" Oct 08 19:57:45 crc kubenswrapper[4988]: I1008 19:57:45.406919 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/d625ac6d-4570-4a73-8e1f-14ea64510864-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-6b789f9dcf-cs98t\" (UID: \"d625ac6d-4570-4a73-8e1f-14ea64510864\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-6b789f9dcf-cs98t" Oct 08 19:57:45 crc kubenswrapper[4988]: I1008 19:57:45.409862 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/3a9085fc-8d56-486f-a107-3e03f9ae0e64-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-6b789f9dcf-d7jwn\" (UID: \"3a9085fc-8d56-486f-a107-3e03f9ae0e64\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-6b789f9dcf-d7jwn" Oct 08 19:57:45 crc kubenswrapper[4988]: I1008 19:57:45.485038 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-6b789f9dcf-d7jwn" Oct 08 19:57:45 crc kubenswrapper[4988]: I1008 19:57:45.492577 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"observability-operator-tls\" (UniqueName: \"kubernetes.io/secret/86cf1f5c-ab13-4b35-ad2f-9529ccdf96db-observability-operator-tls\") pod \"observability-operator-cc5f78dfc-7qjkj\" (UID: \"86cf1f5c-ab13-4b35-ad2f-9529ccdf96db\") " pod="openshift-operators/observability-operator-cc5f78dfc-7qjkj" Oct 08 19:57:45 crc kubenswrapper[4988]: I1008 19:57:45.492782 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zjcrc\" (UniqueName: \"kubernetes.io/projected/86cf1f5c-ab13-4b35-ad2f-9529ccdf96db-kube-api-access-zjcrc\") pod \"observability-operator-cc5f78dfc-7qjkj\" (UID: \"86cf1f5c-ab13-4b35-ad2f-9529ccdf96db\") " pod="openshift-operators/observability-operator-cc5f78dfc-7qjkj" Oct 08 19:57:45 crc kubenswrapper[4988]: I1008 19:57:45.504646 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-6b789f9dcf-cs98t" Oct 08 19:57:45 crc kubenswrapper[4988]: I1008 19:57:45.555360 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/perses-operator-54bc95c9fb-ln8dz"] Oct 08 19:57:45 crc kubenswrapper[4988]: I1008 19:57:45.571544 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/perses-operator-54bc95c9fb-ln8dz" Oct 08 19:57:45 crc kubenswrapper[4988]: I1008 19:57:45.577659 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"perses-operator-dockercfg-j2wfj" Oct 08 19:57:45 crc kubenswrapper[4988]: I1008 19:57:45.579639 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/perses-operator-54bc95c9fb-ln8dz"] Oct 08 19:57:45 crc kubenswrapper[4988]: I1008 19:57:45.598781 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"observability-operator-tls\" (UniqueName: \"kubernetes.io/secret/86cf1f5c-ab13-4b35-ad2f-9529ccdf96db-observability-operator-tls\") pod \"observability-operator-cc5f78dfc-7qjkj\" (UID: \"86cf1f5c-ab13-4b35-ad2f-9529ccdf96db\") " pod="openshift-operators/observability-operator-cc5f78dfc-7qjkj" Oct 08 19:57:45 crc kubenswrapper[4988]: I1008 19:57:45.598940 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zjcrc\" (UniqueName: \"kubernetes.io/projected/86cf1f5c-ab13-4b35-ad2f-9529ccdf96db-kube-api-access-zjcrc\") pod \"observability-operator-cc5f78dfc-7qjkj\" (UID: \"86cf1f5c-ab13-4b35-ad2f-9529ccdf96db\") " pod="openshift-operators/observability-operator-cc5f78dfc-7qjkj" Oct 08 19:57:45 crc kubenswrapper[4988]: I1008 19:57:45.605543 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"observability-operator-tls\" (UniqueName: \"kubernetes.io/secret/86cf1f5c-ab13-4b35-ad2f-9529ccdf96db-observability-operator-tls\") pod \"observability-operator-cc5f78dfc-7qjkj\" (UID: \"86cf1f5c-ab13-4b35-ad2f-9529ccdf96db\") " pod="openshift-operators/observability-operator-cc5f78dfc-7qjkj" Oct 08 19:57:45 crc kubenswrapper[4988]: I1008 19:57:45.622905 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zjcrc\" (UniqueName: \"kubernetes.io/projected/86cf1f5c-ab13-4b35-ad2f-9529ccdf96db-kube-api-access-zjcrc\") pod \"observability-operator-cc5f78dfc-7qjkj\" (UID: \"86cf1f5c-ab13-4b35-ad2f-9529ccdf96db\") " pod="openshift-operators/observability-operator-cc5f78dfc-7qjkj" Oct 08 19:57:45 crc kubenswrapper[4988]: I1008 19:57:45.683793 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/observability-operator-cc5f78dfc-7qjkj" Oct 08 19:57:45 crc kubenswrapper[4988]: I1008 19:57:45.701727 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openshift-service-ca\" (UniqueName: \"kubernetes.io/configmap/8cf809bc-9b01-4426-a099-be47b7b711ba-openshift-service-ca\") pod \"perses-operator-54bc95c9fb-ln8dz\" (UID: \"8cf809bc-9b01-4426-a099-be47b7b711ba\") " pod="openshift-operators/perses-operator-54bc95c9fb-ln8dz" Oct 08 19:57:45 crc kubenswrapper[4988]: I1008 19:57:45.701787 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gnt72\" (UniqueName: \"kubernetes.io/projected/8cf809bc-9b01-4426-a099-be47b7b711ba-kube-api-access-gnt72\") pod \"perses-operator-54bc95c9fb-ln8dz\" (UID: \"8cf809bc-9b01-4426-a099-be47b7b711ba\") " pod="openshift-operators/perses-operator-54bc95c9fb-ln8dz" Oct 08 19:57:45 crc kubenswrapper[4988]: I1008 19:57:45.804830 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openshift-service-ca\" (UniqueName: \"kubernetes.io/configmap/8cf809bc-9b01-4426-a099-be47b7b711ba-openshift-service-ca\") pod \"perses-operator-54bc95c9fb-ln8dz\" (UID: \"8cf809bc-9b01-4426-a099-be47b7b711ba\") " pod="openshift-operators/perses-operator-54bc95c9fb-ln8dz" Oct 08 19:57:45 crc kubenswrapper[4988]: I1008 19:57:45.805184 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gnt72\" (UniqueName: \"kubernetes.io/projected/8cf809bc-9b01-4426-a099-be47b7b711ba-kube-api-access-gnt72\") pod \"perses-operator-54bc95c9fb-ln8dz\" (UID: \"8cf809bc-9b01-4426-a099-be47b7b711ba\") " pod="openshift-operators/perses-operator-54bc95c9fb-ln8dz" Oct 08 19:57:45 crc kubenswrapper[4988]: I1008 19:57:45.806625 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openshift-service-ca\" (UniqueName: \"kubernetes.io/configmap/8cf809bc-9b01-4426-a099-be47b7b711ba-openshift-service-ca\") pod \"perses-operator-54bc95c9fb-ln8dz\" (UID: \"8cf809bc-9b01-4426-a099-be47b7b711ba\") " pod="openshift-operators/perses-operator-54bc95c9fb-ln8dz" Oct 08 19:57:45 crc kubenswrapper[4988]: I1008 19:57:45.836135 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gnt72\" (UniqueName: \"kubernetes.io/projected/8cf809bc-9b01-4426-a099-be47b7b711ba-kube-api-access-gnt72\") pod \"perses-operator-54bc95c9fb-ln8dz\" (UID: \"8cf809bc-9b01-4426-a099-be47b7b711ba\") " pod="openshift-operators/perses-operator-54bc95c9fb-ln8dz" Oct 08 19:57:45 crc kubenswrapper[4988]: I1008 19:57:45.933909 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/perses-operator-54bc95c9fb-ln8dz" Oct 08 19:57:46 crc kubenswrapper[4988]: I1008 19:57:46.097464 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-7c8cf85677-pwbvx"] Oct 08 19:57:46 crc kubenswrapper[4988]: I1008 19:57:46.209396 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-7c8cf85677-pwbvx" event={"ID":"38d89d99-8b9b-43e1-85ad-d928aad3d3d1","Type":"ContainerStarted","Data":"50092882eb8c85b5f55f7df347fd993757c1e88246b1986dfd8fab5b9b465f2b"} Oct 08 19:57:46 crc kubenswrapper[4988]: I1008 19:57:46.233611 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/observability-operator-cc5f78dfc-7qjkj"] Oct 08 19:57:46 crc kubenswrapper[4988]: I1008 19:57:46.253082 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-6b789f9dcf-cs98t"] Oct 08 19:57:46 crc kubenswrapper[4988]: I1008 19:57:46.279757 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-6b789f9dcf-d7jwn"] Oct 08 19:57:46 crc kubenswrapper[4988]: W1008 19:57:46.283101 4988 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3a9085fc_8d56_486f_a107_3e03f9ae0e64.slice/crio-d934271999380858e209b950d1213d87f96a8a1898a3b701e9d7d5f206836a79 WatchSource:0}: Error finding container d934271999380858e209b950d1213d87f96a8a1898a3b701e9d7d5f206836a79: Status 404 returned error can't find the container with id d934271999380858e209b950d1213d87f96a8a1898a3b701e9d7d5f206836a79 Oct 08 19:57:46 crc kubenswrapper[4988]: I1008 19:57:46.505481 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/perses-operator-54bc95c9fb-ln8dz"] Oct 08 19:57:46 crc kubenswrapper[4988]: W1008 19:57:46.509959 4988 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8cf809bc_9b01_4426_a099_be47b7b711ba.slice/crio-bd9bbcb5d3b928cb2174aa8a7a83c40afa1f4ed58d03ec47d90e092e16f23aeb WatchSource:0}: Error finding container bd9bbcb5d3b928cb2174aa8a7a83c40afa1f4ed58d03ec47d90e092e16f23aeb: Status 404 returned error can't find the container with id bd9bbcb5d3b928cb2174aa8a7a83c40afa1f4ed58d03ec47d90e092e16f23aeb Oct 08 19:57:47 crc kubenswrapper[4988]: I1008 19:57:47.219215 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-admission-webhook-6b789f9dcf-d7jwn" event={"ID":"3a9085fc-8d56-486f-a107-3e03f9ae0e64","Type":"ContainerStarted","Data":"d934271999380858e209b950d1213d87f96a8a1898a3b701e9d7d5f206836a79"} Oct 08 19:57:47 crc kubenswrapper[4988]: I1008 19:57:47.220785 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/perses-operator-54bc95c9fb-ln8dz" event={"ID":"8cf809bc-9b01-4426-a099-be47b7b711ba","Type":"ContainerStarted","Data":"bd9bbcb5d3b928cb2174aa8a7a83c40afa1f4ed58d03ec47d90e092e16f23aeb"} Oct 08 19:57:47 crc kubenswrapper[4988]: I1008 19:57:47.222328 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-admission-webhook-6b789f9dcf-cs98t" event={"ID":"d625ac6d-4570-4a73-8e1f-14ea64510864","Type":"ContainerStarted","Data":"0612e086bfdf90b2efe88571c37657f8d50bc79f00fabd1d1535df6f54492a13"} Oct 08 19:57:47 crc kubenswrapper[4988]: I1008 19:57:47.223964 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/observability-operator-cc5f78dfc-7qjkj" event={"ID":"86cf1f5c-ab13-4b35-ad2f-9529ccdf96db","Type":"ContainerStarted","Data":"64180996560a7e75fd89b4d44b0cfa7ca9a437d1918ae18831b32ab363c95f34"} Oct 08 19:57:51 crc kubenswrapper[4988]: I1008 19:57:51.297663 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/perses-operator-54bc95c9fb-ln8dz" event={"ID":"8cf809bc-9b01-4426-a099-be47b7b711ba","Type":"ContainerStarted","Data":"69f603f21f586d4440913a4f5e5c24e914e662bbcdb31852eaeb1ded9a1c6fc8"} Oct 08 19:57:51 crc kubenswrapper[4988]: I1008 19:57:51.298340 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operators/perses-operator-54bc95c9fb-ln8dz" Oct 08 19:57:51 crc kubenswrapper[4988]: I1008 19:57:51.302750 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-admission-webhook-6b789f9dcf-cs98t" event={"ID":"d625ac6d-4570-4a73-8e1f-14ea64510864","Type":"ContainerStarted","Data":"e355d7c057cefb495cb1d4a4f01e32488cf7d4ff426b12e0472e7850d5de3d91"} Oct 08 19:57:51 crc kubenswrapper[4988]: I1008 19:57:51.308594 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-admission-webhook-6b789f9dcf-d7jwn" event={"ID":"3a9085fc-8d56-486f-a107-3e03f9ae0e64","Type":"ContainerStarted","Data":"84dbf35bcfa05046f213b06981961d9be0cc2781b9b8a56da1647360c821ee6a"} Oct 08 19:57:51 crc kubenswrapper[4988]: I1008 19:57:51.326132 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/perses-operator-54bc95c9fb-ln8dz" podStartSLOduration=2.160518427 podStartE2EDuration="6.326111273s" podCreationTimestamp="2025-10-08 19:57:45 +0000 UTC" firstStartedPulling="2025-10-08 19:57:46.512461144 +0000 UTC m=+6411.962303914" lastFinishedPulling="2025-10-08 19:57:50.67805399 +0000 UTC m=+6416.127896760" observedRunningTime="2025-10-08 19:57:51.312897672 +0000 UTC m=+6416.762740462" watchObservedRunningTime="2025-10-08 19:57:51.326111273 +0000 UTC m=+6416.775954043" Oct 08 19:57:51 crc kubenswrapper[4988]: I1008 19:57:51.337526 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/obo-prometheus-operator-admission-webhook-6b789f9dcf-cs98t" podStartSLOduration=4.060426724 podStartE2EDuration="6.337500255s" podCreationTimestamp="2025-10-08 19:57:45 +0000 UTC" firstStartedPulling="2025-10-08 19:57:46.283143649 +0000 UTC m=+6411.732986419" lastFinishedPulling="2025-10-08 19:57:48.56021718 +0000 UTC m=+6414.010059950" observedRunningTime="2025-10-08 19:57:51.332837567 +0000 UTC m=+6416.782680347" watchObservedRunningTime="2025-10-08 19:57:51.337500255 +0000 UTC m=+6416.787343025" Oct 08 19:57:51 crc kubenswrapper[4988]: I1008 19:57:51.454241 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/obo-prometheus-operator-admission-webhook-6b789f9dcf-d7jwn" podStartSLOduration=4.17425725 podStartE2EDuration="6.454223234s" podCreationTimestamp="2025-10-08 19:57:45 +0000 UTC" firstStartedPulling="2025-10-08 19:57:46.288123188 +0000 UTC m=+6411.737965958" lastFinishedPulling="2025-10-08 19:57:48.568089182 +0000 UTC m=+6414.017931942" observedRunningTime="2025-10-08 19:57:51.370201788 +0000 UTC m=+6416.820044578" watchObservedRunningTime="2025-10-08 19:57:51.454223234 +0000 UTC m=+6416.904066004" Oct 08 19:57:52 crc kubenswrapper[4988]: I1008 19:57:52.331471 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-7c8cf85677-pwbvx" event={"ID":"38d89d99-8b9b-43e1-85ad-d928aad3d3d1","Type":"ContainerStarted","Data":"6af6c1cc27161171fd9bfdb4787b9f5d190e1cd27a1f7932f27441deccac43ad"} Oct 08 19:57:52 crc kubenswrapper[4988]: I1008 19:57:52.360789 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/obo-prometheus-operator-7c8cf85677-pwbvx" podStartSLOduration=3.7941582990000002 podStartE2EDuration="8.36077111s" podCreationTimestamp="2025-10-08 19:57:44 +0000 UTC" firstStartedPulling="2025-10-08 19:57:46.10704772 +0000 UTC m=+6411.556890480" lastFinishedPulling="2025-10-08 19:57:50.673660521 +0000 UTC m=+6416.123503291" observedRunningTime="2025-10-08 19:57:52.354099927 +0000 UTC m=+6417.803942707" watchObservedRunningTime="2025-10-08 19:57:52.36077111 +0000 UTC m=+6417.810613870" Oct 08 19:57:53 crc kubenswrapper[4988]: I1008 19:57:53.337800 4988 patch_prober.go:28] interesting pod/machine-config-daemon-nm54f container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 19:57:53 crc kubenswrapper[4988]: I1008 19:57:53.337884 4988 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 19:57:53 crc kubenswrapper[4988]: I1008 19:57:53.337932 4988 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" Oct 08 19:57:53 crc kubenswrapper[4988]: I1008 19:57:53.338982 4988 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"439af713ff941f973b8758bd6af3c097808f76ff520b96ff10bd3268811393f7"} pod="openshift-machine-config-operator/machine-config-daemon-nm54f" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 08 19:57:53 crc kubenswrapper[4988]: I1008 19:57:53.339054 4988 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" containerName="machine-config-daemon" containerID="cri-o://439af713ff941f973b8758bd6af3c097808f76ff520b96ff10bd3268811393f7" gracePeriod=600 Oct 08 19:57:53 crc kubenswrapper[4988]: I1008 19:57:53.495667 4988 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-rrm85" podUID="45843c4e-b4ba-4b4f-b1d9-ceeafb52e794" containerName="registry-server" probeResult="failure" output=< Oct 08 19:57:53 crc kubenswrapper[4988]: timeout: failed to connect service ":50051" within 1s Oct 08 19:57:53 crc kubenswrapper[4988]: > Oct 08 19:57:54 crc kubenswrapper[4988]: E1008 19:57:54.067005 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nm54f_openshift-machine-config-operator(f328a20b-5283-407e-8a93-2ab2da599f18)\"" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" Oct 08 19:57:54 crc kubenswrapper[4988]: I1008 19:57:54.356845 4988 generic.go:334] "Generic (PLEG): container finished" podID="f328a20b-5283-407e-8a93-2ab2da599f18" containerID="439af713ff941f973b8758bd6af3c097808f76ff520b96ff10bd3268811393f7" exitCode=0 Oct 08 19:57:54 crc kubenswrapper[4988]: I1008 19:57:54.356899 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" event={"ID":"f328a20b-5283-407e-8a93-2ab2da599f18","Type":"ContainerDied","Data":"439af713ff941f973b8758bd6af3c097808f76ff520b96ff10bd3268811393f7"} Oct 08 19:57:54 crc kubenswrapper[4988]: I1008 19:57:54.356973 4988 scope.go:117] "RemoveContainer" containerID="07bebacafd1c8af09c722a5994e896b519bd87cf3a710207584b6b2c926371f1" Oct 08 19:57:54 crc kubenswrapper[4988]: I1008 19:57:54.357618 4988 scope.go:117] "RemoveContainer" containerID="439af713ff941f973b8758bd6af3c097808f76ff520b96ff10bd3268811393f7" Oct 08 19:57:54 crc kubenswrapper[4988]: E1008 19:57:54.357943 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nm54f_openshift-machine-config-operator(f328a20b-5283-407e-8a93-2ab2da599f18)\"" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" Oct 08 19:57:55 crc kubenswrapper[4988]: I1008 19:57:55.367633 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/observability-operator-cc5f78dfc-7qjkj" event={"ID":"86cf1f5c-ab13-4b35-ad2f-9529ccdf96db","Type":"ContainerStarted","Data":"b912c4bffca8f2282d02529eee2efe765692bdaec066b4ff04526d09d92e89fd"} Oct 08 19:57:55 crc kubenswrapper[4988]: I1008 19:57:55.368610 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operators/observability-operator-cc5f78dfc-7qjkj" Oct 08 19:57:55 crc kubenswrapper[4988]: I1008 19:57:55.370240 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operators/observability-operator-cc5f78dfc-7qjkj" Oct 08 19:57:55 crc kubenswrapper[4988]: I1008 19:57:55.439861 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/observability-operator-cc5f78dfc-7qjkj" podStartSLOduration=2.549120384 podStartE2EDuration="10.439842158s" podCreationTimestamp="2025-10-08 19:57:45 +0000 UTC" firstStartedPulling="2025-10-08 19:57:46.258127582 +0000 UTC m=+6411.707970352" lastFinishedPulling="2025-10-08 19:57:54.148849356 +0000 UTC m=+6419.598692126" observedRunningTime="2025-10-08 19:57:55.396794116 +0000 UTC m=+6420.846636896" watchObservedRunningTime="2025-10-08 19:57:55.439842158 +0000 UTC m=+6420.889684928" Oct 08 19:57:57 crc kubenswrapper[4988]: I1008 19:57:57.026621 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-create-smwmq"] Oct 08 19:57:57 crc kubenswrapper[4988]: I1008 19:57:57.035899 4988 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-create-smwmq"] Oct 08 19:57:57 crc kubenswrapper[4988]: I1008 19:57:57.248563 4988 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5965ce16-2d47-4658-8429-eb4bcff30a21" path="/var/lib/kubelet/pods/5965ce16-2d47-4658-8429-eb4bcff30a21/volumes" Oct 08 19:58:03 crc kubenswrapper[4988]: I1008 19:58:03.483848 4988 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-rrm85" podUID="45843c4e-b4ba-4b4f-b1d9-ceeafb52e794" containerName="registry-server" probeResult="failure" output=< Oct 08 19:58:03 crc kubenswrapper[4988]: timeout: failed to connect service ":50051" within 1s Oct 08 19:58:03 crc kubenswrapper[4988]: > Oct 08 19:58:05 crc kubenswrapper[4988]: I1008 19:58:05.939242 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operators/perses-operator-54bc95c9fb-ln8dz" Oct 08 19:58:06 crc kubenswrapper[4988]: I1008 19:58:06.238309 4988 scope.go:117] "RemoveContainer" containerID="439af713ff941f973b8758bd6af3c097808f76ff520b96ff10bd3268811393f7" Oct 08 19:58:06 crc kubenswrapper[4988]: E1008 19:58:06.238657 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nm54f_openshift-machine-config-operator(f328a20b-5283-407e-8a93-2ab2da599f18)\"" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" Oct 08 19:58:07 crc kubenswrapper[4988]: I1008 19:58:07.032803 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-16fa-account-create-vxrv2"] Oct 08 19:58:07 crc kubenswrapper[4988]: I1008 19:58:07.042550 4988 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-16fa-account-create-vxrv2"] Oct 08 19:58:07 crc kubenswrapper[4988]: I1008 19:58:07.248754 4988 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5388dded-ae47-41e5-934d-7b196fb9ce79" path="/var/lib/kubelet/pods/5388dded-ae47-41e5-934d-7b196fb9ce79/volumes" Oct 08 19:58:07 crc kubenswrapper[4988]: I1008 19:58:07.689402 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/openstackclient"] Oct 08 19:58:07 crc kubenswrapper[4988]: I1008 19:58:07.689667 4988 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/openstackclient" podUID="cd32ee99-e811-4f6a-8672-3836b7527cc9" containerName="openstackclient" containerID="cri-o://f20ab6daa874b15266bd84f68d85b3b1bb9566ce7795057403cb7229b3424127" gracePeriod=2 Oct 08 19:58:07 crc kubenswrapper[4988]: I1008 19:58:07.700973 4988 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/openstackclient"] Oct 08 19:58:07 crc kubenswrapper[4988]: I1008 19:58:07.732161 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstackclient"] Oct 08 19:58:07 crc kubenswrapper[4988]: E1008 19:58:07.732956 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cd32ee99-e811-4f6a-8672-3836b7527cc9" containerName="openstackclient" Oct 08 19:58:07 crc kubenswrapper[4988]: I1008 19:58:07.733065 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="cd32ee99-e811-4f6a-8672-3836b7527cc9" containerName="openstackclient" Oct 08 19:58:07 crc kubenswrapper[4988]: I1008 19:58:07.733463 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="cd32ee99-e811-4f6a-8672-3836b7527cc9" containerName="openstackclient" Oct 08 19:58:07 crc kubenswrapper[4988]: I1008 19:58:07.734416 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Oct 08 19:58:07 crc kubenswrapper[4988]: I1008 19:58:07.742783 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Oct 08 19:58:07 crc kubenswrapper[4988]: I1008 19:58:07.760748 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/openstackclient"] Oct 08 19:58:07 crc kubenswrapper[4988]: E1008 19:58:07.761724 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="unmounted volumes=[combined-ca-bundle kube-api-access-fblj9 openstack-config openstack-config-secret], unattached volumes=[], failed to process volumes=[]: context canceled" pod="openstack/openstackclient" podUID="cacfef0a-f26f-4f7a-861b-403b4e1fc1f8" Oct 08 19:58:07 crc kubenswrapper[4988]: I1008 19:58:07.784404 4988 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/openstackclient"] Oct 08 19:58:07 crc kubenswrapper[4988]: I1008 19:58:07.792794 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cacfef0a-f26f-4f7a-861b-403b4e1fc1f8-combined-ca-bundle\") pod \"openstackclient\" (UID: \"cacfef0a-f26f-4f7a-861b-403b4e1fc1f8\") " pod="openstack/openstackclient" Oct 08 19:58:07 crc kubenswrapper[4988]: I1008 19:58:07.792843 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/cacfef0a-f26f-4f7a-861b-403b4e1fc1f8-openstack-config-secret\") pod \"openstackclient\" (UID: \"cacfef0a-f26f-4f7a-861b-403b4e1fc1f8\") " pod="openstack/openstackclient" Oct 08 19:58:07 crc kubenswrapper[4988]: I1008 19:58:07.792888 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/cacfef0a-f26f-4f7a-861b-403b4e1fc1f8-openstack-config\") pod \"openstackclient\" (UID: \"cacfef0a-f26f-4f7a-861b-403b4e1fc1f8\") " pod="openstack/openstackclient" Oct 08 19:58:07 crc kubenswrapper[4988]: I1008 19:58:07.792954 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fblj9\" (UniqueName: \"kubernetes.io/projected/cacfef0a-f26f-4f7a-861b-403b4e1fc1f8-kube-api-access-fblj9\") pod \"openstackclient\" (UID: \"cacfef0a-f26f-4f7a-861b-403b4e1fc1f8\") " pod="openstack/openstackclient" Oct 08 19:58:07 crc kubenswrapper[4988]: I1008 19:58:07.799928 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstackclient"] Oct 08 19:58:07 crc kubenswrapper[4988]: I1008 19:58:07.801871 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Oct 08 19:58:07 crc kubenswrapper[4988]: I1008 19:58:07.805764 4988 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openstack/openstackclient" oldPodUID="cd32ee99-e811-4f6a-8672-3836b7527cc9" podUID="2d58f5a2-1a02-4152-95ec-22cc390d8ae5" Oct 08 19:58:07 crc kubenswrapper[4988]: I1008 19:58:07.824104 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Oct 08 19:58:07 crc kubenswrapper[4988]: I1008 19:58:07.894335 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/2d58f5a2-1a02-4152-95ec-22cc390d8ae5-openstack-config\") pod \"openstackclient\" (UID: \"2d58f5a2-1a02-4152-95ec-22cc390d8ae5\") " pod="openstack/openstackclient" Oct 08 19:58:07 crc kubenswrapper[4988]: I1008 19:58:07.894464 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/2d58f5a2-1a02-4152-95ec-22cc390d8ae5-openstack-config-secret\") pod \"openstackclient\" (UID: \"2d58f5a2-1a02-4152-95ec-22cc390d8ae5\") " pod="openstack/openstackclient" Oct 08 19:58:07 crc kubenswrapper[4988]: I1008 19:58:07.894599 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cacfef0a-f26f-4f7a-861b-403b4e1fc1f8-combined-ca-bundle\") pod \"openstackclient\" (UID: \"cacfef0a-f26f-4f7a-861b-403b4e1fc1f8\") " pod="openstack/openstackclient" Oct 08 19:58:07 crc kubenswrapper[4988]: I1008 19:58:07.894638 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/cacfef0a-f26f-4f7a-861b-403b4e1fc1f8-openstack-config-secret\") pod \"openstackclient\" (UID: \"cacfef0a-f26f-4f7a-861b-403b4e1fc1f8\") " pod="openstack/openstackclient" Oct 08 19:58:07 crc kubenswrapper[4988]: I1008 19:58:07.894698 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/cacfef0a-f26f-4f7a-861b-403b4e1fc1f8-openstack-config\") pod \"openstackclient\" (UID: \"cacfef0a-f26f-4f7a-861b-403b4e1fc1f8\") " pod="openstack/openstackclient" Oct 08 19:58:07 crc kubenswrapper[4988]: I1008 19:58:07.894760 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2d58f5a2-1a02-4152-95ec-22cc390d8ae5-combined-ca-bundle\") pod \"openstackclient\" (UID: \"2d58f5a2-1a02-4152-95ec-22cc390d8ae5\") " pod="openstack/openstackclient" Oct 08 19:58:07 crc kubenswrapper[4988]: I1008 19:58:07.894794 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fblj9\" (UniqueName: \"kubernetes.io/projected/cacfef0a-f26f-4f7a-861b-403b4e1fc1f8-kube-api-access-fblj9\") pod \"openstackclient\" (UID: \"cacfef0a-f26f-4f7a-861b-403b4e1fc1f8\") " pod="openstack/openstackclient" Oct 08 19:58:07 crc kubenswrapper[4988]: I1008 19:58:07.894868 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lc784\" (UniqueName: \"kubernetes.io/projected/2d58f5a2-1a02-4152-95ec-22cc390d8ae5-kube-api-access-lc784\") pod \"openstackclient\" (UID: \"2d58f5a2-1a02-4152-95ec-22cc390d8ae5\") " pod="openstack/openstackclient" Oct 08 19:58:07 crc kubenswrapper[4988]: I1008 19:58:07.896660 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/cacfef0a-f26f-4f7a-861b-403b4e1fc1f8-openstack-config\") pod \"openstackclient\" (UID: \"cacfef0a-f26f-4f7a-861b-403b4e1fc1f8\") " pod="openstack/openstackclient" Oct 08 19:58:07 crc kubenswrapper[4988]: E1008 19:58:07.898759 4988 projected.go:194] Error preparing data for projected volume kube-api-access-fblj9 for pod openstack/openstackclient: failed to fetch token: serviceaccounts "openstackclient-openstackclient" is forbidden: the UID in the bound object reference (cacfef0a-f26f-4f7a-861b-403b4e1fc1f8) does not match the UID in record. The object might have been deleted and then recreated Oct 08 19:58:07 crc kubenswrapper[4988]: E1008 19:58:07.898840 4988 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/cacfef0a-f26f-4f7a-861b-403b4e1fc1f8-kube-api-access-fblj9 podName:cacfef0a-f26f-4f7a-861b-403b4e1fc1f8 nodeName:}" failed. No retries permitted until 2025-10-08 19:58:08.398816194 +0000 UTC m=+6433.848659084 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-fblj9" (UniqueName: "kubernetes.io/projected/cacfef0a-f26f-4f7a-861b-403b4e1fc1f8-kube-api-access-fblj9") pod "openstackclient" (UID: "cacfef0a-f26f-4f7a-861b-403b4e1fc1f8") : failed to fetch token: serviceaccounts "openstackclient-openstackclient" is forbidden: the UID in the bound object reference (cacfef0a-f26f-4f7a-861b-403b4e1fc1f8) does not match the UID in record. The object might have been deleted and then recreated Oct 08 19:58:07 crc kubenswrapper[4988]: I1008 19:58:07.901478 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cacfef0a-f26f-4f7a-861b-403b4e1fc1f8-combined-ca-bundle\") pod \"openstackclient\" (UID: \"cacfef0a-f26f-4f7a-861b-403b4e1fc1f8\") " pod="openstack/openstackclient" Oct 08 19:58:07 crc kubenswrapper[4988]: I1008 19:58:07.903842 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/cacfef0a-f26f-4f7a-861b-403b4e1fc1f8-openstack-config-secret\") pod \"openstackclient\" (UID: \"cacfef0a-f26f-4f7a-861b-403b4e1fc1f8\") " pod="openstack/openstackclient" Oct 08 19:58:08 crc kubenswrapper[4988]: I1008 19:58:08.015912 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2d58f5a2-1a02-4152-95ec-22cc390d8ae5-combined-ca-bundle\") pod \"openstackclient\" (UID: \"2d58f5a2-1a02-4152-95ec-22cc390d8ae5\") " pod="openstack/openstackclient" Oct 08 19:58:08 crc kubenswrapper[4988]: I1008 19:58:08.016107 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lc784\" (UniqueName: \"kubernetes.io/projected/2d58f5a2-1a02-4152-95ec-22cc390d8ae5-kube-api-access-lc784\") pod \"openstackclient\" (UID: \"2d58f5a2-1a02-4152-95ec-22cc390d8ae5\") " pod="openstack/openstackclient" Oct 08 19:58:08 crc kubenswrapper[4988]: I1008 19:58:08.016206 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/2d58f5a2-1a02-4152-95ec-22cc390d8ae5-openstack-config\") pod \"openstackclient\" (UID: \"2d58f5a2-1a02-4152-95ec-22cc390d8ae5\") " pod="openstack/openstackclient" Oct 08 19:58:08 crc kubenswrapper[4988]: I1008 19:58:08.016330 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/2d58f5a2-1a02-4152-95ec-22cc390d8ae5-openstack-config-secret\") pod \"openstackclient\" (UID: \"2d58f5a2-1a02-4152-95ec-22cc390d8ae5\") " pod="openstack/openstackclient" Oct 08 19:58:08 crc kubenswrapper[4988]: I1008 19:58:08.021369 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/2d58f5a2-1a02-4152-95ec-22cc390d8ae5-openstack-config\") pod \"openstackclient\" (UID: \"2d58f5a2-1a02-4152-95ec-22cc390d8ae5\") " pod="openstack/openstackclient" Oct 08 19:58:08 crc kubenswrapper[4988]: I1008 19:58:08.042033 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/2d58f5a2-1a02-4152-95ec-22cc390d8ae5-openstack-config-secret\") pod \"openstackclient\" (UID: \"2d58f5a2-1a02-4152-95ec-22cc390d8ae5\") " pod="openstack/openstackclient" Oct 08 19:58:08 crc kubenswrapper[4988]: I1008 19:58:08.047487 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Oct 08 19:58:08 crc kubenswrapper[4988]: I1008 19:58:08.089699 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Oct 08 19:58:08 crc kubenswrapper[4988]: I1008 19:58:08.090849 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lc784\" (UniqueName: \"kubernetes.io/projected/2d58f5a2-1a02-4152-95ec-22cc390d8ae5-kube-api-access-lc784\") pod \"openstackclient\" (UID: \"2d58f5a2-1a02-4152-95ec-22cc390d8ae5\") " pod="openstack/openstackclient" Oct 08 19:58:08 crc kubenswrapper[4988]: I1008 19:58:08.092934 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"telemetry-ceilometer-dockercfg-9z2b4" Oct 08 19:58:08 crc kubenswrapper[4988]: I1008 19:58:08.098206 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 08 19:58:08 crc kubenswrapper[4988]: I1008 19:58:08.104011 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2d58f5a2-1a02-4152-95ec-22cc390d8ae5-combined-ca-bundle\") pod \"openstackclient\" (UID: \"2d58f5a2-1a02-4152-95ec-22cc390d8ae5\") " pod="openstack/openstackclient" Oct 08 19:58:08 crc kubenswrapper[4988]: I1008 19:58:08.134035 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Oct 08 19:58:08 crc kubenswrapper[4988]: I1008 19:58:08.224372 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k4gdm\" (UniqueName: \"kubernetes.io/projected/d9fdb832-bcf7-4e2a-afe0-a01a5881cfb3-kube-api-access-k4gdm\") pod \"kube-state-metrics-0\" (UID: \"d9fdb832-bcf7-4e2a-afe0-a01a5881cfb3\") " pod="openstack/kube-state-metrics-0" Oct 08 19:58:08 crc kubenswrapper[4988]: I1008 19:58:08.326332 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k4gdm\" (UniqueName: \"kubernetes.io/projected/d9fdb832-bcf7-4e2a-afe0-a01a5881cfb3-kube-api-access-k4gdm\") pod \"kube-state-metrics-0\" (UID: \"d9fdb832-bcf7-4e2a-afe0-a01a5881cfb3\") " pod="openstack/kube-state-metrics-0" Oct 08 19:58:08 crc kubenswrapper[4988]: I1008 19:58:08.365827 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k4gdm\" (UniqueName: \"kubernetes.io/projected/d9fdb832-bcf7-4e2a-afe0-a01a5881cfb3-kube-api-access-k4gdm\") pod \"kube-state-metrics-0\" (UID: \"d9fdb832-bcf7-4e2a-afe0-a01a5881cfb3\") " pod="openstack/kube-state-metrics-0" Oct 08 19:58:08 crc kubenswrapper[4988]: I1008 19:58:08.448416 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fblj9\" (UniqueName: \"kubernetes.io/projected/cacfef0a-f26f-4f7a-861b-403b4e1fc1f8-kube-api-access-fblj9\") pod \"openstackclient\" (UID: \"cacfef0a-f26f-4f7a-861b-403b4e1fc1f8\") " pod="openstack/openstackclient" Oct 08 19:58:08 crc kubenswrapper[4988]: E1008 19:58:08.467180 4988 projected.go:194] Error preparing data for projected volume kube-api-access-fblj9 for pod openstack/openstackclient: failed to fetch token: serviceaccounts "openstackclient-openstackclient" is forbidden: the UID in the bound object reference (cacfef0a-f26f-4f7a-861b-403b4e1fc1f8) does not match the UID in record. The object might have been deleted and then recreated Oct 08 19:58:08 crc kubenswrapper[4988]: E1008 19:58:08.467272 4988 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/cacfef0a-f26f-4f7a-861b-403b4e1fc1f8-kube-api-access-fblj9 podName:cacfef0a-f26f-4f7a-861b-403b4e1fc1f8 nodeName:}" failed. No retries permitted until 2025-10-08 19:58:09.467247501 +0000 UTC m=+6434.917090271 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-fblj9" (UniqueName: "kubernetes.io/projected/cacfef0a-f26f-4f7a-861b-403b4e1fc1f8-kube-api-access-fblj9") pod "openstackclient" (UID: "cacfef0a-f26f-4f7a-861b-403b4e1fc1f8") : failed to fetch token: serviceaccounts "openstackclient-openstackclient" is forbidden: the UID in the bound object reference (cacfef0a-f26f-4f7a-861b-403b4e1fc1f8) does not match the UID in record. The object might have been deleted and then recreated Oct 08 19:58:08 crc kubenswrapper[4988]: I1008 19:58:08.467802 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Oct 08 19:58:08 crc kubenswrapper[4988]: I1008 19:58:08.507897 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Oct 08 19:58:08 crc kubenswrapper[4988]: I1008 19:58:08.511440 4988 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openstack/openstackclient" oldPodUID="cacfef0a-f26f-4f7a-861b-403b4e1fc1f8" podUID="2d58f5a2-1a02-4152-95ec-22cc390d8ae5" Oct 08 19:58:08 crc kubenswrapper[4988]: I1008 19:58:08.526154 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Oct 08 19:58:08 crc kubenswrapper[4988]: I1008 19:58:08.652316 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/cacfef0a-f26f-4f7a-861b-403b4e1fc1f8-openstack-config\") pod \"cacfef0a-f26f-4f7a-861b-403b4e1fc1f8\" (UID: \"cacfef0a-f26f-4f7a-861b-403b4e1fc1f8\") " Oct 08 19:58:08 crc kubenswrapper[4988]: I1008 19:58:08.652431 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cacfef0a-f26f-4f7a-861b-403b4e1fc1f8-combined-ca-bundle\") pod \"cacfef0a-f26f-4f7a-861b-403b4e1fc1f8\" (UID: \"cacfef0a-f26f-4f7a-861b-403b4e1fc1f8\") " Oct 08 19:58:08 crc kubenswrapper[4988]: I1008 19:58:08.652502 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/cacfef0a-f26f-4f7a-861b-403b4e1fc1f8-openstack-config-secret\") pod \"cacfef0a-f26f-4f7a-861b-403b4e1fc1f8\" (UID: \"cacfef0a-f26f-4f7a-861b-403b4e1fc1f8\") " Oct 08 19:58:08 crc kubenswrapper[4988]: I1008 19:58:08.652824 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cacfef0a-f26f-4f7a-861b-403b4e1fc1f8-openstack-config" (OuterVolumeSpecName: "openstack-config") pod "cacfef0a-f26f-4f7a-861b-403b4e1fc1f8" (UID: "cacfef0a-f26f-4f7a-861b-403b4e1fc1f8"). InnerVolumeSpecName "openstack-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 19:58:08 crc kubenswrapper[4988]: I1008 19:58:08.653219 4988 reconciler_common.go:293] "Volume detached for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/cacfef0a-f26f-4f7a-861b-403b4e1fc1f8-openstack-config\") on node \"crc\" DevicePath \"\"" Oct 08 19:58:08 crc kubenswrapper[4988]: I1008 19:58:08.653236 4988 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fblj9\" (UniqueName: \"kubernetes.io/projected/cacfef0a-f26f-4f7a-861b-403b4e1fc1f8-kube-api-access-fblj9\") on node \"crc\" DevicePath \"\"" Oct 08 19:58:08 crc kubenswrapper[4988]: I1008 19:58:08.657572 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cacfef0a-f26f-4f7a-861b-403b4e1fc1f8-openstack-config-secret" (OuterVolumeSpecName: "openstack-config-secret") pod "cacfef0a-f26f-4f7a-861b-403b4e1fc1f8" (UID: "cacfef0a-f26f-4f7a-861b-403b4e1fc1f8"). InnerVolumeSpecName "openstack-config-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 19:58:08 crc kubenswrapper[4988]: I1008 19:58:08.659590 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cacfef0a-f26f-4f7a-861b-403b4e1fc1f8-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "cacfef0a-f26f-4f7a-861b-403b4e1fc1f8" (UID: "cacfef0a-f26f-4f7a-861b-403b4e1fc1f8"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 19:58:08 crc kubenswrapper[4988]: I1008 19:58:08.759932 4988 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cacfef0a-f26f-4f7a-861b-403b4e1fc1f8-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 19:58:08 crc kubenswrapper[4988]: I1008 19:58:08.759985 4988 reconciler_common.go:293] "Volume detached for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/cacfef0a-f26f-4f7a-861b-403b4e1fc1f8-openstack-config-secret\") on node \"crc\" DevicePath \"\"" Oct 08 19:58:08 crc kubenswrapper[4988]: I1008 19:58:08.771463 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Oct 08 19:58:09 crc kubenswrapper[4988]: I1008 19:58:09.047878 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 08 19:58:09 crc kubenswrapper[4988]: I1008 19:58:09.081733 4988 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 08 19:58:09 crc kubenswrapper[4988]: I1008 19:58:09.261896 4988 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cacfef0a-f26f-4f7a-861b-403b4e1fc1f8" path="/var/lib/kubelet/pods/cacfef0a-f26f-4f7a-861b-403b4e1fc1f8/volumes" Oct 08 19:58:09 crc kubenswrapper[4988]: I1008 19:58:09.517293 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"2d58f5a2-1a02-4152-95ec-22cc390d8ae5","Type":"ContainerStarted","Data":"b7ea1ae1a08e4188d78ecf340ffa61f27ee87f49871711bb009130301dbdf138"} Oct 08 19:58:09 crc kubenswrapper[4988]: I1008 19:58:09.517337 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"2d58f5a2-1a02-4152-95ec-22cc390d8ae5","Type":"ContainerStarted","Data":"cef585e7d4181047159340e2cd19f2f4e368cc6dee7d5fd45e98d2a86bdc0185"} Oct 08 19:58:09 crc kubenswrapper[4988]: I1008 19:58:09.519108 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"d9fdb832-bcf7-4e2a-afe0-a01a5881cfb3","Type":"ContainerStarted","Data":"46bf7e2db5e29acd04805f971eda135c48ae4c40840889a78ecf852d40b43667"} Oct 08 19:58:09 crc kubenswrapper[4988]: I1008 19:58:09.519118 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Oct 08 19:58:09 crc kubenswrapper[4988]: I1008 19:58:09.532680 4988 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openstack/openstackclient" oldPodUID="cacfef0a-f26f-4f7a-861b-403b4e1fc1f8" podUID="2d58f5a2-1a02-4152-95ec-22cc390d8ae5" Oct 08 19:58:09 crc kubenswrapper[4988]: I1008 19:58:09.536805 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstackclient" podStartSLOduration=2.536786659 podStartE2EDuration="2.536786659s" podCreationTimestamp="2025-10-08 19:58:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 19:58:09.529431354 +0000 UTC m=+6434.979274124" watchObservedRunningTime="2025-10-08 19:58:09.536786659 +0000 UTC m=+6434.986629429" Oct 08 19:58:10 crc kubenswrapper[4988]: I1008 19:58:10.045481 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Oct 08 19:58:10 crc kubenswrapper[4988]: I1008 19:58:10.191821 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/cd32ee99-e811-4f6a-8672-3836b7527cc9-openstack-config-secret\") pod \"cd32ee99-e811-4f6a-8672-3836b7527cc9\" (UID: \"cd32ee99-e811-4f6a-8672-3836b7527cc9\") " Oct 08 19:58:10 crc kubenswrapper[4988]: I1008 19:58:10.191932 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5lw74\" (UniqueName: \"kubernetes.io/projected/cd32ee99-e811-4f6a-8672-3836b7527cc9-kube-api-access-5lw74\") pod \"cd32ee99-e811-4f6a-8672-3836b7527cc9\" (UID: \"cd32ee99-e811-4f6a-8672-3836b7527cc9\") " Oct 08 19:58:10 crc kubenswrapper[4988]: I1008 19:58:10.192050 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/cd32ee99-e811-4f6a-8672-3836b7527cc9-openstack-config\") pod \"cd32ee99-e811-4f6a-8672-3836b7527cc9\" (UID: \"cd32ee99-e811-4f6a-8672-3836b7527cc9\") " Oct 08 19:58:10 crc kubenswrapper[4988]: I1008 19:58:10.192138 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cd32ee99-e811-4f6a-8672-3836b7527cc9-combined-ca-bundle\") pod \"cd32ee99-e811-4f6a-8672-3836b7527cc9\" (UID: \"cd32ee99-e811-4f6a-8672-3836b7527cc9\") " Oct 08 19:58:10 crc kubenswrapper[4988]: I1008 19:58:10.203690 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cd32ee99-e811-4f6a-8672-3836b7527cc9-kube-api-access-5lw74" (OuterVolumeSpecName: "kube-api-access-5lw74") pod "cd32ee99-e811-4f6a-8672-3836b7527cc9" (UID: "cd32ee99-e811-4f6a-8672-3836b7527cc9"). InnerVolumeSpecName "kube-api-access-5lw74". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 19:58:10 crc kubenswrapper[4988]: I1008 19:58:10.231565 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cd32ee99-e811-4f6a-8672-3836b7527cc9-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "cd32ee99-e811-4f6a-8672-3836b7527cc9" (UID: "cd32ee99-e811-4f6a-8672-3836b7527cc9"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 19:58:10 crc kubenswrapper[4988]: I1008 19:58:10.261327 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/alertmanager-metric-storage-0"] Oct 08 19:58:10 crc kubenswrapper[4988]: I1008 19:58:10.263321 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/alertmanager-metric-storage-0" Oct 08 19:58:10 crc kubenswrapper[4988]: I1008 19:58:10.265976 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"alertmanager-metric-storage-tls-assets-0" Oct 08 19:58:10 crc kubenswrapper[4988]: I1008 19:58:10.266266 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"alertmanager-metric-storage-generated" Oct 08 19:58:10 crc kubenswrapper[4988]: I1008 19:58:10.266396 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"alertmanager-metric-storage-web-config" Oct 08 19:58:10 crc kubenswrapper[4988]: I1008 19:58:10.271114 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"metric-storage-alertmanager-dockercfg-5bvvd" Oct 08 19:58:10 crc kubenswrapper[4988]: I1008 19:58:10.273907 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cd32ee99-e811-4f6a-8672-3836b7527cc9-openstack-config" (OuterVolumeSpecName: "openstack-config") pod "cd32ee99-e811-4f6a-8672-3836b7527cc9" (UID: "cd32ee99-e811-4f6a-8672-3836b7527cc9"). InnerVolumeSpecName "openstack-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 19:58:10 crc kubenswrapper[4988]: I1008 19:58:10.279902 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/alertmanager-metric-storage-0"] Oct 08 19:58:10 crc kubenswrapper[4988]: I1008 19:58:10.296151 4988 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5lw74\" (UniqueName: \"kubernetes.io/projected/cd32ee99-e811-4f6a-8672-3836b7527cc9-kube-api-access-5lw74\") on node \"crc\" DevicePath \"\"" Oct 08 19:58:10 crc kubenswrapper[4988]: I1008 19:58:10.296191 4988 reconciler_common.go:293] "Volume detached for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/cd32ee99-e811-4f6a-8672-3836b7527cc9-openstack-config\") on node \"crc\" DevicePath \"\"" Oct 08 19:58:10 crc kubenswrapper[4988]: I1008 19:58:10.296203 4988 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cd32ee99-e811-4f6a-8672-3836b7527cc9-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 19:58:10 crc kubenswrapper[4988]: I1008 19:58:10.320495 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cd32ee99-e811-4f6a-8672-3836b7527cc9-openstack-config-secret" (OuterVolumeSpecName: "openstack-config-secret") pod "cd32ee99-e811-4f6a-8672-3836b7527cc9" (UID: "cd32ee99-e811-4f6a-8672-3836b7527cc9"). InnerVolumeSpecName "openstack-config-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 19:58:10 crc kubenswrapper[4988]: I1008 19:58:10.397508 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5trw7\" (UniqueName: \"kubernetes.io/projected/64157199-c32a-461e-9a59-7d5a782c4838-kube-api-access-5trw7\") pod \"alertmanager-metric-storage-0\" (UID: \"64157199-c32a-461e-9a59-7d5a782c4838\") " pod="openstack/alertmanager-metric-storage-0" Oct 08 19:58:10 crc kubenswrapper[4988]: I1008 19:58:10.398256 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/64157199-c32a-461e-9a59-7d5a782c4838-config-out\") pod \"alertmanager-metric-storage-0\" (UID: \"64157199-c32a-461e-9a59-7d5a782c4838\") " pod="openstack/alertmanager-metric-storage-0" Oct 08 19:58:10 crc kubenswrapper[4988]: I1008 19:58:10.398444 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/64157199-c32a-461e-9a59-7d5a782c4838-tls-assets\") pod \"alertmanager-metric-storage-0\" (UID: \"64157199-c32a-461e-9a59-7d5a782c4838\") " pod="openstack/alertmanager-metric-storage-0" Oct 08 19:58:10 crc kubenswrapper[4988]: I1008 19:58:10.398557 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/64157199-c32a-461e-9a59-7d5a782c4838-web-config\") pod \"alertmanager-metric-storage-0\" (UID: \"64157199-c32a-461e-9a59-7d5a782c4838\") " pod="openstack/alertmanager-metric-storage-0" Oct 08 19:58:10 crc kubenswrapper[4988]: I1008 19:58:10.398712 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/secret/64157199-c32a-461e-9a59-7d5a782c4838-config-volume\") pod \"alertmanager-metric-storage-0\" (UID: \"64157199-c32a-461e-9a59-7d5a782c4838\") " pod="openstack/alertmanager-metric-storage-0" Oct 08 19:58:10 crc kubenswrapper[4988]: I1008 19:58:10.398801 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"alertmanager-metric-storage-db\" (UniqueName: \"kubernetes.io/empty-dir/64157199-c32a-461e-9a59-7d5a782c4838-alertmanager-metric-storage-db\") pod \"alertmanager-metric-storage-0\" (UID: \"64157199-c32a-461e-9a59-7d5a782c4838\") " pod="openstack/alertmanager-metric-storage-0" Oct 08 19:58:10 crc kubenswrapper[4988]: I1008 19:58:10.398917 4988 reconciler_common.go:293] "Volume detached for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/cd32ee99-e811-4f6a-8672-3836b7527cc9-openstack-config-secret\") on node \"crc\" DevicePath \"\"" Oct 08 19:58:10 crc kubenswrapper[4988]: I1008 19:58:10.501133 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/64157199-c32a-461e-9a59-7d5a782c4838-config-out\") pod \"alertmanager-metric-storage-0\" (UID: \"64157199-c32a-461e-9a59-7d5a782c4838\") " pod="openstack/alertmanager-metric-storage-0" Oct 08 19:58:10 crc kubenswrapper[4988]: I1008 19:58:10.501230 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/64157199-c32a-461e-9a59-7d5a782c4838-tls-assets\") pod \"alertmanager-metric-storage-0\" (UID: \"64157199-c32a-461e-9a59-7d5a782c4838\") " pod="openstack/alertmanager-metric-storage-0" Oct 08 19:58:10 crc kubenswrapper[4988]: I1008 19:58:10.501275 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/64157199-c32a-461e-9a59-7d5a782c4838-web-config\") pod \"alertmanager-metric-storage-0\" (UID: \"64157199-c32a-461e-9a59-7d5a782c4838\") " pod="openstack/alertmanager-metric-storage-0" Oct 08 19:58:10 crc kubenswrapper[4988]: I1008 19:58:10.501316 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/secret/64157199-c32a-461e-9a59-7d5a782c4838-config-volume\") pod \"alertmanager-metric-storage-0\" (UID: \"64157199-c32a-461e-9a59-7d5a782c4838\") " pod="openstack/alertmanager-metric-storage-0" Oct 08 19:58:10 crc kubenswrapper[4988]: I1008 19:58:10.501336 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"alertmanager-metric-storage-db\" (UniqueName: \"kubernetes.io/empty-dir/64157199-c32a-461e-9a59-7d5a782c4838-alertmanager-metric-storage-db\") pod \"alertmanager-metric-storage-0\" (UID: \"64157199-c32a-461e-9a59-7d5a782c4838\") " pod="openstack/alertmanager-metric-storage-0" Oct 08 19:58:10 crc kubenswrapper[4988]: I1008 19:58:10.501367 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5trw7\" (UniqueName: \"kubernetes.io/projected/64157199-c32a-461e-9a59-7d5a782c4838-kube-api-access-5trw7\") pod \"alertmanager-metric-storage-0\" (UID: \"64157199-c32a-461e-9a59-7d5a782c4838\") " pod="openstack/alertmanager-metric-storage-0" Oct 08 19:58:10 crc kubenswrapper[4988]: I1008 19:58:10.502063 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"alertmanager-metric-storage-db\" (UniqueName: \"kubernetes.io/empty-dir/64157199-c32a-461e-9a59-7d5a782c4838-alertmanager-metric-storage-db\") pod \"alertmanager-metric-storage-0\" (UID: \"64157199-c32a-461e-9a59-7d5a782c4838\") " pod="openstack/alertmanager-metric-storage-0" Oct 08 19:58:10 crc kubenswrapper[4988]: I1008 19:58:10.506595 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/64157199-c32a-461e-9a59-7d5a782c4838-config-out\") pod \"alertmanager-metric-storage-0\" (UID: \"64157199-c32a-461e-9a59-7d5a782c4838\") " pod="openstack/alertmanager-metric-storage-0" Oct 08 19:58:10 crc kubenswrapper[4988]: I1008 19:58:10.507355 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/secret/64157199-c32a-461e-9a59-7d5a782c4838-config-volume\") pod \"alertmanager-metric-storage-0\" (UID: \"64157199-c32a-461e-9a59-7d5a782c4838\") " pod="openstack/alertmanager-metric-storage-0" Oct 08 19:58:10 crc kubenswrapper[4988]: I1008 19:58:10.507853 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/64157199-c32a-461e-9a59-7d5a782c4838-web-config\") pod \"alertmanager-metric-storage-0\" (UID: \"64157199-c32a-461e-9a59-7d5a782c4838\") " pod="openstack/alertmanager-metric-storage-0" Oct 08 19:58:10 crc kubenswrapper[4988]: I1008 19:58:10.508870 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/64157199-c32a-461e-9a59-7d5a782c4838-tls-assets\") pod \"alertmanager-metric-storage-0\" (UID: \"64157199-c32a-461e-9a59-7d5a782c4838\") " pod="openstack/alertmanager-metric-storage-0" Oct 08 19:58:10 crc kubenswrapper[4988]: I1008 19:58:10.527722 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5trw7\" (UniqueName: \"kubernetes.io/projected/64157199-c32a-461e-9a59-7d5a782c4838-kube-api-access-5trw7\") pod \"alertmanager-metric-storage-0\" (UID: \"64157199-c32a-461e-9a59-7d5a782c4838\") " pod="openstack/alertmanager-metric-storage-0" Oct 08 19:58:10 crc kubenswrapper[4988]: I1008 19:58:10.529496 4988 generic.go:334] "Generic (PLEG): container finished" podID="cd32ee99-e811-4f6a-8672-3836b7527cc9" containerID="f20ab6daa874b15266bd84f68d85b3b1bb9566ce7795057403cb7229b3424127" exitCode=137 Oct 08 19:58:10 crc kubenswrapper[4988]: I1008 19:58:10.529551 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Oct 08 19:58:10 crc kubenswrapper[4988]: I1008 19:58:10.529588 4988 scope.go:117] "RemoveContainer" containerID="f20ab6daa874b15266bd84f68d85b3b1bb9566ce7795057403cb7229b3424127" Oct 08 19:58:10 crc kubenswrapper[4988]: I1008 19:58:10.532006 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"d9fdb832-bcf7-4e2a-afe0-a01a5881cfb3","Type":"ContainerStarted","Data":"d526c948e42b32ee91470150e213b56a919cb4fc6e9b6480ba2a38e3d49dd02e"} Oct 08 19:58:10 crc kubenswrapper[4988]: I1008 19:58:10.553818 4988 scope.go:117] "RemoveContainer" containerID="f20ab6daa874b15266bd84f68d85b3b1bb9566ce7795057403cb7229b3424127" Oct 08 19:58:10 crc kubenswrapper[4988]: E1008 19:58:10.554263 4988 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f20ab6daa874b15266bd84f68d85b3b1bb9566ce7795057403cb7229b3424127\": container with ID starting with f20ab6daa874b15266bd84f68d85b3b1bb9566ce7795057403cb7229b3424127 not found: ID does not exist" containerID="f20ab6daa874b15266bd84f68d85b3b1bb9566ce7795057403cb7229b3424127" Oct 08 19:58:10 crc kubenswrapper[4988]: I1008 19:58:10.554305 4988 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f20ab6daa874b15266bd84f68d85b3b1bb9566ce7795057403cb7229b3424127"} err="failed to get container status \"f20ab6daa874b15266bd84f68d85b3b1bb9566ce7795057403cb7229b3424127\": rpc error: code = NotFound desc = could not find container \"f20ab6daa874b15266bd84f68d85b3b1bb9566ce7795057403cb7229b3424127\": container with ID starting with f20ab6daa874b15266bd84f68d85b3b1bb9566ce7795057403cb7229b3424127 not found: ID does not exist" Oct 08 19:58:10 crc kubenswrapper[4988]: I1008 19:58:10.554813 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=3.059630723 podStartE2EDuration="3.554803335s" podCreationTimestamp="2025-10-08 19:58:07 +0000 UTC" firstStartedPulling="2025-10-08 19:58:09.081530747 +0000 UTC m=+6434.531373517" lastFinishedPulling="2025-10-08 19:58:09.576703359 +0000 UTC m=+6435.026546129" observedRunningTime="2025-10-08 19:58:10.554280739 +0000 UTC m=+6436.004123509" watchObservedRunningTime="2025-10-08 19:58:10.554803335 +0000 UTC m=+6436.004646105" Oct 08 19:58:10 crc kubenswrapper[4988]: I1008 19:58:10.560410 4988 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openstack/openstackclient" oldPodUID="cd32ee99-e811-4f6a-8672-3836b7527cc9" podUID="2d58f5a2-1a02-4152-95ec-22cc390d8ae5" Oct 08 19:58:10 crc kubenswrapper[4988]: I1008 19:58:10.659448 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/alertmanager-metric-storage-0" Oct 08 19:58:11 crc kubenswrapper[4988]: I1008 19:58:11.251096 4988 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cd32ee99-e811-4f6a-8672-3836b7527cc9" path="/var/lib/kubelet/pods/cd32ee99-e811-4f6a-8672-3836b7527cc9/volumes" Oct 08 19:58:11 crc kubenswrapper[4988]: W1008 19:58:11.313994 4988 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod64157199_c32a_461e_9a59_7d5a782c4838.slice/crio-a69c3dc9f6383bab07ca0479f0f7a9f5fa90662132a7ba8b05a31cc42697066b WatchSource:0}: Error finding container a69c3dc9f6383bab07ca0479f0f7a9f5fa90662132a7ba8b05a31cc42697066b: Status 404 returned error can't find the container with id a69c3dc9f6383bab07ca0479f0f7a9f5fa90662132a7ba8b05a31cc42697066b Oct 08 19:58:11 crc kubenswrapper[4988]: I1008 19:58:11.315014 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/alertmanager-metric-storage-0"] Oct 08 19:58:11 crc kubenswrapper[4988]: I1008 19:58:11.542015 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/alertmanager-metric-storage-0" event={"ID":"64157199-c32a-461e-9a59-7d5a782c4838","Type":"ContainerStarted","Data":"a69c3dc9f6383bab07ca0479f0f7a9f5fa90662132a7ba8b05a31cc42697066b"} Oct 08 19:58:11 crc kubenswrapper[4988]: I1008 19:58:11.543432 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Oct 08 19:58:11 crc kubenswrapper[4988]: I1008 19:58:11.547362 4988 scope.go:117] "RemoveContainer" containerID="4fafc9509e524fbea5bed90e036340c69e0b654aeab3cdda3204b4864cd69ba7" Oct 08 19:58:11 crc kubenswrapper[4988]: I1008 19:58:11.575273 4988 scope.go:117] "RemoveContainer" containerID="d7106d9f01fda802281545842274b315267e733b064e24d743fcae62e7dd0bde" Oct 08 19:58:11 crc kubenswrapper[4988]: I1008 19:58:11.649404 4988 scope.go:117] "RemoveContainer" containerID="2d47400cbad5a6a81cdb1dd2bc01f4bcff2cf43f37c997940b5f61fdee846285" Oct 08 19:58:11 crc kubenswrapper[4988]: I1008 19:58:11.672901 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/prometheus-metric-storage-0"] Oct 08 19:58:11 crc kubenswrapper[4988]: I1008 19:58:11.675368 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Oct 08 19:58:11 crc kubenswrapper[4988]: I1008 19:58:11.683204 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-web-config" Oct 08 19:58:11 crc kubenswrapper[4988]: I1008 19:58:11.683563 4988 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"prometheus-metric-storage-rulefiles-0" Oct 08 19:58:11 crc kubenswrapper[4988]: I1008 19:58:11.683725 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-thanos-prometheus-http-client-file" Oct 08 19:58:11 crc kubenswrapper[4988]: I1008 19:58:11.683871 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage" Oct 08 19:58:11 crc kubenswrapper[4988]: I1008 19:58:11.684073 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"metric-storage-prometheus-dockercfg-fhmvc" Oct 08 19:58:11 crc kubenswrapper[4988]: I1008 19:58:11.685469 4988 scope.go:117] "RemoveContainer" containerID="3c4b8767c0a6bc569abb1d4ca4d565a13e4a771dbf0b8eedc8117cf869c70d65" Oct 08 19:58:11 crc kubenswrapper[4988]: I1008 19:58:11.688273 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/prometheus-metric-storage-0"] Oct 08 19:58:11 crc kubenswrapper[4988]: I1008 19:58:11.699215 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-tls-assets-0" Oct 08 19:58:11 crc kubenswrapper[4988]: I1008 19:58:11.742587 4988 scope.go:117] "RemoveContainer" containerID="f9fadee6bd841947a3cdc2d9f5693e3a99900cf16e24e4936647bdb1faa035ba" Oct 08 19:58:11 crc kubenswrapper[4988]: I1008 19:58:11.932741 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/5cd96092-93f6-43a6-bec4-cceaba7bb8af-config\") pod \"prometheus-metric-storage-0\" (UID: \"5cd96092-93f6-43a6-bec4-cceaba7bb8af\") " pod="openstack/prometheus-metric-storage-0" Oct 08 19:58:11 crc kubenswrapper[4988]: I1008 19:58:11.932799 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/5cd96092-93f6-43a6-bec4-cceaba7bb8af-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"5cd96092-93f6-43a6-bec4-cceaba7bb8af\") " pod="openstack/prometheus-metric-storage-0" Oct 08 19:58:11 crc kubenswrapper[4988]: I1008 19:58:11.932820 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zmzrj\" (UniqueName: \"kubernetes.io/projected/5cd96092-93f6-43a6-bec4-cceaba7bb8af-kube-api-access-zmzrj\") pod \"prometheus-metric-storage-0\" (UID: \"5cd96092-93f6-43a6-bec4-cceaba7bb8af\") " pod="openstack/prometheus-metric-storage-0" Oct 08 19:58:11 crc kubenswrapper[4988]: I1008 19:58:11.932852 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/5cd96092-93f6-43a6-bec4-cceaba7bb8af-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"5cd96092-93f6-43a6-bec4-cceaba7bb8af\") " pod="openstack/prometheus-metric-storage-0" Oct 08 19:58:11 crc kubenswrapper[4988]: I1008 19:58:11.932886 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/5cd96092-93f6-43a6-bec4-cceaba7bb8af-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"5cd96092-93f6-43a6-bec4-cceaba7bb8af\") " pod="openstack/prometheus-metric-storage-0" Oct 08 19:58:11 crc kubenswrapper[4988]: I1008 19:58:11.932953 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/5cd96092-93f6-43a6-bec4-cceaba7bb8af-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"5cd96092-93f6-43a6-bec4-cceaba7bb8af\") " pod="openstack/prometheus-metric-storage-0" Oct 08 19:58:11 crc kubenswrapper[4988]: I1008 19:58:11.932973 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-e634a8ba-6c66-4bab-9b9c-79158ccf2cf8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-e634a8ba-6c66-4bab-9b9c-79158ccf2cf8\") pod \"prometheus-metric-storage-0\" (UID: \"5cd96092-93f6-43a6-bec4-cceaba7bb8af\") " pod="openstack/prometheus-metric-storage-0" Oct 08 19:58:11 crc kubenswrapper[4988]: I1008 19:58:11.933033 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/5cd96092-93f6-43a6-bec4-cceaba7bb8af-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"5cd96092-93f6-43a6-bec4-cceaba7bb8af\") " pod="openstack/prometheus-metric-storage-0" Oct 08 19:58:12 crc kubenswrapper[4988]: I1008 19:58:12.034587 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/5cd96092-93f6-43a6-bec4-cceaba7bb8af-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"5cd96092-93f6-43a6-bec4-cceaba7bb8af\") " pod="openstack/prometheus-metric-storage-0" Oct 08 19:58:12 crc kubenswrapper[4988]: I1008 19:58:12.034636 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-e634a8ba-6c66-4bab-9b9c-79158ccf2cf8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-e634a8ba-6c66-4bab-9b9c-79158ccf2cf8\") pod \"prometheus-metric-storage-0\" (UID: \"5cd96092-93f6-43a6-bec4-cceaba7bb8af\") " pod="openstack/prometheus-metric-storage-0" Oct 08 19:58:12 crc kubenswrapper[4988]: I1008 19:58:12.034693 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/5cd96092-93f6-43a6-bec4-cceaba7bb8af-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"5cd96092-93f6-43a6-bec4-cceaba7bb8af\") " pod="openstack/prometheus-metric-storage-0" Oct 08 19:58:12 crc kubenswrapper[4988]: I1008 19:58:12.034752 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/5cd96092-93f6-43a6-bec4-cceaba7bb8af-config\") pod \"prometheus-metric-storage-0\" (UID: \"5cd96092-93f6-43a6-bec4-cceaba7bb8af\") " pod="openstack/prometheus-metric-storage-0" Oct 08 19:58:12 crc kubenswrapper[4988]: I1008 19:58:12.034783 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/5cd96092-93f6-43a6-bec4-cceaba7bb8af-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"5cd96092-93f6-43a6-bec4-cceaba7bb8af\") " pod="openstack/prometheus-metric-storage-0" Oct 08 19:58:12 crc kubenswrapper[4988]: I1008 19:58:12.034805 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zmzrj\" (UniqueName: \"kubernetes.io/projected/5cd96092-93f6-43a6-bec4-cceaba7bb8af-kube-api-access-zmzrj\") pod \"prometheus-metric-storage-0\" (UID: \"5cd96092-93f6-43a6-bec4-cceaba7bb8af\") " pod="openstack/prometheus-metric-storage-0" Oct 08 19:58:12 crc kubenswrapper[4988]: I1008 19:58:12.034834 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/5cd96092-93f6-43a6-bec4-cceaba7bb8af-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"5cd96092-93f6-43a6-bec4-cceaba7bb8af\") " pod="openstack/prometheus-metric-storage-0" Oct 08 19:58:12 crc kubenswrapper[4988]: I1008 19:58:12.034860 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/5cd96092-93f6-43a6-bec4-cceaba7bb8af-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"5cd96092-93f6-43a6-bec4-cceaba7bb8af\") " pod="openstack/prometheus-metric-storage-0" Oct 08 19:58:12 crc kubenswrapper[4988]: I1008 19:58:12.052773 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/5cd96092-93f6-43a6-bec4-cceaba7bb8af-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"5cd96092-93f6-43a6-bec4-cceaba7bb8af\") " pod="openstack/prometheus-metric-storage-0" Oct 08 19:58:12 crc kubenswrapper[4988]: I1008 19:58:12.065588 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/5cd96092-93f6-43a6-bec4-cceaba7bb8af-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"5cd96092-93f6-43a6-bec4-cceaba7bb8af\") " pod="openstack/prometheus-metric-storage-0" Oct 08 19:58:12 crc kubenswrapper[4988]: I1008 19:58:12.066096 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/5cd96092-93f6-43a6-bec4-cceaba7bb8af-config\") pod \"prometheus-metric-storage-0\" (UID: \"5cd96092-93f6-43a6-bec4-cceaba7bb8af\") " pod="openstack/prometheus-metric-storage-0" Oct 08 19:58:12 crc kubenswrapper[4988]: I1008 19:58:12.066282 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/5cd96092-93f6-43a6-bec4-cceaba7bb8af-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"5cd96092-93f6-43a6-bec4-cceaba7bb8af\") " pod="openstack/prometheus-metric-storage-0" Oct 08 19:58:12 crc kubenswrapper[4988]: I1008 19:58:12.066931 4988 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Oct 08 19:58:12 crc kubenswrapper[4988]: I1008 19:58:12.066959 4988 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-e634a8ba-6c66-4bab-9b9c-79158ccf2cf8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-e634a8ba-6c66-4bab-9b9c-79158ccf2cf8\") pod \"prometheus-metric-storage-0\" (UID: \"5cd96092-93f6-43a6-bec4-cceaba7bb8af\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/36bee47a16e8fb587b88b775bafd572b359ef09bb7992247ae065de916eddc2b/globalmount\"" pod="openstack/prometheus-metric-storage-0" Oct 08 19:58:12 crc kubenswrapper[4988]: I1008 19:58:12.073026 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/5cd96092-93f6-43a6-bec4-cceaba7bb8af-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"5cd96092-93f6-43a6-bec4-cceaba7bb8af\") " pod="openstack/prometheus-metric-storage-0" Oct 08 19:58:12 crc kubenswrapper[4988]: I1008 19:58:12.083750 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/5cd96092-93f6-43a6-bec4-cceaba7bb8af-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"5cd96092-93f6-43a6-bec4-cceaba7bb8af\") " pod="openstack/prometheus-metric-storage-0" Oct 08 19:58:12 crc kubenswrapper[4988]: I1008 19:58:12.100156 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zmzrj\" (UniqueName: \"kubernetes.io/projected/5cd96092-93f6-43a6-bec4-cceaba7bb8af-kube-api-access-zmzrj\") pod \"prometheus-metric-storage-0\" (UID: \"5cd96092-93f6-43a6-bec4-cceaba7bb8af\") " pod="openstack/prometheus-metric-storage-0" Oct 08 19:58:12 crc kubenswrapper[4988]: I1008 19:58:12.208947 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-e634a8ba-6c66-4bab-9b9c-79158ccf2cf8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-e634a8ba-6c66-4bab-9b9c-79158ccf2cf8\") pod \"prometheus-metric-storage-0\" (UID: \"5cd96092-93f6-43a6-bec4-cceaba7bb8af\") " pod="openstack/prometheus-metric-storage-0" Oct 08 19:58:12 crc kubenswrapper[4988]: I1008 19:58:12.344026 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Oct 08 19:58:12 crc kubenswrapper[4988]: I1008 19:58:12.527347 4988 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-rrm85" Oct 08 19:58:12 crc kubenswrapper[4988]: I1008 19:58:12.601550 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-rrm85" Oct 08 19:58:12 crc kubenswrapper[4988]: I1008 19:58:12.969843 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/prometheus-metric-storage-0"] Oct 08 19:58:12 crc kubenswrapper[4988]: W1008 19:58:12.978111 4988 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5cd96092_93f6_43a6_bec4_cceaba7bb8af.slice/crio-de94102c766711eedd4c2ecd78f0b44c44a097ab60e33aa977f617bfd8afe66c WatchSource:0}: Error finding container de94102c766711eedd4c2ecd78f0b44c44a097ab60e33aa977f617bfd8afe66c: Status 404 returned error can't find the container with id de94102c766711eedd4c2ecd78f0b44c44a097ab60e33aa977f617bfd8afe66c Oct 08 19:58:13 crc kubenswrapper[4988]: I1008 19:58:13.043742 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-rrm85"] Oct 08 19:58:13 crc kubenswrapper[4988]: I1008 19:58:13.064179 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-sync-hsz6g"] Oct 08 19:58:13 crc kubenswrapper[4988]: I1008 19:58:13.074579 4988 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-sync-hsz6g"] Oct 08 19:58:13 crc kubenswrapper[4988]: I1008 19:58:13.248629 4988 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cc2cdc3d-5598-4c70-b47a-790abb8ecd2d" path="/var/lib/kubelet/pods/cc2cdc3d-5598-4c70-b47a-790abb8ecd2d/volumes" Oct 08 19:58:13 crc kubenswrapper[4988]: I1008 19:58:13.596410 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"5cd96092-93f6-43a6-bec4-cceaba7bb8af","Type":"ContainerStarted","Data":"de94102c766711eedd4c2ecd78f0b44c44a097ab60e33aa977f617bfd8afe66c"} Oct 08 19:58:13 crc kubenswrapper[4988]: I1008 19:58:13.596865 4988 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-rrm85" podUID="45843c4e-b4ba-4b4f-b1d9-ceeafb52e794" containerName="registry-server" containerID="cri-o://d620feb5bdb67de79dc31481e4abe24b19e6e7779e0417024d4655eb274bba06" gracePeriod=2 Oct 08 19:58:14 crc kubenswrapper[4988]: I1008 19:58:14.204242 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-rrm85" Oct 08 19:58:14 crc kubenswrapper[4988]: I1008 19:58:14.295859 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/45843c4e-b4ba-4b4f-b1d9-ceeafb52e794-catalog-content\") pod \"45843c4e-b4ba-4b4f-b1d9-ceeafb52e794\" (UID: \"45843c4e-b4ba-4b4f-b1d9-ceeafb52e794\") " Oct 08 19:58:14 crc kubenswrapper[4988]: I1008 19:58:14.296028 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6zzc4\" (UniqueName: \"kubernetes.io/projected/45843c4e-b4ba-4b4f-b1d9-ceeafb52e794-kube-api-access-6zzc4\") pod \"45843c4e-b4ba-4b4f-b1d9-ceeafb52e794\" (UID: \"45843c4e-b4ba-4b4f-b1d9-ceeafb52e794\") " Oct 08 19:58:14 crc kubenswrapper[4988]: I1008 19:58:14.296055 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/45843c4e-b4ba-4b4f-b1d9-ceeafb52e794-utilities\") pod \"45843c4e-b4ba-4b4f-b1d9-ceeafb52e794\" (UID: \"45843c4e-b4ba-4b4f-b1d9-ceeafb52e794\") " Oct 08 19:58:14 crc kubenswrapper[4988]: I1008 19:58:14.296841 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/45843c4e-b4ba-4b4f-b1d9-ceeafb52e794-utilities" (OuterVolumeSpecName: "utilities") pod "45843c4e-b4ba-4b4f-b1d9-ceeafb52e794" (UID: "45843c4e-b4ba-4b4f-b1d9-ceeafb52e794"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 19:58:14 crc kubenswrapper[4988]: I1008 19:58:14.301241 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/45843c4e-b4ba-4b4f-b1d9-ceeafb52e794-kube-api-access-6zzc4" (OuterVolumeSpecName: "kube-api-access-6zzc4") pod "45843c4e-b4ba-4b4f-b1d9-ceeafb52e794" (UID: "45843c4e-b4ba-4b4f-b1d9-ceeafb52e794"). InnerVolumeSpecName "kube-api-access-6zzc4". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 19:58:14 crc kubenswrapper[4988]: I1008 19:58:14.366298 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/45843c4e-b4ba-4b4f-b1d9-ceeafb52e794-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "45843c4e-b4ba-4b4f-b1d9-ceeafb52e794" (UID: "45843c4e-b4ba-4b4f-b1d9-ceeafb52e794"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 19:58:14 crc kubenswrapper[4988]: I1008 19:58:14.398005 4988 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/45843c4e-b4ba-4b4f-b1d9-ceeafb52e794-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 08 19:58:14 crc kubenswrapper[4988]: I1008 19:58:14.398037 4988 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6zzc4\" (UniqueName: \"kubernetes.io/projected/45843c4e-b4ba-4b4f-b1d9-ceeafb52e794-kube-api-access-6zzc4\") on node \"crc\" DevicePath \"\"" Oct 08 19:58:14 crc kubenswrapper[4988]: I1008 19:58:14.398047 4988 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/45843c4e-b4ba-4b4f-b1d9-ceeafb52e794-utilities\") on node \"crc\" DevicePath \"\"" Oct 08 19:58:14 crc kubenswrapper[4988]: I1008 19:58:14.661830 4988 generic.go:334] "Generic (PLEG): container finished" podID="45843c4e-b4ba-4b4f-b1d9-ceeafb52e794" containerID="d620feb5bdb67de79dc31481e4abe24b19e6e7779e0417024d4655eb274bba06" exitCode=0 Oct 08 19:58:14 crc kubenswrapper[4988]: I1008 19:58:14.662136 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-rrm85" Oct 08 19:58:14 crc kubenswrapper[4988]: I1008 19:58:14.662133 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rrm85" event={"ID":"45843c4e-b4ba-4b4f-b1d9-ceeafb52e794","Type":"ContainerDied","Data":"d620feb5bdb67de79dc31481e4abe24b19e6e7779e0417024d4655eb274bba06"} Oct 08 19:58:14 crc kubenswrapper[4988]: I1008 19:58:14.664508 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rrm85" event={"ID":"45843c4e-b4ba-4b4f-b1d9-ceeafb52e794","Type":"ContainerDied","Data":"f5e3a7a148fd5096b9e2cf8261a6ba9764d6f911839c5ef7179d7bdba26a54e8"} Oct 08 19:58:14 crc kubenswrapper[4988]: I1008 19:58:14.664564 4988 scope.go:117] "RemoveContainer" containerID="d620feb5bdb67de79dc31481e4abe24b19e6e7779e0417024d4655eb274bba06" Oct 08 19:58:14 crc kubenswrapper[4988]: I1008 19:58:14.726319 4988 scope.go:117] "RemoveContainer" containerID="f4b64dc676d09b73bd0e5e7c83afa900e5dc94774ba7d3446540e7650d58659b" Oct 08 19:58:14 crc kubenswrapper[4988]: I1008 19:58:14.770263 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-rrm85"] Oct 08 19:58:14 crc kubenswrapper[4988]: I1008 19:58:14.783927 4988 scope.go:117] "RemoveContainer" containerID="6881bff31c8fd6399e7581d8a3047ad71ff216cdc5694e04cd78641657858591" Oct 08 19:58:14 crc kubenswrapper[4988]: I1008 19:58:14.795604 4988 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-rrm85"] Oct 08 19:58:14 crc kubenswrapper[4988]: I1008 19:58:14.803448 4988 scope.go:117] "RemoveContainer" containerID="d620feb5bdb67de79dc31481e4abe24b19e6e7779e0417024d4655eb274bba06" Oct 08 19:58:14 crc kubenswrapper[4988]: E1008 19:58:14.803916 4988 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d620feb5bdb67de79dc31481e4abe24b19e6e7779e0417024d4655eb274bba06\": container with ID starting with d620feb5bdb67de79dc31481e4abe24b19e6e7779e0417024d4655eb274bba06 not found: ID does not exist" containerID="d620feb5bdb67de79dc31481e4abe24b19e6e7779e0417024d4655eb274bba06" Oct 08 19:58:14 crc kubenswrapper[4988]: I1008 19:58:14.803969 4988 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d620feb5bdb67de79dc31481e4abe24b19e6e7779e0417024d4655eb274bba06"} err="failed to get container status \"d620feb5bdb67de79dc31481e4abe24b19e6e7779e0417024d4655eb274bba06\": rpc error: code = NotFound desc = could not find container \"d620feb5bdb67de79dc31481e4abe24b19e6e7779e0417024d4655eb274bba06\": container with ID starting with d620feb5bdb67de79dc31481e4abe24b19e6e7779e0417024d4655eb274bba06 not found: ID does not exist" Oct 08 19:58:14 crc kubenswrapper[4988]: I1008 19:58:14.804001 4988 scope.go:117] "RemoveContainer" containerID="f4b64dc676d09b73bd0e5e7c83afa900e5dc94774ba7d3446540e7650d58659b" Oct 08 19:58:14 crc kubenswrapper[4988]: E1008 19:58:14.806356 4988 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f4b64dc676d09b73bd0e5e7c83afa900e5dc94774ba7d3446540e7650d58659b\": container with ID starting with f4b64dc676d09b73bd0e5e7c83afa900e5dc94774ba7d3446540e7650d58659b not found: ID does not exist" containerID="f4b64dc676d09b73bd0e5e7c83afa900e5dc94774ba7d3446540e7650d58659b" Oct 08 19:58:14 crc kubenswrapper[4988]: I1008 19:58:14.806416 4988 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f4b64dc676d09b73bd0e5e7c83afa900e5dc94774ba7d3446540e7650d58659b"} err="failed to get container status \"f4b64dc676d09b73bd0e5e7c83afa900e5dc94774ba7d3446540e7650d58659b\": rpc error: code = NotFound desc = could not find container \"f4b64dc676d09b73bd0e5e7c83afa900e5dc94774ba7d3446540e7650d58659b\": container with ID starting with f4b64dc676d09b73bd0e5e7c83afa900e5dc94774ba7d3446540e7650d58659b not found: ID does not exist" Oct 08 19:58:14 crc kubenswrapper[4988]: I1008 19:58:14.806443 4988 scope.go:117] "RemoveContainer" containerID="6881bff31c8fd6399e7581d8a3047ad71ff216cdc5694e04cd78641657858591" Oct 08 19:58:14 crc kubenswrapper[4988]: E1008 19:58:14.806701 4988 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6881bff31c8fd6399e7581d8a3047ad71ff216cdc5694e04cd78641657858591\": container with ID starting with 6881bff31c8fd6399e7581d8a3047ad71ff216cdc5694e04cd78641657858591 not found: ID does not exist" containerID="6881bff31c8fd6399e7581d8a3047ad71ff216cdc5694e04cd78641657858591" Oct 08 19:58:14 crc kubenswrapper[4988]: I1008 19:58:14.806727 4988 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6881bff31c8fd6399e7581d8a3047ad71ff216cdc5694e04cd78641657858591"} err="failed to get container status \"6881bff31c8fd6399e7581d8a3047ad71ff216cdc5694e04cd78641657858591\": rpc error: code = NotFound desc = could not find container \"6881bff31c8fd6399e7581d8a3047ad71ff216cdc5694e04cd78641657858591\": container with ID starting with 6881bff31c8fd6399e7581d8a3047ad71ff216cdc5694e04cd78641657858591 not found: ID does not exist" Oct 08 19:58:15 crc kubenswrapper[4988]: I1008 19:58:15.248701 4988 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="45843c4e-b4ba-4b4f-b1d9-ceeafb52e794" path="/var/lib/kubelet/pods/45843c4e-b4ba-4b4f-b1d9-ceeafb52e794/volumes" Oct 08 19:58:16 crc kubenswrapper[4988]: I1008 19:58:16.671300 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-xhzzn"] Oct 08 19:58:16 crc kubenswrapper[4988]: E1008 19:58:16.672757 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="45843c4e-b4ba-4b4f-b1d9-ceeafb52e794" containerName="extract-content" Oct 08 19:58:16 crc kubenswrapper[4988]: I1008 19:58:16.672781 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="45843c4e-b4ba-4b4f-b1d9-ceeafb52e794" containerName="extract-content" Oct 08 19:58:16 crc kubenswrapper[4988]: E1008 19:58:16.672823 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="45843c4e-b4ba-4b4f-b1d9-ceeafb52e794" containerName="extract-utilities" Oct 08 19:58:16 crc kubenswrapper[4988]: I1008 19:58:16.672836 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="45843c4e-b4ba-4b4f-b1d9-ceeafb52e794" containerName="extract-utilities" Oct 08 19:58:16 crc kubenswrapper[4988]: E1008 19:58:16.672853 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="45843c4e-b4ba-4b4f-b1d9-ceeafb52e794" containerName="registry-server" Oct 08 19:58:16 crc kubenswrapper[4988]: I1008 19:58:16.672865 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="45843c4e-b4ba-4b4f-b1d9-ceeafb52e794" containerName="registry-server" Oct 08 19:58:16 crc kubenswrapper[4988]: I1008 19:58:16.673262 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="45843c4e-b4ba-4b4f-b1d9-ceeafb52e794" containerName="registry-server" Oct 08 19:58:16 crc kubenswrapper[4988]: I1008 19:58:16.676102 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-xhzzn" Oct 08 19:58:16 crc kubenswrapper[4988]: I1008 19:58:16.703934 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-xhzzn"] Oct 08 19:58:16 crc kubenswrapper[4988]: I1008 19:58:16.849988 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/70948643-9439-4c7d-ac9d-6cee3d7d58e0-catalog-content\") pod \"redhat-marketplace-xhzzn\" (UID: \"70948643-9439-4c7d-ac9d-6cee3d7d58e0\") " pod="openshift-marketplace/redhat-marketplace-xhzzn" Oct 08 19:58:16 crc kubenswrapper[4988]: I1008 19:58:16.850139 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/70948643-9439-4c7d-ac9d-6cee3d7d58e0-utilities\") pod \"redhat-marketplace-xhzzn\" (UID: \"70948643-9439-4c7d-ac9d-6cee3d7d58e0\") " pod="openshift-marketplace/redhat-marketplace-xhzzn" Oct 08 19:58:16 crc kubenswrapper[4988]: I1008 19:58:16.850179 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tdh2t\" (UniqueName: \"kubernetes.io/projected/70948643-9439-4c7d-ac9d-6cee3d7d58e0-kube-api-access-tdh2t\") pod \"redhat-marketplace-xhzzn\" (UID: \"70948643-9439-4c7d-ac9d-6cee3d7d58e0\") " pod="openshift-marketplace/redhat-marketplace-xhzzn" Oct 08 19:58:16 crc kubenswrapper[4988]: I1008 19:58:16.952149 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/70948643-9439-4c7d-ac9d-6cee3d7d58e0-utilities\") pod \"redhat-marketplace-xhzzn\" (UID: \"70948643-9439-4c7d-ac9d-6cee3d7d58e0\") " pod="openshift-marketplace/redhat-marketplace-xhzzn" Oct 08 19:58:16 crc kubenswrapper[4988]: I1008 19:58:16.952231 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tdh2t\" (UniqueName: \"kubernetes.io/projected/70948643-9439-4c7d-ac9d-6cee3d7d58e0-kube-api-access-tdh2t\") pod \"redhat-marketplace-xhzzn\" (UID: \"70948643-9439-4c7d-ac9d-6cee3d7d58e0\") " pod="openshift-marketplace/redhat-marketplace-xhzzn" Oct 08 19:58:16 crc kubenswrapper[4988]: I1008 19:58:16.952718 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/70948643-9439-4c7d-ac9d-6cee3d7d58e0-catalog-content\") pod \"redhat-marketplace-xhzzn\" (UID: \"70948643-9439-4c7d-ac9d-6cee3d7d58e0\") " pod="openshift-marketplace/redhat-marketplace-xhzzn" Oct 08 19:58:16 crc kubenswrapper[4988]: I1008 19:58:16.952731 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/70948643-9439-4c7d-ac9d-6cee3d7d58e0-utilities\") pod \"redhat-marketplace-xhzzn\" (UID: \"70948643-9439-4c7d-ac9d-6cee3d7d58e0\") " pod="openshift-marketplace/redhat-marketplace-xhzzn" Oct 08 19:58:16 crc kubenswrapper[4988]: I1008 19:58:16.953148 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/70948643-9439-4c7d-ac9d-6cee3d7d58e0-catalog-content\") pod \"redhat-marketplace-xhzzn\" (UID: \"70948643-9439-4c7d-ac9d-6cee3d7d58e0\") " pod="openshift-marketplace/redhat-marketplace-xhzzn" Oct 08 19:58:16 crc kubenswrapper[4988]: I1008 19:58:16.978677 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tdh2t\" (UniqueName: \"kubernetes.io/projected/70948643-9439-4c7d-ac9d-6cee3d7d58e0-kube-api-access-tdh2t\") pod \"redhat-marketplace-xhzzn\" (UID: \"70948643-9439-4c7d-ac9d-6cee3d7d58e0\") " pod="openshift-marketplace/redhat-marketplace-xhzzn" Oct 08 19:58:17 crc kubenswrapper[4988]: I1008 19:58:17.011589 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-xhzzn" Oct 08 19:58:17 crc kubenswrapper[4988]: I1008 19:58:17.243716 4988 scope.go:117] "RemoveContainer" containerID="439af713ff941f973b8758bd6af3c097808f76ff520b96ff10bd3268811393f7" Oct 08 19:58:17 crc kubenswrapper[4988]: E1008 19:58:17.244561 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nm54f_openshift-machine-config-operator(f328a20b-5283-407e-8a93-2ab2da599f18)\"" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" Oct 08 19:58:17 crc kubenswrapper[4988]: I1008 19:58:17.475844 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-xhzzn"] Oct 08 19:58:17 crc kubenswrapper[4988]: I1008 19:58:17.734094 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"5cd96092-93f6-43a6-bec4-cceaba7bb8af","Type":"ContainerStarted","Data":"37f7075f94acc42853a6f73932fe8a7d9236f35da680b42a93d850eb38e3614e"} Oct 08 19:58:17 crc kubenswrapper[4988]: I1008 19:58:17.735560 4988 generic.go:334] "Generic (PLEG): container finished" podID="70948643-9439-4c7d-ac9d-6cee3d7d58e0" containerID="410573a7cbc60306e168a6c560c14da081413fd5b2be2163ddd0d22686130b06" exitCode=0 Oct 08 19:58:17 crc kubenswrapper[4988]: I1008 19:58:17.735600 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xhzzn" event={"ID":"70948643-9439-4c7d-ac9d-6cee3d7d58e0","Type":"ContainerDied","Data":"410573a7cbc60306e168a6c560c14da081413fd5b2be2163ddd0d22686130b06"} Oct 08 19:58:17 crc kubenswrapper[4988]: I1008 19:58:17.735615 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xhzzn" event={"ID":"70948643-9439-4c7d-ac9d-6cee3d7d58e0","Type":"ContainerStarted","Data":"fdad2110775351674c278640d66dd0460065182605e2225f88807db5633cb8cb"} Oct 08 19:58:18 crc kubenswrapper[4988]: I1008 19:58:18.483118 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Oct 08 19:58:18 crc kubenswrapper[4988]: I1008 19:58:18.749420 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/alertmanager-metric-storage-0" event={"ID":"64157199-c32a-461e-9a59-7d5a782c4838","Type":"ContainerStarted","Data":"7ccc91f6cf3efbb465fd30194e4b07d7acbe891413102b7a65d9b28f204b1723"} Oct 08 19:58:19 crc kubenswrapper[4988]: I1008 19:58:19.083680 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-dchph"] Oct 08 19:58:19 crc kubenswrapper[4988]: I1008 19:58:19.088926 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-dchph" Oct 08 19:58:19 crc kubenswrapper[4988]: I1008 19:58:19.099432 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-dchph"] Oct 08 19:58:19 crc kubenswrapper[4988]: I1008 19:58:19.212231 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7be16926-48be-47f1-a3f7-85be297d1b04-utilities\") pod \"certified-operators-dchph\" (UID: \"7be16926-48be-47f1-a3f7-85be297d1b04\") " pod="openshift-marketplace/certified-operators-dchph" Oct 08 19:58:19 crc kubenswrapper[4988]: I1008 19:58:19.212314 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7be16926-48be-47f1-a3f7-85be297d1b04-catalog-content\") pod \"certified-operators-dchph\" (UID: \"7be16926-48be-47f1-a3f7-85be297d1b04\") " pod="openshift-marketplace/certified-operators-dchph" Oct 08 19:58:19 crc kubenswrapper[4988]: I1008 19:58:19.212454 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rqrrv\" (UniqueName: \"kubernetes.io/projected/7be16926-48be-47f1-a3f7-85be297d1b04-kube-api-access-rqrrv\") pod \"certified-operators-dchph\" (UID: \"7be16926-48be-47f1-a3f7-85be297d1b04\") " pod="openshift-marketplace/certified-operators-dchph" Oct 08 19:58:19 crc kubenswrapper[4988]: I1008 19:58:19.314023 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rqrrv\" (UniqueName: \"kubernetes.io/projected/7be16926-48be-47f1-a3f7-85be297d1b04-kube-api-access-rqrrv\") pod \"certified-operators-dchph\" (UID: \"7be16926-48be-47f1-a3f7-85be297d1b04\") " pod="openshift-marketplace/certified-operators-dchph" Oct 08 19:58:19 crc kubenswrapper[4988]: I1008 19:58:19.314138 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7be16926-48be-47f1-a3f7-85be297d1b04-utilities\") pod \"certified-operators-dchph\" (UID: \"7be16926-48be-47f1-a3f7-85be297d1b04\") " pod="openshift-marketplace/certified-operators-dchph" Oct 08 19:58:19 crc kubenswrapper[4988]: I1008 19:58:19.314191 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7be16926-48be-47f1-a3f7-85be297d1b04-catalog-content\") pod \"certified-operators-dchph\" (UID: \"7be16926-48be-47f1-a3f7-85be297d1b04\") " pod="openshift-marketplace/certified-operators-dchph" Oct 08 19:58:19 crc kubenswrapper[4988]: I1008 19:58:19.314727 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7be16926-48be-47f1-a3f7-85be297d1b04-utilities\") pod \"certified-operators-dchph\" (UID: \"7be16926-48be-47f1-a3f7-85be297d1b04\") " pod="openshift-marketplace/certified-operators-dchph" Oct 08 19:58:19 crc kubenswrapper[4988]: I1008 19:58:19.314759 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7be16926-48be-47f1-a3f7-85be297d1b04-catalog-content\") pod \"certified-operators-dchph\" (UID: \"7be16926-48be-47f1-a3f7-85be297d1b04\") " pod="openshift-marketplace/certified-operators-dchph" Oct 08 19:58:19 crc kubenswrapper[4988]: I1008 19:58:19.331597 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rqrrv\" (UniqueName: \"kubernetes.io/projected/7be16926-48be-47f1-a3f7-85be297d1b04-kube-api-access-rqrrv\") pod \"certified-operators-dchph\" (UID: \"7be16926-48be-47f1-a3f7-85be297d1b04\") " pod="openshift-marketplace/certified-operators-dchph" Oct 08 19:58:19 crc kubenswrapper[4988]: I1008 19:58:19.454603 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-dchph" Oct 08 19:58:19 crc kubenswrapper[4988]: I1008 19:58:19.787578 4988 generic.go:334] "Generic (PLEG): container finished" podID="70948643-9439-4c7d-ac9d-6cee3d7d58e0" containerID="8233c96e7e79afee7550f0d57fbc23ce00d9fda895eeba5b6bb17fc7651c8df2" exitCode=0 Oct 08 19:58:19 crc kubenswrapper[4988]: I1008 19:58:19.787800 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xhzzn" event={"ID":"70948643-9439-4c7d-ac9d-6cee3d7d58e0","Type":"ContainerDied","Data":"8233c96e7e79afee7550f0d57fbc23ce00d9fda895eeba5b6bb17fc7651c8df2"} Oct 08 19:58:20 crc kubenswrapper[4988]: I1008 19:58:20.039329 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-dchph"] Oct 08 19:58:20 crc kubenswrapper[4988]: I1008 19:58:20.809760 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xhzzn" event={"ID":"70948643-9439-4c7d-ac9d-6cee3d7d58e0","Type":"ContainerStarted","Data":"c0d82438675b24fdf9b331b3c150cf910b5ea6d0a19851cf4126daae6dd77a20"} Oct 08 19:58:20 crc kubenswrapper[4988]: I1008 19:58:20.837934 4988 generic.go:334] "Generic (PLEG): container finished" podID="7be16926-48be-47f1-a3f7-85be297d1b04" containerID="b1ecb87dc146e7461c900e757dbfe41c5ce3052dd04c542d3b533db23b17ec76" exitCode=0 Oct 08 19:58:20 crc kubenswrapper[4988]: I1008 19:58:20.837992 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dchph" event={"ID":"7be16926-48be-47f1-a3f7-85be297d1b04","Type":"ContainerDied","Data":"b1ecb87dc146e7461c900e757dbfe41c5ce3052dd04c542d3b533db23b17ec76"} Oct 08 19:58:20 crc kubenswrapper[4988]: I1008 19:58:20.838039 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dchph" event={"ID":"7be16926-48be-47f1-a3f7-85be297d1b04","Type":"ContainerStarted","Data":"a1acc7b413543495ce1fb1d83ba9dec29a19c43187d19f7d4e01ed5a9e50a78b"} Oct 08 19:58:20 crc kubenswrapper[4988]: I1008 19:58:20.854324 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-xhzzn" podStartSLOduration=2.145829075 podStartE2EDuration="4.854302307s" podCreationTimestamp="2025-10-08 19:58:16 +0000 UTC" firstStartedPulling="2025-10-08 19:58:17.736988922 +0000 UTC m=+6443.186831732" lastFinishedPulling="2025-10-08 19:58:20.445462164 +0000 UTC m=+6445.895304964" observedRunningTime="2025-10-08 19:58:20.832188023 +0000 UTC m=+6446.282030803" watchObservedRunningTime="2025-10-08 19:58:20.854302307 +0000 UTC m=+6446.304145067" Oct 08 19:58:22 crc kubenswrapper[4988]: I1008 19:58:22.863142 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dchph" event={"ID":"7be16926-48be-47f1-a3f7-85be297d1b04","Type":"ContainerStarted","Data":"093f1ba3c9cf64932f473a3dffd9de69329fd9acdb4e133dc89034c17616dc20"} Oct 08 19:58:23 crc kubenswrapper[4988]: I1008 19:58:23.876875 4988 generic.go:334] "Generic (PLEG): container finished" podID="7be16926-48be-47f1-a3f7-85be297d1b04" containerID="093f1ba3c9cf64932f473a3dffd9de69329fd9acdb4e133dc89034c17616dc20" exitCode=0 Oct 08 19:58:23 crc kubenswrapper[4988]: I1008 19:58:23.877423 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dchph" event={"ID":"7be16926-48be-47f1-a3f7-85be297d1b04","Type":"ContainerDied","Data":"093f1ba3c9cf64932f473a3dffd9de69329fd9acdb4e133dc89034c17616dc20"} Oct 08 19:58:24 crc kubenswrapper[4988]: I1008 19:58:24.890440 4988 generic.go:334] "Generic (PLEG): container finished" podID="64157199-c32a-461e-9a59-7d5a782c4838" containerID="7ccc91f6cf3efbb465fd30194e4b07d7acbe891413102b7a65d9b28f204b1723" exitCode=0 Oct 08 19:58:24 crc kubenswrapper[4988]: I1008 19:58:24.890698 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/alertmanager-metric-storage-0" event={"ID":"64157199-c32a-461e-9a59-7d5a782c4838","Type":"ContainerDied","Data":"7ccc91f6cf3efbb465fd30194e4b07d7acbe891413102b7a65d9b28f204b1723"} Oct 08 19:58:24 crc kubenswrapper[4988]: I1008 19:58:24.893763 4988 generic.go:334] "Generic (PLEG): container finished" podID="5cd96092-93f6-43a6-bec4-cceaba7bb8af" containerID="37f7075f94acc42853a6f73932fe8a7d9236f35da680b42a93d850eb38e3614e" exitCode=0 Oct 08 19:58:24 crc kubenswrapper[4988]: I1008 19:58:24.893836 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"5cd96092-93f6-43a6-bec4-cceaba7bb8af","Type":"ContainerDied","Data":"37f7075f94acc42853a6f73932fe8a7d9236f35da680b42a93d850eb38e3614e"} Oct 08 19:58:24 crc kubenswrapper[4988]: I1008 19:58:24.898527 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dchph" event={"ID":"7be16926-48be-47f1-a3f7-85be297d1b04","Type":"ContainerStarted","Data":"efd368a715977378f17f8fb0e5a710dae0acaff2d682b57ac1847ad160ba34c7"} Oct 08 19:58:24 crc kubenswrapper[4988]: I1008 19:58:24.998928 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-dchph" podStartSLOduration=2.487733464 podStartE2EDuration="5.998899606s" podCreationTimestamp="2025-10-08 19:58:19 +0000 UTC" firstStartedPulling="2025-10-08 19:58:20.843128401 +0000 UTC m=+6446.292971161" lastFinishedPulling="2025-10-08 19:58:24.354294533 +0000 UTC m=+6449.804137303" observedRunningTime="2025-10-08 19:58:24.94472856 +0000 UTC m=+6450.394571340" watchObservedRunningTime="2025-10-08 19:58:24.998899606 +0000 UTC m=+6450.448742416" Oct 08 19:58:27 crc kubenswrapper[4988]: I1008 19:58:27.012648 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-xhzzn" Oct 08 19:58:27 crc kubenswrapper[4988]: I1008 19:58:27.013514 4988 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-xhzzn" Oct 08 19:58:27 crc kubenswrapper[4988]: I1008 19:58:27.082351 4988 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-xhzzn" Oct 08 19:58:28 crc kubenswrapper[4988]: I1008 19:58:28.002931 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-xhzzn" Oct 08 19:58:28 crc kubenswrapper[4988]: I1008 19:58:28.241446 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-xhzzn"] Oct 08 19:58:28 crc kubenswrapper[4988]: I1008 19:58:28.243139 4988 scope.go:117] "RemoveContainer" containerID="439af713ff941f973b8758bd6af3c097808f76ff520b96ff10bd3268811393f7" Oct 08 19:58:28 crc kubenswrapper[4988]: E1008 19:58:28.243510 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nm54f_openshift-machine-config-operator(f328a20b-5283-407e-8a93-2ab2da599f18)\"" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" Oct 08 19:58:28 crc kubenswrapper[4988]: I1008 19:58:28.954846 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/alertmanager-metric-storage-0" event={"ID":"64157199-c32a-461e-9a59-7d5a782c4838","Type":"ContainerStarted","Data":"d7b90859e15146e58074c5437438d3692b40e3e8104b691b3bc8959c5ccb3b08"} Oct 08 19:58:29 crc kubenswrapper[4988]: I1008 19:58:29.455103 4988 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-dchph" Oct 08 19:58:29 crc kubenswrapper[4988]: I1008 19:58:29.455142 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-dchph" Oct 08 19:58:29 crc kubenswrapper[4988]: I1008 19:58:29.504750 4988 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-dchph" Oct 08 19:58:29 crc kubenswrapper[4988]: I1008 19:58:29.985270 4988 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-xhzzn" podUID="70948643-9439-4c7d-ac9d-6cee3d7d58e0" containerName="registry-server" containerID="cri-o://c0d82438675b24fdf9b331b3c150cf910b5ea6d0a19851cf4126daae6dd77a20" gracePeriod=2 Oct 08 19:58:30 crc kubenswrapper[4988]: I1008 19:58:30.057229 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-dchph" Oct 08 19:58:30 crc kubenswrapper[4988]: I1008 19:58:30.700595 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-xhzzn" Oct 08 19:58:30 crc kubenswrapper[4988]: I1008 19:58:30.768965 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/70948643-9439-4c7d-ac9d-6cee3d7d58e0-catalog-content\") pod \"70948643-9439-4c7d-ac9d-6cee3d7d58e0\" (UID: \"70948643-9439-4c7d-ac9d-6cee3d7d58e0\") " Oct 08 19:58:30 crc kubenswrapper[4988]: I1008 19:58:30.769113 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tdh2t\" (UniqueName: \"kubernetes.io/projected/70948643-9439-4c7d-ac9d-6cee3d7d58e0-kube-api-access-tdh2t\") pod \"70948643-9439-4c7d-ac9d-6cee3d7d58e0\" (UID: \"70948643-9439-4c7d-ac9d-6cee3d7d58e0\") " Oct 08 19:58:30 crc kubenswrapper[4988]: I1008 19:58:30.769326 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/70948643-9439-4c7d-ac9d-6cee3d7d58e0-utilities\") pod \"70948643-9439-4c7d-ac9d-6cee3d7d58e0\" (UID: \"70948643-9439-4c7d-ac9d-6cee3d7d58e0\") " Oct 08 19:58:30 crc kubenswrapper[4988]: I1008 19:58:30.769925 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/70948643-9439-4c7d-ac9d-6cee3d7d58e0-utilities" (OuterVolumeSpecName: "utilities") pod "70948643-9439-4c7d-ac9d-6cee3d7d58e0" (UID: "70948643-9439-4c7d-ac9d-6cee3d7d58e0"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 19:58:30 crc kubenswrapper[4988]: I1008 19:58:30.774955 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/70948643-9439-4c7d-ac9d-6cee3d7d58e0-kube-api-access-tdh2t" (OuterVolumeSpecName: "kube-api-access-tdh2t") pod "70948643-9439-4c7d-ac9d-6cee3d7d58e0" (UID: "70948643-9439-4c7d-ac9d-6cee3d7d58e0"). InnerVolumeSpecName "kube-api-access-tdh2t". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 19:58:30 crc kubenswrapper[4988]: I1008 19:58:30.780401 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/70948643-9439-4c7d-ac9d-6cee3d7d58e0-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "70948643-9439-4c7d-ac9d-6cee3d7d58e0" (UID: "70948643-9439-4c7d-ac9d-6cee3d7d58e0"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 19:58:30 crc kubenswrapper[4988]: I1008 19:58:30.871871 4988 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tdh2t\" (UniqueName: \"kubernetes.io/projected/70948643-9439-4c7d-ac9d-6cee3d7d58e0-kube-api-access-tdh2t\") on node \"crc\" DevicePath \"\"" Oct 08 19:58:30 crc kubenswrapper[4988]: I1008 19:58:30.871914 4988 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/70948643-9439-4c7d-ac9d-6cee3d7d58e0-utilities\") on node \"crc\" DevicePath \"\"" Oct 08 19:58:30 crc kubenswrapper[4988]: I1008 19:58:30.871926 4988 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/70948643-9439-4c7d-ac9d-6cee3d7d58e0-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 08 19:58:31 crc kubenswrapper[4988]: I1008 19:58:31.002909 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"5cd96092-93f6-43a6-bec4-cceaba7bb8af","Type":"ContainerStarted","Data":"5e78b0fe3541a2320e39f5e23ea42803ccc4bbfe2e9239e7b173d033f105ae28"} Oct 08 19:58:31 crc kubenswrapper[4988]: I1008 19:58:31.004993 4988 generic.go:334] "Generic (PLEG): container finished" podID="70948643-9439-4c7d-ac9d-6cee3d7d58e0" containerID="c0d82438675b24fdf9b331b3c150cf910b5ea6d0a19851cf4126daae6dd77a20" exitCode=0 Oct 08 19:58:31 crc kubenswrapper[4988]: I1008 19:58:31.005611 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-xhzzn" Oct 08 19:58:31 crc kubenswrapper[4988]: I1008 19:58:31.005630 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xhzzn" event={"ID":"70948643-9439-4c7d-ac9d-6cee3d7d58e0","Type":"ContainerDied","Data":"c0d82438675b24fdf9b331b3c150cf910b5ea6d0a19851cf4126daae6dd77a20"} Oct 08 19:58:31 crc kubenswrapper[4988]: I1008 19:58:31.005716 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xhzzn" event={"ID":"70948643-9439-4c7d-ac9d-6cee3d7d58e0","Type":"ContainerDied","Data":"fdad2110775351674c278640d66dd0460065182605e2225f88807db5633cb8cb"} Oct 08 19:58:31 crc kubenswrapper[4988]: I1008 19:58:31.005762 4988 scope.go:117] "RemoveContainer" containerID="c0d82438675b24fdf9b331b3c150cf910b5ea6d0a19851cf4126daae6dd77a20" Oct 08 19:58:31 crc kubenswrapper[4988]: I1008 19:58:31.047266 4988 scope.go:117] "RemoveContainer" containerID="8233c96e7e79afee7550f0d57fbc23ce00d9fda895eeba5b6bb17fc7651c8df2" Oct 08 19:58:31 crc kubenswrapper[4988]: I1008 19:58:31.058543 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-xhzzn"] Oct 08 19:58:31 crc kubenswrapper[4988]: I1008 19:58:31.070280 4988 scope.go:117] "RemoveContainer" containerID="410573a7cbc60306e168a6c560c14da081413fd5b2be2163ddd0d22686130b06" Oct 08 19:58:31 crc kubenswrapper[4988]: I1008 19:58:31.076242 4988 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-xhzzn"] Oct 08 19:58:31 crc kubenswrapper[4988]: I1008 19:58:31.097817 4988 scope.go:117] "RemoveContainer" containerID="c0d82438675b24fdf9b331b3c150cf910b5ea6d0a19851cf4126daae6dd77a20" Oct 08 19:58:31 crc kubenswrapper[4988]: E1008 19:58:31.098301 4988 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c0d82438675b24fdf9b331b3c150cf910b5ea6d0a19851cf4126daae6dd77a20\": container with ID starting with c0d82438675b24fdf9b331b3c150cf910b5ea6d0a19851cf4126daae6dd77a20 not found: ID does not exist" containerID="c0d82438675b24fdf9b331b3c150cf910b5ea6d0a19851cf4126daae6dd77a20" Oct 08 19:58:31 crc kubenswrapper[4988]: I1008 19:58:31.098356 4988 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c0d82438675b24fdf9b331b3c150cf910b5ea6d0a19851cf4126daae6dd77a20"} err="failed to get container status \"c0d82438675b24fdf9b331b3c150cf910b5ea6d0a19851cf4126daae6dd77a20\": rpc error: code = NotFound desc = could not find container \"c0d82438675b24fdf9b331b3c150cf910b5ea6d0a19851cf4126daae6dd77a20\": container with ID starting with c0d82438675b24fdf9b331b3c150cf910b5ea6d0a19851cf4126daae6dd77a20 not found: ID does not exist" Oct 08 19:58:31 crc kubenswrapper[4988]: I1008 19:58:31.098414 4988 scope.go:117] "RemoveContainer" containerID="8233c96e7e79afee7550f0d57fbc23ce00d9fda895eeba5b6bb17fc7651c8df2" Oct 08 19:58:31 crc kubenswrapper[4988]: E1008 19:58:31.098757 4988 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8233c96e7e79afee7550f0d57fbc23ce00d9fda895eeba5b6bb17fc7651c8df2\": container with ID starting with 8233c96e7e79afee7550f0d57fbc23ce00d9fda895eeba5b6bb17fc7651c8df2 not found: ID does not exist" containerID="8233c96e7e79afee7550f0d57fbc23ce00d9fda895eeba5b6bb17fc7651c8df2" Oct 08 19:58:31 crc kubenswrapper[4988]: I1008 19:58:31.098796 4988 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8233c96e7e79afee7550f0d57fbc23ce00d9fda895eeba5b6bb17fc7651c8df2"} err="failed to get container status \"8233c96e7e79afee7550f0d57fbc23ce00d9fda895eeba5b6bb17fc7651c8df2\": rpc error: code = NotFound desc = could not find container \"8233c96e7e79afee7550f0d57fbc23ce00d9fda895eeba5b6bb17fc7651c8df2\": container with ID starting with 8233c96e7e79afee7550f0d57fbc23ce00d9fda895eeba5b6bb17fc7651c8df2 not found: ID does not exist" Oct 08 19:58:31 crc kubenswrapper[4988]: I1008 19:58:31.098824 4988 scope.go:117] "RemoveContainer" containerID="410573a7cbc60306e168a6c560c14da081413fd5b2be2163ddd0d22686130b06" Oct 08 19:58:31 crc kubenswrapper[4988]: E1008 19:58:31.099126 4988 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"410573a7cbc60306e168a6c560c14da081413fd5b2be2163ddd0d22686130b06\": container with ID starting with 410573a7cbc60306e168a6c560c14da081413fd5b2be2163ddd0d22686130b06 not found: ID does not exist" containerID="410573a7cbc60306e168a6c560c14da081413fd5b2be2163ddd0d22686130b06" Oct 08 19:58:31 crc kubenswrapper[4988]: I1008 19:58:31.099154 4988 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"410573a7cbc60306e168a6c560c14da081413fd5b2be2163ddd0d22686130b06"} err="failed to get container status \"410573a7cbc60306e168a6c560c14da081413fd5b2be2163ddd0d22686130b06\": rpc error: code = NotFound desc = could not find container \"410573a7cbc60306e168a6c560c14da081413fd5b2be2163ddd0d22686130b06\": container with ID starting with 410573a7cbc60306e168a6c560c14da081413fd5b2be2163ddd0d22686130b06 not found: ID does not exist" Oct 08 19:58:31 crc kubenswrapper[4988]: I1008 19:58:31.258091 4988 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="70948643-9439-4c7d-ac9d-6cee3d7d58e0" path="/var/lib/kubelet/pods/70948643-9439-4c7d-ac9d-6cee3d7d58e0/volumes" Oct 08 19:58:31 crc kubenswrapper[4988]: I1008 19:58:31.442018 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-dchph"] Oct 08 19:58:32 crc kubenswrapper[4988]: I1008 19:58:32.019028 4988 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-dchph" podUID="7be16926-48be-47f1-a3f7-85be297d1b04" containerName="registry-server" containerID="cri-o://efd368a715977378f17f8fb0e5a710dae0acaff2d682b57ac1847ad160ba34c7" gracePeriod=2 Oct 08 19:58:32 crc kubenswrapper[4988]: I1008 19:58:32.699555 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-dchph" Oct 08 19:58:32 crc kubenswrapper[4988]: I1008 19:58:32.819692 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7be16926-48be-47f1-a3f7-85be297d1b04-utilities\") pod \"7be16926-48be-47f1-a3f7-85be297d1b04\" (UID: \"7be16926-48be-47f1-a3f7-85be297d1b04\") " Oct 08 19:58:32 crc kubenswrapper[4988]: I1008 19:58:32.820052 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rqrrv\" (UniqueName: \"kubernetes.io/projected/7be16926-48be-47f1-a3f7-85be297d1b04-kube-api-access-rqrrv\") pod \"7be16926-48be-47f1-a3f7-85be297d1b04\" (UID: \"7be16926-48be-47f1-a3f7-85be297d1b04\") " Oct 08 19:58:32 crc kubenswrapper[4988]: I1008 19:58:32.820108 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7be16926-48be-47f1-a3f7-85be297d1b04-catalog-content\") pod \"7be16926-48be-47f1-a3f7-85be297d1b04\" (UID: \"7be16926-48be-47f1-a3f7-85be297d1b04\") " Oct 08 19:58:32 crc kubenswrapper[4988]: I1008 19:58:32.820570 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7be16926-48be-47f1-a3f7-85be297d1b04-utilities" (OuterVolumeSpecName: "utilities") pod "7be16926-48be-47f1-a3f7-85be297d1b04" (UID: "7be16926-48be-47f1-a3f7-85be297d1b04"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 19:58:32 crc kubenswrapper[4988]: I1008 19:58:32.821114 4988 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7be16926-48be-47f1-a3f7-85be297d1b04-utilities\") on node \"crc\" DevicePath \"\"" Oct 08 19:58:32 crc kubenswrapper[4988]: I1008 19:58:32.863279 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7be16926-48be-47f1-a3f7-85be297d1b04-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "7be16926-48be-47f1-a3f7-85be297d1b04" (UID: "7be16926-48be-47f1-a3f7-85be297d1b04"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 19:58:32 crc kubenswrapper[4988]: I1008 19:58:32.879881 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7be16926-48be-47f1-a3f7-85be297d1b04-kube-api-access-rqrrv" (OuterVolumeSpecName: "kube-api-access-rqrrv") pod "7be16926-48be-47f1-a3f7-85be297d1b04" (UID: "7be16926-48be-47f1-a3f7-85be297d1b04"). InnerVolumeSpecName "kube-api-access-rqrrv". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 19:58:32 crc kubenswrapper[4988]: I1008 19:58:32.924122 4988 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rqrrv\" (UniqueName: \"kubernetes.io/projected/7be16926-48be-47f1-a3f7-85be297d1b04-kube-api-access-rqrrv\") on node \"crc\" DevicePath \"\"" Oct 08 19:58:32 crc kubenswrapper[4988]: I1008 19:58:32.924177 4988 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7be16926-48be-47f1-a3f7-85be297d1b04-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 08 19:58:33 crc kubenswrapper[4988]: I1008 19:58:33.056040 4988 generic.go:334] "Generic (PLEG): container finished" podID="7be16926-48be-47f1-a3f7-85be297d1b04" containerID="efd368a715977378f17f8fb0e5a710dae0acaff2d682b57ac1847ad160ba34c7" exitCode=0 Oct 08 19:58:33 crc kubenswrapper[4988]: I1008 19:58:33.056413 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dchph" event={"ID":"7be16926-48be-47f1-a3f7-85be297d1b04","Type":"ContainerDied","Data":"efd368a715977378f17f8fb0e5a710dae0acaff2d682b57ac1847ad160ba34c7"} Oct 08 19:58:33 crc kubenswrapper[4988]: I1008 19:58:33.056449 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dchph" event={"ID":"7be16926-48be-47f1-a3f7-85be297d1b04","Type":"ContainerDied","Data":"a1acc7b413543495ce1fb1d83ba9dec29a19c43187d19f7d4e01ed5a9e50a78b"} Oct 08 19:58:33 crc kubenswrapper[4988]: I1008 19:58:33.056470 4988 scope.go:117] "RemoveContainer" containerID="efd368a715977378f17f8fb0e5a710dae0acaff2d682b57ac1847ad160ba34c7" Oct 08 19:58:33 crc kubenswrapper[4988]: I1008 19:58:33.056582 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-dchph" Oct 08 19:58:34 crc kubenswrapper[4988]: I1008 19:58:34.400220 4988 scope.go:117] "RemoveContainer" containerID="093f1ba3c9cf64932f473a3dffd9de69329fd9acdb4e133dc89034c17616dc20" Oct 08 19:58:34 crc kubenswrapper[4988]: I1008 19:58:34.429347 4988 scope.go:117] "RemoveContainer" containerID="b1ecb87dc146e7461c900e757dbfe41c5ce3052dd04c542d3b533db23b17ec76" Oct 08 19:58:34 crc kubenswrapper[4988]: I1008 19:58:34.480981 4988 scope.go:117] "RemoveContainer" containerID="efd368a715977378f17f8fb0e5a710dae0acaff2d682b57ac1847ad160ba34c7" Oct 08 19:58:34 crc kubenswrapper[4988]: E1008 19:58:34.481465 4988 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"efd368a715977378f17f8fb0e5a710dae0acaff2d682b57ac1847ad160ba34c7\": container with ID starting with efd368a715977378f17f8fb0e5a710dae0acaff2d682b57ac1847ad160ba34c7 not found: ID does not exist" containerID="efd368a715977378f17f8fb0e5a710dae0acaff2d682b57ac1847ad160ba34c7" Oct 08 19:58:34 crc kubenswrapper[4988]: I1008 19:58:34.481501 4988 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"efd368a715977378f17f8fb0e5a710dae0acaff2d682b57ac1847ad160ba34c7"} err="failed to get container status \"efd368a715977378f17f8fb0e5a710dae0acaff2d682b57ac1847ad160ba34c7\": rpc error: code = NotFound desc = could not find container \"efd368a715977378f17f8fb0e5a710dae0acaff2d682b57ac1847ad160ba34c7\": container with ID starting with efd368a715977378f17f8fb0e5a710dae0acaff2d682b57ac1847ad160ba34c7 not found: ID does not exist" Oct 08 19:58:34 crc kubenswrapper[4988]: I1008 19:58:34.481531 4988 scope.go:117] "RemoveContainer" containerID="093f1ba3c9cf64932f473a3dffd9de69329fd9acdb4e133dc89034c17616dc20" Oct 08 19:58:34 crc kubenswrapper[4988]: E1008 19:58:34.481821 4988 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"093f1ba3c9cf64932f473a3dffd9de69329fd9acdb4e133dc89034c17616dc20\": container with ID starting with 093f1ba3c9cf64932f473a3dffd9de69329fd9acdb4e133dc89034c17616dc20 not found: ID does not exist" containerID="093f1ba3c9cf64932f473a3dffd9de69329fd9acdb4e133dc89034c17616dc20" Oct 08 19:58:34 crc kubenswrapper[4988]: I1008 19:58:34.481869 4988 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"093f1ba3c9cf64932f473a3dffd9de69329fd9acdb4e133dc89034c17616dc20"} err="failed to get container status \"093f1ba3c9cf64932f473a3dffd9de69329fd9acdb4e133dc89034c17616dc20\": rpc error: code = NotFound desc = could not find container \"093f1ba3c9cf64932f473a3dffd9de69329fd9acdb4e133dc89034c17616dc20\": container with ID starting with 093f1ba3c9cf64932f473a3dffd9de69329fd9acdb4e133dc89034c17616dc20 not found: ID does not exist" Oct 08 19:58:34 crc kubenswrapper[4988]: I1008 19:58:34.481895 4988 scope.go:117] "RemoveContainer" containerID="b1ecb87dc146e7461c900e757dbfe41c5ce3052dd04c542d3b533db23b17ec76" Oct 08 19:58:34 crc kubenswrapper[4988]: E1008 19:58:34.482422 4988 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b1ecb87dc146e7461c900e757dbfe41c5ce3052dd04c542d3b533db23b17ec76\": container with ID starting with b1ecb87dc146e7461c900e757dbfe41c5ce3052dd04c542d3b533db23b17ec76 not found: ID does not exist" containerID="b1ecb87dc146e7461c900e757dbfe41c5ce3052dd04c542d3b533db23b17ec76" Oct 08 19:58:34 crc kubenswrapper[4988]: I1008 19:58:34.482447 4988 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b1ecb87dc146e7461c900e757dbfe41c5ce3052dd04c542d3b533db23b17ec76"} err="failed to get container status \"b1ecb87dc146e7461c900e757dbfe41c5ce3052dd04c542d3b533db23b17ec76\": rpc error: code = NotFound desc = could not find container \"b1ecb87dc146e7461c900e757dbfe41c5ce3052dd04c542d3b533db23b17ec76\": container with ID starting with b1ecb87dc146e7461c900e757dbfe41c5ce3052dd04c542d3b533db23b17ec76 not found: ID does not exist" Oct 08 19:58:35 crc kubenswrapper[4988]: I1008 19:58:35.080423 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/alertmanager-metric-storage-0" event={"ID":"64157199-c32a-461e-9a59-7d5a782c4838","Type":"ContainerStarted","Data":"35cc322c853957a87515faae366e05a34d053dc97ab63afd58e6cb5bd4a908a9"} Oct 08 19:58:35 crc kubenswrapper[4988]: I1008 19:58:35.080652 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/alertmanager-metric-storage-0" Oct 08 19:58:35 crc kubenswrapper[4988]: I1008 19:58:35.084101 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"5cd96092-93f6-43a6-bec4-cceaba7bb8af","Type":"ContainerStarted","Data":"bdf26b13144c9af7cb42452ac56fbb0dd91448ac6b82f422d8ff9d3477f4d477"} Oct 08 19:58:35 crc kubenswrapper[4988]: I1008 19:58:35.085615 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/alertmanager-metric-storage-0" Oct 08 19:58:35 crc kubenswrapper[4988]: I1008 19:58:35.115033 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/alertmanager-metric-storage-0" podStartSLOduration=8.381618553 podStartE2EDuration="25.115011184s" podCreationTimestamp="2025-10-08 19:58:10 +0000 UTC" firstStartedPulling="2025-10-08 19:58:11.316137406 +0000 UTC m=+6436.765980176" lastFinishedPulling="2025-10-08 19:58:28.049530037 +0000 UTC m=+6453.499372807" observedRunningTime="2025-10-08 19:58:35.109451377 +0000 UTC m=+6460.559294177" watchObservedRunningTime="2025-10-08 19:58:35.115011184 +0000 UTC m=+6460.564853954" Oct 08 19:58:37 crc kubenswrapper[4988]: I1008 19:58:37.113875 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"5cd96092-93f6-43a6-bec4-cceaba7bb8af","Type":"ContainerStarted","Data":"286424932a0992c4d6fbc7f99d9874252b3d97e9e5d669ba59a770288e4f635d"} Oct 08 19:58:37 crc kubenswrapper[4988]: I1008 19:58:37.157431 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/prometheus-metric-storage-0" podStartSLOduration=3.500844096 podStartE2EDuration="27.1573562s" podCreationTimestamp="2025-10-08 19:58:10 +0000 UTC" firstStartedPulling="2025-10-08 19:58:12.982596678 +0000 UTC m=+6438.432439448" lastFinishedPulling="2025-10-08 19:58:36.639108752 +0000 UTC m=+6462.088951552" observedRunningTime="2025-10-08 19:58:37.146117022 +0000 UTC m=+6462.595959822" watchObservedRunningTime="2025-10-08 19:58:37.1573562 +0000 UTC m=+6462.607199010" Oct 08 19:58:37 crc kubenswrapper[4988]: I1008 19:58:37.345207 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/prometheus-metric-storage-0" Oct 08 19:58:42 crc kubenswrapper[4988]: I1008 19:58:42.344555 4988 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/prometheus-metric-storage-0" Oct 08 19:58:42 crc kubenswrapper[4988]: I1008 19:58:42.347018 4988 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/prometheus-metric-storage-0" Oct 08 19:58:43 crc kubenswrapper[4988]: I1008 19:58:43.194022 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/prometheus-metric-storage-0" Oct 08 19:58:43 crc kubenswrapper[4988]: I1008 19:58:43.237578 4988 scope.go:117] "RemoveContainer" containerID="439af713ff941f973b8758bd6af3c097808f76ff520b96ff10bd3268811393f7" Oct 08 19:58:43 crc kubenswrapper[4988]: E1008 19:58:43.237865 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nm54f_openshift-machine-config-operator(f328a20b-5283-407e-8a93-2ab2da599f18)\"" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" Oct 08 19:58:44 crc kubenswrapper[4988]: I1008 19:58:44.550896 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/openstackclient"] Oct 08 19:58:44 crc kubenswrapper[4988]: I1008 19:58:44.551292 4988 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/openstackclient" podUID="2d58f5a2-1a02-4152-95ec-22cc390d8ae5" containerName="openstackclient" containerID="cri-o://b7ea1ae1a08e4188d78ecf340ffa61f27ee87f49871711bb009130301dbdf138" gracePeriod=2 Oct 08 19:58:44 crc kubenswrapper[4988]: I1008 19:58:44.627633 4988 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/openstackclient"] Oct 08 19:58:44 crc kubenswrapper[4988]: I1008 19:58:44.629960 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstackclient"] Oct 08 19:58:44 crc kubenswrapper[4988]: E1008 19:58:44.630446 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2d58f5a2-1a02-4152-95ec-22cc390d8ae5" containerName="openstackclient" Oct 08 19:58:44 crc kubenswrapper[4988]: I1008 19:58:44.630464 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="2d58f5a2-1a02-4152-95ec-22cc390d8ae5" containerName="openstackclient" Oct 08 19:58:44 crc kubenswrapper[4988]: E1008 19:58:44.630486 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7be16926-48be-47f1-a3f7-85be297d1b04" containerName="extract-utilities" Oct 08 19:58:44 crc kubenswrapper[4988]: I1008 19:58:44.630493 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="7be16926-48be-47f1-a3f7-85be297d1b04" containerName="extract-utilities" Oct 08 19:58:44 crc kubenswrapper[4988]: E1008 19:58:44.630508 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7be16926-48be-47f1-a3f7-85be297d1b04" containerName="extract-content" Oct 08 19:58:44 crc kubenswrapper[4988]: I1008 19:58:44.630514 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="7be16926-48be-47f1-a3f7-85be297d1b04" containerName="extract-content" Oct 08 19:58:44 crc kubenswrapper[4988]: E1008 19:58:44.630527 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7be16926-48be-47f1-a3f7-85be297d1b04" containerName="registry-server" Oct 08 19:58:44 crc kubenswrapper[4988]: I1008 19:58:44.630532 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="7be16926-48be-47f1-a3f7-85be297d1b04" containerName="registry-server" Oct 08 19:58:44 crc kubenswrapper[4988]: E1008 19:58:44.630542 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="70948643-9439-4c7d-ac9d-6cee3d7d58e0" containerName="extract-utilities" Oct 08 19:58:44 crc kubenswrapper[4988]: I1008 19:58:44.630555 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="70948643-9439-4c7d-ac9d-6cee3d7d58e0" containerName="extract-utilities" Oct 08 19:58:44 crc kubenswrapper[4988]: E1008 19:58:44.630573 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="70948643-9439-4c7d-ac9d-6cee3d7d58e0" containerName="registry-server" Oct 08 19:58:44 crc kubenswrapper[4988]: I1008 19:58:44.630579 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="70948643-9439-4c7d-ac9d-6cee3d7d58e0" containerName="registry-server" Oct 08 19:58:44 crc kubenswrapper[4988]: E1008 19:58:44.630596 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="70948643-9439-4c7d-ac9d-6cee3d7d58e0" containerName="extract-content" Oct 08 19:58:44 crc kubenswrapper[4988]: I1008 19:58:44.630602 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="70948643-9439-4c7d-ac9d-6cee3d7d58e0" containerName="extract-content" Oct 08 19:58:44 crc kubenswrapper[4988]: I1008 19:58:44.630796 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="70948643-9439-4c7d-ac9d-6cee3d7d58e0" containerName="registry-server" Oct 08 19:58:44 crc kubenswrapper[4988]: I1008 19:58:44.630807 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="2d58f5a2-1a02-4152-95ec-22cc390d8ae5" containerName="openstackclient" Oct 08 19:58:44 crc kubenswrapper[4988]: I1008 19:58:44.630831 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="7be16926-48be-47f1-a3f7-85be297d1b04" containerName="registry-server" Oct 08 19:58:44 crc kubenswrapper[4988]: I1008 19:58:44.631611 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Oct 08 19:58:44 crc kubenswrapper[4988]: I1008 19:58:44.641776 4988 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openstack/openstackclient" oldPodUID="52335cb7-2ea6-4571-a595-fc6c8c071f50" podUID="efec0d27-7968-44f2-a1f4-733078747791" Oct 08 19:58:44 crc kubenswrapper[4988]: I1008 19:58:44.644598 4988 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openstack/openstackclient" oldPodUID="2d58f5a2-1a02-4152-95ec-22cc390d8ae5" podUID="efec0d27-7968-44f2-a1f4-733078747791" Oct 08 19:58:44 crc kubenswrapper[4988]: I1008 19:58:44.651005 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Oct 08 19:58:44 crc kubenswrapper[4988]: I1008 19:58:44.661340 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/openstackclient"] Oct 08 19:58:44 crc kubenswrapper[4988]: E1008 19:58:44.662114 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="unmounted volumes=[combined-ca-bundle kube-api-access-dbs4g openstack-config openstack-config-secret], unattached volumes=[], failed to process volumes=[]: context canceled" pod="openstack/openstackclient" podUID="52335cb7-2ea6-4571-a595-fc6c8c071f50" Oct 08 19:58:44 crc kubenswrapper[4988]: I1008 19:58:44.669470 4988 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/openstackclient"] Oct 08 19:58:44 crc kubenswrapper[4988]: I1008 19:58:44.681247 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstackclient"] Oct 08 19:58:44 crc kubenswrapper[4988]: I1008 19:58:44.683011 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Oct 08 19:58:44 crc kubenswrapper[4988]: I1008 19:58:44.689320 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Oct 08 19:58:44 crc kubenswrapper[4988]: I1008 19:58:44.705999 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/52335cb7-2ea6-4571-a595-fc6c8c071f50-openstack-config\") pod \"openstackclient\" (UID: \"52335cb7-2ea6-4571-a595-fc6c8c071f50\") " pod="openstack/openstackclient" Oct 08 19:58:44 crc kubenswrapper[4988]: I1008 19:58:44.706217 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dbs4g\" (UniqueName: \"kubernetes.io/projected/52335cb7-2ea6-4571-a595-fc6c8c071f50-kube-api-access-dbs4g\") pod \"openstackclient\" (UID: \"52335cb7-2ea6-4571-a595-fc6c8c071f50\") " pod="openstack/openstackclient" Oct 08 19:58:44 crc kubenswrapper[4988]: I1008 19:58:44.706261 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/52335cb7-2ea6-4571-a595-fc6c8c071f50-combined-ca-bundle\") pod \"openstackclient\" (UID: \"52335cb7-2ea6-4571-a595-fc6c8c071f50\") " pod="openstack/openstackclient" Oct 08 19:58:44 crc kubenswrapper[4988]: I1008 19:58:44.706585 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/52335cb7-2ea6-4571-a595-fc6c8c071f50-openstack-config-secret\") pod \"openstackclient\" (UID: \"52335cb7-2ea6-4571-a595-fc6c8c071f50\") " pod="openstack/openstackclient" Oct 08 19:58:44 crc kubenswrapper[4988]: I1008 19:58:44.808014 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dbs4g\" (UniqueName: \"kubernetes.io/projected/52335cb7-2ea6-4571-a595-fc6c8c071f50-kube-api-access-dbs4g\") pod \"openstackclient\" (UID: \"52335cb7-2ea6-4571-a595-fc6c8c071f50\") " pod="openstack/openstackclient" Oct 08 19:58:44 crc kubenswrapper[4988]: I1008 19:58:44.808070 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/52335cb7-2ea6-4571-a595-fc6c8c071f50-combined-ca-bundle\") pod \"openstackclient\" (UID: \"52335cb7-2ea6-4571-a595-fc6c8c071f50\") " pod="openstack/openstackclient" Oct 08 19:58:44 crc kubenswrapper[4988]: I1008 19:58:44.808156 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8ztjk\" (UniqueName: \"kubernetes.io/projected/efec0d27-7968-44f2-a1f4-733078747791-kube-api-access-8ztjk\") pod \"openstackclient\" (UID: \"efec0d27-7968-44f2-a1f4-733078747791\") " pod="openstack/openstackclient" Oct 08 19:58:44 crc kubenswrapper[4988]: I1008 19:58:44.808196 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/52335cb7-2ea6-4571-a595-fc6c8c071f50-openstack-config-secret\") pod \"openstackclient\" (UID: \"52335cb7-2ea6-4571-a595-fc6c8c071f50\") " pod="openstack/openstackclient" Oct 08 19:58:44 crc kubenswrapper[4988]: I1008 19:58:44.808261 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/52335cb7-2ea6-4571-a595-fc6c8c071f50-openstack-config\") pod \"openstackclient\" (UID: \"52335cb7-2ea6-4571-a595-fc6c8c071f50\") " pod="openstack/openstackclient" Oct 08 19:58:44 crc kubenswrapper[4988]: I1008 19:58:44.808288 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/efec0d27-7968-44f2-a1f4-733078747791-openstack-config\") pod \"openstackclient\" (UID: \"efec0d27-7968-44f2-a1f4-733078747791\") " pod="openstack/openstackclient" Oct 08 19:58:44 crc kubenswrapper[4988]: I1008 19:58:44.808307 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/efec0d27-7968-44f2-a1f4-733078747791-combined-ca-bundle\") pod \"openstackclient\" (UID: \"efec0d27-7968-44f2-a1f4-733078747791\") " pod="openstack/openstackclient" Oct 08 19:58:44 crc kubenswrapper[4988]: I1008 19:58:44.808361 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/efec0d27-7968-44f2-a1f4-733078747791-openstack-config-secret\") pod \"openstackclient\" (UID: \"efec0d27-7968-44f2-a1f4-733078747791\") " pod="openstack/openstackclient" Oct 08 19:58:44 crc kubenswrapper[4988]: I1008 19:58:44.809191 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/52335cb7-2ea6-4571-a595-fc6c8c071f50-openstack-config\") pod \"openstackclient\" (UID: \"52335cb7-2ea6-4571-a595-fc6c8c071f50\") " pod="openstack/openstackclient" Oct 08 19:58:44 crc kubenswrapper[4988]: E1008 19:58:44.814018 4988 projected.go:194] Error preparing data for projected volume kube-api-access-dbs4g for pod openstack/openstackclient: failed to fetch token: serviceaccounts "openstackclient-openstackclient" is forbidden: the UID in the bound object reference (52335cb7-2ea6-4571-a595-fc6c8c071f50) does not match the UID in record. The object might have been deleted and then recreated Oct 08 19:58:44 crc kubenswrapper[4988]: E1008 19:58:44.814275 4988 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/52335cb7-2ea6-4571-a595-fc6c8c071f50-kube-api-access-dbs4g podName:52335cb7-2ea6-4571-a595-fc6c8c071f50 nodeName:}" failed. No retries permitted until 2025-10-08 19:58:45.314253237 +0000 UTC m=+6470.764096117 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-dbs4g" (UniqueName: "kubernetes.io/projected/52335cb7-2ea6-4571-a595-fc6c8c071f50-kube-api-access-dbs4g") pod "openstackclient" (UID: "52335cb7-2ea6-4571-a595-fc6c8c071f50") : failed to fetch token: serviceaccounts "openstackclient-openstackclient" is forbidden: the UID in the bound object reference (52335cb7-2ea6-4571-a595-fc6c8c071f50) does not match the UID in record. The object might have been deleted and then recreated Oct 08 19:58:44 crc kubenswrapper[4988]: I1008 19:58:44.814358 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/52335cb7-2ea6-4571-a595-fc6c8c071f50-openstack-config-secret\") pod \"openstackclient\" (UID: \"52335cb7-2ea6-4571-a595-fc6c8c071f50\") " pod="openstack/openstackclient" Oct 08 19:58:44 crc kubenswrapper[4988]: I1008 19:58:44.814939 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/52335cb7-2ea6-4571-a595-fc6c8c071f50-combined-ca-bundle\") pod \"openstackclient\" (UID: \"52335cb7-2ea6-4571-a595-fc6c8c071f50\") " pod="openstack/openstackclient" Oct 08 19:58:44 crc kubenswrapper[4988]: I1008 19:58:44.910589 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8ztjk\" (UniqueName: \"kubernetes.io/projected/efec0d27-7968-44f2-a1f4-733078747791-kube-api-access-8ztjk\") pod \"openstackclient\" (UID: \"efec0d27-7968-44f2-a1f4-733078747791\") " pod="openstack/openstackclient" Oct 08 19:58:44 crc kubenswrapper[4988]: I1008 19:58:44.910710 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/efec0d27-7968-44f2-a1f4-733078747791-openstack-config\") pod \"openstackclient\" (UID: \"efec0d27-7968-44f2-a1f4-733078747791\") " pod="openstack/openstackclient" Oct 08 19:58:44 crc kubenswrapper[4988]: I1008 19:58:44.910747 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/efec0d27-7968-44f2-a1f4-733078747791-combined-ca-bundle\") pod \"openstackclient\" (UID: \"efec0d27-7968-44f2-a1f4-733078747791\") " pod="openstack/openstackclient" Oct 08 19:58:44 crc kubenswrapper[4988]: I1008 19:58:44.910784 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/efec0d27-7968-44f2-a1f4-733078747791-openstack-config-secret\") pod \"openstackclient\" (UID: \"efec0d27-7968-44f2-a1f4-733078747791\") " pod="openstack/openstackclient" Oct 08 19:58:44 crc kubenswrapper[4988]: I1008 19:58:44.911540 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/efec0d27-7968-44f2-a1f4-733078747791-openstack-config\") pod \"openstackclient\" (UID: \"efec0d27-7968-44f2-a1f4-733078747791\") " pod="openstack/openstackclient" Oct 08 19:58:44 crc kubenswrapper[4988]: I1008 19:58:44.916169 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/efec0d27-7968-44f2-a1f4-733078747791-combined-ca-bundle\") pod \"openstackclient\" (UID: \"efec0d27-7968-44f2-a1f4-733078747791\") " pod="openstack/openstackclient" Oct 08 19:58:44 crc kubenswrapper[4988]: I1008 19:58:44.920696 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/efec0d27-7968-44f2-a1f4-733078747791-openstack-config-secret\") pod \"openstackclient\" (UID: \"efec0d27-7968-44f2-a1f4-733078747791\") " pod="openstack/openstackclient" Oct 08 19:58:44 crc kubenswrapper[4988]: I1008 19:58:44.927246 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8ztjk\" (UniqueName: \"kubernetes.io/projected/efec0d27-7968-44f2-a1f4-733078747791-kube-api-access-8ztjk\") pod \"openstackclient\" (UID: \"efec0d27-7968-44f2-a1f4-733078747791\") " pod="openstack/openstackclient" Oct 08 19:58:45 crc kubenswrapper[4988]: I1008 19:58:45.002928 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Oct 08 19:58:45 crc kubenswrapper[4988]: I1008 19:58:45.228419 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Oct 08 19:58:45 crc kubenswrapper[4988]: I1008 19:58:45.231033 4988 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openstack/openstackclient" oldPodUID="52335cb7-2ea6-4571-a595-fc6c8c071f50" podUID="efec0d27-7968-44f2-a1f4-733078747791" Oct 08 19:58:45 crc kubenswrapper[4988]: I1008 19:58:45.269484 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Oct 08 19:58:45 crc kubenswrapper[4988]: I1008 19:58:45.273017 4988 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openstack/openstackclient" oldPodUID="52335cb7-2ea6-4571-a595-fc6c8c071f50" podUID="efec0d27-7968-44f2-a1f4-733078747791" Oct 08 19:58:45 crc kubenswrapper[4988]: I1008 19:58:45.320753 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dbs4g\" (UniqueName: \"kubernetes.io/projected/52335cb7-2ea6-4571-a595-fc6c8c071f50-kube-api-access-dbs4g\") pod \"openstackclient\" (UID: \"52335cb7-2ea6-4571-a595-fc6c8c071f50\") " pod="openstack/openstackclient" Oct 08 19:58:45 crc kubenswrapper[4988]: E1008 19:58:45.324868 4988 projected.go:194] Error preparing data for projected volume kube-api-access-dbs4g for pod openstack/openstackclient: failed to fetch token: serviceaccounts "openstackclient-openstackclient" is forbidden: the UID in the bound object reference (52335cb7-2ea6-4571-a595-fc6c8c071f50) does not match the UID in record. The object might have been deleted and then recreated Oct 08 19:58:45 crc kubenswrapper[4988]: E1008 19:58:45.324932 4988 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/52335cb7-2ea6-4571-a595-fc6c8c071f50-kube-api-access-dbs4g podName:52335cb7-2ea6-4571-a595-fc6c8c071f50 nodeName:}" failed. No retries permitted until 2025-10-08 19:58:46.324913442 +0000 UTC m=+6471.774756212 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-dbs4g" (UniqueName: "kubernetes.io/projected/52335cb7-2ea6-4571-a595-fc6c8c071f50-kube-api-access-dbs4g") pod "openstackclient" (UID: "52335cb7-2ea6-4571-a595-fc6c8c071f50") : failed to fetch token: serviceaccounts "openstackclient-openstackclient" is forbidden: the UID in the bound object reference (52335cb7-2ea6-4571-a595-fc6c8c071f50) does not match the UID in record. The object might have been deleted and then recreated Oct 08 19:58:45 crc kubenswrapper[4988]: I1008 19:58:45.422695 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/52335cb7-2ea6-4571-a595-fc6c8c071f50-openstack-config\") pod \"52335cb7-2ea6-4571-a595-fc6c8c071f50\" (UID: \"52335cb7-2ea6-4571-a595-fc6c8c071f50\") " Oct 08 19:58:45 crc kubenswrapper[4988]: I1008 19:58:45.423177 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/52335cb7-2ea6-4571-a595-fc6c8c071f50-openstack-config-secret\") pod \"52335cb7-2ea6-4571-a595-fc6c8c071f50\" (UID: \"52335cb7-2ea6-4571-a595-fc6c8c071f50\") " Oct 08 19:58:45 crc kubenswrapper[4988]: I1008 19:58:45.423221 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/52335cb7-2ea6-4571-a595-fc6c8c071f50-combined-ca-bundle\") pod \"52335cb7-2ea6-4571-a595-fc6c8c071f50\" (UID: \"52335cb7-2ea6-4571-a595-fc6c8c071f50\") " Oct 08 19:58:45 crc kubenswrapper[4988]: I1008 19:58:45.423226 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/52335cb7-2ea6-4571-a595-fc6c8c071f50-openstack-config" (OuterVolumeSpecName: "openstack-config") pod "52335cb7-2ea6-4571-a595-fc6c8c071f50" (UID: "52335cb7-2ea6-4571-a595-fc6c8c071f50"). InnerVolumeSpecName "openstack-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 19:58:45 crc kubenswrapper[4988]: I1008 19:58:45.423935 4988 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dbs4g\" (UniqueName: \"kubernetes.io/projected/52335cb7-2ea6-4571-a595-fc6c8c071f50-kube-api-access-dbs4g\") on node \"crc\" DevicePath \"\"" Oct 08 19:58:45 crc kubenswrapper[4988]: I1008 19:58:45.423948 4988 reconciler_common.go:293] "Volume detached for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/52335cb7-2ea6-4571-a595-fc6c8c071f50-openstack-config\") on node \"crc\" DevicePath \"\"" Oct 08 19:58:45 crc kubenswrapper[4988]: I1008 19:58:45.429644 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/52335cb7-2ea6-4571-a595-fc6c8c071f50-openstack-config-secret" (OuterVolumeSpecName: "openstack-config-secret") pod "52335cb7-2ea6-4571-a595-fc6c8c071f50" (UID: "52335cb7-2ea6-4571-a595-fc6c8c071f50"). InnerVolumeSpecName "openstack-config-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 19:58:45 crc kubenswrapper[4988]: I1008 19:58:45.444576 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/52335cb7-2ea6-4571-a595-fc6c8c071f50-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "52335cb7-2ea6-4571-a595-fc6c8c071f50" (UID: "52335cb7-2ea6-4571-a595-fc6c8c071f50"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 19:58:45 crc kubenswrapper[4988]: I1008 19:58:45.525260 4988 reconciler_common.go:293] "Volume detached for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/52335cb7-2ea6-4571-a595-fc6c8c071f50-openstack-config-secret\") on node \"crc\" DevicePath \"\"" Oct 08 19:58:45 crc kubenswrapper[4988]: I1008 19:58:45.525291 4988 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/52335cb7-2ea6-4571-a595-fc6c8c071f50-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 19:58:45 crc kubenswrapper[4988]: I1008 19:58:45.597711 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Oct 08 19:58:45 crc kubenswrapper[4988]: I1008 19:58:45.833207 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/prometheus-metric-storage-0"] Oct 08 19:58:45 crc kubenswrapper[4988]: I1008 19:58:45.833840 4988 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/prometheus-metric-storage-0" podUID="5cd96092-93f6-43a6-bec4-cceaba7bb8af" containerName="prometheus" containerID="cri-o://5e78b0fe3541a2320e39f5e23ea42803ccc4bbfe2e9239e7b173d033f105ae28" gracePeriod=600 Oct 08 19:58:45 crc kubenswrapper[4988]: I1008 19:58:45.833938 4988 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/prometheus-metric-storage-0" podUID="5cd96092-93f6-43a6-bec4-cceaba7bb8af" containerName="config-reloader" containerID="cri-o://bdf26b13144c9af7cb42452ac56fbb0dd91448ac6b82f422d8ff9d3477f4d477" gracePeriod=600 Oct 08 19:58:45 crc kubenswrapper[4988]: I1008 19:58:45.833932 4988 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/prometheus-metric-storage-0" podUID="5cd96092-93f6-43a6-bec4-cceaba7bb8af" containerName="thanos-sidecar" containerID="cri-o://286424932a0992c4d6fbc7f99d9874252b3d97e9e5d669ba59a770288e4f635d" gracePeriod=600 Oct 08 19:58:46 crc kubenswrapper[4988]: I1008 19:58:46.242041 4988 generic.go:334] "Generic (PLEG): container finished" podID="5cd96092-93f6-43a6-bec4-cceaba7bb8af" containerID="286424932a0992c4d6fbc7f99d9874252b3d97e9e5d669ba59a770288e4f635d" exitCode=0 Oct 08 19:58:46 crc kubenswrapper[4988]: I1008 19:58:46.242078 4988 generic.go:334] "Generic (PLEG): container finished" podID="5cd96092-93f6-43a6-bec4-cceaba7bb8af" containerID="bdf26b13144c9af7cb42452ac56fbb0dd91448ac6b82f422d8ff9d3477f4d477" exitCode=0 Oct 08 19:58:46 crc kubenswrapper[4988]: I1008 19:58:46.242090 4988 generic.go:334] "Generic (PLEG): container finished" podID="5cd96092-93f6-43a6-bec4-cceaba7bb8af" containerID="5e78b0fe3541a2320e39f5e23ea42803ccc4bbfe2e9239e7b173d033f105ae28" exitCode=0 Oct 08 19:58:46 crc kubenswrapper[4988]: I1008 19:58:46.242118 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"5cd96092-93f6-43a6-bec4-cceaba7bb8af","Type":"ContainerDied","Data":"286424932a0992c4d6fbc7f99d9874252b3d97e9e5d669ba59a770288e4f635d"} Oct 08 19:58:46 crc kubenswrapper[4988]: I1008 19:58:46.242169 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"5cd96092-93f6-43a6-bec4-cceaba7bb8af","Type":"ContainerDied","Data":"bdf26b13144c9af7cb42452ac56fbb0dd91448ac6b82f422d8ff9d3477f4d477"} Oct 08 19:58:46 crc kubenswrapper[4988]: I1008 19:58:46.242184 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"5cd96092-93f6-43a6-bec4-cceaba7bb8af","Type":"ContainerDied","Data":"5e78b0fe3541a2320e39f5e23ea42803ccc4bbfe2e9239e7b173d033f105ae28"} Oct 08 19:58:46 crc kubenswrapper[4988]: I1008 19:58:46.243862 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Oct 08 19:58:46 crc kubenswrapper[4988]: I1008 19:58:46.243864 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"efec0d27-7968-44f2-a1f4-733078747791","Type":"ContainerStarted","Data":"62f3756e8d2829e44ef985c0306739ea72295ffd5603e8633318fcec75c5e007"} Oct 08 19:58:46 crc kubenswrapper[4988]: I1008 19:58:46.243929 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"efec0d27-7968-44f2-a1f4-733078747791","Type":"ContainerStarted","Data":"cd3f578456c615d2fd1ce44abe1d9d801ab46d9900dca48e3ef5a2e0721a3c8a"} Oct 08 19:58:46 crc kubenswrapper[4988]: I1008 19:58:46.263706 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstackclient" podStartSLOduration=2.263684285 podStartE2EDuration="2.263684285s" podCreationTimestamp="2025-10-08 19:58:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 19:58:46.262794476 +0000 UTC m=+6471.712637436" watchObservedRunningTime="2025-10-08 19:58:46.263684285 +0000 UTC m=+6471.713527055" Oct 08 19:58:46 crc kubenswrapper[4988]: I1008 19:58:46.266714 4988 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openstack/openstackclient" oldPodUID="52335cb7-2ea6-4571-a595-fc6c8c071f50" podUID="efec0d27-7968-44f2-a1f4-733078747791" Oct 08 19:58:46 crc kubenswrapper[4988]: I1008 19:58:46.542437 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 08 19:58:46 crc kubenswrapper[4988]: I1008 19:58:46.549219 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 08 19:58:46 crc kubenswrapper[4988]: I1008 19:58:46.553368 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 08 19:58:46 crc kubenswrapper[4988]: I1008 19:58:46.555430 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 08 19:58:46 crc kubenswrapper[4988]: I1008 19:58:46.577221 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 08 19:58:46 crc kubenswrapper[4988]: I1008 19:58:46.654845 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1e77ab25-7879-4cb9-ae44-0e750ecb8123-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"1e77ab25-7879-4cb9-ae44-0e750ecb8123\") " pod="openstack/ceilometer-0" Oct 08 19:58:46 crc kubenswrapper[4988]: I1008 19:58:46.655240 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1e77ab25-7879-4cb9-ae44-0e750ecb8123-scripts\") pod \"ceilometer-0\" (UID: \"1e77ab25-7879-4cb9-ae44-0e750ecb8123\") " pod="openstack/ceilometer-0" Oct 08 19:58:46 crc kubenswrapper[4988]: I1008 19:58:46.655306 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jgwqx\" (UniqueName: \"kubernetes.io/projected/1e77ab25-7879-4cb9-ae44-0e750ecb8123-kube-api-access-jgwqx\") pod \"ceilometer-0\" (UID: \"1e77ab25-7879-4cb9-ae44-0e750ecb8123\") " pod="openstack/ceilometer-0" Oct 08 19:58:46 crc kubenswrapper[4988]: I1008 19:58:46.655429 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1e77ab25-7879-4cb9-ae44-0e750ecb8123-config-data\") pod \"ceilometer-0\" (UID: \"1e77ab25-7879-4cb9-ae44-0e750ecb8123\") " pod="openstack/ceilometer-0" Oct 08 19:58:46 crc kubenswrapper[4988]: I1008 19:58:46.655482 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/1e77ab25-7879-4cb9-ae44-0e750ecb8123-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"1e77ab25-7879-4cb9-ae44-0e750ecb8123\") " pod="openstack/ceilometer-0" Oct 08 19:58:46 crc kubenswrapper[4988]: I1008 19:58:46.655504 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1e77ab25-7879-4cb9-ae44-0e750ecb8123-run-httpd\") pod \"ceilometer-0\" (UID: \"1e77ab25-7879-4cb9-ae44-0e750ecb8123\") " pod="openstack/ceilometer-0" Oct 08 19:58:46 crc kubenswrapper[4988]: I1008 19:58:46.655527 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1e77ab25-7879-4cb9-ae44-0e750ecb8123-log-httpd\") pod \"ceilometer-0\" (UID: \"1e77ab25-7879-4cb9-ae44-0e750ecb8123\") " pod="openstack/ceilometer-0" Oct 08 19:58:46 crc kubenswrapper[4988]: I1008 19:58:46.757343 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jgwqx\" (UniqueName: \"kubernetes.io/projected/1e77ab25-7879-4cb9-ae44-0e750ecb8123-kube-api-access-jgwqx\") pod \"ceilometer-0\" (UID: \"1e77ab25-7879-4cb9-ae44-0e750ecb8123\") " pod="openstack/ceilometer-0" Oct 08 19:58:46 crc kubenswrapper[4988]: I1008 19:58:46.757447 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1e77ab25-7879-4cb9-ae44-0e750ecb8123-config-data\") pod \"ceilometer-0\" (UID: \"1e77ab25-7879-4cb9-ae44-0e750ecb8123\") " pod="openstack/ceilometer-0" Oct 08 19:58:46 crc kubenswrapper[4988]: I1008 19:58:46.757496 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/1e77ab25-7879-4cb9-ae44-0e750ecb8123-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"1e77ab25-7879-4cb9-ae44-0e750ecb8123\") " pod="openstack/ceilometer-0" Oct 08 19:58:46 crc kubenswrapper[4988]: I1008 19:58:46.757519 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1e77ab25-7879-4cb9-ae44-0e750ecb8123-run-httpd\") pod \"ceilometer-0\" (UID: \"1e77ab25-7879-4cb9-ae44-0e750ecb8123\") " pod="openstack/ceilometer-0" Oct 08 19:58:46 crc kubenswrapper[4988]: I1008 19:58:46.757539 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1e77ab25-7879-4cb9-ae44-0e750ecb8123-log-httpd\") pod \"ceilometer-0\" (UID: \"1e77ab25-7879-4cb9-ae44-0e750ecb8123\") " pod="openstack/ceilometer-0" Oct 08 19:58:46 crc kubenswrapper[4988]: I1008 19:58:46.757585 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1e77ab25-7879-4cb9-ae44-0e750ecb8123-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"1e77ab25-7879-4cb9-ae44-0e750ecb8123\") " pod="openstack/ceilometer-0" Oct 08 19:58:46 crc kubenswrapper[4988]: I1008 19:58:46.757606 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1e77ab25-7879-4cb9-ae44-0e750ecb8123-scripts\") pod \"ceilometer-0\" (UID: \"1e77ab25-7879-4cb9-ae44-0e750ecb8123\") " pod="openstack/ceilometer-0" Oct 08 19:58:46 crc kubenswrapper[4988]: I1008 19:58:46.758729 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1e77ab25-7879-4cb9-ae44-0e750ecb8123-run-httpd\") pod \"ceilometer-0\" (UID: \"1e77ab25-7879-4cb9-ae44-0e750ecb8123\") " pod="openstack/ceilometer-0" Oct 08 19:58:46 crc kubenswrapper[4988]: I1008 19:58:46.759058 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1e77ab25-7879-4cb9-ae44-0e750ecb8123-log-httpd\") pod \"ceilometer-0\" (UID: \"1e77ab25-7879-4cb9-ae44-0e750ecb8123\") " pod="openstack/ceilometer-0" Oct 08 19:58:46 crc kubenswrapper[4988]: I1008 19:58:46.763204 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/1e77ab25-7879-4cb9-ae44-0e750ecb8123-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"1e77ab25-7879-4cb9-ae44-0e750ecb8123\") " pod="openstack/ceilometer-0" Oct 08 19:58:46 crc kubenswrapper[4988]: I1008 19:58:46.763222 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1e77ab25-7879-4cb9-ae44-0e750ecb8123-config-data\") pod \"ceilometer-0\" (UID: \"1e77ab25-7879-4cb9-ae44-0e750ecb8123\") " pod="openstack/ceilometer-0" Oct 08 19:58:46 crc kubenswrapper[4988]: I1008 19:58:46.764186 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1e77ab25-7879-4cb9-ae44-0e750ecb8123-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"1e77ab25-7879-4cb9-ae44-0e750ecb8123\") " pod="openstack/ceilometer-0" Oct 08 19:58:46 crc kubenswrapper[4988]: I1008 19:58:46.777653 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1e77ab25-7879-4cb9-ae44-0e750ecb8123-scripts\") pod \"ceilometer-0\" (UID: \"1e77ab25-7879-4cb9-ae44-0e750ecb8123\") " pod="openstack/ceilometer-0" Oct 08 19:58:46 crc kubenswrapper[4988]: I1008 19:58:46.779992 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jgwqx\" (UniqueName: \"kubernetes.io/projected/1e77ab25-7879-4cb9-ae44-0e750ecb8123-kube-api-access-jgwqx\") pod \"ceilometer-0\" (UID: \"1e77ab25-7879-4cb9-ae44-0e750ecb8123\") " pod="openstack/ceilometer-0" Oct 08 19:58:46 crc kubenswrapper[4988]: I1008 19:58:46.872076 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 08 19:58:47 crc kubenswrapper[4988]: I1008 19:58:47.014476 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Oct 08 19:58:47 crc kubenswrapper[4988]: I1008 19:58:47.018339 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Oct 08 19:58:47 crc kubenswrapper[4988]: I1008 19:58:47.072247 4988 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openstack/openstackclient" oldPodUID="2d58f5a2-1a02-4152-95ec-22cc390d8ae5" podUID="efec0d27-7968-44f2-a1f4-733078747791" Oct 08 19:58:47 crc kubenswrapper[4988]: I1008 19:58:47.174996 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/5cd96092-93f6-43a6-bec4-cceaba7bb8af-tls-assets\") pod \"5cd96092-93f6-43a6-bec4-cceaba7bb8af\" (UID: \"5cd96092-93f6-43a6-bec4-cceaba7bb8af\") " Oct 08 19:58:47 crc kubenswrapper[4988]: I1008 19:58:47.175173 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"prometheus-metric-storage-db\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-e634a8ba-6c66-4bab-9b9c-79158ccf2cf8\") pod \"5cd96092-93f6-43a6-bec4-cceaba7bb8af\" (UID: \"5cd96092-93f6-43a6-bec4-cceaba7bb8af\") " Oct 08 19:58:47 crc kubenswrapper[4988]: I1008 19:58:47.175275 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lc784\" (UniqueName: \"kubernetes.io/projected/2d58f5a2-1a02-4152-95ec-22cc390d8ae5-kube-api-access-lc784\") pod \"2d58f5a2-1a02-4152-95ec-22cc390d8ae5\" (UID: \"2d58f5a2-1a02-4152-95ec-22cc390d8ae5\") " Oct 08 19:58:47 crc kubenswrapper[4988]: I1008 19:58:47.175378 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/5cd96092-93f6-43a6-bec4-cceaba7bb8af-config\") pod \"5cd96092-93f6-43a6-bec4-cceaba7bb8af\" (UID: \"5cd96092-93f6-43a6-bec4-cceaba7bb8af\") " Oct 08 19:58:47 crc kubenswrapper[4988]: I1008 19:58:47.175614 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/2d58f5a2-1a02-4152-95ec-22cc390d8ae5-openstack-config-secret\") pod \"2d58f5a2-1a02-4152-95ec-22cc390d8ae5\" (UID: \"2d58f5a2-1a02-4152-95ec-22cc390d8ae5\") " Oct 08 19:58:47 crc kubenswrapper[4988]: I1008 19:58:47.175662 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/5cd96092-93f6-43a6-bec4-cceaba7bb8af-prometheus-metric-storage-rulefiles-0\") pod \"5cd96092-93f6-43a6-bec4-cceaba7bb8af\" (UID: \"5cd96092-93f6-43a6-bec4-cceaba7bb8af\") " Oct 08 19:58:47 crc kubenswrapper[4988]: I1008 19:58:47.175684 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/5cd96092-93f6-43a6-bec4-cceaba7bb8af-config-out\") pod \"5cd96092-93f6-43a6-bec4-cceaba7bb8af\" (UID: \"5cd96092-93f6-43a6-bec4-cceaba7bb8af\") " Oct 08 19:58:47 crc kubenswrapper[4988]: I1008 19:58:47.175700 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/5cd96092-93f6-43a6-bec4-cceaba7bb8af-thanos-prometheus-http-client-file\") pod \"5cd96092-93f6-43a6-bec4-cceaba7bb8af\" (UID: \"5cd96092-93f6-43a6-bec4-cceaba7bb8af\") " Oct 08 19:58:47 crc kubenswrapper[4988]: I1008 19:58:47.175719 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/2d58f5a2-1a02-4152-95ec-22cc390d8ae5-openstack-config\") pod \"2d58f5a2-1a02-4152-95ec-22cc390d8ae5\" (UID: \"2d58f5a2-1a02-4152-95ec-22cc390d8ae5\") " Oct 08 19:58:47 crc kubenswrapper[4988]: I1008 19:58:47.175743 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zmzrj\" (UniqueName: \"kubernetes.io/projected/5cd96092-93f6-43a6-bec4-cceaba7bb8af-kube-api-access-zmzrj\") pod \"5cd96092-93f6-43a6-bec4-cceaba7bb8af\" (UID: \"5cd96092-93f6-43a6-bec4-cceaba7bb8af\") " Oct 08 19:58:47 crc kubenswrapper[4988]: I1008 19:58:47.175793 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2d58f5a2-1a02-4152-95ec-22cc390d8ae5-combined-ca-bundle\") pod \"2d58f5a2-1a02-4152-95ec-22cc390d8ae5\" (UID: \"2d58f5a2-1a02-4152-95ec-22cc390d8ae5\") " Oct 08 19:58:47 crc kubenswrapper[4988]: I1008 19:58:47.175842 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/5cd96092-93f6-43a6-bec4-cceaba7bb8af-web-config\") pod \"5cd96092-93f6-43a6-bec4-cceaba7bb8af\" (UID: \"5cd96092-93f6-43a6-bec4-cceaba7bb8af\") " Oct 08 19:58:47 crc kubenswrapper[4988]: I1008 19:58:47.177013 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5cd96092-93f6-43a6-bec4-cceaba7bb8af-prometheus-metric-storage-rulefiles-0" (OuterVolumeSpecName: "prometheus-metric-storage-rulefiles-0") pod "5cd96092-93f6-43a6-bec4-cceaba7bb8af" (UID: "5cd96092-93f6-43a6-bec4-cceaba7bb8af"). InnerVolumeSpecName "prometheus-metric-storage-rulefiles-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 19:58:47 crc kubenswrapper[4988]: I1008 19:58:47.180476 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5cd96092-93f6-43a6-bec4-cceaba7bb8af-thanos-prometheus-http-client-file" (OuterVolumeSpecName: "thanos-prometheus-http-client-file") pod "5cd96092-93f6-43a6-bec4-cceaba7bb8af" (UID: "5cd96092-93f6-43a6-bec4-cceaba7bb8af"). InnerVolumeSpecName "thanos-prometheus-http-client-file". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 19:58:47 crc kubenswrapper[4988]: I1008 19:58:47.180671 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5cd96092-93f6-43a6-bec4-cceaba7bb8af-kube-api-access-zmzrj" (OuterVolumeSpecName: "kube-api-access-zmzrj") pod "5cd96092-93f6-43a6-bec4-cceaba7bb8af" (UID: "5cd96092-93f6-43a6-bec4-cceaba7bb8af"). InnerVolumeSpecName "kube-api-access-zmzrj". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 19:58:47 crc kubenswrapper[4988]: I1008 19:58:47.180730 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5cd96092-93f6-43a6-bec4-cceaba7bb8af-tls-assets" (OuterVolumeSpecName: "tls-assets") pod "5cd96092-93f6-43a6-bec4-cceaba7bb8af" (UID: "5cd96092-93f6-43a6-bec4-cceaba7bb8af"). InnerVolumeSpecName "tls-assets". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 19:58:47 crc kubenswrapper[4988]: I1008 19:58:47.181165 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5cd96092-93f6-43a6-bec4-cceaba7bb8af-config" (OuterVolumeSpecName: "config") pod "5cd96092-93f6-43a6-bec4-cceaba7bb8af" (UID: "5cd96092-93f6-43a6-bec4-cceaba7bb8af"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 19:58:47 crc kubenswrapper[4988]: I1008 19:58:47.182808 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2d58f5a2-1a02-4152-95ec-22cc390d8ae5-kube-api-access-lc784" (OuterVolumeSpecName: "kube-api-access-lc784") pod "2d58f5a2-1a02-4152-95ec-22cc390d8ae5" (UID: "2d58f5a2-1a02-4152-95ec-22cc390d8ae5"). InnerVolumeSpecName "kube-api-access-lc784". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 19:58:47 crc kubenswrapper[4988]: I1008 19:58:47.196629 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5cd96092-93f6-43a6-bec4-cceaba7bb8af-config-out" (OuterVolumeSpecName: "config-out") pod "5cd96092-93f6-43a6-bec4-cceaba7bb8af" (UID: "5cd96092-93f6-43a6-bec4-cceaba7bb8af"). InnerVolumeSpecName "config-out". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 19:58:47 crc kubenswrapper[4988]: I1008 19:58:47.214277 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2d58f5a2-1a02-4152-95ec-22cc390d8ae5-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "2d58f5a2-1a02-4152-95ec-22cc390d8ae5" (UID: "2d58f5a2-1a02-4152-95ec-22cc390d8ae5"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 19:58:47 crc kubenswrapper[4988]: I1008 19:58:47.226519 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-e634a8ba-6c66-4bab-9b9c-79158ccf2cf8" (OuterVolumeSpecName: "prometheus-metric-storage-db") pod "5cd96092-93f6-43a6-bec4-cceaba7bb8af" (UID: "5cd96092-93f6-43a6-bec4-cceaba7bb8af"). InnerVolumeSpecName "pvc-e634a8ba-6c66-4bab-9b9c-79158ccf2cf8". PluginName "kubernetes.io/csi", VolumeGidValue "" Oct 08 19:58:47 crc kubenswrapper[4988]: I1008 19:58:47.231643 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2d58f5a2-1a02-4152-95ec-22cc390d8ae5-openstack-config" (OuterVolumeSpecName: "openstack-config") pod "2d58f5a2-1a02-4152-95ec-22cc390d8ae5" (UID: "2d58f5a2-1a02-4152-95ec-22cc390d8ae5"). InnerVolumeSpecName "openstack-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 19:58:47 crc kubenswrapper[4988]: I1008 19:58:47.243540 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5cd96092-93f6-43a6-bec4-cceaba7bb8af-web-config" (OuterVolumeSpecName: "web-config") pod "5cd96092-93f6-43a6-bec4-cceaba7bb8af" (UID: "5cd96092-93f6-43a6-bec4-cceaba7bb8af"). InnerVolumeSpecName "web-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 19:58:47 crc kubenswrapper[4988]: I1008 19:58:47.257460 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2d58f5a2-1a02-4152-95ec-22cc390d8ae5-openstack-config-secret" (OuterVolumeSpecName: "openstack-config-secret") pod "2d58f5a2-1a02-4152-95ec-22cc390d8ae5" (UID: "2d58f5a2-1a02-4152-95ec-22cc390d8ae5"). InnerVolumeSpecName "openstack-config-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 19:58:47 crc kubenswrapper[4988]: I1008 19:58:47.259111 4988 generic.go:334] "Generic (PLEG): container finished" podID="2d58f5a2-1a02-4152-95ec-22cc390d8ae5" containerID="b7ea1ae1a08e4188d78ecf340ffa61f27ee87f49871711bb009130301dbdf138" exitCode=137 Oct 08 19:58:47 crc kubenswrapper[4988]: I1008 19:58:47.259262 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Oct 08 19:58:47 crc kubenswrapper[4988]: I1008 19:58:47.267873 4988 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openstack/openstackclient" oldPodUID="2d58f5a2-1a02-4152-95ec-22cc390d8ae5" podUID="efec0d27-7968-44f2-a1f4-733078747791" Oct 08 19:58:47 crc kubenswrapper[4988]: I1008 19:58:47.268125 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Oct 08 19:58:47 crc kubenswrapper[4988]: I1008 19:58:47.278218 4988 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lc784\" (UniqueName: \"kubernetes.io/projected/2d58f5a2-1a02-4152-95ec-22cc390d8ae5-kube-api-access-lc784\") on node \"crc\" DevicePath \"\"" Oct 08 19:58:47 crc kubenswrapper[4988]: I1008 19:58:47.278260 4988 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/5cd96092-93f6-43a6-bec4-cceaba7bb8af-config\") on node \"crc\" DevicePath \"\"" Oct 08 19:58:47 crc kubenswrapper[4988]: I1008 19:58:47.278271 4988 reconciler_common.go:293] "Volume detached for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/2d58f5a2-1a02-4152-95ec-22cc390d8ae5-openstack-config-secret\") on node \"crc\" DevicePath \"\"" Oct 08 19:58:47 crc kubenswrapper[4988]: I1008 19:58:47.278280 4988 reconciler_common.go:293] "Volume detached for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/2d58f5a2-1a02-4152-95ec-22cc390d8ae5-openstack-config\") on node \"crc\" DevicePath \"\"" Oct 08 19:58:47 crc kubenswrapper[4988]: I1008 19:58:47.278293 4988 reconciler_common.go:293] "Volume detached for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/5cd96092-93f6-43a6-bec4-cceaba7bb8af-prometheus-metric-storage-rulefiles-0\") on node \"crc\" DevicePath \"\"" Oct 08 19:58:47 crc kubenswrapper[4988]: I1008 19:58:47.278307 4988 reconciler_common.go:293] "Volume detached for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/5cd96092-93f6-43a6-bec4-cceaba7bb8af-config-out\") on node \"crc\" DevicePath \"\"" Oct 08 19:58:47 crc kubenswrapper[4988]: I1008 19:58:47.278318 4988 reconciler_common.go:293] "Volume detached for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/5cd96092-93f6-43a6-bec4-cceaba7bb8af-thanos-prometheus-http-client-file\") on node \"crc\" DevicePath \"\"" Oct 08 19:58:47 crc kubenswrapper[4988]: I1008 19:58:47.278331 4988 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zmzrj\" (UniqueName: \"kubernetes.io/projected/5cd96092-93f6-43a6-bec4-cceaba7bb8af-kube-api-access-zmzrj\") on node \"crc\" DevicePath \"\"" Oct 08 19:58:47 crc kubenswrapper[4988]: I1008 19:58:47.278343 4988 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2d58f5a2-1a02-4152-95ec-22cc390d8ae5-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 19:58:47 crc kubenswrapper[4988]: I1008 19:58:47.278354 4988 reconciler_common.go:293] "Volume detached for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/5cd96092-93f6-43a6-bec4-cceaba7bb8af-web-config\") on node \"crc\" DevicePath \"\"" Oct 08 19:58:47 crc kubenswrapper[4988]: I1008 19:58:47.278367 4988 reconciler_common.go:293] "Volume detached for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/5cd96092-93f6-43a6-bec4-cceaba7bb8af-tls-assets\") on node \"crc\" DevicePath \"\"" Oct 08 19:58:47 crc kubenswrapper[4988]: I1008 19:58:47.278415 4988 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"pvc-e634a8ba-6c66-4bab-9b9c-79158ccf2cf8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-e634a8ba-6c66-4bab-9b9c-79158ccf2cf8\") on node \"crc\" " Oct 08 19:58:47 crc kubenswrapper[4988]: I1008 19:58:47.280018 4988 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2d58f5a2-1a02-4152-95ec-22cc390d8ae5" path="/var/lib/kubelet/pods/2d58f5a2-1a02-4152-95ec-22cc390d8ae5/volumes" Oct 08 19:58:47 crc kubenswrapper[4988]: I1008 19:58:47.281668 4988 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="52335cb7-2ea6-4571-a595-fc6c8c071f50" path="/var/lib/kubelet/pods/52335cb7-2ea6-4571-a595-fc6c8c071f50/volumes" Oct 08 19:58:47 crc kubenswrapper[4988]: I1008 19:58:47.283350 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"5cd96092-93f6-43a6-bec4-cceaba7bb8af","Type":"ContainerDied","Data":"de94102c766711eedd4c2ecd78f0b44c44a097ab60e33aa977f617bfd8afe66c"} Oct 08 19:58:47 crc kubenswrapper[4988]: I1008 19:58:47.295018 4988 scope.go:117] "RemoveContainer" containerID="b7ea1ae1a08e4188d78ecf340ffa61f27ee87f49871711bb009130301dbdf138" Oct 08 19:58:47 crc kubenswrapper[4988]: I1008 19:58:47.321550 4988 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openstack/openstackclient" oldPodUID="2d58f5a2-1a02-4152-95ec-22cc390d8ae5" podUID="efec0d27-7968-44f2-a1f4-733078747791" Oct 08 19:58:47 crc kubenswrapper[4988]: I1008 19:58:47.324260 4988 scope.go:117] "RemoveContainer" containerID="b7ea1ae1a08e4188d78ecf340ffa61f27ee87f49871711bb009130301dbdf138" Oct 08 19:58:47 crc kubenswrapper[4988]: E1008 19:58:47.324742 4988 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b7ea1ae1a08e4188d78ecf340ffa61f27ee87f49871711bb009130301dbdf138\": container with ID starting with b7ea1ae1a08e4188d78ecf340ffa61f27ee87f49871711bb009130301dbdf138 not found: ID does not exist" containerID="b7ea1ae1a08e4188d78ecf340ffa61f27ee87f49871711bb009130301dbdf138" Oct 08 19:58:47 crc kubenswrapper[4988]: I1008 19:58:47.324796 4988 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b7ea1ae1a08e4188d78ecf340ffa61f27ee87f49871711bb009130301dbdf138"} err="failed to get container status \"b7ea1ae1a08e4188d78ecf340ffa61f27ee87f49871711bb009130301dbdf138\": rpc error: code = NotFound desc = could not find container \"b7ea1ae1a08e4188d78ecf340ffa61f27ee87f49871711bb009130301dbdf138\": container with ID starting with b7ea1ae1a08e4188d78ecf340ffa61f27ee87f49871711bb009130301dbdf138 not found: ID does not exist" Oct 08 19:58:47 crc kubenswrapper[4988]: I1008 19:58:47.324870 4988 scope.go:117] "RemoveContainer" containerID="286424932a0992c4d6fbc7f99d9874252b3d97e9e5d669ba59a770288e4f635d" Oct 08 19:58:47 crc kubenswrapper[4988]: I1008 19:58:47.367223 4988 csi_attacher.go:630] kubernetes.io/csi: attacher.UnmountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping UnmountDevice... Oct 08 19:58:47 crc kubenswrapper[4988]: I1008 19:58:47.367580 4988 operation_generator.go:917] UnmountDevice succeeded for volume "pvc-e634a8ba-6c66-4bab-9b9c-79158ccf2cf8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-e634a8ba-6c66-4bab-9b9c-79158ccf2cf8") on node "crc" Oct 08 19:58:47 crc kubenswrapper[4988]: I1008 19:58:47.384926 4988 reconciler_common.go:293] "Volume detached for volume \"pvc-e634a8ba-6c66-4bab-9b9c-79158ccf2cf8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-e634a8ba-6c66-4bab-9b9c-79158ccf2cf8\") on node \"crc\" DevicePath \"\"" Oct 08 19:58:47 crc kubenswrapper[4988]: I1008 19:58:47.400817 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/prometheus-metric-storage-0"] Oct 08 19:58:47 crc kubenswrapper[4988]: I1008 19:58:47.403089 4988 scope.go:117] "RemoveContainer" containerID="bdf26b13144c9af7cb42452ac56fbb0dd91448ac6b82f422d8ff9d3477f4d477" Oct 08 19:58:47 crc kubenswrapper[4988]: I1008 19:58:47.431824 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 08 19:58:47 crc kubenswrapper[4988]: I1008 19:58:47.435423 4988 scope.go:117] "RemoveContainer" containerID="5e78b0fe3541a2320e39f5e23ea42803ccc4bbfe2e9239e7b173d033f105ae28" Oct 08 19:58:47 crc kubenswrapper[4988]: I1008 19:58:47.441763 4988 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/prometheus-metric-storage-0"] Oct 08 19:58:47 crc kubenswrapper[4988]: I1008 19:58:47.447685 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/prometheus-metric-storage-0"] Oct 08 19:58:47 crc kubenswrapper[4988]: E1008 19:58:47.448179 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5cd96092-93f6-43a6-bec4-cceaba7bb8af" containerName="init-config-reloader" Oct 08 19:58:47 crc kubenswrapper[4988]: I1008 19:58:47.448192 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="5cd96092-93f6-43a6-bec4-cceaba7bb8af" containerName="init-config-reloader" Oct 08 19:58:47 crc kubenswrapper[4988]: E1008 19:58:47.448217 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5cd96092-93f6-43a6-bec4-cceaba7bb8af" containerName="thanos-sidecar" Oct 08 19:58:47 crc kubenswrapper[4988]: I1008 19:58:47.448223 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="5cd96092-93f6-43a6-bec4-cceaba7bb8af" containerName="thanos-sidecar" Oct 08 19:58:47 crc kubenswrapper[4988]: E1008 19:58:47.448238 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5cd96092-93f6-43a6-bec4-cceaba7bb8af" containerName="config-reloader" Oct 08 19:58:47 crc kubenswrapper[4988]: I1008 19:58:47.448244 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="5cd96092-93f6-43a6-bec4-cceaba7bb8af" containerName="config-reloader" Oct 08 19:58:47 crc kubenswrapper[4988]: E1008 19:58:47.448263 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5cd96092-93f6-43a6-bec4-cceaba7bb8af" containerName="prometheus" Oct 08 19:58:47 crc kubenswrapper[4988]: I1008 19:58:47.448268 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="5cd96092-93f6-43a6-bec4-cceaba7bb8af" containerName="prometheus" Oct 08 19:58:47 crc kubenswrapper[4988]: I1008 19:58:47.448478 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="5cd96092-93f6-43a6-bec4-cceaba7bb8af" containerName="config-reloader" Oct 08 19:58:47 crc kubenswrapper[4988]: I1008 19:58:47.448501 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="5cd96092-93f6-43a6-bec4-cceaba7bb8af" containerName="prometheus" Oct 08 19:58:47 crc kubenswrapper[4988]: I1008 19:58:47.448514 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="5cd96092-93f6-43a6-bec4-cceaba7bb8af" containerName="thanos-sidecar" Oct 08 19:58:47 crc kubenswrapper[4988]: I1008 19:58:47.450462 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Oct 08 19:58:47 crc kubenswrapper[4988]: I1008 19:58:47.453185 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"metric-storage-prometheus-dockercfg-fhmvc" Oct 08 19:58:47 crc kubenswrapper[4988]: I1008 19:58:47.453425 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-metric-storage-prometheus-svc" Oct 08 19:58:47 crc kubenswrapper[4988]: I1008 19:58:47.453767 4988 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"prometheus-metric-storage-rulefiles-0" Oct 08 19:58:47 crc kubenswrapper[4988]: I1008 19:58:47.454009 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage" Oct 08 19:58:47 crc kubenswrapper[4988]: I1008 19:58:47.454211 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-web-config" Oct 08 19:58:47 crc kubenswrapper[4988]: I1008 19:58:47.454358 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/prometheus-metric-storage-0"] Oct 08 19:58:47 crc kubenswrapper[4988]: I1008 19:58:47.454386 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-thanos-prometheus-http-client-file" Oct 08 19:58:47 crc kubenswrapper[4988]: I1008 19:58:47.461496 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-tls-assets-0" Oct 08 19:58:47 crc kubenswrapper[4988]: I1008 19:58:47.471773 4988 scope.go:117] "RemoveContainer" containerID="37f7075f94acc42853a6f73932fe8a7d9236f35da680b42a93d850eb38e3614e" Oct 08 19:58:47 crc kubenswrapper[4988]: I1008 19:58:47.590035 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/9ee564e0-b7af-4905-9957-75cc0d0e50c5-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"9ee564e0-b7af-4905-9957-75cc0d0e50c5\") " pod="openstack/prometheus-metric-storage-0" Oct 08 19:58:47 crc kubenswrapper[4988]: I1008 19:58:47.590116 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/9ee564e0-b7af-4905-9957-75cc0d0e50c5-config\") pod \"prometheus-metric-storage-0\" (UID: \"9ee564e0-b7af-4905-9957-75cc0d0e50c5\") " pod="openstack/prometheus-metric-storage-0" Oct 08 19:58:47 crc kubenswrapper[4988]: I1008 19:58:47.590157 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/9ee564e0-b7af-4905-9957-75cc0d0e50c5-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"9ee564e0-b7af-4905-9957-75cc0d0e50c5\") " pod="openstack/prometheus-metric-storage-0" Oct 08 19:58:47 crc kubenswrapper[4988]: I1008 19:58:47.590220 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9ee564e0-b7af-4905-9957-75cc0d0e50c5-secret-combined-ca-bundle\") pod \"prometheus-metric-storage-0\" (UID: \"9ee564e0-b7af-4905-9957-75cc0d0e50c5\") " pod="openstack/prometheus-metric-storage-0" Oct 08 19:58:47 crc kubenswrapper[4988]: I1008 19:58:47.590258 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\" (UniqueName: \"kubernetes.io/secret/9ee564e0-b7af-4905-9957-75cc0d0e50c5-web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"9ee564e0-b7af-4905-9957-75cc0d0e50c5\") " pod="openstack/prometheus-metric-storage-0" Oct 08 19:58:47 crc kubenswrapper[4988]: I1008 19:58:47.590304 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wndm5\" (UniqueName: \"kubernetes.io/projected/9ee564e0-b7af-4905-9957-75cc0d0e50c5-kube-api-access-wndm5\") pod \"prometheus-metric-storage-0\" (UID: \"9ee564e0-b7af-4905-9957-75cc0d0e50c5\") " pod="openstack/prometheus-metric-storage-0" Oct 08 19:58:47 crc kubenswrapper[4988]: I1008 19:58:47.590341 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/9ee564e0-b7af-4905-9957-75cc0d0e50c5-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"9ee564e0-b7af-4905-9957-75cc0d0e50c5\") " pod="openstack/prometheus-metric-storage-0" Oct 08 19:58:47 crc kubenswrapper[4988]: I1008 19:58:47.590430 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/9ee564e0-b7af-4905-9957-75cc0d0e50c5-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"9ee564e0-b7af-4905-9957-75cc0d0e50c5\") " pod="openstack/prometheus-metric-storage-0" Oct 08 19:58:47 crc kubenswrapper[4988]: I1008 19:58:47.590477 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\" (UniqueName: \"kubernetes.io/secret/9ee564e0-b7af-4905-9957-75cc0d0e50c5-web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"9ee564e0-b7af-4905-9957-75cc0d0e50c5\") " pod="openstack/prometheus-metric-storage-0" Oct 08 19:58:47 crc kubenswrapper[4988]: I1008 19:58:47.590554 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-e634a8ba-6c66-4bab-9b9c-79158ccf2cf8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-e634a8ba-6c66-4bab-9b9c-79158ccf2cf8\") pod \"prometheus-metric-storage-0\" (UID: \"9ee564e0-b7af-4905-9957-75cc0d0e50c5\") " pod="openstack/prometheus-metric-storage-0" Oct 08 19:58:47 crc kubenswrapper[4988]: I1008 19:58:47.590612 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/9ee564e0-b7af-4905-9957-75cc0d0e50c5-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"9ee564e0-b7af-4905-9957-75cc0d0e50c5\") " pod="openstack/prometheus-metric-storage-0" Oct 08 19:58:47 crc kubenswrapper[4988]: I1008 19:58:47.692340 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/9ee564e0-b7af-4905-9957-75cc0d0e50c5-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"9ee564e0-b7af-4905-9957-75cc0d0e50c5\") " pod="openstack/prometheus-metric-storage-0" Oct 08 19:58:47 crc kubenswrapper[4988]: I1008 19:58:47.692424 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/9ee564e0-b7af-4905-9957-75cc0d0e50c5-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"9ee564e0-b7af-4905-9957-75cc0d0e50c5\") " pod="openstack/prometheus-metric-storage-0" Oct 08 19:58:47 crc kubenswrapper[4988]: I1008 19:58:47.692457 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\" (UniqueName: \"kubernetes.io/secret/9ee564e0-b7af-4905-9957-75cc0d0e50c5-web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"9ee564e0-b7af-4905-9957-75cc0d0e50c5\") " pod="openstack/prometheus-metric-storage-0" Oct 08 19:58:47 crc kubenswrapper[4988]: I1008 19:58:47.692509 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-e634a8ba-6c66-4bab-9b9c-79158ccf2cf8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-e634a8ba-6c66-4bab-9b9c-79158ccf2cf8\") pod \"prometheus-metric-storage-0\" (UID: \"9ee564e0-b7af-4905-9957-75cc0d0e50c5\") " pod="openstack/prometheus-metric-storage-0" Oct 08 19:58:47 crc kubenswrapper[4988]: I1008 19:58:47.692546 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/9ee564e0-b7af-4905-9957-75cc0d0e50c5-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"9ee564e0-b7af-4905-9957-75cc0d0e50c5\") " pod="openstack/prometheus-metric-storage-0" Oct 08 19:58:47 crc kubenswrapper[4988]: I1008 19:58:47.692584 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/9ee564e0-b7af-4905-9957-75cc0d0e50c5-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"9ee564e0-b7af-4905-9957-75cc0d0e50c5\") " pod="openstack/prometheus-metric-storage-0" Oct 08 19:58:47 crc kubenswrapper[4988]: I1008 19:58:47.692611 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/9ee564e0-b7af-4905-9957-75cc0d0e50c5-config\") pod \"prometheus-metric-storage-0\" (UID: \"9ee564e0-b7af-4905-9957-75cc0d0e50c5\") " pod="openstack/prometheus-metric-storage-0" Oct 08 19:58:47 crc kubenswrapper[4988]: I1008 19:58:47.692637 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/9ee564e0-b7af-4905-9957-75cc0d0e50c5-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"9ee564e0-b7af-4905-9957-75cc0d0e50c5\") " pod="openstack/prometheus-metric-storage-0" Oct 08 19:58:47 crc kubenswrapper[4988]: I1008 19:58:47.692670 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9ee564e0-b7af-4905-9957-75cc0d0e50c5-secret-combined-ca-bundle\") pod \"prometheus-metric-storage-0\" (UID: \"9ee564e0-b7af-4905-9957-75cc0d0e50c5\") " pod="openstack/prometheus-metric-storage-0" Oct 08 19:58:47 crc kubenswrapper[4988]: I1008 19:58:47.692695 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\" (UniqueName: \"kubernetes.io/secret/9ee564e0-b7af-4905-9957-75cc0d0e50c5-web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"9ee564e0-b7af-4905-9957-75cc0d0e50c5\") " pod="openstack/prometheus-metric-storage-0" Oct 08 19:58:47 crc kubenswrapper[4988]: I1008 19:58:47.692725 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wndm5\" (UniqueName: \"kubernetes.io/projected/9ee564e0-b7af-4905-9957-75cc0d0e50c5-kube-api-access-wndm5\") pod \"prometheus-metric-storage-0\" (UID: \"9ee564e0-b7af-4905-9957-75cc0d0e50c5\") " pod="openstack/prometheus-metric-storage-0" Oct 08 19:58:47 crc kubenswrapper[4988]: I1008 19:58:47.696026 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/9ee564e0-b7af-4905-9957-75cc0d0e50c5-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"9ee564e0-b7af-4905-9957-75cc0d0e50c5\") " pod="openstack/prometheus-metric-storage-0" Oct 08 19:58:47 crc kubenswrapper[4988]: I1008 19:58:47.697002 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/9ee564e0-b7af-4905-9957-75cc0d0e50c5-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"9ee564e0-b7af-4905-9957-75cc0d0e50c5\") " pod="openstack/prometheus-metric-storage-0" Oct 08 19:58:47 crc kubenswrapper[4988]: I1008 19:58:47.697817 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/9ee564e0-b7af-4905-9957-75cc0d0e50c5-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"9ee564e0-b7af-4905-9957-75cc0d0e50c5\") " pod="openstack/prometheus-metric-storage-0" Oct 08 19:58:47 crc kubenswrapper[4988]: I1008 19:58:47.698077 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\" (UniqueName: \"kubernetes.io/secret/9ee564e0-b7af-4905-9957-75cc0d0e50c5-web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"9ee564e0-b7af-4905-9957-75cc0d0e50c5\") " pod="openstack/prometheus-metric-storage-0" Oct 08 19:58:47 crc kubenswrapper[4988]: I1008 19:58:47.698776 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\" (UniqueName: \"kubernetes.io/secret/9ee564e0-b7af-4905-9957-75cc0d0e50c5-web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"9ee564e0-b7af-4905-9957-75cc0d0e50c5\") " pod="openstack/prometheus-metric-storage-0" Oct 08 19:58:47 crc kubenswrapper[4988]: I1008 19:58:47.699055 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/9ee564e0-b7af-4905-9957-75cc0d0e50c5-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"9ee564e0-b7af-4905-9957-75cc0d0e50c5\") " pod="openstack/prometheus-metric-storage-0" Oct 08 19:58:47 crc kubenswrapper[4988]: I1008 19:58:47.699578 4988 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Oct 08 19:58:47 crc kubenswrapper[4988]: I1008 19:58:47.699616 4988 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-e634a8ba-6c66-4bab-9b9c-79158ccf2cf8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-e634a8ba-6c66-4bab-9b9c-79158ccf2cf8\") pod \"prometheus-metric-storage-0\" (UID: \"9ee564e0-b7af-4905-9957-75cc0d0e50c5\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/36bee47a16e8fb587b88b775bafd572b359ef09bb7992247ae065de916eddc2b/globalmount\"" pod="openstack/prometheus-metric-storage-0" Oct 08 19:58:47 crc kubenswrapper[4988]: I1008 19:58:47.707192 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9ee564e0-b7af-4905-9957-75cc0d0e50c5-secret-combined-ca-bundle\") pod \"prometheus-metric-storage-0\" (UID: \"9ee564e0-b7af-4905-9957-75cc0d0e50c5\") " pod="openstack/prometheus-metric-storage-0" Oct 08 19:58:47 crc kubenswrapper[4988]: I1008 19:58:47.708776 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/9ee564e0-b7af-4905-9957-75cc0d0e50c5-config\") pod \"prometheus-metric-storage-0\" (UID: \"9ee564e0-b7af-4905-9957-75cc0d0e50c5\") " pod="openstack/prometheus-metric-storage-0" Oct 08 19:58:47 crc kubenswrapper[4988]: I1008 19:58:47.708776 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/9ee564e0-b7af-4905-9957-75cc0d0e50c5-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"9ee564e0-b7af-4905-9957-75cc0d0e50c5\") " pod="openstack/prometheus-metric-storage-0" Oct 08 19:58:47 crc kubenswrapper[4988]: I1008 19:58:47.714671 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wndm5\" (UniqueName: \"kubernetes.io/projected/9ee564e0-b7af-4905-9957-75cc0d0e50c5-kube-api-access-wndm5\") pod \"prometheus-metric-storage-0\" (UID: \"9ee564e0-b7af-4905-9957-75cc0d0e50c5\") " pod="openstack/prometheus-metric-storage-0" Oct 08 19:58:47 crc kubenswrapper[4988]: I1008 19:58:47.745374 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-e634a8ba-6c66-4bab-9b9c-79158ccf2cf8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-e634a8ba-6c66-4bab-9b9c-79158ccf2cf8\") pod \"prometheus-metric-storage-0\" (UID: \"9ee564e0-b7af-4905-9957-75cc0d0e50c5\") " pod="openstack/prometheus-metric-storage-0" Oct 08 19:58:47 crc kubenswrapper[4988]: I1008 19:58:47.770880 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Oct 08 19:58:48 crc kubenswrapper[4988]: I1008 19:58:48.263083 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/prometheus-metric-storage-0"] Oct 08 19:58:48 crc kubenswrapper[4988]: I1008 19:58:48.290483 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"1e77ab25-7879-4cb9-ae44-0e750ecb8123","Type":"ContainerStarted","Data":"ac2490682666324232b8c023704986f0ffde028943fec0e1ead165a9f4c5de8a"} Oct 08 19:58:48 crc kubenswrapper[4988]: I1008 19:58:48.290531 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"1e77ab25-7879-4cb9-ae44-0e750ecb8123","Type":"ContainerStarted","Data":"b14e617ae178204c88d73af29c00f473cbfa7b0ae2f371a60c78c19096e814c5"} Oct 08 19:58:48 crc kubenswrapper[4988]: I1008 19:58:48.291427 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"9ee564e0-b7af-4905-9957-75cc0d0e50c5","Type":"ContainerStarted","Data":"c4f76d8d97d1c300f8dd4f7828981be9d8559cae5abb236b955e44f4b499f9a7"} Oct 08 19:58:49 crc kubenswrapper[4988]: I1008 19:58:49.250030 4988 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5cd96092-93f6-43a6-bec4-cceaba7bb8af" path="/var/lib/kubelet/pods/5cd96092-93f6-43a6-bec4-cceaba7bb8af/volumes" Oct 08 19:58:49 crc kubenswrapper[4988]: I1008 19:58:49.301342 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"1e77ab25-7879-4cb9-ae44-0e750ecb8123","Type":"ContainerStarted","Data":"d031d5e7030a10c93109bcd69cd9a46d3247a354089a945e8aae7857a0f37496"} Oct 08 19:58:50 crc kubenswrapper[4988]: I1008 19:58:50.315345 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"1e77ab25-7879-4cb9-ae44-0e750ecb8123","Type":"ContainerStarted","Data":"10fa00c70e81451d93287c16175fa45fddd697cfdebee79ca89057501f37fc27"} Oct 08 19:58:51 crc kubenswrapper[4988]: I1008 19:58:51.357958 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"1e77ab25-7879-4cb9-ae44-0e750ecb8123","Type":"ContainerStarted","Data":"831719eba6f00e3b87d88748a721fc9eef82072118af5e7aa35f9f14fe094b44"} Oct 08 19:58:51 crc kubenswrapper[4988]: I1008 19:58:51.358904 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 08 19:58:51 crc kubenswrapper[4988]: I1008 19:58:51.404471 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.070139777 podStartE2EDuration="5.404448154s" podCreationTimestamp="2025-10-08 19:58:46 +0000 UTC" firstStartedPulling="2025-10-08 19:58:47.436062699 +0000 UTC m=+6472.885905469" lastFinishedPulling="2025-10-08 19:58:50.770371056 +0000 UTC m=+6476.220213846" observedRunningTime="2025-10-08 19:58:51.393375111 +0000 UTC m=+6476.843217891" watchObservedRunningTime="2025-10-08 19:58:51.404448154 +0000 UTC m=+6476.854290924" Oct 08 19:58:52 crc kubenswrapper[4988]: I1008 19:58:52.373548 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"9ee564e0-b7af-4905-9957-75cc0d0e50c5","Type":"ContainerStarted","Data":"4b6834eeedb6684ce3b585fa7089ff8db951cb0cf8e9f99e6384ba1ec98a4f6f"} Oct 08 19:58:54 crc kubenswrapper[4988]: I1008 19:58:54.955902 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/aodh-db-create-9qhjh"] Oct 08 19:58:54 crc kubenswrapper[4988]: I1008 19:58:54.957684 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-create-9qhjh" Oct 08 19:58:54 crc kubenswrapper[4988]: I1008 19:58:54.976251 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-db-create-9qhjh"] Oct 08 19:58:55 crc kubenswrapper[4988]: I1008 19:58:55.094156 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-62mtt\" (UniqueName: \"kubernetes.io/projected/792349a5-4d15-4349-9e44-159f8d54da1b-kube-api-access-62mtt\") pod \"aodh-db-create-9qhjh\" (UID: \"792349a5-4d15-4349-9e44-159f8d54da1b\") " pod="openstack/aodh-db-create-9qhjh" Oct 08 19:58:55 crc kubenswrapper[4988]: I1008 19:58:55.198411 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-62mtt\" (UniqueName: \"kubernetes.io/projected/792349a5-4d15-4349-9e44-159f8d54da1b-kube-api-access-62mtt\") pod \"aodh-db-create-9qhjh\" (UID: \"792349a5-4d15-4349-9e44-159f8d54da1b\") " pod="openstack/aodh-db-create-9qhjh" Oct 08 19:58:55 crc kubenswrapper[4988]: I1008 19:58:55.245254 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-62mtt\" (UniqueName: \"kubernetes.io/projected/792349a5-4d15-4349-9e44-159f8d54da1b-kube-api-access-62mtt\") pod \"aodh-db-create-9qhjh\" (UID: \"792349a5-4d15-4349-9e44-159f8d54da1b\") " pod="openstack/aodh-db-create-9qhjh" Oct 08 19:58:55 crc kubenswrapper[4988]: I1008 19:58:55.282431 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-create-9qhjh" Oct 08 19:58:55 crc kubenswrapper[4988]: I1008 19:58:55.686824 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-db-create-9qhjh"] Oct 08 19:58:56 crc kubenswrapper[4988]: I1008 19:58:56.426506 4988 generic.go:334] "Generic (PLEG): container finished" podID="792349a5-4d15-4349-9e44-159f8d54da1b" containerID="75efc3cdf2e5081df1f948e64db0c7baded2ceace55e40b0550cad5cb2f81e4f" exitCode=0 Oct 08 19:58:56 crc kubenswrapper[4988]: I1008 19:58:56.426611 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-create-9qhjh" event={"ID":"792349a5-4d15-4349-9e44-159f8d54da1b","Type":"ContainerDied","Data":"75efc3cdf2e5081df1f948e64db0c7baded2ceace55e40b0550cad5cb2f81e4f"} Oct 08 19:58:56 crc kubenswrapper[4988]: I1008 19:58:56.426944 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-create-9qhjh" event={"ID":"792349a5-4d15-4349-9e44-159f8d54da1b","Type":"ContainerStarted","Data":"75bc664422bbac8207438c12cd8bbb22d736a4e7ea150d8bb15433d6fbac1066"} Oct 08 19:58:57 crc kubenswrapper[4988]: I1008 19:58:57.920475 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-create-9qhjh" Oct 08 19:58:58 crc kubenswrapper[4988]: I1008 19:58:58.073451 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-62mtt\" (UniqueName: \"kubernetes.io/projected/792349a5-4d15-4349-9e44-159f8d54da1b-kube-api-access-62mtt\") pod \"792349a5-4d15-4349-9e44-159f8d54da1b\" (UID: \"792349a5-4d15-4349-9e44-159f8d54da1b\") " Oct 08 19:58:58 crc kubenswrapper[4988]: I1008 19:58:58.085958 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/792349a5-4d15-4349-9e44-159f8d54da1b-kube-api-access-62mtt" (OuterVolumeSpecName: "kube-api-access-62mtt") pod "792349a5-4d15-4349-9e44-159f8d54da1b" (UID: "792349a5-4d15-4349-9e44-159f8d54da1b"). InnerVolumeSpecName "kube-api-access-62mtt". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 19:58:58 crc kubenswrapper[4988]: I1008 19:58:58.176268 4988 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-62mtt\" (UniqueName: \"kubernetes.io/projected/792349a5-4d15-4349-9e44-159f8d54da1b-kube-api-access-62mtt\") on node \"crc\" DevicePath \"\"" Oct 08 19:58:58 crc kubenswrapper[4988]: I1008 19:58:58.239296 4988 scope.go:117] "RemoveContainer" containerID="439af713ff941f973b8758bd6af3c097808f76ff520b96ff10bd3268811393f7" Oct 08 19:58:58 crc kubenswrapper[4988]: E1008 19:58:58.239725 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nm54f_openshift-machine-config-operator(f328a20b-5283-407e-8a93-2ab2da599f18)\"" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" Oct 08 19:58:58 crc kubenswrapper[4988]: I1008 19:58:58.453803 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-create-9qhjh" event={"ID":"792349a5-4d15-4349-9e44-159f8d54da1b","Type":"ContainerDied","Data":"75bc664422bbac8207438c12cd8bbb22d736a4e7ea150d8bb15433d6fbac1066"} Oct 08 19:58:58 crc kubenswrapper[4988]: I1008 19:58:58.453867 4988 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="75bc664422bbac8207438c12cd8bbb22d736a4e7ea150d8bb15433d6fbac1066" Oct 08 19:58:58 crc kubenswrapper[4988]: I1008 19:58:58.453947 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-create-9qhjh" Oct 08 19:58:59 crc kubenswrapper[4988]: I1008 19:58:59.468768 4988 generic.go:334] "Generic (PLEG): container finished" podID="9ee564e0-b7af-4905-9957-75cc0d0e50c5" containerID="4b6834eeedb6684ce3b585fa7089ff8db951cb0cf8e9f99e6384ba1ec98a4f6f" exitCode=0 Oct 08 19:58:59 crc kubenswrapper[4988]: I1008 19:58:59.468873 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"9ee564e0-b7af-4905-9957-75cc0d0e50c5","Type":"ContainerDied","Data":"4b6834eeedb6684ce3b585fa7089ff8db951cb0cf8e9f99e6384ba1ec98a4f6f"} Oct 08 19:59:00 crc kubenswrapper[4988]: I1008 19:59:00.485359 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"9ee564e0-b7af-4905-9957-75cc0d0e50c5","Type":"ContainerStarted","Data":"35439b9e4195a4b2e0497c333d6eb08713021f2aa50a3ff81727dab8ae249d33"} Oct 08 19:59:04 crc kubenswrapper[4988]: I1008 19:59:04.365190 4988 pod_container_manager_linux.go:210] "Failed to delete cgroup paths" cgroupName=["kubepods","burstable","pod7be16926-48be-47f1-a3f7-85be297d1b04"] err="unable to destroy cgroup paths for cgroup [kubepods burstable pod7be16926-48be-47f1-a3f7-85be297d1b04] : Timed out while waiting for systemd to remove kubepods-burstable-pod7be16926_48be_47f1_a3f7_85be297d1b04.slice" Oct 08 19:59:04 crc kubenswrapper[4988]: E1008 19:59:04.365869 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to delete cgroup paths for [kubepods burstable pod7be16926-48be-47f1-a3f7-85be297d1b04] : unable to destroy cgroup paths for cgroup [kubepods burstable pod7be16926-48be-47f1-a3f7-85be297d1b04] : Timed out while waiting for systemd to remove kubepods-burstable-pod7be16926_48be_47f1_a3f7_85be297d1b04.slice" pod="openshift-marketplace/certified-operators-dchph" podUID="7be16926-48be-47f1-a3f7-85be297d1b04" Oct 08 19:59:04 crc kubenswrapper[4988]: I1008 19:59:04.547864 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-dchph" Oct 08 19:59:04 crc kubenswrapper[4988]: I1008 19:59:04.593066 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-dchph"] Oct 08 19:59:04 crc kubenswrapper[4988]: I1008 19:59:04.625618 4988 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-dchph"] Oct 08 19:59:04 crc kubenswrapper[4988]: I1008 19:59:04.965816 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/aodh-5f5f-account-create-bz4fj"] Oct 08 19:59:04 crc kubenswrapper[4988]: E1008 19:59:04.966634 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="792349a5-4d15-4349-9e44-159f8d54da1b" containerName="mariadb-database-create" Oct 08 19:59:04 crc kubenswrapper[4988]: I1008 19:59:04.966654 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="792349a5-4d15-4349-9e44-159f8d54da1b" containerName="mariadb-database-create" Oct 08 19:59:04 crc kubenswrapper[4988]: I1008 19:59:04.966965 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="792349a5-4d15-4349-9e44-159f8d54da1b" containerName="mariadb-database-create" Oct 08 19:59:04 crc kubenswrapper[4988]: I1008 19:59:04.967926 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-5f5f-account-create-bz4fj" Oct 08 19:59:04 crc kubenswrapper[4988]: I1008 19:59:04.977712 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"aodh-db-secret" Oct 08 19:59:04 crc kubenswrapper[4988]: I1008 19:59:04.979181 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-5f5f-account-create-bz4fj"] Oct 08 19:59:05 crc kubenswrapper[4988]: I1008 19:59:05.047455 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xhcdv\" (UniqueName: \"kubernetes.io/projected/a6617968-21ed-42a4-9f84-c21402f8e5e8-kube-api-access-xhcdv\") pod \"aodh-5f5f-account-create-bz4fj\" (UID: \"a6617968-21ed-42a4-9f84-c21402f8e5e8\") " pod="openstack/aodh-5f5f-account-create-bz4fj" Oct 08 19:59:05 crc kubenswrapper[4988]: I1008 19:59:05.151061 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xhcdv\" (UniqueName: \"kubernetes.io/projected/a6617968-21ed-42a4-9f84-c21402f8e5e8-kube-api-access-xhcdv\") pod \"aodh-5f5f-account-create-bz4fj\" (UID: \"a6617968-21ed-42a4-9f84-c21402f8e5e8\") " pod="openstack/aodh-5f5f-account-create-bz4fj" Oct 08 19:59:05 crc kubenswrapper[4988]: I1008 19:59:05.180494 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xhcdv\" (UniqueName: \"kubernetes.io/projected/a6617968-21ed-42a4-9f84-c21402f8e5e8-kube-api-access-xhcdv\") pod \"aodh-5f5f-account-create-bz4fj\" (UID: \"a6617968-21ed-42a4-9f84-c21402f8e5e8\") " pod="openstack/aodh-5f5f-account-create-bz4fj" Oct 08 19:59:05 crc kubenswrapper[4988]: I1008 19:59:05.250835 4988 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7be16926-48be-47f1-a3f7-85be297d1b04" path="/var/lib/kubelet/pods/7be16926-48be-47f1-a3f7-85be297d1b04/volumes" Oct 08 19:59:05 crc kubenswrapper[4988]: I1008 19:59:05.306854 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-5f5f-account-create-bz4fj" Oct 08 19:59:05 crc kubenswrapper[4988]: I1008 19:59:05.563832 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"9ee564e0-b7af-4905-9957-75cc0d0e50c5","Type":"ContainerStarted","Data":"62e29fb8cebe8aaed793e79520b7dc1d8a983c63405ae6c1ab8de8c05394e02d"} Oct 08 19:59:05 crc kubenswrapper[4988]: I1008 19:59:05.564195 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"9ee564e0-b7af-4905-9957-75cc0d0e50c5","Type":"ContainerStarted","Data":"9b762e882572cdceecc240ed6870090d1b27faca71effb99e1ebd162379f7050"} Oct 08 19:59:05 crc kubenswrapper[4988]: I1008 19:59:05.595316 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/prometheus-metric-storage-0" podStartSLOduration=18.595298076 podStartE2EDuration="18.595298076s" podCreationTimestamp="2025-10-08 19:58:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 19:59:05.585823834 +0000 UTC m=+6491.035666594" watchObservedRunningTime="2025-10-08 19:59:05.595298076 +0000 UTC m=+6491.045140846" Oct 08 19:59:05 crc kubenswrapper[4988]: I1008 19:59:05.781143 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-5f5f-account-create-bz4fj"] Oct 08 19:59:05 crc kubenswrapper[4988]: W1008 19:59:05.786475 4988 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda6617968_21ed_42a4_9f84_c21402f8e5e8.slice/crio-ca92b598563277bdc92208bd73f6c7fa70fd6d2c79f155a46a5b0396c759f99f WatchSource:0}: Error finding container ca92b598563277bdc92208bd73f6c7fa70fd6d2c79f155a46a5b0396c759f99f: Status 404 returned error can't find the container with id ca92b598563277bdc92208bd73f6c7fa70fd6d2c79f155a46a5b0396c759f99f Oct 08 19:59:06 crc kubenswrapper[4988]: I1008 19:59:06.580892 4988 generic.go:334] "Generic (PLEG): container finished" podID="a6617968-21ed-42a4-9f84-c21402f8e5e8" containerID="dc43e822f0ef3124c6c1a70e5ff27eb7f9445265f8a1319b0705871c4682b3e6" exitCode=0 Oct 08 19:59:06 crc kubenswrapper[4988]: I1008 19:59:06.581014 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-5f5f-account-create-bz4fj" event={"ID":"a6617968-21ed-42a4-9f84-c21402f8e5e8","Type":"ContainerDied","Data":"dc43e822f0ef3124c6c1a70e5ff27eb7f9445265f8a1319b0705871c4682b3e6"} Oct 08 19:59:06 crc kubenswrapper[4988]: I1008 19:59:06.581337 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-5f5f-account-create-bz4fj" event={"ID":"a6617968-21ed-42a4-9f84-c21402f8e5e8","Type":"ContainerStarted","Data":"ca92b598563277bdc92208bd73f6c7fa70fd6d2c79f155a46a5b0396c759f99f"} Oct 08 19:59:07 crc kubenswrapper[4988]: I1008 19:59:07.771820 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/prometheus-metric-storage-0" Oct 08 19:59:08 crc kubenswrapper[4988]: I1008 19:59:08.090746 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-5f5f-account-create-bz4fj" Oct 08 19:59:08 crc kubenswrapper[4988]: I1008 19:59:08.130800 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xhcdv\" (UniqueName: \"kubernetes.io/projected/a6617968-21ed-42a4-9f84-c21402f8e5e8-kube-api-access-xhcdv\") pod \"a6617968-21ed-42a4-9f84-c21402f8e5e8\" (UID: \"a6617968-21ed-42a4-9f84-c21402f8e5e8\") " Oct 08 19:59:08 crc kubenswrapper[4988]: I1008 19:59:08.137638 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a6617968-21ed-42a4-9f84-c21402f8e5e8-kube-api-access-xhcdv" (OuterVolumeSpecName: "kube-api-access-xhcdv") pod "a6617968-21ed-42a4-9f84-c21402f8e5e8" (UID: "a6617968-21ed-42a4-9f84-c21402f8e5e8"). InnerVolumeSpecName "kube-api-access-xhcdv". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 19:59:08 crc kubenswrapper[4988]: I1008 19:59:08.234253 4988 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xhcdv\" (UniqueName: \"kubernetes.io/projected/a6617968-21ed-42a4-9f84-c21402f8e5e8-kube-api-access-xhcdv\") on node \"crc\" DevicePath \"\"" Oct 08 19:59:08 crc kubenswrapper[4988]: I1008 19:59:08.622455 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-5f5f-account-create-bz4fj" event={"ID":"a6617968-21ed-42a4-9f84-c21402f8e5e8","Type":"ContainerDied","Data":"ca92b598563277bdc92208bd73f6c7fa70fd6d2c79f155a46a5b0396c759f99f"} Oct 08 19:59:08 crc kubenswrapper[4988]: I1008 19:59:08.622514 4988 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ca92b598563277bdc92208bd73f6c7fa70fd6d2c79f155a46a5b0396c759f99f" Oct 08 19:59:08 crc kubenswrapper[4988]: I1008 19:59:08.622594 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-5f5f-account-create-bz4fj" Oct 08 19:59:10 crc kubenswrapper[4988]: I1008 19:59:10.238287 4988 scope.go:117] "RemoveContainer" containerID="439af713ff941f973b8758bd6af3c097808f76ff520b96ff10bd3268811393f7" Oct 08 19:59:10 crc kubenswrapper[4988]: E1008 19:59:10.238911 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nm54f_openshift-machine-config-operator(f328a20b-5283-407e-8a93-2ab2da599f18)\"" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" Oct 08 19:59:10 crc kubenswrapper[4988]: I1008 19:59:10.411884 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/aodh-db-sync-lxj49"] Oct 08 19:59:10 crc kubenswrapper[4988]: E1008 19:59:10.412584 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a6617968-21ed-42a4-9f84-c21402f8e5e8" containerName="mariadb-account-create" Oct 08 19:59:10 crc kubenswrapper[4988]: I1008 19:59:10.412612 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="a6617968-21ed-42a4-9f84-c21402f8e5e8" containerName="mariadb-account-create" Oct 08 19:59:10 crc kubenswrapper[4988]: I1008 19:59:10.413053 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="a6617968-21ed-42a4-9f84-c21402f8e5e8" containerName="mariadb-account-create" Oct 08 19:59:10 crc kubenswrapper[4988]: I1008 19:59:10.414311 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-sync-lxj49" Oct 08 19:59:10 crc kubenswrapper[4988]: I1008 19:59:10.417435 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"telemetry-autoscaling-dockercfg-s848r" Oct 08 19:59:10 crc kubenswrapper[4988]: I1008 19:59:10.417501 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"aodh-config-data" Oct 08 19:59:10 crc kubenswrapper[4988]: I1008 19:59:10.418319 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"aodh-scripts" Oct 08 19:59:10 crc kubenswrapper[4988]: I1008 19:59:10.434675 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-db-sync-lxj49"] Oct 08 19:59:10 crc kubenswrapper[4988]: I1008 19:59:10.496350 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7qb6r\" (UniqueName: \"kubernetes.io/projected/0f268dff-3b76-4bbc-b5db-b2c910aae2c3-kube-api-access-7qb6r\") pod \"aodh-db-sync-lxj49\" (UID: \"0f268dff-3b76-4bbc-b5db-b2c910aae2c3\") " pod="openstack/aodh-db-sync-lxj49" Oct 08 19:59:10 crc kubenswrapper[4988]: I1008 19:59:10.496471 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0f268dff-3b76-4bbc-b5db-b2c910aae2c3-combined-ca-bundle\") pod \"aodh-db-sync-lxj49\" (UID: \"0f268dff-3b76-4bbc-b5db-b2c910aae2c3\") " pod="openstack/aodh-db-sync-lxj49" Oct 08 19:59:10 crc kubenswrapper[4988]: I1008 19:59:10.496516 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0f268dff-3b76-4bbc-b5db-b2c910aae2c3-config-data\") pod \"aodh-db-sync-lxj49\" (UID: \"0f268dff-3b76-4bbc-b5db-b2c910aae2c3\") " pod="openstack/aodh-db-sync-lxj49" Oct 08 19:59:10 crc kubenswrapper[4988]: I1008 19:59:10.496752 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0f268dff-3b76-4bbc-b5db-b2c910aae2c3-scripts\") pod \"aodh-db-sync-lxj49\" (UID: \"0f268dff-3b76-4bbc-b5db-b2c910aae2c3\") " pod="openstack/aodh-db-sync-lxj49" Oct 08 19:59:10 crc kubenswrapper[4988]: I1008 19:59:10.599469 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7qb6r\" (UniqueName: \"kubernetes.io/projected/0f268dff-3b76-4bbc-b5db-b2c910aae2c3-kube-api-access-7qb6r\") pod \"aodh-db-sync-lxj49\" (UID: \"0f268dff-3b76-4bbc-b5db-b2c910aae2c3\") " pod="openstack/aodh-db-sync-lxj49" Oct 08 19:59:10 crc kubenswrapper[4988]: I1008 19:59:10.599553 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0f268dff-3b76-4bbc-b5db-b2c910aae2c3-combined-ca-bundle\") pod \"aodh-db-sync-lxj49\" (UID: \"0f268dff-3b76-4bbc-b5db-b2c910aae2c3\") " pod="openstack/aodh-db-sync-lxj49" Oct 08 19:59:10 crc kubenswrapper[4988]: I1008 19:59:10.599582 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0f268dff-3b76-4bbc-b5db-b2c910aae2c3-config-data\") pod \"aodh-db-sync-lxj49\" (UID: \"0f268dff-3b76-4bbc-b5db-b2c910aae2c3\") " pod="openstack/aodh-db-sync-lxj49" Oct 08 19:59:10 crc kubenswrapper[4988]: I1008 19:59:10.599709 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0f268dff-3b76-4bbc-b5db-b2c910aae2c3-scripts\") pod \"aodh-db-sync-lxj49\" (UID: \"0f268dff-3b76-4bbc-b5db-b2c910aae2c3\") " pod="openstack/aodh-db-sync-lxj49" Oct 08 19:59:10 crc kubenswrapper[4988]: I1008 19:59:10.610021 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0f268dff-3b76-4bbc-b5db-b2c910aae2c3-config-data\") pod \"aodh-db-sync-lxj49\" (UID: \"0f268dff-3b76-4bbc-b5db-b2c910aae2c3\") " pod="openstack/aodh-db-sync-lxj49" Oct 08 19:59:10 crc kubenswrapper[4988]: I1008 19:59:10.611999 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0f268dff-3b76-4bbc-b5db-b2c910aae2c3-combined-ca-bundle\") pod \"aodh-db-sync-lxj49\" (UID: \"0f268dff-3b76-4bbc-b5db-b2c910aae2c3\") " pod="openstack/aodh-db-sync-lxj49" Oct 08 19:59:10 crc kubenswrapper[4988]: I1008 19:59:10.612470 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0f268dff-3b76-4bbc-b5db-b2c910aae2c3-scripts\") pod \"aodh-db-sync-lxj49\" (UID: \"0f268dff-3b76-4bbc-b5db-b2c910aae2c3\") " pod="openstack/aodh-db-sync-lxj49" Oct 08 19:59:10 crc kubenswrapper[4988]: I1008 19:59:10.635420 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7qb6r\" (UniqueName: \"kubernetes.io/projected/0f268dff-3b76-4bbc-b5db-b2c910aae2c3-kube-api-access-7qb6r\") pod \"aodh-db-sync-lxj49\" (UID: \"0f268dff-3b76-4bbc-b5db-b2c910aae2c3\") " pod="openstack/aodh-db-sync-lxj49" Oct 08 19:59:10 crc kubenswrapper[4988]: I1008 19:59:10.733446 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-sync-lxj49" Oct 08 19:59:11 crc kubenswrapper[4988]: I1008 19:59:11.035072 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-db-create-g2pmg"] Oct 08 19:59:11 crc kubenswrapper[4988]: I1008 19:59:11.053349 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-db-create-ctbvx"] Oct 08 19:59:11 crc kubenswrapper[4988]: I1008 19:59:11.059818 4988 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-db-create-g2pmg"] Oct 08 19:59:11 crc kubenswrapper[4988]: I1008 19:59:11.069105 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-db-create-mdbqg"] Oct 08 19:59:11 crc kubenswrapper[4988]: I1008 19:59:11.077092 4988 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-db-create-ctbvx"] Oct 08 19:59:11 crc kubenswrapper[4988]: I1008 19:59:11.086057 4988 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-db-create-mdbqg"] Oct 08 19:59:11 crc kubenswrapper[4988]: I1008 19:59:11.218099 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-db-sync-lxj49"] Oct 08 19:59:11 crc kubenswrapper[4988]: I1008 19:59:11.248946 4988 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="42622d8d-30ea-4b08-b87c-3c77397d9c0e" path="/var/lib/kubelet/pods/42622d8d-30ea-4b08-b87c-3c77397d9c0e/volumes" Oct 08 19:59:11 crc kubenswrapper[4988]: I1008 19:59:11.249638 4988 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8a3a3ca1-d84a-45f0-8879-cdc313823841" path="/var/lib/kubelet/pods/8a3a3ca1-d84a-45f0-8879-cdc313823841/volumes" Oct 08 19:59:11 crc kubenswrapper[4988]: I1008 19:59:11.250129 4988 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fd9a23d7-0671-49d5-b63b-2d632743a81c" path="/var/lib/kubelet/pods/fd9a23d7-0671-49d5-b63b-2d632743a81c/volumes" Oct 08 19:59:11 crc kubenswrapper[4988]: I1008 19:59:11.663950 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-sync-lxj49" event={"ID":"0f268dff-3b76-4bbc-b5db-b2c910aae2c3","Type":"ContainerStarted","Data":"6b3819d4f2fe0d75444c25a6c30fa6475915f063b2c5d2b9dd79ebf15e76a1a8"} Oct 08 19:59:12 crc kubenswrapper[4988]: I1008 19:59:12.167262 4988 scope.go:117] "RemoveContainer" containerID="95c7fc4b25b5bddeed68db71ab1b187e3958e71ccbdf4c0e924fb8a3fc37b446" Oct 08 19:59:12 crc kubenswrapper[4988]: I1008 19:59:12.197875 4988 scope.go:117] "RemoveContainer" containerID="0c0821c37b7f2e7b751b73b3e7164c71c0e4721c7cc79b3d9bf529bc9019a275" Oct 08 19:59:12 crc kubenswrapper[4988]: I1008 19:59:12.273821 4988 scope.go:117] "RemoveContainer" containerID="89036c30ed9153f070b806de127d546654637116b052af9ceb17921797fabcc7" Oct 08 19:59:12 crc kubenswrapper[4988]: I1008 19:59:12.328904 4988 scope.go:117] "RemoveContainer" containerID="97202d3b51d442b5a41c083e3250d51fc15b8e002a829755567476a33bfba747" Oct 08 19:59:15 crc kubenswrapper[4988]: I1008 19:59:15.720349 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-sync-lxj49" event={"ID":"0f268dff-3b76-4bbc-b5db-b2c910aae2c3","Type":"ContainerStarted","Data":"7da83ede41b90de86590af7942c1bfeb0901902308e1fe5ae5b1ff61d501fd6c"} Oct 08 19:59:15 crc kubenswrapper[4988]: I1008 19:59:15.748820 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/aodh-db-sync-lxj49" podStartSLOduration=1.898000846 podStartE2EDuration="5.748786756s" podCreationTimestamp="2025-10-08 19:59:10 +0000 UTC" firstStartedPulling="2025-10-08 19:59:11.215066692 +0000 UTC m=+6496.664909462" lastFinishedPulling="2025-10-08 19:59:15.065852602 +0000 UTC m=+6500.515695372" observedRunningTime="2025-10-08 19:59:15.737780446 +0000 UTC m=+6501.187623236" watchObservedRunningTime="2025-10-08 19:59:15.748786756 +0000 UTC m=+6501.198629546" Oct 08 19:59:16 crc kubenswrapper[4988]: I1008 19:59:16.881886 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Oct 08 19:59:17 crc kubenswrapper[4988]: I1008 19:59:17.743819 4988 generic.go:334] "Generic (PLEG): container finished" podID="0f268dff-3b76-4bbc-b5db-b2c910aae2c3" containerID="7da83ede41b90de86590af7942c1bfeb0901902308e1fe5ae5b1ff61d501fd6c" exitCode=0 Oct 08 19:59:17 crc kubenswrapper[4988]: I1008 19:59:17.743965 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-sync-lxj49" event={"ID":"0f268dff-3b76-4bbc-b5db-b2c910aae2c3","Type":"ContainerDied","Data":"7da83ede41b90de86590af7942c1bfeb0901902308e1fe5ae5b1ff61d501fd6c"} Oct 08 19:59:17 crc kubenswrapper[4988]: I1008 19:59:17.771625 4988 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/prometheus-metric-storage-0" Oct 08 19:59:17 crc kubenswrapper[4988]: I1008 19:59:17.781245 4988 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/prometheus-metric-storage-0" Oct 08 19:59:18 crc kubenswrapper[4988]: I1008 19:59:18.758520 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/prometheus-metric-storage-0" Oct 08 19:59:19 crc kubenswrapper[4988]: I1008 19:59:19.284631 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-sync-lxj49" Oct 08 19:59:19 crc kubenswrapper[4988]: I1008 19:59:19.427287 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7qb6r\" (UniqueName: \"kubernetes.io/projected/0f268dff-3b76-4bbc-b5db-b2c910aae2c3-kube-api-access-7qb6r\") pod \"0f268dff-3b76-4bbc-b5db-b2c910aae2c3\" (UID: \"0f268dff-3b76-4bbc-b5db-b2c910aae2c3\") " Oct 08 19:59:19 crc kubenswrapper[4988]: I1008 19:59:19.427619 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0f268dff-3b76-4bbc-b5db-b2c910aae2c3-scripts\") pod \"0f268dff-3b76-4bbc-b5db-b2c910aae2c3\" (UID: \"0f268dff-3b76-4bbc-b5db-b2c910aae2c3\") " Oct 08 19:59:19 crc kubenswrapper[4988]: I1008 19:59:19.427780 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0f268dff-3b76-4bbc-b5db-b2c910aae2c3-config-data\") pod \"0f268dff-3b76-4bbc-b5db-b2c910aae2c3\" (UID: \"0f268dff-3b76-4bbc-b5db-b2c910aae2c3\") " Oct 08 19:59:19 crc kubenswrapper[4988]: I1008 19:59:19.427914 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0f268dff-3b76-4bbc-b5db-b2c910aae2c3-combined-ca-bundle\") pod \"0f268dff-3b76-4bbc-b5db-b2c910aae2c3\" (UID: \"0f268dff-3b76-4bbc-b5db-b2c910aae2c3\") " Oct 08 19:59:19 crc kubenswrapper[4988]: I1008 19:59:19.435531 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0f268dff-3b76-4bbc-b5db-b2c910aae2c3-scripts" (OuterVolumeSpecName: "scripts") pod "0f268dff-3b76-4bbc-b5db-b2c910aae2c3" (UID: "0f268dff-3b76-4bbc-b5db-b2c910aae2c3"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 19:59:19 crc kubenswrapper[4988]: I1008 19:59:19.436042 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0f268dff-3b76-4bbc-b5db-b2c910aae2c3-kube-api-access-7qb6r" (OuterVolumeSpecName: "kube-api-access-7qb6r") pod "0f268dff-3b76-4bbc-b5db-b2c910aae2c3" (UID: "0f268dff-3b76-4bbc-b5db-b2c910aae2c3"). InnerVolumeSpecName "kube-api-access-7qb6r". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 19:59:19 crc kubenswrapper[4988]: I1008 19:59:19.461081 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0f268dff-3b76-4bbc-b5db-b2c910aae2c3-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "0f268dff-3b76-4bbc-b5db-b2c910aae2c3" (UID: "0f268dff-3b76-4bbc-b5db-b2c910aae2c3"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 19:59:19 crc kubenswrapper[4988]: I1008 19:59:19.467617 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0f268dff-3b76-4bbc-b5db-b2c910aae2c3-config-data" (OuterVolumeSpecName: "config-data") pod "0f268dff-3b76-4bbc-b5db-b2c910aae2c3" (UID: "0f268dff-3b76-4bbc-b5db-b2c910aae2c3"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 19:59:19 crc kubenswrapper[4988]: I1008 19:59:19.530360 4988 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7qb6r\" (UniqueName: \"kubernetes.io/projected/0f268dff-3b76-4bbc-b5db-b2c910aae2c3-kube-api-access-7qb6r\") on node \"crc\" DevicePath \"\"" Oct 08 19:59:19 crc kubenswrapper[4988]: I1008 19:59:19.530501 4988 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0f268dff-3b76-4bbc-b5db-b2c910aae2c3-scripts\") on node \"crc\" DevicePath \"\"" Oct 08 19:59:19 crc kubenswrapper[4988]: I1008 19:59:19.530565 4988 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0f268dff-3b76-4bbc-b5db-b2c910aae2c3-config-data\") on node \"crc\" DevicePath \"\"" Oct 08 19:59:19 crc kubenswrapper[4988]: I1008 19:59:19.530624 4988 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0f268dff-3b76-4bbc-b5db-b2c910aae2c3-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 19:59:19 crc kubenswrapper[4988]: I1008 19:59:19.768100 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-sync-lxj49" event={"ID":"0f268dff-3b76-4bbc-b5db-b2c910aae2c3","Type":"ContainerDied","Data":"6b3819d4f2fe0d75444c25a6c30fa6475915f063b2c5d2b9dd79ebf15e76a1a8"} Oct 08 19:59:19 crc kubenswrapper[4988]: I1008 19:59:19.768482 4988 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6b3819d4f2fe0d75444c25a6c30fa6475915f063b2c5d2b9dd79ebf15e76a1a8" Oct 08 19:59:19 crc kubenswrapper[4988]: I1008 19:59:19.768149 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-sync-lxj49" Oct 08 19:59:19 crc kubenswrapper[4988]: I1008 19:59:19.966574 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/aodh-0"] Oct 08 19:59:19 crc kubenswrapper[4988]: E1008 19:59:19.967083 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0f268dff-3b76-4bbc-b5db-b2c910aae2c3" containerName="aodh-db-sync" Oct 08 19:59:19 crc kubenswrapper[4988]: I1008 19:59:19.967109 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="0f268dff-3b76-4bbc-b5db-b2c910aae2c3" containerName="aodh-db-sync" Oct 08 19:59:19 crc kubenswrapper[4988]: I1008 19:59:19.967440 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="0f268dff-3b76-4bbc-b5db-b2c910aae2c3" containerName="aodh-db-sync" Oct 08 19:59:19 crc kubenswrapper[4988]: I1008 19:59:19.969792 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-0" Oct 08 19:59:19 crc kubenswrapper[4988]: I1008 19:59:19.976627 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"aodh-scripts" Oct 08 19:59:19 crc kubenswrapper[4988]: I1008 19:59:19.976960 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"telemetry-autoscaling-dockercfg-s848r" Oct 08 19:59:19 crc kubenswrapper[4988]: I1008 19:59:19.979256 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"aodh-config-data" Oct 08 19:59:19 crc kubenswrapper[4988]: I1008 19:59:19.997404 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-0"] Oct 08 19:59:20 crc kubenswrapper[4988]: I1008 19:59:20.042712 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6prdj\" (UniqueName: \"kubernetes.io/projected/6a1bc0e7-3209-4fa2-9a2a-c8b18121da5a-kube-api-access-6prdj\") pod \"aodh-0\" (UID: \"6a1bc0e7-3209-4fa2-9a2a-c8b18121da5a\") " pod="openstack/aodh-0" Oct 08 19:59:20 crc kubenswrapper[4988]: I1008 19:59:20.042787 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6a1bc0e7-3209-4fa2-9a2a-c8b18121da5a-scripts\") pod \"aodh-0\" (UID: \"6a1bc0e7-3209-4fa2-9a2a-c8b18121da5a\") " pod="openstack/aodh-0" Oct 08 19:59:20 crc kubenswrapper[4988]: I1008 19:59:20.042809 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6a1bc0e7-3209-4fa2-9a2a-c8b18121da5a-combined-ca-bundle\") pod \"aodh-0\" (UID: \"6a1bc0e7-3209-4fa2-9a2a-c8b18121da5a\") " pod="openstack/aodh-0" Oct 08 19:59:20 crc kubenswrapper[4988]: I1008 19:59:20.043164 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6a1bc0e7-3209-4fa2-9a2a-c8b18121da5a-config-data\") pod \"aodh-0\" (UID: \"6a1bc0e7-3209-4fa2-9a2a-c8b18121da5a\") " pod="openstack/aodh-0" Oct 08 19:59:20 crc kubenswrapper[4988]: I1008 19:59:20.148898 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6a1bc0e7-3209-4fa2-9a2a-c8b18121da5a-config-data\") pod \"aodh-0\" (UID: \"6a1bc0e7-3209-4fa2-9a2a-c8b18121da5a\") " pod="openstack/aodh-0" Oct 08 19:59:20 crc kubenswrapper[4988]: I1008 19:59:20.149070 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6prdj\" (UniqueName: \"kubernetes.io/projected/6a1bc0e7-3209-4fa2-9a2a-c8b18121da5a-kube-api-access-6prdj\") pod \"aodh-0\" (UID: \"6a1bc0e7-3209-4fa2-9a2a-c8b18121da5a\") " pod="openstack/aodh-0" Oct 08 19:59:20 crc kubenswrapper[4988]: I1008 19:59:20.149146 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6a1bc0e7-3209-4fa2-9a2a-c8b18121da5a-scripts\") pod \"aodh-0\" (UID: \"6a1bc0e7-3209-4fa2-9a2a-c8b18121da5a\") " pod="openstack/aodh-0" Oct 08 19:59:20 crc kubenswrapper[4988]: I1008 19:59:20.149170 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6a1bc0e7-3209-4fa2-9a2a-c8b18121da5a-combined-ca-bundle\") pod \"aodh-0\" (UID: \"6a1bc0e7-3209-4fa2-9a2a-c8b18121da5a\") " pod="openstack/aodh-0" Oct 08 19:59:20 crc kubenswrapper[4988]: I1008 19:59:20.169203 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6a1bc0e7-3209-4fa2-9a2a-c8b18121da5a-combined-ca-bundle\") pod \"aodh-0\" (UID: \"6a1bc0e7-3209-4fa2-9a2a-c8b18121da5a\") " pod="openstack/aodh-0" Oct 08 19:59:20 crc kubenswrapper[4988]: I1008 19:59:20.170322 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6a1bc0e7-3209-4fa2-9a2a-c8b18121da5a-config-data\") pod \"aodh-0\" (UID: \"6a1bc0e7-3209-4fa2-9a2a-c8b18121da5a\") " pod="openstack/aodh-0" Oct 08 19:59:20 crc kubenswrapper[4988]: I1008 19:59:20.176064 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6prdj\" (UniqueName: \"kubernetes.io/projected/6a1bc0e7-3209-4fa2-9a2a-c8b18121da5a-kube-api-access-6prdj\") pod \"aodh-0\" (UID: \"6a1bc0e7-3209-4fa2-9a2a-c8b18121da5a\") " pod="openstack/aodh-0" Oct 08 19:59:20 crc kubenswrapper[4988]: I1008 19:59:20.177121 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6a1bc0e7-3209-4fa2-9a2a-c8b18121da5a-scripts\") pod \"aodh-0\" (UID: \"6a1bc0e7-3209-4fa2-9a2a-c8b18121da5a\") " pod="openstack/aodh-0" Oct 08 19:59:20 crc kubenswrapper[4988]: I1008 19:59:20.316593 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-0" Oct 08 19:59:20 crc kubenswrapper[4988]: I1008 19:59:20.821315 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-0"] Oct 08 19:59:20 crc kubenswrapper[4988]: W1008 19:59:20.859096 4988 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6a1bc0e7_3209_4fa2_9a2a_c8b18121da5a.slice/crio-f7630bfffa89fc3a81a120034a59d2e5c8e4f920e056a9b9251b972ddc7c9ff8 WatchSource:0}: Error finding container f7630bfffa89fc3a81a120034a59d2e5c8e4f920e056a9b9251b972ddc7c9ff8: Status 404 returned error can't find the container with id f7630bfffa89fc3a81a120034a59d2e5c8e4f920e056a9b9251b972ddc7c9ff8 Oct 08 19:59:21 crc kubenswrapper[4988]: I1008 19:59:21.053451 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-015a-account-create-t8ks5"] Oct 08 19:59:21 crc kubenswrapper[4988]: I1008 19:59:21.062809 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-ef5f-account-create-mn8dp"] Oct 08 19:59:21 crc kubenswrapper[4988]: I1008 19:59:21.075234 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-0c51-account-create-frwnd"] Oct 08 19:59:21 crc kubenswrapper[4988]: I1008 19:59:21.086014 4988 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-015a-account-create-t8ks5"] Oct 08 19:59:21 crc kubenswrapper[4988]: I1008 19:59:21.094958 4988 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-ef5f-account-create-mn8dp"] Oct 08 19:59:21 crc kubenswrapper[4988]: I1008 19:59:21.103970 4988 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-0c51-account-create-frwnd"] Oct 08 19:59:21 crc kubenswrapper[4988]: I1008 19:59:21.248594 4988 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="19e6cdb4-5c6f-4e69-a918-9f88dd304d9e" path="/var/lib/kubelet/pods/19e6cdb4-5c6f-4e69-a918-9f88dd304d9e/volumes" Oct 08 19:59:21 crc kubenswrapper[4988]: I1008 19:59:21.249628 4988 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4cde7ab7-3097-45ad-a312-98ddc074114c" path="/var/lib/kubelet/pods/4cde7ab7-3097-45ad-a312-98ddc074114c/volumes" Oct 08 19:59:21 crc kubenswrapper[4988]: I1008 19:59:21.250330 4988 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f3f270ce-b57a-4d31-a77a-5ba66dabab78" path="/var/lib/kubelet/pods/f3f270ce-b57a-4d31-a77a-5ba66dabab78/volumes" Oct 08 19:59:21 crc kubenswrapper[4988]: I1008 19:59:21.789981 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"6a1bc0e7-3209-4fa2-9a2a-c8b18121da5a","Type":"ContainerStarted","Data":"de0e0960e8cf3d4761ad85747537742d917d43f51156f5145c593fb2bc682af3"} Oct 08 19:59:21 crc kubenswrapper[4988]: I1008 19:59:21.790560 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"6a1bc0e7-3209-4fa2-9a2a-c8b18121da5a","Type":"ContainerStarted","Data":"f7630bfffa89fc3a81a120034a59d2e5c8e4f920e056a9b9251b972ddc7c9ff8"} Oct 08 19:59:22 crc kubenswrapper[4988]: I1008 19:59:22.241309 4988 scope.go:117] "RemoveContainer" containerID="439af713ff941f973b8758bd6af3c097808f76ff520b96ff10bd3268811393f7" Oct 08 19:59:22 crc kubenswrapper[4988]: E1008 19:59:22.241755 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nm54f_openshift-machine-config-operator(f328a20b-5283-407e-8a93-2ab2da599f18)\"" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" Oct 08 19:59:23 crc kubenswrapper[4988]: I1008 19:59:23.363160 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 08 19:59:23 crc kubenswrapper[4988]: I1008 19:59:23.364710 4988 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="1e77ab25-7879-4cb9-ae44-0e750ecb8123" containerName="sg-core" containerID="cri-o://10fa00c70e81451d93287c16175fa45fddd697cfdebee79ca89057501f37fc27" gracePeriod=30 Oct 08 19:59:23 crc kubenswrapper[4988]: I1008 19:59:23.364755 4988 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="1e77ab25-7879-4cb9-ae44-0e750ecb8123" containerName="proxy-httpd" containerID="cri-o://831719eba6f00e3b87d88748a721fc9eef82072118af5e7aa35f9f14fe094b44" gracePeriod=30 Oct 08 19:59:23 crc kubenswrapper[4988]: I1008 19:59:23.364729 4988 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="1e77ab25-7879-4cb9-ae44-0e750ecb8123" containerName="ceilometer-notification-agent" containerID="cri-o://d031d5e7030a10c93109bcd69cd9a46d3247a354089a945e8aae7857a0f37496" gracePeriod=30 Oct 08 19:59:23 crc kubenswrapper[4988]: I1008 19:59:23.364651 4988 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="1e77ab25-7879-4cb9-ae44-0e750ecb8123" containerName="ceilometer-central-agent" containerID="cri-o://ac2490682666324232b8c023704986f0ffde028943fec0e1ead165a9f4c5de8a" gracePeriod=30 Oct 08 19:59:23 crc kubenswrapper[4988]: I1008 19:59:23.811499 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"6a1bc0e7-3209-4fa2-9a2a-c8b18121da5a","Type":"ContainerStarted","Data":"1a9c83e0bd2e4118cc11a9d647a155cb5f97ba61361dc0ba0e080435aaa48c98"} Oct 08 19:59:23 crc kubenswrapper[4988]: I1008 19:59:23.813940 4988 generic.go:334] "Generic (PLEG): container finished" podID="1e77ab25-7879-4cb9-ae44-0e750ecb8123" containerID="831719eba6f00e3b87d88748a721fc9eef82072118af5e7aa35f9f14fe094b44" exitCode=0 Oct 08 19:59:23 crc kubenswrapper[4988]: I1008 19:59:23.813972 4988 generic.go:334] "Generic (PLEG): container finished" podID="1e77ab25-7879-4cb9-ae44-0e750ecb8123" containerID="10fa00c70e81451d93287c16175fa45fddd697cfdebee79ca89057501f37fc27" exitCode=2 Oct 08 19:59:23 crc kubenswrapper[4988]: I1008 19:59:23.813981 4988 generic.go:334] "Generic (PLEG): container finished" podID="1e77ab25-7879-4cb9-ae44-0e750ecb8123" containerID="ac2490682666324232b8c023704986f0ffde028943fec0e1ead165a9f4c5de8a" exitCode=0 Oct 08 19:59:23 crc kubenswrapper[4988]: I1008 19:59:23.814000 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"1e77ab25-7879-4cb9-ae44-0e750ecb8123","Type":"ContainerDied","Data":"831719eba6f00e3b87d88748a721fc9eef82072118af5e7aa35f9f14fe094b44"} Oct 08 19:59:23 crc kubenswrapper[4988]: I1008 19:59:23.814051 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"1e77ab25-7879-4cb9-ae44-0e750ecb8123","Type":"ContainerDied","Data":"10fa00c70e81451d93287c16175fa45fddd697cfdebee79ca89057501f37fc27"} Oct 08 19:59:23 crc kubenswrapper[4988]: I1008 19:59:23.814069 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"1e77ab25-7879-4cb9-ae44-0e750ecb8123","Type":"ContainerDied","Data":"ac2490682666324232b8c023704986f0ffde028943fec0e1ead165a9f4c5de8a"} Oct 08 19:59:24 crc kubenswrapper[4988]: I1008 19:59:24.658731 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/aodh-0"] Oct 08 19:59:25 crc kubenswrapper[4988]: I1008 19:59:25.447511 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 08 19:59:25 crc kubenswrapper[4988]: I1008 19:59:25.474192 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1e77ab25-7879-4cb9-ae44-0e750ecb8123-config-data\") pod \"1e77ab25-7879-4cb9-ae44-0e750ecb8123\" (UID: \"1e77ab25-7879-4cb9-ae44-0e750ecb8123\") " Oct 08 19:59:25 crc kubenswrapper[4988]: I1008 19:59:25.474271 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1e77ab25-7879-4cb9-ae44-0e750ecb8123-combined-ca-bundle\") pod \"1e77ab25-7879-4cb9-ae44-0e750ecb8123\" (UID: \"1e77ab25-7879-4cb9-ae44-0e750ecb8123\") " Oct 08 19:59:25 crc kubenswrapper[4988]: I1008 19:59:25.474363 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1e77ab25-7879-4cb9-ae44-0e750ecb8123-log-httpd\") pod \"1e77ab25-7879-4cb9-ae44-0e750ecb8123\" (UID: \"1e77ab25-7879-4cb9-ae44-0e750ecb8123\") " Oct 08 19:59:25 crc kubenswrapper[4988]: I1008 19:59:25.474450 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jgwqx\" (UniqueName: \"kubernetes.io/projected/1e77ab25-7879-4cb9-ae44-0e750ecb8123-kube-api-access-jgwqx\") pod \"1e77ab25-7879-4cb9-ae44-0e750ecb8123\" (UID: \"1e77ab25-7879-4cb9-ae44-0e750ecb8123\") " Oct 08 19:59:25 crc kubenswrapper[4988]: I1008 19:59:25.474565 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/1e77ab25-7879-4cb9-ae44-0e750ecb8123-sg-core-conf-yaml\") pod \"1e77ab25-7879-4cb9-ae44-0e750ecb8123\" (UID: \"1e77ab25-7879-4cb9-ae44-0e750ecb8123\") " Oct 08 19:59:25 crc kubenswrapper[4988]: I1008 19:59:25.474650 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1e77ab25-7879-4cb9-ae44-0e750ecb8123-run-httpd\") pod \"1e77ab25-7879-4cb9-ae44-0e750ecb8123\" (UID: \"1e77ab25-7879-4cb9-ae44-0e750ecb8123\") " Oct 08 19:59:25 crc kubenswrapper[4988]: I1008 19:59:25.474765 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1e77ab25-7879-4cb9-ae44-0e750ecb8123-scripts\") pod \"1e77ab25-7879-4cb9-ae44-0e750ecb8123\" (UID: \"1e77ab25-7879-4cb9-ae44-0e750ecb8123\") " Oct 08 19:59:25 crc kubenswrapper[4988]: I1008 19:59:25.479596 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1e77ab25-7879-4cb9-ae44-0e750ecb8123-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "1e77ab25-7879-4cb9-ae44-0e750ecb8123" (UID: "1e77ab25-7879-4cb9-ae44-0e750ecb8123"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 19:59:25 crc kubenswrapper[4988]: I1008 19:59:25.480396 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1e77ab25-7879-4cb9-ae44-0e750ecb8123-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "1e77ab25-7879-4cb9-ae44-0e750ecb8123" (UID: "1e77ab25-7879-4cb9-ae44-0e750ecb8123"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 19:59:25 crc kubenswrapper[4988]: I1008 19:59:25.482627 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1e77ab25-7879-4cb9-ae44-0e750ecb8123-kube-api-access-jgwqx" (OuterVolumeSpecName: "kube-api-access-jgwqx") pod "1e77ab25-7879-4cb9-ae44-0e750ecb8123" (UID: "1e77ab25-7879-4cb9-ae44-0e750ecb8123"). InnerVolumeSpecName "kube-api-access-jgwqx". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 19:59:25 crc kubenswrapper[4988]: I1008 19:59:25.502800 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1e77ab25-7879-4cb9-ae44-0e750ecb8123-scripts" (OuterVolumeSpecName: "scripts") pod "1e77ab25-7879-4cb9-ae44-0e750ecb8123" (UID: "1e77ab25-7879-4cb9-ae44-0e750ecb8123"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 19:59:25 crc kubenswrapper[4988]: I1008 19:59:25.528529 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1e77ab25-7879-4cb9-ae44-0e750ecb8123-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "1e77ab25-7879-4cb9-ae44-0e750ecb8123" (UID: "1e77ab25-7879-4cb9-ae44-0e750ecb8123"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 19:59:25 crc kubenswrapper[4988]: I1008 19:59:25.582917 4988 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1e77ab25-7879-4cb9-ae44-0e750ecb8123-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 08 19:59:25 crc kubenswrapper[4988]: I1008 19:59:25.583206 4988 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jgwqx\" (UniqueName: \"kubernetes.io/projected/1e77ab25-7879-4cb9-ae44-0e750ecb8123-kube-api-access-jgwqx\") on node \"crc\" DevicePath \"\"" Oct 08 19:59:25 crc kubenswrapper[4988]: I1008 19:59:25.583217 4988 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/1e77ab25-7879-4cb9-ae44-0e750ecb8123-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Oct 08 19:59:25 crc kubenswrapper[4988]: I1008 19:59:25.583227 4988 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1e77ab25-7879-4cb9-ae44-0e750ecb8123-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 08 19:59:25 crc kubenswrapper[4988]: I1008 19:59:25.583235 4988 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1e77ab25-7879-4cb9-ae44-0e750ecb8123-scripts\") on node \"crc\" DevicePath \"\"" Oct 08 19:59:25 crc kubenswrapper[4988]: I1008 19:59:25.640800 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1e77ab25-7879-4cb9-ae44-0e750ecb8123-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "1e77ab25-7879-4cb9-ae44-0e750ecb8123" (UID: "1e77ab25-7879-4cb9-ae44-0e750ecb8123"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 19:59:25 crc kubenswrapper[4988]: I1008 19:59:25.643680 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1e77ab25-7879-4cb9-ae44-0e750ecb8123-config-data" (OuterVolumeSpecName: "config-data") pod "1e77ab25-7879-4cb9-ae44-0e750ecb8123" (UID: "1e77ab25-7879-4cb9-ae44-0e750ecb8123"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 19:59:25 crc kubenswrapper[4988]: I1008 19:59:25.657657 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 08 19:59:25 crc kubenswrapper[4988]: I1008 19:59:25.657856 4988 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/kube-state-metrics-0" podUID="d9fdb832-bcf7-4e2a-afe0-a01a5881cfb3" containerName="kube-state-metrics" containerID="cri-o://d526c948e42b32ee91470150e213b56a919cb4fc6e9b6480ba2a38e3d49dd02e" gracePeriod=30 Oct 08 19:59:25 crc kubenswrapper[4988]: I1008 19:59:25.685576 4988 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1e77ab25-7879-4cb9-ae44-0e750ecb8123-config-data\") on node \"crc\" DevicePath \"\"" Oct 08 19:59:25 crc kubenswrapper[4988]: I1008 19:59:25.685604 4988 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1e77ab25-7879-4cb9-ae44-0e750ecb8123-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 19:59:25 crc kubenswrapper[4988]: I1008 19:59:25.836526 4988 generic.go:334] "Generic (PLEG): container finished" podID="1e77ab25-7879-4cb9-ae44-0e750ecb8123" containerID="d031d5e7030a10c93109bcd69cd9a46d3247a354089a945e8aae7857a0f37496" exitCode=0 Oct 08 19:59:25 crc kubenswrapper[4988]: I1008 19:59:25.836598 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 08 19:59:25 crc kubenswrapper[4988]: I1008 19:59:25.836584 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"1e77ab25-7879-4cb9-ae44-0e750ecb8123","Type":"ContainerDied","Data":"d031d5e7030a10c93109bcd69cd9a46d3247a354089a945e8aae7857a0f37496"} Oct 08 19:59:25 crc kubenswrapper[4988]: I1008 19:59:25.836728 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"1e77ab25-7879-4cb9-ae44-0e750ecb8123","Type":"ContainerDied","Data":"b14e617ae178204c88d73af29c00f473cbfa7b0ae2f371a60c78c19096e814c5"} Oct 08 19:59:25 crc kubenswrapper[4988]: I1008 19:59:25.836753 4988 scope.go:117] "RemoveContainer" containerID="831719eba6f00e3b87d88748a721fc9eef82072118af5e7aa35f9f14fe094b44" Oct 08 19:59:25 crc kubenswrapper[4988]: I1008 19:59:25.838410 4988 generic.go:334] "Generic (PLEG): container finished" podID="d9fdb832-bcf7-4e2a-afe0-a01a5881cfb3" containerID="d526c948e42b32ee91470150e213b56a919cb4fc6e9b6480ba2a38e3d49dd02e" exitCode=2 Oct 08 19:59:25 crc kubenswrapper[4988]: I1008 19:59:25.838475 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"d9fdb832-bcf7-4e2a-afe0-a01a5881cfb3","Type":"ContainerDied","Data":"d526c948e42b32ee91470150e213b56a919cb4fc6e9b6480ba2a38e3d49dd02e"} Oct 08 19:59:25 crc kubenswrapper[4988]: I1008 19:59:25.840185 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"6a1bc0e7-3209-4fa2-9a2a-c8b18121da5a","Type":"ContainerStarted","Data":"01e730efa7fc209fb2f43a464a61ae5e2d7bfaa803be6f3d5211a851dc991bd8"} Oct 08 19:59:25 crc kubenswrapper[4988]: I1008 19:59:25.877338 4988 scope.go:117] "RemoveContainer" containerID="10fa00c70e81451d93287c16175fa45fddd697cfdebee79ca89057501f37fc27" Oct 08 19:59:25 crc kubenswrapper[4988]: I1008 19:59:25.886938 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 08 19:59:25 crc kubenswrapper[4988]: I1008 19:59:25.916994 4988 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Oct 08 19:59:25 crc kubenswrapper[4988]: I1008 19:59:25.927545 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 08 19:59:25 crc kubenswrapper[4988]: E1008 19:59:25.927956 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1e77ab25-7879-4cb9-ae44-0e750ecb8123" containerName="ceilometer-central-agent" Oct 08 19:59:25 crc kubenswrapper[4988]: I1008 19:59:25.927972 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="1e77ab25-7879-4cb9-ae44-0e750ecb8123" containerName="ceilometer-central-agent" Oct 08 19:59:25 crc kubenswrapper[4988]: E1008 19:59:25.928000 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1e77ab25-7879-4cb9-ae44-0e750ecb8123" containerName="ceilometer-notification-agent" Oct 08 19:59:25 crc kubenswrapper[4988]: I1008 19:59:25.928007 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="1e77ab25-7879-4cb9-ae44-0e750ecb8123" containerName="ceilometer-notification-agent" Oct 08 19:59:25 crc kubenswrapper[4988]: E1008 19:59:25.928027 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1e77ab25-7879-4cb9-ae44-0e750ecb8123" containerName="sg-core" Oct 08 19:59:25 crc kubenswrapper[4988]: I1008 19:59:25.928033 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="1e77ab25-7879-4cb9-ae44-0e750ecb8123" containerName="sg-core" Oct 08 19:59:25 crc kubenswrapper[4988]: E1008 19:59:25.928044 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1e77ab25-7879-4cb9-ae44-0e750ecb8123" containerName="proxy-httpd" Oct 08 19:59:25 crc kubenswrapper[4988]: I1008 19:59:25.928049 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="1e77ab25-7879-4cb9-ae44-0e750ecb8123" containerName="proxy-httpd" Oct 08 19:59:25 crc kubenswrapper[4988]: I1008 19:59:25.928269 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="1e77ab25-7879-4cb9-ae44-0e750ecb8123" containerName="sg-core" Oct 08 19:59:25 crc kubenswrapper[4988]: I1008 19:59:25.928285 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="1e77ab25-7879-4cb9-ae44-0e750ecb8123" containerName="proxy-httpd" Oct 08 19:59:25 crc kubenswrapper[4988]: I1008 19:59:25.928303 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="1e77ab25-7879-4cb9-ae44-0e750ecb8123" containerName="ceilometer-notification-agent" Oct 08 19:59:25 crc kubenswrapper[4988]: I1008 19:59:25.928312 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="1e77ab25-7879-4cb9-ae44-0e750ecb8123" containerName="ceilometer-central-agent" Oct 08 19:59:25 crc kubenswrapper[4988]: I1008 19:59:25.930135 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 08 19:59:25 crc kubenswrapper[4988]: I1008 19:59:25.932647 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 08 19:59:25 crc kubenswrapper[4988]: I1008 19:59:25.933461 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 08 19:59:25 crc kubenswrapper[4988]: I1008 19:59:25.971139 4988 scope.go:117] "RemoveContainer" containerID="d031d5e7030a10c93109bcd69cd9a46d3247a354089a945e8aae7857a0f37496" Oct 08 19:59:25 crc kubenswrapper[4988]: I1008 19:59:25.971304 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 08 19:59:25 crc kubenswrapper[4988]: I1008 19:59:25.994606 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/9c249a33-8704-4532-8000-279d5ba3a9dc-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"9c249a33-8704-4532-8000-279d5ba3a9dc\") " pod="openstack/ceilometer-0" Oct 08 19:59:25 crc kubenswrapper[4988]: I1008 19:59:25.994643 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9c249a33-8704-4532-8000-279d5ba3a9dc-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"9c249a33-8704-4532-8000-279d5ba3a9dc\") " pod="openstack/ceilometer-0" Oct 08 19:59:25 crc kubenswrapper[4988]: I1008 19:59:25.994714 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9c249a33-8704-4532-8000-279d5ba3a9dc-scripts\") pod \"ceilometer-0\" (UID: \"9c249a33-8704-4532-8000-279d5ba3a9dc\") " pod="openstack/ceilometer-0" Oct 08 19:59:25 crc kubenswrapper[4988]: I1008 19:59:25.994772 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z56ql\" (UniqueName: \"kubernetes.io/projected/9c249a33-8704-4532-8000-279d5ba3a9dc-kube-api-access-z56ql\") pod \"ceilometer-0\" (UID: \"9c249a33-8704-4532-8000-279d5ba3a9dc\") " pod="openstack/ceilometer-0" Oct 08 19:59:25 crc kubenswrapper[4988]: I1008 19:59:25.994841 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9c249a33-8704-4532-8000-279d5ba3a9dc-log-httpd\") pod \"ceilometer-0\" (UID: \"9c249a33-8704-4532-8000-279d5ba3a9dc\") " pod="openstack/ceilometer-0" Oct 08 19:59:25 crc kubenswrapper[4988]: I1008 19:59:25.994871 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9c249a33-8704-4532-8000-279d5ba3a9dc-config-data\") pod \"ceilometer-0\" (UID: \"9c249a33-8704-4532-8000-279d5ba3a9dc\") " pod="openstack/ceilometer-0" Oct 08 19:59:25 crc kubenswrapper[4988]: I1008 19:59:25.994914 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9c249a33-8704-4532-8000-279d5ba3a9dc-run-httpd\") pod \"ceilometer-0\" (UID: \"9c249a33-8704-4532-8000-279d5ba3a9dc\") " pod="openstack/ceilometer-0" Oct 08 19:59:26 crc kubenswrapper[4988]: I1008 19:59:26.039480 4988 scope.go:117] "RemoveContainer" containerID="ac2490682666324232b8c023704986f0ffde028943fec0e1ead165a9f4c5de8a" Oct 08 19:59:26 crc kubenswrapper[4988]: I1008 19:59:26.083545 4988 scope.go:117] "RemoveContainer" containerID="831719eba6f00e3b87d88748a721fc9eef82072118af5e7aa35f9f14fe094b44" Oct 08 19:59:26 crc kubenswrapper[4988]: E1008 19:59:26.085881 4988 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"831719eba6f00e3b87d88748a721fc9eef82072118af5e7aa35f9f14fe094b44\": container with ID starting with 831719eba6f00e3b87d88748a721fc9eef82072118af5e7aa35f9f14fe094b44 not found: ID does not exist" containerID="831719eba6f00e3b87d88748a721fc9eef82072118af5e7aa35f9f14fe094b44" Oct 08 19:59:26 crc kubenswrapper[4988]: I1008 19:59:26.085910 4988 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"831719eba6f00e3b87d88748a721fc9eef82072118af5e7aa35f9f14fe094b44"} err="failed to get container status \"831719eba6f00e3b87d88748a721fc9eef82072118af5e7aa35f9f14fe094b44\": rpc error: code = NotFound desc = could not find container \"831719eba6f00e3b87d88748a721fc9eef82072118af5e7aa35f9f14fe094b44\": container with ID starting with 831719eba6f00e3b87d88748a721fc9eef82072118af5e7aa35f9f14fe094b44 not found: ID does not exist" Oct 08 19:59:26 crc kubenswrapper[4988]: I1008 19:59:26.085935 4988 scope.go:117] "RemoveContainer" containerID="10fa00c70e81451d93287c16175fa45fddd697cfdebee79ca89057501f37fc27" Oct 08 19:59:26 crc kubenswrapper[4988]: E1008 19:59:26.090038 4988 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"10fa00c70e81451d93287c16175fa45fddd697cfdebee79ca89057501f37fc27\": container with ID starting with 10fa00c70e81451d93287c16175fa45fddd697cfdebee79ca89057501f37fc27 not found: ID does not exist" containerID="10fa00c70e81451d93287c16175fa45fddd697cfdebee79ca89057501f37fc27" Oct 08 19:59:26 crc kubenswrapper[4988]: I1008 19:59:26.090116 4988 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"10fa00c70e81451d93287c16175fa45fddd697cfdebee79ca89057501f37fc27"} err="failed to get container status \"10fa00c70e81451d93287c16175fa45fddd697cfdebee79ca89057501f37fc27\": rpc error: code = NotFound desc = could not find container \"10fa00c70e81451d93287c16175fa45fddd697cfdebee79ca89057501f37fc27\": container with ID starting with 10fa00c70e81451d93287c16175fa45fddd697cfdebee79ca89057501f37fc27 not found: ID does not exist" Oct 08 19:59:26 crc kubenswrapper[4988]: I1008 19:59:26.090145 4988 scope.go:117] "RemoveContainer" containerID="d031d5e7030a10c93109bcd69cd9a46d3247a354089a945e8aae7857a0f37496" Oct 08 19:59:26 crc kubenswrapper[4988]: E1008 19:59:26.093739 4988 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d031d5e7030a10c93109bcd69cd9a46d3247a354089a945e8aae7857a0f37496\": container with ID starting with d031d5e7030a10c93109bcd69cd9a46d3247a354089a945e8aae7857a0f37496 not found: ID does not exist" containerID="d031d5e7030a10c93109bcd69cd9a46d3247a354089a945e8aae7857a0f37496" Oct 08 19:59:26 crc kubenswrapper[4988]: I1008 19:59:26.093780 4988 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d031d5e7030a10c93109bcd69cd9a46d3247a354089a945e8aae7857a0f37496"} err="failed to get container status \"d031d5e7030a10c93109bcd69cd9a46d3247a354089a945e8aae7857a0f37496\": rpc error: code = NotFound desc = could not find container \"d031d5e7030a10c93109bcd69cd9a46d3247a354089a945e8aae7857a0f37496\": container with ID starting with d031d5e7030a10c93109bcd69cd9a46d3247a354089a945e8aae7857a0f37496 not found: ID does not exist" Oct 08 19:59:26 crc kubenswrapper[4988]: I1008 19:59:26.093812 4988 scope.go:117] "RemoveContainer" containerID="ac2490682666324232b8c023704986f0ffde028943fec0e1ead165a9f4c5de8a" Oct 08 19:59:26 crc kubenswrapper[4988]: E1008 19:59:26.094635 4988 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ac2490682666324232b8c023704986f0ffde028943fec0e1ead165a9f4c5de8a\": container with ID starting with ac2490682666324232b8c023704986f0ffde028943fec0e1ead165a9f4c5de8a not found: ID does not exist" containerID="ac2490682666324232b8c023704986f0ffde028943fec0e1ead165a9f4c5de8a" Oct 08 19:59:26 crc kubenswrapper[4988]: I1008 19:59:26.094660 4988 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ac2490682666324232b8c023704986f0ffde028943fec0e1ead165a9f4c5de8a"} err="failed to get container status \"ac2490682666324232b8c023704986f0ffde028943fec0e1ead165a9f4c5de8a\": rpc error: code = NotFound desc = could not find container \"ac2490682666324232b8c023704986f0ffde028943fec0e1ead165a9f4c5de8a\": container with ID starting with ac2490682666324232b8c023704986f0ffde028943fec0e1ead165a9f4c5de8a not found: ID does not exist" Oct 08 19:59:26 crc kubenswrapper[4988]: I1008 19:59:26.096911 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z56ql\" (UniqueName: \"kubernetes.io/projected/9c249a33-8704-4532-8000-279d5ba3a9dc-kube-api-access-z56ql\") pod \"ceilometer-0\" (UID: \"9c249a33-8704-4532-8000-279d5ba3a9dc\") " pod="openstack/ceilometer-0" Oct 08 19:59:26 crc kubenswrapper[4988]: I1008 19:59:26.097004 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9c249a33-8704-4532-8000-279d5ba3a9dc-log-httpd\") pod \"ceilometer-0\" (UID: \"9c249a33-8704-4532-8000-279d5ba3a9dc\") " pod="openstack/ceilometer-0" Oct 08 19:59:26 crc kubenswrapper[4988]: I1008 19:59:26.097034 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9c249a33-8704-4532-8000-279d5ba3a9dc-config-data\") pod \"ceilometer-0\" (UID: \"9c249a33-8704-4532-8000-279d5ba3a9dc\") " pod="openstack/ceilometer-0" Oct 08 19:59:26 crc kubenswrapper[4988]: I1008 19:59:26.097080 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9c249a33-8704-4532-8000-279d5ba3a9dc-run-httpd\") pod \"ceilometer-0\" (UID: \"9c249a33-8704-4532-8000-279d5ba3a9dc\") " pod="openstack/ceilometer-0" Oct 08 19:59:26 crc kubenswrapper[4988]: I1008 19:59:26.097115 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/9c249a33-8704-4532-8000-279d5ba3a9dc-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"9c249a33-8704-4532-8000-279d5ba3a9dc\") " pod="openstack/ceilometer-0" Oct 08 19:59:26 crc kubenswrapper[4988]: I1008 19:59:26.097130 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9c249a33-8704-4532-8000-279d5ba3a9dc-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"9c249a33-8704-4532-8000-279d5ba3a9dc\") " pod="openstack/ceilometer-0" Oct 08 19:59:26 crc kubenswrapper[4988]: I1008 19:59:26.097171 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9c249a33-8704-4532-8000-279d5ba3a9dc-scripts\") pod \"ceilometer-0\" (UID: \"9c249a33-8704-4532-8000-279d5ba3a9dc\") " pod="openstack/ceilometer-0" Oct 08 19:59:26 crc kubenswrapper[4988]: I1008 19:59:26.097617 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9c249a33-8704-4532-8000-279d5ba3a9dc-log-httpd\") pod \"ceilometer-0\" (UID: \"9c249a33-8704-4532-8000-279d5ba3a9dc\") " pod="openstack/ceilometer-0" Oct 08 19:59:26 crc kubenswrapper[4988]: I1008 19:59:26.099476 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9c249a33-8704-4532-8000-279d5ba3a9dc-run-httpd\") pod \"ceilometer-0\" (UID: \"9c249a33-8704-4532-8000-279d5ba3a9dc\") " pod="openstack/ceilometer-0" Oct 08 19:59:26 crc kubenswrapper[4988]: I1008 19:59:26.103106 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9c249a33-8704-4532-8000-279d5ba3a9dc-config-data\") pod \"ceilometer-0\" (UID: \"9c249a33-8704-4532-8000-279d5ba3a9dc\") " pod="openstack/ceilometer-0" Oct 08 19:59:26 crc kubenswrapper[4988]: I1008 19:59:26.106851 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9c249a33-8704-4532-8000-279d5ba3a9dc-scripts\") pod \"ceilometer-0\" (UID: \"9c249a33-8704-4532-8000-279d5ba3a9dc\") " pod="openstack/ceilometer-0" Oct 08 19:59:26 crc kubenswrapper[4988]: I1008 19:59:26.110953 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9c249a33-8704-4532-8000-279d5ba3a9dc-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"9c249a33-8704-4532-8000-279d5ba3a9dc\") " pod="openstack/ceilometer-0" Oct 08 19:59:26 crc kubenswrapper[4988]: I1008 19:59:26.111218 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/9c249a33-8704-4532-8000-279d5ba3a9dc-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"9c249a33-8704-4532-8000-279d5ba3a9dc\") " pod="openstack/ceilometer-0" Oct 08 19:59:26 crc kubenswrapper[4988]: I1008 19:59:26.124094 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z56ql\" (UniqueName: \"kubernetes.io/projected/9c249a33-8704-4532-8000-279d5ba3a9dc-kube-api-access-z56ql\") pod \"ceilometer-0\" (UID: \"9c249a33-8704-4532-8000-279d5ba3a9dc\") " pod="openstack/ceilometer-0" Oct 08 19:59:26 crc kubenswrapper[4988]: I1008 19:59:26.201216 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Oct 08 19:59:26 crc kubenswrapper[4988]: I1008 19:59:26.278607 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 08 19:59:26 crc kubenswrapper[4988]: I1008 19:59:26.402134 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-k4gdm\" (UniqueName: \"kubernetes.io/projected/d9fdb832-bcf7-4e2a-afe0-a01a5881cfb3-kube-api-access-k4gdm\") pod \"d9fdb832-bcf7-4e2a-afe0-a01a5881cfb3\" (UID: \"d9fdb832-bcf7-4e2a-afe0-a01a5881cfb3\") " Oct 08 19:59:26 crc kubenswrapper[4988]: I1008 19:59:26.440432 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d9fdb832-bcf7-4e2a-afe0-a01a5881cfb3-kube-api-access-k4gdm" (OuterVolumeSpecName: "kube-api-access-k4gdm") pod "d9fdb832-bcf7-4e2a-afe0-a01a5881cfb3" (UID: "d9fdb832-bcf7-4e2a-afe0-a01a5881cfb3"). InnerVolumeSpecName "kube-api-access-k4gdm". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 19:59:26 crc kubenswrapper[4988]: I1008 19:59:26.505772 4988 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-k4gdm\" (UniqueName: \"kubernetes.io/projected/d9fdb832-bcf7-4e2a-afe0-a01a5881cfb3-kube-api-access-k4gdm\") on node \"crc\" DevicePath \"\"" Oct 08 19:59:26 crc kubenswrapper[4988]: I1008 19:59:26.822001 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 08 19:59:26 crc kubenswrapper[4988]: I1008 19:59:26.856040 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"d9fdb832-bcf7-4e2a-afe0-a01a5881cfb3","Type":"ContainerDied","Data":"46bf7e2db5e29acd04805f971eda135c48ae4c40840889a78ecf852d40b43667"} Oct 08 19:59:26 crc kubenswrapper[4988]: I1008 19:59:26.856085 4988 scope.go:117] "RemoveContainer" containerID="d526c948e42b32ee91470150e213b56a919cb4fc6e9b6480ba2a38e3d49dd02e" Oct 08 19:59:26 crc kubenswrapper[4988]: I1008 19:59:26.856175 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Oct 08 19:59:26 crc kubenswrapper[4988]: I1008 19:59:26.891065 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 08 19:59:26 crc kubenswrapper[4988]: I1008 19:59:26.902810 4988 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 08 19:59:26 crc kubenswrapper[4988]: I1008 19:59:26.920287 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Oct 08 19:59:26 crc kubenswrapper[4988]: E1008 19:59:26.920892 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d9fdb832-bcf7-4e2a-afe0-a01a5881cfb3" containerName="kube-state-metrics" Oct 08 19:59:26 crc kubenswrapper[4988]: I1008 19:59:26.920915 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="d9fdb832-bcf7-4e2a-afe0-a01a5881cfb3" containerName="kube-state-metrics" Oct 08 19:59:26 crc kubenswrapper[4988]: I1008 19:59:26.921204 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="d9fdb832-bcf7-4e2a-afe0-a01a5881cfb3" containerName="kube-state-metrics" Oct 08 19:59:26 crc kubenswrapper[4988]: I1008 19:59:26.921970 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Oct 08 19:59:26 crc kubenswrapper[4988]: I1008 19:59:26.923828 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-kube-state-metrics-svc" Oct 08 19:59:26 crc kubenswrapper[4988]: I1008 19:59:26.924080 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"kube-state-metrics-tls-config" Oct 08 19:59:26 crc kubenswrapper[4988]: I1008 19:59:26.960716 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 08 19:59:27 crc kubenswrapper[4988]: I1008 19:59:27.122774 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/5e4fdf4f-092f-40a0-b38b-ad93f5bad528-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"5e4fdf4f-092f-40a0-b38b-ad93f5bad528\") " pod="openstack/kube-state-metrics-0" Oct 08 19:59:27 crc kubenswrapper[4988]: I1008 19:59:27.123498 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5e4fdf4f-092f-40a0-b38b-ad93f5bad528-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"5e4fdf4f-092f-40a0-b38b-ad93f5bad528\") " pod="openstack/kube-state-metrics-0" Oct 08 19:59:27 crc kubenswrapper[4988]: I1008 19:59:27.123657 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/5e4fdf4f-092f-40a0-b38b-ad93f5bad528-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"5e4fdf4f-092f-40a0-b38b-ad93f5bad528\") " pod="openstack/kube-state-metrics-0" Oct 08 19:59:27 crc kubenswrapper[4988]: I1008 19:59:27.123737 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ffkgn\" (UniqueName: \"kubernetes.io/projected/5e4fdf4f-092f-40a0-b38b-ad93f5bad528-kube-api-access-ffkgn\") pod \"kube-state-metrics-0\" (UID: \"5e4fdf4f-092f-40a0-b38b-ad93f5bad528\") " pod="openstack/kube-state-metrics-0" Oct 08 19:59:27 crc kubenswrapper[4988]: I1008 19:59:27.227489 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/5e4fdf4f-092f-40a0-b38b-ad93f5bad528-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"5e4fdf4f-092f-40a0-b38b-ad93f5bad528\") " pod="openstack/kube-state-metrics-0" Oct 08 19:59:27 crc kubenswrapper[4988]: I1008 19:59:27.227597 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5e4fdf4f-092f-40a0-b38b-ad93f5bad528-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"5e4fdf4f-092f-40a0-b38b-ad93f5bad528\") " pod="openstack/kube-state-metrics-0" Oct 08 19:59:27 crc kubenswrapper[4988]: I1008 19:59:27.227680 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/5e4fdf4f-092f-40a0-b38b-ad93f5bad528-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"5e4fdf4f-092f-40a0-b38b-ad93f5bad528\") " pod="openstack/kube-state-metrics-0" Oct 08 19:59:27 crc kubenswrapper[4988]: I1008 19:59:27.227740 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ffkgn\" (UniqueName: \"kubernetes.io/projected/5e4fdf4f-092f-40a0-b38b-ad93f5bad528-kube-api-access-ffkgn\") pod \"kube-state-metrics-0\" (UID: \"5e4fdf4f-092f-40a0-b38b-ad93f5bad528\") " pod="openstack/kube-state-metrics-0" Oct 08 19:59:27 crc kubenswrapper[4988]: I1008 19:59:27.235722 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/5e4fdf4f-092f-40a0-b38b-ad93f5bad528-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"5e4fdf4f-092f-40a0-b38b-ad93f5bad528\") " pod="openstack/kube-state-metrics-0" Oct 08 19:59:27 crc kubenswrapper[4988]: I1008 19:59:27.237809 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/5e4fdf4f-092f-40a0-b38b-ad93f5bad528-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"5e4fdf4f-092f-40a0-b38b-ad93f5bad528\") " pod="openstack/kube-state-metrics-0" Oct 08 19:59:27 crc kubenswrapper[4988]: I1008 19:59:27.238834 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5e4fdf4f-092f-40a0-b38b-ad93f5bad528-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"5e4fdf4f-092f-40a0-b38b-ad93f5bad528\") " pod="openstack/kube-state-metrics-0" Oct 08 19:59:27 crc kubenswrapper[4988]: I1008 19:59:27.245665 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ffkgn\" (UniqueName: \"kubernetes.io/projected/5e4fdf4f-092f-40a0-b38b-ad93f5bad528-kube-api-access-ffkgn\") pod \"kube-state-metrics-0\" (UID: \"5e4fdf4f-092f-40a0-b38b-ad93f5bad528\") " pod="openstack/kube-state-metrics-0" Oct 08 19:59:27 crc kubenswrapper[4988]: I1008 19:59:27.264944 4988 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1e77ab25-7879-4cb9-ae44-0e750ecb8123" path="/var/lib/kubelet/pods/1e77ab25-7879-4cb9-ae44-0e750ecb8123/volumes" Oct 08 19:59:27 crc kubenswrapper[4988]: I1008 19:59:27.265732 4988 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d9fdb832-bcf7-4e2a-afe0-a01a5881cfb3" path="/var/lib/kubelet/pods/d9fdb832-bcf7-4e2a-afe0-a01a5881cfb3/volumes" Oct 08 19:59:27 crc kubenswrapper[4988]: I1008 19:59:27.271266 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Oct 08 19:59:27 crc kubenswrapper[4988]: I1008 19:59:27.823605 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 08 19:59:27 crc kubenswrapper[4988]: W1008 19:59:27.834294 4988 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5e4fdf4f_092f_40a0_b38b_ad93f5bad528.slice/crio-950c8d2fd75aa1a6ae93b5dfcd02df29382a99d7ff637f273978f427b37ca956 WatchSource:0}: Error finding container 950c8d2fd75aa1a6ae93b5dfcd02df29382a99d7ff637f273978f427b37ca956: Status 404 returned error can't find the container with id 950c8d2fd75aa1a6ae93b5dfcd02df29382a99d7ff637f273978f427b37ca956 Oct 08 19:59:27 crc kubenswrapper[4988]: I1008 19:59:27.873766 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"6a1bc0e7-3209-4fa2-9a2a-c8b18121da5a","Type":"ContainerStarted","Data":"56698d6d5b5c0b4c4e8cea85cd63e9319099dd99593db5f129194b68561150dd"} Oct 08 19:59:27 crc kubenswrapper[4988]: I1008 19:59:27.873857 4988 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/aodh-0" podUID="6a1bc0e7-3209-4fa2-9a2a-c8b18121da5a" containerName="aodh-api" containerID="cri-o://de0e0960e8cf3d4761ad85747537742d917d43f51156f5145c593fb2bc682af3" gracePeriod=30 Oct 08 19:59:27 crc kubenswrapper[4988]: I1008 19:59:27.873898 4988 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/aodh-0" podUID="6a1bc0e7-3209-4fa2-9a2a-c8b18121da5a" containerName="aodh-listener" containerID="cri-o://56698d6d5b5c0b4c4e8cea85cd63e9319099dd99593db5f129194b68561150dd" gracePeriod=30 Oct 08 19:59:27 crc kubenswrapper[4988]: I1008 19:59:27.873954 4988 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/aodh-0" podUID="6a1bc0e7-3209-4fa2-9a2a-c8b18121da5a" containerName="aodh-notifier" containerID="cri-o://01e730efa7fc209fb2f43a464a61ae5e2d7bfaa803be6f3d5211a851dc991bd8" gracePeriod=30 Oct 08 19:59:27 crc kubenswrapper[4988]: I1008 19:59:27.874000 4988 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/aodh-0" podUID="6a1bc0e7-3209-4fa2-9a2a-c8b18121da5a" containerName="aodh-evaluator" containerID="cri-o://1a9c83e0bd2e4118cc11a9d647a155cb5f97ba61361dc0ba0e080435aaa48c98" gracePeriod=30 Oct 08 19:59:27 crc kubenswrapper[4988]: I1008 19:59:27.880293 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9c249a33-8704-4532-8000-279d5ba3a9dc","Type":"ContainerStarted","Data":"b22976a232053882f0cdc341913dfc95ecd1faa6a0161ed81be56e29e12079c1"} Oct 08 19:59:27 crc kubenswrapper[4988]: I1008 19:59:27.895789 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"5e4fdf4f-092f-40a0-b38b-ad93f5bad528","Type":"ContainerStarted","Data":"950c8d2fd75aa1a6ae93b5dfcd02df29382a99d7ff637f273978f427b37ca956"} Oct 08 19:59:27 crc kubenswrapper[4988]: I1008 19:59:27.899081 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/aodh-0" podStartSLOduration=2.462724422 podStartE2EDuration="8.89907054s" podCreationTimestamp="2025-10-08 19:59:19 +0000 UTC" firstStartedPulling="2025-10-08 19:59:20.870452826 +0000 UTC m=+6506.320295596" lastFinishedPulling="2025-10-08 19:59:27.306798944 +0000 UTC m=+6512.756641714" observedRunningTime="2025-10-08 19:59:27.897856301 +0000 UTC m=+6513.347699071" watchObservedRunningTime="2025-10-08 19:59:27.89907054 +0000 UTC m=+6513.348913310" Oct 08 19:59:28 crc kubenswrapper[4988]: I1008 19:59:28.166473 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 08 19:59:28 crc kubenswrapper[4988]: I1008 19:59:28.952279 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"5e4fdf4f-092f-40a0-b38b-ad93f5bad528","Type":"ContainerStarted","Data":"8da70093e66e0aa2969cd51ed3b30ae12ce0b2663a80662d69ef3fcf3d0b2cf3"} Oct 08 19:59:28 crc kubenswrapper[4988]: I1008 19:59:28.953993 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Oct 08 19:59:28 crc kubenswrapper[4988]: I1008 19:59:28.986899 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=2.572208432 podStartE2EDuration="2.986878029s" podCreationTimestamp="2025-10-08 19:59:26 +0000 UTC" firstStartedPulling="2025-10-08 19:59:27.838338076 +0000 UTC m=+6513.288180846" lastFinishedPulling="2025-10-08 19:59:28.253007673 +0000 UTC m=+6513.702850443" observedRunningTime="2025-10-08 19:59:28.982013565 +0000 UTC m=+6514.431856335" watchObservedRunningTime="2025-10-08 19:59:28.986878029 +0000 UTC m=+6514.436720799" Oct 08 19:59:29 crc kubenswrapper[4988]: I1008 19:59:29.002711 4988 generic.go:334] "Generic (PLEG): container finished" podID="6a1bc0e7-3209-4fa2-9a2a-c8b18121da5a" containerID="1a9c83e0bd2e4118cc11a9d647a155cb5f97ba61361dc0ba0e080435aaa48c98" exitCode=0 Oct 08 19:59:29 crc kubenswrapper[4988]: I1008 19:59:29.002752 4988 generic.go:334] "Generic (PLEG): container finished" podID="6a1bc0e7-3209-4fa2-9a2a-c8b18121da5a" containerID="de0e0960e8cf3d4761ad85747537742d917d43f51156f5145c593fb2bc682af3" exitCode=0 Oct 08 19:59:29 crc kubenswrapper[4988]: I1008 19:59:29.002793 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"6a1bc0e7-3209-4fa2-9a2a-c8b18121da5a","Type":"ContainerDied","Data":"1a9c83e0bd2e4118cc11a9d647a155cb5f97ba61361dc0ba0e080435aaa48c98"} Oct 08 19:59:29 crc kubenswrapper[4988]: I1008 19:59:29.002821 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"6a1bc0e7-3209-4fa2-9a2a-c8b18121da5a","Type":"ContainerDied","Data":"de0e0960e8cf3d4761ad85747537742d917d43f51156f5145c593fb2bc682af3"} Oct 08 19:59:29 crc kubenswrapper[4988]: I1008 19:59:29.027593 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9c249a33-8704-4532-8000-279d5ba3a9dc","Type":"ContainerStarted","Data":"c32d80732d725a57668990f1669c0fcade6c2b048c6d7e6b963a776b7cbdd34d"} Oct 08 19:59:29 crc kubenswrapper[4988]: I1008 19:59:29.027645 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9c249a33-8704-4532-8000-279d5ba3a9dc","Type":"ContainerStarted","Data":"9758669f34376a70ba0c1b2db31f16f4aaafdc6dc38fc47d14d6c1f969c101b7"} Oct 08 19:59:30 crc kubenswrapper[4988]: I1008 19:59:30.041101 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9c249a33-8704-4532-8000-279d5ba3a9dc","Type":"ContainerStarted","Data":"a5872364755c9a578d7cd0ce37d05553d41ccbd77915653f95a854c7b1fd6bc9"} Oct 08 19:59:31 crc kubenswrapper[4988]: I1008 19:59:31.043279 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-fscm7"] Oct 08 19:59:31 crc kubenswrapper[4988]: I1008 19:59:31.052812 4988 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-fscm7"] Oct 08 19:59:31 crc kubenswrapper[4988]: I1008 19:59:31.057748 4988 generic.go:334] "Generic (PLEG): container finished" podID="6a1bc0e7-3209-4fa2-9a2a-c8b18121da5a" containerID="01e730efa7fc209fb2f43a464a61ae5e2d7bfaa803be6f3d5211a851dc991bd8" exitCode=0 Oct 08 19:59:31 crc kubenswrapper[4988]: I1008 19:59:31.062928 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"6a1bc0e7-3209-4fa2-9a2a-c8b18121da5a","Type":"ContainerDied","Data":"01e730efa7fc209fb2f43a464a61ae5e2d7bfaa803be6f3d5211a851dc991bd8"} Oct 08 19:59:31 crc kubenswrapper[4988]: I1008 19:59:31.249197 4988 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f95e96b8-7f14-47a8-9d7e-1dca47b4b468" path="/var/lib/kubelet/pods/f95e96b8-7f14-47a8-9d7e-1dca47b4b468/volumes" Oct 08 19:59:32 crc kubenswrapper[4988]: I1008 19:59:32.070337 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9c249a33-8704-4532-8000-279d5ba3a9dc","Type":"ContainerStarted","Data":"b1b18d2ffab25b26a8894864b830724a3797eb4726b69c4d2347efa7e7a6adff"} Oct 08 19:59:32 crc kubenswrapper[4988]: I1008 19:59:32.070602 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 08 19:59:32 crc kubenswrapper[4988]: I1008 19:59:32.070587 4988 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="9c249a33-8704-4532-8000-279d5ba3a9dc" containerName="ceilometer-central-agent" containerID="cri-o://c32d80732d725a57668990f1669c0fcade6c2b048c6d7e6b963a776b7cbdd34d" gracePeriod=30 Oct 08 19:59:32 crc kubenswrapper[4988]: I1008 19:59:32.070718 4988 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="9c249a33-8704-4532-8000-279d5ba3a9dc" containerName="proxy-httpd" containerID="cri-o://b1b18d2ffab25b26a8894864b830724a3797eb4726b69c4d2347efa7e7a6adff" gracePeriod=30 Oct 08 19:59:32 crc kubenswrapper[4988]: I1008 19:59:32.070776 4988 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="9c249a33-8704-4532-8000-279d5ba3a9dc" containerName="sg-core" containerID="cri-o://a5872364755c9a578d7cd0ce37d05553d41ccbd77915653f95a854c7b1fd6bc9" gracePeriod=30 Oct 08 19:59:32 crc kubenswrapper[4988]: I1008 19:59:32.070832 4988 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="9c249a33-8704-4532-8000-279d5ba3a9dc" containerName="ceilometer-notification-agent" containerID="cri-o://9758669f34376a70ba0c1b2db31f16f4aaafdc6dc38fc47d14d6c1f969c101b7" gracePeriod=30 Oct 08 19:59:32 crc kubenswrapper[4988]: I1008 19:59:32.108207 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=3.54705821 podStartE2EDuration="7.108190773s" podCreationTimestamp="2025-10-08 19:59:25 +0000 UTC" firstStartedPulling="2025-10-08 19:59:27.217284082 +0000 UTC m=+6512.667126892" lastFinishedPulling="2025-10-08 19:59:30.778416675 +0000 UTC m=+6516.228259455" observedRunningTime="2025-10-08 19:59:32.099503306 +0000 UTC m=+6517.549346086" watchObservedRunningTime="2025-10-08 19:59:32.108190773 +0000 UTC m=+6517.558033543" Oct 08 19:59:33 crc kubenswrapper[4988]: I1008 19:59:33.089187 4988 generic.go:334] "Generic (PLEG): container finished" podID="9c249a33-8704-4532-8000-279d5ba3a9dc" containerID="b1b18d2ffab25b26a8894864b830724a3797eb4726b69c4d2347efa7e7a6adff" exitCode=0 Oct 08 19:59:33 crc kubenswrapper[4988]: I1008 19:59:33.089417 4988 generic.go:334] "Generic (PLEG): container finished" podID="9c249a33-8704-4532-8000-279d5ba3a9dc" containerID="a5872364755c9a578d7cd0ce37d05553d41ccbd77915653f95a854c7b1fd6bc9" exitCode=2 Oct 08 19:59:33 crc kubenswrapper[4988]: I1008 19:59:33.089426 4988 generic.go:334] "Generic (PLEG): container finished" podID="9c249a33-8704-4532-8000-279d5ba3a9dc" containerID="9758669f34376a70ba0c1b2db31f16f4aaafdc6dc38fc47d14d6c1f969c101b7" exitCode=0 Oct 08 19:59:33 crc kubenswrapper[4988]: I1008 19:59:33.089444 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9c249a33-8704-4532-8000-279d5ba3a9dc","Type":"ContainerDied","Data":"b1b18d2ffab25b26a8894864b830724a3797eb4726b69c4d2347efa7e7a6adff"} Oct 08 19:59:33 crc kubenswrapper[4988]: I1008 19:59:33.089470 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9c249a33-8704-4532-8000-279d5ba3a9dc","Type":"ContainerDied","Data":"a5872364755c9a578d7cd0ce37d05553d41ccbd77915653f95a854c7b1fd6bc9"} Oct 08 19:59:33 crc kubenswrapper[4988]: I1008 19:59:33.089479 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9c249a33-8704-4532-8000-279d5ba3a9dc","Type":"ContainerDied","Data":"9758669f34376a70ba0c1b2db31f16f4aaafdc6dc38fc47d14d6c1f969c101b7"} Oct 08 19:59:34 crc kubenswrapper[4988]: I1008 19:59:34.238737 4988 scope.go:117] "RemoveContainer" containerID="439af713ff941f973b8758bd6af3c097808f76ff520b96ff10bd3268811393f7" Oct 08 19:59:34 crc kubenswrapper[4988]: E1008 19:59:34.239538 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nm54f_openshift-machine-config-operator(f328a20b-5283-407e-8a93-2ab2da599f18)\"" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" Oct 08 19:59:35 crc kubenswrapper[4988]: I1008 19:59:35.124173 4988 generic.go:334] "Generic (PLEG): container finished" podID="9c249a33-8704-4532-8000-279d5ba3a9dc" containerID="c32d80732d725a57668990f1669c0fcade6c2b048c6d7e6b963a776b7cbdd34d" exitCode=0 Oct 08 19:59:35 crc kubenswrapper[4988]: I1008 19:59:35.124578 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9c249a33-8704-4532-8000-279d5ba3a9dc","Type":"ContainerDied","Data":"c32d80732d725a57668990f1669c0fcade6c2b048c6d7e6b963a776b7cbdd34d"} Oct 08 19:59:35 crc kubenswrapper[4988]: I1008 19:59:35.278913 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 08 19:59:35 crc kubenswrapper[4988]: I1008 19:59:35.366968 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9c249a33-8704-4532-8000-279d5ba3a9dc-config-data\") pod \"9c249a33-8704-4532-8000-279d5ba3a9dc\" (UID: \"9c249a33-8704-4532-8000-279d5ba3a9dc\") " Oct 08 19:59:35 crc kubenswrapper[4988]: I1008 19:59:35.367092 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9c249a33-8704-4532-8000-279d5ba3a9dc-log-httpd\") pod \"9c249a33-8704-4532-8000-279d5ba3a9dc\" (UID: \"9c249a33-8704-4532-8000-279d5ba3a9dc\") " Oct 08 19:59:35 crc kubenswrapper[4988]: I1008 19:59:35.367167 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z56ql\" (UniqueName: \"kubernetes.io/projected/9c249a33-8704-4532-8000-279d5ba3a9dc-kube-api-access-z56ql\") pod \"9c249a33-8704-4532-8000-279d5ba3a9dc\" (UID: \"9c249a33-8704-4532-8000-279d5ba3a9dc\") " Oct 08 19:59:35 crc kubenswrapper[4988]: I1008 19:59:35.367206 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9c249a33-8704-4532-8000-279d5ba3a9dc-scripts\") pod \"9c249a33-8704-4532-8000-279d5ba3a9dc\" (UID: \"9c249a33-8704-4532-8000-279d5ba3a9dc\") " Oct 08 19:59:35 crc kubenswrapper[4988]: I1008 19:59:35.367237 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9c249a33-8704-4532-8000-279d5ba3a9dc-run-httpd\") pod \"9c249a33-8704-4532-8000-279d5ba3a9dc\" (UID: \"9c249a33-8704-4532-8000-279d5ba3a9dc\") " Oct 08 19:59:35 crc kubenswrapper[4988]: I1008 19:59:35.367278 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9c249a33-8704-4532-8000-279d5ba3a9dc-combined-ca-bundle\") pod \"9c249a33-8704-4532-8000-279d5ba3a9dc\" (UID: \"9c249a33-8704-4532-8000-279d5ba3a9dc\") " Oct 08 19:59:35 crc kubenswrapper[4988]: I1008 19:59:35.367468 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/9c249a33-8704-4532-8000-279d5ba3a9dc-sg-core-conf-yaml\") pod \"9c249a33-8704-4532-8000-279d5ba3a9dc\" (UID: \"9c249a33-8704-4532-8000-279d5ba3a9dc\") " Oct 08 19:59:35 crc kubenswrapper[4988]: I1008 19:59:35.369557 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9c249a33-8704-4532-8000-279d5ba3a9dc-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "9c249a33-8704-4532-8000-279d5ba3a9dc" (UID: "9c249a33-8704-4532-8000-279d5ba3a9dc"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 19:59:35 crc kubenswrapper[4988]: I1008 19:59:35.369603 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9c249a33-8704-4532-8000-279d5ba3a9dc-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "9c249a33-8704-4532-8000-279d5ba3a9dc" (UID: "9c249a33-8704-4532-8000-279d5ba3a9dc"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 19:59:35 crc kubenswrapper[4988]: I1008 19:59:35.381158 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9c249a33-8704-4532-8000-279d5ba3a9dc-kube-api-access-z56ql" (OuterVolumeSpecName: "kube-api-access-z56ql") pod "9c249a33-8704-4532-8000-279d5ba3a9dc" (UID: "9c249a33-8704-4532-8000-279d5ba3a9dc"). InnerVolumeSpecName "kube-api-access-z56ql". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 19:59:35 crc kubenswrapper[4988]: I1008 19:59:35.381290 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9c249a33-8704-4532-8000-279d5ba3a9dc-scripts" (OuterVolumeSpecName: "scripts") pod "9c249a33-8704-4532-8000-279d5ba3a9dc" (UID: "9c249a33-8704-4532-8000-279d5ba3a9dc"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 19:59:35 crc kubenswrapper[4988]: I1008 19:59:35.427966 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9c249a33-8704-4532-8000-279d5ba3a9dc-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "9c249a33-8704-4532-8000-279d5ba3a9dc" (UID: "9c249a33-8704-4532-8000-279d5ba3a9dc"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 19:59:35 crc kubenswrapper[4988]: I1008 19:59:35.470171 4988 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9c249a33-8704-4532-8000-279d5ba3a9dc-scripts\") on node \"crc\" DevicePath \"\"" Oct 08 19:59:35 crc kubenswrapper[4988]: I1008 19:59:35.470458 4988 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9c249a33-8704-4532-8000-279d5ba3a9dc-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 08 19:59:35 crc kubenswrapper[4988]: I1008 19:59:35.470536 4988 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/9c249a33-8704-4532-8000-279d5ba3a9dc-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Oct 08 19:59:35 crc kubenswrapper[4988]: I1008 19:59:35.470603 4988 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9c249a33-8704-4532-8000-279d5ba3a9dc-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 08 19:59:35 crc kubenswrapper[4988]: I1008 19:59:35.470667 4988 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z56ql\" (UniqueName: \"kubernetes.io/projected/9c249a33-8704-4532-8000-279d5ba3a9dc-kube-api-access-z56ql\") on node \"crc\" DevicePath \"\"" Oct 08 19:59:35 crc kubenswrapper[4988]: I1008 19:59:35.483538 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9c249a33-8704-4532-8000-279d5ba3a9dc-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "9c249a33-8704-4532-8000-279d5ba3a9dc" (UID: "9c249a33-8704-4532-8000-279d5ba3a9dc"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 19:59:35 crc kubenswrapper[4988]: I1008 19:59:35.506932 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9c249a33-8704-4532-8000-279d5ba3a9dc-config-data" (OuterVolumeSpecName: "config-data") pod "9c249a33-8704-4532-8000-279d5ba3a9dc" (UID: "9c249a33-8704-4532-8000-279d5ba3a9dc"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 19:59:35 crc kubenswrapper[4988]: I1008 19:59:35.572941 4988 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9c249a33-8704-4532-8000-279d5ba3a9dc-config-data\") on node \"crc\" DevicePath \"\"" Oct 08 19:59:35 crc kubenswrapper[4988]: I1008 19:59:35.572984 4988 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9c249a33-8704-4532-8000-279d5ba3a9dc-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 19:59:36 crc kubenswrapper[4988]: I1008 19:59:36.148142 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9c249a33-8704-4532-8000-279d5ba3a9dc","Type":"ContainerDied","Data":"b22976a232053882f0cdc341913dfc95ecd1faa6a0161ed81be56e29e12079c1"} Oct 08 19:59:36 crc kubenswrapper[4988]: I1008 19:59:36.148538 4988 scope.go:117] "RemoveContainer" containerID="b1b18d2ffab25b26a8894864b830724a3797eb4726b69c4d2347efa7e7a6adff" Oct 08 19:59:36 crc kubenswrapper[4988]: I1008 19:59:36.148246 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 08 19:59:36 crc kubenswrapper[4988]: I1008 19:59:36.194932 4988 scope.go:117] "RemoveContainer" containerID="a5872364755c9a578d7cd0ce37d05553d41ccbd77915653f95a854c7b1fd6bc9" Oct 08 19:59:36 crc kubenswrapper[4988]: I1008 19:59:36.209406 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 08 19:59:36 crc kubenswrapper[4988]: I1008 19:59:36.220333 4988 scope.go:117] "RemoveContainer" containerID="9758669f34376a70ba0c1b2db31f16f4aaafdc6dc38fc47d14d6c1f969c101b7" Oct 08 19:59:36 crc kubenswrapper[4988]: I1008 19:59:36.230575 4988 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Oct 08 19:59:36 crc kubenswrapper[4988]: I1008 19:59:36.238765 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 08 19:59:36 crc kubenswrapper[4988]: E1008 19:59:36.239190 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9c249a33-8704-4532-8000-279d5ba3a9dc" containerName="ceilometer-notification-agent" Oct 08 19:59:36 crc kubenswrapper[4988]: I1008 19:59:36.239209 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="9c249a33-8704-4532-8000-279d5ba3a9dc" containerName="ceilometer-notification-agent" Oct 08 19:59:36 crc kubenswrapper[4988]: E1008 19:59:36.239227 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9c249a33-8704-4532-8000-279d5ba3a9dc" containerName="proxy-httpd" Oct 08 19:59:36 crc kubenswrapper[4988]: I1008 19:59:36.239235 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="9c249a33-8704-4532-8000-279d5ba3a9dc" containerName="proxy-httpd" Oct 08 19:59:36 crc kubenswrapper[4988]: E1008 19:59:36.239257 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9c249a33-8704-4532-8000-279d5ba3a9dc" containerName="ceilometer-central-agent" Oct 08 19:59:36 crc kubenswrapper[4988]: I1008 19:59:36.239263 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="9c249a33-8704-4532-8000-279d5ba3a9dc" containerName="ceilometer-central-agent" Oct 08 19:59:36 crc kubenswrapper[4988]: E1008 19:59:36.239288 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9c249a33-8704-4532-8000-279d5ba3a9dc" containerName="sg-core" Oct 08 19:59:36 crc kubenswrapper[4988]: I1008 19:59:36.239293 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="9c249a33-8704-4532-8000-279d5ba3a9dc" containerName="sg-core" Oct 08 19:59:36 crc kubenswrapper[4988]: I1008 19:59:36.239496 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="9c249a33-8704-4532-8000-279d5ba3a9dc" containerName="sg-core" Oct 08 19:59:36 crc kubenswrapper[4988]: I1008 19:59:36.239515 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="9c249a33-8704-4532-8000-279d5ba3a9dc" containerName="ceilometer-central-agent" Oct 08 19:59:36 crc kubenswrapper[4988]: I1008 19:59:36.239525 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="9c249a33-8704-4532-8000-279d5ba3a9dc" containerName="proxy-httpd" Oct 08 19:59:36 crc kubenswrapper[4988]: I1008 19:59:36.239542 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="9c249a33-8704-4532-8000-279d5ba3a9dc" containerName="ceilometer-notification-agent" Oct 08 19:59:36 crc kubenswrapper[4988]: I1008 19:59:36.241403 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 08 19:59:36 crc kubenswrapper[4988]: I1008 19:59:36.243860 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 08 19:59:36 crc kubenswrapper[4988]: I1008 19:59:36.244193 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Oct 08 19:59:36 crc kubenswrapper[4988]: I1008 19:59:36.244576 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 08 19:59:36 crc kubenswrapper[4988]: I1008 19:59:36.246751 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 08 19:59:36 crc kubenswrapper[4988]: I1008 19:59:36.258819 4988 scope.go:117] "RemoveContainer" containerID="c32d80732d725a57668990f1669c0fcade6c2b048c6d7e6b963a776b7cbdd34d" Oct 08 19:59:36 crc kubenswrapper[4988]: I1008 19:59:36.302755 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a7b11b95-f581-4db1-ac20-67200519b7bc-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"a7b11b95-f581-4db1-ac20-67200519b7bc\") " pod="openstack/ceilometer-0" Oct 08 19:59:36 crc kubenswrapper[4988]: I1008 19:59:36.303746 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a7b11b95-f581-4db1-ac20-67200519b7bc-config-data\") pod \"ceilometer-0\" (UID: \"a7b11b95-f581-4db1-ac20-67200519b7bc\") " pod="openstack/ceilometer-0" Oct 08 19:59:36 crc kubenswrapper[4988]: I1008 19:59:36.303804 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-77qt9\" (UniqueName: \"kubernetes.io/projected/a7b11b95-f581-4db1-ac20-67200519b7bc-kube-api-access-77qt9\") pod \"ceilometer-0\" (UID: \"a7b11b95-f581-4db1-ac20-67200519b7bc\") " pod="openstack/ceilometer-0" Oct 08 19:59:36 crc kubenswrapper[4988]: I1008 19:59:36.303914 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a7b11b95-f581-4db1-ac20-67200519b7bc-log-httpd\") pod \"ceilometer-0\" (UID: \"a7b11b95-f581-4db1-ac20-67200519b7bc\") " pod="openstack/ceilometer-0" Oct 08 19:59:36 crc kubenswrapper[4988]: I1008 19:59:36.303948 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/a7b11b95-f581-4db1-ac20-67200519b7bc-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"a7b11b95-f581-4db1-ac20-67200519b7bc\") " pod="openstack/ceilometer-0" Oct 08 19:59:36 crc kubenswrapper[4988]: I1008 19:59:36.304072 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a7b11b95-f581-4db1-ac20-67200519b7bc-run-httpd\") pod \"ceilometer-0\" (UID: \"a7b11b95-f581-4db1-ac20-67200519b7bc\") " pod="openstack/ceilometer-0" Oct 08 19:59:36 crc kubenswrapper[4988]: I1008 19:59:36.304198 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a7b11b95-f581-4db1-ac20-67200519b7bc-scripts\") pod \"ceilometer-0\" (UID: \"a7b11b95-f581-4db1-ac20-67200519b7bc\") " pod="openstack/ceilometer-0" Oct 08 19:59:36 crc kubenswrapper[4988]: I1008 19:59:36.304229 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/a7b11b95-f581-4db1-ac20-67200519b7bc-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"a7b11b95-f581-4db1-ac20-67200519b7bc\") " pod="openstack/ceilometer-0" Oct 08 19:59:36 crc kubenswrapper[4988]: I1008 19:59:36.406171 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a7b11b95-f581-4db1-ac20-67200519b7bc-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"a7b11b95-f581-4db1-ac20-67200519b7bc\") " pod="openstack/ceilometer-0" Oct 08 19:59:36 crc kubenswrapper[4988]: I1008 19:59:36.406602 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a7b11b95-f581-4db1-ac20-67200519b7bc-config-data\") pod \"ceilometer-0\" (UID: \"a7b11b95-f581-4db1-ac20-67200519b7bc\") " pod="openstack/ceilometer-0" Oct 08 19:59:36 crc kubenswrapper[4988]: I1008 19:59:36.407858 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-77qt9\" (UniqueName: \"kubernetes.io/projected/a7b11b95-f581-4db1-ac20-67200519b7bc-kube-api-access-77qt9\") pod \"ceilometer-0\" (UID: \"a7b11b95-f581-4db1-ac20-67200519b7bc\") " pod="openstack/ceilometer-0" Oct 08 19:59:36 crc kubenswrapper[4988]: I1008 19:59:36.408086 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a7b11b95-f581-4db1-ac20-67200519b7bc-log-httpd\") pod \"ceilometer-0\" (UID: \"a7b11b95-f581-4db1-ac20-67200519b7bc\") " pod="openstack/ceilometer-0" Oct 08 19:59:36 crc kubenswrapper[4988]: I1008 19:59:36.408297 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/a7b11b95-f581-4db1-ac20-67200519b7bc-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"a7b11b95-f581-4db1-ac20-67200519b7bc\") " pod="openstack/ceilometer-0" Oct 08 19:59:36 crc kubenswrapper[4988]: I1008 19:59:36.408567 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a7b11b95-f581-4db1-ac20-67200519b7bc-run-httpd\") pod \"ceilometer-0\" (UID: \"a7b11b95-f581-4db1-ac20-67200519b7bc\") " pod="openstack/ceilometer-0" Oct 08 19:59:36 crc kubenswrapper[4988]: I1008 19:59:36.408788 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a7b11b95-f581-4db1-ac20-67200519b7bc-scripts\") pod \"ceilometer-0\" (UID: \"a7b11b95-f581-4db1-ac20-67200519b7bc\") " pod="openstack/ceilometer-0" Oct 08 19:59:36 crc kubenswrapper[4988]: I1008 19:59:36.408918 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a7b11b95-f581-4db1-ac20-67200519b7bc-run-httpd\") pod \"ceilometer-0\" (UID: \"a7b11b95-f581-4db1-ac20-67200519b7bc\") " pod="openstack/ceilometer-0" Oct 08 19:59:36 crc kubenswrapper[4988]: I1008 19:59:36.408834 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a7b11b95-f581-4db1-ac20-67200519b7bc-log-httpd\") pod \"ceilometer-0\" (UID: \"a7b11b95-f581-4db1-ac20-67200519b7bc\") " pod="openstack/ceilometer-0" Oct 08 19:59:36 crc kubenswrapper[4988]: I1008 19:59:36.409150 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/a7b11b95-f581-4db1-ac20-67200519b7bc-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"a7b11b95-f581-4db1-ac20-67200519b7bc\") " pod="openstack/ceilometer-0" Oct 08 19:59:36 crc kubenswrapper[4988]: I1008 19:59:36.412717 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a7b11b95-f581-4db1-ac20-67200519b7bc-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"a7b11b95-f581-4db1-ac20-67200519b7bc\") " pod="openstack/ceilometer-0" Oct 08 19:59:36 crc kubenswrapper[4988]: I1008 19:59:36.413923 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/a7b11b95-f581-4db1-ac20-67200519b7bc-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"a7b11b95-f581-4db1-ac20-67200519b7bc\") " pod="openstack/ceilometer-0" Oct 08 19:59:36 crc kubenswrapper[4988]: I1008 19:59:36.418088 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/a7b11b95-f581-4db1-ac20-67200519b7bc-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"a7b11b95-f581-4db1-ac20-67200519b7bc\") " pod="openstack/ceilometer-0" Oct 08 19:59:36 crc kubenswrapper[4988]: I1008 19:59:36.418794 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a7b11b95-f581-4db1-ac20-67200519b7bc-config-data\") pod \"ceilometer-0\" (UID: \"a7b11b95-f581-4db1-ac20-67200519b7bc\") " pod="openstack/ceilometer-0" Oct 08 19:59:36 crc kubenswrapper[4988]: I1008 19:59:36.430001 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a7b11b95-f581-4db1-ac20-67200519b7bc-scripts\") pod \"ceilometer-0\" (UID: \"a7b11b95-f581-4db1-ac20-67200519b7bc\") " pod="openstack/ceilometer-0" Oct 08 19:59:36 crc kubenswrapper[4988]: I1008 19:59:36.437202 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-77qt9\" (UniqueName: \"kubernetes.io/projected/a7b11b95-f581-4db1-ac20-67200519b7bc-kube-api-access-77qt9\") pod \"ceilometer-0\" (UID: \"a7b11b95-f581-4db1-ac20-67200519b7bc\") " pod="openstack/ceilometer-0" Oct 08 19:59:36 crc kubenswrapper[4988]: I1008 19:59:36.563493 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 08 19:59:37 crc kubenswrapper[4988]: I1008 19:59:37.143829 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 08 19:59:37 crc kubenswrapper[4988]: I1008 19:59:37.173884 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a7b11b95-f581-4db1-ac20-67200519b7bc","Type":"ContainerStarted","Data":"71c1719f86b22f0bed5fd44da08c493d0629a469467591e24b6df0ea3af18aba"} Oct 08 19:59:37 crc kubenswrapper[4988]: I1008 19:59:37.250250 4988 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9c249a33-8704-4532-8000-279d5ba3a9dc" path="/var/lib/kubelet/pods/9c249a33-8704-4532-8000-279d5ba3a9dc/volumes" Oct 08 19:59:37 crc kubenswrapper[4988]: I1008 19:59:37.285231 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Oct 08 19:59:38 crc kubenswrapper[4988]: I1008 19:59:38.184456 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a7b11b95-f581-4db1-ac20-67200519b7bc","Type":"ContainerStarted","Data":"e36954d1ad1e400538183bd7e28db40df6287c0a186880b2b0d47a4dbd751b7f"} Oct 08 19:59:39 crc kubenswrapper[4988]: I1008 19:59:39.195096 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a7b11b95-f581-4db1-ac20-67200519b7bc","Type":"ContainerStarted","Data":"d75e727aac5086c60b88d6324af0cd07b722c3b382b0160b07279d3dccde1377"} Oct 08 19:59:40 crc kubenswrapper[4988]: I1008 19:59:40.210078 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a7b11b95-f581-4db1-ac20-67200519b7bc","Type":"ContainerStarted","Data":"52f279bc271928f43604447e60d1400be38cf70fa3203780db08b745737f04ba"} Oct 08 19:59:42 crc kubenswrapper[4988]: I1008 19:59:42.228951 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a7b11b95-f581-4db1-ac20-67200519b7bc","Type":"ContainerStarted","Data":"1be207131ebbf86d7aba8da49ac4966c64eaca3e643044afdac23f9354033a53"} Oct 08 19:59:42 crc kubenswrapper[4988]: I1008 19:59:42.231211 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 08 19:59:42 crc kubenswrapper[4988]: I1008 19:59:42.265941 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=1.949069883 podStartE2EDuration="6.265867886s" podCreationTimestamp="2025-10-08 19:59:36 +0000 UTC" firstStartedPulling="2025-10-08 19:59:37.148760421 +0000 UTC m=+6522.598603201" lastFinishedPulling="2025-10-08 19:59:41.465558394 +0000 UTC m=+6526.915401204" observedRunningTime="2025-10-08 19:59:42.250324121 +0000 UTC m=+6527.700166911" watchObservedRunningTime="2025-10-08 19:59:42.265867886 +0000 UTC m=+6527.715710696" Oct 08 19:59:46 crc kubenswrapper[4988]: I1008 19:59:46.238962 4988 scope.go:117] "RemoveContainer" containerID="439af713ff941f973b8758bd6af3c097808f76ff520b96ff10bd3268811393f7" Oct 08 19:59:46 crc kubenswrapper[4988]: E1008 19:59:46.239861 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nm54f_openshift-machine-config-operator(f328a20b-5283-407e-8a93-2ab2da599f18)\"" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" Oct 08 19:59:50 crc kubenswrapper[4988]: I1008 19:59:50.054933 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-tmx75"] Oct 08 19:59:50 crc kubenswrapper[4988]: I1008 19:59:50.067079 4988 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-tmx75"] Oct 08 19:59:51 crc kubenswrapper[4988]: I1008 19:59:51.036373 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-cell-mapping-jtj56"] Oct 08 19:59:51 crc kubenswrapper[4988]: I1008 19:59:51.047826 4988 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-cell-mapping-jtj56"] Oct 08 19:59:51 crc kubenswrapper[4988]: I1008 19:59:51.251955 4988 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="586b0be5-17fd-4984-87ba-a8c2862ddbc0" path="/var/lib/kubelet/pods/586b0be5-17fd-4984-87ba-a8c2862ddbc0/volumes" Oct 08 19:59:51 crc kubenswrapper[4988]: I1008 19:59:51.252945 4988 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f6f69dea-85ca-4d7e-9f1c-05f11cadd3b9" path="/var/lib/kubelet/pods/f6f69dea-85ca-4d7e-9f1c-05f11cadd3b9/volumes" Oct 08 19:59:58 crc kubenswrapper[4988]: I1008 19:59:58.238189 4988 scope.go:117] "RemoveContainer" containerID="439af713ff941f973b8758bd6af3c097808f76ff520b96ff10bd3268811393f7" Oct 08 19:59:58 crc kubenswrapper[4988]: E1008 19:59:58.239157 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nm54f_openshift-machine-config-operator(f328a20b-5283-407e-8a93-2ab2da599f18)\"" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" Oct 08 19:59:58 crc kubenswrapper[4988]: I1008 19:59:58.486954 4988 generic.go:334] "Generic (PLEG): container finished" podID="6a1bc0e7-3209-4fa2-9a2a-c8b18121da5a" containerID="56698d6d5b5c0b4c4e8cea85cd63e9319099dd99593db5f129194b68561150dd" exitCode=137 Oct 08 19:59:58 crc kubenswrapper[4988]: I1008 19:59:58.487012 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"6a1bc0e7-3209-4fa2-9a2a-c8b18121da5a","Type":"ContainerDied","Data":"56698d6d5b5c0b4c4e8cea85cd63e9319099dd99593db5f129194b68561150dd"} Oct 08 19:59:58 crc kubenswrapper[4988]: I1008 19:59:58.549034 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-0" Oct 08 19:59:58 crc kubenswrapper[4988]: I1008 19:59:58.663267 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6a1bc0e7-3209-4fa2-9a2a-c8b18121da5a-scripts\") pod \"6a1bc0e7-3209-4fa2-9a2a-c8b18121da5a\" (UID: \"6a1bc0e7-3209-4fa2-9a2a-c8b18121da5a\") " Oct 08 19:59:58 crc kubenswrapper[4988]: I1008 19:59:58.663416 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6a1bc0e7-3209-4fa2-9a2a-c8b18121da5a-config-data\") pod \"6a1bc0e7-3209-4fa2-9a2a-c8b18121da5a\" (UID: \"6a1bc0e7-3209-4fa2-9a2a-c8b18121da5a\") " Oct 08 19:59:58 crc kubenswrapper[4988]: I1008 19:59:58.663518 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6prdj\" (UniqueName: \"kubernetes.io/projected/6a1bc0e7-3209-4fa2-9a2a-c8b18121da5a-kube-api-access-6prdj\") pod \"6a1bc0e7-3209-4fa2-9a2a-c8b18121da5a\" (UID: \"6a1bc0e7-3209-4fa2-9a2a-c8b18121da5a\") " Oct 08 19:59:58 crc kubenswrapper[4988]: I1008 19:59:58.663545 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6a1bc0e7-3209-4fa2-9a2a-c8b18121da5a-combined-ca-bundle\") pod \"6a1bc0e7-3209-4fa2-9a2a-c8b18121da5a\" (UID: \"6a1bc0e7-3209-4fa2-9a2a-c8b18121da5a\") " Oct 08 19:59:58 crc kubenswrapper[4988]: I1008 19:59:58.668696 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6a1bc0e7-3209-4fa2-9a2a-c8b18121da5a-scripts" (OuterVolumeSpecName: "scripts") pod "6a1bc0e7-3209-4fa2-9a2a-c8b18121da5a" (UID: "6a1bc0e7-3209-4fa2-9a2a-c8b18121da5a"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 19:59:58 crc kubenswrapper[4988]: I1008 19:59:58.677716 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6a1bc0e7-3209-4fa2-9a2a-c8b18121da5a-kube-api-access-6prdj" (OuterVolumeSpecName: "kube-api-access-6prdj") pod "6a1bc0e7-3209-4fa2-9a2a-c8b18121da5a" (UID: "6a1bc0e7-3209-4fa2-9a2a-c8b18121da5a"). InnerVolumeSpecName "kube-api-access-6prdj". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 19:59:58 crc kubenswrapper[4988]: I1008 19:59:58.768966 4988 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6prdj\" (UniqueName: \"kubernetes.io/projected/6a1bc0e7-3209-4fa2-9a2a-c8b18121da5a-kube-api-access-6prdj\") on node \"crc\" DevicePath \"\"" Oct 08 19:59:58 crc kubenswrapper[4988]: I1008 19:59:58.769021 4988 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6a1bc0e7-3209-4fa2-9a2a-c8b18121da5a-scripts\") on node \"crc\" DevicePath \"\"" Oct 08 19:59:58 crc kubenswrapper[4988]: I1008 19:59:58.790574 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6a1bc0e7-3209-4fa2-9a2a-c8b18121da5a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "6a1bc0e7-3209-4fa2-9a2a-c8b18121da5a" (UID: "6a1bc0e7-3209-4fa2-9a2a-c8b18121da5a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 19:59:58 crc kubenswrapper[4988]: I1008 19:59:58.790639 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6a1bc0e7-3209-4fa2-9a2a-c8b18121da5a-config-data" (OuterVolumeSpecName: "config-data") pod "6a1bc0e7-3209-4fa2-9a2a-c8b18121da5a" (UID: "6a1bc0e7-3209-4fa2-9a2a-c8b18121da5a"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 19:59:58 crc kubenswrapper[4988]: I1008 19:59:58.870791 4988 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6a1bc0e7-3209-4fa2-9a2a-c8b18121da5a-config-data\") on node \"crc\" DevicePath \"\"" Oct 08 19:59:58 crc kubenswrapper[4988]: I1008 19:59:58.870823 4988 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6a1bc0e7-3209-4fa2-9a2a-c8b18121da5a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 19:59:59 crc kubenswrapper[4988]: I1008 19:59:59.506866 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"6a1bc0e7-3209-4fa2-9a2a-c8b18121da5a","Type":"ContainerDied","Data":"f7630bfffa89fc3a81a120034a59d2e5c8e4f920e056a9b9251b972ddc7c9ff8"} Oct 08 19:59:59 crc kubenswrapper[4988]: I1008 19:59:59.506929 4988 scope.go:117] "RemoveContainer" containerID="56698d6d5b5c0b4c4e8cea85cd63e9319099dd99593db5f129194b68561150dd" Oct 08 19:59:59 crc kubenswrapper[4988]: I1008 19:59:59.506990 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-0" Oct 08 19:59:59 crc kubenswrapper[4988]: I1008 19:59:59.537316 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/aodh-0"] Oct 08 19:59:59 crc kubenswrapper[4988]: I1008 19:59:59.542769 4988 scope.go:117] "RemoveContainer" containerID="01e730efa7fc209fb2f43a464a61ae5e2d7bfaa803be6f3d5211a851dc991bd8" Oct 08 19:59:59 crc kubenswrapper[4988]: I1008 19:59:59.571673 4988 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/aodh-0"] Oct 08 19:59:59 crc kubenswrapper[4988]: I1008 19:59:59.588427 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/aodh-0"] Oct 08 19:59:59 crc kubenswrapper[4988]: E1008 19:59:59.588946 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6a1bc0e7-3209-4fa2-9a2a-c8b18121da5a" containerName="aodh-evaluator" Oct 08 19:59:59 crc kubenswrapper[4988]: I1008 19:59:59.588966 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="6a1bc0e7-3209-4fa2-9a2a-c8b18121da5a" containerName="aodh-evaluator" Oct 08 19:59:59 crc kubenswrapper[4988]: E1008 19:59:59.588988 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6a1bc0e7-3209-4fa2-9a2a-c8b18121da5a" containerName="aodh-listener" Oct 08 19:59:59 crc kubenswrapper[4988]: I1008 19:59:59.588997 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="6a1bc0e7-3209-4fa2-9a2a-c8b18121da5a" containerName="aodh-listener" Oct 08 19:59:59 crc kubenswrapper[4988]: E1008 19:59:59.589010 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6a1bc0e7-3209-4fa2-9a2a-c8b18121da5a" containerName="aodh-notifier" Oct 08 19:59:59 crc kubenswrapper[4988]: I1008 19:59:59.589021 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="6a1bc0e7-3209-4fa2-9a2a-c8b18121da5a" containerName="aodh-notifier" Oct 08 19:59:59 crc kubenswrapper[4988]: E1008 19:59:59.589060 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6a1bc0e7-3209-4fa2-9a2a-c8b18121da5a" containerName="aodh-api" Oct 08 19:59:59 crc kubenswrapper[4988]: I1008 19:59:59.589067 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="6a1bc0e7-3209-4fa2-9a2a-c8b18121da5a" containerName="aodh-api" Oct 08 19:59:59 crc kubenswrapper[4988]: I1008 19:59:59.589322 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="6a1bc0e7-3209-4fa2-9a2a-c8b18121da5a" containerName="aodh-evaluator" Oct 08 19:59:59 crc kubenswrapper[4988]: I1008 19:59:59.589341 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="6a1bc0e7-3209-4fa2-9a2a-c8b18121da5a" containerName="aodh-listener" Oct 08 19:59:59 crc kubenswrapper[4988]: I1008 19:59:59.589370 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="6a1bc0e7-3209-4fa2-9a2a-c8b18121da5a" containerName="aodh-notifier" Oct 08 19:59:59 crc kubenswrapper[4988]: I1008 19:59:59.589381 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="6a1bc0e7-3209-4fa2-9a2a-c8b18121da5a" containerName="aodh-api" Oct 08 19:59:59 crc kubenswrapper[4988]: I1008 19:59:59.591809 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-0" Oct 08 19:59:59 crc kubenswrapper[4988]: I1008 19:59:59.595600 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"aodh-scripts" Oct 08 19:59:59 crc kubenswrapper[4988]: I1008 19:59:59.595897 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"aodh-config-data" Oct 08 19:59:59 crc kubenswrapper[4988]: I1008 19:59:59.596109 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"telemetry-autoscaling-dockercfg-s848r" Oct 08 19:59:59 crc kubenswrapper[4988]: I1008 19:59:59.596264 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-aodh-internal-svc" Oct 08 19:59:59 crc kubenswrapper[4988]: I1008 19:59:59.596411 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-aodh-public-svc" Oct 08 19:59:59 crc kubenswrapper[4988]: I1008 19:59:59.599781 4988 scope.go:117] "RemoveContainer" containerID="1a9c83e0bd2e4118cc11a9d647a155cb5f97ba61361dc0ba0e080435aaa48c98" Oct 08 19:59:59 crc kubenswrapper[4988]: I1008 19:59:59.607857 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-0"] Oct 08 19:59:59 crc kubenswrapper[4988]: I1008 19:59:59.685650 4988 scope.go:117] "RemoveContainer" containerID="de0e0960e8cf3d4761ad85747537742d917d43f51156f5145c593fb2bc682af3" Oct 08 19:59:59 crc kubenswrapper[4988]: I1008 19:59:59.697292 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/78aab14f-ea46-4fdb-b1e4-8f8d6c3c54d6-internal-tls-certs\") pod \"aodh-0\" (UID: \"78aab14f-ea46-4fdb-b1e4-8f8d6c3c54d6\") " pod="openstack/aodh-0" Oct 08 19:59:59 crc kubenswrapper[4988]: I1008 19:59:59.697349 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/78aab14f-ea46-4fdb-b1e4-8f8d6c3c54d6-scripts\") pod \"aodh-0\" (UID: \"78aab14f-ea46-4fdb-b1e4-8f8d6c3c54d6\") " pod="openstack/aodh-0" Oct 08 19:59:59 crc kubenswrapper[4988]: I1008 19:59:59.697418 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/78aab14f-ea46-4fdb-b1e4-8f8d6c3c54d6-combined-ca-bundle\") pod \"aodh-0\" (UID: \"78aab14f-ea46-4fdb-b1e4-8f8d6c3c54d6\") " pod="openstack/aodh-0" Oct 08 19:59:59 crc kubenswrapper[4988]: I1008 19:59:59.697450 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gtd52\" (UniqueName: \"kubernetes.io/projected/78aab14f-ea46-4fdb-b1e4-8f8d6c3c54d6-kube-api-access-gtd52\") pod \"aodh-0\" (UID: \"78aab14f-ea46-4fdb-b1e4-8f8d6c3c54d6\") " pod="openstack/aodh-0" Oct 08 19:59:59 crc kubenswrapper[4988]: I1008 19:59:59.697561 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/78aab14f-ea46-4fdb-b1e4-8f8d6c3c54d6-public-tls-certs\") pod \"aodh-0\" (UID: \"78aab14f-ea46-4fdb-b1e4-8f8d6c3c54d6\") " pod="openstack/aodh-0" Oct 08 19:59:59 crc kubenswrapper[4988]: I1008 19:59:59.697596 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/78aab14f-ea46-4fdb-b1e4-8f8d6c3c54d6-config-data\") pod \"aodh-0\" (UID: \"78aab14f-ea46-4fdb-b1e4-8f8d6c3c54d6\") " pod="openstack/aodh-0" Oct 08 19:59:59 crc kubenswrapper[4988]: I1008 19:59:59.799702 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/78aab14f-ea46-4fdb-b1e4-8f8d6c3c54d6-internal-tls-certs\") pod \"aodh-0\" (UID: \"78aab14f-ea46-4fdb-b1e4-8f8d6c3c54d6\") " pod="openstack/aodh-0" Oct 08 19:59:59 crc kubenswrapper[4988]: I1008 19:59:59.800012 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/78aab14f-ea46-4fdb-b1e4-8f8d6c3c54d6-scripts\") pod \"aodh-0\" (UID: \"78aab14f-ea46-4fdb-b1e4-8f8d6c3c54d6\") " pod="openstack/aodh-0" Oct 08 19:59:59 crc kubenswrapper[4988]: I1008 19:59:59.800059 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/78aab14f-ea46-4fdb-b1e4-8f8d6c3c54d6-combined-ca-bundle\") pod \"aodh-0\" (UID: \"78aab14f-ea46-4fdb-b1e4-8f8d6c3c54d6\") " pod="openstack/aodh-0" Oct 08 19:59:59 crc kubenswrapper[4988]: I1008 19:59:59.800087 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gtd52\" (UniqueName: \"kubernetes.io/projected/78aab14f-ea46-4fdb-b1e4-8f8d6c3c54d6-kube-api-access-gtd52\") pod \"aodh-0\" (UID: \"78aab14f-ea46-4fdb-b1e4-8f8d6c3c54d6\") " pod="openstack/aodh-0" Oct 08 19:59:59 crc kubenswrapper[4988]: I1008 19:59:59.800162 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/78aab14f-ea46-4fdb-b1e4-8f8d6c3c54d6-public-tls-certs\") pod \"aodh-0\" (UID: \"78aab14f-ea46-4fdb-b1e4-8f8d6c3c54d6\") " pod="openstack/aodh-0" Oct 08 19:59:59 crc kubenswrapper[4988]: I1008 19:59:59.800187 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/78aab14f-ea46-4fdb-b1e4-8f8d6c3c54d6-config-data\") pod \"aodh-0\" (UID: \"78aab14f-ea46-4fdb-b1e4-8f8d6c3c54d6\") " pod="openstack/aodh-0" Oct 08 19:59:59 crc kubenswrapper[4988]: I1008 19:59:59.803547 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/78aab14f-ea46-4fdb-b1e4-8f8d6c3c54d6-internal-tls-certs\") pod \"aodh-0\" (UID: \"78aab14f-ea46-4fdb-b1e4-8f8d6c3c54d6\") " pod="openstack/aodh-0" Oct 08 19:59:59 crc kubenswrapper[4988]: I1008 19:59:59.803929 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/78aab14f-ea46-4fdb-b1e4-8f8d6c3c54d6-combined-ca-bundle\") pod \"aodh-0\" (UID: \"78aab14f-ea46-4fdb-b1e4-8f8d6c3c54d6\") " pod="openstack/aodh-0" Oct 08 19:59:59 crc kubenswrapper[4988]: I1008 19:59:59.809150 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/78aab14f-ea46-4fdb-b1e4-8f8d6c3c54d6-config-data\") pod \"aodh-0\" (UID: \"78aab14f-ea46-4fdb-b1e4-8f8d6c3c54d6\") " pod="openstack/aodh-0" Oct 08 19:59:59 crc kubenswrapper[4988]: I1008 19:59:59.812221 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/78aab14f-ea46-4fdb-b1e4-8f8d6c3c54d6-scripts\") pod \"aodh-0\" (UID: \"78aab14f-ea46-4fdb-b1e4-8f8d6c3c54d6\") " pod="openstack/aodh-0" Oct 08 19:59:59 crc kubenswrapper[4988]: I1008 19:59:59.814920 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/78aab14f-ea46-4fdb-b1e4-8f8d6c3c54d6-public-tls-certs\") pod \"aodh-0\" (UID: \"78aab14f-ea46-4fdb-b1e4-8f8d6c3c54d6\") " pod="openstack/aodh-0" Oct 08 19:59:59 crc kubenswrapper[4988]: I1008 19:59:59.817745 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gtd52\" (UniqueName: \"kubernetes.io/projected/78aab14f-ea46-4fdb-b1e4-8f8d6c3c54d6-kube-api-access-gtd52\") pod \"aodh-0\" (UID: \"78aab14f-ea46-4fdb-b1e4-8f8d6c3c54d6\") " pod="openstack/aodh-0" Oct 08 19:59:59 crc kubenswrapper[4988]: I1008 19:59:59.960875 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-0" Oct 08 20:00:00 crc kubenswrapper[4988]: I1008 20:00:00.179592 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29332560-bhvtp"] Oct 08 20:00:00 crc kubenswrapper[4988]: I1008 20:00:00.181635 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29332560-bhvtp" Oct 08 20:00:00 crc kubenswrapper[4988]: I1008 20:00:00.184402 4988 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 08 20:00:00 crc kubenswrapper[4988]: I1008 20:00:00.184729 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 08 20:00:00 crc kubenswrapper[4988]: I1008 20:00:00.206904 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29332560-bhvtp"] Oct 08 20:00:00 crc kubenswrapper[4988]: I1008 20:00:00.310897 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/4a463591-730f-4fd6-bd9d-25ed05409362-config-volume\") pod \"collect-profiles-29332560-bhvtp\" (UID: \"4a463591-730f-4fd6-bd9d-25ed05409362\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29332560-bhvtp" Oct 08 20:00:00 crc kubenswrapper[4988]: I1008 20:00:00.310983 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jlqsc\" (UniqueName: \"kubernetes.io/projected/4a463591-730f-4fd6-bd9d-25ed05409362-kube-api-access-jlqsc\") pod \"collect-profiles-29332560-bhvtp\" (UID: \"4a463591-730f-4fd6-bd9d-25ed05409362\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29332560-bhvtp" Oct 08 20:00:00 crc kubenswrapper[4988]: I1008 20:00:00.311147 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/4a463591-730f-4fd6-bd9d-25ed05409362-secret-volume\") pod \"collect-profiles-29332560-bhvtp\" (UID: \"4a463591-730f-4fd6-bd9d-25ed05409362\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29332560-bhvtp" Oct 08 20:00:00 crc kubenswrapper[4988]: I1008 20:00:00.413527 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/4a463591-730f-4fd6-bd9d-25ed05409362-config-volume\") pod \"collect-profiles-29332560-bhvtp\" (UID: \"4a463591-730f-4fd6-bd9d-25ed05409362\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29332560-bhvtp" Oct 08 20:00:00 crc kubenswrapper[4988]: I1008 20:00:00.413598 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jlqsc\" (UniqueName: \"kubernetes.io/projected/4a463591-730f-4fd6-bd9d-25ed05409362-kube-api-access-jlqsc\") pod \"collect-profiles-29332560-bhvtp\" (UID: \"4a463591-730f-4fd6-bd9d-25ed05409362\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29332560-bhvtp" Oct 08 20:00:00 crc kubenswrapper[4988]: I1008 20:00:00.413698 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/4a463591-730f-4fd6-bd9d-25ed05409362-secret-volume\") pod \"collect-profiles-29332560-bhvtp\" (UID: \"4a463591-730f-4fd6-bd9d-25ed05409362\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29332560-bhvtp" Oct 08 20:00:00 crc kubenswrapper[4988]: I1008 20:00:00.415007 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/4a463591-730f-4fd6-bd9d-25ed05409362-config-volume\") pod \"collect-profiles-29332560-bhvtp\" (UID: \"4a463591-730f-4fd6-bd9d-25ed05409362\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29332560-bhvtp" Oct 08 20:00:00 crc kubenswrapper[4988]: I1008 20:00:00.421426 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/4a463591-730f-4fd6-bd9d-25ed05409362-secret-volume\") pod \"collect-profiles-29332560-bhvtp\" (UID: \"4a463591-730f-4fd6-bd9d-25ed05409362\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29332560-bhvtp" Oct 08 20:00:00 crc kubenswrapper[4988]: I1008 20:00:00.436371 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jlqsc\" (UniqueName: \"kubernetes.io/projected/4a463591-730f-4fd6-bd9d-25ed05409362-kube-api-access-jlqsc\") pod \"collect-profiles-29332560-bhvtp\" (UID: \"4a463591-730f-4fd6-bd9d-25ed05409362\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29332560-bhvtp" Oct 08 20:00:00 crc kubenswrapper[4988]: I1008 20:00:00.486589 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-0"] Oct 08 20:00:00 crc kubenswrapper[4988]: W1008 20:00:00.494181 4988 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod78aab14f_ea46_4fdb_b1e4_8f8d6c3c54d6.slice/crio-31fbd5998006cc094befff29d18df64636be259d196cb75afc50be16d1c8369b WatchSource:0}: Error finding container 31fbd5998006cc094befff29d18df64636be259d196cb75afc50be16d1c8369b: Status 404 returned error can't find the container with id 31fbd5998006cc094befff29d18df64636be259d196cb75afc50be16d1c8369b Oct 08 20:00:00 crc kubenswrapper[4988]: I1008 20:00:00.509120 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29332560-bhvtp" Oct 08 20:00:00 crc kubenswrapper[4988]: I1008 20:00:00.517613 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"78aab14f-ea46-4fdb-b1e4-8f8d6c3c54d6","Type":"ContainerStarted","Data":"31fbd5998006cc094befff29d18df64636be259d196cb75afc50be16d1c8369b"} Oct 08 20:00:00 crc kubenswrapper[4988]: I1008 20:00:00.992719 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29332560-bhvtp"] Oct 08 20:00:01 crc kubenswrapper[4988]: I1008 20:00:01.253992 4988 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6a1bc0e7-3209-4fa2-9a2a-c8b18121da5a" path="/var/lib/kubelet/pods/6a1bc0e7-3209-4fa2-9a2a-c8b18121da5a/volumes" Oct 08 20:00:01 crc kubenswrapper[4988]: I1008 20:00:01.532097 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"78aab14f-ea46-4fdb-b1e4-8f8d6c3c54d6","Type":"ContainerStarted","Data":"fac5e5d0b8bec08e01f2fb30cf6658649205f47b5b325c4a383f121e62744669"} Oct 08 20:00:01 crc kubenswrapper[4988]: I1008 20:00:01.533454 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29332560-bhvtp" event={"ID":"4a463591-730f-4fd6-bd9d-25ed05409362","Type":"ContainerStarted","Data":"28ef619b8a9aa7a07c0c27b92232eb6389ba4b462f56f10e8ad03f80ed6e3f20"} Oct 08 20:00:01 crc kubenswrapper[4988]: I1008 20:00:01.533473 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29332560-bhvtp" event={"ID":"4a463591-730f-4fd6-bd9d-25ed05409362","Type":"ContainerStarted","Data":"c36ec6e43e030416b39bb6cc7b9749bdc91cc0424785afdb27f7430e89cfe583"} Oct 08 20:00:01 crc kubenswrapper[4988]: I1008 20:00:01.562660 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29332560-bhvtp" podStartSLOduration=1.562639028 podStartE2EDuration="1.562639028s" podCreationTimestamp="2025-10-08 20:00:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 20:00:01.552564578 +0000 UTC m=+6547.002407358" watchObservedRunningTime="2025-10-08 20:00:01.562639028 +0000 UTC m=+6547.012481818" Oct 08 20:00:02 crc kubenswrapper[4988]: I1008 20:00:02.543196 4988 generic.go:334] "Generic (PLEG): container finished" podID="4a463591-730f-4fd6-bd9d-25ed05409362" containerID="28ef619b8a9aa7a07c0c27b92232eb6389ba4b462f56f10e8ad03f80ed6e3f20" exitCode=0 Oct 08 20:00:02 crc kubenswrapper[4988]: I1008 20:00:02.543316 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29332560-bhvtp" event={"ID":"4a463591-730f-4fd6-bd9d-25ed05409362","Type":"ContainerDied","Data":"28ef619b8a9aa7a07c0c27b92232eb6389ba4b462f56f10e8ad03f80ed6e3f20"} Oct 08 20:00:02 crc kubenswrapper[4988]: I1008 20:00:02.545703 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"78aab14f-ea46-4fdb-b1e4-8f8d6c3c54d6","Type":"ContainerStarted","Data":"594467cb74eb7e5a201adcf178b2f0b49eb9851a15d6f09aee73630be34b8e6a"} Oct 08 20:00:03 crc kubenswrapper[4988]: I1008 20:00:03.564538 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"78aab14f-ea46-4fdb-b1e4-8f8d6c3c54d6","Type":"ContainerStarted","Data":"f92ebd8b7b4277a17a0e36350c73c66c28543e23845c6839027cea101a4b7a53"} Oct 08 20:00:03 crc kubenswrapper[4988]: I1008 20:00:03.963001 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29332560-bhvtp" Oct 08 20:00:04 crc kubenswrapper[4988]: I1008 20:00:04.106840 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/4a463591-730f-4fd6-bd9d-25ed05409362-secret-volume\") pod \"4a463591-730f-4fd6-bd9d-25ed05409362\" (UID: \"4a463591-730f-4fd6-bd9d-25ed05409362\") " Oct 08 20:00:04 crc kubenswrapper[4988]: I1008 20:00:04.108878 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jlqsc\" (UniqueName: \"kubernetes.io/projected/4a463591-730f-4fd6-bd9d-25ed05409362-kube-api-access-jlqsc\") pod \"4a463591-730f-4fd6-bd9d-25ed05409362\" (UID: \"4a463591-730f-4fd6-bd9d-25ed05409362\") " Oct 08 20:00:04 crc kubenswrapper[4988]: I1008 20:00:04.108976 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/4a463591-730f-4fd6-bd9d-25ed05409362-config-volume\") pod \"4a463591-730f-4fd6-bd9d-25ed05409362\" (UID: \"4a463591-730f-4fd6-bd9d-25ed05409362\") " Oct 08 20:00:04 crc kubenswrapper[4988]: I1008 20:00:04.110356 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4a463591-730f-4fd6-bd9d-25ed05409362-config-volume" (OuterVolumeSpecName: "config-volume") pod "4a463591-730f-4fd6-bd9d-25ed05409362" (UID: "4a463591-730f-4fd6-bd9d-25ed05409362"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 20:00:04 crc kubenswrapper[4988]: I1008 20:00:04.115725 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4a463591-730f-4fd6-bd9d-25ed05409362-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "4a463591-730f-4fd6-bd9d-25ed05409362" (UID: "4a463591-730f-4fd6-bd9d-25ed05409362"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 20:00:04 crc kubenswrapper[4988]: I1008 20:00:04.125815 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4a463591-730f-4fd6-bd9d-25ed05409362-kube-api-access-jlqsc" (OuterVolumeSpecName: "kube-api-access-jlqsc") pod "4a463591-730f-4fd6-bd9d-25ed05409362" (UID: "4a463591-730f-4fd6-bd9d-25ed05409362"). InnerVolumeSpecName "kube-api-access-jlqsc". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 20:00:04 crc kubenswrapper[4988]: I1008 20:00:04.211726 4988 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jlqsc\" (UniqueName: \"kubernetes.io/projected/4a463591-730f-4fd6-bd9d-25ed05409362-kube-api-access-jlqsc\") on node \"crc\" DevicePath \"\"" Oct 08 20:00:04 crc kubenswrapper[4988]: I1008 20:00:04.211767 4988 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/4a463591-730f-4fd6-bd9d-25ed05409362-config-volume\") on node \"crc\" DevicePath \"\"" Oct 08 20:00:04 crc kubenswrapper[4988]: I1008 20:00:04.211778 4988 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/4a463591-730f-4fd6-bd9d-25ed05409362-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 08 20:00:04 crc kubenswrapper[4988]: I1008 20:00:04.592094 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"78aab14f-ea46-4fdb-b1e4-8f8d6c3c54d6","Type":"ContainerStarted","Data":"f3d79e5533b3659430eeaefc814a10aa6ca23b7b4e36843c5098dc265501cd17"} Oct 08 20:00:04 crc kubenswrapper[4988]: I1008 20:00:04.600002 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29332560-bhvtp" event={"ID":"4a463591-730f-4fd6-bd9d-25ed05409362","Type":"ContainerDied","Data":"c36ec6e43e030416b39bb6cc7b9749bdc91cc0424785afdb27f7430e89cfe583"} Oct 08 20:00:04 crc kubenswrapper[4988]: I1008 20:00:04.600044 4988 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c36ec6e43e030416b39bb6cc7b9749bdc91cc0424785afdb27f7430e89cfe583" Oct 08 20:00:04 crc kubenswrapper[4988]: I1008 20:00:04.600098 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29332560-bhvtp" Oct 08 20:00:05 crc kubenswrapper[4988]: I1008 20:00:05.007775 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/aodh-0" podStartSLOduration=2.484420597 podStartE2EDuration="6.007746406s" podCreationTimestamp="2025-10-08 19:59:59 +0000 UTC" firstStartedPulling="2025-10-08 20:00:00.496682345 +0000 UTC m=+6545.946525105" lastFinishedPulling="2025-10-08 20:00:04.020008134 +0000 UTC m=+6549.469850914" observedRunningTime="2025-10-08 20:00:04.636091828 +0000 UTC m=+6550.085934598" watchObservedRunningTime="2025-10-08 20:00:05.007746406 +0000 UTC m=+6550.457589216" Oct 08 20:00:05 crc kubenswrapper[4988]: I1008 20:00:05.044084 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29332515-zjccp"] Oct 08 20:00:05 crc kubenswrapper[4988]: I1008 20:00:05.051946 4988 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29332515-zjccp"] Oct 08 20:00:05 crc kubenswrapper[4988]: I1008 20:00:05.254169 4988 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4f49819e-fea9-4f31-bd01-f34ddaa425cf" path="/var/lib/kubelet/pods/4f49819e-fea9-4f31-bd01-f34ddaa425cf/volumes" Oct 08 20:00:06 crc kubenswrapper[4988]: I1008 20:00:06.039693 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-cell-mapping-lmgq7"] Oct 08 20:00:06 crc kubenswrapper[4988]: I1008 20:00:06.052552 4988 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-cell-mapping-lmgq7"] Oct 08 20:00:06 crc kubenswrapper[4988]: I1008 20:00:06.420139 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-687f9c4c69-rnsd7"] Oct 08 20:00:06 crc kubenswrapper[4988]: E1008 20:00:06.420956 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4a463591-730f-4fd6-bd9d-25ed05409362" containerName="collect-profiles" Oct 08 20:00:06 crc kubenswrapper[4988]: I1008 20:00:06.420979 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="4a463591-730f-4fd6-bd9d-25ed05409362" containerName="collect-profiles" Oct 08 20:00:06 crc kubenswrapper[4988]: I1008 20:00:06.421270 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="4a463591-730f-4fd6-bd9d-25ed05409362" containerName="collect-profiles" Oct 08 20:00:06 crc kubenswrapper[4988]: I1008 20:00:06.422921 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-687f9c4c69-rnsd7" Oct 08 20:00:06 crc kubenswrapper[4988]: I1008 20:00:06.425156 4988 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1" Oct 08 20:00:06 crc kubenswrapper[4988]: I1008 20:00:06.430344 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-687f9c4c69-rnsd7"] Oct 08 20:00:06 crc kubenswrapper[4988]: I1008 20:00:06.462651 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ed68d8aa-f1fe-4db4-bf11-033919ca1449-dns-svc\") pod \"dnsmasq-dns-687f9c4c69-rnsd7\" (UID: \"ed68d8aa-f1fe-4db4-bf11-033919ca1449\") " pod="openstack/dnsmasq-dns-687f9c4c69-rnsd7" Oct 08 20:00:06 crc kubenswrapper[4988]: I1008 20:00:06.462753 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ed68d8aa-f1fe-4db4-bf11-033919ca1449-config\") pod \"dnsmasq-dns-687f9c4c69-rnsd7\" (UID: \"ed68d8aa-f1fe-4db4-bf11-033919ca1449\") " pod="openstack/dnsmasq-dns-687f9c4c69-rnsd7" Oct 08 20:00:06 crc kubenswrapper[4988]: I1008 20:00:06.462822 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-cell1\" (UniqueName: \"kubernetes.io/configmap/ed68d8aa-f1fe-4db4-bf11-033919ca1449-openstack-cell1\") pod \"dnsmasq-dns-687f9c4c69-rnsd7\" (UID: \"ed68d8aa-f1fe-4db4-bf11-033919ca1449\") " pod="openstack/dnsmasq-dns-687f9c4c69-rnsd7" Oct 08 20:00:06 crc kubenswrapper[4988]: I1008 20:00:06.462912 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ed68d8aa-f1fe-4db4-bf11-033919ca1449-ovsdbserver-nb\") pod \"dnsmasq-dns-687f9c4c69-rnsd7\" (UID: \"ed68d8aa-f1fe-4db4-bf11-033919ca1449\") " pod="openstack/dnsmasq-dns-687f9c4c69-rnsd7" Oct 08 20:00:06 crc kubenswrapper[4988]: I1008 20:00:06.463023 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sb8x6\" (UniqueName: \"kubernetes.io/projected/ed68d8aa-f1fe-4db4-bf11-033919ca1449-kube-api-access-sb8x6\") pod \"dnsmasq-dns-687f9c4c69-rnsd7\" (UID: \"ed68d8aa-f1fe-4db4-bf11-033919ca1449\") " pod="openstack/dnsmasq-dns-687f9c4c69-rnsd7" Oct 08 20:00:06 crc kubenswrapper[4988]: I1008 20:00:06.463084 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ed68d8aa-f1fe-4db4-bf11-033919ca1449-ovsdbserver-sb\") pod \"dnsmasq-dns-687f9c4c69-rnsd7\" (UID: \"ed68d8aa-f1fe-4db4-bf11-033919ca1449\") " pod="openstack/dnsmasq-dns-687f9c4c69-rnsd7" Oct 08 20:00:06 crc kubenswrapper[4988]: I1008 20:00:06.564912 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sb8x6\" (UniqueName: \"kubernetes.io/projected/ed68d8aa-f1fe-4db4-bf11-033919ca1449-kube-api-access-sb8x6\") pod \"dnsmasq-dns-687f9c4c69-rnsd7\" (UID: \"ed68d8aa-f1fe-4db4-bf11-033919ca1449\") " pod="openstack/dnsmasq-dns-687f9c4c69-rnsd7" Oct 08 20:00:06 crc kubenswrapper[4988]: I1008 20:00:06.564964 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ed68d8aa-f1fe-4db4-bf11-033919ca1449-ovsdbserver-sb\") pod \"dnsmasq-dns-687f9c4c69-rnsd7\" (UID: \"ed68d8aa-f1fe-4db4-bf11-033919ca1449\") " pod="openstack/dnsmasq-dns-687f9c4c69-rnsd7" Oct 08 20:00:06 crc kubenswrapper[4988]: I1008 20:00:06.565041 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ed68d8aa-f1fe-4db4-bf11-033919ca1449-dns-svc\") pod \"dnsmasq-dns-687f9c4c69-rnsd7\" (UID: \"ed68d8aa-f1fe-4db4-bf11-033919ca1449\") " pod="openstack/dnsmasq-dns-687f9c4c69-rnsd7" Oct 08 20:00:06 crc kubenswrapper[4988]: I1008 20:00:06.565076 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ed68d8aa-f1fe-4db4-bf11-033919ca1449-config\") pod \"dnsmasq-dns-687f9c4c69-rnsd7\" (UID: \"ed68d8aa-f1fe-4db4-bf11-033919ca1449\") " pod="openstack/dnsmasq-dns-687f9c4c69-rnsd7" Oct 08 20:00:06 crc kubenswrapper[4988]: I1008 20:00:06.565107 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-cell1\" (UniqueName: \"kubernetes.io/configmap/ed68d8aa-f1fe-4db4-bf11-033919ca1449-openstack-cell1\") pod \"dnsmasq-dns-687f9c4c69-rnsd7\" (UID: \"ed68d8aa-f1fe-4db4-bf11-033919ca1449\") " pod="openstack/dnsmasq-dns-687f9c4c69-rnsd7" Oct 08 20:00:06 crc kubenswrapper[4988]: I1008 20:00:06.565148 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ed68d8aa-f1fe-4db4-bf11-033919ca1449-ovsdbserver-nb\") pod \"dnsmasq-dns-687f9c4c69-rnsd7\" (UID: \"ed68d8aa-f1fe-4db4-bf11-033919ca1449\") " pod="openstack/dnsmasq-dns-687f9c4c69-rnsd7" Oct 08 20:00:06 crc kubenswrapper[4988]: I1008 20:00:06.566237 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ed68d8aa-f1fe-4db4-bf11-033919ca1449-ovsdbserver-nb\") pod \"dnsmasq-dns-687f9c4c69-rnsd7\" (UID: \"ed68d8aa-f1fe-4db4-bf11-033919ca1449\") " pod="openstack/dnsmasq-dns-687f9c4c69-rnsd7" Oct 08 20:00:06 crc kubenswrapper[4988]: I1008 20:00:06.566533 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ed68d8aa-f1fe-4db4-bf11-033919ca1449-config\") pod \"dnsmasq-dns-687f9c4c69-rnsd7\" (UID: \"ed68d8aa-f1fe-4db4-bf11-033919ca1449\") " pod="openstack/dnsmasq-dns-687f9c4c69-rnsd7" Oct 08 20:00:06 crc kubenswrapper[4988]: I1008 20:00:06.566628 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ed68d8aa-f1fe-4db4-bf11-033919ca1449-dns-svc\") pod \"dnsmasq-dns-687f9c4c69-rnsd7\" (UID: \"ed68d8aa-f1fe-4db4-bf11-033919ca1449\") " pod="openstack/dnsmasq-dns-687f9c4c69-rnsd7" Oct 08 20:00:06 crc kubenswrapper[4988]: I1008 20:00:06.566996 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-cell1\" (UniqueName: \"kubernetes.io/configmap/ed68d8aa-f1fe-4db4-bf11-033919ca1449-openstack-cell1\") pod \"dnsmasq-dns-687f9c4c69-rnsd7\" (UID: \"ed68d8aa-f1fe-4db4-bf11-033919ca1449\") " pod="openstack/dnsmasq-dns-687f9c4c69-rnsd7" Oct 08 20:00:06 crc kubenswrapper[4988]: I1008 20:00:06.567519 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ed68d8aa-f1fe-4db4-bf11-033919ca1449-ovsdbserver-sb\") pod \"dnsmasq-dns-687f9c4c69-rnsd7\" (UID: \"ed68d8aa-f1fe-4db4-bf11-033919ca1449\") " pod="openstack/dnsmasq-dns-687f9c4c69-rnsd7" Oct 08 20:00:06 crc kubenswrapper[4988]: I1008 20:00:06.588230 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Oct 08 20:00:06 crc kubenswrapper[4988]: I1008 20:00:06.589903 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sb8x6\" (UniqueName: \"kubernetes.io/projected/ed68d8aa-f1fe-4db4-bf11-033919ca1449-kube-api-access-sb8x6\") pod \"dnsmasq-dns-687f9c4c69-rnsd7\" (UID: \"ed68d8aa-f1fe-4db4-bf11-033919ca1449\") " pod="openstack/dnsmasq-dns-687f9c4c69-rnsd7" Oct 08 20:00:06 crc kubenswrapper[4988]: I1008 20:00:06.741467 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-687f9c4c69-rnsd7" Oct 08 20:00:07 crc kubenswrapper[4988]: I1008 20:00:07.249866 4988 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fa6bc882-bc50-4842-91a2-626c0e0329f4" path="/var/lib/kubelet/pods/fa6bc882-bc50-4842-91a2-626c0e0329f4/volumes" Oct 08 20:00:07 crc kubenswrapper[4988]: I1008 20:00:07.273032 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-687f9c4c69-rnsd7"] Oct 08 20:00:07 crc kubenswrapper[4988]: I1008 20:00:07.635491 4988 generic.go:334] "Generic (PLEG): container finished" podID="ed68d8aa-f1fe-4db4-bf11-033919ca1449" containerID="53df725348367f4fbaeb73973f8c30353e8a4091e5b8c61ed716a5144718acdb" exitCode=0 Oct 08 20:00:07 crc kubenswrapper[4988]: I1008 20:00:07.635703 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-687f9c4c69-rnsd7" event={"ID":"ed68d8aa-f1fe-4db4-bf11-033919ca1449","Type":"ContainerDied","Data":"53df725348367f4fbaeb73973f8c30353e8a4091e5b8c61ed716a5144718acdb"} Oct 08 20:00:07 crc kubenswrapper[4988]: I1008 20:00:07.635879 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-687f9c4c69-rnsd7" event={"ID":"ed68d8aa-f1fe-4db4-bf11-033919ca1449","Type":"ContainerStarted","Data":"d429e156a98089ee6af2466914c0f989abcad1d09ea0fb5ea6ff7a256b43226d"} Oct 08 20:00:08 crc kubenswrapper[4988]: I1008 20:00:08.665589 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-687f9c4c69-rnsd7" event={"ID":"ed68d8aa-f1fe-4db4-bf11-033919ca1449","Type":"ContainerStarted","Data":"4076efaf6dd043f83ecaca99250372d807f2155e36c84cce51f39baaf5b4ef9f"} Oct 08 20:00:08 crc kubenswrapper[4988]: I1008 20:00:08.666055 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-687f9c4c69-rnsd7" Oct 08 20:00:08 crc kubenswrapper[4988]: I1008 20:00:08.687205 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-687f9c4c69-rnsd7" podStartSLOduration=2.687185307 podStartE2EDuration="2.687185307s" podCreationTimestamp="2025-10-08 20:00:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 20:00:08.682371224 +0000 UTC m=+6554.132213984" watchObservedRunningTime="2025-10-08 20:00:08.687185307 +0000 UTC m=+6554.137028077" Oct 08 20:00:12 crc kubenswrapper[4988]: I1008 20:00:12.585797 4988 scope.go:117] "RemoveContainer" containerID="17bffefafaad07ac2d268dde4cf519ffef397e22b0bb24d9fc9fd17189af4020" Oct 08 20:00:12 crc kubenswrapper[4988]: I1008 20:00:12.623993 4988 scope.go:117] "RemoveContainer" containerID="4ff1fec466874c2f74cd25cc619db431c83691c072385c72f632aceaa00e88ec" Oct 08 20:00:12 crc kubenswrapper[4988]: I1008 20:00:12.695876 4988 scope.go:117] "RemoveContainer" containerID="9b8bbc55b526532b8819a7ac1a54a4c1aa8fc31d5cca19ae4663664bf7137162" Oct 08 20:00:12 crc kubenswrapper[4988]: I1008 20:00:12.787101 4988 scope.go:117] "RemoveContainer" containerID="30d556b219e1e75ea900a02038621eaa6233a071efd905b622e3fc38cb801be5" Oct 08 20:00:12 crc kubenswrapper[4988]: I1008 20:00:12.830359 4988 scope.go:117] "RemoveContainer" containerID="b4459fd38af0cc29ba01f889cf7c7cfa630d4be8878f893f3431416f7fa443d3" Oct 08 20:00:12 crc kubenswrapper[4988]: I1008 20:00:12.889471 4988 scope.go:117] "RemoveContainer" containerID="521bb2202c7abb16ee78ce44796f153e3a190098a143f6b2bcd30634e128b462" Oct 08 20:00:12 crc kubenswrapper[4988]: I1008 20:00:12.955142 4988 scope.go:117] "RemoveContainer" containerID="2a2987792e85f659046e9d2bfb64d4bb2f1b90ded6eb9bc60cc5b8fe63002c6f" Oct 08 20:00:13 crc kubenswrapper[4988]: I1008 20:00:13.006318 4988 scope.go:117] "RemoveContainer" containerID="a156581bd3e8962fa43eb67139042f378b710e919ab897fb5ab73789cc40f5ad" Oct 08 20:00:13 crc kubenswrapper[4988]: I1008 20:00:13.237860 4988 scope.go:117] "RemoveContainer" containerID="439af713ff941f973b8758bd6af3c097808f76ff520b96ff10bd3268811393f7" Oct 08 20:00:13 crc kubenswrapper[4988]: E1008 20:00:13.238363 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nm54f_openshift-machine-config-operator(f328a20b-5283-407e-8a93-2ab2da599f18)\"" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" Oct 08 20:00:16 crc kubenswrapper[4988]: I1008 20:00:16.743709 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-687f9c4c69-rnsd7" Oct 08 20:00:16 crc kubenswrapper[4988]: I1008 20:00:16.840346 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-65d478dd7c-lc957"] Oct 08 20:00:16 crc kubenswrapper[4988]: I1008 20:00:16.843784 4988 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-65d478dd7c-lc957" podUID="5c03ce9a-8cfe-4830-a131-e5b6353bef7c" containerName="dnsmasq-dns" containerID="cri-o://a2613649e8a5b46817e3a16690ad3f809127cf8b260464db7c2c75c1cf186c5a" gracePeriod=10 Oct 08 20:00:17 crc kubenswrapper[4988]: I1008 20:00:17.066234 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-76dfc7c65-crgzj"] Oct 08 20:00:17 crc kubenswrapper[4988]: I1008 20:00:17.068527 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-76dfc7c65-crgzj" Oct 08 20:00:17 crc kubenswrapper[4988]: I1008 20:00:17.095510 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-76dfc7c65-crgzj"] Oct 08 20:00:17 crc kubenswrapper[4988]: I1008 20:00:17.241377 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pbtsb\" (UniqueName: \"kubernetes.io/projected/9e2ebad6-add9-4567-8f39-8e51f614e50f-kube-api-access-pbtsb\") pod \"dnsmasq-dns-76dfc7c65-crgzj\" (UID: \"9e2ebad6-add9-4567-8f39-8e51f614e50f\") " pod="openstack/dnsmasq-dns-76dfc7c65-crgzj" Oct 08 20:00:17 crc kubenswrapper[4988]: I1008 20:00:17.241536 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9e2ebad6-add9-4567-8f39-8e51f614e50f-config\") pod \"dnsmasq-dns-76dfc7c65-crgzj\" (UID: \"9e2ebad6-add9-4567-8f39-8e51f614e50f\") " pod="openstack/dnsmasq-dns-76dfc7c65-crgzj" Oct 08 20:00:17 crc kubenswrapper[4988]: I1008 20:00:17.241567 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9e2ebad6-add9-4567-8f39-8e51f614e50f-ovsdbserver-sb\") pod \"dnsmasq-dns-76dfc7c65-crgzj\" (UID: \"9e2ebad6-add9-4567-8f39-8e51f614e50f\") " pod="openstack/dnsmasq-dns-76dfc7c65-crgzj" Oct 08 20:00:17 crc kubenswrapper[4988]: I1008 20:00:17.241593 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9e2ebad6-add9-4567-8f39-8e51f614e50f-dns-svc\") pod \"dnsmasq-dns-76dfc7c65-crgzj\" (UID: \"9e2ebad6-add9-4567-8f39-8e51f614e50f\") " pod="openstack/dnsmasq-dns-76dfc7c65-crgzj" Oct 08 20:00:17 crc kubenswrapper[4988]: I1008 20:00:17.241810 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9e2ebad6-add9-4567-8f39-8e51f614e50f-ovsdbserver-nb\") pod \"dnsmasq-dns-76dfc7c65-crgzj\" (UID: \"9e2ebad6-add9-4567-8f39-8e51f614e50f\") " pod="openstack/dnsmasq-dns-76dfc7c65-crgzj" Oct 08 20:00:17 crc kubenswrapper[4988]: I1008 20:00:17.241902 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-cell1\" (UniqueName: \"kubernetes.io/configmap/9e2ebad6-add9-4567-8f39-8e51f614e50f-openstack-cell1\") pod \"dnsmasq-dns-76dfc7c65-crgzj\" (UID: \"9e2ebad6-add9-4567-8f39-8e51f614e50f\") " pod="openstack/dnsmasq-dns-76dfc7c65-crgzj" Oct 08 20:00:17 crc kubenswrapper[4988]: I1008 20:00:17.343865 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9e2ebad6-add9-4567-8f39-8e51f614e50f-ovsdbserver-nb\") pod \"dnsmasq-dns-76dfc7c65-crgzj\" (UID: \"9e2ebad6-add9-4567-8f39-8e51f614e50f\") " pod="openstack/dnsmasq-dns-76dfc7c65-crgzj" Oct 08 20:00:17 crc kubenswrapper[4988]: I1008 20:00:17.343918 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-cell1\" (UniqueName: \"kubernetes.io/configmap/9e2ebad6-add9-4567-8f39-8e51f614e50f-openstack-cell1\") pod \"dnsmasq-dns-76dfc7c65-crgzj\" (UID: \"9e2ebad6-add9-4567-8f39-8e51f614e50f\") " pod="openstack/dnsmasq-dns-76dfc7c65-crgzj" Oct 08 20:00:17 crc kubenswrapper[4988]: I1008 20:00:17.344011 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pbtsb\" (UniqueName: \"kubernetes.io/projected/9e2ebad6-add9-4567-8f39-8e51f614e50f-kube-api-access-pbtsb\") pod \"dnsmasq-dns-76dfc7c65-crgzj\" (UID: \"9e2ebad6-add9-4567-8f39-8e51f614e50f\") " pod="openstack/dnsmasq-dns-76dfc7c65-crgzj" Oct 08 20:00:17 crc kubenswrapper[4988]: I1008 20:00:17.344070 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9e2ebad6-add9-4567-8f39-8e51f614e50f-config\") pod \"dnsmasq-dns-76dfc7c65-crgzj\" (UID: \"9e2ebad6-add9-4567-8f39-8e51f614e50f\") " pod="openstack/dnsmasq-dns-76dfc7c65-crgzj" Oct 08 20:00:17 crc kubenswrapper[4988]: I1008 20:00:17.344092 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9e2ebad6-add9-4567-8f39-8e51f614e50f-ovsdbserver-sb\") pod \"dnsmasq-dns-76dfc7c65-crgzj\" (UID: \"9e2ebad6-add9-4567-8f39-8e51f614e50f\") " pod="openstack/dnsmasq-dns-76dfc7c65-crgzj" Oct 08 20:00:17 crc kubenswrapper[4988]: I1008 20:00:17.344117 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9e2ebad6-add9-4567-8f39-8e51f614e50f-dns-svc\") pod \"dnsmasq-dns-76dfc7c65-crgzj\" (UID: \"9e2ebad6-add9-4567-8f39-8e51f614e50f\") " pod="openstack/dnsmasq-dns-76dfc7c65-crgzj" Oct 08 20:00:17 crc kubenswrapper[4988]: I1008 20:00:17.345504 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-cell1\" (UniqueName: \"kubernetes.io/configmap/9e2ebad6-add9-4567-8f39-8e51f614e50f-openstack-cell1\") pod \"dnsmasq-dns-76dfc7c65-crgzj\" (UID: \"9e2ebad6-add9-4567-8f39-8e51f614e50f\") " pod="openstack/dnsmasq-dns-76dfc7c65-crgzj" Oct 08 20:00:17 crc kubenswrapper[4988]: I1008 20:00:17.346670 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9e2ebad6-add9-4567-8f39-8e51f614e50f-ovsdbserver-sb\") pod \"dnsmasq-dns-76dfc7c65-crgzj\" (UID: \"9e2ebad6-add9-4567-8f39-8e51f614e50f\") " pod="openstack/dnsmasq-dns-76dfc7c65-crgzj" Oct 08 20:00:17 crc kubenswrapper[4988]: I1008 20:00:17.346859 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9e2ebad6-add9-4567-8f39-8e51f614e50f-ovsdbserver-nb\") pod \"dnsmasq-dns-76dfc7c65-crgzj\" (UID: \"9e2ebad6-add9-4567-8f39-8e51f614e50f\") " pod="openstack/dnsmasq-dns-76dfc7c65-crgzj" Oct 08 20:00:17 crc kubenswrapper[4988]: I1008 20:00:17.347106 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9e2ebad6-add9-4567-8f39-8e51f614e50f-dns-svc\") pod \"dnsmasq-dns-76dfc7c65-crgzj\" (UID: \"9e2ebad6-add9-4567-8f39-8e51f614e50f\") " pod="openstack/dnsmasq-dns-76dfc7c65-crgzj" Oct 08 20:00:17 crc kubenswrapper[4988]: I1008 20:00:17.347137 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9e2ebad6-add9-4567-8f39-8e51f614e50f-config\") pod \"dnsmasq-dns-76dfc7c65-crgzj\" (UID: \"9e2ebad6-add9-4567-8f39-8e51f614e50f\") " pod="openstack/dnsmasq-dns-76dfc7c65-crgzj" Oct 08 20:00:17 crc kubenswrapper[4988]: I1008 20:00:17.362514 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pbtsb\" (UniqueName: \"kubernetes.io/projected/9e2ebad6-add9-4567-8f39-8e51f614e50f-kube-api-access-pbtsb\") pod \"dnsmasq-dns-76dfc7c65-crgzj\" (UID: \"9e2ebad6-add9-4567-8f39-8e51f614e50f\") " pod="openstack/dnsmasq-dns-76dfc7c65-crgzj" Oct 08 20:00:17 crc kubenswrapper[4988]: I1008 20:00:17.396504 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-76dfc7c65-crgzj" Oct 08 20:00:17 crc kubenswrapper[4988]: I1008 20:00:17.577090 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-65d478dd7c-lc957" Oct 08 20:00:17 crc kubenswrapper[4988]: I1008 20:00:17.650289 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5c03ce9a-8cfe-4830-a131-e5b6353bef7c-dns-svc\") pod \"5c03ce9a-8cfe-4830-a131-e5b6353bef7c\" (UID: \"5c03ce9a-8cfe-4830-a131-e5b6353bef7c\") " Oct 08 20:00:17 crc kubenswrapper[4988]: I1008 20:00:17.650341 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5c03ce9a-8cfe-4830-a131-e5b6353bef7c-config\") pod \"5c03ce9a-8cfe-4830-a131-e5b6353bef7c\" (UID: \"5c03ce9a-8cfe-4830-a131-e5b6353bef7c\") " Oct 08 20:00:17 crc kubenswrapper[4988]: I1008 20:00:17.650374 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5c03ce9a-8cfe-4830-a131-e5b6353bef7c-ovsdbserver-nb\") pod \"5c03ce9a-8cfe-4830-a131-e5b6353bef7c\" (UID: \"5c03ce9a-8cfe-4830-a131-e5b6353bef7c\") " Oct 08 20:00:17 crc kubenswrapper[4988]: I1008 20:00:17.650523 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lcfxp\" (UniqueName: \"kubernetes.io/projected/5c03ce9a-8cfe-4830-a131-e5b6353bef7c-kube-api-access-lcfxp\") pod \"5c03ce9a-8cfe-4830-a131-e5b6353bef7c\" (UID: \"5c03ce9a-8cfe-4830-a131-e5b6353bef7c\") " Oct 08 20:00:17 crc kubenswrapper[4988]: I1008 20:00:17.650567 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5c03ce9a-8cfe-4830-a131-e5b6353bef7c-ovsdbserver-sb\") pod \"5c03ce9a-8cfe-4830-a131-e5b6353bef7c\" (UID: \"5c03ce9a-8cfe-4830-a131-e5b6353bef7c\") " Oct 08 20:00:17 crc kubenswrapper[4988]: I1008 20:00:17.660917 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5c03ce9a-8cfe-4830-a131-e5b6353bef7c-kube-api-access-lcfxp" (OuterVolumeSpecName: "kube-api-access-lcfxp") pod "5c03ce9a-8cfe-4830-a131-e5b6353bef7c" (UID: "5c03ce9a-8cfe-4830-a131-e5b6353bef7c"). InnerVolumeSpecName "kube-api-access-lcfxp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 20:00:17 crc kubenswrapper[4988]: I1008 20:00:17.721154 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5c03ce9a-8cfe-4830-a131-e5b6353bef7c-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "5c03ce9a-8cfe-4830-a131-e5b6353bef7c" (UID: "5c03ce9a-8cfe-4830-a131-e5b6353bef7c"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 20:00:17 crc kubenswrapper[4988]: I1008 20:00:17.727579 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5c03ce9a-8cfe-4830-a131-e5b6353bef7c-config" (OuterVolumeSpecName: "config") pod "5c03ce9a-8cfe-4830-a131-e5b6353bef7c" (UID: "5c03ce9a-8cfe-4830-a131-e5b6353bef7c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 20:00:17 crc kubenswrapper[4988]: I1008 20:00:17.738435 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5c03ce9a-8cfe-4830-a131-e5b6353bef7c-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "5c03ce9a-8cfe-4830-a131-e5b6353bef7c" (UID: "5c03ce9a-8cfe-4830-a131-e5b6353bef7c"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 20:00:17 crc kubenswrapper[4988]: I1008 20:00:17.748707 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5c03ce9a-8cfe-4830-a131-e5b6353bef7c-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "5c03ce9a-8cfe-4830-a131-e5b6353bef7c" (UID: "5c03ce9a-8cfe-4830-a131-e5b6353bef7c"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 20:00:17 crc kubenswrapper[4988]: I1008 20:00:17.753178 4988 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5c03ce9a-8cfe-4830-a131-e5b6353bef7c-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 08 20:00:17 crc kubenswrapper[4988]: I1008 20:00:17.753212 4988 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5c03ce9a-8cfe-4830-a131-e5b6353bef7c-config\") on node \"crc\" DevicePath \"\"" Oct 08 20:00:17 crc kubenswrapper[4988]: I1008 20:00:17.753222 4988 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5c03ce9a-8cfe-4830-a131-e5b6353bef7c-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 08 20:00:17 crc kubenswrapper[4988]: I1008 20:00:17.753235 4988 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lcfxp\" (UniqueName: \"kubernetes.io/projected/5c03ce9a-8cfe-4830-a131-e5b6353bef7c-kube-api-access-lcfxp\") on node \"crc\" DevicePath \"\"" Oct 08 20:00:17 crc kubenswrapper[4988]: I1008 20:00:17.753245 4988 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5c03ce9a-8cfe-4830-a131-e5b6353bef7c-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 08 20:00:17 crc kubenswrapper[4988]: I1008 20:00:17.791620 4988 generic.go:334] "Generic (PLEG): container finished" podID="5c03ce9a-8cfe-4830-a131-e5b6353bef7c" containerID="a2613649e8a5b46817e3a16690ad3f809127cf8b260464db7c2c75c1cf186c5a" exitCode=0 Oct 08 20:00:17 crc kubenswrapper[4988]: I1008 20:00:17.791687 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-65d478dd7c-lc957" Oct 08 20:00:17 crc kubenswrapper[4988]: I1008 20:00:17.791692 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-65d478dd7c-lc957" event={"ID":"5c03ce9a-8cfe-4830-a131-e5b6353bef7c","Type":"ContainerDied","Data":"a2613649e8a5b46817e3a16690ad3f809127cf8b260464db7c2c75c1cf186c5a"} Oct 08 20:00:17 crc kubenswrapper[4988]: I1008 20:00:17.792152 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-65d478dd7c-lc957" event={"ID":"5c03ce9a-8cfe-4830-a131-e5b6353bef7c","Type":"ContainerDied","Data":"8e2da6caaa60825f6ac114028a3bde9ad140aad58d71713ccf7cff4d35497bdf"} Oct 08 20:00:17 crc kubenswrapper[4988]: I1008 20:00:17.792176 4988 scope.go:117] "RemoveContainer" containerID="a2613649e8a5b46817e3a16690ad3f809127cf8b260464db7c2c75c1cf186c5a" Oct 08 20:00:17 crc kubenswrapper[4988]: I1008 20:00:17.824867 4988 scope.go:117] "RemoveContainer" containerID="de44ce794ae6a6705f6329d861517cd9df6c73b5e3feec35b602075e9037d78e" Oct 08 20:00:17 crc kubenswrapper[4988]: I1008 20:00:17.846744 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-65d478dd7c-lc957"] Oct 08 20:00:17 crc kubenswrapper[4988]: I1008 20:00:17.864523 4988 scope.go:117] "RemoveContainer" containerID="a2613649e8a5b46817e3a16690ad3f809127cf8b260464db7c2c75c1cf186c5a" Oct 08 20:00:17 crc kubenswrapper[4988]: E1008 20:00:17.864912 4988 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a2613649e8a5b46817e3a16690ad3f809127cf8b260464db7c2c75c1cf186c5a\": container with ID starting with a2613649e8a5b46817e3a16690ad3f809127cf8b260464db7c2c75c1cf186c5a not found: ID does not exist" containerID="a2613649e8a5b46817e3a16690ad3f809127cf8b260464db7c2c75c1cf186c5a" Oct 08 20:00:17 crc kubenswrapper[4988]: I1008 20:00:17.864972 4988 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a2613649e8a5b46817e3a16690ad3f809127cf8b260464db7c2c75c1cf186c5a"} err="failed to get container status \"a2613649e8a5b46817e3a16690ad3f809127cf8b260464db7c2c75c1cf186c5a\": rpc error: code = NotFound desc = could not find container \"a2613649e8a5b46817e3a16690ad3f809127cf8b260464db7c2c75c1cf186c5a\": container with ID starting with a2613649e8a5b46817e3a16690ad3f809127cf8b260464db7c2c75c1cf186c5a not found: ID does not exist" Oct 08 20:00:17 crc kubenswrapper[4988]: I1008 20:00:17.864999 4988 scope.go:117] "RemoveContainer" containerID="de44ce794ae6a6705f6329d861517cd9df6c73b5e3feec35b602075e9037d78e" Oct 08 20:00:17 crc kubenswrapper[4988]: E1008 20:00:17.867022 4988 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"de44ce794ae6a6705f6329d861517cd9df6c73b5e3feec35b602075e9037d78e\": container with ID starting with de44ce794ae6a6705f6329d861517cd9df6c73b5e3feec35b602075e9037d78e not found: ID does not exist" containerID="de44ce794ae6a6705f6329d861517cd9df6c73b5e3feec35b602075e9037d78e" Oct 08 20:00:17 crc kubenswrapper[4988]: I1008 20:00:17.867055 4988 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"de44ce794ae6a6705f6329d861517cd9df6c73b5e3feec35b602075e9037d78e"} err="failed to get container status \"de44ce794ae6a6705f6329d861517cd9df6c73b5e3feec35b602075e9037d78e\": rpc error: code = NotFound desc = could not find container \"de44ce794ae6a6705f6329d861517cd9df6c73b5e3feec35b602075e9037d78e\": container with ID starting with de44ce794ae6a6705f6329d861517cd9df6c73b5e3feec35b602075e9037d78e not found: ID does not exist" Oct 08 20:00:17 crc kubenswrapper[4988]: I1008 20:00:17.869722 4988 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-65d478dd7c-lc957"] Oct 08 20:00:17 crc kubenswrapper[4988]: I1008 20:00:17.941774 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-76dfc7c65-crgzj"] Oct 08 20:00:18 crc kubenswrapper[4988]: I1008 20:00:18.811250 4988 generic.go:334] "Generic (PLEG): container finished" podID="9e2ebad6-add9-4567-8f39-8e51f614e50f" containerID="475045834e16b81e656694559f3e4c79dcc5fa78361c926263cb80e632b0c494" exitCode=0 Oct 08 20:00:18 crc kubenswrapper[4988]: I1008 20:00:18.811418 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-76dfc7c65-crgzj" event={"ID":"9e2ebad6-add9-4567-8f39-8e51f614e50f","Type":"ContainerDied","Data":"475045834e16b81e656694559f3e4c79dcc5fa78361c926263cb80e632b0c494"} Oct 08 20:00:18 crc kubenswrapper[4988]: I1008 20:00:18.812163 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-76dfc7c65-crgzj" event={"ID":"9e2ebad6-add9-4567-8f39-8e51f614e50f","Type":"ContainerStarted","Data":"1673e208a77b0600837079043d07728bc66110378d9dcdd30dd2a09cce2a9e69"} Oct 08 20:00:19 crc kubenswrapper[4988]: I1008 20:00:19.267318 4988 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5c03ce9a-8cfe-4830-a131-e5b6353bef7c" path="/var/lib/kubelet/pods/5c03ce9a-8cfe-4830-a131-e5b6353bef7c/volumes" Oct 08 20:00:19 crc kubenswrapper[4988]: I1008 20:00:19.834544 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-76dfc7c65-crgzj" event={"ID":"9e2ebad6-add9-4567-8f39-8e51f614e50f","Type":"ContainerStarted","Data":"7e0ddf846b8b676e47d5f2b2e0614b3d7d0989dc6b8ecf033a6080831eef7f57"} Oct 08 20:00:19 crc kubenswrapper[4988]: I1008 20:00:19.834943 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-76dfc7c65-crgzj" Oct 08 20:00:19 crc kubenswrapper[4988]: I1008 20:00:19.872933 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-76dfc7c65-crgzj" podStartSLOduration=2.872913648 podStartE2EDuration="2.872913648s" podCreationTimestamp="2025-10-08 20:00:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 20:00:19.862117874 +0000 UTC m=+6565.311960654" watchObservedRunningTime="2025-10-08 20:00:19.872913648 +0000 UTC m=+6565.322756418" Oct 08 20:00:27 crc kubenswrapper[4988]: I1008 20:00:27.398618 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-76dfc7c65-crgzj" Oct 08 20:00:27 crc kubenswrapper[4988]: I1008 20:00:27.500618 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-687f9c4c69-rnsd7"] Oct 08 20:00:27 crc kubenswrapper[4988]: I1008 20:00:27.500888 4988 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-687f9c4c69-rnsd7" podUID="ed68d8aa-f1fe-4db4-bf11-033919ca1449" containerName="dnsmasq-dns" containerID="cri-o://4076efaf6dd043f83ecaca99250372d807f2155e36c84cce51f39baaf5b4ef9f" gracePeriod=10 Oct 08 20:00:27 crc kubenswrapper[4988]: I1008 20:00:27.931047 4988 generic.go:334] "Generic (PLEG): container finished" podID="ed68d8aa-f1fe-4db4-bf11-033919ca1449" containerID="4076efaf6dd043f83ecaca99250372d807f2155e36c84cce51f39baaf5b4ef9f" exitCode=0 Oct 08 20:00:27 crc kubenswrapper[4988]: I1008 20:00:27.931218 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-687f9c4c69-rnsd7" event={"ID":"ed68d8aa-f1fe-4db4-bf11-033919ca1449","Type":"ContainerDied","Data":"4076efaf6dd043f83ecaca99250372d807f2155e36c84cce51f39baaf5b4ef9f"} Oct 08 20:00:28 crc kubenswrapper[4988]: I1008 20:00:28.058076 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-687f9c4c69-rnsd7" Oct 08 20:00:28 crc kubenswrapper[4988]: I1008 20:00:28.235220 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ed68d8aa-f1fe-4db4-bf11-033919ca1449-config\") pod \"ed68d8aa-f1fe-4db4-bf11-033919ca1449\" (UID: \"ed68d8aa-f1fe-4db4-bf11-033919ca1449\") " Oct 08 20:00:28 crc kubenswrapper[4988]: I1008 20:00:28.235638 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-cell1\" (UniqueName: \"kubernetes.io/configmap/ed68d8aa-f1fe-4db4-bf11-033919ca1449-openstack-cell1\") pod \"ed68d8aa-f1fe-4db4-bf11-033919ca1449\" (UID: \"ed68d8aa-f1fe-4db4-bf11-033919ca1449\") " Oct 08 20:00:28 crc kubenswrapper[4988]: I1008 20:00:28.235685 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sb8x6\" (UniqueName: \"kubernetes.io/projected/ed68d8aa-f1fe-4db4-bf11-033919ca1449-kube-api-access-sb8x6\") pod \"ed68d8aa-f1fe-4db4-bf11-033919ca1449\" (UID: \"ed68d8aa-f1fe-4db4-bf11-033919ca1449\") " Oct 08 20:00:28 crc kubenswrapper[4988]: I1008 20:00:28.235776 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ed68d8aa-f1fe-4db4-bf11-033919ca1449-dns-svc\") pod \"ed68d8aa-f1fe-4db4-bf11-033919ca1449\" (UID: \"ed68d8aa-f1fe-4db4-bf11-033919ca1449\") " Oct 08 20:00:28 crc kubenswrapper[4988]: I1008 20:00:28.235840 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ed68d8aa-f1fe-4db4-bf11-033919ca1449-ovsdbserver-nb\") pod \"ed68d8aa-f1fe-4db4-bf11-033919ca1449\" (UID: \"ed68d8aa-f1fe-4db4-bf11-033919ca1449\") " Oct 08 20:00:28 crc kubenswrapper[4988]: I1008 20:00:28.235884 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ed68d8aa-f1fe-4db4-bf11-033919ca1449-ovsdbserver-sb\") pod \"ed68d8aa-f1fe-4db4-bf11-033919ca1449\" (UID: \"ed68d8aa-f1fe-4db4-bf11-033919ca1449\") " Oct 08 20:00:28 crc kubenswrapper[4988]: I1008 20:00:28.237503 4988 scope.go:117] "RemoveContainer" containerID="439af713ff941f973b8758bd6af3c097808f76ff520b96ff10bd3268811393f7" Oct 08 20:00:28 crc kubenswrapper[4988]: E1008 20:00:28.238034 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nm54f_openshift-machine-config-operator(f328a20b-5283-407e-8a93-2ab2da599f18)\"" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" Oct 08 20:00:28 crc kubenswrapper[4988]: I1008 20:00:28.241633 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ed68d8aa-f1fe-4db4-bf11-033919ca1449-kube-api-access-sb8x6" (OuterVolumeSpecName: "kube-api-access-sb8x6") pod "ed68d8aa-f1fe-4db4-bf11-033919ca1449" (UID: "ed68d8aa-f1fe-4db4-bf11-033919ca1449"). InnerVolumeSpecName "kube-api-access-sb8x6". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 20:00:28 crc kubenswrapper[4988]: I1008 20:00:28.290608 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ed68d8aa-f1fe-4db4-bf11-033919ca1449-config" (OuterVolumeSpecName: "config") pod "ed68d8aa-f1fe-4db4-bf11-033919ca1449" (UID: "ed68d8aa-f1fe-4db4-bf11-033919ca1449"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 20:00:28 crc kubenswrapper[4988]: I1008 20:00:28.291000 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ed68d8aa-f1fe-4db4-bf11-033919ca1449-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "ed68d8aa-f1fe-4db4-bf11-033919ca1449" (UID: "ed68d8aa-f1fe-4db4-bf11-033919ca1449"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 20:00:28 crc kubenswrapper[4988]: I1008 20:00:28.291809 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ed68d8aa-f1fe-4db4-bf11-033919ca1449-openstack-cell1" (OuterVolumeSpecName: "openstack-cell1") pod "ed68d8aa-f1fe-4db4-bf11-033919ca1449" (UID: "ed68d8aa-f1fe-4db4-bf11-033919ca1449"). InnerVolumeSpecName "openstack-cell1". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 20:00:28 crc kubenswrapper[4988]: I1008 20:00:28.317916 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ed68d8aa-f1fe-4db4-bf11-033919ca1449-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "ed68d8aa-f1fe-4db4-bf11-033919ca1449" (UID: "ed68d8aa-f1fe-4db4-bf11-033919ca1449"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 20:00:28 crc kubenswrapper[4988]: I1008 20:00:28.323583 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ed68d8aa-f1fe-4db4-bf11-033919ca1449-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "ed68d8aa-f1fe-4db4-bf11-033919ca1449" (UID: "ed68d8aa-f1fe-4db4-bf11-033919ca1449"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 20:00:28 crc kubenswrapper[4988]: I1008 20:00:28.339133 4988 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ed68d8aa-f1fe-4db4-bf11-033919ca1449-config\") on node \"crc\" DevicePath \"\"" Oct 08 20:00:28 crc kubenswrapper[4988]: I1008 20:00:28.339416 4988 reconciler_common.go:293] "Volume detached for volume \"openstack-cell1\" (UniqueName: \"kubernetes.io/configmap/ed68d8aa-f1fe-4db4-bf11-033919ca1449-openstack-cell1\") on node \"crc\" DevicePath \"\"" Oct 08 20:00:28 crc kubenswrapper[4988]: I1008 20:00:28.339501 4988 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sb8x6\" (UniqueName: \"kubernetes.io/projected/ed68d8aa-f1fe-4db4-bf11-033919ca1449-kube-api-access-sb8x6\") on node \"crc\" DevicePath \"\"" Oct 08 20:00:28 crc kubenswrapper[4988]: I1008 20:00:28.339566 4988 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ed68d8aa-f1fe-4db4-bf11-033919ca1449-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 08 20:00:28 crc kubenswrapper[4988]: I1008 20:00:28.339656 4988 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ed68d8aa-f1fe-4db4-bf11-033919ca1449-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 08 20:00:28 crc kubenswrapper[4988]: I1008 20:00:28.342557 4988 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ed68d8aa-f1fe-4db4-bf11-033919ca1449-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 08 20:00:28 crc kubenswrapper[4988]: I1008 20:00:28.956590 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-687f9c4c69-rnsd7" event={"ID":"ed68d8aa-f1fe-4db4-bf11-033919ca1449","Type":"ContainerDied","Data":"d429e156a98089ee6af2466914c0f989abcad1d09ea0fb5ea6ff7a256b43226d"} Oct 08 20:00:28 crc kubenswrapper[4988]: I1008 20:00:28.956927 4988 scope.go:117] "RemoveContainer" containerID="4076efaf6dd043f83ecaca99250372d807f2155e36c84cce51f39baaf5b4ef9f" Oct 08 20:00:28 crc kubenswrapper[4988]: I1008 20:00:28.957199 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-687f9c4c69-rnsd7" Oct 08 20:00:28 crc kubenswrapper[4988]: I1008 20:00:28.978210 4988 scope.go:117] "RemoveContainer" containerID="53df725348367f4fbaeb73973f8c30353e8a4091e5b8c61ed716a5144718acdb" Oct 08 20:00:29 crc kubenswrapper[4988]: I1008 20:00:29.010324 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-687f9c4c69-rnsd7"] Oct 08 20:00:29 crc kubenswrapper[4988]: I1008 20:00:29.020251 4988 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-687f9c4c69-rnsd7"] Oct 08 20:00:29 crc kubenswrapper[4988]: I1008 20:00:29.257666 4988 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ed68d8aa-f1fe-4db4-bf11-033919ca1449" path="/var/lib/kubelet/pods/ed68d8aa-f1fe-4db4-bf11-033919ca1449/volumes" Oct 08 20:00:37 crc kubenswrapper[4988]: I1008 20:00:37.689777 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/pre-adoption-validation-openstack-pre-adoption-openstack-cpgsv4"] Oct 08 20:00:37 crc kubenswrapper[4988]: E1008 20:00:37.690711 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ed68d8aa-f1fe-4db4-bf11-033919ca1449" containerName="dnsmasq-dns" Oct 08 20:00:37 crc kubenswrapper[4988]: I1008 20:00:37.690726 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="ed68d8aa-f1fe-4db4-bf11-033919ca1449" containerName="dnsmasq-dns" Oct 08 20:00:37 crc kubenswrapper[4988]: E1008 20:00:37.690746 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ed68d8aa-f1fe-4db4-bf11-033919ca1449" containerName="init" Oct 08 20:00:37 crc kubenswrapper[4988]: I1008 20:00:37.690751 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="ed68d8aa-f1fe-4db4-bf11-033919ca1449" containerName="init" Oct 08 20:00:37 crc kubenswrapper[4988]: E1008 20:00:37.690762 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5c03ce9a-8cfe-4830-a131-e5b6353bef7c" containerName="dnsmasq-dns" Oct 08 20:00:37 crc kubenswrapper[4988]: I1008 20:00:37.690769 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="5c03ce9a-8cfe-4830-a131-e5b6353bef7c" containerName="dnsmasq-dns" Oct 08 20:00:37 crc kubenswrapper[4988]: E1008 20:00:37.690797 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5c03ce9a-8cfe-4830-a131-e5b6353bef7c" containerName="init" Oct 08 20:00:37 crc kubenswrapper[4988]: I1008 20:00:37.690802 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="5c03ce9a-8cfe-4830-a131-e5b6353bef7c" containerName="init" Oct 08 20:00:37 crc kubenswrapper[4988]: I1008 20:00:37.691172 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="ed68d8aa-f1fe-4db4-bf11-033919ca1449" containerName="dnsmasq-dns" Oct 08 20:00:37 crc kubenswrapper[4988]: I1008 20:00:37.691186 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="5c03ce9a-8cfe-4830-a131-e5b6353bef7c" containerName="dnsmasq-dns" Oct 08 20:00:37 crc kubenswrapper[4988]: I1008 20:00:37.691897 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-cpgsv4" Oct 08 20:00:37 crc kubenswrapper[4988]: I1008 20:00:37.694271 4988 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 08 20:00:37 crc kubenswrapper[4988]: I1008 20:00:37.694465 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Oct 08 20:00:37 crc kubenswrapper[4988]: I1008 20:00:37.694521 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Oct 08 20:00:37 crc kubenswrapper[4988]: I1008 20:00:37.695774 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-4sdzr" Oct 08 20:00:37 crc kubenswrapper[4988]: I1008 20:00:37.711219 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/pre-adoption-validation-openstack-pre-adoption-openstack-cpgsv4"] Oct 08 20:00:37 crc kubenswrapper[4988]: I1008 20:00:37.773198 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/93e6e1ae-4ef7-43f3-87a4-14a000232170-inventory\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-cpgsv4\" (UID: \"93e6e1ae-4ef7-43f3-87a4-14a000232170\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-cpgsv4" Oct 08 20:00:37 crc kubenswrapper[4988]: I1008 20:00:37.773381 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pre-adoption-validation-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/93e6e1ae-4ef7-43f3-87a4-14a000232170-pre-adoption-validation-combined-ca-bundle\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-cpgsv4\" (UID: \"93e6e1ae-4ef7-43f3-87a4-14a000232170\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-cpgsv4" Oct 08 20:00:37 crc kubenswrapper[4988]: I1008 20:00:37.773629 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/93e6e1ae-4ef7-43f3-87a4-14a000232170-ssh-key\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-cpgsv4\" (UID: \"93e6e1ae-4ef7-43f3-87a4-14a000232170\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-cpgsv4" Oct 08 20:00:37 crc kubenswrapper[4988]: I1008 20:00:37.773830 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-699sq\" (UniqueName: \"kubernetes.io/projected/93e6e1ae-4ef7-43f3-87a4-14a000232170-kube-api-access-699sq\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-cpgsv4\" (UID: \"93e6e1ae-4ef7-43f3-87a4-14a000232170\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-cpgsv4" Oct 08 20:00:37 crc kubenswrapper[4988]: I1008 20:00:37.876514 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-699sq\" (UniqueName: \"kubernetes.io/projected/93e6e1ae-4ef7-43f3-87a4-14a000232170-kube-api-access-699sq\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-cpgsv4\" (UID: \"93e6e1ae-4ef7-43f3-87a4-14a000232170\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-cpgsv4" Oct 08 20:00:37 crc kubenswrapper[4988]: I1008 20:00:37.876868 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/93e6e1ae-4ef7-43f3-87a4-14a000232170-inventory\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-cpgsv4\" (UID: \"93e6e1ae-4ef7-43f3-87a4-14a000232170\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-cpgsv4" Oct 08 20:00:37 crc kubenswrapper[4988]: I1008 20:00:37.877054 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pre-adoption-validation-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/93e6e1ae-4ef7-43f3-87a4-14a000232170-pre-adoption-validation-combined-ca-bundle\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-cpgsv4\" (UID: \"93e6e1ae-4ef7-43f3-87a4-14a000232170\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-cpgsv4" Oct 08 20:00:37 crc kubenswrapper[4988]: I1008 20:00:37.877222 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/93e6e1ae-4ef7-43f3-87a4-14a000232170-ssh-key\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-cpgsv4\" (UID: \"93e6e1ae-4ef7-43f3-87a4-14a000232170\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-cpgsv4" Oct 08 20:00:37 crc kubenswrapper[4988]: I1008 20:00:37.883413 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/93e6e1ae-4ef7-43f3-87a4-14a000232170-ssh-key\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-cpgsv4\" (UID: \"93e6e1ae-4ef7-43f3-87a4-14a000232170\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-cpgsv4" Oct 08 20:00:37 crc kubenswrapper[4988]: I1008 20:00:37.883523 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pre-adoption-validation-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/93e6e1ae-4ef7-43f3-87a4-14a000232170-pre-adoption-validation-combined-ca-bundle\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-cpgsv4\" (UID: \"93e6e1ae-4ef7-43f3-87a4-14a000232170\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-cpgsv4" Oct 08 20:00:37 crc kubenswrapper[4988]: I1008 20:00:37.883974 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/93e6e1ae-4ef7-43f3-87a4-14a000232170-inventory\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-cpgsv4\" (UID: \"93e6e1ae-4ef7-43f3-87a4-14a000232170\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-cpgsv4" Oct 08 20:00:37 crc kubenswrapper[4988]: I1008 20:00:37.902084 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-699sq\" (UniqueName: \"kubernetes.io/projected/93e6e1ae-4ef7-43f3-87a4-14a000232170-kube-api-access-699sq\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-cpgsv4\" (UID: \"93e6e1ae-4ef7-43f3-87a4-14a000232170\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-cpgsv4" Oct 08 20:00:38 crc kubenswrapper[4988]: I1008 20:00:38.012620 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-cpgsv4" Oct 08 20:00:38 crc kubenswrapper[4988]: I1008 20:00:38.847348 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/pre-adoption-validation-openstack-pre-adoption-openstack-cpgsv4"] Oct 08 20:00:38 crc kubenswrapper[4988]: W1008 20:00:38.861724 4988 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod93e6e1ae_4ef7_43f3_87a4_14a000232170.slice/crio-324b356b9671d6daca39923af259e0514ef4f643ef4a184e9d9440d77d595d86 WatchSource:0}: Error finding container 324b356b9671d6daca39923af259e0514ef4f643ef4a184e9d9440d77d595d86: Status 404 returned error can't find the container with id 324b356b9671d6daca39923af259e0514ef4f643ef4a184e9d9440d77d595d86 Oct 08 20:00:39 crc kubenswrapper[4988]: I1008 20:00:39.125676 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-cpgsv4" event={"ID":"93e6e1ae-4ef7-43f3-87a4-14a000232170","Type":"ContainerStarted","Data":"324b356b9671d6daca39923af259e0514ef4f643ef4a184e9d9440d77d595d86"} Oct 08 20:00:39 crc kubenswrapper[4988]: I1008 20:00:39.238716 4988 scope.go:117] "RemoveContainer" containerID="439af713ff941f973b8758bd6af3c097808f76ff520b96ff10bd3268811393f7" Oct 08 20:00:39 crc kubenswrapper[4988]: E1008 20:00:39.239568 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nm54f_openshift-machine-config-operator(f328a20b-5283-407e-8a93-2ab2da599f18)\"" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" Oct 08 20:00:48 crc kubenswrapper[4988]: I1008 20:00:48.235090 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-cpgsv4" event={"ID":"93e6e1ae-4ef7-43f3-87a4-14a000232170","Type":"ContainerStarted","Data":"3aa315ce653f790e26ec0ae9bd158cd2f5d1cccc7d2a1310654952a4ad644bf0"} Oct 08 20:00:48 crc kubenswrapper[4988]: I1008 20:00:48.267206 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-cpgsv4" podStartSLOduration=2.777849222 podStartE2EDuration="11.267186293s" podCreationTimestamp="2025-10-08 20:00:37 +0000 UTC" firstStartedPulling="2025-10-08 20:00:38.864787168 +0000 UTC m=+6584.314629938" lastFinishedPulling="2025-10-08 20:00:47.354124199 +0000 UTC m=+6592.803967009" observedRunningTime="2025-10-08 20:00:48.257849426 +0000 UTC m=+6593.707692236" watchObservedRunningTime="2025-10-08 20:00:48.267186293 +0000 UTC m=+6593.717029073" Oct 08 20:00:51 crc kubenswrapper[4988]: I1008 20:00:51.238730 4988 scope.go:117] "RemoveContainer" containerID="439af713ff941f973b8758bd6af3c097808f76ff520b96ff10bd3268811393f7" Oct 08 20:00:51 crc kubenswrapper[4988]: E1008 20:00:51.239706 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nm54f_openshift-machine-config-operator(f328a20b-5283-407e-8a93-2ab2da599f18)\"" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" Oct 08 20:01:00 crc kubenswrapper[4988]: I1008 20:01:00.156231 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-cron-29332561-46rm8"] Oct 08 20:01:00 crc kubenswrapper[4988]: I1008 20:01:00.159329 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29332561-46rm8" Oct 08 20:01:00 crc kubenswrapper[4988]: I1008 20:01:00.190737 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29332561-46rm8"] Oct 08 20:01:00 crc kubenswrapper[4988]: I1008 20:01:00.336519 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-94s8k\" (UniqueName: \"kubernetes.io/projected/9855c453-5d95-4cc9-ad10-b418ae54ff46-kube-api-access-94s8k\") pod \"keystone-cron-29332561-46rm8\" (UID: \"9855c453-5d95-4cc9-ad10-b418ae54ff46\") " pod="openstack/keystone-cron-29332561-46rm8" Oct 08 20:01:00 crc kubenswrapper[4988]: I1008 20:01:00.336744 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9855c453-5d95-4cc9-ad10-b418ae54ff46-combined-ca-bundle\") pod \"keystone-cron-29332561-46rm8\" (UID: \"9855c453-5d95-4cc9-ad10-b418ae54ff46\") " pod="openstack/keystone-cron-29332561-46rm8" Oct 08 20:01:00 crc kubenswrapper[4988]: I1008 20:01:00.336941 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9855c453-5d95-4cc9-ad10-b418ae54ff46-config-data\") pod \"keystone-cron-29332561-46rm8\" (UID: \"9855c453-5d95-4cc9-ad10-b418ae54ff46\") " pod="openstack/keystone-cron-29332561-46rm8" Oct 08 20:01:00 crc kubenswrapper[4988]: I1008 20:01:00.337262 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/9855c453-5d95-4cc9-ad10-b418ae54ff46-fernet-keys\") pod \"keystone-cron-29332561-46rm8\" (UID: \"9855c453-5d95-4cc9-ad10-b418ae54ff46\") " pod="openstack/keystone-cron-29332561-46rm8" Oct 08 20:01:00 crc kubenswrapper[4988]: I1008 20:01:00.441018 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/9855c453-5d95-4cc9-ad10-b418ae54ff46-fernet-keys\") pod \"keystone-cron-29332561-46rm8\" (UID: \"9855c453-5d95-4cc9-ad10-b418ae54ff46\") " pod="openstack/keystone-cron-29332561-46rm8" Oct 08 20:01:00 crc kubenswrapper[4988]: I1008 20:01:00.441295 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-94s8k\" (UniqueName: \"kubernetes.io/projected/9855c453-5d95-4cc9-ad10-b418ae54ff46-kube-api-access-94s8k\") pod \"keystone-cron-29332561-46rm8\" (UID: \"9855c453-5d95-4cc9-ad10-b418ae54ff46\") " pod="openstack/keystone-cron-29332561-46rm8" Oct 08 20:01:00 crc kubenswrapper[4988]: I1008 20:01:00.441480 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9855c453-5d95-4cc9-ad10-b418ae54ff46-combined-ca-bundle\") pod \"keystone-cron-29332561-46rm8\" (UID: \"9855c453-5d95-4cc9-ad10-b418ae54ff46\") " pod="openstack/keystone-cron-29332561-46rm8" Oct 08 20:01:00 crc kubenswrapper[4988]: I1008 20:01:00.441596 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9855c453-5d95-4cc9-ad10-b418ae54ff46-config-data\") pod \"keystone-cron-29332561-46rm8\" (UID: \"9855c453-5d95-4cc9-ad10-b418ae54ff46\") " pod="openstack/keystone-cron-29332561-46rm8" Oct 08 20:01:00 crc kubenswrapper[4988]: I1008 20:01:00.450594 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9855c453-5d95-4cc9-ad10-b418ae54ff46-combined-ca-bundle\") pod \"keystone-cron-29332561-46rm8\" (UID: \"9855c453-5d95-4cc9-ad10-b418ae54ff46\") " pod="openstack/keystone-cron-29332561-46rm8" Oct 08 20:01:00 crc kubenswrapper[4988]: I1008 20:01:00.460250 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/9855c453-5d95-4cc9-ad10-b418ae54ff46-fernet-keys\") pod \"keystone-cron-29332561-46rm8\" (UID: \"9855c453-5d95-4cc9-ad10-b418ae54ff46\") " pod="openstack/keystone-cron-29332561-46rm8" Oct 08 20:01:00 crc kubenswrapper[4988]: I1008 20:01:00.463379 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9855c453-5d95-4cc9-ad10-b418ae54ff46-config-data\") pod \"keystone-cron-29332561-46rm8\" (UID: \"9855c453-5d95-4cc9-ad10-b418ae54ff46\") " pod="openstack/keystone-cron-29332561-46rm8" Oct 08 20:01:00 crc kubenswrapper[4988]: I1008 20:01:00.465284 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-94s8k\" (UniqueName: \"kubernetes.io/projected/9855c453-5d95-4cc9-ad10-b418ae54ff46-kube-api-access-94s8k\") pod \"keystone-cron-29332561-46rm8\" (UID: \"9855c453-5d95-4cc9-ad10-b418ae54ff46\") " pod="openstack/keystone-cron-29332561-46rm8" Oct 08 20:01:00 crc kubenswrapper[4988]: I1008 20:01:00.479436 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29332561-46rm8" Oct 08 20:01:01 crc kubenswrapper[4988]: I1008 20:01:01.003941 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29332561-46rm8"] Oct 08 20:01:01 crc kubenswrapper[4988]: I1008 20:01:01.426253 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29332561-46rm8" event={"ID":"9855c453-5d95-4cc9-ad10-b418ae54ff46","Type":"ContainerStarted","Data":"90db655c04a92381782a4794bb0f21cded6d946a1f4ea97f7d213ddae1f879f4"} Oct 08 20:01:01 crc kubenswrapper[4988]: I1008 20:01:01.426636 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29332561-46rm8" event={"ID":"9855c453-5d95-4cc9-ad10-b418ae54ff46","Type":"ContainerStarted","Data":"e1d47f38a20e3033f906e131adb51fa8928ed10ecde0158373af448447ff2a8c"} Oct 08 20:01:01 crc kubenswrapper[4988]: I1008 20:01:01.430941 4988 generic.go:334] "Generic (PLEG): container finished" podID="93e6e1ae-4ef7-43f3-87a4-14a000232170" containerID="3aa315ce653f790e26ec0ae9bd158cd2f5d1cccc7d2a1310654952a4ad644bf0" exitCode=0 Oct 08 20:01:01 crc kubenswrapper[4988]: I1008 20:01:01.430977 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-cpgsv4" event={"ID":"93e6e1ae-4ef7-43f3-87a4-14a000232170","Type":"ContainerDied","Data":"3aa315ce653f790e26ec0ae9bd158cd2f5d1cccc7d2a1310654952a4ad644bf0"} Oct 08 20:01:01 crc kubenswrapper[4988]: I1008 20:01:01.450444 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-cron-29332561-46rm8" podStartSLOduration=1.450426372 podStartE2EDuration="1.450426372s" podCreationTimestamp="2025-10-08 20:01:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 20:01:01.44349252 +0000 UTC m=+6606.893335310" watchObservedRunningTime="2025-10-08 20:01:01.450426372 +0000 UTC m=+6606.900269152" Oct 08 20:01:02 crc kubenswrapper[4988]: I1008 20:01:02.238797 4988 scope.go:117] "RemoveContainer" containerID="439af713ff941f973b8758bd6af3c097808f76ff520b96ff10bd3268811393f7" Oct 08 20:01:02 crc kubenswrapper[4988]: E1008 20:01:02.239613 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nm54f_openshift-machine-config-operator(f328a20b-5283-407e-8a93-2ab2da599f18)\"" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" Oct 08 20:01:02 crc kubenswrapper[4988]: I1008 20:01:02.962734 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-cpgsv4" Oct 08 20:01:03 crc kubenswrapper[4988]: I1008 20:01:03.103695 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/93e6e1ae-4ef7-43f3-87a4-14a000232170-inventory\") pod \"93e6e1ae-4ef7-43f3-87a4-14a000232170\" (UID: \"93e6e1ae-4ef7-43f3-87a4-14a000232170\") " Oct 08 20:01:03 crc kubenswrapper[4988]: I1008 20:01:03.103754 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pre-adoption-validation-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/93e6e1ae-4ef7-43f3-87a4-14a000232170-pre-adoption-validation-combined-ca-bundle\") pod \"93e6e1ae-4ef7-43f3-87a4-14a000232170\" (UID: \"93e6e1ae-4ef7-43f3-87a4-14a000232170\") " Oct 08 20:01:03 crc kubenswrapper[4988]: I1008 20:01:03.103876 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/93e6e1ae-4ef7-43f3-87a4-14a000232170-ssh-key\") pod \"93e6e1ae-4ef7-43f3-87a4-14a000232170\" (UID: \"93e6e1ae-4ef7-43f3-87a4-14a000232170\") " Oct 08 20:01:03 crc kubenswrapper[4988]: I1008 20:01:03.104034 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-699sq\" (UniqueName: \"kubernetes.io/projected/93e6e1ae-4ef7-43f3-87a4-14a000232170-kube-api-access-699sq\") pod \"93e6e1ae-4ef7-43f3-87a4-14a000232170\" (UID: \"93e6e1ae-4ef7-43f3-87a4-14a000232170\") " Oct 08 20:01:03 crc kubenswrapper[4988]: I1008 20:01:03.110822 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/93e6e1ae-4ef7-43f3-87a4-14a000232170-pre-adoption-validation-combined-ca-bundle" (OuterVolumeSpecName: "pre-adoption-validation-combined-ca-bundle") pod "93e6e1ae-4ef7-43f3-87a4-14a000232170" (UID: "93e6e1ae-4ef7-43f3-87a4-14a000232170"). InnerVolumeSpecName "pre-adoption-validation-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 20:01:03 crc kubenswrapper[4988]: I1008 20:01:03.121619 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/93e6e1ae-4ef7-43f3-87a4-14a000232170-kube-api-access-699sq" (OuterVolumeSpecName: "kube-api-access-699sq") pod "93e6e1ae-4ef7-43f3-87a4-14a000232170" (UID: "93e6e1ae-4ef7-43f3-87a4-14a000232170"). InnerVolumeSpecName "kube-api-access-699sq". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 20:01:03 crc kubenswrapper[4988]: I1008 20:01:03.140444 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/93e6e1ae-4ef7-43f3-87a4-14a000232170-inventory" (OuterVolumeSpecName: "inventory") pod "93e6e1ae-4ef7-43f3-87a4-14a000232170" (UID: "93e6e1ae-4ef7-43f3-87a4-14a000232170"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 20:01:03 crc kubenswrapper[4988]: I1008 20:01:03.144276 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/93e6e1ae-4ef7-43f3-87a4-14a000232170-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "93e6e1ae-4ef7-43f3-87a4-14a000232170" (UID: "93e6e1ae-4ef7-43f3-87a4-14a000232170"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 20:01:03 crc kubenswrapper[4988]: I1008 20:01:03.207113 4988 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/93e6e1ae-4ef7-43f3-87a4-14a000232170-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 08 20:01:03 crc kubenswrapper[4988]: I1008 20:01:03.207242 4988 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-699sq\" (UniqueName: \"kubernetes.io/projected/93e6e1ae-4ef7-43f3-87a4-14a000232170-kube-api-access-699sq\") on node \"crc\" DevicePath \"\"" Oct 08 20:01:03 crc kubenswrapper[4988]: I1008 20:01:03.207372 4988 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/93e6e1ae-4ef7-43f3-87a4-14a000232170-inventory\") on node \"crc\" DevicePath \"\"" Oct 08 20:01:03 crc kubenswrapper[4988]: I1008 20:01:03.207500 4988 reconciler_common.go:293] "Volume detached for volume \"pre-adoption-validation-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/93e6e1ae-4ef7-43f3-87a4-14a000232170-pre-adoption-validation-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 20:01:03 crc kubenswrapper[4988]: I1008 20:01:03.459163 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-cpgsv4" event={"ID":"93e6e1ae-4ef7-43f3-87a4-14a000232170","Type":"ContainerDied","Data":"324b356b9671d6daca39923af259e0514ef4f643ef4a184e9d9440d77d595d86"} Oct 08 20:01:03 crc kubenswrapper[4988]: I1008 20:01:03.459236 4988 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="324b356b9671d6daca39923af259e0514ef4f643ef4a184e9d9440d77d595d86" Oct 08 20:01:03 crc kubenswrapper[4988]: I1008 20:01:03.459321 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-cpgsv4" Oct 08 20:01:04 crc kubenswrapper[4988]: I1008 20:01:04.476480 4988 generic.go:334] "Generic (PLEG): container finished" podID="9855c453-5d95-4cc9-ad10-b418ae54ff46" containerID="90db655c04a92381782a4794bb0f21cded6d946a1f4ea97f7d213ddae1f879f4" exitCode=0 Oct 08 20:01:04 crc kubenswrapper[4988]: I1008 20:01:04.476586 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29332561-46rm8" event={"ID":"9855c453-5d95-4cc9-ad10-b418ae54ff46","Type":"ContainerDied","Data":"90db655c04a92381782a4794bb0f21cded6d946a1f4ea97f7d213ddae1f879f4"} Oct 08 20:01:05 crc kubenswrapper[4988]: I1008 20:01:05.931505 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29332561-46rm8" Oct 08 20:01:06 crc kubenswrapper[4988]: I1008 20:01:06.088452 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9855c453-5d95-4cc9-ad10-b418ae54ff46-config-data\") pod \"9855c453-5d95-4cc9-ad10-b418ae54ff46\" (UID: \"9855c453-5d95-4cc9-ad10-b418ae54ff46\") " Oct 08 20:01:06 crc kubenswrapper[4988]: I1008 20:01:06.088625 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9855c453-5d95-4cc9-ad10-b418ae54ff46-combined-ca-bundle\") pod \"9855c453-5d95-4cc9-ad10-b418ae54ff46\" (UID: \"9855c453-5d95-4cc9-ad10-b418ae54ff46\") " Oct 08 20:01:06 crc kubenswrapper[4988]: I1008 20:01:06.088814 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-94s8k\" (UniqueName: \"kubernetes.io/projected/9855c453-5d95-4cc9-ad10-b418ae54ff46-kube-api-access-94s8k\") pod \"9855c453-5d95-4cc9-ad10-b418ae54ff46\" (UID: \"9855c453-5d95-4cc9-ad10-b418ae54ff46\") " Oct 08 20:01:06 crc kubenswrapper[4988]: I1008 20:01:06.088940 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/9855c453-5d95-4cc9-ad10-b418ae54ff46-fernet-keys\") pod \"9855c453-5d95-4cc9-ad10-b418ae54ff46\" (UID: \"9855c453-5d95-4cc9-ad10-b418ae54ff46\") " Oct 08 20:01:06 crc kubenswrapper[4988]: I1008 20:01:06.096192 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9855c453-5d95-4cc9-ad10-b418ae54ff46-kube-api-access-94s8k" (OuterVolumeSpecName: "kube-api-access-94s8k") pod "9855c453-5d95-4cc9-ad10-b418ae54ff46" (UID: "9855c453-5d95-4cc9-ad10-b418ae54ff46"). InnerVolumeSpecName "kube-api-access-94s8k". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 20:01:06 crc kubenswrapper[4988]: I1008 20:01:06.096691 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9855c453-5d95-4cc9-ad10-b418ae54ff46-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "9855c453-5d95-4cc9-ad10-b418ae54ff46" (UID: "9855c453-5d95-4cc9-ad10-b418ae54ff46"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 20:01:06 crc kubenswrapper[4988]: I1008 20:01:06.133925 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9855c453-5d95-4cc9-ad10-b418ae54ff46-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "9855c453-5d95-4cc9-ad10-b418ae54ff46" (UID: "9855c453-5d95-4cc9-ad10-b418ae54ff46"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 20:01:06 crc kubenswrapper[4988]: I1008 20:01:06.178792 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9855c453-5d95-4cc9-ad10-b418ae54ff46-config-data" (OuterVolumeSpecName: "config-data") pod "9855c453-5d95-4cc9-ad10-b418ae54ff46" (UID: "9855c453-5d95-4cc9-ad10-b418ae54ff46"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 20:01:06 crc kubenswrapper[4988]: I1008 20:01:06.191513 4988 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9855c453-5d95-4cc9-ad10-b418ae54ff46-config-data\") on node \"crc\" DevicePath \"\"" Oct 08 20:01:06 crc kubenswrapper[4988]: I1008 20:01:06.191549 4988 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9855c453-5d95-4cc9-ad10-b418ae54ff46-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 20:01:06 crc kubenswrapper[4988]: I1008 20:01:06.191567 4988 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-94s8k\" (UniqueName: \"kubernetes.io/projected/9855c453-5d95-4cc9-ad10-b418ae54ff46-kube-api-access-94s8k\") on node \"crc\" DevicePath \"\"" Oct 08 20:01:06 crc kubenswrapper[4988]: I1008 20:01:06.191580 4988 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/9855c453-5d95-4cc9-ad10-b418ae54ff46-fernet-keys\") on node \"crc\" DevicePath \"\"" Oct 08 20:01:06 crc kubenswrapper[4988]: I1008 20:01:06.510341 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29332561-46rm8" event={"ID":"9855c453-5d95-4cc9-ad10-b418ae54ff46","Type":"ContainerDied","Data":"e1d47f38a20e3033f906e131adb51fa8928ed10ecde0158373af448447ff2a8c"} Oct 08 20:01:06 crc kubenswrapper[4988]: I1008 20:01:06.510442 4988 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e1d47f38a20e3033f906e131adb51fa8928ed10ecde0158373af448447ff2a8c" Oct 08 20:01:06 crc kubenswrapper[4988]: I1008 20:01:06.510472 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29332561-46rm8" Oct 08 20:01:10 crc kubenswrapper[4988]: I1008 20:01:10.650251 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-2xbjd"] Oct 08 20:01:10 crc kubenswrapper[4988]: E1008 20:01:10.651701 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9855c453-5d95-4cc9-ad10-b418ae54ff46" containerName="keystone-cron" Oct 08 20:01:10 crc kubenswrapper[4988]: I1008 20:01:10.651735 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="9855c453-5d95-4cc9-ad10-b418ae54ff46" containerName="keystone-cron" Oct 08 20:01:10 crc kubenswrapper[4988]: E1008 20:01:10.651783 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="93e6e1ae-4ef7-43f3-87a4-14a000232170" containerName="pre-adoption-validation-openstack-pre-adoption-openstack-cell1" Oct 08 20:01:10 crc kubenswrapper[4988]: I1008 20:01:10.651792 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="93e6e1ae-4ef7-43f3-87a4-14a000232170" containerName="pre-adoption-validation-openstack-pre-adoption-openstack-cell1" Oct 08 20:01:10 crc kubenswrapper[4988]: I1008 20:01:10.652259 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="9855c453-5d95-4cc9-ad10-b418ae54ff46" containerName="keystone-cron" Oct 08 20:01:10 crc kubenswrapper[4988]: I1008 20:01:10.652285 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="93e6e1ae-4ef7-43f3-87a4-14a000232170" containerName="pre-adoption-validation-openstack-pre-adoption-openstack-cell1" Oct 08 20:01:10 crc kubenswrapper[4988]: I1008 20:01:10.654266 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-2xbjd" Oct 08 20:01:10 crc kubenswrapper[4988]: I1008 20:01:10.657673 4988 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 08 20:01:10 crc kubenswrapper[4988]: I1008 20:01:10.658015 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Oct 08 20:01:10 crc kubenswrapper[4988]: I1008 20:01:10.658211 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-4sdzr" Oct 08 20:01:10 crc kubenswrapper[4988]: I1008 20:01:10.659845 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Oct 08 20:01:10 crc kubenswrapper[4988]: I1008 20:01:10.661713 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-2xbjd"] Oct 08 20:01:10 crc kubenswrapper[4988]: I1008 20:01:10.815168 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mp2pc\" (UniqueName: \"kubernetes.io/projected/7dd8825a-ca2d-4361-a6e0-cf8a68ba8078-kube-api-access-mp2pc\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-cell1-2xbjd\" (UID: \"7dd8825a-ca2d-4361-a6e0-cf8a68ba8078\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-2xbjd" Oct 08 20:01:10 crc kubenswrapper[4988]: I1008 20:01:10.815626 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7dd8825a-ca2d-4361-a6e0-cf8a68ba8078-ssh-key\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-cell1-2xbjd\" (UID: \"7dd8825a-ca2d-4361-a6e0-cf8a68ba8078\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-2xbjd" Oct 08 20:01:10 crc kubenswrapper[4988]: I1008 20:01:10.815801 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7dd8825a-ca2d-4361-a6e0-cf8a68ba8078-inventory\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-cell1-2xbjd\" (UID: \"7dd8825a-ca2d-4361-a6e0-cf8a68ba8078\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-2xbjd" Oct 08 20:01:10 crc kubenswrapper[4988]: I1008 20:01:10.816068 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tripleo-cleanup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7dd8825a-ca2d-4361-a6e0-cf8a68ba8078-tripleo-cleanup-combined-ca-bundle\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-cell1-2xbjd\" (UID: \"7dd8825a-ca2d-4361-a6e0-cf8a68ba8078\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-2xbjd" Oct 08 20:01:10 crc kubenswrapper[4988]: I1008 20:01:10.918088 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mp2pc\" (UniqueName: \"kubernetes.io/projected/7dd8825a-ca2d-4361-a6e0-cf8a68ba8078-kube-api-access-mp2pc\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-cell1-2xbjd\" (UID: \"7dd8825a-ca2d-4361-a6e0-cf8a68ba8078\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-2xbjd" Oct 08 20:01:10 crc kubenswrapper[4988]: I1008 20:01:10.918159 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7dd8825a-ca2d-4361-a6e0-cf8a68ba8078-ssh-key\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-cell1-2xbjd\" (UID: \"7dd8825a-ca2d-4361-a6e0-cf8a68ba8078\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-2xbjd" Oct 08 20:01:10 crc kubenswrapper[4988]: I1008 20:01:10.918200 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7dd8825a-ca2d-4361-a6e0-cf8a68ba8078-inventory\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-cell1-2xbjd\" (UID: \"7dd8825a-ca2d-4361-a6e0-cf8a68ba8078\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-2xbjd" Oct 08 20:01:10 crc kubenswrapper[4988]: I1008 20:01:10.918274 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tripleo-cleanup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7dd8825a-ca2d-4361-a6e0-cf8a68ba8078-tripleo-cleanup-combined-ca-bundle\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-cell1-2xbjd\" (UID: \"7dd8825a-ca2d-4361-a6e0-cf8a68ba8078\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-2xbjd" Oct 08 20:01:10 crc kubenswrapper[4988]: I1008 20:01:10.923546 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7dd8825a-ca2d-4361-a6e0-cf8a68ba8078-inventory\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-cell1-2xbjd\" (UID: \"7dd8825a-ca2d-4361-a6e0-cf8a68ba8078\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-2xbjd" Oct 08 20:01:10 crc kubenswrapper[4988]: I1008 20:01:10.925243 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tripleo-cleanup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7dd8825a-ca2d-4361-a6e0-cf8a68ba8078-tripleo-cleanup-combined-ca-bundle\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-cell1-2xbjd\" (UID: \"7dd8825a-ca2d-4361-a6e0-cf8a68ba8078\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-2xbjd" Oct 08 20:01:10 crc kubenswrapper[4988]: I1008 20:01:10.928437 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7dd8825a-ca2d-4361-a6e0-cf8a68ba8078-ssh-key\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-cell1-2xbjd\" (UID: \"7dd8825a-ca2d-4361-a6e0-cf8a68ba8078\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-2xbjd" Oct 08 20:01:10 crc kubenswrapper[4988]: I1008 20:01:10.936453 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mp2pc\" (UniqueName: \"kubernetes.io/projected/7dd8825a-ca2d-4361-a6e0-cf8a68ba8078-kube-api-access-mp2pc\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-cell1-2xbjd\" (UID: \"7dd8825a-ca2d-4361-a6e0-cf8a68ba8078\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-2xbjd" Oct 08 20:01:11 crc kubenswrapper[4988]: I1008 20:01:11.036967 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-2xbjd" Oct 08 20:01:11 crc kubenswrapper[4988]: I1008 20:01:11.616954 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-2xbjd"] Oct 08 20:01:12 crc kubenswrapper[4988]: I1008 20:01:12.606146 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-2xbjd" event={"ID":"7dd8825a-ca2d-4361-a6e0-cf8a68ba8078","Type":"ContainerStarted","Data":"2670b20aeeba35ee5c9d7282e9e1548b7b36e66ec32c3f84e10edf683b879b73"} Oct 08 20:01:12 crc kubenswrapper[4988]: I1008 20:01:12.607010 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-2xbjd" event={"ID":"7dd8825a-ca2d-4361-a6e0-cf8a68ba8078","Type":"ContainerStarted","Data":"5f2129db33f905fc4ee9ecabf8ea2a7e549b0e7c471ae49068a782f871eaf3ed"} Oct 08 20:01:12 crc kubenswrapper[4988]: I1008 20:01:12.644068 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-2xbjd" podStartSLOduration=2.011817213 podStartE2EDuration="2.644047972s" podCreationTimestamp="2025-10-08 20:01:10 +0000 UTC" firstStartedPulling="2025-10-08 20:01:11.62555608 +0000 UTC m=+6617.075398850" lastFinishedPulling="2025-10-08 20:01:12.257786849 +0000 UTC m=+6617.707629609" observedRunningTime="2025-10-08 20:01:12.62858217 +0000 UTC m=+6618.078424980" watchObservedRunningTime="2025-10-08 20:01:12.644047972 +0000 UTC m=+6618.093890752" Oct 08 20:01:13 crc kubenswrapper[4988]: I1008 20:01:13.337591 4988 scope.go:117] "RemoveContainer" containerID="1d4c823b49d5e59e790c219ce79ca123a64288d8978ef4875e9ade7e28b36d25" Oct 08 20:01:13 crc kubenswrapper[4988]: I1008 20:01:13.632091 4988 scope.go:117] "RemoveContainer" containerID="1ae46144f520666a647fe5640b334d415fb0e6714b2c54837415df1f1b0a681d" Oct 08 20:01:15 crc kubenswrapper[4988]: I1008 20:01:15.255344 4988 scope.go:117] "RemoveContainer" containerID="439af713ff941f973b8758bd6af3c097808f76ff520b96ff10bd3268811393f7" Oct 08 20:01:15 crc kubenswrapper[4988]: E1008 20:01:15.255770 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nm54f_openshift-machine-config-operator(f328a20b-5283-407e-8a93-2ab2da599f18)\"" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" Oct 08 20:01:27 crc kubenswrapper[4988]: I1008 20:01:27.238537 4988 scope.go:117] "RemoveContainer" containerID="439af713ff941f973b8758bd6af3c097808f76ff520b96ff10bd3268811393f7" Oct 08 20:01:27 crc kubenswrapper[4988]: E1008 20:01:27.239993 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nm54f_openshift-machine-config-operator(f328a20b-5283-407e-8a93-2ab2da599f18)\"" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" Oct 08 20:01:31 crc kubenswrapper[4988]: I1008 20:01:31.063266 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/octavia-db-create-lfm67"] Oct 08 20:01:31 crc kubenswrapper[4988]: I1008 20:01:31.073714 4988 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/octavia-db-create-lfm67"] Oct 08 20:01:31 crc kubenswrapper[4988]: I1008 20:01:31.254675 4988 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0dcba654-49b6-4faf-9013-cca23d157c74" path="/var/lib/kubelet/pods/0dcba654-49b6-4faf-9013-cca23d157c74/volumes" Oct 08 20:01:41 crc kubenswrapper[4988]: I1008 20:01:41.038596 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/octavia-70cf-account-create-66n9b"] Oct 08 20:01:41 crc kubenswrapper[4988]: I1008 20:01:41.048775 4988 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/octavia-70cf-account-create-66n9b"] Oct 08 20:01:41 crc kubenswrapper[4988]: I1008 20:01:41.238316 4988 scope.go:117] "RemoveContainer" containerID="439af713ff941f973b8758bd6af3c097808f76ff520b96ff10bd3268811393f7" Oct 08 20:01:41 crc kubenswrapper[4988]: E1008 20:01:41.238879 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nm54f_openshift-machine-config-operator(f328a20b-5283-407e-8a93-2ab2da599f18)\"" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" Oct 08 20:01:41 crc kubenswrapper[4988]: I1008 20:01:41.249255 4988 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="75cd79f3-6abb-4eb2-9e2a-02c6da1eb074" path="/var/lib/kubelet/pods/75cd79f3-6abb-4eb2-9e2a-02c6da1eb074/volumes" Oct 08 20:01:47 crc kubenswrapper[4988]: I1008 20:01:47.072994 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/octavia-persistence-db-create-7f7s2"] Oct 08 20:01:47 crc kubenswrapper[4988]: I1008 20:01:47.085176 4988 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/octavia-persistence-db-create-7f7s2"] Oct 08 20:01:47 crc kubenswrapper[4988]: I1008 20:01:47.251480 4988 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="039811dd-898b-4b0f-bf09-6530a670233e" path="/var/lib/kubelet/pods/039811dd-898b-4b0f-bf09-6530a670233e/volumes" Oct 08 20:01:54 crc kubenswrapper[4988]: I1008 20:01:54.238174 4988 scope.go:117] "RemoveContainer" containerID="439af713ff941f973b8758bd6af3c097808f76ff520b96ff10bd3268811393f7" Oct 08 20:01:54 crc kubenswrapper[4988]: E1008 20:01:54.239206 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nm54f_openshift-machine-config-operator(f328a20b-5283-407e-8a93-2ab2da599f18)\"" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" Oct 08 20:01:58 crc kubenswrapper[4988]: I1008 20:01:58.053105 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/octavia-0b5f-account-create-mvdkn"] Oct 08 20:01:58 crc kubenswrapper[4988]: I1008 20:01:58.067757 4988 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/octavia-0b5f-account-create-mvdkn"] Oct 08 20:01:59 crc kubenswrapper[4988]: I1008 20:01:59.260501 4988 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="868eea53-6829-45cf-9ebd-c504223fe1ae" path="/var/lib/kubelet/pods/868eea53-6829-45cf-9ebd-c504223fe1ae/volumes" Oct 08 20:02:07 crc kubenswrapper[4988]: I1008 20:02:07.240734 4988 scope.go:117] "RemoveContainer" containerID="439af713ff941f973b8758bd6af3c097808f76ff520b96ff10bd3268811393f7" Oct 08 20:02:07 crc kubenswrapper[4988]: E1008 20:02:07.241794 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nm54f_openshift-machine-config-operator(f328a20b-5283-407e-8a93-2ab2da599f18)\"" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" Oct 08 20:02:13 crc kubenswrapper[4988]: I1008 20:02:13.728247 4988 scope.go:117] "RemoveContainer" containerID="c154f8ad7eb6341238ad4b33fbbc84fb4d6309bbbd4d2aab84058e8f4371b9c8" Oct 08 20:02:13 crc kubenswrapper[4988]: I1008 20:02:13.767341 4988 scope.go:117] "RemoveContainer" containerID="8b3820b3ba038da4c218186f9a302029947854aaaaac288e9947ee6eb75efb33" Oct 08 20:02:13 crc kubenswrapper[4988]: I1008 20:02:13.826518 4988 scope.go:117] "RemoveContainer" containerID="fcb652f27029f5c4ec2dad99a9ccd16e58b1b8bc4b61fe19dd9f66678a1e730f" Oct 08 20:02:13 crc kubenswrapper[4988]: I1008 20:02:13.899303 4988 scope.go:117] "RemoveContainer" containerID="4aa0d55b4f3d154ace085d1fa54fbcd4ff8e3d440884648b943272a9e4954e63" Oct 08 20:02:20 crc kubenswrapper[4988]: I1008 20:02:20.246878 4988 scope.go:117] "RemoveContainer" containerID="439af713ff941f973b8758bd6af3c097808f76ff520b96ff10bd3268811393f7" Oct 08 20:02:20 crc kubenswrapper[4988]: E1008 20:02:20.248188 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nm54f_openshift-machine-config-operator(f328a20b-5283-407e-8a93-2ab2da599f18)\"" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" Oct 08 20:02:33 crc kubenswrapper[4988]: I1008 20:02:33.060120 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/octavia-db-sync-r8g9v"] Oct 08 20:02:33 crc kubenswrapper[4988]: I1008 20:02:33.072963 4988 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/octavia-db-sync-r8g9v"] Oct 08 20:02:33 crc kubenswrapper[4988]: I1008 20:02:33.252024 4988 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f75923d6-45dc-42fa-9f2b-2a64fb7490b1" path="/var/lib/kubelet/pods/f75923d6-45dc-42fa-9f2b-2a64fb7490b1/volumes" Oct 08 20:02:34 crc kubenswrapper[4988]: I1008 20:02:34.238047 4988 scope.go:117] "RemoveContainer" containerID="439af713ff941f973b8758bd6af3c097808f76ff520b96ff10bd3268811393f7" Oct 08 20:02:34 crc kubenswrapper[4988]: E1008 20:02:34.239086 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nm54f_openshift-machine-config-operator(f328a20b-5283-407e-8a93-2ab2da599f18)\"" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" Oct 08 20:02:47 crc kubenswrapper[4988]: I1008 20:02:47.238650 4988 scope.go:117] "RemoveContainer" containerID="439af713ff941f973b8758bd6af3c097808f76ff520b96ff10bd3268811393f7" Oct 08 20:02:47 crc kubenswrapper[4988]: E1008 20:02:47.239641 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nm54f_openshift-machine-config-operator(f328a20b-5283-407e-8a93-2ab2da599f18)\"" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" Oct 08 20:02:59 crc kubenswrapper[4988]: I1008 20:02:59.238518 4988 scope.go:117] "RemoveContainer" containerID="439af713ff941f973b8758bd6af3c097808f76ff520b96ff10bd3268811393f7" Oct 08 20:02:59 crc kubenswrapper[4988]: I1008 20:02:59.996575 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" event={"ID":"f328a20b-5283-407e-8a93-2ab2da599f18","Type":"ContainerStarted","Data":"9598dff28ed6b3b3fb6f6e42092378b208d30297c11600b02f78ebfa113063ea"} Oct 08 20:03:14 crc kubenswrapper[4988]: I1008 20:03:14.074553 4988 scope.go:117] "RemoveContainer" containerID="ea11f6f5c59b0839c5743a2f38ec67a1da94e3ef717c1e36f809ce296a109d4e" Oct 08 20:03:14 crc kubenswrapper[4988]: I1008 20:03:14.117711 4988 scope.go:117] "RemoveContainer" containerID="fd73f70995f5184740422e93c2520de613e32ba386667d673b23f5aaf4dba04b" Oct 08 20:03:14 crc kubenswrapper[4988]: I1008 20:03:14.161758 4988 scope.go:117] "RemoveContainer" containerID="35d074e20a1fa24b5d08bbd8fd08ca66a673a3e894293a2e998ac660f4dc7f21" Oct 08 20:03:14 crc kubenswrapper[4988]: I1008 20:03:14.198281 4988 scope.go:117] "RemoveContainer" containerID="0eb81c8aa96066cadf521105121ac612d65857e5ef8c6e6fe7e132225a6a4e9f" Oct 08 20:03:14 crc kubenswrapper[4988]: I1008 20:03:14.263734 4988 scope.go:117] "RemoveContainer" containerID="33c6f5043911fe4478477ea2b8457d1abcff028ec754adec057bfb6a524ecb3b" Oct 08 20:05:23 crc kubenswrapper[4988]: I1008 20:05:23.337976 4988 patch_prober.go:28] interesting pod/machine-config-daemon-nm54f container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 20:05:23 crc kubenswrapper[4988]: I1008 20:05:23.338718 4988 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 20:05:53 crc kubenswrapper[4988]: I1008 20:05:53.338301 4988 patch_prober.go:28] interesting pod/machine-config-daemon-nm54f container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 20:05:53 crc kubenswrapper[4988]: I1008 20:05:53.339038 4988 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 20:06:01 crc kubenswrapper[4988]: I1008 20:06:01.068638 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-db-create-rz6jz"] Oct 08 20:06:01 crc kubenswrapper[4988]: I1008 20:06:01.077083 4988 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-db-create-rz6jz"] Oct 08 20:06:01 crc kubenswrapper[4988]: I1008 20:06:01.252331 4988 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="23ccee36-ec99-4075-95f8-19cb2f402c3b" path="/var/lib/kubelet/pods/23ccee36-ec99-4075-95f8-19cb2f402c3b/volumes" Oct 08 20:06:11 crc kubenswrapper[4988]: I1008 20:06:11.037707 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-6119-account-create-fcfhx"] Oct 08 20:06:11 crc kubenswrapper[4988]: I1008 20:06:11.049497 4988 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-6119-account-create-fcfhx"] Oct 08 20:06:11 crc kubenswrapper[4988]: I1008 20:06:11.255715 4988 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d4b96fb0-1c91-4514-938b-5dbffa503fc1" path="/var/lib/kubelet/pods/d4b96fb0-1c91-4514-938b-5dbffa503fc1/volumes" Oct 08 20:06:14 crc kubenswrapper[4988]: I1008 20:06:14.485335 4988 scope.go:117] "RemoveContainer" containerID="18c2bea687f803c935f24db50662fe3c14af212c5ec3794b7cd3defdea79da15" Oct 08 20:06:14 crc kubenswrapper[4988]: I1008 20:06:14.549033 4988 scope.go:117] "RemoveContainer" containerID="2dd58f0f20f70d265cd0868a92e6a83410886f5cfaa37d2fb006a899a6b89420" Oct 08 20:06:23 crc kubenswrapper[4988]: I1008 20:06:23.337925 4988 patch_prober.go:28] interesting pod/machine-config-daemon-nm54f container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 20:06:23 crc kubenswrapper[4988]: I1008 20:06:23.338715 4988 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 20:06:23 crc kubenswrapper[4988]: I1008 20:06:23.338796 4988 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" Oct 08 20:06:23 crc kubenswrapper[4988]: I1008 20:06:23.340078 4988 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"9598dff28ed6b3b3fb6f6e42092378b208d30297c11600b02f78ebfa113063ea"} pod="openshift-machine-config-operator/machine-config-daemon-nm54f" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 08 20:06:23 crc kubenswrapper[4988]: I1008 20:06:23.340210 4988 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" containerName="machine-config-daemon" containerID="cri-o://9598dff28ed6b3b3fb6f6e42092378b208d30297c11600b02f78ebfa113063ea" gracePeriod=600 Oct 08 20:06:23 crc kubenswrapper[4988]: I1008 20:06:23.560471 4988 generic.go:334] "Generic (PLEG): container finished" podID="f328a20b-5283-407e-8a93-2ab2da599f18" containerID="9598dff28ed6b3b3fb6f6e42092378b208d30297c11600b02f78ebfa113063ea" exitCode=0 Oct 08 20:06:23 crc kubenswrapper[4988]: I1008 20:06:23.560500 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" event={"ID":"f328a20b-5283-407e-8a93-2ab2da599f18","Type":"ContainerDied","Data":"9598dff28ed6b3b3fb6f6e42092378b208d30297c11600b02f78ebfa113063ea"} Oct 08 20:06:23 crc kubenswrapper[4988]: I1008 20:06:23.560938 4988 scope.go:117] "RemoveContainer" containerID="439af713ff941f973b8758bd6af3c097808f76ff520b96ff10bd3268811393f7" Oct 08 20:06:24 crc kubenswrapper[4988]: I1008 20:06:24.580137 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" event={"ID":"f328a20b-5283-407e-8a93-2ab2da599f18","Type":"ContainerStarted","Data":"c7dc147e5c4549b85b8a906334a50f44db496ead5160780b469fe7774d8aa2fa"} Oct 08 20:06:25 crc kubenswrapper[4988]: I1008 20:06:25.058147 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-db-sync-xs9q8"] Oct 08 20:06:25 crc kubenswrapper[4988]: I1008 20:06:25.066567 4988 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-db-sync-xs9q8"] Oct 08 20:06:25 crc kubenswrapper[4988]: I1008 20:06:25.253758 4988 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="162b2195-0c02-4cfc-ae52-ac4cdc1e1a9c" path="/var/lib/kubelet/pods/162b2195-0c02-4cfc-ae52-ac4cdc1e1a9c/volumes" Oct 08 20:07:14 crc kubenswrapper[4988]: I1008 20:07:14.700670 4988 scope.go:117] "RemoveContainer" containerID="2f3511be8d5426549a31710170af0dbe435492910a7fc827500070e46b3487ec" Oct 08 20:08:23 crc kubenswrapper[4988]: I1008 20:08:23.338278 4988 patch_prober.go:28] interesting pod/machine-config-daemon-nm54f container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 20:08:23 crc kubenswrapper[4988]: I1008 20:08:23.338985 4988 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 20:08:52 crc kubenswrapper[4988]: I1008 20:08:52.632620 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-fs7vh"] Oct 08 20:08:52 crc kubenswrapper[4988]: I1008 20:08:52.636356 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-fs7vh" Oct 08 20:08:52 crc kubenswrapper[4988]: I1008 20:08:52.661694 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-fs7vh"] Oct 08 20:08:52 crc kubenswrapper[4988]: I1008 20:08:52.752563 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/350a6fe4-c0ca-4d1f-930b-721ee95b5e37-catalog-content\") pod \"redhat-operators-fs7vh\" (UID: \"350a6fe4-c0ca-4d1f-930b-721ee95b5e37\") " pod="openshift-marketplace/redhat-operators-fs7vh" Oct 08 20:08:52 crc kubenswrapper[4988]: I1008 20:08:52.752741 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/350a6fe4-c0ca-4d1f-930b-721ee95b5e37-utilities\") pod \"redhat-operators-fs7vh\" (UID: \"350a6fe4-c0ca-4d1f-930b-721ee95b5e37\") " pod="openshift-marketplace/redhat-operators-fs7vh" Oct 08 20:08:52 crc kubenswrapper[4988]: I1008 20:08:52.752807 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4jvdz\" (UniqueName: \"kubernetes.io/projected/350a6fe4-c0ca-4d1f-930b-721ee95b5e37-kube-api-access-4jvdz\") pod \"redhat-operators-fs7vh\" (UID: \"350a6fe4-c0ca-4d1f-930b-721ee95b5e37\") " pod="openshift-marketplace/redhat-operators-fs7vh" Oct 08 20:08:52 crc kubenswrapper[4988]: I1008 20:08:52.855467 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/350a6fe4-c0ca-4d1f-930b-721ee95b5e37-catalog-content\") pod \"redhat-operators-fs7vh\" (UID: \"350a6fe4-c0ca-4d1f-930b-721ee95b5e37\") " pod="openshift-marketplace/redhat-operators-fs7vh" Oct 08 20:08:52 crc kubenswrapper[4988]: I1008 20:08:52.855604 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/350a6fe4-c0ca-4d1f-930b-721ee95b5e37-utilities\") pod \"redhat-operators-fs7vh\" (UID: \"350a6fe4-c0ca-4d1f-930b-721ee95b5e37\") " pod="openshift-marketplace/redhat-operators-fs7vh" Oct 08 20:08:52 crc kubenswrapper[4988]: I1008 20:08:52.855655 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4jvdz\" (UniqueName: \"kubernetes.io/projected/350a6fe4-c0ca-4d1f-930b-721ee95b5e37-kube-api-access-4jvdz\") pod \"redhat-operators-fs7vh\" (UID: \"350a6fe4-c0ca-4d1f-930b-721ee95b5e37\") " pod="openshift-marketplace/redhat-operators-fs7vh" Oct 08 20:08:52 crc kubenswrapper[4988]: I1008 20:08:52.856215 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/350a6fe4-c0ca-4d1f-930b-721ee95b5e37-catalog-content\") pod \"redhat-operators-fs7vh\" (UID: \"350a6fe4-c0ca-4d1f-930b-721ee95b5e37\") " pod="openshift-marketplace/redhat-operators-fs7vh" Oct 08 20:08:52 crc kubenswrapper[4988]: I1008 20:08:52.856231 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/350a6fe4-c0ca-4d1f-930b-721ee95b5e37-utilities\") pod \"redhat-operators-fs7vh\" (UID: \"350a6fe4-c0ca-4d1f-930b-721ee95b5e37\") " pod="openshift-marketplace/redhat-operators-fs7vh" Oct 08 20:08:52 crc kubenswrapper[4988]: I1008 20:08:52.879455 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4jvdz\" (UniqueName: \"kubernetes.io/projected/350a6fe4-c0ca-4d1f-930b-721ee95b5e37-kube-api-access-4jvdz\") pod \"redhat-operators-fs7vh\" (UID: \"350a6fe4-c0ca-4d1f-930b-721ee95b5e37\") " pod="openshift-marketplace/redhat-operators-fs7vh" Oct 08 20:08:52 crc kubenswrapper[4988]: I1008 20:08:52.980324 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-fs7vh" Oct 08 20:08:53 crc kubenswrapper[4988]: I1008 20:08:53.337828 4988 patch_prober.go:28] interesting pod/machine-config-daemon-nm54f container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 20:08:53 crc kubenswrapper[4988]: I1008 20:08:53.338163 4988 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 20:08:53 crc kubenswrapper[4988]: I1008 20:08:53.486962 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-fs7vh"] Oct 08 20:08:54 crc kubenswrapper[4988]: I1008 20:08:54.429176 4988 generic.go:334] "Generic (PLEG): container finished" podID="350a6fe4-c0ca-4d1f-930b-721ee95b5e37" containerID="54e6da52b21e4f6c1116e7edae966df988e7fecacebf8cd96ad38901aee0b7d8" exitCode=0 Oct 08 20:08:54 crc kubenswrapper[4988]: I1008 20:08:54.429605 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-fs7vh" event={"ID":"350a6fe4-c0ca-4d1f-930b-721ee95b5e37","Type":"ContainerDied","Data":"54e6da52b21e4f6c1116e7edae966df988e7fecacebf8cd96ad38901aee0b7d8"} Oct 08 20:08:54 crc kubenswrapper[4988]: I1008 20:08:54.429670 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-fs7vh" event={"ID":"350a6fe4-c0ca-4d1f-930b-721ee95b5e37","Type":"ContainerStarted","Data":"f76ccbdb2f88df2fe8aa27f9052b49c842e57e615fb177d369152a74177e52f8"} Oct 08 20:08:54 crc kubenswrapper[4988]: I1008 20:08:54.434896 4988 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 08 20:08:55 crc kubenswrapper[4988]: I1008 20:08:55.441599 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-fs7vh" event={"ID":"350a6fe4-c0ca-4d1f-930b-721ee95b5e37","Type":"ContainerStarted","Data":"20a5c8985bfa4986366e0dd4fe5771a1c757d338627627aaa1e6e1dacdb8859f"} Oct 08 20:08:58 crc kubenswrapper[4988]: I1008 20:08:58.038297 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/aodh-db-create-9qhjh"] Oct 08 20:08:58 crc kubenswrapper[4988]: I1008 20:08:58.047948 4988 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/aodh-db-create-9qhjh"] Oct 08 20:08:59 crc kubenswrapper[4988]: I1008 20:08:59.255505 4988 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="792349a5-4d15-4349-9e44-159f8d54da1b" path="/var/lib/kubelet/pods/792349a5-4d15-4349-9e44-159f8d54da1b/volumes" Oct 08 20:08:59 crc kubenswrapper[4988]: I1008 20:08:59.489636 4988 generic.go:334] "Generic (PLEG): container finished" podID="350a6fe4-c0ca-4d1f-930b-721ee95b5e37" containerID="20a5c8985bfa4986366e0dd4fe5771a1c757d338627627aaa1e6e1dacdb8859f" exitCode=0 Oct 08 20:08:59 crc kubenswrapper[4988]: I1008 20:08:59.490109 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-fs7vh" event={"ID":"350a6fe4-c0ca-4d1f-930b-721ee95b5e37","Type":"ContainerDied","Data":"20a5c8985bfa4986366e0dd4fe5771a1c757d338627627aaa1e6e1dacdb8859f"} Oct 08 20:09:00 crc kubenswrapper[4988]: I1008 20:09:00.504880 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-fs7vh" event={"ID":"350a6fe4-c0ca-4d1f-930b-721ee95b5e37","Type":"ContainerStarted","Data":"e512032138022cd97da737fb923ae455feb8f68bb38bb053bd7045ea20322940"} Oct 08 20:09:00 crc kubenswrapper[4988]: I1008 20:09:00.533463 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-fs7vh" podStartSLOduration=3.061188374 podStartE2EDuration="8.533445322s" podCreationTimestamp="2025-10-08 20:08:52 +0000 UTC" firstStartedPulling="2025-10-08 20:08:54.434650107 +0000 UTC m=+7079.884492877" lastFinishedPulling="2025-10-08 20:08:59.906907015 +0000 UTC m=+7085.356749825" observedRunningTime="2025-10-08 20:09:00.531916883 +0000 UTC m=+7085.981759663" watchObservedRunningTime="2025-10-08 20:09:00.533445322 +0000 UTC m=+7085.983288092" Oct 08 20:09:02 crc kubenswrapper[4988]: I1008 20:09:02.980751 4988 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-fs7vh" Oct 08 20:09:02 crc kubenswrapper[4988]: I1008 20:09:02.981132 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-fs7vh" Oct 08 20:09:04 crc kubenswrapper[4988]: I1008 20:09:04.054324 4988 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-fs7vh" podUID="350a6fe4-c0ca-4d1f-930b-721ee95b5e37" containerName="registry-server" probeResult="failure" output=< Oct 08 20:09:04 crc kubenswrapper[4988]: timeout: failed to connect service ":50051" within 1s Oct 08 20:09:04 crc kubenswrapper[4988]: > Oct 08 20:09:09 crc kubenswrapper[4988]: I1008 20:09:09.036562 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/aodh-5f5f-account-create-bz4fj"] Oct 08 20:09:09 crc kubenswrapper[4988]: I1008 20:09:09.047736 4988 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/aodh-5f5f-account-create-bz4fj"] Oct 08 20:09:09 crc kubenswrapper[4988]: I1008 20:09:09.254318 4988 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a6617968-21ed-42a4-9f84-c21402f8e5e8" path="/var/lib/kubelet/pods/a6617968-21ed-42a4-9f84-c21402f8e5e8/volumes" Oct 08 20:09:14 crc kubenswrapper[4988]: I1008 20:09:14.066949 4988 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-fs7vh" podUID="350a6fe4-c0ca-4d1f-930b-721ee95b5e37" containerName="registry-server" probeResult="failure" output=< Oct 08 20:09:14 crc kubenswrapper[4988]: timeout: failed to connect service ":50051" within 1s Oct 08 20:09:14 crc kubenswrapper[4988]: > Oct 08 20:09:14 crc kubenswrapper[4988]: I1008 20:09:14.841406 4988 scope.go:117] "RemoveContainer" containerID="75efc3cdf2e5081df1f948e64db0c7baded2ceace55e40b0550cad5cb2f81e4f" Oct 08 20:09:14 crc kubenswrapper[4988]: I1008 20:09:14.876908 4988 scope.go:117] "RemoveContainer" containerID="dc43e822f0ef3124c6c1a70e5ff27eb7f9445265f8a1319b0705871c4682b3e6" Oct 08 20:09:19 crc kubenswrapper[4988]: I1008 20:09:19.052995 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/aodh-db-sync-lxj49"] Oct 08 20:09:19 crc kubenswrapper[4988]: I1008 20:09:19.069138 4988 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/aodh-db-sync-lxj49"] Oct 08 20:09:19 crc kubenswrapper[4988]: I1008 20:09:19.254769 4988 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0f268dff-3b76-4bbc-b5db-b2c910aae2c3" path="/var/lib/kubelet/pods/0f268dff-3b76-4bbc-b5db-b2c910aae2c3/volumes" Oct 08 20:09:23 crc kubenswrapper[4988]: I1008 20:09:23.075446 4988 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-fs7vh" Oct 08 20:09:23 crc kubenswrapper[4988]: I1008 20:09:23.166934 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-fs7vh" Oct 08 20:09:23 crc kubenswrapper[4988]: I1008 20:09:23.338364 4988 patch_prober.go:28] interesting pod/machine-config-daemon-nm54f container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 20:09:23 crc kubenswrapper[4988]: I1008 20:09:23.338522 4988 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 20:09:23 crc kubenswrapper[4988]: I1008 20:09:23.339307 4988 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" Oct 08 20:09:23 crc kubenswrapper[4988]: I1008 20:09:23.340910 4988 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"c7dc147e5c4549b85b8a906334a50f44db496ead5160780b469fe7774d8aa2fa"} pod="openshift-machine-config-operator/machine-config-daemon-nm54f" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 08 20:09:23 crc kubenswrapper[4988]: I1008 20:09:23.341069 4988 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" containerName="machine-config-daemon" containerID="cri-o://c7dc147e5c4549b85b8a906334a50f44db496ead5160780b469fe7774d8aa2fa" gracePeriod=600 Oct 08 20:09:23 crc kubenswrapper[4988]: E1008 20:09:23.471902 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nm54f_openshift-machine-config-operator(f328a20b-5283-407e-8a93-2ab2da599f18)\"" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" Oct 08 20:09:23 crc kubenswrapper[4988]: I1008 20:09:23.809871 4988 generic.go:334] "Generic (PLEG): container finished" podID="f328a20b-5283-407e-8a93-2ab2da599f18" containerID="c7dc147e5c4549b85b8a906334a50f44db496ead5160780b469fe7774d8aa2fa" exitCode=0 Oct 08 20:09:23 crc kubenswrapper[4988]: I1008 20:09:23.809930 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" event={"ID":"f328a20b-5283-407e-8a93-2ab2da599f18","Type":"ContainerDied","Data":"c7dc147e5c4549b85b8a906334a50f44db496ead5160780b469fe7774d8aa2fa"} Oct 08 20:09:23 crc kubenswrapper[4988]: I1008 20:09:23.810338 4988 scope.go:117] "RemoveContainer" containerID="9598dff28ed6b3b3fb6f6e42092378b208d30297c11600b02f78ebfa113063ea" Oct 08 20:09:23 crc kubenswrapper[4988]: I1008 20:09:23.811124 4988 scope.go:117] "RemoveContainer" containerID="c7dc147e5c4549b85b8a906334a50f44db496ead5160780b469fe7774d8aa2fa" Oct 08 20:09:23 crc kubenswrapper[4988]: E1008 20:09:23.811401 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nm54f_openshift-machine-config-operator(f328a20b-5283-407e-8a93-2ab2da599f18)\"" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" Oct 08 20:09:23 crc kubenswrapper[4988]: I1008 20:09:23.820692 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-fs7vh"] Oct 08 20:09:24 crc kubenswrapper[4988]: I1008 20:09:24.825420 4988 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-fs7vh" podUID="350a6fe4-c0ca-4d1f-930b-721ee95b5e37" containerName="registry-server" containerID="cri-o://e512032138022cd97da737fb923ae455feb8f68bb38bb053bd7045ea20322940" gracePeriod=2 Oct 08 20:09:25 crc kubenswrapper[4988]: I1008 20:09:25.328554 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-fs7vh" Oct 08 20:09:25 crc kubenswrapper[4988]: I1008 20:09:25.521169 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4jvdz\" (UniqueName: \"kubernetes.io/projected/350a6fe4-c0ca-4d1f-930b-721ee95b5e37-kube-api-access-4jvdz\") pod \"350a6fe4-c0ca-4d1f-930b-721ee95b5e37\" (UID: \"350a6fe4-c0ca-4d1f-930b-721ee95b5e37\") " Oct 08 20:09:25 crc kubenswrapper[4988]: I1008 20:09:25.521603 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/350a6fe4-c0ca-4d1f-930b-721ee95b5e37-utilities\") pod \"350a6fe4-c0ca-4d1f-930b-721ee95b5e37\" (UID: \"350a6fe4-c0ca-4d1f-930b-721ee95b5e37\") " Oct 08 20:09:25 crc kubenswrapper[4988]: I1008 20:09:25.521668 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/350a6fe4-c0ca-4d1f-930b-721ee95b5e37-catalog-content\") pod \"350a6fe4-c0ca-4d1f-930b-721ee95b5e37\" (UID: \"350a6fe4-c0ca-4d1f-930b-721ee95b5e37\") " Oct 08 20:09:25 crc kubenswrapper[4988]: I1008 20:09:25.522346 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/350a6fe4-c0ca-4d1f-930b-721ee95b5e37-utilities" (OuterVolumeSpecName: "utilities") pod "350a6fe4-c0ca-4d1f-930b-721ee95b5e37" (UID: "350a6fe4-c0ca-4d1f-930b-721ee95b5e37"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 20:09:25 crc kubenswrapper[4988]: I1008 20:09:25.529721 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/350a6fe4-c0ca-4d1f-930b-721ee95b5e37-kube-api-access-4jvdz" (OuterVolumeSpecName: "kube-api-access-4jvdz") pod "350a6fe4-c0ca-4d1f-930b-721ee95b5e37" (UID: "350a6fe4-c0ca-4d1f-930b-721ee95b5e37"). InnerVolumeSpecName "kube-api-access-4jvdz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 20:09:25 crc kubenswrapper[4988]: I1008 20:09:25.625447 4988 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/350a6fe4-c0ca-4d1f-930b-721ee95b5e37-utilities\") on node \"crc\" DevicePath \"\"" Oct 08 20:09:25 crc kubenswrapper[4988]: I1008 20:09:25.625519 4988 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4jvdz\" (UniqueName: \"kubernetes.io/projected/350a6fe4-c0ca-4d1f-930b-721ee95b5e37-kube-api-access-4jvdz\") on node \"crc\" DevicePath \"\"" Oct 08 20:09:25 crc kubenswrapper[4988]: I1008 20:09:25.639965 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/350a6fe4-c0ca-4d1f-930b-721ee95b5e37-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "350a6fe4-c0ca-4d1f-930b-721ee95b5e37" (UID: "350a6fe4-c0ca-4d1f-930b-721ee95b5e37"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 20:09:25 crc kubenswrapper[4988]: I1008 20:09:25.727379 4988 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/350a6fe4-c0ca-4d1f-930b-721ee95b5e37-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 08 20:09:25 crc kubenswrapper[4988]: I1008 20:09:25.838125 4988 generic.go:334] "Generic (PLEG): container finished" podID="350a6fe4-c0ca-4d1f-930b-721ee95b5e37" containerID="e512032138022cd97da737fb923ae455feb8f68bb38bb053bd7045ea20322940" exitCode=0 Oct 08 20:09:25 crc kubenswrapper[4988]: I1008 20:09:25.838578 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-fs7vh" event={"ID":"350a6fe4-c0ca-4d1f-930b-721ee95b5e37","Type":"ContainerDied","Data":"e512032138022cd97da737fb923ae455feb8f68bb38bb053bd7045ea20322940"} Oct 08 20:09:25 crc kubenswrapper[4988]: I1008 20:09:25.838653 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-fs7vh" event={"ID":"350a6fe4-c0ca-4d1f-930b-721ee95b5e37","Type":"ContainerDied","Data":"f76ccbdb2f88df2fe8aa27f9052b49c842e57e615fb177d369152a74177e52f8"} Oct 08 20:09:25 crc kubenswrapper[4988]: I1008 20:09:25.838682 4988 scope.go:117] "RemoveContainer" containerID="e512032138022cd97da737fb923ae455feb8f68bb38bb053bd7045ea20322940" Oct 08 20:09:25 crc kubenswrapper[4988]: I1008 20:09:25.838891 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-fs7vh" Oct 08 20:09:25 crc kubenswrapper[4988]: I1008 20:09:25.890548 4988 scope.go:117] "RemoveContainer" containerID="20a5c8985bfa4986366e0dd4fe5771a1c757d338627627aaa1e6e1dacdb8859f" Oct 08 20:09:25 crc kubenswrapper[4988]: I1008 20:09:25.896234 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-fs7vh"] Oct 08 20:09:25 crc kubenswrapper[4988]: I1008 20:09:25.905835 4988 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-fs7vh"] Oct 08 20:09:25 crc kubenswrapper[4988]: I1008 20:09:25.924600 4988 scope.go:117] "RemoveContainer" containerID="54e6da52b21e4f6c1116e7edae966df988e7fecacebf8cd96ad38901aee0b7d8" Oct 08 20:09:25 crc kubenswrapper[4988]: I1008 20:09:25.998476 4988 scope.go:117] "RemoveContainer" containerID="e512032138022cd97da737fb923ae455feb8f68bb38bb053bd7045ea20322940" Oct 08 20:09:25 crc kubenswrapper[4988]: E1008 20:09:25.999434 4988 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e512032138022cd97da737fb923ae455feb8f68bb38bb053bd7045ea20322940\": container with ID starting with e512032138022cd97da737fb923ae455feb8f68bb38bb053bd7045ea20322940 not found: ID does not exist" containerID="e512032138022cd97da737fb923ae455feb8f68bb38bb053bd7045ea20322940" Oct 08 20:09:25 crc kubenswrapper[4988]: I1008 20:09:25.999499 4988 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e512032138022cd97da737fb923ae455feb8f68bb38bb053bd7045ea20322940"} err="failed to get container status \"e512032138022cd97da737fb923ae455feb8f68bb38bb053bd7045ea20322940\": rpc error: code = NotFound desc = could not find container \"e512032138022cd97da737fb923ae455feb8f68bb38bb053bd7045ea20322940\": container with ID starting with e512032138022cd97da737fb923ae455feb8f68bb38bb053bd7045ea20322940 not found: ID does not exist" Oct 08 20:09:25 crc kubenswrapper[4988]: I1008 20:09:25.999539 4988 scope.go:117] "RemoveContainer" containerID="20a5c8985bfa4986366e0dd4fe5771a1c757d338627627aaa1e6e1dacdb8859f" Oct 08 20:09:26 crc kubenswrapper[4988]: E1008 20:09:26.000346 4988 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"20a5c8985bfa4986366e0dd4fe5771a1c757d338627627aaa1e6e1dacdb8859f\": container with ID starting with 20a5c8985bfa4986366e0dd4fe5771a1c757d338627627aaa1e6e1dacdb8859f not found: ID does not exist" containerID="20a5c8985bfa4986366e0dd4fe5771a1c757d338627627aaa1e6e1dacdb8859f" Oct 08 20:09:26 crc kubenswrapper[4988]: I1008 20:09:26.000505 4988 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"20a5c8985bfa4986366e0dd4fe5771a1c757d338627627aaa1e6e1dacdb8859f"} err="failed to get container status \"20a5c8985bfa4986366e0dd4fe5771a1c757d338627627aaa1e6e1dacdb8859f\": rpc error: code = NotFound desc = could not find container \"20a5c8985bfa4986366e0dd4fe5771a1c757d338627627aaa1e6e1dacdb8859f\": container with ID starting with 20a5c8985bfa4986366e0dd4fe5771a1c757d338627627aaa1e6e1dacdb8859f not found: ID does not exist" Oct 08 20:09:26 crc kubenswrapper[4988]: I1008 20:09:26.000681 4988 scope.go:117] "RemoveContainer" containerID="54e6da52b21e4f6c1116e7edae966df988e7fecacebf8cd96ad38901aee0b7d8" Oct 08 20:09:26 crc kubenswrapper[4988]: E1008 20:09:26.001071 4988 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"54e6da52b21e4f6c1116e7edae966df988e7fecacebf8cd96ad38901aee0b7d8\": container with ID starting with 54e6da52b21e4f6c1116e7edae966df988e7fecacebf8cd96ad38901aee0b7d8 not found: ID does not exist" containerID="54e6da52b21e4f6c1116e7edae966df988e7fecacebf8cd96ad38901aee0b7d8" Oct 08 20:09:26 crc kubenswrapper[4988]: I1008 20:09:26.001116 4988 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"54e6da52b21e4f6c1116e7edae966df988e7fecacebf8cd96ad38901aee0b7d8"} err="failed to get container status \"54e6da52b21e4f6c1116e7edae966df988e7fecacebf8cd96ad38901aee0b7d8\": rpc error: code = NotFound desc = could not find container \"54e6da52b21e4f6c1116e7edae966df988e7fecacebf8cd96ad38901aee0b7d8\": container with ID starting with 54e6da52b21e4f6c1116e7edae966df988e7fecacebf8cd96ad38901aee0b7d8 not found: ID does not exist" Oct 08 20:09:27 crc kubenswrapper[4988]: I1008 20:09:27.255312 4988 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="350a6fe4-c0ca-4d1f-930b-721ee95b5e37" path="/var/lib/kubelet/pods/350a6fe4-c0ca-4d1f-930b-721ee95b5e37/volumes" Oct 08 20:09:38 crc kubenswrapper[4988]: I1008 20:09:38.237771 4988 scope.go:117] "RemoveContainer" containerID="c7dc147e5c4549b85b8a906334a50f44db496ead5160780b469fe7774d8aa2fa" Oct 08 20:09:38 crc kubenswrapper[4988]: E1008 20:09:38.238680 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nm54f_openshift-machine-config-operator(f328a20b-5283-407e-8a93-2ab2da599f18)\"" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" Oct 08 20:09:40 crc kubenswrapper[4988]: I1008 20:09:40.486304 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-lhtgf"] Oct 08 20:09:40 crc kubenswrapper[4988]: E1008 20:09:40.488623 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="350a6fe4-c0ca-4d1f-930b-721ee95b5e37" containerName="extract-content" Oct 08 20:09:40 crc kubenswrapper[4988]: I1008 20:09:40.488638 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="350a6fe4-c0ca-4d1f-930b-721ee95b5e37" containerName="extract-content" Oct 08 20:09:40 crc kubenswrapper[4988]: E1008 20:09:40.488670 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="350a6fe4-c0ca-4d1f-930b-721ee95b5e37" containerName="extract-utilities" Oct 08 20:09:40 crc kubenswrapper[4988]: I1008 20:09:40.488677 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="350a6fe4-c0ca-4d1f-930b-721ee95b5e37" containerName="extract-utilities" Oct 08 20:09:40 crc kubenswrapper[4988]: E1008 20:09:40.488694 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="350a6fe4-c0ca-4d1f-930b-721ee95b5e37" containerName="registry-server" Oct 08 20:09:40 crc kubenswrapper[4988]: I1008 20:09:40.488699 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="350a6fe4-c0ca-4d1f-930b-721ee95b5e37" containerName="registry-server" Oct 08 20:09:40 crc kubenswrapper[4988]: I1008 20:09:40.488903 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="350a6fe4-c0ca-4d1f-930b-721ee95b5e37" containerName="registry-server" Oct 08 20:09:40 crc kubenswrapper[4988]: I1008 20:09:40.493645 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-lhtgf" Oct 08 20:09:40 crc kubenswrapper[4988]: I1008 20:09:40.517536 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-lhtgf"] Oct 08 20:09:40 crc kubenswrapper[4988]: I1008 20:09:40.634023 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/491de4c7-9179-4477-a0ad-044978c19ad4-utilities\") pod \"certified-operators-lhtgf\" (UID: \"491de4c7-9179-4477-a0ad-044978c19ad4\") " pod="openshift-marketplace/certified-operators-lhtgf" Oct 08 20:09:40 crc kubenswrapper[4988]: I1008 20:09:40.634079 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fkcqg\" (UniqueName: \"kubernetes.io/projected/491de4c7-9179-4477-a0ad-044978c19ad4-kube-api-access-fkcqg\") pod \"certified-operators-lhtgf\" (UID: \"491de4c7-9179-4477-a0ad-044978c19ad4\") " pod="openshift-marketplace/certified-operators-lhtgf" Oct 08 20:09:40 crc kubenswrapper[4988]: I1008 20:09:40.634113 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/491de4c7-9179-4477-a0ad-044978c19ad4-catalog-content\") pod \"certified-operators-lhtgf\" (UID: \"491de4c7-9179-4477-a0ad-044978c19ad4\") " pod="openshift-marketplace/certified-operators-lhtgf" Oct 08 20:09:40 crc kubenswrapper[4988]: I1008 20:09:40.740653 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/491de4c7-9179-4477-a0ad-044978c19ad4-utilities\") pod \"certified-operators-lhtgf\" (UID: \"491de4c7-9179-4477-a0ad-044978c19ad4\") " pod="openshift-marketplace/certified-operators-lhtgf" Oct 08 20:09:40 crc kubenswrapper[4988]: I1008 20:09:40.740732 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fkcqg\" (UniqueName: \"kubernetes.io/projected/491de4c7-9179-4477-a0ad-044978c19ad4-kube-api-access-fkcqg\") pod \"certified-operators-lhtgf\" (UID: \"491de4c7-9179-4477-a0ad-044978c19ad4\") " pod="openshift-marketplace/certified-operators-lhtgf" Oct 08 20:09:40 crc kubenswrapper[4988]: I1008 20:09:40.740776 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/491de4c7-9179-4477-a0ad-044978c19ad4-catalog-content\") pod \"certified-operators-lhtgf\" (UID: \"491de4c7-9179-4477-a0ad-044978c19ad4\") " pod="openshift-marketplace/certified-operators-lhtgf" Oct 08 20:09:40 crc kubenswrapper[4988]: I1008 20:09:40.741310 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/491de4c7-9179-4477-a0ad-044978c19ad4-catalog-content\") pod \"certified-operators-lhtgf\" (UID: \"491de4c7-9179-4477-a0ad-044978c19ad4\") " pod="openshift-marketplace/certified-operators-lhtgf" Oct 08 20:09:40 crc kubenswrapper[4988]: I1008 20:09:40.741611 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/491de4c7-9179-4477-a0ad-044978c19ad4-utilities\") pod \"certified-operators-lhtgf\" (UID: \"491de4c7-9179-4477-a0ad-044978c19ad4\") " pod="openshift-marketplace/certified-operators-lhtgf" Oct 08 20:09:40 crc kubenswrapper[4988]: I1008 20:09:40.780062 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fkcqg\" (UniqueName: \"kubernetes.io/projected/491de4c7-9179-4477-a0ad-044978c19ad4-kube-api-access-fkcqg\") pod \"certified-operators-lhtgf\" (UID: \"491de4c7-9179-4477-a0ad-044978c19ad4\") " pod="openshift-marketplace/certified-operators-lhtgf" Oct 08 20:09:40 crc kubenswrapper[4988]: I1008 20:09:40.823981 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-lhtgf" Oct 08 20:09:41 crc kubenswrapper[4988]: I1008 20:09:41.331981 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-lhtgf"] Oct 08 20:09:42 crc kubenswrapper[4988]: I1008 20:09:42.038280 4988 generic.go:334] "Generic (PLEG): container finished" podID="491de4c7-9179-4477-a0ad-044978c19ad4" containerID="903a117a26d8e1041eb62bb40c2e1a3d3f971c06a0b5f3db516aea31c982759f" exitCode=0 Oct 08 20:09:42 crc kubenswrapper[4988]: I1008 20:09:42.038329 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-lhtgf" event={"ID":"491de4c7-9179-4477-a0ad-044978c19ad4","Type":"ContainerDied","Data":"903a117a26d8e1041eb62bb40c2e1a3d3f971c06a0b5f3db516aea31c982759f"} Oct 08 20:09:42 crc kubenswrapper[4988]: I1008 20:09:42.039548 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-lhtgf" event={"ID":"491de4c7-9179-4477-a0ad-044978c19ad4","Type":"ContainerStarted","Data":"4552f5e74577b31ad65cdb339a3fcd4ff81bdeaeb7a406da1210fb2db38fbb53"} Oct 08 20:09:44 crc kubenswrapper[4988]: I1008 20:09:44.068482 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-lhtgf" event={"ID":"491de4c7-9179-4477-a0ad-044978c19ad4","Type":"ContainerStarted","Data":"ecfe020e694ecca07c989054a53163cf018d473755f1e85fae625d0c62504477"} Oct 08 20:09:45 crc kubenswrapper[4988]: I1008 20:09:45.083750 4988 generic.go:334] "Generic (PLEG): container finished" podID="491de4c7-9179-4477-a0ad-044978c19ad4" containerID="ecfe020e694ecca07c989054a53163cf018d473755f1e85fae625d0c62504477" exitCode=0 Oct 08 20:09:45 crc kubenswrapper[4988]: I1008 20:09:45.083883 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-lhtgf" event={"ID":"491de4c7-9179-4477-a0ad-044978c19ad4","Type":"ContainerDied","Data":"ecfe020e694ecca07c989054a53163cf018d473755f1e85fae625d0c62504477"} Oct 08 20:09:46 crc kubenswrapper[4988]: I1008 20:09:46.095806 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-lhtgf" event={"ID":"491de4c7-9179-4477-a0ad-044978c19ad4","Type":"ContainerStarted","Data":"7eca1731207007de2ddad694e501b73e0227ea3a352caa0ab000e5603cb60baa"} Oct 08 20:09:46 crc kubenswrapper[4988]: I1008 20:09:46.124648 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-lhtgf" podStartSLOduration=2.654155489 podStartE2EDuration="6.124630404s" podCreationTimestamp="2025-10-08 20:09:40 +0000 UTC" firstStartedPulling="2025-10-08 20:09:42.041787823 +0000 UTC m=+7127.491630593" lastFinishedPulling="2025-10-08 20:09:45.512262718 +0000 UTC m=+7130.962105508" observedRunningTime="2025-10-08 20:09:46.117864989 +0000 UTC m=+7131.567707819" watchObservedRunningTime="2025-10-08 20:09:46.124630404 +0000 UTC m=+7131.574473174" Oct 08 20:09:50 crc kubenswrapper[4988]: I1008 20:09:50.824833 4988 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-lhtgf" Oct 08 20:09:50 crc kubenswrapper[4988]: I1008 20:09:50.825966 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-lhtgf" Oct 08 20:09:50 crc kubenswrapper[4988]: I1008 20:09:50.943479 4988 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-lhtgf" Oct 08 20:09:51 crc kubenswrapper[4988]: I1008 20:09:51.256506 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-lhtgf" Oct 08 20:09:51 crc kubenswrapper[4988]: I1008 20:09:51.317001 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-lhtgf"] Oct 08 20:09:53 crc kubenswrapper[4988]: I1008 20:09:53.182751 4988 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-lhtgf" podUID="491de4c7-9179-4477-a0ad-044978c19ad4" containerName="registry-server" containerID="cri-o://7eca1731207007de2ddad694e501b73e0227ea3a352caa0ab000e5603cb60baa" gracePeriod=2 Oct 08 20:09:53 crc kubenswrapper[4988]: I1008 20:09:53.240461 4988 scope.go:117] "RemoveContainer" containerID="c7dc147e5c4549b85b8a906334a50f44db496ead5160780b469fe7774d8aa2fa" Oct 08 20:09:53 crc kubenswrapper[4988]: E1008 20:09:53.240823 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nm54f_openshift-machine-config-operator(f328a20b-5283-407e-8a93-2ab2da599f18)\"" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" Oct 08 20:09:53 crc kubenswrapper[4988]: I1008 20:09:53.728579 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-lhtgf" Oct 08 20:09:53 crc kubenswrapper[4988]: I1008 20:09:53.884020 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/491de4c7-9179-4477-a0ad-044978c19ad4-utilities\") pod \"491de4c7-9179-4477-a0ad-044978c19ad4\" (UID: \"491de4c7-9179-4477-a0ad-044978c19ad4\") " Oct 08 20:09:53 crc kubenswrapper[4988]: I1008 20:09:53.885207 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/491de4c7-9179-4477-a0ad-044978c19ad4-utilities" (OuterVolumeSpecName: "utilities") pod "491de4c7-9179-4477-a0ad-044978c19ad4" (UID: "491de4c7-9179-4477-a0ad-044978c19ad4"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 20:09:53 crc kubenswrapper[4988]: I1008 20:09:53.887562 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fkcqg\" (UniqueName: \"kubernetes.io/projected/491de4c7-9179-4477-a0ad-044978c19ad4-kube-api-access-fkcqg\") pod \"491de4c7-9179-4477-a0ad-044978c19ad4\" (UID: \"491de4c7-9179-4477-a0ad-044978c19ad4\") " Oct 08 20:09:53 crc kubenswrapper[4988]: I1008 20:09:53.887881 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/491de4c7-9179-4477-a0ad-044978c19ad4-catalog-content\") pod \"491de4c7-9179-4477-a0ad-044978c19ad4\" (UID: \"491de4c7-9179-4477-a0ad-044978c19ad4\") " Oct 08 20:09:53 crc kubenswrapper[4988]: I1008 20:09:53.888824 4988 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/491de4c7-9179-4477-a0ad-044978c19ad4-utilities\") on node \"crc\" DevicePath \"\"" Oct 08 20:09:53 crc kubenswrapper[4988]: I1008 20:09:53.896736 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/491de4c7-9179-4477-a0ad-044978c19ad4-kube-api-access-fkcqg" (OuterVolumeSpecName: "kube-api-access-fkcqg") pod "491de4c7-9179-4477-a0ad-044978c19ad4" (UID: "491de4c7-9179-4477-a0ad-044978c19ad4"). InnerVolumeSpecName "kube-api-access-fkcqg". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 20:09:53 crc kubenswrapper[4988]: I1008 20:09:53.970822 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/491de4c7-9179-4477-a0ad-044978c19ad4-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "491de4c7-9179-4477-a0ad-044978c19ad4" (UID: "491de4c7-9179-4477-a0ad-044978c19ad4"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 20:09:53 crc kubenswrapper[4988]: I1008 20:09:53.990997 4988 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fkcqg\" (UniqueName: \"kubernetes.io/projected/491de4c7-9179-4477-a0ad-044978c19ad4-kube-api-access-fkcqg\") on node \"crc\" DevicePath \"\"" Oct 08 20:09:53 crc kubenswrapper[4988]: I1008 20:09:53.991043 4988 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/491de4c7-9179-4477-a0ad-044978c19ad4-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 08 20:09:54 crc kubenswrapper[4988]: I1008 20:09:54.205252 4988 generic.go:334] "Generic (PLEG): container finished" podID="491de4c7-9179-4477-a0ad-044978c19ad4" containerID="7eca1731207007de2ddad694e501b73e0227ea3a352caa0ab000e5603cb60baa" exitCode=0 Oct 08 20:09:54 crc kubenswrapper[4988]: I1008 20:09:54.205328 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-lhtgf" event={"ID":"491de4c7-9179-4477-a0ad-044978c19ad4","Type":"ContainerDied","Data":"7eca1731207007de2ddad694e501b73e0227ea3a352caa0ab000e5603cb60baa"} Oct 08 20:09:54 crc kubenswrapper[4988]: I1008 20:09:54.205371 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-lhtgf" Oct 08 20:09:54 crc kubenswrapper[4988]: I1008 20:09:54.205425 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-lhtgf" event={"ID":"491de4c7-9179-4477-a0ad-044978c19ad4","Type":"ContainerDied","Data":"4552f5e74577b31ad65cdb339a3fcd4ff81bdeaeb7a406da1210fb2db38fbb53"} Oct 08 20:09:54 crc kubenswrapper[4988]: I1008 20:09:54.205469 4988 scope.go:117] "RemoveContainer" containerID="7eca1731207007de2ddad694e501b73e0227ea3a352caa0ab000e5603cb60baa" Oct 08 20:09:54 crc kubenswrapper[4988]: I1008 20:09:54.260189 4988 scope.go:117] "RemoveContainer" containerID="ecfe020e694ecca07c989054a53163cf018d473755f1e85fae625d0c62504477" Oct 08 20:09:54 crc kubenswrapper[4988]: I1008 20:09:54.263592 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-lhtgf"] Oct 08 20:09:54 crc kubenswrapper[4988]: I1008 20:09:54.274017 4988 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-lhtgf"] Oct 08 20:09:54 crc kubenswrapper[4988]: I1008 20:09:54.299854 4988 scope.go:117] "RemoveContainer" containerID="903a117a26d8e1041eb62bb40c2e1a3d3f971c06a0b5f3db516aea31c982759f" Oct 08 20:09:54 crc kubenswrapper[4988]: I1008 20:09:54.366160 4988 scope.go:117] "RemoveContainer" containerID="7eca1731207007de2ddad694e501b73e0227ea3a352caa0ab000e5603cb60baa" Oct 08 20:09:54 crc kubenswrapper[4988]: E1008 20:09:54.366891 4988 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7eca1731207007de2ddad694e501b73e0227ea3a352caa0ab000e5603cb60baa\": container with ID starting with 7eca1731207007de2ddad694e501b73e0227ea3a352caa0ab000e5603cb60baa not found: ID does not exist" containerID="7eca1731207007de2ddad694e501b73e0227ea3a352caa0ab000e5603cb60baa" Oct 08 20:09:54 crc kubenswrapper[4988]: I1008 20:09:54.366953 4988 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7eca1731207007de2ddad694e501b73e0227ea3a352caa0ab000e5603cb60baa"} err="failed to get container status \"7eca1731207007de2ddad694e501b73e0227ea3a352caa0ab000e5603cb60baa\": rpc error: code = NotFound desc = could not find container \"7eca1731207007de2ddad694e501b73e0227ea3a352caa0ab000e5603cb60baa\": container with ID starting with 7eca1731207007de2ddad694e501b73e0227ea3a352caa0ab000e5603cb60baa not found: ID does not exist" Oct 08 20:09:54 crc kubenswrapper[4988]: I1008 20:09:54.366993 4988 scope.go:117] "RemoveContainer" containerID="ecfe020e694ecca07c989054a53163cf018d473755f1e85fae625d0c62504477" Oct 08 20:09:54 crc kubenswrapper[4988]: E1008 20:09:54.367669 4988 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ecfe020e694ecca07c989054a53163cf018d473755f1e85fae625d0c62504477\": container with ID starting with ecfe020e694ecca07c989054a53163cf018d473755f1e85fae625d0c62504477 not found: ID does not exist" containerID="ecfe020e694ecca07c989054a53163cf018d473755f1e85fae625d0c62504477" Oct 08 20:09:54 crc kubenswrapper[4988]: I1008 20:09:54.367736 4988 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ecfe020e694ecca07c989054a53163cf018d473755f1e85fae625d0c62504477"} err="failed to get container status \"ecfe020e694ecca07c989054a53163cf018d473755f1e85fae625d0c62504477\": rpc error: code = NotFound desc = could not find container \"ecfe020e694ecca07c989054a53163cf018d473755f1e85fae625d0c62504477\": container with ID starting with ecfe020e694ecca07c989054a53163cf018d473755f1e85fae625d0c62504477 not found: ID does not exist" Oct 08 20:09:54 crc kubenswrapper[4988]: I1008 20:09:54.367785 4988 scope.go:117] "RemoveContainer" containerID="903a117a26d8e1041eb62bb40c2e1a3d3f971c06a0b5f3db516aea31c982759f" Oct 08 20:09:54 crc kubenswrapper[4988]: E1008 20:09:54.368276 4988 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"903a117a26d8e1041eb62bb40c2e1a3d3f971c06a0b5f3db516aea31c982759f\": container with ID starting with 903a117a26d8e1041eb62bb40c2e1a3d3f971c06a0b5f3db516aea31c982759f not found: ID does not exist" containerID="903a117a26d8e1041eb62bb40c2e1a3d3f971c06a0b5f3db516aea31c982759f" Oct 08 20:09:54 crc kubenswrapper[4988]: I1008 20:09:54.368335 4988 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"903a117a26d8e1041eb62bb40c2e1a3d3f971c06a0b5f3db516aea31c982759f"} err="failed to get container status \"903a117a26d8e1041eb62bb40c2e1a3d3f971c06a0b5f3db516aea31c982759f\": rpc error: code = NotFound desc = could not find container \"903a117a26d8e1041eb62bb40c2e1a3d3f971c06a0b5f3db516aea31c982759f\": container with ID starting with 903a117a26d8e1041eb62bb40c2e1a3d3f971c06a0b5f3db516aea31c982759f not found: ID does not exist" Oct 08 20:09:55 crc kubenswrapper[4988]: I1008 20:09:55.258941 4988 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="491de4c7-9179-4477-a0ad-044978c19ad4" path="/var/lib/kubelet/pods/491de4c7-9179-4477-a0ad-044978c19ad4/volumes" Oct 08 20:10:06 crc kubenswrapper[4988]: I1008 20:10:06.237076 4988 scope.go:117] "RemoveContainer" containerID="c7dc147e5c4549b85b8a906334a50f44db496ead5160780b469fe7774d8aa2fa" Oct 08 20:10:06 crc kubenswrapper[4988]: E1008 20:10:06.237788 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nm54f_openshift-machine-config-operator(f328a20b-5283-407e-8a93-2ab2da599f18)\"" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" Oct 08 20:10:14 crc kubenswrapper[4988]: I1008 20:10:14.996060 4988 scope.go:117] "RemoveContainer" containerID="7da83ede41b90de86590af7942c1bfeb0901902308e1fe5ae5b1ff61d501fd6c" Oct 08 20:10:18 crc kubenswrapper[4988]: I1008 20:10:18.238749 4988 scope.go:117] "RemoveContainer" containerID="c7dc147e5c4549b85b8a906334a50f44db496ead5160780b469fe7774d8aa2fa" Oct 08 20:10:18 crc kubenswrapper[4988]: E1008 20:10:18.240180 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nm54f_openshift-machine-config-operator(f328a20b-5283-407e-8a93-2ab2da599f18)\"" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" Oct 08 20:10:29 crc kubenswrapper[4988]: I1008 20:10:29.238920 4988 scope.go:117] "RemoveContainer" containerID="c7dc147e5c4549b85b8a906334a50f44db496ead5160780b469fe7774d8aa2fa" Oct 08 20:10:29 crc kubenswrapper[4988]: E1008 20:10:29.240664 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nm54f_openshift-machine-config-operator(f328a20b-5283-407e-8a93-2ab2da599f18)\"" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" Oct 08 20:10:44 crc kubenswrapper[4988]: I1008 20:10:44.238954 4988 scope.go:117] "RemoveContainer" containerID="c7dc147e5c4549b85b8a906334a50f44db496ead5160780b469fe7774d8aa2fa" Oct 08 20:10:44 crc kubenswrapper[4988]: E1008 20:10:44.240244 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nm54f_openshift-machine-config-operator(f328a20b-5283-407e-8a93-2ab2da599f18)\"" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" Oct 08 20:10:57 crc kubenswrapper[4988]: I1008 20:10:57.246475 4988 scope.go:117] "RemoveContainer" containerID="c7dc147e5c4549b85b8a906334a50f44db496ead5160780b469fe7774d8aa2fa" Oct 08 20:10:57 crc kubenswrapper[4988]: E1008 20:10:57.247624 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nm54f_openshift-machine-config-operator(f328a20b-5283-407e-8a93-2ab2da599f18)\"" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" Oct 08 20:11:10 crc kubenswrapper[4988]: I1008 20:11:10.239304 4988 scope.go:117] "RemoveContainer" containerID="c7dc147e5c4549b85b8a906334a50f44db496ead5160780b469fe7774d8aa2fa" Oct 08 20:11:10 crc kubenswrapper[4988]: E1008 20:11:10.240056 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nm54f_openshift-machine-config-operator(f328a20b-5283-407e-8a93-2ab2da599f18)\"" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" Oct 08 20:11:22 crc kubenswrapper[4988]: I1008 20:11:22.238800 4988 scope.go:117] "RemoveContainer" containerID="c7dc147e5c4549b85b8a906334a50f44db496ead5160780b469fe7774d8aa2fa" Oct 08 20:11:22 crc kubenswrapper[4988]: E1008 20:11:22.240212 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nm54f_openshift-machine-config-operator(f328a20b-5283-407e-8a93-2ab2da599f18)\"" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" Oct 08 20:11:33 crc kubenswrapper[4988]: I1008 20:11:33.237269 4988 scope.go:117] "RemoveContainer" containerID="c7dc147e5c4549b85b8a906334a50f44db496ead5160780b469fe7774d8aa2fa" Oct 08 20:11:33 crc kubenswrapper[4988]: E1008 20:11:33.238027 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nm54f_openshift-machine-config-operator(f328a20b-5283-407e-8a93-2ab2da599f18)\"" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" Oct 08 20:11:47 crc kubenswrapper[4988]: I1008 20:11:47.238467 4988 scope.go:117] "RemoveContainer" containerID="c7dc147e5c4549b85b8a906334a50f44db496ead5160780b469fe7774d8aa2fa" Oct 08 20:11:47 crc kubenswrapper[4988]: E1008 20:11:47.239569 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nm54f_openshift-machine-config-operator(f328a20b-5283-407e-8a93-2ab2da599f18)\"" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" Oct 08 20:12:02 crc kubenswrapper[4988]: I1008 20:12:02.238451 4988 scope.go:117] "RemoveContainer" containerID="c7dc147e5c4549b85b8a906334a50f44db496ead5160780b469fe7774d8aa2fa" Oct 08 20:12:02 crc kubenswrapper[4988]: E1008 20:12:02.239676 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nm54f_openshift-machine-config-operator(f328a20b-5283-407e-8a93-2ab2da599f18)\"" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" Oct 08 20:12:15 crc kubenswrapper[4988]: I1008 20:12:15.254771 4988 scope.go:117] "RemoveContainer" containerID="c7dc147e5c4549b85b8a906334a50f44db496ead5160780b469fe7774d8aa2fa" Oct 08 20:12:15 crc kubenswrapper[4988]: E1008 20:12:15.255893 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nm54f_openshift-machine-config-operator(f328a20b-5283-407e-8a93-2ab2da599f18)\"" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" Oct 08 20:12:30 crc kubenswrapper[4988]: I1008 20:12:30.238915 4988 scope.go:117] "RemoveContainer" containerID="c7dc147e5c4549b85b8a906334a50f44db496ead5160780b469fe7774d8aa2fa" Oct 08 20:12:30 crc kubenswrapper[4988]: E1008 20:12:30.240349 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nm54f_openshift-machine-config-operator(f328a20b-5283-407e-8a93-2ab2da599f18)\"" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" Oct 08 20:12:43 crc kubenswrapper[4988]: I1008 20:12:43.237551 4988 scope.go:117] "RemoveContainer" containerID="c7dc147e5c4549b85b8a906334a50f44db496ead5160780b469fe7774d8aa2fa" Oct 08 20:12:43 crc kubenswrapper[4988]: E1008 20:12:43.238564 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nm54f_openshift-machine-config-operator(f328a20b-5283-407e-8a93-2ab2da599f18)\"" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" Oct 08 20:12:54 crc kubenswrapper[4988]: I1008 20:12:54.238737 4988 scope.go:117] "RemoveContainer" containerID="c7dc147e5c4549b85b8a906334a50f44db496ead5160780b469fe7774d8aa2fa" Oct 08 20:12:54 crc kubenswrapper[4988]: E1008 20:12:54.240131 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nm54f_openshift-machine-config-operator(f328a20b-5283-407e-8a93-2ab2da599f18)\"" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" Oct 08 20:13:06 crc kubenswrapper[4988]: I1008 20:13:06.239090 4988 scope.go:117] "RemoveContainer" containerID="c7dc147e5c4549b85b8a906334a50f44db496ead5160780b469fe7774d8aa2fa" Oct 08 20:13:06 crc kubenswrapper[4988]: E1008 20:13:06.239982 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nm54f_openshift-machine-config-operator(f328a20b-5283-407e-8a93-2ab2da599f18)\"" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" Oct 08 20:13:13 crc kubenswrapper[4988]: I1008 20:13:13.721193 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-n2smn"] Oct 08 20:13:13 crc kubenswrapper[4988]: E1008 20:13:13.722577 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="491de4c7-9179-4477-a0ad-044978c19ad4" containerName="extract-utilities" Oct 08 20:13:13 crc kubenswrapper[4988]: I1008 20:13:13.722594 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="491de4c7-9179-4477-a0ad-044978c19ad4" containerName="extract-utilities" Oct 08 20:13:13 crc kubenswrapper[4988]: E1008 20:13:13.722636 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="491de4c7-9179-4477-a0ad-044978c19ad4" containerName="extract-content" Oct 08 20:13:13 crc kubenswrapper[4988]: I1008 20:13:13.722644 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="491de4c7-9179-4477-a0ad-044978c19ad4" containerName="extract-content" Oct 08 20:13:13 crc kubenswrapper[4988]: E1008 20:13:13.722691 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="491de4c7-9179-4477-a0ad-044978c19ad4" containerName="registry-server" Oct 08 20:13:13 crc kubenswrapper[4988]: I1008 20:13:13.722699 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="491de4c7-9179-4477-a0ad-044978c19ad4" containerName="registry-server" Oct 08 20:13:13 crc kubenswrapper[4988]: I1008 20:13:13.722988 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="491de4c7-9179-4477-a0ad-044978c19ad4" containerName="registry-server" Oct 08 20:13:13 crc kubenswrapper[4988]: I1008 20:13:13.725073 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-n2smn" Oct 08 20:13:13 crc kubenswrapper[4988]: I1008 20:13:13.735169 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-n2smn"] Oct 08 20:13:13 crc kubenswrapper[4988]: I1008 20:13:13.889779 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-774gx\" (UniqueName: \"kubernetes.io/projected/7b719c42-3547-44a4-ba4d-4929583bd21a-kube-api-access-774gx\") pod \"redhat-marketplace-n2smn\" (UID: \"7b719c42-3547-44a4-ba4d-4929583bd21a\") " pod="openshift-marketplace/redhat-marketplace-n2smn" Oct 08 20:13:13 crc kubenswrapper[4988]: I1008 20:13:13.890198 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7b719c42-3547-44a4-ba4d-4929583bd21a-utilities\") pod \"redhat-marketplace-n2smn\" (UID: \"7b719c42-3547-44a4-ba4d-4929583bd21a\") " pod="openshift-marketplace/redhat-marketplace-n2smn" Oct 08 20:13:13 crc kubenswrapper[4988]: I1008 20:13:13.890350 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7b719c42-3547-44a4-ba4d-4929583bd21a-catalog-content\") pod \"redhat-marketplace-n2smn\" (UID: \"7b719c42-3547-44a4-ba4d-4929583bd21a\") " pod="openshift-marketplace/redhat-marketplace-n2smn" Oct 08 20:13:13 crc kubenswrapper[4988]: I1008 20:13:13.992839 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7b719c42-3547-44a4-ba4d-4929583bd21a-utilities\") pod \"redhat-marketplace-n2smn\" (UID: \"7b719c42-3547-44a4-ba4d-4929583bd21a\") " pod="openshift-marketplace/redhat-marketplace-n2smn" Oct 08 20:13:13 crc kubenswrapper[4988]: I1008 20:13:13.992905 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7b719c42-3547-44a4-ba4d-4929583bd21a-catalog-content\") pod \"redhat-marketplace-n2smn\" (UID: \"7b719c42-3547-44a4-ba4d-4929583bd21a\") " pod="openshift-marketplace/redhat-marketplace-n2smn" Oct 08 20:13:13 crc kubenswrapper[4988]: I1008 20:13:13.993058 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-774gx\" (UniqueName: \"kubernetes.io/projected/7b719c42-3547-44a4-ba4d-4929583bd21a-kube-api-access-774gx\") pod \"redhat-marketplace-n2smn\" (UID: \"7b719c42-3547-44a4-ba4d-4929583bd21a\") " pod="openshift-marketplace/redhat-marketplace-n2smn" Oct 08 20:13:13 crc kubenswrapper[4988]: I1008 20:13:13.993544 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7b719c42-3547-44a4-ba4d-4929583bd21a-utilities\") pod \"redhat-marketplace-n2smn\" (UID: \"7b719c42-3547-44a4-ba4d-4929583bd21a\") " pod="openshift-marketplace/redhat-marketplace-n2smn" Oct 08 20:13:13 crc kubenswrapper[4988]: I1008 20:13:13.993758 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7b719c42-3547-44a4-ba4d-4929583bd21a-catalog-content\") pod \"redhat-marketplace-n2smn\" (UID: \"7b719c42-3547-44a4-ba4d-4929583bd21a\") " pod="openshift-marketplace/redhat-marketplace-n2smn" Oct 08 20:13:14 crc kubenswrapper[4988]: I1008 20:13:14.021122 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-774gx\" (UniqueName: \"kubernetes.io/projected/7b719c42-3547-44a4-ba4d-4929583bd21a-kube-api-access-774gx\") pod \"redhat-marketplace-n2smn\" (UID: \"7b719c42-3547-44a4-ba4d-4929583bd21a\") " pod="openshift-marketplace/redhat-marketplace-n2smn" Oct 08 20:13:14 crc kubenswrapper[4988]: I1008 20:13:14.047003 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-n2smn" Oct 08 20:13:14 crc kubenswrapper[4988]: I1008 20:13:14.601832 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-n2smn"] Oct 08 20:13:14 crc kubenswrapper[4988]: I1008 20:13:14.691306 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-n2smn" event={"ID":"7b719c42-3547-44a4-ba4d-4929583bd21a","Type":"ContainerStarted","Data":"fea3cac48196fde1c225f00ace66e315cc5a5e173ffde7ffc500e2a22b4538d2"} Oct 08 20:13:15 crc kubenswrapper[4988]: I1008 20:13:15.707347 4988 generic.go:334] "Generic (PLEG): container finished" podID="7b719c42-3547-44a4-ba4d-4929583bd21a" containerID="29ef7bad568f8d3620bdafeea217a73dbab236b2da0b6fed9571e1a7c35ad637" exitCode=0 Oct 08 20:13:15 crc kubenswrapper[4988]: I1008 20:13:15.708010 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-n2smn" event={"ID":"7b719c42-3547-44a4-ba4d-4929583bd21a","Type":"ContainerDied","Data":"29ef7bad568f8d3620bdafeea217a73dbab236b2da0b6fed9571e1a7c35ad637"} Oct 08 20:13:16 crc kubenswrapper[4988]: I1008 20:13:16.118972 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-lxm7v"] Oct 08 20:13:16 crc kubenswrapper[4988]: I1008 20:13:16.121694 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-lxm7v" Oct 08 20:13:16 crc kubenswrapper[4988]: I1008 20:13:16.136805 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-lxm7v"] Oct 08 20:13:16 crc kubenswrapper[4988]: I1008 20:13:16.268337 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tk9c5\" (UniqueName: \"kubernetes.io/projected/9edf6678-96ef-43aa-b2e0-314948fd9283-kube-api-access-tk9c5\") pod \"community-operators-lxm7v\" (UID: \"9edf6678-96ef-43aa-b2e0-314948fd9283\") " pod="openshift-marketplace/community-operators-lxm7v" Oct 08 20:13:16 crc kubenswrapper[4988]: I1008 20:13:16.268416 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9edf6678-96ef-43aa-b2e0-314948fd9283-utilities\") pod \"community-operators-lxm7v\" (UID: \"9edf6678-96ef-43aa-b2e0-314948fd9283\") " pod="openshift-marketplace/community-operators-lxm7v" Oct 08 20:13:16 crc kubenswrapper[4988]: I1008 20:13:16.268772 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9edf6678-96ef-43aa-b2e0-314948fd9283-catalog-content\") pod \"community-operators-lxm7v\" (UID: \"9edf6678-96ef-43aa-b2e0-314948fd9283\") " pod="openshift-marketplace/community-operators-lxm7v" Oct 08 20:13:16 crc kubenswrapper[4988]: I1008 20:13:16.370732 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9edf6678-96ef-43aa-b2e0-314948fd9283-catalog-content\") pod \"community-operators-lxm7v\" (UID: \"9edf6678-96ef-43aa-b2e0-314948fd9283\") " pod="openshift-marketplace/community-operators-lxm7v" Oct 08 20:13:16 crc kubenswrapper[4988]: I1008 20:13:16.370866 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tk9c5\" (UniqueName: \"kubernetes.io/projected/9edf6678-96ef-43aa-b2e0-314948fd9283-kube-api-access-tk9c5\") pod \"community-operators-lxm7v\" (UID: \"9edf6678-96ef-43aa-b2e0-314948fd9283\") " pod="openshift-marketplace/community-operators-lxm7v" Oct 08 20:13:16 crc kubenswrapper[4988]: I1008 20:13:16.370901 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9edf6678-96ef-43aa-b2e0-314948fd9283-utilities\") pod \"community-operators-lxm7v\" (UID: \"9edf6678-96ef-43aa-b2e0-314948fd9283\") " pod="openshift-marketplace/community-operators-lxm7v" Oct 08 20:13:16 crc kubenswrapper[4988]: I1008 20:13:16.371406 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9edf6678-96ef-43aa-b2e0-314948fd9283-utilities\") pod \"community-operators-lxm7v\" (UID: \"9edf6678-96ef-43aa-b2e0-314948fd9283\") " pod="openshift-marketplace/community-operators-lxm7v" Oct 08 20:13:16 crc kubenswrapper[4988]: I1008 20:13:16.371504 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9edf6678-96ef-43aa-b2e0-314948fd9283-catalog-content\") pod \"community-operators-lxm7v\" (UID: \"9edf6678-96ef-43aa-b2e0-314948fd9283\") " pod="openshift-marketplace/community-operators-lxm7v" Oct 08 20:13:16 crc kubenswrapper[4988]: I1008 20:13:16.389481 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tk9c5\" (UniqueName: \"kubernetes.io/projected/9edf6678-96ef-43aa-b2e0-314948fd9283-kube-api-access-tk9c5\") pod \"community-operators-lxm7v\" (UID: \"9edf6678-96ef-43aa-b2e0-314948fd9283\") " pod="openshift-marketplace/community-operators-lxm7v" Oct 08 20:13:16 crc kubenswrapper[4988]: I1008 20:13:16.451070 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-lxm7v" Oct 08 20:13:16 crc kubenswrapper[4988]: I1008 20:13:16.986053 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-lxm7v"] Oct 08 20:13:16 crc kubenswrapper[4988]: W1008 20:13:16.993119 4988 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9edf6678_96ef_43aa_b2e0_314948fd9283.slice/crio-e623eaf60f18320b5563e72823bb7daa5d0c2b217f8178a02d7ff82e73e12fb6 WatchSource:0}: Error finding container e623eaf60f18320b5563e72823bb7daa5d0c2b217f8178a02d7ff82e73e12fb6: Status 404 returned error can't find the container with id e623eaf60f18320b5563e72823bb7daa5d0c2b217f8178a02d7ff82e73e12fb6 Oct 08 20:13:17 crc kubenswrapper[4988]: I1008 20:13:17.728775 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-n2smn" event={"ID":"7b719c42-3547-44a4-ba4d-4929583bd21a","Type":"ContainerStarted","Data":"f703ea98652895c8114f97e86aba56640caf428370938254536a597fcbae528c"} Oct 08 20:13:17 crc kubenswrapper[4988]: I1008 20:13:17.730933 4988 generic.go:334] "Generic (PLEG): container finished" podID="9edf6678-96ef-43aa-b2e0-314948fd9283" containerID="41814fefb97f4596d2df83358e697fef29d469c88686198163995d3f9ccefd76" exitCode=0 Oct 08 20:13:17 crc kubenswrapper[4988]: I1008 20:13:17.730992 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-lxm7v" event={"ID":"9edf6678-96ef-43aa-b2e0-314948fd9283","Type":"ContainerDied","Data":"41814fefb97f4596d2df83358e697fef29d469c88686198163995d3f9ccefd76"} Oct 08 20:13:17 crc kubenswrapper[4988]: I1008 20:13:17.731023 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-lxm7v" event={"ID":"9edf6678-96ef-43aa-b2e0-314948fd9283","Type":"ContainerStarted","Data":"e623eaf60f18320b5563e72823bb7daa5d0c2b217f8178a02d7ff82e73e12fb6"} Oct 08 20:13:18 crc kubenswrapper[4988]: I1008 20:13:18.743671 4988 generic.go:334] "Generic (PLEG): container finished" podID="7b719c42-3547-44a4-ba4d-4929583bd21a" containerID="f703ea98652895c8114f97e86aba56640caf428370938254536a597fcbae528c" exitCode=0 Oct 08 20:13:18 crc kubenswrapper[4988]: I1008 20:13:18.743781 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-n2smn" event={"ID":"7b719c42-3547-44a4-ba4d-4929583bd21a","Type":"ContainerDied","Data":"f703ea98652895c8114f97e86aba56640caf428370938254536a597fcbae528c"} Oct 08 20:13:19 crc kubenswrapper[4988]: I1008 20:13:19.757241 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-lxm7v" event={"ID":"9edf6678-96ef-43aa-b2e0-314948fd9283","Type":"ContainerStarted","Data":"52f3f34820a268574ef0ff44fdef5d2f3b65ebae4be6574f13dd38ffdc8c3334"} Oct 08 20:13:20 crc kubenswrapper[4988]: I1008 20:13:20.778381 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-n2smn" event={"ID":"7b719c42-3547-44a4-ba4d-4929583bd21a","Type":"ContainerStarted","Data":"49aa3f4bf62627d261f098e8e6c5d8d1e5f1b75ed8449c2fd82650a430432c8b"} Oct 08 20:13:20 crc kubenswrapper[4988]: I1008 20:13:20.813320 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-n2smn" podStartSLOduration=3.644302942 podStartE2EDuration="7.81330487s" podCreationTimestamp="2025-10-08 20:13:13 +0000 UTC" firstStartedPulling="2025-10-08 20:13:15.713628272 +0000 UTC m=+7341.163471052" lastFinishedPulling="2025-10-08 20:13:19.8826302 +0000 UTC m=+7345.332472980" observedRunningTime="2025-10-08 20:13:20.799373116 +0000 UTC m=+7346.249215886" watchObservedRunningTime="2025-10-08 20:13:20.81330487 +0000 UTC m=+7346.263147630" Oct 08 20:13:21 crc kubenswrapper[4988]: I1008 20:13:21.238335 4988 scope.go:117] "RemoveContainer" containerID="c7dc147e5c4549b85b8a906334a50f44db496ead5160780b469fe7774d8aa2fa" Oct 08 20:13:21 crc kubenswrapper[4988]: E1008 20:13:21.238997 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nm54f_openshift-machine-config-operator(f328a20b-5283-407e-8a93-2ab2da599f18)\"" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" Oct 08 20:13:23 crc kubenswrapper[4988]: I1008 20:13:23.811054 4988 generic.go:334] "Generic (PLEG): container finished" podID="9edf6678-96ef-43aa-b2e0-314948fd9283" containerID="52f3f34820a268574ef0ff44fdef5d2f3b65ebae4be6574f13dd38ffdc8c3334" exitCode=0 Oct 08 20:13:23 crc kubenswrapper[4988]: I1008 20:13:23.811125 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-lxm7v" event={"ID":"9edf6678-96ef-43aa-b2e0-314948fd9283","Type":"ContainerDied","Data":"52f3f34820a268574ef0ff44fdef5d2f3b65ebae4be6574f13dd38ffdc8c3334"} Oct 08 20:13:24 crc kubenswrapper[4988]: I1008 20:13:24.048039 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-n2smn" Oct 08 20:13:24 crc kubenswrapper[4988]: I1008 20:13:24.050100 4988 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-n2smn" Oct 08 20:13:24 crc kubenswrapper[4988]: I1008 20:13:24.109046 4988 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-n2smn" Oct 08 20:13:25 crc kubenswrapper[4988]: I1008 20:13:25.901217 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-n2smn" Oct 08 20:13:26 crc kubenswrapper[4988]: I1008 20:13:26.305089 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-n2smn"] Oct 08 20:13:27 crc kubenswrapper[4988]: I1008 20:13:27.848270 4988 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-n2smn" podUID="7b719c42-3547-44a4-ba4d-4929583bd21a" containerName="registry-server" containerID="cri-o://49aa3f4bf62627d261f098e8e6c5d8d1e5f1b75ed8449c2fd82650a430432c8b" gracePeriod=2 Oct 08 20:13:28 crc kubenswrapper[4988]: I1008 20:13:28.587831 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-n2smn" Oct 08 20:13:28 crc kubenswrapper[4988]: I1008 20:13:28.730260 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7b719c42-3547-44a4-ba4d-4929583bd21a-utilities\") pod \"7b719c42-3547-44a4-ba4d-4929583bd21a\" (UID: \"7b719c42-3547-44a4-ba4d-4929583bd21a\") " Oct 08 20:13:28 crc kubenswrapper[4988]: I1008 20:13:28.730746 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7b719c42-3547-44a4-ba4d-4929583bd21a-catalog-content\") pod \"7b719c42-3547-44a4-ba4d-4929583bd21a\" (UID: \"7b719c42-3547-44a4-ba4d-4929583bd21a\") " Oct 08 20:13:28 crc kubenswrapper[4988]: I1008 20:13:28.730860 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-774gx\" (UniqueName: \"kubernetes.io/projected/7b719c42-3547-44a4-ba4d-4929583bd21a-kube-api-access-774gx\") pod \"7b719c42-3547-44a4-ba4d-4929583bd21a\" (UID: \"7b719c42-3547-44a4-ba4d-4929583bd21a\") " Oct 08 20:13:28 crc kubenswrapper[4988]: I1008 20:13:28.731225 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7b719c42-3547-44a4-ba4d-4929583bd21a-utilities" (OuterVolumeSpecName: "utilities") pod "7b719c42-3547-44a4-ba4d-4929583bd21a" (UID: "7b719c42-3547-44a4-ba4d-4929583bd21a"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 20:13:28 crc kubenswrapper[4988]: I1008 20:13:28.731452 4988 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7b719c42-3547-44a4-ba4d-4929583bd21a-utilities\") on node \"crc\" DevicePath \"\"" Oct 08 20:13:28 crc kubenswrapper[4988]: I1008 20:13:28.737623 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7b719c42-3547-44a4-ba4d-4929583bd21a-kube-api-access-774gx" (OuterVolumeSpecName: "kube-api-access-774gx") pod "7b719c42-3547-44a4-ba4d-4929583bd21a" (UID: "7b719c42-3547-44a4-ba4d-4929583bd21a"). InnerVolumeSpecName "kube-api-access-774gx". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 20:13:28 crc kubenswrapper[4988]: I1008 20:13:28.743465 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7b719c42-3547-44a4-ba4d-4929583bd21a-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "7b719c42-3547-44a4-ba4d-4929583bd21a" (UID: "7b719c42-3547-44a4-ba4d-4929583bd21a"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 20:13:28 crc kubenswrapper[4988]: I1008 20:13:28.833852 4988 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-774gx\" (UniqueName: \"kubernetes.io/projected/7b719c42-3547-44a4-ba4d-4929583bd21a-kube-api-access-774gx\") on node \"crc\" DevicePath \"\"" Oct 08 20:13:28 crc kubenswrapper[4988]: I1008 20:13:28.833901 4988 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7b719c42-3547-44a4-ba4d-4929583bd21a-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 08 20:13:28 crc kubenswrapper[4988]: I1008 20:13:28.862308 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-lxm7v" event={"ID":"9edf6678-96ef-43aa-b2e0-314948fd9283","Type":"ContainerStarted","Data":"262680c5f0a75394a90edadbeb2c57a80307246940f164dc7e66a28275a48456"} Oct 08 20:13:28 crc kubenswrapper[4988]: I1008 20:13:28.866788 4988 generic.go:334] "Generic (PLEG): container finished" podID="7b719c42-3547-44a4-ba4d-4929583bd21a" containerID="49aa3f4bf62627d261f098e8e6c5d8d1e5f1b75ed8449c2fd82650a430432c8b" exitCode=0 Oct 08 20:13:28 crc kubenswrapper[4988]: I1008 20:13:28.866846 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-n2smn" event={"ID":"7b719c42-3547-44a4-ba4d-4929583bd21a","Type":"ContainerDied","Data":"49aa3f4bf62627d261f098e8e6c5d8d1e5f1b75ed8449c2fd82650a430432c8b"} Oct 08 20:13:28 crc kubenswrapper[4988]: I1008 20:13:28.866914 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-n2smn" event={"ID":"7b719c42-3547-44a4-ba4d-4929583bd21a","Type":"ContainerDied","Data":"fea3cac48196fde1c225f00ace66e315cc5a5e173ffde7ffc500e2a22b4538d2"} Oct 08 20:13:28 crc kubenswrapper[4988]: I1008 20:13:28.866941 4988 scope.go:117] "RemoveContainer" containerID="49aa3f4bf62627d261f098e8e6c5d8d1e5f1b75ed8449c2fd82650a430432c8b" Oct 08 20:13:28 crc kubenswrapper[4988]: I1008 20:13:28.867163 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-n2smn" Oct 08 20:13:28 crc kubenswrapper[4988]: I1008 20:13:28.896044 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-lxm7v" podStartSLOduration=2.904168207 podStartE2EDuration="12.896014418s" podCreationTimestamp="2025-10-08 20:13:16 +0000 UTC" firstStartedPulling="2025-10-08 20:13:17.733073095 +0000 UTC m=+7343.182915865" lastFinishedPulling="2025-10-08 20:13:27.724919266 +0000 UTC m=+7353.174762076" observedRunningTime="2025-10-08 20:13:28.884061148 +0000 UTC m=+7354.333903918" watchObservedRunningTime="2025-10-08 20:13:28.896014418 +0000 UTC m=+7354.345857188" Oct 08 20:13:28 crc kubenswrapper[4988]: I1008 20:13:28.900634 4988 scope.go:117] "RemoveContainer" containerID="f703ea98652895c8114f97e86aba56640caf428370938254536a597fcbae528c" Oct 08 20:13:28 crc kubenswrapper[4988]: I1008 20:13:28.918227 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-n2smn"] Oct 08 20:13:28 crc kubenswrapper[4988]: I1008 20:13:28.926688 4988 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-n2smn"] Oct 08 20:13:28 crc kubenswrapper[4988]: I1008 20:13:28.936845 4988 scope.go:117] "RemoveContainer" containerID="29ef7bad568f8d3620bdafeea217a73dbab236b2da0b6fed9571e1a7c35ad637" Oct 08 20:13:28 crc kubenswrapper[4988]: I1008 20:13:28.989544 4988 scope.go:117] "RemoveContainer" containerID="49aa3f4bf62627d261f098e8e6c5d8d1e5f1b75ed8449c2fd82650a430432c8b" Oct 08 20:13:28 crc kubenswrapper[4988]: E1008 20:13:28.998526 4988 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"49aa3f4bf62627d261f098e8e6c5d8d1e5f1b75ed8449c2fd82650a430432c8b\": container with ID starting with 49aa3f4bf62627d261f098e8e6c5d8d1e5f1b75ed8449c2fd82650a430432c8b not found: ID does not exist" containerID="49aa3f4bf62627d261f098e8e6c5d8d1e5f1b75ed8449c2fd82650a430432c8b" Oct 08 20:13:28 crc kubenswrapper[4988]: I1008 20:13:28.998575 4988 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"49aa3f4bf62627d261f098e8e6c5d8d1e5f1b75ed8449c2fd82650a430432c8b"} err="failed to get container status \"49aa3f4bf62627d261f098e8e6c5d8d1e5f1b75ed8449c2fd82650a430432c8b\": rpc error: code = NotFound desc = could not find container \"49aa3f4bf62627d261f098e8e6c5d8d1e5f1b75ed8449c2fd82650a430432c8b\": container with ID starting with 49aa3f4bf62627d261f098e8e6c5d8d1e5f1b75ed8449c2fd82650a430432c8b not found: ID does not exist" Oct 08 20:13:28 crc kubenswrapper[4988]: I1008 20:13:28.998610 4988 scope.go:117] "RemoveContainer" containerID="f703ea98652895c8114f97e86aba56640caf428370938254536a597fcbae528c" Oct 08 20:13:28 crc kubenswrapper[4988]: E1008 20:13:28.999103 4988 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f703ea98652895c8114f97e86aba56640caf428370938254536a597fcbae528c\": container with ID starting with f703ea98652895c8114f97e86aba56640caf428370938254536a597fcbae528c not found: ID does not exist" containerID="f703ea98652895c8114f97e86aba56640caf428370938254536a597fcbae528c" Oct 08 20:13:28 crc kubenswrapper[4988]: I1008 20:13:28.999128 4988 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f703ea98652895c8114f97e86aba56640caf428370938254536a597fcbae528c"} err="failed to get container status \"f703ea98652895c8114f97e86aba56640caf428370938254536a597fcbae528c\": rpc error: code = NotFound desc = could not find container \"f703ea98652895c8114f97e86aba56640caf428370938254536a597fcbae528c\": container with ID starting with f703ea98652895c8114f97e86aba56640caf428370938254536a597fcbae528c not found: ID does not exist" Oct 08 20:13:28 crc kubenswrapper[4988]: I1008 20:13:28.999140 4988 scope.go:117] "RemoveContainer" containerID="29ef7bad568f8d3620bdafeea217a73dbab236b2da0b6fed9571e1a7c35ad637" Oct 08 20:13:28 crc kubenswrapper[4988]: E1008 20:13:28.999343 4988 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"29ef7bad568f8d3620bdafeea217a73dbab236b2da0b6fed9571e1a7c35ad637\": container with ID starting with 29ef7bad568f8d3620bdafeea217a73dbab236b2da0b6fed9571e1a7c35ad637 not found: ID does not exist" containerID="29ef7bad568f8d3620bdafeea217a73dbab236b2da0b6fed9571e1a7c35ad637" Oct 08 20:13:28 crc kubenswrapper[4988]: I1008 20:13:28.999374 4988 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"29ef7bad568f8d3620bdafeea217a73dbab236b2da0b6fed9571e1a7c35ad637"} err="failed to get container status \"29ef7bad568f8d3620bdafeea217a73dbab236b2da0b6fed9571e1a7c35ad637\": rpc error: code = NotFound desc = could not find container \"29ef7bad568f8d3620bdafeea217a73dbab236b2da0b6fed9571e1a7c35ad637\": container with ID starting with 29ef7bad568f8d3620bdafeea217a73dbab236b2da0b6fed9571e1a7c35ad637 not found: ID does not exist" Oct 08 20:13:29 crc kubenswrapper[4988]: I1008 20:13:29.261968 4988 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7b719c42-3547-44a4-ba4d-4929583bd21a" path="/var/lib/kubelet/pods/7b719c42-3547-44a4-ba4d-4929583bd21a/volumes" Oct 08 20:13:33 crc kubenswrapper[4988]: I1008 20:13:33.237827 4988 scope.go:117] "RemoveContainer" containerID="c7dc147e5c4549b85b8a906334a50f44db496ead5160780b469fe7774d8aa2fa" Oct 08 20:13:33 crc kubenswrapper[4988]: E1008 20:13:33.238869 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nm54f_openshift-machine-config-operator(f328a20b-5283-407e-8a93-2ab2da599f18)\"" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" Oct 08 20:13:36 crc kubenswrapper[4988]: I1008 20:13:36.451604 4988 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-lxm7v" Oct 08 20:13:36 crc kubenswrapper[4988]: I1008 20:13:36.452011 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-lxm7v" Oct 08 20:13:36 crc kubenswrapper[4988]: I1008 20:13:36.525698 4988 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-lxm7v" Oct 08 20:13:37 crc kubenswrapper[4988]: I1008 20:13:37.033596 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-lxm7v" Oct 08 20:13:37 crc kubenswrapper[4988]: I1008 20:13:37.099002 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-lxm7v"] Oct 08 20:13:38 crc kubenswrapper[4988]: I1008 20:13:38.975240 4988 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-lxm7v" podUID="9edf6678-96ef-43aa-b2e0-314948fd9283" containerName="registry-server" containerID="cri-o://262680c5f0a75394a90edadbeb2c57a80307246940f164dc7e66a28275a48456" gracePeriod=2 Oct 08 20:13:39 crc kubenswrapper[4988]: I1008 20:13:39.622800 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-lxm7v" Oct 08 20:13:39 crc kubenswrapper[4988]: I1008 20:13:39.810329 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9edf6678-96ef-43aa-b2e0-314948fd9283-catalog-content\") pod \"9edf6678-96ef-43aa-b2e0-314948fd9283\" (UID: \"9edf6678-96ef-43aa-b2e0-314948fd9283\") " Oct 08 20:13:39 crc kubenswrapper[4988]: I1008 20:13:39.810554 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tk9c5\" (UniqueName: \"kubernetes.io/projected/9edf6678-96ef-43aa-b2e0-314948fd9283-kube-api-access-tk9c5\") pod \"9edf6678-96ef-43aa-b2e0-314948fd9283\" (UID: \"9edf6678-96ef-43aa-b2e0-314948fd9283\") " Oct 08 20:13:39 crc kubenswrapper[4988]: I1008 20:13:39.810718 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9edf6678-96ef-43aa-b2e0-314948fd9283-utilities\") pod \"9edf6678-96ef-43aa-b2e0-314948fd9283\" (UID: \"9edf6678-96ef-43aa-b2e0-314948fd9283\") " Oct 08 20:13:39 crc kubenswrapper[4988]: I1008 20:13:39.812229 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9edf6678-96ef-43aa-b2e0-314948fd9283-utilities" (OuterVolumeSpecName: "utilities") pod "9edf6678-96ef-43aa-b2e0-314948fd9283" (UID: "9edf6678-96ef-43aa-b2e0-314948fd9283"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 20:13:39 crc kubenswrapper[4988]: I1008 20:13:39.847607 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9edf6678-96ef-43aa-b2e0-314948fd9283-kube-api-access-tk9c5" (OuterVolumeSpecName: "kube-api-access-tk9c5") pod "9edf6678-96ef-43aa-b2e0-314948fd9283" (UID: "9edf6678-96ef-43aa-b2e0-314948fd9283"). InnerVolumeSpecName "kube-api-access-tk9c5". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 20:13:39 crc kubenswrapper[4988]: I1008 20:13:39.884452 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9edf6678-96ef-43aa-b2e0-314948fd9283-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "9edf6678-96ef-43aa-b2e0-314948fd9283" (UID: "9edf6678-96ef-43aa-b2e0-314948fd9283"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 20:13:39 crc kubenswrapper[4988]: I1008 20:13:39.914428 4988 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tk9c5\" (UniqueName: \"kubernetes.io/projected/9edf6678-96ef-43aa-b2e0-314948fd9283-kube-api-access-tk9c5\") on node \"crc\" DevicePath \"\"" Oct 08 20:13:39 crc kubenswrapper[4988]: I1008 20:13:39.914473 4988 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9edf6678-96ef-43aa-b2e0-314948fd9283-utilities\") on node \"crc\" DevicePath \"\"" Oct 08 20:13:39 crc kubenswrapper[4988]: I1008 20:13:39.914490 4988 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9edf6678-96ef-43aa-b2e0-314948fd9283-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 08 20:13:39 crc kubenswrapper[4988]: I1008 20:13:39.999417 4988 generic.go:334] "Generic (PLEG): container finished" podID="9edf6678-96ef-43aa-b2e0-314948fd9283" containerID="262680c5f0a75394a90edadbeb2c57a80307246940f164dc7e66a28275a48456" exitCode=0 Oct 08 20:13:39 crc kubenswrapper[4988]: I1008 20:13:39.999468 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-lxm7v" event={"ID":"9edf6678-96ef-43aa-b2e0-314948fd9283","Type":"ContainerDied","Data":"262680c5f0a75394a90edadbeb2c57a80307246940f164dc7e66a28275a48456"} Oct 08 20:13:40 crc kubenswrapper[4988]: I1008 20:13:39.999501 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-lxm7v" event={"ID":"9edf6678-96ef-43aa-b2e0-314948fd9283","Type":"ContainerDied","Data":"e623eaf60f18320b5563e72823bb7daa5d0c2b217f8178a02d7ff82e73e12fb6"} Oct 08 20:13:40 crc kubenswrapper[4988]: I1008 20:13:39.999524 4988 scope.go:117] "RemoveContainer" containerID="262680c5f0a75394a90edadbeb2c57a80307246940f164dc7e66a28275a48456" Oct 08 20:13:40 crc kubenswrapper[4988]: I1008 20:13:39.999542 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-lxm7v" Oct 08 20:13:40 crc kubenswrapper[4988]: I1008 20:13:40.041211 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-lxm7v"] Oct 08 20:13:40 crc kubenswrapper[4988]: I1008 20:13:40.043040 4988 scope.go:117] "RemoveContainer" containerID="52f3f34820a268574ef0ff44fdef5d2f3b65ebae4be6574f13dd38ffdc8c3334" Oct 08 20:13:40 crc kubenswrapper[4988]: I1008 20:13:40.051575 4988 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-lxm7v"] Oct 08 20:13:40 crc kubenswrapper[4988]: I1008 20:13:40.083755 4988 scope.go:117] "RemoveContainer" containerID="41814fefb97f4596d2df83358e697fef29d469c88686198163995d3f9ccefd76" Oct 08 20:13:40 crc kubenswrapper[4988]: I1008 20:13:40.131022 4988 scope.go:117] "RemoveContainer" containerID="262680c5f0a75394a90edadbeb2c57a80307246940f164dc7e66a28275a48456" Oct 08 20:13:40 crc kubenswrapper[4988]: E1008 20:13:40.131660 4988 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"262680c5f0a75394a90edadbeb2c57a80307246940f164dc7e66a28275a48456\": container with ID starting with 262680c5f0a75394a90edadbeb2c57a80307246940f164dc7e66a28275a48456 not found: ID does not exist" containerID="262680c5f0a75394a90edadbeb2c57a80307246940f164dc7e66a28275a48456" Oct 08 20:13:40 crc kubenswrapper[4988]: I1008 20:13:40.131834 4988 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"262680c5f0a75394a90edadbeb2c57a80307246940f164dc7e66a28275a48456"} err="failed to get container status \"262680c5f0a75394a90edadbeb2c57a80307246940f164dc7e66a28275a48456\": rpc error: code = NotFound desc = could not find container \"262680c5f0a75394a90edadbeb2c57a80307246940f164dc7e66a28275a48456\": container with ID starting with 262680c5f0a75394a90edadbeb2c57a80307246940f164dc7e66a28275a48456 not found: ID does not exist" Oct 08 20:13:40 crc kubenswrapper[4988]: I1008 20:13:40.131995 4988 scope.go:117] "RemoveContainer" containerID="52f3f34820a268574ef0ff44fdef5d2f3b65ebae4be6574f13dd38ffdc8c3334" Oct 08 20:13:40 crc kubenswrapper[4988]: E1008 20:13:40.132986 4988 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"52f3f34820a268574ef0ff44fdef5d2f3b65ebae4be6574f13dd38ffdc8c3334\": container with ID starting with 52f3f34820a268574ef0ff44fdef5d2f3b65ebae4be6574f13dd38ffdc8c3334 not found: ID does not exist" containerID="52f3f34820a268574ef0ff44fdef5d2f3b65ebae4be6574f13dd38ffdc8c3334" Oct 08 20:13:40 crc kubenswrapper[4988]: I1008 20:13:40.133038 4988 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"52f3f34820a268574ef0ff44fdef5d2f3b65ebae4be6574f13dd38ffdc8c3334"} err="failed to get container status \"52f3f34820a268574ef0ff44fdef5d2f3b65ebae4be6574f13dd38ffdc8c3334\": rpc error: code = NotFound desc = could not find container \"52f3f34820a268574ef0ff44fdef5d2f3b65ebae4be6574f13dd38ffdc8c3334\": container with ID starting with 52f3f34820a268574ef0ff44fdef5d2f3b65ebae4be6574f13dd38ffdc8c3334 not found: ID does not exist" Oct 08 20:13:40 crc kubenswrapper[4988]: I1008 20:13:40.133076 4988 scope.go:117] "RemoveContainer" containerID="41814fefb97f4596d2df83358e697fef29d469c88686198163995d3f9ccefd76" Oct 08 20:13:40 crc kubenswrapper[4988]: E1008 20:13:40.133483 4988 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"41814fefb97f4596d2df83358e697fef29d469c88686198163995d3f9ccefd76\": container with ID starting with 41814fefb97f4596d2df83358e697fef29d469c88686198163995d3f9ccefd76 not found: ID does not exist" containerID="41814fefb97f4596d2df83358e697fef29d469c88686198163995d3f9ccefd76" Oct 08 20:13:40 crc kubenswrapper[4988]: I1008 20:13:40.133522 4988 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"41814fefb97f4596d2df83358e697fef29d469c88686198163995d3f9ccefd76"} err="failed to get container status \"41814fefb97f4596d2df83358e697fef29d469c88686198163995d3f9ccefd76\": rpc error: code = NotFound desc = could not find container \"41814fefb97f4596d2df83358e697fef29d469c88686198163995d3f9ccefd76\": container with ID starting with 41814fefb97f4596d2df83358e697fef29d469c88686198163995d3f9ccefd76 not found: ID does not exist" Oct 08 20:13:41 crc kubenswrapper[4988]: I1008 20:13:41.258938 4988 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9edf6678-96ef-43aa-b2e0-314948fd9283" path="/var/lib/kubelet/pods/9edf6678-96ef-43aa-b2e0-314948fd9283/volumes" Oct 08 20:13:48 crc kubenswrapper[4988]: I1008 20:13:48.238230 4988 scope.go:117] "RemoveContainer" containerID="c7dc147e5c4549b85b8a906334a50f44db496ead5160780b469fe7774d8aa2fa" Oct 08 20:13:48 crc kubenswrapper[4988]: E1008 20:13:48.239243 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nm54f_openshift-machine-config-operator(f328a20b-5283-407e-8a93-2ab2da599f18)\"" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" Oct 08 20:13:56 crc kubenswrapper[4988]: I1008 20:13:56.226935 4988 generic.go:334] "Generic (PLEG): container finished" podID="7dd8825a-ca2d-4361-a6e0-cf8a68ba8078" containerID="2670b20aeeba35ee5c9d7282e9e1548b7b36e66ec32c3f84e10edf683b879b73" exitCode=0 Oct 08 20:13:56 crc kubenswrapper[4988]: I1008 20:13:56.227005 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-2xbjd" event={"ID":"7dd8825a-ca2d-4361-a6e0-cf8a68ba8078","Type":"ContainerDied","Data":"2670b20aeeba35ee5c9d7282e9e1548b7b36e66ec32c3f84e10edf683b879b73"} Oct 08 20:13:57 crc kubenswrapper[4988]: I1008 20:13:57.861235 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-2xbjd" Oct 08 20:13:57 crc kubenswrapper[4988]: I1008 20:13:57.953331 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7dd8825a-ca2d-4361-a6e0-cf8a68ba8078-ssh-key\") pod \"7dd8825a-ca2d-4361-a6e0-cf8a68ba8078\" (UID: \"7dd8825a-ca2d-4361-a6e0-cf8a68ba8078\") " Oct 08 20:13:57 crc kubenswrapper[4988]: I1008 20:13:57.953570 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mp2pc\" (UniqueName: \"kubernetes.io/projected/7dd8825a-ca2d-4361-a6e0-cf8a68ba8078-kube-api-access-mp2pc\") pod \"7dd8825a-ca2d-4361-a6e0-cf8a68ba8078\" (UID: \"7dd8825a-ca2d-4361-a6e0-cf8a68ba8078\") " Oct 08 20:13:57 crc kubenswrapper[4988]: I1008 20:13:57.953729 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7dd8825a-ca2d-4361-a6e0-cf8a68ba8078-inventory\") pod \"7dd8825a-ca2d-4361-a6e0-cf8a68ba8078\" (UID: \"7dd8825a-ca2d-4361-a6e0-cf8a68ba8078\") " Oct 08 20:13:57 crc kubenswrapper[4988]: I1008 20:13:57.953782 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tripleo-cleanup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7dd8825a-ca2d-4361-a6e0-cf8a68ba8078-tripleo-cleanup-combined-ca-bundle\") pod \"7dd8825a-ca2d-4361-a6e0-cf8a68ba8078\" (UID: \"7dd8825a-ca2d-4361-a6e0-cf8a68ba8078\") " Oct 08 20:13:57 crc kubenswrapper[4988]: I1008 20:13:57.961719 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7dd8825a-ca2d-4361-a6e0-cf8a68ba8078-kube-api-access-mp2pc" (OuterVolumeSpecName: "kube-api-access-mp2pc") pod "7dd8825a-ca2d-4361-a6e0-cf8a68ba8078" (UID: "7dd8825a-ca2d-4361-a6e0-cf8a68ba8078"). InnerVolumeSpecName "kube-api-access-mp2pc". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 20:13:57 crc kubenswrapper[4988]: I1008 20:13:57.963263 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7dd8825a-ca2d-4361-a6e0-cf8a68ba8078-tripleo-cleanup-combined-ca-bundle" (OuterVolumeSpecName: "tripleo-cleanup-combined-ca-bundle") pod "7dd8825a-ca2d-4361-a6e0-cf8a68ba8078" (UID: "7dd8825a-ca2d-4361-a6e0-cf8a68ba8078"). InnerVolumeSpecName "tripleo-cleanup-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 20:13:57 crc kubenswrapper[4988]: I1008 20:13:57.989779 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7dd8825a-ca2d-4361-a6e0-cf8a68ba8078-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "7dd8825a-ca2d-4361-a6e0-cf8a68ba8078" (UID: "7dd8825a-ca2d-4361-a6e0-cf8a68ba8078"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 20:13:57 crc kubenswrapper[4988]: I1008 20:13:57.997663 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7dd8825a-ca2d-4361-a6e0-cf8a68ba8078-inventory" (OuterVolumeSpecName: "inventory") pod "7dd8825a-ca2d-4361-a6e0-cf8a68ba8078" (UID: "7dd8825a-ca2d-4361-a6e0-cf8a68ba8078"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 20:13:58 crc kubenswrapper[4988]: I1008 20:13:58.057873 4988 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mp2pc\" (UniqueName: \"kubernetes.io/projected/7dd8825a-ca2d-4361-a6e0-cf8a68ba8078-kube-api-access-mp2pc\") on node \"crc\" DevicePath \"\"" Oct 08 20:13:58 crc kubenswrapper[4988]: I1008 20:13:58.057924 4988 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7dd8825a-ca2d-4361-a6e0-cf8a68ba8078-inventory\") on node \"crc\" DevicePath \"\"" Oct 08 20:13:58 crc kubenswrapper[4988]: I1008 20:13:58.057943 4988 reconciler_common.go:293] "Volume detached for volume \"tripleo-cleanup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7dd8825a-ca2d-4361-a6e0-cf8a68ba8078-tripleo-cleanup-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 20:13:58 crc kubenswrapper[4988]: I1008 20:13:58.057959 4988 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7dd8825a-ca2d-4361-a6e0-cf8a68ba8078-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 08 20:13:58 crc kubenswrapper[4988]: I1008 20:13:58.254845 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-2xbjd" event={"ID":"7dd8825a-ca2d-4361-a6e0-cf8a68ba8078","Type":"ContainerDied","Data":"5f2129db33f905fc4ee9ecabf8ea2a7e549b0e7c471ae49068a782f871eaf3ed"} Oct 08 20:13:58 crc kubenswrapper[4988]: I1008 20:13:58.254895 4988 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5f2129db33f905fc4ee9ecabf8ea2a7e549b0e7c471ae49068a782f871eaf3ed" Oct 08 20:13:58 crc kubenswrapper[4988]: I1008 20:13:58.254887 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-2xbjd" Oct 08 20:14:03 crc kubenswrapper[4988]: I1008 20:14:03.239109 4988 scope.go:117] "RemoveContainer" containerID="c7dc147e5c4549b85b8a906334a50f44db496ead5160780b469fe7774d8aa2fa" Oct 08 20:14:03 crc kubenswrapper[4988]: E1008 20:14:03.239986 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nm54f_openshift-machine-config-operator(f328a20b-5283-407e-8a93-2ab2da599f18)\"" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" Oct 08 20:14:08 crc kubenswrapper[4988]: I1008 20:14:08.797740 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/bootstrap-openstack-openstack-cell1-t5z85"] Oct 08 20:14:08 crc kubenswrapper[4988]: E1008 20:14:08.801722 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7dd8825a-ca2d-4361-a6e0-cf8a68ba8078" containerName="tripleo-cleanup-tripleo-cleanup-openstack-cell1" Oct 08 20:14:08 crc kubenswrapper[4988]: I1008 20:14:08.801759 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="7dd8825a-ca2d-4361-a6e0-cf8a68ba8078" containerName="tripleo-cleanup-tripleo-cleanup-openstack-cell1" Oct 08 20:14:08 crc kubenswrapper[4988]: E1008 20:14:08.801817 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7b719c42-3547-44a4-ba4d-4929583bd21a" containerName="extract-content" Oct 08 20:14:08 crc kubenswrapper[4988]: I1008 20:14:08.801826 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="7b719c42-3547-44a4-ba4d-4929583bd21a" containerName="extract-content" Oct 08 20:14:08 crc kubenswrapper[4988]: E1008 20:14:08.801849 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9edf6678-96ef-43aa-b2e0-314948fd9283" containerName="extract-utilities" Oct 08 20:14:08 crc kubenswrapper[4988]: I1008 20:14:08.801859 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="9edf6678-96ef-43aa-b2e0-314948fd9283" containerName="extract-utilities" Oct 08 20:14:08 crc kubenswrapper[4988]: E1008 20:14:08.801892 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9edf6678-96ef-43aa-b2e0-314948fd9283" containerName="extract-content" Oct 08 20:14:08 crc kubenswrapper[4988]: I1008 20:14:08.801901 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="9edf6678-96ef-43aa-b2e0-314948fd9283" containerName="extract-content" Oct 08 20:14:08 crc kubenswrapper[4988]: E1008 20:14:08.801925 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7b719c42-3547-44a4-ba4d-4929583bd21a" containerName="extract-utilities" Oct 08 20:14:08 crc kubenswrapper[4988]: I1008 20:14:08.801932 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="7b719c42-3547-44a4-ba4d-4929583bd21a" containerName="extract-utilities" Oct 08 20:14:08 crc kubenswrapper[4988]: E1008 20:14:08.801965 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7b719c42-3547-44a4-ba4d-4929583bd21a" containerName="registry-server" Oct 08 20:14:08 crc kubenswrapper[4988]: I1008 20:14:08.801974 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="7b719c42-3547-44a4-ba4d-4929583bd21a" containerName="registry-server" Oct 08 20:14:08 crc kubenswrapper[4988]: E1008 20:14:08.801993 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9edf6678-96ef-43aa-b2e0-314948fd9283" containerName="registry-server" Oct 08 20:14:08 crc kubenswrapper[4988]: I1008 20:14:08.802000 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="9edf6678-96ef-43aa-b2e0-314948fd9283" containerName="registry-server" Oct 08 20:14:08 crc kubenswrapper[4988]: I1008 20:14:08.804655 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="9edf6678-96ef-43aa-b2e0-314948fd9283" containerName="registry-server" Oct 08 20:14:08 crc kubenswrapper[4988]: I1008 20:14:08.804706 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="7b719c42-3547-44a4-ba4d-4929583bd21a" containerName="registry-server" Oct 08 20:14:08 crc kubenswrapper[4988]: I1008 20:14:08.804751 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="7dd8825a-ca2d-4361-a6e0-cf8a68ba8078" containerName="tripleo-cleanup-tripleo-cleanup-openstack-cell1" Oct 08 20:14:08 crc kubenswrapper[4988]: I1008 20:14:08.809922 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-openstack-openstack-cell1-t5z85" Oct 08 20:14:08 crc kubenswrapper[4988]: I1008 20:14:08.819063 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-4sdzr" Oct 08 20:14:08 crc kubenswrapper[4988]: I1008 20:14:08.820159 4988 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 08 20:14:08 crc kubenswrapper[4988]: I1008 20:14:08.820167 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Oct 08 20:14:08 crc kubenswrapper[4988]: I1008 20:14:08.820550 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Oct 08 20:14:08 crc kubenswrapper[4988]: I1008 20:14:08.843935 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/bootstrap-openstack-openstack-cell1-t5z85"] Oct 08 20:14:08 crc kubenswrapper[4988]: I1008 20:14:08.847519 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9e3f73e4-4a15-4ed9-a899-e7c4d6837b03-inventory\") pod \"bootstrap-openstack-openstack-cell1-t5z85\" (UID: \"9e3f73e4-4a15-4ed9-a899-e7c4d6837b03\") " pod="openstack/bootstrap-openstack-openstack-cell1-t5z85" Oct 08 20:14:08 crc kubenswrapper[4988]: I1008 20:14:08.847641 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9e3f73e4-4a15-4ed9-a899-e7c4d6837b03-bootstrap-combined-ca-bundle\") pod \"bootstrap-openstack-openstack-cell1-t5z85\" (UID: \"9e3f73e4-4a15-4ed9-a899-e7c4d6837b03\") " pod="openstack/bootstrap-openstack-openstack-cell1-t5z85" Oct 08 20:14:08 crc kubenswrapper[4988]: I1008 20:14:08.848966 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5wp62\" (UniqueName: \"kubernetes.io/projected/9e3f73e4-4a15-4ed9-a899-e7c4d6837b03-kube-api-access-5wp62\") pod \"bootstrap-openstack-openstack-cell1-t5z85\" (UID: \"9e3f73e4-4a15-4ed9-a899-e7c4d6837b03\") " pod="openstack/bootstrap-openstack-openstack-cell1-t5z85" Oct 08 20:14:08 crc kubenswrapper[4988]: I1008 20:14:08.849130 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9e3f73e4-4a15-4ed9-a899-e7c4d6837b03-ssh-key\") pod \"bootstrap-openstack-openstack-cell1-t5z85\" (UID: \"9e3f73e4-4a15-4ed9-a899-e7c4d6837b03\") " pod="openstack/bootstrap-openstack-openstack-cell1-t5z85" Oct 08 20:14:08 crc kubenswrapper[4988]: I1008 20:14:08.951239 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9e3f73e4-4a15-4ed9-a899-e7c4d6837b03-inventory\") pod \"bootstrap-openstack-openstack-cell1-t5z85\" (UID: \"9e3f73e4-4a15-4ed9-a899-e7c4d6837b03\") " pod="openstack/bootstrap-openstack-openstack-cell1-t5z85" Oct 08 20:14:08 crc kubenswrapper[4988]: I1008 20:14:08.951346 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9e3f73e4-4a15-4ed9-a899-e7c4d6837b03-bootstrap-combined-ca-bundle\") pod \"bootstrap-openstack-openstack-cell1-t5z85\" (UID: \"9e3f73e4-4a15-4ed9-a899-e7c4d6837b03\") " pod="openstack/bootstrap-openstack-openstack-cell1-t5z85" Oct 08 20:14:08 crc kubenswrapper[4988]: I1008 20:14:08.951472 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5wp62\" (UniqueName: \"kubernetes.io/projected/9e3f73e4-4a15-4ed9-a899-e7c4d6837b03-kube-api-access-5wp62\") pod \"bootstrap-openstack-openstack-cell1-t5z85\" (UID: \"9e3f73e4-4a15-4ed9-a899-e7c4d6837b03\") " pod="openstack/bootstrap-openstack-openstack-cell1-t5z85" Oct 08 20:14:08 crc kubenswrapper[4988]: I1008 20:14:08.951554 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9e3f73e4-4a15-4ed9-a899-e7c4d6837b03-ssh-key\") pod \"bootstrap-openstack-openstack-cell1-t5z85\" (UID: \"9e3f73e4-4a15-4ed9-a899-e7c4d6837b03\") " pod="openstack/bootstrap-openstack-openstack-cell1-t5z85" Oct 08 20:14:08 crc kubenswrapper[4988]: I1008 20:14:08.956939 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9e3f73e4-4a15-4ed9-a899-e7c4d6837b03-inventory\") pod \"bootstrap-openstack-openstack-cell1-t5z85\" (UID: \"9e3f73e4-4a15-4ed9-a899-e7c4d6837b03\") " pod="openstack/bootstrap-openstack-openstack-cell1-t5z85" Oct 08 20:14:08 crc kubenswrapper[4988]: I1008 20:14:08.957270 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9e3f73e4-4a15-4ed9-a899-e7c4d6837b03-ssh-key\") pod \"bootstrap-openstack-openstack-cell1-t5z85\" (UID: \"9e3f73e4-4a15-4ed9-a899-e7c4d6837b03\") " pod="openstack/bootstrap-openstack-openstack-cell1-t5z85" Oct 08 20:14:08 crc kubenswrapper[4988]: I1008 20:14:08.957351 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9e3f73e4-4a15-4ed9-a899-e7c4d6837b03-bootstrap-combined-ca-bundle\") pod \"bootstrap-openstack-openstack-cell1-t5z85\" (UID: \"9e3f73e4-4a15-4ed9-a899-e7c4d6837b03\") " pod="openstack/bootstrap-openstack-openstack-cell1-t5z85" Oct 08 20:14:08 crc kubenswrapper[4988]: I1008 20:14:08.983321 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5wp62\" (UniqueName: \"kubernetes.io/projected/9e3f73e4-4a15-4ed9-a899-e7c4d6837b03-kube-api-access-5wp62\") pod \"bootstrap-openstack-openstack-cell1-t5z85\" (UID: \"9e3f73e4-4a15-4ed9-a899-e7c4d6837b03\") " pod="openstack/bootstrap-openstack-openstack-cell1-t5z85" Oct 08 20:14:09 crc kubenswrapper[4988]: I1008 20:14:09.136986 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-openstack-openstack-cell1-t5z85" Oct 08 20:14:09 crc kubenswrapper[4988]: I1008 20:14:09.771421 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/bootstrap-openstack-openstack-cell1-t5z85"] Oct 08 20:14:09 crc kubenswrapper[4988]: I1008 20:14:09.774565 4988 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 08 20:14:10 crc kubenswrapper[4988]: I1008 20:14:10.410858 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-openstack-openstack-cell1-t5z85" event={"ID":"9e3f73e4-4a15-4ed9-a899-e7c4d6837b03","Type":"ContainerStarted","Data":"e4cc8f86b4ba53cd20d704e4554d94e11a34402c0456602ea7902079495b1798"} Oct 08 20:14:11 crc kubenswrapper[4988]: I1008 20:14:11.423296 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-openstack-openstack-cell1-t5z85" event={"ID":"9e3f73e4-4a15-4ed9-a899-e7c4d6837b03","Type":"ContainerStarted","Data":"4fc5075925ce5784d8887201206627270d55041d7a803cc26bd534df807508f2"} Oct 08 20:14:11 crc kubenswrapper[4988]: I1008 20:14:11.456466 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/bootstrap-openstack-openstack-cell1-t5z85" podStartSLOduration=2.975753762 podStartE2EDuration="3.456441441s" podCreationTimestamp="2025-10-08 20:14:08 +0000 UTC" firstStartedPulling="2025-10-08 20:14:09.773986514 +0000 UTC m=+7395.223829324" lastFinishedPulling="2025-10-08 20:14:10.254674233 +0000 UTC m=+7395.704517003" observedRunningTime="2025-10-08 20:14:11.445764902 +0000 UTC m=+7396.895607732" watchObservedRunningTime="2025-10-08 20:14:11.456441441 +0000 UTC m=+7396.906284251" Oct 08 20:14:14 crc kubenswrapper[4988]: I1008 20:14:14.237941 4988 scope.go:117] "RemoveContainer" containerID="c7dc147e5c4549b85b8a906334a50f44db496ead5160780b469fe7774d8aa2fa" Oct 08 20:14:14 crc kubenswrapper[4988]: E1008 20:14:14.238692 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nm54f_openshift-machine-config-operator(f328a20b-5283-407e-8a93-2ab2da599f18)\"" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" Oct 08 20:14:26 crc kubenswrapper[4988]: I1008 20:14:26.238306 4988 scope.go:117] "RemoveContainer" containerID="c7dc147e5c4549b85b8a906334a50f44db496ead5160780b469fe7774d8aa2fa" Oct 08 20:14:26 crc kubenswrapper[4988]: I1008 20:14:26.628770 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" event={"ID":"f328a20b-5283-407e-8a93-2ab2da599f18","Type":"ContainerStarted","Data":"f42539de8c8a890b6e35787c0007fb9c8dd69cabc2b007044b887c7ef2f0610b"} Oct 08 20:15:00 crc kubenswrapper[4988]: I1008 20:15:00.183476 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29332575-svtg7"] Oct 08 20:15:00 crc kubenswrapper[4988]: I1008 20:15:00.187081 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29332575-svtg7" Oct 08 20:15:00 crc kubenswrapper[4988]: I1008 20:15:00.189542 4988 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 08 20:15:00 crc kubenswrapper[4988]: I1008 20:15:00.190478 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 08 20:15:00 crc kubenswrapper[4988]: I1008 20:15:00.200185 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29332575-svtg7"] Oct 08 20:15:00 crc kubenswrapper[4988]: I1008 20:15:00.365313 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z96dm\" (UniqueName: \"kubernetes.io/projected/32593c4e-949e-40be-884a-3074d923b4f9-kube-api-access-z96dm\") pod \"collect-profiles-29332575-svtg7\" (UID: \"32593c4e-949e-40be-884a-3074d923b4f9\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29332575-svtg7" Oct 08 20:15:00 crc kubenswrapper[4988]: I1008 20:15:00.366115 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/32593c4e-949e-40be-884a-3074d923b4f9-secret-volume\") pod \"collect-profiles-29332575-svtg7\" (UID: \"32593c4e-949e-40be-884a-3074d923b4f9\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29332575-svtg7" Oct 08 20:15:00 crc kubenswrapper[4988]: I1008 20:15:00.366181 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/32593c4e-949e-40be-884a-3074d923b4f9-config-volume\") pod \"collect-profiles-29332575-svtg7\" (UID: \"32593c4e-949e-40be-884a-3074d923b4f9\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29332575-svtg7" Oct 08 20:15:00 crc kubenswrapper[4988]: I1008 20:15:00.468920 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/32593c4e-949e-40be-884a-3074d923b4f9-secret-volume\") pod \"collect-profiles-29332575-svtg7\" (UID: \"32593c4e-949e-40be-884a-3074d923b4f9\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29332575-svtg7" Oct 08 20:15:00 crc kubenswrapper[4988]: I1008 20:15:00.469031 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/32593c4e-949e-40be-884a-3074d923b4f9-config-volume\") pod \"collect-profiles-29332575-svtg7\" (UID: \"32593c4e-949e-40be-884a-3074d923b4f9\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29332575-svtg7" Oct 08 20:15:00 crc kubenswrapper[4988]: I1008 20:15:00.469258 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z96dm\" (UniqueName: \"kubernetes.io/projected/32593c4e-949e-40be-884a-3074d923b4f9-kube-api-access-z96dm\") pod \"collect-profiles-29332575-svtg7\" (UID: \"32593c4e-949e-40be-884a-3074d923b4f9\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29332575-svtg7" Oct 08 20:15:00 crc kubenswrapper[4988]: I1008 20:15:00.470912 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/32593c4e-949e-40be-884a-3074d923b4f9-config-volume\") pod \"collect-profiles-29332575-svtg7\" (UID: \"32593c4e-949e-40be-884a-3074d923b4f9\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29332575-svtg7" Oct 08 20:15:00 crc kubenswrapper[4988]: I1008 20:15:00.487524 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/32593c4e-949e-40be-884a-3074d923b4f9-secret-volume\") pod \"collect-profiles-29332575-svtg7\" (UID: \"32593c4e-949e-40be-884a-3074d923b4f9\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29332575-svtg7" Oct 08 20:15:00 crc kubenswrapper[4988]: I1008 20:15:00.490590 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z96dm\" (UniqueName: \"kubernetes.io/projected/32593c4e-949e-40be-884a-3074d923b4f9-kube-api-access-z96dm\") pod \"collect-profiles-29332575-svtg7\" (UID: \"32593c4e-949e-40be-884a-3074d923b4f9\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29332575-svtg7" Oct 08 20:15:00 crc kubenswrapper[4988]: I1008 20:15:00.522008 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29332575-svtg7" Oct 08 20:15:01 crc kubenswrapper[4988]: I1008 20:15:01.022674 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29332575-svtg7"] Oct 08 20:15:01 crc kubenswrapper[4988]: I1008 20:15:01.083868 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29332575-svtg7" event={"ID":"32593c4e-949e-40be-884a-3074d923b4f9","Type":"ContainerStarted","Data":"f2cd58eca4a05fce99747cb1e6f6c97631867adbb0cf9211e67d30567849e0a8"} Oct 08 20:15:02 crc kubenswrapper[4988]: I1008 20:15:02.098275 4988 generic.go:334] "Generic (PLEG): container finished" podID="32593c4e-949e-40be-884a-3074d923b4f9" containerID="2bf6c14a16c4c0ca339e89944a1fe637b5c9f8431fa1acad3b682a250598484f" exitCode=0 Oct 08 20:15:02 crc kubenswrapper[4988]: I1008 20:15:02.098411 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29332575-svtg7" event={"ID":"32593c4e-949e-40be-884a-3074d923b4f9","Type":"ContainerDied","Data":"2bf6c14a16c4c0ca339e89944a1fe637b5c9f8431fa1acad3b682a250598484f"} Oct 08 20:15:03 crc kubenswrapper[4988]: I1008 20:15:03.545188 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29332575-svtg7" Oct 08 20:15:03 crc kubenswrapper[4988]: I1008 20:15:03.553562 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/32593c4e-949e-40be-884a-3074d923b4f9-config-volume\") pod \"32593c4e-949e-40be-884a-3074d923b4f9\" (UID: \"32593c4e-949e-40be-884a-3074d923b4f9\") " Oct 08 20:15:03 crc kubenswrapper[4988]: I1008 20:15:03.553616 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/32593c4e-949e-40be-884a-3074d923b4f9-secret-volume\") pod \"32593c4e-949e-40be-884a-3074d923b4f9\" (UID: \"32593c4e-949e-40be-884a-3074d923b4f9\") " Oct 08 20:15:03 crc kubenswrapper[4988]: I1008 20:15:03.553698 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z96dm\" (UniqueName: \"kubernetes.io/projected/32593c4e-949e-40be-884a-3074d923b4f9-kube-api-access-z96dm\") pod \"32593c4e-949e-40be-884a-3074d923b4f9\" (UID: \"32593c4e-949e-40be-884a-3074d923b4f9\") " Oct 08 20:15:03 crc kubenswrapper[4988]: I1008 20:15:03.554612 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/32593c4e-949e-40be-884a-3074d923b4f9-config-volume" (OuterVolumeSpecName: "config-volume") pod "32593c4e-949e-40be-884a-3074d923b4f9" (UID: "32593c4e-949e-40be-884a-3074d923b4f9"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 20:15:03 crc kubenswrapper[4988]: I1008 20:15:03.559555 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/32593c4e-949e-40be-884a-3074d923b4f9-kube-api-access-z96dm" (OuterVolumeSpecName: "kube-api-access-z96dm") pod "32593c4e-949e-40be-884a-3074d923b4f9" (UID: "32593c4e-949e-40be-884a-3074d923b4f9"). InnerVolumeSpecName "kube-api-access-z96dm". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 20:15:03 crc kubenswrapper[4988]: I1008 20:15:03.560868 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/32593c4e-949e-40be-884a-3074d923b4f9-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "32593c4e-949e-40be-884a-3074d923b4f9" (UID: "32593c4e-949e-40be-884a-3074d923b4f9"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 20:15:03 crc kubenswrapper[4988]: I1008 20:15:03.656645 4988 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/32593c4e-949e-40be-884a-3074d923b4f9-config-volume\") on node \"crc\" DevicePath \"\"" Oct 08 20:15:03 crc kubenswrapper[4988]: I1008 20:15:03.656891 4988 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/32593c4e-949e-40be-884a-3074d923b4f9-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 08 20:15:03 crc kubenswrapper[4988]: I1008 20:15:03.656960 4988 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z96dm\" (UniqueName: \"kubernetes.io/projected/32593c4e-949e-40be-884a-3074d923b4f9-kube-api-access-z96dm\") on node \"crc\" DevicePath \"\"" Oct 08 20:15:04 crc kubenswrapper[4988]: I1008 20:15:04.125460 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29332575-svtg7" event={"ID":"32593c4e-949e-40be-884a-3074d923b4f9","Type":"ContainerDied","Data":"f2cd58eca4a05fce99747cb1e6f6c97631867adbb0cf9211e67d30567849e0a8"} Oct 08 20:15:04 crc kubenswrapper[4988]: I1008 20:15:04.125507 4988 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f2cd58eca4a05fce99747cb1e6f6c97631867adbb0cf9211e67d30567849e0a8" Oct 08 20:15:04 crc kubenswrapper[4988]: I1008 20:15:04.126597 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29332575-svtg7" Oct 08 20:15:04 crc kubenswrapper[4988]: I1008 20:15:04.637006 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29332530-rgqrs"] Oct 08 20:15:04 crc kubenswrapper[4988]: I1008 20:15:04.653377 4988 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29332530-rgqrs"] Oct 08 20:15:05 crc kubenswrapper[4988]: I1008 20:15:05.260862 4988 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="501d2535-94b2-41bf-93e1-262f008d88a6" path="/var/lib/kubelet/pods/501d2535-94b2-41bf-93e1-262f008d88a6/volumes" Oct 08 20:15:15 crc kubenswrapper[4988]: I1008 20:15:15.284039 4988 scope.go:117] "RemoveContainer" containerID="5416125fd85a323f78420b1ef7049c539da1430e9d5b0a9596588cb33b829447" Oct 08 20:16:53 crc kubenswrapper[4988]: I1008 20:16:53.338091 4988 patch_prober.go:28] interesting pod/machine-config-daemon-nm54f container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 20:16:53 crc kubenswrapper[4988]: I1008 20:16:53.338959 4988 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 20:17:18 crc kubenswrapper[4988]: I1008 20:17:18.807432 4988 generic.go:334] "Generic (PLEG): container finished" podID="9e3f73e4-4a15-4ed9-a899-e7c4d6837b03" containerID="4fc5075925ce5784d8887201206627270d55041d7a803cc26bd534df807508f2" exitCode=0 Oct 08 20:17:18 crc kubenswrapper[4988]: I1008 20:17:18.807527 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-openstack-openstack-cell1-t5z85" event={"ID":"9e3f73e4-4a15-4ed9-a899-e7c4d6837b03","Type":"ContainerDied","Data":"4fc5075925ce5784d8887201206627270d55041d7a803cc26bd534df807508f2"} Oct 08 20:17:20 crc kubenswrapper[4988]: I1008 20:17:20.304678 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-openstack-openstack-cell1-t5z85" Oct 08 20:17:20 crc kubenswrapper[4988]: I1008 20:17:20.425425 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9e3f73e4-4a15-4ed9-a899-e7c4d6837b03-ssh-key\") pod \"9e3f73e4-4a15-4ed9-a899-e7c4d6837b03\" (UID: \"9e3f73e4-4a15-4ed9-a899-e7c4d6837b03\") " Oct 08 20:17:20 crc kubenswrapper[4988]: I1008 20:17:20.425498 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9e3f73e4-4a15-4ed9-a899-e7c4d6837b03-bootstrap-combined-ca-bundle\") pod \"9e3f73e4-4a15-4ed9-a899-e7c4d6837b03\" (UID: \"9e3f73e4-4a15-4ed9-a899-e7c4d6837b03\") " Oct 08 20:17:20 crc kubenswrapper[4988]: I1008 20:17:20.425859 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5wp62\" (UniqueName: \"kubernetes.io/projected/9e3f73e4-4a15-4ed9-a899-e7c4d6837b03-kube-api-access-5wp62\") pod \"9e3f73e4-4a15-4ed9-a899-e7c4d6837b03\" (UID: \"9e3f73e4-4a15-4ed9-a899-e7c4d6837b03\") " Oct 08 20:17:20 crc kubenswrapper[4988]: I1008 20:17:20.425921 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9e3f73e4-4a15-4ed9-a899-e7c4d6837b03-inventory\") pod \"9e3f73e4-4a15-4ed9-a899-e7c4d6837b03\" (UID: \"9e3f73e4-4a15-4ed9-a899-e7c4d6837b03\") " Oct 08 20:17:20 crc kubenswrapper[4988]: I1008 20:17:20.432831 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9e3f73e4-4a15-4ed9-a899-e7c4d6837b03-kube-api-access-5wp62" (OuterVolumeSpecName: "kube-api-access-5wp62") pod "9e3f73e4-4a15-4ed9-a899-e7c4d6837b03" (UID: "9e3f73e4-4a15-4ed9-a899-e7c4d6837b03"). InnerVolumeSpecName "kube-api-access-5wp62". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 20:17:20 crc kubenswrapper[4988]: I1008 20:17:20.433029 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9e3f73e4-4a15-4ed9-a899-e7c4d6837b03-bootstrap-combined-ca-bundle" (OuterVolumeSpecName: "bootstrap-combined-ca-bundle") pod "9e3f73e4-4a15-4ed9-a899-e7c4d6837b03" (UID: "9e3f73e4-4a15-4ed9-a899-e7c4d6837b03"). InnerVolumeSpecName "bootstrap-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 20:17:20 crc kubenswrapper[4988]: I1008 20:17:20.455507 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9e3f73e4-4a15-4ed9-a899-e7c4d6837b03-inventory" (OuterVolumeSpecName: "inventory") pod "9e3f73e4-4a15-4ed9-a899-e7c4d6837b03" (UID: "9e3f73e4-4a15-4ed9-a899-e7c4d6837b03"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 20:17:20 crc kubenswrapper[4988]: I1008 20:17:20.468546 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9e3f73e4-4a15-4ed9-a899-e7c4d6837b03-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "9e3f73e4-4a15-4ed9-a899-e7c4d6837b03" (UID: "9e3f73e4-4a15-4ed9-a899-e7c4d6837b03"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 20:17:20 crc kubenswrapper[4988]: I1008 20:17:20.529028 4988 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5wp62\" (UniqueName: \"kubernetes.io/projected/9e3f73e4-4a15-4ed9-a899-e7c4d6837b03-kube-api-access-5wp62\") on node \"crc\" DevicePath \"\"" Oct 08 20:17:20 crc kubenswrapper[4988]: I1008 20:17:20.529082 4988 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9e3f73e4-4a15-4ed9-a899-e7c4d6837b03-inventory\") on node \"crc\" DevicePath \"\"" Oct 08 20:17:20 crc kubenswrapper[4988]: I1008 20:17:20.529104 4988 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9e3f73e4-4a15-4ed9-a899-e7c4d6837b03-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 08 20:17:20 crc kubenswrapper[4988]: I1008 20:17:20.529124 4988 reconciler_common.go:293] "Volume detached for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9e3f73e4-4a15-4ed9-a899-e7c4d6837b03-bootstrap-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 20:17:20 crc kubenswrapper[4988]: I1008 20:17:20.829397 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-openstack-openstack-cell1-t5z85" event={"ID":"9e3f73e4-4a15-4ed9-a899-e7c4d6837b03","Type":"ContainerDied","Data":"e4cc8f86b4ba53cd20d704e4554d94e11a34402c0456602ea7902079495b1798"} Oct 08 20:17:20 crc kubenswrapper[4988]: I1008 20:17:20.829774 4988 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e4cc8f86b4ba53cd20d704e4554d94e11a34402c0456602ea7902079495b1798" Oct 08 20:17:20 crc kubenswrapper[4988]: I1008 20:17:20.829459 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-openstack-openstack-cell1-t5z85" Oct 08 20:17:20 crc kubenswrapper[4988]: I1008 20:17:20.926583 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/download-cache-openstack-openstack-cell1-fjdm2"] Oct 08 20:17:20 crc kubenswrapper[4988]: E1008 20:17:20.927124 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9e3f73e4-4a15-4ed9-a899-e7c4d6837b03" containerName="bootstrap-openstack-openstack-cell1" Oct 08 20:17:20 crc kubenswrapper[4988]: I1008 20:17:20.927140 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="9e3f73e4-4a15-4ed9-a899-e7c4d6837b03" containerName="bootstrap-openstack-openstack-cell1" Oct 08 20:17:20 crc kubenswrapper[4988]: E1008 20:17:20.927164 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="32593c4e-949e-40be-884a-3074d923b4f9" containerName="collect-profiles" Oct 08 20:17:20 crc kubenswrapper[4988]: I1008 20:17:20.927171 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="32593c4e-949e-40be-884a-3074d923b4f9" containerName="collect-profiles" Oct 08 20:17:20 crc kubenswrapper[4988]: I1008 20:17:20.927374 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="9e3f73e4-4a15-4ed9-a899-e7c4d6837b03" containerName="bootstrap-openstack-openstack-cell1" Oct 08 20:17:20 crc kubenswrapper[4988]: I1008 20:17:20.927417 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="32593c4e-949e-40be-884a-3074d923b4f9" containerName="collect-profiles" Oct 08 20:17:20 crc kubenswrapper[4988]: I1008 20:17:20.928114 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-openstack-openstack-cell1-fjdm2" Oct 08 20:17:20 crc kubenswrapper[4988]: I1008 20:17:20.930298 4988 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 08 20:17:20 crc kubenswrapper[4988]: I1008 20:17:20.930720 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Oct 08 20:17:20 crc kubenswrapper[4988]: I1008 20:17:20.932010 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Oct 08 20:17:20 crc kubenswrapper[4988]: I1008 20:17:20.932938 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-4sdzr" Oct 08 20:17:20 crc kubenswrapper[4988]: I1008 20:17:20.944953 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/download-cache-openstack-openstack-cell1-fjdm2"] Oct 08 20:17:21 crc kubenswrapper[4988]: I1008 20:17:21.039349 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5ab4ca0a-e3b3-418e-8640-13828852f6fb-inventory\") pod \"download-cache-openstack-openstack-cell1-fjdm2\" (UID: \"5ab4ca0a-e3b3-418e-8640-13828852f6fb\") " pod="openstack/download-cache-openstack-openstack-cell1-fjdm2" Oct 08 20:17:21 crc kubenswrapper[4988]: I1008 20:17:21.039633 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-njv46\" (UniqueName: \"kubernetes.io/projected/5ab4ca0a-e3b3-418e-8640-13828852f6fb-kube-api-access-njv46\") pod \"download-cache-openstack-openstack-cell1-fjdm2\" (UID: \"5ab4ca0a-e3b3-418e-8640-13828852f6fb\") " pod="openstack/download-cache-openstack-openstack-cell1-fjdm2" Oct 08 20:17:21 crc kubenswrapper[4988]: I1008 20:17:21.039670 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5ab4ca0a-e3b3-418e-8640-13828852f6fb-ssh-key\") pod \"download-cache-openstack-openstack-cell1-fjdm2\" (UID: \"5ab4ca0a-e3b3-418e-8640-13828852f6fb\") " pod="openstack/download-cache-openstack-openstack-cell1-fjdm2" Oct 08 20:17:21 crc kubenswrapper[4988]: I1008 20:17:21.141690 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-njv46\" (UniqueName: \"kubernetes.io/projected/5ab4ca0a-e3b3-418e-8640-13828852f6fb-kube-api-access-njv46\") pod \"download-cache-openstack-openstack-cell1-fjdm2\" (UID: \"5ab4ca0a-e3b3-418e-8640-13828852f6fb\") " pod="openstack/download-cache-openstack-openstack-cell1-fjdm2" Oct 08 20:17:21 crc kubenswrapper[4988]: I1008 20:17:21.141737 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5ab4ca0a-e3b3-418e-8640-13828852f6fb-ssh-key\") pod \"download-cache-openstack-openstack-cell1-fjdm2\" (UID: \"5ab4ca0a-e3b3-418e-8640-13828852f6fb\") " pod="openstack/download-cache-openstack-openstack-cell1-fjdm2" Oct 08 20:17:21 crc kubenswrapper[4988]: I1008 20:17:21.141780 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5ab4ca0a-e3b3-418e-8640-13828852f6fb-inventory\") pod \"download-cache-openstack-openstack-cell1-fjdm2\" (UID: \"5ab4ca0a-e3b3-418e-8640-13828852f6fb\") " pod="openstack/download-cache-openstack-openstack-cell1-fjdm2" Oct 08 20:17:21 crc kubenswrapper[4988]: I1008 20:17:21.146879 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5ab4ca0a-e3b3-418e-8640-13828852f6fb-inventory\") pod \"download-cache-openstack-openstack-cell1-fjdm2\" (UID: \"5ab4ca0a-e3b3-418e-8640-13828852f6fb\") " pod="openstack/download-cache-openstack-openstack-cell1-fjdm2" Oct 08 20:17:21 crc kubenswrapper[4988]: I1008 20:17:21.157131 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5ab4ca0a-e3b3-418e-8640-13828852f6fb-ssh-key\") pod \"download-cache-openstack-openstack-cell1-fjdm2\" (UID: \"5ab4ca0a-e3b3-418e-8640-13828852f6fb\") " pod="openstack/download-cache-openstack-openstack-cell1-fjdm2" Oct 08 20:17:21 crc kubenswrapper[4988]: I1008 20:17:21.159186 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-njv46\" (UniqueName: \"kubernetes.io/projected/5ab4ca0a-e3b3-418e-8640-13828852f6fb-kube-api-access-njv46\") pod \"download-cache-openstack-openstack-cell1-fjdm2\" (UID: \"5ab4ca0a-e3b3-418e-8640-13828852f6fb\") " pod="openstack/download-cache-openstack-openstack-cell1-fjdm2" Oct 08 20:17:21 crc kubenswrapper[4988]: I1008 20:17:21.253842 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-openstack-openstack-cell1-fjdm2" Oct 08 20:17:21 crc kubenswrapper[4988]: I1008 20:17:21.898164 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/download-cache-openstack-openstack-cell1-fjdm2"] Oct 08 20:17:22 crc kubenswrapper[4988]: I1008 20:17:22.859117 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-openstack-openstack-cell1-fjdm2" event={"ID":"5ab4ca0a-e3b3-418e-8640-13828852f6fb","Type":"ContainerStarted","Data":"cec846fa594f9dd4bda380aa31cfc3ac943ea23f3edc98405de4b2e8e54f0a48"} Oct 08 20:17:22 crc kubenswrapper[4988]: I1008 20:17:22.860158 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-openstack-openstack-cell1-fjdm2" event={"ID":"5ab4ca0a-e3b3-418e-8640-13828852f6fb","Type":"ContainerStarted","Data":"468a24e2aceba349325a35412246e319fe6d88f1436fcfad51382785de97d333"} Oct 08 20:17:22 crc kubenswrapper[4988]: I1008 20:17:22.887099 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/download-cache-openstack-openstack-cell1-fjdm2" podStartSLOduration=2.375313717 podStartE2EDuration="2.887066464s" podCreationTimestamp="2025-10-08 20:17:20 +0000 UTC" firstStartedPulling="2025-10-08 20:17:21.90283335 +0000 UTC m=+7587.352676120" lastFinishedPulling="2025-10-08 20:17:22.414586097 +0000 UTC m=+7587.864428867" observedRunningTime="2025-10-08 20:17:22.880282398 +0000 UTC m=+7588.330125238" watchObservedRunningTime="2025-10-08 20:17:22.887066464 +0000 UTC m=+7588.336909284" Oct 08 20:17:23 crc kubenswrapper[4988]: I1008 20:17:23.338836 4988 patch_prober.go:28] interesting pod/machine-config-daemon-nm54f container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 20:17:23 crc kubenswrapper[4988]: I1008 20:17:23.338918 4988 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 20:17:53 crc kubenswrapper[4988]: I1008 20:17:53.338579 4988 patch_prober.go:28] interesting pod/machine-config-daemon-nm54f container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 20:17:53 crc kubenswrapper[4988]: I1008 20:17:53.339274 4988 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 20:17:53 crc kubenswrapper[4988]: I1008 20:17:53.339351 4988 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" Oct 08 20:17:53 crc kubenswrapper[4988]: I1008 20:17:53.340647 4988 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"f42539de8c8a890b6e35787c0007fb9c8dd69cabc2b007044b887c7ef2f0610b"} pod="openshift-machine-config-operator/machine-config-daemon-nm54f" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 08 20:17:53 crc kubenswrapper[4988]: I1008 20:17:53.340767 4988 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" containerName="machine-config-daemon" containerID="cri-o://f42539de8c8a890b6e35787c0007fb9c8dd69cabc2b007044b887c7ef2f0610b" gracePeriod=600 Oct 08 20:17:54 crc kubenswrapper[4988]: I1008 20:17:54.225168 4988 generic.go:334] "Generic (PLEG): container finished" podID="f328a20b-5283-407e-8a93-2ab2da599f18" containerID="f42539de8c8a890b6e35787c0007fb9c8dd69cabc2b007044b887c7ef2f0610b" exitCode=0 Oct 08 20:17:54 crc kubenswrapper[4988]: I1008 20:17:54.225235 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" event={"ID":"f328a20b-5283-407e-8a93-2ab2da599f18","Type":"ContainerDied","Data":"f42539de8c8a890b6e35787c0007fb9c8dd69cabc2b007044b887c7ef2f0610b"} Oct 08 20:17:54 crc kubenswrapper[4988]: I1008 20:17:54.225479 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" event={"ID":"f328a20b-5283-407e-8a93-2ab2da599f18","Type":"ContainerStarted","Data":"7541195e86b98682ea101ebff2b19f543aa9bab45dd1665e8f710922ad18c5e7"} Oct 08 20:17:54 crc kubenswrapper[4988]: I1008 20:17:54.225495 4988 scope.go:117] "RemoveContainer" containerID="c7dc147e5c4549b85b8a906334a50f44db496ead5160780b469fe7774d8aa2fa" Oct 08 20:19:00 crc kubenswrapper[4988]: I1008 20:19:00.096952 4988 generic.go:334] "Generic (PLEG): container finished" podID="5ab4ca0a-e3b3-418e-8640-13828852f6fb" containerID="cec846fa594f9dd4bda380aa31cfc3ac943ea23f3edc98405de4b2e8e54f0a48" exitCode=0 Oct 08 20:19:00 crc kubenswrapper[4988]: I1008 20:19:00.097081 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-openstack-openstack-cell1-fjdm2" event={"ID":"5ab4ca0a-e3b3-418e-8640-13828852f6fb","Type":"ContainerDied","Data":"cec846fa594f9dd4bda380aa31cfc3ac943ea23f3edc98405de4b2e8e54f0a48"} Oct 08 20:19:01 crc kubenswrapper[4988]: I1008 20:19:01.628322 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-openstack-openstack-cell1-fjdm2" Oct 08 20:19:01 crc kubenswrapper[4988]: I1008 20:19:01.748976 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5ab4ca0a-e3b3-418e-8640-13828852f6fb-inventory\") pod \"5ab4ca0a-e3b3-418e-8640-13828852f6fb\" (UID: \"5ab4ca0a-e3b3-418e-8640-13828852f6fb\") " Oct 08 20:19:01 crc kubenswrapper[4988]: I1008 20:19:01.749246 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5ab4ca0a-e3b3-418e-8640-13828852f6fb-ssh-key\") pod \"5ab4ca0a-e3b3-418e-8640-13828852f6fb\" (UID: \"5ab4ca0a-e3b3-418e-8640-13828852f6fb\") " Oct 08 20:19:01 crc kubenswrapper[4988]: I1008 20:19:01.749303 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-njv46\" (UniqueName: \"kubernetes.io/projected/5ab4ca0a-e3b3-418e-8640-13828852f6fb-kube-api-access-njv46\") pod \"5ab4ca0a-e3b3-418e-8640-13828852f6fb\" (UID: \"5ab4ca0a-e3b3-418e-8640-13828852f6fb\") " Oct 08 20:19:01 crc kubenswrapper[4988]: I1008 20:19:01.764215 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5ab4ca0a-e3b3-418e-8640-13828852f6fb-kube-api-access-njv46" (OuterVolumeSpecName: "kube-api-access-njv46") pod "5ab4ca0a-e3b3-418e-8640-13828852f6fb" (UID: "5ab4ca0a-e3b3-418e-8640-13828852f6fb"). InnerVolumeSpecName "kube-api-access-njv46". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 20:19:01 crc kubenswrapper[4988]: I1008 20:19:01.786268 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5ab4ca0a-e3b3-418e-8640-13828852f6fb-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "5ab4ca0a-e3b3-418e-8640-13828852f6fb" (UID: "5ab4ca0a-e3b3-418e-8640-13828852f6fb"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 20:19:01 crc kubenswrapper[4988]: I1008 20:19:01.804407 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5ab4ca0a-e3b3-418e-8640-13828852f6fb-inventory" (OuterVolumeSpecName: "inventory") pod "5ab4ca0a-e3b3-418e-8640-13828852f6fb" (UID: "5ab4ca0a-e3b3-418e-8640-13828852f6fb"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 20:19:01 crc kubenswrapper[4988]: I1008 20:19:01.851704 4988 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5ab4ca0a-e3b3-418e-8640-13828852f6fb-inventory\") on node \"crc\" DevicePath \"\"" Oct 08 20:19:01 crc kubenswrapper[4988]: I1008 20:19:01.851737 4988 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5ab4ca0a-e3b3-418e-8640-13828852f6fb-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 08 20:19:01 crc kubenswrapper[4988]: I1008 20:19:01.851748 4988 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-njv46\" (UniqueName: \"kubernetes.io/projected/5ab4ca0a-e3b3-418e-8640-13828852f6fb-kube-api-access-njv46\") on node \"crc\" DevicePath \"\"" Oct 08 20:19:02 crc kubenswrapper[4988]: I1008 20:19:02.129935 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-openstack-openstack-cell1-fjdm2" event={"ID":"5ab4ca0a-e3b3-418e-8640-13828852f6fb","Type":"ContainerDied","Data":"468a24e2aceba349325a35412246e319fe6d88f1436fcfad51382785de97d333"} Oct 08 20:19:02 crc kubenswrapper[4988]: I1008 20:19:02.130264 4988 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="468a24e2aceba349325a35412246e319fe6d88f1436fcfad51382785de97d333" Oct 08 20:19:02 crc kubenswrapper[4988]: I1008 20:19:02.130050 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-openstack-openstack-cell1-fjdm2" Oct 08 20:19:02 crc kubenswrapper[4988]: I1008 20:19:02.242717 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-network-openstack-openstack-cell1-cpds6"] Oct 08 20:19:02 crc kubenswrapper[4988]: E1008 20:19:02.243667 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5ab4ca0a-e3b3-418e-8640-13828852f6fb" containerName="download-cache-openstack-openstack-cell1" Oct 08 20:19:02 crc kubenswrapper[4988]: I1008 20:19:02.243712 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="5ab4ca0a-e3b3-418e-8640-13828852f6fb" containerName="download-cache-openstack-openstack-cell1" Oct 08 20:19:02 crc kubenswrapper[4988]: I1008 20:19:02.244215 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="5ab4ca0a-e3b3-418e-8640-13828852f6fb" containerName="download-cache-openstack-openstack-cell1" Oct 08 20:19:02 crc kubenswrapper[4988]: I1008 20:19:02.245971 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-openstack-openstack-cell1-cpds6" Oct 08 20:19:02 crc kubenswrapper[4988]: I1008 20:19:02.247846 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Oct 08 20:19:02 crc kubenswrapper[4988]: I1008 20:19:02.248788 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-4sdzr" Oct 08 20:19:02 crc kubenswrapper[4988]: I1008 20:19:02.248869 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Oct 08 20:19:02 crc kubenswrapper[4988]: I1008 20:19:02.249182 4988 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 08 20:19:02 crc kubenswrapper[4988]: I1008 20:19:02.254433 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-network-openstack-openstack-cell1-cpds6"] Oct 08 20:19:02 crc kubenswrapper[4988]: I1008 20:19:02.365137 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pczwn\" (UniqueName: \"kubernetes.io/projected/a5ae4d27-798c-4d58-a9d9-ea145d68d6fc-kube-api-access-pczwn\") pod \"configure-network-openstack-openstack-cell1-cpds6\" (UID: \"a5ae4d27-798c-4d58-a9d9-ea145d68d6fc\") " pod="openstack/configure-network-openstack-openstack-cell1-cpds6" Oct 08 20:19:02 crc kubenswrapper[4988]: I1008 20:19:02.365583 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a5ae4d27-798c-4d58-a9d9-ea145d68d6fc-inventory\") pod \"configure-network-openstack-openstack-cell1-cpds6\" (UID: \"a5ae4d27-798c-4d58-a9d9-ea145d68d6fc\") " pod="openstack/configure-network-openstack-openstack-cell1-cpds6" Oct 08 20:19:02 crc kubenswrapper[4988]: I1008 20:19:02.365732 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a5ae4d27-798c-4d58-a9d9-ea145d68d6fc-ssh-key\") pod \"configure-network-openstack-openstack-cell1-cpds6\" (UID: \"a5ae4d27-798c-4d58-a9d9-ea145d68d6fc\") " pod="openstack/configure-network-openstack-openstack-cell1-cpds6" Oct 08 20:19:02 crc kubenswrapper[4988]: I1008 20:19:02.468508 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pczwn\" (UniqueName: \"kubernetes.io/projected/a5ae4d27-798c-4d58-a9d9-ea145d68d6fc-kube-api-access-pczwn\") pod \"configure-network-openstack-openstack-cell1-cpds6\" (UID: \"a5ae4d27-798c-4d58-a9d9-ea145d68d6fc\") " pod="openstack/configure-network-openstack-openstack-cell1-cpds6" Oct 08 20:19:02 crc kubenswrapper[4988]: I1008 20:19:02.468814 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a5ae4d27-798c-4d58-a9d9-ea145d68d6fc-inventory\") pod \"configure-network-openstack-openstack-cell1-cpds6\" (UID: \"a5ae4d27-798c-4d58-a9d9-ea145d68d6fc\") " pod="openstack/configure-network-openstack-openstack-cell1-cpds6" Oct 08 20:19:02 crc kubenswrapper[4988]: I1008 20:19:02.468859 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a5ae4d27-798c-4d58-a9d9-ea145d68d6fc-ssh-key\") pod \"configure-network-openstack-openstack-cell1-cpds6\" (UID: \"a5ae4d27-798c-4d58-a9d9-ea145d68d6fc\") " pod="openstack/configure-network-openstack-openstack-cell1-cpds6" Oct 08 20:19:02 crc kubenswrapper[4988]: I1008 20:19:02.477349 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a5ae4d27-798c-4d58-a9d9-ea145d68d6fc-ssh-key\") pod \"configure-network-openstack-openstack-cell1-cpds6\" (UID: \"a5ae4d27-798c-4d58-a9d9-ea145d68d6fc\") " pod="openstack/configure-network-openstack-openstack-cell1-cpds6" Oct 08 20:19:02 crc kubenswrapper[4988]: I1008 20:19:02.477364 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a5ae4d27-798c-4d58-a9d9-ea145d68d6fc-inventory\") pod \"configure-network-openstack-openstack-cell1-cpds6\" (UID: \"a5ae4d27-798c-4d58-a9d9-ea145d68d6fc\") " pod="openstack/configure-network-openstack-openstack-cell1-cpds6" Oct 08 20:19:02 crc kubenswrapper[4988]: I1008 20:19:02.499483 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pczwn\" (UniqueName: \"kubernetes.io/projected/a5ae4d27-798c-4d58-a9d9-ea145d68d6fc-kube-api-access-pczwn\") pod \"configure-network-openstack-openstack-cell1-cpds6\" (UID: \"a5ae4d27-798c-4d58-a9d9-ea145d68d6fc\") " pod="openstack/configure-network-openstack-openstack-cell1-cpds6" Oct 08 20:19:02 crc kubenswrapper[4988]: I1008 20:19:02.566456 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-openstack-openstack-cell1-cpds6" Oct 08 20:19:03 crc kubenswrapper[4988]: I1008 20:19:03.132627 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-network-openstack-openstack-cell1-cpds6"] Oct 08 20:19:04 crc kubenswrapper[4988]: I1008 20:19:04.151306 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-openstack-openstack-cell1-cpds6" event={"ID":"a5ae4d27-798c-4d58-a9d9-ea145d68d6fc","Type":"ContainerStarted","Data":"1287b21ffc98890757c354add091c2fba49fbf20ce75d9c5f31d0378d6b34efa"} Oct 08 20:19:04 crc kubenswrapper[4988]: I1008 20:19:04.151700 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-openstack-openstack-cell1-cpds6" event={"ID":"a5ae4d27-798c-4d58-a9d9-ea145d68d6fc","Type":"ContainerStarted","Data":"fecd35d513f4c888d3df24c7eb1163f3142a7f71f0d46fa0c38fab44c6c8b23d"} Oct 08 20:19:04 crc kubenswrapper[4988]: I1008 20:19:04.179839 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-network-openstack-openstack-cell1-cpds6" podStartSLOduration=1.6390142220000001 podStartE2EDuration="2.179820484s" podCreationTimestamp="2025-10-08 20:19:02 +0000 UTC" firstStartedPulling="2025-10-08 20:19:03.139801164 +0000 UTC m=+7688.589643934" lastFinishedPulling="2025-10-08 20:19:03.680607426 +0000 UTC m=+7689.130450196" observedRunningTime="2025-10-08 20:19:04.170408135 +0000 UTC m=+7689.620250905" watchObservedRunningTime="2025-10-08 20:19:04.179820484 +0000 UTC m=+7689.629663254" Oct 08 20:19:48 crc kubenswrapper[4988]: I1008 20:19:48.973553 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-6s6t2"] Oct 08 20:19:48 crc kubenswrapper[4988]: I1008 20:19:48.977779 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-6s6t2" Oct 08 20:19:48 crc kubenswrapper[4988]: I1008 20:19:48.991066 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-6s6t2"] Oct 08 20:19:49 crc kubenswrapper[4988]: I1008 20:19:49.086524 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dczxg\" (UniqueName: \"kubernetes.io/projected/69b806a7-b7ef-4f6a-b543-18712f6df72c-kube-api-access-dczxg\") pod \"certified-operators-6s6t2\" (UID: \"69b806a7-b7ef-4f6a-b543-18712f6df72c\") " pod="openshift-marketplace/certified-operators-6s6t2" Oct 08 20:19:49 crc kubenswrapper[4988]: I1008 20:19:49.086603 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/69b806a7-b7ef-4f6a-b543-18712f6df72c-utilities\") pod \"certified-operators-6s6t2\" (UID: \"69b806a7-b7ef-4f6a-b543-18712f6df72c\") " pod="openshift-marketplace/certified-operators-6s6t2" Oct 08 20:19:49 crc kubenswrapper[4988]: I1008 20:19:49.086733 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/69b806a7-b7ef-4f6a-b543-18712f6df72c-catalog-content\") pod \"certified-operators-6s6t2\" (UID: \"69b806a7-b7ef-4f6a-b543-18712f6df72c\") " pod="openshift-marketplace/certified-operators-6s6t2" Oct 08 20:19:49 crc kubenswrapper[4988]: I1008 20:19:49.189418 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dczxg\" (UniqueName: \"kubernetes.io/projected/69b806a7-b7ef-4f6a-b543-18712f6df72c-kube-api-access-dczxg\") pod \"certified-operators-6s6t2\" (UID: \"69b806a7-b7ef-4f6a-b543-18712f6df72c\") " pod="openshift-marketplace/certified-operators-6s6t2" Oct 08 20:19:49 crc kubenswrapper[4988]: I1008 20:19:49.189520 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/69b806a7-b7ef-4f6a-b543-18712f6df72c-utilities\") pod \"certified-operators-6s6t2\" (UID: \"69b806a7-b7ef-4f6a-b543-18712f6df72c\") " pod="openshift-marketplace/certified-operators-6s6t2" Oct 08 20:19:49 crc kubenswrapper[4988]: I1008 20:19:49.189678 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/69b806a7-b7ef-4f6a-b543-18712f6df72c-catalog-content\") pod \"certified-operators-6s6t2\" (UID: \"69b806a7-b7ef-4f6a-b543-18712f6df72c\") " pod="openshift-marketplace/certified-operators-6s6t2" Oct 08 20:19:49 crc kubenswrapper[4988]: I1008 20:19:49.190072 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/69b806a7-b7ef-4f6a-b543-18712f6df72c-utilities\") pod \"certified-operators-6s6t2\" (UID: \"69b806a7-b7ef-4f6a-b543-18712f6df72c\") " pod="openshift-marketplace/certified-operators-6s6t2" Oct 08 20:19:49 crc kubenswrapper[4988]: I1008 20:19:49.190268 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/69b806a7-b7ef-4f6a-b543-18712f6df72c-catalog-content\") pod \"certified-operators-6s6t2\" (UID: \"69b806a7-b7ef-4f6a-b543-18712f6df72c\") " pod="openshift-marketplace/certified-operators-6s6t2" Oct 08 20:19:49 crc kubenswrapper[4988]: I1008 20:19:49.214404 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dczxg\" (UniqueName: \"kubernetes.io/projected/69b806a7-b7ef-4f6a-b543-18712f6df72c-kube-api-access-dczxg\") pod \"certified-operators-6s6t2\" (UID: \"69b806a7-b7ef-4f6a-b543-18712f6df72c\") " pod="openshift-marketplace/certified-operators-6s6t2" Oct 08 20:19:49 crc kubenswrapper[4988]: I1008 20:19:49.317949 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-6s6t2" Oct 08 20:19:49 crc kubenswrapper[4988]: I1008 20:19:49.843733 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-6s6t2"] Oct 08 20:19:50 crc kubenswrapper[4988]: I1008 20:19:50.738823 4988 generic.go:334] "Generic (PLEG): container finished" podID="69b806a7-b7ef-4f6a-b543-18712f6df72c" containerID="7c60a2a102250598d3662fa06b114b3f8b1b157c8543bb35e01986a47b482fbe" exitCode=0 Oct 08 20:19:50 crc kubenswrapper[4988]: I1008 20:19:50.739236 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6s6t2" event={"ID":"69b806a7-b7ef-4f6a-b543-18712f6df72c","Type":"ContainerDied","Data":"7c60a2a102250598d3662fa06b114b3f8b1b157c8543bb35e01986a47b482fbe"} Oct 08 20:19:50 crc kubenswrapper[4988]: I1008 20:19:50.739278 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6s6t2" event={"ID":"69b806a7-b7ef-4f6a-b543-18712f6df72c","Type":"ContainerStarted","Data":"5c392583844d5ccdf446c88c7b519de3327ec9b242255422f4ecc5c5e2698ca8"} Oct 08 20:19:50 crc kubenswrapper[4988]: I1008 20:19:50.743678 4988 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 08 20:19:52 crc kubenswrapper[4988]: I1008 20:19:52.766184 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6s6t2" event={"ID":"69b806a7-b7ef-4f6a-b543-18712f6df72c","Type":"ContainerStarted","Data":"94fe47f266fa5debacc92eaf9f78269b8b40111e9f5cdd4a7d3cf23d34638552"} Oct 08 20:19:53 crc kubenswrapper[4988]: I1008 20:19:53.338241 4988 patch_prober.go:28] interesting pod/machine-config-daemon-nm54f container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 20:19:53 crc kubenswrapper[4988]: I1008 20:19:53.338318 4988 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 20:19:53 crc kubenswrapper[4988]: I1008 20:19:53.786650 4988 generic.go:334] "Generic (PLEG): container finished" podID="69b806a7-b7ef-4f6a-b543-18712f6df72c" containerID="94fe47f266fa5debacc92eaf9f78269b8b40111e9f5cdd4a7d3cf23d34638552" exitCode=0 Oct 08 20:19:53 crc kubenswrapper[4988]: I1008 20:19:53.786733 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6s6t2" event={"ID":"69b806a7-b7ef-4f6a-b543-18712f6df72c","Type":"ContainerDied","Data":"94fe47f266fa5debacc92eaf9f78269b8b40111e9f5cdd4a7d3cf23d34638552"} Oct 08 20:19:54 crc kubenswrapper[4988]: I1008 20:19:54.808733 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6s6t2" event={"ID":"69b806a7-b7ef-4f6a-b543-18712f6df72c","Type":"ContainerStarted","Data":"086e397a02d2c8532417b7c9d25f3653f41d361b2f6308ca5d6327b90b6edab5"} Oct 08 20:19:54 crc kubenswrapper[4988]: I1008 20:19:54.841971 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-6s6t2" podStartSLOduration=3.3287152239999998 podStartE2EDuration="6.84193994s" podCreationTimestamp="2025-10-08 20:19:48 +0000 UTC" firstStartedPulling="2025-10-08 20:19:50.743031483 +0000 UTC m=+7736.192874273" lastFinishedPulling="2025-10-08 20:19:54.256256179 +0000 UTC m=+7739.706098989" observedRunningTime="2025-10-08 20:19:54.827661275 +0000 UTC m=+7740.277504055" watchObservedRunningTime="2025-10-08 20:19:54.84193994 +0000 UTC m=+7740.291782750" Oct 08 20:19:56 crc kubenswrapper[4988]: I1008 20:19:56.336957 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-nx4wh"] Oct 08 20:19:56 crc kubenswrapper[4988]: I1008 20:19:56.339985 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-nx4wh" Oct 08 20:19:56 crc kubenswrapper[4988]: I1008 20:19:56.356894 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-nx4wh"] Oct 08 20:19:56 crc kubenswrapper[4988]: I1008 20:19:56.470124 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r6mzx\" (UniqueName: \"kubernetes.io/projected/c9b1ff7b-bd13-4bac-af05-c4fb2fd22f90-kube-api-access-r6mzx\") pod \"redhat-operators-nx4wh\" (UID: \"c9b1ff7b-bd13-4bac-af05-c4fb2fd22f90\") " pod="openshift-marketplace/redhat-operators-nx4wh" Oct 08 20:19:56 crc kubenswrapper[4988]: I1008 20:19:56.470280 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c9b1ff7b-bd13-4bac-af05-c4fb2fd22f90-utilities\") pod \"redhat-operators-nx4wh\" (UID: \"c9b1ff7b-bd13-4bac-af05-c4fb2fd22f90\") " pod="openshift-marketplace/redhat-operators-nx4wh" Oct 08 20:19:56 crc kubenswrapper[4988]: I1008 20:19:56.470419 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c9b1ff7b-bd13-4bac-af05-c4fb2fd22f90-catalog-content\") pod \"redhat-operators-nx4wh\" (UID: \"c9b1ff7b-bd13-4bac-af05-c4fb2fd22f90\") " pod="openshift-marketplace/redhat-operators-nx4wh" Oct 08 20:19:56 crc kubenswrapper[4988]: I1008 20:19:56.572878 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r6mzx\" (UniqueName: \"kubernetes.io/projected/c9b1ff7b-bd13-4bac-af05-c4fb2fd22f90-kube-api-access-r6mzx\") pod \"redhat-operators-nx4wh\" (UID: \"c9b1ff7b-bd13-4bac-af05-c4fb2fd22f90\") " pod="openshift-marketplace/redhat-operators-nx4wh" Oct 08 20:19:56 crc kubenswrapper[4988]: I1008 20:19:56.573000 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c9b1ff7b-bd13-4bac-af05-c4fb2fd22f90-utilities\") pod \"redhat-operators-nx4wh\" (UID: \"c9b1ff7b-bd13-4bac-af05-c4fb2fd22f90\") " pod="openshift-marketplace/redhat-operators-nx4wh" Oct 08 20:19:56 crc kubenswrapper[4988]: I1008 20:19:56.573070 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c9b1ff7b-bd13-4bac-af05-c4fb2fd22f90-catalog-content\") pod \"redhat-operators-nx4wh\" (UID: \"c9b1ff7b-bd13-4bac-af05-c4fb2fd22f90\") " pod="openshift-marketplace/redhat-operators-nx4wh" Oct 08 20:19:56 crc kubenswrapper[4988]: I1008 20:19:56.573698 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c9b1ff7b-bd13-4bac-af05-c4fb2fd22f90-utilities\") pod \"redhat-operators-nx4wh\" (UID: \"c9b1ff7b-bd13-4bac-af05-c4fb2fd22f90\") " pod="openshift-marketplace/redhat-operators-nx4wh" Oct 08 20:19:56 crc kubenswrapper[4988]: I1008 20:19:56.573737 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c9b1ff7b-bd13-4bac-af05-c4fb2fd22f90-catalog-content\") pod \"redhat-operators-nx4wh\" (UID: \"c9b1ff7b-bd13-4bac-af05-c4fb2fd22f90\") " pod="openshift-marketplace/redhat-operators-nx4wh" Oct 08 20:19:56 crc kubenswrapper[4988]: I1008 20:19:56.594413 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r6mzx\" (UniqueName: \"kubernetes.io/projected/c9b1ff7b-bd13-4bac-af05-c4fb2fd22f90-kube-api-access-r6mzx\") pod \"redhat-operators-nx4wh\" (UID: \"c9b1ff7b-bd13-4bac-af05-c4fb2fd22f90\") " pod="openshift-marketplace/redhat-operators-nx4wh" Oct 08 20:19:56 crc kubenswrapper[4988]: I1008 20:19:56.684941 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-nx4wh" Oct 08 20:19:57 crc kubenswrapper[4988]: I1008 20:19:57.178546 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-nx4wh"] Oct 08 20:19:57 crc kubenswrapper[4988]: I1008 20:19:57.848307 4988 generic.go:334] "Generic (PLEG): container finished" podID="c9b1ff7b-bd13-4bac-af05-c4fb2fd22f90" containerID="e024b6428d727a5f171671364c5c41e1a8f21b8302fa1c21b0475ba22bd68440" exitCode=0 Oct 08 20:19:57 crc kubenswrapper[4988]: I1008 20:19:57.848408 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-nx4wh" event={"ID":"c9b1ff7b-bd13-4bac-af05-c4fb2fd22f90","Type":"ContainerDied","Data":"e024b6428d727a5f171671364c5c41e1a8f21b8302fa1c21b0475ba22bd68440"} Oct 08 20:19:57 crc kubenswrapper[4988]: I1008 20:19:57.848750 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-nx4wh" event={"ID":"c9b1ff7b-bd13-4bac-af05-c4fb2fd22f90","Type":"ContainerStarted","Data":"b93dff10ecb664d24939b4ea7d9ae0bc07f0b3360fa1eaf94208d4cbd9929d9e"} Oct 08 20:19:59 crc kubenswrapper[4988]: I1008 20:19:59.318655 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-6s6t2" Oct 08 20:19:59 crc kubenswrapper[4988]: I1008 20:19:59.319313 4988 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-6s6t2" Oct 08 20:19:59 crc kubenswrapper[4988]: I1008 20:19:59.413582 4988 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-6s6t2" Oct 08 20:19:59 crc kubenswrapper[4988]: I1008 20:19:59.880556 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-nx4wh" event={"ID":"c9b1ff7b-bd13-4bac-af05-c4fb2fd22f90","Type":"ContainerStarted","Data":"4b9ba564a185c7c405110ada11d0892dde1471799faad0f4e5a94a509513fc7e"} Oct 08 20:19:59 crc kubenswrapper[4988]: I1008 20:19:59.978794 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-6s6t2" Oct 08 20:20:01 crc kubenswrapper[4988]: I1008 20:20:01.136779 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-6s6t2"] Oct 08 20:20:01 crc kubenswrapper[4988]: I1008 20:20:01.900498 4988 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-6s6t2" podUID="69b806a7-b7ef-4f6a-b543-18712f6df72c" containerName="registry-server" containerID="cri-o://086e397a02d2c8532417b7c9d25f3653f41d361b2f6308ca5d6327b90b6edab5" gracePeriod=2 Oct 08 20:20:02 crc kubenswrapper[4988]: I1008 20:20:02.477412 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-6s6t2" Oct 08 20:20:02 crc kubenswrapper[4988]: I1008 20:20:02.514724 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dczxg\" (UniqueName: \"kubernetes.io/projected/69b806a7-b7ef-4f6a-b543-18712f6df72c-kube-api-access-dczxg\") pod \"69b806a7-b7ef-4f6a-b543-18712f6df72c\" (UID: \"69b806a7-b7ef-4f6a-b543-18712f6df72c\") " Oct 08 20:20:02 crc kubenswrapper[4988]: I1008 20:20:02.514911 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/69b806a7-b7ef-4f6a-b543-18712f6df72c-catalog-content\") pod \"69b806a7-b7ef-4f6a-b543-18712f6df72c\" (UID: \"69b806a7-b7ef-4f6a-b543-18712f6df72c\") " Oct 08 20:20:02 crc kubenswrapper[4988]: I1008 20:20:02.514974 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/69b806a7-b7ef-4f6a-b543-18712f6df72c-utilities\") pod \"69b806a7-b7ef-4f6a-b543-18712f6df72c\" (UID: \"69b806a7-b7ef-4f6a-b543-18712f6df72c\") " Oct 08 20:20:02 crc kubenswrapper[4988]: I1008 20:20:02.516422 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/69b806a7-b7ef-4f6a-b543-18712f6df72c-utilities" (OuterVolumeSpecName: "utilities") pod "69b806a7-b7ef-4f6a-b543-18712f6df72c" (UID: "69b806a7-b7ef-4f6a-b543-18712f6df72c"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 20:20:02 crc kubenswrapper[4988]: I1008 20:20:02.524733 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/69b806a7-b7ef-4f6a-b543-18712f6df72c-kube-api-access-dczxg" (OuterVolumeSpecName: "kube-api-access-dczxg") pod "69b806a7-b7ef-4f6a-b543-18712f6df72c" (UID: "69b806a7-b7ef-4f6a-b543-18712f6df72c"). InnerVolumeSpecName "kube-api-access-dczxg". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 20:20:02 crc kubenswrapper[4988]: I1008 20:20:02.593901 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/69b806a7-b7ef-4f6a-b543-18712f6df72c-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "69b806a7-b7ef-4f6a-b543-18712f6df72c" (UID: "69b806a7-b7ef-4f6a-b543-18712f6df72c"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 20:20:02 crc kubenswrapper[4988]: I1008 20:20:02.617672 4988 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/69b806a7-b7ef-4f6a-b543-18712f6df72c-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 08 20:20:02 crc kubenswrapper[4988]: I1008 20:20:02.617700 4988 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/69b806a7-b7ef-4f6a-b543-18712f6df72c-utilities\") on node \"crc\" DevicePath \"\"" Oct 08 20:20:02 crc kubenswrapper[4988]: I1008 20:20:02.617711 4988 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dczxg\" (UniqueName: \"kubernetes.io/projected/69b806a7-b7ef-4f6a-b543-18712f6df72c-kube-api-access-dczxg\") on node \"crc\" DevicePath \"\"" Oct 08 20:20:02 crc kubenswrapper[4988]: I1008 20:20:02.914456 4988 generic.go:334] "Generic (PLEG): container finished" podID="c9b1ff7b-bd13-4bac-af05-c4fb2fd22f90" containerID="4b9ba564a185c7c405110ada11d0892dde1471799faad0f4e5a94a509513fc7e" exitCode=0 Oct 08 20:20:02 crc kubenswrapper[4988]: I1008 20:20:02.914571 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-nx4wh" event={"ID":"c9b1ff7b-bd13-4bac-af05-c4fb2fd22f90","Type":"ContainerDied","Data":"4b9ba564a185c7c405110ada11d0892dde1471799faad0f4e5a94a509513fc7e"} Oct 08 20:20:02 crc kubenswrapper[4988]: I1008 20:20:02.919584 4988 generic.go:334] "Generic (PLEG): container finished" podID="69b806a7-b7ef-4f6a-b543-18712f6df72c" containerID="086e397a02d2c8532417b7c9d25f3653f41d361b2f6308ca5d6327b90b6edab5" exitCode=0 Oct 08 20:20:02 crc kubenswrapper[4988]: I1008 20:20:02.919663 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6s6t2" event={"ID":"69b806a7-b7ef-4f6a-b543-18712f6df72c","Type":"ContainerDied","Data":"086e397a02d2c8532417b7c9d25f3653f41d361b2f6308ca5d6327b90b6edab5"} Oct 08 20:20:02 crc kubenswrapper[4988]: I1008 20:20:02.919722 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-6s6t2" Oct 08 20:20:02 crc kubenswrapper[4988]: I1008 20:20:02.920680 4988 scope.go:117] "RemoveContainer" containerID="086e397a02d2c8532417b7c9d25f3653f41d361b2f6308ca5d6327b90b6edab5" Oct 08 20:20:02 crc kubenswrapper[4988]: I1008 20:20:02.920586 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6s6t2" event={"ID":"69b806a7-b7ef-4f6a-b543-18712f6df72c","Type":"ContainerDied","Data":"5c392583844d5ccdf446c88c7b519de3327ec9b242255422f4ecc5c5e2698ca8"} Oct 08 20:20:02 crc kubenswrapper[4988]: I1008 20:20:02.963919 4988 scope.go:117] "RemoveContainer" containerID="94fe47f266fa5debacc92eaf9f78269b8b40111e9f5cdd4a7d3cf23d34638552" Oct 08 20:20:02 crc kubenswrapper[4988]: I1008 20:20:02.987740 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-6s6t2"] Oct 08 20:20:02 crc kubenswrapper[4988]: I1008 20:20:02.993596 4988 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-6s6t2"] Oct 08 20:20:02 crc kubenswrapper[4988]: I1008 20:20:02.994712 4988 scope.go:117] "RemoveContainer" containerID="7c60a2a102250598d3662fa06b114b3f8b1b157c8543bb35e01986a47b482fbe" Oct 08 20:20:03 crc kubenswrapper[4988]: I1008 20:20:03.067299 4988 scope.go:117] "RemoveContainer" containerID="086e397a02d2c8532417b7c9d25f3653f41d361b2f6308ca5d6327b90b6edab5" Oct 08 20:20:03 crc kubenswrapper[4988]: E1008 20:20:03.068048 4988 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"086e397a02d2c8532417b7c9d25f3653f41d361b2f6308ca5d6327b90b6edab5\": container with ID starting with 086e397a02d2c8532417b7c9d25f3653f41d361b2f6308ca5d6327b90b6edab5 not found: ID does not exist" containerID="086e397a02d2c8532417b7c9d25f3653f41d361b2f6308ca5d6327b90b6edab5" Oct 08 20:20:03 crc kubenswrapper[4988]: I1008 20:20:03.068093 4988 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"086e397a02d2c8532417b7c9d25f3653f41d361b2f6308ca5d6327b90b6edab5"} err="failed to get container status \"086e397a02d2c8532417b7c9d25f3653f41d361b2f6308ca5d6327b90b6edab5\": rpc error: code = NotFound desc = could not find container \"086e397a02d2c8532417b7c9d25f3653f41d361b2f6308ca5d6327b90b6edab5\": container with ID starting with 086e397a02d2c8532417b7c9d25f3653f41d361b2f6308ca5d6327b90b6edab5 not found: ID does not exist" Oct 08 20:20:03 crc kubenswrapper[4988]: I1008 20:20:03.068123 4988 scope.go:117] "RemoveContainer" containerID="94fe47f266fa5debacc92eaf9f78269b8b40111e9f5cdd4a7d3cf23d34638552" Oct 08 20:20:03 crc kubenswrapper[4988]: E1008 20:20:03.068741 4988 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"94fe47f266fa5debacc92eaf9f78269b8b40111e9f5cdd4a7d3cf23d34638552\": container with ID starting with 94fe47f266fa5debacc92eaf9f78269b8b40111e9f5cdd4a7d3cf23d34638552 not found: ID does not exist" containerID="94fe47f266fa5debacc92eaf9f78269b8b40111e9f5cdd4a7d3cf23d34638552" Oct 08 20:20:03 crc kubenswrapper[4988]: I1008 20:20:03.068799 4988 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"94fe47f266fa5debacc92eaf9f78269b8b40111e9f5cdd4a7d3cf23d34638552"} err="failed to get container status \"94fe47f266fa5debacc92eaf9f78269b8b40111e9f5cdd4a7d3cf23d34638552\": rpc error: code = NotFound desc = could not find container \"94fe47f266fa5debacc92eaf9f78269b8b40111e9f5cdd4a7d3cf23d34638552\": container with ID starting with 94fe47f266fa5debacc92eaf9f78269b8b40111e9f5cdd4a7d3cf23d34638552 not found: ID does not exist" Oct 08 20:20:03 crc kubenswrapper[4988]: I1008 20:20:03.068834 4988 scope.go:117] "RemoveContainer" containerID="7c60a2a102250598d3662fa06b114b3f8b1b157c8543bb35e01986a47b482fbe" Oct 08 20:20:03 crc kubenswrapper[4988]: E1008 20:20:03.069249 4988 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7c60a2a102250598d3662fa06b114b3f8b1b157c8543bb35e01986a47b482fbe\": container with ID starting with 7c60a2a102250598d3662fa06b114b3f8b1b157c8543bb35e01986a47b482fbe not found: ID does not exist" containerID="7c60a2a102250598d3662fa06b114b3f8b1b157c8543bb35e01986a47b482fbe" Oct 08 20:20:03 crc kubenswrapper[4988]: I1008 20:20:03.069296 4988 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7c60a2a102250598d3662fa06b114b3f8b1b157c8543bb35e01986a47b482fbe"} err="failed to get container status \"7c60a2a102250598d3662fa06b114b3f8b1b157c8543bb35e01986a47b482fbe\": rpc error: code = NotFound desc = could not find container \"7c60a2a102250598d3662fa06b114b3f8b1b157c8543bb35e01986a47b482fbe\": container with ID starting with 7c60a2a102250598d3662fa06b114b3f8b1b157c8543bb35e01986a47b482fbe not found: ID does not exist" Oct 08 20:20:03 crc kubenswrapper[4988]: I1008 20:20:03.271860 4988 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="69b806a7-b7ef-4f6a-b543-18712f6df72c" path="/var/lib/kubelet/pods/69b806a7-b7ef-4f6a-b543-18712f6df72c/volumes" Oct 08 20:20:04 crc kubenswrapper[4988]: I1008 20:20:04.945912 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-nx4wh" event={"ID":"c9b1ff7b-bd13-4bac-af05-c4fb2fd22f90","Type":"ContainerStarted","Data":"14c73d37c66233c907d3c7d8166c9f8b4138ed9703e5bea2c46b7cfec613373a"} Oct 08 20:20:04 crc kubenswrapper[4988]: I1008 20:20:04.963708 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-nx4wh" podStartSLOduration=2.916330836 podStartE2EDuration="8.963691606s" podCreationTimestamp="2025-10-08 20:19:56 +0000 UTC" firstStartedPulling="2025-10-08 20:19:57.852368973 +0000 UTC m=+7743.302211743" lastFinishedPulling="2025-10-08 20:20:03.899729713 +0000 UTC m=+7749.349572513" observedRunningTime="2025-10-08 20:20:04.961611089 +0000 UTC m=+7750.411453869" watchObservedRunningTime="2025-10-08 20:20:04.963691606 +0000 UTC m=+7750.413534376" Oct 08 20:20:06 crc kubenswrapper[4988]: I1008 20:20:06.685411 4988 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-nx4wh" Oct 08 20:20:06 crc kubenswrapper[4988]: I1008 20:20:06.686610 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-nx4wh" Oct 08 20:20:07 crc kubenswrapper[4988]: I1008 20:20:07.737883 4988 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-nx4wh" podUID="c9b1ff7b-bd13-4bac-af05-c4fb2fd22f90" containerName="registry-server" probeResult="failure" output=< Oct 08 20:20:07 crc kubenswrapper[4988]: timeout: failed to connect service ":50051" within 1s Oct 08 20:20:07 crc kubenswrapper[4988]: > Oct 08 20:20:16 crc kubenswrapper[4988]: I1008 20:20:16.765140 4988 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-nx4wh" Oct 08 20:20:16 crc kubenswrapper[4988]: I1008 20:20:16.892628 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-nx4wh" Oct 08 20:20:17 crc kubenswrapper[4988]: I1008 20:20:17.031065 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-nx4wh"] Oct 08 20:20:18 crc kubenswrapper[4988]: I1008 20:20:18.124952 4988 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-nx4wh" podUID="c9b1ff7b-bd13-4bac-af05-c4fb2fd22f90" containerName="registry-server" containerID="cri-o://14c73d37c66233c907d3c7d8166c9f8b4138ed9703e5bea2c46b7cfec613373a" gracePeriod=2 Oct 08 20:20:18 crc kubenswrapper[4988]: I1008 20:20:18.652830 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-nx4wh" Oct 08 20:20:18 crc kubenswrapper[4988]: I1008 20:20:18.713828 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r6mzx\" (UniqueName: \"kubernetes.io/projected/c9b1ff7b-bd13-4bac-af05-c4fb2fd22f90-kube-api-access-r6mzx\") pod \"c9b1ff7b-bd13-4bac-af05-c4fb2fd22f90\" (UID: \"c9b1ff7b-bd13-4bac-af05-c4fb2fd22f90\") " Oct 08 20:20:18 crc kubenswrapper[4988]: I1008 20:20:18.714003 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c9b1ff7b-bd13-4bac-af05-c4fb2fd22f90-catalog-content\") pod \"c9b1ff7b-bd13-4bac-af05-c4fb2fd22f90\" (UID: \"c9b1ff7b-bd13-4bac-af05-c4fb2fd22f90\") " Oct 08 20:20:18 crc kubenswrapper[4988]: I1008 20:20:18.714038 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c9b1ff7b-bd13-4bac-af05-c4fb2fd22f90-utilities\") pod \"c9b1ff7b-bd13-4bac-af05-c4fb2fd22f90\" (UID: \"c9b1ff7b-bd13-4bac-af05-c4fb2fd22f90\") " Oct 08 20:20:18 crc kubenswrapper[4988]: I1008 20:20:18.715243 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c9b1ff7b-bd13-4bac-af05-c4fb2fd22f90-utilities" (OuterVolumeSpecName: "utilities") pod "c9b1ff7b-bd13-4bac-af05-c4fb2fd22f90" (UID: "c9b1ff7b-bd13-4bac-af05-c4fb2fd22f90"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 20:20:18 crc kubenswrapper[4988]: I1008 20:20:18.726611 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c9b1ff7b-bd13-4bac-af05-c4fb2fd22f90-kube-api-access-r6mzx" (OuterVolumeSpecName: "kube-api-access-r6mzx") pod "c9b1ff7b-bd13-4bac-af05-c4fb2fd22f90" (UID: "c9b1ff7b-bd13-4bac-af05-c4fb2fd22f90"). InnerVolumeSpecName "kube-api-access-r6mzx". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 20:20:18 crc kubenswrapper[4988]: I1008 20:20:18.816906 4988 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r6mzx\" (UniqueName: \"kubernetes.io/projected/c9b1ff7b-bd13-4bac-af05-c4fb2fd22f90-kube-api-access-r6mzx\") on node \"crc\" DevicePath \"\"" Oct 08 20:20:18 crc kubenswrapper[4988]: I1008 20:20:18.816955 4988 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c9b1ff7b-bd13-4bac-af05-c4fb2fd22f90-utilities\") on node \"crc\" DevicePath \"\"" Oct 08 20:20:18 crc kubenswrapper[4988]: I1008 20:20:18.817347 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c9b1ff7b-bd13-4bac-af05-c4fb2fd22f90-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "c9b1ff7b-bd13-4bac-af05-c4fb2fd22f90" (UID: "c9b1ff7b-bd13-4bac-af05-c4fb2fd22f90"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 20:20:18 crc kubenswrapper[4988]: I1008 20:20:18.919435 4988 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c9b1ff7b-bd13-4bac-af05-c4fb2fd22f90-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 08 20:20:19 crc kubenswrapper[4988]: I1008 20:20:19.139121 4988 generic.go:334] "Generic (PLEG): container finished" podID="c9b1ff7b-bd13-4bac-af05-c4fb2fd22f90" containerID="14c73d37c66233c907d3c7d8166c9f8b4138ed9703e5bea2c46b7cfec613373a" exitCode=0 Oct 08 20:20:19 crc kubenswrapper[4988]: I1008 20:20:19.139626 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-nx4wh" event={"ID":"c9b1ff7b-bd13-4bac-af05-c4fb2fd22f90","Type":"ContainerDied","Data":"14c73d37c66233c907d3c7d8166c9f8b4138ed9703e5bea2c46b7cfec613373a"} Oct 08 20:20:19 crc kubenswrapper[4988]: I1008 20:20:19.139664 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-nx4wh" event={"ID":"c9b1ff7b-bd13-4bac-af05-c4fb2fd22f90","Type":"ContainerDied","Data":"b93dff10ecb664d24939b4ea7d9ae0bc07f0b3360fa1eaf94208d4cbd9929d9e"} Oct 08 20:20:19 crc kubenswrapper[4988]: I1008 20:20:19.139688 4988 scope.go:117] "RemoveContainer" containerID="14c73d37c66233c907d3c7d8166c9f8b4138ed9703e5bea2c46b7cfec613373a" Oct 08 20:20:19 crc kubenswrapper[4988]: I1008 20:20:19.139928 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-nx4wh" Oct 08 20:20:19 crc kubenswrapper[4988]: I1008 20:20:19.183270 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-nx4wh"] Oct 08 20:20:19 crc kubenswrapper[4988]: I1008 20:20:19.184494 4988 scope.go:117] "RemoveContainer" containerID="4b9ba564a185c7c405110ada11d0892dde1471799faad0f4e5a94a509513fc7e" Oct 08 20:20:19 crc kubenswrapper[4988]: I1008 20:20:19.192148 4988 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-nx4wh"] Oct 08 20:20:19 crc kubenswrapper[4988]: I1008 20:20:19.205785 4988 scope.go:117] "RemoveContainer" containerID="e024b6428d727a5f171671364c5c41e1a8f21b8302fa1c21b0475ba22bd68440" Oct 08 20:20:19 crc kubenswrapper[4988]: I1008 20:20:19.253692 4988 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c9b1ff7b-bd13-4bac-af05-c4fb2fd22f90" path="/var/lib/kubelet/pods/c9b1ff7b-bd13-4bac-af05-c4fb2fd22f90/volumes" Oct 08 20:20:19 crc kubenswrapper[4988]: I1008 20:20:19.271631 4988 scope.go:117] "RemoveContainer" containerID="14c73d37c66233c907d3c7d8166c9f8b4138ed9703e5bea2c46b7cfec613373a" Oct 08 20:20:19 crc kubenswrapper[4988]: E1008 20:20:19.272082 4988 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"14c73d37c66233c907d3c7d8166c9f8b4138ed9703e5bea2c46b7cfec613373a\": container with ID starting with 14c73d37c66233c907d3c7d8166c9f8b4138ed9703e5bea2c46b7cfec613373a not found: ID does not exist" containerID="14c73d37c66233c907d3c7d8166c9f8b4138ed9703e5bea2c46b7cfec613373a" Oct 08 20:20:19 crc kubenswrapper[4988]: I1008 20:20:19.272115 4988 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"14c73d37c66233c907d3c7d8166c9f8b4138ed9703e5bea2c46b7cfec613373a"} err="failed to get container status \"14c73d37c66233c907d3c7d8166c9f8b4138ed9703e5bea2c46b7cfec613373a\": rpc error: code = NotFound desc = could not find container \"14c73d37c66233c907d3c7d8166c9f8b4138ed9703e5bea2c46b7cfec613373a\": container with ID starting with 14c73d37c66233c907d3c7d8166c9f8b4138ed9703e5bea2c46b7cfec613373a not found: ID does not exist" Oct 08 20:20:19 crc kubenswrapper[4988]: I1008 20:20:19.272142 4988 scope.go:117] "RemoveContainer" containerID="4b9ba564a185c7c405110ada11d0892dde1471799faad0f4e5a94a509513fc7e" Oct 08 20:20:19 crc kubenswrapper[4988]: E1008 20:20:19.272762 4988 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4b9ba564a185c7c405110ada11d0892dde1471799faad0f4e5a94a509513fc7e\": container with ID starting with 4b9ba564a185c7c405110ada11d0892dde1471799faad0f4e5a94a509513fc7e not found: ID does not exist" containerID="4b9ba564a185c7c405110ada11d0892dde1471799faad0f4e5a94a509513fc7e" Oct 08 20:20:19 crc kubenswrapper[4988]: I1008 20:20:19.272794 4988 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4b9ba564a185c7c405110ada11d0892dde1471799faad0f4e5a94a509513fc7e"} err="failed to get container status \"4b9ba564a185c7c405110ada11d0892dde1471799faad0f4e5a94a509513fc7e\": rpc error: code = NotFound desc = could not find container \"4b9ba564a185c7c405110ada11d0892dde1471799faad0f4e5a94a509513fc7e\": container with ID starting with 4b9ba564a185c7c405110ada11d0892dde1471799faad0f4e5a94a509513fc7e not found: ID does not exist" Oct 08 20:20:19 crc kubenswrapper[4988]: I1008 20:20:19.272813 4988 scope.go:117] "RemoveContainer" containerID="e024b6428d727a5f171671364c5c41e1a8f21b8302fa1c21b0475ba22bd68440" Oct 08 20:20:19 crc kubenswrapper[4988]: E1008 20:20:19.273140 4988 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e024b6428d727a5f171671364c5c41e1a8f21b8302fa1c21b0475ba22bd68440\": container with ID starting with e024b6428d727a5f171671364c5c41e1a8f21b8302fa1c21b0475ba22bd68440 not found: ID does not exist" containerID="e024b6428d727a5f171671364c5c41e1a8f21b8302fa1c21b0475ba22bd68440" Oct 08 20:20:19 crc kubenswrapper[4988]: I1008 20:20:19.273181 4988 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e024b6428d727a5f171671364c5c41e1a8f21b8302fa1c21b0475ba22bd68440"} err="failed to get container status \"e024b6428d727a5f171671364c5c41e1a8f21b8302fa1c21b0475ba22bd68440\": rpc error: code = NotFound desc = could not find container \"e024b6428d727a5f171671364c5c41e1a8f21b8302fa1c21b0475ba22bd68440\": container with ID starting with e024b6428d727a5f171671364c5c41e1a8f21b8302fa1c21b0475ba22bd68440 not found: ID does not exist" Oct 08 20:20:22 crc kubenswrapper[4988]: I1008 20:20:22.184356 4988 generic.go:334] "Generic (PLEG): container finished" podID="a5ae4d27-798c-4d58-a9d9-ea145d68d6fc" containerID="1287b21ffc98890757c354add091c2fba49fbf20ce75d9c5f31d0378d6b34efa" exitCode=0 Oct 08 20:20:22 crc kubenswrapper[4988]: I1008 20:20:22.184495 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-openstack-openstack-cell1-cpds6" event={"ID":"a5ae4d27-798c-4d58-a9d9-ea145d68d6fc","Type":"ContainerDied","Data":"1287b21ffc98890757c354add091c2fba49fbf20ce75d9c5f31d0378d6b34efa"} Oct 08 20:20:23 crc kubenswrapper[4988]: I1008 20:20:23.338325 4988 patch_prober.go:28] interesting pod/machine-config-daemon-nm54f container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 20:20:23 crc kubenswrapper[4988]: I1008 20:20:23.338720 4988 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 20:20:23 crc kubenswrapper[4988]: I1008 20:20:23.690128 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-openstack-openstack-cell1-cpds6" Oct 08 20:20:23 crc kubenswrapper[4988]: I1008 20:20:23.744045 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a5ae4d27-798c-4d58-a9d9-ea145d68d6fc-inventory\") pod \"a5ae4d27-798c-4d58-a9d9-ea145d68d6fc\" (UID: \"a5ae4d27-798c-4d58-a9d9-ea145d68d6fc\") " Oct 08 20:20:23 crc kubenswrapper[4988]: I1008 20:20:23.744133 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a5ae4d27-798c-4d58-a9d9-ea145d68d6fc-ssh-key\") pod \"a5ae4d27-798c-4d58-a9d9-ea145d68d6fc\" (UID: \"a5ae4d27-798c-4d58-a9d9-ea145d68d6fc\") " Oct 08 20:20:23 crc kubenswrapper[4988]: I1008 20:20:23.744245 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pczwn\" (UniqueName: \"kubernetes.io/projected/a5ae4d27-798c-4d58-a9d9-ea145d68d6fc-kube-api-access-pczwn\") pod \"a5ae4d27-798c-4d58-a9d9-ea145d68d6fc\" (UID: \"a5ae4d27-798c-4d58-a9d9-ea145d68d6fc\") " Oct 08 20:20:23 crc kubenswrapper[4988]: I1008 20:20:23.754776 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a5ae4d27-798c-4d58-a9d9-ea145d68d6fc-kube-api-access-pczwn" (OuterVolumeSpecName: "kube-api-access-pczwn") pod "a5ae4d27-798c-4d58-a9d9-ea145d68d6fc" (UID: "a5ae4d27-798c-4d58-a9d9-ea145d68d6fc"). InnerVolumeSpecName "kube-api-access-pczwn". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 20:20:23 crc kubenswrapper[4988]: I1008 20:20:23.781687 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a5ae4d27-798c-4d58-a9d9-ea145d68d6fc-inventory" (OuterVolumeSpecName: "inventory") pod "a5ae4d27-798c-4d58-a9d9-ea145d68d6fc" (UID: "a5ae4d27-798c-4d58-a9d9-ea145d68d6fc"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 20:20:23 crc kubenswrapper[4988]: I1008 20:20:23.789828 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a5ae4d27-798c-4d58-a9d9-ea145d68d6fc-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "a5ae4d27-798c-4d58-a9d9-ea145d68d6fc" (UID: "a5ae4d27-798c-4d58-a9d9-ea145d68d6fc"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 20:20:23 crc kubenswrapper[4988]: I1008 20:20:23.847448 4988 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a5ae4d27-798c-4d58-a9d9-ea145d68d6fc-inventory\") on node \"crc\" DevicePath \"\"" Oct 08 20:20:23 crc kubenswrapper[4988]: I1008 20:20:23.847481 4988 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a5ae4d27-798c-4d58-a9d9-ea145d68d6fc-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 08 20:20:23 crc kubenswrapper[4988]: I1008 20:20:23.847492 4988 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pczwn\" (UniqueName: \"kubernetes.io/projected/a5ae4d27-798c-4d58-a9d9-ea145d68d6fc-kube-api-access-pczwn\") on node \"crc\" DevicePath \"\"" Oct 08 20:20:24 crc kubenswrapper[4988]: I1008 20:20:24.208478 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-openstack-openstack-cell1-cpds6" event={"ID":"a5ae4d27-798c-4d58-a9d9-ea145d68d6fc","Type":"ContainerDied","Data":"fecd35d513f4c888d3df24c7eb1163f3142a7f71f0d46fa0c38fab44c6c8b23d"} Oct 08 20:20:24 crc kubenswrapper[4988]: I1008 20:20:24.208530 4988 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fecd35d513f4c888d3df24c7eb1163f3142a7f71f0d46fa0c38fab44c6c8b23d" Oct 08 20:20:24 crc kubenswrapper[4988]: I1008 20:20:24.208582 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-openstack-openstack-cell1-cpds6" Oct 08 20:20:24 crc kubenswrapper[4988]: I1008 20:20:24.332866 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/validate-network-openstack-openstack-cell1-c6dg8"] Oct 08 20:20:24 crc kubenswrapper[4988]: E1008 20:20:24.333835 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c9b1ff7b-bd13-4bac-af05-c4fb2fd22f90" containerName="extract-utilities" Oct 08 20:20:24 crc kubenswrapper[4988]: I1008 20:20:24.333861 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="c9b1ff7b-bd13-4bac-af05-c4fb2fd22f90" containerName="extract-utilities" Oct 08 20:20:24 crc kubenswrapper[4988]: E1008 20:20:24.333885 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="69b806a7-b7ef-4f6a-b543-18712f6df72c" containerName="extract-content" Oct 08 20:20:24 crc kubenswrapper[4988]: I1008 20:20:24.333898 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="69b806a7-b7ef-4f6a-b543-18712f6df72c" containerName="extract-content" Oct 08 20:20:24 crc kubenswrapper[4988]: E1008 20:20:24.333913 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c9b1ff7b-bd13-4bac-af05-c4fb2fd22f90" containerName="registry-server" Oct 08 20:20:24 crc kubenswrapper[4988]: I1008 20:20:24.333925 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="c9b1ff7b-bd13-4bac-af05-c4fb2fd22f90" containerName="registry-server" Oct 08 20:20:24 crc kubenswrapper[4988]: E1008 20:20:24.333954 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a5ae4d27-798c-4d58-a9d9-ea145d68d6fc" containerName="configure-network-openstack-openstack-cell1" Oct 08 20:20:24 crc kubenswrapper[4988]: I1008 20:20:24.333967 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="a5ae4d27-798c-4d58-a9d9-ea145d68d6fc" containerName="configure-network-openstack-openstack-cell1" Oct 08 20:20:24 crc kubenswrapper[4988]: E1008 20:20:24.333992 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="69b806a7-b7ef-4f6a-b543-18712f6df72c" containerName="extract-utilities" Oct 08 20:20:24 crc kubenswrapper[4988]: I1008 20:20:24.334004 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="69b806a7-b7ef-4f6a-b543-18712f6df72c" containerName="extract-utilities" Oct 08 20:20:24 crc kubenswrapper[4988]: E1008 20:20:24.334025 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c9b1ff7b-bd13-4bac-af05-c4fb2fd22f90" containerName="extract-content" Oct 08 20:20:24 crc kubenswrapper[4988]: I1008 20:20:24.334038 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="c9b1ff7b-bd13-4bac-af05-c4fb2fd22f90" containerName="extract-content" Oct 08 20:20:24 crc kubenswrapper[4988]: E1008 20:20:24.334094 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="69b806a7-b7ef-4f6a-b543-18712f6df72c" containerName="registry-server" Oct 08 20:20:24 crc kubenswrapper[4988]: I1008 20:20:24.334106 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="69b806a7-b7ef-4f6a-b543-18712f6df72c" containerName="registry-server" Oct 08 20:20:24 crc kubenswrapper[4988]: I1008 20:20:24.334483 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="69b806a7-b7ef-4f6a-b543-18712f6df72c" containerName="registry-server" Oct 08 20:20:24 crc kubenswrapper[4988]: I1008 20:20:24.334523 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="c9b1ff7b-bd13-4bac-af05-c4fb2fd22f90" containerName="registry-server" Oct 08 20:20:24 crc kubenswrapper[4988]: I1008 20:20:24.334564 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="a5ae4d27-798c-4d58-a9d9-ea145d68d6fc" containerName="configure-network-openstack-openstack-cell1" Oct 08 20:20:24 crc kubenswrapper[4988]: I1008 20:20:24.335734 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-openstack-openstack-cell1-c6dg8" Oct 08 20:20:24 crc kubenswrapper[4988]: I1008 20:20:24.337810 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-4sdzr" Oct 08 20:20:24 crc kubenswrapper[4988]: I1008 20:20:24.338664 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Oct 08 20:20:24 crc kubenswrapper[4988]: I1008 20:20:24.346595 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Oct 08 20:20:24 crc kubenswrapper[4988]: I1008 20:20:24.350313 4988 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 08 20:20:24 crc kubenswrapper[4988]: I1008 20:20:24.355733 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/validate-network-openstack-openstack-cell1-c6dg8"] Oct 08 20:20:24 crc kubenswrapper[4988]: I1008 20:20:24.468269 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/848dd699-16fa-4254-b3cb-bc506dae3229-inventory\") pod \"validate-network-openstack-openstack-cell1-c6dg8\" (UID: \"848dd699-16fa-4254-b3cb-bc506dae3229\") " pod="openstack/validate-network-openstack-openstack-cell1-c6dg8" Oct 08 20:20:24 crc kubenswrapper[4988]: I1008 20:20:24.468704 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/848dd699-16fa-4254-b3cb-bc506dae3229-ssh-key\") pod \"validate-network-openstack-openstack-cell1-c6dg8\" (UID: \"848dd699-16fa-4254-b3cb-bc506dae3229\") " pod="openstack/validate-network-openstack-openstack-cell1-c6dg8" Oct 08 20:20:24 crc kubenswrapper[4988]: I1008 20:20:24.468943 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vbvwd\" (UniqueName: \"kubernetes.io/projected/848dd699-16fa-4254-b3cb-bc506dae3229-kube-api-access-vbvwd\") pod \"validate-network-openstack-openstack-cell1-c6dg8\" (UID: \"848dd699-16fa-4254-b3cb-bc506dae3229\") " pod="openstack/validate-network-openstack-openstack-cell1-c6dg8" Oct 08 20:20:24 crc kubenswrapper[4988]: I1008 20:20:24.571365 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vbvwd\" (UniqueName: \"kubernetes.io/projected/848dd699-16fa-4254-b3cb-bc506dae3229-kube-api-access-vbvwd\") pod \"validate-network-openstack-openstack-cell1-c6dg8\" (UID: \"848dd699-16fa-4254-b3cb-bc506dae3229\") " pod="openstack/validate-network-openstack-openstack-cell1-c6dg8" Oct 08 20:20:24 crc kubenswrapper[4988]: I1008 20:20:24.571521 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/848dd699-16fa-4254-b3cb-bc506dae3229-inventory\") pod \"validate-network-openstack-openstack-cell1-c6dg8\" (UID: \"848dd699-16fa-4254-b3cb-bc506dae3229\") " pod="openstack/validate-network-openstack-openstack-cell1-c6dg8" Oct 08 20:20:24 crc kubenswrapper[4988]: I1008 20:20:24.571747 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/848dd699-16fa-4254-b3cb-bc506dae3229-ssh-key\") pod \"validate-network-openstack-openstack-cell1-c6dg8\" (UID: \"848dd699-16fa-4254-b3cb-bc506dae3229\") " pod="openstack/validate-network-openstack-openstack-cell1-c6dg8" Oct 08 20:20:24 crc kubenswrapper[4988]: I1008 20:20:24.577909 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/848dd699-16fa-4254-b3cb-bc506dae3229-ssh-key\") pod \"validate-network-openstack-openstack-cell1-c6dg8\" (UID: \"848dd699-16fa-4254-b3cb-bc506dae3229\") " pod="openstack/validate-network-openstack-openstack-cell1-c6dg8" Oct 08 20:20:24 crc kubenswrapper[4988]: I1008 20:20:24.586906 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/848dd699-16fa-4254-b3cb-bc506dae3229-inventory\") pod \"validate-network-openstack-openstack-cell1-c6dg8\" (UID: \"848dd699-16fa-4254-b3cb-bc506dae3229\") " pod="openstack/validate-network-openstack-openstack-cell1-c6dg8" Oct 08 20:20:24 crc kubenswrapper[4988]: I1008 20:20:24.598721 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vbvwd\" (UniqueName: \"kubernetes.io/projected/848dd699-16fa-4254-b3cb-bc506dae3229-kube-api-access-vbvwd\") pod \"validate-network-openstack-openstack-cell1-c6dg8\" (UID: \"848dd699-16fa-4254-b3cb-bc506dae3229\") " pod="openstack/validate-network-openstack-openstack-cell1-c6dg8" Oct 08 20:20:24 crc kubenswrapper[4988]: I1008 20:20:24.656764 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-openstack-openstack-cell1-c6dg8" Oct 08 20:20:25 crc kubenswrapper[4988]: I1008 20:20:25.279662 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/validate-network-openstack-openstack-cell1-c6dg8"] Oct 08 20:20:25 crc kubenswrapper[4988]: W1008 20:20:25.281161 4988 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod848dd699_16fa_4254_b3cb_bc506dae3229.slice/crio-7f3096561cc0230fc9c11cc0dc5fde7c6d0184a84410eb2d96dfb6c88b7d53d4 WatchSource:0}: Error finding container 7f3096561cc0230fc9c11cc0dc5fde7c6d0184a84410eb2d96dfb6c88b7d53d4: Status 404 returned error can't find the container with id 7f3096561cc0230fc9c11cc0dc5fde7c6d0184a84410eb2d96dfb6c88b7d53d4 Oct 08 20:20:26 crc kubenswrapper[4988]: I1008 20:20:26.241059 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-openstack-openstack-cell1-c6dg8" event={"ID":"848dd699-16fa-4254-b3cb-bc506dae3229","Type":"ContainerStarted","Data":"cf45edd226a55a04a00cd38c9c0fdbbb58d42776f432bc66d58688c1bd83e568"} Oct 08 20:20:26 crc kubenswrapper[4988]: I1008 20:20:26.241813 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-openstack-openstack-cell1-c6dg8" event={"ID":"848dd699-16fa-4254-b3cb-bc506dae3229","Type":"ContainerStarted","Data":"7f3096561cc0230fc9c11cc0dc5fde7c6d0184a84410eb2d96dfb6c88b7d53d4"} Oct 08 20:20:26 crc kubenswrapper[4988]: I1008 20:20:26.280736 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/validate-network-openstack-openstack-cell1-c6dg8" podStartSLOduration=1.719870303 podStartE2EDuration="2.280702493s" podCreationTimestamp="2025-10-08 20:20:24 +0000 UTC" firstStartedPulling="2025-10-08 20:20:25.284725244 +0000 UTC m=+7770.734568014" lastFinishedPulling="2025-10-08 20:20:25.845557434 +0000 UTC m=+7771.295400204" observedRunningTime="2025-10-08 20:20:26.267507823 +0000 UTC m=+7771.717350643" watchObservedRunningTime="2025-10-08 20:20:26.280702493 +0000 UTC m=+7771.730545303" Oct 08 20:20:33 crc kubenswrapper[4988]: I1008 20:20:33.334669 4988 generic.go:334] "Generic (PLEG): container finished" podID="848dd699-16fa-4254-b3cb-bc506dae3229" containerID="cf45edd226a55a04a00cd38c9c0fdbbb58d42776f432bc66d58688c1bd83e568" exitCode=0 Oct 08 20:20:33 crc kubenswrapper[4988]: I1008 20:20:33.334796 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-openstack-openstack-cell1-c6dg8" event={"ID":"848dd699-16fa-4254-b3cb-bc506dae3229","Type":"ContainerDied","Data":"cf45edd226a55a04a00cd38c9c0fdbbb58d42776f432bc66d58688c1bd83e568"} Oct 08 20:20:34 crc kubenswrapper[4988]: I1008 20:20:34.892878 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-openstack-openstack-cell1-c6dg8" Oct 08 20:20:35 crc kubenswrapper[4988]: I1008 20:20:35.031102 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vbvwd\" (UniqueName: \"kubernetes.io/projected/848dd699-16fa-4254-b3cb-bc506dae3229-kube-api-access-vbvwd\") pod \"848dd699-16fa-4254-b3cb-bc506dae3229\" (UID: \"848dd699-16fa-4254-b3cb-bc506dae3229\") " Oct 08 20:20:35 crc kubenswrapper[4988]: I1008 20:20:35.031218 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/848dd699-16fa-4254-b3cb-bc506dae3229-inventory\") pod \"848dd699-16fa-4254-b3cb-bc506dae3229\" (UID: \"848dd699-16fa-4254-b3cb-bc506dae3229\") " Oct 08 20:20:35 crc kubenswrapper[4988]: I1008 20:20:35.031259 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/848dd699-16fa-4254-b3cb-bc506dae3229-ssh-key\") pod \"848dd699-16fa-4254-b3cb-bc506dae3229\" (UID: \"848dd699-16fa-4254-b3cb-bc506dae3229\") " Oct 08 20:20:35 crc kubenswrapper[4988]: I1008 20:20:35.084202 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/848dd699-16fa-4254-b3cb-bc506dae3229-inventory" (OuterVolumeSpecName: "inventory") pod "848dd699-16fa-4254-b3cb-bc506dae3229" (UID: "848dd699-16fa-4254-b3cb-bc506dae3229"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 20:20:35 crc kubenswrapper[4988]: I1008 20:20:35.092156 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/848dd699-16fa-4254-b3cb-bc506dae3229-kube-api-access-vbvwd" (OuterVolumeSpecName: "kube-api-access-vbvwd") pod "848dd699-16fa-4254-b3cb-bc506dae3229" (UID: "848dd699-16fa-4254-b3cb-bc506dae3229"). InnerVolumeSpecName "kube-api-access-vbvwd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 20:20:35 crc kubenswrapper[4988]: I1008 20:20:35.098883 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/848dd699-16fa-4254-b3cb-bc506dae3229-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "848dd699-16fa-4254-b3cb-bc506dae3229" (UID: "848dd699-16fa-4254-b3cb-bc506dae3229"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 20:20:35 crc kubenswrapper[4988]: I1008 20:20:35.133666 4988 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vbvwd\" (UniqueName: \"kubernetes.io/projected/848dd699-16fa-4254-b3cb-bc506dae3229-kube-api-access-vbvwd\") on node \"crc\" DevicePath \"\"" Oct 08 20:20:35 crc kubenswrapper[4988]: I1008 20:20:35.133884 4988 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/848dd699-16fa-4254-b3cb-bc506dae3229-inventory\") on node \"crc\" DevicePath \"\"" Oct 08 20:20:35 crc kubenswrapper[4988]: I1008 20:20:35.133950 4988 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/848dd699-16fa-4254-b3cb-bc506dae3229-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 08 20:20:35 crc kubenswrapper[4988]: I1008 20:20:35.362060 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-openstack-openstack-cell1-c6dg8" event={"ID":"848dd699-16fa-4254-b3cb-bc506dae3229","Type":"ContainerDied","Data":"7f3096561cc0230fc9c11cc0dc5fde7c6d0184a84410eb2d96dfb6c88b7d53d4"} Oct 08 20:20:35 crc kubenswrapper[4988]: I1008 20:20:35.362364 4988 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7f3096561cc0230fc9c11cc0dc5fde7c6d0184a84410eb2d96dfb6c88b7d53d4" Oct 08 20:20:35 crc kubenswrapper[4988]: I1008 20:20:35.362125 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-openstack-openstack-cell1-c6dg8" Oct 08 20:20:35 crc kubenswrapper[4988]: I1008 20:20:35.446824 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/install-os-openstack-openstack-cell1-pc97n"] Oct 08 20:20:35 crc kubenswrapper[4988]: E1008 20:20:35.447270 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="848dd699-16fa-4254-b3cb-bc506dae3229" containerName="validate-network-openstack-openstack-cell1" Oct 08 20:20:35 crc kubenswrapper[4988]: I1008 20:20:35.447289 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="848dd699-16fa-4254-b3cb-bc506dae3229" containerName="validate-network-openstack-openstack-cell1" Oct 08 20:20:35 crc kubenswrapper[4988]: I1008 20:20:35.447533 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="848dd699-16fa-4254-b3cb-bc506dae3229" containerName="validate-network-openstack-openstack-cell1" Oct 08 20:20:35 crc kubenswrapper[4988]: I1008 20:20:35.448237 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-openstack-openstack-cell1-pc97n" Oct 08 20:20:35 crc kubenswrapper[4988]: I1008 20:20:35.455757 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Oct 08 20:20:35 crc kubenswrapper[4988]: I1008 20:20:35.455952 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Oct 08 20:20:35 crc kubenswrapper[4988]: I1008 20:20:35.455754 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-4sdzr" Oct 08 20:20:35 crc kubenswrapper[4988]: I1008 20:20:35.455839 4988 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 08 20:20:35 crc kubenswrapper[4988]: I1008 20:20:35.472372 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-os-openstack-openstack-cell1-pc97n"] Oct 08 20:20:35 crc kubenswrapper[4988]: I1008 20:20:35.542142 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-knqps\" (UniqueName: \"kubernetes.io/projected/469f9f63-7654-468b-9ad8-c620f0c1b18d-kube-api-access-knqps\") pod \"install-os-openstack-openstack-cell1-pc97n\" (UID: \"469f9f63-7654-468b-9ad8-c620f0c1b18d\") " pod="openstack/install-os-openstack-openstack-cell1-pc97n" Oct 08 20:20:35 crc kubenswrapper[4988]: I1008 20:20:35.542228 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/469f9f63-7654-468b-9ad8-c620f0c1b18d-inventory\") pod \"install-os-openstack-openstack-cell1-pc97n\" (UID: \"469f9f63-7654-468b-9ad8-c620f0c1b18d\") " pod="openstack/install-os-openstack-openstack-cell1-pc97n" Oct 08 20:20:35 crc kubenswrapper[4988]: I1008 20:20:35.542311 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/469f9f63-7654-468b-9ad8-c620f0c1b18d-ssh-key\") pod \"install-os-openstack-openstack-cell1-pc97n\" (UID: \"469f9f63-7654-468b-9ad8-c620f0c1b18d\") " pod="openstack/install-os-openstack-openstack-cell1-pc97n" Oct 08 20:20:35 crc kubenswrapper[4988]: I1008 20:20:35.643954 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-knqps\" (UniqueName: \"kubernetes.io/projected/469f9f63-7654-468b-9ad8-c620f0c1b18d-kube-api-access-knqps\") pod \"install-os-openstack-openstack-cell1-pc97n\" (UID: \"469f9f63-7654-468b-9ad8-c620f0c1b18d\") " pod="openstack/install-os-openstack-openstack-cell1-pc97n" Oct 08 20:20:35 crc kubenswrapper[4988]: I1008 20:20:35.644055 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/469f9f63-7654-468b-9ad8-c620f0c1b18d-inventory\") pod \"install-os-openstack-openstack-cell1-pc97n\" (UID: \"469f9f63-7654-468b-9ad8-c620f0c1b18d\") " pod="openstack/install-os-openstack-openstack-cell1-pc97n" Oct 08 20:20:35 crc kubenswrapper[4988]: I1008 20:20:35.644151 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/469f9f63-7654-468b-9ad8-c620f0c1b18d-ssh-key\") pod \"install-os-openstack-openstack-cell1-pc97n\" (UID: \"469f9f63-7654-468b-9ad8-c620f0c1b18d\") " pod="openstack/install-os-openstack-openstack-cell1-pc97n" Oct 08 20:20:35 crc kubenswrapper[4988]: I1008 20:20:35.650055 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/469f9f63-7654-468b-9ad8-c620f0c1b18d-ssh-key\") pod \"install-os-openstack-openstack-cell1-pc97n\" (UID: \"469f9f63-7654-468b-9ad8-c620f0c1b18d\") " pod="openstack/install-os-openstack-openstack-cell1-pc97n" Oct 08 20:20:35 crc kubenswrapper[4988]: I1008 20:20:35.652444 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/469f9f63-7654-468b-9ad8-c620f0c1b18d-inventory\") pod \"install-os-openstack-openstack-cell1-pc97n\" (UID: \"469f9f63-7654-468b-9ad8-c620f0c1b18d\") " pod="openstack/install-os-openstack-openstack-cell1-pc97n" Oct 08 20:20:35 crc kubenswrapper[4988]: I1008 20:20:35.663082 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-knqps\" (UniqueName: \"kubernetes.io/projected/469f9f63-7654-468b-9ad8-c620f0c1b18d-kube-api-access-knqps\") pod \"install-os-openstack-openstack-cell1-pc97n\" (UID: \"469f9f63-7654-468b-9ad8-c620f0c1b18d\") " pod="openstack/install-os-openstack-openstack-cell1-pc97n" Oct 08 20:20:35 crc kubenswrapper[4988]: I1008 20:20:35.772191 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-openstack-openstack-cell1-pc97n" Oct 08 20:20:36 crc kubenswrapper[4988]: I1008 20:20:36.136904 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-os-openstack-openstack-cell1-pc97n"] Oct 08 20:20:36 crc kubenswrapper[4988]: W1008 20:20:36.141692 4988 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod469f9f63_7654_468b_9ad8_c620f0c1b18d.slice/crio-6c75be4503ad3a2a667cba75f14421852e5eb605d74aa8891361b834fae5e7f6 WatchSource:0}: Error finding container 6c75be4503ad3a2a667cba75f14421852e5eb605d74aa8891361b834fae5e7f6: Status 404 returned error can't find the container with id 6c75be4503ad3a2a667cba75f14421852e5eb605d74aa8891361b834fae5e7f6 Oct 08 20:20:36 crc kubenswrapper[4988]: I1008 20:20:36.373600 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-openstack-openstack-cell1-pc97n" event={"ID":"469f9f63-7654-468b-9ad8-c620f0c1b18d","Type":"ContainerStarted","Data":"6c75be4503ad3a2a667cba75f14421852e5eb605d74aa8891361b834fae5e7f6"} Oct 08 20:20:37 crc kubenswrapper[4988]: I1008 20:20:37.388325 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-openstack-openstack-cell1-pc97n" event={"ID":"469f9f63-7654-468b-9ad8-c620f0c1b18d","Type":"ContainerStarted","Data":"ce605aa3321cfb2cb7915cb37c9f1f3d128ed86ab4fbd6c33d4761b84838a4a7"} Oct 08 20:20:37 crc kubenswrapper[4988]: I1008 20:20:37.421307 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/install-os-openstack-openstack-cell1-pc97n" podStartSLOduration=1.657134125 podStartE2EDuration="2.421277985s" podCreationTimestamp="2025-10-08 20:20:35 +0000 UTC" firstStartedPulling="2025-10-08 20:20:36.14484197 +0000 UTC m=+7781.594684740" lastFinishedPulling="2025-10-08 20:20:36.9089858 +0000 UTC m=+7782.358828600" observedRunningTime="2025-10-08 20:20:37.408539539 +0000 UTC m=+7782.858382369" watchObservedRunningTime="2025-10-08 20:20:37.421277985 +0000 UTC m=+7782.871120785" Oct 08 20:20:53 crc kubenswrapper[4988]: I1008 20:20:53.338494 4988 patch_prober.go:28] interesting pod/machine-config-daemon-nm54f container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 20:20:53 crc kubenswrapper[4988]: I1008 20:20:53.339130 4988 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 20:20:53 crc kubenswrapper[4988]: I1008 20:20:53.339202 4988 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" Oct 08 20:20:53 crc kubenswrapper[4988]: I1008 20:20:53.340610 4988 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"7541195e86b98682ea101ebff2b19f543aa9bab45dd1665e8f710922ad18c5e7"} pod="openshift-machine-config-operator/machine-config-daemon-nm54f" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 08 20:20:53 crc kubenswrapper[4988]: I1008 20:20:53.340717 4988 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" containerName="machine-config-daemon" containerID="cri-o://7541195e86b98682ea101ebff2b19f543aa9bab45dd1665e8f710922ad18c5e7" gracePeriod=600 Oct 08 20:20:53 crc kubenswrapper[4988]: E1008 20:20:53.476550 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nm54f_openshift-machine-config-operator(f328a20b-5283-407e-8a93-2ab2da599f18)\"" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" Oct 08 20:20:53 crc kubenswrapper[4988]: I1008 20:20:53.610336 4988 generic.go:334] "Generic (PLEG): container finished" podID="f328a20b-5283-407e-8a93-2ab2da599f18" containerID="7541195e86b98682ea101ebff2b19f543aa9bab45dd1665e8f710922ad18c5e7" exitCode=0 Oct 08 20:20:53 crc kubenswrapper[4988]: I1008 20:20:53.610418 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" event={"ID":"f328a20b-5283-407e-8a93-2ab2da599f18","Type":"ContainerDied","Data":"7541195e86b98682ea101ebff2b19f543aa9bab45dd1665e8f710922ad18c5e7"} Oct 08 20:20:53 crc kubenswrapper[4988]: I1008 20:20:53.610473 4988 scope.go:117] "RemoveContainer" containerID="f42539de8c8a890b6e35787c0007fb9c8dd69cabc2b007044b887c7ef2f0610b" Oct 08 20:20:53 crc kubenswrapper[4988]: I1008 20:20:53.611101 4988 scope.go:117] "RemoveContainer" containerID="7541195e86b98682ea101ebff2b19f543aa9bab45dd1665e8f710922ad18c5e7" Oct 08 20:20:53 crc kubenswrapper[4988]: E1008 20:20:53.611406 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nm54f_openshift-machine-config-operator(f328a20b-5283-407e-8a93-2ab2da599f18)\"" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" Oct 08 20:21:05 crc kubenswrapper[4988]: I1008 20:21:05.248289 4988 scope.go:117] "RemoveContainer" containerID="7541195e86b98682ea101ebff2b19f543aa9bab45dd1665e8f710922ad18c5e7" Oct 08 20:21:05 crc kubenswrapper[4988]: E1008 20:21:05.249071 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nm54f_openshift-machine-config-operator(f328a20b-5283-407e-8a93-2ab2da599f18)\"" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" Oct 08 20:21:17 crc kubenswrapper[4988]: I1008 20:21:17.239186 4988 scope.go:117] "RemoveContainer" containerID="7541195e86b98682ea101ebff2b19f543aa9bab45dd1665e8f710922ad18c5e7" Oct 08 20:21:17 crc kubenswrapper[4988]: E1008 20:21:17.240176 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nm54f_openshift-machine-config-operator(f328a20b-5283-407e-8a93-2ab2da599f18)\"" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" Oct 08 20:21:23 crc kubenswrapper[4988]: I1008 20:21:23.979120 4988 generic.go:334] "Generic (PLEG): container finished" podID="469f9f63-7654-468b-9ad8-c620f0c1b18d" containerID="ce605aa3321cfb2cb7915cb37c9f1f3d128ed86ab4fbd6c33d4761b84838a4a7" exitCode=0 Oct 08 20:21:23 crc kubenswrapper[4988]: I1008 20:21:23.979625 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-openstack-openstack-cell1-pc97n" event={"ID":"469f9f63-7654-468b-9ad8-c620f0c1b18d","Type":"ContainerDied","Data":"ce605aa3321cfb2cb7915cb37c9f1f3d128ed86ab4fbd6c33d4761b84838a4a7"} Oct 08 20:21:25 crc kubenswrapper[4988]: I1008 20:21:25.524734 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-openstack-openstack-cell1-pc97n" Oct 08 20:21:25 crc kubenswrapper[4988]: I1008 20:21:25.606149 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-knqps\" (UniqueName: \"kubernetes.io/projected/469f9f63-7654-468b-9ad8-c620f0c1b18d-kube-api-access-knqps\") pod \"469f9f63-7654-468b-9ad8-c620f0c1b18d\" (UID: \"469f9f63-7654-468b-9ad8-c620f0c1b18d\") " Oct 08 20:21:25 crc kubenswrapper[4988]: I1008 20:21:25.606217 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/469f9f63-7654-468b-9ad8-c620f0c1b18d-ssh-key\") pod \"469f9f63-7654-468b-9ad8-c620f0c1b18d\" (UID: \"469f9f63-7654-468b-9ad8-c620f0c1b18d\") " Oct 08 20:21:25 crc kubenswrapper[4988]: I1008 20:21:25.606326 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/469f9f63-7654-468b-9ad8-c620f0c1b18d-inventory\") pod \"469f9f63-7654-468b-9ad8-c620f0c1b18d\" (UID: \"469f9f63-7654-468b-9ad8-c620f0c1b18d\") " Oct 08 20:21:25 crc kubenswrapper[4988]: I1008 20:21:25.614821 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/469f9f63-7654-468b-9ad8-c620f0c1b18d-kube-api-access-knqps" (OuterVolumeSpecName: "kube-api-access-knqps") pod "469f9f63-7654-468b-9ad8-c620f0c1b18d" (UID: "469f9f63-7654-468b-9ad8-c620f0c1b18d"). InnerVolumeSpecName "kube-api-access-knqps". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 20:21:25 crc kubenswrapper[4988]: I1008 20:21:25.655242 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/469f9f63-7654-468b-9ad8-c620f0c1b18d-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "469f9f63-7654-468b-9ad8-c620f0c1b18d" (UID: "469f9f63-7654-468b-9ad8-c620f0c1b18d"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 20:21:25 crc kubenswrapper[4988]: I1008 20:21:25.657748 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/469f9f63-7654-468b-9ad8-c620f0c1b18d-inventory" (OuterVolumeSpecName: "inventory") pod "469f9f63-7654-468b-9ad8-c620f0c1b18d" (UID: "469f9f63-7654-468b-9ad8-c620f0c1b18d"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 20:21:25 crc kubenswrapper[4988]: I1008 20:21:25.710463 4988 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-knqps\" (UniqueName: \"kubernetes.io/projected/469f9f63-7654-468b-9ad8-c620f0c1b18d-kube-api-access-knqps\") on node \"crc\" DevicePath \"\"" Oct 08 20:21:25 crc kubenswrapper[4988]: I1008 20:21:25.710491 4988 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/469f9f63-7654-468b-9ad8-c620f0c1b18d-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 08 20:21:25 crc kubenswrapper[4988]: I1008 20:21:25.710502 4988 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/469f9f63-7654-468b-9ad8-c620f0c1b18d-inventory\") on node \"crc\" DevicePath \"\"" Oct 08 20:21:26 crc kubenswrapper[4988]: I1008 20:21:26.007922 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-openstack-openstack-cell1-pc97n" event={"ID":"469f9f63-7654-468b-9ad8-c620f0c1b18d","Type":"ContainerDied","Data":"6c75be4503ad3a2a667cba75f14421852e5eb605d74aa8891361b834fae5e7f6"} Oct 08 20:21:26 crc kubenswrapper[4988]: I1008 20:21:26.008465 4988 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6c75be4503ad3a2a667cba75f14421852e5eb605d74aa8891361b834fae5e7f6" Oct 08 20:21:26 crc kubenswrapper[4988]: I1008 20:21:26.007980 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-openstack-openstack-cell1-pc97n" Oct 08 20:21:26 crc kubenswrapper[4988]: I1008 20:21:26.138291 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-os-openstack-openstack-cell1-2k4qc"] Oct 08 20:21:26 crc kubenswrapper[4988]: E1008 20:21:26.138886 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="469f9f63-7654-468b-9ad8-c620f0c1b18d" containerName="install-os-openstack-openstack-cell1" Oct 08 20:21:26 crc kubenswrapper[4988]: I1008 20:21:26.138908 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="469f9f63-7654-468b-9ad8-c620f0c1b18d" containerName="install-os-openstack-openstack-cell1" Oct 08 20:21:26 crc kubenswrapper[4988]: I1008 20:21:26.139200 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="469f9f63-7654-468b-9ad8-c620f0c1b18d" containerName="install-os-openstack-openstack-cell1" Oct 08 20:21:26 crc kubenswrapper[4988]: I1008 20:21:26.140132 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-openstack-openstack-cell1-2k4qc" Oct 08 20:21:26 crc kubenswrapper[4988]: I1008 20:21:26.143237 4988 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 08 20:21:26 crc kubenswrapper[4988]: I1008 20:21:26.143409 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Oct 08 20:21:26 crc kubenswrapper[4988]: I1008 20:21:26.143887 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Oct 08 20:21:26 crc kubenswrapper[4988]: I1008 20:21:26.144347 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-4sdzr" Oct 08 20:21:26 crc kubenswrapper[4988]: I1008 20:21:26.160222 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-openstack-openstack-cell1-2k4qc"] Oct 08 20:21:26 crc kubenswrapper[4988]: I1008 20:21:26.221379 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5fc3787f-07bb-4c4b-b97d-e3057a1107e0-inventory\") pod \"configure-os-openstack-openstack-cell1-2k4qc\" (UID: \"5fc3787f-07bb-4c4b-b97d-e3057a1107e0\") " pod="openstack/configure-os-openstack-openstack-cell1-2k4qc" Oct 08 20:21:26 crc kubenswrapper[4988]: I1008 20:21:26.221576 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5fc3787f-07bb-4c4b-b97d-e3057a1107e0-ssh-key\") pod \"configure-os-openstack-openstack-cell1-2k4qc\" (UID: \"5fc3787f-07bb-4c4b-b97d-e3057a1107e0\") " pod="openstack/configure-os-openstack-openstack-cell1-2k4qc" Oct 08 20:21:26 crc kubenswrapper[4988]: I1008 20:21:26.221743 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t7rnj\" (UniqueName: \"kubernetes.io/projected/5fc3787f-07bb-4c4b-b97d-e3057a1107e0-kube-api-access-t7rnj\") pod \"configure-os-openstack-openstack-cell1-2k4qc\" (UID: \"5fc3787f-07bb-4c4b-b97d-e3057a1107e0\") " pod="openstack/configure-os-openstack-openstack-cell1-2k4qc" Oct 08 20:21:26 crc kubenswrapper[4988]: I1008 20:21:26.324613 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5fc3787f-07bb-4c4b-b97d-e3057a1107e0-inventory\") pod \"configure-os-openstack-openstack-cell1-2k4qc\" (UID: \"5fc3787f-07bb-4c4b-b97d-e3057a1107e0\") " pod="openstack/configure-os-openstack-openstack-cell1-2k4qc" Oct 08 20:21:26 crc kubenswrapper[4988]: I1008 20:21:26.324819 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5fc3787f-07bb-4c4b-b97d-e3057a1107e0-ssh-key\") pod \"configure-os-openstack-openstack-cell1-2k4qc\" (UID: \"5fc3787f-07bb-4c4b-b97d-e3057a1107e0\") " pod="openstack/configure-os-openstack-openstack-cell1-2k4qc" Oct 08 20:21:26 crc kubenswrapper[4988]: I1008 20:21:26.324921 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t7rnj\" (UniqueName: \"kubernetes.io/projected/5fc3787f-07bb-4c4b-b97d-e3057a1107e0-kube-api-access-t7rnj\") pod \"configure-os-openstack-openstack-cell1-2k4qc\" (UID: \"5fc3787f-07bb-4c4b-b97d-e3057a1107e0\") " pod="openstack/configure-os-openstack-openstack-cell1-2k4qc" Oct 08 20:21:26 crc kubenswrapper[4988]: I1008 20:21:26.329092 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5fc3787f-07bb-4c4b-b97d-e3057a1107e0-inventory\") pod \"configure-os-openstack-openstack-cell1-2k4qc\" (UID: \"5fc3787f-07bb-4c4b-b97d-e3057a1107e0\") " pod="openstack/configure-os-openstack-openstack-cell1-2k4qc" Oct 08 20:21:26 crc kubenswrapper[4988]: I1008 20:21:26.330715 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5fc3787f-07bb-4c4b-b97d-e3057a1107e0-ssh-key\") pod \"configure-os-openstack-openstack-cell1-2k4qc\" (UID: \"5fc3787f-07bb-4c4b-b97d-e3057a1107e0\") " pod="openstack/configure-os-openstack-openstack-cell1-2k4qc" Oct 08 20:21:26 crc kubenswrapper[4988]: I1008 20:21:26.346125 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t7rnj\" (UniqueName: \"kubernetes.io/projected/5fc3787f-07bb-4c4b-b97d-e3057a1107e0-kube-api-access-t7rnj\") pod \"configure-os-openstack-openstack-cell1-2k4qc\" (UID: \"5fc3787f-07bb-4c4b-b97d-e3057a1107e0\") " pod="openstack/configure-os-openstack-openstack-cell1-2k4qc" Oct 08 20:21:26 crc kubenswrapper[4988]: I1008 20:21:26.471761 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-openstack-openstack-cell1-2k4qc" Oct 08 20:21:27 crc kubenswrapper[4988]: I1008 20:21:27.054270 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-openstack-openstack-cell1-2k4qc"] Oct 08 20:21:28 crc kubenswrapper[4988]: I1008 20:21:28.041305 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-openstack-openstack-cell1-2k4qc" event={"ID":"5fc3787f-07bb-4c4b-b97d-e3057a1107e0","Type":"ContainerStarted","Data":"04dbe5de68e0f86120dd6c73b0b71bc57e86aa21ae9863313feeff9ab3125d54"} Oct 08 20:21:28 crc kubenswrapper[4988]: I1008 20:21:28.041923 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-openstack-openstack-cell1-2k4qc" event={"ID":"5fc3787f-07bb-4c4b-b97d-e3057a1107e0","Type":"ContainerStarted","Data":"66d62c9e8282ba760d52a03642b1c63dc0e6404a03e4ab87f1c6ba1dbd3a550d"} Oct 08 20:21:28 crc kubenswrapper[4988]: I1008 20:21:28.072974 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-os-openstack-openstack-cell1-2k4qc" podStartSLOduration=1.669960372 podStartE2EDuration="2.072941188s" podCreationTimestamp="2025-10-08 20:21:26 +0000 UTC" firstStartedPulling="2025-10-08 20:21:27.069074368 +0000 UTC m=+7832.518917138" lastFinishedPulling="2025-10-08 20:21:27.472055144 +0000 UTC m=+7832.921897954" observedRunningTime="2025-10-08 20:21:28.061481383 +0000 UTC m=+7833.511324213" watchObservedRunningTime="2025-10-08 20:21:28.072941188 +0000 UTC m=+7833.522783998" Oct 08 20:21:30 crc kubenswrapper[4988]: I1008 20:21:30.239626 4988 scope.go:117] "RemoveContainer" containerID="7541195e86b98682ea101ebff2b19f543aa9bab45dd1665e8f710922ad18c5e7" Oct 08 20:21:30 crc kubenswrapper[4988]: E1008 20:21:30.240327 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nm54f_openshift-machine-config-operator(f328a20b-5283-407e-8a93-2ab2da599f18)\"" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" Oct 08 20:21:43 crc kubenswrapper[4988]: I1008 20:21:43.238494 4988 scope.go:117] "RemoveContainer" containerID="7541195e86b98682ea101ebff2b19f543aa9bab45dd1665e8f710922ad18c5e7" Oct 08 20:21:43 crc kubenswrapper[4988]: E1008 20:21:43.239247 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nm54f_openshift-machine-config-operator(f328a20b-5283-407e-8a93-2ab2da599f18)\"" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" Oct 08 20:21:58 crc kubenswrapper[4988]: I1008 20:21:58.237833 4988 scope.go:117] "RemoveContainer" containerID="7541195e86b98682ea101ebff2b19f543aa9bab45dd1665e8f710922ad18c5e7" Oct 08 20:21:58 crc kubenswrapper[4988]: E1008 20:21:58.238656 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nm54f_openshift-machine-config-operator(f328a20b-5283-407e-8a93-2ab2da599f18)\"" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" Oct 08 20:22:13 crc kubenswrapper[4988]: I1008 20:22:13.240520 4988 scope.go:117] "RemoveContainer" containerID="7541195e86b98682ea101ebff2b19f543aa9bab45dd1665e8f710922ad18c5e7" Oct 08 20:22:13 crc kubenswrapper[4988]: E1008 20:22:13.241942 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nm54f_openshift-machine-config-operator(f328a20b-5283-407e-8a93-2ab2da599f18)\"" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" Oct 08 20:22:16 crc kubenswrapper[4988]: I1008 20:22:16.570503 4988 generic.go:334] "Generic (PLEG): container finished" podID="5fc3787f-07bb-4c4b-b97d-e3057a1107e0" containerID="04dbe5de68e0f86120dd6c73b0b71bc57e86aa21ae9863313feeff9ab3125d54" exitCode=0 Oct 08 20:22:16 crc kubenswrapper[4988]: I1008 20:22:16.570545 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-openstack-openstack-cell1-2k4qc" event={"ID":"5fc3787f-07bb-4c4b-b97d-e3057a1107e0","Type":"ContainerDied","Data":"04dbe5de68e0f86120dd6c73b0b71bc57e86aa21ae9863313feeff9ab3125d54"} Oct 08 20:22:18 crc kubenswrapper[4988]: I1008 20:22:18.125429 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-openstack-openstack-cell1-2k4qc" Oct 08 20:22:18 crc kubenswrapper[4988]: I1008 20:22:18.178606 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-t7rnj\" (UniqueName: \"kubernetes.io/projected/5fc3787f-07bb-4c4b-b97d-e3057a1107e0-kube-api-access-t7rnj\") pod \"5fc3787f-07bb-4c4b-b97d-e3057a1107e0\" (UID: \"5fc3787f-07bb-4c4b-b97d-e3057a1107e0\") " Oct 08 20:22:18 crc kubenswrapper[4988]: I1008 20:22:18.178840 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5fc3787f-07bb-4c4b-b97d-e3057a1107e0-ssh-key\") pod \"5fc3787f-07bb-4c4b-b97d-e3057a1107e0\" (UID: \"5fc3787f-07bb-4c4b-b97d-e3057a1107e0\") " Oct 08 20:22:18 crc kubenswrapper[4988]: I1008 20:22:18.178895 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5fc3787f-07bb-4c4b-b97d-e3057a1107e0-inventory\") pod \"5fc3787f-07bb-4c4b-b97d-e3057a1107e0\" (UID: \"5fc3787f-07bb-4c4b-b97d-e3057a1107e0\") " Oct 08 20:22:18 crc kubenswrapper[4988]: I1008 20:22:18.186725 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5fc3787f-07bb-4c4b-b97d-e3057a1107e0-kube-api-access-t7rnj" (OuterVolumeSpecName: "kube-api-access-t7rnj") pod "5fc3787f-07bb-4c4b-b97d-e3057a1107e0" (UID: "5fc3787f-07bb-4c4b-b97d-e3057a1107e0"). InnerVolumeSpecName "kube-api-access-t7rnj". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 20:22:18 crc kubenswrapper[4988]: I1008 20:22:18.211005 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fc3787f-07bb-4c4b-b97d-e3057a1107e0-inventory" (OuterVolumeSpecName: "inventory") pod "5fc3787f-07bb-4c4b-b97d-e3057a1107e0" (UID: "5fc3787f-07bb-4c4b-b97d-e3057a1107e0"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 20:22:18 crc kubenswrapper[4988]: I1008 20:22:18.225739 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fc3787f-07bb-4c4b-b97d-e3057a1107e0-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "5fc3787f-07bb-4c4b-b97d-e3057a1107e0" (UID: "5fc3787f-07bb-4c4b-b97d-e3057a1107e0"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 20:22:18 crc kubenswrapper[4988]: I1008 20:22:18.281898 4988 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5fc3787f-07bb-4c4b-b97d-e3057a1107e0-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 08 20:22:18 crc kubenswrapper[4988]: I1008 20:22:18.281943 4988 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5fc3787f-07bb-4c4b-b97d-e3057a1107e0-inventory\") on node \"crc\" DevicePath \"\"" Oct 08 20:22:18 crc kubenswrapper[4988]: I1008 20:22:18.281959 4988 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-t7rnj\" (UniqueName: \"kubernetes.io/projected/5fc3787f-07bb-4c4b-b97d-e3057a1107e0-kube-api-access-t7rnj\") on node \"crc\" DevicePath \"\"" Oct 08 20:22:18 crc kubenswrapper[4988]: I1008 20:22:18.632161 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-openstack-openstack-cell1-2k4qc" Oct 08 20:22:18 crc kubenswrapper[4988]: I1008 20:22:18.632635 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-openstack-openstack-cell1-2k4qc" event={"ID":"5fc3787f-07bb-4c4b-b97d-e3057a1107e0","Type":"ContainerDied","Data":"66d62c9e8282ba760d52a03642b1c63dc0e6404a03e4ab87f1c6ba1dbd3a550d"} Oct 08 20:22:18 crc kubenswrapper[4988]: I1008 20:22:18.632711 4988 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="66d62c9e8282ba760d52a03642b1c63dc0e6404a03e4ab87f1c6ba1dbd3a550d" Oct 08 20:22:18 crc kubenswrapper[4988]: I1008 20:22:18.701007 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ssh-known-hosts-openstack-2qjln"] Oct 08 20:22:18 crc kubenswrapper[4988]: E1008 20:22:18.701721 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5fc3787f-07bb-4c4b-b97d-e3057a1107e0" containerName="configure-os-openstack-openstack-cell1" Oct 08 20:22:18 crc kubenswrapper[4988]: I1008 20:22:18.701742 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="5fc3787f-07bb-4c4b-b97d-e3057a1107e0" containerName="configure-os-openstack-openstack-cell1" Oct 08 20:22:18 crc kubenswrapper[4988]: I1008 20:22:18.702007 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="5fc3787f-07bb-4c4b-b97d-e3057a1107e0" containerName="configure-os-openstack-openstack-cell1" Oct 08 20:22:18 crc kubenswrapper[4988]: I1008 20:22:18.702770 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-openstack-2qjln" Oct 08 20:22:18 crc kubenswrapper[4988]: I1008 20:22:18.709170 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Oct 08 20:22:18 crc kubenswrapper[4988]: I1008 20:22:18.709414 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-4sdzr" Oct 08 20:22:18 crc kubenswrapper[4988]: I1008 20:22:18.709619 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Oct 08 20:22:18 crc kubenswrapper[4988]: I1008 20:22:18.709957 4988 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 08 20:22:18 crc kubenswrapper[4988]: E1008 20:22:18.737347 4988 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5fc3787f_07bb_4c4b_b97d_e3057a1107e0.slice/crio-66d62c9e8282ba760d52a03642b1c63dc0e6404a03e4ab87f1c6ba1dbd3a550d\": RecentStats: unable to find data in memory cache]" Oct 08 20:22:18 crc kubenswrapper[4988]: I1008 20:22:18.753698 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ssh-known-hosts-openstack-2qjln"] Oct 08 20:22:18 crc kubenswrapper[4988]: I1008 20:22:18.806950 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/b1cf5326-c7d5-4a5f-9a48-159b6482aa0a-ssh-key-openstack-cell1\") pod \"ssh-known-hosts-openstack-2qjln\" (UID: \"b1cf5326-c7d5-4a5f-9a48-159b6482aa0a\") " pod="openstack/ssh-known-hosts-openstack-2qjln" Oct 08 20:22:18 crc kubenswrapper[4988]: I1008 20:22:18.807066 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/b1cf5326-c7d5-4a5f-9a48-159b6482aa0a-inventory-0\") pod \"ssh-known-hosts-openstack-2qjln\" (UID: \"b1cf5326-c7d5-4a5f-9a48-159b6482aa0a\") " pod="openstack/ssh-known-hosts-openstack-2qjln" Oct 08 20:22:18 crc kubenswrapper[4988]: I1008 20:22:18.807084 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d497f\" (UniqueName: \"kubernetes.io/projected/b1cf5326-c7d5-4a5f-9a48-159b6482aa0a-kube-api-access-d497f\") pod \"ssh-known-hosts-openstack-2qjln\" (UID: \"b1cf5326-c7d5-4a5f-9a48-159b6482aa0a\") " pod="openstack/ssh-known-hosts-openstack-2qjln" Oct 08 20:22:18 crc kubenswrapper[4988]: I1008 20:22:18.908377 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/b1cf5326-c7d5-4a5f-9a48-159b6482aa0a-inventory-0\") pod \"ssh-known-hosts-openstack-2qjln\" (UID: \"b1cf5326-c7d5-4a5f-9a48-159b6482aa0a\") " pod="openstack/ssh-known-hosts-openstack-2qjln" Oct 08 20:22:18 crc kubenswrapper[4988]: I1008 20:22:18.908445 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d497f\" (UniqueName: \"kubernetes.io/projected/b1cf5326-c7d5-4a5f-9a48-159b6482aa0a-kube-api-access-d497f\") pod \"ssh-known-hosts-openstack-2qjln\" (UID: \"b1cf5326-c7d5-4a5f-9a48-159b6482aa0a\") " pod="openstack/ssh-known-hosts-openstack-2qjln" Oct 08 20:22:18 crc kubenswrapper[4988]: I1008 20:22:18.908577 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/b1cf5326-c7d5-4a5f-9a48-159b6482aa0a-ssh-key-openstack-cell1\") pod \"ssh-known-hosts-openstack-2qjln\" (UID: \"b1cf5326-c7d5-4a5f-9a48-159b6482aa0a\") " pod="openstack/ssh-known-hosts-openstack-2qjln" Oct 08 20:22:18 crc kubenswrapper[4988]: I1008 20:22:18.912507 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/b1cf5326-c7d5-4a5f-9a48-159b6482aa0a-inventory-0\") pod \"ssh-known-hosts-openstack-2qjln\" (UID: \"b1cf5326-c7d5-4a5f-9a48-159b6482aa0a\") " pod="openstack/ssh-known-hosts-openstack-2qjln" Oct 08 20:22:18 crc kubenswrapper[4988]: I1008 20:22:18.912719 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/b1cf5326-c7d5-4a5f-9a48-159b6482aa0a-ssh-key-openstack-cell1\") pod \"ssh-known-hosts-openstack-2qjln\" (UID: \"b1cf5326-c7d5-4a5f-9a48-159b6482aa0a\") " pod="openstack/ssh-known-hosts-openstack-2qjln" Oct 08 20:22:18 crc kubenswrapper[4988]: I1008 20:22:18.927323 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d497f\" (UniqueName: \"kubernetes.io/projected/b1cf5326-c7d5-4a5f-9a48-159b6482aa0a-kube-api-access-d497f\") pod \"ssh-known-hosts-openstack-2qjln\" (UID: \"b1cf5326-c7d5-4a5f-9a48-159b6482aa0a\") " pod="openstack/ssh-known-hosts-openstack-2qjln" Oct 08 20:22:19 crc kubenswrapper[4988]: I1008 20:22:19.058097 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-openstack-2qjln" Oct 08 20:22:19 crc kubenswrapper[4988]: I1008 20:22:19.695565 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ssh-known-hosts-openstack-2qjln"] Oct 08 20:22:20 crc kubenswrapper[4988]: I1008 20:22:20.657770 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-openstack-2qjln" event={"ID":"b1cf5326-c7d5-4a5f-9a48-159b6482aa0a","Type":"ContainerStarted","Data":"ce6d20be2e9fb627053abc2a03f5dbd1c2bdc007874fb9e4023b05b79ea73bf1"} Oct 08 20:22:20 crc kubenswrapper[4988]: I1008 20:22:20.659065 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-openstack-2qjln" event={"ID":"b1cf5326-c7d5-4a5f-9a48-159b6482aa0a","Type":"ContainerStarted","Data":"13ac4b7d42e2a44ecaeadaea187b84810102344e41e4dd6cfc54306d503818ec"} Oct 08 20:22:20 crc kubenswrapper[4988]: I1008 20:22:20.687477 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ssh-known-hosts-openstack-2qjln" podStartSLOduration=2.108129444 podStartE2EDuration="2.687448102s" podCreationTimestamp="2025-10-08 20:22:18 +0000 UTC" firstStartedPulling="2025-10-08 20:22:19.702957259 +0000 UTC m=+7885.152800069" lastFinishedPulling="2025-10-08 20:22:20.282275907 +0000 UTC m=+7885.732118727" observedRunningTime="2025-10-08 20:22:20.685246682 +0000 UTC m=+7886.135089502" watchObservedRunningTime="2025-10-08 20:22:20.687448102 +0000 UTC m=+7886.137290912" Oct 08 20:22:27 crc kubenswrapper[4988]: I1008 20:22:27.238533 4988 scope.go:117] "RemoveContainer" containerID="7541195e86b98682ea101ebff2b19f543aa9bab45dd1665e8f710922ad18c5e7" Oct 08 20:22:27 crc kubenswrapper[4988]: E1008 20:22:27.239221 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nm54f_openshift-machine-config-operator(f328a20b-5283-407e-8a93-2ab2da599f18)\"" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" Oct 08 20:22:29 crc kubenswrapper[4988]: I1008 20:22:29.776508 4988 generic.go:334] "Generic (PLEG): container finished" podID="b1cf5326-c7d5-4a5f-9a48-159b6482aa0a" containerID="ce6d20be2e9fb627053abc2a03f5dbd1c2bdc007874fb9e4023b05b79ea73bf1" exitCode=0 Oct 08 20:22:29 crc kubenswrapper[4988]: I1008 20:22:29.776609 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-openstack-2qjln" event={"ID":"b1cf5326-c7d5-4a5f-9a48-159b6482aa0a","Type":"ContainerDied","Data":"ce6d20be2e9fb627053abc2a03f5dbd1c2bdc007874fb9e4023b05b79ea73bf1"} Oct 08 20:22:31 crc kubenswrapper[4988]: I1008 20:22:31.283883 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-openstack-2qjln" Oct 08 20:22:31 crc kubenswrapper[4988]: I1008 20:22:31.351243 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/b1cf5326-c7d5-4a5f-9a48-159b6482aa0a-ssh-key-openstack-cell1\") pod \"b1cf5326-c7d5-4a5f-9a48-159b6482aa0a\" (UID: \"b1cf5326-c7d5-4a5f-9a48-159b6482aa0a\") " Oct 08 20:22:31 crc kubenswrapper[4988]: I1008 20:22:31.351348 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d497f\" (UniqueName: \"kubernetes.io/projected/b1cf5326-c7d5-4a5f-9a48-159b6482aa0a-kube-api-access-d497f\") pod \"b1cf5326-c7d5-4a5f-9a48-159b6482aa0a\" (UID: \"b1cf5326-c7d5-4a5f-9a48-159b6482aa0a\") " Oct 08 20:22:31 crc kubenswrapper[4988]: I1008 20:22:31.351649 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/b1cf5326-c7d5-4a5f-9a48-159b6482aa0a-inventory-0\") pod \"b1cf5326-c7d5-4a5f-9a48-159b6482aa0a\" (UID: \"b1cf5326-c7d5-4a5f-9a48-159b6482aa0a\") " Oct 08 20:22:31 crc kubenswrapper[4988]: I1008 20:22:31.357942 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b1cf5326-c7d5-4a5f-9a48-159b6482aa0a-kube-api-access-d497f" (OuterVolumeSpecName: "kube-api-access-d497f") pod "b1cf5326-c7d5-4a5f-9a48-159b6482aa0a" (UID: "b1cf5326-c7d5-4a5f-9a48-159b6482aa0a"). InnerVolumeSpecName "kube-api-access-d497f". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 20:22:31 crc kubenswrapper[4988]: I1008 20:22:31.381838 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b1cf5326-c7d5-4a5f-9a48-159b6482aa0a-ssh-key-openstack-cell1" (OuterVolumeSpecName: "ssh-key-openstack-cell1") pod "b1cf5326-c7d5-4a5f-9a48-159b6482aa0a" (UID: "b1cf5326-c7d5-4a5f-9a48-159b6482aa0a"). InnerVolumeSpecName "ssh-key-openstack-cell1". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 20:22:31 crc kubenswrapper[4988]: I1008 20:22:31.383837 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b1cf5326-c7d5-4a5f-9a48-159b6482aa0a-inventory-0" (OuterVolumeSpecName: "inventory-0") pod "b1cf5326-c7d5-4a5f-9a48-159b6482aa0a" (UID: "b1cf5326-c7d5-4a5f-9a48-159b6482aa0a"). InnerVolumeSpecName "inventory-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 20:22:31 crc kubenswrapper[4988]: I1008 20:22:31.455199 4988 reconciler_common.go:293] "Volume detached for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/b1cf5326-c7d5-4a5f-9a48-159b6482aa0a-inventory-0\") on node \"crc\" DevicePath \"\"" Oct 08 20:22:31 crc kubenswrapper[4988]: I1008 20:22:31.455329 4988 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/b1cf5326-c7d5-4a5f-9a48-159b6482aa0a-ssh-key-openstack-cell1\") on node \"crc\" DevicePath \"\"" Oct 08 20:22:31 crc kubenswrapper[4988]: I1008 20:22:31.455361 4988 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d497f\" (UniqueName: \"kubernetes.io/projected/b1cf5326-c7d5-4a5f-9a48-159b6482aa0a-kube-api-access-d497f\") on node \"crc\" DevicePath \"\"" Oct 08 20:22:31 crc kubenswrapper[4988]: I1008 20:22:31.802859 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-openstack-2qjln" event={"ID":"b1cf5326-c7d5-4a5f-9a48-159b6482aa0a","Type":"ContainerDied","Data":"13ac4b7d42e2a44ecaeadaea187b84810102344e41e4dd6cfc54306d503818ec"} Oct 08 20:22:31 crc kubenswrapper[4988]: I1008 20:22:31.802927 4988 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="13ac4b7d42e2a44ecaeadaea187b84810102344e41e4dd6cfc54306d503818ec" Oct 08 20:22:31 crc kubenswrapper[4988]: I1008 20:22:31.802948 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-openstack-2qjln" Oct 08 20:22:31 crc kubenswrapper[4988]: I1008 20:22:31.908027 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/run-os-openstack-openstack-cell1-2m44z"] Oct 08 20:22:31 crc kubenswrapper[4988]: E1008 20:22:31.908568 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b1cf5326-c7d5-4a5f-9a48-159b6482aa0a" containerName="ssh-known-hosts-openstack" Oct 08 20:22:31 crc kubenswrapper[4988]: I1008 20:22:31.908585 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="b1cf5326-c7d5-4a5f-9a48-159b6482aa0a" containerName="ssh-known-hosts-openstack" Oct 08 20:22:31 crc kubenswrapper[4988]: I1008 20:22:31.908875 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="b1cf5326-c7d5-4a5f-9a48-159b6482aa0a" containerName="ssh-known-hosts-openstack" Oct 08 20:22:31 crc kubenswrapper[4988]: I1008 20:22:31.909756 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-openstack-openstack-cell1-2m44z" Oct 08 20:22:31 crc kubenswrapper[4988]: I1008 20:22:31.916179 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-4sdzr" Oct 08 20:22:31 crc kubenswrapper[4988]: I1008 20:22:31.916766 4988 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 08 20:22:31 crc kubenswrapper[4988]: I1008 20:22:31.916777 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Oct 08 20:22:31 crc kubenswrapper[4988]: I1008 20:22:31.917259 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Oct 08 20:22:31 crc kubenswrapper[4988]: I1008 20:22:31.928122 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/run-os-openstack-openstack-cell1-2m44z"] Oct 08 20:22:31 crc kubenswrapper[4988]: I1008 20:22:31.968859 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8f284658-1b89-4ace-8914-bfc709e071e5-inventory\") pod \"run-os-openstack-openstack-cell1-2m44z\" (UID: \"8f284658-1b89-4ace-8914-bfc709e071e5\") " pod="openstack/run-os-openstack-openstack-cell1-2m44z" Oct 08 20:22:31 crc kubenswrapper[4988]: I1008 20:22:31.969130 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/8f284658-1b89-4ace-8914-bfc709e071e5-ssh-key\") pod \"run-os-openstack-openstack-cell1-2m44z\" (UID: \"8f284658-1b89-4ace-8914-bfc709e071e5\") " pod="openstack/run-os-openstack-openstack-cell1-2m44z" Oct 08 20:22:31 crc kubenswrapper[4988]: I1008 20:22:31.969199 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cjbk2\" (UniqueName: \"kubernetes.io/projected/8f284658-1b89-4ace-8914-bfc709e071e5-kube-api-access-cjbk2\") pod \"run-os-openstack-openstack-cell1-2m44z\" (UID: \"8f284658-1b89-4ace-8914-bfc709e071e5\") " pod="openstack/run-os-openstack-openstack-cell1-2m44z" Oct 08 20:22:32 crc kubenswrapper[4988]: I1008 20:22:32.071797 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8f284658-1b89-4ace-8914-bfc709e071e5-inventory\") pod \"run-os-openstack-openstack-cell1-2m44z\" (UID: \"8f284658-1b89-4ace-8914-bfc709e071e5\") " pod="openstack/run-os-openstack-openstack-cell1-2m44z" Oct 08 20:22:32 crc kubenswrapper[4988]: I1008 20:22:32.072311 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/8f284658-1b89-4ace-8914-bfc709e071e5-ssh-key\") pod \"run-os-openstack-openstack-cell1-2m44z\" (UID: \"8f284658-1b89-4ace-8914-bfc709e071e5\") " pod="openstack/run-os-openstack-openstack-cell1-2m44z" Oct 08 20:22:32 crc kubenswrapper[4988]: I1008 20:22:32.072365 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cjbk2\" (UniqueName: \"kubernetes.io/projected/8f284658-1b89-4ace-8914-bfc709e071e5-kube-api-access-cjbk2\") pod \"run-os-openstack-openstack-cell1-2m44z\" (UID: \"8f284658-1b89-4ace-8914-bfc709e071e5\") " pod="openstack/run-os-openstack-openstack-cell1-2m44z" Oct 08 20:22:32 crc kubenswrapper[4988]: I1008 20:22:32.077740 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8f284658-1b89-4ace-8914-bfc709e071e5-inventory\") pod \"run-os-openstack-openstack-cell1-2m44z\" (UID: \"8f284658-1b89-4ace-8914-bfc709e071e5\") " pod="openstack/run-os-openstack-openstack-cell1-2m44z" Oct 08 20:22:32 crc kubenswrapper[4988]: I1008 20:22:32.078238 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/8f284658-1b89-4ace-8914-bfc709e071e5-ssh-key\") pod \"run-os-openstack-openstack-cell1-2m44z\" (UID: \"8f284658-1b89-4ace-8914-bfc709e071e5\") " pod="openstack/run-os-openstack-openstack-cell1-2m44z" Oct 08 20:22:32 crc kubenswrapper[4988]: I1008 20:22:32.103098 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cjbk2\" (UniqueName: \"kubernetes.io/projected/8f284658-1b89-4ace-8914-bfc709e071e5-kube-api-access-cjbk2\") pod \"run-os-openstack-openstack-cell1-2m44z\" (UID: \"8f284658-1b89-4ace-8914-bfc709e071e5\") " pod="openstack/run-os-openstack-openstack-cell1-2m44z" Oct 08 20:22:32 crc kubenswrapper[4988]: I1008 20:22:32.248182 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-openstack-openstack-cell1-2m44z" Oct 08 20:22:32 crc kubenswrapper[4988]: I1008 20:22:32.894195 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/run-os-openstack-openstack-cell1-2m44z"] Oct 08 20:22:32 crc kubenswrapper[4988]: W1008 20:22:32.896030 4988 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8f284658_1b89_4ace_8914_bfc709e071e5.slice/crio-34d7c574cdc3302617b04fdb158964e5b2220cb5821b562f1911839eb96b2ffe WatchSource:0}: Error finding container 34d7c574cdc3302617b04fdb158964e5b2220cb5821b562f1911839eb96b2ffe: Status 404 returned error can't find the container with id 34d7c574cdc3302617b04fdb158964e5b2220cb5821b562f1911839eb96b2ffe Oct 08 20:22:33 crc kubenswrapper[4988]: I1008 20:22:33.842499 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-openstack-openstack-cell1-2m44z" event={"ID":"8f284658-1b89-4ace-8914-bfc709e071e5","Type":"ContainerStarted","Data":"673a844e4d3d22c4e0dbab09c36b32249ac3faaf64dade9b543de280d873c1ed"} Oct 08 20:22:33 crc kubenswrapper[4988]: I1008 20:22:33.842736 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-openstack-openstack-cell1-2m44z" event={"ID":"8f284658-1b89-4ace-8914-bfc709e071e5","Type":"ContainerStarted","Data":"34d7c574cdc3302617b04fdb158964e5b2220cb5821b562f1911839eb96b2ffe"} Oct 08 20:22:33 crc kubenswrapper[4988]: I1008 20:22:33.868234 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/run-os-openstack-openstack-cell1-2m44z" podStartSLOduration=2.268113547 podStartE2EDuration="2.868159335s" podCreationTimestamp="2025-10-08 20:22:31 +0000 UTC" firstStartedPulling="2025-10-08 20:22:32.900053453 +0000 UTC m=+7898.349896253" lastFinishedPulling="2025-10-08 20:22:33.500099221 +0000 UTC m=+7898.949942041" observedRunningTime="2025-10-08 20:22:33.864657724 +0000 UTC m=+7899.314500524" watchObservedRunningTime="2025-10-08 20:22:33.868159335 +0000 UTC m=+7899.318002145" Oct 08 20:22:42 crc kubenswrapper[4988]: I1008 20:22:42.238674 4988 scope.go:117] "RemoveContainer" containerID="7541195e86b98682ea101ebff2b19f543aa9bab45dd1665e8f710922ad18c5e7" Oct 08 20:22:42 crc kubenswrapper[4988]: E1008 20:22:42.239692 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nm54f_openshift-machine-config-operator(f328a20b-5283-407e-8a93-2ab2da599f18)\"" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" Oct 08 20:22:42 crc kubenswrapper[4988]: I1008 20:22:42.953933 4988 generic.go:334] "Generic (PLEG): container finished" podID="8f284658-1b89-4ace-8914-bfc709e071e5" containerID="673a844e4d3d22c4e0dbab09c36b32249ac3faaf64dade9b543de280d873c1ed" exitCode=0 Oct 08 20:22:42 crc kubenswrapper[4988]: I1008 20:22:42.954038 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-openstack-openstack-cell1-2m44z" event={"ID":"8f284658-1b89-4ace-8914-bfc709e071e5","Type":"ContainerDied","Data":"673a844e4d3d22c4e0dbab09c36b32249ac3faaf64dade9b543de280d873c1ed"} Oct 08 20:22:44 crc kubenswrapper[4988]: I1008 20:22:44.501669 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-openstack-openstack-cell1-2m44z" Oct 08 20:22:44 crc kubenswrapper[4988]: I1008 20:22:44.563726 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cjbk2\" (UniqueName: \"kubernetes.io/projected/8f284658-1b89-4ace-8914-bfc709e071e5-kube-api-access-cjbk2\") pod \"8f284658-1b89-4ace-8914-bfc709e071e5\" (UID: \"8f284658-1b89-4ace-8914-bfc709e071e5\") " Oct 08 20:22:44 crc kubenswrapper[4988]: I1008 20:22:44.563891 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/8f284658-1b89-4ace-8914-bfc709e071e5-ssh-key\") pod \"8f284658-1b89-4ace-8914-bfc709e071e5\" (UID: \"8f284658-1b89-4ace-8914-bfc709e071e5\") " Oct 08 20:22:44 crc kubenswrapper[4988]: I1008 20:22:44.563919 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8f284658-1b89-4ace-8914-bfc709e071e5-inventory\") pod \"8f284658-1b89-4ace-8914-bfc709e071e5\" (UID: \"8f284658-1b89-4ace-8914-bfc709e071e5\") " Oct 08 20:22:44 crc kubenswrapper[4988]: I1008 20:22:44.571006 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f284658-1b89-4ace-8914-bfc709e071e5-kube-api-access-cjbk2" (OuterVolumeSpecName: "kube-api-access-cjbk2") pod "8f284658-1b89-4ace-8914-bfc709e071e5" (UID: "8f284658-1b89-4ace-8914-bfc709e071e5"). InnerVolumeSpecName "kube-api-access-cjbk2". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 20:22:44 crc kubenswrapper[4988]: I1008 20:22:44.593486 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8f284658-1b89-4ace-8914-bfc709e071e5-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "8f284658-1b89-4ace-8914-bfc709e071e5" (UID: "8f284658-1b89-4ace-8914-bfc709e071e5"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 20:22:44 crc kubenswrapper[4988]: I1008 20:22:44.595784 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8f284658-1b89-4ace-8914-bfc709e071e5-inventory" (OuterVolumeSpecName: "inventory") pod "8f284658-1b89-4ace-8914-bfc709e071e5" (UID: "8f284658-1b89-4ace-8914-bfc709e071e5"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 20:22:44 crc kubenswrapper[4988]: I1008 20:22:44.666148 4988 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/8f284658-1b89-4ace-8914-bfc709e071e5-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 08 20:22:44 crc kubenswrapper[4988]: I1008 20:22:44.666181 4988 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8f284658-1b89-4ace-8914-bfc709e071e5-inventory\") on node \"crc\" DevicePath \"\"" Oct 08 20:22:44 crc kubenswrapper[4988]: I1008 20:22:44.666192 4988 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cjbk2\" (UniqueName: \"kubernetes.io/projected/8f284658-1b89-4ace-8914-bfc709e071e5-kube-api-access-cjbk2\") on node \"crc\" DevicePath \"\"" Oct 08 20:22:44 crc kubenswrapper[4988]: I1008 20:22:44.979524 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-openstack-openstack-cell1-2m44z" event={"ID":"8f284658-1b89-4ace-8914-bfc709e071e5","Type":"ContainerDied","Data":"34d7c574cdc3302617b04fdb158964e5b2220cb5821b562f1911839eb96b2ffe"} Oct 08 20:22:44 crc kubenswrapper[4988]: I1008 20:22:44.979581 4988 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="34d7c574cdc3302617b04fdb158964e5b2220cb5821b562f1911839eb96b2ffe" Oct 08 20:22:44 crc kubenswrapper[4988]: I1008 20:22:44.979595 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-openstack-openstack-cell1-2m44z" Oct 08 20:22:45 crc kubenswrapper[4988]: I1008 20:22:45.091186 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/reboot-os-openstack-openstack-cell1-rxm8h"] Oct 08 20:22:45 crc kubenswrapper[4988]: E1008 20:22:45.091742 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8f284658-1b89-4ace-8914-bfc709e071e5" containerName="run-os-openstack-openstack-cell1" Oct 08 20:22:45 crc kubenswrapper[4988]: I1008 20:22:45.091765 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="8f284658-1b89-4ace-8914-bfc709e071e5" containerName="run-os-openstack-openstack-cell1" Oct 08 20:22:45 crc kubenswrapper[4988]: I1008 20:22:45.092022 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="8f284658-1b89-4ace-8914-bfc709e071e5" containerName="run-os-openstack-openstack-cell1" Oct 08 20:22:45 crc kubenswrapper[4988]: I1008 20:22:45.092955 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-openstack-openstack-cell1-rxm8h" Oct 08 20:22:45 crc kubenswrapper[4988]: I1008 20:22:45.097116 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-4sdzr" Oct 08 20:22:45 crc kubenswrapper[4988]: I1008 20:22:45.097448 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Oct 08 20:22:45 crc kubenswrapper[4988]: I1008 20:22:45.097924 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Oct 08 20:22:45 crc kubenswrapper[4988]: I1008 20:22:45.099814 4988 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 08 20:22:45 crc kubenswrapper[4988]: I1008 20:22:45.105655 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/reboot-os-openstack-openstack-cell1-rxm8h"] Oct 08 20:22:45 crc kubenswrapper[4988]: I1008 20:22:45.176225 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d4085938-9772-43b4-bd58-7044e2e3c6e0-ssh-key\") pod \"reboot-os-openstack-openstack-cell1-rxm8h\" (UID: \"d4085938-9772-43b4-bd58-7044e2e3c6e0\") " pod="openstack/reboot-os-openstack-openstack-cell1-rxm8h" Oct 08 20:22:45 crc kubenswrapper[4988]: I1008 20:22:45.176312 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d4085938-9772-43b4-bd58-7044e2e3c6e0-inventory\") pod \"reboot-os-openstack-openstack-cell1-rxm8h\" (UID: \"d4085938-9772-43b4-bd58-7044e2e3c6e0\") " pod="openstack/reboot-os-openstack-openstack-cell1-rxm8h" Oct 08 20:22:45 crc kubenswrapper[4988]: I1008 20:22:45.176451 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hxzvl\" (UniqueName: \"kubernetes.io/projected/d4085938-9772-43b4-bd58-7044e2e3c6e0-kube-api-access-hxzvl\") pod \"reboot-os-openstack-openstack-cell1-rxm8h\" (UID: \"d4085938-9772-43b4-bd58-7044e2e3c6e0\") " pod="openstack/reboot-os-openstack-openstack-cell1-rxm8h" Oct 08 20:22:45 crc kubenswrapper[4988]: I1008 20:22:45.279364 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d4085938-9772-43b4-bd58-7044e2e3c6e0-inventory\") pod \"reboot-os-openstack-openstack-cell1-rxm8h\" (UID: \"d4085938-9772-43b4-bd58-7044e2e3c6e0\") " pod="openstack/reboot-os-openstack-openstack-cell1-rxm8h" Oct 08 20:22:45 crc kubenswrapper[4988]: I1008 20:22:45.279678 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hxzvl\" (UniqueName: \"kubernetes.io/projected/d4085938-9772-43b4-bd58-7044e2e3c6e0-kube-api-access-hxzvl\") pod \"reboot-os-openstack-openstack-cell1-rxm8h\" (UID: \"d4085938-9772-43b4-bd58-7044e2e3c6e0\") " pod="openstack/reboot-os-openstack-openstack-cell1-rxm8h" Oct 08 20:22:45 crc kubenswrapper[4988]: I1008 20:22:45.279971 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d4085938-9772-43b4-bd58-7044e2e3c6e0-ssh-key\") pod \"reboot-os-openstack-openstack-cell1-rxm8h\" (UID: \"d4085938-9772-43b4-bd58-7044e2e3c6e0\") " pod="openstack/reboot-os-openstack-openstack-cell1-rxm8h" Oct 08 20:22:45 crc kubenswrapper[4988]: I1008 20:22:45.283798 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d4085938-9772-43b4-bd58-7044e2e3c6e0-inventory\") pod \"reboot-os-openstack-openstack-cell1-rxm8h\" (UID: \"d4085938-9772-43b4-bd58-7044e2e3c6e0\") " pod="openstack/reboot-os-openstack-openstack-cell1-rxm8h" Oct 08 20:22:45 crc kubenswrapper[4988]: I1008 20:22:45.291467 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d4085938-9772-43b4-bd58-7044e2e3c6e0-ssh-key\") pod \"reboot-os-openstack-openstack-cell1-rxm8h\" (UID: \"d4085938-9772-43b4-bd58-7044e2e3c6e0\") " pod="openstack/reboot-os-openstack-openstack-cell1-rxm8h" Oct 08 20:22:45 crc kubenswrapper[4988]: I1008 20:22:45.297523 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hxzvl\" (UniqueName: \"kubernetes.io/projected/d4085938-9772-43b4-bd58-7044e2e3c6e0-kube-api-access-hxzvl\") pod \"reboot-os-openstack-openstack-cell1-rxm8h\" (UID: \"d4085938-9772-43b4-bd58-7044e2e3c6e0\") " pod="openstack/reboot-os-openstack-openstack-cell1-rxm8h" Oct 08 20:22:45 crc kubenswrapper[4988]: I1008 20:22:45.418965 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-openstack-openstack-cell1-rxm8h" Oct 08 20:22:46 crc kubenswrapper[4988]: I1008 20:22:46.034293 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/reboot-os-openstack-openstack-cell1-rxm8h"] Oct 08 20:22:47 crc kubenswrapper[4988]: I1008 20:22:47.008248 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-openstack-openstack-cell1-rxm8h" event={"ID":"d4085938-9772-43b4-bd58-7044e2e3c6e0","Type":"ContainerStarted","Data":"fe885d370bfa33cca9f48c37e742cbea8e65c00185e391b4473c46288ddc943a"} Oct 08 20:22:47 crc kubenswrapper[4988]: I1008 20:22:47.008679 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-openstack-openstack-cell1-rxm8h" event={"ID":"d4085938-9772-43b4-bd58-7044e2e3c6e0","Type":"ContainerStarted","Data":"4e675ffc9eacdc4f6f99eae8495ee2af59411fbdb2304a5bca981a870c36f6fd"} Oct 08 20:22:47 crc kubenswrapper[4988]: I1008 20:22:47.026138 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/reboot-os-openstack-openstack-cell1-rxm8h" podStartSLOduration=1.447408985 podStartE2EDuration="2.026118944s" podCreationTimestamp="2025-10-08 20:22:45 +0000 UTC" firstStartedPulling="2025-10-08 20:22:46.057098063 +0000 UTC m=+7911.506940863" lastFinishedPulling="2025-10-08 20:22:46.635808002 +0000 UTC m=+7912.085650822" observedRunningTime="2025-10-08 20:22:47.024217764 +0000 UTC m=+7912.474060554" watchObservedRunningTime="2025-10-08 20:22:47.026118944 +0000 UTC m=+7912.475961714" Oct 08 20:22:56 crc kubenswrapper[4988]: I1008 20:22:56.237902 4988 scope.go:117] "RemoveContainer" containerID="7541195e86b98682ea101ebff2b19f543aa9bab45dd1665e8f710922ad18c5e7" Oct 08 20:22:56 crc kubenswrapper[4988]: E1008 20:22:56.239221 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nm54f_openshift-machine-config-operator(f328a20b-5283-407e-8a93-2ab2da599f18)\"" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" Oct 08 20:23:03 crc kubenswrapper[4988]: I1008 20:23:03.198128 4988 generic.go:334] "Generic (PLEG): container finished" podID="d4085938-9772-43b4-bd58-7044e2e3c6e0" containerID="fe885d370bfa33cca9f48c37e742cbea8e65c00185e391b4473c46288ddc943a" exitCode=0 Oct 08 20:23:03 crc kubenswrapper[4988]: I1008 20:23:03.198596 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-openstack-openstack-cell1-rxm8h" event={"ID":"d4085938-9772-43b4-bd58-7044e2e3c6e0","Type":"ContainerDied","Data":"fe885d370bfa33cca9f48c37e742cbea8e65c00185e391b4473c46288ddc943a"} Oct 08 20:23:04 crc kubenswrapper[4988]: I1008 20:23:04.716854 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-openstack-openstack-cell1-rxm8h" Oct 08 20:23:04 crc kubenswrapper[4988]: I1008 20:23:04.777797 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d4085938-9772-43b4-bd58-7044e2e3c6e0-inventory\") pod \"d4085938-9772-43b4-bd58-7044e2e3c6e0\" (UID: \"d4085938-9772-43b4-bd58-7044e2e3c6e0\") " Oct 08 20:23:04 crc kubenswrapper[4988]: I1008 20:23:04.777934 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d4085938-9772-43b4-bd58-7044e2e3c6e0-ssh-key\") pod \"d4085938-9772-43b4-bd58-7044e2e3c6e0\" (UID: \"d4085938-9772-43b4-bd58-7044e2e3c6e0\") " Oct 08 20:23:04 crc kubenswrapper[4988]: I1008 20:23:04.777962 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hxzvl\" (UniqueName: \"kubernetes.io/projected/d4085938-9772-43b4-bd58-7044e2e3c6e0-kube-api-access-hxzvl\") pod \"d4085938-9772-43b4-bd58-7044e2e3c6e0\" (UID: \"d4085938-9772-43b4-bd58-7044e2e3c6e0\") " Oct 08 20:23:04 crc kubenswrapper[4988]: I1008 20:23:04.789807 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d4085938-9772-43b4-bd58-7044e2e3c6e0-kube-api-access-hxzvl" (OuterVolumeSpecName: "kube-api-access-hxzvl") pod "d4085938-9772-43b4-bd58-7044e2e3c6e0" (UID: "d4085938-9772-43b4-bd58-7044e2e3c6e0"). InnerVolumeSpecName "kube-api-access-hxzvl". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 20:23:04 crc kubenswrapper[4988]: I1008 20:23:04.829603 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d4085938-9772-43b4-bd58-7044e2e3c6e0-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "d4085938-9772-43b4-bd58-7044e2e3c6e0" (UID: "d4085938-9772-43b4-bd58-7044e2e3c6e0"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 20:23:04 crc kubenswrapper[4988]: I1008 20:23:04.830356 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d4085938-9772-43b4-bd58-7044e2e3c6e0-inventory" (OuterVolumeSpecName: "inventory") pod "d4085938-9772-43b4-bd58-7044e2e3c6e0" (UID: "d4085938-9772-43b4-bd58-7044e2e3c6e0"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 20:23:04 crc kubenswrapper[4988]: I1008 20:23:04.882201 4988 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d4085938-9772-43b4-bd58-7044e2e3c6e0-inventory\") on node \"crc\" DevicePath \"\"" Oct 08 20:23:04 crc kubenswrapper[4988]: I1008 20:23:04.882254 4988 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d4085938-9772-43b4-bd58-7044e2e3c6e0-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 08 20:23:04 crc kubenswrapper[4988]: I1008 20:23:04.882276 4988 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hxzvl\" (UniqueName: \"kubernetes.io/projected/d4085938-9772-43b4-bd58-7044e2e3c6e0-kube-api-access-hxzvl\") on node \"crc\" DevicePath \"\"" Oct 08 20:23:05 crc kubenswrapper[4988]: I1008 20:23:05.223889 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-openstack-openstack-cell1-rxm8h" event={"ID":"d4085938-9772-43b4-bd58-7044e2e3c6e0","Type":"ContainerDied","Data":"4e675ffc9eacdc4f6f99eae8495ee2af59411fbdb2304a5bca981a870c36f6fd"} Oct 08 20:23:05 crc kubenswrapper[4988]: I1008 20:23:05.223926 4988 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4e675ffc9eacdc4f6f99eae8495ee2af59411fbdb2304a5bca981a870c36f6fd" Oct 08 20:23:05 crc kubenswrapper[4988]: I1008 20:23:05.223979 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-openstack-openstack-cell1-rxm8h" Oct 08 20:23:05 crc kubenswrapper[4988]: I1008 20:23:05.331621 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/install-certs-openstack-openstack-cell1-n45qh"] Oct 08 20:23:05 crc kubenswrapper[4988]: E1008 20:23:05.332241 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d4085938-9772-43b4-bd58-7044e2e3c6e0" containerName="reboot-os-openstack-openstack-cell1" Oct 08 20:23:05 crc kubenswrapper[4988]: I1008 20:23:05.332270 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="d4085938-9772-43b4-bd58-7044e2e3c6e0" containerName="reboot-os-openstack-openstack-cell1" Oct 08 20:23:05 crc kubenswrapper[4988]: I1008 20:23:05.332645 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="d4085938-9772-43b4-bd58-7044e2e3c6e0" containerName="reboot-os-openstack-openstack-cell1" Oct 08 20:23:05 crc kubenswrapper[4988]: I1008 20:23:05.333987 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-openstack-openstack-cell1-n45qh" Oct 08 20:23:05 crc kubenswrapper[4988]: I1008 20:23:05.337690 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-ovn-default-certs-0" Oct 08 20:23:05 crc kubenswrapper[4988]: I1008 20:23:05.337760 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-libvirt-default-certs-0" Oct 08 20:23:05 crc kubenswrapper[4988]: I1008 20:23:05.338139 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-telemetry-default-certs-0" Oct 08 20:23:05 crc kubenswrapper[4988]: I1008 20:23:05.338299 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-neutron-metadata-default-certs-0" Oct 08 20:23:05 crc kubenswrapper[4988]: I1008 20:23:05.338363 4988 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 08 20:23:05 crc kubenswrapper[4988]: I1008 20:23:05.338903 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Oct 08 20:23:05 crc kubenswrapper[4988]: I1008 20:23:05.339796 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Oct 08 20:23:05 crc kubenswrapper[4988]: I1008 20:23:05.349818 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-4sdzr" Oct 08 20:23:05 crc kubenswrapper[4988]: I1008 20:23:05.367085 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-certs-openstack-openstack-cell1-n45qh"] Oct 08 20:23:05 crc kubenswrapper[4988]: I1008 20:23:05.394822 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c6b7f3f9-1ffb-454d-8dc6-f697eac681ae-libvirt-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-n45qh\" (UID: \"c6b7f3f9-1ffb-454d-8dc6-f697eac681ae\") " pod="openstack/install-certs-openstack-openstack-cell1-n45qh" Oct 08 20:23:05 crc kubenswrapper[4988]: I1008 20:23:05.394867 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-sriov-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c6b7f3f9-1ffb-454d-8dc6-f697eac681ae-neutron-sriov-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-n45qh\" (UID: \"c6b7f3f9-1ffb-454d-8dc6-f697eac681ae\") " pod="openstack/install-certs-openstack-openstack-cell1-n45qh" Oct 08 20:23:05 crc kubenswrapper[4988]: I1008 20:23:05.394905 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c6b7f3f9-1ffb-454d-8dc6-f697eac681ae-ssh-key\") pod \"install-certs-openstack-openstack-cell1-n45qh\" (UID: \"c6b7f3f9-1ffb-454d-8dc6-f697eac681ae\") " pod="openstack/install-certs-openstack-openstack-cell1-n45qh" Oct 08 20:23:05 crc kubenswrapper[4988]: I1008 20:23:05.394933 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c6b7f3f9-1ffb-454d-8dc6-f697eac681ae-bootstrap-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-n45qh\" (UID: \"c6b7f3f9-1ffb-454d-8dc6-f697eac681ae\") " pod="openstack/install-certs-openstack-openstack-cell1-n45qh" Oct 08 20:23:05 crc kubenswrapper[4988]: I1008 20:23:05.394949 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-cell1-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/c6b7f3f9-1ffb-454d-8dc6-f697eac681ae-openstack-cell1-telemetry-default-certs-0\") pod \"install-certs-openstack-openstack-cell1-n45qh\" (UID: \"c6b7f3f9-1ffb-454d-8dc6-f697eac681ae\") " pod="openstack/install-certs-openstack-openstack-cell1-n45qh" Oct 08 20:23:05 crc kubenswrapper[4988]: I1008 20:23:05.394977 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6mkgw\" (UniqueName: \"kubernetes.io/projected/c6b7f3f9-1ffb-454d-8dc6-f697eac681ae-kube-api-access-6mkgw\") pod \"install-certs-openstack-openstack-cell1-n45qh\" (UID: \"c6b7f3f9-1ffb-454d-8dc6-f697eac681ae\") " pod="openstack/install-certs-openstack-openstack-cell1-n45qh" Oct 08 20:23:05 crc kubenswrapper[4988]: I1008 20:23:05.395034 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c6b7f3f9-1ffb-454d-8dc6-f697eac681ae-ovn-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-n45qh\" (UID: \"c6b7f3f9-1ffb-454d-8dc6-f697eac681ae\") " pod="openstack/install-certs-openstack-openstack-cell1-n45qh" Oct 08 20:23:05 crc kubenswrapper[4988]: I1008 20:23:05.395076 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c6b7f3f9-1ffb-454d-8dc6-f697eac681ae-neutron-metadata-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-n45qh\" (UID: \"c6b7f3f9-1ffb-454d-8dc6-f697eac681ae\") " pod="openstack/install-certs-openstack-openstack-cell1-n45qh" Oct 08 20:23:05 crc kubenswrapper[4988]: I1008 20:23:05.395446 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-cell1-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/c6b7f3f9-1ffb-454d-8dc6-f697eac681ae-openstack-cell1-neutron-metadata-default-certs-0\") pod \"install-certs-openstack-openstack-cell1-n45qh\" (UID: \"c6b7f3f9-1ffb-454d-8dc6-f697eac681ae\") " pod="openstack/install-certs-openstack-openstack-cell1-n45qh" Oct 08 20:23:05 crc kubenswrapper[4988]: I1008 20:23:05.395478 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-cell1-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/c6b7f3f9-1ffb-454d-8dc6-f697eac681ae-openstack-cell1-ovn-default-certs-0\") pod \"install-certs-openstack-openstack-cell1-n45qh\" (UID: \"c6b7f3f9-1ffb-454d-8dc6-f697eac681ae\") " pod="openstack/install-certs-openstack-openstack-cell1-n45qh" Oct 08 20:23:05 crc kubenswrapper[4988]: I1008 20:23:05.395502 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-cell1-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/c6b7f3f9-1ffb-454d-8dc6-f697eac681ae-openstack-cell1-libvirt-default-certs-0\") pod \"install-certs-openstack-openstack-cell1-n45qh\" (UID: \"c6b7f3f9-1ffb-454d-8dc6-f697eac681ae\") " pod="openstack/install-certs-openstack-openstack-cell1-n45qh" Oct 08 20:23:05 crc kubenswrapper[4988]: I1008 20:23:05.395682 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-dhcp-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c6b7f3f9-1ffb-454d-8dc6-f697eac681ae-neutron-dhcp-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-n45qh\" (UID: \"c6b7f3f9-1ffb-454d-8dc6-f697eac681ae\") " pod="openstack/install-certs-openstack-openstack-cell1-n45qh" Oct 08 20:23:05 crc kubenswrapper[4988]: I1008 20:23:05.395750 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c6b7f3f9-1ffb-454d-8dc6-f697eac681ae-telemetry-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-n45qh\" (UID: \"c6b7f3f9-1ffb-454d-8dc6-f697eac681ae\") " pod="openstack/install-certs-openstack-openstack-cell1-n45qh" Oct 08 20:23:05 crc kubenswrapper[4988]: I1008 20:23:05.395779 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c6b7f3f9-1ffb-454d-8dc6-f697eac681ae-inventory\") pod \"install-certs-openstack-openstack-cell1-n45qh\" (UID: \"c6b7f3f9-1ffb-454d-8dc6-f697eac681ae\") " pod="openstack/install-certs-openstack-openstack-cell1-n45qh" Oct 08 20:23:05 crc kubenswrapper[4988]: I1008 20:23:05.395829 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c6b7f3f9-1ffb-454d-8dc6-f697eac681ae-nova-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-n45qh\" (UID: \"c6b7f3f9-1ffb-454d-8dc6-f697eac681ae\") " pod="openstack/install-certs-openstack-openstack-cell1-n45qh" Oct 08 20:23:05 crc kubenswrapper[4988]: I1008 20:23:05.497604 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c6b7f3f9-1ffb-454d-8dc6-f697eac681ae-nova-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-n45qh\" (UID: \"c6b7f3f9-1ffb-454d-8dc6-f697eac681ae\") " pod="openstack/install-certs-openstack-openstack-cell1-n45qh" Oct 08 20:23:05 crc kubenswrapper[4988]: I1008 20:23:05.497674 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c6b7f3f9-1ffb-454d-8dc6-f697eac681ae-libvirt-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-n45qh\" (UID: \"c6b7f3f9-1ffb-454d-8dc6-f697eac681ae\") " pod="openstack/install-certs-openstack-openstack-cell1-n45qh" Oct 08 20:23:05 crc kubenswrapper[4988]: I1008 20:23:05.497698 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-sriov-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c6b7f3f9-1ffb-454d-8dc6-f697eac681ae-neutron-sriov-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-n45qh\" (UID: \"c6b7f3f9-1ffb-454d-8dc6-f697eac681ae\") " pod="openstack/install-certs-openstack-openstack-cell1-n45qh" Oct 08 20:23:05 crc kubenswrapper[4988]: I1008 20:23:05.497723 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c6b7f3f9-1ffb-454d-8dc6-f697eac681ae-ssh-key\") pod \"install-certs-openstack-openstack-cell1-n45qh\" (UID: \"c6b7f3f9-1ffb-454d-8dc6-f697eac681ae\") " pod="openstack/install-certs-openstack-openstack-cell1-n45qh" Oct 08 20:23:05 crc kubenswrapper[4988]: I1008 20:23:05.497748 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c6b7f3f9-1ffb-454d-8dc6-f697eac681ae-bootstrap-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-n45qh\" (UID: \"c6b7f3f9-1ffb-454d-8dc6-f697eac681ae\") " pod="openstack/install-certs-openstack-openstack-cell1-n45qh" Oct 08 20:23:05 crc kubenswrapper[4988]: I1008 20:23:05.497766 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-cell1-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/c6b7f3f9-1ffb-454d-8dc6-f697eac681ae-openstack-cell1-telemetry-default-certs-0\") pod \"install-certs-openstack-openstack-cell1-n45qh\" (UID: \"c6b7f3f9-1ffb-454d-8dc6-f697eac681ae\") " pod="openstack/install-certs-openstack-openstack-cell1-n45qh" Oct 08 20:23:05 crc kubenswrapper[4988]: I1008 20:23:05.497792 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6mkgw\" (UniqueName: \"kubernetes.io/projected/c6b7f3f9-1ffb-454d-8dc6-f697eac681ae-kube-api-access-6mkgw\") pod \"install-certs-openstack-openstack-cell1-n45qh\" (UID: \"c6b7f3f9-1ffb-454d-8dc6-f697eac681ae\") " pod="openstack/install-certs-openstack-openstack-cell1-n45qh" Oct 08 20:23:05 crc kubenswrapper[4988]: I1008 20:23:05.497846 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c6b7f3f9-1ffb-454d-8dc6-f697eac681ae-ovn-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-n45qh\" (UID: \"c6b7f3f9-1ffb-454d-8dc6-f697eac681ae\") " pod="openstack/install-certs-openstack-openstack-cell1-n45qh" Oct 08 20:23:05 crc kubenswrapper[4988]: I1008 20:23:05.497888 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c6b7f3f9-1ffb-454d-8dc6-f697eac681ae-neutron-metadata-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-n45qh\" (UID: \"c6b7f3f9-1ffb-454d-8dc6-f697eac681ae\") " pod="openstack/install-certs-openstack-openstack-cell1-n45qh" Oct 08 20:23:05 crc kubenswrapper[4988]: I1008 20:23:05.497919 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-cell1-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/c6b7f3f9-1ffb-454d-8dc6-f697eac681ae-openstack-cell1-neutron-metadata-default-certs-0\") pod \"install-certs-openstack-openstack-cell1-n45qh\" (UID: \"c6b7f3f9-1ffb-454d-8dc6-f697eac681ae\") " pod="openstack/install-certs-openstack-openstack-cell1-n45qh" Oct 08 20:23:05 crc kubenswrapper[4988]: I1008 20:23:05.497948 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-cell1-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/c6b7f3f9-1ffb-454d-8dc6-f697eac681ae-openstack-cell1-ovn-default-certs-0\") pod \"install-certs-openstack-openstack-cell1-n45qh\" (UID: \"c6b7f3f9-1ffb-454d-8dc6-f697eac681ae\") " pod="openstack/install-certs-openstack-openstack-cell1-n45qh" Oct 08 20:23:05 crc kubenswrapper[4988]: I1008 20:23:05.497985 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-cell1-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/c6b7f3f9-1ffb-454d-8dc6-f697eac681ae-openstack-cell1-libvirt-default-certs-0\") pod \"install-certs-openstack-openstack-cell1-n45qh\" (UID: \"c6b7f3f9-1ffb-454d-8dc6-f697eac681ae\") " pod="openstack/install-certs-openstack-openstack-cell1-n45qh" Oct 08 20:23:05 crc kubenswrapper[4988]: I1008 20:23:05.498027 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-dhcp-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c6b7f3f9-1ffb-454d-8dc6-f697eac681ae-neutron-dhcp-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-n45qh\" (UID: \"c6b7f3f9-1ffb-454d-8dc6-f697eac681ae\") " pod="openstack/install-certs-openstack-openstack-cell1-n45qh" Oct 08 20:23:05 crc kubenswrapper[4988]: I1008 20:23:05.498051 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c6b7f3f9-1ffb-454d-8dc6-f697eac681ae-telemetry-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-n45qh\" (UID: \"c6b7f3f9-1ffb-454d-8dc6-f697eac681ae\") " pod="openstack/install-certs-openstack-openstack-cell1-n45qh" Oct 08 20:23:05 crc kubenswrapper[4988]: I1008 20:23:05.498089 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c6b7f3f9-1ffb-454d-8dc6-f697eac681ae-inventory\") pod \"install-certs-openstack-openstack-cell1-n45qh\" (UID: \"c6b7f3f9-1ffb-454d-8dc6-f697eac681ae\") " pod="openstack/install-certs-openstack-openstack-cell1-n45qh" Oct 08 20:23:05 crc kubenswrapper[4988]: I1008 20:23:05.501260 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c6b7f3f9-1ffb-454d-8dc6-f697eac681ae-inventory\") pod \"install-certs-openstack-openstack-cell1-n45qh\" (UID: \"c6b7f3f9-1ffb-454d-8dc6-f697eac681ae\") " pod="openstack/install-certs-openstack-openstack-cell1-n45qh" Oct 08 20:23:05 crc kubenswrapper[4988]: I1008 20:23:05.504553 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c6b7f3f9-1ffb-454d-8dc6-f697eac681ae-ssh-key\") pod \"install-certs-openstack-openstack-cell1-n45qh\" (UID: \"c6b7f3f9-1ffb-454d-8dc6-f697eac681ae\") " pod="openstack/install-certs-openstack-openstack-cell1-n45qh" Oct 08 20:23:05 crc kubenswrapper[4988]: I1008 20:23:05.506914 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c6b7f3f9-1ffb-454d-8dc6-f697eac681ae-bootstrap-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-n45qh\" (UID: \"c6b7f3f9-1ffb-454d-8dc6-f697eac681ae\") " pod="openstack/install-certs-openstack-openstack-cell1-n45qh" Oct 08 20:23:05 crc kubenswrapper[4988]: I1008 20:23:05.507101 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c6b7f3f9-1ffb-454d-8dc6-f697eac681ae-ovn-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-n45qh\" (UID: \"c6b7f3f9-1ffb-454d-8dc6-f697eac681ae\") " pod="openstack/install-certs-openstack-openstack-cell1-n45qh" Oct 08 20:23:05 crc kubenswrapper[4988]: I1008 20:23:05.508259 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-cell1-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/c6b7f3f9-1ffb-454d-8dc6-f697eac681ae-openstack-cell1-neutron-metadata-default-certs-0\") pod \"install-certs-openstack-openstack-cell1-n45qh\" (UID: \"c6b7f3f9-1ffb-454d-8dc6-f697eac681ae\") " pod="openstack/install-certs-openstack-openstack-cell1-n45qh" Oct 08 20:23:05 crc kubenswrapper[4988]: I1008 20:23:05.509448 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-dhcp-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c6b7f3f9-1ffb-454d-8dc6-f697eac681ae-neutron-dhcp-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-n45qh\" (UID: \"c6b7f3f9-1ffb-454d-8dc6-f697eac681ae\") " pod="openstack/install-certs-openstack-openstack-cell1-n45qh" Oct 08 20:23:05 crc kubenswrapper[4988]: I1008 20:23:05.509740 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c6b7f3f9-1ffb-454d-8dc6-f697eac681ae-neutron-metadata-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-n45qh\" (UID: \"c6b7f3f9-1ffb-454d-8dc6-f697eac681ae\") " pod="openstack/install-certs-openstack-openstack-cell1-n45qh" Oct 08 20:23:05 crc kubenswrapper[4988]: I1008 20:23:05.511376 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c6b7f3f9-1ffb-454d-8dc6-f697eac681ae-libvirt-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-n45qh\" (UID: \"c6b7f3f9-1ffb-454d-8dc6-f697eac681ae\") " pod="openstack/install-certs-openstack-openstack-cell1-n45qh" Oct 08 20:23:05 crc kubenswrapper[4988]: I1008 20:23:05.511896 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c6b7f3f9-1ffb-454d-8dc6-f697eac681ae-nova-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-n45qh\" (UID: \"c6b7f3f9-1ffb-454d-8dc6-f697eac681ae\") " pod="openstack/install-certs-openstack-openstack-cell1-n45qh" Oct 08 20:23:05 crc kubenswrapper[4988]: I1008 20:23:05.514161 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-cell1-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/c6b7f3f9-1ffb-454d-8dc6-f697eac681ae-openstack-cell1-ovn-default-certs-0\") pod \"install-certs-openstack-openstack-cell1-n45qh\" (UID: \"c6b7f3f9-1ffb-454d-8dc6-f697eac681ae\") " pod="openstack/install-certs-openstack-openstack-cell1-n45qh" Oct 08 20:23:05 crc kubenswrapper[4988]: I1008 20:23:05.515960 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-cell1-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/c6b7f3f9-1ffb-454d-8dc6-f697eac681ae-openstack-cell1-libvirt-default-certs-0\") pod \"install-certs-openstack-openstack-cell1-n45qh\" (UID: \"c6b7f3f9-1ffb-454d-8dc6-f697eac681ae\") " pod="openstack/install-certs-openstack-openstack-cell1-n45qh" Oct 08 20:23:05 crc kubenswrapper[4988]: I1008 20:23:05.516831 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-cell1-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/c6b7f3f9-1ffb-454d-8dc6-f697eac681ae-openstack-cell1-telemetry-default-certs-0\") pod \"install-certs-openstack-openstack-cell1-n45qh\" (UID: \"c6b7f3f9-1ffb-454d-8dc6-f697eac681ae\") " pod="openstack/install-certs-openstack-openstack-cell1-n45qh" Oct 08 20:23:05 crc kubenswrapper[4988]: I1008 20:23:05.516924 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c6b7f3f9-1ffb-454d-8dc6-f697eac681ae-telemetry-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-n45qh\" (UID: \"c6b7f3f9-1ffb-454d-8dc6-f697eac681ae\") " pod="openstack/install-certs-openstack-openstack-cell1-n45qh" Oct 08 20:23:05 crc kubenswrapper[4988]: I1008 20:23:05.517058 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-sriov-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c6b7f3f9-1ffb-454d-8dc6-f697eac681ae-neutron-sriov-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-n45qh\" (UID: \"c6b7f3f9-1ffb-454d-8dc6-f697eac681ae\") " pod="openstack/install-certs-openstack-openstack-cell1-n45qh" Oct 08 20:23:05 crc kubenswrapper[4988]: I1008 20:23:05.520287 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6mkgw\" (UniqueName: \"kubernetes.io/projected/c6b7f3f9-1ffb-454d-8dc6-f697eac681ae-kube-api-access-6mkgw\") pod \"install-certs-openstack-openstack-cell1-n45qh\" (UID: \"c6b7f3f9-1ffb-454d-8dc6-f697eac681ae\") " pod="openstack/install-certs-openstack-openstack-cell1-n45qh" Oct 08 20:23:05 crc kubenswrapper[4988]: I1008 20:23:05.665342 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-openstack-openstack-cell1-n45qh" Oct 08 20:23:06 crc kubenswrapper[4988]: I1008 20:23:06.266741 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-certs-openstack-openstack-cell1-n45qh"] Oct 08 20:23:07 crc kubenswrapper[4988]: I1008 20:23:07.261599 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-openstack-openstack-cell1-n45qh" event={"ID":"c6b7f3f9-1ffb-454d-8dc6-f697eac681ae","Type":"ContainerStarted","Data":"993ebda6a262be6763b8f41fa45bdfdfa275b35a32dbe98cf2b9dfb2044bec99"} Oct 08 20:23:07 crc kubenswrapper[4988]: I1008 20:23:07.262100 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-openstack-openstack-cell1-n45qh" event={"ID":"c6b7f3f9-1ffb-454d-8dc6-f697eac681ae","Type":"ContainerStarted","Data":"fa35e734579e60c1d171f42dbde3dc5237bd19aacd2c9e01ad79e7bc742bff56"} Oct 08 20:23:07 crc kubenswrapper[4988]: I1008 20:23:07.295829 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/install-certs-openstack-openstack-cell1-n45qh" podStartSLOduration=1.776265013 podStartE2EDuration="2.295800829s" podCreationTimestamp="2025-10-08 20:23:05 +0000 UTC" firstStartedPulling="2025-10-08 20:23:06.273741719 +0000 UTC m=+7931.723584499" lastFinishedPulling="2025-10-08 20:23:06.793277545 +0000 UTC m=+7932.243120315" observedRunningTime="2025-10-08 20:23:07.293221507 +0000 UTC m=+7932.743064367" watchObservedRunningTime="2025-10-08 20:23:07.295800829 +0000 UTC m=+7932.745643629" Oct 08 20:23:09 crc kubenswrapper[4988]: I1008 20:23:09.240048 4988 scope.go:117] "RemoveContainer" containerID="7541195e86b98682ea101ebff2b19f543aa9bab45dd1665e8f710922ad18c5e7" Oct 08 20:23:09 crc kubenswrapper[4988]: E1008 20:23:09.241067 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nm54f_openshift-machine-config-operator(f328a20b-5283-407e-8a93-2ab2da599f18)\"" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" Oct 08 20:23:22 crc kubenswrapper[4988]: I1008 20:23:22.238334 4988 scope.go:117] "RemoveContainer" containerID="7541195e86b98682ea101ebff2b19f543aa9bab45dd1665e8f710922ad18c5e7" Oct 08 20:23:22 crc kubenswrapper[4988]: E1008 20:23:22.239449 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nm54f_openshift-machine-config-operator(f328a20b-5283-407e-8a93-2ab2da599f18)\"" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" Oct 08 20:23:34 crc kubenswrapper[4988]: I1008 20:23:34.238638 4988 scope.go:117] "RemoveContainer" containerID="7541195e86b98682ea101ebff2b19f543aa9bab45dd1665e8f710922ad18c5e7" Oct 08 20:23:34 crc kubenswrapper[4988]: E1008 20:23:34.240080 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nm54f_openshift-machine-config-operator(f328a20b-5283-407e-8a93-2ab2da599f18)\"" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" Oct 08 20:23:46 crc kubenswrapper[4988]: I1008 20:23:46.238520 4988 scope.go:117] "RemoveContainer" containerID="7541195e86b98682ea101ebff2b19f543aa9bab45dd1665e8f710922ad18c5e7" Oct 08 20:23:46 crc kubenswrapper[4988]: E1008 20:23:46.239839 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nm54f_openshift-machine-config-operator(f328a20b-5283-407e-8a93-2ab2da599f18)\"" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" Oct 08 20:23:46 crc kubenswrapper[4988]: I1008 20:23:46.773997 4988 generic.go:334] "Generic (PLEG): container finished" podID="c6b7f3f9-1ffb-454d-8dc6-f697eac681ae" containerID="993ebda6a262be6763b8f41fa45bdfdfa275b35a32dbe98cf2b9dfb2044bec99" exitCode=0 Oct 08 20:23:46 crc kubenswrapper[4988]: I1008 20:23:46.774063 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-openstack-openstack-cell1-n45qh" event={"ID":"c6b7f3f9-1ffb-454d-8dc6-f697eac681ae","Type":"ContainerDied","Data":"993ebda6a262be6763b8f41fa45bdfdfa275b35a32dbe98cf2b9dfb2044bec99"} Oct 08 20:23:48 crc kubenswrapper[4988]: I1008 20:23:48.497765 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-openstack-openstack-cell1-n45qh" Oct 08 20:23:48 crc kubenswrapper[4988]: I1008 20:23:48.632703 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c6b7f3f9-1ffb-454d-8dc6-f697eac681ae-ssh-key\") pod \"c6b7f3f9-1ffb-454d-8dc6-f697eac681ae\" (UID: \"c6b7f3f9-1ffb-454d-8dc6-f697eac681ae\") " Oct 08 20:23:48 crc kubenswrapper[4988]: I1008 20:23:48.632759 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c6b7f3f9-1ffb-454d-8dc6-f697eac681ae-inventory\") pod \"c6b7f3f9-1ffb-454d-8dc6-f697eac681ae\" (UID: \"c6b7f3f9-1ffb-454d-8dc6-f697eac681ae\") " Oct 08 20:23:48 crc kubenswrapper[4988]: I1008 20:23:48.632799 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c6b7f3f9-1ffb-454d-8dc6-f697eac681ae-telemetry-combined-ca-bundle\") pod \"c6b7f3f9-1ffb-454d-8dc6-f697eac681ae\" (UID: \"c6b7f3f9-1ffb-454d-8dc6-f697eac681ae\") " Oct 08 20:23:48 crc kubenswrapper[4988]: I1008 20:23:48.632830 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-cell1-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/c6b7f3f9-1ffb-454d-8dc6-f697eac681ae-openstack-cell1-telemetry-default-certs-0\") pod \"c6b7f3f9-1ffb-454d-8dc6-f697eac681ae\" (UID: \"c6b7f3f9-1ffb-454d-8dc6-f697eac681ae\") " Oct 08 20:23:48 crc kubenswrapper[4988]: I1008 20:23:48.632879 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c6b7f3f9-1ffb-454d-8dc6-f697eac681ae-libvirt-combined-ca-bundle\") pod \"c6b7f3f9-1ffb-454d-8dc6-f697eac681ae\" (UID: \"c6b7f3f9-1ffb-454d-8dc6-f697eac681ae\") " Oct 08 20:23:48 crc kubenswrapper[4988]: I1008 20:23:48.632917 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-cell1-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/c6b7f3f9-1ffb-454d-8dc6-f697eac681ae-openstack-cell1-neutron-metadata-default-certs-0\") pod \"c6b7f3f9-1ffb-454d-8dc6-f697eac681ae\" (UID: \"c6b7f3f9-1ffb-454d-8dc6-f697eac681ae\") " Oct 08 20:23:48 crc kubenswrapper[4988]: I1008 20:23:48.632989 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-sriov-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c6b7f3f9-1ffb-454d-8dc6-f697eac681ae-neutron-sriov-combined-ca-bundle\") pod \"c6b7f3f9-1ffb-454d-8dc6-f697eac681ae\" (UID: \"c6b7f3f9-1ffb-454d-8dc6-f697eac681ae\") " Oct 08 20:23:48 crc kubenswrapper[4988]: I1008 20:23:48.633098 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-cell1-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/c6b7f3f9-1ffb-454d-8dc6-f697eac681ae-openstack-cell1-ovn-default-certs-0\") pod \"c6b7f3f9-1ffb-454d-8dc6-f697eac681ae\" (UID: \"c6b7f3f9-1ffb-454d-8dc6-f697eac681ae\") " Oct 08 20:23:48 crc kubenswrapper[4988]: I1008 20:23:48.633134 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c6b7f3f9-1ffb-454d-8dc6-f697eac681ae-neutron-metadata-combined-ca-bundle\") pod \"c6b7f3f9-1ffb-454d-8dc6-f697eac681ae\" (UID: \"c6b7f3f9-1ffb-454d-8dc6-f697eac681ae\") " Oct 08 20:23:48 crc kubenswrapper[4988]: I1008 20:23:48.633188 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6mkgw\" (UniqueName: \"kubernetes.io/projected/c6b7f3f9-1ffb-454d-8dc6-f697eac681ae-kube-api-access-6mkgw\") pod \"c6b7f3f9-1ffb-454d-8dc6-f697eac681ae\" (UID: \"c6b7f3f9-1ffb-454d-8dc6-f697eac681ae\") " Oct 08 20:23:48 crc kubenswrapper[4988]: I1008 20:23:48.633235 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c6b7f3f9-1ffb-454d-8dc6-f697eac681ae-nova-combined-ca-bundle\") pod \"c6b7f3f9-1ffb-454d-8dc6-f697eac681ae\" (UID: \"c6b7f3f9-1ffb-454d-8dc6-f697eac681ae\") " Oct 08 20:23:48 crc kubenswrapper[4988]: I1008 20:23:48.633301 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-cell1-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/c6b7f3f9-1ffb-454d-8dc6-f697eac681ae-openstack-cell1-libvirt-default-certs-0\") pod \"c6b7f3f9-1ffb-454d-8dc6-f697eac681ae\" (UID: \"c6b7f3f9-1ffb-454d-8dc6-f697eac681ae\") " Oct 08 20:23:48 crc kubenswrapper[4988]: I1008 20:23:48.633347 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c6b7f3f9-1ffb-454d-8dc6-f697eac681ae-bootstrap-combined-ca-bundle\") pod \"c6b7f3f9-1ffb-454d-8dc6-f697eac681ae\" (UID: \"c6b7f3f9-1ffb-454d-8dc6-f697eac681ae\") " Oct 08 20:23:48 crc kubenswrapper[4988]: I1008 20:23:48.633381 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c6b7f3f9-1ffb-454d-8dc6-f697eac681ae-ovn-combined-ca-bundle\") pod \"c6b7f3f9-1ffb-454d-8dc6-f697eac681ae\" (UID: \"c6b7f3f9-1ffb-454d-8dc6-f697eac681ae\") " Oct 08 20:23:48 crc kubenswrapper[4988]: I1008 20:23:48.633429 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-dhcp-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c6b7f3f9-1ffb-454d-8dc6-f697eac681ae-neutron-dhcp-combined-ca-bundle\") pod \"c6b7f3f9-1ffb-454d-8dc6-f697eac681ae\" (UID: \"c6b7f3f9-1ffb-454d-8dc6-f697eac681ae\") " Oct 08 20:23:48 crc kubenswrapper[4988]: I1008 20:23:48.639431 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c6b7f3f9-1ffb-454d-8dc6-f697eac681ae-nova-combined-ca-bundle" (OuterVolumeSpecName: "nova-combined-ca-bundle") pod "c6b7f3f9-1ffb-454d-8dc6-f697eac681ae" (UID: "c6b7f3f9-1ffb-454d-8dc6-f697eac681ae"). InnerVolumeSpecName "nova-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 20:23:48 crc kubenswrapper[4988]: I1008 20:23:48.641572 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c6b7f3f9-1ffb-454d-8dc6-f697eac681ae-neutron-metadata-combined-ca-bundle" (OuterVolumeSpecName: "neutron-metadata-combined-ca-bundle") pod "c6b7f3f9-1ffb-454d-8dc6-f697eac681ae" (UID: "c6b7f3f9-1ffb-454d-8dc6-f697eac681ae"). InnerVolumeSpecName "neutron-metadata-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 20:23:48 crc kubenswrapper[4988]: I1008 20:23:48.642190 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c6b7f3f9-1ffb-454d-8dc6-f697eac681ae-neutron-dhcp-combined-ca-bundle" (OuterVolumeSpecName: "neutron-dhcp-combined-ca-bundle") pod "c6b7f3f9-1ffb-454d-8dc6-f697eac681ae" (UID: "c6b7f3f9-1ffb-454d-8dc6-f697eac681ae"). InnerVolumeSpecName "neutron-dhcp-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 20:23:48 crc kubenswrapper[4988]: I1008 20:23:48.642587 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c6b7f3f9-1ffb-454d-8dc6-f697eac681ae-openstack-cell1-ovn-default-certs-0" (OuterVolumeSpecName: "openstack-cell1-ovn-default-certs-0") pod "c6b7f3f9-1ffb-454d-8dc6-f697eac681ae" (UID: "c6b7f3f9-1ffb-454d-8dc6-f697eac681ae"). InnerVolumeSpecName "openstack-cell1-ovn-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 20:23:48 crc kubenswrapper[4988]: I1008 20:23:48.644003 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c6b7f3f9-1ffb-454d-8dc6-f697eac681ae-openstack-cell1-telemetry-default-certs-0" (OuterVolumeSpecName: "openstack-cell1-telemetry-default-certs-0") pod "c6b7f3f9-1ffb-454d-8dc6-f697eac681ae" (UID: "c6b7f3f9-1ffb-454d-8dc6-f697eac681ae"). InnerVolumeSpecName "openstack-cell1-telemetry-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 20:23:48 crc kubenswrapper[4988]: I1008 20:23:48.646808 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c6b7f3f9-1ffb-454d-8dc6-f697eac681ae-bootstrap-combined-ca-bundle" (OuterVolumeSpecName: "bootstrap-combined-ca-bundle") pod "c6b7f3f9-1ffb-454d-8dc6-f697eac681ae" (UID: "c6b7f3f9-1ffb-454d-8dc6-f697eac681ae"). InnerVolumeSpecName "bootstrap-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 20:23:48 crc kubenswrapper[4988]: I1008 20:23:48.646819 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c6b7f3f9-1ffb-454d-8dc6-f697eac681ae-openstack-cell1-libvirt-default-certs-0" (OuterVolumeSpecName: "openstack-cell1-libvirt-default-certs-0") pod "c6b7f3f9-1ffb-454d-8dc6-f697eac681ae" (UID: "c6b7f3f9-1ffb-454d-8dc6-f697eac681ae"). InnerVolumeSpecName "openstack-cell1-libvirt-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 20:23:48 crc kubenswrapper[4988]: I1008 20:23:48.646853 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c6b7f3f9-1ffb-454d-8dc6-f697eac681ae-telemetry-combined-ca-bundle" (OuterVolumeSpecName: "telemetry-combined-ca-bundle") pod "c6b7f3f9-1ffb-454d-8dc6-f697eac681ae" (UID: "c6b7f3f9-1ffb-454d-8dc6-f697eac681ae"). InnerVolumeSpecName "telemetry-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 20:23:48 crc kubenswrapper[4988]: I1008 20:23:48.647523 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c6b7f3f9-1ffb-454d-8dc6-f697eac681ae-openstack-cell1-neutron-metadata-default-certs-0" (OuterVolumeSpecName: "openstack-cell1-neutron-metadata-default-certs-0") pod "c6b7f3f9-1ffb-454d-8dc6-f697eac681ae" (UID: "c6b7f3f9-1ffb-454d-8dc6-f697eac681ae"). InnerVolumeSpecName "openstack-cell1-neutron-metadata-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 20:23:48 crc kubenswrapper[4988]: I1008 20:23:48.647559 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c6b7f3f9-1ffb-454d-8dc6-f697eac681ae-libvirt-combined-ca-bundle" (OuterVolumeSpecName: "libvirt-combined-ca-bundle") pod "c6b7f3f9-1ffb-454d-8dc6-f697eac681ae" (UID: "c6b7f3f9-1ffb-454d-8dc6-f697eac681ae"). InnerVolumeSpecName "libvirt-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 20:23:48 crc kubenswrapper[4988]: I1008 20:23:48.656085 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c6b7f3f9-1ffb-454d-8dc6-f697eac681ae-ovn-combined-ca-bundle" (OuterVolumeSpecName: "ovn-combined-ca-bundle") pod "c6b7f3f9-1ffb-454d-8dc6-f697eac681ae" (UID: "c6b7f3f9-1ffb-454d-8dc6-f697eac681ae"). InnerVolumeSpecName "ovn-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 20:23:48 crc kubenswrapper[4988]: I1008 20:23:48.656237 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c6b7f3f9-1ffb-454d-8dc6-f697eac681ae-neutron-sriov-combined-ca-bundle" (OuterVolumeSpecName: "neutron-sriov-combined-ca-bundle") pod "c6b7f3f9-1ffb-454d-8dc6-f697eac681ae" (UID: "c6b7f3f9-1ffb-454d-8dc6-f697eac681ae"). InnerVolumeSpecName "neutron-sriov-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 20:23:48 crc kubenswrapper[4988]: I1008 20:23:48.656614 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c6b7f3f9-1ffb-454d-8dc6-f697eac681ae-kube-api-access-6mkgw" (OuterVolumeSpecName: "kube-api-access-6mkgw") pod "c6b7f3f9-1ffb-454d-8dc6-f697eac681ae" (UID: "c6b7f3f9-1ffb-454d-8dc6-f697eac681ae"). InnerVolumeSpecName "kube-api-access-6mkgw". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 20:23:48 crc kubenswrapper[4988]: I1008 20:23:48.680228 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c6b7f3f9-1ffb-454d-8dc6-f697eac681ae-inventory" (OuterVolumeSpecName: "inventory") pod "c6b7f3f9-1ffb-454d-8dc6-f697eac681ae" (UID: "c6b7f3f9-1ffb-454d-8dc6-f697eac681ae"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 20:23:48 crc kubenswrapper[4988]: I1008 20:23:48.680370 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c6b7f3f9-1ffb-454d-8dc6-f697eac681ae-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "c6b7f3f9-1ffb-454d-8dc6-f697eac681ae" (UID: "c6b7f3f9-1ffb-454d-8dc6-f697eac681ae"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 20:23:48 crc kubenswrapper[4988]: I1008 20:23:48.735717 4988 reconciler_common.go:293] "Volume detached for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c6b7f3f9-1ffb-454d-8dc6-f697eac681ae-telemetry-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 20:23:48 crc kubenswrapper[4988]: I1008 20:23:48.735750 4988 reconciler_common.go:293] "Volume detached for volume \"openstack-cell1-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/c6b7f3f9-1ffb-454d-8dc6-f697eac681ae-openstack-cell1-telemetry-default-certs-0\") on node \"crc\" DevicePath \"\"" Oct 08 20:23:48 crc kubenswrapper[4988]: I1008 20:23:48.735762 4988 reconciler_common.go:293] "Volume detached for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c6b7f3f9-1ffb-454d-8dc6-f697eac681ae-libvirt-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 20:23:48 crc kubenswrapper[4988]: I1008 20:23:48.735773 4988 reconciler_common.go:293] "Volume detached for volume \"openstack-cell1-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/c6b7f3f9-1ffb-454d-8dc6-f697eac681ae-openstack-cell1-neutron-metadata-default-certs-0\") on node \"crc\" DevicePath \"\"" Oct 08 20:23:48 crc kubenswrapper[4988]: I1008 20:23:48.735784 4988 reconciler_common.go:293] "Volume detached for volume \"neutron-sriov-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c6b7f3f9-1ffb-454d-8dc6-f697eac681ae-neutron-sriov-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 20:23:48 crc kubenswrapper[4988]: I1008 20:23:48.735794 4988 reconciler_common.go:293] "Volume detached for volume \"openstack-cell1-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/c6b7f3f9-1ffb-454d-8dc6-f697eac681ae-openstack-cell1-ovn-default-certs-0\") on node \"crc\" DevicePath \"\"" Oct 08 20:23:48 crc kubenswrapper[4988]: I1008 20:23:48.735804 4988 reconciler_common.go:293] "Volume detached for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c6b7f3f9-1ffb-454d-8dc6-f697eac681ae-neutron-metadata-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 20:23:48 crc kubenswrapper[4988]: I1008 20:23:48.735814 4988 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6mkgw\" (UniqueName: \"kubernetes.io/projected/c6b7f3f9-1ffb-454d-8dc6-f697eac681ae-kube-api-access-6mkgw\") on node \"crc\" DevicePath \"\"" Oct 08 20:23:48 crc kubenswrapper[4988]: I1008 20:23:48.735823 4988 reconciler_common.go:293] "Volume detached for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c6b7f3f9-1ffb-454d-8dc6-f697eac681ae-nova-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 20:23:48 crc kubenswrapper[4988]: I1008 20:23:48.735832 4988 reconciler_common.go:293] "Volume detached for volume \"openstack-cell1-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/c6b7f3f9-1ffb-454d-8dc6-f697eac681ae-openstack-cell1-libvirt-default-certs-0\") on node \"crc\" DevicePath \"\"" Oct 08 20:23:48 crc kubenswrapper[4988]: I1008 20:23:48.735841 4988 reconciler_common.go:293] "Volume detached for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c6b7f3f9-1ffb-454d-8dc6-f697eac681ae-bootstrap-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 20:23:48 crc kubenswrapper[4988]: I1008 20:23:48.735850 4988 reconciler_common.go:293] "Volume detached for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c6b7f3f9-1ffb-454d-8dc6-f697eac681ae-ovn-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 20:23:48 crc kubenswrapper[4988]: I1008 20:23:48.735859 4988 reconciler_common.go:293] "Volume detached for volume \"neutron-dhcp-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c6b7f3f9-1ffb-454d-8dc6-f697eac681ae-neutron-dhcp-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 20:23:48 crc kubenswrapper[4988]: I1008 20:23:48.735870 4988 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c6b7f3f9-1ffb-454d-8dc6-f697eac681ae-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 08 20:23:48 crc kubenswrapper[4988]: I1008 20:23:48.735878 4988 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c6b7f3f9-1ffb-454d-8dc6-f697eac681ae-inventory\") on node \"crc\" DevicePath \"\"" Oct 08 20:23:48 crc kubenswrapper[4988]: I1008 20:23:48.805158 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-openstack-openstack-cell1-n45qh" event={"ID":"c6b7f3f9-1ffb-454d-8dc6-f697eac681ae","Type":"ContainerDied","Data":"fa35e734579e60c1d171f42dbde3dc5237bd19aacd2c9e01ad79e7bc742bff56"} Oct 08 20:23:48 crc kubenswrapper[4988]: I1008 20:23:48.805235 4988 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fa35e734579e60c1d171f42dbde3dc5237bd19aacd2c9e01ad79e7bc742bff56" Oct 08 20:23:48 crc kubenswrapper[4988]: I1008 20:23:48.805346 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-openstack-openstack-cell1-n45qh" Oct 08 20:23:48 crc kubenswrapper[4988]: I1008 20:23:48.937238 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-openstack-openstack-cell1-5h2h6"] Oct 08 20:23:48 crc kubenswrapper[4988]: E1008 20:23:48.937970 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c6b7f3f9-1ffb-454d-8dc6-f697eac681ae" containerName="install-certs-openstack-openstack-cell1" Oct 08 20:23:48 crc kubenswrapper[4988]: I1008 20:23:48.938003 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="c6b7f3f9-1ffb-454d-8dc6-f697eac681ae" containerName="install-certs-openstack-openstack-cell1" Oct 08 20:23:48 crc kubenswrapper[4988]: I1008 20:23:48.938405 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="c6b7f3f9-1ffb-454d-8dc6-f697eac681ae" containerName="install-certs-openstack-openstack-cell1" Oct 08 20:23:48 crc kubenswrapper[4988]: I1008 20:23:48.939721 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-openstack-openstack-cell1-5h2h6" Oct 08 20:23:48 crc kubenswrapper[4988]: I1008 20:23:48.942152 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-4sdzr" Oct 08 20:23:48 crc kubenswrapper[4988]: I1008 20:23:48.942168 4988 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 08 20:23:48 crc kubenswrapper[4988]: I1008 20:23:48.942306 4988 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-config" Oct 08 20:23:48 crc kubenswrapper[4988]: I1008 20:23:48.942373 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Oct 08 20:23:48 crc kubenswrapper[4988]: I1008 20:23:48.944494 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Oct 08 20:23:48 crc kubenswrapper[4988]: I1008 20:23:48.951292 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-openstack-openstack-cell1-5h2h6"] Oct 08 20:23:49 crc kubenswrapper[4988]: I1008 20:23:49.042819 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/de67964d-9527-467b-8ca4-364cdec72a43-ssh-key\") pod \"ovn-openstack-openstack-cell1-5h2h6\" (UID: \"de67964d-9527-467b-8ca4-364cdec72a43\") " pod="openstack/ovn-openstack-openstack-cell1-5h2h6" Oct 08 20:23:49 crc kubenswrapper[4988]: I1008 20:23:49.042952 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/de67964d-9527-467b-8ca4-364cdec72a43-ovn-combined-ca-bundle\") pod \"ovn-openstack-openstack-cell1-5h2h6\" (UID: \"de67964d-9527-467b-8ca4-364cdec72a43\") " pod="openstack/ovn-openstack-openstack-cell1-5h2h6" Oct 08 20:23:49 crc kubenswrapper[4988]: I1008 20:23:49.043060 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bn29g\" (UniqueName: \"kubernetes.io/projected/de67964d-9527-467b-8ca4-364cdec72a43-kube-api-access-bn29g\") pod \"ovn-openstack-openstack-cell1-5h2h6\" (UID: \"de67964d-9527-467b-8ca4-364cdec72a43\") " pod="openstack/ovn-openstack-openstack-cell1-5h2h6" Oct 08 20:23:49 crc kubenswrapper[4988]: I1008 20:23:49.043596 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/de67964d-9527-467b-8ca4-364cdec72a43-ovncontroller-config-0\") pod \"ovn-openstack-openstack-cell1-5h2h6\" (UID: \"de67964d-9527-467b-8ca4-364cdec72a43\") " pod="openstack/ovn-openstack-openstack-cell1-5h2h6" Oct 08 20:23:49 crc kubenswrapper[4988]: I1008 20:23:49.043898 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/de67964d-9527-467b-8ca4-364cdec72a43-inventory\") pod \"ovn-openstack-openstack-cell1-5h2h6\" (UID: \"de67964d-9527-467b-8ca4-364cdec72a43\") " pod="openstack/ovn-openstack-openstack-cell1-5h2h6" Oct 08 20:23:49 crc kubenswrapper[4988]: I1008 20:23:49.146169 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/de67964d-9527-467b-8ca4-364cdec72a43-inventory\") pod \"ovn-openstack-openstack-cell1-5h2h6\" (UID: \"de67964d-9527-467b-8ca4-364cdec72a43\") " pod="openstack/ovn-openstack-openstack-cell1-5h2h6" Oct 08 20:23:49 crc kubenswrapper[4988]: I1008 20:23:49.146258 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/de67964d-9527-467b-8ca4-364cdec72a43-ssh-key\") pod \"ovn-openstack-openstack-cell1-5h2h6\" (UID: \"de67964d-9527-467b-8ca4-364cdec72a43\") " pod="openstack/ovn-openstack-openstack-cell1-5h2h6" Oct 08 20:23:49 crc kubenswrapper[4988]: I1008 20:23:49.146306 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/de67964d-9527-467b-8ca4-364cdec72a43-ovn-combined-ca-bundle\") pod \"ovn-openstack-openstack-cell1-5h2h6\" (UID: \"de67964d-9527-467b-8ca4-364cdec72a43\") " pod="openstack/ovn-openstack-openstack-cell1-5h2h6" Oct 08 20:23:49 crc kubenswrapper[4988]: I1008 20:23:49.146352 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bn29g\" (UniqueName: \"kubernetes.io/projected/de67964d-9527-467b-8ca4-364cdec72a43-kube-api-access-bn29g\") pod \"ovn-openstack-openstack-cell1-5h2h6\" (UID: \"de67964d-9527-467b-8ca4-364cdec72a43\") " pod="openstack/ovn-openstack-openstack-cell1-5h2h6" Oct 08 20:23:49 crc kubenswrapper[4988]: I1008 20:23:49.146487 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/de67964d-9527-467b-8ca4-364cdec72a43-ovncontroller-config-0\") pod \"ovn-openstack-openstack-cell1-5h2h6\" (UID: \"de67964d-9527-467b-8ca4-364cdec72a43\") " pod="openstack/ovn-openstack-openstack-cell1-5h2h6" Oct 08 20:23:49 crc kubenswrapper[4988]: I1008 20:23:49.147657 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/de67964d-9527-467b-8ca4-364cdec72a43-ovncontroller-config-0\") pod \"ovn-openstack-openstack-cell1-5h2h6\" (UID: \"de67964d-9527-467b-8ca4-364cdec72a43\") " pod="openstack/ovn-openstack-openstack-cell1-5h2h6" Oct 08 20:23:49 crc kubenswrapper[4988]: I1008 20:23:49.152361 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/de67964d-9527-467b-8ca4-364cdec72a43-ssh-key\") pod \"ovn-openstack-openstack-cell1-5h2h6\" (UID: \"de67964d-9527-467b-8ca4-364cdec72a43\") " pod="openstack/ovn-openstack-openstack-cell1-5h2h6" Oct 08 20:23:49 crc kubenswrapper[4988]: I1008 20:23:49.154121 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/de67964d-9527-467b-8ca4-364cdec72a43-inventory\") pod \"ovn-openstack-openstack-cell1-5h2h6\" (UID: \"de67964d-9527-467b-8ca4-364cdec72a43\") " pod="openstack/ovn-openstack-openstack-cell1-5h2h6" Oct 08 20:23:49 crc kubenswrapper[4988]: I1008 20:23:49.162758 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/de67964d-9527-467b-8ca4-364cdec72a43-ovn-combined-ca-bundle\") pod \"ovn-openstack-openstack-cell1-5h2h6\" (UID: \"de67964d-9527-467b-8ca4-364cdec72a43\") " pod="openstack/ovn-openstack-openstack-cell1-5h2h6" Oct 08 20:23:49 crc kubenswrapper[4988]: I1008 20:23:49.166726 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bn29g\" (UniqueName: \"kubernetes.io/projected/de67964d-9527-467b-8ca4-364cdec72a43-kube-api-access-bn29g\") pod \"ovn-openstack-openstack-cell1-5h2h6\" (UID: \"de67964d-9527-467b-8ca4-364cdec72a43\") " pod="openstack/ovn-openstack-openstack-cell1-5h2h6" Oct 08 20:23:49 crc kubenswrapper[4988]: I1008 20:23:49.266709 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-openstack-openstack-cell1-5h2h6" Oct 08 20:23:49 crc kubenswrapper[4988]: I1008 20:23:49.869790 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-openstack-openstack-cell1-5h2h6"] Oct 08 20:23:50 crc kubenswrapper[4988]: I1008 20:23:50.835269 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-openstack-openstack-cell1-5h2h6" event={"ID":"de67964d-9527-467b-8ca4-364cdec72a43","Type":"ContainerStarted","Data":"f85d6ccd518a86c5871ce5a23e37da114cd5032f2c6a64cb894e769d3cb957e5"} Oct 08 20:23:50 crc kubenswrapper[4988]: I1008 20:23:50.835691 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-openstack-openstack-cell1-5h2h6" event={"ID":"de67964d-9527-467b-8ca4-364cdec72a43","Type":"ContainerStarted","Data":"5dd310eab7f5e71305b60454536195367b908cf78370ed6f52e039b52262f377"} Oct 08 20:23:50 crc kubenswrapper[4988]: I1008 20:23:50.866566 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-openstack-openstack-cell1-5h2h6" podStartSLOduration=2.349290074 podStartE2EDuration="2.866542887s" podCreationTimestamp="2025-10-08 20:23:48 +0000 UTC" firstStartedPulling="2025-10-08 20:23:49.878936944 +0000 UTC m=+7975.328779714" lastFinishedPulling="2025-10-08 20:23:50.396189717 +0000 UTC m=+7975.846032527" observedRunningTime="2025-10-08 20:23:50.859293276 +0000 UTC m=+7976.309136086" watchObservedRunningTime="2025-10-08 20:23:50.866542887 +0000 UTC m=+7976.316385677" Oct 08 20:23:55 crc kubenswrapper[4988]: I1008 20:23:55.670870 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-5hl5q"] Oct 08 20:23:55 crc kubenswrapper[4988]: I1008 20:23:55.675680 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-5hl5q" Oct 08 20:23:55 crc kubenswrapper[4988]: I1008 20:23:55.683894 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-5hl5q"] Oct 08 20:23:55 crc kubenswrapper[4988]: I1008 20:23:55.831656 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-787t2\" (UniqueName: \"kubernetes.io/projected/80bb8388-273a-405d-bffd-1ca1bb6d5999-kube-api-access-787t2\") pod \"redhat-marketplace-5hl5q\" (UID: \"80bb8388-273a-405d-bffd-1ca1bb6d5999\") " pod="openshift-marketplace/redhat-marketplace-5hl5q" Oct 08 20:23:55 crc kubenswrapper[4988]: I1008 20:23:55.831883 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/80bb8388-273a-405d-bffd-1ca1bb6d5999-catalog-content\") pod \"redhat-marketplace-5hl5q\" (UID: \"80bb8388-273a-405d-bffd-1ca1bb6d5999\") " pod="openshift-marketplace/redhat-marketplace-5hl5q" Oct 08 20:23:55 crc kubenswrapper[4988]: I1008 20:23:55.831924 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/80bb8388-273a-405d-bffd-1ca1bb6d5999-utilities\") pod \"redhat-marketplace-5hl5q\" (UID: \"80bb8388-273a-405d-bffd-1ca1bb6d5999\") " pod="openshift-marketplace/redhat-marketplace-5hl5q" Oct 08 20:23:55 crc kubenswrapper[4988]: I1008 20:23:55.934431 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/80bb8388-273a-405d-bffd-1ca1bb6d5999-catalog-content\") pod \"redhat-marketplace-5hl5q\" (UID: \"80bb8388-273a-405d-bffd-1ca1bb6d5999\") " pod="openshift-marketplace/redhat-marketplace-5hl5q" Oct 08 20:23:55 crc kubenswrapper[4988]: I1008 20:23:55.934797 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/80bb8388-273a-405d-bffd-1ca1bb6d5999-utilities\") pod \"redhat-marketplace-5hl5q\" (UID: \"80bb8388-273a-405d-bffd-1ca1bb6d5999\") " pod="openshift-marketplace/redhat-marketplace-5hl5q" Oct 08 20:23:55 crc kubenswrapper[4988]: I1008 20:23:55.934819 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/80bb8388-273a-405d-bffd-1ca1bb6d5999-catalog-content\") pod \"redhat-marketplace-5hl5q\" (UID: \"80bb8388-273a-405d-bffd-1ca1bb6d5999\") " pod="openshift-marketplace/redhat-marketplace-5hl5q" Oct 08 20:23:55 crc kubenswrapper[4988]: I1008 20:23:55.934929 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-787t2\" (UniqueName: \"kubernetes.io/projected/80bb8388-273a-405d-bffd-1ca1bb6d5999-kube-api-access-787t2\") pod \"redhat-marketplace-5hl5q\" (UID: \"80bb8388-273a-405d-bffd-1ca1bb6d5999\") " pod="openshift-marketplace/redhat-marketplace-5hl5q" Oct 08 20:23:55 crc kubenswrapper[4988]: I1008 20:23:55.935129 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/80bb8388-273a-405d-bffd-1ca1bb6d5999-utilities\") pod \"redhat-marketplace-5hl5q\" (UID: \"80bb8388-273a-405d-bffd-1ca1bb6d5999\") " pod="openshift-marketplace/redhat-marketplace-5hl5q" Oct 08 20:23:55 crc kubenswrapper[4988]: I1008 20:23:55.969644 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-787t2\" (UniqueName: \"kubernetes.io/projected/80bb8388-273a-405d-bffd-1ca1bb6d5999-kube-api-access-787t2\") pod \"redhat-marketplace-5hl5q\" (UID: \"80bb8388-273a-405d-bffd-1ca1bb6d5999\") " pod="openshift-marketplace/redhat-marketplace-5hl5q" Oct 08 20:23:56 crc kubenswrapper[4988]: I1008 20:23:56.022405 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-5hl5q" Oct 08 20:23:56 crc kubenswrapper[4988]: I1008 20:23:56.512604 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-5hl5q"] Oct 08 20:23:56 crc kubenswrapper[4988]: I1008 20:23:56.913143 4988 generic.go:334] "Generic (PLEG): container finished" podID="80bb8388-273a-405d-bffd-1ca1bb6d5999" containerID="4383334a7dfa7470ba3bedd6ccc6bead9d8a338395570934006931f51e337489" exitCode=0 Oct 08 20:23:56 crc kubenswrapper[4988]: I1008 20:23:56.913223 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5hl5q" event={"ID":"80bb8388-273a-405d-bffd-1ca1bb6d5999","Type":"ContainerDied","Data":"4383334a7dfa7470ba3bedd6ccc6bead9d8a338395570934006931f51e337489"} Oct 08 20:23:56 crc kubenswrapper[4988]: I1008 20:23:56.913731 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5hl5q" event={"ID":"80bb8388-273a-405d-bffd-1ca1bb6d5999","Type":"ContainerStarted","Data":"e4730a2b2bd458011b018920fee049a801c1901c6f0d02ce39e762f5570be83d"} Oct 08 20:23:57 crc kubenswrapper[4988]: I1008 20:23:57.238653 4988 scope.go:117] "RemoveContainer" containerID="7541195e86b98682ea101ebff2b19f543aa9bab45dd1665e8f710922ad18c5e7" Oct 08 20:23:57 crc kubenswrapper[4988]: E1008 20:23:57.238934 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nm54f_openshift-machine-config-operator(f328a20b-5283-407e-8a93-2ab2da599f18)\"" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" Oct 08 20:23:57 crc kubenswrapper[4988]: I1008 20:23:57.929654 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5hl5q" event={"ID":"80bb8388-273a-405d-bffd-1ca1bb6d5999","Type":"ContainerStarted","Data":"997eb2f074fda40e4ebada5f637a1cc7dfb22e99d75c137433f169bcd3aff1cc"} Oct 08 20:23:58 crc kubenswrapper[4988]: I1008 20:23:58.943654 4988 generic.go:334] "Generic (PLEG): container finished" podID="80bb8388-273a-405d-bffd-1ca1bb6d5999" containerID="997eb2f074fda40e4ebada5f637a1cc7dfb22e99d75c137433f169bcd3aff1cc" exitCode=0 Oct 08 20:23:58 crc kubenswrapper[4988]: I1008 20:23:58.943720 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5hl5q" event={"ID":"80bb8388-273a-405d-bffd-1ca1bb6d5999","Type":"ContainerDied","Data":"997eb2f074fda40e4ebada5f637a1cc7dfb22e99d75c137433f169bcd3aff1cc"} Oct 08 20:23:59 crc kubenswrapper[4988]: I1008 20:23:59.956579 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5hl5q" event={"ID":"80bb8388-273a-405d-bffd-1ca1bb6d5999","Type":"ContainerStarted","Data":"3fb7312f2a74c975d50de97ef35903a3fc92fcf630d2e4f689378713ae3eda99"} Oct 08 20:23:59 crc kubenswrapper[4988]: I1008 20:23:59.983046 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-5hl5q" podStartSLOduration=2.42300077 podStartE2EDuration="4.983028068s" podCreationTimestamp="2025-10-08 20:23:55 +0000 UTC" firstStartedPulling="2025-10-08 20:23:56.915283931 +0000 UTC m=+7982.365126701" lastFinishedPulling="2025-10-08 20:23:59.475311239 +0000 UTC m=+7984.925153999" observedRunningTime="2025-10-08 20:23:59.981868641 +0000 UTC m=+7985.431711431" watchObservedRunningTime="2025-10-08 20:23:59.983028068 +0000 UTC m=+7985.432870838" Oct 08 20:24:03 crc kubenswrapper[4988]: I1008 20:24:03.070258 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-m44zj"] Oct 08 20:24:03 crc kubenswrapper[4988]: I1008 20:24:03.074754 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-m44zj" Oct 08 20:24:03 crc kubenswrapper[4988]: I1008 20:24:03.110933 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-m44zj"] Oct 08 20:24:03 crc kubenswrapper[4988]: I1008 20:24:03.217664 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gn9zm\" (UniqueName: \"kubernetes.io/projected/ad57069f-f141-4bdc-85d9-2969822b24c7-kube-api-access-gn9zm\") pod \"community-operators-m44zj\" (UID: \"ad57069f-f141-4bdc-85d9-2969822b24c7\") " pod="openshift-marketplace/community-operators-m44zj" Oct 08 20:24:03 crc kubenswrapper[4988]: I1008 20:24:03.217893 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ad57069f-f141-4bdc-85d9-2969822b24c7-catalog-content\") pod \"community-operators-m44zj\" (UID: \"ad57069f-f141-4bdc-85d9-2969822b24c7\") " pod="openshift-marketplace/community-operators-m44zj" Oct 08 20:24:03 crc kubenswrapper[4988]: I1008 20:24:03.217943 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ad57069f-f141-4bdc-85d9-2969822b24c7-utilities\") pod \"community-operators-m44zj\" (UID: \"ad57069f-f141-4bdc-85d9-2969822b24c7\") " pod="openshift-marketplace/community-operators-m44zj" Oct 08 20:24:03 crc kubenswrapper[4988]: I1008 20:24:03.319843 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ad57069f-f141-4bdc-85d9-2969822b24c7-catalog-content\") pod \"community-operators-m44zj\" (UID: \"ad57069f-f141-4bdc-85d9-2969822b24c7\") " pod="openshift-marketplace/community-operators-m44zj" Oct 08 20:24:03 crc kubenswrapper[4988]: I1008 20:24:03.319900 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ad57069f-f141-4bdc-85d9-2969822b24c7-utilities\") pod \"community-operators-m44zj\" (UID: \"ad57069f-f141-4bdc-85d9-2969822b24c7\") " pod="openshift-marketplace/community-operators-m44zj" Oct 08 20:24:03 crc kubenswrapper[4988]: I1008 20:24:03.319982 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gn9zm\" (UniqueName: \"kubernetes.io/projected/ad57069f-f141-4bdc-85d9-2969822b24c7-kube-api-access-gn9zm\") pod \"community-operators-m44zj\" (UID: \"ad57069f-f141-4bdc-85d9-2969822b24c7\") " pod="openshift-marketplace/community-operators-m44zj" Oct 08 20:24:03 crc kubenswrapper[4988]: I1008 20:24:03.320818 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ad57069f-f141-4bdc-85d9-2969822b24c7-catalog-content\") pod \"community-operators-m44zj\" (UID: \"ad57069f-f141-4bdc-85d9-2969822b24c7\") " pod="openshift-marketplace/community-operators-m44zj" Oct 08 20:24:03 crc kubenswrapper[4988]: I1008 20:24:03.320919 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ad57069f-f141-4bdc-85d9-2969822b24c7-utilities\") pod \"community-operators-m44zj\" (UID: \"ad57069f-f141-4bdc-85d9-2969822b24c7\") " pod="openshift-marketplace/community-operators-m44zj" Oct 08 20:24:03 crc kubenswrapper[4988]: I1008 20:24:03.349331 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gn9zm\" (UniqueName: \"kubernetes.io/projected/ad57069f-f141-4bdc-85d9-2969822b24c7-kube-api-access-gn9zm\") pod \"community-operators-m44zj\" (UID: \"ad57069f-f141-4bdc-85d9-2969822b24c7\") " pod="openshift-marketplace/community-operators-m44zj" Oct 08 20:24:03 crc kubenswrapper[4988]: I1008 20:24:03.414536 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-m44zj" Oct 08 20:24:03 crc kubenswrapper[4988]: I1008 20:24:03.995110 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-m44zj"] Oct 08 20:24:03 crc kubenswrapper[4988]: W1008 20:24:03.998200 4988 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podad57069f_f141_4bdc_85d9_2969822b24c7.slice/crio-133572d393e92fc9f0ffd0c465b6e25cd94c1dc19fbbbf6faf095f40ac129c9a WatchSource:0}: Error finding container 133572d393e92fc9f0ffd0c465b6e25cd94c1dc19fbbbf6faf095f40ac129c9a: Status 404 returned error can't find the container with id 133572d393e92fc9f0ffd0c465b6e25cd94c1dc19fbbbf6faf095f40ac129c9a Oct 08 20:24:05 crc kubenswrapper[4988]: I1008 20:24:05.029590 4988 generic.go:334] "Generic (PLEG): container finished" podID="ad57069f-f141-4bdc-85d9-2969822b24c7" containerID="017832e7ed93d2bc3656858fbd275f926271a6a56701b1deb4ea65e8076b08a0" exitCode=0 Oct 08 20:24:05 crc kubenswrapper[4988]: I1008 20:24:05.029669 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-m44zj" event={"ID":"ad57069f-f141-4bdc-85d9-2969822b24c7","Type":"ContainerDied","Data":"017832e7ed93d2bc3656858fbd275f926271a6a56701b1deb4ea65e8076b08a0"} Oct 08 20:24:05 crc kubenswrapper[4988]: I1008 20:24:05.029960 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-m44zj" event={"ID":"ad57069f-f141-4bdc-85d9-2969822b24c7","Type":"ContainerStarted","Data":"133572d393e92fc9f0ffd0c465b6e25cd94c1dc19fbbbf6faf095f40ac129c9a"} Oct 08 20:24:06 crc kubenswrapper[4988]: I1008 20:24:06.022557 4988 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-5hl5q" Oct 08 20:24:06 crc kubenswrapper[4988]: I1008 20:24:06.022936 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-5hl5q" Oct 08 20:24:06 crc kubenswrapper[4988]: I1008 20:24:06.101755 4988 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-5hl5q" Oct 08 20:24:06 crc kubenswrapper[4988]: I1008 20:24:06.166983 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-5hl5q" Oct 08 20:24:07 crc kubenswrapper[4988]: I1008 20:24:07.059117 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-m44zj" event={"ID":"ad57069f-f141-4bdc-85d9-2969822b24c7","Type":"ContainerStarted","Data":"ad70f8ce56e85f32e184ef80518b705f333f2b6e313416c343562bc05a3716d9"} Oct 08 20:24:07 crc kubenswrapper[4988]: I1008 20:24:07.221083 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-5hl5q"] Oct 08 20:24:08 crc kubenswrapper[4988]: I1008 20:24:08.076253 4988 generic.go:334] "Generic (PLEG): container finished" podID="ad57069f-f141-4bdc-85d9-2969822b24c7" containerID="ad70f8ce56e85f32e184ef80518b705f333f2b6e313416c343562bc05a3716d9" exitCode=0 Oct 08 20:24:08 crc kubenswrapper[4988]: I1008 20:24:08.076426 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-m44zj" event={"ID":"ad57069f-f141-4bdc-85d9-2969822b24c7","Type":"ContainerDied","Data":"ad70f8ce56e85f32e184ef80518b705f333f2b6e313416c343562bc05a3716d9"} Oct 08 20:24:08 crc kubenswrapper[4988]: I1008 20:24:08.076615 4988 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-5hl5q" podUID="80bb8388-273a-405d-bffd-1ca1bb6d5999" containerName="registry-server" containerID="cri-o://3fb7312f2a74c975d50de97ef35903a3fc92fcf630d2e4f689378713ae3eda99" gracePeriod=2 Oct 08 20:24:08 crc kubenswrapper[4988]: I1008 20:24:08.666264 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-5hl5q" Oct 08 20:24:08 crc kubenswrapper[4988]: I1008 20:24:08.765480 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-787t2\" (UniqueName: \"kubernetes.io/projected/80bb8388-273a-405d-bffd-1ca1bb6d5999-kube-api-access-787t2\") pod \"80bb8388-273a-405d-bffd-1ca1bb6d5999\" (UID: \"80bb8388-273a-405d-bffd-1ca1bb6d5999\") " Oct 08 20:24:08 crc kubenswrapper[4988]: I1008 20:24:08.765731 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/80bb8388-273a-405d-bffd-1ca1bb6d5999-utilities\") pod \"80bb8388-273a-405d-bffd-1ca1bb6d5999\" (UID: \"80bb8388-273a-405d-bffd-1ca1bb6d5999\") " Oct 08 20:24:08 crc kubenswrapper[4988]: I1008 20:24:08.765788 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/80bb8388-273a-405d-bffd-1ca1bb6d5999-catalog-content\") pod \"80bb8388-273a-405d-bffd-1ca1bb6d5999\" (UID: \"80bb8388-273a-405d-bffd-1ca1bb6d5999\") " Oct 08 20:24:08 crc kubenswrapper[4988]: I1008 20:24:08.767163 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/80bb8388-273a-405d-bffd-1ca1bb6d5999-utilities" (OuterVolumeSpecName: "utilities") pod "80bb8388-273a-405d-bffd-1ca1bb6d5999" (UID: "80bb8388-273a-405d-bffd-1ca1bb6d5999"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 20:24:08 crc kubenswrapper[4988]: I1008 20:24:08.772099 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/80bb8388-273a-405d-bffd-1ca1bb6d5999-kube-api-access-787t2" (OuterVolumeSpecName: "kube-api-access-787t2") pod "80bb8388-273a-405d-bffd-1ca1bb6d5999" (UID: "80bb8388-273a-405d-bffd-1ca1bb6d5999"). InnerVolumeSpecName "kube-api-access-787t2". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 20:24:08 crc kubenswrapper[4988]: I1008 20:24:08.778376 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/80bb8388-273a-405d-bffd-1ca1bb6d5999-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "80bb8388-273a-405d-bffd-1ca1bb6d5999" (UID: "80bb8388-273a-405d-bffd-1ca1bb6d5999"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 20:24:08 crc kubenswrapper[4988]: I1008 20:24:08.869054 4988 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/80bb8388-273a-405d-bffd-1ca1bb6d5999-utilities\") on node \"crc\" DevicePath \"\"" Oct 08 20:24:08 crc kubenswrapper[4988]: I1008 20:24:08.869096 4988 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/80bb8388-273a-405d-bffd-1ca1bb6d5999-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 08 20:24:08 crc kubenswrapper[4988]: I1008 20:24:08.869111 4988 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-787t2\" (UniqueName: \"kubernetes.io/projected/80bb8388-273a-405d-bffd-1ca1bb6d5999-kube-api-access-787t2\") on node \"crc\" DevicePath \"\"" Oct 08 20:24:09 crc kubenswrapper[4988]: I1008 20:24:09.088718 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-m44zj" event={"ID":"ad57069f-f141-4bdc-85d9-2969822b24c7","Type":"ContainerStarted","Data":"9ed7d8e4857ee617837546884998a9f99ee38cc54d87c829d709deabbf7d04d8"} Oct 08 20:24:09 crc kubenswrapper[4988]: I1008 20:24:09.091624 4988 generic.go:334] "Generic (PLEG): container finished" podID="80bb8388-273a-405d-bffd-1ca1bb6d5999" containerID="3fb7312f2a74c975d50de97ef35903a3fc92fcf630d2e4f689378713ae3eda99" exitCode=0 Oct 08 20:24:09 crc kubenswrapper[4988]: I1008 20:24:09.091675 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5hl5q" event={"ID":"80bb8388-273a-405d-bffd-1ca1bb6d5999","Type":"ContainerDied","Data":"3fb7312f2a74c975d50de97ef35903a3fc92fcf630d2e4f689378713ae3eda99"} Oct 08 20:24:09 crc kubenswrapper[4988]: I1008 20:24:09.091706 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5hl5q" event={"ID":"80bb8388-273a-405d-bffd-1ca1bb6d5999","Type":"ContainerDied","Data":"e4730a2b2bd458011b018920fee049a801c1901c6f0d02ce39e762f5570be83d"} Oct 08 20:24:09 crc kubenswrapper[4988]: I1008 20:24:09.091726 4988 scope.go:117] "RemoveContainer" containerID="3fb7312f2a74c975d50de97ef35903a3fc92fcf630d2e4f689378713ae3eda99" Oct 08 20:24:09 crc kubenswrapper[4988]: I1008 20:24:09.091781 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-5hl5q" Oct 08 20:24:09 crc kubenswrapper[4988]: I1008 20:24:09.118371 4988 scope.go:117] "RemoveContainer" containerID="997eb2f074fda40e4ebada5f637a1cc7dfb22e99d75c137433f169bcd3aff1cc" Oct 08 20:24:09 crc kubenswrapper[4988]: I1008 20:24:09.127784 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-m44zj" podStartSLOduration=2.491765795 podStartE2EDuration="6.127753728s" podCreationTimestamp="2025-10-08 20:24:03 +0000 UTC" firstStartedPulling="2025-10-08 20:24:05.032168367 +0000 UTC m=+7990.482011177" lastFinishedPulling="2025-10-08 20:24:08.66815634 +0000 UTC m=+7994.117999110" observedRunningTime="2025-10-08 20:24:09.110011903 +0000 UTC m=+7994.559854763" watchObservedRunningTime="2025-10-08 20:24:09.127753728 +0000 UTC m=+7994.577596538" Oct 08 20:24:09 crc kubenswrapper[4988]: I1008 20:24:09.149257 4988 scope.go:117] "RemoveContainer" containerID="4383334a7dfa7470ba3bedd6ccc6bead9d8a338395570934006931f51e337489" Oct 08 20:24:09 crc kubenswrapper[4988]: I1008 20:24:09.164555 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-5hl5q"] Oct 08 20:24:09 crc kubenswrapper[4988]: I1008 20:24:09.186623 4988 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-5hl5q"] Oct 08 20:24:09 crc kubenswrapper[4988]: I1008 20:24:09.191109 4988 scope.go:117] "RemoveContainer" containerID="3fb7312f2a74c975d50de97ef35903a3fc92fcf630d2e4f689378713ae3eda99" Oct 08 20:24:09 crc kubenswrapper[4988]: E1008 20:24:09.191644 4988 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3fb7312f2a74c975d50de97ef35903a3fc92fcf630d2e4f689378713ae3eda99\": container with ID starting with 3fb7312f2a74c975d50de97ef35903a3fc92fcf630d2e4f689378713ae3eda99 not found: ID does not exist" containerID="3fb7312f2a74c975d50de97ef35903a3fc92fcf630d2e4f689378713ae3eda99" Oct 08 20:24:09 crc kubenswrapper[4988]: I1008 20:24:09.191709 4988 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3fb7312f2a74c975d50de97ef35903a3fc92fcf630d2e4f689378713ae3eda99"} err="failed to get container status \"3fb7312f2a74c975d50de97ef35903a3fc92fcf630d2e4f689378713ae3eda99\": rpc error: code = NotFound desc = could not find container \"3fb7312f2a74c975d50de97ef35903a3fc92fcf630d2e4f689378713ae3eda99\": container with ID starting with 3fb7312f2a74c975d50de97ef35903a3fc92fcf630d2e4f689378713ae3eda99 not found: ID does not exist" Oct 08 20:24:09 crc kubenswrapper[4988]: I1008 20:24:09.191751 4988 scope.go:117] "RemoveContainer" containerID="997eb2f074fda40e4ebada5f637a1cc7dfb22e99d75c137433f169bcd3aff1cc" Oct 08 20:24:09 crc kubenswrapper[4988]: E1008 20:24:09.192075 4988 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"997eb2f074fda40e4ebada5f637a1cc7dfb22e99d75c137433f169bcd3aff1cc\": container with ID starting with 997eb2f074fda40e4ebada5f637a1cc7dfb22e99d75c137433f169bcd3aff1cc not found: ID does not exist" containerID="997eb2f074fda40e4ebada5f637a1cc7dfb22e99d75c137433f169bcd3aff1cc" Oct 08 20:24:09 crc kubenswrapper[4988]: I1008 20:24:09.192106 4988 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"997eb2f074fda40e4ebada5f637a1cc7dfb22e99d75c137433f169bcd3aff1cc"} err="failed to get container status \"997eb2f074fda40e4ebada5f637a1cc7dfb22e99d75c137433f169bcd3aff1cc\": rpc error: code = NotFound desc = could not find container \"997eb2f074fda40e4ebada5f637a1cc7dfb22e99d75c137433f169bcd3aff1cc\": container with ID starting with 997eb2f074fda40e4ebada5f637a1cc7dfb22e99d75c137433f169bcd3aff1cc not found: ID does not exist" Oct 08 20:24:09 crc kubenswrapper[4988]: I1008 20:24:09.192175 4988 scope.go:117] "RemoveContainer" containerID="4383334a7dfa7470ba3bedd6ccc6bead9d8a338395570934006931f51e337489" Oct 08 20:24:09 crc kubenswrapper[4988]: E1008 20:24:09.192594 4988 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4383334a7dfa7470ba3bedd6ccc6bead9d8a338395570934006931f51e337489\": container with ID starting with 4383334a7dfa7470ba3bedd6ccc6bead9d8a338395570934006931f51e337489 not found: ID does not exist" containerID="4383334a7dfa7470ba3bedd6ccc6bead9d8a338395570934006931f51e337489" Oct 08 20:24:09 crc kubenswrapper[4988]: I1008 20:24:09.192621 4988 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4383334a7dfa7470ba3bedd6ccc6bead9d8a338395570934006931f51e337489"} err="failed to get container status \"4383334a7dfa7470ba3bedd6ccc6bead9d8a338395570934006931f51e337489\": rpc error: code = NotFound desc = could not find container \"4383334a7dfa7470ba3bedd6ccc6bead9d8a338395570934006931f51e337489\": container with ID starting with 4383334a7dfa7470ba3bedd6ccc6bead9d8a338395570934006931f51e337489 not found: ID does not exist" Oct 08 20:24:09 crc kubenswrapper[4988]: I1008 20:24:09.243148 4988 scope.go:117] "RemoveContainer" containerID="7541195e86b98682ea101ebff2b19f543aa9bab45dd1665e8f710922ad18c5e7" Oct 08 20:24:09 crc kubenswrapper[4988]: E1008 20:24:09.243425 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nm54f_openshift-machine-config-operator(f328a20b-5283-407e-8a93-2ab2da599f18)\"" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" Oct 08 20:24:09 crc kubenswrapper[4988]: I1008 20:24:09.249221 4988 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="80bb8388-273a-405d-bffd-1ca1bb6d5999" path="/var/lib/kubelet/pods/80bb8388-273a-405d-bffd-1ca1bb6d5999/volumes" Oct 08 20:24:13 crc kubenswrapper[4988]: I1008 20:24:13.415690 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-m44zj" Oct 08 20:24:13 crc kubenswrapper[4988]: I1008 20:24:13.416494 4988 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-m44zj" Oct 08 20:24:13 crc kubenswrapper[4988]: I1008 20:24:13.509031 4988 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-m44zj" Oct 08 20:24:14 crc kubenswrapper[4988]: I1008 20:24:14.239880 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-m44zj" Oct 08 20:24:14 crc kubenswrapper[4988]: I1008 20:24:14.425944 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-m44zj"] Oct 08 20:24:16 crc kubenswrapper[4988]: I1008 20:24:16.170318 4988 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-m44zj" podUID="ad57069f-f141-4bdc-85d9-2969822b24c7" containerName="registry-server" containerID="cri-o://9ed7d8e4857ee617837546884998a9f99ee38cc54d87c829d709deabbf7d04d8" gracePeriod=2 Oct 08 20:24:16 crc kubenswrapper[4988]: I1008 20:24:16.736369 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-m44zj" Oct 08 20:24:16 crc kubenswrapper[4988]: I1008 20:24:16.859103 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ad57069f-f141-4bdc-85d9-2969822b24c7-catalog-content\") pod \"ad57069f-f141-4bdc-85d9-2969822b24c7\" (UID: \"ad57069f-f141-4bdc-85d9-2969822b24c7\") " Oct 08 20:24:16 crc kubenswrapper[4988]: I1008 20:24:16.859159 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gn9zm\" (UniqueName: \"kubernetes.io/projected/ad57069f-f141-4bdc-85d9-2969822b24c7-kube-api-access-gn9zm\") pod \"ad57069f-f141-4bdc-85d9-2969822b24c7\" (UID: \"ad57069f-f141-4bdc-85d9-2969822b24c7\") " Oct 08 20:24:16 crc kubenswrapper[4988]: I1008 20:24:16.859298 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ad57069f-f141-4bdc-85d9-2969822b24c7-utilities\") pod \"ad57069f-f141-4bdc-85d9-2969822b24c7\" (UID: \"ad57069f-f141-4bdc-85d9-2969822b24c7\") " Oct 08 20:24:16 crc kubenswrapper[4988]: I1008 20:24:16.860262 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ad57069f-f141-4bdc-85d9-2969822b24c7-utilities" (OuterVolumeSpecName: "utilities") pod "ad57069f-f141-4bdc-85d9-2969822b24c7" (UID: "ad57069f-f141-4bdc-85d9-2969822b24c7"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 20:24:16 crc kubenswrapper[4988]: I1008 20:24:16.866517 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ad57069f-f141-4bdc-85d9-2969822b24c7-kube-api-access-gn9zm" (OuterVolumeSpecName: "kube-api-access-gn9zm") pod "ad57069f-f141-4bdc-85d9-2969822b24c7" (UID: "ad57069f-f141-4bdc-85d9-2969822b24c7"). InnerVolumeSpecName "kube-api-access-gn9zm". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 20:24:16 crc kubenswrapper[4988]: I1008 20:24:16.939056 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ad57069f-f141-4bdc-85d9-2969822b24c7-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "ad57069f-f141-4bdc-85d9-2969822b24c7" (UID: "ad57069f-f141-4bdc-85d9-2969822b24c7"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 20:24:16 crc kubenswrapper[4988]: I1008 20:24:16.962355 4988 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ad57069f-f141-4bdc-85d9-2969822b24c7-utilities\") on node \"crc\" DevicePath \"\"" Oct 08 20:24:16 crc kubenswrapper[4988]: I1008 20:24:16.962437 4988 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ad57069f-f141-4bdc-85d9-2969822b24c7-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 08 20:24:16 crc kubenswrapper[4988]: I1008 20:24:16.962462 4988 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gn9zm\" (UniqueName: \"kubernetes.io/projected/ad57069f-f141-4bdc-85d9-2969822b24c7-kube-api-access-gn9zm\") on node \"crc\" DevicePath \"\"" Oct 08 20:24:17 crc kubenswrapper[4988]: I1008 20:24:17.189174 4988 generic.go:334] "Generic (PLEG): container finished" podID="ad57069f-f141-4bdc-85d9-2969822b24c7" containerID="9ed7d8e4857ee617837546884998a9f99ee38cc54d87c829d709deabbf7d04d8" exitCode=0 Oct 08 20:24:17 crc kubenswrapper[4988]: I1008 20:24:17.189260 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-m44zj" event={"ID":"ad57069f-f141-4bdc-85d9-2969822b24c7","Type":"ContainerDied","Data":"9ed7d8e4857ee617837546884998a9f99ee38cc54d87c829d709deabbf7d04d8"} Oct 08 20:24:17 crc kubenswrapper[4988]: I1008 20:24:17.189318 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-m44zj" event={"ID":"ad57069f-f141-4bdc-85d9-2969822b24c7","Type":"ContainerDied","Data":"133572d393e92fc9f0ffd0c465b6e25cd94c1dc19fbbbf6faf095f40ac129c9a"} Oct 08 20:24:17 crc kubenswrapper[4988]: I1008 20:24:17.189335 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-m44zj" Oct 08 20:24:17 crc kubenswrapper[4988]: I1008 20:24:17.189348 4988 scope.go:117] "RemoveContainer" containerID="9ed7d8e4857ee617837546884998a9f99ee38cc54d87c829d709deabbf7d04d8" Oct 08 20:24:17 crc kubenswrapper[4988]: I1008 20:24:17.231049 4988 scope.go:117] "RemoveContainer" containerID="ad70f8ce56e85f32e184ef80518b705f333f2b6e313416c343562bc05a3716d9" Oct 08 20:24:17 crc kubenswrapper[4988]: I1008 20:24:17.285847 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-m44zj"] Oct 08 20:24:17 crc kubenswrapper[4988]: I1008 20:24:17.287000 4988 scope.go:117] "RemoveContainer" containerID="017832e7ed93d2bc3656858fbd275f926271a6a56701b1deb4ea65e8076b08a0" Oct 08 20:24:17 crc kubenswrapper[4988]: I1008 20:24:17.298732 4988 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-m44zj"] Oct 08 20:24:17 crc kubenswrapper[4988]: I1008 20:24:17.341665 4988 scope.go:117] "RemoveContainer" containerID="9ed7d8e4857ee617837546884998a9f99ee38cc54d87c829d709deabbf7d04d8" Oct 08 20:24:17 crc kubenswrapper[4988]: E1008 20:24:17.342230 4988 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9ed7d8e4857ee617837546884998a9f99ee38cc54d87c829d709deabbf7d04d8\": container with ID starting with 9ed7d8e4857ee617837546884998a9f99ee38cc54d87c829d709deabbf7d04d8 not found: ID does not exist" containerID="9ed7d8e4857ee617837546884998a9f99ee38cc54d87c829d709deabbf7d04d8" Oct 08 20:24:17 crc kubenswrapper[4988]: I1008 20:24:17.342287 4988 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9ed7d8e4857ee617837546884998a9f99ee38cc54d87c829d709deabbf7d04d8"} err="failed to get container status \"9ed7d8e4857ee617837546884998a9f99ee38cc54d87c829d709deabbf7d04d8\": rpc error: code = NotFound desc = could not find container \"9ed7d8e4857ee617837546884998a9f99ee38cc54d87c829d709deabbf7d04d8\": container with ID starting with 9ed7d8e4857ee617837546884998a9f99ee38cc54d87c829d709deabbf7d04d8 not found: ID does not exist" Oct 08 20:24:17 crc kubenswrapper[4988]: I1008 20:24:17.342327 4988 scope.go:117] "RemoveContainer" containerID="ad70f8ce56e85f32e184ef80518b705f333f2b6e313416c343562bc05a3716d9" Oct 08 20:24:17 crc kubenswrapper[4988]: E1008 20:24:17.342695 4988 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ad70f8ce56e85f32e184ef80518b705f333f2b6e313416c343562bc05a3716d9\": container with ID starting with ad70f8ce56e85f32e184ef80518b705f333f2b6e313416c343562bc05a3716d9 not found: ID does not exist" containerID="ad70f8ce56e85f32e184ef80518b705f333f2b6e313416c343562bc05a3716d9" Oct 08 20:24:17 crc kubenswrapper[4988]: I1008 20:24:17.342737 4988 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ad70f8ce56e85f32e184ef80518b705f333f2b6e313416c343562bc05a3716d9"} err="failed to get container status \"ad70f8ce56e85f32e184ef80518b705f333f2b6e313416c343562bc05a3716d9\": rpc error: code = NotFound desc = could not find container \"ad70f8ce56e85f32e184ef80518b705f333f2b6e313416c343562bc05a3716d9\": container with ID starting with ad70f8ce56e85f32e184ef80518b705f333f2b6e313416c343562bc05a3716d9 not found: ID does not exist" Oct 08 20:24:17 crc kubenswrapper[4988]: I1008 20:24:17.342772 4988 scope.go:117] "RemoveContainer" containerID="017832e7ed93d2bc3656858fbd275f926271a6a56701b1deb4ea65e8076b08a0" Oct 08 20:24:17 crc kubenswrapper[4988]: E1008 20:24:17.343048 4988 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"017832e7ed93d2bc3656858fbd275f926271a6a56701b1deb4ea65e8076b08a0\": container with ID starting with 017832e7ed93d2bc3656858fbd275f926271a6a56701b1deb4ea65e8076b08a0 not found: ID does not exist" containerID="017832e7ed93d2bc3656858fbd275f926271a6a56701b1deb4ea65e8076b08a0" Oct 08 20:24:17 crc kubenswrapper[4988]: I1008 20:24:17.343081 4988 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"017832e7ed93d2bc3656858fbd275f926271a6a56701b1deb4ea65e8076b08a0"} err="failed to get container status \"017832e7ed93d2bc3656858fbd275f926271a6a56701b1deb4ea65e8076b08a0\": rpc error: code = NotFound desc = could not find container \"017832e7ed93d2bc3656858fbd275f926271a6a56701b1deb4ea65e8076b08a0\": container with ID starting with 017832e7ed93d2bc3656858fbd275f926271a6a56701b1deb4ea65e8076b08a0 not found: ID does not exist" Oct 08 20:24:19 crc kubenswrapper[4988]: I1008 20:24:19.249081 4988 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ad57069f-f141-4bdc-85d9-2969822b24c7" path="/var/lib/kubelet/pods/ad57069f-f141-4bdc-85d9-2969822b24c7/volumes" Oct 08 20:24:22 crc kubenswrapper[4988]: I1008 20:24:22.237449 4988 scope.go:117] "RemoveContainer" containerID="7541195e86b98682ea101ebff2b19f543aa9bab45dd1665e8f710922ad18c5e7" Oct 08 20:24:22 crc kubenswrapper[4988]: E1008 20:24:22.238191 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nm54f_openshift-machine-config-operator(f328a20b-5283-407e-8a93-2ab2da599f18)\"" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" Oct 08 20:24:35 crc kubenswrapper[4988]: I1008 20:24:35.252086 4988 scope.go:117] "RemoveContainer" containerID="7541195e86b98682ea101ebff2b19f543aa9bab45dd1665e8f710922ad18c5e7" Oct 08 20:24:35 crc kubenswrapper[4988]: E1008 20:24:35.253955 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nm54f_openshift-machine-config-operator(f328a20b-5283-407e-8a93-2ab2da599f18)\"" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" Oct 08 20:24:48 crc kubenswrapper[4988]: I1008 20:24:48.237645 4988 scope.go:117] "RemoveContainer" containerID="7541195e86b98682ea101ebff2b19f543aa9bab45dd1665e8f710922ad18c5e7" Oct 08 20:24:48 crc kubenswrapper[4988]: E1008 20:24:48.238284 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nm54f_openshift-machine-config-operator(f328a20b-5283-407e-8a93-2ab2da599f18)\"" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" Oct 08 20:25:01 crc kubenswrapper[4988]: I1008 20:25:01.238407 4988 scope.go:117] "RemoveContainer" containerID="7541195e86b98682ea101ebff2b19f543aa9bab45dd1665e8f710922ad18c5e7" Oct 08 20:25:01 crc kubenswrapper[4988]: E1008 20:25:01.239436 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nm54f_openshift-machine-config-operator(f328a20b-5283-407e-8a93-2ab2da599f18)\"" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" Oct 08 20:25:01 crc kubenswrapper[4988]: I1008 20:25:01.741035 4988 generic.go:334] "Generic (PLEG): container finished" podID="de67964d-9527-467b-8ca4-364cdec72a43" containerID="f85d6ccd518a86c5871ce5a23e37da114cd5032f2c6a64cb894e769d3cb957e5" exitCode=0 Oct 08 20:25:01 crc kubenswrapper[4988]: I1008 20:25:01.741088 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-openstack-openstack-cell1-5h2h6" event={"ID":"de67964d-9527-467b-8ca4-364cdec72a43","Type":"ContainerDied","Data":"f85d6ccd518a86c5871ce5a23e37da114cd5032f2c6a64cb894e769d3cb957e5"} Oct 08 20:25:03 crc kubenswrapper[4988]: I1008 20:25:03.383058 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-openstack-openstack-cell1-5h2h6" Oct 08 20:25:03 crc kubenswrapper[4988]: I1008 20:25:03.431738 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/de67964d-9527-467b-8ca4-364cdec72a43-ovn-combined-ca-bundle\") pod \"de67964d-9527-467b-8ca4-364cdec72a43\" (UID: \"de67964d-9527-467b-8ca4-364cdec72a43\") " Oct 08 20:25:03 crc kubenswrapper[4988]: I1008 20:25:03.431792 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/de67964d-9527-467b-8ca4-364cdec72a43-inventory\") pod \"de67964d-9527-467b-8ca4-364cdec72a43\" (UID: \"de67964d-9527-467b-8ca4-364cdec72a43\") " Oct 08 20:25:03 crc kubenswrapper[4988]: I1008 20:25:03.431910 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/de67964d-9527-467b-8ca4-364cdec72a43-ovncontroller-config-0\") pod \"de67964d-9527-467b-8ca4-364cdec72a43\" (UID: \"de67964d-9527-467b-8ca4-364cdec72a43\") " Oct 08 20:25:03 crc kubenswrapper[4988]: I1008 20:25:03.431953 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bn29g\" (UniqueName: \"kubernetes.io/projected/de67964d-9527-467b-8ca4-364cdec72a43-kube-api-access-bn29g\") pod \"de67964d-9527-467b-8ca4-364cdec72a43\" (UID: \"de67964d-9527-467b-8ca4-364cdec72a43\") " Oct 08 20:25:03 crc kubenswrapper[4988]: I1008 20:25:03.432070 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/de67964d-9527-467b-8ca4-364cdec72a43-ssh-key\") pod \"de67964d-9527-467b-8ca4-364cdec72a43\" (UID: \"de67964d-9527-467b-8ca4-364cdec72a43\") " Oct 08 20:25:03 crc kubenswrapper[4988]: I1008 20:25:03.437337 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/de67964d-9527-467b-8ca4-364cdec72a43-ovn-combined-ca-bundle" (OuterVolumeSpecName: "ovn-combined-ca-bundle") pod "de67964d-9527-467b-8ca4-364cdec72a43" (UID: "de67964d-9527-467b-8ca4-364cdec72a43"). InnerVolumeSpecName "ovn-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 20:25:03 crc kubenswrapper[4988]: I1008 20:25:03.438885 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/de67964d-9527-467b-8ca4-364cdec72a43-kube-api-access-bn29g" (OuterVolumeSpecName: "kube-api-access-bn29g") pod "de67964d-9527-467b-8ca4-364cdec72a43" (UID: "de67964d-9527-467b-8ca4-364cdec72a43"). InnerVolumeSpecName "kube-api-access-bn29g". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 20:25:03 crc kubenswrapper[4988]: I1008 20:25:03.460467 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/de67964d-9527-467b-8ca4-364cdec72a43-ovncontroller-config-0" (OuterVolumeSpecName: "ovncontroller-config-0") pod "de67964d-9527-467b-8ca4-364cdec72a43" (UID: "de67964d-9527-467b-8ca4-364cdec72a43"). InnerVolumeSpecName "ovncontroller-config-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 20:25:03 crc kubenswrapper[4988]: I1008 20:25:03.467663 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/de67964d-9527-467b-8ca4-364cdec72a43-inventory" (OuterVolumeSpecName: "inventory") pod "de67964d-9527-467b-8ca4-364cdec72a43" (UID: "de67964d-9527-467b-8ca4-364cdec72a43"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 20:25:03 crc kubenswrapper[4988]: I1008 20:25:03.487927 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/de67964d-9527-467b-8ca4-364cdec72a43-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "de67964d-9527-467b-8ca4-364cdec72a43" (UID: "de67964d-9527-467b-8ca4-364cdec72a43"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 20:25:03 crc kubenswrapper[4988]: I1008 20:25:03.534705 4988 reconciler_common.go:293] "Volume detached for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/de67964d-9527-467b-8ca4-364cdec72a43-ovn-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 20:25:03 crc kubenswrapper[4988]: I1008 20:25:03.535247 4988 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/de67964d-9527-467b-8ca4-364cdec72a43-inventory\") on node \"crc\" DevicePath \"\"" Oct 08 20:25:03 crc kubenswrapper[4988]: I1008 20:25:03.535362 4988 reconciler_common.go:293] "Volume detached for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/de67964d-9527-467b-8ca4-364cdec72a43-ovncontroller-config-0\") on node \"crc\" DevicePath \"\"" Oct 08 20:25:03 crc kubenswrapper[4988]: I1008 20:25:03.535453 4988 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bn29g\" (UniqueName: \"kubernetes.io/projected/de67964d-9527-467b-8ca4-364cdec72a43-kube-api-access-bn29g\") on node \"crc\" DevicePath \"\"" Oct 08 20:25:03 crc kubenswrapper[4988]: I1008 20:25:03.535531 4988 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/de67964d-9527-467b-8ca4-364cdec72a43-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 08 20:25:03 crc kubenswrapper[4988]: I1008 20:25:03.766196 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-openstack-openstack-cell1-5h2h6" event={"ID":"de67964d-9527-467b-8ca4-364cdec72a43","Type":"ContainerDied","Data":"5dd310eab7f5e71305b60454536195367b908cf78370ed6f52e039b52262f377"} Oct 08 20:25:03 crc kubenswrapper[4988]: I1008 20:25:03.766462 4988 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5dd310eab7f5e71305b60454536195367b908cf78370ed6f52e039b52262f377" Oct 08 20:25:03 crc kubenswrapper[4988]: I1008 20:25:03.766344 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-openstack-openstack-cell1-5h2h6" Oct 08 20:25:03 crc kubenswrapper[4988]: I1008 20:25:03.871897 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-metadata-openstack-openstack-cell1-b5mrs"] Oct 08 20:25:03 crc kubenswrapper[4988]: E1008 20:25:03.872361 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="80bb8388-273a-405d-bffd-1ca1bb6d5999" containerName="extract-content" Oct 08 20:25:03 crc kubenswrapper[4988]: I1008 20:25:03.872377 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="80bb8388-273a-405d-bffd-1ca1bb6d5999" containerName="extract-content" Oct 08 20:25:03 crc kubenswrapper[4988]: E1008 20:25:03.872413 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ad57069f-f141-4bdc-85d9-2969822b24c7" containerName="extract-utilities" Oct 08 20:25:03 crc kubenswrapper[4988]: I1008 20:25:03.872424 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="ad57069f-f141-4bdc-85d9-2969822b24c7" containerName="extract-utilities" Oct 08 20:25:03 crc kubenswrapper[4988]: E1008 20:25:03.872438 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ad57069f-f141-4bdc-85d9-2969822b24c7" containerName="extract-content" Oct 08 20:25:03 crc kubenswrapper[4988]: I1008 20:25:03.872444 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="ad57069f-f141-4bdc-85d9-2969822b24c7" containerName="extract-content" Oct 08 20:25:03 crc kubenswrapper[4988]: E1008 20:25:03.872453 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="de67964d-9527-467b-8ca4-364cdec72a43" containerName="ovn-openstack-openstack-cell1" Oct 08 20:25:03 crc kubenswrapper[4988]: I1008 20:25:03.872458 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="de67964d-9527-467b-8ca4-364cdec72a43" containerName="ovn-openstack-openstack-cell1" Oct 08 20:25:03 crc kubenswrapper[4988]: E1008 20:25:03.872498 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ad57069f-f141-4bdc-85d9-2969822b24c7" containerName="registry-server" Oct 08 20:25:03 crc kubenswrapper[4988]: I1008 20:25:03.872504 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="ad57069f-f141-4bdc-85d9-2969822b24c7" containerName="registry-server" Oct 08 20:25:03 crc kubenswrapper[4988]: E1008 20:25:03.872519 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="80bb8388-273a-405d-bffd-1ca1bb6d5999" containerName="registry-server" Oct 08 20:25:03 crc kubenswrapper[4988]: I1008 20:25:03.872524 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="80bb8388-273a-405d-bffd-1ca1bb6d5999" containerName="registry-server" Oct 08 20:25:03 crc kubenswrapper[4988]: E1008 20:25:03.872536 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="80bb8388-273a-405d-bffd-1ca1bb6d5999" containerName="extract-utilities" Oct 08 20:25:03 crc kubenswrapper[4988]: I1008 20:25:03.872542 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="80bb8388-273a-405d-bffd-1ca1bb6d5999" containerName="extract-utilities" Oct 08 20:25:03 crc kubenswrapper[4988]: I1008 20:25:03.872746 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="de67964d-9527-467b-8ca4-364cdec72a43" containerName="ovn-openstack-openstack-cell1" Oct 08 20:25:03 crc kubenswrapper[4988]: I1008 20:25:03.872768 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="ad57069f-f141-4bdc-85d9-2969822b24c7" containerName="registry-server" Oct 08 20:25:03 crc kubenswrapper[4988]: I1008 20:25:03.872778 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="80bb8388-273a-405d-bffd-1ca1bb6d5999" containerName="registry-server" Oct 08 20:25:03 crc kubenswrapper[4988]: I1008 20:25:03.873605 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-openstack-openstack-cell1-b5mrs" Oct 08 20:25:03 crc kubenswrapper[4988]: I1008 20:25:03.876406 4988 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 08 20:25:03 crc kubenswrapper[4988]: I1008 20:25:03.876480 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-ovn-metadata-agent-neutron-config" Oct 08 20:25:03 crc kubenswrapper[4988]: I1008 20:25:03.876480 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-neutron-config" Oct 08 20:25:03 crc kubenswrapper[4988]: I1008 20:25:03.876523 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-4sdzr" Oct 08 20:25:03 crc kubenswrapper[4988]: I1008 20:25:03.876681 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Oct 08 20:25:03 crc kubenswrapper[4988]: I1008 20:25:03.876758 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Oct 08 20:25:03 crc kubenswrapper[4988]: I1008 20:25:03.882598 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-metadata-openstack-openstack-cell1-b5mrs"] Oct 08 20:25:03 crc kubenswrapper[4988]: I1008 20:25:03.946268 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dfca5e3a-fa8b-4a83-b6b8-f9694bbb8506-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-openstack-openstack-cell1-b5mrs\" (UID: \"dfca5e3a-fa8b-4a83-b6b8-f9694bbb8506\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-b5mrs" Oct 08 20:25:03 crc kubenswrapper[4988]: I1008 20:25:03.946330 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-54gb7\" (UniqueName: \"kubernetes.io/projected/dfca5e3a-fa8b-4a83-b6b8-f9694bbb8506-kube-api-access-54gb7\") pod \"neutron-metadata-openstack-openstack-cell1-b5mrs\" (UID: \"dfca5e3a-fa8b-4a83-b6b8-f9694bbb8506\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-b5mrs" Oct 08 20:25:03 crc kubenswrapper[4988]: I1008 20:25:03.946471 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/dfca5e3a-fa8b-4a83-b6b8-f9694bbb8506-ssh-key\") pod \"neutron-metadata-openstack-openstack-cell1-b5mrs\" (UID: \"dfca5e3a-fa8b-4a83-b6b8-f9694bbb8506\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-b5mrs" Oct 08 20:25:03 crc kubenswrapper[4988]: I1008 20:25:03.946511 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/dfca5e3a-fa8b-4a83-b6b8-f9694bbb8506-inventory\") pod \"neutron-metadata-openstack-openstack-cell1-b5mrs\" (UID: \"dfca5e3a-fa8b-4a83-b6b8-f9694bbb8506\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-b5mrs" Oct 08 20:25:03 crc kubenswrapper[4988]: I1008 20:25:03.946554 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/dfca5e3a-fa8b-4a83-b6b8-f9694bbb8506-nova-metadata-neutron-config-0\") pod \"neutron-metadata-openstack-openstack-cell1-b5mrs\" (UID: \"dfca5e3a-fa8b-4a83-b6b8-f9694bbb8506\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-b5mrs" Oct 08 20:25:03 crc kubenswrapper[4988]: I1008 20:25:03.946589 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/dfca5e3a-fa8b-4a83-b6b8-f9694bbb8506-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-openstack-openstack-cell1-b5mrs\" (UID: \"dfca5e3a-fa8b-4a83-b6b8-f9694bbb8506\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-b5mrs" Oct 08 20:25:04 crc kubenswrapper[4988]: I1008 20:25:04.048786 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/dfca5e3a-fa8b-4a83-b6b8-f9694bbb8506-ssh-key\") pod \"neutron-metadata-openstack-openstack-cell1-b5mrs\" (UID: \"dfca5e3a-fa8b-4a83-b6b8-f9694bbb8506\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-b5mrs" Oct 08 20:25:04 crc kubenswrapper[4988]: I1008 20:25:04.049011 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/dfca5e3a-fa8b-4a83-b6b8-f9694bbb8506-inventory\") pod \"neutron-metadata-openstack-openstack-cell1-b5mrs\" (UID: \"dfca5e3a-fa8b-4a83-b6b8-f9694bbb8506\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-b5mrs" Oct 08 20:25:04 crc kubenswrapper[4988]: I1008 20:25:04.049145 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/dfca5e3a-fa8b-4a83-b6b8-f9694bbb8506-nova-metadata-neutron-config-0\") pod \"neutron-metadata-openstack-openstack-cell1-b5mrs\" (UID: \"dfca5e3a-fa8b-4a83-b6b8-f9694bbb8506\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-b5mrs" Oct 08 20:25:04 crc kubenswrapper[4988]: I1008 20:25:04.049241 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/dfca5e3a-fa8b-4a83-b6b8-f9694bbb8506-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-openstack-openstack-cell1-b5mrs\" (UID: \"dfca5e3a-fa8b-4a83-b6b8-f9694bbb8506\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-b5mrs" Oct 08 20:25:04 crc kubenswrapper[4988]: I1008 20:25:04.049347 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dfca5e3a-fa8b-4a83-b6b8-f9694bbb8506-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-openstack-openstack-cell1-b5mrs\" (UID: \"dfca5e3a-fa8b-4a83-b6b8-f9694bbb8506\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-b5mrs" Oct 08 20:25:04 crc kubenswrapper[4988]: I1008 20:25:04.049445 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-54gb7\" (UniqueName: \"kubernetes.io/projected/dfca5e3a-fa8b-4a83-b6b8-f9694bbb8506-kube-api-access-54gb7\") pod \"neutron-metadata-openstack-openstack-cell1-b5mrs\" (UID: \"dfca5e3a-fa8b-4a83-b6b8-f9694bbb8506\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-b5mrs" Oct 08 20:25:04 crc kubenswrapper[4988]: I1008 20:25:04.053283 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/dfca5e3a-fa8b-4a83-b6b8-f9694bbb8506-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-openstack-openstack-cell1-b5mrs\" (UID: \"dfca5e3a-fa8b-4a83-b6b8-f9694bbb8506\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-b5mrs" Oct 08 20:25:04 crc kubenswrapper[4988]: I1008 20:25:04.055114 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/dfca5e3a-fa8b-4a83-b6b8-f9694bbb8506-ssh-key\") pod \"neutron-metadata-openstack-openstack-cell1-b5mrs\" (UID: \"dfca5e3a-fa8b-4a83-b6b8-f9694bbb8506\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-b5mrs" Oct 08 20:25:04 crc kubenswrapper[4988]: I1008 20:25:04.056833 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/dfca5e3a-fa8b-4a83-b6b8-f9694bbb8506-inventory\") pod \"neutron-metadata-openstack-openstack-cell1-b5mrs\" (UID: \"dfca5e3a-fa8b-4a83-b6b8-f9694bbb8506\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-b5mrs" Oct 08 20:25:04 crc kubenswrapper[4988]: I1008 20:25:04.058224 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/dfca5e3a-fa8b-4a83-b6b8-f9694bbb8506-nova-metadata-neutron-config-0\") pod \"neutron-metadata-openstack-openstack-cell1-b5mrs\" (UID: \"dfca5e3a-fa8b-4a83-b6b8-f9694bbb8506\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-b5mrs" Oct 08 20:25:04 crc kubenswrapper[4988]: I1008 20:25:04.062603 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dfca5e3a-fa8b-4a83-b6b8-f9694bbb8506-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-openstack-openstack-cell1-b5mrs\" (UID: \"dfca5e3a-fa8b-4a83-b6b8-f9694bbb8506\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-b5mrs" Oct 08 20:25:04 crc kubenswrapper[4988]: I1008 20:25:04.080689 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-54gb7\" (UniqueName: \"kubernetes.io/projected/dfca5e3a-fa8b-4a83-b6b8-f9694bbb8506-kube-api-access-54gb7\") pod \"neutron-metadata-openstack-openstack-cell1-b5mrs\" (UID: \"dfca5e3a-fa8b-4a83-b6b8-f9694bbb8506\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-b5mrs" Oct 08 20:25:04 crc kubenswrapper[4988]: I1008 20:25:04.196290 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-openstack-openstack-cell1-b5mrs" Oct 08 20:25:04 crc kubenswrapper[4988]: I1008 20:25:04.811244 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-metadata-openstack-openstack-cell1-b5mrs"] Oct 08 20:25:04 crc kubenswrapper[4988]: W1008 20:25:04.821609 4988 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poddfca5e3a_fa8b_4a83_b6b8_f9694bbb8506.slice/crio-ad620bef824cc7adec285c467115212ecbbaff33d9f3e41fc6e747cbd20c652e WatchSource:0}: Error finding container ad620bef824cc7adec285c467115212ecbbaff33d9f3e41fc6e747cbd20c652e: Status 404 returned error can't find the container with id ad620bef824cc7adec285c467115212ecbbaff33d9f3e41fc6e747cbd20c652e Oct 08 20:25:04 crc kubenswrapper[4988]: I1008 20:25:04.823995 4988 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 08 20:25:05 crc kubenswrapper[4988]: I1008 20:25:05.786626 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-openstack-openstack-cell1-b5mrs" event={"ID":"dfca5e3a-fa8b-4a83-b6b8-f9694bbb8506","Type":"ContainerStarted","Data":"fbcaa5fd5bf3d399eb397156bb58d0d82b287ed0775d1ab688ec6a606e578022"} Oct 08 20:25:05 crc kubenswrapper[4988]: I1008 20:25:05.786916 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-openstack-openstack-cell1-b5mrs" event={"ID":"dfca5e3a-fa8b-4a83-b6b8-f9694bbb8506","Type":"ContainerStarted","Data":"ad620bef824cc7adec285c467115212ecbbaff33d9f3e41fc6e747cbd20c652e"} Oct 08 20:25:05 crc kubenswrapper[4988]: I1008 20:25:05.810604 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-metadata-openstack-openstack-cell1-b5mrs" podStartSLOduration=2.211047754 podStartE2EDuration="2.810575765s" podCreationTimestamp="2025-10-08 20:25:03 +0000 UTC" firstStartedPulling="2025-10-08 20:25:04.823653144 +0000 UTC m=+8050.273495904" lastFinishedPulling="2025-10-08 20:25:05.423181135 +0000 UTC m=+8050.873023915" observedRunningTime="2025-10-08 20:25:05.805639518 +0000 UTC m=+8051.255482308" watchObservedRunningTime="2025-10-08 20:25:05.810575765 +0000 UTC m=+8051.260418575" Oct 08 20:25:14 crc kubenswrapper[4988]: I1008 20:25:14.238954 4988 scope.go:117] "RemoveContainer" containerID="7541195e86b98682ea101ebff2b19f543aa9bab45dd1665e8f710922ad18c5e7" Oct 08 20:25:14 crc kubenswrapper[4988]: E1008 20:25:14.240134 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nm54f_openshift-machine-config-operator(f328a20b-5283-407e-8a93-2ab2da599f18)\"" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" Oct 08 20:25:25 crc kubenswrapper[4988]: I1008 20:25:25.254701 4988 scope.go:117] "RemoveContainer" containerID="7541195e86b98682ea101ebff2b19f543aa9bab45dd1665e8f710922ad18c5e7" Oct 08 20:25:25 crc kubenswrapper[4988]: E1008 20:25:25.255736 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nm54f_openshift-machine-config-operator(f328a20b-5283-407e-8a93-2ab2da599f18)\"" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" Oct 08 20:25:40 crc kubenswrapper[4988]: I1008 20:25:40.238989 4988 scope.go:117] "RemoveContainer" containerID="7541195e86b98682ea101ebff2b19f543aa9bab45dd1665e8f710922ad18c5e7" Oct 08 20:25:40 crc kubenswrapper[4988]: E1008 20:25:40.239881 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nm54f_openshift-machine-config-operator(f328a20b-5283-407e-8a93-2ab2da599f18)\"" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" Oct 08 20:25:54 crc kubenswrapper[4988]: I1008 20:25:54.237737 4988 scope.go:117] "RemoveContainer" containerID="7541195e86b98682ea101ebff2b19f543aa9bab45dd1665e8f710922ad18c5e7" Oct 08 20:25:55 crc kubenswrapper[4988]: I1008 20:25:55.396761 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" event={"ID":"f328a20b-5283-407e-8a93-2ab2da599f18","Type":"ContainerStarted","Data":"7cbd759a42c747ea6a2391dd004be2e8c08240e1bbe09cf06de3384164b67bb4"} Oct 08 20:26:01 crc kubenswrapper[4988]: I1008 20:26:01.471831 4988 generic.go:334] "Generic (PLEG): container finished" podID="dfca5e3a-fa8b-4a83-b6b8-f9694bbb8506" containerID="fbcaa5fd5bf3d399eb397156bb58d0d82b287ed0775d1ab688ec6a606e578022" exitCode=0 Oct 08 20:26:01 crc kubenswrapper[4988]: I1008 20:26:01.471929 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-openstack-openstack-cell1-b5mrs" event={"ID":"dfca5e3a-fa8b-4a83-b6b8-f9694bbb8506","Type":"ContainerDied","Data":"fbcaa5fd5bf3d399eb397156bb58d0d82b287ed0775d1ab688ec6a606e578022"} Oct 08 20:26:03 crc kubenswrapper[4988]: I1008 20:26:03.033128 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-openstack-openstack-cell1-b5mrs" Oct 08 20:26:03 crc kubenswrapper[4988]: I1008 20:26:03.071116 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-54gb7\" (UniqueName: \"kubernetes.io/projected/dfca5e3a-fa8b-4a83-b6b8-f9694bbb8506-kube-api-access-54gb7\") pod \"dfca5e3a-fa8b-4a83-b6b8-f9694bbb8506\" (UID: \"dfca5e3a-fa8b-4a83-b6b8-f9694bbb8506\") " Oct 08 20:26:03 crc kubenswrapper[4988]: I1008 20:26:03.071206 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dfca5e3a-fa8b-4a83-b6b8-f9694bbb8506-neutron-metadata-combined-ca-bundle\") pod \"dfca5e3a-fa8b-4a83-b6b8-f9694bbb8506\" (UID: \"dfca5e3a-fa8b-4a83-b6b8-f9694bbb8506\") " Oct 08 20:26:03 crc kubenswrapper[4988]: I1008 20:26:03.071230 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/dfca5e3a-fa8b-4a83-b6b8-f9694bbb8506-ssh-key\") pod \"dfca5e3a-fa8b-4a83-b6b8-f9694bbb8506\" (UID: \"dfca5e3a-fa8b-4a83-b6b8-f9694bbb8506\") " Oct 08 20:26:03 crc kubenswrapper[4988]: I1008 20:26:03.072235 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/dfca5e3a-fa8b-4a83-b6b8-f9694bbb8506-inventory\") pod \"dfca5e3a-fa8b-4a83-b6b8-f9694bbb8506\" (UID: \"dfca5e3a-fa8b-4a83-b6b8-f9694bbb8506\") " Oct 08 20:26:03 crc kubenswrapper[4988]: I1008 20:26:03.072270 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/dfca5e3a-fa8b-4a83-b6b8-f9694bbb8506-nova-metadata-neutron-config-0\") pod \"dfca5e3a-fa8b-4a83-b6b8-f9694bbb8506\" (UID: \"dfca5e3a-fa8b-4a83-b6b8-f9694bbb8506\") " Oct 08 20:26:03 crc kubenswrapper[4988]: I1008 20:26:03.072326 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/dfca5e3a-fa8b-4a83-b6b8-f9694bbb8506-neutron-ovn-metadata-agent-neutron-config-0\") pod \"dfca5e3a-fa8b-4a83-b6b8-f9694bbb8506\" (UID: \"dfca5e3a-fa8b-4a83-b6b8-f9694bbb8506\") " Oct 08 20:26:03 crc kubenswrapper[4988]: I1008 20:26:03.095717 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dfca5e3a-fa8b-4a83-b6b8-f9694bbb8506-kube-api-access-54gb7" (OuterVolumeSpecName: "kube-api-access-54gb7") pod "dfca5e3a-fa8b-4a83-b6b8-f9694bbb8506" (UID: "dfca5e3a-fa8b-4a83-b6b8-f9694bbb8506"). InnerVolumeSpecName "kube-api-access-54gb7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 20:26:03 crc kubenswrapper[4988]: I1008 20:26:03.095854 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dfca5e3a-fa8b-4a83-b6b8-f9694bbb8506-neutron-metadata-combined-ca-bundle" (OuterVolumeSpecName: "neutron-metadata-combined-ca-bundle") pod "dfca5e3a-fa8b-4a83-b6b8-f9694bbb8506" (UID: "dfca5e3a-fa8b-4a83-b6b8-f9694bbb8506"). InnerVolumeSpecName "neutron-metadata-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 20:26:03 crc kubenswrapper[4988]: I1008 20:26:03.115104 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dfca5e3a-fa8b-4a83-b6b8-f9694bbb8506-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "dfca5e3a-fa8b-4a83-b6b8-f9694bbb8506" (UID: "dfca5e3a-fa8b-4a83-b6b8-f9694bbb8506"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 20:26:03 crc kubenswrapper[4988]: I1008 20:26:03.130161 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dfca5e3a-fa8b-4a83-b6b8-f9694bbb8506-nova-metadata-neutron-config-0" (OuterVolumeSpecName: "nova-metadata-neutron-config-0") pod "dfca5e3a-fa8b-4a83-b6b8-f9694bbb8506" (UID: "dfca5e3a-fa8b-4a83-b6b8-f9694bbb8506"). InnerVolumeSpecName "nova-metadata-neutron-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 20:26:03 crc kubenswrapper[4988]: I1008 20:26:03.135606 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dfca5e3a-fa8b-4a83-b6b8-f9694bbb8506-neutron-ovn-metadata-agent-neutron-config-0" (OuterVolumeSpecName: "neutron-ovn-metadata-agent-neutron-config-0") pod "dfca5e3a-fa8b-4a83-b6b8-f9694bbb8506" (UID: "dfca5e3a-fa8b-4a83-b6b8-f9694bbb8506"). InnerVolumeSpecName "neutron-ovn-metadata-agent-neutron-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 20:26:03 crc kubenswrapper[4988]: I1008 20:26:03.150153 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dfca5e3a-fa8b-4a83-b6b8-f9694bbb8506-inventory" (OuterVolumeSpecName: "inventory") pod "dfca5e3a-fa8b-4a83-b6b8-f9694bbb8506" (UID: "dfca5e3a-fa8b-4a83-b6b8-f9694bbb8506"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 20:26:03 crc kubenswrapper[4988]: I1008 20:26:03.173661 4988 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/dfca5e3a-fa8b-4a83-b6b8-f9694bbb8506-inventory\") on node \"crc\" DevicePath \"\"" Oct 08 20:26:03 crc kubenswrapper[4988]: I1008 20:26:03.173813 4988 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/dfca5e3a-fa8b-4a83-b6b8-f9694bbb8506-nova-metadata-neutron-config-0\") on node \"crc\" DevicePath \"\"" Oct 08 20:26:03 crc kubenswrapper[4988]: I1008 20:26:03.173894 4988 reconciler_common.go:293] "Volume detached for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/dfca5e3a-fa8b-4a83-b6b8-f9694bbb8506-neutron-ovn-metadata-agent-neutron-config-0\") on node \"crc\" DevicePath \"\"" Oct 08 20:26:03 crc kubenswrapper[4988]: I1008 20:26:03.173969 4988 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-54gb7\" (UniqueName: \"kubernetes.io/projected/dfca5e3a-fa8b-4a83-b6b8-f9694bbb8506-kube-api-access-54gb7\") on node \"crc\" DevicePath \"\"" Oct 08 20:26:03 crc kubenswrapper[4988]: I1008 20:26:03.174040 4988 reconciler_common.go:293] "Volume detached for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dfca5e3a-fa8b-4a83-b6b8-f9694bbb8506-neutron-metadata-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 20:26:03 crc kubenswrapper[4988]: I1008 20:26:03.174111 4988 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/dfca5e3a-fa8b-4a83-b6b8-f9694bbb8506-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 08 20:26:03 crc kubenswrapper[4988]: I1008 20:26:03.499641 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-openstack-openstack-cell1-b5mrs" event={"ID":"dfca5e3a-fa8b-4a83-b6b8-f9694bbb8506","Type":"ContainerDied","Data":"ad620bef824cc7adec285c467115212ecbbaff33d9f3e41fc6e747cbd20c652e"} Oct 08 20:26:03 crc kubenswrapper[4988]: I1008 20:26:03.500198 4988 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ad620bef824cc7adec285c467115212ecbbaff33d9f3e41fc6e747cbd20c652e" Oct 08 20:26:03 crc kubenswrapper[4988]: I1008 20:26:03.499747 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-openstack-openstack-cell1-b5mrs" Oct 08 20:26:03 crc kubenswrapper[4988]: I1008 20:26:03.590068 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/libvirt-openstack-openstack-cell1-mzn6v"] Oct 08 20:26:03 crc kubenswrapper[4988]: E1008 20:26:03.590564 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dfca5e3a-fa8b-4a83-b6b8-f9694bbb8506" containerName="neutron-metadata-openstack-openstack-cell1" Oct 08 20:26:03 crc kubenswrapper[4988]: I1008 20:26:03.590587 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="dfca5e3a-fa8b-4a83-b6b8-f9694bbb8506" containerName="neutron-metadata-openstack-openstack-cell1" Oct 08 20:26:03 crc kubenswrapper[4988]: I1008 20:26:03.590861 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="dfca5e3a-fa8b-4a83-b6b8-f9694bbb8506" containerName="neutron-metadata-openstack-openstack-cell1" Oct 08 20:26:03 crc kubenswrapper[4988]: I1008 20:26:03.591623 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-openstack-openstack-cell1-mzn6v" Oct 08 20:26:03 crc kubenswrapper[4988]: I1008 20:26:03.595848 4988 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 08 20:26:03 crc kubenswrapper[4988]: I1008 20:26:03.596425 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Oct 08 20:26:03 crc kubenswrapper[4988]: I1008 20:26:03.596648 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"libvirt-secret" Oct 08 20:26:03 crc kubenswrapper[4988]: I1008 20:26:03.597462 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-4sdzr" Oct 08 20:26:03 crc kubenswrapper[4988]: I1008 20:26:03.597833 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Oct 08 20:26:03 crc kubenswrapper[4988]: I1008 20:26:03.605885 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/libvirt-openstack-openstack-cell1-mzn6v"] Oct 08 20:26:03 crc kubenswrapper[4988]: I1008 20:26:03.686952 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/aef3b7a1-6628-4895-a380-153a230bd0a3-inventory\") pod \"libvirt-openstack-openstack-cell1-mzn6v\" (UID: \"aef3b7a1-6628-4895-a380-153a230bd0a3\") " pod="openstack/libvirt-openstack-openstack-cell1-mzn6v" Oct 08 20:26:03 crc kubenswrapper[4988]: I1008 20:26:03.687032 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cmb68\" (UniqueName: \"kubernetes.io/projected/aef3b7a1-6628-4895-a380-153a230bd0a3-kube-api-access-cmb68\") pod \"libvirt-openstack-openstack-cell1-mzn6v\" (UID: \"aef3b7a1-6628-4895-a380-153a230bd0a3\") " pod="openstack/libvirt-openstack-openstack-cell1-mzn6v" Oct 08 20:26:03 crc kubenswrapper[4988]: I1008 20:26:03.687075 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/aef3b7a1-6628-4895-a380-153a230bd0a3-ssh-key\") pod \"libvirt-openstack-openstack-cell1-mzn6v\" (UID: \"aef3b7a1-6628-4895-a380-153a230bd0a3\") " pod="openstack/libvirt-openstack-openstack-cell1-mzn6v" Oct 08 20:26:03 crc kubenswrapper[4988]: I1008 20:26:03.687238 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/aef3b7a1-6628-4895-a380-153a230bd0a3-libvirt-secret-0\") pod \"libvirt-openstack-openstack-cell1-mzn6v\" (UID: \"aef3b7a1-6628-4895-a380-153a230bd0a3\") " pod="openstack/libvirt-openstack-openstack-cell1-mzn6v" Oct 08 20:26:03 crc kubenswrapper[4988]: I1008 20:26:03.687295 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aef3b7a1-6628-4895-a380-153a230bd0a3-libvirt-combined-ca-bundle\") pod \"libvirt-openstack-openstack-cell1-mzn6v\" (UID: \"aef3b7a1-6628-4895-a380-153a230bd0a3\") " pod="openstack/libvirt-openstack-openstack-cell1-mzn6v" Oct 08 20:26:03 crc kubenswrapper[4988]: I1008 20:26:03.789019 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/aef3b7a1-6628-4895-a380-153a230bd0a3-inventory\") pod \"libvirt-openstack-openstack-cell1-mzn6v\" (UID: \"aef3b7a1-6628-4895-a380-153a230bd0a3\") " pod="openstack/libvirt-openstack-openstack-cell1-mzn6v" Oct 08 20:26:03 crc kubenswrapper[4988]: I1008 20:26:03.789086 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cmb68\" (UniqueName: \"kubernetes.io/projected/aef3b7a1-6628-4895-a380-153a230bd0a3-kube-api-access-cmb68\") pod \"libvirt-openstack-openstack-cell1-mzn6v\" (UID: \"aef3b7a1-6628-4895-a380-153a230bd0a3\") " pod="openstack/libvirt-openstack-openstack-cell1-mzn6v" Oct 08 20:26:03 crc kubenswrapper[4988]: I1008 20:26:03.789138 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/aef3b7a1-6628-4895-a380-153a230bd0a3-ssh-key\") pod \"libvirt-openstack-openstack-cell1-mzn6v\" (UID: \"aef3b7a1-6628-4895-a380-153a230bd0a3\") " pod="openstack/libvirt-openstack-openstack-cell1-mzn6v" Oct 08 20:26:03 crc kubenswrapper[4988]: I1008 20:26:03.789263 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/aef3b7a1-6628-4895-a380-153a230bd0a3-libvirt-secret-0\") pod \"libvirt-openstack-openstack-cell1-mzn6v\" (UID: \"aef3b7a1-6628-4895-a380-153a230bd0a3\") " pod="openstack/libvirt-openstack-openstack-cell1-mzn6v" Oct 08 20:26:03 crc kubenswrapper[4988]: I1008 20:26:03.789298 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aef3b7a1-6628-4895-a380-153a230bd0a3-libvirt-combined-ca-bundle\") pod \"libvirt-openstack-openstack-cell1-mzn6v\" (UID: \"aef3b7a1-6628-4895-a380-153a230bd0a3\") " pod="openstack/libvirt-openstack-openstack-cell1-mzn6v" Oct 08 20:26:03 crc kubenswrapper[4988]: I1008 20:26:03.792961 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/aef3b7a1-6628-4895-a380-153a230bd0a3-inventory\") pod \"libvirt-openstack-openstack-cell1-mzn6v\" (UID: \"aef3b7a1-6628-4895-a380-153a230bd0a3\") " pod="openstack/libvirt-openstack-openstack-cell1-mzn6v" Oct 08 20:26:03 crc kubenswrapper[4988]: I1008 20:26:03.792997 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/aef3b7a1-6628-4895-a380-153a230bd0a3-libvirt-secret-0\") pod \"libvirt-openstack-openstack-cell1-mzn6v\" (UID: \"aef3b7a1-6628-4895-a380-153a230bd0a3\") " pod="openstack/libvirt-openstack-openstack-cell1-mzn6v" Oct 08 20:26:03 crc kubenswrapper[4988]: I1008 20:26:03.793350 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/aef3b7a1-6628-4895-a380-153a230bd0a3-ssh-key\") pod \"libvirt-openstack-openstack-cell1-mzn6v\" (UID: \"aef3b7a1-6628-4895-a380-153a230bd0a3\") " pod="openstack/libvirt-openstack-openstack-cell1-mzn6v" Oct 08 20:26:03 crc kubenswrapper[4988]: I1008 20:26:03.794977 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aef3b7a1-6628-4895-a380-153a230bd0a3-libvirt-combined-ca-bundle\") pod \"libvirt-openstack-openstack-cell1-mzn6v\" (UID: \"aef3b7a1-6628-4895-a380-153a230bd0a3\") " pod="openstack/libvirt-openstack-openstack-cell1-mzn6v" Oct 08 20:26:03 crc kubenswrapper[4988]: I1008 20:26:03.811529 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cmb68\" (UniqueName: \"kubernetes.io/projected/aef3b7a1-6628-4895-a380-153a230bd0a3-kube-api-access-cmb68\") pod \"libvirt-openstack-openstack-cell1-mzn6v\" (UID: \"aef3b7a1-6628-4895-a380-153a230bd0a3\") " pod="openstack/libvirt-openstack-openstack-cell1-mzn6v" Oct 08 20:26:03 crc kubenswrapper[4988]: I1008 20:26:03.931329 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-openstack-openstack-cell1-mzn6v" Oct 08 20:26:04 crc kubenswrapper[4988]: I1008 20:26:04.503374 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/libvirt-openstack-openstack-cell1-mzn6v"] Oct 08 20:26:04 crc kubenswrapper[4988]: W1008 20:26:04.512355 4988 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podaef3b7a1_6628_4895_a380_153a230bd0a3.slice/crio-657950ad28784152be6e4e5e6b331f8e7d80ea51ec12cb9901b686cb68cbc47b WatchSource:0}: Error finding container 657950ad28784152be6e4e5e6b331f8e7d80ea51ec12cb9901b686cb68cbc47b: Status 404 returned error can't find the container with id 657950ad28784152be6e4e5e6b331f8e7d80ea51ec12cb9901b686cb68cbc47b Oct 08 20:26:05 crc kubenswrapper[4988]: I1008 20:26:05.523204 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-openstack-openstack-cell1-mzn6v" event={"ID":"aef3b7a1-6628-4895-a380-153a230bd0a3","Type":"ContainerStarted","Data":"affd5206fa13d0036c43bfaaf072688f3755710f77bc251632479521b90a3e46"} Oct 08 20:26:05 crc kubenswrapper[4988]: I1008 20:26:05.523708 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-openstack-openstack-cell1-mzn6v" event={"ID":"aef3b7a1-6628-4895-a380-153a230bd0a3","Type":"ContainerStarted","Data":"657950ad28784152be6e4e5e6b331f8e7d80ea51ec12cb9901b686cb68cbc47b"} Oct 08 20:26:05 crc kubenswrapper[4988]: I1008 20:26:05.562220 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/libvirt-openstack-openstack-cell1-mzn6v" podStartSLOduration=2.121203288 podStartE2EDuration="2.562199883s" podCreationTimestamp="2025-10-08 20:26:03 +0000 UTC" firstStartedPulling="2025-10-08 20:26:04.514995874 +0000 UTC m=+8109.964838654" lastFinishedPulling="2025-10-08 20:26:04.955992459 +0000 UTC m=+8110.405835249" observedRunningTime="2025-10-08 20:26:05.553740964 +0000 UTC m=+8111.003583754" watchObservedRunningTime="2025-10-08 20:26:05.562199883 +0000 UTC m=+8111.012042653" Oct 08 20:28:23 crc kubenswrapper[4988]: I1008 20:28:23.338111 4988 patch_prober.go:28] interesting pod/machine-config-daemon-nm54f container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 20:28:23 crc kubenswrapper[4988]: I1008 20:28:23.338901 4988 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 20:28:53 crc kubenswrapper[4988]: I1008 20:28:53.337791 4988 patch_prober.go:28] interesting pod/machine-config-daemon-nm54f container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 20:28:53 crc kubenswrapper[4988]: I1008 20:28:53.338440 4988 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 20:29:23 crc kubenswrapper[4988]: I1008 20:29:23.337800 4988 patch_prober.go:28] interesting pod/machine-config-daemon-nm54f container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 20:29:23 crc kubenswrapper[4988]: I1008 20:29:23.338338 4988 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 20:29:23 crc kubenswrapper[4988]: I1008 20:29:23.338412 4988 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" Oct 08 20:29:23 crc kubenswrapper[4988]: I1008 20:29:23.339344 4988 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"7cbd759a42c747ea6a2391dd004be2e8c08240e1bbe09cf06de3384164b67bb4"} pod="openshift-machine-config-operator/machine-config-daemon-nm54f" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 08 20:29:23 crc kubenswrapper[4988]: I1008 20:29:23.339426 4988 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" containerName="machine-config-daemon" containerID="cri-o://7cbd759a42c747ea6a2391dd004be2e8c08240e1bbe09cf06de3384164b67bb4" gracePeriod=600 Oct 08 20:29:23 crc kubenswrapper[4988]: I1008 20:29:23.999186 4988 generic.go:334] "Generic (PLEG): container finished" podID="f328a20b-5283-407e-8a93-2ab2da599f18" containerID="7cbd759a42c747ea6a2391dd004be2e8c08240e1bbe09cf06de3384164b67bb4" exitCode=0 Oct 08 20:29:23 crc kubenswrapper[4988]: I1008 20:29:23.999231 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" event={"ID":"f328a20b-5283-407e-8a93-2ab2da599f18","Type":"ContainerDied","Data":"7cbd759a42c747ea6a2391dd004be2e8c08240e1bbe09cf06de3384164b67bb4"} Oct 08 20:29:23 crc kubenswrapper[4988]: I1008 20:29:23.999761 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" event={"ID":"f328a20b-5283-407e-8a93-2ab2da599f18","Type":"ContainerStarted","Data":"413fd9f07623b7fb95cefbc48df19e6c85a31a0236c333177ea0591f24a99a71"} Oct 08 20:29:23 crc kubenswrapper[4988]: I1008 20:29:23.999803 4988 scope.go:117] "RemoveContainer" containerID="7541195e86b98682ea101ebff2b19f543aa9bab45dd1665e8f710922ad18c5e7" Oct 08 20:30:00 crc kubenswrapper[4988]: I1008 20:30:00.173447 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29332590-6fvwv"] Oct 08 20:30:00 crc kubenswrapper[4988]: I1008 20:30:00.186588 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29332590-6fvwv" Oct 08 20:30:00 crc kubenswrapper[4988]: I1008 20:30:00.191634 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29332590-6fvwv"] Oct 08 20:30:00 crc kubenswrapper[4988]: I1008 20:30:00.192201 4988 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 08 20:30:00 crc kubenswrapper[4988]: I1008 20:30:00.192522 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 08 20:30:00 crc kubenswrapper[4988]: I1008 20:30:00.365347 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/9bb32c8b-83c8-4be6-93c0-3dd925d46665-secret-volume\") pod \"collect-profiles-29332590-6fvwv\" (UID: \"9bb32c8b-83c8-4be6-93c0-3dd925d46665\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29332590-6fvwv" Oct 08 20:30:00 crc kubenswrapper[4988]: I1008 20:30:00.365570 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cd6zc\" (UniqueName: \"kubernetes.io/projected/9bb32c8b-83c8-4be6-93c0-3dd925d46665-kube-api-access-cd6zc\") pod \"collect-profiles-29332590-6fvwv\" (UID: \"9bb32c8b-83c8-4be6-93c0-3dd925d46665\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29332590-6fvwv" Oct 08 20:30:00 crc kubenswrapper[4988]: I1008 20:30:00.365680 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/9bb32c8b-83c8-4be6-93c0-3dd925d46665-config-volume\") pod \"collect-profiles-29332590-6fvwv\" (UID: \"9bb32c8b-83c8-4be6-93c0-3dd925d46665\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29332590-6fvwv" Oct 08 20:30:00 crc kubenswrapper[4988]: I1008 20:30:00.468168 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/9bb32c8b-83c8-4be6-93c0-3dd925d46665-config-volume\") pod \"collect-profiles-29332590-6fvwv\" (UID: \"9bb32c8b-83c8-4be6-93c0-3dd925d46665\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29332590-6fvwv" Oct 08 20:30:00 crc kubenswrapper[4988]: I1008 20:30:00.468330 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/9bb32c8b-83c8-4be6-93c0-3dd925d46665-secret-volume\") pod \"collect-profiles-29332590-6fvwv\" (UID: \"9bb32c8b-83c8-4be6-93c0-3dd925d46665\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29332590-6fvwv" Oct 08 20:30:00 crc kubenswrapper[4988]: I1008 20:30:00.469208 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/9bb32c8b-83c8-4be6-93c0-3dd925d46665-config-volume\") pod \"collect-profiles-29332590-6fvwv\" (UID: \"9bb32c8b-83c8-4be6-93c0-3dd925d46665\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29332590-6fvwv" Oct 08 20:30:00 crc kubenswrapper[4988]: I1008 20:30:00.469477 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cd6zc\" (UniqueName: \"kubernetes.io/projected/9bb32c8b-83c8-4be6-93c0-3dd925d46665-kube-api-access-cd6zc\") pod \"collect-profiles-29332590-6fvwv\" (UID: \"9bb32c8b-83c8-4be6-93c0-3dd925d46665\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29332590-6fvwv" Oct 08 20:30:00 crc kubenswrapper[4988]: I1008 20:30:00.480837 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/9bb32c8b-83c8-4be6-93c0-3dd925d46665-secret-volume\") pod \"collect-profiles-29332590-6fvwv\" (UID: \"9bb32c8b-83c8-4be6-93c0-3dd925d46665\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29332590-6fvwv" Oct 08 20:30:00 crc kubenswrapper[4988]: I1008 20:30:00.497517 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cd6zc\" (UniqueName: \"kubernetes.io/projected/9bb32c8b-83c8-4be6-93c0-3dd925d46665-kube-api-access-cd6zc\") pod \"collect-profiles-29332590-6fvwv\" (UID: \"9bb32c8b-83c8-4be6-93c0-3dd925d46665\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29332590-6fvwv" Oct 08 20:30:00 crc kubenswrapper[4988]: I1008 20:30:00.516663 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29332590-6fvwv" Oct 08 20:30:01 crc kubenswrapper[4988]: I1008 20:30:01.064455 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29332590-6fvwv"] Oct 08 20:30:01 crc kubenswrapper[4988]: I1008 20:30:01.458701 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29332590-6fvwv" event={"ID":"9bb32c8b-83c8-4be6-93c0-3dd925d46665","Type":"ContainerStarted","Data":"a7b3e1a2096824631fa9b0c09961d35e34a8b154a743f4e9a0f923d0f1cfd9f5"} Oct 08 20:30:01 crc kubenswrapper[4988]: I1008 20:30:01.459053 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29332590-6fvwv" event={"ID":"9bb32c8b-83c8-4be6-93c0-3dd925d46665","Type":"ContainerStarted","Data":"e5ea41d0935b6f29f789c16296be03317c867e444a05b8707d3f7e30cc617773"} Oct 08 20:30:01 crc kubenswrapper[4988]: I1008 20:30:01.481200 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29332590-6fvwv" podStartSLOduration=1.481177673 podStartE2EDuration="1.481177673s" podCreationTimestamp="2025-10-08 20:30:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 20:30:01.477313572 +0000 UTC m=+8346.927156342" watchObservedRunningTime="2025-10-08 20:30:01.481177673 +0000 UTC m=+8346.931020463" Oct 08 20:30:02 crc kubenswrapper[4988]: I1008 20:30:02.470736 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29332590-6fvwv" event={"ID":"9bb32c8b-83c8-4be6-93c0-3dd925d46665","Type":"ContainerDied","Data":"a7b3e1a2096824631fa9b0c09961d35e34a8b154a743f4e9a0f923d0f1cfd9f5"} Oct 08 20:30:02 crc kubenswrapper[4988]: I1008 20:30:02.470595 4988 generic.go:334] "Generic (PLEG): container finished" podID="9bb32c8b-83c8-4be6-93c0-3dd925d46665" containerID="a7b3e1a2096824631fa9b0c09961d35e34a8b154a743f4e9a0f923d0f1cfd9f5" exitCode=0 Oct 08 20:30:03 crc kubenswrapper[4988]: I1008 20:30:03.948242 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29332590-6fvwv" Oct 08 20:30:04 crc kubenswrapper[4988]: I1008 20:30:04.054460 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/9bb32c8b-83c8-4be6-93c0-3dd925d46665-config-volume\") pod \"9bb32c8b-83c8-4be6-93c0-3dd925d46665\" (UID: \"9bb32c8b-83c8-4be6-93c0-3dd925d46665\") " Oct 08 20:30:04 crc kubenswrapper[4988]: I1008 20:30:04.054587 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cd6zc\" (UniqueName: \"kubernetes.io/projected/9bb32c8b-83c8-4be6-93c0-3dd925d46665-kube-api-access-cd6zc\") pod \"9bb32c8b-83c8-4be6-93c0-3dd925d46665\" (UID: \"9bb32c8b-83c8-4be6-93c0-3dd925d46665\") " Oct 08 20:30:04 crc kubenswrapper[4988]: I1008 20:30:04.055228 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9bb32c8b-83c8-4be6-93c0-3dd925d46665-config-volume" (OuterVolumeSpecName: "config-volume") pod "9bb32c8b-83c8-4be6-93c0-3dd925d46665" (UID: "9bb32c8b-83c8-4be6-93c0-3dd925d46665"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 20:30:04 crc kubenswrapper[4988]: I1008 20:30:04.056083 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/9bb32c8b-83c8-4be6-93c0-3dd925d46665-secret-volume\") pod \"9bb32c8b-83c8-4be6-93c0-3dd925d46665\" (UID: \"9bb32c8b-83c8-4be6-93c0-3dd925d46665\") " Oct 08 20:30:04 crc kubenswrapper[4988]: I1008 20:30:04.056862 4988 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/9bb32c8b-83c8-4be6-93c0-3dd925d46665-config-volume\") on node \"crc\" DevicePath \"\"" Oct 08 20:30:04 crc kubenswrapper[4988]: I1008 20:30:04.061503 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9bb32c8b-83c8-4be6-93c0-3dd925d46665-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "9bb32c8b-83c8-4be6-93c0-3dd925d46665" (UID: "9bb32c8b-83c8-4be6-93c0-3dd925d46665"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 20:30:04 crc kubenswrapper[4988]: I1008 20:30:04.077148 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9bb32c8b-83c8-4be6-93c0-3dd925d46665-kube-api-access-cd6zc" (OuterVolumeSpecName: "kube-api-access-cd6zc") pod "9bb32c8b-83c8-4be6-93c0-3dd925d46665" (UID: "9bb32c8b-83c8-4be6-93c0-3dd925d46665"). InnerVolumeSpecName "kube-api-access-cd6zc". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 20:30:04 crc kubenswrapper[4988]: I1008 20:30:04.158290 4988 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/9bb32c8b-83c8-4be6-93c0-3dd925d46665-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 08 20:30:04 crc kubenswrapper[4988]: I1008 20:30:04.158319 4988 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cd6zc\" (UniqueName: \"kubernetes.io/projected/9bb32c8b-83c8-4be6-93c0-3dd925d46665-kube-api-access-cd6zc\") on node \"crc\" DevicePath \"\"" Oct 08 20:30:04 crc kubenswrapper[4988]: I1008 20:30:04.512469 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29332590-6fvwv" event={"ID":"9bb32c8b-83c8-4be6-93c0-3dd925d46665","Type":"ContainerDied","Data":"e5ea41d0935b6f29f789c16296be03317c867e444a05b8707d3f7e30cc617773"} Oct 08 20:30:04 crc kubenswrapper[4988]: I1008 20:30:04.512531 4988 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e5ea41d0935b6f29f789c16296be03317c867e444a05b8707d3f7e30cc617773" Oct 08 20:30:04 crc kubenswrapper[4988]: I1008 20:30:04.512617 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29332590-6fvwv" Oct 08 20:30:04 crc kubenswrapper[4988]: I1008 20:30:04.588975 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29332545-d245p"] Oct 08 20:30:04 crc kubenswrapper[4988]: I1008 20:30:04.600907 4988 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29332545-d245p"] Oct 08 20:30:05 crc kubenswrapper[4988]: I1008 20:30:05.255278 4988 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c88f1935-f363-4596-9b74-fe8b3ee2946a" path="/var/lib/kubelet/pods/c88f1935-f363-4596-9b74-fe8b3ee2946a/volumes" Oct 08 20:30:15 crc kubenswrapper[4988]: I1008 20:30:15.833458 4988 scope.go:117] "RemoveContainer" containerID="0e7dce3df7eb9938242996e6b4d550500a33eb43c94bed56498f93c0669e77f1" Oct 08 20:30:36 crc kubenswrapper[4988]: I1008 20:30:36.362443 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-c4cz2"] Oct 08 20:30:36 crc kubenswrapper[4988]: E1008 20:30:36.365056 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9bb32c8b-83c8-4be6-93c0-3dd925d46665" containerName="collect-profiles" Oct 08 20:30:36 crc kubenswrapper[4988]: I1008 20:30:36.366217 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="9bb32c8b-83c8-4be6-93c0-3dd925d46665" containerName="collect-profiles" Oct 08 20:30:36 crc kubenswrapper[4988]: I1008 20:30:36.366741 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="9bb32c8b-83c8-4be6-93c0-3dd925d46665" containerName="collect-profiles" Oct 08 20:30:36 crc kubenswrapper[4988]: I1008 20:30:36.369358 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-c4cz2" Oct 08 20:30:36 crc kubenswrapper[4988]: I1008 20:30:36.385240 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-c4cz2"] Oct 08 20:30:36 crc kubenswrapper[4988]: I1008 20:30:36.483880 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/34c2181b-c702-4807-a4bf-4d5186bde6e6-catalog-content\") pod \"redhat-operators-c4cz2\" (UID: \"34c2181b-c702-4807-a4bf-4d5186bde6e6\") " pod="openshift-marketplace/redhat-operators-c4cz2" Oct 08 20:30:36 crc kubenswrapper[4988]: I1008 20:30:36.484114 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bnqtq\" (UniqueName: \"kubernetes.io/projected/34c2181b-c702-4807-a4bf-4d5186bde6e6-kube-api-access-bnqtq\") pod \"redhat-operators-c4cz2\" (UID: \"34c2181b-c702-4807-a4bf-4d5186bde6e6\") " pod="openshift-marketplace/redhat-operators-c4cz2" Oct 08 20:30:36 crc kubenswrapper[4988]: I1008 20:30:36.484323 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/34c2181b-c702-4807-a4bf-4d5186bde6e6-utilities\") pod \"redhat-operators-c4cz2\" (UID: \"34c2181b-c702-4807-a4bf-4d5186bde6e6\") " pod="openshift-marketplace/redhat-operators-c4cz2" Oct 08 20:30:36 crc kubenswrapper[4988]: I1008 20:30:36.586086 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/34c2181b-c702-4807-a4bf-4d5186bde6e6-utilities\") pod \"redhat-operators-c4cz2\" (UID: \"34c2181b-c702-4807-a4bf-4d5186bde6e6\") " pod="openshift-marketplace/redhat-operators-c4cz2" Oct 08 20:30:36 crc kubenswrapper[4988]: I1008 20:30:36.586156 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/34c2181b-c702-4807-a4bf-4d5186bde6e6-catalog-content\") pod \"redhat-operators-c4cz2\" (UID: \"34c2181b-c702-4807-a4bf-4d5186bde6e6\") " pod="openshift-marketplace/redhat-operators-c4cz2" Oct 08 20:30:36 crc kubenswrapper[4988]: I1008 20:30:36.586297 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bnqtq\" (UniqueName: \"kubernetes.io/projected/34c2181b-c702-4807-a4bf-4d5186bde6e6-kube-api-access-bnqtq\") pod \"redhat-operators-c4cz2\" (UID: \"34c2181b-c702-4807-a4bf-4d5186bde6e6\") " pod="openshift-marketplace/redhat-operators-c4cz2" Oct 08 20:30:36 crc kubenswrapper[4988]: I1008 20:30:36.586741 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/34c2181b-c702-4807-a4bf-4d5186bde6e6-catalog-content\") pod \"redhat-operators-c4cz2\" (UID: \"34c2181b-c702-4807-a4bf-4d5186bde6e6\") " pod="openshift-marketplace/redhat-operators-c4cz2" Oct 08 20:30:36 crc kubenswrapper[4988]: I1008 20:30:36.586920 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/34c2181b-c702-4807-a4bf-4d5186bde6e6-utilities\") pod \"redhat-operators-c4cz2\" (UID: \"34c2181b-c702-4807-a4bf-4d5186bde6e6\") " pod="openshift-marketplace/redhat-operators-c4cz2" Oct 08 20:30:36 crc kubenswrapper[4988]: I1008 20:30:36.608542 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bnqtq\" (UniqueName: \"kubernetes.io/projected/34c2181b-c702-4807-a4bf-4d5186bde6e6-kube-api-access-bnqtq\") pod \"redhat-operators-c4cz2\" (UID: \"34c2181b-c702-4807-a4bf-4d5186bde6e6\") " pod="openshift-marketplace/redhat-operators-c4cz2" Oct 08 20:30:36 crc kubenswrapper[4988]: I1008 20:30:36.698308 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-c4cz2" Oct 08 20:30:37 crc kubenswrapper[4988]: I1008 20:30:37.173362 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-c4cz2"] Oct 08 20:30:37 crc kubenswrapper[4988]: I1008 20:30:37.924821 4988 generic.go:334] "Generic (PLEG): container finished" podID="34c2181b-c702-4807-a4bf-4d5186bde6e6" containerID="86d35aa0f130f6f353b79b89ac6f0443b7e4bcbaf065bac182fccf61174e514a" exitCode=0 Oct 08 20:30:37 crc kubenswrapper[4988]: I1008 20:30:37.924923 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-c4cz2" event={"ID":"34c2181b-c702-4807-a4bf-4d5186bde6e6","Type":"ContainerDied","Data":"86d35aa0f130f6f353b79b89ac6f0443b7e4bcbaf065bac182fccf61174e514a"} Oct 08 20:30:37 crc kubenswrapper[4988]: I1008 20:30:37.925791 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-c4cz2" event={"ID":"34c2181b-c702-4807-a4bf-4d5186bde6e6","Type":"ContainerStarted","Data":"257178f51055ea90d48e11388339577bbccecd5479eb68ee02352649f6c768fb"} Oct 08 20:30:37 crc kubenswrapper[4988]: I1008 20:30:37.927812 4988 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 08 20:30:39 crc kubenswrapper[4988]: I1008 20:30:39.950432 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-c4cz2" event={"ID":"34c2181b-c702-4807-a4bf-4d5186bde6e6","Type":"ContainerStarted","Data":"d7cf9d127f684b25182152f3ae77a1c7c15ce2590b77f80b0f09898906e56362"} Oct 08 20:30:41 crc kubenswrapper[4988]: I1008 20:30:41.971162 4988 generic.go:334] "Generic (PLEG): container finished" podID="34c2181b-c702-4807-a4bf-4d5186bde6e6" containerID="d7cf9d127f684b25182152f3ae77a1c7c15ce2590b77f80b0f09898906e56362" exitCode=0 Oct 08 20:30:41 crc kubenswrapper[4988]: I1008 20:30:41.971551 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-c4cz2" event={"ID":"34c2181b-c702-4807-a4bf-4d5186bde6e6","Type":"ContainerDied","Data":"d7cf9d127f684b25182152f3ae77a1c7c15ce2590b77f80b0f09898906e56362"} Oct 08 20:30:42 crc kubenswrapper[4988]: I1008 20:30:42.994819 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-c4cz2" event={"ID":"34c2181b-c702-4807-a4bf-4d5186bde6e6","Type":"ContainerStarted","Data":"0fa339bb17905f8ed14fbaab0f3303a90e2d91211ca50bac16e9b004c1893f4f"} Oct 08 20:30:43 crc kubenswrapper[4988]: I1008 20:30:43.047148 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-c4cz2" podStartSLOduration=2.560385424 podStartE2EDuration="7.047126445s" podCreationTimestamp="2025-10-08 20:30:36 +0000 UTC" firstStartedPulling="2025-10-08 20:30:37.9272135 +0000 UTC m=+8383.377056300" lastFinishedPulling="2025-10-08 20:30:42.413954551 +0000 UTC m=+8387.863797321" observedRunningTime="2025-10-08 20:30:43.024868678 +0000 UTC m=+8388.474711548" watchObservedRunningTime="2025-10-08 20:30:43.047126445 +0000 UTC m=+8388.496969225" Oct 08 20:30:46 crc kubenswrapper[4988]: I1008 20:30:46.698713 4988 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-c4cz2" Oct 08 20:30:46 crc kubenswrapper[4988]: I1008 20:30:46.699172 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-c4cz2" Oct 08 20:30:47 crc kubenswrapper[4988]: I1008 20:30:47.740740 4988 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-c4cz2" podUID="34c2181b-c702-4807-a4bf-4d5186bde6e6" containerName="registry-server" probeResult="failure" output=< Oct 08 20:30:47 crc kubenswrapper[4988]: timeout: failed to connect service ":50051" within 1s Oct 08 20:30:47 crc kubenswrapper[4988]: > Oct 08 20:30:52 crc kubenswrapper[4988]: I1008 20:30:52.800865 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-tg6lc"] Oct 08 20:30:52 crc kubenswrapper[4988]: I1008 20:30:52.804227 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-tg6lc" Oct 08 20:30:52 crc kubenswrapper[4988]: I1008 20:30:52.815086 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-tg6lc"] Oct 08 20:30:52 crc kubenswrapper[4988]: I1008 20:30:52.913644 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0be1269f-0a8b-48e2-a60b-98e83628f6d0-utilities\") pod \"certified-operators-tg6lc\" (UID: \"0be1269f-0a8b-48e2-a60b-98e83628f6d0\") " pod="openshift-marketplace/certified-operators-tg6lc" Oct 08 20:30:52 crc kubenswrapper[4988]: I1008 20:30:52.913755 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r9hzc\" (UniqueName: \"kubernetes.io/projected/0be1269f-0a8b-48e2-a60b-98e83628f6d0-kube-api-access-r9hzc\") pod \"certified-operators-tg6lc\" (UID: \"0be1269f-0a8b-48e2-a60b-98e83628f6d0\") " pod="openshift-marketplace/certified-operators-tg6lc" Oct 08 20:30:52 crc kubenswrapper[4988]: I1008 20:30:52.913813 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0be1269f-0a8b-48e2-a60b-98e83628f6d0-catalog-content\") pod \"certified-operators-tg6lc\" (UID: \"0be1269f-0a8b-48e2-a60b-98e83628f6d0\") " pod="openshift-marketplace/certified-operators-tg6lc" Oct 08 20:30:53 crc kubenswrapper[4988]: I1008 20:30:53.016454 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0be1269f-0a8b-48e2-a60b-98e83628f6d0-utilities\") pod \"certified-operators-tg6lc\" (UID: \"0be1269f-0a8b-48e2-a60b-98e83628f6d0\") " pod="openshift-marketplace/certified-operators-tg6lc" Oct 08 20:30:53 crc kubenswrapper[4988]: I1008 20:30:53.016906 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r9hzc\" (UniqueName: \"kubernetes.io/projected/0be1269f-0a8b-48e2-a60b-98e83628f6d0-kube-api-access-r9hzc\") pod \"certified-operators-tg6lc\" (UID: \"0be1269f-0a8b-48e2-a60b-98e83628f6d0\") " pod="openshift-marketplace/certified-operators-tg6lc" Oct 08 20:30:53 crc kubenswrapper[4988]: I1008 20:30:53.016993 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0be1269f-0a8b-48e2-a60b-98e83628f6d0-catalog-content\") pod \"certified-operators-tg6lc\" (UID: \"0be1269f-0a8b-48e2-a60b-98e83628f6d0\") " pod="openshift-marketplace/certified-operators-tg6lc" Oct 08 20:30:53 crc kubenswrapper[4988]: I1008 20:30:53.017903 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0be1269f-0a8b-48e2-a60b-98e83628f6d0-utilities\") pod \"certified-operators-tg6lc\" (UID: \"0be1269f-0a8b-48e2-a60b-98e83628f6d0\") " pod="openshift-marketplace/certified-operators-tg6lc" Oct 08 20:30:53 crc kubenswrapper[4988]: I1008 20:30:53.017972 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0be1269f-0a8b-48e2-a60b-98e83628f6d0-catalog-content\") pod \"certified-operators-tg6lc\" (UID: \"0be1269f-0a8b-48e2-a60b-98e83628f6d0\") " pod="openshift-marketplace/certified-operators-tg6lc" Oct 08 20:30:53 crc kubenswrapper[4988]: I1008 20:30:53.043829 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r9hzc\" (UniqueName: \"kubernetes.io/projected/0be1269f-0a8b-48e2-a60b-98e83628f6d0-kube-api-access-r9hzc\") pod \"certified-operators-tg6lc\" (UID: \"0be1269f-0a8b-48e2-a60b-98e83628f6d0\") " pod="openshift-marketplace/certified-operators-tg6lc" Oct 08 20:30:53 crc kubenswrapper[4988]: I1008 20:30:53.131709 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-tg6lc" Oct 08 20:30:53 crc kubenswrapper[4988]: I1008 20:30:53.697023 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-tg6lc"] Oct 08 20:30:53 crc kubenswrapper[4988]: W1008 20:30:53.704769 4988 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0be1269f_0a8b_48e2_a60b_98e83628f6d0.slice/crio-9e2e160c4303804248a81f6e29e08f72585c0059d50866af8312504db1a96396 WatchSource:0}: Error finding container 9e2e160c4303804248a81f6e29e08f72585c0059d50866af8312504db1a96396: Status 404 returned error can't find the container with id 9e2e160c4303804248a81f6e29e08f72585c0059d50866af8312504db1a96396 Oct 08 20:30:54 crc kubenswrapper[4988]: I1008 20:30:54.128341 4988 generic.go:334] "Generic (PLEG): container finished" podID="0be1269f-0a8b-48e2-a60b-98e83628f6d0" containerID="d7e9ac863efe5ea10fc02b61b17ec5a4979a7a3d42997f96b5ec798a5ed34747" exitCode=0 Oct 08 20:30:54 crc kubenswrapper[4988]: I1008 20:30:54.128460 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-tg6lc" event={"ID":"0be1269f-0a8b-48e2-a60b-98e83628f6d0","Type":"ContainerDied","Data":"d7e9ac863efe5ea10fc02b61b17ec5a4979a7a3d42997f96b5ec798a5ed34747"} Oct 08 20:30:54 crc kubenswrapper[4988]: I1008 20:30:54.128744 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-tg6lc" event={"ID":"0be1269f-0a8b-48e2-a60b-98e83628f6d0","Type":"ContainerStarted","Data":"9e2e160c4303804248a81f6e29e08f72585c0059d50866af8312504db1a96396"} Oct 08 20:30:54 crc kubenswrapper[4988]: E1008 20:30:54.181607 4988 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0be1269f_0a8b_48e2_a60b_98e83628f6d0.slice/crio-d7e9ac863efe5ea10fc02b61b17ec5a4979a7a3d42997f96b5ec798a5ed34747.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0be1269f_0a8b_48e2_a60b_98e83628f6d0.slice/crio-conmon-d7e9ac863efe5ea10fc02b61b17ec5a4979a7a3d42997f96b5ec798a5ed34747.scope\": RecentStats: unable to find data in memory cache]" Oct 08 20:30:56 crc kubenswrapper[4988]: I1008 20:30:56.163500 4988 generic.go:334] "Generic (PLEG): container finished" podID="aef3b7a1-6628-4895-a380-153a230bd0a3" containerID="affd5206fa13d0036c43bfaaf072688f3755710f77bc251632479521b90a3e46" exitCode=0 Oct 08 20:30:56 crc kubenswrapper[4988]: I1008 20:30:56.163550 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-openstack-openstack-cell1-mzn6v" event={"ID":"aef3b7a1-6628-4895-a380-153a230bd0a3","Type":"ContainerDied","Data":"affd5206fa13d0036c43bfaaf072688f3755710f77bc251632479521b90a3e46"} Oct 08 20:30:56 crc kubenswrapper[4988]: I1008 20:30:56.167249 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-tg6lc" event={"ID":"0be1269f-0a8b-48e2-a60b-98e83628f6d0","Type":"ContainerStarted","Data":"20bc2b70475e6be093e53a6a34575fe27b87e6882a36507437eb2b9001c30680"} Oct 08 20:30:56 crc kubenswrapper[4988]: I1008 20:30:56.785521 4988 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-c4cz2" Oct 08 20:30:56 crc kubenswrapper[4988]: I1008 20:30:56.879551 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-c4cz2" Oct 08 20:30:57 crc kubenswrapper[4988]: I1008 20:30:57.182837 4988 generic.go:334] "Generic (PLEG): container finished" podID="0be1269f-0a8b-48e2-a60b-98e83628f6d0" containerID="20bc2b70475e6be093e53a6a34575fe27b87e6882a36507437eb2b9001c30680" exitCode=0 Oct 08 20:30:57 crc kubenswrapper[4988]: I1008 20:30:57.182928 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-tg6lc" event={"ID":"0be1269f-0a8b-48e2-a60b-98e83628f6d0","Type":"ContainerDied","Data":"20bc2b70475e6be093e53a6a34575fe27b87e6882a36507437eb2b9001c30680"} Oct 08 20:30:57 crc kubenswrapper[4988]: I1008 20:30:57.777933 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-openstack-openstack-cell1-mzn6v" Oct 08 20:30:57 crc kubenswrapper[4988]: I1008 20:30:57.848072 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/aef3b7a1-6628-4895-a380-153a230bd0a3-ssh-key\") pod \"aef3b7a1-6628-4895-a380-153a230bd0a3\" (UID: \"aef3b7a1-6628-4895-a380-153a230bd0a3\") " Oct 08 20:30:57 crc kubenswrapper[4988]: I1008 20:30:57.848236 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/aef3b7a1-6628-4895-a380-153a230bd0a3-libvirt-secret-0\") pod \"aef3b7a1-6628-4895-a380-153a230bd0a3\" (UID: \"aef3b7a1-6628-4895-a380-153a230bd0a3\") " Oct 08 20:30:57 crc kubenswrapper[4988]: I1008 20:30:57.848279 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/aef3b7a1-6628-4895-a380-153a230bd0a3-inventory\") pod \"aef3b7a1-6628-4895-a380-153a230bd0a3\" (UID: \"aef3b7a1-6628-4895-a380-153a230bd0a3\") " Oct 08 20:30:57 crc kubenswrapper[4988]: I1008 20:30:57.848346 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cmb68\" (UniqueName: \"kubernetes.io/projected/aef3b7a1-6628-4895-a380-153a230bd0a3-kube-api-access-cmb68\") pod \"aef3b7a1-6628-4895-a380-153a230bd0a3\" (UID: \"aef3b7a1-6628-4895-a380-153a230bd0a3\") " Oct 08 20:30:57 crc kubenswrapper[4988]: I1008 20:30:57.848482 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aef3b7a1-6628-4895-a380-153a230bd0a3-libvirt-combined-ca-bundle\") pod \"aef3b7a1-6628-4895-a380-153a230bd0a3\" (UID: \"aef3b7a1-6628-4895-a380-153a230bd0a3\") " Oct 08 20:30:57 crc kubenswrapper[4988]: I1008 20:30:57.853977 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aef3b7a1-6628-4895-a380-153a230bd0a3-libvirt-combined-ca-bundle" (OuterVolumeSpecName: "libvirt-combined-ca-bundle") pod "aef3b7a1-6628-4895-a380-153a230bd0a3" (UID: "aef3b7a1-6628-4895-a380-153a230bd0a3"). InnerVolumeSpecName "libvirt-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 20:30:57 crc kubenswrapper[4988]: I1008 20:30:57.856156 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/aef3b7a1-6628-4895-a380-153a230bd0a3-kube-api-access-cmb68" (OuterVolumeSpecName: "kube-api-access-cmb68") pod "aef3b7a1-6628-4895-a380-153a230bd0a3" (UID: "aef3b7a1-6628-4895-a380-153a230bd0a3"). InnerVolumeSpecName "kube-api-access-cmb68". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 20:30:57 crc kubenswrapper[4988]: I1008 20:30:57.895895 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aef3b7a1-6628-4895-a380-153a230bd0a3-inventory" (OuterVolumeSpecName: "inventory") pod "aef3b7a1-6628-4895-a380-153a230bd0a3" (UID: "aef3b7a1-6628-4895-a380-153a230bd0a3"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 20:30:57 crc kubenswrapper[4988]: I1008 20:30:57.906935 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aef3b7a1-6628-4895-a380-153a230bd0a3-libvirt-secret-0" (OuterVolumeSpecName: "libvirt-secret-0") pod "aef3b7a1-6628-4895-a380-153a230bd0a3" (UID: "aef3b7a1-6628-4895-a380-153a230bd0a3"). InnerVolumeSpecName "libvirt-secret-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 20:30:57 crc kubenswrapper[4988]: I1008 20:30:57.909156 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aef3b7a1-6628-4895-a380-153a230bd0a3-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "aef3b7a1-6628-4895-a380-153a230bd0a3" (UID: "aef3b7a1-6628-4895-a380-153a230bd0a3"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 20:30:57 crc kubenswrapper[4988]: I1008 20:30:57.951293 4988 reconciler_common.go:293] "Volume detached for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/aef3b7a1-6628-4895-a380-153a230bd0a3-libvirt-secret-0\") on node \"crc\" DevicePath \"\"" Oct 08 20:30:57 crc kubenswrapper[4988]: I1008 20:30:57.951334 4988 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/aef3b7a1-6628-4895-a380-153a230bd0a3-inventory\") on node \"crc\" DevicePath \"\"" Oct 08 20:30:57 crc kubenswrapper[4988]: I1008 20:30:57.951348 4988 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cmb68\" (UniqueName: \"kubernetes.io/projected/aef3b7a1-6628-4895-a380-153a230bd0a3-kube-api-access-cmb68\") on node \"crc\" DevicePath \"\"" Oct 08 20:30:57 crc kubenswrapper[4988]: I1008 20:30:57.951363 4988 reconciler_common.go:293] "Volume detached for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aef3b7a1-6628-4895-a380-153a230bd0a3-libvirt-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 20:30:57 crc kubenswrapper[4988]: I1008 20:30:57.951377 4988 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/aef3b7a1-6628-4895-a380-153a230bd0a3-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 08 20:30:58 crc kubenswrapper[4988]: I1008 20:30:58.196414 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-openstack-openstack-cell1-mzn6v" event={"ID":"aef3b7a1-6628-4895-a380-153a230bd0a3","Type":"ContainerDied","Data":"657950ad28784152be6e4e5e6b331f8e7d80ea51ec12cb9901b686cb68cbc47b"} Oct 08 20:30:58 crc kubenswrapper[4988]: I1008 20:30:58.196453 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-openstack-openstack-cell1-mzn6v" Oct 08 20:30:58 crc kubenswrapper[4988]: I1008 20:30:58.196473 4988 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="657950ad28784152be6e4e5e6b331f8e7d80ea51ec12cb9901b686cb68cbc47b" Oct 08 20:30:58 crc kubenswrapper[4988]: I1008 20:30:58.199267 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-tg6lc" event={"ID":"0be1269f-0a8b-48e2-a60b-98e83628f6d0","Type":"ContainerStarted","Data":"3d5f531ac4ee8366686db21aab2e80fab04a00fb15c25c673c71d25d56eb8478"} Oct 08 20:30:58 crc kubenswrapper[4988]: I1008 20:30:58.236275 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-tg6lc" podStartSLOduration=2.539929311 podStartE2EDuration="6.236250246s" podCreationTimestamp="2025-10-08 20:30:52 +0000 UTC" firstStartedPulling="2025-10-08 20:30:54.134582181 +0000 UTC m=+8399.584424951" lastFinishedPulling="2025-10-08 20:30:57.830903086 +0000 UTC m=+8403.280745886" observedRunningTime="2025-10-08 20:30:58.229122634 +0000 UTC m=+8403.678965414" watchObservedRunningTime="2025-10-08 20:30:58.236250246 +0000 UTC m=+8403.686093026" Oct 08 20:30:58 crc kubenswrapper[4988]: I1008 20:30:58.288962 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-openstack-openstack-cell1-j9bg2"] Oct 08 20:30:58 crc kubenswrapper[4988]: E1008 20:30:58.289473 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aef3b7a1-6628-4895-a380-153a230bd0a3" containerName="libvirt-openstack-openstack-cell1" Oct 08 20:30:58 crc kubenswrapper[4988]: I1008 20:30:58.289495 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="aef3b7a1-6628-4895-a380-153a230bd0a3" containerName="libvirt-openstack-openstack-cell1" Oct 08 20:30:58 crc kubenswrapper[4988]: I1008 20:30:58.289809 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="aef3b7a1-6628-4895-a380-153a230bd0a3" containerName="libvirt-openstack-openstack-cell1" Oct 08 20:30:58 crc kubenswrapper[4988]: I1008 20:30:58.290734 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-openstack-openstack-cell1-j9bg2" Oct 08 20:30:58 crc kubenswrapper[4988]: I1008 20:30:58.294046 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-compute-config" Oct 08 20:30:58 crc kubenswrapper[4988]: I1008 20:30:58.294313 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Oct 08 20:30:58 crc kubenswrapper[4988]: I1008 20:30:58.294354 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-migration-ssh-key" Oct 08 20:30:58 crc kubenswrapper[4988]: I1008 20:30:58.294371 4988 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 08 20:30:58 crc kubenswrapper[4988]: I1008 20:30:58.294835 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Oct 08 20:30:58 crc kubenswrapper[4988]: I1008 20:30:58.296229 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-4sdzr" Oct 08 20:30:58 crc kubenswrapper[4988]: I1008 20:30:58.301153 4988 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"nova-cells-global-config" Oct 08 20:30:58 crc kubenswrapper[4988]: I1008 20:30:58.302370 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-openstack-openstack-cell1-j9bg2"] Oct 08 20:30:58 crc kubenswrapper[4988]: I1008 20:30:58.360182 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cells-global-config-0\" (UniqueName: \"kubernetes.io/configmap/08584161-d03c-4354-9e07-60eae54b233d-nova-cells-global-config-0\") pod \"nova-cell1-openstack-openstack-cell1-j9bg2\" (UID: \"08584161-d03c-4354-9e07-60eae54b233d\") " pod="openstack/nova-cell1-openstack-openstack-cell1-j9bg2" Oct 08 20:30:58 crc kubenswrapper[4988]: I1008 20:30:58.360230 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/08584161-d03c-4354-9e07-60eae54b233d-ssh-key\") pod \"nova-cell1-openstack-openstack-cell1-j9bg2\" (UID: \"08584161-d03c-4354-9e07-60eae54b233d\") " pod="openstack/nova-cell1-openstack-openstack-cell1-j9bg2" Oct 08 20:30:58 crc kubenswrapper[4988]: I1008 20:30:58.360274 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/08584161-d03c-4354-9e07-60eae54b233d-nova-cell1-combined-ca-bundle\") pod \"nova-cell1-openstack-openstack-cell1-j9bg2\" (UID: \"08584161-d03c-4354-9e07-60eae54b233d\") " pod="openstack/nova-cell1-openstack-openstack-cell1-j9bg2" Oct 08 20:30:58 crc kubenswrapper[4988]: I1008 20:30:58.360515 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/08584161-d03c-4354-9e07-60eae54b233d-nova-migration-ssh-key-0\") pod \"nova-cell1-openstack-openstack-cell1-j9bg2\" (UID: \"08584161-d03c-4354-9e07-60eae54b233d\") " pod="openstack/nova-cell1-openstack-openstack-cell1-j9bg2" Oct 08 20:30:58 crc kubenswrapper[4988]: I1008 20:30:58.360553 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/08584161-d03c-4354-9e07-60eae54b233d-inventory\") pod \"nova-cell1-openstack-openstack-cell1-j9bg2\" (UID: \"08584161-d03c-4354-9e07-60eae54b233d\") " pod="openstack/nova-cell1-openstack-openstack-cell1-j9bg2" Oct 08 20:30:58 crc kubenswrapper[4988]: I1008 20:30:58.360910 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/08584161-d03c-4354-9e07-60eae54b233d-nova-cell1-compute-config-0\") pod \"nova-cell1-openstack-openstack-cell1-j9bg2\" (UID: \"08584161-d03c-4354-9e07-60eae54b233d\") " pod="openstack/nova-cell1-openstack-openstack-cell1-j9bg2" Oct 08 20:30:58 crc kubenswrapper[4988]: I1008 20:30:58.361011 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8j8gm\" (UniqueName: \"kubernetes.io/projected/08584161-d03c-4354-9e07-60eae54b233d-kube-api-access-8j8gm\") pod \"nova-cell1-openstack-openstack-cell1-j9bg2\" (UID: \"08584161-d03c-4354-9e07-60eae54b233d\") " pod="openstack/nova-cell1-openstack-openstack-cell1-j9bg2" Oct 08 20:30:58 crc kubenswrapper[4988]: I1008 20:30:58.361190 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/08584161-d03c-4354-9e07-60eae54b233d-nova-migration-ssh-key-1\") pod \"nova-cell1-openstack-openstack-cell1-j9bg2\" (UID: \"08584161-d03c-4354-9e07-60eae54b233d\") " pod="openstack/nova-cell1-openstack-openstack-cell1-j9bg2" Oct 08 20:30:58 crc kubenswrapper[4988]: I1008 20:30:58.361247 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/08584161-d03c-4354-9e07-60eae54b233d-nova-cell1-compute-config-1\") pod \"nova-cell1-openstack-openstack-cell1-j9bg2\" (UID: \"08584161-d03c-4354-9e07-60eae54b233d\") " pod="openstack/nova-cell1-openstack-openstack-cell1-j9bg2" Oct 08 20:30:58 crc kubenswrapper[4988]: I1008 20:30:58.462945 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/08584161-d03c-4354-9e07-60eae54b233d-nova-cell1-compute-config-1\") pod \"nova-cell1-openstack-openstack-cell1-j9bg2\" (UID: \"08584161-d03c-4354-9e07-60eae54b233d\") " pod="openstack/nova-cell1-openstack-openstack-cell1-j9bg2" Oct 08 20:30:58 crc kubenswrapper[4988]: I1008 20:30:58.463292 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cells-global-config-0\" (UniqueName: \"kubernetes.io/configmap/08584161-d03c-4354-9e07-60eae54b233d-nova-cells-global-config-0\") pod \"nova-cell1-openstack-openstack-cell1-j9bg2\" (UID: \"08584161-d03c-4354-9e07-60eae54b233d\") " pod="openstack/nova-cell1-openstack-openstack-cell1-j9bg2" Oct 08 20:30:58 crc kubenswrapper[4988]: I1008 20:30:58.463324 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/08584161-d03c-4354-9e07-60eae54b233d-ssh-key\") pod \"nova-cell1-openstack-openstack-cell1-j9bg2\" (UID: \"08584161-d03c-4354-9e07-60eae54b233d\") " pod="openstack/nova-cell1-openstack-openstack-cell1-j9bg2" Oct 08 20:30:58 crc kubenswrapper[4988]: I1008 20:30:58.463373 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/08584161-d03c-4354-9e07-60eae54b233d-nova-cell1-combined-ca-bundle\") pod \"nova-cell1-openstack-openstack-cell1-j9bg2\" (UID: \"08584161-d03c-4354-9e07-60eae54b233d\") " pod="openstack/nova-cell1-openstack-openstack-cell1-j9bg2" Oct 08 20:30:58 crc kubenswrapper[4988]: I1008 20:30:58.463458 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/08584161-d03c-4354-9e07-60eae54b233d-nova-migration-ssh-key-0\") pod \"nova-cell1-openstack-openstack-cell1-j9bg2\" (UID: \"08584161-d03c-4354-9e07-60eae54b233d\") " pod="openstack/nova-cell1-openstack-openstack-cell1-j9bg2" Oct 08 20:30:58 crc kubenswrapper[4988]: I1008 20:30:58.463490 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/08584161-d03c-4354-9e07-60eae54b233d-inventory\") pod \"nova-cell1-openstack-openstack-cell1-j9bg2\" (UID: \"08584161-d03c-4354-9e07-60eae54b233d\") " pod="openstack/nova-cell1-openstack-openstack-cell1-j9bg2" Oct 08 20:30:58 crc kubenswrapper[4988]: I1008 20:30:58.463611 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/08584161-d03c-4354-9e07-60eae54b233d-nova-cell1-compute-config-0\") pod \"nova-cell1-openstack-openstack-cell1-j9bg2\" (UID: \"08584161-d03c-4354-9e07-60eae54b233d\") " pod="openstack/nova-cell1-openstack-openstack-cell1-j9bg2" Oct 08 20:30:58 crc kubenswrapper[4988]: I1008 20:30:58.463653 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8j8gm\" (UniqueName: \"kubernetes.io/projected/08584161-d03c-4354-9e07-60eae54b233d-kube-api-access-8j8gm\") pod \"nova-cell1-openstack-openstack-cell1-j9bg2\" (UID: \"08584161-d03c-4354-9e07-60eae54b233d\") " pod="openstack/nova-cell1-openstack-openstack-cell1-j9bg2" Oct 08 20:30:58 crc kubenswrapper[4988]: I1008 20:30:58.463725 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/08584161-d03c-4354-9e07-60eae54b233d-nova-migration-ssh-key-1\") pod \"nova-cell1-openstack-openstack-cell1-j9bg2\" (UID: \"08584161-d03c-4354-9e07-60eae54b233d\") " pod="openstack/nova-cell1-openstack-openstack-cell1-j9bg2" Oct 08 20:30:58 crc kubenswrapper[4988]: I1008 20:30:58.464236 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cells-global-config-0\" (UniqueName: \"kubernetes.io/configmap/08584161-d03c-4354-9e07-60eae54b233d-nova-cells-global-config-0\") pod \"nova-cell1-openstack-openstack-cell1-j9bg2\" (UID: \"08584161-d03c-4354-9e07-60eae54b233d\") " pod="openstack/nova-cell1-openstack-openstack-cell1-j9bg2" Oct 08 20:30:58 crc kubenswrapper[4988]: I1008 20:30:58.467617 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/08584161-d03c-4354-9e07-60eae54b233d-nova-cell1-compute-config-1\") pod \"nova-cell1-openstack-openstack-cell1-j9bg2\" (UID: \"08584161-d03c-4354-9e07-60eae54b233d\") " pod="openstack/nova-cell1-openstack-openstack-cell1-j9bg2" Oct 08 20:30:58 crc kubenswrapper[4988]: I1008 20:30:58.467775 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/08584161-d03c-4354-9e07-60eae54b233d-nova-migration-ssh-key-0\") pod \"nova-cell1-openstack-openstack-cell1-j9bg2\" (UID: \"08584161-d03c-4354-9e07-60eae54b233d\") " pod="openstack/nova-cell1-openstack-openstack-cell1-j9bg2" Oct 08 20:30:58 crc kubenswrapper[4988]: I1008 20:30:58.468020 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/08584161-d03c-4354-9e07-60eae54b233d-inventory\") pod \"nova-cell1-openstack-openstack-cell1-j9bg2\" (UID: \"08584161-d03c-4354-9e07-60eae54b233d\") " pod="openstack/nova-cell1-openstack-openstack-cell1-j9bg2" Oct 08 20:30:58 crc kubenswrapper[4988]: I1008 20:30:58.468085 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/08584161-d03c-4354-9e07-60eae54b233d-nova-cell1-combined-ca-bundle\") pod \"nova-cell1-openstack-openstack-cell1-j9bg2\" (UID: \"08584161-d03c-4354-9e07-60eae54b233d\") " pod="openstack/nova-cell1-openstack-openstack-cell1-j9bg2" Oct 08 20:30:58 crc kubenswrapper[4988]: I1008 20:30:58.468253 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/08584161-d03c-4354-9e07-60eae54b233d-ssh-key\") pod \"nova-cell1-openstack-openstack-cell1-j9bg2\" (UID: \"08584161-d03c-4354-9e07-60eae54b233d\") " pod="openstack/nova-cell1-openstack-openstack-cell1-j9bg2" Oct 08 20:30:58 crc kubenswrapper[4988]: I1008 20:30:58.468575 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/08584161-d03c-4354-9e07-60eae54b233d-nova-migration-ssh-key-1\") pod \"nova-cell1-openstack-openstack-cell1-j9bg2\" (UID: \"08584161-d03c-4354-9e07-60eae54b233d\") " pod="openstack/nova-cell1-openstack-openstack-cell1-j9bg2" Oct 08 20:30:58 crc kubenswrapper[4988]: I1008 20:30:58.469142 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/08584161-d03c-4354-9e07-60eae54b233d-nova-cell1-compute-config-0\") pod \"nova-cell1-openstack-openstack-cell1-j9bg2\" (UID: \"08584161-d03c-4354-9e07-60eae54b233d\") " pod="openstack/nova-cell1-openstack-openstack-cell1-j9bg2" Oct 08 20:30:58 crc kubenswrapper[4988]: I1008 20:30:58.480338 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8j8gm\" (UniqueName: \"kubernetes.io/projected/08584161-d03c-4354-9e07-60eae54b233d-kube-api-access-8j8gm\") pod \"nova-cell1-openstack-openstack-cell1-j9bg2\" (UID: \"08584161-d03c-4354-9e07-60eae54b233d\") " pod="openstack/nova-cell1-openstack-openstack-cell1-j9bg2" Oct 08 20:30:58 crc kubenswrapper[4988]: I1008 20:30:58.616637 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-openstack-openstack-cell1-j9bg2" Oct 08 20:30:58 crc kubenswrapper[4988]: I1008 20:30:58.984859 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-c4cz2"] Oct 08 20:30:58 crc kubenswrapper[4988]: I1008 20:30:58.985362 4988 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-c4cz2" podUID="34c2181b-c702-4807-a4bf-4d5186bde6e6" containerName="registry-server" containerID="cri-o://0fa339bb17905f8ed14fbaab0f3303a90e2d91211ca50bac16e9b004c1893f4f" gracePeriod=2 Oct 08 20:30:59 crc kubenswrapper[4988]: I1008 20:30:59.213673 4988 generic.go:334] "Generic (PLEG): container finished" podID="34c2181b-c702-4807-a4bf-4d5186bde6e6" containerID="0fa339bb17905f8ed14fbaab0f3303a90e2d91211ca50bac16e9b004c1893f4f" exitCode=0 Oct 08 20:30:59 crc kubenswrapper[4988]: I1008 20:30:59.214749 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-c4cz2" event={"ID":"34c2181b-c702-4807-a4bf-4d5186bde6e6","Type":"ContainerDied","Data":"0fa339bb17905f8ed14fbaab0f3303a90e2d91211ca50bac16e9b004c1893f4f"} Oct 08 20:30:59 crc kubenswrapper[4988]: I1008 20:30:59.259674 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-openstack-openstack-cell1-j9bg2"] Oct 08 20:30:59 crc kubenswrapper[4988]: I1008 20:30:59.429315 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-c4cz2" Oct 08 20:30:59 crc kubenswrapper[4988]: I1008 20:30:59.492377 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bnqtq\" (UniqueName: \"kubernetes.io/projected/34c2181b-c702-4807-a4bf-4d5186bde6e6-kube-api-access-bnqtq\") pod \"34c2181b-c702-4807-a4bf-4d5186bde6e6\" (UID: \"34c2181b-c702-4807-a4bf-4d5186bde6e6\") " Oct 08 20:30:59 crc kubenswrapper[4988]: I1008 20:30:59.492673 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/34c2181b-c702-4807-a4bf-4d5186bde6e6-catalog-content\") pod \"34c2181b-c702-4807-a4bf-4d5186bde6e6\" (UID: \"34c2181b-c702-4807-a4bf-4d5186bde6e6\") " Oct 08 20:30:59 crc kubenswrapper[4988]: I1008 20:30:59.492765 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/34c2181b-c702-4807-a4bf-4d5186bde6e6-utilities\") pod \"34c2181b-c702-4807-a4bf-4d5186bde6e6\" (UID: \"34c2181b-c702-4807-a4bf-4d5186bde6e6\") " Oct 08 20:30:59 crc kubenswrapper[4988]: I1008 20:30:59.494064 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/34c2181b-c702-4807-a4bf-4d5186bde6e6-utilities" (OuterVolumeSpecName: "utilities") pod "34c2181b-c702-4807-a4bf-4d5186bde6e6" (UID: "34c2181b-c702-4807-a4bf-4d5186bde6e6"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 20:30:59 crc kubenswrapper[4988]: I1008 20:30:59.497846 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/34c2181b-c702-4807-a4bf-4d5186bde6e6-kube-api-access-bnqtq" (OuterVolumeSpecName: "kube-api-access-bnqtq") pod "34c2181b-c702-4807-a4bf-4d5186bde6e6" (UID: "34c2181b-c702-4807-a4bf-4d5186bde6e6"). InnerVolumeSpecName "kube-api-access-bnqtq". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 20:30:59 crc kubenswrapper[4988]: I1008 20:30:59.589054 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/34c2181b-c702-4807-a4bf-4d5186bde6e6-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "34c2181b-c702-4807-a4bf-4d5186bde6e6" (UID: "34c2181b-c702-4807-a4bf-4d5186bde6e6"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 20:30:59 crc kubenswrapper[4988]: I1008 20:30:59.595665 4988 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/34c2181b-c702-4807-a4bf-4d5186bde6e6-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 08 20:30:59 crc kubenswrapper[4988]: I1008 20:30:59.595704 4988 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/34c2181b-c702-4807-a4bf-4d5186bde6e6-utilities\") on node \"crc\" DevicePath \"\"" Oct 08 20:30:59 crc kubenswrapper[4988]: I1008 20:30:59.595718 4988 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bnqtq\" (UniqueName: \"kubernetes.io/projected/34c2181b-c702-4807-a4bf-4d5186bde6e6-kube-api-access-bnqtq\") on node \"crc\" DevicePath \"\"" Oct 08 20:31:00 crc kubenswrapper[4988]: I1008 20:31:00.227512 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-openstack-openstack-cell1-j9bg2" event={"ID":"08584161-d03c-4354-9e07-60eae54b233d","Type":"ContainerStarted","Data":"1ba94e8733236f5d0012ac7e04358930fa0b77a52342b6661a7dc2130074158c"} Oct 08 20:31:00 crc kubenswrapper[4988]: I1008 20:31:00.228005 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-openstack-openstack-cell1-j9bg2" event={"ID":"08584161-d03c-4354-9e07-60eae54b233d","Type":"ContainerStarted","Data":"44a14a52e88048512dba5fb30a306fd4ea1f39ca7cd88ffe8404a1a844fc21e4"} Oct 08 20:31:00 crc kubenswrapper[4988]: I1008 20:31:00.230346 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-c4cz2" event={"ID":"34c2181b-c702-4807-a4bf-4d5186bde6e6","Type":"ContainerDied","Data":"257178f51055ea90d48e11388339577bbccecd5479eb68ee02352649f6c768fb"} Oct 08 20:31:00 crc kubenswrapper[4988]: I1008 20:31:00.230448 4988 scope.go:117] "RemoveContainer" containerID="0fa339bb17905f8ed14fbaab0f3303a90e2d91211ca50bac16e9b004c1893f4f" Oct 08 20:31:00 crc kubenswrapper[4988]: I1008 20:31:00.230747 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-c4cz2" Oct 08 20:31:00 crc kubenswrapper[4988]: I1008 20:31:00.276141 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-openstack-openstack-cell1-j9bg2" podStartSLOduration=1.746311794 podStartE2EDuration="2.276120051s" podCreationTimestamp="2025-10-08 20:30:58 +0000 UTC" firstStartedPulling="2025-10-08 20:30:59.286506708 +0000 UTC m=+8404.736349478" lastFinishedPulling="2025-10-08 20:30:59.816314965 +0000 UTC m=+8405.266157735" observedRunningTime="2025-10-08 20:31:00.251621274 +0000 UTC m=+8405.701464064" watchObservedRunningTime="2025-10-08 20:31:00.276120051 +0000 UTC m=+8405.725962831" Oct 08 20:31:00 crc kubenswrapper[4988]: I1008 20:31:00.285204 4988 scope.go:117] "RemoveContainer" containerID="d7cf9d127f684b25182152f3ae77a1c7c15ce2590b77f80b0f09898906e56362" Oct 08 20:31:00 crc kubenswrapper[4988]: I1008 20:31:00.297578 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-c4cz2"] Oct 08 20:31:00 crc kubenswrapper[4988]: I1008 20:31:00.325373 4988 scope.go:117] "RemoveContainer" containerID="86d35aa0f130f6f353b79b89ac6f0443b7e4bcbaf065bac182fccf61174e514a" Oct 08 20:31:00 crc kubenswrapper[4988]: I1008 20:31:00.335670 4988 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-c4cz2"] Oct 08 20:31:01 crc kubenswrapper[4988]: I1008 20:31:01.264934 4988 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="34c2181b-c702-4807-a4bf-4d5186bde6e6" path="/var/lib/kubelet/pods/34c2181b-c702-4807-a4bf-4d5186bde6e6/volumes" Oct 08 20:31:03 crc kubenswrapper[4988]: I1008 20:31:03.132412 4988 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-tg6lc" Oct 08 20:31:03 crc kubenswrapper[4988]: I1008 20:31:03.132767 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-tg6lc" Oct 08 20:31:03 crc kubenswrapper[4988]: I1008 20:31:03.198466 4988 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-tg6lc" Oct 08 20:31:03 crc kubenswrapper[4988]: I1008 20:31:03.320229 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-tg6lc" Oct 08 20:31:04 crc kubenswrapper[4988]: I1008 20:31:04.384378 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-tg6lc"] Oct 08 20:31:05 crc kubenswrapper[4988]: I1008 20:31:05.277663 4988 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-tg6lc" podUID="0be1269f-0a8b-48e2-a60b-98e83628f6d0" containerName="registry-server" containerID="cri-o://3d5f531ac4ee8366686db21aab2e80fab04a00fb15c25c673c71d25d56eb8478" gracePeriod=2 Oct 08 20:31:05 crc kubenswrapper[4988]: I1008 20:31:05.867897 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-tg6lc" Oct 08 20:31:05 crc kubenswrapper[4988]: I1008 20:31:05.962954 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0be1269f-0a8b-48e2-a60b-98e83628f6d0-utilities\") pod \"0be1269f-0a8b-48e2-a60b-98e83628f6d0\" (UID: \"0be1269f-0a8b-48e2-a60b-98e83628f6d0\") " Oct 08 20:31:05 crc kubenswrapper[4988]: I1008 20:31:05.963371 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r9hzc\" (UniqueName: \"kubernetes.io/projected/0be1269f-0a8b-48e2-a60b-98e83628f6d0-kube-api-access-r9hzc\") pod \"0be1269f-0a8b-48e2-a60b-98e83628f6d0\" (UID: \"0be1269f-0a8b-48e2-a60b-98e83628f6d0\") " Oct 08 20:31:05 crc kubenswrapper[4988]: I1008 20:31:05.963761 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0be1269f-0a8b-48e2-a60b-98e83628f6d0-utilities" (OuterVolumeSpecName: "utilities") pod "0be1269f-0a8b-48e2-a60b-98e83628f6d0" (UID: "0be1269f-0a8b-48e2-a60b-98e83628f6d0"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 20:31:05 crc kubenswrapper[4988]: I1008 20:31:05.963792 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0be1269f-0a8b-48e2-a60b-98e83628f6d0-catalog-content\") pod \"0be1269f-0a8b-48e2-a60b-98e83628f6d0\" (UID: \"0be1269f-0a8b-48e2-a60b-98e83628f6d0\") " Oct 08 20:31:05 crc kubenswrapper[4988]: I1008 20:31:05.964718 4988 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0be1269f-0a8b-48e2-a60b-98e83628f6d0-utilities\") on node \"crc\" DevicePath \"\"" Oct 08 20:31:05 crc kubenswrapper[4988]: I1008 20:31:05.972526 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0be1269f-0a8b-48e2-a60b-98e83628f6d0-kube-api-access-r9hzc" (OuterVolumeSpecName: "kube-api-access-r9hzc") pod "0be1269f-0a8b-48e2-a60b-98e83628f6d0" (UID: "0be1269f-0a8b-48e2-a60b-98e83628f6d0"). InnerVolumeSpecName "kube-api-access-r9hzc". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 20:31:06 crc kubenswrapper[4988]: I1008 20:31:06.053597 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0be1269f-0a8b-48e2-a60b-98e83628f6d0-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "0be1269f-0a8b-48e2-a60b-98e83628f6d0" (UID: "0be1269f-0a8b-48e2-a60b-98e83628f6d0"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 20:31:06 crc kubenswrapper[4988]: I1008 20:31:06.067818 4988 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0be1269f-0a8b-48e2-a60b-98e83628f6d0-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 08 20:31:06 crc kubenswrapper[4988]: I1008 20:31:06.067872 4988 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r9hzc\" (UniqueName: \"kubernetes.io/projected/0be1269f-0a8b-48e2-a60b-98e83628f6d0-kube-api-access-r9hzc\") on node \"crc\" DevicePath \"\"" Oct 08 20:31:06 crc kubenswrapper[4988]: I1008 20:31:06.294227 4988 generic.go:334] "Generic (PLEG): container finished" podID="0be1269f-0a8b-48e2-a60b-98e83628f6d0" containerID="3d5f531ac4ee8366686db21aab2e80fab04a00fb15c25c673c71d25d56eb8478" exitCode=0 Oct 08 20:31:06 crc kubenswrapper[4988]: I1008 20:31:06.294311 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-tg6lc" Oct 08 20:31:06 crc kubenswrapper[4988]: I1008 20:31:06.294311 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-tg6lc" event={"ID":"0be1269f-0a8b-48e2-a60b-98e83628f6d0","Type":"ContainerDied","Data":"3d5f531ac4ee8366686db21aab2e80fab04a00fb15c25c673c71d25d56eb8478"} Oct 08 20:31:06 crc kubenswrapper[4988]: I1008 20:31:06.294525 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-tg6lc" event={"ID":"0be1269f-0a8b-48e2-a60b-98e83628f6d0","Type":"ContainerDied","Data":"9e2e160c4303804248a81f6e29e08f72585c0059d50866af8312504db1a96396"} Oct 08 20:31:06 crc kubenswrapper[4988]: I1008 20:31:06.294611 4988 scope.go:117] "RemoveContainer" containerID="3d5f531ac4ee8366686db21aab2e80fab04a00fb15c25c673c71d25d56eb8478" Oct 08 20:31:06 crc kubenswrapper[4988]: I1008 20:31:06.330376 4988 scope.go:117] "RemoveContainer" containerID="20bc2b70475e6be093e53a6a34575fe27b87e6882a36507437eb2b9001c30680" Oct 08 20:31:06 crc kubenswrapper[4988]: I1008 20:31:06.363363 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-tg6lc"] Oct 08 20:31:06 crc kubenswrapper[4988]: I1008 20:31:06.371958 4988 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-tg6lc"] Oct 08 20:31:06 crc kubenswrapper[4988]: I1008 20:31:06.384467 4988 scope.go:117] "RemoveContainer" containerID="d7e9ac863efe5ea10fc02b61b17ec5a4979a7a3d42997f96b5ec798a5ed34747" Oct 08 20:31:06 crc kubenswrapper[4988]: I1008 20:31:06.447080 4988 scope.go:117] "RemoveContainer" containerID="3d5f531ac4ee8366686db21aab2e80fab04a00fb15c25c673c71d25d56eb8478" Oct 08 20:31:06 crc kubenswrapper[4988]: E1008 20:31:06.447452 4988 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3d5f531ac4ee8366686db21aab2e80fab04a00fb15c25c673c71d25d56eb8478\": container with ID starting with 3d5f531ac4ee8366686db21aab2e80fab04a00fb15c25c673c71d25d56eb8478 not found: ID does not exist" containerID="3d5f531ac4ee8366686db21aab2e80fab04a00fb15c25c673c71d25d56eb8478" Oct 08 20:31:06 crc kubenswrapper[4988]: I1008 20:31:06.447491 4988 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3d5f531ac4ee8366686db21aab2e80fab04a00fb15c25c673c71d25d56eb8478"} err="failed to get container status \"3d5f531ac4ee8366686db21aab2e80fab04a00fb15c25c673c71d25d56eb8478\": rpc error: code = NotFound desc = could not find container \"3d5f531ac4ee8366686db21aab2e80fab04a00fb15c25c673c71d25d56eb8478\": container with ID starting with 3d5f531ac4ee8366686db21aab2e80fab04a00fb15c25c673c71d25d56eb8478 not found: ID does not exist" Oct 08 20:31:06 crc kubenswrapper[4988]: I1008 20:31:06.447520 4988 scope.go:117] "RemoveContainer" containerID="20bc2b70475e6be093e53a6a34575fe27b87e6882a36507437eb2b9001c30680" Oct 08 20:31:06 crc kubenswrapper[4988]: E1008 20:31:06.447707 4988 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"20bc2b70475e6be093e53a6a34575fe27b87e6882a36507437eb2b9001c30680\": container with ID starting with 20bc2b70475e6be093e53a6a34575fe27b87e6882a36507437eb2b9001c30680 not found: ID does not exist" containerID="20bc2b70475e6be093e53a6a34575fe27b87e6882a36507437eb2b9001c30680" Oct 08 20:31:06 crc kubenswrapper[4988]: I1008 20:31:06.447733 4988 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"20bc2b70475e6be093e53a6a34575fe27b87e6882a36507437eb2b9001c30680"} err="failed to get container status \"20bc2b70475e6be093e53a6a34575fe27b87e6882a36507437eb2b9001c30680\": rpc error: code = NotFound desc = could not find container \"20bc2b70475e6be093e53a6a34575fe27b87e6882a36507437eb2b9001c30680\": container with ID starting with 20bc2b70475e6be093e53a6a34575fe27b87e6882a36507437eb2b9001c30680 not found: ID does not exist" Oct 08 20:31:06 crc kubenswrapper[4988]: I1008 20:31:06.447750 4988 scope.go:117] "RemoveContainer" containerID="d7e9ac863efe5ea10fc02b61b17ec5a4979a7a3d42997f96b5ec798a5ed34747" Oct 08 20:31:06 crc kubenswrapper[4988]: E1008 20:31:06.448022 4988 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d7e9ac863efe5ea10fc02b61b17ec5a4979a7a3d42997f96b5ec798a5ed34747\": container with ID starting with d7e9ac863efe5ea10fc02b61b17ec5a4979a7a3d42997f96b5ec798a5ed34747 not found: ID does not exist" containerID="d7e9ac863efe5ea10fc02b61b17ec5a4979a7a3d42997f96b5ec798a5ed34747" Oct 08 20:31:06 crc kubenswrapper[4988]: I1008 20:31:06.448099 4988 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d7e9ac863efe5ea10fc02b61b17ec5a4979a7a3d42997f96b5ec798a5ed34747"} err="failed to get container status \"d7e9ac863efe5ea10fc02b61b17ec5a4979a7a3d42997f96b5ec798a5ed34747\": rpc error: code = NotFound desc = could not find container \"d7e9ac863efe5ea10fc02b61b17ec5a4979a7a3d42997f96b5ec798a5ed34747\": container with ID starting with d7e9ac863efe5ea10fc02b61b17ec5a4979a7a3d42997f96b5ec798a5ed34747 not found: ID does not exist" Oct 08 20:31:07 crc kubenswrapper[4988]: I1008 20:31:07.259730 4988 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0be1269f-0a8b-48e2-a60b-98e83628f6d0" path="/var/lib/kubelet/pods/0be1269f-0a8b-48e2-a60b-98e83628f6d0/volumes" Oct 08 20:31:23 crc kubenswrapper[4988]: I1008 20:31:23.338491 4988 patch_prober.go:28] interesting pod/machine-config-daemon-nm54f container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 20:31:23 crc kubenswrapper[4988]: I1008 20:31:23.339183 4988 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 20:31:53 crc kubenswrapper[4988]: I1008 20:31:53.338728 4988 patch_prober.go:28] interesting pod/machine-config-daemon-nm54f container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 20:31:53 crc kubenswrapper[4988]: I1008 20:31:53.339378 4988 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 20:32:23 crc kubenswrapper[4988]: I1008 20:32:23.338888 4988 patch_prober.go:28] interesting pod/machine-config-daemon-nm54f container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 20:32:23 crc kubenswrapper[4988]: I1008 20:32:23.341680 4988 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 20:32:23 crc kubenswrapper[4988]: I1008 20:32:23.341836 4988 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" Oct 08 20:32:23 crc kubenswrapper[4988]: I1008 20:32:23.343068 4988 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"413fd9f07623b7fb95cefbc48df19e6c85a31a0236c333177ea0591f24a99a71"} pod="openshift-machine-config-operator/machine-config-daemon-nm54f" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 08 20:32:23 crc kubenswrapper[4988]: I1008 20:32:23.343968 4988 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" containerName="machine-config-daemon" containerID="cri-o://413fd9f07623b7fb95cefbc48df19e6c85a31a0236c333177ea0591f24a99a71" gracePeriod=600 Oct 08 20:32:23 crc kubenswrapper[4988]: E1008 20:32:23.475476 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nm54f_openshift-machine-config-operator(f328a20b-5283-407e-8a93-2ab2da599f18)\"" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" Oct 08 20:32:24 crc kubenswrapper[4988]: I1008 20:32:24.274338 4988 generic.go:334] "Generic (PLEG): container finished" podID="f328a20b-5283-407e-8a93-2ab2da599f18" containerID="413fd9f07623b7fb95cefbc48df19e6c85a31a0236c333177ea0591f24a99a71" exitCode=0 Oct 08 20:32:24 crc kubenswrapper[4988]: I1008 20:32:24.274446 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" event={"ID":"f328a20b-5283-407e-8a93-2ab2da599f18","Type":"ContainerDied","Data":"413fd9f07623b7fb95cefbc48df19e6c85a31a0236c333177ea0591f24a99a71"} Oct 08 20:32:24 crc kubenswrapper[4988]: I1008 20:32:24.274798 4988 scope.go:117] "RemoveContainer" containerID="7cbd759a42c747ea6a2391dd004be2e8c08240e1bbe09cf06de3384164b67bb4" Oct 08 20:32:24 crc kubenswrapper[4988]: I1008 20:32:24.275943 4988 scope.go:117] "RemoveContainer" containerID="413fd9f07623b7fb95cefbc48df19e6c85a31a0236c333177ea0591f24a99a71" Oct 08 20:32:24 crc kubenswrapper[4988]: E1008 20:32:24.276512 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nm54f_openshift-machine-config-operator(f328a20b-5283-407e-8a93-2ab2da599f18)\"" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" Oct 08 20:32:37 crc kubenswrapper[4988]: I1008 20:32:37.237702 4988 scope.go:117] "RemoveContainer" containerID="413fd9f07623b7fb95cefbc48df19e6c85a31a0236c333177ea0591f24a99a71" Oct 08 20:32:37 crc kubenswrapper[4988]: E1008 20:32:37.239302 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nm54f_openshift-machine-config-operator(f328a20b-5283-407e-8a93-2ab2da599f18)\"" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" Oct 08 20:32:49 crc kubenswrapper[4988]: I1008 20:32:49.238288 4988 scope.go:117] "RemoveContainer" containerID="413fd9f07623b7fb95cefbc48df19e6c85a31a0236c333177ea0591f24a99a71" Oct 08 20:32:49 crc kubenswrapper[4988]: E1008 20:32:49.239405 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nm54f_openshift-machine-config-operator(f328a20b-5283-407e-8a93-2ab2da599f18)\"" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" Oct 08 20:33:01 crc kubenswrapper[4988]: I1008 20:33:01.238620 4988 scope.go:117] "RemoveContainer" containerID="413fd9f07623b7fb95cefbc48df19e6c85a31a0236c333177ea0591f24a99a71" Oct 08 20:33:01 crc kubenswrapper[4988]: E1008 20:33:01.239773 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nm54f_openshift-machine-config-operator(f328a20b-5283-407e-8a93-2ab2da599f18)\"" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" Oct 08 20:33:13 crc kubenswrapper[4988]: I1008 20:33:13.238949 4988 scope.go:117] "RemoveContainer" containerID="413fd9f07623b7fb95cefbc48df19e6c85a31a0236c333177ea0591f24a99a71" Oct 08 20:33:13 crc kubenswrapper[4988]: E1008 20:33:13.240016 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nm54f_openshift-machine-config-operator(f328a20b-5283-407e-8a93-2ab2da599f18)\"" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" Oct 08 20:33:27 crc kubenswrapper[4988]: I1008 20:33:27.239591 4988 scope.go:117] "RemoveContainer" containerID="413fd9f07623b7fb95cefbc48df19e6c85a31a0236c333177ea0591f24a99a71" Oct 08 20:33:27 crc kubenswrapper[4988]: E1008 20:33:27.240835 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nm54f_openshift-machine-config-operator(f328a20b-5283-407e-8a93-2ab2da599f18)\"" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" Oct 08 20:33:42 crc kubenswrapper[4988]: I1008 20:33:42.239000 4988 scope.go:117] "RemoveContainer" containerID="413fd9f07623b7fb95cefbc48df19e6c85a31a0236c333177ea0591f24a99a71" Oct 08 20:33:42 crc kubenswrapper[4988]: E1008 20:33:42.240093 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nm54f_openshift-machine-config-operator(f328a20b-5283-407e-8a93-2ab2da599f18)\"" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" Oct 08 20:33:56 crc kubenswrapper[4988]: I1008 20:33:56.239565 4988 scope.go:117] "RemoveContainer" containerID="413fd9f07623b7fb95cefbc48df19e6c85a31a0236c333177ea0591f24a99a71" Oct 08 20:33:56 crc kubenswrapper[4988]: E1008 20:33:56.240519 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nm54f_openshift-machine-config-operator(f328a20b-5283-407e-8a93-2ab2da599f18)\"" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" Oct 08 20:34:11 crc kubenswrapper[4988]: I1008 20:34:11.238959 4988 scope.go:117] "RemoveContainer" containerID="413fd9f07623b7fb95cefbc48df19e6c85a31a0236c333177ea0591f24a99a71" Oct 08 20:34:11 crc kubenswrapper[4988]: E1008 20:34:11.239776 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nm54f_openshift-machine-config-operator(f328a20b-5283-407e-8a93-2ab2da599f18)\"" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" Oct 08 20:34:23 crc kubenswrapper[4988]: I1008 20:34:23.238828 4988 scope.go:117] "RemoveContainer" containerID="413fd9f07623b7fb95cefbc48df19e6c85a31a0236c333177ea0591f24a99a71" Oct 08 20:34:23 crc kubenswrapper[4988]: E1008 20:34:23.239827 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nm54f_openshift-machine-config-operator(f328a20b-5283-407e-8a93-2ab2da599f18)\"" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" Oct 08 20:34:28 crc kubenswrapper[4988]: I1008 20:34:28.494536 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-bd6p5"] Oct 08 20:34:28 crc kubenswrapper[4988]: E1008 20:34:28.496320 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0be1269f-0a8b-48e2-a60b-98e83628f6d0" containerName="extract-utilities" Oct 08 20:34:28 crc kubenswrapper[4988]: I1008 20:34:28.496351 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="0be1269f-0a8b-48e2-a60b-98e83628f6d0" containerName="extract-utilities" Oct 08 20:34:28 crc kubenswrapper[4988]: E1008 20:34:28.496529 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="34c2181b-c702-4807-a4bf-4d5186bde6e6" containerName="extract-utilities" Oct 08 20:34:28 crc kubenswrapper[4988]: I1008 20:34:28.496543 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="34c2181b-c702-4807-a4bf-4d5186bde6e6" containerName="extract-utilities" Oct 08 20:34:28 crc kubenswrapper[4988]: E1008 20:34:28.496584 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="34c2181b-c702-4807-a4bf-4d5186bde6e6" containerName="registry-server" Oct 08 20:34:28 crc kubenswrapper[4988]: I1008 20:34:28.496594 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="34c2181b-c702-4807-a4bf-4d5186bde6e6" containerName="registry-server" Oct 08 20:34:28 crc kubenswrapper[4988]: E1008 20:34:28.496621 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="34c2181b-c702-4807-a4bf-4d5186bde6e6" containerName="extract-content" Oct 08 20:34:28 crc kubenswrapper[4988]: I1008 20:34:28.496629 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="34c2181b-c702-4807-a4bf-4d5186bde6e6" containerName="extract-content" Oct 08 20:34:28 crc kubenswrapper[4988]: E1008 20:34:28.496649 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0be1269f-0a8b-48e2-a60b-98e83628f6d0" containerName="extract-content" Oct 08 20:34:28 crc kubenswrapper[4988]: I1008 20:34:28.496659 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="0be1269f-0a8b-48e2-a60b-98e83628f6d0" containerName="extract-content" Oct 08 20:34:28 crc kubenswrapper[4988]: E1008 20:34:28.496676 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0be1269f-0a8b-48e2-a60b-98e83628f6d0" containerName="registry-server" Oct 08 20:34:28 crc kubenswrapper[4988]: I1008 20:34:28.496686 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="0be1269f-0a8b-48e2-a60b-98e83628f6d0" containerName="registry-server" Oct 08 20:34:28 crc kubenswrapper[4988]: I1008 20:34:28.497011 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="0be1269f-0a8b-48e2-a60b-98e83628f6d0" containerName="registry-server" Oct 08 20:34:28 crc kubenswrapper[4988]: I1008 20:34:28.497052 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="34c2181b-c702-4807-a4bf-4d5186bde6e6" containerName="registry-server" Oct 08 20:34:28 crc kubenswrapper[4988]: I1008 20:34:28.499769 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-bd6p5" Oct 08 20:34:28 crc kubenswrapper[4988]: I1008 20:34:28.510009 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-bd6p5"] Oct 08 20:34:28 crc kubenswrapper[4988]: I1008 20:34:28.607225 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0581d264-a08f-4f5b-9bdb-a284f3ce1415-catalog-content\") pod \"redhat-marketplace-bd6p5\" (UID: \"0581d264-a08f-4f5b-9bdb-a284f3ce1415\") " pod="openshift-marketplace/redhat-marketplace-bd6p5" Oct 08 20:34:28 crc kubenswrapper[4988]: I1008 20:34:28.607638 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0581d264-a08f-4f5b-9bdb-a284f3ce1415-utilities\") pod \"redhat-marketplace-bd6p5\" (UID: \"0581d264-a08f-4f5b-9bdb-a284f3ce1415\") " pod="openshift-marketplace/redhat-marketplace-bd6p5" Oct 08 20:34:28 crc kubenswrapper[4988]: I1008 20:34:28.607700 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pf7c7\" (UniqueName: \"kubernetes.io/projected/0581d264-a08f-4f5b-9bdb-a284f3ce1415-kube-api-access-pf7c7\") pod \"redhat-marketplace-bd6p5\" (UID: \"0581d264-a08f-4f5b-9bdb-a284f3ce1415\") " pod="openshift-marketplace/redhat-marketplace-bd6p5" Oct 08 20:34:28 crc kubenswrapper[4988]: I1008 20:34:28.709017 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0581d264-a08f-4f5b-9bdb-a284f3ce1415-catalog-content\") pod \"redhat-marketplace-bd6p5\" (UID: \"0581d264-a08f-4f5b-9bdb-a284f3ce1415\") " pod="openshift-marketplace/redhat-marketplace-bd6p5" Oct 08 20:34:28 crc kubenswrapper[4988]: I1008 20:34:28.709338 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0581d264-a08f-4f5b-9bdb-a284f3ce1415-utilities\") pod \"redhat-marketplace-bd6p5\" (UID: \"0581d264-a08f-4f5b-9bdb-a284f3ce1415\") " pod="openshift-marketplace/redhat-marketplace-bd6p5" Oct 08 20:34:28 crc kubenswrapper[4988]: I1008 20:34:28.709464 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pf7c7\" (UniqueName: \"kubernetes.io/projected/0581d264-a08f-4f5b-9bdb-a284f3ce1415-kube-api-access-pf7c7\") pod \"redhat-marketplace-bd6p5\" (UID: \"0581d264-a08f-4f5b-9bdb-a284f3ce1415\") " pod="openshift-marketplace/redhat-marketplace-bd6p5" Oct 08 20:34:28 crc kubenswrapper[4988]: I1008 20:34:28.713927 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0581d264-a08f-4f5b-9bdb-a284f3ce1415-catalog-content\") pod \"redhat-marketplace-bd6p5\" (UID: \"0581d264-a08f-4f5b-9bdb-a284f3ce1415\") " pod="openshift-marketplace/redhat-marketplace-bd6p5" Oct 08 20:34:28 crc kubenswrapper[4988]: I1008 20:34:28.714058 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0581d264-a08f-4f5b-9bdb-a284f3ce1415-utilities\") pod \"redhat-marketplace-bd6p5\" (UID: \"0581d264-a08f-4f5b-9bdb-a284f3ce1415\") " pod="openshift-marketplace/redhat-marketplace-bd6p5" Oct 08 20:34:28 crc kubenswrapper[4988]: I1008 20:34:28.757555 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pf7c7\" (UniqueName: \"kubernetes.io/projected/0581d264-a08f-4f5b-9bdb-a284f3ce1415-kube-api-access-pf7c7\") pod \"redhat-marketplace-bd6p5\" (UID: \"0581d264-a08f-4f5b-9bdb-a284f3ce1415\") " pod="openshift-marketplace/redhat-marketplace-bd6p5" Oct 08 20:34:28 crc kubenswrapper[4988]: I1008 20:34:28.839040 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-bd6p5" Oct 08 20:34:29 crc kubenswrapper[4988]: I1008 20:34:29.300657 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-bd6p5"] Oct 08 20:34:29 crc kubenswrapper[4988]: W1008 20:34:29.311530 4988 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0581d264_a08f_4f5b_9bdb_a284f3ce1415.slice/crio-d71562e802d41dac1b289c625e53a5fbb451773251c85c283e9029401f5ae22f WatchSource:0}: Error finding container d71562e802d41dac1b289c625e53a5fbb451773251c85c283e9029401f5ae22f: Status 404 returned error can't find the container with id d71562e802d41dac1b289c625e53a5fbb451773251c85c283e9029401f5ae22f Oct 08 20:34:29 crc kubenswrapper[4988]: I1008 20:34:29.842487 4988 generic.go:334] "Generic (PLEG): container finished" podID="0581d264-a08f-4f5b-9bdb-a284f3ce1415" containerID="2e977a6a047941036ee8920f4f4f5dda8e918bcb10a7d711195dbe1de1799d74" exitCode=0 Oct 08 20:34:29 crc kubenswrapper[4988]: I1008 20:34:29.842608 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bd6p5" event={"ID":"0581d264-a08f-4f5b-9bdb-a284f3ce1415","Type":"ContainerDied","Data":"2e977a6a047941036ee8920f4f4f5dda8e918bcb10a7d711195dbe1de1799d74"} Oct 08 20:34:29 crc kubenswrapper[4988]: I1008 20:34:29.842796 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bd6p5" event={"ID":"0581d264-a08f-4f5b-9bdb-a284f3ce1415","Type":"ContainerStarted","Data":"d71562e802d41dac1b289c625e53a5fbb451773251c85c283e9029401f5ae22f"} Oct 08 20:34:30 crc kubenswrapper[4988]: I1008 20:34:30.853673 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bd6p5" event={"ID":"0581d264-a08f-4f5b-9bdb-a284f3ce1415","Type":"ContainerStarted","Data":"1ae794c2fc86be27705309069fbf77b041e068dd0c0c6f06e0b68299c6a52588"} Oct 08 20:34:31 crc kubenswrapper[4988]: I1008 20:34:31.868412 4988 generic.go:334] "Generic (PLEG): container finished" podID="0581d264-a08f-4f5b-9bdb-a284f3ce1415" containerID="1ae794c2fc86be27705309069fbf77b041e068dd0c0c6f06e0b68299c6a52588" exitCode=0 Oct 08 20:34:31 crc kubenswrapper[4988]: I1008 20:34:31.868516 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bd6p5" event={"ID":"0581d264-a08f-4f5b-9bdb-a284f3ce1415","Type":"ContainerDied","Data":"1ae794c2fc86be27705309069fbf77b041e068dd0c0c6f06e0b68299c6a52588"} Oct 08 20:34:32 crc kubenswrapper[4988]: I1008 20:34:32.888434 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bd6p5" event={"ID":"0581d264-a08f-4f5b-9bdb-a284f3ce1415","Type":"ContainerStarted","Data":"d211580e2fbad523443dd8d31e9dc978cb2492feaa338e767282b4c991792d6a"} Oct 08 20:34:32 crc kubenswrapper[4988]: I1008 20:34:32.928824 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-bd6p5" podStartSLOduration=2.4527460469999998 podStartE2EDuration="4.928793288s" podCreationTimestamp="2025-10-08 20:34:28 +0000 UTC" firstStartedPulling="2025-10-08 20:34:29.845497676 +0000 UTC m=+8615.295340456" lastFinishedPulling="2025-10-08 20:34:32.321544887 +0000 UTC m=+8617.771387697" observedRunningTime="2025-10-08 20:34:32.915518492 +0000 UTC m=+8618.365361312" watchObservedRunningTime="2025-10-08 20:34:32.928793288 +0000 UTC m=+8618.378636068" Oct 08 20:34:37 crc kubenswrapper[4988]: I1008 20:34:37.238958 4988 scope.go:117] "RemoveContainer" containerID="413fd9f07623b7fb95cefbc48df19e6c85a31a0236c333177ea0591f24a99a71" Oct 08 20:34:37 crc kubenswrapper[4988]: E1008 20:34:37.239728 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nm54f_openshift-machine-config-operator(f328a20b-5283-407e-8a93-2ab2da599f18)\"" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" Oct 08 20:34:38 crc kubenswrapper[4988]: I1008 20:34:38.839699 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-bd6p5" Oct 08 20:34:38 crc kubenswrapper[4988]: I1008 20:34:38.840020 4988 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-bd6p5" Oct 08 20:34:38 crc kubenswrapper[4988]: I1008 20:34:38.930191 4988 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-bd6p5" Oct 08 20:34:39 crc kubenswrapper[4988]: I1008 20:34:39.054447 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-bd6p5" Oct 08 20:34:39 crc kubenswrapper[4988]: I1008 20:34:39.191675 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-bd6p5"] Oct 08 20:34:40 crc kubenswrapper[4988]: I1008 20:34:40.992186 4988 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-bd6p5" podUID="0581d264-a08f-4f5b-9bdb-a284f3ce1415" containerName="registry-server" containerID="cri-o://d211580e2fbad523443dd8d31e9dc978cb2492feaa338e767282b4c991792d6a" gracePeriod=2 Oct 08 20:34:41 crc kubenswrapper[4988]: E1008 20:34:41.114765 4988 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0581d264_a08f_4f5b_9bdb_a284f3ce1415.slice/crio-conmon-d211580e2fbad523443dd8d31e9dc978cb2492feaa338e767282b4c991792d6a.scope\": RecentStats: unable to find data in memory cache]" Oct 08 20:34:41 crc kubenswrapper[4988]: I1008 20:34:41.529932 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-bd6p5" Oct 08 20:34:41 crc kubenswrapper[4988]: I1008 20:34:41.651669 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0581d264-a08f-4f5b-9bdb-a284f3ce1415-utilities\") pod \"0581d264-a08f-4f5b-9bdb-a284f3ce1415\" (UID: \"0581d264-a08f-4f5b-9bdb-a284f3ce1415\") " Oct 08 20:34:41 crc kubenswrapper[4988]: I1008 20:34:41.652089 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pf7c7\" (UniqueName: \"kubernetes.io/projected/0581d264-a08f-4f5b-9bdb-a284f3ce1415-kube-api-access-pf7c7\") pod \"0581d264-a08f-4f5b-9bdb-a284f3ce1415\" (UID: \"0581d264-a08f-4f5b-9bdb-a284f3ce1415\") " Oct 08 20:34:41 crc kubenswrapper[4988]: I1008 20:34:41.654659 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0581d264-a08f-4f5b-9bdb-a284f3ce1415-catalog-content\") pod \"0581d264-a08f-4f5b-9bdb-a284f3ce1415\" (UID: \"0581d264-a08f-4f5b-9bdb-a284f3ce1415\") " Oct 08 20:34:41 crc kubenswrapper[4988]: I1008 20:34:41.655145 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0581d264-a08f-4f5b-9bdb-a284f3ce1415-utilities" (OuterVolumeSpecName: "utilities") pod "0581d264-a08f-4f5b-9bdb-a284f3ce1415" (UID: "0581d264-a08f-4f5b-9bdb-a284f3ce1415"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 20:34:41 crc kubenswrapper[4988]: I1008 20:34:41.655782 4988 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0581d264-a08f-4f5b-9bdb-a284f3ce1415-utilities\") on node \"crc\" DevicePath \"\"" Oct 08 20:34:41 crc kubenswrapper[4988]: I1008 20:34:41.659870 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0581d264-a08f-4f5b-9bdb-a284f3ce1415-kube-api-access-pf7c7" (OuterVolumeSpecName: "kube-api-access-pf7c7") pod "0581d264-a08f-4f5b-9bdb-a284f3ce1415" (UID: "0581d264-a08f-4f5b-9bdb-a284f3ce1415"). InnerVolumeSpecName "kube-api-access-pf7c7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 20:34:41 crc kubenswrapper[4988]: I1008 20:34:41.681249 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0581d264-a08f-4f5b-9bdb-a284f3ce1415-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "0581d264-a08f-4f5b-9bdb-a284f3ce1415" (UID: "0581d264-a08f-4f5b-9bdb-a284f3ce1415"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 20:34:41 crc kubenswrapper[4988]: I1008 20:34:41.757722 4988 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pf7c7\" (UniqueName: \"kubernetes.io/projected/0581d264-a08f-4f5b-9bdb-a284f3ce1415-kube-api-access-pf7c7\") on node \"crc\" DevicePath \"\"" Oct 08 20:34:41 crc kubenswrapper[4988]: I1008 20:34:41.757771 4988 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0581d264-a08f-4f5b-9bdb-a284f3ce1415-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 08 20:34:42 crc kubenswrapper[4988]: I1008 20:34:42.013455 4988 generic.go:334] "Generic (PLEG): container finished" podID="0581d264-a08f-4f5b-9bdb-a284f3ce1415" containerID="d211580e2fbad523443dd8d31e9dc978cb2492feaa338e767282b4c991792d6a" exitCode=0 Oct 08 20:34:42 crc kubenswrapper[4988]: I1008 20:34:42.013522 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bd6p5" event={"ID":"0581d264-a08f-4f5b-9bdb-a284f3ce1415","Type":"ContainerDied","Data":"d211580e2fbad523443dd8d31e9dc978cb2492feaa338e767282b4c991792d6a"} Oct 08 20:34:42 crc kubenswrapper[4988]: I1008 20:34:42.013566 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bd6p5" event={"ID":"0581d264-a08f-4f5b-9bdb-a284f3ce1415","Type":"ContainerDied","Data":"d71562e802d41dac1b289c625e53a5fbb451773251c85c283e9029401f5ae22f"} Oct 08 20:34:42 crc kubenswrapper[4988]: I1008 20:34:42.013597 4988 scope.go:117] "RemoveContainer" containerID="d211580e2fbad523443dd8d31e9dc978cb2492feaa338e767282b4c991792d6a" Oct 08 20:34:42 crc kubenswrapper[4988]: I1008 20:34:42.013590 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-bd6p5" Oct 08 20:34:42 crc kubenswrapper[4988]: I1008 20:34:42.052936 4988 scope.go:117] "RemoveContainer" containerID="1ae794c2fc86be27705309069fbf77b041e068dd0c0c6f06e0b68299c6a52588" Oct 08 20:34:42 crc kubenswrapper[4988]: I1008 20:34:42.069840 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-bd6p5"] Oct 08 20:34:42 crc kubenswrapper[4988]: I1008 20:34:42.080892 4988 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-bd6p5"] Oct 08 20:34:42 crc kubenswrapper[4988]: I1008 20:34:42.101318 4988 scope.go:117] "RemoveContainer" containerID="2e977a6a047941036ee8920f4f4f5dda8e918bcb10a7d711195dbe1de1799d74" Oct 08 20:34:42 crc kubenswrapper[4988]: I1008 20:34:42.153731 4988 scope.go:117] "RemoveContainer" containerID="d211580e2fbad523443dd8d31e9dc978cb2492feaa338e767282b4c991792d6a" Oct 08 20:34:42 crc kubenswrapper[4988]: E1008 20:34:42.154571 4988 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d211580e2fbad523443dd8d31e9dc978cb2492feaa338e767282b4c991792d6a\": container with ID starting with d211580e2fbad523443dd8d31e9dc978cb2492feaa338e767282b4c991792d6a not found: ID does not exist" containerID="d211580e2fbad523443dd8d31e9dc978cb2492feaa338e767282b4c991792d6a" Oct 08 20:34:42 crc kubenswrapper[4988]: I1008 20:34:42.154814 4988 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d211580e2fbad523443dd8d31e9dc978cb2492feaa338e767282b4c991792d6a"} err="failed to get container status \"d211580e2fbad523443dd8d31e9dc978cb2492feaa338e767282b4c991792d6a\": rpc error: code = NotFound desc = could not find container \"d211580e2fbad523443dd8d31e9dc978cb2492feaa338e767282b4c991792d6a\": container with ID starting with d211580e2fbad523443dd8d31e9dc978cb2492feaa338e767282b4c991792d6a not found: ID does not exist" Oct 08 20:34:42 crc kubenswrapper[4988]: I1008 20:34:42.154973 4988 scope.go:117] "RemoveContainer" containerID="1ae794c2fc86be27705309069fbf77b041e068dd0c0c6f06e0b68299c6a52588" Oct 08 20:34:42 crc kubenswrapper[4988]: E1008 20:34:42.155667 4988 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1ae794c2fc86be27705309069fbf77b041e068dd0c0c6f06e0b68299c6a52588\": container with ID starting with 1ae794c2fc86be27705309069fbf77b041e068dd0c0c6f06e0b68299c6a52588 not found: ID does not exist" containerID="1ae794c2fc86be27705309069fbf77b041e068dd0c0c6f06e0b68299c6a52588" Oct 08 20:34:42 crc kubenswrapper[4988]: I1008 20:34:42.155832 4988 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1ae794c2fc86be27705309069fbf77b041e068dd0c0c6f06e0b68299c6a52588"} err="failed to get container status \"1ae794c2fc86be27705309069fbf77b041e068dd0c0c6f06e0b68299c6a52588\": rpc error: code = NotFound desc = could not find container \"1ae794c2fc86be27705309069fbf77b041e068dd0c0c6f06e0b68299c6a52588\": container with ID starting with 1ae794c2fc86be27705309069fbf77b041e068dd0c0c6f06e0b68299c6a52588 not found: ID does not exist" Oct 08 20:34:42 crc kubenswrapper[4988]: I1008 20:34:42.155987 4988 scope.go:117] "RemoveContainer" containerID="2e977a6a047941036ee8920f4f4f5dda8e918bcb10a7d711195dbe1de1799d74" Oct 08 20:34:42 crc kubenswrapper[4988]: E1008 20:34:42.156682 4988 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2e977a6a047941036ee8920f4f4f5dda8e918bcb10a7d711195dbe1de1799d74\": container with ID starting with 2e977a6a047941036ee8920f4f4f5dda8e918bcb10a7d711195dbe1de1799d74 not found: ID does not exist" containerID="2e977a6a047941036ee8920f4f4f5dda8e918bcb10a7d711195dbe1de1799d74" Oct 08 20:34:42 crc kubenswrapper[4988]: I1008 20:34:42.156862 4988 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2e977a6a047941036ee8920f4f4f5dda8e918bcb10a7d711195dbe1de1799d74"} err="failed to get container status \"2e977a6a047941036ee8920f4f4f5dda8e918bcb10a7d711195dbe1de1799d74\": rpc error: code = NotFound desc = could not find container \"2e977a6a047941036ee8920f4f4f5dda8e918bcb10a7d711195dbe1de1799d74\": container with ID starting with 2e977a6a047941036ee8920f4f4f5dda8e918bcb10a7d711195dbe1de1799d74 not found: ID does not exist" Oct 08 20:34:43 crc kubenswrapper[4988]: I1008 20:34:43.263557 4988 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0581d264-a08f-4f5b-9bdb-a284f3ce1415" path="/var/lib/kubelet/pods/0581d264-a08f-4f5b-9bdb-a284f3ce1415/volumes" Oct 08 20:34:49 crc kubenswrapper[4988]: I1008 20:34:49.114244 4988 generic.go:334] "Generic (PLEG): container finished" podID="08584161-d03c-4354-9e07-60eae54b233d" containerID="1ba94e8733236f5d0012ac7e04358930fa0b77a52342b6661a7dc2130074158c" exitCode=0 Oct 08 20:34:49 crc kubenswrapper[4988]: I1008 20:34:49.114407 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-openstack-openstack-cell1-j9bg2" event={"ID":"08584161-d03c-4354-9e07-60eae54b233d","Type":"ContainerDied","Data":"1ba94e8733236f5d0012ac7e04358930fa0b77a52342b6661a7dc2130074158c"} Oct 08 20:34:49 crc kubenswrapper[4988]: I1008 20:34:49.238632 4988 scope.go:117] "RemoveContainer" containerID="413fd9f07623b7fb95cefbc48df19e6c85a31a0236c333177ea0591f24a99a71" Oct 08 20:34:49 crc kubenswrapper[4988]: E1008 20:34:49.239038 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nm54f_openshift-machine-config-operator(f328a20b-5283-407e-8a93-2ab2da599f18)\"" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" Oct 08 20:34:50 crc kubenswrapper[4988]: I1008 20:34:50.751910 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-openstack-openstack-cell1-j9bg2" Oct 08 20:34:50 crc kubenswrapper[4988]: I1008 20:34:50.889630 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/08584161-d03c-4354-9e07-60eae54b233d-nova-cell1-compute-config-0\") pod \"08584161-d03c-4354-9e07-60eae54b233d\" (UID: \"08584161-d03c-4354-9e07-60eae54b233d\") " Oct 08 20:34:50 crc kubenswrapper[4988]: I1008 20:34:50.889853 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/08584161-d03c-4354-9e07-60eae54b233d-nova-cell1-combined-ca-bundle\") pod \"08584161-d03c-4354-9e07-60eae54b233d\" (UID: \"08584161-d03c-4354-9e07-60eae54b233d\") " Oct 08 20:34:50 crc kubenswrapper[4988]: I1008 20:34:50.889903 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/08584161-d03c-4354-9e07-60eae54b233d-inventory\") pod \"08584161-d03c-4354-9e07-60eae54b233d\" (UID: \"08584161-d03c-4354-9e07-60eae54b233d\") " Oct 08 20:34:50 crc kubenswrapper[4988]: I1008 20:34:50.889954 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/08584161-d03c-4354-9e07-60eae54b233d-ssh-key\") pod \"08584161-d03c-4354-9e07-60eae54b233d\" (UID: \"08584161-d03c-4354-9e07-60eae54b233d\") " Oct 08 20:34:50 crc kubenswrapper[4988]: I1008 20:34:50.890021 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8j8gm\" (UniqueName: \"kubernetes.io/projected/08584161-d03c-4354-9e07-60eae54b233d-kube-api-access-8j8gm\") pod \"08584161-d03c-4354-9e07-60eae54b233d\" (UID: \"08584161-d03c-4354-9e07-60eae54b233d\") " Oct 08 20:34:50 crc kubenswrapper[4988]: I1008 20:34:50.890067 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/08584161-d03c-4354-9e07-60eae54b233d-nova-migration-ssh-key-1\") pod \"08584161-d03c-4354-9e07-60eae54b233d\" (UID: \"08584161-d03c-4354-9e07-60eae54b233d\") " Oct 08 20:34:50 crc kubenswrapper[4988]: I1008 20:34:50.890096 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/08584161-d03c-4354-9e07-60eae54b233d-nova-cell1-compute-config-1\") pod \"08584161-d03c-4354-9e07-60eae54b233d\" (UID: \"08584161-d03c-4354-9e07-60eae54b233d\") " Oct 08 20:34:50 crc kubenswrapper[4988]: I1008 20:34:50.890191 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cells-global-config-0\" (UniqueName: \"kubernetes.io/configmap/08584161-d03c-4354-9e07-60eae54b233d-nova-cells-global-config-0\") pod \"08584161-d03c-4354-9e07-60eae54b233d\" (UID: \"08584161-d03c-4354-9e07-60eae54b233d\") " Oct 08 20:34:50 crc kubenswrapper[4988]: I1008 20:34:50.890255 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/08584161-d03c-4354-9e07-60eae54b233d-nova-migration-ssh-key-0\") pod \"08584161-d03c-4354-9e07-60eae54b233d\" (UID: \"08584161-d03c-4354-9e07-60eae54b233d\") " Oct 08 20:34:50 crc kubenswrapper[4988]: I1008 20:34:50.895886 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/08584161-d03c-4354-9e07-60eae54b233d-nova-cell1-combined-ca-bundle" (OuterVolumeSpecName: "nova-cell1-combined-ca-bundle") pod "08584161-d03c-4354-9e07-60eae54b233d" (UID: "08584161-d03c-4354-9e07-60eae54b233d"). InnerVolumeSpecName "nova-cell1-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 20:34:50 crc kubenswrapper[4988]: I1008 20:34:50.899682 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/08584161-d03c-4354-9e07-60eae54b233d-kube-api-access-8j8gm" (OuterVolumeSpecName: "kube-api-access-8j8gm") pod "08584161-d03c-4354-9e07-60eae54b233d" (UID: "08584161-d03c-4354-9e07-60eae54b233d"). InnerVolumeSpecName "kube-api-access-8j8gm". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 20:34:50 crc kubenswrapper[4988]: I1008 20:34:50.923321 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/08584161-d03c-4354-9e07-60eae54b233d-nova-cells-global-config-0" (OuterVolumeSpecName: "nova-cells-global-config-0") pod "08584161-d03c-4354-9e07-60eae54b233d" (UID: "08584161-d03c-4354-9e07-60eae54b233d"). InnerVolumeSpecName "nova-cells-global-config-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 20:34:50 crc kubenswrapper[4988]: I1008 20:34:50.923545 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/08584161-d03c-4354-9e07-60eae54b233d-nova-migration-ssh-key-0" (OuterVolumeSpecName: "nova-migration-ssh-key-0") pod "08584161-d03c-4354-9e07-60eae54b233d" (UID: "08584161-d03c-4354-9e07-60eae54b233d"). InnerVolumeSpecName "nova-migration-ssh-key-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 20:34:50 crc kubenswrapper[4988]: I1008 20:34:50.925833 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/08584161-d03c-4354-9e07-60eae54b233d-inventory" (OuterVolumeSpecName: "inventory") pod "08584161-d03c-4354-9e07-60eae54b233d" (UID: "08584161-d03c-4354-9e07-60eae54b233d"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 20:34:50 crc kubenswrapper[4988]: I1008 20:34:50.928853 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/08584161-d03c-4354-9e07-60eae54b233d-nova-migration-ssh-key-1" (OuterVolumeSpecName: "nova-migration-ssh-key-1") pod "08584161-d03c-4354-9e07-60eae54b233d" (UID: "08584161-d03c-4354-9e07-60eae54b233d"). InnerVolumeSpecName "nova-migration-ssh-key-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 20:34:50 crc kubenswrapper[4988]: I1008 20:34:50.932024 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/08584161-d03c-4354-9e07-60eae54b233d-nova-cell1-compute-config-1" (OuterVolumeSpecName: "nova-cell1-compute-config-1") pod "08584161-d03c-4354-9e07-60eae54b233d" (UID: "08584161-d03c-4354-9e07-60eae54b233d"). InnerVolumeSpecName "nova-cell1-compute-config-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 20:34:50 crc kubenswrapper[4988]: I1008 20:34:50.934724 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/08584161-d03c-4354-9e07-60eae54b233d-nova-cell1-compute-config-0" (OuterVolumeSpecName: "nova-cell1-compute-config-0") pod "08584161-d03c-4354-9e07-60eae54b233d" (UID: "08584161-d03c-4354-9e07-60eae54b233d"). InnerVolumeSpecName "nova-cell1-compute-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 20:34:50 crc kubenswrapper[4988]: I1008 20:34:50.942779 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/08584161-d03c-4354-9e07-60eae54b233d-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "08584161-d03c-4354-9e07-60eae54b233d" (UID: "08584161-d03c-4354-9e07-60eae54b233d"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 20:34:50 crc kubenswrapper[4988]: I1008 20:34:50.993090 4988 reconciler_common.go:293] "Volume detached for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/08584161-d03c-4354-9e07-60eae54b233d-nova-migration-ssh-key-1\") on node \"crc\" DevicePath \"\"" Oct 08 20:34:50 crc kubenswrapper[4988]: I1008 20:34:50.993151 4988 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/08584161-d03c-4354-9e07-60eae54b233d-nova-cell1-compute-config-1\") on node \"crc\" DevicePath \"\"" Oct 08 20:34:50 crc kubenswrapper[4988]: I1008 20:34:50.993163 4988 reconciler_common.go:293] "Volume detached for volume \"nova-cells-global-config-0\" (UniqueName: \"kubernetes.io/configmap/08584161-d03c-4354-9e07-60eae54b233d-nova-cells-global-config-0\") on node \"crc\" DevicePath \"\"" Oct 08 20:34:50 crc kubenswrapper[4988]: I1008 20:34:50.993172 4988 reconciler_common.go:293] "Volume detached for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/08584161-d03c-4354-9e07-60eae54b233d-nova-migration-ssh-key-0\") on node \"crc\" DevicePath \"\"" Oct 08 20:34:50 crc kubenswrapper[4988]: I1008 20:34:50.993205 4988 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/08584161-d03c-4354-9e07-60eae54b233d-nova-cell1-compute-config-0\") on node \"crc\" DevicePath \"\"" Oct 08 20:34:50 crc kubenswrapper[4988]: I1008 20:34:50.993214 4988 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/08584161-d03c-4354-9e07-60eae54b233d-nova-cell1-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 20:34:50 crc kubenswrapper[4988]: I1008 20:34:50.993223 4988 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/08584161-d03c-4354-9e07-60eae54b233d-inventory\") on node \"crc\" DevicePath \"\"" Oct 08 20:34:50 crc kubenswrapper[4988]: I1008 20:34:50.993231 4988 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/08584161-d03c-4354-9e07-60eae54b233d-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 08 20:34:50 crc kubenswrapper[4988]: I1008 20:34:50.993239 4988 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8j8gm\" (UniqueName: \"kubernetes.io/projected/08584161-d03c-4354-9e07-60eae54b233d-kube-api-access-8j8gm\") on node \"crc\" DevicePath \"\"" Oct 08 20:34:51 crc kubenswrapper[4988]: I1008 20:34:51.142842 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-openstack-openstack-cell1-j9bg2" event={"ID":"08584161-d03c-4354-9e07-60eae54b233d","Type":"ContainerDied","Data":"44a14a52e88048512dba5fb30a306fd4ea1f39ca7cd88ffe8404a1a844fc21e4"} Oct 08 20:34:51 crc kubenswrapper[4988]: I1008 20:34:51.142904 4988 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="44a14a52e88048512dba5fb30a306fd4ea1f39ca7cd88ffe8404a1a844fc21e4" Oct 08 20:34:51 crc kubenswrapper[4988]: I1008 20:34:51.142926 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-openstack-openstack-cell1-j9bg2" Oct 08 20:34:51 crc kubenswrapper[4988]: I1008 20:34:51.252536 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/telemetry-openstack-openstack-cell1-qmvh4"] Oct 08 20:34:51 crc kubenswrapper[4988]: E1008 20:34:51.252934 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0581d264-a08f-4f5b-9bdb-a284f3ce1415" containerName="extract-utilities" Oct 08 20:34:51 crc kubenswrapper[4988]: I1008 20:34:51.252950 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="0581d264-a08f-4f5b-9bdb-a284f3ce1415" containerName="extract-utilities" Oct 08 20:34:51 crc kubenswrapper[4988]: E1008 20:34:51.252965 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0581d264-a08f-4f5b-9bdb-a284f3ce1415" containerName="extract-content" Oct 08 20:34:51 crc kubenswrapper[4988]: I1008 20:34:51.252973 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="0581d264-a08f-4f5b-9bdb-a284f3ce1415" containerName="extract-content" Oct 08 20:34:51 crc kubenswrapper[4988]: E1008 20:34:51.252992 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0581d264-a08f-4f5b-9bdb-a284f3ce1415" containerName="registry-server" Oct 08 20:34:51 crc kubenswrapper[4988]: I1008 20:34:51.253002 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="0581d264-a08f-4f5b-9bdb-a284f3ce1415" containerName="registry-server" Oct 08 20:34:51 crc kubenswrapper[4988]: E1008 20:34:51.253035 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="08584161-d03c-4354-9e07-60eae54b233d" containerName="nova-cell1-openstack-openstack-cell1" Oct 08 20:34:51 crc kubenswrapper[4988]: I1008 20:34:51.253042 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="08584161-d03c-4354-9e07-60eae54b233d" containerName="nova-cell1-openstack-openstack-cell1" Oct 08 20:34:51 crc kubenswrapper[4988]: I1008 20:34:51.253316 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="0581d264-a08f-4f5b-9bdb-a284f3ce1415" containerName="registry-server" Oct 08 20:34:51 crc kubenswrapper[4988]: I1008 20:34:51.253349 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="08584161-d03c-4354-9e07-60eae54b233d" containerName="nova-cell1-openstack-openstack-cell1" Oct 08 20:34:51 crc kubenswrapper[4988]: I1008 20:34:51.254270 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-openstack-openstack-cell1-qmvh4" Oct 08 20:34:51 crc kubenswrapper[4988]: I1008 20:34:51.258293 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Oct 08 20:34:51 crc kubenswrapper[4988]: I1008 20:34:51.258347 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-4sdzr" Oct 08 20:34:51 crc kubenswrapper[4988]: I1008 20:34:51.258418 4988 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 08 20:34:51 crc kubenswrapper[4988]: I1008 20:34:51.258638 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-compute-config-data" Oct 08 20:34:51 crc kubenswrapper[4988]: I1008 20:34:51.258753 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Oct 08 20:34:51 crc kubenswrapper[4988]: I1008 20:34:51.267722 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/telemetry-openstack-openstack-cell1-qmvh4"] Oct 08 20:34:51 crc kubenswrapper[4988]: I1008 20:34:51.402227 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/54d6a0c7-be73-43b6-b857-a621d1a1c794-telemetry-combined-ca-bundle\") pod \"telemetry-openstack-openstack-cell1-qmvh4\" (UID: \"54d6a0c7-be73-43b6-b857-a621d1a1c794\") " pod="openstack/telemetry-openstack-openstack-cell1-qmvh4" Oct 08 20:34:51 crc kubenswrapper[4988]: I1008 20:34:51.402289 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9f67j\" (UniqueName: \"kubernetes.io/projected/54d6a0c7-be73-43b6-b857-a621d1a1c794-kube-api-access-9f67j\") pod \"telemetry-openstack-openstack-cell1-qmvh4\" (UID: \"54d6a0c7-be73-43b6-b857-a621d1a1c794\") " pod="openstack/telemetry-openstack-openstack-cell1-qmvh4" Oct 08 20:34:51 crc kubenswrapper[4988]: I1008 20:34:51.402311 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/54d6a0c7-be73-43b6-b857-a621d1a1c794-ceilometer-compute-config-data-2\") pod \"telemetry-openstack-openstack-cell1-qmvh4\" (UID: \"54d6a0c7-be73-43b6-b857-a621d1a1c794\") " pod="openstack/telemetry-openstack-openstack-cell1-qmvh4" Oct 08 20:34:51 crc kubenswrapper[4988]: I1008 20:34:51.402353 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/54d6a0c7-be73-43b6-b857-a621d1a1c794-inventory\") pod \"telemetry-openstack-openstack-cell1-qmvh4\" (UID: \"54d6a0c7-be73-43b6-b857-a621d1a1c794\") " pod="openstack/telemetry-openstack-openstack-cell1-qmvh4" Oct 08 20:34:51 crc kubenswrapper[4988]: I1008 20:34:51.402470 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/54d6a0c7-be73-43b6-b857-a621d1a1c794-ceilometer-compute-config-data-1\") pod \"telemetry-openstack-openstack-cell1-qmvh4\" (UID: \"54d6a0c7-be73-43b6-b857-a621d1a1c794\") " pod="openstack/telemetry-openstack-openstack-cell1-qmvh4" Oct 08 20:34:51 crc kubenswrapper[4988]: I1008 20:34:51.402523 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/54d6a0c7-be73-43b6-b857-a621d1a1c794-ceilometer-compute-config-data-0\") pod \"telemetry-openstack-openstack-cell1-qmvh4\" (UID: \"54d6a0c7-be73-43b6-b857-a621d1a1c794\") " pod="openstack/telemetry-openstack-openstack-cell1-qmvh4" Oct 08 20:34:51 crc kubenswrapper[4988]: I1008 20:34:51.402546 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/54d6a0c7-be73-43b6-b857-a621d1a1c794-ssh-key\") pod \"telemetry-openstack-openstack-cell1-qmvh4\" (UID: \"54d6a0c7-be73-43b6-b857-a621d1a1c794\") " pod="openstack/telemetry-openstack-openstack-cell1-qmvh4" Oct 08 20:34:51 crc kubenswrapper[4988]: I1008 20:34:51.504456 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/54d6a0c7-be73-43b6-b857-a621d1a1c794-ceilometer-compute-config-data-1\") pod \"telemetry-openstack-openstack-cell1-qmvh4\" (UID: \"54d6a0c7-be73-43b6-b857-a621d1a1c794\") " pod="openstack/telemetry-openstack-openstack-cell1-qmvh4" Oct 08 20:34:51 crc kubenswrapper[4988]: I1008 20:34:51.504786 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/54d6a0c7-be73-43b6-b857-a621d1a1c794-ceilometer-compute-config-data-0\") pod \"telemetry-openstack-openstack-cell1-qmvh4\" (UID: \"54d6a0c7-be73-43b6-b857-a621d1a1c794\") " pod="openstack/telemetry-openstack-openstack-cell1-qmvh4" Oct 08 20:34:51 crc kubenswrapper[4988]: I1008 20:34:51.504812 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/54d6a0c7-be73-43b6-b857-a621d1a1c794-ssh-key\") pod \"telemetry-openstack-openstack-cell1-qmvh4\" (UID: \"54d6a0c7-be73-43b6-b857-a621d1a1c794\") " pod="openstack/telemetry-openstack-openstack-cell1-qmvh4" Oct 08 20:34:51 crc kubenswrapper[4988]: I1008 20:34:51.504881 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/54d6a0c7-be73-43b6-b857-a621d1a1c794-telemetry-combined-ca-bundle\") pod \"telemetry-openstack-openstack-cell1-qmvh4\" (UID: \"54d6a0c7-be73-43b6-b857-a621d1a1c794\") " pod="openstack/telemetry-openstack-openstack-cell1-qmvh4" Oct 08 20:34:51 crc kubenswrapper[4988]: I1008 20:34:51.504920 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9f67j\" (UniqueName: \"kubernetes.io/projected/54d6a0c7-be73-43b6-b857-a621d1a1c794-kube-api-access-9f67j\") pod \"telemetry-openstack-openstack-cell1-qmvh4\" (UID: \"54d6a0c7-be73-43b6-b857-a621d1a1c794\") " pod="openstack/telemetry-openstack-openstack-cell1-qmvh4" Oct 08 20:34:51 crc kubenswrapper[4988]: I1008 20:34:51.504940 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/54d6a0c7-be73-43b6-b857-a621d1a1c794-ceilometer-compute-config-data-2\") pod \"telemetry-openstack-openstack-cell1-qmvh4\" (UID: \"54d6a0c7-be73-43b6-b857-a621d1a1c794\") " pod="openstack/telemetry-openstack-openstack-cell1-qmvh4" Oct 08 20:34:51 crc kubenswrapper[4988]: I1008 20:34:51.504975 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/54d6a0c7-be73-43b6-b857-a621d1a1c794-inventory\") pod \"telemetry-openstack-openstack-cell1-qmvh4\" (UID: \"54d6a0c7-be73-43b6-b857-a621d1a1c794\") " pod="openstack/telemetry-openstack-openstack-cell1-qmvh4" Oct 08 20:34:51 crc kubenswrapper[4988]: I1008 20:34:51.509175 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/54d6a0c7-be73-43b6-b857-a621d1a1c794-telemetry-combined-ca-bundle\") pod \"telemetry-openstack-openstack-cell1-qmvh4\" (UID: \"54d6a0c7-be73-43b6-b857-a621d1a1c794\") " pod="openstack/telemetry-openstack-openstack-cell1-qmvh4" Oct 08 20:34:51 crc kubenswrapper[4988]: I1008 20:34:51.509325 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/54d6a0c7-be73-43b6-b857-a621d1a1c794-ssh-key\") pod \"telemetry-openstack-openstack-cell1-qmvh4\" (UID: \"54d6a0c7-be73-43b6-b857-a621d1a1c794\") " pod="openstack/telemetry-openstack-openstack-cell1-qmvh4" Oct 08 20:34:51 crc kubenswrapper[4988]: I1008 20:34:51.510316 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/54d6a0c7-be73-43b6-b857-a621d1a1c794-inventory\") pod \"telemetry-openstack-openstack-cell1-qmvh4\" (UID: \"54d6a0c7-be73-43b6-b857-a621d1a1c794\") " pod="openstack/telemetry-openstack-openstack-cell1-qmvh4" Oct 08 20:34:51 crc kubenswrapper[4988]: I1008 20:34:51.510761 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/54d6a0c7-be73-43b6-b857-a621d1a1c794-ceilometer-compute-config-data-0\") pod \"telemetry-openstack-openstack-cell1-qmvh4\" (UID: \"54d6a0c7-be73-43b6-b857-a621d1a1c794\") " pod="openstack/telemetry-openstack-openstack-cell1-qmvh4" Oct 08 20:34:51 crc kubenswrapper[4988]: I1008 20:34:51.512913 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/54d6a0c7-be73-43b6-b857-a621d1a1c794-ceilometer-compute-config-data-1\") pod \"telemetry-openstack-openstack-cell1-qmvh4\" (UID: \"54d6a0c7-be73-43b6-b857-a621d1a1c794\") " pod="openstack/telemetry-openstack-openstack-cell1-qmvh4" Oct 08 20:34:51 crc kubenswrapper[4988]: I1008 20:34:51.513149 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/54d6a0c7-be73-43b6-b857-a621d1a1c794-ceilometer-compute-config-data-2\") pod \"telemetry-openstack-openstack-cell1-qmvh4\" (UID: \"54d6a0c7-be73-43b6-b857-a621d1a1c794\") " pod="openstack/telemetry-openstack-openstack-cell1-qmvh4" Oct 08 20:34:51 crc kubenswrapper[4988]: I1008 20:34:51.524542 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9f67j\" (UniqueName: \"kubernetes.io/projected/54d6a0c7-be73-43b6-b857-a621d1a1c794-kube-api-access-9f67j\") pod \"telemetry-openstack-openstack-cell1-qmvh4\" (UID: \"54d6a0c7-be73-43b6-b857-a621d1a1c794\") " pod="openstack/telemetry-openstack-openstack-cell1-qmvh4" Oct 08 20:34:51 crc kubenswrapper[4988]: I1008 20:34:51.598121 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-openstack-openstack-cell1-qmvh4" Oct 08 20:34:52 crc kubenswrapper[4988]: I1008 20:34:52.216565 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/telemetry-openstack-openstack-cell1-qmvh4"] Oct 08 20:34:53 crc kubenswrapper[4988]: I1008 20:34:53.161343 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-openstack-openstack-cell1-qmvh4" event={"ID":"54d6a0c7-be73-43b6-b857-a621d1a1c794","Type":"ContainerStarted","Data":"205da578d0eea36af26a70a87732f6999199f75a56061bad3f558de41d4411b5"} Oct 08 20:34:53 crc kubenswrapper[4988]: I1008 20:34:53.161876 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-openstack-openstack-cell1-qmvh4" event={"ID":"54d6a0c7-be73-43b6-b857-a621d1a1c794","Type":"ContainerStarted","Data":"926d10a8119d1c781b7f22978def0ec33ce02b3ef470c6a8ec791dded140558c"} Oct 08 20:34:53 crc kubenswrapper[4988]: I1008 20:34:53.186550 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/telemetry-openstack-openstack-cell1-qmvh4" podStartSLOduration=1.5810456309999998 podStartE2EDuration="2.186526237s" podCreationTimestamp="2025-10-08 20:34:51 +0000 UTC" firstStartedPulling="2025-10-08 20:34:52.221727551 +0000 UTC m=+8637.671570321" lastFinishedPulling="2025-10-08 20:34:52.827208157 +0000 UTC m=+8638.277050927" observedRunningTime="2025-10-08 20:34:53.180343454 +0000 UTC m=+8638.630186244" watchObservedRunningTime="2025-10-08 20:34:53.186526237 +0000 UTC m=+8638.636369047" Oct 08 20:35:01 crc kubenswrapper[4988]: I1008 20:35:01.238947 4988 scope.go:117] "RemoveContainer" containerID="413fd9f07623b7fb95cefbc48df19e6c85a31a0236c333177ea0591f24a99a71" Oct 08 20:35:01 crc kubenswrapper[4988]: E1008 20:35:01.239995 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nm54f_openshift-machine-config-operator(f328a20b-5283-407e-8a93-2ab2da599f18)\"" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" Oct 08 20:35:13 crc kubenswrapper[4988]: I1008 20:35:13.238527 4988 scope.go:117] "RemoveContainer" containerID="413fd9f07623b7fb95cefbc48df19e6c85a31a0236c333177ea0591f24a99a71" Oct 08 20:35:13 crc kubenswrapper[4988]: E1008 20:35:13.239362 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nm54f_openshift-machine-config-operator(f328a20b-5283-407e-8a93-2ab2da599f18)\"" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" Oct 08 20:35:28 crc kubenswrapper[4988]: I1008 20:35:28.238787 4988 scope.go:117] "RemoveContainer" containerID="413fd9f07623b7fb95cefbc48df19e6c85a31a0236c333177ea0591f24a99a71" Oct 08 20:35:28 crc kubenswrapper[4988]: E1008 20:35:28.239787 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nm54f_openshift-machine-config-operator(f328a20b-5283-407e-8a93-2ab2da599f18)\"" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" Oct 08 20:35:43 crc kubenswrapper[4988]: I1008 20:35:43.239041 4988 scope.go:117] "RemoveContainer" containerID="413fd9f07623b7fb95cefbc48df19e6c85a31a0236c333177ea0591f24a99a71" Oct 08 20:35:43 crc kubenswrapper[4988]: E1008 20:35:43.240230 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nm54f_openshift-machine-config-operator(f328a20b-5283-407e-8a93-2ab2da599f18)\"" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" Oct 08 20:35:58 crc kubenswrapper[4988]: I1008 20:35:58.238407 4988 scope.go:117] "RemoveContainer" containerID="413fd9f07623b7fb95cefbc48df19e6c85a31a0236c333177ea0591f24a99a71" Oct 08 20:35:58 crc kubenswrapper[4988]: E1008 20:35:58.239413 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nm54f_openshift-machine-config-operator(f328a20b-5283-407e-8a93-2ab2da599f18)\"" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" Oct 08 20:36:12 crc kubenswrapper[4988]: I1008 20:36:12.238882 4988 scope.go:117] "RemoveContainer" containerID="413fd9f07623b7fb95cefbc48df19e6c85a31a0236c333177ea0591f24a99a71" Oct 08 20:36:12 crc kubenswrapper[4988]: E1008 20:36:12.240859 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nm54f_openshift-machine-config-operator(f328a20b-5283-407e-8a93-2ab2da599f18)\"" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" Oct 08 20:36:26 crc kubenswrapper[4988]: I1008 20:36:26.238231 4988 scope.go:117] "RemoveContainer" containerID="413fd9f07623b7fb95cefbc48df19e6c85a31a0236c333177ea0591f24a99a71" Oct 08 20:36:26 crc kubenswrapper[4988]: E1008 20:36:26.239723 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nm54f_openshift-machine-config-operator(f328a20b-5283-407e-8a93-2ab2da599f18)\"" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" Oct 08 20:36:39 crc kubenswrapper[4988]: I1008 20:36:39.239115 4988 scope.go:117] "RemoveContainer" containerID="413fd9f07623b7fb95cefbc48df19e6c85a31a0236c333177ea0591f24a99a71" Oct 08 20:36:39 crc kubenswrapper[4988]: E1008 20:36:39.240651 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nm54f_openshift-machine-config-operator(f328a20b-5283-407e-8a93-2ab2da599f18)\"" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" Oct 08 20:36:52 crc kubenswrapper[4988]: I1008 20:36:52.238007 4988 scope.go:117] "RemoveContainer" containerID="413fd9f07623b7fb95cefbc48df19e6c85a31a0236c333177ea0591f24a99a71" Oct 08 20:36:52 crc kubenswrapper[4988]: E1008 20:36:52.239276 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nm54f_openshift-machine-config-operator(f328a20b-5283-407e-8a93-2ab2da599f18)\"" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" Oct 08 20:37:03 crc kubenswrapper[4988]: I1008 20:37:03.238355 4988 scope.go:117] "RemoveContainer" containerID="413fd9f07623b7fb95cefbc48df19e6c85a31a0236c333177ea0591f24a99a71" Oct 08 20:37:03 crc kubenswrapper[4988]: E1008 20:37:03.239747 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nm54f_openshift-machine-config-operator(f328a20b-5283-407e-8a93-2ab2da599f18)\"" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" Oct 08 20:37:14 crc kubenswrapper[4988]: I1008 20:37:14.237809 4988 scope.go:117] "RemoveContainer" containerID="413fd9f07623b7fb95cefbc48df19e6c85a31a0236c333177ea0591f24a99a71" Oct 08 20:37:14 crc kubenswrapper[4988]: E1008 20:37:14.238621 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nm54f_openshift-machine-config-operator(f328a20b-5283-407e-8a93-2ab2da599f18)\"" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" Oct 08 20:37:28 crc kubenswrapper[4988]: I1008 20:37:28.239165 4988 scope.go:117] "RemoveContainer" containerID="413fd9f07623b7fb95cefbc48df19e6c85a31a0236c333177ea0591f24a99a71" Oct 08 20:37:29 crc kubenswrapper[4988]: I1008 20:37:29.158123 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" event={"ID":"f328a20b-5283-407e-8a93-2ab2da599f18","Type":"ContainerStarted","Data":"f21c43696ae96fab4aca044e2adbdde2d544edfcba427c7247d210f2d406c66d"} Oct 08 20:39:53 crc kubenswrapper[4988]: I1008 20:39:53.338340 4988 patch_prober.go:28] interesting pod/machine-config-daemon-nm54f container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 20:39:53 crc kubenswrapper[4988]: I1008 20:39:53.339011 4988 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 20:39:56 crc kubenswrapper[4988]: I1008 20:39:56.049953 4988 generic.go:334] "Generic (PLEG): container finished" podID="54d6a0c7-be73-43b6-b857-a621d1a1c794" containerID="205da578d0eea36af26a70a87732f6999199f75a56061bad3f558de41d4411b5" exitCode=0 Oct 08 20:39:56 crc kubenswrapper[4988]: I1008 20:39:56.050468 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-openstack-openstack-cell1-qmvh4" event={"ID":"54d6a0c7-be73-43b6-b857-a621d1a1c794","Type":"ContainerDied","Data":"205da578d0eea36af26a70a87732f6999199f75a56061bad3f558de41d4411b5"} Oct 08 20:39:57 crc kubenswrapper[4988]: I1008 20:39:57.510425 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-openstack-openstack-cell1-qmvh4" Oct 08 20:39:57 crc kubenswrapper[4988]: I1008 20:39:57.628457 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/54d6a0c7-be73-43b6-b857-a621d1a1c794-ceilometer-compute-config-data-0\") pod \"54d6a0c7-be73-43b6-b857-a621d1a1c794\" (UID: \"54d6a0c7-be73-43b6-b857-a621d1a1c794\") " Oct 08 20:39:57 crc kubenswrapper[4988]: I1008 20:39:57.628527 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/54d6a0c7-be73-43b6-b857-a621d1a1c794-ssh-key\") pod \"54d6a0c7-be73-43b6-b857-a621d1a1c794\" (UID: \"54d6a0c7-be73-43b6-b857-a621d1a1c794\") " Oct 08 20:39:57 crc kubenswrapper[4988]: I1008 20:39:57.628634 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/54d6a0c7-be73-43b6-b857-a621d1a1c794-telemetry-combined-ca-bundle\") pod \"54d6a0c7-be73-43b6-b857-a621d1a1c794\" (UID: \"54d6a0c7-be73-43b6-b857-a621d1a1c794\") " Oct 08 20:39:57 crc kubenswrapper[4988]: I1008 20:39:57.628676 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/54d6a0c7-be73-43b6-b857-a621d1a1c794-ceilometer-compute-config-data-1\") pod \"54d6a0c7-be73-43b6-b857-a621d1a1c794\" (UID: \"54d6a0c7-be73-43b6-b857-a621d1a1c794\") " Oct 08 20:39:57 crc kubenswrapper[4988]: I1008 20:39:57.628792 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9f67j\" (UniqueName: \"kubernetes.io/projected/54d6a0c7-be73-43b6-b857-a621d1a1c794-kube-api-access-9f67j\") pod \"54d6a0c7-be73-43b6-b857-a621d1a1c794\" (UID: \"54d6a0c7-be73-43b6-b857-a621d1a1c794\") " Oct 08 20:39:57 crc kubenswrapper[4988]: I1008 20:39:57.628864 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/54d6a0c7-be73-43b6-b857-a621d1a1c794-inventory\") pod \"54d6a0c7-be73-43b6-b857-a621d1a1c794\" (UID: \"54d6a0c7-be73-43b6-b857-a621d1a1c794\") " Oct 08 20:39:57 crc kubenswrapper[4988]: I1008 20:39:57.628884 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/54d6a0c7-be73-43b6-b857-a621d1a1c794-ceilometer-compute-config-data-2\") pod \"54d6a0c7-be73-43b6-b857-a621d1a1c794\" (UID: \"54d6a0c7-be73-43b6-b857-a621d1a1c794\") " Oct 08 20:39:57 crc kubenswrapper[4988]: I1008 20:39:57.634768 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/54d6a0c7-be73-43b6-b857-a621d1a1c794-telemetry-combined-ca-bundle" (OuterVolumeSpecName: "telemetry-combined-ca-bundle") pod "54d6a0c7-be73-43b6-b857-a621d1a1c794" (UID: "54d6a0c7-be73-43b6-b857-a621d1a1c794"). InnerVolumeSpecName "telemetry-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 20:39:57 crc kubenswrapper[4988]: I1008 20:39:57.635149 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/54d6a0c7-be73-43b6-b857-a621d1a1c794-kube-api-access-9f67j" (OuterVolumeSpecName: "kube-api-access-9f67j") pod "54d6a0c7-be73-43b6-b857-a621d1a1c794" (UID: "54d6a0c7-be73-43b6-b857-a621d1a1c794"). InnerVolumeSpecName "kube-api-access-9f67j". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 20:39:57 crc kubenswrapper[4988]: I1008 20:39:57.658760 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/54d6a0c7-be73-43b6-b857-a621d1a1c794-ceilometer-compute-config-data-2" (OuterVolumeSpecName: "ceilometer-compute-config-data-2") pod "54d6a0c7-be73-43b6-b857-a621d1a1c794" (UID: "54d6a0c7-be73-43b6-b857-a621d1a1c794"). InnerVolumeSpecName "ceilometer-compute-config-data-2". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 20:39:57 crc kubenswrapper[4988]: I1008 20:39:57.661714 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/54d6a0c7-be73-43b6-b857-a621d1a1c794-inventory" (OuterVolumeSpecName: "inventory") pod "54d6a0c7-be73-43b6-b857-a621d1a1c794" (UID: "54d6a0c7-be73-43b6-b857-a621d1a1c794"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 20:39:57 crc kubenswrapper[4988]: I1008 20:39:57.665863 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/54d6a0c7-be73-43b6-b857-a621d1a1c794-ceilometer-compute-config-data-1" (OuterVolumeSpecName: "ceilometer-compute-config-data-1") pod "54d6a0c7-be73-43b6-b857-a621d1a1c794" (UID: "54d6a0c7-be73-43b6-b857-a621d1a1c794"). InnerVolumeSpecName "ceilometer-compute-config-data-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 20:39:57 crc kubenswrapper[4988]: I1008 20:39:57.666877 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/54d6a0c7-be73-43b6-b857-a621d1a1c794-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "54d6a0c7-be73-43b6-b857-a621d1a1c794" (UID: "54d6a0c7-be73-43b6-b857-a621d1a1c794"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 20:39:57 crc kubenswrapper[4988]: I1008 20:39:57.698151 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/54d6a0c7-be73-43b6-b857-a621d1a1c794-ceilometer-compute-config-data-0" (OuterVolumeSpecName: "ceilometer-compute-config-data-0") pod "54d6a0c7-be73-43b6-b857-a621d1a1c794" (UID: "54d6a0c7-be73-43b6-b857-a621d1a1c794"). InnerVolumeSpecName "ceilometer-compute-config-data-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 20:39:57 crc kubenswrapper[4988]: I1008 20:39:57.732198 4988 reconciler_common.go:293] "Volume detached for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/54d6a0c7-be73-43b6-b857-a621d1a1c794-telemetry-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 20:39:57 crc kubenswrapper[4988]: I1008 20:39:57.732245 4988 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/54d6a0c7-be73-43b6-b857-a621d1a1c794-ceilometer-compute-config-data-1\") on node \"crc\" DevicePath \"\"" Oct 08 20:39:57 crc kubenswrapper[4988]: I1008 20:39:57.732263 4988 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9f67j\" (UniqueName: \"kubernetes.io/projected/54d6a0c7-be73-43b6-b857-a621d1a1c794-kube-api-access-9f67j\") on node \"crc\" DevicePath \"\"" Oct 08 20:39:57 crc kubenswrapper[4988]: I1008 20:39:57.732276 4988 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/54d6a0c7-be73-43b6-b857-a621d1a1c794-ceilometer-compute-config-data-2\") on node \"crc\" DevicePath \"\"" Oct 08 20:39:57 crc kubenswrapper[4988]: I1008 20:39:57.732290 4988 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/54d6a0c7-be73-43b6-b857-a621d1a1c794-inventory\") on node \"crc\" DevicePath \"\"" Oct 08 20:39:57 crc kubenswrapper[4988]: I1008 20:39:57.732303 4988 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/54d6a0c7-be73-43b6-b857-a621d1a1c794-ceilometer-compute-config-data-0\") on node \"crc\" DevicePath \"\"" Oct 08 20:39:57 crc kubenswrapper[4988]: I1008 20:39:57.732315 4988 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/54d6a0c7-be73-43b6-b857-a621d1a1c794-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 08 20:39:58 crc kubenswrapper[4988]: I1008 20:39:58.078678 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-openstack-openstack-cell1-qmvh4" event={"ID":"54d6a0c7-be73-43b6-b857-a621d1a1c794","Type":"ContainerDied","Data":"926d10a8119d1c781b7f22978def0ec33ce02b3ef470c6a8ec791dded140558c"} Oct 08 20:39:58 crc kubenswrapper[4988]: I1008 20:39:58.078981 4988 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="926d10a8119d1c781b7f22978def0ec33ce02b3ef470c6a8ec791dded140558c" Oct 08 20:39:58 crc kubenswrapper[4988]: I1008 20:39:58.078780 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-openstack-openstack-cell1-qmvh4" Oct 08 20:39:58 crc kubenswrapper[4988]: I1008 20:39:58.219569 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-sriov-openstack-openstack-cell1-8vk4f"] Oct 08 20:39:58 crc kubenswrapper[4988]: E1008 20:39:58.220011 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="54d6a0c7-be73-43b6-b857-a621d1a1c794" containerName="telemetry-openstack-openstack-cell1" Oct 08 20:39:58 crc kubenswrapper[4988]: I1008 20:39:58.220029 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="54d6a0c7-be73-43b6-b857-a621d1a1c794" containerName="telemetry-openstack-openstack-cell1" Oct 08 20:39:58 crc kubenswrapper[4988]: I1008 20:39:58.220245 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="54d6a0c7-be73-43b6-b857-a621d1a1c794" containerName="telemetry-openstack-openstack-cell1" Oct 08 20:39:58 crc kubenswrapper[4988]: I1008 20:39:58.221009 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-sriov-openstack-openstack-cell1-8vk4f" Oct 08 20:39:58 crc kubenswrapper[4988]: I1008 20:39:58.225063 4988 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 08 20:39:58 crc kubenswrapper[4988]: I1008 20:39:58.225062 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-sriov-agent-neutron-config" Oct 08 20:39:58 crc kubenswrapper[4988]: I1008 20:39:58.226072 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Oct 08 20:39:58 crc kubenswrapper[4988]: I1008 20:39:58.226247 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-4sdzr" Oct 08 20:39:58 crc kubenswrapper[4988]: I1008 20:39:58.226717 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Oct 08 20:39:58 crc kubenswrapper[4988]: I1008 20:39:58.233097 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-sriov-openstack-openstack-cell1-8vk4f"] Oct 08 20:39:58 crc kubenswrapper[4988]: I1008 20:39:58.345811 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-sriov-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/96659c9c-7996-4e91-81f2-3edbaed544d3-neutron-sriov-combined-ca-bundle\") pod \"neutron-sriov-openstack-openstack-cell1-8vk4f\" (UID: \"96659c9c-7996-4e91-81f2-3edbaed544d3\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-8vk4f" Oct 08 20:39:58 crc kubenswrapper[4988]: I1008 20:39:58.345888 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/96659c9c-7996-4e91-81f2-3edbaed544d3-inventory\") pod \"neutron-sriov-openstack-openstack-cell1-8vk4f\" (UID: \"96659c9c-7996-4e91-81f2-3edbaed544d3\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-8vk4f" Oct 08 20:39:58 crc kubenswrapper[4988]: I1008 20:39:58.345935 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qzh46\" (UniqueName: \"kubernetes.io/projected/96659c9c-7996-4e91-81f2-3edbaed544d3-kube-api-access-qzh46\") pod \"neutron-sriov-openstack-openstack-cell1-8vk4f\" (UID: \"96659c9c-7996-4e91-81f2-3edbaed544d3\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-8vk4f" Oct 08 20:39:58 crc kubenswrapper[4988]: I1008 20:39:58.346296 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-sriov-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/96659c9c-7996-4e91-81f2-3edbaed544d3-neutron-sriov-agent-neutron-config-0\") pod \"neutron-sriov-openstack-openstack-cell1-8vk4f\" (UID: \"96659c9c-7996-4e91-81f2-3edbaed544d3\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-8vk4f" Oct 08 20:39:58 crc kubenswrapper[4988]: I1008 20:39:58.346343 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/96659c9c-7996-4e91-81f2-3edbaed544d3-ssh-key\") pod \"neutron-sriov-openstack-openstack-cell1-8vk4f\" (UID: \"96659c9c-7996-4e91-81f2-3edbaed544d3\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-8vk4f" Oct 08 20:39:58 crc kubenswrapper[4988]: I1008 20:39:58.449377 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-sriov-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/96659c9c-7996-4e91-81f2-3edbaed544d3-neutron-sriov-agent-neutron-config-0\") pod \"neutron-sriov-openstack-openstack-cell1-8vk4f\" (UID: \"96659c9c-7996-4e91-81f2-3edbaed544d3\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-8vk4f" Oct 08 20:39:58 crc kubenswrapper[4988]: I1008 20:39:58.449446 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/96659c9c-7996-4e91-81f2-3edbaed544d3-ssh-key\") pod \"neutron-sriov-openstack-openstack-cell1-8vk4f\" (UID: \"96659c9c-7996-4e91-81f2-3edbaed544d3\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-8vk4f" Oct 08 20:39:58 crc kubenswrapper[4988]: I1008 20:39:58.449536 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-sriov-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/96659c9c-7996-4e91-81f2-3edbaed544d3-neutron-sriov-combined-ca-bundle\") pod \"neutron-sriov-openstack-openstack-cell1-8vk4f\" (UID: \"96659c9c-7996-4e91-81f2-3edbaed544d3\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-8vk4f" Oct 08 20:39:58 crc kubenswrapper[4988]: I1008 20:39:58.449559 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/96659c9c-7996-4e91-81f2-3edbaed544d3-inventory\") pod \"neutron-sriov-openstack-openstack-cell1-8vk4f\" (UID: \"96659c9c-7996-4e91-81f2-3edbaed544d3\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-8vk4f" Oct 08 20:39:58 crc kubenswrapper[4988]: I1008 20:39:58.449587 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qzh46\" (UniqueName: \"kubernetes.io/projected/96659c9c-7996-4e91-81f2-3edbaed544d3-kube-api-access-qzh46\") pod \"neutron-sriov-openstack-openstack-cell1-8vk4f\" (UID: \"96659c9c-7996-4e91-81f2-3edbaed544d3\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-8vk4f" Oct 08 20:39:58 crc kubenswrapper[4988]: I1008 20:39:58.460289 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-sriov-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/96659c9c-7996-4e91-81f2-3edbaed544d3-neutron-sriov-agent-neutron-config-0\") pod \"neutron-sriov-openstack-openstack-cell1-8vk4f\" (UID: \"96659c9c-7996-4e91-81f2-3edbaed544d3\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-8vk4f" Oct 08 20:39:58 crc kubenswrapper[4988]: I1008 20:39:58.460578 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/96659c9c-7996-4e91-81f2-3edbaed544d3-ssh-key\") pod \"neutron-sriov-openstack-openstack-cell1-8vk4f\" (UID: \"96659c9c-7996-4e91-81f2-3edbaed544d3\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-8vk4f" Oct 08 20:39:58 crc kubenswrapper[4988]: I1008 20:39:58.462286 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/96659c9c-7996-4e91-81f2-3edbaed544d3-inventory\") pod \"neutron-sriov-openstack-openstack-cell1-8vk4f\" (UID: \"96659c9c-7996-4e91-81f2-3edbaed544d3\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-8vk4f" Oct 08 20:39:58 crc kubenswrapper[4988]: I1008 20:39:58.467236 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-sriov-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/96659c9c-7996-4e91-81f2-3edbaed544d3-neutron-sriov-combined-ca-bundle\") pod \"neutron-sriov-openstack-openstack-cell1-8vk4f\" (UID: \"96659c9c-7996-4e91-81f2-3edbaed544d3\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-8vk4f" Oct 08 20:39:58 crc kubenswrapper[4988]: I1008 20:39:58.481701 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qzh46\" (UniqueName: \"kubernetes.io/projected/96659c9c-7996-4e91-81f2-3edbaed544d3-kube-api-access-qzh46\") pod \"neutron-sriov-openstack-openstack-cell1-8vk4f\" (UID: \"96659c9c-7996-4e91-81f2-3edbaed544d3\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-8vk4f" Oct 08 20:39:58 crc kubenswrapper[4988]: I1008 20:39:58.549121 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-sriov-openstack-openstack-cell1-8vk4f" Oct 08 20:39:59 crc kubenswrapper[4988]: I1008 20:39:59.171227 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-sriov-openstack-openstack-cell1-8vk4f"] Oct 08 20:39:59 crc kubenswrapper[4988]: I1008 20:39:59.183926 4988 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 08 20:40:00 crc kubenswrapper[4988]: I1008 20:40:00.102610 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-sriov-openstack-openstack-cell1-8vk4f" event={"ID":"96659c9c-7996-4e91-81f2-3edbaed544d3","Type":"ContainerStarted","Data":"7f0bd9c77706cb11aca0f07c5f2ba3809180052d5e6ded6f50e874a05d1e7fd1"} Oct 08 20:40:00 crc kubenswrapper[4988]: I1008 20:40:00.103174 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-sriov-openstack-openstack-cell1-8vk4f" event={"ID":"96659c9c-7996-4e91-81f2-3edbaed544d3","Type":"ContainerStarted","Data":"82d09d33d5d56b664296a49ca11ca33cdbff3aaaa271699245df0d266b74a497"} Oct 08 20:40:00 crc kubenswrapper[4988]: I1008 20:40:00.131635 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-sriov-openstack-openstack-cell1-8vk4f" podStartSLOduration=1.566053886 podStartE2EDuration="2.131617893s" podCreationTimestamp="2025-10-08 20:39:58 +0000 UTC" firstStartedPulling="2025-10-08 20:39:59.183696315 +0000 UTC m=+8944.633539085" lastFinishedPulling="2025-10-08 20:39:59.749260282 +0000 UTC m=+8945.199103092" observedRunningTime="2025-10-08 20:40:00.128226047 +0000 UTC m=+8945.578068837" watchObservedRunningTime="2025-10-08 20:40:00.131617893 +0000 UTC m=+8945.581460683" Oct 08 20:40:23 crc kubenswrapper[4988]: I1008 20:40:23.338237 4988 patch_prober.go:28] interesting pod/machine-config-daemon-nm54f container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 20:40:23 crc kubenswrapper[4988]: I1008 20:40:23.339124 4988 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 20:40:53 crc kubenswrapper[4988]: I1008 20:40:53.338189 4988 patch_prober.go:28] interesting pod/machine-config-daemon-nm54f container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 20:40:53 crc kubenswrapper[4988]: I1008 20:40:53.338863 4988 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 20:40:53 crc kubenswrapper[4988]: I1008 20:40:53.338933 4988 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" Oct 08 20:40:53 crc kubenswrapper[4988]: I1008 20:40:53.340310 4988 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"f21c43696ae96fab4aca044e2adbdde2d544edfcba427c7247d210f2d406c66d"} pod="openshift-machine-config-operator/machine-config-daemon-nm54f" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 08 20:40:53 crc kubenswrapper[4988]: I1008 20:40:53.340468 4988 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" containerName="machine-config-daemon" containerID="cri-o://f21c43696ae96fab4aca044e2adbdde2d544edfcba427c7247d210f2d406c66d" gracePeriod=600 Oct 08 20:40:53 crc kubenswrapper[4988]: I1008 20:40:53.784735 4988 generic.go:334] "Generic (PLEG): container finished" podID="f328a20b-5283-407e-8a93-2ab2da599f18" containerID="f21c43696ae96fab4aca044e2adbdde2d544edfcba427c7247d210f2d406c66d" exitCode=0 Oct 08 20:40:53 crc kubenswrapper[4988]: I1008 20:40:53.784818 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" event={"ID":"f328a20b-5283-407e-8a93-2ab2da599f18","Type":"ContainerDied","Data":"f21c43696ae96fab4aca044e2adbdde2d544edfcba427c7247d210f2d406c66d"} Oct 08 20:40:53 crc kubenswrapper[4988]: I1008 20:40:53.785091 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" event={"ID":"f328a20b-5283-407e-8a93-2ab2da599f18","Type":"ContainerStarted","Data":"6382c087e916f3958809279cab4d35e45e3b3ccd0cb9d8f0da5a015741773071"} Oct 08 20:40:53 crc kubenswrapper[4988]: I1008 20:40:53.785118 4988 scope.go:117] "RemoveContainer" containerID="413fd9f07623b7fb95cefbc48df19e6c85a31a0236c333177ea0591f24a99a71" Oct 08 20:41:34 crc kubenswrapper[4988]: I1008 20:41:34.559862 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-4nskl"] Oct 08 20:41:34 crc kubenswrapper[4988]: I1008 20:41:34.568846 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-4nskl" Oct 08 20:41:34 crc kubenswrapper[4988]: I1008 20:41:34.591062 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-4nskl"] Oct 08 20:41:34 crc kubenswrapper[4988]: I1008 20:41:34.686928 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6nq5b\" (UniqueName: \"kubernetes.io/projected/6fd57a5f-1920-430f-8127-683a8cb4614c-kube-api-access-6nq5b\") pod \"redhat-operators-4nskl\" (UID: \"6fd57a5f-1920-430f-8127-683a8cb4614c\") " pod="openshift-marketplace/redhat-operators-4nskl" Oct 08 20:41:34 crc kubenswrapper[4988]: I1008 20:41:34.687098 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6fd57a5f-1920-430f-8127-683a8cb4614c-catalog-content\") pod \"redhat-operators-4nskl\" (UID: \"6fd57a5f-1920-430f-8127-683a8cb4614c\") " pod="openshift-marketplace/redhat-operators-4nskl" Oct 08 20:41:34 crc kubenswrapper[4988]: I1008 20:41:34.687192 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6fd57a5f-1920-430f-8127-683a8cb4614c-utilities\") pod \"redhat-operators-4nskl\" (UID: \"6fd57a5f-1920-430f-8127-683a8cb4614c\") " pod="openshift-marketplace/redhat-operators-4nskl" Oct 08 20:41:34 crc kubenswrapper[4988]: I1008 20:41:34.789524 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6nq5b\" (UniqueName: \"kubernetes.io/projected/6fd57a5f-1920-430f-8127-683a8cb4614c-kube-api-access-6nq5b\") pod \"redhat-operators-4nskl\" (UID: \"6fd57a5f-1920-430f-8127-683a8cb4614c\") " pod="openshift-marketplace/redhat-operators-4nskl" Oct 08 20:41:34 crc kubenswrapper[4988]: I1008 20:41:34.789638 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6fd57a5f-1920-430f-8127-683a8cb4614c-catalog-content\") pod \"redhat-operators-4nskl\" (UID: \"6fd57a5f-1920-430f-8127-683a8cb4614c\") " pod="openshift-marketplace/redhat-operators-4nskl" Oct 08 20:41:34 crc kubenswrapper[4988]: I1008 20:41:34.789707 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6fd57a5f-1920-430f-8127-683a8cb4614c-utilities\") pod \"redhat-operators-4nskl\" (UID: \"6fd57a5f-1920-430f-8127-683a8cb4614c\") " pod="openshift-marketplace/redhat-operators-4nskl" Oct 08 20:41:34 crc kubenswrapper[4988]: I1008 20:41:34.790231 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6fd57a5f-1920-430f-8127-683a8cb4614c-utilities\") pod \"redhat-operators-4nskl\" (UID: \"6fd57a5f-1920-430f-8127-683a8cb4614c\") " pod="openshift-marketplace/redhat-operators-4nskl" Oct 08 20:41:34 crc kubenswrapper[4988]: I1008 20:41:34.790323 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6fd57a5f-1920-430f-8127-683a8cb4614c-catalog-content\") pod \"redhat-operators-4nskl\" (UID: \"6fd57a5f-1920-430f-8127-683a8cb4614c\") " pod="openshift-marketplace/redhat-operators-4nskl" Oct 08 20:41:34 crc kubenswrapper[4988]: I1008 20:41:34.819295 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6nq5b\" (UniqueName: \"kubernetes.io/projected/6fd57a5f-1920-430f-8127-683a8cb4614c-kube-api-access-6nq5b\") pod \"redhat-operators-4nskl\" (UID: \"6fd57a5f-1920-430f-8127-683a8cb4614c\") " pod="openshift-marketplace/redhat-operators-4nskl" Oct 08 20:41:34 crc kubenswrapper[4988]: I1008 20:41:34.939910 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-4nskl" Oct 08 20:41:35 crc kubenswrapper[4988]: I1008 20:41:35.438577 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-4nskl"] Oct 08 20:41:36 crc kubenswrapper[4988]: I1008 20:41:36.355127 4988 generic.go:334] "Generic (PLEG): container finished" podID="6fd57a5f-1920-430f-8127-683a8cb4614c" containerID="aa7249f6063c1e6869e4fdcfeaaf0b0e6b37ec8572f54a57da9ac12f46c93fe7" exitCode=0 Oct 08 20:41:36 crc kubenswrapper[4988]: I1008 20:41:36.355245 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4nskl" event={"ID":"6fd57a5f-1920-430f-8127-683a8cb4614c","Type":"ContainerDied","Data":"aa7249f6063c1e6869e4fdcfeaaf0b0e6b37ec8572f54a57da9ac12f46c93fe7"} Oct 08 20:41:36 crc kubenswrapper[4988]: I1008 20:41:36.355519 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4nskl" event={"ID":"6fd57a5f-1920-430f-8127-683a8cb4614c","Type":"ContainerStarted","Data":"d901d4aec35d3bdbc33337cd4f953db49b6e0439fbc822641a19534017795c96"} Oct 08 20:41:38 crc kubenswrapper[4988]: I1008 20:41:38.392900 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4nskl" event={"ID":"6fd57a5f-1920-430f-8127-683a8cb4614c","Type":"ContainerStarted","Data":"db2e7a3c706f2b0c4fc11175116c382211f05279108f18edd0b2292a7d0b917a"} Oct 08 20:41:39 crc kubenswrapper[4988]: I1008 20:41:39.406649 4988 generic.go:334] "Generic (PLEG): container finished" podID="6fd57a5f-1920-430f-8127-683a8cb4614c" containerID="db2e7a3c706f2b0c4fc11175116c382211f05279108f18edd0b2292a7d0b917a" exitCode=0 Oct 08 20:41:39 crc kubenswrapper[4988]: I1008 20:41:39.406734 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4nskl" event={"ID":"6fd57a5f-1920-430f-8127-683a8cb4614c","Type":"ContainerDied","Data":"db2e7a3c706f2b0c4fc11175116c382211f05279108f18edd0b2292a7d0b917a"} Oct 08 20:41:40 crc kubenswrapper[4988]: I1008 20:41:40.423771 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4nskl" event={"ID":"6fd57a5f-1920-430f-8127-683a8cb4614c","Type":"ContainerStarted","Data":"434f39d7dc8acc96a32e8e956dc7c1ef3fe23769765a877a365d3abb8a34f61d"} Oct 08 20:41:40 crc kubenswrapper[4988]: I1008 20:41:40.467258 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-4nskl" podStartSLOduration=2.969126937 podStartE2EDuration="6.467243496s" podCreationTimestamp="2025-10-08 20:41:34 +0000 UTC" firstStartedPulling="2025-10-08 20:41:36.358546311 +0000 UTC m=+9041.808389121" lastFinishedPulling="2025-10-08 20:41:39.8566629 +0000 UTC m=+9045.306505680" observedRunningTime="2025-10-08 20:41:40.450659487 +0000 UTC m=+9045.900502247" watchObservedRunningTime="2025-10-08 20:41:40.467243496 +0000 UTC m=+9045.917086266" Oct 08 20:41:44 crc kubenswrapper[4988]: I1008 20:41:44.941040 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-4nskl" Oct 08 20:41:44 crc kubenswrapper[4988]: I1008 20:41:44.941771 4988 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-4nskl" Oct 08 20:41:46 crc kubenswrapper[4988]: I1008 20:41:46.029005 4988 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-4nskl" podUID="6fd57a5f-1920-430f-8127-683a8cb4614c" containerName="registry-server" probeResult="failure" output=< Oct 08 20:41:46 crc kubenswrapper[4988]: timeout: failed to connect service ":50051" within 1s Oct 08 20:41:46 crc kubenswrapper[4988]: > Oct 08 20:41:55 crc kubenswrapper[4988]: I1008 20:41:55.026965 4988 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-4nskl" Oct 08 20:41:55 crc kubenswrapper[4988]: I1008 20:41:55.111630 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-4nskl" Oct 08 20:41:55 crc kubenswrapper[4988]: I1008 20:41:55.281353 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-4nskl"] Oct 08 20:41:56 crc kubenswrapper[4988]: I1008 20:41:56.606755 4988 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-4nskl" podUID="6fd57a5f-1920-430f-8127-683a8cb4614c" containerName="registry-server" containerID="cri-o://434f39d7dc8acc96a32e8e956dc7c1ef3fe23769765a877a365d3abb8a34f61d" gracePeriod=2 Oct 08 20:41:57 crc kubenswrapper[4988]: I1008 20:41:57.185109 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-4nskl" Oct 08 20:41:57 crc kubenswrapper[4988]: I1008 20:41:57.233557 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6fd57a5f-1920-430f-8127-683a8cb4614c-utilities\") pod \"6fd57a5f-1920-430f-8127-683a8cb4614c\" (UID: \"6fd57a5f-1920-430f-8127-683a8cb4614c\") " Oct 08 20:41:57 crc kubenswrapper[4988]: I1008 20:41:57.233663 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6nq5b\" (UniqueName: \"kubernetes.io/projected/6fd57a5f-1920-430f-8127-683a8cb4614c-kube-api-access-6nq5b\") pod \"6fd57a5f-1920-430f-8127-683a8cb4614c\" (UID: \"6fd57a5f-1920-430f-8127-683a8cb4614c\") " Oct 08 20:41:57 crc kubenswrapper[4988]: I1008 20:41:57.233936 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6fd57a5f-1920-430f-8127-683a8cb4614c-catalog-content\") pod \"6fd57a5f-1920-430f-8127-683a8cb4614c\" (UID: \"6fd57a5f-1920-430f-8127-683a8cb4614c\") " Oct 08 20:41:57 crc kubenswrapper[4988]: I1008 20:41:57.242128 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6fd57a5f-1920-430f-8127-683a8cb4614c-utilities" (OuterVolumeSpecName: "utilities") pod "6fd57a5f-1920-430f-8127-683a8cb4614c" (UID: "6fd57a5f-1920-430f-8127-683a8cb4614c"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 20:41:57 crc kubenswrapper[4988]: I1008 20:41:57.246478 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6fd57a5f-1920-430f-8127-683a8cb4614c-kube-api-access-6nq5b" (OuterVolumeSpecName: "kube-api-access-6nq5b") pod "6fd57a5f-1920-430f-8127-683a8cb4614c" (UID: "6fd57a5f-1920-430f-8127-683a8cb4614c"). InnerVolumeSpecName "kube-api-access-6nq5b". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 20:41:57 crc kubenswrapper[4988]: I1008 20:41:57.323853 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6fd57a5f-1920-430f-8127-683a8cb4614c-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "6fd57a5f-1920-430f-8127-683a8cb4614c" (UID: "6fd57a5f-1920-430f-8127-683a8cb4614c"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 20:41:57 crc kubenswrapper[4988]: I1008 20:41:57.336729 4988 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6fd57a5f-1920-430f-8127-683a8cb4614c-utilities\") on node \"crc\" DevicePath \"\"" Oct 08 20:41:57 crc kubenswrapper[4988]: I1008 20:41:57.336764 4988 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6nq5b\" (UniqueName: \"kubernetes.io/projected/6fd57a5f-1920-430f-8127-683a8cb4614c-kube-api-access-6nq5b\") on node \"crc\" DevicePath \"\"" Oct 08 20:41:57 crc kubenswrapper[4988]: I1008 20:41:57.336779 4988 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6fd57a5f-1920-430f-8127-683a8cb4614c-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 08 20:41:57 crc kubenswrapper[4988]: I1008 20:41:57.619563 4988 generic.go:334] "Generic (PLEG): container finished" podID="6fd57a5f-1920-430f-8127-683a8cb4614c" containerID="434f39d7dc8acc96a32e8e956dc7c1ef3fe23769765a877a365d3abb8a34f61d" exitCode=0 Oct 08 20:41:57 crc kubenswrapper[4988]: I1008 20:41:57.619601 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4nskl" event={"ID":"6fd57a5f-1920-430f-8127-683a8cb4614c","Type":"ContainerDied","Data":"434f39d7dc8acc96a32e8e956dc7c1ef3fe23769765a877a365d3abb8a34f61d"} Oct 08 20:41:57 crc kubenswrapper[4988]: I1008 20:41:57.619627 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4nskl" event={"ID":"6fd57a5f-1920-430f-8127-683a8cb4614c","Type":"ContainerDied","Data":"d901d4aec35d3bdbc33337cd4f953db49b6e0439fbc822641a19534017795c96"} Oct 08 20:41:57 crc kubenswrapper[4988]: I1008 20:41:57.619651 4988 scope.go:117] "RemoveContainer" containerID="434f39d7dc8acc96a32e8e956dc7c1ef3fe23769765a877a365d3abb8a34f61d" Oct 08 20:41:57 crc kubenswrapper[4988]: I1008 20:41:57.619720 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-4nskl" Oct 08 20:41:57 crc kubenswrapper[4988]: I1008 20:41:57.652657 4988 scope.go:117] "RemoveContainer" containerID="db2e7a3c706f2b0c4fc11175116c382211f05279108f18edd0b2292a7d0b917a" Oct 08 20:41:57 crc kubenswrapper[4988]: I1008 20:41:57.677744 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-4nskl"] Oct 08 20:41:57 crc kubenswrapper[4988]: I1008 20:41:57.683484 4988 scope.go:117] "RemoveContainer" containerID="aa7249f6063c1e6869e4fdcfeaaf0b0e6b37ec8572f54a57da9ac12f46c93fe7" Oct 08 20:41:57 crc kubenswrapper[4988]: I1008 20:41:57.694303 4988 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-4nskl"] Oct 08 20:41:57 crc kubenswrapper[4988]: I1008 20:41:57.742940 4988 scope.go:117] "RemoveContainer" containerID="434f39d7dc8acc96a32e8e956dc7c1ef3fe23769765a877a365d3abb8a34f61d" Oct 08 20:41:57 crc kubenswrapper[4988]: E1008 20:41:57.743404 4988 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"434f39d7dc8acc96a32e8e956dc7c1ef3fe23769765a877a365d3abb8a34f61d\": container with ID starting with 434f39d7dc8acc96a32e8e956dc7c1ef3fe23769765a877a365d3abb8a34f61d not found: ID does not exist" containerID="434f39d7dc8acc96a32e8e956dc7c1ef3fe23769765a877a365d3abb8a34f61d" Oct 08 20:41:57 crc kubenswrapper[4988]: I1008 20:41:57.743495 4988 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"434f39d7dc8acc96a32e8e956dc7c1ef3fe23769765a877a365d3abb8a34f61d"} err="failed to get container status \"434f39d7dc8acc96a32e8e956dc7c1ef3fe23769765a877a365d3abb8a34f61d\": rpc error: code = NotFound desc = could not find container \"434f39d7dc8acc96a32e8e956dc7c1ef3fe23769765a877a365d3abb8a34f61d\": container with ID starting with 434f39d7dc8acc96a32e8e956dc7c1ef3fe23769765a877a365d3abb8a34f61d not found: ID does not exist" Oct 08 20:41:57 crc kubenswrapper[4988]: I1008 20:41:57.743590 4988 scope.go:117] "RemoveContainer" containerID="db2e7a3c706f2b0c4fc11175116c382211f05279108f18edd0b2292a7d0b917a" Oct 08 20:41:57 crc kubenswrapper[4988]: E1008 20:41:57.743952 4988 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"db2e7a3c706f2b0c4fc11175116c382211f05279108f18edd0b2292a7d0b917a\": container with ID starting with db2e7a3c706f2b0c4fc11175116c382211f05279108f18edd0b2292a7d0b917a not found: ID does not exist" containerID="db2e7a3c706f2b0c4fc11175116c382211f05279108f18edd0b2292a7d0b917a" Oct 08 20:41:57 crc kubenswrapper[4988]: I1008 20:41:57.744026 4988 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"db2e7a3c706f2b0c4fc11175116c382211f05279108f18edd0b2292a7d0b917a"} err="failed to get container status \"db2e7a3c706f2b0c4fc11175116c382211f05279108f18edd0b2292a7d0b917a\": rpc error: code = NotFound desc = could not find container \"db2e7a3c706f2b0c4fc11175116c382211f05279108f18edd0b2292a7d0b917a\": container with ID starting with db2e7a3c706f2b0c4fc11175116c382211f05279108f18edd0b2292a7d0b917a not found: ID does not exist" Oct 08 20:41:57 crc kubenswrapper[4988]: I1008 20:41:57.744139 4988 scope.go:117] "RemoveContainer" containerID="aa7249f6063c1e6869e4fdcfeaaf0b0e6b37ec8572f54a57da9ac12f46c93fe7" Oct 08 20:41:57 crc kubenswrapper[4988]: E1008 20:41:57.744509 4988 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"aa7249f6063c1e6869e4fdcfeaaf0b0e6b37ec8572f54a57da9ac12f46c93fe7\": container with ID starting with aa7249f6063c1e6869e4fdcfeaaf0b0e6b37ec8572f54a57da9ac12f46c93fe7 not found: ID does not exist" containerID="aa7249f6063c1e6869e4fdcfeaaf0b0e6b37ec8572f54a57da9ac12f46c93fe7" Oct 08 20:41:57 crc kubenswrapper[4988]: I1008 20:41:57.744542 4988 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"aa7249f6063c1e6869e4fdcfeaaf0b0e6b37ec8572f54a57da9ac12f46c93fe7"} err="failed to get container status \"aa7249f6063c1e6869e4fdcfeaaf0b0e6b37ec8572f54a57da9ac12f46c93fe7\": rpc error: code = NotFound desc = could not find container \"aa7249f6063c1e6869e4fdcfeaaf0b0e6b37ec8572f54a57da9ac12f46c93fe7\": container with ID starting with aa7249f6063c1e6869e4fdcfeaaf0b0e6b37ec8572f54a57da9ac12f46c93fe7 not found: ID does not exist" Oct 08 20:41:59 crc kubenswrapper[4988]: I1008 20:41:59.260546 4988 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6fd57a5f-1920-430f-8127-683a8cb4614c" path="/var/lib/kubelet/pods/6fd57a5f-1920-430f-8127-683a8cb4614c/volumes" Oct 08 20:42:53 crc kubenswrapper[4988]: I1008 20:42:53.337707 4988 patch_prober.go:28] interesting pod/machine-config-daemon-nm54f container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 20:42:53 crc kubenswrapper[4988]: I1008 20:42:53.338267 4988 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 20:42:55 crc kubenswrapper[4988]: I1008 20:42:55.399567 4988 generic.go:334] "Generic (PLEG): container finished" podID="96659c9c-7996-4e91-81f2-3edbaed544d3" containerID="7f0bd9c77706cb11aca0f07c5f2ba3809180052d5e6ded6f50e874a05d1e7fd1" exitCode=0 Oct 08 20:42:55 crc kubenswrapper[4988]: I1008 20:42:55.399625 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-sriov-openstack-openstack-cell1-8vk4f" event={"ID":"96659c9c-7996-4e91-81f2-3edbaed544d3","Type":"ContainerDied","Data":"7f0bd9c77706cb11aca0f07c5f2ba3809180052d5e6ded6f50e874a05d1e7fd1"} Oct 08 20:42:56 crc kubenswrapper[4988]: I1008 20:42:56.979982 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-sriov-openstack-openstack-cell1-8vk4f" Oct 08 20:42:57 crc kubenswrapper[4988]: I1008 20:42:57.085200 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/96659c9c-7996-4e91-81f2-3edbaed544d3-inventory\") pod \"96659c9c-7996-4e91-81f2-3edbaed544d3\" (UID: \"96659c9c-7996-4e91-81f2-3edbaed544d3\") " Oct 08 20:42:57 crc kubenswrapper[4988]: I1008 20:42:57.085316 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/96659c9c-7996-4e91-81f2-3edbaed544d3-ssh-key\") pod \"96659c9c-7996-4e91-81f2-3edbaed544d3\" (UID: \"96659c9c-7996-4e91-81f2-3edbaed544d3\") " Oct 08 20:42:57 crc kubenswrapper[4988]: I1008 20:42:57.085468 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-sriov-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/96659c9c-7996-4e91-81f2-3edbaed544d3-neutron-sriov-agent-neutron-config-0\") pod \"96659c9c-7996-4e91-81f2-3edbaed544d3\" (UID: \"96659c9c-7996-4e91-81f2-3edbaed544d3\") " Oct 08 20:42:57 crc kubenswrapper[4988]: I1008 20:42:57.085581 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-sriov-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/96659c9c-7996-4e91-81f2-3edbaed544d3-neutron-sriov-combined-ca-bundle\") pod \"96659c9c-7996-4e91-81f2-3edbaed544d3\" (UID: \"96659c9c-7996-4e91-81f2-3edbaed544d3\") " Oct 08 20:42:57 crc kubenswrapper[4988]: I1008 20:42:57.085626 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qzh46\" (UniqueName: \"kubernetes.io/projected/96659c9c-7996-4e91-81f2-3edbaed544d3-kube-api-access-qzh46\") pod \"96659c9c-7996-4e91-81f2-3edbaed544d3\" (UID: \"96659c9c-7996-4e91-81f2-3edbaed544d3\") " Oct 08 20:42:57 crc kubenswrapper[4988]: I1008 20:42:57.092268 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/96659c9c-7996-4e91-81f2-3edbaed544d3-neutron-sriov-combined-ca-bundle" (OuterVolumeSpecName: "neutron-sriov-combined-ca-bundle") pod "96659c9c-7996-4e91-81f2-3edbaed544d3" (UID: "96659c9c-7996-4e91-81f2-3edbaed544d3"). InnerVolumeSpecName "neutron-sriov-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 20:42:57 crc kubenswrapper[4988]: I1008 20:42:57.093566 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/96659c9c-7996-4e91-81f2-3edbaed544d3-kube-api-access-qzh46" (OuterVolumeSpecName: "kube-api-access-qzh46") pod "96659c9c-7996-4e91-81f2-3edbaed544d3" (UID: "96659c9c-7996-4e91-81f2-3edbaed544d3"). InnerVolumeSpecName "kube-api-access-qzh46". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 20:42:57 crc kubenswrapper[4988]: I1008 20:42:57.127024 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/96659c9c-7996-4e91-81f2-3edbaed544d3-inventory" (OuterVolumeSpecName: "inventory") pod "96659c9c-7996-4e91-81f2-3edbaed544d3" (UID: "96659c9c-7996-4e91-81f2-3edbaed544d3"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 20:42:57 crc kubenswrapper[4988]: I1008 20:42:57.127582 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/96659c9c-7996-4e91-81f2-3edbaed544d3-neutron-sriov-agent-neutron-config-0" (OuterVolumeSpecName: "neutron-sriov-agent-neutron-config-0") pod "96659c9c-7996-4e91-81f2-3edbaed544d3" (UID: "96659c9c-7996-4e91-81f2-3edbaed544d3"). InnerVolumeSpecName "neutron-sriov-agent-neutron-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 20:42:57 crc kubenswrapper[4988]: I1008 20:42:57.137587 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/96659c9c-7996-4e91-81f2-3edbaed544d3-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "96659c9c-7996-4e91-81f2-3edbaed544d3" (UID: "96659c9c-7996-4e91-81f2-3edbaed544d3"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 20:42:57 crc kubenswrapper[4988]: I1008 20:42:57.188696 4988 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/96659c9c-7996-4e91-81f2-3edbaed544d3-inventory\") on node \"crc\" DevicePath \"\"" Oct 08 20:42:57 crc kubenswrapper[4988]: I1008 20:42:57.188735 4988 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/96659c9c-7996-4e91-81f2-3edbaed544d3-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 08 20:42:57 crc kubenswrapper[4988]: I1008 20:42:57.188749 4988 reconciler_common.go:293] "Volume detached for volume \"neutron-sriov-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/96659c9c-7996-4e91-81f2-3edbaed544d3-neutron-sriov-agent-neutron-config-0\") on node \"crc\" DevicePath \"\"" Oct 08 20:42:57 crc kubenswrapper[4988]: I1008 20:42:57.188763 4988 reconciler_common.go:293] "Volume detached for volume \"neutron-sriov-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/96659c9c-7996-4e91-81f2-3edbaed544d3-neutron-sriov-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 20:42:57 crc kubenswrapper[4988]: I1008 20:42:57.188779 4988 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qzh46\" (UniqueName: \"kubernetes.io/projected/96659c9c-7996-4e91-81f2-3edbaed544d3-kube-api-access-qzh46\") on node \"crc\" DevicePath \"\"" Oct 08 20:42:57 crc kubenswrapper[4988]: I1008 20:42:57.425478 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-sriov-openstack-openstack-cell1-8vk4f" event={"ID":"96659c9c-7996-4e91-81f2-3edbaed544d3","Type":"ContainerDied","Data":"82d09d33d5d56b664296a49ca11ca33cdbff3aaaa271699245df0d266b74a497"} Oct 08 20:42:57 crc kubenswrapper[4988]: I1008 20:42:57.425512 4988 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="82d09d33d5d56b664296a49ca11ca33cdbff3aaaa271699245df0d266b74a497" Oct 08 20:42:57 crc kubenswrapper[4988]: I1008 20:42:57.425574 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-sriov-openstack-openstack-cell1-8vk4f" Oct 08 20:42:57 crc kubenswrapper[4988]: I1008 20:42:57.565327 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-dhcp-openstack-openstack-cell1-mbrs2"] Oct 08 20:42:57 crc kubenswrapper[4988]: E1008 20:42:57.568008 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6fd57a5f-1920-430f-8127-683a8cb4614c" containerName="extract-content" Oct 08 20:42:57 crc kubenswrapper[4988]: I1008 20:42:57.568162 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="6fd57a5f-1920-430f-8127-683a8cb4614c" containerName="extract-content" Oct 08 20:42:57 crc kubenswrapper[4988]: E1008 20:42:57.568330 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="96659c9c-7996-4e91-81f2-3edbaed544d3" containerName="neutron-sriov-openstack-openstack-cell1" Oct 08 20:42:57 crc kubenswrapper[4988]: I1008 20:42:57.568476 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="96659c9c-7996-4e91-81f2-3edbaed544d3" containerName="neutron-sriov-openstack-openstack-cell1" Oct 08 20:42:57 crc kubenswrapper[4988]: E1008 20:42:57.568640 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6fd57a5f-1920-430f-8127-683a8cb4614c" containerName="registry-server" Oct 08 20:42:57 crc kubenswrapper[4988]: I1008 20:42:57.568738 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="6fd57a5f-1920-430f-8127-683a8cb4614c" containerName="registry-server" Oct 08 20:42:57 crc kubenswrapper[4988]: E1008 20:42:57.568886 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6fd57a5f-1920-430f-8127-683a8cb4614c" containerName="extract-utilities" Oct 08 20:42:57 crc kubenswrapper[4988]: I1008 20:42:57.569002 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="6fd57a5f-1920-430f-8127-683a8cb4614c" containerName="extract-utilities" Oct 08 20:42:57 crc kubenswrapper[4988]: I1008 20:42:57.570402 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="96659c9c-7996-4e91-81f2-3edbaed544d3" containerName="neutron-sriov-openstack-openstack-cell1" Oct 08 20:42:57 crc kubenswrapper[4988]: I1008 20:42:57.570752 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="6fd57a5f-1920-430f-8127-683a8cb4614c" containerName="registry-server" Oct 08 20:42:57 crc kubenswrapper[4988]: I1008 20:42:57.572660 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-dhcp-openstack-openstack-cell1-mbrs2" Oct 08 20:42:57 crc kubenswrapper[4988]: I1008 20:42:57.585639 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-dhcp-agent-neutron-config" Oct 08 20:42:57 crc kubenswrapper[4988]: I1008 20:42:57.585913 4988 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 08 20:42:57 crc kubenswrapper[4988]: I1008 20:42:57.586259 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Oct 08 20:42:57 crc kubenswrapper[4988]: I1008 20:42:57.586538 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-4sdzr" Oct 08 20:42:57 crc kubenswrapper[4988]: I1008 20:42:57.586800 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Oct 08 20:42:57 crc kubenswrapper[4988]: I1008 20:42:57.617939 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-dhcp-openstack-openstack-cell1-mbrs2"] Oct 08 20:42:57 crc kubenswrapper[4988]: I1008 20:42:57.706495 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-dhcp-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7a339659-4ba8-4fb5-99fa-760e2bef90a2-neutron-dhcp-combined-ca-bundle\") pod \"neutron-dhcp-openstack-openstack-cell1-mbrs2\" (UID: \"7a339659-4ba8-4fb5-99fa-760e2bef90a2\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-mbrs2" Oct 08 20:42:57 crc kubenswrapper[4988]: I1008 20:42:57.706781 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7a339659-4ba8-4fb5-99fa-760e2bef90a2-ssh-key\") pod \"neutron-dhcp-openstack-openstack-cell1-mbrs2\" (UID: \"7a339659-4ba8-4fb5-99fa-760e2bef90a2\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-mbrs2" Oct 08 20:42:57 crc kubenswrapper[4988]: I1008 20:42:57.706873 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7a339659-4ba8-4fb5-99fa-760e2bef90a2-inventory\") pod \"neutron-dhcp-openstack-openstack-cell1-mbrs2\" (UID: \"7a339659-4ba8-4fb5-99fa-760e2bef90a2\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-mbrs2" Oct 08 20:42:57 crc kubenswrapper[4988]: I1008 20:42:57.707038 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zrrc2\" (UniqueName: \"kubernetes.io/projected/7a339659-4ba8-4fb5-99fa-760e2bef90a2-kube-api-access-zrrc2\") pod \"neutron-dhcp-openstack-openstack-cell1-mbrs2\" (UID: \"7a339659-4ba8-4fb5-99fa-760e2bef90a2\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-mbrs2" Oct 08 20:42:57 crc kubenswrapper[4988]: I1008 20:42:57.707157 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-dhcp-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/7a339659-4ba8-4fb5-99fa-760e2bef90a2-neutron-dhcp-agent-neutron-config-0\") pod \"neutron-dhcp-openstack-openstack-cell1-mbrs2\" (UID: \"7a339659-4ba8-4fb5-99fa-760e2bef90a2\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-mbrs2" Oct 08 20:42:57 crc kubenswrapper[4988]: I1008 20:42:57.810433 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-dhcp-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/7a339659-4ba8-4fb5-99fa-760e2bef90a2-neutron-dhcp-agent-neutron-config-0\") pod \"neutron-dhcp-openstack-openstack-cell1-mbrs2\" (UID: \"7a339659-4ba8-4fb5-99fa-760e2bef90a2\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-mbrs2" Oct 08 20:42:57 crc kubenswrapper[4988]: I1008 20:42:57.810630 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-dhcp-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7a339659-4ba8-4fb5-99fa-760e2bef90a2-neutron-dhcp-combined-ca-bundle\") pod \"neutron-dhcp-openstack-openstack-cell1-mbrs2\" (UID: \"7a339659-4ba8-4fb5-99fa-760e2bef90a2\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-mbrs2" Oct 08 20:42:57 crc kubenswrapper[4988]: I1008 20:42:57.810689 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7a339659-4ba8-4fb5-99fa-760e2bef90a2-ssh-key\") pod \"neutron-dhcp-openstack-openstack-cell1-mbrs2\" (UID: \"7a339659-4ba8-4fb5-99fa-760e2bef90a2\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-mbrs2" Oct 08 20:42:57 crc kubenswrapper[4988]: I1008 20:42:57.810725 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7a339659-4ba8-4fb5-99fa-760e2bef90a2-inventory\") pod \"neutron-dhcp-openstack-openstack-cell1-mbrs2\" (UID: \"7a339659-4ba8-4fb5-99fa-760e2bef90a2\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-mbrs2" Oct 08 20:42:57 crc kubenswrapper[4988]: I1008 20:42:57.810885 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zrrc2\" (UniqueName: \"kubernetes.io/projected/7a339659-4ba8-4fb5-99fa-760e2bef90a2-kube-api-access-zrrc2\") pod \"neutron-dhcp-openstack-openstack-cell1-mbrs2\" (UID: \"7a339659-4ba8-4fb5-99fa-760e2bef90a2\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-mbrs2" Oct 08 20:42:57 crc kubenswrapper[4988]: I1008 20:42:57.816944 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-dhcp-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/7a339659-4ba8-4fb5-99fa-760e2bef90a2-neutron-dhcp-agent-neutron-config-0\") pod \"neutron-dhcp-openstack-openstack-cell1-mbrs2\" (UID: \"7a339659-4ba8-4fb5-99fa-760e2bef90a2\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-mbrs2" Oct 08 20:42:57 crc kubenswrapper[4988]: I1008 20:42:57.817485 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7a339659-4ba8-4fb5-99fa-760e2bef90a2-ssh-key\") pod \"neutron-dhcp-openstack-openstack-cell1-mbrs2\" (UID: \"7a339659-4ba8-4fb5-99fa-760e2bef90a2\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-mbrs2" Oct 08 20:42:57 crc kubenswrapper[4988]: I1008 20:42:57.817533 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7a339659-4ba8-4fb5-99fa-760e2bef90a2-inventory\") pod \"neutron-dhcp-openstack-openstack-cell1-mbrs2\" (UID: \"7a339659-4ba8-4fb5-99fa-760e2bef90a2\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-mbrs2" Oct 08 20:42:57 crc kubenswrapper[4988]: I1008 20:42:57.818622 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-dhcp-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7a339659-4ba8-4fb5-99fa-760e2bef90a2-neutron-dhcp-combined-ca-bundle\") pod \"neutron-dhcp-openstack-openstack-cell1-mbrs2\" (UID: \"7a339659-4ba8-4fb5-99fa-760e2bef90a2\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-mbrs2" Oct 08 20:42:57 crc kubenswrapper[4988]: I1008 20:42:57.844087 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zrrc2\" (UniqueName: \"kubernetes.io/projected/7a339659-4ba8-4fb5-99fa-760e2bef90a2-kube-api-access-zrrc2\") pod \"neutron-dhcp-openstack-openstack-cell1-mbrs2\" (UID: \"7a339659-4ba8-4fb5-99fa-760e2bef90a2\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-mbrs2" Oct 08 20:42:57 crc kubenswrapper[4988]: I1008 20:42:57.908840 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-dhcp-openstack-openstack-cell1-mbrs2" Oct 08 20:42:58 crc kubenswrapper[4988]: I1008 20:42:58.541741 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-dhcp-openstack-openstack-cell1-mbrs2"] Oct 08 20:42:58 crc kubenswrapper[4988]: W1008 20:42:58.550167 4988 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7a339659_4ba8_4fb5_99fa_760e2bef90a2.slice/crio-b87a33dfcd4e4bf0e14580593e48b609161a530d038d93c31ddc44ea7e9f5711 WatchSource:0}: Error finding container b87a33dfcd4e4bf0e14580593e48b609161a530d038d93c31ddc44ea7e9f5711: Status 404 returned error can't find the container with id b87a33dfcd4e4bf0e14580593e48b609161a530d038d93c31ddc44ea7e9f5711 Oct 08 20:42:59 crc kubenswrapper[4988]: I1008 20:42:59.460056 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-dhcp-openstack-openstack-cell1-mbrs2" event={"ID":"7a339659-4ba8-4fb5-99fa-760e2bef90a2","Type":"ContainerStarted","Data":"4cbf674f1eceac170f689ec46f8d4cc5c78451b4027db0171cc8ea8329456885"} Oct 08 20:42:59 crc kubenswrapper[4988]: I1008 20:42:59.460867 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-dhcp-openstack-openstack-cell1-mbrs2" event={"ID":"7a339659-4ba8-4fb5-99fa-760e2bef90a2","Type":"ContainerStarted","Data":"b87a33dfcd4e4bf0e14580593e48b609161a530d038d93c31ddc44ea7e9f5711"} Oct 08 20:42:59 crc kubenswrapper[4988]: I1008 20:42:59.483353 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-dhcp-openstack-openstack-cell1-mbrs2" podStartSLOduration=1.904271401 podStartE2EDuration="2.483331289s" podCreationTimestamp="2025-10-08 20:42:57 +0000 UTC" firstStartedPulling="2025-10-08 20:42:58.553000603 +0000 UTC m=+9124.002843373" lastFinishedPulling="2025-10-08 20:42:59.132060451 +0000 UTC m=+9124.581903261" observedRunningTime="2025-10-08 20:42:59.479553361 +0000 UTC m=+9124.929396161" watchObservedRunningTime="2025-10-08 20:42:59.483331289 +0000 UTC m=+9124.933174070" Oct 08 20:43:13 crc kubenswrapper[4988]: I1008 20:43:13.559085 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-zfxhf"] Oct 08 20:43:13 crc kubenswrapper[4988]: I1008 20:43:13.564633 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-zfxhf" Oct 08 20:43:13 crc kubenswrapper[4988]: I1008 20:43:13.574273 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-zfxhf"] Oct 08 20:43:13 crc kubenswrapper[4988]: I1008 20:43:13.667464 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f8c17a29-6d49-4db3-a59c-dfc9651863cf-utilities\") pod \"community-operators-zfxhf\" (UID: \"f8c17a29-6d49-4db3-a59c-dfc9651863cf\") " pod="openshift-marketplace/community-operators-zfxhf" Oct 08 20:43:13 crc kubenswrapper[4988]: I1008 20:43:13.667545 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v9l75\" (UniqueName: \"kubernetes.io/projected/f8c17a29-6d49-4db3-a59c-dfc9651863cf-kube-api-access-v9l75\") pod \"community-operators-zfxhf\" (UID: \"f8c17a29-6d49-4db3-a59c-dfc9651863cf\") " pod="openshift-marketplace/community-operators-zfxhf" Oct 08 20:43:13 crc kubenswrapper[4988]: I1008 20:43:13.667631 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f8c17a29-6d49-4db3-a59c-dfc9651863cf-catalog-content\") pod \"community-operators-zfxhf\" (UID: \"f8c17a29-6d49-4db3-a59c-dfc9651863cf\") " pod="openshift-marketplace/community-operators-zfxhf" Oct 08 20:43:13 crc kubenswrapper[4988]: I1008 20:43:13.769732 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f8c17a29-6d49-4db3-a59c-dfc9651863cf-catalog-content\") pod \"community-operators-zfxhf\" (UID: \"f8c17a29-6d49-4db3-a59c-dfc9651863cf\") " pod="openshift-marketplace/community-operators-zfxhf" Oct 08 20:43:13 crc kubenswrapper[4988]: I1008 20:43:13.769874 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f8c17a29-6d49-4db3-a59c-dfc9651863cf-utilities\") pod \"community-operators-zfxhf\" (UID: \"f8c17a29-6d49-4db3-a59c-dfc9651863cf\") " pod="openshift-marketplace/community-operators-zfxhf" Oct 08 20:43:13 crc kubenswrapper[4988]: I1008 20:43:13.769953 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v9l75\" (UniqueName: \"kubernetes.io/projected/f8c17a29-6d49-4db3-a59c-dfc9651863cf-kube-api-access-v9l75\") pod \"community-operators-zfxhf\" (UID: \"f8c17a29-6d49-4db3-a59c-dfc9651863cf\") " pod="openshift-marketplace/community-operators-zfxhf" Oct 08 20:43:13 crc kubenswrapper[4988]: I1008 20:43:13.770306 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f8c17a29-6d49-4db3-a59c-dfc9651863cf-catalog-content\") pod \"community-operators-zfxhf\" (UID: \"f8c17a29-6d49-4db3-a59c-dfc9651863cf\") " pod="openshift-marketplace/community-operators-zfxhf" Oct 08 20:43:13 crc kubenswrapper[4988]: I1008 20:43:13.770485 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f8c17a29-6d49-4db3-a59c-dfc9651863cf-utilities\") pod \"community-operators-zfxhf\" (UID: \"f8c17a29-6d49-4db3-a59c-dfc9651863cf\") " pod="openshift-marketplace/community-operators-zfxhf" Oct 08 20:43:13 crc kubenswrapper[4988]: I1008 20:43:13.787444 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v9l75\" (UniqueName: \"kubernetes.io/projected/f8c17a29-6d49-4db3-a59c-dfc9651863cf-kube-api-access-v9l75\") pod \"community-operators-zfxhf\" (UID: \"f8c17a29-6d49-4db3-a59c-dfc9651863cf\") " pod="openshift-marketplace/community-operators-zfxhf" Oct 08 20:43:13 crc kubenswrapper[4988]: I1008 20:43:13.903348 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-zfxhf" Oct 08 20:43:14 crc kubenswrapper[4988]: I1008 20:43:14.478492 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-zfxhf"] Oct 08 20:43:14 crc kubenswrapper[4988]: I1008 20:43:14.570617 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-kqnnh"] Oct 08 20:43:14 crc kubenswrapper[4988]: I1008 20:43:14.577481 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-kqnnh" Oct 08 20:43:14 crc kubenswrapper[4988]: I1008 20:43:14.597573 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-kqnnh"] Oct 08 20:43:14 crc kubenswrapper[4988]: I1008 20:43:14.663511 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zfxhf" event={"ID":"f8c17a29-6d49-4db3-a59c-dfc9651863cf","Type":"ContainerStarted","Data":"e746168bee1f4d88a6b75655aedb61ea31deb628b45c5794ecf749b83023a7dc"} Oct 08 20:43:14 crc kubenswrapper[4988]: I1008 20:43:14.696805 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b900c85d-f9a1-4ebc-a34d-472935b581a1-utilities\") pod \"certified-operators-kqnnh\" (UID: \"b900c85d-f9a1-4ebc-a34d-472935b581a1\") " pod="openshift-marketplace/certified-operators-kqnnh" Oct 08 20:43:14 crc kubenswrapper[4988]: I1008 20:43:14.696912 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6q696\" (UniqueName: \"kubernetes.io/projected/b900c85d-f9a1-4ebc-a34d-472935b581a1-kube-api-access-6q696\") pod \"certified-operators-kqnnh\" (UID: \"b900c85d-f9a1-4ebc-a34d-472935b581a1\") " pod="openshift-marketplace/certified-operators-kqnnh" Oct 08 20:43:14 crc kubenswrapper[4988]: I1008 20:43:14.696953 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b900c85d-f9a1-4ebc-a34d-472935b581a1-catalog-content\") pod \"certified-operators-kqnnh\" (UID: \"b900c85d-f9a1-4ebc-a34d-472935b581a1\") " pod="openshift-marketplace/certified-operators-kqnnh" Oct 08 20:43:14 crc kubenswrapper[4988]: I1008 20:43:14.799969 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b900c85d-f9a1-4ebc-a34d-472935b581a1-utilities\") pod \"certified-operators-kqnnh\" (UID: \"b900c85d-f9a1-4ebc-a34d-472935b581a1\") " pod="openshift-marketplace/certified-operators-kqnnh" Oct 08 20:43:14 crc kubenswrapper[4988]: I1008 20:43:14.800130 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6q696\" (UniqueName: \"kubernetes.io/projected/b900c85d-f9a1-4ebc-a34d-472935b581a1-kube-api-access-6q696\") pod \"certified-operators-kqnnh\" (UID: \"b900c85d-f9a1-4ebc-a34d-472935b581a1\") " pod="openshift-marketplace/certified-operators-kqnnh" Oct 08 20:43:14 crc kubenswrapper[4988]: I1008 20:43:14.800200 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b900c85d-f9a1-4ebc-a34d-472935b581a1-catalog-content\") pod \"certified-operators-kqnnh\" (UID: \"b900c85d-f9a1-4ebc-a34d-472935b581a1\") " pod="openshift-marketplace/certified-operators-kqnnh" Oct 08 20:43:14 crc kubenswrapper[4988]: I1008 20:43:14.800886 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b900c85d-f9a1-4ebc-a34d-472935b581a1-catalog-content\") pod \"certified-operators-kqnnh\" (UID: \"b900c85d-f9a1-4ebc-a34d-472935b581a1\") " pod="openshift-marketplace/certified-operators-kqnnh" Oct 08 20:43:14 crc kubenswrapper[4988]: I1008 20:43:14.800993 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b900c85d-f9a1-4ebc-a34d-472935b581a1-utilities\") pod \"certified-operators-kqnnh\" (UID: \"b900c85d-f9a1-4ebc-a34d-472935b581a1\") " pod="openshift-marketplace/certified-operators-kqnnh" Oct 08 20:43:14 crc kubenswrapper[4988]: I1008 20:43:14.822748 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6q696\" (UniqueName: \"kubernetes.io/projected/b900c85d-f9a1-4ebc-a34d-472935b581a1-kube-api-access-6q696\") pod \"certified-operators-kqnnh\" (UID: \"b900c85d-f9a1-4ebc-a34d-472935b581a1\") " pod="openshift-marketplace/certified-operators-kqnnh" Oct 08 20:43:14 crc kubenswrapper[4988]: I1008 20:43:14.943365 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-kqnnh" Oct 08 20:43:15 crc kubenswrapper[4988]: I1008 20:43:15.442644 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-kqnnh"] Oct 08 20:43:15 crc kubenswrapper[4988]: W1008 20:43:15.455742 4988 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb900c85d_f9a1_4ebc_a34d_472935b581a1.slice/crio-4e68db86d47e3c0b291c66aa94429bd2f30c7b7aab84d06524fda8b7e68a6eaa WatchSource:0}: Error finding container 4e68db86d47e3c0b291c66aa94429bd2f30c7b7aab84d06524fda8b7e68a6eaa: Status 404 returned error can't find the container with id 4e68db86d47e3c0b291c66aa94429bd2f30c7b7aab84d06524fda8b7e68a6eaa Oct 08 20:43:15 crc kubenswrapper[4988]: I1008 20:43:15.685770 4988 generic.go:334] "Generic (PLEG): container finished" podID="f8c17a29-6d49-4db3-a59c-dfc9651863cf" containerID="e9fd5a349342ec9a381f86557fba4f6863ad6467e63a77ed3af054ecc04d633c" exitCode=0 Oct 08 20:43:15 crc kubenswrapper[4988]: I1008 20:43:15.685874 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zfxhf" event={"ID":"f8c17a29-6d49-4db3-a59c-dfc9651863cf","Type":"ContainerDied","Data":"e9fd5a349342ec9a381f86557fba4f6863ad6467e63a77ed3af054ecc04d633c"} Oct 08 20:43:15 crc kubenswrapper[4988]: I1008 20:43:15.689304 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kqnnh" event={"ID":"b900c85d-f9a1-4ebc-a34d-472935b581a1","Type":"ContainerStarted","Data":"4e68db86d47e3c0b291c66aa94429bd2f30c7b7aab84d06524fda8b7e68a6eaa"} Oct 08 20:43:16 crc kubenswrapper[4988]: I1008 20:43:16.700667 4988 generic.go:334] "Generic (PLEG): container finished" podID="b900c85d-f9a1-4ebc-a34d-472935b581a1" containerID="fdd709eb8c42932c3fe4b77d8949e81d77cb9cda035a2896f6469cde8c9c18b7" exitCode=0 Oct 08 20:43:16 crc kubenswrapper[4988]: I1008 20:43:16.702069 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kqnnh" event={"ID":"b900c85d-f9a1-4ebc-a34d-472935b581a1","Type":"ContainerDied","Data":"fdd709eb8c42932c3fe4b77d8949e81d77cb9cda035a2896f6469cde8c9c18b7"} Oct 08 20:43:17 crc kubenswrapper[4988]: I1008 20:43:17.713336 4988 generic.go:334] "Generic (PLEG): container finished" podID="f8c17a29-6d49-4db3-a59c-dfc9651863cf" containerID="1e1c65e2621e63ec31302970f53bc04889693827324c2f8a2f5aba2f05632bf0" exitCode=0 Oct 08 20:43:17 crc kubenswrapper[4988]: I1008 20:43:17.715543 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zfxhf" event={"ID":"f8c17a29-6d49-4db3-a59c-dfc9651863cf","Type":"ContainerDied","Data":"1e1c65e2621e63ec31302970f53bc04889693827324c2f8a2f5aba2f05632bf0"} Oct 08 20:43:17 crc kubenswrapper[4988]: I1008 20:43:17.718064 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kqnnh" event={"ID":"b900c85d-f9a1-4ebc-a34d-472935b581a1","Type":"ContainerStarted","Data":"2949743f49a88ff215afb97bc6cc603fed1204fe9baf1035f2478954f032e501"} Oct 08 20:43:18 crc kubenswrapper[4988]: I1008 20:43:18.730289 4988 generic.go:334] "Generic (PLEG): container finished" podID="b900c85d-f9a1-4ebc-a34d-472935b581a1" containerID="2949743f49a88ff215afb97bc6cc603fed1204fe9baf1035f2478954f032e501" exitCode=0 Oct 08 20:43:18 crc kubenswrapper[4988]: I1008 20:43:18.730420 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kqnnh" event={"ID":"b900c85d-f9a1-4ebc-a34d-472935b581a1","Type":"ContainerDied","Data":"2949743f49a88ff215afb97bc6cc603fed1204fe9baf1035f2478954f032e501"} Oct 08 20:43:18 crc kubenswrapper[4988]: I1008 20:43:18.738506 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zfxhf" event={"ID":"f8c17a29-6d49-4db3-a59c-dfc9651863cf","Type":"ContainerStarted","Data":"6ed6cfd1bc25ea5765376cd9f3d48a33a19934bfaa03fda2a1f80b3ff225bb1e"} Oct 08 20:43:18 crc kubenswrapper[4988]: I1008 20:43:18.796056 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-zfxhf" podStartSLOduration=3.292644376 podStartE2EDuration="5.796036412s" podCreationTimestamp="2025-10-08 20:43:13 +0000 UTC" firstStartedPulling="2025-10-08 20:43:15.688609675 +0000 UTC m=+9141.138452445" lastFinishedPulling="2025-10-08 20:43:18.192001701 +0000 UTC m=+9143.641844481" observedRunningTime="2025-10-08 20:43:18.7892445 +0000 UTC m=+9144.239087290" watchObservedRunningTime="2025-10-08 20:43:18.796036412 +0000 UTC m=+9144.245879182" Oct 08 20:43:19 crc kubenswrapper[4988]: I1008 20:43:19.748603 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kqnnh" event={"ID":"b900c85d-f9a1-4ebc-a34d-472935b581a1","Type":"ContainerStarted","Data":"c5440984bd37480bd7d7409f7f36560243cdb295b62724b62f6a405fc45f51d0"} Oct 08 20:43:19 crc kubenswrapper[4988]: I1008 20:43:19.771863 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-kqnnh" podStartSLOduration=3.25608404 podStartE2EDuration="5.771844373s" podCreationTimestamp="2025-10-08 20:43:14 +0000 UTC" firstStartedPulling="2025-10-08 20:43:16.702471127 +0000 UTC m=+9142.152313897" lastFinishedPulling="2025-10-08 20:43:19.21823145 +0000 UTC m=+9144.668074230" observedRunningTime="2025-10-08 20:43:19.761889531 +0000 UTC m=+9145.211732301" watchObservedRunningTime="2025-10-08 20:43:19.771844373 +0000 UTC m=+9145.221687143" Oct 08 20:43:23 crc kubenswrapper[4988]: I1008 20:43:23.337659 4988 patch_prober.go:28] interesting pod/machine-config-daemon-nm54f container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 20:43:23 crc kubenswrapper[4988]: I1008 20:43:23.338136 4988 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 20:43:23 crc kubenswrapper[4988]: I1008 20:43:23.903827 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-zfxhf" Oct 08 20:43:23 crc kubenswrapper[4988]: I1008 20:43:23.903889 4988 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-zfxhf" Oct 08 20:43:23 crc kubenswrapper[4988]: I1008 20:43:23.971566 4988 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-zfxhf" Oct 08 20:43:24 crc kubenswrapper[4988]: I1008 20:43:24.914799 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-zfxhf" Oct 08 20:43:24 crc kubenswrapper[4988]: I1008 20:43:24.943651 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-kqnnh" Oct 08 20:43:24 crc kubenswrapper[4988]: I1008 20:43:24.943709 4988 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-kqnnh" Oct 08 20:43:24 crc kubenswrapper[4988]: I1008 20:43:24.956339 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-zfxhf"] Oct 08 20:43:24 crc kubenswrapper[4988]: I1008 20:43:24.987464 4988 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-kqnnh" Oct 08 20:43:25 crc kubenswrapper[4988]: I1008 20:43:25.866751 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-kqnnh" Oct 08 20:43:26 crc kubenswrapper[4988]: I1008 20:43:26.732584 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-kqnnh"] Oct 08 20:43:26 crc kubenswrapper[4988]: I1008 20:43:26.820484 4988 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-zfxhf" podUID="f8c17a29-6d49-4db3-a59c-dfc9651863cf" containerName="registry-server" containerID="cri-o://6ed6cfd1bc25ea5765376cd9f3d48a33a19934bfaa03fda2a1f80b3ff225bb1e" gracePeriod=2 Oct 08 20:43:27 crc kubenswrapper[4988]: I1008 20:43:27.331508 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-zfxhf" Oct 08 20:43:27 crc kubenswrapper[4988]: I1008 20:43:27.430482 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f8c17a29-6d49-4db3-a59c-dfc9651863cf-utilities\") pod \"f8c17a29-6d49-4db3-a59c-dfc9651863cf\" (UID: \"f8c17a29-6d49-4db3-a59c-dfc9651863cf\") " Oct 08 20:43:27 crc kubenswrapper[4988]: I1008 20:43:27.430759 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v9l75\" (UniqueName: \"kubernetes.io/projected/f8c17a29-6d49-4db3-a59c-dfc9651863cf-kube-api-access-v9l75\") pod \"f8c17a29-6d49-4db3-a59c-dfc9651863cf\" (UID: \"f8c17a29-6d49-4db3-a59c-dfc9651863cf\") " Oct 08 20:43:27 crc kubenswrapper[4988]: I1008 20:43:27.430817 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f8c17a29-6d49-4db3-a59c-dfc9651863cf-catalog-content\") pod \"f8c17a29-6d49-4db3-a59c-dfc9651863cf\" (UID: \"f8c17a29-6d49-4db3-a59c-dfc9651863cf\") " Oct 08 20:43:27 crc kubenswrapper[4988]: I1008 20:43:27.431172 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f8c17a29-6d49-4db3-a59c-dfc9651863cf-utilities" (OuterVolumeSpecName: "utilities") pod "f8c17a29-6d49-4db3-a59c-dfc9651863cf" (UID: "f8c17a29-6d49-4db3-a59c-dfc9651863cf"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 20:43:27 crc kubenswrapper[4988]: I1008 20:43:27.436627 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f8c17a29-6d49-4db3-a59c-dfc9651863cf-kube-api-access-v9l75" (OuterVolumeSpecName: "kube-api-access-v9l75") pod "f8c17a29-6d49-4db3-a59c-dfc9651863cf" (UID: "f8c17a29-6d49-4db3-a59c-dfc9651863cf"). InnerVolumeSpecName "kube-api-access-v9l75". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 20:43:27 crc kubenswrapper[4988]: I1008 20:43:27.484969 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f8c17a29-6d49-4db3-a59c-dfc9651863cf-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "f8c17a29-6d49-4db3-a59c-dfc9651863cf" (UID: "f8c17a29-6d49-4db3-a59c-dfc9651863cf"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 20:43:27 crc kubenswrapper[4988]: I1008 20:43:27.533556 4988 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v9l75\" (UniqueName: \"kubernetes.io/projected/f8c17a29-6d49-4db3-a59c-dfc9651863cf-kube-api-access-v9l75\") on node \"crc\" DevicePath \"\"" Oct 08 20:43:27 crc kubenswrapper[4988]: I1008 20:43:27.533605 4988 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f8c17a29-6d49-4db3-a59c-dfc9651863cf-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 08 20:43:27 crc kubenswrapper[4988]: I1008 20:43:27.533621 4988 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f8c17a29-6d49-4db3-a59c-dfc9651863cf-utilities\") on node \"crc\" DevicePath \"\"" Oct 08 20:43:27 crc kubenswrapper[4988]: I1008 20:43:27.837606 4988 generic.go:334] "Generic (PLEG): container finished" podID="f8c17a29-6d49-4db3-a59c-dfc9651863cf" containerID="6ed6cfd1bc25ea5765376cd9f3d48a33a19934bfaa03fda2a1f80b3ff225bb1e" exitCode=0 Oct 08 20:43:27 crc kubenswrapper[4988]: I1008 20:43:27.837680 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zfxhf" event={"ID":"f8c17a29-6d49-4db3-a59c-dfc9651863cf","Type":"ContainerDied","Data":"6ed6cfd1bc25ea5765376cd9f3d48a33a19934bfaa03fda2a1f80b3ff225bb1e"} Oct 08 20:43:27 crc kubenswrapper[4988]: I1008 20:43:27.837778 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zfxhf" event={"ID":"f8c17a29-6d49-4db3-a59c-dfc9651863cf","Type":"ContainerDied","Data":"e746168bee1f4d88a6b75655aedb61ea31deb628b45c5794ecf749b83023a7dc"} Oct 08 20:43:27 crc kubenswrapper[4988]: I1008 20:43:27.837810 4988 scope.go:117] "RemoveContainer" containerID="6ed6cfd1bc25ea5765376cd9f3d48a33a19934bfaa03fda2a1f80b3ff225bb1e" Oct 08 20:43:27 crc kubenswrapper[4988]: I1008 20:43:27.838003 4988 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-kqnnh" podUID="b900c85d-f9a1-4ebc-a34d-472935b581a1" containerName="registry-server" containerID="cri-o://c5440984bd37480bd7d7409f7f36560243cdb295b62724b62f6a405fc45f51d0" gracePeriod=2 Oct 08 20:43:27 crc kubenswrapper[4988]: I1008 20:43:27.838789 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-zfxhf" Oct 08 20:43:27 crc kubenswrapper[4988]: I1008 20:43:27.890657 4988 scope.go:117] "RemoveContainer" containerID="1e1c65e2621e63ec31302970f53bc04889693827324c2f8a2f5aba2f05632bf0" Oct 08 20:43:27 crc kubenswrapper[4988]: I1008 20:43:27.915172 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-zfxhf"] Oct 08 20:43:27 crc kubenswrapper[4988]: I1008 20:43:27.926431 4988 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-zfxhf"] Oct 08 20:43:27 crc kubenswrapper[4988]: I1008 20:43:27.936912 4988 scope.go:117] "RemoveContainer" containerID="e9fd5a349342ec9a381f86557fba4f6863ad6467e63a77ed3af054ecc04d633c" Oct 08 20:43:28 crc kubenswrapper[4988]: I1008 20:43:28.112538 4988 scope.go:117] "RemoveContainer" containerID="6ed6cfd1bc25ea5765376cd9f3d48a33a19934bfaa03fda2a1f80b3ff225bb1e" Oct 08 20:43:28 crc kubenswrapper[4988]: E1008 20:43:28.113277 4988 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6ed6cfd1bc25ea5765376cd9f3d48a33a19934bfaa03fda2a1f80b3ff225bb1e\": container with ID starting with 6ed6cfd1bc25ea5765376cd9f3d48a33a19934bfaa03fda2a1f80b3ff225bb1e not found: ID does not exist" containerID="6ed6cfd1bc25ea5765376cd9f3d48a33a19934bfaa03fda2a1f80b3ff225bb1e" Oct 08 20:43:28 crc kubenswrapper[4988]: I1008 20:43:28.113317 4988 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6ed6cfd1bc25ea5765376cd9f3d48a33a19934bfaa03fda2a1f80b3ff225bb1e"} err="failed to get container status \"6ed6cfd1bc25ea5765376cd9f3d48a33a19934bfaa03fda2a1f80b3ff225bb1e\": rpc error: code = NotFound desc = could not find container \"6ed6cfd1bc25ea5765376cd9f3d48a33a19934bfaa03fda2a1f80b3ff225bb1e\": container with ID starting with 6ed6cfd1bc25ea5765376cd9f3d48a33a19934bfaa03fda2a1f80b3ff225bb1e not found: ID does not exist" Oct 08 20:43:28 crc kubenswrapper[4988]: I1008 20:43:28.113344 4988 scope.go:117] "RemoveContainer" containerID="1e1c65e2621e63ec31302970f53bc04889693827324c2f8a2f5aba2f05632bf0" Oct 08 20:43:28 crc kubenswrapper[4988]: E1008 20:43:28.113813 4988 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1e1c65e2621e63ec31302970f53bc04889693827324c2f8a2f5aba2f05632bf0\": container with ID starting with 1e1c65e2621e63ec31302970f53bc04889693827324c2f8a2f5aba2f05632bf0 not found: ID does not exist" containerID="1e1c65e2621e63ec31302970f53bc04889693827324c2f8a2f5aba2f05632bf0" Oct 08 20:43:28 crc kubenswrapper[4988]: I1008 20:43:28.113861 4988 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1e1c65e2621e63ec31302970f53bc04889693827324c2f8a2f5aba2f05632bf0"} err="failed to get container status \"1e1c65e2621e63ec31302970f53bc04889693827324c2f8a2f5aba2f05632bf0\": rpc error: code = NotFound desc = could not find container \"1e1c65e2621e63ec31302970f53bc04889693827324c2f8a2f5aba2f05632bf0\": container with ID starting with 1e1c65e2621e63ec31302970f53bc04889693827324c2f8a2f5aba2f05632bf0 not found: ID does not exist" Oct 08 20:43:28 crc kubenswrapper[4988]: I1008 20:43:28.113894 4988 scope.go:117] "RemoveContainer" containerID="e9fd5a349342ec9a381f86557fba4f6863ad6467e63a77ed3af054ecc04d633c" Oct 08 20:43:28 crc kubenswrapper[4988]: E1008 20:43:28.114360 4988 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e9fd5a349342ec9a381f86557fba4f6863ad6467e63a77ed3af054ecc04d633c\": container with ID starting with e9fd5a349342ec9a381f86557fba4f6863ad6467e63a77ed3af054ecc04d633c not found: ID does not exist" containerID="e9fd5a349342ec9a381f86557fba4f6863ad6467e63a77ed3af054ecc04d633c" Oct 08 20:43:28 crc kubenswrapper[4988]: I1008 20:43:28.114415 4988 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e9fd5a349342ec9a381f86557fba4f6863ad6467e63a77ed3af054ecc04d633c"} err="failed to get container status \"e9fd5a349342ec9a381f86557fba4f6863ad6467e63a77ed3af054ecc04d633c\": rpc error: code = NotFound desc = could not find container \"e9fd5a349342ec9a381f86557fba4f6863ad6467e63a77ed3af054ecc04d633c\": container with ID starting with e9fd5a349342ec9a381f86557fba4f6863ad6467e63a77ed3af054ecc04d633c not found: ID does not exist" Oct 08 20:43:28 crc kubenswrapper[4988]: I1008 20:43:28.421080 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-kqnnh" Oct 08 20:43:28 crc kubenswrapper[4988]: I1008 20:43:28.559634 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6q696\" (UniqueName: \"kubernetes.io/projected/b900c85d-f9a1-4ebc-a34d-472935b581a1-kube-api-access-6q696\") pod \"b900c85d-f9a1-4ebc-a34d-472935b581a1\" (UID: \"b900c85d-f9a1-4ebc-a34d-472935b581a1\") " Oct 08 20:43:28 crc kubenswrapper[4988]: I1008 20:43:28.559701 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b900c85d-f9a1-4ebc-a34d-472935b581a1-utilities\") pod \"b900c85d-f9a1-4ebc-a34d-472935b581a1\" (UID: \"b900c85d-f9a1-4ebc-a34d-472935b581a1\") " Oct 08 20:43:28 crc kubenswrapper[4988]: I1008 20:43:28.559756 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b900c85d-f9a1-4ebc-a34d-472935b581a1-catalog-content\") pod \"b900c85d-f9a1-4ebc-a34d-472935b581a1\" (UID: \"b900c85d-f9a1-4ebc-a34d-472935b581a1\") " Oct 08 20:43:28 crc kubenswrapper[4988]: I1008 20:43:28.560761 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b900c85d-f9a1-4ebc-a34d-472935b581a1-utilities" (OuterVolumeSpecName: "utilities") pod "b900c85d-f9a1-4ebc-a34d-472935b581a1" (UID: "b900c85d-f9a1-4ebc-a34d-472935b581a1"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 20:43:28 crc kubenswrapper[4988]: I1008 20:43:28.569824 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b900c85d-f9a1-4ebc-a34d-472935b581a1-kube-api-access-6q696" (OuterVolumeSpecName: "kube-api-access-6q696") pod "b900c85d-f9a1-4ebc-a34d-472935b581a1" (UID: "b900c85d-f9a1-4ebc-a34d-472935b581a1"). InnerVolumeSpecName "kube-api-access-6q696". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 20:43:28 crc kubenswrapper[4988]: I1008 20:43:28.602796 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b900c85d-f9a1-4ebc-a34d-472935b581a1-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b900c85d-f9a1-4ebc-a34d-472935b581a1" (UID: "b900c85d-f9a1-4ebc-a34d-472935b581a1"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 20:43:28 crc kubenswrapper[4988]: I1008 20:43:28.661912 4988 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6q696\" (UniqueName: \"kubernetes.io/projected/b900c85d-f9a1-4ebc-a34d-472935b581a1-kube-api-access-6q696\") on node \"crc\" DevicePath \"\"" Oct 08 20:43:28 crc kubenswrapper[4988]: I1008 20:43:28.662271 4988 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b900c85d-f9a1-4ebc-a34d-472935b581a1-utilities\") on node \"crc\" DevicePath \"\"" Oct 08 20:43:28 crc kubenswrapper[4988]: I1008 20:43:28.662281 4988 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b900c85d-f9a1-4ebc-a34d-472935b581a1-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 08 20:43:28 crc kubenswrapper[4988]: I1008 20:43:28.852762 4988 generic.go:334] "Generic (PLEG): container finished" podID="b900c85d-f9a1-4ebc-a34d-472935b581a1" containerID="c5440984bd37480bd7d7409f7f36560243cdb295b62724b62f6a405fc45f51d0" exitCode=0 Oct 08 20:43:28 crc kubenswrapper[4988]: I1008 20:43:28.852814 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-kqnnh" Oct 08 20:43:28 crc kubenswrapper[4988]: I1008 20:43:28.852811 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kqnnh" event={"ID":"b900c85d-f9a1-4ebc-a34d-472935b581a1","Type":"ContainerDied","Data":"c5440984bd37480bd7d7409f7f36560243cdb295b62724b62f6a405fc45f51d0"} Oct 08 20:43:28 crc kubenswrapper[4988]: I1008 20:43:28.853009 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kqnnh" event={"ID":"b900c85d-f9a1-4ebc-a34d-472935b581a1","Type":"ContainerDied","Data":"4e68db86d47e3c0b291c66aa94429bd2f30c7b7aab84d06524fda8b7e68a6eaa"} Oct 08 20:43:28 crc kubenswrapper[4988]: I1008 20:43:28.853083 4988 scope.go:117] "RemoveContainer" containerID="c5440984bd37480bd7d7409f7f36560243cdb295b62724b62f6a405fc45f51d0" Oct 08 20:43:28 crc kubenswrapper[4988]: I1008 20:43:28.888911 4988 scope.go:117] "RemoveContainer" containerID="2949743f49a88ff215afb97bc6cc603fed1204fe9baf1035f2478954f032e501" Oct 08 20:43:28 crc kubenswrapper[4988]: I1008 20:43:28.895085 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-kqnnh"] Oct 08 20:43:28 crc kubenswrapper[4988]: I1008 20:43:28.906957 4988 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-kqnnh"] Oct 08 20:43:28 crc kubenswrapper[4988]: I1008 20:43:28.931668 4988 scope.go:117] "RemoveContainer" containerID="fdd709eb8c42932c3fe4b77d8949e81d77cb9cda035a2896f6469cde8c9c18b7" Oct 08 20:43:28 crc kubenswrapper[4988]: I1008 20:43:28.964590 4988 scope.go:117] "RemoveContainer" containerID="c5440984bd37480bd7d7409f7f36560243cdb295b62724b62f6a405fc45f51d0" Oct 08 20:43:28 crc kubenswrapper[4988]: E1008 20:43:28.965134 4988 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c5440984bd37480bd7d7409f7f36560243cdb295b62724b62f6a405fc45f51d0\": container with ID starting with c5440984bd37480bd7d7409f7f36560243cdb295b62724b62f6a405fc45f51d0 not found: ID does not exist" containerID="c5440984bd37480bd7d7409f7f36560243cdb295b62724b62f6a405fc45f51d0" Oct 08 20:43:28 crc kubenswrapper[4988]: I1008 20:43:28.965163 4988 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c5440984bd37480bd7d7409f7f36560243cdb295b62724b62f6a405fc45f51d0"} err="failed to get container status \"c5440984bd37480bd7d7409f7f36560243cdb295b62724b62f6a405fc45f51d0\": rpc error: code = NotFound desc = could not find container \"c5440984bd37480bd7d7409f7f36560243cdb295b62724b62f6a405fc45f51d0\": container with ID starting with c5440984bd37480bd7d7409f7f36560243cdb295b62724b62f6a405fc45f51d0 not found: ID does not exist" Oct 08 20:43:28 crc kubenswrapper[4988]: I1008 20:43:28.965183 4988 scope.go:117] "RemoveContainer" containerID="2949743f49a88ff215afb97bc6cc603fed1204fe9baf1035f2478954f032e501" Oct 08 20:43:28 crc kubenswrapper[4988]: E1008 20:43:28.965581 4988 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2949743f49a88ff215afb97bc6cc603fed1204fe9baf1035f2478954f032e501\": container with ID starting with 2949743f49a88ff215afb97bc6cc603fed1204fe9baf1035f2478954f032e501 not found: ID does not exist" containerID="2949743f49a88ff215afb97bc6cc603fed1204fe9baf1035f2478954f032e501" Oct 08 20:43:28 crc kubenswrapper[4988]: I1008 20:43:28.965610 4988 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2949743f49a88ff215afb97bc6cc603fed1204fe9baf1035f2478954f032e501"} err="failed to get container status \"2949743f49a88ff215afb97bc6cc603fed1204fe9baf1035f2478954f032e501\": rpc error: code = NotFound desc = could not find container \"2949743f49a88ff215afb97bc6cc603fed1204fe9baf1035f2478954f032e501\": container with ID starting with 2949743f49a88ff215afb97bc6cc603fed1204fe9baf1035f2478954f032e501 not found: ID does not exist" Oct 08 20:43:28 crc kubenswrapper[4988]: I1008 20:43:28.965631 4988 scope.go:117] "RemoveContainer" containerID="fdd709eb8c42932c3fe4b77d8949e81d77cb9cda035a2896f6469cde8c9c18b7" Oct 08 20:43:28 crc kubenswrapper[4988]: E1008 20:43:28.965976 4988 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fdd709eb8c42932c3fe4b77d8949e81d77cb9cda035a2896f6469cde8c9c18b7\": container with ID starting with fdd709eb8c42932c3fe4b77d8949e81d77cb9cda035a2896f6469cde8c9c18b7 not found: ID does not exist" containerID="fdd709eb8c42932c3fe4b77d8949e81d77cb9cda035a2896f6469cde8c9c18b7" Oct 08 20:43:28 crc kubenswrapper[4988]: I1008 20:43:28.966004 4988 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fdd709eb8c42932c3fe4b77d8949e81d77cb9cda035a2896f6469cde8c9c18b7"} err="failed to get container status \"fdd709eb8c42932c3fe4b77d8949e81d77cb9cda035a2896f6469cde8c9c18b7\": rpc error: code = NotFound desc = could not find container \"fdd709eb8c42932c3fe4b77d8949e81d77cb9cda035a2896f6469cde8c9c18b7\": container with ID starting with fdd709eb8c42932c3fe4b77d8949e81d77cb9cda035a2896f6469cde8c9c18b7 not found: ID does not exist" Oct 08 20:43:29 crc kubenswrapper[4988]: I1008 20:43:29.260683 4988 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b900c85d-f9a1-4ebc-a34d-472935b581a1" path="/var/lib/kubelet/pods/b900c85d-f9a1-4ebc-a34d-472935b581a1/volumes" Oct 08 20:43:29 crc kubenswrapper[4988]: I1008 20:43:29.261460 4988 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f8c17a29-6d49-4db3-a59c-dfc9651863cf" path="/var/lib/kubelet/pods/f8c17a29-6d49-4db3-a59c-dfc9651863cf/volumes" Oct 08 20:43:53 crc kubenswrapper[4988]: I1008 20:43:53.338412 4988 patch_prober.go:28] interesting pod/machine-config-daemon-nm54f container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 20:43:53 crc kubenswrapper[4988]: I1008 20:43:53.338841 4988 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 20:43:53 crc kubenswrapper[4988]: I1008 20:43:53.338883 4988 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" Oct 08 20:43:53 crc kubenswrapper[4988]: I1008 20:43:53.339638 4988 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"6382c087e916f3958809279cab4d35e45e3b3ccd0cb9d8f0da5a015741773071"} pod="openshift-machine-config-operator/machine-config-daemon-nm54f" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 08 20:43:53 crc kubenswrapper[4988]: I1008 20:43:53.339686 4988 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" containerName="machine-config-daemon" containerID="cri-o://6382c087e916f3958809279cab4d35e45e3b3ccd0cb9d8f0da5a015741773071" gracePeriod=600 Oct 08 20:43:53 crc kubenswrapper[4988]: E1008 20:43:53.460035 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nm54f_openshift-machine-config-operator(f328a20b-5283-407e-8a93-2ab2da599f18)\"" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" Oct 08 20:43:54 crc kubenswrapper[4988]: I1008 20:43:54.172499 4988 generic.go:334] "Generic (PLEG): container finished" podID="f328a20b-5283-407e-8a93-2ab2da599f18" containerID="6382c087e916f3958809279cab4d35e45e3b3ccd0cb9d8f0da5a015741773071" exitCode=0 Oct 08 20:43:54 crc kubenswrapper[4988]: I1008 20:43:54.172628 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" event={"ID":"f328a20b-5283-407e-8a93-2ab2da599f18","Type":"ContainerDied","Data":"6382c087e916f3958809279cab4d35e45e3b3ccd0cb9d8f0da5a015741773071"} Oct 08 20:43:54 crc kubenswrapper[4988]: I1008 20:43:54.172912 4988 scope.go:117] "RemoveContainer" containerID="f21c43696ae96fab4aca044e2adbdde2d544edfcba427c7247d210f2d406c66d" Oct 08 20:43:54 crc kubenswrapper[4988]: I1008 20:43:54.174256 4988 scope.go:117] "RemoveContainer" containerID="6382c087e916f3958809279cab4d35e45e3b3ccd0cb9d8f0da5a015741773071" Oct 08 20:43:54 crc kubenswrapper[4988]: E1008 20:43:54.174892 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nm54f_openshift-machine-config-operator(f328a20b-5283-407e-8a93-2ab2da599f18)\"" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" Oct 08 20:44:05 crc kubenswrapper[4988]: I1008 20:44:05.244166 4988 scope.go:117] "RemoveContainer" containerID="6382c087e916f3958809279cab4d35e45e3b3ccd0cb9d8f0da5a015741773071" Oct 08 20:44:05 crc kubenswrapper[4988]: E1008 20:44:05.244842 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nm54f_openshift-machine-config-operator(f328a20b-5283-407e-8a93-2ab2da599f18)\"" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" Oct 08 20:44:17 crc kubenswrapper[4988]: I1008 20:44:17.239337 4988 scope.go:117] "RemoveContainer" containerID="6382c087e916f3958809279cab4d35e45e3b3ccd0cb9d8f0da5a015741773071" Oct 08 20:44:17 crc kubenswrapper[4988]: E1008 20:44:17.240766 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nm54f_openshift-machine-config-operator(f328a20b-5283-407e-8a93-2ab2da599f18)\"" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" Oct 08 20:44:30 crc kubenswrapper[4988]: I1008 20:44:30.238472 4988 scope.go:117] "RemoveContainer" containerID="6382c087e916f3958809279cab4d35e45e3b3ccd0cb9d8f0da5a015741773071" Oct 08 20:44:30 crc kubenswrapper[4988]: E1008 20:44:30.239501 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nm54f_openshift-machine-config-operator(f328a20b-5283-407e-8a93-2ab2da599f18)\"" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" Oct 08 20:44:45 crc kubenswrapper[4988]: I1008 20:44:45.288580 4988 scope.go:117] "RemoveContainer" containerID="6382c087e916f3958809279cab4d35e45e3b3ccd0cb9d8f0da5a015741773071" Oct 08 20:44:45 crc kubenswrapper[4988]: E1008 20:44:45.289593 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nm54f_openshift-machine-config-operator(f328a20b-5283-407e-8a93-2ab2da599f18)\"" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" Oct 08 20:44:58 crc kubenswrapper[4988]: I1008 20:44:58.237513 4988 scope.go:117] "RemoveContainer" containerID="6382c087e916f3958809279cab4d35e45e3b3ccd0cb9d8f0da5a015741773071" Oct 08 20:44:58 crc kubenswrapper[4988]: E1008 20:44:58.238126 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nm54f_openshift-machine-config-operator(f328a20b-5283-407e-8a93-2ab2da599f18)\"" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" Oct 08 20:45:00 crc kubenswrapper[4988]: I1008 20:45:00.171072 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29332605-jgdxc"] Oct 08 20:45:00 crc kubenswrapper[4988]: E1008 20:45:00.172648 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f8c17a29-6d49-4db3-a59c-dfc9651863cf" containerName="extract-content" Oct 08 20:45:00 crc kubenswrapper[4988]: I1008 20:45:00.172671 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="f8c17a29-6d49-4db3-a59c-dfc9651863cf" containerName="extract-content" Oct 08 20:45:00 crc kubenswrapper[4988]: E1008 20:45:00.172698 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b900c85d-f9a1-4ebc-a34d-472935b581a1" containerName="extract-utilities" Oct 08 20:45:00 crc kubenswrapper[4988]: I1008 20:45:00.172707 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="b900c85d-f9a1-4ebc-a34d-472935b581a1" containerName="extract-utilities" Oct 08 20:45:00 crc kubenswrapper[4988]: E1008 20:45:00.172725 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b900c85d-f9a1-4ebc-a34d-472935b581a1" containerName="extract-content" Oct 08 20:45:00 crc kubenswrapper[4988]: I1008 20:45:00.172734 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="b900c85d-f9a1-4ebc-a34d-472935b581a1" containerName="extract-content" Oct 08 20:45:00 crc kubenswrapper[4988]: E1008 20:45:00.172759 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f8c17a29-6d49-4db3-a59c-dfc9651863cf" containerName="registry-server" Oct 08 20:45:00 crc kubenswrapper[4988]: I1008 20:45:00.172765 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="f8c17a29-6d49-4db3-a59c-dfc9651863cf" containerName="registry-server" Oct 08 20:45:00 crc kubenswrapper[4988]: E1008 20:45:00.172799 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f8c17a29-6d49-4db3-a59c-dfc9651863cf" containerName="extract-utilities" Oct 08 20:45:00 crc kubenswrapper[4988]: I1008 20:45:00.172805 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="f8c17a29-6d49-4db3-a59c-dfc9651863cf" containerName="extract-utilities" Oct 08 20:45:00 crc kubenswrapper[4988]: E1008 20:45:00.172825 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b900c85d-f9a1-4ebc-a34d-472935b581a1" containerName="registry-server" Oct 08 20:45:00 crc kubenswrapper[4988]: I1008 20:45:00.172837 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="b900c85d-f9a1-4ebc-a34d-472935b581a1" containerName="registry-server" Oct 08 20:45:00 crc kubenswrapper[4988]: I1008 20:45:00.173129 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="b900c85d-f9a1-4ebc-a34d-472935b581a1" containerName="registry-server" Oct 08 20:45:00 crc kubenswrapper[4988]: I1008 20:45:00.173149 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="f8c17a29-6d49-4db3-a59c-dfc9651863cf" containerName="registry-server" Oct 08 20:45:00 crc kubenswrapper[4988]: I1008 20:45:00.174432 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29332605-jgdxc" Oct 08 20:45:00 crc kubenswrapper[4988]: I1008 20:45:00.201571 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29332605-jgdxc"] Oct 08 20:45:00 crc kubenswrapper[4988]: I1008 20:45:00.206822 4988 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 08 20:45:00 crc kubenswrapper[4988]: I1008 20:45:00.207306 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 08 20:45:00 crc kubenswrapper[4988]: I1008 20:45:00.272214 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/faa56149-e2dc-4cb0-8038-9ce9c8332beb-config-volume\") pod \"collect-profiles-29332605-jgdxc\" (UID: \"faa56149-e2dc-4cb0-8038-9ce9c8332beb\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29332605-jgdxc" Oct 08 20:45:00 crc kubenswrapper[4988]: I1008 20:45:00.272311 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-58hp2\" (UniqueName: \"kubernetes.io/projected/faa56149-e2dc-4cb0-8038-9ce9c8332beb-kube-api-access-58hp2\") pod \"collect-profiles-29332605-jgdxc\" (UID: \"faa56149-e2dc-4cb0-8038-9ce9c8332beb\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29332605-jgdxc" Oct 08 20:45:00 crc kubenswrapper[4988]: I1008 20:45:00.272458 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/faa56149-e2dc-4cb0-8038-9ce9c8332beb-secret-volume\") pod \"collect-profiles-29332605-jgdxc\" (UID: \"faa56149-e2dc-4cb0-8038-9ce9c8332beb\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29332605-jgdxc" Oct 08 20:45:00 crc kubenswrapper[4988]: I1008 20:45:00.375375 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/faa56149-e2dc-4cb0-8038-9ce9c8332beb-config-volume\") pod \"collect-profiles-29332605-jgdxc\" (UID: \"faa56149-e2dc-4cb0-8038-9ce9c8332beb\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29332605-jgdxc" Oct 08 20:45:00 crc kubenswrapper[4988]: I1008 20:45:00.375462 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-58hp2\" (UniqueName: \"kubernetes.io/projected/faa56149-e2dc-4cb0-8038-9ce9c8332beb-kube-api-access-58hp2\") pod \"collect-profiles-29332605-jgdxc\" (UID: \"faa56149-e2dc-4cb0-8038-9ce9c8332beb\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29332605-jgdxc" Oct 08 20:45:00 crc kubenswrapper[4988]: I1008 20:45:00.375611 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/faa56149-e2dc-4cb0-8038-9ce9c8332beb-secret-volume\") pod \"collect-profiles-29332605-jgdxc\" (UID: \"faa56149-e2dc-4cb0-8038-9ce9c8332beb\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29332605-jgdxc" Oct 08 20:45:00 crc kubenswrapper[4988]: I1008 20:45:00.376514 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/faa56149-e2dc-4cb0-8038-9ce9c8332beb-config-volume\") pod \"collect-profiles-29332605-jgdxc\" (UID: \"faa56149-e2dc-4cb0-8038-9ce9c8332beb\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29332605-jgdxc" Oct 08 20:45:00 crc kubenswrapper[4988]: I1008 20:45:00.480747 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-58hp2\" (UniqueName: \"kubernetes.io/projected/faa56149-e2dc-4cb0-8038-9ce9c8332beb-kube-api-access-58hp2\") pod \"collect-profiles-29332605-jgdxc\" (UID: \"faa56149-e2dc-4cb0-8038-9ce9c8332beb\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29332605-jgdxc" Oct 08 20:45:00 crc kubenswrapper[4988]: I1008 20:45:00.481659 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/faa56149-e2dc-4cb0-8038-9ce9c8332beb-secret-volume\") pod \"collect-profiles-29332605-jgdxc\" (UID: \"faa56149-e2dc-4cb0-8038-9ce9c8332beb\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29332605-jgdxc" Oct 08 20:45:00 crc kubenswrapper[4988]: I1008 20:45:00.529511 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29332605-jgdxc" Oct 08 20:45:01 crc kubenswrapper[4988]: I1008 20:45:01.046361 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29332605-jgdxc"] Oct 08 20:45:02 crc kubenswrapper[4988]: I1008 20:45:02.073828 4988 generic.go:334] "Generic (PLEG): container finished" podID="faa56149-e2dc-4cb0-8038-9ce9c8332beb" containerID="c296a5a4150e29d06a2e8d66cc7f9e41c3d0bb916cb680644919741f81bb2c2b" exitCode=0 Oct 08 20:45:02 crc kubenswrapper[4988]: I1008 20:45:02.073902 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29332605-jgdxc" event={"ID":"faa56149-e2dc-4cb0-8038-9ce9c8332beb","Type":"ContainerDied","Data":"c296a5a4150e29d06a2e8d66cc7f9e41c3d0bb916cb680644919741f81bb2c2b"} Oct 08 20:45:02 crc kubenswrapper[4988]: I1008 20:45:02.074204 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29332605-jgdxc" event={"ID":"faa56149-e2dc-4cb0-8038-9ce9c8332beb","Type":"ContainerStarted","Data":"2fb96f8cafe887221560a61e1d1f2127ddc855d686b318a2e1637804cceeea58"} Oct 08 20:45:03 crc kubenswrapper[4988]: I1008 20:45:03.560256 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29332605-jgdxc" Oct 08 20:45:03 crc kubenswrapper[4988]: I1008 20:45:03.661789 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/faa56149-e2dc-4cb0-8038-9ce9c8332beb-config-volume\") pod \"faa56149-e2dc-4cb0-8038-9ce9c8332beb\" (UID: \"faa56149-e2dc-4cb0-8038-9ce9c8332beb\") " Oct 08 20:45:03 crc kubenswrapper[4988]: I1008 20:45:03.662309 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-58hp2\" (UniqueName: \"kubernetes.io/projected/faa56149-e2dc-4cb0-8038-9ce9c8332beb-kube-api-access-58hp2\") pod \"faa56149-e2dc-4cb0-8038-9ce9c8332beb\" (UID: \"faa56149-e2dc-4cb0-8038-9ce9c8332beb\") " Oct 08 20:45:03 crc kubenswrapper[4988]: I1008 20:45:03.662465 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/faa56149-e2dc-4cb0-8038-9ce9c8332beb-secret-volume\") pod \"faa56149-e2dc-4cb0-8038-9ce9c8332beb\" (UID: \"faa56149-e2dc-4cb0-8038-9ce9c8332beb\") " Oct 08 20:45:03 crc kubenswrapper[4988]: I1008 20:45:03.662605 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/faa56149-e2dc-4cb0-8038-9ce9c8332beb-config-volume" (OuterVolumeSpecName: "config-volume") pod "faa56149-e2dc-4cb0-8038-9ce9c8332beb" (UID: "faa56149-e2dc-4cb0-8038-9ce9c8332beb"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 20:45:03 crc kubenswrapper[4988]: I1008 20:45:03.663165 4988 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/faa56149-e2dc-4cb0-8038-9ce9c8332beb-config-volume\") on node \"crc\" DevicePath \"\"" Oct 08 20:45:03 crc kubenswrapper[4988]: I1008 20:45:03.670521 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/faa56149-e2dc-4cb0-8038-9ce9c8332beb-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "faa56149-e2dc-4cb0-8038-9ce9c8332beb" (UID: "faa56149-e2dc-4cb0-8038-9ce9c8332beb"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 20:45:03 crc kubenswrapper[4988]: I1008 20:45:03.671510 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/faa56149-e2dc-4cb0-8038-9ce9c8332beb-kube-api-access-58hp2" (OuterVolumeSpecName: "kube-api-access-58hp2") pod "faa56149-e2dc-4cb0-8038-9ce9c8332beb" (UID: "faa56149-e2dc-4cb0-8038-9ce9c8332beb"). InnerVolumeSpecName "kube-api-access-58hp2". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 20:45:03 crc kubenswrapper[4988]: I1008 20:45:03.765239 4988 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-58hp2\" (UniqueName: \"kubernetes.io/projected/faa56149-e2dc-4cb0-8038-9ce9c8332beb-kube-api-access-58hp2\") on node \"crc\" DevicePath \"\"" Oct 08 20:45:03 crc kubenswrapper[4988]: I1008 20:45:03.765270 4988 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/faa56149-e2dc-4cb0-8038-9ce9c8332beb-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 08 20:45:04 crc kubenswrapper[4988]: I1008 20:45:04.097589 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29332605-jgdxc" event={"ID":"faa56149-e2dc-4cb0-8038-9ce9c8332beb","Type":"ContainerDied","Data":"2fb96f8cafe887221560a61e1d1f2127ddc855d686b318a2e1637804cceeea58"} Oct 08 20:45:04 crc kubenswrapper[4988]: I1008 20:45:04.097662 4988 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2fb96f8cafe887221560a61e1d1f2127ddc855d686b318a2e1637804cceeea58" Oct 08 20:45:04 crc kubenswrapper[4988]: I1008 20:45:04.097666 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29332605-jgdxc" Oct 08 20:45:04 crc kubenswrapper[4988]: I1008 20:45:04.645516 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29332560-bhvtp"] Oct 08 20:45:04 crc kubenswrapper[4988]: I1008 20:45:04.654569 4988 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29332560-bhvtp"] Oct 08 20:45:05 crc kubenswrapper[4988]: I1008 20:45:05.256805 4988 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4a463591-730f-4fd6-bd9d-25ed05409362" path="/var/lib/kubelet/pods/4a463591-730f-4fd6-bd9d-25ed05409362/volumes" Oct 08 20:45:12 crc kubenswrapper[4988]: I1008 20:45:12.238438 4988 scope.go:117] "RemoveContainer" containerID="6382c087e916f3958809279cab4d35e45e3b3ccd0cb9d8f0da5a015741773071" Oct 08 20:45:12 crc kubenswrapper[4988]: E1008 20:45:12.241142 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nm54f_openshift-machine-config-operator(f328a20b-5283-407e-8a93-2ab2da599f18)\"" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" Oct 08 20:45:16 crc kubenswrapper[4988]: I1008 20:45:16.383536 4988 scope.go:117] "RemoveContainer" containerID="28ef619b8a9aa7a07c0c27b92232eb6389ba4b462f56f10e8ad03f80ed6e3f20" Oct 08 20:45:22 crc kubenswrapper[4988]: I1008 20:45:22.843857 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-79j4p"] Oct 08 20:45:22 crc kubenswrapper[4988]: E1008 20:45:22.845160 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="faa56149-e2dc-4cb0-8038-9ce9c8332beb" containerName="collect-profiles" Oct 08 20:45:22 crc kubenswrapper[4988]: I1008 20:45:22.845183 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="faa56149-e2dc-4cb0-8038-9ce9c8332beb" containerName="collect-profiles" Oct 08 20:45:22 crc kubenswrapper[4988]: I1008 20:45:22.845618 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="faa56149-e2dc-4cb0-8038-9ce9c8332beb" containerName="collect-profiles" Oct 08 20:45:22 crc kubenswrapper[4988]: I1008 20:45:22.849729 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-79j4p" Oct 08 20:45:22 crc kubenswrapper[4988]: I1008 20:45:22.867963 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-79j4p"] Oct 08 20:45:22 crc kubenswrapper[4988]: I1008 20:45:22.989495 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/877cfa88-03f3-4326-ac89-d1c932cf6b05-catalog-content\") pod \"redhat-marketplace-79j4p\" (UID: \"877cfa88-03f3-4326-ac89-d1c932cf6b05\") " pod="openshift-marketplace/redhat-marketplace-79j4p" Oct 08 20:45:22 crc kubenswrapper[4988]: I1008 20:45:22.989578 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/877cfa88-03f3-4326-ac89-d1c932cf6b05-utilities\") pod \"redhat-marketplace-79j4p\" (UID: \"877cfa88-03f3-4326-ac89-d1c932cf6b05\") " pod="openshift-marketplace/redhat-marketplace-79j4p" Oct 08 20:45:22 crc kubenswrapper[4988]: I1008 20:45:22.990424 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cw6z9\" (UniqueName: \"kubernetes.io/projected/877cfa88-03f3-4326-ac89-d1c932cf6b05-kube-api-access-cw6z9\") pod \"redhat-marketplace-79j4p\" (UID: \"877cfa88-03f3-4326-ac89-d1c932cf6b05\") " pod="openshift-marketplace/redhat-marketplace-79j4p" Oct 08 20:45:23 crc kubenswrapper[4988]: I1008 20:45:23.093344 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/877cfa88-03f3-4326-ac89-d1c932cf6b05-catalog-content\") pod \"redhat-marketplace-79j4p\" (UID: \"877cfa88-03f3-4326-ac89-d1c932cf6b05\") " pod="openshift-marketplace/redhat-marketplace-79j4p" Oct 08 20:45:23 crc kubenswrapper[4988]: I1008 20:45:23.093473 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/877cfa88-03f3-4326-ac89-d1c932cf6b05-utilities\") pod \"redhat-marketplace-79j4p\" (UID: \"877cfa88-03f3-4326-ac89-d1c932cf6b05\") " pod="openshift-marketplace/redhat-marketplace-79j4p" Oct 08 20:45:23 crc kubenswrapper[4988]: I1008 20:45:23.093606 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cw6z9\" (UniqueName: \"kubernetes.io/projected/877cfa88-03f3-4326-ac89-d1c932cf6b05-kube-api-access-cw6z9\") pod \"redhat-marketplace-79j4p\" (UID: \"877cfa88-03f3-4326-ac89-d1c932cf6b05\") " pod="openshift-marketplace/redhat-marketplace-79j4p" Oct 08 20:45:23 crc kubenswrapper[4988]: I1008 20:45:23.094016 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/877cfa88-03f3-4326-ac89-d1c932cf6b05-catalog-content\") pod \"redhat-marketplace-79j4p\" (UID: \"877cfa88-03f3-4326-ac89-d1c932cf6b05\") " pod="openshift-marketplace/redhat-marketplace-79j4p" Oct 08 20:45:23 crc kubenswrapper[4988]: I1008 20:45:23.094628 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/877cfa88-03f3-4326-ac89-d1c932cf6b05-utilities\") pod \"redhat-marketplace-79j4p\" (UID: \"877cfa88-03f3-4326-ac89-d1c932cf6b05\") " pod="openshift-marketplace/redhat-marketplace-79j4p" Oct 08 20:45:23 crc kubenswrapper[4988]: I1008 20:45:23.111789 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cw6z9\" (UniqueName: \"kubernetes.io/projected/877cfa88-03f3-4326-ac89-d1c932cf6b05-kube-api-access-cw6z9\") pod \"redhat-marketplace-79j4p\" (UID: \"877cfa88-03f3-4326-ac89-d1c932cf6b05\") " pod="openshift-marketplace/redhat-marketplace-79j4p" Oct 08 20:45:23 crc kubenswrapper[4988]: I1008 20:45:23.178788 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-79j4p" Oct 08 20:45:23 crc kubenswrapper[4988]: I1008 20:45:23.662805 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-79j4p"] Oct 08 20:45:24 crc kubenswrapper[4988]: I1008 20:45:24.238197 4988 scope.go:117] "RemoveContainer" containerID="6382c087e916f3958809279cab4d35e45e3b3ccd0cb9d8f0da5a015741773071" Oct 08 20:45:24 crc kubenswrapper[4988]: E1008 20:45:24.239076 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nm54f_openshift-machine-config-operator(f328a20b-5283-407e-8a93-2ab2da599f18)\"" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" Oct 08 20:45:24 crc kubenswrapper[4988]: I1008 20:45:24.351917 4988 generic.go:334] "Generic (PLEG): container finished" podID="877cfa88-03f3-4326-ac89-d1c932cf6b05" containerID="4332a34c48bb6d980af84fb7a29cc4ccc15adbcd3a18be66df7ce3c33f20c482" exitCode=0 Oct 08 20:45:24 crc kubenswrapper[4988]: I1008 20:45:24.352010 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-79j4p" event={"ID":"877cfa88-03f3-4326-ac89-d1c932cf6b05","Type":"ContainerDied","Data":"4332a34c48bb6d980af84fb7a29cc4ccc15adbcd3a18be66df7ce3c33f20c482"} Oct 08 20:45:24 crc kubenswrapper[4988]: I1008 20:45:24.352069 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-79j4p" event={"ID":"877cfa88-03f3-4326-ac89-d1c932cf6b05","Type":"ContainerStarted","Data":"0b26ef5e5f6eb7a4e0f2c5046c78396b00cda63b12e4912477bef6889f1e1aaf"} Oct 08 20:45:24 crc kubenswrapper[4988]: I1008 20:45:24.356189 4988 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 08 20:45:26 crc kubenswrapper[4988]: I1008 20:45:26.379957 4988 generic.go:334] "Generic (PLEG): container finished" podID="877cfa88-03f3-4326-ac89-d1c932cf6b05" containerID="aef13304db848f156bea69c94303a0cdc72fd78c4260fd097336386d6dfa71cd" exitCode=0 Oct 08 20:45:26 crc kubenswrapper[4988]: I1008 20:45:26.380010 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-79j4p" event={"ID":"877cfa88-03f3-4326-ac89-d1c932cf6b05","Type":"ContainerDied","Data":"aef13304db848f156bea69c94303a0cdc72fd78c4260fd097336386d6dfa71cd"} Oct 08 20:45:28 crc kubenswrapper[4988]: I1008 20:45:28.432495 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-79j4p" event={"ID":"877cfa88-03f3-4326-ac89-d1c932cf6b05","Type":"ContainerStarted","Data":"3e8f0ce60cdd064665bd679463ba6b7cc8bbca588949a5f15c629945d5b03836"} Oct 08 20:45:28 crc kubenswrapper[4988]: I1008 20:45:28.470243 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-79j4p" podStartSLOduration=3.975156255 podStartE2EDuration="6.470213162s" podCreationTimestamp="2025-10-08 20:45:22 +0000 UTC" firstStartedPulling="2025-10-08 20:45:24.355623306 +0000 UTC m=+9269.805466116" lastFinishedPulling="2025-10-08 20:45:26.850680213 +0000 UTC m=+9272.300523023" observedRunningTime="2025-10-08 20:45:28.461587677 +0000 UTC m=+9273.911430487" watchObservedRunningTime="2025-10-08 20:45:28.470213162 +0000 UTC m=+9273.920055932" Oct 08 20:45:33 crc kubenswrapper[4988]: I1008 20:45:33.179021 4988 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-79j4p" Oct 08 20:45:33 crc kubenswrapper[4988]: I1008 20:45:33.179654 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-79j4p" Oct 08 20:45:33 crc kubenswrapper[4988]: I1008 20:45:33.273313 4988 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-79j4p" Oct 08 20:45:33 crc kubenswrapper[4988]: I1008 20:45:33.566414 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-79j4p" Oct 08 20:45:33 crc kubenswrapper[4988]: I1008 20:45:33.655469 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-79j4p"] Oct 08 20:45:35 crc kubenswrapper[4988]: I1008 20:45:35.251434 4988 scope.go:117] "RemoveContainer" containerID="6382c087e916f3958809279cab4d35e45e3b3ccd0cb9d8f0da5a015741773071" Oct 08 20:45:35 crc kubenswrapper[4988]: E1008 20:45:35.252354 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nm54f_openshift-machine-config-operator(f328a20b-5283-407e-8a93-2ab2da599f18)\"" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" Oct 08 20:45:35 crc kubenswrapper[4988]: I1008 20:45:35.523735 4988 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-79j4p" podUID="877cfa88-03f3-4326-ac89-d1c932cf6b05" containerName="registry-server" containerID="cri-o://3e8f0ce60cdd064665bd679463ba6b7cc8bbca588949a5f15c629945d5b03836" gracePeriod=2 Oct 08 20:45:36 crc kubenswrapper[4988]: I1008 20:45:36.036564 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-79j4p" Oct 08 20:45:36 crc kubenswrapper[4988]: I1008 20:45:36.132542 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/877cfa88-03f3-4326-ac89-d1c932cf6b05-catalog-content\") pod \"877cfa88-03f3-4326-ac89-d1c932cf6b05\" (UID: \"877cfa88-03f3-4326-ac89-d1c932cf6b05\") " Oct 08 20:45:36 crc kubenswrapper[4988]: I1008 20:45:36.132628 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cw6z9\" (UniqueName: \"kubernetes.io/projected/877cfa88-03f3-4326-ac89-d1c932cf6b05-kube-api-access-cw6z9\") pod \"877cfa88-03f3-4326-ac89-d1c932cf6b05\" (UID: \"877cfa88-03f3-4326-ac89-d1c932cf6b05\") " Oct 08 20:45:36 crc kubenswrapper[4988]: I1008 20:45:36.132684 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/877cfa88-03f3-4326-ac89-d1c932cf6b05-utilities\") pod \"877cfa88-03f3-4326-ac89-d1c932cf6b05\" (UID: \"877cfa88-03f3-4326-ac89-d1c932cf6b05\") " Oct 08 20:45:36 crc kubenswrapper[4988]: I1008 20:45:36.134187 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/877cfa88-03f3-4326-ac89-d1c932cf6b05-utilities" (OuterVolumeSpecName: "utilities") pod "877cfa88-03f3-4326-ac89-d1c932cf6b05" (UID: "877cfa88-03f3-4326-ac89-d1c932cf6b05"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 20:45:36 crc kubenswrapper[4988]: I1008 20:45:36.141866 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/877cfa88-03f3-4326-ac89-d1c932cf6b05-kube-api-access-cw6z9" (OuterVolumeSpecName: "kube-api-access-cw6z9") pod "877cfa88-03f3-4326-ac89-d1c932cf6b05" (UID: "877cfa88-03f3-4326-ac89-d1c932cf6b05"). InnerVolumeSpecName "kube-api-access-cw6z9". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 20:45:36 crc kubenswrapper[4988]: I1008 20:45:36.146290 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/877cfa88-03f3-4326-ac89-d1c932cf6b05-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "877cfa88-03f3-4326-ac89-d1c932cf6b05" (UID: "877cfa88-03f3-4326-ac89-d1c932cf6b05"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 20:45:36 crc kubenswrapper[4988]: I1008 20:45:36.235435 4988 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/877cfa88-03f3-4326-ac89-d1c932cf6b05-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 08 20:45:36 crc kubenswrapper[4988]: I1008 20:45:36.236143 4988 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cw6z9\" (UniqueName: \"kubernetes.io/projected/877cfa88-03f3-4326-ac89-d1c932cf6b05-kube-api-access-cw6z9\") on node \"crc\" DevicePath \"\"" Oct 08 20:45:36 crc kubenswrapper[4988]: I1008 20:45:36.236289 4988 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/877cfa88-03f3-4326-ac89-d1c932cf6b05-utilities\") on node \"crc\" DevicePath \"\"" Oct 08 20:45:36 crc kubenswrapper[4988]: I1008 20:45:36.538918 4988 generic.go:334] "Generic (PLEG): container finished" podID="877cfa88-03f3-4326-ac89-d1c932cf6b05" containerID="3e8f0ce60cdd064665bd679463ba6b7cc8bbca588949a5f15c629945d5b03836" exitCode=0 Oct 08 20:45:36 crc kubenswrapper[4988]: I1008 20:45:36.538979 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-79j4p" event={"ID":"877cfa88-03f3-4326-ac89-d1c932cf6b05","Type":"ContainerDied","Data":"3e8f0ce60cdd064665bd679463ba6b7cc8bbca588949a5f15c629945d5b03836"} Oct 08 20:45:36 crc kubenswrapper[4988]: I1008 20:45:36.539024 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-79j4p" event={"ID":"877cfa88-03f3-4326-ac89-d1c932cf6b05","Type":"ContainerDied","Data":"0b26ef5e5f6eb7a4e0f2c5046c78396b00cda63b12e4912477bef6889f1e1aaf"} Oct 08 20:45:36 crc kubenswrapper[4988]: I1008 20:45:36.539029 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-79j4p" Oct 08 20:45:36 crc kubenswrapper[4988]: I1008 20:45:36.539051 4988 scope.go:117] "RemoveContainer" containerID="3e8f0ce60cdd064665bd679463ba6b7cc8bbca588949a5f15c629945d5b03836" Oct 08 20:45:36 crc kubenswrapper[4988]: I1008 20:45:36.593083 4988 scope.go:117] "RemoveContainer" containerID="aef13304db848f156bea69c94303a0cdc72fd78c4260fd097336386d6dfa71cd" Oct 08 20:45:36 crc kubenswrapper[4988]: I1008 20:45:36.614577 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-79j4p"] Oct 08 20:45:36 crc kubenswrapper[4988]: I1008 20:45:36.631817 4988 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-79j4p"] Oct 08 20:45:36 crc kubenswrapper[4988]: I1008 20:45:36.637988 4988 scope.go:117] "RemoveContainer" containerID="4332a34c48bb6d980af84fb7a29cc4ccc15adbcd3a18be66df7ce3c33f20c482" Oct 08 20:45:36 crc kubenswrapper[4988]: I1008 20:45:36.700135 4988 scope.go:117] "RemoveContainer" containerID="3e8f0ce60cdd064665bd679463ba6b7cc8bbca588949a5f15c629945d5b03836" Oct 08 20:45:36 crc kubenswrapper[4988]: E1008 20:45:36.700706 4988 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3e8f0ce60cdd064665bd679463ba6b7cc8bbca588949a5f15c629945d5b03836\": container with ID starting with 3e8f0ce60cdd064665bd679463ba6b7cc8bbca588949a5f15c629945d5b03836 not found: ID does not exist" containerID="3e8f0ce60cdd064665bd679463ba6b7cc8bbca588949a5f15c629945d5b03836" Oct 08 20:45:36 crc kubenswrapper[4988]: I1008 20:45:36.700775 4988 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3e8f0ce60cdd064665bd679463ba6b7cc8bbca588949a5f15c629945d5b03836"} err="failed to get container status \"3e8f0ce60cdd064665bd679463ba6b7cc8bbca588949a5f15c629945d5b03836\": rpc error: code = NotFound desc = could not find container \"3e8f0ce60cdd064665bd679463ba6b7cc8bbca588949a5f15c629945d5b03836\": container with ID starting with 3e8f0ce60cdd064665bd679463ba6b7cc8bbca588949a5f15c629945d5b03836 not found: ID does not exist" Oct 08 20:45:36 crc kubenswrapper[4988]: I1008 20:45:36.700810 4988 scope.go:117] "RemoveContainer" containerID="aef13304db848f156bea69c94303a0cdc72fd78c4260fd097336386d6dfa71cd" Oct 08 20:45:36 crc kubenswrapper[4988]: E1008 20:45:36.701500 4988 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"aef13304db848f156bea69c94303a0cdc72fd78c4260fd097336386d6dfa71cd\": container with ID starting with aef13304db848f156bea69c94303a0cdc72fd78c4260fd097336386d6dfa71cd not found: ID does not exist" containerID="aef13304db848f156bea69c94303a0cdc72fd78c4260fd097336386d6dfa71cd" Oct 08 20:45:36 crc kubenswrapper[4988]: I1008 20:45:36.701552 4988 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"aef13304db848f156bea69c94303a0cdc72fd78c4260fd097336386d6dfa71cd"} err="failed to get container status \"aef13304db848f156bea69c94303a0cdc72fd78c4260fd097336386d6dfa71cd\": rpc error: code = NotFound desc = could not find container \"aef13304db848f156bea69c94303a0cdc72fd78c4260fd097336386d6dfa71cd\": container with ID starting with aef13304db848f156bea69c94303a0cdc72fd78c4260fd097336386d6dfa71cd not found: ID does not exist" Oct 08 20:45:36 crc kubenswrapper[4988]: I1008 20:45:36.701582 4988 scope.go:117] "RemoveContainer" containerID="4332a34c48bb6d980af84fb7a29cc4ccc15adbcd3a18be66df7ce3c33f20c482" Oct 08 20:45:36 crc kubenswrapper[4988]: E1008 20:45:36.702069 4988 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4332a34c48bb6d980af84fb7a29cc4ccc15adbcd3a18be66df7ce3c33f20c482\": container with ID starting with 4332a34c48bb6d980af84fb7a29cc4ccc15adbcd3a18be66df7ce3c33f20c482 not found: ID does not exist" containerID="4332a34c48bb6d980af84fb7a29cc4ccc15adbcd3a18be66df7ce3c33f20c482" Oct 08 20:45:36 crc kubenswrapper[4988]: I1008 20:45:36.702112 4988 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4332a34c48bb6d980af84fb7a29cc4ccc15adbcd3a18be66df7ce3c33f20c482"} err="failed to get container status \"4332a34c48bb6d980af84fb7a29cc4ccc15adbcd3a18be66df7ce3c33f20c482\": rpc error: code = NotFound desc = could not find container \"4332a34c48bb6d980af84fb7a29cc4ccc15adbcd3a18be66df7ce3c33f20c482\": container with ID starting with 4332a34c48bb6d980af84fb7a29cc4ccc15adbcd3a18be66df7ce3c33f20c482 not found: ID does not exist" Oct 08 20:45:37 crc kubenswrapper[4988]: I1008 20:45:37.268788 4988 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="877cfa88-03f3-4326-ac89-d1c932cf6b05" path="/var/lib/kubelet/pods/877cfa88-03f3-4326-ac89-d1c932cf6b05/volumes" Oct 08 20:45:49 crc kubenswrapper[4988]: I1008 20:45:49.238529 4988 scope.go:117] "RemoveContainer" containerID="6382c087e916f3958809279cab4d35e45e3b3ccd0cb9d8f0da5a015741773071" Oct 08 20:45:49 crc kubenswrapper[4988]: E1008 20:45:49.239549 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nm54f_openshift-machine-config-operator(f328a20b-5283-407e-8a93-2ab2da599f18)\"" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" Oct 08 20:46:04 crc kubenswrapper[4988]: I1008 20:46:04.239256 4988 scope.go:117] "RemoveContainer" containerID="6382c087e916f3958809279cab4d35e45e3b3ccd0cb9d8f0da5a015741773071" Oct 08 20:46:04 crc kubenswrapper[4988]: E1008 20:46:04.240511 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nm54f_openshift-machine-config-operator(f328a20b-5283-407e-8a93-2ab2da599f18)\"" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" Oct 08 20:46:16 crc kubenswrapper[4988]: I1008 20:46:16.238452 4988 scope.go:117] "RemoveContainer" containerID="6382c087e916f3958809279cab4d35e45e3b3ccd0cb9d8f0da5a015741773071" Oct 08 20:46:16 crc kubenswrapper[4988]: E1008 20:46:16.239207 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nm54f_openshift-machine-config-operator(f328a20b-5283-407e-8a93-2ab2da599f18)\"" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" Oct 08 20:46:31 crc kubenswrapper[4988]: I1008 20:46:31.238500 4988 scope.go:117] "RemoveContainer" containerID="6382c087e916f3958809279cab4d35e45e3b3ccd0cb9d8f0da5a015741773071" Oct 08 20:46:31 crc kubenswrapper[4988]: E1008 20:46:31.239177 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nm54f_openshift-machine-config-operator(f328a20b-5283-407e-8a93-2ab2da599f18)\"" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" Oct 08 20:46:43 crc kubenswrapper[4988]: I1008 20:46:43.238167 4988 scope.go:117] "RemoveContainer" containerID="6382c087e916f3958809279cab4d35e45e3b3ccd0cb9d8f0da5a015741773071" Oct 08 20:46:43 crc kubenswrapper[4988]: E1008 20:46:43.239058 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nm54f_openshift-machine-config-operator(f328a20b-5283-407e-8a93-2ab2da599f18)\"" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" Oct 08 20:46:54 crc kubenswrapper[4988]: I1008 20:46:54.238878 4988 scope.go:117] "RemoveContainer" containerID="6382c087e916f3958809279cab4d35e45e3b3ccd0cb9d8f0da5a015741773071" Oct 08 20:46:54 crc kubenswrapper[4988]: E1008 20:46:54.240441 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nm54f_openshift-machine-config-operator(f328a20b-5283-407e-8a93-2ab2da599f18)\"" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" Oct 08 20:47:05 crc kubenswrapper[4988]: I1008 20:47:05.649696 4988 generic.go:334] "Generic (PLEG): container finished" podID="7a339659-4ba8-4fb5-99fa-760e2bef90a2" containerID="4cbf674f1eceac170f689ec46f8d4cc5c78451b4027db0171cc8ea8329456885" exitCode=0 Oct 08 20:47:05 crc kubenswrapper[4988]: I1008 20:47:05.649813 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-dhcp-openstack-openstack-cell1-mbrs2" event={"ID":"7a339659-4ba8-4fb5-99fa-760e2bef90a2","Type":"ContainerDied","Data":"4cbf674f1eceac170f689ec46f8d4cc5c78451b4027db0171cc8ea8329456885"} Oct 08 20:47:07 crc kubenswrapper[4988]: I1008 20:47:07.218519 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-dhcp-openstack-openstack-cell1-mbrs2" Oct 08 20:47:07 crc kubenswrapper[4988]: I1008 20:47:07.339474 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7a339659-4ba8-4fb5-99fa-760e2bef90a2-inventory\") pod \"7a339659-4ba8-4fb5-99fa-760e2bef90a2\" (UID: \"7a339659-4ba8-4fb5-99fa-760e2bef90a2\") " Oct 08 20:47:07 crc kubenswrapper[4988]: I1008 20:47:07.340005 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-dhcp-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/7a339659-4ba8-4fb5-99fa-760e2bef90a2-neutron-dhcp-agent-neutron-config-0\") pod \"7a339659-4ba8-4fb5-99fa-760e2bef90a2\" (UID: \"7a339659-4ba8-4fb5-99fa-760e2bef90a2\") " Oct 08 20:47:07 crc kubenswrapper[4988]: I1008 20:47:07.340261 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7a339659-4ba8-4fb5-99fa-760e2bef90a2-ssh-key\") pod \"7a339659-4ba8-4fb5-99fa-760e2bef90a2\" (UID: \"7a339659-4ba8-4fb5-99fa-760e2bef90a2\") " Oct 08 20:47:07 crc kubenswrapper[4988]: I1008 20:47:07.340371 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-dhcp-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7a339659-4ba8-4fb5-99fa-760e2bef90a2-neutron-dhcp-combined-ca-bundle\") pod \"7a339659-4ba8-4fb5-99fa-760e2bef90a2\" (UID: \"7a339659-4ba8-4fb5-99fa-760e2bef90a2\") " Oct 08 20:47:07 crc kubenswrapper[4988]: I1008 20:47:07.340446 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zrrc2\" (UniqueName: \"kubernetes.io/projected/7a339659-4ba8-4fb5-99fa-760e2bef90a2-kube-api-access-zrrc2\") pod \"7a339659-4ba8-4fb5-99fa-760e2bef90a2\" (UID: \"7a339659-4ba8-4fb5-99fa-760e2bef90a2\") " Oct 08 20:47:07 crc kubenswrapper[4988]: I1008 20:47:07.344755 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7a339659-4ba8-4fb5-99fa-760e2bef90a2-kube-api-access-zrrc2" (OuterVolumeSpecName: "kube-api-access-zrrc2") pod "7a339659-4ba8-4fb5-99fa-760e2bef90a2" (UID: "7a339659-4ba8-4fb5-99fa-760e2bef90a2"). InnerVolumeSpecName "kube-api-access-zrrc2". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 20:47:07 crc kubenswrapper[4988]: I1008 20:47:07.345287 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7a339659-4ba8-4fb5-99fa-760e2bef90a2-neutron-dhcp-combined-ca-bundle" (OuterVolumeSpecName: "neutron-dhcp-combined-ca-bundle") pod "7a339659-4ba8-4fb5-99fa-760e2bef90a2" (UID: "7a339659-4ba8-4fb5-99fa-760e2bef90a2"). InnerVolumeSpecName "neutron-dhcp-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 20:47:07 crc kubenswrapper[4988]: I1008 20:47:07.368560 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7a339659-4ba8-4fb5-99fa-760e2bef90a2-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "7a339659-4ba8-4fb5-99fa-760e2bef90a2" (UID: "7a339659-4ba8-4fb5-99fa-760e2bef90a2"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 20:47:07 crc kubenswrapper[4988]: I1008 20:47:07.381250 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7a339659-4ba8-4fb5-99fa-760e2bef90a2-neutron-dhcp-agent-neutron-config-0" (OuterVolumeSpecName: "neutron-dhcp-agent-neutron-config-0") pod "7a339659-4ba8-4fb5-99fa-760e2bef90a2" (UID: "7a339659-4ba8-4fb5-99fa-760e2bef90a2"). InnerVolumeSpecName "neutron-dhcp-agent-neutron-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 20:47:07 crc kubenswrapper[4988]: I1008 20:47:07.395778 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7a339659-4ba8-4fb5-99fa-760e2bef90a2-inventory" (OuterVolumeSpecName: "inventory") pod "7a339659-4ba8-4fb5-99fa-760e2bef90a2" (UID: "7a339659-4ba8-4fb5-99fa-760e2bef90a2"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 20:47:07 crc kubenswrapper[4988]: I1008 20:47:07.444576 4988 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7a339659-4ba8-4fb5-99fa-760e2bef90a2-inventory\") on node \"crc\" DevicePath \"\"" Oct 08 20:47:07 crc kubenswrapper[4988]: I1008 20:47:07.444650 4988 reconciler_common.go:293] "Volume detached for volume \"neutron-dhcp-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/7a339659-4ba8-4fb5-99fa-760e2bef90a2-neutron-dhcp-agent-neutron-config-0\") on node \"crc\" DevicePath \"\"" Oct 08 20:47:07 crc kubenswrapper[4988]: I1008 20:47:07.444666 4988 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7a339659-4ba8-4fb5-99fa-760e2bef90a2-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 08 20:47:07 crc kubenswrapper[4988]: I1008 20:47:07.444702 4988 reconciler_common.go:293] "Volume detached for volume \"neutron-dhcp-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7a339659-4ba8-4fb5-99fa-760e2bef90a2-neutron-dhcp-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 20:47:07 crc kubenswrapper[4988]: I1008 20:47:07.444719 4988 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zrrc2\" (UniqueName: \"kubernetes.io/projected/7a339659-4ba8-4fb5-99fa-760e2bef90a2-kube-api-access-zrrc2\") on node \"crc\" DevicePath \"\"" Oct 08 20:47:07 crc kubenswrapper[4988]: I1008 20:47:07.680115 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-dhcp-openstack-openstack-cell1-mbrs2" event={"ID":"7a339659-4ba8-4fb5-99fa-760e2bef90a2","Type":"ContainerDied","Data":"b87a33dfcd4e4bf0e14580593e48b609161a530d038d93c31ddc44ea7e9f5711"} Oct 08 20:47:07 crc kubenswrapper[4988]: I1008 20:47:07.680170 4988 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b87a33dfcd4e4bf0e14580593e48b609161a530d038d93c31ddc44ea7e9f5711" Oct 08 20:47:07 crc kubenswrapper[4988]: I1008 20:47:07.680266 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-dhcp-openstack-openstack-cell1-mbrs2" Oct 08 20:47:09 crc kubenswrapper[4988]: I1008 20:47:09.238869 4988 scope.go:117] "RemoveContainer" containerID="6382c087e916f3958809279cab4d35e45e3b3ccd0cb9d8f0da5a015741773071" Oct 08 20:47:09 crc kubenswrapper[4988]: E1008 20:47:09.239889 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nm54f_openshift-machine-config-operator(f328a20b-5283-407e-8a93-2ab2da599f18)\"" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" Oct 08 20:47:22 crc kubenswrapper[4988]: I1008 20:47:22.238441 4988 scope.go:117] "RemoveContainer" containerID="6382c087e916f3958809279cab4d35e45e3b3ccd0cb9d8f0da5a015741773071" Oct 08 20:47:22 crc kubenswrapper[4988]: E1008 20:47:22.240894 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nm54f_openshift-machine-config-operator(f328a20b-5283-407e-8a93-2ab2da599f18)\"" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" Oct 08 20:47:30 crc kubenswrapper[4988]: I1008 20:47:30.533410 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-0"] Oct 08 20:47:30 crc kubenswrapper[4988]: I1008 20:47:30.534081 4988 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell0-conductor-0" podUID="667c78af-4a60-4271-88c6-ef5ac33982a6" containerName="nova-cell0-conductor-conductor" containerID="cri-o://c2884809446b412404e04b0517c34a161c686d7e35b437b35238d006fdbe1300" gracePeriod=30 Oct 08 20:47:30 crc kubenswrapper[4988]: I1008 20:47:30.553793 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-conductor-0"] Oct 08 20:47:30 crc kubenswrapper[4988]: I1008 20:47:30.554003 4988 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell1-conductor-0" podUID="730ab813-7314-4ff9-bd44-1eeeff96e912" containerName="nova-cell1-conductor-conductor" containerID="cri-o://f14ec0486f35029bb67d558c8eb8d27090175a66965eb8a80d480f8632247426" gracePeriod=30 Oct 08 20:47:31 crc kubenswrapper[4988]: E1008 20:47:31.031114 4988 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="f14ec0486f35029bb67d558c8eb8d27090175a66965eb8a80d480f8632247426" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Oct 08 20:47:31 crc kubenswrapper[4988]: E1008 20:47:31.032978 4988 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="f14ec0486f35029bb67d558c8eb8d27090175a66965eb8a80d480f8632247426" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Oct 08 20:47:31 crc kubenswrapper[4988]: E1008 20:47:31.034491 4988 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="f14ec0486f35029bb67d558c8eb8d27090175a66965eb8a80d480f8632247426" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Oct 08 20:47:31 crc kubenswrapper[4988]: E1008 20:47:31.034532 4988 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-cell1-conductor-0" podUID="730ab813-7314-4ff9-bd44-1eeeff96e912" containerName="nova-cell1-conductor-conductor" Oct 08 20:47:31 crc kubenswrapper[4988]: I1008 20:47:31.344117 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 08 20:47:31 crc kubenswrapper[4988]: I1008 20:47:31.344353 4988 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="60e4b83d-3db4-4f5a-aca8-06ee1ee85ea7" containerName="nova-api-log" containerID="cri-o://cb80a17c24a9f12755e932ef41d6a5bb75e4e04a3d0ba21999c59874b0080b6a" gracePeriod=30 Oct 08 20:47:31 crc kubenswrapper[4988]: I1008 20:47:31.344509 4988 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="60e4b83d-3db4-4f5a-aca8-06ee1ee85ea7" containerName="nova-api-api" containerID="cri-o://607110804b1f15f95b3a567b056ee8c4c1b1f154bbc886dc1d8716bf75980d29" gracePeriod=30 Oct 08 20:47:31 crc kubenswrapper[4988]: I1008 20:47:31.355026 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Oct 08 20:47:31 crc kubenswrapper[4988]: I1008 20:47:31.355257 4988 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="ca7e4fbf-efe7-4ae8-b85c-104d24f83f95" containerName="nova-scheduler-scheduler" containerID="cri-o://5c2dfbd0d5dd3874131abf5da252d865ed64608cd12ff2aad33f7474101e0229" gracePeriod=30 Oct 08 20:47:31 crc kubenswrapper[4988]: I1008 20:47:31.367293 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Oct 08 20:47:31 crc kubenswrapper[4988]: I1008 20:47:31.367603 4988 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="96d0365d-bfac-45d3-85e6-a89e540e433f" containerName="nova-metadata-log" containerID="cri-o://183c4a8fb4867be027b4eaa6827471b499f33dddf4c08c32851630fbe21ef74c" gracePeriod=30 Oct 08 20:47:31 crc kubenswrapper[4988]: I1008 20:47:31.367638 4988 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="96d0365d-bfac-45d3-85e6-a89e540e433f" containerName="nova-metadata-metadata" containerID="cri-o://f3c33b39236a5e6ebb4d7f65e628002c11c1202f89454a370f1f038359f92254" gracePeriod=30 Oct 08 20:47:31 crc kubenswrapper[4988]: I1008 20:47:31.438233 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellgcg5p"] Oct 08 20:47:31 crc kubenswrapper[4988]: E1008 20:47:31.442557 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="877cfa88-03f3-4326-ac89-d1c932cf6b05" containerName="extract-utilities" Oct 08 20:47:31 crc kubenswrapper[4988]: I1008 20:47:31.442584 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="877cfa88-03f3-4326-ac89-d1c932cf6b05" containerName="extract-utilities" Oct 08 20:47:31 crc kubenswrapper[4988]: E1008 20:47:31.442730 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7a339659-4ba8-4fb5-99fa-760e2bef90a2" containerName="neutron-dhcp-openstack-openstack-cell1" Oct 08 20:47:31 crc kubenswrapper[4988]: I1008 20:47:31.442779 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="7a339659-4ba8-4fb5-99fa-760e2bef90a2" containerName="neutron-dhcp-openstack-openstack-cell1" Oct 08 20:47:31 crc kubenswrapper[4988]: E1008 20:47:31.442840 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="877cfa88-03f3-4326-ac89-d1c932cf6b05" containerName="registry-server" Oct 08 20:47:31 crc kubenswrapper[4988]: I1008 20:47:31.442849 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="877cfa88-03f3-4326-ac89-d1c932cf6b05" containerName="registry-server" Oct 08 20:47:31 crc kubenswrapper[4988]: E1008 20:47:31.442901 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="877cfa88-03f3-4326-ac89-d1c932cf6b05" containerName="extract-content" Oct 08 20:47:31 crc kubenswrapper[4988]: I1008 20:47:31.442909 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="877cfa88-03f3-4326-ac89-d1c932cf6b05" containerName="extract-content" Oct 08 20:47:31 crc kubenswrapper[4988]: I1008 20:47:31.443866 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="7a339659-4ba8-4fb5-99fa-760e2bef90a2" containerName="neutron-dhcp-openstack-openstack-cell1" Oct 08 20:47:31 crc kubenswrapper[4988]: I1008 20:47:31.443911 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="877cfa88-03f3-4326-ac89-d1c932cf6b05" containerName="registry-server" Oct 08 20:47:31 crc kubenswrapper[4988]: I1008 20:47:31.453096 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellgcg5p" Oct 08 20:47:31 crc kubenswrapper[4988]: I1008 20:47:31.456936 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellgcg5p"] Oct 08 20:47:31 crc kubenswrapper[4988]: I1008 20:47:31.457907 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-migration-ssh-key" Oct 08 20:47:31 crc kubenswrapper[4988]: I1008 20:47:31.458104 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-4sdzr" Oct 08 20:47:31 crc kubenswrapper[4988]: I1008 20:47:31.458525 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Oct 08 20:47:31 crc kubenswrapper[4988]: I1008 20:47:31.458671 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Oct 08 20:47:31 crc kubenswrapper[4988]: I1008 20:47:31.458993 4988 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"nova-cells-global-config" Oct 08 20:47:31 crc kubenswrapper[4988]: I1008 20:47:31.459236 4988 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 08 20:47:31 crc kubenswrapper[4988]: I1008 20:47:31.459455 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-compute-config" Oct 08 20:47:31 crc kubenswrapper[4988]: I1008 20:47:31.537698 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a2f34e97-3b21-419b-883b-bdd5108b7dae-inventory\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellgcg5p\" (UID: \"a2f34e97-3b21-419b-883b-bdd5108b7dae\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellgcg5p" Oct 08 20:47:31 crc kubenswrapper[4988]: I1008 20:47:31.537770 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/a2f34e97-3b21-419b-883b-bdd5108b7dae-nova-cell1-compute-config-1\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellgcg5p\" (UID: \"a2f34e97-3b21-419b-883b-bdd5108b7dae\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellgcg5p" Oct 08 20:47:31 crc kubenswrapper[4988]: I1008 20:47:31.537964 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/a2f34e97-3b21-419b-883b-bdd5108b7dae-nova-migration-ssh-key-0\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellgcg5p\" (UID: \"a2f34e97-3b21-419b-883b-bdd5108b7dae\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellgcg5p" Oct 08 20:47:31 crc kubenswrapper[4988]: I1008 20:47:31.538040 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rfs7g\" (UniqueName: \"kubernetes.io/projected/a2f34e97-3b21-419b-883b-bdd5108b7dae-kube-api-access-rfs7g\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellgcg5p\" (UID: \"a2f34e97-3b21-419b-883b-bdd5108b7dae\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellgcg5p" Oct 08 20:47:31 crc kubenswrapper[4988]: I1008 20:47:31.538139 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/a2f34e97-3b21-419b-883b-bdd5108b7dae-nova-migration-ssh-key-1\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellgcg5p\" (UID: \"a2f34e97-3b21-419b-883b-bdd5108b7dae\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellgcg5p" Oct 08 20:47:31 crc kubenswrapper[4988]: I1008 20:47:31.538203 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a2f34e97-3b21-419b-883b-bdd5108b7dae-ssh-key\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellgcg5p\" (UID: \"a2f34e97-3b21-419b-883b-bdd5108b7dae\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellgcg5p" Oct 08 20:47:31 crc kubenswrapper[4988]: I1008 20:47:31.538234 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a2f34e97-3b21-419b-883b-bdd5108b7dae-nova-cell1-combined-ca-bundle\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellgcg5p\" (UID: \"a2f34e97-3b21-419b-883b-bdd5108b7dae\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellgcg5p" Oct 08 20:47:31 crc kubenswrapper[4988]: I1008 20:47:31.538318 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cells-global-config-0\" (UniqueName: \"kubernetes.io/configmap/a2f34e97-3b21-419b-883b-bdd5108b7dae-nova-cells-global-config-0\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellgcg5p\" (UID: \"a2f34e97-3b21-419b-883b-bdd5108b7dae\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellgcg5p" Oct 08 20:47:31 crc kubenswrapper[4988]: I1008 20:47:31.538348 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/a2f34e97-3b21-419b-883b-bdd5108b7dae-nova-cell1-compute-config-0\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellgcg5p\" (UID: \"a2f34e97-3b21-419b-883b-bdd5108b7dae\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellgcg5p" Oct 08 20:47:31 crc kubenswrapper[4988]: E1008 20:47:31.617588 4988 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="c2884809446b412404e04b0517c34a161c686d7e35b437b35238d006fdbe1300" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Oct 08 20:47:31 crc kubenswrapper[4988]: E1008 20:47:31.625752 4988 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="c2884809446b412404e04b0517c34a161c686d7e35b437b35238d006fdbe1300" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Oct 08 20:47:31 crc kubenswrapper[4988]: E1008 20:47:31.629736 4988 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="c2884809446b412404e04b0517c34a161c686d7e35b437b35238d006fdbe1300" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Oct 08 20:47:31 crc kubenswrapper[4988]: E1008 20:47:31.629788 4988 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-cell0-conductor-0" podUID="667c78af-4a60-4271-88c6-ef5ac33982a6" containerName="nova-cell0-conductor-conductor" Oct 08 20:47:31 crc kubenswrapper[4988]: I1008 20:47:31.640547 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/a2f34e97-3b21-419b-883b-bdd5108b7dae-nova-cell1-compute-config-0\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellgcg5p\" (UID: \"a2f34e97-3b21-419b-883b-bdd5108b7dae\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellgcg5p" Oct 08 20:47:31 crc kubenswrapper[4988]: I1008 20:47:31.640630 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a2f34e97-3b21-419b-883b-bdd5108b7dae-inventory\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellgcg5p\" (UID: \"a2f34e97-3b21-419b-883b-bdd5108b7dae\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellgcg5p" Oct 08 20:47:31 crc kubenswrapper[4988]: I1008 20:47:31.640656 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/a2f34e97-3b21-419b-883b-bdd5108b7dae-nova-cell1-compute-config-1\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellgcg5p\" (UID: \"a2f34e97-3b21-419b-883b-bdd5108b7dae\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellgcg5p" Oct 08 20:47:31 crc kubenswrapper[4988]: I1008 20:47:31.640724 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/a2f34e97-3b21-419b-883b-bdd5108b7dae-nova-migration-ssh-key-0\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellgcg5p\" (UID: \"a2f34e97-3b21-419b-883b-bdd5108b7dae\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellgcg5p" Oct 08 20:47:31 crc kubenswrapper[4988]: I1008 20:47:31.640752 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rfs7g\" (UniqueName: \"kubernetes.io/projected/a2f34e97-3b21-419b-883b-bdd5108b7dae-kube-api-access-rfs7g\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellgcg5p\" (UID: \"a2f34e97-3b21-419b-883b-bdd5108b7dae\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellgcg5p" Oct 08 20:47:31 crc kubenswrapper[4988]: I1008 20:47:31.640817 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/a2f34e97-3b21-419b-883b-bdd5108b7dae-nova-migration-ssh-key-1\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellgcg5p\" (UID: \"a2f34e97-3b21-419b-883b-bdd5108b7dae\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellgcg5p" Oct 08 20:47:31 crc kubenswrapper[4988]: I1008 20:47:31.640849 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a2f34e97-3b21-419b-883b-bdd5108b7dae-ssh-key\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellgcg5p\" (UID: \"a2f34e97-3b21-419b-883b-bdd5108b7dae\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellgcg5p" Oct 08 20:47:31 crc kubenswrapper[4988]: I1008 20:47:31.640871 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a2f34e97-3b21-419b-883b-bdd5108b7dae-nova-cell1-combined-ca-bundle\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellgcg5p\" (UID: \"a2f34e97-3b21-419b-883b-bdd5108b7dae\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellgcg5p" Oct 08 20:47:31 crc kubenswrapper[4988]: I1008 20:47:31.640909 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cells-global-config-0\" (UniqueName: \"kubernetes.io/configmap/a2f34e97-3b21-419b-883b-bdd5108b7dae-nova-cells-global-config-0\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellgcg5p\" (UID: \"a2f34e97-3b21-419b-883b-bdd5108b7dae\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellgcg5p" Oct 08 20:47:31 crc kubenswrapper[4988]: I1008 20:47:31.641717 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cells-global-config-0\" (UniqueName: \"kubernetes.io/configmap/a2f34e97-3b21-419b-883b-bdd5108b7dae-nova-cells-global-config-0\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellgcg5p\" (UID: \"a2f34e97-3b21-419b-883b-bdd5108b7dae\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellgcg5p" Oct 08 20:47:31 crc kubenswrapper[4988]: I1008 20:47:31.653782 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/a2f34e97-3b21-419b-883b-bdd5108b7dae-nova-cell1-compute-config-1\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellgcg5p\" (UID: \"a2f34e97-3b21-419b-883b-bdd5108b7dae\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellgcg5p" Oct 08 20:47:31 crc kubenswrapper[4988]: I1008 20:47:31.654158 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a2f34e97-3b21-419b-883b-bdd5108b7dae-inventory\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellgcg5p\" (UID: \"a2f34e97-3b21-419b-883b-bdd5108b7dae\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellgcg5p" Oct 08 20:47:31 crc kubenswrapper[4988]: I1008 20:47:31.654269 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/a2f34e97-3b21-419b-883b-bdd5108b7dae-nova-cell1-compute-config-0\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellgcg5p\" (UID: \"a2f34e97-3b21-419b-883b-bdd5108b7dae\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellgcg5p" Oct 08 20:47:31 crc kubenswrapper[4988]: I1008 20:47:31.655281 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a2f34e97-3b21-419b-883b-bdd5108b7dae-ssh-key\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellgcg5p\" (UID: \"a2f34e97-3b21-419b-883b-bdd5108b7dae\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellgcg5p" Oct 08 20:47:31 crc kubenswrapper[4988]: I1008 20:47:31.657423 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a2f34e97-3b21-419b-883b-bdd5108b7dae-nova-cell1-combined-ca-bundle\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellgcg5p\" (UID: \"a2f34e97-3b21-419b-883b-bdd5108b7dae\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellgcg5p" Oct 08 20:47:31 crc kubenswrapper[4988]: I1008 20:47:31.658640 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/a2f34e97-3b21-419b-883b-bdd5108b7dae-nova-migration-ssh-key-0\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellgcg5p\" (UID: \"a2f34e97-3b21-419b-883b-bdd5108b7dae\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellgcg5p" Oct 08 20:47:31 crc kubenswrapper[4988]: I1008 20:47:31.660298 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rfs7g\" (UniqueName: \"kubernetes.io/projected/a2f34e97-3b21-419b-883b-bdd5108b7dae-kube-api-access-rfs7g\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellgcg5p\" (UID: \"a2f34e97-3b21-419b-883b-bdd5108b7dae\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellgcg5p" Oct 08 20:47:31 crc kubenswrapper[4988]: I1008 20:47:31.678137 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/a2f34e97-3b21-419b-883b-bdd5108b7dae-nova-migration-ssh-key-1\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellgcg5p\" (UID: \"a2f34e97-3b21-419b-883b-bdd5108b7dae\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellgcg5p" Oct 08 20:47:31 crc kubenswrapper[4988]: I1008 20:47:31.897498 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellgcg5p" Oct 08 20:47:31 crc kubenswrapper[4988]: I1008 20:47:31.967656 4988 generic.go:334] "Generic (PLEG): container finished" podID="60e4b83d-3db4-4f5a-aca8-06ee1ee85ea7" containerID="cb80a17c24a9f12755e932ef41d6a5bb75e4e04a3d0ba21999c59874b0080b6a" exitCode=143 Oct 08 20:47:31 crc kubenswrapper[4988]: I1008 20:47:31.967746 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"60e4b83d-3db4-4f5a-aca8-06ee1ee85ea7","Type":"ContainerDied","Data":"cb80a17c24a9f12755e932ef41d6a5bb75e4e04a3d0ba21999c59874b0080b6a"} Oct 08 20:47:31 crc kubenswrapper[4988]: I1008 20:47:31.971000 4988 generic.go:334] "Generic (PLEG): container finished" podID="96d0365d-bfac-45d3-85e6-a89e540e433f" containerID="183c4a8fb4867be027b4eaa6827471b499f33dddf4c08c32851630fbe21ef74c" exitCode=143 Oct 08 20:47:31 crc kubenswrapper[4988]: I1008 20:47:31.971145 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"96d0365d-bfac-45d3-85e6-a89e540e433f","Type":"ContainerDied","Data":"183c4a8fb4867be027b4eaa6827471b499f33dddf4c08c32851630fbe21ef74c"} Oct 08 20:47:32 crc kubenswrapper[4988]: I1008 20:47:32.554801 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellgcg5p"] Oct 08 20:47:32 crc kubenswrapper[4988]: I1008 20:47:32.645690 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 08 20:47:32 crc kubenswrapper[4988]: I1008 20:47:32.777053 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ca7e4fbf-efe7-4ae8-b85c-104d24f83f95-config-data\") pod \"ca7e4fbf-efe7-4ae8-b85c-104d24f83f95\" (UID: \"ca7e4fbf-efe7-4ae8-b85c-104d24f83f95\") " Oct 08 20:47:32 crc kubenswrapper[4988]: I1008 20:47:32.777090 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gmtwd\" (UniqueName: \"kubernetes.io/projected/ca7e4fbf-efe7-4ae8-b85c-104d24f83f95-kube-api-access-gmtwd\") pod \"ca7e4fbf-efe7-4ae8-b85c-104d24f83f95\" (UID: \"ca7e4fbf-efe7-4ae8-b85c-104d24f83f95\") " Oct 08 20:47:32 crc kubenswrapper[4988]: I1008 20:47:32.777176 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ca7e4fbf-efe7-4ae8-b85c-104d24f83f95-combined-ca-bundle\") pod \"ca7e4fbf-efe7-4ae8-b85c-104d24f83f95\" (UID: \"ca7e4fbf-efe7-4ae8-b85c-104d24f83f95\") " Oct 08 20:47:32 crc kubenswrapper[4988]: I1008 20:47:32.987825 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellgcg5p" event={"ID":"a2f34e97-3b21-419b-883b-bdd5108b7dae","Type":"ContainerStarted","Data":"c40d76919f87192452426e0318f1e15643e7a3c1c2dd04aaf0ce773fa961c2c5"} Oct 08 20:47:32 crc kubenswrapper[4988]: I1008 20:47:32.990593 4988 generic.go:334] "Generic (PLEG): container finished" podID="730ab813-7314-4ff9-bd44-1eeeff96e912" containerID="f14ec0486f35029bb67d558c8eb8d27090175a66965eb8a80d480f8632247426" exitCode=0 Oct 08 20:47:32 crc kubenswrapper[4988]: I1008 20:47:32.990720 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"730ab813-7314-4ff9-bd44-1eeeff96e912","Type":"ContainerDied","Data":"f14ec0486f35029bb67d558c8eb8d27090175a66965eb8a80d480f8632247426"} Oct 08 20:47:32 crc kubenswrapper[4988]: I1008 20:47:32.994443 4988 generic.go:334] "Generic (PLEG): container finished" podID="ca7e4fbf-efe7-4ae8-b85c-104d24f83f95" containerID="5c2dfbd0d5dd3874131abf5da252d865ed64608cd12ff2aad33f7474101e0229" exitCode=0 Oct 08 20:47:32 crc kubenswrapper[4988]: I1008 20:47:32.994498 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"ca7e4fbf-efe7-4ae8-b85c-104d24f83f95","Type":"ContainerDied","Data":"5c2dfbd0d5dd3874131abf5da252d865ed64608cd12ff2aad33f7474101e0229"} Oct 08 20:47:32 crc kubenswrapper[4988]: I1008 20:47:32.994540 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"ca7e4fbf-efe7-4ae8-b85c-104d24f83f95","Type":"ContainerDied","Data":"b681f0306d45f7110985bb8e895a96f594a212828bde833223fca8c5337ef990"} Oct 08 20:47:32 crc kubenswrapper[4988]: I1008 20:47:32.994563 4988 scope.go:117] "RemoveContainer" containerID="5c2dfbd0d5dd3874131abf5da252d865ed64608cd12ff2aad33f7474101e0229" Oct 08 20:47:32 crc kubenswrapper[4988]: I1008 20:47:32.994769 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 08 20:47:33 crc kubenswrapper[4988]: I1008 20:47:33.480798 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ca7e4fbf-efe7-4ae8-b85c-104d24f83f95-kube-api-access-gmtwd" (OuterVolumeSpecName: "kube-api-access-gmtwd") pod "ca7e4fbf-efe7-4ae8-b85c-104d24f83f95" (UID: "ca7e4fbf-efe7-4ae8-b85c-104d24f83f95"). InnerVolumeSpecName "kube-api-access-gmtwd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 20:47:33 crc kubenswrapper[4988]: I1008 20:47:33.492839 4988 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gmtwd\" (UniqueName: \"kubernetes.io/projected/ca7e4fbf-efe7-4ae8-b85c-104d24f83f95-kube-api-access-gmtwd\") on node \"crc\" DevicePath \"\"" Oct 08 20:47:33 crc kubenswrapper[4988]: I1008 20:47:33.504864 4988 scope.go:117] "RemoveContainer" containerID="5c2dfbd0d5dd3874131abf5da252d865ed64608cd12ff2aad33f7474101e0229" Oct 08 20:47:33 crc kubenswrapper[4988]: E1008 20:47:33.505475 4988 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5c2dfbd0d5dd3874131abf5da252d865ed64608cd12ff2aad33f7474101e0229\": container with ID starting with 5c2dfbd0d5dd3874131abf5da252d865ed64608cd12ff2aad33f7474101e0229 not found: ID does not exist" containerID="5c2dfbd0d5dd3874131abf5da252d865ed64608cd12ff2aad33f7474101e0229" Oct 08 20:47:33 crc kubenswrapper[4988]: I1008 20:47:33.505547 4988 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5c2dfbd0d5dd3874131abf5da252d865ed64608cd12ff2aad33f7474101e0229"} err="failed to get container status \"5c2dfbd0d5dd3874131abf5da252d865ed64608cd12ff2aad33f7474101e0229\": rpc error: code = NotFound desc = could not find container \"5c2dfbd0d5dd3874131abf5da252d865ed64608cd12ff2aad33f7474101e0229\": container with ID starting with 5c2dfbd0d5dd3874131abf5da252d865ed64608cd12ff2aad33f7474101e0229 not found: ID does not exist" Oct 08 20:47:33 crc kubenswrapper[4988]: I1008 20:47:33.765521 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ca7e4fbf-efe7-4ae8-b85c-104d24f83f95-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ca7e4fbf-efe7-4ae8-b85c-104d24f83f95" (UID: "ca7e4fbf-efe7-4ae8-b85c-104d24f83f95"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 20:47:33 crc kubenswrapper[4988]: I1008 20:47:33.778621 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ca7e4fbf-efe7-4ae8-b85c-104d24f83f95-config-data" (OuterVolumeSpecName: "config-data") pod "ca7e4fbf-efe7-4ae8-b85c-104d24f83f95" (UID: "ca7e4fbf-efe7-4ae8-b85c-104d24f83f95"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 20:47:33 crc kubenswrapper[4988]: I1008 20:47:33.806888 4988 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ca7e4fbf-efe7-4ae8-b85c-104d24f83f95-config-data\") on node \"crc\" DevicePath \"\"" Oct 08 20:47:33 crc kubenswrapper[4988]: I1008 20:47:33.807009 4988 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ca7e4fbf-efe7-4ae8-b85c-104d24f83f95-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 20:47:33 crc kubenswrapper[4988]: I1008 20:47:33.888091 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Oct 08 20:47:33 crc kubenswrapper[4988]: I1008 20:47:33.971024 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Oct 08 20:47:33 crc kubenswrapper[4988]: I1008 20:47:33.986676 4988 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Oct 08 20:47:34 crc kubenswrapper[4988]: I1008 20:47:34.016660 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Oct 08 20:47:34 crc kubenswrapper[4988]: E1008 20:47:34.017253 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ca7e4fbf-efe7-4ae8-b85c-104d24f83f95" containerName="nova-scheduler-scheduler" Oct 08 20:47:34 crc kubenswrapper[4988]: I1008 20:47:34.017281 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="ca7e4fbf-efe7-4ae8-b85c-104d24f83f95" containerName="nova-scheduler-scheduler" Oct 08 20:47:34 crc kubenswrapper[4988]: E1008 20:47:34.017291 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="730ab813-7314-4ff9-bd44-1eeeff96e912" containerName="nova-cell1-conductor-conductor" Oct 08 20:47:34 crc kubenswrapper[4988]: I1008 20:47:34.017298 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="730ab813-7314-4ff9-bd44-1eeeff96e912" containerName="nova-cell1-conductor-conductor" Oct 08 20:47:34 crc kubenswrapper[4988]: I1008 20:47:34.017403 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/730ab813-7314-4ff9-bd44-1eeeff96e912-combined-ca-bundle\") pod \"730ab813-7314-4ff9-bd44-1eeeff96e912\" (UID: \"730ab813-7314-4ff9-bd44-1eeeff96e912\") " Oct 08 20:47:34 crc kubenswrapper[4988]: I1008 20:47:34.017570 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="730ab813-7314-4ff9-bd44-1eeeff96e912" containerName="nova-cell1-conductor-conductor" Oct 08 20:47:34 crc kubenswrapper[4988]: I1008 20:47:34.017589 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9htsf\" (UniqueName: \"kubernetes.io/projected/730ab813-7314-4ff9-bd44-1eeeff96e912-kube-api-access-9htsf\") pod \"730ab813-7314-4ff9-bd44-1eeeff96e912\" (UID: \"730ab813-7314-4ff9-bd44-1eeeff96e912\") " Oct 08 20:47:34 crc kubenswrapper[4988]: I1008 20:47:34.017598 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="ca7e4fbf-efe7-4ae8-b85c-104d24f83f95" containerName="nova-scheduler-scheduler" Oct 08 20:47:34 crc kubenswrapper[4988]: I1008 20:47:34.017736 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/730ab813-7314-4ff9-bd44-1eeeff96e912-config-data\") pod \"730ab813-7314-4ff9-bd44-1eeeff96e912\" (UID: \"730ab813-7314-4ff9-bd44-1eeeff96e912\") " Oct 08 20:47:34 crc kubenswrapper[4988]: I1008 20:47:34.018533 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 08 20:47:34 crc kubenswrapper[4988]: I1008 20:47:34.024568 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/730ab813-7314-4ff9-bd44-1eeeff96e912-kube-api-access-9htsf" (OuterVolumeSpecName: "kube-api-access-9htsf") pod "730ab813-7314-4ff9-bd44-1eeeff96e912" (UID: "730ab813-7314-4ff9-bd44-1eeeff96e912"). InnerVolumeSpecName "kube-api-access-9htsf". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 20:47:34 crc kubenswrapper[4988]: I1008 20:47:34.025006 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Oct 08 20:47:34 crc kubenswrapper[4988]: I1008 20:47:34.027506 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Oct 08 20:47:34 crc kubenswrapper[4988]: I1008 20:47:34.029078 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"730ab813-7314-4ff9-bd44-1eeeff96e912","Type":"ContainerDied","Data":"53af335385340a15cf262e6df1d2842c11d4629fab2cc1d9081351bdd45830dc"} Oct 08 20:47:34 crc kubenswrapper[4988]: I1008 20:47:34.029187 4988 scope.go:117] "RemoveContainer" containerID="f14ec0486f35029bb67d558c8eb8d27090175a66965eb8a80d480f8632247426" Oct 08 20:47:34 crc kubenswrapper[4988]: I1008 20:47:34.029398 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Oct 08 20:47:34 crc kubenswrapper[4988]: I1008 20:47:34.067233 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/730ab813-7314-4ff9-bd44-1eeeff96e912-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "730ab813-7314-4ff9-bd44-1eeeff96e912" (UID: "730ab813-7314-4ff9-bd44-1eeeff96e912"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 20:47:34 crc kubenswrapper[4988]: I1008 20:47:34.074794 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellgcg5p" podStartSLOduration=2.116895166 podStartE2EDuration="3.074776203s" podCreationTimestamp="2025-10-08 20:47:31 +0000 UTC" firstStartedPulling="2025-10-08 20:47:32.580919697 +0000 UTC m=+9398.030762477" lastFinishedPulling="2025-10-08 20:47:33.538800704 +0000 UTC m=+9398.988643514" observedRunningTime="2025-10-08 20:47:34.062600015 +0000 UTC m=+9399.512442785" watchObservedRunningTime="2025-10-08 20:47:34.074776203 +0000 UTC m=+9399.524618973" Oct 08 20:47:34 crc kubenswrapper[4988]: I1008 20:47:34.085515 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/730ab813-7314-4ff9-bd44-1eeeff96e912-config-data" (OuterVolumeSpecName: "config-data") pod "730ab813-7314-4ff9-bd44-1eeeff96e912" (UID: "730ab813-7314-4ff9-bd44-1eeeff96e912"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 20:47:34 crc kubenswrapper[4988]: E1008 20:47:34.117878 4988 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podca7e4fbf_efe7_4ae8_b85c_104d24f83f95.slice/crio-b681f0306d45f7110985bb8e895a96f594a212828bde833223fca8c5337ef990\": RecentStats: unable to find data in memory cache]" Oct 08 20:47:34 crc kubenswrapper[4988]: I1008 20:47:34.120994 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/88e44fb8-0012-4f95-9845-38d86ea84918-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"88e44fb8-0012-4f95-9845-38d86ea84918\") " pod="openstack/nova-scheduler-0" Oct 08 20:47:34 crc kubenswrapper[4988]: I1008 20:47:34.121089 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/88e44fb8-0012-4f95-9845-38d86ea84918-config-data\") pod \"nova-scheduler-0\" (UID: \"88e44fb8-0012-4f95-9845-38d86ea84918\") " pod="openstack/nova-scheduler-0" Oct 08 20:47:34 crc kubenswrapper[4988]: I1008 20:47:34.121163 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-92tf7\" (UniqueName: \"kubernetes.io/projected/88e44fb8-0012-4f95-9845-38d86ea84918-kube-api-access-92tf7\") pod \"nova-scheduler-0\" (UID: \"88e44fb8-0012-4f95-9845-38d86ea84918\") " pod="openstack/nova-scheduler-0" Oct 08 20:47:34 crc kubenswrapper[4988]: I1008 20:47:34.121432 4988 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/730ab813-7314-4ff9-bd44-1eeeff96e912-config-data\") on node \"crc\" DevicePath \"\"" Oct 08 20:47:34 crc kubenswrapper[4988]: I1008 20:47:34.121768 4988 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/730ab813-7314-4ff9-bd44-1eeeff96e912-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 20:47:34 crc kubenswrapper[4988]: I1008 20:47:34.121842 4988 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9htsf\" (UniqueName: \"kubernetes.io/projected/730ab813-7314-4ff9-bd44-1eeeff96e912-kube-api-access-9htsf\") on node \"crc\" DevicePath \"\"" Oct 08 20:47:34 crc kubenswrapper[4988]: I1008 20:47:34.223828 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/88e44fb8-0012-4f95-9845-38d86ea84918-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"88e44fb8-0012-4f95-9845-38d86ea84918\") " pod="openstack/nova-scheduler-0" Oct 08 20:47:34 crc kubenswrapper[4988]: I1008 20:47:34.224171 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/88e44fb8-0012-4f95-9845-38d86ea84918-config-data\") pod \"nova-scheduler-0\" (UID: \"88e44fb8-0012-4f95-9845-38d86ea84918\") " pod="openstack/nova-scheduler-0" Oct 08 20:47:34 crc kubenswrapper[4988]: I1008 20:47:34.224214 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-92tf7\" (UniqueName: \"kubernetes.io/projected/88e44fb8-0012-4f95-9845-38d86ea84918-kube-api-access-92tf7\") pod \"nova-scheduler-0\" (UID: \"88e44fb8-0012-4f95-9845-38d86ea84918\") " pod="openstack/nova-scheduler-0" Oct 08 20:47:34 crc kubenswrapper[4988]: I1008 20:47:34.235165 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/88e44fb8-0012-4f95-9845-38d86ea84918-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"88e44fb8-0012-4f95-9845-38d86ea84918\") " pod="openstack/nova-scheduler-0" Oct 08 20:47:34 crc kubenswrapper[4988]: I1008 20:47:34.235929 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/88e44fb8-0012-4f95-9845-38d86ea84918-config-data\") pod \"nova-scheduler-0\" (UID: \"88e44fb8-0012-4f95-9845-38d86ea84918\") " pod="openstack/nova-scheduler-0" Oct 08 20:47:34 crc kubenswrapper[4988]: I1008 20:47:34.241404 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-92tf7\" (UniqueName: \"kubernetes.io/projected/88e44fb8-0012-4f95-9845-38d86ea84918-kube-api-access-92tf7\") pod \"nova-scheduler-0\" (UID: \"88e44fb8-0012-4f95-9845-38d86ea84918\") " pod="openstack/nova-scheduler-0" Oct 08 20:47:34 crc kubenswrapper[4988]: I1008 20:47:34.360677 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-conductor-0"] Oct 08 20:47:34 crc kubenswrapper[4988]: I1008 20:47:34.373791 4988 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-conductor-0"] Oct 08 20:47:34 crc kubenswrapper[4988]: I1008 20:47:34.395485 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-0"] Oct 08 20:47:34 crc kubenswrapper[4988]: I1008 20:47:34.398197 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Oct 08 20:47:34 crc kubenswrapper[4988]: I1008 20:47:34.407590 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Oct 08 20:47:34 crc kubenswrapper[4988]: I1008 20:47:34.430734 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Oct 08 20:47:34 crc kubenswrapper[4988]: I1008 20:47:34.452507 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 08 20:47:34 crc kubenswrapper[4988]: I1008 20:47:34.539802 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/115dfff7-cb2b-4638-b40a-28a11aa2aaf9-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"115dfff7-cb2b-4638-b40a-28a11aa2aaf9\") " pod="openstack/nova-cell1-conductor-0" Oct 08 20:47:34 crc kubenswrapper[4988]: I1008 20:47:34.540495 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/115dfff7-cb2b-4638-b40a-28a11aa2aaf9-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"115dfff7-cb2b-4638-b40a-28a11aa2aaf9\") " pod="openstack/nova-cell1-conductor-0" Oct 08 20:47:34 crc kubenswrapper[4988]: I1008 20:47:34.540555 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gcwdl\" (UniqueName: \"kubernetes.io/projected/115dfff7-cb2b-4638-b40a-28a11aa2aaf9-kube-api-access-gcwdl\") pod \"nova-cell1-conductor-0\" (UID: \"115dfff7-cb2b-4638-b40a-28a11aa2aaf9\") " pod="openstack/nova-cell1-conductor-0" Oct 08 20:47:34 crc kubenswrapper[4988]: I1008 20:47:34.642409 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/115dfff7-cb2b-4638-b40a-28a11aa2aaf9-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"115dfff7-cb2b-4638-b40a-28a11aa2aaf9\") " pod="openstack/nova-cell1-conductor-0" Oct 08 20:47:34 crc kubenswrapper[4988]: I1008 20:47:34.642462 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gcwdl\" (UniqueName: \"kubernetes.io/projected/115dfff7-cb2b-4638-b40a-28a11aa2aaf9-kube-api-access-gcwdl\") pod \"nova-cell1-conductor-0\" (UID: \"115dfff7-cb2b-4638-b40a-28a11aa2aaf9\") " pod="openstack/nova-cell1-conductor-0" Oct 08 20:47:34 crc kubenswrapper[4988]: I1008 20:47:34.642507 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/115dfff7-cb2b-4638-b40a-28a11aa2aaf9-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"115dfff7-cb2b-4638-b40a-28a11aa2aaf9\") " pod="openstack/nova-cell1-conductor-0" Oct 08 20:47:34 crc kubenswrapper[4988]: I1008 20:47:34.651528 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/115dfff7-cb2b-4638-b40a-28a11aa2aaf9-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"115dfff7-cb2b-4638-b40a-28a11aa2aaf9\") " pod="openstack/nova-cell1-conductor-0" Oct 08 20:47:34 crc kubenswrapper[4988]: I1008 20:47:34.658785 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/115dfff7-cb2b-4638-b40a-28a11aa2aaf9-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"115dfff7-cb2b-4638-b40a-28a11aa2aaf9\") " pod="openstack/nova-cell1-conductor-0" Oct 08 20:47:34 crc kubenswrapper[4988]: I1008 20:47:34.662551 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gcwdl\" (UniqueName: \"kubernetes.io/projected/115dfff7-cb2b-4638-b40a-28a11aa2aaf9-kube-api-access-gcwdl\") pod \"nova-cell1-conductor-0\" (UID: \"115dfff7-cb2b-4638-b40a-28a11aa2aaf9\") " pod="openstack/nova-cell1-conductor-0" Oct 08 20:47:34 crc kubenswrapper[4988]: I1008 20:47:34.754917 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Oct 08 20:47:34 crc kubenswrapper[4988]: I1008 20:47:34.804281 4988 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="96d0365d-bfac-45d3-85e6-a89e540e433f" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.1.96:8775/\": read tcp 10.217.0.2:35282->10.217.1.96:8775: read: connection reset by peer" Oct 08 20:47:34 crc kubenswrapper[4988]: I1008 20:47:34.804779 4988 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="96d0365d-bfac-45d3-85e6-a89e540e433f" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.1.96:8775/\": read tcp 10.217.0.2:35290->10.217.1.96:8775: read: connection reset by peer" Oct 08 20:47:34 crc kubenswrapper[4988]: I1008 20:47:34.921746 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Oct 08 20:47:35 crc kubenswrapper[4988]: I1008 20:47:35.047066 4988 generic.go:334] "Generic (PLEG): container finished" podID="96d0365d-bfac-45d3-85e6-a89e540e433f" containerID="f3c33b39236a5e6ebb4d7f65e628002c11c1202f89454a370f1f038359f92254" exitCode=0 Oct 08 20:47:35 crc kubenswrapper[4988]: I1008 20:47:35.047141 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"96d0365d-bfac-45d3-85e6-a89e540e433f","Type":"ContainerDied","Data":"f3c33b39236a5e6ebb4d7f65e628002c11c1202f89454a370f1f038359f92254"} Oct 08 20:47:35 crc kubenswrapper[4988]: I1008 20:47:35.048939 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellgcg5p" event={"ID":"a2f34e97-3b21-419b-883b-bdd5108b7dae","Type":"ContainerStarted","Data":"886f78c41ab3b0c677fc4cb14d4cec225016b181fd83dbe7cf8adfdf2dac37f8"} Oct 08 20:47:35 crc kubenswrapper[4988]: I1008 20:47:35.050762 4988 generic.go:334] "Generic (PLEG): container finished" podID="60e4b83d-3db4-4f5a-aca8-06ee1ee85ea7" containerID="607110804b1f15f95b3a567b056ee8c4c1b1f154bbc886dc1d8716bf75980d29" exitCode=0 Oct 08 20:47:35 crc kubenswrapper[4988]: I1008 20:47:35.050788 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"60e4b83d-3db4-4f5a-aca8-06ee1ee85ea7","Type":"ContainerDied","Data":"607110804b1f15f95b3a567b056ee8c4c1b1f154bbc886dc1d8716bf75980d29"} Oct 08 20:47:35 crc kubenswrapper[4988]: I1008 20:47:35.051988 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"88e44fb8-0012-4f95-9845-38d86ea84918","Type":"ContainerStarted","Data":"c63a09e84cd1bb8dd883be41147e5894b4fd86c8d93709a04802e0f2d00d8b23"} Oct 08 20:47:35 crc kubenswrapper[4988]: I1008 20:47:35.249990 4988 scope.go:117] "RemoveContainer" containerID="6382c087e916f3958809279cab4d35e45e3b3ccd0cb9d8f0da5a015741773071" Oct 08 20:47:35 crc kubenswrapper[4988]: E1008 20:47:35.250296 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nm54f_openshift-machine-config-operator(f328a20b-5283-407e-8a93-2ab2da599f18)\"" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" Oct 08 20:47:35 crc kubenswrapper[4988]: I1008 20:47:35.257935 4988 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="730ab813-7314-4ff9-bd44-1eeeff96e912" path="/var/lib/kubelet/pods/730ab813-7314-4ff9-bd44-1eeeff96e912/volumes" Oct 08 20:47:35 crc kubenswrapper[4988]: I1008 20:47:35.258519 4988 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ca7e4fbf-efe7-4ae8-b85c-104d24f83f95" path="/var/lib/kubelet/pods/ca7e4fbf-efe7-4ae8-b85c-104d24f83f95/volumes" Oct 08 20:47:35 crc kubenswrapper[4988]: I1008 20:47:35.495031 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Oct 08 20:47:35 crc kubenswrapper[4988]: I1008 20:47:35.786721 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 08 20:47:35 crc kubenswrapper[4988]: I1008 20:47:35.826466 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 08 20:47:35 crc kubenswrapper[4988]: I1008 20:47:35.889705 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/96d0365d-bfac-45d3-85e6-a89e540e433f-config-data\") pod \"96d0365d-bfac-45d3-85e6-a89e540e433f\" (UID: \"96d0365d-bfac-45d3-85e6-a89e540e433f\") " Oct 08 20:47:35 crc kubenswrapper[4988]: I1008 20:47:35.890003 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/96d0365d-bfac-45d3-85e6-a89e540e433f-combined-ca-bundle\") pod \"96d0365d-bfac-45d3-85e6-a89e540e433f\" (UID: \"96d0365d-bfac-45d3-85e6-a89e540e433f\") " Oct 08 20:47:35 crc kubenswrapper[4988]: I1008 20:47:35.890201 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qwzhn\" (UniqueName: \"kubernetes.io/projected/96d0365d-bfac-45d3-85e6-a89e540e433f-kube-api-access-qwzhn\") pod \"96d0365d-bfac-45d3-85e6-a89e540e433f\" (UID: \"96d0365d-bfac-45d3-85e6-a89e540e433f\") " Oct 08 20:47:35 crc kubenswrapper[4988]: I1008 20:47:35.890280 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/96d0365d-bfac-45d3-85e6-a89e540e433f-logs\") pod \"96d0365d-bfac-45d3-85e6-a89e540e433f\" (UID: \"96d0365d-bfac-45d3-85e6-a89e540e433f\") " Oct 08 20:47:35 crc kubenswrapper[4988]: I1008 20:47:35.890430 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/96d0365d-bfac-45d3-85e6-a89e540e433f-nova-metadata-tls-certs\") pod \"96d0365d-bfac-45d3-85e6-a89e540e433f\" (UID: \"96d0365d-bfac-45d3-85e6-a89e540e433f\") " Oct 08 20:47:35 crc kubenswrapper[4988]: I1008 20:47:35.890814 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/96d0365d-bfac-45d3-85e6-a89e540e433f-logs" (OuterVolumeSpecName: "logs") pod "96d0365d-bfac-45d3-85e6-a89e540e433f" (UID: "96d0365d-bfac-45d3-85e6-a89e540e433f"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 20:47:35 crc kubenswrapper[4988]: I1008 20:47:35.891468 4988 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/96d0365d-bfac-45d3-85e6-a89e540e433f-logs\") on node \"crc\" DevicePath \"\"" Oct 08 20:47:35 crc kubenswrapper[4988]: I1008 20:47:35.897353 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/96d0365d-bfac-45d3-85e6-a89e540e433f-kube-api-access-qwzhn" (OuterVolumeSpecName: "kube-api-access-qwzhn") pod "96d0365d-bfac-45d3-85e6-a89e540e433f" (UID: "96d0365d-bfac-45d3-85e6-a89e540e433f"). InnerVolumeSpecName "kube-api-access-qwzhn". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 20:47:35 crc kubenswrapper[4988]: I1008 20:47:35.961749 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/96d0365d-bfac-45d3-85e6-a89e540e433f-config-data" (OuterVolumeSpecName: "config-data") pod "96d0365d-bfac-45d3-85e6-a89e540e433f" (UID: "96d0365d-bfac-45d3-85e6-a89e540e433f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 20:47:35 crc kubenswrapper[4988]: I1008 20:47:35.985433 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/96d0365d-bfac-45d3-85e6-a89e540e433f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "96d0365d-bfac-45d3-85e6-a89e540e433f" (UID: "96d0365d-bfac-45d3-85e6-a89e540e433f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 20:47:35 crc kubenswrapper[4988]: I1008 20:47:35.994104 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/60e4b83d-3db4-4f5a-aca8-06ee1ee85ea7-internal-tls-certs\") pod \"60e4b83d-3db4-4f5a-aca8-06ee1ee85ea7\" (UID: \"60e4b83d-3db4-4f5a-aca8-06ee1ee85ea7\") " Oct 08 20:47:35 crc kubenswrapper[4988]: I1008 20:47:35.994948 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/60e4b83d-3db4-4f5a-aca8-06ee1ee85ea7-combined-ca-bundle\") pod \"60e4b83d-3db4-4f5a-aca8-06ee1ee85ea7\" (UID: \"60e4b83d-3db4-4f5a-aca8-06ee1ee85ea7\") " Oct 08 20:47:35 crc kubenswrapper[4988]: I1008 20:47:35.995145 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/60e4b83d-3db4-4f5a-aca8-06ee1ee85ea7-public-tls-certs\") pod \"60e4b83d-3db4-4f5a-aca8-06ee1ee85ea7\" (UID: \"60e4b83d-3db4-4f5a-aca8-06ee1ee85ea7\") " Oct 08 20:47:35 crc kubenswrapper[4988]: I1008 20:47:35.995182 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/60e4b83d-3db4-4f5a-aca8-06ee1ee85ea7-logs\") pod \"60e4b83d-3db4-4f5a-aca8-06ee1ee85ea7\" (UID: \"60e4b83d-3db4-4f5a-aca8-06ee1ee85ea7\") " Oct 08 20:47:35 crc kubenswrapper[4988]: I1008 20:47:35.995341 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7fm97\" (UniqueName: \"kubernetes.io/projected/60e4b83d-3db4-4f5a-aca8-06ee1ee85ea7-kube-api-access-7fm97\") pod \"60e4b83d-3db4-4f5a-aca8-06ee1ee85ea7\" (UID: \"60e4b83d-3db4-4f5a-aca8-06ee1ee85ea7\") " Oct 08 20:47:35 crc kubenswrapper[4988]: I1008 20:47:35.995460 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/60e4b83d-3db4-4f5a-aca8-06ee1ee85ea7-config-data\") pod \"60e4b83d-3db4-4f5a-aca8-06ee1ee85ea7\" (UID: \"60e4b83d-3db4-4f5a-aca8-06ee1ee85ea7\") " Oct 08 20:47:35 crc kubenswrapper[4988]: I1008 20:47:35.996236 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/60e4b83d-3db4-4f5a-aca8-06ee1ee85ea7-logs" (OuterVolumeSpecName: "logs") pod "60e4b83d-3db4-4f5a-aca8-06ee1ee85ea7" (UID: "60e4b83d-3db4-4f5a-aca8-06ee1ee85ea7"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 20:47:35 crc kubenswrapper[4988]: I1008 20:47:35.997378 4988 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/96d0365d-bfac-45d3-85e6-a89e540e433f-config-data\") on node \"crc\" DevicePath \"\"" Oct 08 20:47:35 crc kubenswrapper[4988]: I1008 20:47:35.997415 4988 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/96d0365d-bfac-45d3-85e6-a89e540e433f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 20:47:35 crc kubenswrapper[4988]: I1008 20:47:35.997432 4988 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qwzhn\" (UniqueName: \"kubernetes.io/projected/96d0365d-bfac-45d3-85e6-a89e540e433f-kube-api-access-qwzhn\") on node \"crc\" DevicePath \"\"" Oct 08 20:47:35 crc kubenswrapper[4988]: I1008 20:47:35.997445 4988 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/60e4b83d-3db4-4f5a-aca8-06ee1ee85ea7-logs\") on node \"crc\" DevicePath \"\"" Oct 08 20:47:36 crc kubenswrapper[4988]: I1008 20:47:36.011672 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/60e4b83d-3db4-4f5a-aca8-06ee1ee85ea7-kube-api-access-7fm97" (OuterVolumeSpecName: "kube-api-access-7fm97") pod "60e4b83d-3db4-4f5a-aca8-06ee1ee85ea7" (UID: "60e4b83d-3db4-4f5a-aca8-06ee1ee85ea7"). InnerVolumeSpecName "kube-api-access-7fm97". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 20:47:36 crc kubenswrapper[4988]: I1008 20:47:36.036604 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/60e4b83d-3db4-4f5a-aca8-06ee1ee85ea7-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "60e4b83d-3db4-4f5a-aca8-06ee1ee85ea7" (UID: "60e4b83d-3db4-4f5a-aca8-06ee1ee85ea7"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 20:47:36 crc kubenswrapper[4988]: I1008 20:47:36.038863 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/96d0365d-bfac-45d3-85e6-a89e540e433f-nova-metadata-tls-certs" (OuterVolumeSpecName: "nova-metadata-tls-certs") pod "96d0365d-bfac-45d3-85e6-a89e540e433f" (UID: "96d0365d-bfac-45d3-85e6-a89e540e433f"). InnerVolumeSpecName "nova-metadata-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 20:47:36 crc kubenswrapper[4988]: I1008 20:47:36.062637 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/60e4b83d-3db4-4f5a-aca8-06ee1ee85ea7-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "60e4b83d-3db4-4f5a-aca8-06ee1ee85ea7" (UID: "60e4b83d-3db4-4f5a-aca8-06ee1ee85ea7"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 20:47:36 crc kubenswrapper[4988]: I1008 20:47:36.068241 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/60e4b83d-3db4-4f5a-aca8-06ee1ee85ea7-config-data" (OuterVolumeSpecName: "config-data") pod "60e4b83d-3db4-4f5a-aca8-06ee1ee85ea7" (UID: "60e4b83d-3db4-4f5a-aca8-06ee1ee85ea7"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 20:47:36 crc kubenswrapper[4988]: I1008 20:47:36.073200 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"88e44fb8-0012-4f95-9845-38d86ea84918","Type":"ContainerStarted","Data":"71c20f314a3e188b0cbe03eac6647d234e773a4b4ea5a478fdb9765f44a4101a"} Oct 08 20:47:36 crc kubenswrapper[4988]: I1008 20:47:36.087041 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"115dfff7-cb2b-4638-b40a-28a11aa2aaf9","Type":"ContainerStarted","Data":"25e417076fe5e43bbef628e3b3ef741cb689e2530705ca548e7c3bd61c149ace"} Oct 08 20:47:36 crc kubenswrapper[4988]: I1008 20:47:36.087098 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"115dfff7-cb2b-4638-b40a-28a11aa2aaf9","Type":"ContainerStarted","Data":"9416c67de726ecf930dc50752637f1f5dfabafc490ea54a2b780f3b09a373259"} Oct 08 20:47:36 crc kubenswrapper[4988]: I1008 20:47:36.087228 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-conductor-0" Oct 08 20:47:36 crc kubenswrapper[4988]: I1008 20:47:36.101563 4988 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7fm97\" (UniqueName: \"kubernetes.io/projected/60e4b83d-3db4-4f5a-aca8-06ee1ee85ea7-kube-api-access-7fm97\") on node \"crc\" DevicePath \"\"" Oct 08 20:47:36 crc kubenswrapper[4988]: I1008 20:47:36.101844 4988 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/96d0365d-bfac-45d3-85e6-a89e540e433f-nova-metadata-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 08 20:47:36 crc kubenswrapper[4988]: I1008 20:47:36.101857 4988 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/60e4b83d-3db4-4f5a-aca8-06ee1ee85ea7-config-data\") on node \"crc\" DevicePath \"\"" Oct 08 20:47:36 crc kubenswrapper[4988]: I1008 20:47:36.101866 4988 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/60e4b83d-3db4-4f5a-aca8-06ee1ee85ea7-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 08 20:47:36 crc kubenswrapper[4988]: I1008 20:47:36.101874 4988 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/60e4b83d-3db4-4f5a-aca8-06ee1ee85ea7-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 20:47:36 crc kubenswrapper[4988]: I1008 20:47:36.108324 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=3.108307318 podStartE2EDuration="3.108307318s" podCreationTimestamp="2025-10-08 20:47:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 20:47:36.087555756 +0000 UTC m=+9401.537398526" watchObservedRunningTime="2025-10-08 20:47:36.108307318 +0000 UTC m=+9401.558150088" Oct 08 20:47:36 crc kubenswrapper[4988]: I1008 20:47:36.109834 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"96d0365d-bfac-45d3-85e6-a89e540e433f","Type":"ContainerDied","Data":"d7128b4e27ffb9ac708e3296be5bc84babf586304faad2e0d0444570805873b1"} Oct 08 20:47:36 crc kubenswrapper[4988]: I1008 20:47:36.109878 4988 scope.go:117] "RemoveContainer" containerID="f3c33b39236a5e6ebb4d7f65e628002c11c1202f89454a370f1f038359f92254" Oct 08 20:47:36 crc kubenswrapper[4988]: I1008 20:47:36.109999 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 08 20:47:36 crc kubenswrapper[4988]: I1008 20:47:36.117427 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"60e4b83d-3db4-4f5a-aca8-06ee1ee85ea7","Type":"ContainerDied","Data":"a47b5cd9b728e721f1efe6ff52e9a59abff1e7b84636b4af0a4e3b42718087f9"} Oct 08 20:47:36 crc kubenswrapper[4988]: I1008 20:47:36.117529 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 08 20:47:36 crc kubenswrapper[4988]: I1008 20:47:36.140229 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/60e4b83d-3db4-4f5a-aca8-06ee1ee85ea7-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "60e4b83d-3db4-4f5a-aca8-06ee1ee85ea7" (UID: "60e4b83d-3db4-4f5a-aca8-06ee1ee85ea7"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 20:47:36 crc kubenswrapper[4988]: I1008 20:47:36.140614 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-0" podStartSLOduration=2.140595939 podStartE2EDuration="2.140595939s" podCreationTimestamp="2025-10-08 20:47:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 20:47:36.106425719 +0000 UTC m=+9401.556268489" watchObservedRunningTime="2025-10-08 20:47:36.140595939 +0000 UTC m=+9401.590438709" Oct 08 20:47:36 crc kubenswrapper[4988]: I1008 20:47:36.158560 4988 scope.go:117] "RemoveContainer" containerID="183c4a8fb4867be027b4eaa6827471b499f33dddf4c08c32851630fbe21ef74c" Oct 08 20:47:36 crc kubenswrapper[4988]: I1008 20:47:36.193970 4988 scope.go:117] "RemoveContainer" containerID="607110804b1f15f95b3a567b056ee8c4c1b1f154bbc886dc1d8716bf75980d29" Oct 08 20:47:36 crc kubenswrapper[4988]: I1008 20:47:36.203224 4988 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/60e4b83d-3db4-4f5a-aca8-06ee1ee85ea7-public-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 08 20:47:36 crc kubenswrapper[4988]: I1008 20:47:36.216107 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Oct 08 20:47:36 crc kubenswrapper[4988]: I1008 20:47:36.224298 4988 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Oct 08 20:47:36 crc kubenswrapper[4988]: I1008 20:47:36.225596 4988 scope.go:117] "RemoveContainer" containerID="cb80a17c24a9f12755e932ef41d6a5bb75e4e04a3d0ba21999c59874b0080b6a" Oct 08 20:47:36 crc kubenswrapper[4988]: I1008 20:47:36.235373 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Oct 08 20:47:36 crc kubenswrapper[4988]: E1008 20:47:36.235812 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="96d0365d-bfac-45d3-85e6-a89e540e433f" containerName="nova-metadata-metadata" Oct 08 20:47:36 crc kubenswrapper[4988]: I1008 20:47:36.235830 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="96d0365d-bfac-45d3-85e6-a89e540e433f" containerName="nova-metadata-metadata" Oct 08 20:47:36 crc kubenswrapper[4988]: E1008 20:47:36.235854 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="96d0365d-bfac-45d3-85e6-a89e540e433f" containerName="nova-metadata-log" Oct 08 20:47:36 crc kubenswrapper[4988]: I1008 20:47:36.235862 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="96d0365d-bfac-45d3-85e6-a89e540e433f" containerName="nova-metadata-log" Oct 08 20:47:36 crc kubenswrapper[4988]: E1008 20:47:36.235878 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="60e4b83d-3db4-4f5a-aca8-06ee1ee85ea7" containerName="nova-api-api" Oct 08 20:47:36 crc kubenswrapper[4988]: I1008 20:47:36.235885 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="60e4b83d-3db4-4f5a-aca8-06ee1ee85ea7" containerName="nova-api-api" Oct 08 20:47:36 crc kubenswrapper[4988]: E1008 20:47:36.235898 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="60e4b83d-3db4-4f5a-aca8-06ee1ee85ea7" containerName="nova-api-log" Oct 08 20:47:36 crc kubenswrapper[4988]: I1008 20:47:36.235906 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="60e4b83d-3db4-4f5a-aca8-06ee1ee85ea7" containerName="nova-api-log" Oct 08 20:47:36 crc kubenswrapper[4988]: I1008 20:47:36.236138 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="96d0365d-bfac-45d3-85e6-a89e540e433f" containerName="nova-metadata-metadata" Oct 08 20:47:36 crc kubenswrapper[4988]: I1008 20:47:36.236169 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="96d0365d-bfac-45d3-85e6-a89e540e433f" containerName="nova-metadata-log" Oct 08 20:47:36 crc kubenswrapper[4988]: I1008 20:47:36.236193 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="60e4b83d-3db4-4f5a-aca8-06ee1ee85ea7" containerName="nova-api-api" Oct 08 20:47:36 crc kubenswrapper[4988]: I1008 20:47:36.236206 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="60e4b83d-3db4-4f5a-aca8-06ee1ee85ea7" containerName="nova-api-log" Oct 08 20:47:36 crc kubenswrapper[4988]: I1008 20:47:36.237403 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 08 20:47:36 crc kubenswrapper[4988]: I1008 20:47:36.240201 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Oct 08 20:47:36 crc kubenswrapper[4988]: I1008 20:47:36.240329 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Oct 08 20:47:36 crc kubenswrapper[4988]: I1008 20:47:36.248160 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 08 20:47:36 crc kubenswrapper[4988]: I1008 20:47:36.417617 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/59722276-fba4-4321-9c14-696d0c1c98ce-logs\") pod \"nova-metadata-0\" (UID: \"59722276-fba4-4321-9c14-696d0c1c98ce\") " pod="openstack/nova-metadata-0" Oct 08 20:47:36 crc kubenswrapper[4988]: I1008 20:47:36.417664 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/59722276-fba4-4321-9c14-696d0c1c98ce-config-data\") pod \"nova-metadata-0\" (UID: \"59722276-fba4-4321-9c14-696d0c1c98ce\") " pod="openstack/nova-metadata-0" Oct 08 20:47:36 crc kubenswrapper[4988]: I1008 20:47:36.417755 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/59722276-fba4-4321-9c14-696d0c1c98ce-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"59722276-fba4-4321-9c14-696d0c1c98ce\") " pod="openstack/nova-metadata-0" Oct 08 20:47:36 crc kubenswrapper[4988]: I1008 20:47:36.417790 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/59722276-fba4-4321-9c14-696d0c1c98ce-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"59722276-fba4-4321-9c14-696d0c1c98ce\") " pod="openstack/nova-metadata-0" Oct 08 20:47:36 crc kubenswrapper[4988]: I1008 20:47:36.417822 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-trqkr\" (UniqueName: \"kubernetes.io/projected/59722276-fba4-4321-9c14-696d0c1c98ce-kube-api-access-trqkr\") pod \"nova-metadata-0\" (UID: \"59722276-fba4-4321-9c14-696d0c1c98ce\") " pod="openstack/nova-metadata-0" Oct 08 20:47:36 crc kubenswrapper[4988]: I1008 20:47:36.449594 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 08 20:47:36 crc kubenswrapper[4988]: I1008 20:47:36.463196 4988 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Oct 08 20:47:36 crc kubenswrapper[4988]: I1008 20:47:36.473262 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Oct 08 20:47:36 crc kubenswrapper[4988]: I1008 20:47:36.476956 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 08 20:47:36 crc kubenswrapper[4988]: I1008 20:47:36.479935 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Oct 08 20:47:36 crc kubenswrapper[4988]: I1008 20:47:36.480096 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-internal-svc" Oct 08 20:47:36 crc kubenswrapper[4988]: I1008 20:47:36.480195 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-public-svc" Oct 08 20:47:36 crc kubenswrapper[4988]: I1008 20:47:36.493058 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 08 20:47:36 crc kubenswrapper[4988]: I1008 20:47:36.519397 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/59722276-fba4-4321-9c14-696d0c1c98ce-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"59722276-fba4-4321-9c14-696d0c1c98ce\") " pod="openstack/nova-metadata-0" Oct 08 20:47:36 crc kubenswrapper[4988]: I1008 20:47:36.519464 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/59722276-fba4-4321-9c14-696d0c1c98ce-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"59722276-fba4-4321-9c14-696d0c1c98ce\") " pod="openstack/nova-metadata-0" Oct 08 20:47:36 crc kubenswrapper[4988]: I1008 20:47:36.519516 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-trqkr\" (UniqueName: \"kubernetes.io/projected/59722276-fba4-4321-9c14-696d0c1c98ce-kube-api-access-trqkr\") pod \"nova-metadata-0\" (UID: \"59722276-fba4-4321-9c14-696d0c1c98ce\") " pod="openstack/nova-metadata-0" Oct 08 20:47:36 crc kubenswrapper[4988]: I1008 20:47:36.520516 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/59722276-fba4-4321-9c14-696d0c1c98ce-logs\") pod \"nova-metadata-0\" (UID: \"59722276-fba4-4321-9c14-696d0c1c98ce\") " pod="openstack/nova-metadata-0" Oct 08 20:47:36 crc kubenswrapper[4988]: I1008 20:47:36.520669 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/59722276-fba4-4321-9c14-696d0c1c98ce-config-data\") pod \"nova-metadata-0\" (UID: \"59722276-fba4-4321-9c14-696d0c1c98ce\") " pod="openstack/nova-metadata-0" Oct 08 20:47:36 crc kubenswrapper[4988]: I1008 20:47:36.521079 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/59722276-fba4-4321-9c14-696d0c1c98ce-logs\") pod \"nova-metadata-0\" (UID: \"59722276-fba4-4321-9c14-696d0c1c98ce\") " pod="openstack/nova-metadata-0" Oct 08 20:47:36 crc kubenswrapper[4988]: I1008 20:47:36.523686 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/59722276-fba4-4321-9c14-696d0c1c98ce-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"59722276-fba4-4321-9c14-696d0c1c98ce\") " pod="openstack/nova-metadata-0" Oct 08 20:47:36 crc kubenswrapper[4988]: I1008 20:47:36.542879 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/59722276-fba4-4321-9c14-696d0c1c98ce-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"59722276-fba4-4321-9c14-696d0c1c98ce\") " pod="openstack/nova-metadata-0" Oct 08 20:47:36 crc kubenswrapper[4988]: I1008 20:47:36.544016 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/59722276-fba4-4321-9c14-696d0c1c98ce-config-data\") pod \"nova-metadata-0\" (UID: \"59722276-fba4-4321-9c14-696d0c1c98ce\") " pod="openstack/nova-metadata-0" Oct 08 20:47:36 crc kubenswrapper[4988]: I1008 20:47:36.565603 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-trqkr\" (UniqueName: \"kubernetes.io/projected/59722276-fba4-4321-9c14-696d0c1c98ce-kube-api-access-trqkr\") pod \"nova-metadata-0\" (UID: \"59722276-fba4-4321-9c14-696d0c1c98ce\") " pod="openstack/nova-metadata-0" Oct 08 20:47:36 crc kubenswrapper[4988]: E1008 20:47:36.621621 4988 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="c2884809446b412404e04b0517c34a161c686d7e35b437b35238d006fdbe1300" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Oct 08 20:47:36 crc kubenswrapper[4988]: E1008 20:47:36.622903 4988 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="c2884809446b412404e04b0517c34a161c686d7e35b437b35238d006fdbe1300" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Oct 08 20:47:36 crc kubenswrapper[4988]: I1008 20:47:36.623679 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b2s8z\" (UniqueName: \"kubernetes.io/projected/a881e1d1-a353-47e5-83be-b264f9552c26-kube-api-access-b2s8z\") pod \"nova-api-0\" (UID: \"a881e1d1-a353-47e5-83be-b264f9552c26\") " pod="openstack/nova-api-0" Oct 08 20:47:36 crc kubenswrapper[4988]: I1008 20:47:36.623743 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/a881e1d1-a353-47e5-83be-b264f9552c26-public-tls-certs\") pod \"nova-api-0\" (UID: \"a881e1d1-a353-47e5-83be-b264f9552c26\") " pod="openstack/nova-api-0" Oct 08 20:47:36 crc kubenswrapper[4988]: I1008 20:47:36.623910 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/a881e1d1-a353-47e5-83be-b264f9552c26-internal-tls-certs\") pod \"nova-api-0\" (UID: \"a881e1d1-a353-47e5-83be-b264f9552c26\") " pod="openstack/nova-api-0" Oct 08 20:47:36 crc kubenswrapper[4988]: I1008 20:47:36.623976 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a881e1d1-a353-47e5-83be-b264f9552c26-logs\") pod \"nova-api-0\" (UID: \"a881e1d1-a353-47e5-83be-b264f9552c26\") " pod="openstack/nova-api-0" Oct 08 20:47:36 crc kubenswrapper[4988]: I1008 20:47:36.624216 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a881e1d1-a353-47e5-83be-b264f9552c26-config-data\") pod \"nova-api-0\" (UID: \"a881e1d1-a353-47e5-83be-b264f9552c26\") " pod="openstack/nova-api-0" Oct 08 20:47:36 crc kubenswrapper[4988]: I1008 20:47:36.624288 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a881e1d1-a353-47e5-83be-b264f9552c26-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"a881e1d1-a353-47e5-83be-b264f9552c26\") " pod="openstack/nova-api-0" Oct 08 20:47:36 crc kubenswrapper[4988]: E1008 20:47:36.625493 4988 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="c2884809446b412404e04b0517c34a161c686d7e35b437b35238d006fdbe1300" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Oct 08 20:47:36 crc kubenswrapper[4988]: E1008 20:47:36.625545 4988 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-cell0-conductor-0" podUID="667c78af-4a60-4271-88c6-ef5ac33982a6" containerName="nova-cell0-conductor-conductor" Oct 08 20:47:36 crc kubenswrapper[4988]: I1008 20:47:36.726326 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/a881e1d1-a353-47e5-83be-b264f9552c26-internal-tls-certs\") pod \"nova-api-0\" (UID: \"a881e1d1-a353-47e5-83be-b264f9552c26\") " pod="openstack/nova-api-0" Oct 08 20:47:36 crc kubenswrapper[4988]: I1008 20:47:36.726371 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a881e1d1-a353-47e5-83be-b264f9552c26-logs\") pod \"nova-api-0\" (UID: \"a881e1d1-a353-47e5-83be-b264f9552c26\") " pod="openstack/nova-api-0" Oct 08 20:47:36 crc kubenswrapper[4988]: I1008 20:47:36.726472 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a881e1d1-a353-47e5-83be-b264f9552c26-config-data\") pod \"nova-api-0\" (UID: \"a881e1d1-a353-47e5-83be-b264f9552c26\") " pod="openstack/nova-api-0" Oct 08 20:47:36 crc kubenswrapper[4988]: I1008 20:47:36.726506 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a881e1d1-a353-47e5-83be-b264f9552c26-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"a881e1d1-a353-47e5-83be-b264f9552c26\") " pod="openstack/nova-api-0" Oct 08 20:47:36 crc kubenswrapper[4988]: I1008 20:47:36.726533 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b2s8z\" (UniqueName: \"kubernetes.io/projected/a881e1d1-a353-47e5-83be-b264f9552c26-kube-api-access-b2s8z\") pod \"nova-api-0\" (UID: \"a881e1d1-a353-47e5-83be-b264f9552c26\") " pod="openstack/nova-api-0" Oct 08 20:47:36 crc kubenswrapper[4988]: I1008 20:47:36.726553 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/a881e1d1-a353-47e5-83be-b264f9552c26-public-tls-certs\") pod \"nova-api-0\" (UID: \"a881e1d1-a353-47e5-83be-b264f9552c26\") " pod="openstack/nova-api-0" Oct 08 20:47:36 crc kubenswrapper[4988]: I1008 20:47:36.726927 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a881e1d1-a353-47e5-83be-b264f9552c26-logs\") pod \"nova-api-0\" (UID: \"a881e1d1-a353-47e5-83be-b264f9552c26\") " pod="openstack/nova-api-0" Oct 08 20:47:36 crc kubenswrapper[4988]: I1008 20:47:36.729614 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a881e1d1-a353-47e5-83be-b264f9552c26-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"a881e1d1-a353-47e5-83be-b264f9552c26\") " pod="openstack/nova-api-0" Oct 08 20:47:36 crc kubenswrapper[4988]: I1008 20:47:36.730034 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/a881e1d1-a353-47e5-83be-b264f9552c26-public-tls-certs\") pod \"nova-api-0\" (UID: \"a881e1d1-a353-47e5-83be-b264f9552c26\") " pod="openstack/nova-api-0" Oct 08 20:47:36 crc kubenswrapper[4988]: I1008 20:47:36.738046 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a881e1d1-a353-47e5-83be-b264f9552c26-config-data\") pod \"nova-api-0\" (UID: \"a881e1d1-a353-47e5-83be-b264f9552c26\") " pod="openstack/nova-api-0" Oct 08 20:47:36 crc kubenswrapper[4988]: I1008 20:47:36.746802 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/a881e1d1-a353-47e5-83be-b264f9552c26-internal-tls-certs\") pod \"nova-api-0\" (UID: \"a881e1d1-a353-47e5-83be-b264f9552c26\") " pod="openstack/nova-api-0" Oct 08 20:47:36 crc kubenswrapper[4988]: I1008 20:47:36.751605 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b2s8z\" (UniqueName: \"kubernetes.io/projected/a881e1d1-a353-47e5-83be-b264f9552c26-kube-api-access-b2s8z\") pod \"nova-api-0\" (UID: \"a881e1d1-a353-47e5-83be-b264f9552c26\") " pod="openstack/nova-api-0" Oct 08 20:47:36 crc kubenswrapper[4988]: I1008 20:47:36.858752 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 08 20:47:36 crc kubenswrapper[4988]: I1008 20:47:36.961429 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 08 20:47:37 crc kubenswrapper[4988]: I1008 20:47:37.255889 4988 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="60e4b83d-3db4-4f5a-aca8-06ee1ee85ea7" path="/var/lib/kubelet/pods/60e4b83d-3db4-4f5a-aca8-06ee1ee85ea7/volumes" Oct 08 20:47:37 crc kubenswrapper[4988]: I1008 20:47:37.257321 4988 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="96d0365d-bfac-45d3-85e6-a89e540e433f" path="/var/lib/kubelet/pods/96d0365d-bfac-45d3-85e6-a89e540e433f/volumes" Oct 08 20:47:37 crc kubenswrapper[4988]: I1008 20:47:37.350772 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 08 20:47:37 crc kubenswrapper[4988]: I1008 20:47:37.458487 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 08 20:47:38 crc kubenswrapper[4988]: I1008 20:47:38.142665 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"a881e1d1-a353-47e5-83be-b264f9552c26","Type":"ContainerStarted","Data":"b3d4deb00845d945fe170fa5110144cafc110e31197fc91d08ca54f203f2edd5"} Oct 08 20:47:38 crc kubenswrapper[4988]: I1008 20:47:38.143192 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"a881e1d1-a353-47e5-83be-b264f9552c26","Type":"ContainerStarted","Data":"3b3d109dccd1f8e1884998b980422990fc653ef91850b3dfdba1ac77dceb2b99"} Oct 08 20:47:38 crc kubenswrapper[4988]: I1008 20:47:38.143211 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"a881e1d1-a353-47e5-83be-b264f9552c26","Type":"ContainerStarted","Data":"b5fb0189528df41805e6602fe038fd73e507efa867d90694f91d88a716144925"} Oct 08 20:47:38 crc kubenswrapper[4988]: I1008 20:47:38.145899 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"59722276-fba4-4321-9c14-696d0c1c98ce","Type":"ContainerStarted","Data":"cb999629fbc5ab92370714df444247bdf0c412db513a3881c71d3ea04a1080f3"} Oct 08 20:47:38 crc kubenswrapper[4988]: I1008 20:47:38.145959 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"59722276-fba4-4321-9c14-696d0c1c98ce","Type":"ContainerStarted","Data":"293475a209050f5c94447c9f84534c1c2f6654567059603a77ee0e566b2ffdac"} Oct 08 20:47:38 crc kubenswrapper[4988]: I1008 20:47:38.145973 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"59722276-fba4-4321-9c14-696d0c1c98ce","Type":"ContainerStarted","Data":"860525eb08e1d02851962aa9cc93a3321af57d5b9fc001ba07a44a317cec1a7b"} Oct 08 20:47:38 crc kubenswrapper[4988]: I1008 20:47:38.169876 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.169857307 podStartE2EDuration="2.169857307s" podCreationTimestamp="2025-10-08 20:47:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 20:47:38.168832595 +0000 UTC m=+9403.618675375" watchObservedRunningTime="2025-10-08 20:47:38.169857307 +0000 UTC m=+9403.619700087" Oct 08 20:47:38 crc kubenswrapper[4988]: I1008 20:47:38.195334 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.195308069 podStartE2EDuration="2.195308069s" podCreationTimestamp="2025-10-08 20:47:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 20:47:38.189559636 +0000 UTC m=+9403.639402406" watchObservedRunningTime="2025-10-08 20:47:38.195308069 +0000 UTC m=+9403.645150859" Oct 08 20:47:39 crc kubenswrapper[4988]: I1008 20:47:39.454261 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Oct 08 20:47:40 crc kubenswrapper[4988]: I1008 20:47:40.962795 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Oct 08 20:47:41 crc kubenswrapper[4988]: I1008 20:47:41.131840 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/667c78af-4a60-4271-88c6-ef5ac33982a6-config-data\") pod \"667c78af-4a60-4271-88c6-ef5ac33982a6\" (UID: \"667c78af-4a60-4271-88c6-ef5ac33982a6\") " Oct 08 20:47:41 crc kubenswrapper[4988]: I1008 20:47:41.131940 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/667c78af-4a60-4271-88c6-ef5ac33982a6-combined-ca-bundle\") pod \"667c78af-4a60-4271-88c6-ef5ac33982a6\" (UID: \"667c78af-4a60-4271-88c6-ef5ac33982a6\") " Oct 08 20:47:41 crc kubenswrapper[4988]: I1008 20:47:41.132279 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wrwf7\" (UniqueName: \"kubernetes.io/projected/667c78af-4a60-4271-88c6-ef5ac33982a6-kube-api-access-wrwf7\") pod \"667c78af-4a60-4271-88c6-ef5ac33982a6\" (UID: \"667c78af-4a60-4271-88c6-ef5ac33982a6\") " Oct 08 20:47:41 crc kubenswrapper[4988]: I1008 20:47:41.138630 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/667c78af-4a60-4271-88c6-ef5ac33982a6-kube-api-access-wrwf7" (OuterVolumeSpecName: "kube-api-access-wrwf7") pod "667c78af-4a60-4271-88c6-ef5ac33982a6" (UID: "667c78af-4a60-4271-88c6-ef5ac33982a6"). InnerVolumeSpecName "kube-api-access-wrwf7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 20:47:41 crc kubenswrapper[4988]: I1008 20:47:41.169474 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/667c78af-4a60-4271-88c6-ef5ac33982a6-config-data" (OuterVolumeSpecName: "config-data") pod "667c78af-4a60-4271-88c6-ef5ac33982a6" (UID: "667c78af-4a60-4271-88c6-ef5ac33982a6"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 20:47:41 crc kubenswrapper[4988]: I1008 20:47:41.175482 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/667c78af-4a60-4271-88c6-ef5ac33982a6-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "667c78af-4a60-4271-88c6-ef5ac33982a6" (UID: "667c78af-4a60-4271-88c6-ef5ac33982a6"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 20:47:41 crc kubenswrapper[4988]: I1008 20:47:41.180930 4988 generic.go:334] "Generic (PLEG): container finished" podID="667c78af-4a60-4271-88c6-ef5ac33982a6" containerID="c2884809446b412404e04b0517c34a161c686d7e35b437b35238d006fdbe1300" exitCode=0 Oct 08 20:47:41 crc kubenswrapper[4988]: I1008 20:47:41.181002 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"667c78af-4a60-4271-88c6-ef5ac33982a6","Type":"ContainerDied","Data":"c2884809446b412404e04b0517c34a161c686d7e35b437b35238d006fdbe1300"} Oct 08 20:47:41 crc kubenswrapper[4988]: I1008 20:47:41.181038 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"667c78af-4a60-4271-88c6-ef5ac33982a6","Type":"ContainerDied","Data":"cc8d4a55b18c04765f9700f188f9b9acc26b4e0b1269759d871dc3ad10d53413"} Oct 08 20:47:41 crc kubenswrapper[4988]: I1008 20:47:41.181057 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Oct 08 20:47:41 crc kubenswrapper[4988]: I1008 20:47:41.181107 4988 scope.go:117] "RemoveContainer" containerID="c2884809446b412404e04b0517c34a161c686d7e35b437b35238d006fdbe1300" Oct 08 20:47:41 crc kubenswrapper[4988]: I1008 20:47:41.236049 4988 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wrwf7\" (UniqueName: \"kubernetes.io/projected/667c78af-4a60-4271-88c6-ef5ac33982a6-kube-api-access-wrwf7\") on node \"crc\" DevicePath \"\"" Oct 08 20:47:41 crc kubenswrapper[4988]: I1008 20:47:41.236404 4988 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/667c78af-4a60-4271-88c6-ef5ac33982a6-config-data\") on node \"crc\" DevicePath \"\"" Oct 08 20:47:41 crc kubenswrapper[4988]: I1008 20:47:41.236421 4988 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/667c78af-4a60-4271-88c6-ef5ac33982a6-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 20:47:41 crc kubenswrapper[4988]: I1008 20:47:41.293162 4988 scope.go:117] "RemoveContainer" containerID="c2884809446b412404e04b0517c34a161c686d7e35b437b35238d006fdbe1300" Oct 08 20:47:41 crc kubenswrapper[4988]: E1008 20:47:41.293490 4988 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c2884809446b412404e04b0517c34a161c686d7e35b437b35238d006fdbe1300\": container with ID starting with c2884809446b412404e04b0517c34a161c686d7e35b437b35238d006fdbe1300 not found: ID does not exist" containerID="c2884809446b412404e04b0517c34a161c686d7e35b437b35238d006fdbe1300" Oct 08 20:47:41 crc kubenswrapper[4988]: I1008 20:47:41.293513 4988 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c2884809446b412404e04b0517c34a161c686d7e35b437b35238d006fdbe1300"} err="failed to get container status \"c2884809446b412404e04b0517c34a161c686d7e35b437b35238d006fdbe1300\": rpc error: code = NotFound desc = could not find container \"c2884809446b412404e04b0517c34a161c686d7e35b437b35238d006fdbe1300\": container with ID starting with c2884809446b412404e04b0517c34a161c686d7e35b437b35238d006fdbe1300 not found: ID does not exist" Oct 08 20:47:41 crc kubenswrapper[4988]: I1008 20:47:41.310435 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-0"] Oct 08 20:47:41 crc kubenswrapper[4988]: I1008 20:47:41.326868 4988 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-conductor-0"] Oct 08 20:47:41 crc kubenswrapper[4988]: I1008 20:47:41.336325 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-0"] Oct 08 20:47:41 crc kubenswrapper[4988]: E1008 20:47:41.337157 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="667c78af-4a60-4271-88c6-ef5ac33982a6" containerName="nova-cell0-conductor-conductor" Oct 08 20:47:41 crc kubenswrapper[4988]: I1008 20:47:41.337204 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="667c78af-4a60-4271-88c6-ef5ac33982a6" containerName="nova-cell0-conductor-conductor" Oct 08 20:47:41 crc kubenswrapper[4988]: I1008 20:47:41.337668 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="667c78af-4a60-4271-88c6-ef5ac33982a6" containerName="nova-cell0-conductor-conductor" Oct 08 20:47:41 crc kubenswrapper[4988]: I1008 20:47:41.338879 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Oct 08 20:47:41 crc kubenswrapper[4988]: I1008 20:47:41.344706 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Oct 08 20:47:41 crc kubenswrapper[4988]: I1008 20:47:41.345940 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Oct 08 20:47:41 crc kubenswrapper[4988]: I1008 20:47:41.461605 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/02f910a2-e03f-427d-b8be-e6ff9c3d6025-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"02f910a2-e03f-427d-b8be-e6ff9c3d6025\") " pod="openstack/nova-cell0-conductor-0" Oct 08 20:47:41 crc kubenswrapper[4988]: I1008 20:47:41.461679 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9wfzv\" (UniqueName: \"kubernetes.io/projected/02f910a2-e03f-427d-b8be-e6ff9c3d6025-kube-api-access-9wfzv\") pod \"nova-cell0-conductor-0\" (UID: \"02f910a2-e03f-427d-b8be-e6ff9c3d6025\") " pod="openstack/nova-cell0-conductor-0" Oct 08 20:47:41 crc kubenswrapper[4988]: I1008 20:47:41.462001 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/02f910a2-e03f-427d-b8be-e6ff9c3d6025-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"02f910a2-e03f-427d-b8be-e6ff9c3d6025\") " pod="openstack/nova-cell0-conductor-0" Oct 08 20:47:41 crc kubenswrapper[4988]: I1008 20:47:41.564519 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/02f910a2-e03f-427d-b8be-e6ff9c3d6025-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"02f910a2-e03f-427d-b8be-e6ff9c3d6025\") " pod="openstack/nova-cell0-conductor-0" Oct 08 20:47:41 crc kubenswrapper[4988]: I1008 20:47:41.564710 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/02f910a2-e03f-427d-b8be-e6ff9c3d6025-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"02f910a2-e03f-427d-b8be-e6ff9c3d6025\") " pod="openstack/nova-cell0-conductor-0" Oct 08 20:47:41 crc kubenswrapper[4988]: I1008 20:47:41.564771 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9wfzv\" (UniqueName: \"kubernetes.io/projected/02f910a2-e03f-427d-b8be-e6ff9c3d6025-kube-api-access-9wfzv\") pod \"nova-cell0-conductor-0\" (UID: \"02f910a2-e03f-427d-b8be-e6ff9c3d6025\") " pod="openstack/nova-cell0-conductor-0" Oct 08 20:47:41 crc kubenswrapper[4988]: I1008 20:47:41.572871 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/02f910a2-e03f-427d-b8be-e6ff9c3d6025-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"02f910a2-e03f-427d-b8be-e6ff9c3d6025\") " pod="openstack/nova-cell0-conductor-0" Oct 08 20:47:41 crc kubenswrapper[4988]: I1008 20:47:41.573019 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/02f910a2-e03f-427d-b8be-e6ff9c3d6025-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"02f910a2-e03f-427d-b8be-e6ff9c3d6025\") " pod="openstack/nova-cell0-conductor-0" Oct 08 20:47:41 crc kubenswrapper[4988]: I1008 20:47:41.587788 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9wfzv\" (UniqueName: \"kubernetes.io/projected/02f910a2-e03f-427d-b8be-e6ff9c3d6025-kube-api-access-9wfzv\") pod \"nova-cell0-conductor-0\" (UID: \"02f910a2-e03f-427d-b8be-e6ff9c3d6025\") " pod="openstack/nova-cell0-conductor-0" Oct 08 20:47:41 crc kubenswrapper[4988]: I1008 20:47:41.670606 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Oct 08 20:47:41 crc kubenswrapper[4988]: I1008 20:47:41.859537 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Oct 08 20:47:41 crc kubenswrapper[4988]: I1008 20:47:41.861439 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Oct 08 20:47:42 crc kubenswrapper[4988]: I1008 20:47:42.125545 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Oct 08 20:47:42 crc kubenswrapper[4988]: I1008 20:47:42.191637 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"02f910a2-e03f-427d-b8be-e6ff9c3d6025","Type":"ContainerStarted","Data":"fc8a162e77bd2bcc5a7d956b0fc209a3b4458d912eed7a09b77548aacb5e15ee"} Oct 08 20:47:43 crc kubenswrapper[4988]: I1008 20:47:43.211228 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"02f910a2-e03f-427d-b8be-e6ff9c3d6025","Type":"ContainerStarted","Data":"3d5f681e7d114eb78c0ce1f5c2d0d1f147fe196bad5befeedcb01ebcd9a1022d"} Oct 08 20:47:43 crc kubenswrapper[4988]: I1008 20:47:43.211716 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell0-conductor-0" Oct 08 20:47:43 crc kubenswrapper[4988]: I1008 20:47:43.237964 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-0" podStartSLOduration=2.23794107 podStartE2EDuration="2.23794107s" podCreationTimestamp="2025-10-08 20:47:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 20:47:43.235705499 +0000 UTC m=+9408.685548289" watchObservedRunningTime="2025-10-08 20:47:43.23794107 +0000 UTC m=+9408.687783880" Oct 08 20:47:43 crc kubenswrapper[4988]: I1008 20:47:43.252834 4988 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="667c78af-4a60-4271-88c6-ef5ac33982a6" path="/var/lib/kubelet/pods/667c78af-4a60-4271-88c6-ef5ac33982a6/volumes" Oct 08 20:47:44 crc kubenswrapper[4988]: I1008 20:47:44.453678 4988 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Oct 08 20:47:45 crc kubenswrapper[4988]: I1008 20:47:45.125717 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-conductor-0" Oct 08 20:47:45 crc kubenswrapper[4988]: I1008 20:47:45.224919 4988 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Oct 08 20:47:45 crc kubenswrapper[4988]: I1008 20:47:45.287432 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Oct 08 20:47:46 crc kubenswrapper[4988]: I1008 20:47:46.859701 4988 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Oct 08 20:47:46 crc kubenswrapper[4988]: I1008 20:47:46.860100 4988 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Oct 08 20:47:46 crc kubenswrapper[4988]: I1008 20:47:46.961693 4988 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 08 20:47:46 crc kubenswrapper[4988]: I1008 20:47:46.961845 4988 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 08 20:47:47 crc kubenswrapper[4988]: I1008 20:47:47.874562 4988 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="59722276-fba4-4321-9c14-696d0c1c98ce" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.1.207:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 08 20:47:47 crc kubenswrapper[4988]: I1008 20:47:47.874569 4988 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="59722276-fba4-4321-9c14-696d0c1c98ce" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.1.207:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 08 20:47:47 crc kubenswrapper[4988]: I1008 20:47:47.977624 4988 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="a881e1d1-a353-47e5-83be-b264f9552c26" containerName="nova-api-log" probeResult="failure" output="Get \"https://10.217.1.208:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 08 20:47:47 crc kubenswrapper[4988]: I1008 20:47:47.977686 4988 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="a881e1d1-a353-47e5-83be-b264f9552c26" containerName="nova-api-api" probeResult="failure" output="Get \"https://10.217.1.208:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 08 20:47:49 crc kubenswrapper[4988]: I1008 20:47:49.239500 4988 scope.go:117] "RemoveContainer" containerID="6382c087e916f3958809279cab4d35e45e3b3ccd0cb9d8f0da5a015741773071" Oct 08 20:47:49 crc kubenswrapper[4988]: E1008 20:47:49.240312 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nm54f_openshift-machine-config-operator(f328a20b-5283-407e-8a93-2ab2da599f18)\"" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" Oct 08 20:47:51 crc kubenswrapper[4988]: I1008 20:47:51.700679 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell0-conductor-0" Oct 08 20:47:56 crc kubenswrapper[4988]: I1008 20:47:56.867723 4988 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Oct 08 20:47:56 crc kubenswrapper[4988]: I1008 20:47:56.868671 4988 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Oct 08 20:47:56 crc kubenswrapper[4988]: I1008 20:47:56.875589 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Oct 08 20:47:56 crc kubenswrapper[4988]: I1008 20:47:56.877511 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Oct 08 20:47:56 crc kubenswrapper[4988]: I1008 20:47:56.977371 4988 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Oct 08 20:47:56 crc kubenswrapper[4988]: I1008 20:47:56.984892 4988 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Oct 08 20:47:56 crc kubenswrapper[4988]: I1008 20:47:56.985871 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Oct 08 20:47:57 crc kubenswrapper[4988]: I1008 20:47:57.010509 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Oct 08 20:47:57 crc kubenswrapper[4988]: I1008 20:47:57.373676 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Oct 08 20:47:57 crc kubenswrapper[4988]: I1008 20:47:57.382418 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Oct 08 20:48:02 crc kubenswrapper[4988]: I1008 20:48:02.238887 4988 scope.go:117] "RemoveContainer" containerID="6382c087e916f3958809279cab4d35e45e3b3ccd0cb9d8f0da5a015741773071" Oct 08 20:48:02 crc kubenswrapper[4988]: E1008 20:48:02.239829 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nm54f_openshift-machine-config-operator(f328a20b-5283-407e-8a93-2ab2da599f18)\"" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" Oct 08 20:48:15 crc kubenswrapper[4988]: I1008 20:48:15.243705 4988 scope.go:117] "RemoveContainer" containerID="6382c087e916f3958809279cab4d35e45e3b3ccd0cb9d8f0da5a015741773071" Oct 08 20:48:15 crc kubenswrapper[4988]: E1008 20:48:15.244417 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nm54f_openshift-machine-config-operator(f328a20b-5283-407e-8a93-2ab2da599f18)\"" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" Oct 08 20:48:27 crc kubenswrapper[4988]: I1008 20:48:27.238603 4988 scope.go:117] "RemoveContainer" containerID="6382c087e916f3958809279cab4d35e45e3b3ccd0cb9d8f0da5a015741773071" Oct 08 20:48:27 crc kubenswrapper[4988]: E1008 20:48:27.241453 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nm54f_openshift-machine-config-operator(f328a20b-5283-407e-8a93-2ab2da599f18)\"" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" Oct 08 20:48:40 crc kubenswrapper[4988]: I1008 20:48:40.238528 4988 scope.go:117] "RemoveContainer" containerID="6382c087e916f3958809279cab4d35e45e3b3ccd0cb9d8f0da5a015741773071" Oct 08 20:48:40 crc kubenswrapper[4988]: E1008 20:48:40.239831 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nm54f_openshift-machine-config-operator(f328a20b-5283-407e-8a93-2ab2da599f18)\"" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" Oct 08 20:48:55 crc kubenswrapper[4988]: I1008 20:48:55.238376 4988 scope.go:117] "RemoveContainer" containerID="6382c087e916f3958809279cab4d35e45e3b3ccd0cb9d8f0da5a015741773071" Oct 08 20:48:56 crc kubenswrapper[4988]: I1008 20:48:56.149455 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" event={"ID":"f328a20b-5283-407e-8a93-2ab2da599f18","Type":"ContainerStarted","Data":"3a1d870522f1f317cf2b669f777c577892b10a5614c2e13e7030b149ce77fbb8"} Oct 08 20:48:59 crc kubenswrapper[4988]: I1008 20:48:59.403587 4988 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/swift-proxy-b8548dfd6-qblss" podUID="679a7d0a-3ad1-45c8-b1ff-6772a862379a" containerName="proxy-server" probeResult="failure" output="HTTP probe failed with statuscode: 502" Oct 08 20:51:23 crc kubenswrapper[4988]: I1008 20:51:23.338760 4988 patch_prober.go:28] interesting pod/machine-config-daemon-nm54f container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 20:51:23 crc kubenswrapper[4988]: I1008 20:51:23.340213 4988 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 20:51:53 crc kubenswrapper[4988]: I1008 20:51:53.338557 4988 patch_prober.go:28] interesting pod/machine-config-daemon-nm54f container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 20:51:53 crc kubenswrapper[4988]: I1008 20:51:53.339237 4988 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 20:52:23 crc kubenswrapper[4988]: I1008 20:52:23.338291 4988 patch_prober.go:28] interesting pod/machine-config-daemon-nm54f container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 20:52:23 crc kubenswrapper[4988]: I1008 20:52:23.338845 4988 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 20:52:23 crc kubenswrapper[4988]: I1008 20:52:23.338901 4988 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" Oct 08 20:52:23 crc kubenswrapper[4988]: I1008 20:52:23.340194 4988 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"3a1d870522f1f317cf2b669f777c577892b10a5614c2e13e7030b149ce77fbb8"} pod="openshift-machine-config-operator/machine-config-daemon-nm54f" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 08 20:52:23 crc kubenswrapper[4988]: I1008 20:52:23.340410 4988 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" containerName="machine-config-daemon" containerID="cri-o://3a1d870522f1f317cf2b669f777c577892b10a5614c2e13e7030b149ce77fbb8" gracePeriod=600 Oct 08 20:52:23 crc kubenswrapper[4988]: I1008 20:52:23.878658 4988 generic.go:334] "Generic (PLEG): container finished" podID="f328a20b-5283-407e-8a93-2ab2da599f18" containerID="3a1d870522f1f317cf2b669f777c577892b10a5614c2e13e7030b149ce77fbb8" exitCode=0 Oct 08 20:52:23 crc kubenswrapper[4988]: I1008 20:52:23.878919 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" event={"ID":"f328a20b-5283-407e-8a93-2ab2da599f18","Type":"ContainerDied","Data":"3a1d870522f1f317cf2b669f777c577892b10a5614c2e13e7030b149ce77fbb8"} Oct 08 20:52:23 crc kubenswrapper[4988]: I1008 20:52:23.879033 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" event={"ID":"f328a20b-5283-407e-8a93-2ab2da599f18","Type":"ContainerStarted","Data":"fc5906dffdd6ca78fe0b71f3079a04d7d08624d242c87fd146529c035ee33bca"} Oct 08 20:52:23 crc kubenswrapper[4988]: I1008 20:52:23.879050 4988 scope.go:117] "RemoveContainer" containerID="6382c087e916f3958809279cab4d35e45e3b3ccd0cb9d8f0da5a015741773071" Oct 08 20:53:18 crc kubenswrapper[4988]: I1008 20:53:18.129371 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-rspch"] Oct 08 20:53:18 crc kubenswrapper[4988]: I1008 20:53:18.134689 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-rspch" Oct 08 20:53:18 crc kubenswrapper[4988]: I1008 20:53:18.154962 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-rspch"] Oct 08 20:53:18 crc kubenswrapper[4988]: I1008 20:53:18.180479 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9znzs\" (UniqueName: \"kubernetes.io/projected/dba3ff46-0700-4473-9fbc-9ae3f8a6de5c-kube-api-access-9znzs\") pod \"certified-operators-rspch\" (UID: \"dba3ff46-0700-4473-9fbc-9ae3f8a6de5c\") " pod="openshift-marketplace/certified-operators-rspch" Oct 08 20:53:18 crc kubenswrapper[4988]: I1008 20:53:18.180815 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dba3ff46-0700-4473-9fbc-9ae3f8a6de5c-utilities\") pod \"certified-operators-rspch\" (UID: \"dba3ff46-0700-4473-9fbc-9ae3f8a6de5c\") " pod="openshift-marketplace/certified-operators-rspch" Oct 08 20:53:18 crc kubenswrapper[4988]: I1008 20:53:18.181235 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dba3ff46-0700-4473-9fbc-9ae3f8a6de5c-catalog-content\") pod \"certified-operators-rspch\" (UID: \"dba3ff46-0700-4473-9fbc-9ae3f8a6de5c\") " pod="openshift-marketplace/certified-operators-rspch" Oct 08 20:53:18 crc kubenswrapper[4988]: I1008 20:53:18.283586 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dba3ff46-0700-4473-9fbc-9ae3f8a6de5c-catalog-content\") pod \"certified-operators-rspch\" (UID: \"dba3ff46-0700-4473-9fbc-9ae3f8a6de5c\") " pod="openshift-marketplace/certified-operators-rspch" Oct 08 20:53:18 crc kubenswrapper[4988]: I1008 20:53:18.283786 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9znzs\" (UniqueName: \"kubernetes.io/projected/dba3ff46-0700-4473-9fbc-9ae3f8a6de5c-kube-api-access-9znzs\") pod \"certified-operators-rspch\" (UID: \"dba3ff46-0700-4473-9fbc-9ae3f8a6de5c\") " pod="openshift-marketplace/certified-operators-rspch" Oct 08 20:53:18 crc kubenswrapper[4988]: I1008 20:53:18.283813 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dba3ff46-0700-4473-9fbc-9ae3f8a6de5c-utilities\") pod \"certified-operators-rspch\" (UID: \"dba3ff46-0700-4473-9fbc-9ae3f8a6de5c\") " pod="openshift-marketplace/certified-operators-rspch" Oct 08 20:53:18 crc kubenswrapper[4988]: I1008 20:53:18.284250 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dba3ff46-0700-4473-9fbc-9ae3f8a6de5c-utilities\") pod \"certified-operators-rspch\" (UID: \"dba3ff46-0700-4473-9fbc-9ae3f8a6de5c\") " pod="openshift-marketplace/certified-operators-rspch" Oct 08 20:53:18 crc kubenswrapper[4988]: I1008 20:53:18.284665 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dba3ff46-0700-4473-9fbc-9ae3f8a6de5c-catalog-content\") pod \"certified-operators-rspch\" (UID: \"dba3ff46-0700-4473-9fbc-9ae3f8a6de5c\") " pod="openshift-marketplace/certified-operators-rspch" Oct 08 20:53:18 crc kubenswrapper[4988]: I1008 20:53:18.318711 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9znzs\" (UniqueName: \"kubernetes.io/projected/dba3ff46-0700-4473-9fbc-9ae3f8a6de5c-kube-api-access-9znzs\") pod \"certified-operators-rspch\" (UID: \"dba3ff46-0700-4473-9fbc-9ae3f8a6de5c\") " pod="openshift-marketplace/certified-operators-rspch" Oct 08 20:53:18 crc kubenswrapper[4988]: I1008 20:53:18.474421 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-rspch" Oct 08 20:53:19 crc kubenswrapper[4988]: I1008 20:53:19.084041 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-rspch"] Oct 08 20:53:19 crc kubenswrapper[4988]: I1008 20:53:19.592445 4988 generic.go:334] "Generic (PLEG): container finished" podID="dba3ff46-0700-4473-9fbc-9ae3f8a6de5c" containerID="873d82ce9692280390b9a0754a4116a90930dbdd762e44b1606a95e1d9d1bccd" exitCode=0 Oct 08 20:53:19 crc kubenswrapper[4988]: I1008 20:53:19.592519 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rspch" event={"ID":"dba3ff46-0700-4473-9fbc-9ae3f8a6de5c","Type":"ContainerDied","Data":"873d82ce9692280390b9a0754a4116a90930dbdd762e44b1606a95e1d9d1bccd"} Oct 08 20:53:19 crc kubenswrapper[4988]: I1008 20:53:19.592802 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rspch" event={"ID":"dba3ff46-0700-4473-9fbc-9ae3f8a6de5c","Type":"ContainerStarted","Data":"520f0a1321b6ad8bac4027145f1755510fc895dc6377daf37e79383a4e55f736"} Oct 08 20:53:19 crc kubenswrapper[4988]: I1008 20:53:19.595773 4988 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 08 20:53:21 crc kubenswrapper[4988]: I1008 20:53:21.621142 4988 generic.go:334] "Generic (PLEG): container finished" podID="dba3ff46-0700-4473-9fbc-9ae3f8a6de5c" containerID="442f97734a9a0a8e20379e69545cec4b48a8e694f1031ee1676ce13620a1c577" exitCode=0 Oct 08 20:53:21 crc kubenswrapper[4988]: I1008 20:53:21.621287 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rspch" event={"ID":"dba3ff46-0700-4473-9fbc-9ae3f8a6de5c","Type":"ContainerDied","Data":"442f97734a9a0a8e20379e69545cec4b48a8e694f1031ee1676ce13620a1c577"} Oct 08 20:53:22 crc kubenswrapper[4988]: I1008 20:53:22.636508 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rspch" event={"ID":"dba3ff46-0700-4473-9fbc-9ae3f8a6de5c","Type":"ContainerStarted","Data":"4c17de557c8a14d9c35ebcbd2309cd409c11cb970637f6bf6b539eefe4f6d997"} Oct 08 20:53:22 crc kubenswrapper[4988]: I1008 20:53:22.660452 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-rspch" podStartSLOduration=2.036454881 podStartE2EDuration="4.660428604s" podCreationTimestamp="2025-10-08 20:53:18 +0000 UTC" firstStartedPulling="2025-10-08 20:53:19.59532054 +0000 UTC m=+9745.045163330" lastFinishedPulling="2025-10-08 20:53:22.219294273 +0000 UTC m=+9747.669137053" observedRunningTime="2025-10-08 20:53:22.658018877 +0000 UTC m=+9748.107861657" watchObservedRunningTime="2025-10-08 20:53:22.660428604 +0000 UTC m=+9748.110271394" Oct 08 20:53:28 crc kubenswrapper[4988]: I1008 20:53:28.475035 4988 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-rspch" Oct 08 20:53:28 crc kubenswrapper[4988]: I1008 20:53:28.475728 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-rspch" Oct 08 20:53:28 crc kubenswrapper[4988]: I1008 20:53:28.543118 4988 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-rspch" Oct 08 20:53:28 crc kubenswrapper[4988]: I1008 20:53:28.742340 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-rspch" Oct 08 20:53:28 crc kubenswrapper[4988]: I1008 20:53:28.803489 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-rspch"] Oct 08 20:53:30 crc kubenswrapper[4988]: I1008 20:53:30.726247 4988 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-rspch" podUID="dba3ff46-0700-4473-9fbc-9ae3f8a6de5c" containerName="registry-server" containerID="cri-o://4c17de557c8a14d9c35ebcbd2309cd409c11cb970637f6bf6b539eefe4f6d997" gracePeriod=2 Oct 08 20:53:31 crc kubenswrapper[4988]: I1008 20:53:31.258861 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-rspch" Oct 08 20:53:31 crc kubenswrapper[4988]: I1008 20:53:31.330620 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9znzs\" (UniqueName: \"kubernetes.io/projected/dba3ff46-0700-4473-9fbc-9ae3f8a6de5c-kube-api-access-9znzs\") pod \"dba3ff46-0700-4473-9fbc-9ae3f8a6de5c\" (UID: \"dba3ff46-0700-4473-9fbc-9ae3f8a6de5c\") " Oct 08 20:53:31 crc kubenswrapper[4988]: I1008 20:53:31.331222 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dba3ff46-0700-4473-9fbc-9ae3f8a6de5c-catalog-content\") pod \"dba3ff46-0700-4473-9fbc-9ae3f8a6de5c\" (UID: \"dba3ff46-0700-4473-9fbc-9ae3f8a6de5c\") " Oct 08 20:53:31 crc kubenswrapper[4988]: I1008 20:53:31.331293 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dba3ff46-0700-4473-9fbc-9ae3f8a6de5c-utilities\") pod \"dba3ff46-0700-4473-9fbc-9ae3f8a6de5c\" (UID: \"dba3ff46-0700-4473-9fbc-9ae3f8a6de5c\") " Oct 08 20:53:31 crc kubenswrapper[4988]: I1008 20:53:31.332007 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/dba3ff46-0700-4473-9fbc-9ae3f8a6de5c-utilities" (OuterVolumeSpecName: "utilities") pod "dba3ff46-0700-4473-9fbc-9ae3f8a6de5c" (UID: "dba3ff46-0700-4473-9fbc-9ae3f8a6de5c"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 20:53:31 crc kubenswrapper[4988]: I1008 20:53:31.333372 4988 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dba3ff46-0700-4473-9fbc-9ae3f8a6de5c-utilities\") on node \"crc\" DevicePath \"\"" Oct 08 20:53:31 crc kubenswrapper[4988]: I1008 20:53:31.341789 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dba3ff46-0700-4473-9fbc-9ae3f8a6de5c-kube-api-access-9znzs" (OuterVolumeSpecName: "kube-api-access-9znzs") pod "dba3ff46-0700-4473-9fbc-9ae3f8a6de5c" (UID: "dba3ff46-0700-4473-9fbc-9ae3f8a6de5c"). InnerVolumeSpecName "kube-api-access-9znzs". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 20:53:31 crc kubenswrapper[4988]: I1008 20:53:31.387710 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/dba3ff46-0700-4473-9fbc-9ae3f8a6de5c-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "dba3ff46-0700-4473-9fbc-9ae3f8a6de5c" (UID: "dba3ff46-0700-4473-9fbc-9ae3f8a6de5c"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 20:53:31 crc kubenswrapper[4988]: I1008 20:53:31.435427 4988 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9znzs\" (UniqueName: \"kubernetes.io/projected/dba3ff46-0700-4473-9fbc-9ae3f8a6de5c-kube-api-access-9znzs\") on node \"crc\" DevicePath \"\"" Oct 08 20:53:31 crc kubenswrapper[4988]: I1008 20:53:31.435493 4988 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dba3ff46-0700-4473-9fbc-9ae3f8a6de5c-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 08 20:53:31 crc kubenswrapper[4988]: I1008 20:53:31.744615 4988 generic.go:334] "Generic (PLEG): container finished" podID="dba3ff46-0700-4473-9fbc-9ae3f8a6de5c" containerID="4c17de557c8a14d9c35ebcbd2309cd409c11cb970637f6bf6b539eefe4f6d997" exitCode=0 Oct 08 20:53:31 crc kubenswrapper[4988]: I1008 20:53:31.744686 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rspch" event={"ID":"dba3ff46-0700-4473-9fbc-9ae3f8a6de5c","Type":"ContainerDied","Data":"4c17de557c8a14d9c35ebcbd2309cd409c11cb970637f6bf6b539eefe4f6d997"} Oct 08 20:53:31 crc kubenswrapper[4988]: I1008 20:53:31.744739 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rspch" event={"ID":"dba3ff46-0700-4473-9fbc-9ae3f8a6de5c","Type":"ContainerDied","Data":"520f0a1321b6ad8bac4027145f1755510fc895dc6377daf37e79383a4e55f736"} Oct 08 20:53:31 crc kubenswrapper[4988]: I1008 20:53:31.744751 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-rspch" Oct 08 20:53:31 crc kubenswrapper[4988]: I1008 20:53:31.744771 4988 scope.go:117] "RemoveContainer" containerID="4c17de557c8a14d9c35ebcbd2309cd409c11cb970637f6bf6b539eefe4f6d997" Oct 08 20:53:31 crc kubenswrapper[4988]: I1008 20:53:31.781747 4988 scope.go:117] "RemoveContainer" containerID="442f97734a9a0a8e20379e69545cec4b48a8e694f1031ee1676ce13620a1c577" Oct 08 20:53:31 crc kubenswrapper[4988]: I1008 20:53:31.811462 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-rspch"] Oct 08 20:53:31 crc kubenswrapper[4988]: I1008 20:53:31.819824 4988 scope.go:117] "RemoveContainer" containerID="873d82ce9692280390b9a0754a4116a90930dbdd762e44b1606a95e1d9d1bccd" Oct 08 20:53:31 crc kubenswrapper[4988]: I1008 20:53:31.824015 4988 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-rspch"] Oct 08 20:53:31 crc kubenswrapper[4988]: I1008 20:53:31.893147 4988 scope.go:117] "RemoveContainer" containerID="4c17de557c8a14d9c35ebcbd2309cd409c11cb970637f6bf6b539eefe4f6d997" Oct 08 20:53:31 crc kubenswrapper[4988]: E1008 20:53:31.893719 4988 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4c17de557c8a14d9c35ebcbd2309cd409c11cb970637f6bf6b539eefe4f6d997\": container with ID starting with 4c17de557c8a14d9c35ebcbd2309cd409c11cb970637f6bf6b539eefe4f6d997 not found: ID does not exist" containerID="4c17de557c8a14d9c35ebcbd2309cd409c11cb970637f6bf6b539eefe4f6d997" Oct 08 20:53:31 crc kubenswrapper[4988]: I1008 20:53:31.893770 4988 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4c17de557c8a14d9c35ebcbd2309cd409c11cb970637f6bf6b539eefe4f6d997"} err="failed to get container status \"4c17de557c8a14d9c35ebcbd2309cd409c11cb970637f6bf6b539eefe4f6d997\": rpc error: code = NotFound desc = could not find container \"4c17de557c8a14d9c35ebcbd2309cd409c11cb970637f6bf6b539eefe4f6d997\": container with ID starting with 4c17de557c8a14d9c35ebcbd2309cd409c11cb970637f6bf6b539eefe4f6d997 not found: ID does not exist" Oct 08 20:53:31 crc kubenswrapper[4988]: I1008 20:53:31.893803 4988 scope.go:117] "RemoveContainer" containerID="442f97734a9a0a8e20379e69545cec4b48a8e694f1031ee1676ce13620a1c577" Oct 08 20:53:31 crc kubenswrapper[4988]: E1008 20:53:31.894122 4988 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"442f97734a9a0a8e20379e69545cec4b48a8e694f1031ee1676ce13620a1c577\": container with ID starting with 442f97734a9a0a8e20379e69545cec4b48a8e694f1031ee1676ce13620a1c577 not found: ID does not exist" containerID="442f97734a9a0a8e20379e69545cec4b48a8e694f1031ee1676ce13620a1c577" Oct 08 20:53:31 crc kubenswrapper[4988]: I1008 20:53:31.894159 4988 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"442f97734a9a0a8e20379e69545cec4b48a8e694f1031ee1676ce13620a1c577"} err="failed to get container status \"442f97734a9a0a8e20379e69545cec4b48a8e694f1031ee1676ce13620a1c577\": rpc error: code = NotFound desc = could not find container \"442f97734a9a0a8e20379e69545cec4b48a8e694f1031ee1676ce13620a1c577\": container with ID starting with 442f97734a9a0a8e20379e69545cec4b48a8e694f1031ee1676ce13620a1c577 not found: ID does not exist" Oct 08 20:53:31 crc kubenswrapper[4988]: I1008 20:53:31.894183 4988 scope.go:117] "RemoveContainer" containerID="873d82ce9692280390b9a0754a4116a90930dbdd762e44b1606a95e1d9d1bccd" Oct 08 20:53:31 crc kubenswrapper[4988]: E1008 20:53:31.894476 4988 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"873d82ce9692280390b9a0754a4116a90930dbdd762e44b1606a95e1d9d1bccd\": container with ID starting with 873d82ce9692280390b9a0754a4116a90930dbdd762e44b1606a95e1d9d1bccd not found: ID does not exist" containerID="873d82ce9692280390b9a0754a4116a90930dbdd762e44b1606a95e1d9d1bccd" Oct 08 20:53:31 crc kubenswrapper[4988]: I1008 20:53:31.894513 4988 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"873d82ce9692280390b9a0754a4116a90930dbdd762e44b1606a95e1d9d1bccd"} err="failed to get container status \"873d82ce9692280390b9a0754a4116a90930dbdd762e44b1606a95e1d9d1bccd\": rpc error: code = NotFound desc = could not find container \"873d82ce9692280390b9a0754a4116a90930dbdd762e44b1606a95e1d9d1bccd\": container with ID starting with 873d82ce9692280390b9a0754a4116a90930dbdd762e44b1606a95e1d9d1bccd not found: ID does not exist" Oct 08 20:53:33 crc kubenswrapper[4988]: I1008 20:53:33.259077 4988 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="dba3ff46-0700-4473-9fbc-9ae3f8a6de5c" path="/var/lib/kubelet/pods/dba3ff46-0700-4473-9fbc-9ae3f8a6de5c/volumes" Oct 08 20:54:23 crc kubenswrapper[4988]: I1008 20:54:23.338314 4988 patch_prober.go:28] interesting pod/machine-config-daemon-nm54f container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 20:54:23 crc kubenswrapper[4988]: I1008 20:54:23.339175 4988 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 20:54:33 crc kubenswrapper[4988]: I1008 20:54:33.413307 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-6dxkb"] Oct 08 20:54:33 crc kubenswrapper[4988]: E1008 20:54:33.414459 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dba3ff46-0700-4473-9fbc-9ae3f8a6de5c" containerName="extract-utilities" Oct 08 20:54:33 crc kubenswrapper[4988]: I1008 20:54:33.414480 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="dba3ff46-0700-4473-9fbc-9ae3f8a6de5c" containerName="extract-utilities" Oct 08 20:54:33 crc kubenswrapper[4988]: E1008 20:54:33.414509 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dba3ff46-0700-4473-9fbc-9ae3f8a6de5c" containerName="registry-server" Oct 08 20:54:33 crc kubenswrapper[4988]: I1008 20:54:33.414521 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="dba3ff46-0700-4473-9fbc-9ae3f8a6de5c" containerName="registry-server" Oct 08 20:54:33 crc kubenswrapper[4988]: E1008 20:54:33.414560 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dba3ff46-0700-4473-9fbc-9ae3f8a6de5c" containerName="extract-content" Oct 08 20:54:33 crc kubenswrapper[4988]: I1008 20:54:33.414573 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="dba3ff46-0700-4473-9fbc-9ae3f8a6de5c" containerName="extract-content" Oct 08 20:54:33 crc kubenswrapper[4988]: I1008 20:54:33.414979 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="dba3ff46-0700-4473-9fbc-9ae3f8a6de5c" containerName="registry-server" Oct 08 20:54:33 crc kubenswrapper[4988]: I1008 20:54:33.423140 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-6dxkb" Oct 08 20:54:33 crc kubenswrapper[4988]: I1008 20:54:33.434905 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-6dxkb"] Oct 08 20:54:33 crc kubenswrapper[4988]: I1008 20:54:33.459957 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-45jql\" (UniqueName: \"kubernetes.io/projected/ac5d09bc-5c99-4a7c-8632-ea0cfe1cf50a-kube-api-access-45jql\") pod \"community-operators-6dxkb\" (UID: \"ac5d09bc-5c99-4a7c-8632-ea0cfe1cf50a\") " pod="openshift-marketplace/community-operators-6dxkb" Oct 08 20:54:33 crc kubenswrapper[4988]: I1008 20:54:33.460062 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ac5d09bc-5c99-4a7c-8632-ea0cfe1cf50a-utilities\") pod \"community-operators-6dxkb\" (UID: \"ac5d09bc-5c99-4a7c-8632-ea0cfe1cf50a\") " pod="openshift-marketplace/community-operators-6dxkb" Oct 08 20:54:33 crc kubenswrapper[4988]: I1008 20:54:33.460173 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ac5d09bc-5c99-4a7c-8632-ea0cfe1cf50a-catalog-content\") pod \"community-operators-6dxkb\" (UID: \"ac5d09bc-5c99-4a7c-8632-ea0cfe1cf50a\") " pod="openshift-marketplace/community-operators-6dxkb" Oct 08 20:54:33 crc kubenswrapper[4988]: I1008 20:54:33.563174 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-45jql\" (UniqueName: \"kubernetes.io/projected/ac5d09bc-5c99-4a7c-8632-ea0cfe1cf50a-kube-api-access-45jql\") pod \"community-operators-6dxkb\" (UID: \"ac5d09bc-5c99-4a7c-8632-ea0cfe1cf50a\") " pod="openshift-marketplace/community-operators-6dxkb" Oct 08 20:54:33 crc kubenswrapper[4988]: I1008 20:54:33.563361 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ac5d09bc-5c99-4a7c-8632-ea0cfe1cf50a-utilities\") pod \"community-operators-6dxkb\" (UID: \"ac5d09bc-5c99-4a7c-8632-ea0cfe1cf50a\") " pod="openshift-marketplace/community-operators-6dxkb" Oct 08 20:54:33 crc kubenswrapper[4988]: I1008 20:54:33.563420 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ac5d09bc-5c99-4a7c-8632-ea0cfe1cf50a-catalog-content\") pod \"community-operators-6dxkb\" (UID: \"ac5d09bc-5c99-4a7c-8632-ea0cfe1cf50a\") " pod="openshift-marketplace/community-operators-6dxkb" Oct 08 20:54:33 crc kubenswrapper[4988]: I1008 20:54:33.564157 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ac5d09bc-5c99-4a7c-8632-ea0cfe1cf50a-catalog-content\") pod \"community-operators-6dxkb\" (UID: \"ac5d09bc-5c99-4a7c-8632-ea0cfe1cf50a\") " pod="openshift-marketplace/community-operators-6dxkb" Oct 08 20:54:33 crc kubenswrapper[4988]: I1008 20:54:33.564348 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ac5d09bc-5c99-4a7c-8632-ea0cfe1cf50a-utilities\") pod \"community-operators-6dxkb\" (UID: \"ac5d09bc-5c99-4a7c-8632-ea0cfe1cf50a\") " pod="openshift-marketplace/community-operators-6dxkb" Oct 08 20:54:33 crc kubenswrapper[4988]: I1008 20:54:33.584880 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-45jql\" (UniqueName: \"kubernetes.io/projected/ac5d09bc-5c99-4a7c-8632-ea0cfe1cf50a-kube-api-access-45jql\") pod \"community-operators-6dxkb\" (UID: \"ac5d09bc-5c99-4a7c-8632-ea0cfe1cf50a\") " pod="openshift-marketplace/community-operators-6dxkb" Oct 08 20:54:33 crc kubenswrapper[4988]: I1008 20:54:33.773206 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-6dxkb" Oct 08 20:54:34 crc kubenswrapper[4988]: W1008 20:54:34.368786 4988 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podac5d09bc_5c99_4a7c_8632_ea0cfe1cf50a.slice/crio-0a5a9bac2c67184d3bff7f2fb27a16ce8694a5e500ef5d461a22d97290a68ff7 WatchSource:0}: Error finding container 0a5a9bac2c67184d3bff7f2fb27a16ce8694a5e500ef5d461a22d97290a68ff7: Status 404 returned error can't find the container with id 0a5a9bac2c67184d3bff7f2fb27a16ce8694a5e500ef5d461a22d97290a68ff7 Oct 08 20:54:34 crc kubenswrapper[4988]: I1008 20:54:34.369438 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-6dxkb"] Oct 08 20:54:34 crc kubenswrapper[4988]: I1008 20:54:34.559753 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6dxkb" event={"ID":"ac5d09bc-5c99-4a7c-8632-ea0cfe1cf50a","Type":"ContainerStarted","Data":"0a5a9bac2c67184d3bff7f2fb27a16ce8694a5e500ef5d461a22d97290a68ff7"} Oct 08 20:54:35 crc kubenswrapper[4988]: I1008 20:54:35.572882 4988 generic.go:334] "Generic (PLEG): container finished" podID="ac5d09bc-5c99-4a7c-8632-ea0cfe1cf50a" containerID="b5b117598a7d5bfcdf783dcc601bcf241c8e2ec744fad00723a6a10e506e380b" exitCode=0 Oct 08 20:54:35 crc kubenswrapper[4988]: I1008 20:54:35.573001 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6dxkb" event={"ID":"ac5d09bc-5c99-4a7c-8632-ea0cfe1cf50a","Type":"ContainerDied","Data":"b5b117598a7d5bfcdf783dcc601bcf241c8e2ec744fad00723a6a10e506e380b"} Oct 08 20:54:36 crc kubenswrapper[4988]: I1008 20:54:36.587639 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6dxkb" event={"ID":"ac5d09bc-5c99-4a7c-8632-ea0cfe1cf50a","Type":"ContainerStarted","Data":"eac8b358e6759c11532fc8d88c98f54da96b89922db005599d3b23ec3824fc57"} Oct 08 20:54:38 crc kubenswrapper[4988]: I1008 20:54:38.617425 4988 generic.go:334] "Generic (PLEG): container finished" podID="ac5d09bc-5c99-4a7c-8632-ea0cfe1cf50a" containerID="eac8b358e6759c11532fc8d88c98f54da96b89922db005599d3b23ec3824fc57" exitCode=0 Oct 08 20:54:38 crc kubenswrapper[4988]: I1008 20:54:38.617498 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6dxkb" event={"ID":"ac5d09bc-5c99-4a7c-8632-ea0cfe1cf50a","Type":"ContainerDied","Data":"eac8b358e6759c11532fc8d88c98f54da96b89922db005599d3b23ec3824fc57"} Oct 08 20:54:39 crc kubenswrapper[4988]: I1008 20:54:39.630698 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6dxkb" event={"ID":"ac5d09bc-5c99-4a7c-8632-ea0cfe1cf50a","Type":"ContainerStarted","Data":"a5be2570d20b4208bd1e8ed7adae2ee93898e90ba4dbc11caa4c920957b76ada"} Oct 08 20:54:39 crc kubenswrapper[4988]: I1008 20:54:39.667407 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-6dxkb" podStartSLOduration=3.175567563 podStartE2EDuration="6.667376679s" podCreationTimestamp="2025-10-08 20:54:33 +0000 UTC" firstStartedPulling="2025-10-08 20:54:35.576172749 +0000 UTC m=+9821.026015519" lastFinishedPulling="2025-10-08 20:54:39.067981865 +0000 UTC m=+9824.517824635" observedRunningTime="2025-10-08 20:54:39.651561484 +0000 UTC m=+9825.101404294" watchObservedRunningTime="2025-10-08 20:54:39.667376679 +0000 UTC m=+9825.117219489" Oct 08 20:54:43 crc kubenswrapper[4988]: I1008 20:54:43.774241 4988 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-6dxkb" Oct 08 20:54:43 crc kubenswrapper[4988]: I1008 20:54:43.774842 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-6dxkb" Oct 08 20:54:43 crc kubenswrapper[4988]: I1008 20:54:43.825181 4988 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-6dxkb" Oct 08 20:54:44 crc kubenswrapper[4988]: I1008 20:54:44.755693 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-6dxkb" Oct 08 20:54:44 crc kubenswrapper[4988]: I1008 20:54:44.828754 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-6dxkb"] Oct 08 20:54:46 crc kubenswrapper[4988]: I1008 20:54:46.718062 4988 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-6dxkb" podUID="ac5d09bc-5c99-4a7c-8632-ea0cfe1cf50a" containerName="registry-server" containerID="cri-o://a5be2570d20b4208bd1e8ed7adae2ee93898e90ba4dbc11caa4c920957b76ada" gracePeriod=2 Oct 08 20:54:47 crc kubenswrapper[4988]: I1008 20:54:47.221070 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-6dxkb" Oct 08 20:54:47 crc kubenswrapper[4988]: I1008 20:54:47.418084 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ac5d09bc-5c99-4a7c-8632-ea0cfe1cf50a-catalog-content\") pod \"ac5d09bc-5c99-4a7c-8632-ea0cfe1cf50a\" (UID: \"ac5d09bc-5c99-4a7c-8632-ea0cfe1cf50a\") " Oct 08 20:54:47 crc kubenswrapper[4988]: I1008 20:54:47.418240 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ac5d09bc-5c99-4a7c-8632-ea0cfe1cf50a-utilities\") pod \"ac5d09bc-5c99-4a7c-8632-ea0cfe1cf50a\" (UID: \"ac5d09bc-5c99-4a7c-8632-ea0cfe1cf50a\") " Oct 08 20:54:47 crc kubenswrapper[4988]: I1008 20:54:47.418342 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-45jql\" (UniqueName: \"kubernetes.io/projected/ac5d09bc-5c99-4a7c-8632-ea0cfe1cf50a-kube-api-access-45jql\") pod \"ac5d09bc-5c99-4a7c-8632-ea0cfe1cf50a\" (UID: \"ac5d09bc-5c99-4a7c-8632-ea0cfe1cf50a\") " Oct 08 20:54:47 crc kubenswrapper[4988]: I1008 20:54:47.419600 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ac5d09bc-5c99-4a7c-8632-ea0cfe1cf50a-utilities" (OuterVolumeSpecName: "utilities") pod "ac5d09bc-5c99-4a7c-8632-ea0cfe1cf50a" (UID: "ac5d09bc-5c99-4a7c-8632-ea0cfe1cf50a"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 20:54:47 crc kubenswrapper[4988]: I1008 20:54:47.430010 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ac5d09bc-5c99-4a7c-8632-ea0cfe1cf50a-kube-api-access-45jql" (OuterVolumeSpecName: "kube-api-access-45jql") pod "ac5d09bc-5c99-4a7c-8632-ea0cfe1cf50a" (UID: "ac5d09bc-5c99-4a7c-8632-ea0cfe1cf50a"). InnerVolumeSpecName "kube-api-access-45jql". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 20:54:47 crc kubenswrapper[4988]: I1008 20:54:47.462255 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ac5d09bc-5c99-4a7c-8632-ea0cfe1cf50a-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "ac5d09bc-5c99-4a7c-8632-ea0cfe1cf50a" (UID: "ac5d09bc-5c99-4a7c-8632-ea0cfe1cf50a"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 20:54:47 crc kubenswrapper[4988]: I1008 20:54:47.521777 4988 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ac5d09bc-5c99-4a7c-8632-ea0cfe1cf50a-utilities\") on node \"crc\" DevicePath \"\"" Oct 08 20:54:47 crc kubenswrapper[4988]: I1008 20:54:47.521998 4988 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-45jql\" (UniqueName: \"kubernetes.io/projected/ac5d09bc-5c99-4a7c-8632-ea0cfe1cf50a-kube-api-access-45jql\") on node \"crc\" DevicePath \"\"" Oct 08 20:54:47 crc kubenswrapper[4988]: I1008 20:54:47.522152 4988 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ac5d09bc-5c99-4a7c-8632-ea0cfe1cf50a-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 08 20:54:47 crc kubenswrapper[4988]: I1008 20:54:47.734159 4988 generic.go:334] "Generic (PLEG): container finished" podID="ac5d09bc-5c99-4a7c-8632-ea0cfe1cf50a" containerID="a5be2570d20b4208bd1e8ed7adae2ee93898e90ba4dbc11caa4c920957b76ada" exitCode=0 Oct 08 20:54:47 crc kubenswrapper[4988]: I1008 20:54:47.734242 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-6dxkb" Oct 08 20:54:47 crc kubenswrapper[4988]: I1008 20:54:47.734250 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6dxkb" event={"ID":"ac5d09bc-5c99-4a7c-8632-ea0cfe1cf50a","Type":"ContainerDied","Data":"a5be2570d20b4208bd1e8ed7adae2ee93898e90ba4dbc11caa4c920957b76ada"} Oct 08 20:54:47 crc kubenswrapper[4988]: I1008 20:54:47.734496 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6dxkb" event={"ID":"ac5d09bc-5c99-4a7c-8632-ea0cfe1cf50a","Type":"ContainerDied","Data":"0a5a9bac2c67184d3bff7f2fb27a16ce8694a5e500ef5d461a22d97290a68ff7"} Oct 08 20:54:47 crc kubenswrapper[4988]: I1008 20:54:47.734541 4988 scope.go:117] "RemoveContainer" containerID="a5be2570d20b4208bd1e8ed7adae2ee93898e90ba4dbc11caa4c920957b76ada" Oct 08 20:54:47 crc kubenswrapper[4988]: I1008 20:54:47.787788 4988 scope.go:117] "RemoveContainer" containerID="eac8b358e6759c11532fc8d88c98f54da96b89922db005599d3b23ec3824fc57" Oct 08 20:54:47 crc kubenswrapper[4988]: I1008 20:54:47.791963 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-6dxkb"] Oct 08 20:54:47 crc kubenswrapper[4988]: I1008 20:54:47.803590 4988 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-6dxkb"] Oct 08 20:54:47 crc kubenswrapper[4988]: I1008 20:54:47.833646 4988 scope.go:117] "RemoveContainer" containerID="b5b117598a7d5bfcdf783dcc601bcf241c8e2ec744fad00723a6a10e506e380b" Oct 08 20:54:47 crc kubenswrapper[4988]: I1008 20:54:47.886237 4988 scope.go:117] "RemoveContainer" containerID="a5be2570d20b4208bd1e8ed7adae2ee93898e90ba4dbc11caa4c920957b76ada" Oct 08 20:54:47 crc kubenswrapper[4988]: E1008 20:54:47.886886 4988 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a5be2570d20b4208bd1e8ed7adae2ee93898e90ba4dbc11caa4c920957b76ada\": container with ID starting with a5be2570d20b4208bd1e8ed7adae2ee93898e90ba4dbc11caa4c920957b76ada not found: ID does not exist" containerID="a5be2570d20b4208bd1e8ed7adae2ee93898e90ba4dbc11caa4c920957b76ada" Oct 08 20:54:47 crc kubenswrapper[4988]: I1008 20:54:47.886957 4988 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a5be2570d20b4208bd1e8ed7adae2ee93898e90ba4dbc11caa4c920957b76ada"} err="failed to get container status \"a5be2570d20b4208bd1e8ed7adae2ee93898e90ba4dbc11caa4c920957b76ada\": rpc error: code = NotFound desc = could not find container \"a5be2570d20b4208bd1e8ed7adae2ee93898e90ba4dbc11caa4c920957b76ada\": container with ID starting with a5be2570d20b4208bd1e8ed7adae2ee93898e90ba4dbc11caa4c920957b76ada not found: ID does not exist" Oct 08 20:54:47 crc kubenswrapper[4988]: I1008 20:54:47.886984 4988 scope.go:117] "RemoveContainer" containerID="eac8b358e6759c11532fc8d88c98f54da96b89922db005599d3b23ec3824fc57" Oct 08 20:54:47 crc kubenswrapper[4988]: E1008 20:54:47.887519 4988 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"eac8b358e6759c11532fc8d88c98f54da96b89922db005599d3b23ec3824fc57\": container with ID starting with eac8b358e6759c11532fc8d88c98f54da96b89922db005599d3b23ec3824fc57 not found: ID does not exist" containerID="eac8b358e6759c11532fc8d88c98f54da96b89922db005599d3b23ec3824fc57" Oct 08 20:54:47 crc kubenswrapper[4988]: I1008 20:54:47.887567 4988 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"eac8b358e6759c11532fc8d88c98f54da96b89922db005599d3b23ec3824fc57"} err="failed to get container status \"eac8b358e6759c11532fc8d88c98f54da96b89922db005599d3b23ec3824fc57\": rpc error: code = NotFound desc = could not find container \"eac8b358e6759c11532fc8d88c98f54da96b89922db005599d3b23ec3824fc57\": container with ID starting with eac8b358e6759c11532fc8d88c98f54da96b89922db005599d3b23ec3824fc57 not found: ID does not exist" Oct 08 20:54:47 crc kubenswrapper[4988]: I1008 20:54:47.887598 4988 scope.go:117] "RemoveContainer" containerID="b5b117598a7d5bfcdf783dcc601bcf241c8e2ec744fad00723a6a10e506e380b" Oct 08 20:54:47 crc kubenswrapper[4988]: E1008 20:54:47.889128 4988 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b5b117598a7d5bfcdf783dcc601bcf241c8e2ec744fad00723a6a10e506e380b\": container with ID starting with b5b117598a7d5bfcdf783dcc601bcf241c8e2ec744fad00723a6a10e506e380b not found: ID does not exist" containerID="b5b117598a7d5bfcdf783dcc601bcf241c8e2ec744fad00723a6a10e506e380b" Oct 08 20:54:47 crc kubenswrapper[4988]: I1008 20:54:47.889174 4988 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b5b117598a7d5bfcdf783dcc601bcf241c8e2ec744fad00723a6a10e506e380b"} err="failed to get container status \"b5b117598a7d5bfcdf783dcc601bcf241c8e2ec744fad00723a6a10e506e380b\": rpc error: code = NotFound desc = could not find container \"b5b117598a7d5bfcdf783dcc601bcf241c8e2ec744fad00723a6a10e506e380b\": container with ID starting with b5b117598a7d5bfcdf783dcc601bcf241c8e2ec744fad00723a6a10e506e380b not found: ID does not exist" Oct 08 20:54:49 crc kubenswrapper[4988]: I1008 20:54:49.252353 4988 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ac5d09bc-5c99-4a7c-8632-ea0cfe1cf50a" path="/var/lib/kubelet/pods/ac5d09bc-5c99-4a7c-8632-ea0cfe1cf50a/volumes" Oct 08 20:54:53 crc kubenswrapper[4988]: I1008 20:54:53.338078 4988 patch_prober.go:28] interesting pod/machine-config-daemon-nm54f container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 20:54:53 crc kubenswrapper[4988]: I1008 20:54:53.338636 4988 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 20:55:23 crc kubenswrapper[4988]: I1008 20:55:23.338261 4988 patch_prober.go:28] interesting pod/machine-config-daemon-nm54f container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 20:55:23 crc kubenswrapper[4988]: I1008 20:55:23.339092 4988 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 20:55:23 crc kubenswrapper[4988]: I1008 20:55:23.339177 4988 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" Oct 08 20:55:23 crc kubenswrapper[4988]: I1008 20:55:23.340580 4988 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"fc5906dffdd6ca78fe0b71f3079a04d7d08624d242c87fd146529c035ee33bca"} pod="openshift-machine-config-operator/machine-config-daemon-nm54f" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 08 20:55:23 crc kubenswrapper[4988]: I1008 20:55:23.340714 4988 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" containerName="machine-config-daemon" containerID="cri-o://fc5906dffdd6ca78fe0b71f3079a04d7d08624d242c87fd146529c035ee33bca" gracePeriod=600 Oct 08 20:55:23 crc kubenswrapper[4988]: E1008 20:55:23.523103 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nm54f_openshift-machine-config-operator(f328a20b-5283-407e-8a93-2ab2da599f18)\"" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" Oct 08 20:55:24 crc kubenswrapper[4988]: I1008 20:55:24.155745 4988 generic.go:334] "Generic (PLEG): container finished" podID="f328a20b-5283-407e-8a93-2ab2da599f18" containerID="fc5906dffdd6ca78fe0b71f3079a04d7d08624d242c87fd146529c035ee33bca" exitCode=0 Oct 08 20:55:24 crc kubenswrapper[4988]: I1008 20:55:24.155793 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" event={"ID":"f328a20b-5283-407e-8a93-2ab2da599f18","Type":"ContainerDied","Data":"fc5906dffdd6ca78fe0b71f3079a04d7d08624d242c87fd146529c035ee33bca"} Oct 08 20:55:24 crc kubenswrapper[4988]: I1008 20:55:24.155824 4988 scope.go:117] "RemoveContainer" containerID="3a1d870522f1f317cf2b669f777c577892b10a5614c2e13e7030b149ce77fbb8" Oct 08 20:55:24 crc kubenswrapper[4988]: I1008 20:55:24.156506 4988 scope.go:117] "RemoveContainer" containerID="fc5906dffdd6ca78fe0b71f3079a04d7d08624d242c87fd146529c035ee33bca" Oct 08 20:55:24 crc kubenswrapper[4988]: E1008 20:55:24.156815 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nm54f_openshift-machine-config-operator(f328a20b-5283-407e-8a93-2ab2da599f18)\"" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" Oct 08 20:55:39 crc kubenswrapper[4988]: I1008 20:55:39.239080 4988 scope.go:117] "RemoveContainer" containerID="fc5906dffdd6ca78fe0b71f3079a04d7d08624d242c87fd146529c035ee33bca" Oct 08 20:55:39 crc kubenswrapper[4988]: E1008 20:55:39.240167 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nm54f_openshift-machine-config-operator(f328a20b-5283-407e-8a93-2ab2da599f18)\"" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" Oct 08 20:55:49 crc kubenswrapper[4988]: I1008 20:55:49.198828 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-wndj7"] Oct 08 20:55:49 crc kubenswrapper[4988]: E1008 20:55:49.200415 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ac5d09bc-5c99-4a7c-8632-ea0cfe1cf50a" containerName="registry-server" Oct 08 20:55:49 crc kubenswrapper[4988]: I1008 20:55:49.200435 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="ac5d09bc-5c99-4a7c-8632-ea0cfe1cf50a" containerName="registry-server" Oct 08 20:55:49 crc kubenswrapper[4988]: E1008 20:55:49.200456 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ac5d09bc-5c99-4a7c-8632-ea0cfe1cf50a" containerName="extract-content" Oct 08 20:55:49 crc kubenswrapper[4988]: I1008 20:55:49.200463 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="ac5d09bc-5c99-4a7c-8632-ea0cfe1cf50a" containerName="extract-content" Oct 08 20:55:49 crc kubenswrapper[4988]: E1008 20:55:49.200516 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ac5d09bc-5c99-4a7c-8632-ea0cfe1cf50a" containerName="extract-utilities" Oct 08 20:55:49 crc kubenswrapper[4988]: I1008 20:55:49.200527 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="ac5d09bc-5c99-4a7c-8632-ea0cfe1cf50a" containerName="extract-utilities" Oct 08 20:55:49 crc kubenswrapper[4988]: I1008 20:55:49.200977 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="ac5d09bc-5c99-4a7c-8632-ea0cfe1cf50a" containerName="registry-server" Oct 08 20:55:49 crc kubenswrapper[4988]: I1008 20:55:49.205311 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-wndj7" Oct 08 20:55:49 crc kubenswrapper[4988]: I1008 20:55:49.226183 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-wndj7"] Oct 08 20:55:49 crc kubenswrapper[4988]: I1008 20:55:49.335876 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2561bb39-0015-43fa-a91a-149ed08b2807-catalog-content\") pod \"redhat-marketplace-wndj7\" (UID: \"2561bb39-0015-43fa-a91a-149ed08b2807\") " pod="openshift-marketplace/redhat-marketplace-wndj7" Oct 08 20:55:49 crc kubenswrapper[4988]: I1008 20:55:49.336241 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sm9sc\" (UniqueName: \"kubernetes.io/projected/2561bb39-0015-43fa-a91a-149ed08b2807-kube-api-access-sm9sc\") pod \"redhat-marketplace-wndj7\" (UID: \"2561bb39-0015-43fa-a91a-149ed08b2807\") " pod="openshift-marketplace/redhat-marketplace-wndj7" Oct 08 20:55:49 crc kubenswrapper[4988]: I1008 20:55:49.336286 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2561bb39-0015-43fa-a91a-149ed08b2807-utilities\") pod \"redhat-marketplace-wndj7\" (UID: \"2561bb39-0015-43fa-a91a-149ed08b2807\") " pod="openshift-marketplace/redhat-marketplace-wndj7" Oct 08 20:55:49 crc kubenswrapper[4988]: I1008 20:55:49.438077 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sm9sc\" (UniqueName: \"kubernetes.io/projected/2561bb39-0015-43fa-a91a-149ed08b2807-kube-api-access-sm9sc\") pod \"redhat-marketplace-wndj7\" (UID: \"2561bb39-0015-43fa-a91a-149ed08b2807\") " pod="openshift-marketplace/redhat-marketplace-wndj7" Oct 08 20:55:49 crc kubenswrapper[4988]: I1008 20:55:49.438132 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2561bb39-0015-43fa-a91a-149ed08b2807-utilities\") pod \"redhat-marketplace-wndj7\" (UID: \"2561bb39-0015-43fa-a91a-149ed08b2807\") " pod="openshift-marketplace/redhat-marketplace-wndj7" Oct 08 20:55:49 crc kubenswrapper[4988]: I1008 20:55:49.438276 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2561bb39-0015-43fa-a91a-149ed08b2807-catalog-content\") pod \"redhat-marketplace-wndj7\" (UID: \"2561bb39-0015-43fa-a91a-149ed08b2807\") " pod="openshift-marketplace/redhat-marketplace-wndj7" Oct 08 20:55:49 crc kubenswrapper[4988]: I1008 20:55:49.438749 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2561bb39-0015-43fa-a91a-149ed08b2807-catalog-content\") pod \"redhat-marketplace-wndj7\" (UID: \"2561bb39-0015-43fa-a91a-149ed08b2807\") " pod="openshift-marketplace/redhat-marketplace-wndj7" Oct 08 20:55:49 crc kubenswrapper[4988]: I1008 20:55:49.439050 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2561bb39-0015-43fa-a91a-149ed08b2807-utilities\") pod \"redhat-marketplace-wndj7\" (UID: \"2561bb39-0015-43fa-a91a-149ed08b2807\") " pod="openshift-marketplace/redhat-marketplace-wndj7" Oct 08 20:55:49 crc kubenswrapper[4988]: I1008 20:55:49.460638 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sm9sc\" (UniqueName: \"kubernetes.io/projected/2561bb39-0015-43fa-a91a-149ed08b2807-kube-api-access-sm9sc\") pod \"redhat-marketplace-wndj7\" (UID: \"2561bb39-0015-43fa-a91a-149ed08b2807\") " pod="openshift-marketplace/redhat-marketplace-wndj7" Oct 08 20:55:49 crc kubenswrapper[4988]: I1008 20:55:49.540957 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-wndj7" Oct 08 20:55:50 crc kubenswrapper[4988]: I1008 20:55:50.054817 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-wndj7"] Oct 08 20:55:50 crc kubenswrapper[4988]: I1008 20:55:50.495230 4988 generic.go:334] "Generic (PLEG): container finished" podID="2561bb39-0015-43fa-a91a-149ed08b2807" containerID="f83cde596f4f84259ce24d6c42c20eeb0d38ea61c9f1a9fb87db505edb37468a" exitCode=0 Oct 08 20:55:50 crc kubenswrapper[4988]: I1008 20:55:50.495297 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-wndj7" event={"ID":"2561bb39-0015-43fa-a91a-149ed08b2807","Type":"ContainerDied","Data":"f83cde596f4f84259ce24d6c42c20eeb0d38ea61c9f1a9fb87db505edb37468a"} Oct 08 20:55:50 crc kubenswrapper[4988]: I1008 20:55:50.495643 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-wndj7" event={"ID":"2561bb39-0015-43fa-a91a-149ed08b2807","Type":"ContainerStarted","Data":"b590da4b6bb63412dfb9d0f746ca1f5024bd5f49d82bd1b9fea0ac9f59800803"} Oct 08 20:55:51 crc kubenswrapper[4988]: I1008 20:55:51.237701 4988 scope.go:117] "RemoveContainer" containerID="fc5906dffdd6ca78fe0b71f3079a04d7d08624d242c87fd146529c035ee33bca" Oct 08 20:55:51 crc kubenswrapper[4988]: E1008 20:55:51.238263 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nm54f_openshift-machine-config-operator(f328a20b-5283-407e-8a93-2ab2da599f18)\"" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" Oct 08 20:55:52 crc kubenswrapper[4988]: I1008 20:55:52.535649 4988 generic.go:334] "Generic (PLEG): container finished" podID="2561bb39-0015-43fa-a91a-149ed08b2807" containerID="1347967414ff4bcfe5f6966d0e19611b9f1920c09a3b91e1acc8269c923a288d" exitCode=0 Oct 08 20:55:52 crc kubenswrapper[4988]: I1008 20:55:52.535727 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-wndj7" event={"ID":"2561bb39-0015-43fa-a91a-149ed08b2807","Type":"ContainerDied","Data":"1347967414ff4bcfe5f6966d0e19611b9f1920c09a3b91e1acc8269c923a288d"} Oct 08 20:55:53 crc kubenswrapper[4988]: I1008 20:55:53.551878 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-wndj7" event={"ID":"2561bb39-0015-43fa-a91a-149ed08b2807","Type":"ContainerStarted","Data":"f3c6900770e86d49b45987f69d8f67c5c7570ffd91cb642e0352029c32735448"} Oct 08 20:55:53 crc kubenswrapper[4988]: I1008 20:55:53.578733 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-wndj7" podStartSLOduration=2.057761712 podStartE2EDuration="4.578717797s" podCreationTimestamp="2025-10-08 20:55:49 +0000 UTC" firstStartedPulling="2025-10-08 20:55:50.498030724 +0000 UTC m=+9895.947873534" lastFinishedPulling="2025-10-08 20:55:53.018986849 +0000 UTC m=+9898.468829619" observedRunningTime="2025-10-08 20:55:53.574280194 +0000 UTC m=+9899.024122974" watchObservedRunningTime="2025-10-08 20:55:53.578717797 +0000 UTC m=+9899.028560567" Oct 08 20:55:59 crc kubenswrapper[4988]: I1008 20:55:59.546615 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-wndj7" Oct 08 20:55:59 crc kubenswrapper[4988]: I1008 20:55:59.547274 4988 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-wndj7" Oct 08 20:55:59 crc kubenswrapper[4988]: I1008 20:55:59.629696 4988 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-wndj7" Oct 08 20:55:59 crc kubenswrapper[4988]: I1008 20:55:59.703896 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-wndj7" Oct 08 20:55:59 crc kubenswrapper[4988]: I1008 20:55:59.887580 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-wndj7"] Oct 08 20:56:01 crc kubenswrapper[4988]: I1008 20:56:01.668034 4988 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-wndj7" podUID="2561bb39-0015-43fa-a91a-149ed08b2807" containerName="registry-server" containerID="cri-o://f3c6900770e86d49b45987f69d8f67c5c7570ffd91cb642e0352029c32735448" gracePeriod=2 Oct 08 20:56:02 crc kubenswrapper[4988]: I1008 20:56:02.202975 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-wndj7" Oct 08 20:56:02 crc kubenswrapper[4988]: I1008 20:56:02.261324 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2561bb39-0015-43fa-a91a-149ed08b2807-utilities\") pod \"2561bb39-0015-43fa-a91a-149ed08b2807\" (UID: \"2561bb39-0015-43fa-a91a-149ed08b2807\") " Oct 08 20:56:02 crc kubenswrapper[4988]: I1008 20:56:02.265663 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sm9sc\" (UniqueName: \"kubernetes.io/projected/2561bb39-0015-43fa-a91a-149ed08b2807-kube-api-access-sm9sc\") pod \"2561bb39-0015-43fa-a91a-149ed08b2807\" (UID: \"2561bb39-0015-43fa-a91a-149ed08b2807\") " Oct 08 20:56:02 crc kubenswrapper[4988]: I1008 20:56:02.266028 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2561bb39-0015-43fa-a91a-149ed08b2807-catalog-content\") pod \"2561bb39-0015-43fa-a91a-149ed08b2807\" (UID: \"2561bb39-0015-43fa-a91a-149ed08b2807\") " Oct 08 20:56:02 crc kubenswrapper[4988]: I1008 20:56:02.269365 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2561bb39-0015-43fa-a91a-149ed08b2807-utilities" (OuterVolumeSpecName: "utilities") pod "2561bb39-0015-43fa-a91a-149ed08b2807" (UID: "2561bb39-0015-43fa-a91a-149ed08b2807"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 20:56:02 crc kubenswrapper[4988]: I1008 20:56:02.270420 4988 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2561bb39-0015-43fa-a91a-149ed08b2807-utilities\") on node \"crc\" DevicePath \"\"" Oct 08 20:56:02 crc kubenswrapper[4988]: I1008 20:56:02.278932 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2561bb39-0015-43fa-a91a-149ed08b2807-kube-api-access-sm9sc" (OuterVolumeSpecName: "kube-api-access-sm9sc") pod "2561bb39-0015-43fa-a91a-149ed08b2807" (UID: "2561bb39-0015-43fa-a91a-149ed08b2807"). InnerVolumeSpecName "kube-api-access-sm9sc". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 20:56:02 crc kubenswrapper[4988]: I1008 20:56:02.285974 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2561bb39-0015-43fa-a91a-149ed08b2807-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "2561bb39-0015-43fa-a91a-149ed08b2807" (UID: "2561bb39-0015-43fa-a91a-149ed08b2807"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 20:56:02 crc kubenswrapper[4988]: I1008 20:56:02.372749 4988 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2561bb39-0015-43fa-a91a-149ed08b2807-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 08 20:56:02 crc kubenswrapper[4988]: I1008 20:56:02.372815 4988 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sm9sc\" (UniqueName: \"kubernetes.io/projected/2561bb39-0015-43fa-a91a-149ed08b2807-kube-api-access-sm9sc\") on node \"crc\" DevicePath \"\"" Oct 08 20:56:02 crc kubenswrapper[4988]: I1008 20:56:02.686278 4988 generic.go:334] "Generic (PLEG): container finished" podID="2561bb39-0015-43fa-a91a-149ed08b2807" containerID="f3c6900770e86d49b45987f69d8f67c5c7570ffd91cb642e0352029c32735448" exitCode=0 Oct 08 20:56:02 crc kubenswrapper[4988]: I1008 20:56:02.686330 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-wndj7" event={"ID":"2561bb39-0015-43fa-a91a-149ed08b2807","Type":"ContainerDied","Data":"f3c6900770e86d49b45987f69d8f67c5c7570ffd91cb642e0352029c32735448"} Oct 08 20:56:02 crc kubenswrapper[4988]: I1008 20:56:02.686369 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-wndj7" event={"ID":"2561bb39-0015-43fa-a91a-149ed08b2807","Type":"ContainerDied","Data":"b590da4b6bb63412dfb9d0f746ca1f5024bd5f49d82bd1b9fea0ac9f59800803"} Oct 08 20:56:02 crc kubenswrapper[4988]: I1008 20:56:02.686441 4988 scope.go:117] "RemoveContainer" containerID="f3c6900770e86d49b45987f69d8f67c5c7570ffd91cb642e0352029c32735448" Oct 08 20:56:02 crc kubenswrapper[4988]: I1008 20:56:02.686519 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-wndj7" Oct 08 20:56:02 crc kubenswrapper[4988]: I1008 20:56:02.717113 4988 scope.go:117] "RemoveContainer" containerID="1347967414ff4bcfe5f6966d0e19611b9f1920c09a3b91e1acc8269c923a288d" Oct 08 20:56:02 crc kubenswrapper[4988]: I1008 20:56:02.753142 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-wndj7"] Oct 08 20:56:02 crc kubenswrapper[4988]: I1008 20:56:02.764474 4988 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-wndj7"] Oct 08 20:56:02 crc kubenswrapper[4988]: I1008 20:56:02.782102 4988 scope.go:117] "RemoveContainer" containerID="f83cde596f4f84259ce24d6c42c20eeb0d38ea61c9f1a9fb87db505edb37468a" Oct 08 20:56:02 crc kubenswrapper[4988]: I1008 20:56:02.840506 4988 scope.go:117] "RemoveContainer" containerID="f3c6900770e86d49b45987f69d8f67c5c7570ffd91cb642e0352029c32735448" Oct 08 20:56:02 crc kubenswrapper[4988]: E1008 20:56:02.841036 4988 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f3c6900770e86d49b45987f69d8f67c5c7570ffd91cb642e0352029c32735448\": container with ID starting with f3c6900770e86d49b45987f69d8f67c5c7570ffd91cb642e0352029c32735448 not found: ID does not exist" containerID="f3c6900770e86d49b45987f69d8f67c5c7570ffd91cb642e0352029c32735448" Oct 08 20:56:02 crc kubenswrapper[4988]: I1008 20:56:02.841075 4988 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f3c6900770e86d49b45987f69d8f67c5c7570ffd91cb642e0352029c32735448"} err="failed to get container status \"f3c6900770e86d49b45987f69d8f67c5c7570ffd91cb642e0352029c32735448\": rpc error: code = NotFound desc = could not find container \"f3c6900770e86d49b45987f69d8f67c5c7570ffd91cb642e0352029c32735448\": container with ID starting with f3c6900770e86d49b45987f69d8f67c5c7570ffd91cb642e0352029c32735448 not found: ID does not exist" Oct 08 20:56:02 crc kubenswrapper[4988]: I1008 20:56:02.841099 4988 scope.go:117] "RemoveContainer" containerID="1347967414ff4bcfe5f6966d0e19611b9f1920c09a3b91e1acc8269c923a288d" Oct 08 20:56:02 crc kubenswrapper[4988]: E1008 20:56:02.843064 4988 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1347967414ff4bcfe5f6966d0e19611b9f1920c09a3b91e1acc8269c923a288d\": container with ID starting with 1347967414ff4bcfe5f6966d0e19611b9f1920c09a3b91e1acc8269c923a288d not found: ID does not exist" containerID="1347967414ff4bcfe5f6966d0e19611b9f1920c09a3b91e1acc8269c923a288d" Oct 08 20:56:02 crc kubenswrapper[4988]: I1008 20:56:02.843093 4988 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1347967414ff4bcfe5f6966d0e19611b9f1920c09a3b91e1acc8269c923a288d"} err="failed to get container status \"1347967414ff4bcfe5f6966d0e19611b9f1920c09a3b91e1acc8269c923a288d\": rpc error: code = NotFound desc = could not find container \"1347967414ff4bcfe5f6966d0e19611b9f1920c09a3b91e1acc8269c923a288d\": container with ID starting with 1347967414ff4bcfe5f6966d0e19611b9f1920c09a3b91e1acc8269c923a288d not found: ID does not exist" Oct 08 20:56:02 crc kubenswrapper[4988]: I1008 20:56:02.843112 4988 scope.go:117] "RemoveContainer" containerID="f83cde596f4f84259ce24d6c42c20eeb0d38ea61c9f1a9fb87db505edb37468a" Oct 08 20:56:02 crc kubenswrapper[4988]: E1008 20:56:02.844223 4988 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f83cde596f4f84259ce24d6c42c20eeb0d38ea61c9f1a9fb87db505edb37468a\": container with ID starting with f83cde596f4f84259ce24d6c42c20eeb0d38ea61c9f1a9fb87db505edb37468a not found: ID does not exist" containerID="f83cde596f4f84259ce24d6c42c20eeb0d38ea61c9f1a9fb87db505edb37468a" Oct 08 20:56:02 crc kubenswrapper[4988]: I1008 20:56:02.844257 4988 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f83cde596f4f84259ce24d6c42c20eeb0d38ea61c9f1a9fb87db505edb37468a"} err="failed to get container status \"f83cde596f4f84259ce24d6c42c20eeb0d38ea61c9f1a9fb87db505edb37468a\": rpc error: code = NotFound desc = could not find container \"f83cde596f4f84259ce24d6c42c20eeb0d38ea61c9f1a9fb87db505edb37468a\": container with ID starting with f83cde596f4f84259ce24d6c42c20eeb0d38ea61c9f1a9fb87db505edb37468a not found: ID does not exist" Oct 08 20:56:03 crc kubenswrapper[4988]: I1008 20:56:03.238351 4988 scope.go:117] "RemoveContainer" containerID="fc5906dffdd6ca78fe0b71f3079a04d7d08624d242c87fd146529c035ee33bca" Oct 08 20:56:03 crc kubenswrapper[4988]: E1008 20:56:03.238911 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nm54f_openshift-machine-config-operator(f328a20b-5283-407e-8a93-2ab2da599f18)\"" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" Oct 08 20:56:03 crc kubenswrapper[4988]: I1008 20:56:03.259695 4988 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2561bb39-0015-43fa-a91a-149ed08b2807" path="/var/lib/kubelet/pods/2561bb39-0015-43fa-a91a-149ed08b2807/volumes" Oct 08 20:56:15 crc kubenswrapper[4988]: I1008 20:56:15.238586 4988 scope.go:117] "RemoveContainer" containerID="fc5906dffdd6ca78fe0b71f3079a04d7d08624d242c87fd146529c035ee33bca" Oct 08 20:56:15 crc kubenswrapper[4988]: E1008 20:56:15.239570 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nm54f_openshift-machine-config-operator(f328a20b-5283-407e-8a93-2ab2da599f18)\"" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" Oct 08 20:56:29 crc kubenswrapper[4988]: I1008 20:56:29.238028 4988 scope.go:117] "RemoveContainer" containerID="fc5906dffdd6ca78fe0b71f3079a04d7d08624d242c87fd146529c035ee33bca" Oct 08 20:56:29 crc kubenswrapper[4988]: E1008 20:56:29.238841 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nm54f_openshift-machine-config-operator(f328a20b-5283-407e-8a93-2ab2da599f18)\"" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" Oct 08 20:56:44 crc kubenswrapper[4988]: I1008 20:56:44.238326 4988 scope.go:117] "RemoveContainer" containerID="fc5906dffdd6ca78fe0b71f3079a04d7d08624d242c87fd146529c035ee33bca" Oct 08 20:56:44 crc kubenswrapper[4988]: E1008 20:56:44.239221 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nm54f_openshift-machine-config-operator(f328a20b-5283-407e-8a93-2ab2da599f18)\"" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" Oct 08 20:56:58 crc kubenswrapper[4988]: I1008 20:56:58.238723 4988 scope.go:117] "RemoveContainer" containerID="fc5906dffdd6ca78fe0b71f3079a04d7d08624d242c87fd146529c035ee33bca" Oct 08 20:56:58 crc kubenswrapper[4988]: E1008 20:56:58.239476 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nm54f_openshift-machine-config-operator(f328a20b-5283-407e-8a93-2ab2da599f18)\"" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" Oct 08 20:57:10 crc kubenswrapper[4988]: I1008 20:57:10.238158 4988 scope.go:117] "RemoveContainer" containerID="fc5906dffdd6ca78fe0b71f3079a04d7d08624d242c87fd146529c035ee33bca" Oct 08 20:57:10 crc kubenswrapper[4988]: E1008 20:57:10.239043 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nm54f_openshift-machine-config-operator(f328a20b-5283-407e-8a93-2ab2da599f18)\"" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" Oct 08 20:57:22 crc kubenswrapper[4988]: I1008 20:57:22.238437 4988 scope.go:117] "RemoveContainer" containerID="fc5906dffdd6ca78fe0b71f3079a04d7d08624d242c87fd146529c035ee33bca" Oct 08 20:57:22 crc kubenswrapper[4988]: E1008 20:57:22.240206 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nm54f_openshift-machine-config-operator(f328a20b-5283-407e-8a93-2ab2da599f18)\"" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" Oct 08 20:57:37 crc kubenswrapper[4988]: I1008 20:57:37.238521 4988 scope.go:117] "RemoveContainer" containerID="fc5906dffdd6ca78fe0b71f3079a04d7d08624d242c87fd146529c035ee33bca" Oct 08 20:57:37 crc kubenswrapper[4988]: E1008 20:57:37.239784 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nm54f_openshift-machine-config-operator(f328a20b-5283-407e-8a93-2ab2da599f18)\"" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" Oct 08 20:57:42 crc kubenswrapper[4988]: I1008 20:57:42.997147 4988 generic.go:334] "Generic (PLEG): container finished" podID="a2f34e97-3b21-419b-883b-bdd5108b7dae" containerID="886f78c41ab3b0c677fc4cb14d4cec225016b181fd83dbe7cf8adfdf2dac37f8" exitCode=0 Oct 08 20:57:42 crc kubenswrapper[4988]: I1008 20:57:42.997253 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellgcg5p" event={"ID":"a2f34e97-3b21-419b-883b-bdd5108b7dae","Type":"ContainerDied","Data":"886f78c41ab3b0c677fc4cb14d4cec225016b181fd83dbe7cf8adfdf2dac37f8"} Oct 08 20:57:45 crc kubenswrapper[4988]: I1008 20:57:45.424622 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellgcg5p" Oct 08 20:57:45 crc kubenswrapper[4988]: I1008 20:57:45.608976 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/a2f34e97-3b21-419b-883b-bdd5108b7dae-nova-cell1-compute-config-1\") pod \"a2f34e97-3b21-419b-883b-bdd5108b7dae\" (UID: \"a2f34e97-3b21-419b-883b-bdd5108b7dae\") " Oct 08 20:57:45 crc kubenswrapper[4988]: I1008 20:57:45.609223 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/a2f34e97-3b21-419b-883b-bdd5108b7dae-nova-migration-ssh-key-1\") pod \"a2f34e97-3b21-419b-883b-bdd5108b7dae\" (UID: \"a2f34e97-3b21-419b-883b-bdd5108b7dae\") " Oct 08 20:57:45 crc kubenswrapper[4988]: I1008 20:57:45.609417 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/a2f34e97-3b21-419b-883b-bdd5108b7dae-nova-migration-ssh-key-0\") pod \"a2f34e97-3b21-419b-883b-bdd5108b7dae\" (UID: \"a2f34e97-3b21-419b-883b-bdd5108b7dae\") " Oct 08 20:57:45 crc kubenswrapper[4988]: I1008 20:57:45.609541 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/a2f34e97-3b21-419b-883b-bdd5108b7dae-nova-cell1-compute-config-0\") pod \"a2f34e97-3b21-419b-883b-bdd5108b7dae\" (UID: \"a2f34e97-3b21-419b-883b-bdd5108b7dae\") " Oct 08 20:57:45 crc kubenswrapper[4988]: I1008 20:57:45.610559 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a2f34e97-3b21-419b-883b-bdd5108b7dae-ssh-key\") pod \"a2f34e97-3b21-419b-883b-bdd5108b7dae\" (UID: \"a2f34e97-3b21-419b-883b-bdd5108b7dae\") " Oct 08 20:57:45 crc kubenswrapper[4988]: I1008 20:57:45.610657 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a2f34e97-3b21-419b-883b-bdd5108b7dae-nova-cell1-combined-ca-bundle\") pod \"a2f34e97-3b21-419b-883b-bdd5108b7dae\" (UID: \"a2f34e97-3b21-419b-883b-bdd5108b7dae\") " Oct 08 20:57:45 crc kubenswrapper[4988]: I1008 20:57:45.610689 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a2f34e97-3b21-419b-883b-bdd5108b7dae-inventory\") pod \"a2f34e97-3b21-419b-883b-bdd5108b7dae\" (UID: \"a2f34e97-3b21-419b-883b-bdd5108b7dae\") " Oct 08 20:57:45 crc kubenswrapper[4988]: I1008 20:57:45.610751 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cells-global-config-0\" (UniqueName: \"kubernetes.io/configmap/a2f34e97-3b21-419b-883b-bdd5108b7dae-nova-cells-global-config-0\") pod \"a2f34e97-3b21-419b-883b-bdd5108b7dae\" (UID: \"a2f34e97-3b21-419b-883b-bdd5108b7dae\") " Oct 08 20:57:45 crc kubenswrapper[4988]: I1008 20:57:45.610782 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rfs7g\" (UniqueName: \"kubernetes.io/projected/a2f34e97-3b21-419b-883b-bdd5108b7dae-kube-api-access-rfs7g\") pod \"a2f34e97-3b21-419b-883b-bdd5108b7dae\" (UID: \"a2f34e97-3b21-419b-883b-bdd5108b7dae\") " Oct 08 20:57:45 crc kubenswrapper[4988]: I1008 20:57:45.630025 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a2f34e97-3b21-419b-883b-bdd5108b7dae-nova-cell1-combined-ca-bundle" (OuterVolumeSpecName: "nova-cell1-combined-ca-bundle") pod "a2f34e97-3b21-419b-883b-bdd5108b7dae" (UID: "a2f34e97-3b21-419b-883b-bdd5108b7dae"). InnerVolumeSpecName "nova-cell1-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 20:57:45 crc kubenswrapper[4988]: I1008 20:57:45.655658 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a2f34e97-3b21-419b-883b-bdd5108b7dae-kube-api-access-rfs7g" (OuterVolumeSpecName: "kube-api-access-rfs7g") pod "a2f34e97-3b21-419b-883b-bdd5108b7dae" (UID: "a2f34e97-3b21-419b-883b-bdd5108b7dae"). InnerVolumeSpecName "kube-api-access-rfs7g". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 20:57:45 crc kubenswrapper[4988]: I1008 20:57:45.679787 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a2f34e97-3b21-419b-883b-bdd5108b7dae-nova-cell1-compute-config-1" (OuterVolumeSpecName: "nova-cell1-compute-config-1") pod "a2f34e97-3b21-419b-883b-bdd5108b7dae" (UID: "a2f34e97-3b21-419b-883b-bdd5108b7dae"). InnerVolumeSpecName "nova-cell1-compute-config-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 20:57:45 crc kubenswrapper[4988]: I1008 20:57:45.703541 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a2f34e97-3b21-419b-883b-bdd5108b7dae-nova-cell1-compute-config-0" (OuterVolumeSpecName: "nova-cell1-compute-config-0") pod "a2f34e97-3b21-419b-883b-bdd5108b7dae" (UID: "a2f34e97-3b21-419b-883b-bdd5108b7dae"). InnerVolumeSpecName "nova-cell1-compute-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 20:57:45 crc kubenswrapper[4988]: I1008 20:57:45.704194 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a2f34e97-3b21-419b-883b-bdd5108b7dae-nova-cells-global-config-0" (OuterVolumeSpecName: "nova-cells-global-config-0") pod "a2f34e97-3b21-419b-883b-bdd5108b7dae" (UID: "a2f34e97-3b21-419b-883b-bdd5108b7dae"). InnerVolumeSpecName "nova-cells-global-config-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 20:57:45 crc kubenswrapper[4988]: I1008 20:57:45.744826 4988 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/a2f34e97-3b21-419b-883b-bdd5108b7dae-nova-cell1-compute-config-0\") on node \"crc\" DevicePath \"\"" Oct 08 20:57:45 crc kubenswrapper[4988]: I1008 20:57:45.745077 4988 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a2f34e97-3b21-419b-883b-bdd5108b7dae-nova-cell1-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 20:57:45 crc kubenswrapper[4988]: I1008 20:57:45.745148 4988 reconciler_common.go:293] "Volume detached for volume \"nova-cells-global-config-0\" (UniqueName: \"kubernetes.io/configmap/a2f34e97-3b21-419b-883b-bdd5108b7dae-nova-cells-global-config-0\") on node \"crc\" DevicePath \"\"" Oct 08 20:57:45 crc kubenswrapper[4988]: I1008 20:57:45.745209 4988 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rfs7g\" (UniqueName: \"kubernetes.io/projected/a2f34e97-3b21-419b-883b-bdd5108b7dae-kube-api-access-rfs7g\") on node \"crc\" DevicePath \"\"" Oct 08 20:57:45 crc kubenswrapper[4988]: I1008 20:57:45.745278 4988 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/a2f34e97-3b21-419b-883b-bdd5108b7dae-nova-cell1-compute-config-1\") on node \"crc\" DevicePath \"\"" Oct 08 20:57:45 crc kubenswrapper[4988]: I1008 20:57:45.758606 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a2f34e97-3b21-419b-883b-bdd5108b7dae-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "a2f34e97-3b21-419b-883b-bdd5108b7dae" (UID: "a2f34e97-3b21-419b-883b-bdd5108b7dae"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 20:57:45 crc kubenswrapper[4988]: I1008 20:57:45.783946 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a2f34e97-3b21-419b-883b-bdd5108b7dae-nova-migration-ssh-key-0" (OuterVolumeSpecName: "nova-migration-ssh-key-0") pod "a2f34e97-3b21-419b-883b-bdd5108b7dae" (UID: "a2f34e97-3b21-419b-883b-bdd5108b7dae"). InnerVolumeSpecName "nova-migration-ssh-key-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 20:57:45 crc kubenswrapper[4988]: I1008 20:57:45.830800 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a2f34e97-3b21-419b-883b-bdd5108b7dae-inventory" (OuterVolumeSpecName: "inventory") pod "a2f34e97-3b21-419b-883b-bdd5108b7dae" (UID: "a2f34e97-3b21-419b-883b-bdd5108b7dae"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 20:57:45 crc kubenswrapper[4988]: I1008 20:57:45.847913 4988 reconciler_common.go:293] "Volume detached for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/a2f34e97-3b21-419b-883b-bdd5108b7dae-nova-migration-ssh-key-0\") on node \"crc\" DevicePath \"\"" Oct 08 20:57:45 crc kubenswrapper[4988]: I1008 20:57:45.847950 4988 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a2f34e97-3b21-419b-883b-bdd5108b7dae-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 08 20:57:45 crc kubenswrapper[4988]: I1008 20:57:45.847960 4988 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a2f34e97-3b21-419b-883b-bdd5108b7dae-inventory\") on node \"crc\" DevicePath \"\"" Oct 08 20:57:45 crc kubenswrapper[4988]: I1008 20:57:45.849476 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a2f34e97-3b21-419b-883b-bdd5108b7dae-nova-migration-ssh-key-1" (OuterVolumeSpecName: "nova-migration-ssh-key-1") pod "a2f34e97-3b21-419b-883b-bdd5108b7dae" (UID: "a2f34e97-3b21-419b-883b-bdd5108b7dae"). InnerVolumeSpecName "nova-migration-ssh-key-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 20:57:45 crc kubenswrapper[4988]: I1008 20:57:45.949622 4988 reconciler_common.go:293] "Volume detached for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/a2f34e97-3b21-419b-883b-bdd5108b7dae-nova-migration-ssh-key-1\") on node \"crc\" DevicePath \"\"" Oct 08 20:57:46 crc kubenswrapper[4988]: I1008 20:57:46.041865 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellgcg5p" event={"ID":"a2f34e97-3b21-419b-883b-bdd5108b7dae","Type":"ContainerDied","Data":"c40d76919f87192452426e0318f1e15643e7a3c1c2dd04aaf0ce773fa961c2c5"} Oct 08 20:57:46 crc kubenswrapper[4988]: I1008 20:57:46.041920 4988 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c40d76919f87192452426e0318f1e15643e7a3c1c2dd04aaf0ce773fa961c2c5" Oct 08 20:57:46 crc kubenswrapper[4988]: I1008 20:57:46.042295 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellgcg5p" Oct 08 20:57:49 crc kubenswrapper[4988]: I1008 20:57:49.238146 4988 scope.go:117] "RemoveContainer" containerID="fc5906dffdd6ca78fe0b71f3079a04d7d08624d242c87fd146529c035ee33bca" Oct 08 20:57:49 crc kubenswrapper[4988]: E1008 20:57:49.238917 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nm54f_openshift-machine-config-operator(f328a20b-5283-407e-8a93-2ab2da599f18)\"" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" Oct 08 20:58:02 crc kubenswrapper[4988]: I1008 20:58:02.238704 4988 scope.go:117] "RemoveContainer" containerID="fc5906dffdd6ca78fe0b71f3079a04d7d08624d242c87fd146529c035ee33bca" Oct 08 20:58:02 crc kubenswrapper[4988]: E1008 20:58:02.239895 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nm54f_openshift-machine-config-operator(f328a20b-5283-407e-8a93-2ab2da599f18)\"" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" Oct 08 20:58:11 crc kubenswrapper[4988]: I1008 20:58:11.373954 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-lwfs7"] Oct 08 20:58:11 crc kubenswrapper[4988]: E1008 20:58:11.375269 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2561bb39-0015-43fa-a91a-149ed08b2807" containerName="extract-utilities" Oct 08 20:58:11 crc kubenswrapper[4988]: I1008 20:58:11.375294 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="2561bb39-0015-43fa-a91a-149ed08b2807" containerName="extract-utilities" Oct 08 20:58:11 crc kubenswrapper[4988]: E1008 20:58:11.375354 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a2f34e97-3b21-419b-883b-bdd5108b7dae" containerName="nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell1" Oct 08 20:58:11 crc kubenswrapper[4988]: I1008 20:58:11.375368 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="a2f34e97-3b21-419b-883b-bdd5108b7dae" containerName="nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell1" Oct 08 20:58:11 crc kubenswrapper[4988]: E1008 20:58:11.375417 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2561bb39-0015-43fa-a91a-149ed08b2807" containerName="registry-server" Oct 08 20:58:11 crc kubenswrapper[4988]: I1008 20:58:11.375431 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="2561bb39-0015-43fa-a91a-149ed08b2807" containerName="registry-server" Oct 08 20:58:11 crc kubenswrapper[4988]: E1008 20:58:11.375470 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2561bb39-0015-43fa-a91a-149ed08b2807" containerName="extract-content" Oct 08 20:58:11 crc kubenswrapper[4988]: I1008 20:58:11.375483 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="2561bb39-0015-43fa-a91a-149ed08b2807" containerName="extract-content" Oct 08 20:58:11 crc kubenswrapper[4988]: I1008 20:58:11.375876 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="a2f34e97-3b21-419b-883b-bdd5108b7dae" containerName="nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell1" Oct 08 20:58:11 crc kubenswrapper[4988]: I1008 20:58:11.375921 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="2561bb39-0015-43fa-a91a-149ed08b2807" containerName="registry-server" Oct 08 20:58:11 crc kubenswrapper[4988]: I1008 20:58:11.378884 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-lwfs7" Oct 08 20:58:11 crc kubenswrapper[4988]: I1008 20:58:11.402963 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-lwfs7"] Oct 08 20:58:11 crc kubenswrapper[4988]: I1008 20:58:11.415720 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-94lvk\" (UniqueName: \"kubernetes.io/projected/eb46f2ea-20fe-4f17-9af4-5f09120e868d-kube-api-access-94lvk\") pod \"redhat-operators-lwfs7\" (UID: \"eb46f2ea-20fe-4f17-9af4-5f09120e868d\") " pod="openshift-marketplace/redhat-operators-lwfs7" Oct 08 20:58:11 crc kubenswrapper[4988]: I1008 20:58:11.415770 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/eb46f2ea-20fe-4f17-9af4-5f09120e868d-utilities\") pod \"redhat-operators-lwfs7\" (UID: \"eb46f2ea-20fe-4f17-9af4-5f09120e868d\") " pod="openshift-marketplace/redhat-operators-lwfs7" Oct 08 20:58:11 crc kubenswrapper[4988]: I1008 20:58:11.415840 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/eb46f2ea-20fe-4f17-9af4-5f09120e868d-catalog-content\") pod \"redhat-operators-lwfs7\" (UID: \"eb46f2ea-20fe-4f17-9af4-5f09120e868d\") " pod="openshift-marketplace/redhat-operators-lwfs7" Oct 08 20:58:11 crc kubenswrapper[4988]: I1008 20:58:11.518180 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/eb46f2ea-20fe-4f17-9af4-5f09120e868d-catalog-content\") pod \"redhat-operators-lwfs7\" (UID: \"eb46f2ea-20fe-4f17-9af4-5f09120e868d\") " pod="openshift-marketplace/redhat-operators-lwfs7" Oct 08 20:58:11 crc kubenswrapper[4988]: I1008 20:58:11.518426 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-94lvk\" (UniqueName: \"kubernetes.io/projected/eb46f2ea-20fe-4f17-9af4-5f09120e868d-kube-api-access-94lvk\") pod \"redhat-operators-lwfs7\" (UID: \"eb46f2ea-20fe-4f17-9af4-5f09120e868d\") " pod="openshift-marketplace/redhat-operators-lwfs7" Oct 08 20:58:11 crc kubenswrapper[4988]: I1008 20:58:11.518460 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/eb46f2ea-20fe-4f17-9af4-5f09120e868d-utilities\") pod \"redhat-operators-lwfs7\" (UID: \"eb46f2ea-20fe-4f17-9af4-5f09120e868d\") " pod="openshift-marketplace/redhat-operators-lwfs7" Oct 08 20:58:11 crc kubenswrapper[4988]: I1008 20:58:11.519148 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/eb46f2ea-20fe-4f17-9af4-5f09120e868d-utilities\") pod \"redhat-operators-lwfs7\" (UID: \"eb46f2ea-20fe-4f17-9af4-5f09120e868d\") " pod="openshift-marketplace/redhat-operators-lwfs7" Oct 08 20:58:11 crc kubenswrapper[4988]: I1008 20:58:11.519406 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/eb46f2ea-20fe-4f17-9af4-5f09120e868d-catalog-content\") pod \"redhat-operators-lwfs7\" (UID: \"eb46f2ea-20fe-4f17-9af4-5f09120e868d\") " pod="openshift-marketplace/redhat-operators-lwfs7" Oct 08 20:58:11 crc kubenswrapper[4988]: I1008 20:58:11.545223 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-94lvk\" (UniqueName: \"kubernetes.io/projected/eb46f2ea-20fe-4f17-9af4-5f09120e868d-kube-api-access-94lvk\") pod \"redhat-operators-lwfs7\" (UID: \"eb46f2ea-20fe-4f17-9af4-5f09120e868d\") " pod="openshift-marketplace/redhat-operators-lwfs7" Oct 08 20:58:11 crc kubenswrapper[4988]: I1008 20:58:11.732662 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-lwfs7" Oct 08 20:58:12 crc kubenswrapper[4988]: I1008 20:58:12.923822 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-lwfs7"] Oct 08 20:58:13 crc kubenswrapper[4988]: E1008 20:58:13.395931 4988 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podeb46f2ea_20fe_4f17_9af4_5f09120e868d.slice/crio-70e6d5f8bda74e379f816bb9926a0ec665584b37d451f74acbcf42f97c84a59c.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podeb46f2ea_20fe_4f17_9af4_5f09120e868d.slice/crio-conmon-70e6d5f8bda74e379f816bb9926a0ec665584b37d451f74acbcf42f97c84a59c.scope\": RecentStats: unable to find data in memory cache]" Oct 08 20:58:13 crc kubenswrapper[4988]: I1008 20:58:13.441250 4988 generic.go:334] "Generic (PLEG): container finished" podID="eb46f2ea-20fe-4f17-9af4-5f09120e868d" containerID="70e6d5f8bda74e379f816bb9926a0ec665584b37d451f74acbcf42f97c84a59c" exitCode=0 Oct 08 20:58:13 crc kubenswrapper[4988]: I1008 20:58:13.441308 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-lwfs7" event={"ID":"eb46f2ea-20fe-4f17-9af4-5f09120e868d","Type":"ContainerDied","Data":"70e6d5f8bda74e379f816bb9926a0ec665584b37d451f74acbcf42f97c84a59c"} Oct 08 20:58:13 crc kubenswrapper[4988]: I1008 20:58:13.441340 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-lwfs7" event={"ID":"eb46f2ea-20fe-4f17-9af4-5f09120e868d","Type":"ContainerStarted","Data":"58ba85ad11fbf4c62084a22ddaf8ee76be22116123b0707ef5081cca5df33bfe"} Oct 08 20:58:15 crc kubenswrapper[4988]: I1008 20:58:15.467329 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-lwfs7" event={"ID":"eb46f2ea-20fe-4f17-9af4-5f09120e868d","Type":"ContainerStarted","Data":"dde29775677a7c307544f99451e05d01d9bd63215557927bbe36b07871df330a"} Oct 08 20:58:17 crc kubenswrapper[4988]: I1008 20:58:17.239425 4988 scope.go:117] "RemoveContainer" containerID="fc5906dffdd6ca78fe0b71f3079a04d7d08624d242c87fd146529c035ee33bca" Oct 08 20:58:17 crc kubenswrapper[4988]: E1008 20:58:17.240014 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nm54f_openshift-machine-config-operator(f328a20b-5283-407e-8a93-2ab2da599f18)\"" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" Oct 08 20:58:23 crc kubenswrapper[4988]: I1008 20:58:23.593249 4988 generic.go:334] "Generic (PLEG): container finished" podID="eb46f2ea-20fe-4f17-9af4-5f09120e868d" containerID="dde29775677a7c307544f99451e05d01d9bd63215557927bbe36b07871df330a" exitCode=0 Oct 08 20:58:23 crc kubenswrapper[4988]: I1008 20:58:23.593340 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-lwfs7" event={"ID":"eb46f2ea-20fe-4f17-9af4-5f09120e868d","Type":"ContainerDied","Data":"dde29775677a7c307544f99451e05d01d9bd63215557927bbe36b07871df330a"} Oct 08 20:58:23 crc kubenswrapper[4988]: I1008 20:58:23.599651 4988 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 08 20:58:24 crc kubenswrapper[4988]: I1008 20:58:24.622523 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-lwfs7" event={"ID":"eb46f2ea-20fe-4f17-9af4-5f09120e868d","Type":"ContainerStarted","Data":"955556cb40c964e7c16342dbe76413f38b27cb8afdb0bd8bed7c611bbecf0a60"} Oct 08 20:58:24 crc kubenswrapper[4988]: I1008 20:58:24.658084 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-lwfs7" podStartSLOduration=3.030151991 podStartE2EDuration="13.658057136s" podCreationTimestamp="2025-10-08 20:58:11 +0000 UTC" firstStartedPulling="2025-10-08 20:58:13.445351782 +0000 UTC m=+10038.895194552" lastFinishedPulling="2025-10-08 20:58:24.073256917 +0000 UTC m=+10049.523099697" observedRunningTime="2025-10-08 20:58:24.647529489 +0000 UTC m=+10050.097372329" watchObservedRunningTime="2025-10-08 20:58:24.658057136 +0000 UTC m=+10050.107899946" Oct 08 20:58:29 crc kubenswrapper[4988]: I1008 20:58:29.238310 4988 scope.go:117] "RemoveContainer" containerID="fc5906dffdd6ca78fe0b71f3079a04d7d08624d242c87fd146529c035ee33bca" Oct 08 20:58:29 crc kubenswrapper[4988]: E1008 20:58:29.239814 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nm54f_openshift-machine-config-operator(f328a20b-5283-407e-8a93-2ab2da599f18)\"" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" Oct 08 20:58:31 crc kubenswrapper[4988]: I1008 20:58:31.733381 4988 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-lwfs7" Oct 08 20:58:31 crc kubenswrapper[4988]: I1008 20:58:31.736174 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-lwfs7" Oct 08 20:58:31 crc kubenswrapper[4988]: I1008 20:58:31.792781 4988 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-lwfs7" Oct 08 20:58:32 crc kubenswrapper[4988]: I1008 20:58:32.801999 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-lwfs7" Oct 08 20:58:32 crc kubenswrapper[4988]: I1008 20:58:32.885834 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-lwfs7"] Oct 08 20:58:34 crc kubenswrapper[4988]: I1008 20:58:34.751575 4988 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-lwfs7" podUID="eb46f2ea-20fe-4f17-9af4-5f09120e868d" containerName="registry-server" containerID="cri-o://955556cb40c964e7c16342dbe76413f38b27cb8afdb0bd8bed7c611bbecf0a60" gracePeriod=2 Oct 08 20:58:35 crc kubenswrapper[4988]: I1008 20:58:35.275413 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-lwfs7" Oct 08 20:58:35 crc kubenswrapper[4988]: I1008 20:58:35.448100 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/eb46f2ea-20fe-4f17-9af4-5f09120e868d-catalog-content\") pod \"eb46f2ea-20fe-4f17-9af4-5f09120e868d\" (UID: \"eb46f2ea-20fe-4f17-9af4-5f09120e868d\") " Oct 08 20:58:35 crc kubenswrapper[4988]: I1008 20:58:35.448231 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/eb46f2ea-20fe-4f17-9af4-5f09120e868d-utilities\") pod \"eb46f2ea-20fe-4f17-9af4-5f09120e868d\" (UID: \"eb46f2ea-20fe-4f17-9af4-5f09120e868d\") " Oct 08 20:58:35 crc kubenswrapper[4988]: I1008 20:58:35.448482 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-94lvk\" (UniqueName: \"kubernetes.io/projected/eb46f2ea-20fe-4f17-9af4-5f09120e868d-kube-api-access-94lvk\") pod \"eb46f2ea-20fe-4f17-9af4-5f09120e868d\" (UID: \"eb46f2ea-20fe-4f17-9af4-5f09120e868d\") " Oct 08 20:58:35 crc kubenswrapper[4988]: I1008 20:58:35.448928 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/eb46f2ea-20fe-4f17-9af4-5f09120e868d-utilities" (OuterVolumeSpecName: "utilities") pod "eb46f2ea-20fe-4f17-9af4-5f09120e868d" (UID: "eb46f2ea-20fe-4f17-9af4-5f09120e868d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 20:58:35 crc kubenswrapper[4988]: I1008 20:58:35.449673 4988 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/eb46f2ea-20fe-4f17-9af4-5f09120e868d-utilities\") on node \"crc\" DevicePath \"\"" Oct 08 20:58:35 crc kubenswrapper[4988]: I1008 20:58:35.455017 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/eb46f2ea-20fe-4f17-9af4-5f09120e868d-kube-api-access-94lvk" (OuterVolumeSpecName: "kube-api-access-94lvk") pod "eb46f2ea-20fe-4f17-9af4-5f09120e868d" (UID: "eb46f2ea-20fe-4f17-9af4-5f09120e868d"). InnerVolumeSpecName "kube-api-access-94lvk". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 20:58:35 crc kubenswrapper[4988]: I1008 20:58:35.528653 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/eb46f2ea-20fe-4f17-9af4-5f09120e868d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "eb46f2ea-20fe-4f17-9af4-5f09120e868d" (UID: "eb46f2ea-20fe-4f17-9af4-5f09120e868d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 20:58:35 crc kubenswrapper[4988]: I1008 20:58:35.551751 4988 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-94lvk\" (UniqueName: \"kubernetes.io/projected/eb46f2ea-20fe-4f17-9af4-5f09120e868d-kube-api-access-94lvk\") on node \"crc\" DevicePath \"\"" Oct 08 20:58:35 crc kubenswrapper[4988]: I1008 20:58:35.551790 4988 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/eb46f2ea-20fe-4f17-9af4-5f09120e868d-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 08 20:58:35 crc kubenswrapper[4988]: I1008 20:58:35.771151 4988 generic.go:334] "Generic (PLEG): container finished" podID="eb46f2ea-20fe-4f17-9af4-5f09120e868d" containerID="955556cb40c964e7c16342dbe76413f38b27cb8afdb0bd8bed7c611bbecf0a60" exitCode=0 Oct 08 20:58:35 crc kubenswrapper[4988]: I1008 20:58:35.771216 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-lwfs7" event={"ID":"eb46f2ea-20fe-4f17-9af4-5f09120e868d","Type":"ContainerDied","Data":"955556cb40c964e7c16342dbe76413f38b27cb8afdb0bd8bed7c611bbecf0a60"} Oct 08 20:58:35 crc kubenswrapper[4988]: I1008 20:58:35.771613 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-lwfs7" event={"ID":"eb46f2ea-20fe-4f17-9af4-5f09120e868d","Type":"ContainerDied","Data":"58ba85ad11fbf4c62084a22ddaf8ee76be22116123b0707ef5081cca5df33bfe"} Oct 08 20:58:35 crc kubenswrapper[4988]: I1008 20:58:35.771256 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-lwfs7" Oct 08 20:58:35 crc kubenswrapper[4988]: I1008 20:58:35.771651 4988 scope.go:117] "RemoveContainer" containerID="955556cb40c964e7c16342dbe76413f38b27cb8afdb0bd8bed7c611bbecf0a60" Oct 08 20:58:35 crc kubenswrapper[4988]: I1008 20:58:35.805974 4988 scope.go:117] "RemoveContainer" containerID="dde29775677a7c307544f99451e05d01d9bd63215557927bbe36b07871df330a" Oct 08 20:58:35 crc kubenswrapper[4988]: I1008 20:58:35.826921 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-lwfs7"] Oct 08 20:58:35 crc kubenswrapper[4988]: I1008 20:58:35.836689 4988 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-lwfs7"] Oct 08 20:58:35 crc kubenswrapper[4988]: I1008 20:58:35.853594 4988 scope.go:117] "RemoveContainer" containerID="70e6d5f8bda74e379f816bb9926a0ec665584b37d451f74acbcf42f97c84a59c" Oct 08 20:58:35 crc kubenswrapper[4988]: I1008 20:58:35.881556 4988 scope.go:117] "RemoveContainer" containerID="955556cb40c964e7c16342dbe76413f38b27cb8afdb0bd8bed7c611bbecf0a60" Oct 08 20:58:35 crc kubenswrapper[4988]: E1008 20:58:35.881996 4988 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"955556cb40c964e7c16342dbe76413f38b27cb8afdb0bd8bed7c611bbecf0a60\": container with ID starting with 955556cb40c964e7c16342dbe76413f38b27cb8afdb0bd8bed7c611bbecf0a60 not found: ID does not exist" containerID="955556cb40c964e7c16342dbe76413f38b27cb8afdb0bd8bed7c611bbecf0a60" Oct 08 20:58:35 crc kubenswrapper[4988]: I1008 20:58:35.882035 4988 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"955556cb40c964e7c16342dbe76413f38b27cb8afdb0bd8bed7c611bbecf0a60"} err="failed to get container status \"955556cb40c964e7c16342dbe76413f38b27cb8afdb0bd8bed7c611bbecf0a60\": rpc error: code = NotFound desc = could not find container \"955556cb40c964e7c16342dbe76413f38b27cb8afdb0bd8bed7c611bbecf0a60\": container with ID starting with 955556cb40c964e7c16342dbe76413f38b27cb8afdb0bd8bed7c611bbecf0a60 not found: ID does not exist" Oct 08 20:58:35 crc kubenswrapper[4988]: I1008 20:58:35.882063 4988 scope.go:117] "RemoveContainer" containerID="dde29775677a7c307544f99451e05d01d9bd63215557927bbe36b07871df330a" Oct 08 20:58:35 crc kubenswrapper[4988]: E1008 20:58:35.882853 4988 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"dde29775677a7c307544f99451e05d01d9bd63215557927bbe36b07871df330a\": container with ID starting with dde29775677a7c307544f99451e05d01d9bd63215557927bbe36b07871df330a not found: ID does not exist" containerID="dde29775677a7c307544f99451e05d01d9bd63215557927bbe36b07871df330a" Oct 08 20:58:35 crc kubenswrapper[4988]: I1008 20:58:35.882878 4988 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dde29775677a7c307544f99451e05d01d9bd63215557927bbe36b07871df330a"} err="failed to get container status \"dde29775677a7c307544f99451e05d01d9bd63215557927bbe36b07871df330a\": rpc error: code = NotFound desc = could not find container \"dde29775677a7c307544f99451e05d01d9bd63215557927bbe36b07871df330a\": container with ID starting with dde29775677a7c307544f99451e05d01d9bd63215557927bbe36b07871df330a not found: ID does not exist" Oct 08 20:58:35 crc kubenswrapper[4988]: I1008 20:58:35.882895 4988 scope.go:117] "RemoveContainer" containerID="70e6d5f8bda74e379f816bb9926a0ec665584b37d451f74acbcf42f97c84a59c" Oct 08 20:58:35 crc kubenswrapper[4988]: E1008 20:58:35.883289 4988 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"70e6d5f8bda74e379f816bb9926a0ec665584b37d451f74acbcf42f97c84a59c\": container with ID starting with 70e6d5f8bda74e379f816bb9926a0ec665584b37d451f74acbcf42f97c84a59c not found: ID does not exist" containerID="70e6d5f8bda74e379f816bb9926a0ec665584b37d451f74acbcf42f97c84a59c" Oct 08 20:58:35 crc kubenswrapper[4988]: I1008 20:58:35.883316 4988 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"70e6d5f8bda74e379f816bb9926a0ec665584b37d451f74acbcf42f97c84a59c"} err="failed to get container status \"70e6d5f8bda74e379f816bb9926a0ec665584b37d451f74acbcf42f97c84a59c\": rpc error: code = NotFound desc = could not find container \"70e6d5f8bda74e379f816bb9926a0ec665584b37d451f74acbcf42f97c84a59c\": container with ID starting with 70e6d5f8bda74e379f816bb9926a0ec665584b37d451f74acbcf42f97c84a59c not found: ID does not exist" Oct 08 20:58:37 crc kubenswrapper[4988]: I1008 20:58:37.261467 4988 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="eb46f2ea-20fe-4f17-9af4-5f09120e868d" path="/var/lib/kubelet/pods/eb46f2ea-20fe-4f17-9af4-5f09120e868d/volumes" Oct 08 20:58:42 crc kubenswrapper[4988]: I1008 20:58:42.238933 4988 scope.go:117] "RemoveContainer" containerID="fc5906dffdd6ca78fe0b71f3079a04d7d08624d242c87fd146529c035ee33bca" Oct 08 20:58:42 crc kubenswrapper[4988]: E1008 20:58:42.239629 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nm54f_openshift-machine-config-operator(f328a20b-5283-407e-8a93-2ab2da599f18)\"" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" Oct 08 20:58:54 crc kubenswrapper[4988]: I1008 20:58:54.238250 4988 scope.go:117] "RemoveContainer" containerID="fc5906dffdd6ca78fe0b71f3079a04d7d08624d242c87fd146529c035ee33bca" Oct 08 20:58:54 crc kubenswrapper[4988]: E1008 20:58:54.239809 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nm54f_openshift-machine-config-operator(f328a20b-5283-407e-8a93-2ab2da599f18)\"" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" Oct 08 20:59:07 crc kubenswrapper[4988]: I1008 20:59:07.239146 4988 scope.go:117] "RemoveContainer" containerID="fc5906dffdd6ca78fe0b71f3079a04d7d08624d242c87fd146529c035ee33bca" Oct 08 20:59:07 crc kubenswrapper[4988]: E1008 20:59:07.240293 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nm54f_openshift-machine-config-operator(f328a20b-5283-407e-8a93-2ab2da599f18)\"" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" Oct 08 20:59:21 crc kubenswrapper[4988]: I1008 20:59:21.238207 4988 scope.go:117] "RemoveContainer" containerID="fc5906dffdd6ca78fe0b71f3079a04d7d08624d242c87fd146529c035ee33bca" Oct 08 20:59:21 crc kubenswrapper[4988]: E1008 20:59:21.239114 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nm54f_openshift-machine-config-operator(f328a20b-5283-407e-8a93-2ab2da599f18)\"" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" Oct 08 20:59:35 crc kubenswrapper[4988]: I1008 20:59:35.247982 4988 scope.go:117] "RemoveContainer" containerID="fc5906dffdd6ca78fe0b71f3079a04d7d08624d242c87fd146529c035ee33bca" Oct 08 20:59:35 crc kubenswrapper[4988]: E1008 20:59:35.249086 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nm54f_openshift-machine-config-operator(f328a20b-5283-407e-8a93-2ab2da599f18)\"" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" Oct 08 20:59:40 crc kubenswrapper[4988]: I1008 20:59:40.048053 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mariadb-copy-data"] Oct 08 20:59:40 crc kubenswrapper[4988]: I1008 20:59:40.048893 4988 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/mariadb-copy-data" podUID="9e312e3a-e0e2-490b-bb43-cc3b7191c109" containerName="adoption" containerID="cri-o://9f72418b4bb8bbb8075246185311d8f58cdc5981b8ff6897ca7c1eb0dedfa988" gracePeriod=30 Oct 08 20:59:47 crc kubenswrapper[4988]: I1008 20:59:47.239025 4988 scope.go:117] "RemoveContainer" containerID="fc5906dffdd6ca78fe0b71f3079a04d7d08624d242c87fd146529c035ee33bca" Oct 08 20:59:47 crc kubenswrapper[4988]: E1008 20:59:47.242289 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nm54f_openshift-machine-config-operator(f328a20b-5283-407e-8a93-2ab2da599f18)\"" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" Oct 08 21:00:00 crc kubenswrapper[4988]: I1008 21:00:00.200404 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29332620-89w5h"] Oct 08 21:00:00 crc kubenswrapper[4988]: E1008 21:00:00.201448 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eb46f2ea-20fe-4f17-9af4-5f09120e868d" containerName="registry-server" Oct 08 21:00:00 crc kubenswrapper[4988]: I1008 21:00:00.201463 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="eb46f2ea-20fe-4f17-9af4-5f09120e868d" containerName="registry-server" Oct 08 21:00:00 crc kubenswrapper[4988]: E1008 21:00:00.201486 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eb46f2ea-20fe-4f17-9af4-5f09120e868d" containerName="extract-utilities" Oct 08 21:00:00 crc kubenswrapper[4988]: I1008 21:00:00.201493 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="eb46f2ea-20fe-4f17-9af4-5f09120e868d" containerName="extract-utilities" Oct 08 21:00:00 crc kubenswrapper[4988]: E1008 21:00:00.201524 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eb46f2ea-20fe-4f17-9af4-5f09120e868d" containerName="extract-content" Oct 08 21:00:00 crc kubenswrapper[4988]: I1008 21:00:00.201532 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="eb46f2ea-20fe-4f17-9af4-5f09120e868d" containerName="extract-content" Oct 08 21:00:00 crc kubenswrapper[4988]: I1008 21:00:00.201765 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="eb46f2ea-20fe-4f17-9af4-5f09120e868d" containerName="registry-server" Oct 08 21:00:00 crc kubenswrapper[4988]: I1008 21:00:00.202549 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29332620-89w5h" Oct 08 21:00:00 crc kubenswrapper[4988]: I1008 21:00:00.205998 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 08 21:00:00 crc kubenswrapper[4988]: I1008 21:00:00.208439 4988 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 08 21:00:00 crc kubenswrapper[4988]: I1008 21:00:00.219124 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5lmms\" (UniqueName: \"kubernetes.io/projected/a0d00888-ee8d-406a-ab16-7b014aa100e7-kube-api-access-5lmms\") pod \"collect-profiles-29332620-89w5h\" (UID: \"a0d00888-ee8d-406a-ab16-7b014aa100e7\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29332620-89w5h" Oct 08 21:00:00 crc kubenswrapper[4988]: I1008 21:00:00.219258 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/a0d00888-ee8d-406a-ab16-7b014aa100e7-config-volume\") pod \"collect-profiles-29332620-89w5h\" (UID: \"a0d00888-ee8d-406a-ab16-7b014aa100e7\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29332620-89w5h" Oct 08 21:00:00 crc kubenswrapper[4988]: I1008 21:00:00.219553 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/a0d00888-ee8d-406a-ab16-7b014aa100e7-secret-volume\") pod \"collect-profiles-29332620-89w5h\" (UID: \"a0d00888-ee8d-406a-ab16-7b014aa100e7\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29332620-89w5h" Oct 08 21:00:00 crc kubenswrapper[4988]: I1008 21:00:00.228438 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29332620-89w5h"] Oct 08 21:00:00 crc kubenswrapper[4988]: I1008 21:00:00.239528 4988 scope.go:117] "RemoveContainer" containerID="fc5906dffdd6ca78fe0b71f3079a04d7d08624d242c87fd146529c035ee33bca" Oct 08 21:00:00 crc kubenswrapper[4988]: E1008 21:00:00.239921 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nm54f_openshift-machine-config-operator(f328a20b-5283-407e-8a93-2ab2da599f18)\"" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" Oct 08 21:00:00 crc kubenswrapper[4988]: I1008 21:00:00.322089 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5lmms\" (UniqueName: \"kubernetes.io/projected/a0d00888-ee8d-406a-ab16-7b014aa100e7-kube-api-access-5lmms\") pod \"collect-profiles-29332620-89w5h\" (UID: \"a0d00888-ee8d-406a-ab16-7b014aa100e7\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29332620-89w5h" Oct 08 21:00:00 crc kubenswrapper[4988]: I1008 21:00:00.322242 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/a0d00888-ee8d-406a-ab16-7b014aa100e7-config-volume\") pod \"collect-profiles-29332620-89w5h\" (UID: \"a0d00888-ee8d-406a-ab16-7b014aa100e7\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29332620-89w5h" Oct 08 21:00:00 crc kubenswrapper[4988]: I1008 21:00:00.322436 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/a0d00888-ee8d-406a-ab16-7b014aa100e7-secret-volume\") pod \"collect-profiles-29332620-89w5h\" (UID: \"a0d00888-ee8d-406a-ab16-7b014aa100e7\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29332620-89w5h" Oct 08 21:00:00 crc kubenswrapper[4988]: I1008 21:00:00.323538 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/a0d00888-ee8d-406a-ab16-7b014aa100e7-config-volume\") pod \"collect-profiles-29332620-89w5h\" (UID: \"a0d00888-ee8d-406a-ab16-7b014aa100e7\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29332620-89w5h" Oct 08 21:00:00 crc kubenswrapper[4988]: I1008 21:00:00.330747 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/a0d00888-ee8d-406a-ab16-7b014aa100e7-secret-volume\") pod \"collect-profiles-29332620-89w5h\" (UID: \"a0d00888-ee8d-406a-ab16-7b014aa100e7\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29332620-89w5h" Oct 08 21:00:00 crc kubenswrapper[4988]: I1008 21:00:00.340728 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5lmms\" (UniqueName: \"kubernetes.io/projected/a0d00888-ee8d-406a-ab16-7b014aa100e7-kube-api-access-5lmms\") pod \"collect-profiles-29332620-89w5h\" (UID: \"a0d00888-ee8d-406a-ab16-7b014aa100e7\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29332620-89w5h" Oct 08 21:00:00 crc kubenswrapper[4988]: I1008 21:00:00.539578 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29332620-89w5h" Oct 08 21:00:01 crc kubenswrapper[4988]: I1008 21:00:01.106019 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29332620-89w5h"] Oct 08 21:00:01 crc kubenswrapper[4988]: I1008 21:00:01.866288 4988 generic.go:334] "Generic (PLEG): container finished" podID="a0d00888-ee8d-406a-ab16-7b014aa100e7" containerID="81a632ec2dff1f7f72e3ed905b25869e2194dabfe8164fabbcee3220fc7eedcd" exitCode=0 Oct 08 21:00:01 crc kubenswrapper[4988]: I1008 21:00:01.866342 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29332620-89w5h" event={"ID":"a0d00888-ee8d-406a-ab16-7b014aa100e7","Type":"ContainerDied","Data":"81a632ec2dff1f7f72e3ed905b25869e2194dabfe8164fabbcee3220fc7eedcd"} Oct 08 21:00:01 crc kubenswrapper[4988]: I1008 21:00:01.866812 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29332620-89w5h" event={"ID":"a0d00888-ee8d-406a-ab16-7b014aa100e7","Type":"ContainerStarted","Data":"2e64a74e12a49bf81b770ff5a65a78987f3b7feb4064171c283fd6bd698038da"} Oct 08 21:00:03 crc kubenswrapper[4988]: I1008 21:00:03.369425 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29332620-89w5h" Oct 08 21:00:03 crc kubenswrapper[4988]: I1008 21:00:03.437598 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/a0d00888-ee8d-406a-ab16-7b014aa100e7-config-volume\") pod \"a0d00888-ee8d-406a-ab16-7b014aa100e7\" (UID: \"a0d00888-ee8d-406a-ab16-7b014aa100e7\") " Oct 08 21:00:03 crc kubenswrapper[4988]: I1008 21:00:03.437817 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/a0d00888-ee8d-406a-ab16-7b014aa100e7-secret-volume\") pod \"a0d00888-ee8d-406a-ab16-7b014aa100e7\" (UID: \"a0d00888-ee8d-406a-ab16-7b014aa100e7\") " Oct 08 21:00:03 crc kubenswrapper[4988]: I1008 21:00:03.437893 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5lmms\" (UniqueName: \"kubernetes.io/projected/a0d00888-ee8d-406a-ab16-7b014aa100e7-kube-api-access-5lmms\") pod \"a0d00888-ee8d-406a-ab16-7b014aa100e7\" (UID: \"a0d00888-ee8d-406a-ab16-7b014aa100e7\") " Oct 08 21:00:03 crc kubenswrapper[4988]: I1008 21:00:03.438356 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a0d00888-ee8d-406a-ab16-7b014aa100e7-config-volume" (OuterVolumeSpecName: "config-volume") pod "a0d00888-ee8d-406a-ab16-7b014aa100e7" (UID: "a0d00888-ee8d-406a-ab16-7b014aa100e7"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 21:00:03 crc kubenswrapper[4988]: I1008 21:00:03.443198 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a0d00888-ee8d-406a-ab16-7b014aa100e7-kube-api-access-5lmms" (OuterVolumeSpecName: "kube-api-access-5lmms") pod "a0d00888-ee8d-406a-ab16-7b014aa100e7" (UID: "a0d00888-ee8d-406a-ab16-7b014aa100e7"). InnerVolumeSpecName "kube-api-access-5lmms". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 21:00:03 crc kubenswrapper[4988]: I1008 21:00:03.444329 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a0d00888-ee8d-406a-ab16-7b014aa100e7-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "a0d00888-ee8d-406a-ab16-7b014aa100e7" (UID: "a0d00888-ee8d-406a-ab16-7b014aa100e7"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 21:00:03 crc kubenswrapper[4988]: I1008 21:00:03.540214 4988 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/a0d00888-ee8d-406a-ab16-7b014aa100e7-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 08 21:00:03 crc kubenswrapper[4988]: I1008 21:00:03.540257 4988 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5lmms\" (UniqueName: \"kubernetes.io/projected/a0d00888-ee8d-406a-ab16-7b014aa100e7-kube-api-access-5lmms\") on node \"crc\" DevicePath \"\"" Oct 08 21:00:03 crc kubenswrapper[4988]: I1008 21:00:03.540271 4988 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/a0d00888-ee8d-406a-ab16-7b014aa100e7-config-volume\") on node \"crc\" DevicePath \"\"" Oct 08 21:00:03 crc kubenswrapper[4988]: I1008 21:00:03.898795 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29332620-89w5h" event={"ID":"a0d00888-ee8d-406a-ab16-7b014aa100e7","Type":"ContainerDied","Data":"2e64a74e12a49bf81b770ff5a65a78987f3b7feb4064171c283fd6bd698038da"} Oct 08 21:00:03 crc kubenswrapper[4988]: I1008 21:00:03.898835 4988 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2e64a74e12a49bf81b770ff5a65a78987f3b7feb4064171c283fd6bd698038da" Oct 08 21:00:03 crc kubenswrapper[4988]: I1008 21:00:03.898893 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29332620-89w5h" Oct 08 21:00:04 crc kubenswrapper[4988]: I1008 21:00:04.501431 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29332575-svtg7"] Oct 08 21:00:04 crc kubenswrapper[4988]: I1008 21:00:04.517188 4988 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29332575-svtg7"] Oct 08 21:00:05 crc kubenswrapper[4988]: I1008 21:00:05.259190 4988 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="32593c4e-949e-40be-884a-3074d923b4f9" path="/var/lib/kubelet/pods/32593c4e-949e-40be-884a-3074d923b4f9/volumes" Oct 08 21:00:10 crc kubenswrapper[4988]: I1008 21:00:10.669699 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-copy-data" Oct 08 21:00:10 crc kubenswrapper[4988]: I1008 21:00:10.719766 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jw4zs\" (UniqueName: \"kubernetes.io/projected/9e312e3a-e0e2-490b-bb43-cc3b7191c109-kube-api-access-jw4zs\") pod \"9e312e3a-e0e2-490b-bb43-cc3b7191c109\" (UID: \"9e312e3a-e0e2-490b-bb43-cc3b7191c109\") " Oct 08 21:00:10 crc kubenswrapper[4988]: I1008 21:00:10.721974 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mariadb-data\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-81095d3c-890a-4460-a497-531b459f10ce\") pod \"9e312e3a-e0e2-490b-bb43-cc3b7191c109\" (UID: \"9e312e3a-e0e2-490b-bb43-cc3b7191c109\") " Oct 08 21:00:10 crc kubenswrapper[4988]: I1008 21:00:10.737543 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9e312e3a-e0e2-490b-bb43-cc3b7191c109-kube-api-access-jw4zs" (OuterVolumeSpecName: "kube-api-access-jw4zs") pod "9e312e3a-e0e2-490b-bb43-cc3b7191c109" (UID: "9e312e3a-e0e2-490b-bb43-cc3b7191c109"). InnerVolumeSpecName "kube-api-access-jw4zs". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 21:00:10 crc kubenswrapper[4988]: I1008 21:00:10.757173 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-81095d3c-890a-4460-a497-531b459f10ce" (OuterVolumeSpecName: "mariadb-data") pod "9e312e3a-e0e2-490b-bb43-cc3b7191c109" (UID: "9e312e3a-e0e2-490b-bb43-cc3b7191c109"). InnerVolumeSpecName "pvc-81095d3c-890a-4460-a497-531b459f10ce". PluginName "kubernetes.io/csi", VolumeGidValue "" Oct 08 21:00:10 crc kubenswrapper[4988]: I1008 21:00:10.825625 4988 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jw4zs\" (UniqueName: \"kubernetes.io/projected/9e312e3a-e0e2-490b-bb43-cc3b7191c109-kube-api-access-jw4zs\") on node \"crc\" DevicePath \"\"" Oct 08 21:00:10 crc kubenswrapper[4988]: I1008 21:00:10.825693 4988 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"pvc-81095d3c-890a-4460-a497-531b459f10ce\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-81095d3c-890a-4460-a497-531b459f10ce\") on node \"crc\" " Oct 08 21:00:10 crc kubenswrapper[4988]: I1008 21:00:10.854277 4988 csi_attacher.go:630] kubernetes.io/csi: attacher.UnmountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping UnmountDevice... Oct 08 21:00:10 crc kubenswrapper[4988]: I1008 21:00:10.855392 4988 operation_generator.go:917] UnmountDevice succeeded for volume "pvc-81095d3c-890a-4460-a497-531b459f10ce" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-81095d3c-890a-4460-a497-531b459f10ce") on node "crc" Oct 08 21:00:10 crc kubenswrapper[4988]: I1008 21:00:10.927051 4988 reconciler_common.go:293] "Volume detached for volume \"pvc-81095d3c-890a-4460-a497-531b459f10ce\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-81095d3c-890a-4460-a497-531b459f10ce\") on node \"crc\" DevicePath \"\"" Oct 08 21:00:10 crc kubenswrapper[4988]: I1008 21:00:10.998898 4988 generic.go:334] "Generic (PLEG): container finished" podID="9e312e3a-e0e2-490b-bb43-cc3b7191c109" containerID="9f72418b4bb8bbb8075246185311d8f58cdc5981b8ff6897ca7c1eb0dedfa988" exitCode=137 Oct 08 21:00:10 crc kubenswrapper[4988]: I1008 21:00:10.998984 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-copy-data" Oct 08 21:00:10 crc kubenswrapper[4988]: I1008 21:00:10.998996 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-copy-data" event={"ID":"9e312e3a-e0e2-490b-bb43-cc3b7191c109","Type":"ContainerDied","Data":"9f72418b4bb8bbb8075246185311d8f58cdc5981b8ff6897ca7c1eb0dedfa988"} Oct 08 21:00:10 crc kubenswrapper[4988]: I1008 21:00:10.999085 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-copy-data" event={"ID":"9e312e3a-e0e2-490b-bb43-cc3b7191c109","Type":"ContainerDied","Data":"2c41cfac11e6123f9fbbd7a1a71110c6bb83d0cde4a428538571ef8fe94ffa50"} Oct 08 21:00:10 crc kubenswrapper[4988]: I1008 21:00:10.999116 4988 scope.go:117] "RemoveContainer" containerID="9f72418b4bb8bbb8075246185311d8f58cdc5981b8ff6897ca7c1eb0dedfa988" Oct 08 21:00:11 crc kubenswrapper[4988]: I1008 21:00:11.026698 4988 scope.go:117] "RemoveContainer" containerID="9f72418b4bb8bbb8075246185311d8f58cdc5981b8ff6897ca7c1eb0dedfa988" Oct 08 21:00:11 crc kubenswrapper[4988]: E1008 21:00:11.027232 4988 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9f72418b4bb8bbb8075246185311d8f58cdc5981b8ff6897ca7c1eb0dedfa988\": container with ID starting with 9f72418b4bb8bbb8075246185311d8f58cdc5981b8ff6897ca7c1eb0dedfa988 not found: ID does not exist" containerID="9f72418b4bb8bbb8075246185311d8f58cdc5981b8ff6897ca7c1eb0dedfa988" Oct 08 21:00:11 crc kubenswrapper[4988]: I1008 21:00:11.027285 4988 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9f72418b4bb8bbb8075246185311d8f58cdc5981b8ff6897ca7c1eb0dedfa988"} err="failed to get container status \"9f72418b4bb8bbb8075246185311d8f58cdc5981b8ff6897ca7c1eb0dedfa988\": rpc error: code = NotFound desc = could not find container \"9f72418b4bb8bbb8075246185311d8f58cdc5981b8ff6897ca7c1eb0dedfa988\": container with ID starting with 9f72418b4bb8bbb8075246185311d8f58cdc5981b8ff6897ca7c1eb0dedfa988 not found: ID does not exist" Oct 08 21:00:11 crc kubenswrapper[4988]: I1008 21:00:11.055885 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mariadb-copy-data"] Oct 08 21:00:11 crc kubenswrapper[4988]: I1008 21:00:11.068159 4988 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/mariadb-copy-data"] Oct 08 21:00:11 crc kubenswrapper[4988]: I1008 21:00:11.256261 4988 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9e312e3a-e0e2-490b-bb43-cc3b7191c109" path="/var/lib/kubelet/pods/9e312e3a-e0e2-490b-bb43-cc3b7191c109/volumes" Oct 08 21:00:11 crc kubenswrapper[4988]: I1008 21:00:11.777434 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-copy-data"] Oct 08 21:00:11 crc kubenswrapper[4988]: I1008 21:00:11.778912 4988 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ovn-copy-data" podUID="bc523d09-e017-4e91-8ba0-45ca290bd102" containerName="adoption" containerID="cri-o://ec84a95c1676eda81cf66f0edcee8ec4cd2b7b456c53f4e63d241ae3a94c3fd2" gracePeriod=30 Oct 08 21:00:13 crc kubenswrapper[4988]: I1008 21:00:13.238739 4988 scope.go:117] "RemoveContainer" containerID="fc5906dffdd6ca78fe0b71f3079a04d7d08624d242c87fd146529c035ee33bca" Oct 08 21:00:13 crc kubenswrapper[4988]: E1008 21:00:13.239699 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nm54f_openshift-machine-config-operator(f328a20b-5283-407e-8a93-2ab2da599f18)\"" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" Oct 08 21:00:17 crc kubenswrapper[4988]: I1008 21:00:17.116547 4988 scope.go:117] "RemoveContainer" containerID="2bf6c14a16c4c0ca339e89944a1fe637b5c9f8431fa1acad3b682a250598484f" Oct 08 21:00:26 crc kubenswrapper[4988]: I1008 21:00:26.238611 4988 scope.go:117] "RemoveContainer" containerID="fc5906dffdd6ca78fe0b71f3079a04d7d08624d242c87fd146529c035ee33bca" Oct 08 21:00:27 crc kubenswrapper[4988]: I1008 21:00:27.201724 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" event={"ID":"f328a20b-5283-407e-8a93-2ab2da599f18","Type":"ContainerStarted","Data":"6a30956f2db44d788d7c7cdfd2114a016aa50c17f8bee2b54970bedfe0df7317"} Oct 08 21:00:42 crc kubenswrapper[4988]: I1008 21:00:42.388954 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-copy-data" Oct 08 21:00:42 crc kubenswrapper[4988]: I1008 21:00:42.445173 4988 generic.go:334] "Generic (PLEG): container finished" podID="bc523d09-e017-4e91-8ba0-45ca290bd102" containerID="ec84a95c1676eda81cf66f0edcee8ec4cd2b7b456c53f4e63d241ae3a94c3fd2" exitCode=137 Oct 08 21:00:42 crc kubenswrapper[4988]: I1008 21:00:42.445256 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-copy-data" event={"ID":"bc523d09-e017-4e91-8ba0-45ca290bd102","Type":"ContainerDied","Data":"ec84a95c1676eda81cf66f0edcee8ec4cd2b7b456c53f4e63d241ae3a94c3fd2"} Oct 08 21:00:42 crc kubenswrapper[4988]: I1008 21:00:42.445270 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-copy-data" Oct 08 21:00:42 crc kubenswrapper[4988]: I1008 21:00:42.445309 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-copy-data" event={"ID":"bc523d09-e017-4e91-8ba0-45ca290bd102","Type":"ContainerDied","Data":"6f022cf84222d9a831a3fdf675980d24133f09be02587a2014db51b0ae62dc24"} Oct 08 21:00:42 crc kubenswrapper[4988]: I1008 21:00:42.445339 4988 scope.go:117] "RemoveContainer" containerID="ec84a95c1676eda81cf66f0edcee8ec4cd2b7b456c53f4e63d241ae3a94c3fd2" Oct 08 21:00:42 crc kubenswrapper[4988]: I1008 21:00:42.483107 4988 scope.go:117] "RemoveContainer" containerID="ec84a95c1676eda81cf66f0edcee8ec4cd2b7b456c53f4e63d241ae3a94c3fd2" Oct 08 21:00:42 crc kubenswrapper[4988]: E1008 21:00:42.483852 4988 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ec84a95c1676eda81cf66f0edcee8ec4cd2b7b456c53f4e63d241ae3a94c3fd2\": container with ID starting with ec84a95c1676eda81cf66f0edcee8ec4cd2b7b456c53f4e63d241ae3a94c3fd2 not found: ID does not exist" containerID="ec84a95c1676eda81cf66f0edcee8ec4cd2b7b456c53f4e63d241ae3a94c3fd2" Oct 08 21:00:42 crc kubenswrapper[4988]: I1008 21:00:42.483901 4988 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ec84a95c1676eda81cf66f0edcee8ec4cd2b7b456c53f4e63d241ae3a94c3fd2"} err="failed to get container status \"ec84a95c1676eda81cf66f0edcee8ec4cd2b7b456c53f4e63d241ae3a94c3fd2\": rpc error: code = NotFound desc = could not find container \"ec84a95c1676eda81cf66f0edcee8ec4cd2b7b456c53f4e63d241ae3a94c3fd2\": container with ID starting with ec84a95c1676eda81cf66f0edcee8ec4cd2b7b456c53f4e63d241ae3a94c3fd2 not found: ID does not exist" Oct 08 21:00:42 crc kubenswrapper[4988]: I1008 21:00:42.499686 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-data\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-c3809aad-6fef-4719-a8e9-1427666bd89a\") pod \"bc523d09-e017-4e91-8ba0-45ca290bd102\" (UID: \"bc523d09-e017-4e91-8ba0-45ca290bd102\") " Oct 08 21:00:42 crc kubenswrapper[4988]: I1008 21:00:42.499785 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-data-cert\" (UniqueName: \"kubernetes.io/secret/bc523d09-e017-4e91-8ba0-45ca290bd102-ovn-data-cert\") pod \"bc523d09-e017-4e91-8ba0-45ca290bd102\" (UID: \"bc523d09-e017-4e91-8ba0-45ca290bd102\") " Oct 08 21:00:42 crc kubenswrapper[4988]: I1008 21:00:42.499849 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dfqbt\" (UniqueName: \"kubernetes.io/projected/bc523d09-e017-4e91-8ba0-45ca290bd102-kube-api-access-dfqbt\") pod \"bc523d09-e017-4e91-8ba0-45ca290bd102\" (UID: \"bc523d09-e017-4e91-8ba0-45ca290bd102\") " Oct 08 21:00:42 crc kubenswrapper[4988]: I1008 21:00:42.507436 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bc523d09-e017-4e91-8ba0-45ca290bd102-ovn-data-cert" (OuterVolumeSpecName: "ovn-data-cert") pod "bc523d09-e017-4e91-8ba0-45ca290bd102" (UID: "bc523d09-e017-4e91-8ba0-45ca290bd102"). InnerVolumeSpecName "ovn-data-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 21:00:42 crc kubenswrapper[4988]: I1008 21:00:42.510814 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bc523d09-e017-4e91-8ba0-45ca290bd102-kube-api-access-dfqbt" (OuterVolumeSpecName: "kube-api-access-dfqbt") pod "bc523d09-e017-4e91-8ba0-45ca290bd102" (UID: "bc523d09-e017-4e91-8ba0-45ca290bd102"). InnerVolumeSpecName "kube-api-access-dfqbt". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 21:00:42 crc kubenswrapper[4988]: I1008 21:00:42.530113 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-c3809aad-6fef-4719-a8e9-1427666bd89a" (OuterVolumeSpecName: "ovn-data") pod "bc523d09-e017-4e91-8ba0-45ca290bd102" (UID: "bc523d09-e017-4e91-8ba0-45ca290bd102"). InnerVolumeSpecName "pvc-c3809aad-6fef-4719-a8e9-1427666bd89a". PluginName "kubernetes.io/csi", VolumeGidValue "" Oct 08 21:00:42 crc kubenswrapper[4988]: I1008 21:00:42.603479 4988 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"pvc-c3809aad-6fef-4719-a8e9-1427666bd89a\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-c3809aad-6fef-4719-a8e9-1427666bd89a\") on node \"crc\" " Oct 08 21:00:42 crc kubenswrapper[4988]: I1008 21:00:42.603532 4988 reconciler_common.go:293] "Volume detached for volume \"ovn-data-cert\" (UniqueName: \"kubernetes.io/secret/bc523d09-e017-4e91-8ba0-45ca290bd102-ovn-data-cert\") on node \"crc\" DevicePath \"\"" Oct 08 21:00:42 crc kubenswrapper[4988]: I1008 21:00:42.603560 4988 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dfqbt\" (UniqueName: \"kubernetes.io/projected/bc523d09-e017-4e91-8ba0-45ca290bd102-kube-api-access-dfqbt\") on node \"crc\" DevicePath \"\"" Oct 08 21:00:43 crc kubenswrapper[4988]: I1008 21:00:43.628270 4988 csi_attacher.go:630] kubernetes.io/csi: attacher.UnmountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping UnmountDevice... Oct 08 21:00:43 crc kubenswrapper[4988]: I1008 21:00:43.628975 4988 operation_generator.go:917] UnmountDevice succeeded for volume "pvc-c3809aad-6fef-4719-a8e9-1427666bd89a" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-c3809aad-6fef-4719-a8e9-1427666bd89a") on node "crc" Oct 08 21:00:43 crc kubenswrapper[4988]: I1008 21:00:43.640141 4988 reconciler_common.go:293] "Volume detached for volume \"pvc-c3809aad-6fef-4719-a8e9-1427666bd89a\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-c3809aad-6fef-4719-a8e9-1427666bd89a\") on node \"crc\" DevicePath \"\"" Oct 08 21:00:43 crc kubenswrapper[4988]: I1008 21:00:43.682592 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-copy-data"] Oct 08 21:00:43 crc kubenswrapper[4988]: I1008 21:00:43.689826 4988 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-copy-data"] Oct 08 21:00:45 crc kubenswrapper[4988]: I1008 21:00:45.257168 4988 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bc523d09-e017-4e91-8ba0-45ca290bd102" path="/var/lib/kubelet/pods/bc523d09-e017-4e91-8ba0-45ca290bd102/volumes" Oct 08 21:01:00 crc kubenswrapper[4988]: I1008 21:01:00.179508 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-cron-29332621-swt82"] Oct 08 21:01:00 crc kubenswrapper[4988]: E1008 21:01:00.180490 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bc523d09-e017-4e91-8ba0-45ca290bd102" containerName="adoption" Oct 08 21:01:00 crc kubenswrapper[4988]: I1008 21:01:00.180504 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="bc523d09-e017-4e91-8ba0-45ca290bd102" containerName="adoption" Oct 08 21:01:00 crc kubenswrapper[4988]: E1008 21:01:00.180553 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9e312e3a-e0e2-490b-bb43-cc3b7191c109" containerName="adoption" Oct 08 21:01:00 crc kubenswrapper[4988]: I1008 21:01:00.180559 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="9e312e3a-e0e2-490b-bb43-cc3b7191c109" containerName="adoption" Oct 08 21:01:00 crc kubenswrapper[4988]: E1008 21:01:00.180577 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a0d00888-ee8d-406a-ab16-7b014aa100e7" containerName="collect-profiles" Oct 08 21:01:00 crc kubenswrapper[4988]: I1008 21:01:00.180583 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="a0d00888-ee8d-406a-ab16-7b014aa100e7" containerName="collect-profiles" Oct 08 21:01:00 crc kubenswrapper[4988]: I1008 21:01:00.180773 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="bc523d09-e017-4e91-8ba0-45ca290bd102" containerName="adoption" Oct 08 21:01:00 crc kubenswrapper[4988]: I1008 21:01:00.180789 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="a0d00888-ee8d-406a-ab16-7b014aa100e7" containerName="collect-profiles" Oct 08 21:01:00 crc kubenswrapper[4988]: I1008 21:01:00.180811 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="9e312e3a-e0e2-490b-bb43-cc3b7191c109" containerName="adoption" Oct 08 21:01:00 crc kubenswrapper[4988]: I1008 21:01:00.181577 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29332621-swt82" Oct 08 21:01:00 crc kubenswrapper[4988]: I1008 21:01:00.194527 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29332621-swt82"] Oct 08 21:01:00 crc kubenswrapper[4988]: I1008 21:01:00.284865 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tlxqg\" (UniqueName: \"kubernetes.io/projected/8d60984c-fb2d-425a-b6e5-87d20c7c0cc5-kube-api-access-tlxqg\") pod \"keystone-cron-29332621-swt82\" (UID: \"8d60984c-fb2d-425a-b6e5-87d20c7c0cc5\") " pod="openstack/keystone-cron-29332621-swt82" Oct 08 21:01:00 crc kubenswrapper[4988]: I1008 21:01:00.285023 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/8d60984c-fb2d-425a-b6e5-87d20c7c0cc5-fernet-keys\") pod \"keystone-cron-29332621-swt82\" (UID: \"8d60984c-fb2d-425a-b6e5-87d20c7c0cc5\") " pod="openstack/keystone-cron-29332621-swt82" Oct 08 21:01:00 crc kubenswrapper[4988]: I1008 21:01:00.285158 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8d60984c-fb2d-425a-b6e5-87d20c7c0cc5-combined-ca-bundle\") pod \"keystone-cron-29332621-swt82\" (UID: \"8d60984c-fb2d-425a-b6e5-87d20c7c0cc5\") " pod="openstack/keystone-cron-29332621-swt82" Oct 08 21:01:00 crc kubenswrapper[4988]: I1008 21:01:00.285206 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8d60984c-fb2d-425a-b6e5-87d20c7c0cc5-config-data\") pod \"keystone-cron-29332621-swt82\" (UID: \"8d60984c-fb2d-425a-b6e5-87d20c7c0cc5\") " pod="openstack/keystone-cron-29332621-swt82" Oct 08 21:01:00 crc kubenswrapper[4988]: I1008 21:01:00.387421 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/8d60984c-fb2d-425a-b6e5-87d20c7c0cc5-fernet-keys\") pod \"keystone-cron-29332621-swt82\" (UID: \"8d60984c-fb2d-425a-b6e5-87d20c7c0cc5\") " pod="openstack/keystone-cron-29332621-swt82" Oct 08 21:01:00 crc kubenswrapper[4988]: I1008 21:01:00.387781 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8d60984c-fb2d-425a-b6e5-87d20c7c0cc5-combined-ca-bundle\") pod \"keystone-cron-29332621-swt82\" (UID: \"8d60984c-fb2d-425a-b6e5-87d20c7c0cc5\") " pod="openstack/keystone-cron-29332621-swt82" Oct 08 21:01:00 crc kubenswrapper[4988]: I1008 21:01:00.387910 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8d60984c-fb2d-425a-b6e5-87d20c7c0cc5-config-data\") pod \"keystone-cron-29332621-swt82\" (UID: \"8d60984c-fb2d-425a-b6e5-87d20c7c0cc5\") " pod="openstack/keystone-cron-29332621-swt82" Oct 08 21:01:00 crc kubenswrapper[4988]: I1008 21:01:00.388266 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tlxqg\" (UniqueName: \"kubernetes.io/projected/8d60984c-fb2d-425a-b6e5-87d20c7c0cc5-kube-api-access-tlxqg\") pod \"keystone-cron-29332621-swt82\" (UID: \"8d60984c-fb2d-425a-b6e5-87d20c7c0cc5\") " pod="openstack/keystone-cron-29332621-swt82" Oct 08 21:01:00 crc kubenswrapper[4988]: I1008 21:01:00.396753 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8d60984c-fb2d-425a-b6e5-87d20c7c0cc5-combined-ca-bundle\") pod \"keystone-cron-29332621-swt82\" (UID: \"8d60984c-fb2d-425a-b6e5-87d20c7c0cc5\") " pod="openstack/keystone-cron-29332621-swt82" Oct 08 21:01:00 crc kubenswrapper[4988]: I1008 21:01:00.399195 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8d60984c-fb2d-425a-b6e5-87d20c7c0cc5-config-data\") pod \"keystone-cron-29332621-swt82\" (UID: \"8d60984c-fb2d-425a-b6e5-87d20c7c0cc5\") " pod="openstack/keystone-cron-29332621-swt82" Oct 08 21:01:00 crc kubenswrapper[4988]: I1008 21:01:00.399745 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/8d60984c-fb2d-425a-b6e5-87d20c7c0cc5-fernet-keys\") pod \"keystone-cron-29332621-swt82\" (UID: \"8d60984c-fb2d-425a-b6e5-87d20c7c0cc5\") " pod="openstack/keystone-cron-29332621-swt82" Oct 08 21:01:00 crc kubenswrapper[4988]: I1008 21:01:00.415320 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tlxqg\" (UniqueName: \"kubernetes.io/projected/8d60984c-fb2d-425a-b6e5-87d20c7c0cc5-kube-api-access-tlxqg\") pod \"keystone-cron-29332621-swt82\" (UID: \"8d60984c-fb2d-425a-b6e5-87d20c7c0cc5\") " pod="openstack/keystone-cron-29332621-swt82" Oct 08 21:01:00 crc kubenswrapper[4988]: I1008 21:01:00.520830 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29332621-swt82" Oct 08 21:01:01 crc kubenswrapper[4988]: I1008 21:01:01.038376 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29332621-swt82"] Oct 08 21:01:01 crc kubenswrapper[4988]: W1008 21:01:01.887322 4988 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8d60984c_fb2d_425a_b6e5_87d20c7c0cc5.slice/crio-f0e435ec515ab3d0265b78f195a83917ff345444c94c084b6dd758ad00c2cbf5 WatchSource:0}: Error finding container f0e435ec515ab3d0265b78f195a83917ff345444c94c084b6dd758ad00c2cbf5: Status 404 returned error can't find the container with id f0e435ec515ab3d0265b78f195a83917ff345444c94c084b6dd758ad00c2cbf5 Oct 08 21:01:02 crc kubenswrapper[4988]: I1008 21:01:02.765259 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29332621-swt82" event={"ID":"8d60984c-fb2d-425a-b6e5-87d20c7c0cc5","Type":"ContainerStarted","Data":"5640204417d6d1ed4eeb493e01b2a58589d3408801e55b594744c00487548644"} Oct 08 21:01:02 crc kubenswrapper[4988]: I1008 21:01:02.765792 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29332621-swt82" event={"ID":"8d60984c-fb2d-425a-b6e5-87d20c7c0cc5","Type":"ContainerStarted","Data":"f0e435ec515ab3d0265b78f195a83917ff345444c94c084b6dd758ad00c2cbf5"} Oct 08 21:01:02 crc kubenswrapper[4988]: I1008 21:01:02.794101 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-cron-29332621-swt82" podStartSLOduration=2.794074002 podStartE2EDuration="2.794074002s" podCreationTimestamp="2025-10-08 21:01:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 21:01:02.788260246 +0000 UTC m=+10208.238103046" watchObservedRunningTime="2025-10-08 21:01:02.794074002 +0000 UTC m=+10208.243916782" Oct 08 21:01:05 crc kubenswrapper[4988]: I1008 21:01:05.811887 4988 generic.go:334] "Generic (PLEG): container finished" podID="8d60984c-fb2d-425a-b6e5-87d20c7c0cc5" containerID="5640204417d6d1ed4eeb493e01b2a58589d3408801e55b594744c00487548644" exitCode=0 Oct 08 21:01:05 crc kubenswrapper[4988]: I1008 21:01:05.812003 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29332621-swt82" event={"ID":"8d60984c-fb2d-425a-b6e5-87d20c7c0cc5","Type":"ContainerDied","Data":"5640204417d6d1ed4eeb493e01b2a58589d3408801e55b594744c00487548644"} Oct 08 21:01:07 crc kubenswrapper[4988]: I1008 21:01:07.247358 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29332621-swt82" Oct 08 21:01:07 crc kubenswrapper[4988]: I1008 21:01:07.374817 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/8d60984c-fb2d-425a-b6e5-87d20c7c0cc5-fernet-keys\") pod \"8d60984c-fb2d-425a-b6e5-87d20c7c0cc5\" (UID: \"8d60984c-fb2d-425a-b6e5-87d20c7c0cc5\") " Oct 08 21:01:07 crc kubenswrapper[4988]: I1008 21:01:07.374995 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8d60984c-fb2d-425a-b6e5-87d20c7c0cc5-config-data\") pod \"8d60984c-fb2d-425a-b6e5-87d20c7c0cc5\" (UID: \"8d60984c-fb2d-425a-b6e5-87d20c7c0cc5\") " Oct 08 21:01:07 crc kubenswrapper[4988]: I1008 21:01:07.375171 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tlxqg\" (UniqueName: \"kubernetes.io/projected/8d60984c-fb2d-425a-b6e5-87d20c7c0cc5-kube-api-access-tlxqg\") pod \"8d60984c-fb2d-425a-b6e5-87d20c7c0cc5\" (UID: \"8d60984c-fb2d-425a-b6e5-87d20c7c0cc5\") " Oct 08 21:01:07 crc kubenswrapper[4988]: I1008 21:01:07.375276 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8d60984c-fb2d-425a-b6e5-87d20c7c0cc5-combined-ca-bundle\") pod \"8d60984c-fb2d-425a-b6e5-87d20c7c0cc5\" (UID: \"8d60984c-fb2d-425a-b6e5-87d20c7c0cc5\") " Oct 08 21:01:07 crc kubenswrapper[4988]: I1008 21:01:07.381214 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8d60984c-fb2d-425a-b6e5-87d20c7c0cc5-kube-api-access-tlxqg" (OuterVolumeSpecName: "kube-api-access-tlxqg") pod "8d60984c-fb2d-425a-b6e5-87d20c7c0cc5" (UID: "8d60984c-fb2d-425a-b6e5-87d20c7c0cc5"). InnerVolumeSpecName "kube-api-access-tlxqg". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 21:01:07 crc kubenswrapper[4988]: I1008 21:01:07.382449 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8d60984c-fb2d-425a-b6e5-87d20c7c0cc5-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "8d60984c-fb2d-425a-b6e5-87d20c7c0cc5" (UID: "8d60984c-fb2d-425a-b6e5-87d20c7c0cc5"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 21:01:07 crc kubenswrapper[4988]: I1008 21:01:07.410175 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8d60984c-fb2d-425a-b6e5-87d20c7c0cc5-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "8d60984c-fb2d-425a-b6e5-87d20c7c0cc5" (UID: "8d60984c-fb2d-425a-b6e5-87d20c7c0cc5"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 21:01:07 crc kubenswrapper[4988]: I1008 21:01:07.436517 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8d60984c-fb2d-425a-b6e5-87d20c7c0cc5-config-data" (OuterVolumeSpecName: "config-data") pod "8d60984c-fb2d-425a-b6e5-87d20c7c0cc5" (UID: "8d60984c-fb2d-425a-b6e5-87d20c7c0cc5"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 21:01:07 crc kubenswrapper[4988]: I1008 21:01:07.479611 4988 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8d60984c-fb2d-425a-b6e5-87d20c7c0cc5-config-data\") on node \"crc\" DevicePath \"\"" Oct 08 21:01:07 crc kubenswrapper[4988]: I1008 21:01:07.479657 4988 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tlxqg\" (UniqueName: \"kubernetes.io/projected/8d60984c-fb2d-425a-b6e5-87d20c7c0cc5-kube-api-access-tlxqg\") on node \"crc\" DevicePath \"\"" Oct 08 21:01:07 crc kubenswrapper[4988]: I1008 21:01:07.479675 4988 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8d60984c-fb2d-425a-b6e5-87d20c7c0cc5-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 21:01:07 crc kubenswrapper[4988]: I1008 21:01:07.479688 4988 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/8d60984c-fb2d-425a-b6e5-87d20c7c0cc5-fernet-keys\") on node \"crc\" DevicePath \"\"" Oct 08 21:01:07 crc kubenswrapper[4988]: I1008 21:01:07.841214 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29332621-swt82" event={"ID":"8d60984c-fb2d-425a-b6e5-87d20c7c0cc5","Type":"ContainerDied","Data":"f0e435ec515ab3d0265b78f195a83917ff345444c94c084b6dd758ad00c2cbf5"} Oct 08 21:01:07 crc kubenswrapper[4988]: I1008 21:01:07.841469 4988 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f0e435ec515ab3d0265b78f195a83917ff345444c94c084b6dd758ad00c2cbf5" Oct 08 21:01:07 crc kubenswrapper[4988]: I1008 21:01:07.841536 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29332621-swt82" Oct 08 21:01:46 crc kubenswrapper[4988]: I1008 21:01:46.398934 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-dddvr/must-gather-6bgk6"] Oct 08 21:01:46 crc kubenswrapper[4988]: E1008 21:01:46.399739 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8d60984c-fb2d-425a-b6e5-87d20c7c0cc5" containerName="keystone-cron" Oct 08 21:01:46 crc kubenswrapper[4988]: I1008 21:01:46.399751 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="8d60984c-fb2d-425a-b6e5-87d20c7c0cc5" containerName="keystone-cron" Oct 08 21:01:46 crc kubenswrapper[4988]: I1008 21:01:46.399939 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="8d60984c-fb2d-425a-b6e5-87d20c7c0cc5" containerName="keystone-cron" Oct 08 21:01:46 crc kubenswrapper[4988]: I1008 21:01:46.401023 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-dddvr/must-gather-6bgk6" Oct 08 21:01:46 crc kubenswrapper[4988]: I1008 21:01:46.404269 4988 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-dddvr"/"openshift-service-ca.crt" Oct 08 21:01:46 crc kubenswrapper[4988]: I1008 21:01:46.406712 4988 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-dddvr"/"kube-root-ca.crt" Oct 08 21:01:46 crc kubenswrapper[4988]: I1008 21:01:46.406736 4988 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-dddvr"/"default-dockercfg-krwtm" Oct 08 21:01:46 crc kubenswrapper[4988]: I1008 21:01:46.415843 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-dddvr/must-gather-6bgk6"] Oct 08 21:01:46 crc kubenswrapper[4988]: I1008 21:01:46.545254 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-687zg\" (UniqueName: \"kubernetes.io/projected/d157ad3b-3365-4413-a0b0-1da5184fc8ec-kube-api-access-687zg\") pod \"must-gather-6bgk6\" (UID: \"d157ad3b-3365-4413-a0b0-1da5184fc8ec\") " pod="openshift-must-gather-dddvr/must-gather-6bgk6" Oct 08 21:01:46 crc kubenswrapper[4988]: I1008 21:01:46.545305 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/d157ad3b-3365-4413-a0b0-1da5184fc8ec-must-gather-output\") pod \"must-gather-6bgk6\" (UID: \"d157ad3b-3365-4413-a0b0-1da5184fc8ec\") " pod="openshift-must-gather-dddvr/must-gather-6bgk6" Oct 08 21:01:46 crc kubenswrapper[4988]: I1008 21:01:46.647489 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-687zg\" (UniqueName: \"kubernetes.io/projected/d157ad3b-3365-4413-a0b0-1da5184fc8ec-kube-api-access-687zg\") pod \"must-gather-6bgk6\" (UID: \"d157ad3b-3365-4413-a0b0-1da5184fc8ec\") " pod="openshift-must-gather-dddvr/must-gather-6bgk6" Oct 08 21:01:46 crc kubenswrapper[4988]: I1008 21:01:46.647562 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/d157ad3b-3365-4413-a0b0-1da5184fc8ec-must-gather-output\") pod \"must-gather-6bgk6\" (UID: \"d157ad3b-3365-4413-a0b0-1da5184fc8ec\") " pod="openshift-must-gather-dddvr/must-gather-6bgk6" Oct 08 21:01:46 crc kubenswrapper[4988]: I1008 21:01:46.648034 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/d157ad3b-3365-4413-a0b0-1da5184fc8ec-must-gather-output\") pod \"must-gather-6bgk6\" (UID: \"d157ad3b-3365-4413-a0b0-1da5184fc8ec\") " pod="openshift-must-gather-dddvr/must-gather-6bgk6" Oct 08 21:01:46 crc kubenswrapper[4988]: I1008 21:01:46.667497 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-687zg\" (UniqueName: \"kubernetes.io/projected/d157ad3b-3365-4413-a0b0-1da5184fc8ec-kube-api-access-687zg\") pod \"must-gather-6bgk6\" (UID: \"d157ad3b-3365-4413-a0b0-1da5184fc8ec\") " pod="openshift-must-gather-dddvr/must-gather-6bgk6" Oct 08 21:01:46 crc kubenswrapper[4988]: I1008 21:01:46.728069 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-dddvr/must-gather-6bgk6" Oct 08 21:01:47 crc kubenswrapper[4988]: I1008 21:01:47.232501 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-dddvr/must-gather-6bgk6"] Oct 08 21:01:47 crc kubenswrapper[4988]: I1008 21:01:47.294785 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-dddvr/must-gather-6bgk6" event={"ID":"d157ad3b-3365-4413-a0b0-1da5184fc8ec","Type":"ContainerStarted","Data":"53f32dfa4b8cac032efa8e78bbcac16b8b4af1d52bcfc96d6c370c4074379587"} Oct 08 21:01:52 crc kubenswrapper[4988]: I1008 21:01:52.353145 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-dddvr/must-gather-6bgk6" event={"ID":"d157ad3b-3365-4413-a0b0-1da5184fc8ec","Type":"ContainerStarted","Data":"3f0bb832ec6b9f8cd5f93ef06d7767f3fdefb7827677752f6aeb98d3ca730b3d"} Oct 08 21:01:52 crc kubenswrapper[4988]: I1008 21:01:52.353717 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-dddvr/must-gather-6bgk6" event={"ID":"d157ad3b-3365-4413-a0b0-1da5184fc8ec","Type":"ContainerStarted","Data":"27b0386a1f9f60257d0cd7d5e4d6e679c50af8dee65d0626b7ecb0c5d2fd29fb"} Oct 08 21:01:52 crc kubenswrapper[4988]: I1008 21:01:52.394163 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-dddvr/must-gather-6bgk6" podStartSLOduration=2.205425299 podStartE2EDuration="6.394134252s" podCreationTimestamp="2025-10-08 21:01:46 +0000 UTC" firstStartedPulling="2025-10-08 21:01:47.239595248 +0000 UTC m=+10252.689438018" lastFinishedPulling="2025-10-08 21:01:51.428304201 +0000 UTC m=+10256.878146971" observedRunningTime="2025-10-08 21:01:52.381484468 +0000 UTC m=+10257.831327268" watchObservedRunningTime="2025-10-08 21:01:52.394134252 +0000 UTC m=+10257.843977062" Oct 08 21:01:55 crc kubenswrapper[4988]: I1008 21:01:55.343013 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-dddvr/crc-debug-mk6ql"] Oct 08 21:01:55 crc kubenswrapper[4988]: I1008 21:01:55.345015 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-dddvr/crc-debug-mk6ql" Oct 08 21:01:55 crc kubenswrapper[4988]: I1008 21:01:55.374086 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/2f5ac0db-5840-4b3c-b23a-5b12e1938538-host\") pod \"crc-debug-mk6ql\" (UID: \"2f5ac0db-5840-4b3c-b23a-5b12e1938538\") " pod="openshift-must-gather-dddvr/crc-debug-mk6ql" Oct 08 21:01:55 crc kubenswrapper[4988]: I1008 21:01:55.374149 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m4sr6\" (UniqueName: \"kubernetes.io/projected/2f5ac0db-5840-4b3c-b23a-5b12e1938538-kube-api-access-m4sr6\") pod \"crc-debug-mk6ql\" (UID: \"2f5ac0db-5840-4b3c-b23a-5b12e1938538\") " pod="openshift-must-gather-dddvr/crc-debug-mk6ql" Oct 08 21:01:55 crc kubenswrapper[4988]: I1008 21:01:55.476423 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/2f5ac0db-5840-4b3c-b23a-5b12e1938538-host\") pod \"crc-debug-mk6ql\" (UID: \"2f5ac0db-5840-4b3c-b23a-5b12e1938538\") " pod="openshift-must-gather-dddvr/crc-debug-mk6ql" Oct 08 21:01:55 crc kubenswrapper[4988]: I1008 21:01:55.476517 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m4sr6\" (UniqueName: \"kubernetes.io/projected/2f5ac0db-5840-4b3c-b23a-5b12e1938538-kube-api-access-m4sr6\") pod \"crc-debug-mk6ql\" (UID: \"2f5ac0db-5840-4b3c-b23a-5b12e1938538\") " pod="openshift-must-gather-dddvr/crc-debug-mk6ql" Oct 08 21:01:55 crc kubenswrapper[4988]: I1008 21:01:55.477017 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/2f5ac0db-5840-4b3c-b23a-5b12e1938538-host\") pod \"crc-debug-mk6ql\" (UID: \"2f5ac0db-5840-4b3c-b23a-5b12e1938538\") " pod="openshift-must-gather-dddvr/crc-debug-mk6ql" Oct 08 21:01:55 crc kubenswrapper[4988]: I1008 21:01:55.513186 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m4sr6\" (UniqueName: \"kubernetes.io/projected/2f5ac0db-5840-4b3c-b23a-5b12e1938538-kube-api-access-m4sr6\") pod \"crc-debug-mk6ql\" (UID: \"2f5ac0db-5840-4b3c-b23a-5b12e1938538\") " pod="openshift-must-gather-dddvr/crc-debug-mk6ql" Oct 08 21:01:55 crc kubenswrapper[4988]: I1008 21:01:55.663066 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-dddvr/crc-debug-mk6ql" Oct 08 21:01:56 crc kubenswrapper[4988]: I1008 21:01:56.397597 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-dddvr/crc-debug-mk6ql" event={"ID":"2f5ac0db-5840-4b3c-b23a-5b12e1938538","Type":"ContainerStarted","Data":"0508d0be2d1223dae5e3e0d0d5ed48c63b15032dad686257d19fdafc42806ecb"} Oct 08 21:02:10 crc kubenswrapper[4988]: I1008 21:02:10.563730 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-dddvr/crc-debug-mk6ql" event={"ID":"2f5ac0db-5840-4b3c-b23a-5b12e1938538","Type":"ContainerStarted","Data":"ae91a39e56ba749ee033286573db40ffb8cd3d81b67d2675f3c8bf50e3d6e5f6"} Oct 08 21:02:10 crc kubenswrapper[4988]: I1008 21:02:10.585687 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-dddvr/crc-debug-mk6ql" podStartSLOduration=2.005575444 podStartE2EDuration="15.585636345s" podCreationTimestamp="2025-10-08 21:01:55 +0000 UTC" firstStartedPulling="2025-10-08 21:01:55.797169374 +0000 UTC m=+10261.247012144" lastFinishedPulling="2025-10-08 21:02:09.377230275 +0000 UTC m=+10274.827073045" observedRunningTime="2025-10-08 21:02:10.581798223 +0000 UTC m=+10276.031640993" watchObservedRunningTime="2025-10-08 21:02:10.585636345 +0000 UTC m=+10276.035479145" Oct 08 21:02:47 crc kubenswrapper[4988]: I1008 21:02:47.968639 4988 generic.go:334] "Generic (PLEG): container finished" podID="2f5ac0db-5840-4b3c-b23a-5b12e1938538" containerID="ae91a39e56ba749ee033286573db40ffb8cd3d81b67d2675f3c8bf50e3d6e5f6" exitCode=0 Oct 08 21:02:47 crc kubenswrapper[4988]: I1008 21:02:47.968703 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-dddvr/crc-debug-mk6ql" event={"ID":"2f5ac0db-5840-4b3c-b23a-5b12e1938538","Type":"ContainerDied","Data":"ae91a39e56ba749ee033286573db40ffb8cd3d81b67d2675f3c8bf50e3d6e5f6"} Oct 08 21:02:49 crc kubenswrapper[4988]: I1008 21:02:49.127515 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-dddvr/crc-debug-mk6ql" Oct 08 21:02:49 crc kubenswrapper[4988]: I1008 21:02:49.177510 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-dddvr/crc-debug-mk6ql"] Oct 08 21:02:49 crc kubenswrapper[4988]: I1008 21:02:49.189001 4988 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-dddvr/crc-debug-mk6ql"] Oct 08 21:02:49 crc kubenswrapper[4988]: I1008 21:02:49.223125 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m4sr6\" (UniqueName: \"kubernetes.io/projected/2f5ac0db-5840-4b3c-b23a-5b12e1938538-kube-api-access-m4sr6\") pod \"2f5ac0db-5840-4b3c-b23a-5b12e1938538\" (UID: \"2f5ac0db-5840-4b3c-b23a-5b12e1938538\") " Oct 08 21:02:49 crc kubenswrapper[4988]: I1008 21:02:49.223256 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/2f5ac0db-5840-4b3c-b23a-5b12e1938538-host\") pod \"2f5ac0db-5840-4b3c-b23a-5b12e1938538\" (UID: \"2f5ac0db-5840-4b3c-b23a-5b12e1938538\") " Oct 08 21:02:49 crc kubenswrapper[4988]: I1008 21:02:49.223420 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/2f5ac0db-5840-4b3c-b23a-5b12e1938538-host" (OuterVolumeSpecName: "host") pod "2f5ac0db-5840-4b3c-b23a-5b12e1938538" (UID: "2f5ac0db-5840-4b3c-b23a-5b12e1938538"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 08 21:02:49 crc kubenswrapper[4988]: I1008 21:02:49.223949 4988 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/2f5ac0db-5840-4b3c-b23a-5b12e1938538-host\") on node \"crc\" DevicePath \"\"" Oct 08 21:02:49 crc kubenswrapper[4988]: I1008 21:02:49.236666 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2f5ac0db-5840-4b3c-b23a-5b12e1938538-kube-api-access-m4sr6" (OuterVolumeSpecName: "kube-api-access-m4sr6") pod "2f5ac0db-5840-4b3c-b23a-5b12e1938538" (UID: "2f5ac0db-5840-4b3c-b23a-5b12e1938538"). InnerVolumeSpecName "kube-api-access-m4sr6". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 21:02:49 crc kubenswrapper[4988]: I1008 21:02:49.265162 4988 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2f5ac0db-5840-4b3c-b23a-5b12e1938538" path="/var/lib/kubelet/pods/2f5ac0db-5840-4b3c-b23a-5b12e1938538/volumes" Oct 08 21:02:49 crc kubenswrapper[4988]: I1008 21:02:49.326701 4988 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-m4sr6\" (UniqueName: \"kubernetes.io/projected/2f5ac0db-5840-4b3c-b23a-5b12e1938538-kube-api-access-m4sr6\") on node \"crc\" DevicePath \"\"" Oct 08 21:02:50 crc kubenswrapper[4988]: I1008 21:02:50.002087 4988 scope.go:117] "RemoveContainer" containerID="ae91a39e56ba749ee033286573db40ffb8cd3d81b67d2675f3c8bf50e3d6e5f6" Oct 08 21:02:50 crc kubenswrapper[4988]: I1008 21:02:50.002540 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-dddvr/crc-debug-mk6ql" Oct 08 21:02:50 crc kubenswrapper[4988]: I1008 21:02:50.439129 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-dddvr/crc-debug-94mws"] Oct 08 21:02:50 crc kubenswrapper[4988]: E1008 21:02:50.439682 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2f5ac0db-5840-4b3c-b23a-5b12e1938538" containerName="container-00" Oct 08 21:02:50 crc kubenswrapper[4988]: I1008 21:02:50.439698 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="2f5ac0db-5840-4b3c-b23a-5b12e1938538" containerName="container-00" Oct 08 21:02:50 crc kubenswrapper[4988]: I1008 21:02:50.440333 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="2f5ac0db-5840-4b3c-b23a-5b12e1938538" containerName="container-00" Oct 08 21:02:50 crc kubenswrapper[4988]: I1008 21:02:50.441267 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-dddvr/crc-debug-94mws" Oct 08 21:02:50 crc kubenswrapper[4988]: I1008 21:02:50.458728 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b7mzx\" (UniqueName: \"kubernetes.io/projected/2d9f96b5-9e5b-4b95-b96a-80dcd0ad8a37-kube-api-access-b7mzx\") pod \"crc-debug-94mws\" (UID: \"2d9f96b5-9e5b-4b95-b96a-80dcd0ad8a37\") " pod="openshift-must-gather-dddvr/crc-debug-94mws" Oct 08 21:02:50 crc kubenswrapper[4988]: I1008 21:02:50.459038 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/2d9f96b5-9e5b-4b95-b96a-80dcd0ad8a37-host\") pod \"crc-debug-94mws\" (UID: \"2d9f96b5-9e5b-4b95-b96a-80dcd0ad8a37\") " pod="openshift-must-gather-dddvr/crc-debug-94mws" Oct 08 21:02:50 crc kubenswrapper[4988]: I1008 21:02:50.560223 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b7mzx\" (UniqueName: \"kubernetes.io/projected/2d9f96b5-9e5b-4b95-b96a-80dcd0ad8a37-kube-api-access-b7mzx\") pod \"crc-debug-94mws\" (UID: \"2d9f96b5-9e5b-4b95-b96a-80dcd0ad8a37\") " pod="openshift-must-gather-dddvr/crc-debug-94mws" Oct 08 21:02:50 crc kubenswrapper[4988]: I1008 21:02:50.560438 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/2d9f96b5-9e5b-4b95-b96a-80dcd0ad8a37-host\") pod \"crc-debug-94mws\" (UID: \"2d9f96b5-9e5b-4b95-b96a-80dcd0ad8a37\") " pod="openshift-must-gather-dddvr/crc-debug-94mws" Oct 08 21:02:50 crc kubenswrapper[4988]: I1008 21:02:50.560638 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/2d9f96b5-9e5b-4b95-b96a-80dcd0ad8a37-host\") pod \"crc-debug-94mws\" (UID: \"2d9f96b5-9e5b-4b95-b96a-80dcd0ad8a37\") " pod="openshift-must-gather-dddvr/crc-debug-94mws" Oct 08 21:02:50 crc kubenswrapper[4988]: I1008 21:02:50.586831 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b7mzx\" (UniqueName: \"kubernetes.io/projected/2d9f96b5-9e5b-4b95-b96a-80dcd0ad8a37-kube-api-access-b7mzx\") pod \"crc-debug-94mws\" (UID: \"2d9f96b5-9e5b-4b95-b96a-80dcd0ad8a37\") " pod="openshift-must-gather-dddvr/crc-debug-94mws" Oct 08 21:02:50 crc kubenswrapper[4988]: I1008 21:02:50.793539 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-dddvr/crc-debug-94mws" Oct 08 21:02:51 crc kubenswrapper[4988]: I1008 21:02:51.019412 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-dddvr/crc-debug-94mws" event={"ID":"2d9f96b5-9e5b-4b95-b96a-80dcd0ad8a37","Type":"ContainerStarted","Data":"f2e1fc079c59e23c7bbf7a1b803cb28cdd333214be29eb2fcc002058a5258ef2"} Oct 08 21:02:52 crc kubenswrapper[4988]: I1008 21:02:52.032715 4988 generic.go:334] "Generic (PLEG): container finished" podID="2d9f96b5-9e5b-4b95-b96a-80dcd0ad8a37" containerID="3d5d8abc5ea99d0367ffeaee65ef62a4656b2b0bb1b2d5a9df044379aa58d0eb" exitCode=0 Oct 08 21:02:52 crc kubenswrapper[4988]: I1008 21:02:52.032791 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-dddvr/crc-debug-94mws" event={"ID":"2d9f96b5-9e5b-4b95-b96a-80dcd0ad8a37","Type":"ContainerDied","Data":"3d5d8abc5ea99d0367ffeaee65ef62a4656b2b0bb1b2d5a9df044379aa58d0eb"} Oct 08 21:02:52 crc kubenswrapper[4988]: I1008 21:02:52.492352 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-dddvr/crc-debug-94mws"] Oct 08 21:02:52 crc kubenswrapper[4988]: I1008 21:02:52.500947 4988 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-dddvr/crc-debug-94mws"] Oct 08 21:02:53 crc kubenswrapper[4988]: I1008 21:02:53.153988 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-dddvr/crc-debug-94mws" Oct 08 21:02:53 crc kubenswrapper[4988]: I1008 21:02:53.243940 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/2d9f96b5-9e5b-4b95-b96a-80dcd0ad8a37-host\") pod \"2d9f96b5-9e5b-4b95-b96a-80dcd0ad8a37\" (UID: \"2d9f96b5-9e5b-4b95-b96a-80dcd0ad8a37\") " Oct 08 21:02:53 crc kubenswrapper[4988]: I1008 21:02:53.244088 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/2d9f96b5-9e5b-4b95-b96a-80dcd0ad8a37-host" (OuterVolumeSpecName: "host") pod "2d9f96b5-9e5b-4b95-b96a-80dcd0ad8a37" (UID: "2d9f96b5-9e5b-4b95-b96a-80dcd0ad8a37"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 08 21:02:53 crc kubenswrapper[4988]: I1008 21:02:53.244173 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b7mzx\" (UniqueName: \"kubernetes.io/projected/2d9f96b5-9e5b-4b95-b96a-80dcd0ad8a37-kube-api-access-b7mzx\") pod \"2d9f96b5-9e5b-4b95-b96a-80dcd0ad8a37\" (UID: \"2d9f96b5-9e5b-4b95-b96a-80dcd0ad8a37\") " Oct 08 21:02:53 crc kubenswrapper[4988]: I1008 21:02:53.245716 4988 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/2d9f96b5-9e5b-4b95-b96a-80dcd0ad8a37-host\") on node \"crc\" DevicePath \"\"" Oct 08 21:02:53 crc kubenswrapper[4988]: I1008 21:02:53.267003 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2d9f96b5-9e5b-4b95-b96a-80dcd0ad8a37-kube-api-access-b7mzx" (OuterVolumeSpecName: "kube-api-access-b7mzx") pod "2d9f96b5-9e5b-4b95-b96a-80dcd0ad8a37" (UID: "2d9f96b5-9e5b-4b95-b96a-80dcd0ad8a37"). InnerVolumeSpecName "kube-api-access-b7mzx". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 21:02:53 crc kubenswrapper[4988]: I1008 21:02:53.272225 4988 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2d9f96b5-9e5b-4b95-b96a-80dcd0ad8a37" path="/var/lib/kubelet/pods/2d9f96b5-9e5b-4b95-b96a-80dcd0ad8a37/volumes" Oct 08 21:02:53 crc kubenswrapper[4988]: I1008 21:02:53.338112 4988 patch_prober.go:28] interesting pod/machine-config-daemon-nm54f container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 21:02:53 crc kubenswrapper[4988]: I1008 21:02:53.338183 4988 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 21:02:53 crc kubenswrapper[4988]: I1008 21:02:53.348787 4988 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b7mzx\" (UniqueName: \"kubernetes.io/projected/2d9f96b5-9e5b-4b95-b96a-80dcd0ad8a37-kube-api-access-b7mzx\") on node \"crc\" DevicePath \"\"" Oct 08 21:02:53 crc kubenswrapper[4988]: I1008 21:02:53.700756 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-dddvr/crc-debug-tmlj5"] Oct 08 21:02:53 crc kubenswrapper[4988]: E1008 21:02:53.701558 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2d9f96b5-9e5b-4b95-b96a-80dcd0ad8a37" containerName="container-00" Oct 08 21:02:53 crc kubenswrapper[4988]: I1008 21:02:53.701578 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="2d9f96b5-9e5b-4b95-b96a-80dcd0ad8a37" containerName="container-00" Oct 08 21:02:53 crc kubenswrapper[4988]: I1008 21:02:53.702074 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="2d9f96b5-9e5b-4b95-b96a-80dcd0ad8a37" containerName="container-00" Oct 08 21:02:53 crc kubenswrapper[4988]: I1008 21:02:53.702957 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-dddvr/crc-debug-tmlj5" Oct 08 21:02:53 crc kubenswrapper[4988]: I1008 21:02:53.864041 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tfnx4\" (UniqueName: \"kubernetes.io/projected/0e6e85a0-3db8-451b-b388-6d7afe8a2d15-kube-api-access-tfnx4\") pod \"crc-debug-tmlj5\" (UID: \"0e6e85a0-3db8-451b-b388-6d7afe8a2d15\") " pod="openshift-must-gather-dddvr/crc-debug-tmlj5" Oct 08 21:02:53 crc kubenswrapper[4988]: I1008 21:02:53.864186 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/0e6e85a0-3db8-451b-b388-6d7afe8a2d15-host\") pod \"crc-debug-tmlj5\" (UID: \"0e6e85a0-3db8-451b-b388-6d7afe8a2d15\") " pod="openshift-must-gather-dddvr/crc-debug-tmlj5" Oct 08 21:02:53 crc kubenswrapper[4988]: I1008 21:02:53.966004 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/0e6e85a0-3db8-451b-b388-6d7afe8a2d15-host\") pod \"crc-debug-tmlj5\" (UID: \"0e6e85a0-3db8-451b-b388-6d7afe8a2d15\") " pod="openshift-must-gather-dddvr/crc-debug-tmlj5" Oct 08 21:02:53 crc kubenswrapper[4988]: I1008 21:02:53.966145 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/0e6e85a0-3db8-451b-b388-6d7afe8a2d15-host\") pod \"crc-debug-tmlj5\" (UID: \"0e6e85a0-3db8-451b-b388-6d7afe8a2d15\") " pod="openshift-must-gather-dddvr/crc-debug-tmlj5" Oct 08 21:02:53 crc kubenswrapper[4988]: I1008 21:02:53.966187 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tfnx4\" (UniqueName: \"kubernetes.io/projected/0e6e85a0-3db8-451b-b388-6d7afe8a2d15-kube-api-access-tfnx4\") pod \"crc-debug-tmlj5\" (UID: \"0e6e85a0-3db8-451b-b388-6d7afe8a2d15\") " pod="openshift-must-gather-dddvr/crc-debug-tmlj5" Oct 08 21:02:54 crc kubenswrapper[4988]: I1008 21:02:54.001020 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tfnx4\" (UniqueName: \"kubernetes.io/projected/0e6e85a0-3db8-451b-b388-6d7afe8a2d15-kube-api-access-tfnx4\") pod \"crc-debug-tmlj5\" (UID: \"0e6e85a0-3db8-451b-b388-6d7afe8a2d15\") " pod="openshift-must-gather-dddvr/crc-debug-tmlj5" Oct 08 21:02:54 crc kubenswrapper[4988]: I1008 21:02:54.028176 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-dddvr/crc-debug-tmlj5" Oct 08 21:02:54 crc kubenswrapper[4988]: I1008 21:02:54.061201 4988 scope.go:117] "RemoveContainer" containerID="3d5d8abc5ea99d0367ffeaee65ef62a4656b2b0bb1b2d5a9df044379aa58d0eb" Oct 08 21:02:54 crc kubenswrapper[4988]: I1008 21:02:54.061242 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-dddvr/crc-debug-94mws" Oct 08 21:02:54 crc kubenswrapper[4988]: W1008 21:02:54.062025 4988 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0e6e85a0_3db8_451b_b388_6d7afe8a2d15.slice/crio-82eeb0485daab3a2cc162e8d78d15bc02d23944b4e75ef1bf35a05aab244f8aa WatchSource:0}: Error finding container 82eeb0485daab3a2cc162e8d78d15bc02d23944b4e75ef1bf35a05aab244f8aa: Status 404 returned error can't find the container with id 82eeb0485daab3a2cc162e8d78d15bc02d23944b4e75ef1bf35a05aab244f8aa Oct 08 21:02:55 crc kubenswrapper[4988]: I1008 21:02:55.076421 4988 generic.go:334] "Generic (PLEG): container finished" podID="0e6e85a0-3db8-451b-b388-6d7afe8a2d15" containerID="920df3cfd8465e647b2bbd3bb63fd3257f8e26c1623fcceb77bb474efd2dc1ec" exitCode=0 Oct 08 21:02:55 crc kubenswrapper[4988]: I1008 21:02:55.077056 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-dddvr/crc-debug-tmlj5" event={"ID":"0e6e85a0-3db8-451b-b388-6d7afe8a2d15","Type":"ContainerDied","Data":"920df3cfd8465e647b2bbd3bb63fd3257f8e26c1623fcceb77bb474efd2dc1ec"} Oct 08 21:02:55 crc kubenswrapper[4988]: I1008 21:02:55.077092 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-dddvr/crc-debug-tmlj5" event={"ID":"0e6e85a0-3db8-451b-b388-6d7afe8a2d15","Type":"ContainerStarted","Data":"82eeb0485daab3a2cc162e8d78d15bc02d23944b4e75ef1bf35a05aab244f8aa"} Oct 08 21:02:55 crc kubenswrapper[4988]: I1008 21:02:55.138727 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-dddvr/crc-debug-tmlj5"] Oct 08 21:02:55 crc kubenswrapper[4988]: I1008 21:02:55.149568 4988 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-dddvr/crc-debug-tmlj5"] Oct 08 21:02:56 crc kubenswrapper[4988]: I1008 21:02:56.240990 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-dddvr/crc-debug-tmlj5" Oct 08 21:02:56 crc kubenswrapper[4988]: I1008 21:02:56.423873 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tfnx4\" (UniqueName: \"kubernetes.io/projected/0e6e85a0-3db8-451b-b388-6d7afe8a2d15-kube-api-access-tfnx4\") pod \"0e6e85a0-3db8-451b-b388-6d7afe8a2d15\" (UID: \"0e6e85a0-3db8-451b-b388-6d7afe8a2d15\") " Oct 08 21:02:56 crc kubenswrapper[4988]: I1008 21:02:56.423955 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/0e6e85a0-3db8-451b-b388-6d7afe8a2d15-host\") pod \"0e6e85a0-3db8-451b-b388-6d7afe8a2d15\" (UID: \"0e6e85a0-3db8-451b-b388-6d7afe8a2d15\") " Oct 08 21:02:56 crc kubenswrapper[4988]: I1008 21:02:56.424087 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/0e6e85a0-3db8-451b-b388-6d7afe8a2d15-host" (OuterVolumeSpecName: "host") pod "0e6e85a0-3db8-451b-b388-6d7afe8a2d15" (UID: "0e6e85a0-3db8-451b-b388-6d7afe8a2d15"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 08 21:02:56 crc kubenswrapper[4988]: I1008 21:02:56.425135 4988 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/0e6e85a0-3db8-451b-b388-6d7afe8a2d15-host\") on node \"crc\" DevicePath \"\"" Oct 08 21:02:56 crc kubenswrapper[4988]: I1008 21:02:56.432289 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0e6e85a0-3db8-451b-b388-6d7afe8a2d15-kube-api-access-tfnx4" (OuterVolumeSpecName: "kube-api-access-tfnx4") pod "0e6e85a0-3db8-451b-b388-6d7afe8a2d15" (UID: "0e6e85a0-3db8-451b-b388-6d7afe8a2d15"). InnerVolumeSpecName "kube-api-access-tfnx4". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 21:02:56 crc kubenswrapper[4988]: I1008 21:02:56.527504 4988 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tfnx4\" (UniqueName: \"kubernetes.io/projected/0e6e85a0-3db8-451b-b388-6d7afe8a2d15-kube-api-access-tfnx4\") on node \"crc\" DevicePath \"\"" Oct 08 21:02:57 crc kubenswrapper[4988]: I1008 21:02:57.101308 4988 scope.go:117] "RemoveContainer" containerID="920df3cfd8465e647b2bbd3bb63fd3257f8e26c1623fcceb77bb474efd2dc1ec" Oct 08 21:02:57 crc kubenswrapper[4988]: I1008 21:02:57.101522 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-dddvr/crc-debug-tmlj5" Oct 08 21:02:57 crc kubenswrapper[4988]: I1008 21:02:57.250274 4988 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0e6e85a0-3db8-451b-b388-6d7afe8a2d15" path="/var/lib/kubelet/pods/0e6e85a0-3db8-451b-b388-6d7afe8a2d15/volumes" Oct 08 21:03:23 crc kubenswrapper[4988]: I1008 21:03:23.337798 4988 patch_prober.go:28] interesting pod/machine-config-daemon-nm54f container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 21:03:23 crc kubenswrapper[4988]: I1008 21:03:23.338253 4988 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 21:03:24 crc kubenswrapper[4988]: I1008 21:03:24.193473 4988 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_alertmanager-metric-storage-0_64157199-c32a-461e-9a59-7d5a782c4838/init-config-reloader/0.log" Oct 08 21:03:24 crc kubenswrapper[4988]: I1008 21:03:24.425669 4988 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_alertmanager-metric-storage-0_64157199-c32a-461e-9a59-7d5a782c4838/alertmanager/0.log" Oct 08 21:03:24 crc kubenswrapper[4988]: I1008 21:03:24.434280 4988 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_alertmanager-metric-storage-0_64157199-c32a-461e-9a59-7d5a782c4838/init-config-reloader/0.log" Oct 08 21:03:24 crc kubenswrapper[4988]: I1008 21:03:24.435012 4988 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_alertmanager-metric-storage-0_64157199-c32a-461e-9a59-7d5a782c4838/config-reloader/0.log" Oct 08 21:03:24 crc kubenswrapper[4988]: I1008 21:03:24.610951 4988 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_aodh-0_78aab14f-ea46-4fdb-b1e4-8f8d6c3c54d6/aodh-api/0.log" Oct 08 21:03:24 crc kubenswrapper[4988]: I1008 21:03:24.701218 4988 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_aodh-0_78aab14f-ea46-4fdb-b1e4-8f8d6c3c54d6/aodh-evaluator/0.log" Oct 08 21:03:24 crc kubenswrapper[4988]: I1008 21:03:24.799022 4988 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_aodh-0_78aab14f-ea46-4fdb-b1e4-8f8d6c3c54d6/aodh-listener/0.log" Oct 08 21:03:24 crc kubenswrapper[4988]: I1008 21:03:24.877637 4988 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_aodh-0_78aab14f-ea46-4fdb-b1e4-8f8d6c3c54d6/aodh-notifier/0.log" Oct 08 21:03:25 crc kubenswrapper[4988]: I1008 21:03:25.043281 4988 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-5bb9747ff6-msjmt_ab612ec9-381c-4cb0-b608-10a1bb5768bd/barbican-api/0.log" Oct 08 21:03:25 crc kubenswrapper[4988]: I1008 21:03:25.079368 4988 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-5bb9747ff6-msjmt_ab612ec9-381c-4cb0-b608-10a1bb5768bd/barbican-api-log/0.log" Oct 08 21:03:25 crc kubenswrapper[4988]: I1008 21:03:25.275996 4988 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-75d59957f4-c7sp8_04727a82-9855-47c2-9df2-97b9e7a636c1/barbican-keystone-listener/0.log" Oct 08 21:03:25 crc kubenswrapper[4988]: I1008 21:03:25.329911 4988 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-75d59957f4-c7sp8_04727a82-9855-47c2-9df2-97b9e7a636c1/barbican-keystone-listener-log/0.log" Oct 08 21:03:25 crc kubenswrapper[4988]: I1008 21:03:25.503547 4988 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-6d89d4fcc5-rcwl2_6f247168-21ff-44e9-8e52-f49698e93a65/barbican-worker-log/0.log" Oct 08 21:03:25 crc kubenswrapper[4988]: I1008 21:03:25.547877 4988 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-6d89d4fcc5-rcwl2_6f247168-21ff-44e9-8e52-f49698e93a65/barbican-worker/0.log" Oct 08 21:03:25 crc kubenswrapper[4988]: I1008 21:03:25.826511 4988 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_bootstrap-openstack-openstack-cell1-t5z85_9e3f73e4-4a15-4ed9-a899-e7c4d6837b03/bootstrap-openstack-openstack-cell1/0.log" Oct 08 21:03:26 crc kubenswrapper[4988]: I1008 21:03:26.047377 4988 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_a7b11b95-f581-4db1-ac20-67200519b7bc/ceilometer-central-agent/0.log" Oct 08 21:03:26 crc kubenswrapper[4988]: I1008 21:03:26.165736 4988 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_a7b11b95-f581-4db1-ac20-67200519b7bc/ceilometer-notification-agent/0.log" Oct 08 21:03:26 crc kubenswrapper[4988]: I1008 21:03:26.183908 4988 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_a7b11b95-f581-4db1-ac20-67200519b7bc/proxy-httpd/0.log" Oct 08 21:03:26 crc kubenswrapper[4988]: I1008 21:03:26.352122 4988 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_a7b11b95-f581-4db1-ac20-67200519b7bc/sg-core/0.log" Oct 08 21:03:26 crc kubenswrapper[4988]: I1008 21:03:26.456411 4988 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_15e8111d-836a-428e-b58f-b0df3f96c251/cinder-api/0.log" Oct 08 21:03:26 crc kubenswrapper[4988]: I1008 21:03:26.543616 4988 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_15e8111d-836a-428e-b58f-b0df3f96c251/cinder-api-log/0.log" Oct 08 21:03:26 crc kubenswrapper[4988]: I1008 21:03:26.621269 4988 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_95310a64-9176-420b-aca3-e97e6c821eed/cinder-scheduler/0.log" Oct 08 21:03:26 crc kubenswrapper[4988]: I1008 21:03:26.792197 4988 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_95310a64-9176-420b-aca3-e97e6c821eed/probe/0.log" Oct 08 21:03:26 crc kubenswrapper[4988]: I1008 21:03:26.878983 4988 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-network-openstack-openstack-cell1-cpds6_a5ae4d27-798c-4d58-a9d9-ea145d68d6fc/configure-network-openstack-openstack-cell1/0.log" Oct 08 21:03:27 crc kubenswrapper[4988]: I1008 21:03:27.049740 4988 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-os-openstack-openstack-cell1-2k4qc_5fc3787f-07bb-4c4b-b97d-e3057a1107e0/configure-os-openstack-openstack-cell1/0.log" Oct 08 21:03:27 crc kubenswrapper[4988]: I1008 21:03:27.278463 4988 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-76dfc7c65-crgzj_9e2ebad6-add9-4567-8f39-8e51f614e50f/init/0.log" Oct 08 21:03:27 crc kubenswrapper[4988]: I1008 21:03:27.408064 4988 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-76dfc7c65-crgzj_9e2ebad6-add9-4567-8f39-8e51f614e50f/init/0.log" Oct 08 21:03:27 crc kubenswrapper[4988]: I1008 21:03:27.424209 4988 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-76dfc7c65-crgzj_9e2ebad6-add9-4567-8f39-8e51f614e50f/dnsmasq-dns/0.log" Oct 08 21:03:27 crc kubenswrapper[4988]: I1008 21:03:27.466612 4988 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_download-cache-openstack-openstack-cell1-fjdm2_5ab4ca0a-e3b3-418e-8640-13828852f6fb/download-cache-openstack-openstack-cell1/0.log" Oct 08 21:03:27 crc kubenswrapper[4988]: I1008 21:03:27.628296 4988 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_b6e8a219-5fcb-4640-8672-8278e9e628a7/glance-httpd/0.log" Oct 08 21:03:27 crc kubenswrapper[4988]: I1008 21:03:27.705396 4988 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_b6e8a219-5fcb-4640-8672-8278e9e628a7/glance-log/0.log" Oct 08 21:03:27 crc kubenswrapper[4988]: I1008 21:03:27.858292 4988 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_2b071964-bede-4593-bf58-b5deaa573d05/glance-httpd/0.log" Oct 08 21:03:27 crc kubenswrapper[4988]: I1008 21:03:27.891238 4988 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_2b071964-bede-4593-bf58-b5deaa573d05/glance-log/0.log" Oct 08 21:03:28 crc kubenswrapper[4988]: I1008 21:03:28.383575 4988 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_heat-engine-78cc4595db-p87kc_807b4691-54e6-4ae9-abeb-8e95ab06578a/heat-engine/0.log" Oct 08 21:03:28 crc kubenswrapper[4988]: I1008 21:03:28.596634 4988 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_heat-api-594c995686-vt22w_e988e6d2-d2c3-46b6-a6e4-e8524f3fbe8a/heat-api/0.log" Oct 08 21:03:28 crc kubenswrapper[4988]: I1008 21:03:28.678587 4988 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_heat-cfnapi-9587dcd9d-cx57b_e6dd3357-9d84-4c92-8e40-69f19dd64a83/heat-cfnapi/0.log" Oct 08 21:03:28 crc kubenswrapper[4988]: I1008 21:03:28.682994 4988 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-64f8b4fb78-xcwgf_f084c3ca-3754-4e7e-82dd-f0f3029d7102/horizon/0.log" Oct 08 21:03:28 crc kubenswrapper[4988]: I1008 21:03:28.855270 4988 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-certs-openstack-openstack-cell1-n45qh_c6b7f3f9-1ffb-454d-8dc6-f697eac681ae/install-certs-openstack-openstack-cell1/0.log" Oct 08 21:03:29 crc kubenswrapper[4988]: I1008 21:03:29.059024 4988 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-os-openstack-openstack-cell1-pc97n_469f9f63-7654-468b-9ad8-c620f0c1b18d/install-os-openstack-openstack-cell1/0.log" Oct 08 21:03:29 crc kubenswrapper[4988]: I1008 21:03:29.381237 4988 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-7c6d678f8f-dxth8_a92eb88b-08c7-4f05-abdf-87777d7d7170/keystone-api/0.log" Oct 08 21:03:29 crc kubenswrapper[4988]: I1008 21:03:29.419040 4988 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-64f8b4fb78-xcwgf_f084c3ca-3754-4e7e-82dd-f0f3029d7102/horizon-log/0.log" Oct 08 21:03:29 crc kubenswrapper[4988]: I1008 21:03:29.434607 4988 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-cron-29332561-46rm8_9855c453-5d95-4cc9-ad10-b418ae54ff46/keystone-cron/0.log" Oct 08 21:03:29 crc kubenswrapper[4988]: I1008 21:03:29.542179 4988 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-cron-29332621-swt82_8d60984c-fb2d-425a-b6e5-87d20c7c0cc5/keystone-cron/0.log" Oct 08 21:03:29 crc kubenswrapper[4988]: I1008 21:03:29.606709 4988 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_kube-state-metrics-0_5e4fdf4f-092f-40a0-b38b-ad93f5bad528/kube-state-metrics/0.log" Oct 08 21:03:29 crc kubenswrapper[4988]: I1008 21:03:29.771012 4988 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_libvirt-openstack-openstack-cell1-mzn6v_aef3b7a1-6628-4895-a380-153a230bd0a3/libvirt-openstack-openstack-cell1/0.log" Oct 08 21:03:30 crc kubenswrapper[4988]: I1008 21:03:30.040557 4988 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-77f8f8964c-wclf6_bc6e7082-3102-4dfb-af23-396d055e06a0/neutron-api/0.log" Oct 08 21:03:30 crc kubenswrapper[4988]: I1008 21:03:30.126631 4988 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-77f8f8964c-wclf6_bc6e7082-3102-4dfb-af23-396d055e06a0/neutron-httpd/0.log" Oct 08 21:03:30 crc kubenswrapper[4988]: I1008 21:03:30.291094 4988 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-dhcp-openstack-openstack-cell1-mbrs2_7a339659-4ba8-4fb5-99fa-760e2bef90a2/neutron-dhcp-openstack-openstack-cell1/0.log" Oct 08 21:03:30 crc kubenswrapper[4988]: I1008 21:03:30.556665 4988 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-metadata-openstack-openstack-cell1-b5mrs_dfca5e3a-fa8b-4a83-b6b8-f9694bbb8506/neutron-metadata-openstack-openstack-cell1/0.log" Oct 08 21:03:30 crc kubenswrapper[4988]: I1008 21:03:30.658094 4988 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-sriov-openstack-openstack-cell1-8vk4f_96659c9c-7996-4e91-81f2-3edbaed544d3/neutron-sriov-openstack-openstack-cell1/0.log" Oct 08 21:03:30 crc kubenswrapper[4988]: I1008 21:03:30.903993 4988 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_a881e1d1-a353-47e5-83be-b264f9552c26/nova-api-api/0.log" Oct 08 21:03:30 crc kubenswrapper[4988]: I1008 21:03:30.984645 4988 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_a881e1d1-a353-47e5-83be-b264f9552c26/nova-api-log/0.log" Oct 08 21:03:31 crc kubenswrapper[4988]: I1008 21:03:31.233122 4988 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell0-conductor-0_02f910a2-e03f-427d-b8be-e6ff9c3d6025/nova-cell0-conductor-conductor/0.log" Oct 08 21:03:31 crc kubenswrapper[4988]: I1008 21:03:31.414697 4988 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-conductor-0_115dfff7-cb2b-4638-b40a-28a11aa2aaf9/nova-cell1-conductor-conductor/0.log" Oct 08 21:03:31 crc kubenswrapper[4988]: I1008 21:03:31.513799 4988 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-novncproxy-0_48846c2d-55ec-4747-b3ab-4ed54cebc731/nova-cell1-novncproxy-novncproxy/0.log" Oct 08 21:03:31 crc kubenswrapper[4988]: I1008 21:03:31.871905 4988 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellgcg5p_a2f34e97-3b21-419b-883b-bdd5108b7dae/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell1/0.log" Oct 08 21:03:32 crc kubenswrapper[4988]: I1008 21:03:32.523135 4988 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-openstack-openstack-cell1-j9bg2_08584161-d03c-4354-9e07-60eae54b233d/nova-cell1-openstack-openstack-cell1/0.log" Oct 08 21:03:32 crc kubenswrapper[4988]: I1008 21:03:32.654253 4988 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_59722276-fba4-4321-9c14-696d0c1c98ce/nova-metadata-log/0.log" Oct 08 21:03:32 crc kubenswrapper[4988]: I1008 21:03:32.952952 4988 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-scheduler-0_88e44fb8-0012-4f95-9845-38d86ea84918/nova-scheduler-scheduler/0.log" Oct 08 21:03:33 crc kubenswrapper[4988]: I1008 21:03:33.096868 4988 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_59722276-fba4-4321-9c14-696d0c1c98ce/nova-metadata-metadata/0.log" Oct 08 21:03:33 crc kubenswrapper[4988]: I1008 21:03:33.173716 4988 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_octavia-api-764db59bb9-p78lt_a0c334ad-5ffa-41d9-8cdd-c5c50372137d/init/0.log" Oct 08 21:03:33 crc kubenswrapper[4988]: I1008 21:03:33.307570 4988 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_octavia-api-764db59bb9-p78lt_a0c334ad-5ffa-41d9-8cdd-c5c50372137d/init/0.log" Oct 08 21:03:33 crc kubenswrapper[4988]: I1008 21:03:33.387749 4988 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_octavia-api-764db59bb9-p78lt_a0c334ad-5ffa-41d9-8cdd-c5c50372137d/octavia-api-provider-agent/0.log" Oct 08 21:03:33 crc kubenswrapper[4988]: I1008 21:03:33.550722 4988 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_octavia-api-764db59bb9-p78lt_a0c334ad-5ffa-41d9-8cdd-c5c50372137d/octavia-api/0.log" Oct 08 21:03:34 crc kubenswrapper[4988]: I1008 21:03:34.112530 4988 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_octavia-healthmanager-hhkfh_d7a04c8d-1f4c-4dd8-98b6-7afdc47bece7/init/0.log" Oct 08 21:03:34 crc kubenswrapper[4988]: I1008 21:03:34.536258 4988 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_octavia-healthmanager-hhkfh_d7a04c8d-1f4c-4dd8-98b6-7afdc47bece7/init/0.log" Oct 08 21:03:34 crc kubenswrapper[4988]: I1008 21:03:34.667756 4988 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_octavia-healthmanager-hhkfh_d7a04c8d-1f4c-4dd8-98b6-7afdc47bece7/octavia-healthmanager/0.log" Oct 08 21:03:34 crc kubenswrapper[4988]: I1008 21:03:34.800041 4988 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_octavia-housekeeping-zwdsf_19ef63af-3cf2-40c9-9002-3a5536d518b9/init/0.log" Oct 08 21:03:34 crc kubenswrapper[4988]: I1008 21:03:34.946046 4988 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_octavia-housekeeping-zwdsf_19ef63af-3cf2-40c9-9002-3a5536d518b9/init/0.log" Oct 08 21:03:35 crc kubenswrapper[4988]: I1008 21:03:35.001906 4988 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_octavia-housekeeping-zwdsf_19ef63af-3cf2-40c9-9002-3a5536d518b9/octavia-housekeeping/0.log" Oct 08 21:03:35 crc kubenswrapper[4988]: I1008 21:03:35.172410 4988 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_octavia-image-upload-678599687f-2f6h5_68ad708f-db19-445d-a991-1d1a011a4a9f/init/0.log" Oct 08 21:03:35 crc kubenswrapper[4988]: I1008 21:03:35.329182 4988 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_octavia-image-upload-678599687f-2f6h5_68ad708f-db19-445d-a991-1d1a011a4a9f/init/0.log" Oct 08 21:03:35 crc kubenswrapper[4988]: I1008 21:03:35.383594 4988 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_octavia-image-upload-678599687f-2f6h5_68ad708f-db19-445d-a991-1d1a011a4a9f/octavia-amphora-httpd/0.log" Oct 08 21:03:35 crc kubenswrapper[4988]: I1008 21:03:35.519653 4988 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_octavia-rsyslog-xdhtp_2171d012-b738-48cf-a6e7-673d2996c676/init/0.log" Oct 08 21:03:35 crc kubenswrapper[4988]: I1008 21:03:35.718346 4988 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_octavia-rsyslog-xdhtp_2171d012-b738-48cf-a6e7-673d2996c676/init/0.log" Oct 08 21:03:35 crc kubenswrapper[4988]: I1008 21:03:35.815743 4988 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_octavia-rsyslog-xdhtp_2171d012-b738-48cf-a6e7-673d2996c676/octavia-rsyslog/0.log" Oct 08 21:03:35 crc kubenswrapper[4988]: I1008 21:03:35.974973 4988 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_octavia-worker-rtcdp_baa5ab97-2a79-4125-b1f8-5f0e7af21c6e/init/0.log" Oct 08 21:03:36 crc kubenswrapper[4988]: I1008 21:03:36.138326 4988 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_octavia-worker-rtcdp_baa5ab97-2a79-4125-b1f8-5f0e7af21c6e/init/0.log" Oct 08 21:03:36 crc kubenswrapper[4988]: I1008 21:03:36.293413 4988 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_octavia-worker-rtcdp_baa5ab97-2a79-4125-b1f8-5f0e7af21c6e/octavia-worker/0.log" Oct 08 21:03:36 crc kubenswrapper[4988]: I1008 21:03:36.345701 4988 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_bbe1c39b-8f0b-41a4-87a3-20672272af49/mysql-bootstrap/0.log" Oct 08 21:03:36 crc kubenswrapper[4988]: I1008 21:03:36.487423 4988 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_bbe1c39b-8f0b-41a4-87a3-20672272af49/mysql-bootstrap/0.log" Oct 08 21:03:36 crc kubenswrapper[4988]: I1008 21:03:36.511783 4988 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_bbe1c39b-8f0b-41a4-87a3-20672272af49/galera/0.log" Oct 08 21:03:36 crc kubenswrapper[4988]: I1008 21:03:36.711339 4988 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_689b2a7d-38cb-4476-a0ea-27bbd614fb2b/mysql-bootstrap/0.log" Oct 08 21:03:36 crc kubenswrapper[4988]: I1008 21:03:36.911663 4988 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_689b2a7d-38cb-4476-a0ea-27bbd614fb2b/galera/0.log" Oct 08 21:03:36 crc kubenswrapper[4988]: I1008 21:03:36.932256 4988 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_689b2a7d-38cb-4476-a0ea-27bbd614fb2b/mysql-bootstrap/0.log" Oct 08 21:03:37 crc kubenswrapper[4988]: I1008 21:03:37.114903 4988 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstackclient_efec0d27-7968-44f2-a1f4-733078747791/openstackclient/0.log" Oct 08 21:03:37 crc kubenswrapper[4988]: I1008 21:03:37.231922 4988 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-2xddw_a1187bea-e139-42e7-bd42-8086ed52bceb/ovn-controller/0.log" Oct 08 21:03:37 crc kubenswrapper[4988]: I1008 21:03:37.401443 4988 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-metrics-pdhvt_8c2854b9-1c17-48e4-82c7-84a4eb44f407/openstack-network-exporter/0.log" Oct 08 21:03:37 crc kubenswrapper[4988]: I1008 21:03:37.585683 4988 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-4667z_eb6a17a8-576e-4d61-8467-cca342654b2c/ovsdb-server-init/0.log" Oct 08 21:03:37 crc kubenswrapper[4988]: I1008 21:03:37.995362 4988 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-4667z_eb6a17a8-576e-4d61-8467-cca342654b2c/ovsdb-server-init/0.log" Oct 08 21:03:38 crc kubenswrapper[4988]: I1008 21:03:38.016891 4988 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-4667z_eb6a17a8-576e-4d61-8467-cca342654b2c/ovs-vswitchd/0.log" Oct 08 21:03:38 crc kubenswrapper[4988]: I1008 21:03:38.040710 4988 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-4667z_eb6a17a8-576e-4d61-8467-cca342654b2c/ovsdb-server/0.log" Oct 08 21:03:38 crc kubenswrapper[4988]: I1008 21:03:38.238608 4988 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_442d8773-cc55-4b79-945a-7d8c05431385/openstack-network-exporter/0.log" Oct 08 21:03:38 crc kubenswrapper[4988]: I1008 21:03:38.260728 4988 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_442d8773-cc55-4b79-945a-7d8c05431385/ovn-northd/0.log" Oct 08 21:03:38 crc kubenswrapper[4988]: I1008 21:03:38.496716 4988 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-openstack-openstack-cell1-5h2h6_de67964d-9527-467b-8ca4-364cdec72a43/ovn-openstack-openstack-cell1/0.log" Oct 08 21:03:38 crc kubenswrapper[4988]: I1008 21:03:38.694695 4988 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_e00c8be6-0ec3-4045-9a25-9f397cc78dc9/openstack-network-exporter/0.log" Oct 08 21:03:38 crc kubenswrapper[4988]: I1008 21:03:38.725814 4988 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_e00c8be6-0ec3-4045-9a25-9f397cc78dc9/ovsdbserver-nb/0.log" Oct 08 21:03:38 crc kubenswrapper[4988]: I1008 21:03:38.909698 4988 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-1_cbef3512-4ad7-45f0-9e87-d01ea0177708/openstack-network-exporter/0.log" Oct 08 21:03:38 crc kubenswrapper[4988]: I1008 21:03:38.924479 4988 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-1_cbef3512-4ad7-45f0-9e87-d01ea0177708/ovsdbserver-nb/0.log" Oct 08 21:03:39 crc kubenswrapper[4988]: I1008 21:03:39.116055 4988 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-2_40f9f201-93e1-4d27-842f-77040d48d28d/ovsdbserver-nb/0.log" Oct 08 21:03:39 crc kubenswrapper[4988]: I1008 21:03:39.217826 4988 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-2_40f9f201-93e1-4d27-842f-77040d48d28d/openstack-network-exporter/0.log" Oct 08 21:03:39 crc kubenswrapper[4988]: I1008 21:03:39.321890 4988 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_ba4c95b7-edad-454a-86a5-d2edb9a8908d/openstack-network-exporter/0.log" Oct 08 21:03:39 crc kubenswrapper[4988]: I1008 21:03:39.444479 4988 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_ba4c95b7-edad-454a-86a5-d2edb9a8908d/ovsdbserver-sb/0.log" Oct 08 21:03:39 crc kubenswrapper[4988]: I1008 21:03:39.584517 4988 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-1_dad7c6b4-98d7-490b-b17a-c4f13a12cd92/openstack-network-exporter/0.log" Oct 08 21:03:39 crc kubenswrapper[4988]: I1008 21:03:39.620971 4988 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-1_dad7c6b4-98d7-490b-b17a-c4f13a12cd92/ovsdbserver-sb/0.log" Oct 08 21:03:39 crc kubenswrapper[4988]: I1008 21:03:39.811166 4988 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-2_40504de9-4567-46fb-a598-6bbfb722b367/openstack-network-exporter/0.log" Oct 08 21:03:39 crc kubenswrapper[4988]: I1008 21:03:39.853425 4988 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-2_40504de9-4567-46fb-a598-6bbfb722b367/ovsdbserver-sb/0.log" Oct 08 21:03:40 crc kubenswrapper[4988]: I1008 21:03:40.037788 4988 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-766746c8cd-vsm49_dbb701b2-bfa0-4a83-b0db-91bd2d047e68/placement-api/0.log" Oct 08 21:03:40 crc kubenswrapper[4988]: I1008 21:03:40.192782 4988 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-766746c8cd-vsm49_dbb701b2-bfa0-4a83-b0db-91bd2d047e68/placement-log/0.log" Oct 08 21:03:40 crc kubenswrapper[4988]: I1008 21:03:40.266851 4988 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_pre-adoption-validation-openstack-pre-adoption-openstack-cpgsv4_93e6e1ae-4ef7-43f3-87a4-14a000232170/pre-adoption-validation-openstack-pre-adoption-openstack-cell1/0.log" Oct 08 21:03:40 crc kubenswrapper[4988]: I1008 21:03:40.423802 4988 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_9ee564e0-b7af-4905-9957-75cc0d0e50c5/init-config-reloader/0.log" Oct 08 21:03:40 crc kubenswrapper[4988]: I1008 21:03:40.689256 4988 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_9ee564e0-b7af-4905-9957-75cc0d0e50c5/init-config-reloader/0.log" Oct 08 21:03:40 crc kubenswrapper[4988]: I1008 21:03:40.732516 4988 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_9ee564e0-b7af-4905-9957-75cc0d0e50c5/config-reloader/0.log" Oct 08 21:03:40 crc kubenswrapper[4988]: I1008 21:03:40.745556 4988 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_9ee564e0-b7af-4905-9957-75cc0d0e50c5/prometheus/0.log" Oct 08 21:03:40 crc kubenswrapper[4988]: I1008 21:03:40.894262 4988 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_9ee564e0-b7af-4905-9957-75cc0d0e50c5/thanos-sidecar/0.log" Oct 08 21:03:40 crc kubenswrapper[4988]: I1008 21:03:40.948240 4988 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_16348dd8-8383-45c0-910a-fe3cf4abcab0/setup-container/0.log" Oct 08 21:03:41 crc kubenswrapper[4988]: I1008 21:03:41.152380 4988 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_16348dd8-8383-45c0-910a-fe3cf4abcab0/setup-container/0.log" Oct 08 21:03:41 crc kubenswrapper[4988]: I1008 21:03:41.203019 4988 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_16348dd8-8383-45c0-910a-fe3cf4abcab0/rabbitmq/0.log" Oct 08 21:03:41 crc kubenswrapper[4988]: I1008 21:03:41.348791 4988 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_9d6674d9-eacf-4d44-91db-5ace88fa9a01/setup-container/0.log" Oct 08 21:03:41 crc kubenswrapper[4988]: I1008 21:03:41.549228 4988 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_9d6674d9-eacf-4d44-91db-5ace88fa9a01/setup-container/0.log" Oct 08 21:03:41 crc kubenswrapper[4988]: I1008 21:03:41.666827 4988 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_9d6674d9-eacf-4d44-91db-5ace88fa9a01/rabbitmq/0.log" Oct 08 21:03:41 crc kubenswrapper[4988]: I1008 21:03:41.723666 4988 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_reboot-os-openstack-openstack-cell1-rxm8h_d4085938-9772-43b4-bd58-7044e2e3c6e0/reboot-os-openstack-openstack-cell1/0.log" Oct 08 21:03:41 crc kubenswrapper[4988]: I1008 21:03:41.845812 4988 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_run-os-openstack-openstack-cell1-2m44z_8f284658-1b89-4ace-8914-bfc709e071e5/run-os-openstack-openstack-cell1/0.log" Oct 08 21:03:42 crc kubenswrapper[4988]: I1008 21:03:42.260593 4988 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ssh-known-hosts-openstack-2qjln_b1cf5326-c7d5-4a5f-9a48-159b6482aa0a/ssh-known-hosts-openstack/0.log" Oct 08 21:03:42 crc kubenswrapper[4988]: I1008 21:03:42.626879 4988 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-b8548dfd6-qblss_679a7d0a-3ad1-45c8-b1ff-6772a862379a/proxy-server/0.log" Oct 08 21:03:42 crc kubenswrapper[4988]: I1008 21:03:42.882613 4988 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-b8548dfd6-qblss_679a7d0a-3ad1-45c8-b1ff-6772a862379a/proxy-httpd/0.log" Oct 08 21:03:42 crc kubenswrapper[4988]: I1008 21:03:42.899761 4988 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-ring-rebalance-t9tqb_de3434fc-c4fa-4495-b3f3-7f232decfed9/swift-ring-rebalance/0.log" Oct 08 21:03:43 crc kubenswrapper[4988]: I1008 21:03:43.163090 4988 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_telemetry-openstack-openstack-cell1-qmvh4_54d6a0c7-be73-43b6-b857-a621d1a1c794/telemetry-openstack-openstack-cell1/0.log" Oct 08 21:03:43 crc kubenswrapper[4988]: I1008 21:03:43.179223 4988 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_memcached-0_a44ce926-29f2-4851-aabf-65cc7ffd62bc/memcached/0.log" Oct 08 21:03:43 crc kubenswrapper[4988]: I1008 21:03:43.342956 4988 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_tripleo-cleanup-tripleo-cleanup-openstack-cell1-2xbjd_7dd8825a-ca2d-4361-a6e0-cf8a68ba8078/tripleo-cleanup-tripleo-cleanup-openstack-cell1/0.log" Oct 08 21:03:43 crc kubenswrapper[4988]: I1008 21:03:43.391885 4988 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_validate-network-openstack-openstack-cell1-c6dg8_848dd699-16fa-4254-b3cb-bc506dae3229/validate-network-openstack-openstack-cell1/0.log" Oct 08 21:03:53 crc kubenswrapper[4988]: I1008 21:03:53.338159 4988 patch_prober.go:28] interesting pod/machine-config-daemon-nm54f container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 21:03:53 crc kubenswrapper[4988]: I1008 21:03:53.338710 4988 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 21:03:53 crc kubenswrapper[4988]: I1008 21:03:53.338770 4988 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" Oct 08 21:03:53 crc kubenswrapper[4988]: I1008 21:03:53.339690 4988 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"6a30956f2db44d788d7c7cdfd2114a016aa50c17f8bee2b54970bedfe0df7317"} pod="openshift-machine-config-operator/machine-config-daemon-nm54f" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 08 21:03:53 crc kubenswrapper[4988]: I1008 21:03:53.339753 4988 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" containerName="machine-config-daemon" containerID="cri-o://6a30956f2db44d788d7c7cdfd2114a016aa50c17f8bee2b54970bedfe0df7317" gracePeriod=600 Oct 08 21:03:53 crc kubenswrapper[4988]: I1008 21:03:53.696351 4988 generic.go:334] "Generic (PLEG): container finished" podID="f328a20b-5283-407e-8a93-2ab2da599f18" containerID="6a30956f2db44d788d7c7cdfd2114a016aa50c17f8bee2b54970bedfe0df7317" exitCode=0 Oct 08 21:03:53 crc kubenswrapper[4988]: I1008 21:03:53.696872 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" event={"ID":"f328a20b-5283-407e-8a93-2ab2da599f18","Type":"ContainerDied","Data":"6a30956f2db44d788d7c7cdfd2114a016aa50c17f8bee2b54970bedfe0df7317"} Oct 08 21:03:53 crc kubenswrapper[4988]: I1008 21:03:53.696907 4988 scope.go:117] "RemoveContainer" containerID="fc5906dffdd6ca78fe0b71f3079a04d7d08624d242c87fd146529c035ee33bca" Oct 08 21:03:54 crc kubenswrapper[4988]: I1008 21:03:54.709276 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" event={"ID":"f328a20b-5283-407e-8a93-2ab2da599f18","Type":"ContainerStarted","Data":"6c07ed3a23b9e90ec44e0b3107713fd9bedf485fa1562b3bc3a266ffc1308b7c"} Oct 08 21:04:20 crc kubenswrapper[4988]: I1008 21:04:20.940538 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-5rspl"] Oct 08 21:04:20 crc kubenswrapper[4988]: E1008 21:04:20.941759 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0e6e85a0-3db8-451b-b388-6d7afe8a2d15" containerName="container-00" Oct 08 21:04:20 crc kubenswrapper[4988]: I1008 21:04:20.941782 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="0e6e85a0-3db8-451b-b388-6d7afe8a2d15" containerName="container-00" Oct 08 21:04:20 crc kubenswrapper[4988]: I1008 21:04:20.942194 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="0e6e85a0-3db8-451b-b388-6d7afe8a2d15" containerName="container-00" Oct 08 21:04:20 crc kubenswrapper[4988]: I1008 21:04:20.944929 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-5rspl" Oct 08 21:04:20 crc kubenswrapper[4988]: I1008 21:04:20.951696 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-5rspl"] Oct 08 21:04:21 crc kubenswrapper[4988]: I1008 21:04:21.059235 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fd655024-3f37-4279-b15a-790d349e8328-catalog-content\") pod \"certified-operators-5rspl\" (UID: \"fd655024-3f37-4279-b15a-790d349e8328\") " pod="openshift-marketplace/certified-operators-5rspl" Oct 08 21:04:21 crc kubenswrapper[4988]: I1008 21:04:21.059301 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fhl77\" (UniqueName: \"kubernetes.io/projected/fd655024-3f37-4279-b15a-790d349e8328-kube-api-access-fhl77\") pod \"certified-operators-5rspl\" (UID: \"fd655024-3f37-4279-b15a-790d349e8328\") " pod="openshift-marketplace/certified-operators-5rspl" Oct 08 21:04:21 crc kubenswrapper[4988]: I1008 21:04:21.059400 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fd655024-3f37-4279-b15a-790d349e8328-utilities\") pod \"certified-operators-5rspl\" (UID: \"fd655024-3f37-4279-b15a-790d349e8328\") " pod="openshift-marketplace/certified-operators-5rspl" Oct 08 21:04:21 crc kubenswrapper[4988]: I1008 21:04:21.161889 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fd655024-3f37-4279-b15a-790d349e8328-utilities\") pod \"certified-operators-5rspl\" (UID: \"fd655024-3f37-4279-b15a-790d349e8328\") " pod="openshift-marketplace/certified-operators-5rspl" Oct 08 21:04:21 crc kubenswrapper[4988]: I1008 21:04:21.162461 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fd655024-3f37-4279-b15a-790d349e8328-catalog-content\") pod \"certified-operators-5rspl\" (UID: \"fd655024-3f37-4279-b15a-790d349e8328\") " pod="openshift-marketplace/certified-operators-5rspl" Oct 08 21:04:21 crc kubenswrapper[4988]: I1008 21:04:21.162478 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fd655024-3f37-4279-b15a-790d349e8328-utilities\") pod \"certified-operators-5rspl\" (UID: \"fd655024-3f37-4279-b15a-790d349e8328\") " pod="openshift-marketplace/certified-operators-5rspl" Oct 08 21:04:21 crc kubenswrapper[4988]: I1008 21:04:21.162792 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fd655024-3f37-4279-b15a-790d349e8328-catalog-content\") pod \"certified-operators-5rspl\" (UID: \"fd655024-3f37-4279-b15a-790d349e8328\") " pod="openshift-marketplace/certified-operators-5rspl" Oct 08 21:04:21 crc kubenswrapper[4988]: I1008 21:04:21.162860 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fhl77\" (UniqueName: \"kubernetes.io/projected/fd655024-3f37-4279-b15a-790d349e8328-kube-api-access-fhl77\") pod \"certified-operators-5rspl\" (UID: \"fd655024-3f37-4279-b15a-790d349e8328\") " pod="openshift-marketplace/certified-operators-5rspl" Oct 08 21:04:21 crc kubenswrapper[4988]: I1008 21:04:21.192681 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fhl77\" (UniqueName: \"kubernetes.io/projected/fd655024-3f37-4279-b15a-790d349e8328-kube-api-access-fhl77\") pod \"certified-operators-5rspl\" (UID: \"fd655024-3f37-4279-b15a-790d349e8328\") " pod="openshift-marketplace/certified-operators-5rspl" Oct 08 21:04:21 crc kubenswrapper[4988]: I1008 21:04:21.282249 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-5rspl" Oct 08 21:04:22 crc kubenswrapper[4988]: I1008 21:04:22.417464 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-5rspl"] Oct 08 21:04:23 crc kubenswrapper[4988]: I1008 21:04:23.059985 4988 generic.go:334] "Generic (PLEG): container finished" podID="fd655024-3f37-4279-b15a-790d349e8328" containerID="a54175f7fb56acc2088911aeabf45d77f0a2baf977de6b3ab580b11798fe0a66" exitCode=0 Oct 08 21:04:23 crc kubenswrapper[4988]: I1008 21:04:23.060067 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5rspl" event={"ID":"fd655024-3f37-4279-b15a-790d349e8328","Type":"ContainerDied","Data":"a54175f7fb56acc2088911aeabf45d77f0a2baf977de6b3ab580b11798fe0a66"} Oct 08 21:04:23 crc kubenswrapper[4988]: I1008 21:04:23.060360 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5rspl" event={"ID":"fd655024-3f37-4279-b15a-790d349e8328","Type":"ContainerStarted","Data":"920cfceeddbb0ed1c4667b08d1e0fc5b7613e43ac95ed3f6e8c1f41071aea0aa"} Oct 08 21:04:23 crc kubenswrapper[4988]: I1008 21:04:23.062087 4988 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 08 21:04:25 crc kubenswrapper[4988]: I1008 21:04:25.085875 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5rspl" event={"ID":"fd655024-3f37-4279-b15a-790d349e8328","Type":"ContainerStarted","Data":"60e4bd533d39c56619695318d76096d9eef912dd8a528acf7811a97c25051fdf"} Oct 08 21:04:26 crc kubenswrapper[4988]: I1008 21:04:26.100143 4988 generic.go:334] "Generic (PLEG): container finished" podID="fd655024-3f37-4279-b15a-790d349e8328" containerID="60e4bd533d39c56619695318d76096d9eef912dd8a528acf7811a97c25051fdf" exitCode=0 Oct 08 21:04:26 crc kubenswrapper[4988]: I1008 21:04:26.100198 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5rspl" event={"ID":"fd655024-3f37-4279-b15a-790d349e8328","Type":"ContainerDied","Data":"60e4bd533d39c56619695318d76096d9eef912dd8a528acf7811a97c25051fdf"} Oct 08 21:04:28 crc kubenswrapper[4988]: I1008 21:04:28.121212 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5rspl" event={"ID":"fd655024-3f37-4279-b15a-790d349e8328","Type":"ContainerStarted","Data":"7a3a404cf33a4bad099dbcf44a0c7aae982b49aa8c18852d84c47767b43b241e"} Oct 08 21:04:28 crc kubenswrapper[4988]: I1008 21:04:28.143409 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-5rspl" podStartSLOduration=4.267494311 podStartE2EDuration="8.143360179s" podCreationTimestamp="2025-10-08 21:04:20 +0000 UTC" firstStartedPulling="2025-10-08 21:04:23.061826158 +0000 UTC m=+10408.511668938" lastFinishedPulling="2025-10-08 21:04:26.937692036 +0000 UTC m=+10412.387534806" observedRunningTime="2025-10-08 21:04:28.138718741 +0000 UTC m=+10413.588561531" watchObservedRunningTime="2025-10-08 21:04:28.143360179 +0000 UTC m=+10413.593202969" Oct 08 21:04:31 crc kubenswrapper[4988]: I1008 21:04:31.282412 4988 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-5rspl" Oct 08 21:04:31 crc kubenswrapper[4988]: I1008 21:04:31.283018 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-5rspl" Oct 08 21:04:31 crc kubenswrapper[4988]: I1008 21:04:31.367492 4988 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-5rspl" Oct 08 21:04:32 crc kubenswrapper[4988]: I1008 21:04:32.288601 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-5rspl" Oct 08 21:04:32 crc kubenswrapper[4988]: I1008 21:04:32.350400 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-5rspl"] Oct 08 21:04:34 crc kubenswrapper[4988]: I1008 21:04:34.218311 4988 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-5rspl" podUID="fd655024-3f37-4279-b15a-790d349e8328" containerName="registry-server" containerID="cri-o://7a3a404cf33a4bad099dbcf44a0c7aae982b49aa8c18852d84c47767b43b241e" gracePeriod=2 Oct 08 21:04:34 crc kubenswrapper[4988]: I1008 21:04:34.722808 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-5rspl" Oct 08 21:04:34 crc kubenswrapper[4988]: I1008 21:04:34.902754 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fd655024-3f37-4279-b15a-790d349e8328-catalog-content\") pod \"fd655024-3f37-4279-b15a-790d349e8328\" (UID: \"fd655024-3f37-4279-b15a-790d349e8328\") " Oct 08 21:04:34 crc kubenswrapper[4988]: I1008 21:04:34.902867 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fhl77\" (UniqueName: \"kubernetes.io/projected/fd655024-3f37-4279-b15a-790d349e8328-kube-api-access-fhl77\") pod \"fd655024-3f37-4279-b15a-790d349e8328\" (UID: \"fd655024-3f37-4279-b15a-790d349e8328\") " Oct 08 21:04:34 crc kubenswrapper[4988]: I1008 21:04:34.903055 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fd655024-3f37-4279-b15a-790d349e8328-utilities\") pod \"fd655024-3f37-4279-b15a-790d349e8328\" (UID: \"fd655024-3f37-4279-b15a-790d349e8328\") " Oct 08 21:04:34 crc kubenswrapper[4988]: I1008 21:04:34.904551 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fd655024-3f37-4279-b15a-790d349e8328-utilities" (OuterVolumeSpecName: "utilities") pod "fd655024-3f37-4279-b15a-790d349e8328" (UID: "fd655024-3f37-4279-b15a-790d349e8328"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 21:04:34 crc kubenswrapper[4988]: I1008 21:04:34.909943 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fd655024-3f37-4279-b15a-790d349e8328-kube-api-access-fhl77" (OuterVolumeSpecName: "kube-api-access-fhl77") pod "fd655024-3f37-4279-b15a-790d349e8328" (UID: "fd655024-3f37-4279-b15a-790d349e8328"). InnerVolumeSpecName "kube-api-access-fhl77". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 21:04:34 crc kubenswrapper[4988]: I1008 21:04:34.972937 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fd655024-3f37-4279-b15a-790d349e8328-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "fd655024-3f37-4279-b15a-790d349e8328" (UID: "fd655024-3f37-4279-b15a-790d349e8328"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 21:04:35 crc kubenswrapper[4988]: I1008 21:04:35.008580 4988 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fd655024-3f37-4279-b15a-790d349e8328-utilities\") on node \"crc\" DevicePath \"\"" Oct 08 21:04:35 crc kubenswrapper[4988]: I1008 21:04:35.008944 4988 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fd655024-3f37-4279-b15a-790d349e8328-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 08 21:04:35 crc kubenswrapper[4988]: I1008 21:04:35.009118 4988 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fhl77\" (UniqueName: \"kubernetes.io/projected/fd655024-3f37-4279-b15a-790d349e8328-kube-api-access-fhl77\") on node \"crc\" DevicePath \"\"" Oct 08 21:04:35 crc kubenswrapper[4988]: I1008 21:04:35.232065 4988 generic.go:334] "Generic (PLEG): container finished" podID="fd655024-3f37-4279-b15a-790d349e8328" containerID="7a3a404cf33a4bad099dbcf44a0c7aae982b49aa8c18852d84c47767b43b241e" exitCode=0 Oct 08 21:04:35 crc kubenswrapper[4988]: I1008 21:04:35.232106 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5rspl" event={"ID":"fd655024-3f37-4279-b15a-790d349e8328","Type":"ContainerDied","Data":"7a3a404cf33a4bad099dbcf44a0c7aae982b49aa8c18852d84c47767b43b241e"} Oct 08 21:04:35 crc kubenswrapper[4988]: I1008 21:04:35.232130 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5rspl" event={"ID":"fd655024-3f37-4279-b15a-790d349e8328","Type":"ContainerDied","Data":"920cfceeddbb0ed1c4667b08d1e0fc5b7613e43ac95ed3f6e8c1f41071aea0aa"} Oct 08 21:04:35 crc kubenswrapper[4988]: I1008 21:04:35.232145 4988 scope.go:117] "RemoveContainer" containerID="7a3a404cf33a4bad099dbcf44a0c7aae982b49aa8c18852d84c47767b43b241e" Oct 08 21:04:35 crc kubenswrapper[4988]: I1008 21:04:35.232593 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-5rspl" Oct 08 21:04:35 crc kubenswrapper[4988]: I1008 21:04:35.294398 4988 scope.go:117] "RemoveContainer" containerID="60e4bd533d39c56619695318d76096d9eef912dd8a528acf7811a97c25051fdf" Oct 08 21:04:35 crc kubenswrapper[4988]: I1008 21:04:35.309877 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-5rspl"] Oct 08 21:04:35 crc kubenswrapper[4988]: I1008 21:04:35.322064 4988 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-5rspl"] Oct 08 21:04:35 crc kubenswrapper[4988]: I1008 21:04:35.328695 4988 scope.go:117] "RemoveContainer" containerID="a54175f7fb56acc2088911aeabf45d77f0a2baf977de6b3ab580b11798fe0a66" Oct 08 21:04:35 crc kubenswrapper[4988]: I1008 21:04:35.375107 4988 scope.go:117] "RemoveContainer" containerID="7a3a404cf33a4bad099dbcf44a0c7aae982b49aa8c18852d84c47767b43b241e" Oct 08 21:04:35 crc kubenswrapper[4988]: E1008 21:04:35.375577 4988 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7a3a404cf33a4bad099dbcf44a0c7aae982b49aa8c18852d84c47767b43b241e\": container with ID starting with 7a3a404cf33a4bad099dbcf44a0c7aae982b49aa8c18852d84c47767b43b241e not found: ID does not exist" containerID="7a3a404cf33a4bad099dbcf44a0c7aae982b49aa8c18852d84c47767b43b241e" Oct 08 21:04:35 crc kubenswrapper[4988]: I1008 21:04:35.375615 4988 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7a3a404cf33a4bad099dbcf44a0c7aae982b49aa8c18852d84c47767b43b241e"} err="failed to get container status \"7a3a404cf33a4bad099dbcf44a0c7aae982b49aa8c18852d84c47767b43b241e\": rpc error: code = NotFound desc = could not find container \"7a3a404cf33a4bad099dbcf44a0c7aae982b49aa8c18852d84c47767b43b241e\": container with ID starting with 7a3a404cf33a4bad099dbcf44a0c7aae982b49aa8c18852d84c47767b43b241e not found: ID does not exist" Oct 08 21:04:35 crc kubenswrapper[4988]: I1008 21:04:35.375636 4988 scope.go:117] "RemoveContainer" containerID="60e4bd533d39c56619695318d76096d9eef912dd8a528acf7811a97c25051fdf" Oct 08 21:04:35 crc kubenswrapper[4988]: E1008 21:04:35.375872 4988 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"60e4bd533d39c56619695318d76096d9eef912dd8a528acf7811a97c25051fdf\": container with ID starting with 60e4bd533d39c56619695318d76096d9eef912dd8a528acf7811a97c25051fdf not found: ID does not exist" containerID="60e4bd533d39c56619695318d76096d9eef912dd8a528acf7811a97c25051fdf" Oct 08 21:04:35 crc kubenswrapper[4988]: I1008 21:04:35.375908 4988 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"60e4bd533d39c56619695318d76096d9eef912dd8a528acf7811a97c25051fdf"} err="failed to get container status \"60e4bd533d39c56619695318d76096d9eef912dd8a528acf7811a97c25051fdf\": rpc error: code = NotFound desc = could not find container \"60e4bd533d39c56619695318d76096d9eef912dd8a528acf7811a97c25051fdf\": container with ID starting with 60e4bd533d39c56619695318d76096d9eef912dd8a528acf7811a97c25051fdf not found: ID does not exist" Oct 08 21:04:35 crc kubenswrapper[4988]: I1008 21:04:35.375929 4988 scope.go:117] "RemoveContainer" containerID="a54175f7fb56acc2088911aeabf45d77f0a2baf977de6b3ab580b11798fe0a66" Oct 08 21:04:35 crc kubenswrapper[4988]: E1008 21:04:35.376806 4988 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a54175f7fb56acc2088911aeabf45d77f0a2baf977de6b3ab580b11798fe0a66\": container with ID starting with a54175f7fb56acc2088911aeabf45d77f0a2baf977de6b3ab580b11798fe0a66 not found: ID does not exist" containerID="a54175f7fb56acc2088911aeabf45d77f0a2baf977de6b3ab580b11798fe0a66" Oct 08 21:04:35 crc kubenswrapper[4988]: I1008 21:04:35.376829 4988 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a54175f7fb56acc2088911aeabf45d77f0a2baf977de6b3ab580b11798fe0a66"} err="failed to get container status \"a54175f7fb56acc2088911aeabf45d77f0a2baf977de6b3ab580b11798fe0a66\": rpc error: code = NotFound desc = could not find container \"a54175f7fb56acc2088911aeabf45d77f0a2baf977de6b3ab580b11798fe0a66\": container with ID starting with a54175f7fb56acc2088911aeabf45d77f0a2baf977de6b3ab580b11798fe0a66 not found: ID does not exist" Oct 08 21:04:37 crc kubenswrapper[4988]: I1008 21:04:37.253731 4988 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fd655024-3f37-4279-b15a-790d349e8328" path="/var/lib/kubelet/pods/fd655024-3f37-4279-b15a-790d349e8328/volumes" Oct 08 21:05:04 crc kubenswrapper[4988]: I1008 21:05:04.920942 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-h7zpk"] Oct 08 21:05:04 crc kubenswrapper[4988]: E1008 21:05:04.922046 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fd655024-3f37-4279-b15a-790d349e8328" containerName="registry-server" Oct 08 21:05:04 crc kubenswrapper[4988]: I1008 21:05:04.922064 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="fd655024-3f37-4279-b15a-790d349e8328" containerName="registry-server" Oct 08 21:05:04 crc kubenswrapper[4988]: E1008 21:05:04.922099 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fd655024-3f37-4279-b15a-790d349e8328" containerName="extract-content" Oct 08 21:05:04 crc kubenswrapper[4988]: I1008 21:05:04.922107 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="fd655024-3f37-4279-b15a-790d349e8328" containerName="extract-content" Oct 08 21:05:04 crc kubenswrapper[4988]: E1008 21:05:04.922122 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fd655024-3f37-4279-b15a-790d349e8328" containerName="extract-utilities" Oct 08 21:05:04 crc kubenswrapper[4988]: I1008 21:05:04.922131 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="fd655024-3f37-4279-b15a-790d349e8328" containerName="extract-utilities" Oct 08 21:05:04 crc kubenswrapper[4988]: I1008 21:05:04.922449 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="fd655024-3f37-4279-b15a-790d349e8328" containerName="registry-server" Oct 08 21:05:04 crc kubenswrapper[4988]: I1008 21:05:04.924703 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-h7zpk" Oct 08 21:05:04 crc kubenswrapper[4988]: I1008 21:05:04.933708 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-h7zpk"] Oct 08 21:05:05 crc kubenswrapper[4988]: I1008 21:05:05.022999 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jm5vh\" (UniqueName: \"kubernetes.io/projected/2eb2a0fe-0016-4372-abff-484b668e9fc0-kube-api-access-jm5vh\") pod \"community-operators-h7zpk\" (UID: \"2eb2a0fe-0016-4372-abff-484b668e9fc0\") " pod="openshift-marketplace/community-operators-h7zpk" Oct 08 21:05:05 crc kubenswrapper[4988]: I1008 21:05:05.023878 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2eb2a0fe-0016-4372-abff-484b668e9fc0-catalog-content\") pod \"community-operators-h7zpk\" (UID: \"2eb2a0fe-0016-4372-abff-484b668e9fc0\") " pod="openshift-marketplace/community-operators-h7zpk" Oct 08 21:05:05 crc kubenswrapper[4988]: I1008 21:05:05.024199 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2eb2a0fe-0016-4372-abff-484b668e9fc0-utilities\") pod \"community-operators-h7zpk\" (UID: \"2eb2a0fe-0016-4372-abff-484b668e9fc0\") " pod="openshift-marketplace/community-operators-h7zpk" Oct 08 21:05:05 crc kubenswrapper[4988]: I1008 21:05:05.126119 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jm5vh\" (UniqueName: \"kubernetes.io/projected/2eb2a0fe-0016-4372-abff-484b668e9fc0-kube-api-access-jm5vh\") pod \"community-operators-h7zpk\" (UID: \"2eb2a0fe-0016-4372-abff-484b668e9fc0\") " pod="openshift-marketplace/community-operators-h7zpk" Oct 08 21:05:05 crc kubenswrapper[4988]: I1008 21:05:05.126496 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2eb2a0fe-0016-4372-abff-484b668e9fc0-catalog-content\") pod \"community-operators-h7zpk\" (UID: \"2eb2a0fe-0016-4372-abff-484b668e9fc0\") " pod="openshift-marketplace/community-operators-h7zpk" Oct 08 21:05:05 crc kubenswrapper[4988]: I1008 21:05:05.126717 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2eb2a0fe-0016-4372-abff-484b668e9fc0-utilities\") pod \"community-operators-h7zpk\" (UID: \"2eb2a0fe-0016-4372-abff-484b668e9fc0\") " pod="openshift-marketplace/community-operators-h7zpk" Oct 08 21:05:05 crc kubenswrapper[4988]: I1008 21:05:05.126885 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2eb2a0fe-0016-4372-abff-484b668e9fc0-catalog-content\") pod \"community-operators-h7zpk\" (UID: \"2eb2a0fe-0016-4372-abff-484b668e9fc0\") " pod="openshift-marketplace/community-operators-h7zpk" Oct 08 21:05:05 crc kubenswrapper[4988]: I1008 21:05:05.127092 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2eb2a0fe-0016-4372-abff-484b668e9fc0-utilities\") pod \"community-operators-h7zpk\" (UID: \"2eb2a0fe-0016-4372-abff-484b668e9fc0\") " pod="openshift-marketplace/community-operators-h7zpk" Oct 08 21:05:05 crc kubenswrapper[4988]: I1008 21:05:05.153285 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jm5vh\" (UniqueName: \"kubernetes.io/projected/2eb2a0fe-0016-4372-abff-484b668e9fc0-kube-api-access-jm5vh\") pod \"community-operators-h7zpk\" (UID: \"2eb2a0fe-0016-4372-abff-484b668e9fc0\") " pod="openshift-marketplace/community-operators-h7zpk" Oct 08 21:05:05 crc kubenswrapper[4988]: I1008 21:05:05.257355 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-h7zpk" Oct 08 21:05:05 crc kubenswrapper[4988]: I1008 21:05:05.792032 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-h7zpk"] Oct 08 21:05:06 crc kubenswrapper[4988]: I1008 21:05:06.616501 4988 generic.go:334] "Generic (PLEG): container finished" podID="2eb2a0fe-0016-4372-abff-484b668e9fc0" containerID="5099d120d977334dd65322f2b566f73fe4fee67506a0781a2aad1458c6e806fd" exitCode=0 Oct 08 21:05:06 crc kubenswrapper[4988]: I1008 21:05:06.616970 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-h7zpk" event={"ID":"2eb2a0fe-0016-4372-abff-484b668e9fc0","Type":"ContainerDied","Data":"5099d120d977334dd65322f2b566f73fe4fee67506a0781a2aad1458c6e806fd"} Oct 08 21:05:06 crc kubenswrapper[4988]: I1008 21:05:06.617005 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-h7zpk" event={"ID":"2eb2a0fe-0016-4372-abff-484b668e9fc0","Type":"ContainerStarted","Data":"280a68a370c6a726e915933273fb88a43f7f2c48d4f4059ddec3c2e1c556a737"} Oct 08 21:05:08 crc kubenswrapper[4988]: I1008 21:05:08.656007 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-h7zpk" event={"ID":"2eb2a0fe-0016-4372-abff-484b668e9fc0","Type":"ContainerStarted","Data":"4a2133da1f3ec2a1449812877cf81476e56613e1b02669d82d7edfbc565295e2"} Oct 08 21:05:09 crc kubenswrapper[4988]: I1008 21:05:09.669546 4988 generic.go:334] "Generic (PLEG): container finished" podID="2eb2a0fe-0016-4372-abff-484b668e9fc0" containerID="4a2133da1f3ec2a1449812877cf81476e56613e1b02669d82d7edfbc565295e2" exitCode=0 Oct 08 21:05:09 crc kubenswrapper[4988]: I1008 21:05:09.669655 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-h7zpk" event={"ID":"2eb2a0fe-0016-4372-abff-484b668e9fc0","Type":"ContainerDied","Data":"4a2133da1f3ec2a1449812877cf81476e56613e1b02669d82d7edfbc565295e2"} Oct 08 21:05:10 crc kubenswrapper[4988]: I1008 21:05:10.683085 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-h7zpk" event={"ID":"2eb2a0fe-0016-4372-abff-484b668e9fc0","Type":"ContainerStarted","Data":"4541964266bdcfe730c10981582213e1ce36b09f274b04aa56f15f7f9f5123dc"} Oct 08 21:05:10 crc kubenswrapper[4988]: I1008 21:05:10.718502 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-h7zpk" podStartSLOduration=3.236768056 podStartE2EDuration="6.71848146s" podCreationTimestamp="2025-10-08 21:05:04 +0000 UTC" firstStartedPulling="2025-10-08 21:05:06.619232227 +0000 UTC m=+10452.069075027" lastFinishedPulling="2025-10-08 21:05:10.100945651 +0000 UTC m=+10455.550788431" observedRunningTime="2025-10-08 21:05:10.71312748 +0000 UTC m=+10456.162970330" watchObservedRunningTime="2025-10-08 21:05:10.71848146 +0000 UTC m=+10456.168324240" Oct 08 21:05:15 crc kubenswrapper[4988]: I1008 21:05:15.258594 4988 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-h7zpk" Oct 08 21:05:15 crc kubenswrapper[4988]: I1008 21:05:15.259275 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-h7zpk" Oct 08 21:05:15 crc kubenswrapper[4988]: I1008 21:05:15.356377 4988 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-h7zpk" Oct 08 21:05:15 crc kubenswrapper[4988]: I1008 21:05:15.787015 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-h7zpk" Oct 08 21:05:15 crc kubenswrapper[4988]: I1008 21:05:15.854765 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-h7zpk"] Oct 08 21:05:17 crc kubenswrapper[4988]: I1008 21:05:17.761803 4988 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-h7zpk" podUID="2eb2a0fe-0016-4372-abff-484b668e9fc0" containerName="registry-server" containerID="cri-o://4541964266bdcfe730c10981582213e1ce36b09f274b04aa56f15f7f9f5123dc" gracePeriod=2 Oct 08 21:05:18 crc kubenswrapper[4988]: I1008 21:05:18.758512 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-h7zpk" Oct 08 21:05:18 crc kubenswrapper[4988]: I1008 21:05:18.777885 4988 generic.go:334] "Generic (PLEG): container finished" podID="2eb2a0fe-0016-4372-abff-484b668e9fc0" containerID="4541964266bdcfe730c10981582213e1ce36b09f274b04aa56f15f7f9f5123dc" exitCode=0 Oct 08 21:05:18 crc kubenswrapper[4988]: I1008 21:05:18.777936 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-h7zpk" event={"ID":"2eb2a0fe-0016-4372-abff-484b668e9fc0","Type":"ContainerDied","Data":"4541964266bdcfe730c10981582213e1ce36b09f274b04aa56f15f7f9f5123dc"} Oct 08 21:05:18 crc kubenswrapper[4988]: I1008 21:05:18.777968 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-h7zpk" event={"ID":"2eb2a0fe-0016-4372-abff-484b668e9fc0","Type":"ContainerDied","Data":"280a68a370c6a726e915933273fb88a43f7f2c48d4f4059ddec3c2e1c556a737"} Oct 08 21:05:18 crc kubenswrapper[4988]: I1008 21:05:18.777990 4988 scope.go:117] "RemoveContainer" containerID="4541964266bdcfe730c10981582213e1ce36b09f274b04aa56f15f7f9f5123dc" Oct 08 21:05:18 crc kubenswrapper[4988]: I1008 21:05:18.778146 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-h7zpk" Oct 08 21:05:18 crc kubenswrapper[4988]: I1008 21:05:18.852452 4988 scope.go:117] "RemoveContainer" containerID="4a2133da1f3ec2a1449812877cf81476e56613e1b02669d82d7edfbc565295e2" Oct 08 21:05:18 crc kubenswrapper[4988]: I1008 21:05:18.884234 4988 scope.go:117] "RemoveContainer" containerID="5099d120d977334dd65322f2b566f73fe4fee67506a0781a2aad1458c6e806fd" Oct 08 21:05:18 crc kubenswrapper[4988]: I1008 21:05:18.894915 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2eb2a0fe-0016-4372-abff-484b668e9fc0-utilities\") pod \"2eb2a0fe-0016-4372-abff-484b668e9fc0\" (UID: \"2eb2a0fe-0016-4372-abff-484b668e9fc0\") " Oct 08 21:05:18 crc kubenswrapper[4988]: I1008 21:05:18.895237 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2eb2a0fe-0016-4372-abff-484b668e9fc0-catalog-content\") pod \"2eb2a0fe-0016-4372-abff-484b668e9fc0\" (UID: \"2eb2a0fe-0016-4372-abff-484b668e9fc0\") " Oct 08 21:05:18 crc kubenswrapper[4988]: I1008 21:05:18.895284 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jm5vh\" (UniqueName: \"kubernetes.io/projected/2eb2a0fe-0016-4372-abff-484b668e9fc0-kube-api-access-jm5vh\") pod \"2eb2a0fe-0016-4372-abff-484b668e9fc0\" (UID: \"2eb2a0fe-0016-4372-abff-484b668e9fc0\") " Oct 08 21:05:18 crc kubenswrapper[4988]: I1008 21:05:18.896515 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2eb2a0fe-0016-4372-abff-484b668e9fc0-utilities" (OuterVolumeSpecName: "utilities") pod "2eb2a0fe-0016-4372-abff-484b668e9fc0" (UID: "2eb2a0fe-0016-4372-abff-484b668e9fc0"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 21:05:18 crc kubenswrapper[4988]: I1008 21:05:18.902503 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2eb2a0fe-0016-4372-abff-484b668e9fc0-kube-api-access-jm5vh" (OuterVolumeSpecName: "kube-api-access-jm5vh") pod "2eb2a0fe-0016-4372-abff-484b668e9fc0" (UID: "2eb2a0fe-0016-4372-abff-484b668e9fc0"). InnerVolumeSpecName "kube-api-access-jm5vh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 21:05:18 crc kubenswrapper[4988]: I1008 21:05:18.962139 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2eb2a0fe-0016-4372-abff-484b668e9fc0-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "2eb2a0fe-0016-4372-abff-484b668e9fc0" (UID: "2eb2a0fe-0016-4372-abff-484b668e9fc0"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 21:05:18 crc kubenswrapper[4988]: I1008 21:05:18.975774 4988 scope.go:117] "RemoveContainer" containerID="4541964266bdcfe730c10981582213e1ce36b09f274b04aa56f15f7f9f5123dc" Oct 08 21:05:18 crc kubenswrapper[4988]: E1008 21:05:18.976449 4988 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4541964266bdcfe730c10981582213e1ce36b09f274b04aa56f15f7f9f5123dc\": container with ID starting with 4541964266bdcfe730c10981582213e1ce36b09f274b04aa56f15f7f9f5123dc not found: ID does not exist" containerID="4541964266bdcfe730c10981582213e1ce36b09f274b04aa56f15f7f9f5123dc" Oct 08 21:05:18 crc kubenswrapper[4988]: I1008 21:05:18.976507 4988 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4541964266bdcfe730c10981582213e1ce36b09f274b04aa56f15f7f9f5123dc"} err="failed to get container status \"4541964266bdcfe730c10981582213e1ce36b09f274b04aa56f15f7f9f5123dc\": rpc error: code = NotFound desc = could not find container \"4541964266bdcfe730c10981582213e1ce36b09f274b04aa56f15f7f9f5123dc\": container with ID starting with 4541964266bdcfe730c10981582213e1ce36b09f274b04aa56f15f7f9f5123dc not found: ID does not exist" Oct 08 21:05:18 crc kubenswrapper[4988]: I1008 21:05:18.976548 4988 scope.go:117] "RemoveContainer" containerID="4a2133da1f3ec2a1449812877cf81476e56613e1b02669d82d7edfbc565295e2" Oct 08 21:05:18 crc kubenswrapper[4988]: E1008 21:05:18.976981 4988 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4a2133da1f3ec2a1449812877cf81476e56613e1b02669d82d7edfbc565295e2\": container with ID starting with 4a2133da1f3ec2a1449812877cf81476e56613e1b02669d82d7edfbc565295e2 not found: ID does not exist" containerID="4a2133da1f3ec2a1449812877cf81476e56613e1b02669d82d7edfbc565295e2" Oct 08 21:05:18 crc kubenswrapper[4988]: I1008 21:05:18.977019 4988 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4a2133da1f3ec2a1449812877cf81476e56613e1b02669d82d7edfbc565295e2"} err="failed to get container status \"4a2133da1f3ec2a1449812877cf81476e56613e1b02669d82d7edfbc565295e2\": rpc error: code = NotFound desc = could not find container \"4a2133da1f3ec2a1449812877cf81476e56613e1b02669d82d7edfbc565295e2\": container with ID starting with 4a2133da1f3ec2a1449812877cf81476e56613e1b02669d82d7edfbc565295e2 not found: ID does not exist" Oct 08 21:05:18 crc kubenswrapper[4988]: I1008 21:05:18.977038 4988 scope.go:117] "RemoveContainer" containerID="5099d120d977334dd65322f2b566f73fe4fee67506a0781a2aad1458c6e806fd" Oct 08 21:05:18 crc kubenswrapper[4988]: E1008 21:05:18.977322 4988 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5099d120d977334dd65322f2b566f73fe4fee67506a0781a2aad1458c6e806fd\": container with ID starting with 5099d120d977334dd65322f2b566f73fe4fee67506a0781a2aad1458c6e806fd not found: ID does not exist" containerID="5099d120d977334dd65322f2b566f73fe4fee67506a0781a2aad1458c6e806fd" Oct 08 21:05:18 crc kubenswrapper[4988]: I1008 21:05:18.977348 4988 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5099d120d977334dd65322f2b566f73fe4fee67506a0781a2aad1458c6e806fd"} err="failed to get container status \"5099d120d977334dd65322f2b566f73fe4fee67506a0781a2aad1458c6e806fd\": rpc error: code = NotFound desc = could not find container \"5099d120d977334dd65322f2b566f73fe4fee67506a0781a2aad1458c6e806fd\": container with ID starting with 5099d120d977334dd65322f2b566f73fe4fee67506a0781a2aad1458c6e806fd not found: ID does not exist" Oct 08 21:05:18 crc kubenswrapper[4988]: I1008 21:05:18.997868 4988 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2eb2a0fe-0016-4372-abff-484b668e9fc0-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 08 21:05:18 crc kubenswrapper[4988]: I1008 21:05:18.997906 4988 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jm5vh\" (UniqueName: \"kubernetes.io/projected/2eb2a0fe-0016-4372-abff-484b668e9fc0-kube-api-access-jm5vh\") on node \"crc\" DevicePath \"\"" Oct 08 21:05:18 crc kubenswrapper[4988]: I1008 21:05:18.997920 4988 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2eb2a0fe-0016-4372-abff-484b668e9fc0-utilities\") on node \"crc\" DevicePath \"\"" Oct 08 21:05:19 crc kubenswrapper[4988]: I1008 21:05:19.115552 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-h7zpk"] Oct 08 21:05:19 crc kubenswrapper[4988]: I1008 21:05:19.124477 4988 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-h7zpk"] Oct 08 21:05:19 crc kubenswrapper[4988]: I1008 21:05:19.252095 4988 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2eb2a0fe-0016-4372-abff-484b668e9fc0" path="/var/lib/kubelet/pods/2eb2a0fe-0016-4372-abff-484b668e9fc0/volumes" Oct 08 21:05:53 crc kubenswrapper[4988]: I1008 21:05:53.338129 4988 patch_prober.go:28] interesting pod/machine-config-daemon-nm54f container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 21:05:53 crc kubenswrapper[4988]: I1008 21:05:53.338735 4988 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 21:06:12 crc kubenswrapper[4988]: I1008 21:06:12.455879 4988 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_1bfea55cf6f540853d0c0d919aed51b5fd6e2d105cc44a68cb64399f8bdkg77_75e43998-fcba-4311-8021-d102def09b3c/util/0.log" Oct 08 21:06:12 crc kubenswrapper[4988]: I1008 21:06:12.638859 4988 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_1bfea55cf6f540853d0c0d919aed51b5fd6e2d105cc44a68cb64399f8bdkg77_75e43998-fcba-4311-8021-d102def09b3c/pull/0.log" Oct 08 21:06:12 crc kubenswrapper[4988]: I1008 21:06:12.650810 4988 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_1bfea55cf6f540853d0c0d919aed51b5fd6e2d105cc44a68cb64399f8bdkg77_75e43998-fcba-4311-8021-d102def09b3c/util/0.log" Oct 08 21:06:12 crc kubenswrapper[4988]: I1008 21:06:12.668732 4988 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_1bfea55cf6f540853d0c0d919aed51b5fd6e2d105cc44a68cb64399f8bdkg77_75e43998-fcba-4311-8021-d102def09b3c/pull/0.log" Oct 08 21:06:12 crc kubenswrapper[4988]: I1008 21:06:12.830839 4988 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_1bfea55cf6f540853d0c0d919aed51b5fd6e2d105cc44a68cb64399f8bdkg77_75e43998-fcba-4311-8021-d102def09b3c/pull/0.log" Oct 08 21:06:12 crc kubenswrapper[4988]: I1008 21:06:12.834837 4988 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_1bfea55cf6f540853d0c0d919aed51b5fd6e2d105cc44a68cb64399f8bdkg77_75e43998-fcba-4311-8021-d102def09b3c/util/0.log" Oct 08 21:06:12 crc kubenswrapper[4988]: I1008 21:06:12.841353 4988 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_1bfea55cf6f540853d0c0d919aed51b5fd6e2d105cc44a68cb64399f8bdkg77_75e43998-fcba-4311-8021-d102def09b3c/extract/0.log" Oct 08 21:06:13 crc kubenswrapper[4988]: I1008 21:06:13.020457 4988 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-658bdf4b74-cmnkr_c569c6ea-0929-4fc6-8d67-47578fa225c1/kube-rbac-proxy/0.log" Oct 08 21:06:13 crc kubenswrapper[4988]: I1008 21:06:13.099867 4988 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-658bdf4b74-cmnkr_c569c6ea-0929-4fc6-8d67-47578fa225c1/manager/0.log" Oct 08 21:06:13 crc kubenswrapper[4988]: I1008 21:06:13.314805 4988 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-7b7fb68549-7xcz4_420edeb5-79c9-4ece-9e95-d497d4be3b76/kube-rbac-proxy/0.log" Oct 08 21:06:13 crc kubenswrapper[4988]: I1008 21:06:13.469215 4988 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-85d5d9dd78-94rjz_81f789aa-012b-4ee6-a3c4-70bba5d64291/kube-rbac-proxy/0.log" Oct 08 21:06:13 crc kubenswrapper[4988]: I1008 21:06:13.487901 4988 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-7b7fb68549-7xcz4_420edeb5-79c9-4ece-9e95-d497d4be3b76/manager/0.log" Oct 08 21:06:13 crc kubenswrapper[4988]: I1008 21:06:13.514916 4988 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-85d5d9dd78-94rjz_81f789aa-012b-4ee6-a3c4-70bba5d64291/manager/0.log" Oct 08 21:06:13 crc kubenswrapper[4988]: I1008 21:06:13.696316 4988 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-84b9b84486-gpl58_1af0d400-4968-4d6e-af43-0624f7bda6f0/kube-rbac-proxy/0.log" Oct 08 21:06:13 crc kubenswrapper[4988]: I1008 21:06:13.821358 4988 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-84b9b84486-gpl58_1af0d400-4968-4d6e-af43-0624f7bda6f0/manager/0.log" Oct 08 21:06:13 crc kubenswrapper[4988]: I1008 21:06:13.943710 4988 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-858f76bbdd-rjzd5_2afedc0a-a797-4f9e-921f-5160fc437f79/manager/0.log" Oct 08 21:06:13 crc kubenswrapper[4988]: I1008 21:06:13.961622 4988 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-858f76bbdd-rjzd5_2afedc0a-a797-4f9e-921f-5160fc437f79/kube-rbac-proxy/0.log" Oct 08 21:06:14 crc kubenswrapper[4988]: I1008 21:06:14.011102 4988 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-7ffbcb7588-jwr4r_8600e1a1-b22a-4cdb-9e91-04e0a65b3756/kube-rbac-proxy/0.log" Oct 08 21:06:14 crc kubenswrapper[4988]: I1008 21:06:14.183707 4988 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-7ffbcb7588-jwr4r_8600e1a1-b22a-4cdb-9e91-04e0a65b3756/manager/0.log" Oct 08 21:06:14 crc kubenswrapper[4988]: I1008 21:06:14.186337 4988 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-656bcbd775-2xfhc_2ebb5c10-6685-4a61-8b57-bd39d36b3ce5/kube-rbac-proxy/0.log" Oct 08 21:06:14 crc kubenswrapper[4988]: I1008 21:06:14.428458 4988 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-9c5c78d49-fv7xm_70495ce4-a8be-4853-be10-a5fe9e1020a7/manager/0.log" Oct 08 21:06:14 crc kubenswrapper[4988]: I1008 21:06:14.457355 4988 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-9c5c78d49-fv7xm_70495ce4-a8be-4853-be10-a5fe9e1020a7/kube-rbac-proxy/0.log" Oct 08 21:06:14 crc kubenswrapper[4988]: I1008 21:06:14.472084 4988 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-656bcbd775-2xfhc_2ebb5c10-6685-4a61-8b57-bd39d36b3ce5/manager/0.log" Oct 08 21:06:14 crc kubenswrapper[4988]: I1008 21:06:14.703073 4988 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-55b6b7c7b8-4ts62_70347d19-f528-4db3-81e7-6fbca659710e/kube-rbac-proxy/0.log" Oct 08 21:06:14 crc kubenswrapper[4988]: I1008 21:06:14.734610 4988 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-55b6b7c7b8-4ts62_70347d19-f528-4db3-81e7-6fbca659710e/manager/0.log" Oct 08 21:06:14 crc kubenswrapper[4988]: I1008 21:06:14.873904 4988 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-5f67fbc655-m7d7l_128cb2db-23b2-423f-9ec1-a7c4ed246731/kube-rbac-proxy/0.log" Oct 08 21:06:14 crc kubenswrapper[4988]: I1008 21:06:14.905980 4988 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-5f67fbc655-m7d7l_128cb2db-23b2-423f-9ec1-a7c4ed246731/manager/0.log" Oct 08 21:06:14 crc kubenswrapper[4988]: I1008 21:06:14.950811 4988 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-f9fb45f8f-mdxnm_cc3a04bc-795b-49e9-bc53-06b8534244ac/kube-rbac-proxy/0.log" Oct 08 21:06:15 crc kubenswrapper[4988]: I1008 21:06:15.094490 4988 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-f9fb45f8f-mdxnm_cc3a04bc-795b-49e9-bc53-06b8534244ac/manager/0.log" Oct 08 21:06:15 crc kubenswrapper[4988]: I1008 21:06:15.138383 4988 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-79d585cb66-9cfhq_4c6dc76c-9b71-45aa-b1e9-0fd9deecbd39/kube-rbac-proxy/0.log" Oct 08 21:06:15 crc kubenswrapper[4988]: I1008 21:06:15.260009 4988 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-79d585cb66-9cfhq_4c6dc76c-9b71-45aa-b1e9-0fd9deecbd39/manager/0.log" Oct 08 21:06:15 crc kubenswrapper[4988]: I1008 21:06:15.326653 4988 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-5df598886f-8589q_f9ff7c98-5f49-4c45-a367-5ed4471083a8/kube-rbac-proxy/0.log" Oct 08 21:06:15 crc kubenswrapper[4988]: I1008 21:06:15.540520 4988 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-5df598886f-8589q_f9ff7c98-5f49-4c45-a367-5ed4471083a8/manager/0.log" Oct 08 21:06:15 crc kubenswrapper[4988]: I1008 21:06:15.578450 4988 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-69fdcfc5f5-zd6wx_db43233d-abb6-46e3-9b9f-946024ab477a/manager/0.log" Oct 08 21:06:15 crc kubenswrapper[4988]: I1008 21:06:15.583341 4988 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-69fdcfc5f5-zd6wx_db43233d-abb6-46e3-9b9f-946024ab477a/kube-rbac-proxy/0.log" Oct 08 21:06:15 crc kubenswrapper[4988]: I1008 21:06:15.656172 4988 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-677c5f5bff2dshw_361b9b35-62d8-494a-9b8b-7d471adb246f/kube-rbac-proxy/0.log" Oct 08 21:06:15 crc kubenswrapper[4988]: I1008 21:06:15.723747 4988 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-677c5f5bff2dshw_361b9b35-62d8-494a-9b8b-7d471adb246f/manager/0.log" Oct 08 21:06:15 crc kubenswrapper[4988]: I1008 21:06:15.804192 4988 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-598c4c5b5-mxr9w_3ed83b64-1d35-4ace-853f-2d246e27fc72/kube-rbac-proxy/0.log" Oct 08 21:06:15 crc kubenswrapper[4988]: I1008 21:06:15.927790 4988 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-operator-bd6bc67fb-q9rwk_cbb26b26-1ffa-42b5-acce-5bca3a217c69/kube-rbac-proxy/0.log" Oct 08 21:06:16 crc kubenswrapper[4988]: I1008 21:06:16.235032 4988 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-79db49b9fb-75xwr_fc79827a-abba-4525-a705-129531293589/kube-rbac-proxy/0.log" Oct 08 21:06:16 crc kubenswrapper[4988]: I1008 21:06:16.257832 4988 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-operator-bd6bc67fb-q9rwk_cbb26b26-1ffa-42b5-acce-5bca3a217c69/operator/0.log" Oct 08 21:06:16 crc kubenswrapper[4988]: I1008 21:06:16.389679 4988 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-index-5ccsc_c1a9b179-3660-477f-8fdf-1aeb9cbb66df/registry-server/0.log" Oct 08 21:06:16 crc kubenswrapper[4988]: I1008 21:06:16.470884 4988 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-79db49b9fb-75xwr_fc79827a-abba-4525-a705-129531293589/manager/0.log" Oct 08 21:06:16 crc kubenswrapper[4988]: I1008 21:06:16.536246 4988 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-68b6c87b68-rf945_3433daff-1b78-4aa4-8eec-4bebfcbe1320/kube-rbac-proxy/0.log" Oct 08 21:06:16 crc kubenswrapper[4988]: I1008 21:06:16.647582 4988 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-68b6c87b68-rf945_3433daff-1b78-4aa4-8eec-4bebfcbe1320/manager/0.log" Oct 08 21:06:16 crc kubenswrapper[4988]: I1008 21:06:16.720251 4988 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_rabbitmq-cluster-operator-manager-5f97d8c699-25bz8_e4ef3040-d167-453d-aaa4-b807020a6876/operator/0.log" Oct 08 21:06:16 crc kubenswrapper[4988]: I1008 21:06:16.882615 4988 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-db6d7f97b-mxg6q_1154bd89-f7b0-4cce-b02c-9157b7866bf8/kube-rbac-proxy/0.log" Oct 08 21:06:16 crc kubenswrapper[4988]: I1008 21:06:16.989785 4988 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-76796d4c6b-2szmt_c97bc0d9-b498-44ce-9fb5-f4cf7ca7561e/kube-rbac-proxy/0.log" Oct 08 21:06:17 crc kubenswrapper[4988]: I1008 21:06:17.196601 4988 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-56c698c775-2p984_d3afe0af-aa6b-4321-99f3-39b2f64c2451/kube-rbac-proxy/0.log" Oct 08 21:06:17 crc kubenswrapper[4988]: I1008 21:06:17.269873 4988 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-db6d7f97b-mxg6q_1154bd89-f7b0-4cce-b02c-9157b7866bf8/manager/0.log" Oct 08 21:06:17 crc kubenswrapper[4988]: I1008 21:06:17.324998 4988 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-56c698c775-2p984_d3afe0af-aa6b-4321-99f3-39b2f64c2451/manager/0.log" Oct 08 21:06:17 crc kubenswrapper[4988]: I1008 21:06:17.437614 4988 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-7794bc6bd-wfq6r_47379f88-653e-4bda-a37b-a81857b7af36/kube-rbac-proxy/0.log" Oct 08 21:06:17 crc kubenswrapper[4988]: I1008 21:06:17.511036 4988 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-76796d4c6b-2szmt_c97bc0d9-b498-44ce-9fb5-f4cf7ca7561e/manager/0.log" Oct 08 21:06:17 crc kubenswrapper[4988]: I1008 21:06:17.546510 4988 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-7794bc6bd-wfq6r_47379f88-653e-4bda-a37b-a81857b7af36/manager/0.log" Oct 08 21:06:18 crc kubenswrapper[4988]: I1008 21:06:18.341173 4988 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-598c4c5b5-mxr9w_3ed83b64-1d35-4ace-853f-2d246e27fc72/manager/0.log" Oct 08 21:06:23 crc kubenswrapper[4988]: I1008 21:06:23.338477 4988 patch_prober.go:28] interesting pod/machine-config-daemon-nm54f container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 21:06:23 crc kubenswrapper[4988]: I1008 21:06:23.338971 4988 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 21:06:35 crc kubenswrapper[4988]: I1008 21:06:35.348887 4988 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_control-plane-machine-set-operator-78cbb6b69f-hgnvq_31a216f1-4766-4836-b3c5-a245ad42d709/control-plane-machine-set-operator/0.log" Oct 08 21:06:35 crc kubenswrapper[4988]: I1008 21:06:35.524559 4988 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-5mrrc_a760b68f-a1f6-4741-a9e9-51f6d3968748/kube-rbac-proxy/0.log" Oct 08 21:06:35 crc kubenswrapper[4988]: I1008 21:06:35.525512 4988 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-5mrrc_a760b68f-a1f6-4741-a9e9-51f6d3968748/machine-api-operator/0.log" Oct 08 21:06:48 crc kubenswrapper[4988]: I1008 21:06:48.010203 4988 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-7d4cc89fcb-4b68b_d829b439-107e-4276-81dc-12052e1baf08/cert-manager-controller/0.log" Oct 08 21:06:48 crc kubenswrapper[4988]: I1008 21:06:48.204119 4988 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-webhook-d969966f-d8xbq_c718a13a-62b5-43c9-a151-bd3aab9e209d/cert-manager-webhook/0.log" Oct 08 21:06:48 crc kubenswrapper[4988]: I1008 21:06:48.211657 4988 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-cainjector-7d9f95dbf-qrjr7_75c13b54-1bce-4854-ae04-96c03b7b0333/cert-manager-cainjector/0.log" Oct 08 21:06:53 crc kubenswrapper[4988]: I1008 21:06:53.337840 4988 patch_prober.go:28] interesting pod/machine-config-daemon-nm54f container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 21:06:53 crc kubenswrapper[4988]: I1008 21:06:53.338548 4988 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 21:06:53 crc kubenswrapper[4988]: I1008 21:06:53.338603 4988 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" Oct 08 21:06:53 crc kubenswrapper[4988]: I1008 21:06:53.339589 4988 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"6c07ed3a23b9e90ec44e0b3107713fd9bedf485fa1562b3bc3a266ffc1308b7c"} pod="openshift-machine-config-operator/machine-config-daemon-nm54f" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 08 21:06:53 crc kubenswrapper[4988]: I1008 21:06:53.339663 4988 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" containerName="machine-config-daemon" containerID="cri-o://6c07ed3a23b9e90ec44e0b3107713fd9bedf485fa1562b3bc3a266ffc1308b7c" gracePeriod=600 Oct 08 21:06:53 crc kubenswrapper[4988]: E1008 21:06:53.469991 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nm54f_openshift-machine-config-operator(f328a20b-5283-407e-8a93-2ab2da599f18)\"" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" Oct 08 21:06:53 crc kubenswrapper[4988]: I1008 21:06:53.897269 4988 generic.go:334] "Generic (PLEG): container finished" podID="f328a20b-5283-407e-8a93-2ab2da599f18" containerID="6c07ed3a23b9e90ec44e0b3107713fd9bedf485fa1562b3bc3a266ffc1308b7c" exitCode=0 Oct 08 21:06:53 crc kubenswrapper[4988]: I1008 21:06:53.897314 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" event={"ID":"f328a20b-5283-407e-8a93-2ab2da599f18","Type":"ContainerDied","Data":"6c07ed3a23b9e90ec44e0b3107713fd9bedf485fa1562b3bc3a266ffc1308b7c"} Oct 08 21:06:53 crc kubenswrapper[4988]: I1008 21:06:53.897660 4988 scope.go:117] "RemoveContainer" containerID="6a30956f2db44d788d7c7cdfd2114a016aa50c17f8bee2b54970bedfe0df7317" Oct 08 21:06:53 crc kubenswrapper[4988]: I1008 21:06:53.898647 4988 scope.go:117] "RemoveContainer" containerID="6c07ed3a23b9e90ec44e0b3107713fd9bedf485fa1562b3bc3a266ffc1308b7c" Oct 08 21:06:53 crc kubenswrapper[4988]: E1008 21:06:53.899268 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nm54f_openshift-machine-config-operator(f328a20b-5283-407e-8a93-2ab2da599f18)\"" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" Oct 08 21:07:00 crc kubenswrapper[4988]: I1008 21:07:00.590175 4988 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-console-plugin-6b874cbd85-wh7bx_670ba223-6bd1-429c-8739-61aea9074914/nmstate-console-plugin/0.log" Oct 08 21:07:00 crc kubenswrapper[4988]: I1008 21:07:00.785836 4988 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-handler-dv92p_b0d72824-9655-4a95-9ada-e7d7f8938213/nmstate-handler/0.log" Oct 08 21:07:00 crc kubenswrapper[4988]: I1008 21:07:00.827697 4988 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-fdff9cb8d-b574s_09b9ef2d-78cf-4bb7-a5ed-79f63f65fa8c/kube-rbac-proxy/0.log" Oct 08 21:07:00 crc kubenswrapper[4988]: I1008 21:07:00.853857 4988 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-fdff9cb8d-b574s_09b9ef2d-78cf-4bb7-a5ed-79f63f65fa8c/nmstate-metrics/0.log" Oct 08 21:07:01 crc kubenswrapper[4988]: I1008 21:07:01.844755 4988 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-webhook-6cdbc54649-xg4fg_bb27ea19-b9c4-4ea6-b7b1-6de4ae603b43/nmstate-webhook/0.log" Oct 08 21:07:01 crc kubenswrapper[4988]: I1008 21:07:01.849478 4988 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-operator-858ddd8f98-hplwk_cc122836-fc53-4258-8074-f0943ca9c946/nmstate-operator/0.log" Oct 08 21:07:06 crc kubenswrapper[4988]: I1008 21:07:06.237489 4988 scope.go:117] "RemoveContainer" containerID="6c07ed3a23b9e90ec44e0b3107713fd9bedf485fa1562b3bc3a266ffc1308b7c" Oct 08 21:07:06 crc kubenswrapper[4988]: E1008 21:07:06.238366 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nm54f_openshift-machine-config-operator(f328a20b-5283-407e-8a93-2ab2da599f18)\"" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" Oct 08 21:07:12 crc kubenswrapper[4988]: I1008 21:07:12.370355 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-9ql2l"] Oct 08 21:07:12 crc kubenswrapper[4988]: E1008 21:07:12.371801 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2eb2a0fe-0016-4372-abff-484b668e9fc0" containerName="extract-utilities" Oct 08 21:07:12 crc kubenswrapper[4988]: I1008 21:07:12.371825 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="2eb2a0fe-0016-4372-abff-484b668e9fc0" containerName="extract-utilities" Oct 08 21:07:12 crc kubenswrapper[4988]: E1008 21:07:12.371848 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2eb2a0fe-0016-4372-abff-484b668e9fc0" containerName="extract-content" Oct 08 21:07:12 crc kubenswrapper[4988]: I1008 21:07:12.371858 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="2eb2a0fe-0016-4372-abff-484b668e9fc0" containerName="extract-content" Oct 08 21:07:12 crc kubenswrapper[4988]: E1008 21:07:12.371902 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2eb2a0fe-0016-4372-abff-484b668e9fc0" containerName="registry-server" Oct 08 21:07:12 crc kubenswrapper[4988]: I1008 21:07:12.371914 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="2eb2a0fe-0016-4372-abff-484b668e9fc0" containerName="registry-server" Oct 08 21:07:12 crc kubenswrapper[4988]: I1008 21:07:12.372317 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="2eb2a0fe-0016-4372-abff-484b668e9fc0" containerName="registry-server" Oct 08 21:07:12 crc kubenswrapper[4988]: I1008 21:07:12.374971 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-9ql2l" Oct 08 21:07:12 crc kubenswrapper[4988]: I1008 21:07:12.384367 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-9ql2l"] Oct 08 21:07:12 crc kubenswrapper[4988]: I1008 21:07:12.569800 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1bfd72ab-365f-483a-9ecd-3db46219eed9-catalog-content\") pod \"redhat-marketplace-9ql2l\" (UID: \"1bfd72ab-365f-483a-9ecd-3db46219eed9\") " pod="openshift-marketplace/redhat-marketplace-9ql2l" Oct 08 21:07:12 crc kubenswrapper[4988]: I1008 21:07:12.569895 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fxc7d\" (UniqueName: \"kubernetes.io/projected/1bfd72ab-365f-483a-9ecd-3db46219eed9-kube-api-access-fxc7d\") pod \"redhat-marketplace-9ql2l\" (UID: \"1bfd72ab-365f-483a-9ecd-3db46219eed9\") " pod="openshift-marketplace/redhat-marketplace-9ql2l" Oct 08 21:07:12 crc kubenswrapper[4988]: I1008 21:07:12.569926 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1bfd72ab-365f-483a-9ecd-3db46219eed9-utilities\") pod \"redhat-marketplace-9ql2l\" (UID: \"1bfd72ab-365f-483a-9ecd-3db46219eed9\") " pod="openshift-marketplace/redhat-marketplace-9ql2l" Oct 08 21:07:12 crc kubenswrapper[4988]: I1008 21:07:12.672259 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1bfd72ab-365f-483a-9ecd-3db46219eed9-catalog-content\") pod \"redhat-marketplace-9ql2l\" (UID: \"1bfd72ab-365f-483a-9ecd-3db46219eed9\") " pod="openshift-marketplace/redhat-marketplace-9ql2l" Oct 08 21:07:12 crc kubenswrapper[4988]: I1008 21:07:12.672395 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fxc7d\" (UniqueName: \"kubernetes.io/projected/1bfd72ab-365f-483a-9ecd-3db46219eed9-kube-api-access-fxc7d\") pod \"redhat-marketplace-9ql2l\" (UID: \"1bfd72ab-365f-483a-9ecd-3db46219eed9\") " pod="openshift-marketplace/redhat-marketplace-9ql2l" Oct 08 21:07:12 crc kubenswrapper[4988]: I1008 21:07:12.672427 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1bfd72ab-365f-483a-9ecd-3db46219eed9-utilities\") pod \"redhat-marketplace-9ql2l\" (UID: \"1bfd72ab-365f-483a-9ecd-3db46219eed9\") " pod="openshift-marketplace/redhat-marketplace-9ql2l" Oct 08 21:07:12 crc kubenswrapper[4988]: I1008 21:07:12.672830 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1bfd72ab-365f-483a-9ecd-3db46219eed9-catalog-content\") pod \"redhat-marketplace-9ql2l\" (UID: \"1bfd72ab-365f-483a-9ecd-3db46219eed9\") " pod="openshift-marketplace/redhat-marketplace-9ql2l" Oct 08 21:07:12 crc kubenswrapper[4988]: I1008 21:07:12.672937 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1bfd72ab-365f-483a-9ecd-3db46219eed9-utilities\") pod \"redhat-marketplace-9ql2l\" (UID: \"1bfd72ab-365f-483a-9ecd-3db46219eed9\") " pod="openshift-marketplace/redhat-marketplace-9ql2l" Oct 08 21:07:12 crc kubenswrapper[4988]: I1008 21:07:12.695515 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fxc7d\" (UniqueName: \"kubernetes.io/projected/1bfd72ab-365f-483a-9ecd-3db46219eed9-kube-api-access-fxc7d\") pod \"redhat-marketplace-9ql2l\" (UID: \"1bfd72ab-365f-483a-9ecd-3db46219eed9\") " pod="openshift-marketplace/redhat-marketplace-9ql2l" Oct 08 21:07:12 crc kubenswrapper[4988]: I1008 21:07:12.707700 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-9ql2l" Oct 08 21:07:13 crc kubenswrapper[4988]: I1008 21:07:13.231036 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-9ql2l"] Oct 08 21:07:14 crc kubenswrapper[4988]: I1008 21:07:14.137576 4988 generic.go:334] "Generic (PLEG): container finished" podID="1bfd72ab-365f-483a-9ecd-3db46219eed9" containerID="53d685f1941bcdd50ef63c822c8bb7d40c79f09ac01f4e0c6b8eae5b4bd359ce" exitCode=0 Oct 08 21:07:14 crc kubenswrapper[4988]: I1008 21:07:14.137659 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-9ql2l" event={"ID":"1bfd72ab-365f-483a-9ecd-3db46219eed9","Type":"ContainerDied","Data":"53d685f1941bcdd50ef63c822c8bb7d40c79f09ac01f4e0c6b8eae5b4bd359ce"} Oct 08 21:07:14 crc kubenswrapper[4988]: I1008 21:07:14.137883 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-9ql2l" event={"ID":"1bfd72ab-365f-483a-9ecd-3db46219eed9","Type":"ContainerStarted","Data":"2ee6c215314d67114e15d9b03a1c1b13df7c39a4706162b7f175161d7f205aca"} Oct 08 21:07:16 crc kubenswrapper[4988]: I1008 21:07:16.161103 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-9ql2l" event={"ID":"1bfd72ab-365f-483a-9ecd-3db46219eed9","Type":"ContainerStarted","Data":"1b7be8d0ab343d65a145868a8e40da2de086399913a7e03ad0576d50aea2ef11"} Oct 08 21:07:17 crc kubenswrapper[4988]: I1008 21:07:17.127491 4988 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-68d546b9d8-kxkcv_dcae2201-01d1-4f23-bf2d-134d6c01a1a4/kube-rbac-proxy/0.log" Oct 08 21:07:17 crc kubenswrapper[4988]: I1008 21:07:17.175774 4988 generic.go:334] "Generic (PLEG): container finished" podID="1bfd72ab-365f-483a-9ecd-3db46219eed9" containerID="1b7be8d0ab343d65a145868a8e40da2de086399913a7e03ad0576d50aea2ef11" exitCode=0 Oct 08 21:07:17 crc kubenswrapper[4988]: I1008 21:07:17.175827 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-9ql2l" event={"ID":"1bfd72ab-365f-483a-9ecd-3db46219eed9","Type":"ContainerDied","Data":"1b7be8d0ab343d65a145868a8e40da2de086399913a7e03ad0576d50aea2ef11"} Oct 08 21:07:17 crc kubenswrapper[4988]: I1008 21:07:17.319779 4988 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-webhook-server-64bf5d555-66f55_5e1113b9-5d57-427e-9c92-0da0866e8e4d/frr-k8s-webhook-server/0.log" Oct 08 21:07:17 crc kubenswrapper[4988]: I1008 21:07:17.508904 4988 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-xfwvv_f350c4ca-de05-4509-8c61-1941081b8b79/cp-frr-files/0.log" Oct 08 21:07:17 crc kubenswrapper[4988]: I1008 21:07:17.524460 4988 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-68d546b9d8-kxkcv_dcae2201-01d1-4f23-bf2d-134d6c01a1a4/controller/0.log" Oct 08 21:07:17 crc kubenswrapper[4988]: I1008 21:07:17.686752 4988 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-xfwvv_f350c4ca-de05-4509-8c61-1941081b8b79/cp-reloader/0.log" Oct 08 21:07:17 crc kubenswrapper[4988]: I1008 21:07:17.694483 4988 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-xfwvv_f350c4ca-de05-4509-8c61-1941081b8b79/cp-metrics/0.log" Oct 08 21:07:17 crc kubenswrapper[4988]: I1008 21:07:17.729547 4988 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-xfwvv_f350c4ca-de05-4509-8c61-1941081b8b79/cp-frr-files/0.log" Oct 08 21:07:17 crc kubenswrapper[4988]: I1008 21:07:17.780658 4988 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-xfwvv_f350c4ca-de05-4509-8c61-1941081b8b79/cp-reloader/0.log" Oct 08 21:07:17 crc kubenswrapper[4988]: I1008 21:07:17.902583 4988 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-xfwvv_f350c4ca-de05-4509-8c61-1941081b8b79/cp-reloader/0.log" Oct 08 21:07:17 crc kubenswrapper[4988]: I1008 21:07:17.907628 4988 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-xfwvv_f350c4ca-de05-4509-8c61-1941081b8b79/cp-frr-files/0.log" Oct 08 21:07:17 crc kubenswrapper[4988]: I1008 21:07:17.941321 4988 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-xfwvv_f350c4ca-de05-4509-8c61-1941081b8b79/cp-metrics/0.log" Oct 08 21:07:18 crc kubenswrapper[4988]: I1008 21:07:18.002998 4988 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-xfwvv_f350c4ca-de05-4509-8c61-1941081b8b79/cp-metrics/0.log" Oct 08 21:07:18 crc kubenswrapper[4988]: I1008 21:07:18.186663 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-9ql2l" event={"ID":"1bfd72ab-365f-483a-9ecd-3db46219eed9","Type":"ContainerStarted","Data":"792b3a0b87979453b2123549efb0ccea8cc64a257d39971f05377f058a872cc7"} Oct 08 21:07:18 crc kubenswrapper[4988]: I1008 21:07:18.200850 4988 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-xfwvv_f350c4ca-de05-4509-8c61-1941081b8b79/cp-reloader/0.log" Oct 08 21:07:18 crc kubenswrapper[4988]: I1008 21:07:18.204745 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-9ql2l" podStartSLOduration=2.7041038569999998 podStartE2EDuration="6.204726652s" podCreationTimestamp="2025-10-08 21:07:12 +0000 UTC" firstStartedPulling="2025-10-08 21:07:14.141185675 +0000 UTC m=+10579.591028465" lastFinishedPulling="2025-10-08 21:07:17.64180849 +0000 UTC m=+10583.091651260" observedRunningTime="2025-10-08 21:07:18.203810512 +0000 UTC m=+10583.653653302" watchObservedRunningTime="2025-10-08 21:07:18.204726652 +0000 UTC m=+10583.654569422" Oct 08 21:07:18 crc kubenswrapper[4988]: I1008 21:07:18.227554 4988 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-xfwvv_f350c4ca-de05-4509-8c61-1941081b8b79/controller/0.log" Oct 08 21:07:18 crc kubenswrapper[4988]: I1008 21:07:18.259178 4988 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-xfwvv_f350c4ca-de05-4509-8c61-1941081b8b79/cp-metrics/0.log" Oct 08 21:07:18 crc kubenswrapper[4988]: I1008 21:07:18.265621 4988 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-xfwvv_f350c4ca-de05-4509-8c61-1941081b8b79/cp-frr-files/0.log" Oct 08 21:07:18 crc kubenswrapper[4988]: I1008 21:07:18.387962 4988 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-xfwvv_f350c4ca-de05-4509-8c61-1941081b8b79/frr-metrics/0.log" Oct 08 21:07:18 crc kubenswrapper[4988]: I1008 21:07:18.471773 4988 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-xfwvv_f350c4ca-de05-4509-8c61-1941081b8b79/kube-rbac-proxy-frr/0.log" Oct 08 21:07:18 crc kubenswrapper[4988]: I1008 21:07:18.518688 4988 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-xfwvv_f350c4ca-de05-4509-8c61-1941081b8b79/kube-rbac-proxy/0.log" Oct 08 21:07:18 crc kubenswrapper[4988]: I1008 21:07:18.617726 4988 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-xfwvv_f350c4ca-de05-4509-8c61-1941081b8b79/reloader/0.log" Oct 08 21:07:18 crc kubenswrapper[4988]: I1008 21:07:18.876917 4988 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-controller-manager-55b79cb758-ld9ht_4cc757e2-0271-4785-830f-00aa42bd5a6d/manager/0.log" Oct 08 21:07:18 crc kubenswrapper[4988]: I1008 21:07:18.971327 4988 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-webhook-server-6585b9b6b6-zpjq5_11101830-713e-483b-bfae-699e80dbde14/webhook-server/0.log" Oct 08 21:07:19 crc kubenswrapper[4988]: I1008 21:07:19.093985 4988 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-nbhjx_645996a1-4e9d-471e-ad3f-866864472407/kube-rbac-proxy/0.log" Oct 08 21:07:20 crc kubenswrapper[4988]: I1008 21:07:20.102040 4988 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-nbhjx_645996a1-4e9d-471e-ad3f-866864472407/speaker/0.log" Oct 08 21:07:20 crc kubenswrapper[4988]: I1008 21:07:20.238467 4988 scope.go:117] "RemoveContainer" containerID="6c07ed3a23b9e90ec44e0b3107713fd9bedf485fa1562b3bc3a266ffc1308b7c" Oct 08 21:07:20 crc kubenswrapper[4988]: E1008 21:07:20.238715 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nm54f_openshift-machine-config-operator(f328a20b-5283-407e-8a93-2ab2da599f18)\"" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" Oct 08 21:07:21 crc kubenswrapper[4988]: I1008 21:07:21.963066 4988 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-xfwvv_f350c4ca-de05-4509-8c61-1941081b8b79/frr/0.log" Oct 08 21:07:22 crc kubenswrapper[4988]: I1008 21:07:22.707811 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-9ql2l" Oct 08 21:07:22 crc kubenswrapper[4988]: I1008 21:07:22.708130 4988 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-9ql2l" Oct 08 21:07:22 crc kubenswrapper[4988]: I1008 21:07:22.766573 4988 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-9ql2l" Oct 08 21:07:24 crc kubenswrapper[4988]: I1008 21:07:24.149137 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-9ql2l" Oct 08 21:07:24 crc kubenswrapper[4988]: I1008 21:07:24.213098 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-9ql2l"] Oct 08 21:07:25 crc kubenswrapper[4988]: I1008 21:07:25.260046 4988 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-9ql2l" podUID="1bfd72ab-365f-483a-9ecd-3db46219eed9" containerName="registry-server" containerID="cri-o://792b3a0b87979453b2123549efb0ccea8cc64a257d39971f05377f058a872cc7" gracePeriod=2 Oct 08 21:07:25 crc kubenswrapper[4988]: I1008 21:07:25.771632 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-9ql2l" Oct 08 21:07:25 crc kubenswrapper[4988]: I1008 21:07:25.865872 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1bfd72ab-365f-483a-9ecd-3db46219eed9-catalog-content\") pod \"1bfd72ab-365f-483a-9ecd-3db46219eed9\" (UID: \"1bfd72ab-365f-483a-9ecd-3db46219eed9\") " Oct 08 21:07:25 crc kubenswrapper[4988]: I1008 21:07:25.865925 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1bfd72ab-365f-483a-9ecd-3db46219eed9-utilities\") pod \"1bfd72ab-365f-483a-9ecd-3db46219eed9\" (UID: \"1bfd72ab-365f-483a-9ecd-3db46219eed9\") " Oct 08 21:07:25 crc kubenswrapper[4988]: I1008 21:07:25.865972 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fxc7d\" (UniqueName: \"kubernetes.io/projected/1bfd72ab-365f-483a-9ecd-3db46219eed9-kube-api-access-fxc7d\") pod \"1bfd72ab-365f-483a-9ecd-3db46219eed9\" (UID: \"1bfd72ab-365f-483a-9ecd-3db46219eed9\") " Oct 08 21:07:25 crc kubenswrapper[4988]: I1008 21:07:25.867913 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1bfd72ab-365f-483a-9ecd-3db46219eed9-utilities" (OuterVolumeSpecName: "utilities") pod "1bfd72ab-365f-483a-9ecd-3db46219eed9" (UID: "1bfd72ab-365f-483a-9ecd-3db46219eed9"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 21:07:25 crc kubenswrapper[4988]: I1008 21:07:25.876193 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1bfd72ab-365f-483a-9ecd-3db46219eed9-kube-api-access-fxc7d" (OuterVolumeSpecName: "kube-api-access-fxc7d") pod "1bfd72ab-365f-483a-9ecd-3db46219eed9" (UID: "1bfd72ab-365f-483a-9ecd-3db46219eed9"). InnerVolumeSpecName "kube-api-access-fxc7d". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 21:07:25 crc kubenswrapper[4988]: I1008 21:07:25.883452 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1bfd72ab-365f-483a-9ecd-3db46219eed9-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1bfd72ab-365f-483a-9ecd-3db46219eed9" (UID: "1bfd72ab-365f-483a-9ecd-3db46219eed9"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 21:07:25 crc kubenswrapper[4988]: I1008 21:07:25.968293 4988 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1bfd72ab-365f-483a-9ecd-3db46219eed9-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 08 21:07:25 crc kubenswrapper[4988]: I1008 21:07:25.968701 4988 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1bfd72ab-365f-483a-9ecd-3db46219eed9-utilities\") on node \"crc\" DevicePath \"\"" Oct 08 21:07:25 crc kubenswrapper[4988]: I1008 21:07:25.968728 4988 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fxc7d\" (UniqueName: \"kubernetes.io/projected/1bfd72ab-365f-483a-9ecd-3db46219eed9-kube-api-access-fxc7d\") on node \"crc\" DevicePath \"\"" Oct 08 21:07:26 crc kubenswrapper[4988]: I1008 21:07:26.275911 4988 generic.go:334] "Generic (PLEG): container finished" podID="1bfd72ab-365f-483a-9ecd-3db46219eed9" containerID="792b3a0b87979453b2123549efb0ccea8cc64a257d39971f05377f058a872cc7" exitCode=0 Oct 08 21:07:26 crc kubenswrapper[4988]: I1008 21:07:26.275970 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-9ql2l" event={"ID":"1bfd72ab-365f-483a-9ecd-3db46219eed9","Type":"ContainerDied","Data":"792b3a0b87979453b2123549efb0ccea8cc64a257d39971f05377f058a872cc7"} Oct 08 21:07:26 crc kubenswrapper[4988]: I1008 21:07:26.276004 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-9ql2l" event={"ID":"1bfd72ab-365f-483a-9ecd-3db46219eed9","Type":"ContainerDied","Data":"2ee6c215314d67114e15d9b03a1c1b13df7c39a4706162b7f175161d7f205aca"} Oct 08 21:07:26 crc kubenswrapper[4988]: I1008 21:07:26.276028 4988 scope.go:117] "RemoveContainer" containerID="792b3a0b87979453b2123549efb0ccea8cc64a257d39971f05377f058a872cc7" Oct 08 21:07:26 crc kubenswrapper[4988]: I1008 21:07:26.277597 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-9ql2l" Oct 08 21:07:26 crc kubenswrapper[4988]: I1008 21:07:26.667005 4988 scope.go:117] "RemoveContainer" containerID="1b7be8d0ab343d65a145868a8e40da2de086399913a7e03ad0576d50aea2ef11" Oct 08 21:07:26 crc kubenswrapper[4988]: I1008 21:07:26.685776 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-9ql2l"] Oct 08 21:07:26 crc kubenswrapper[4988]: I1008 21:07:26.694905 4988 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-9ql2l"] Oct 08 21:07:26 crc kubenswrapper[4988]: I1008 21:07:26.796871 4988 scope.go:117] "RemoveContainer" containerID="53d685f1941bcdd50ef63c822c8bb7d40c79f09ac01f4e0c6b8eae5b4bd359ce" Oct 08 21:07:26 crc kubenswrapper[4988]: I1008 21:07:26.818836 4988 scope.go:117] "RemoveContainer" containerID="792b3a0b87979453b2123549efb0ccea8cc64a257d39971f05377f058a872cc7" Oct 08 21:07:26 crc kubenswrapper[4988]: E1008 21:07:26.819290 4988 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"792b3a0b87979453b2123549efb0ccea8cc64a257d39971f05377f058a872cc7\": container with ID starting with 792b3a0b87979453b2123549efb0ccea8cc64a257d39971f05377f058a872cc7 not found: ID does not exist" containerID="792b3a0b87979453b2123549efb0ccea8cc64a257d39971f05377f058a872cc7" Oct 08 21:07:26 crc kubenswrapper[4988]: I1008 21:07:26.819339 4988 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"792b3a0b87979453b2123549efb0ccea8cc64a257d39971f05377f058a872cc7"} err="failed to get container status \"792b3a0b87979453b2123549efb0ccea8cc64a257d39971f05377f058a872cc7\": rpc error: code = NotFound desc = could not find container \"792b3a0b87979453b2123549efb0ccea8cc64a257d39971f05377f058a872cc7\": container with ID starting with 792b3a0b87979453b2123549efb0ccea8cc64a257d39971f05377f058a872cc7 not found: ID does not exist" Oct 08 21:07:26 crc kubenswrapper[4988]: I1008 21:07:26.819365 4988 scope.go:117] "RemoveContainer" containerID="1b7be8d0ab343d65a145868a8e40da2de086399913a7e03ad0576d50aea2ef11" Oct 08 21:07:26 crc kubenswrapper[4988]: E1008 21:07:26.820135 4988 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1b7be8d0ab343d65a145868a8e40da2de086399913a7e03ad0576d50aea2ef11\": container with ID starting with 1b7be8d0ab343d65a145868a8e40da2de086399913a7e03ad0576d50aea2ef11 not found: ID does not exist" containerID="1b7be8d0ab343d65a145868a8e40da2de086399913a7e03ad0576d50aea2ef11" Oct 08 21:07:26 crc kubenswrapper[4988]: I1008 21:07:26.820170 4988 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1b7be8d0ab343d65a145868a8e40da2de086399913a7e03ad0576d50aea2ef11"} err="failed to get container status \"1b7be8d0ab343d65a145868a8e40da2de086399913a7e03ad0576d50aea2ef11\": rpc error: code = NotFound desc = could not find container \"1b7be8d0ab343d65a145868a8e40da2de086399913a7e03ad0576d50aea2ef11\": container with ID starting with 1b7be8d0ab343d65a145868a8e40da2de086399913a7e03ad0576d50aea2ef11 not found: ID does not exist" Oct 08 21:07:26 crc kubenswrapper[4988]: I1008 21:07:26.820191 4988 scope.go:117] "RemoveContainer" containerID="53d685f1941bcdd50ef63c822c8bb7d40c79f09ac01f4e0c6b8eae5b4bd359ce" Oct 08 21:07:26 crc kubenswrapper[4988]: E1008 21:07:26.820709 4988 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"53d685f1941bcdd50ef63c822c8bb7d40c79f09ac01f4e0c6b8eae5b4bd359ce\": container with ID starting with 53d685f1941bcdd50ef63c822c8bb7d40c79f09ac01f4e0c6b8eae5b4bd359ce not found: ID does not exist" containerID="53d685f1941bcdd50ef63c822c8bb7d40c79f09ac01f4e0c6b8eae5b4bd359ce" Oct 08 21:07:26 crc kubenswrapper[4988]: I1008 21:07:26.820740 4988 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"53d685f1941bcdd50ef63c822c8bb7d40c79f09ac01f4e0c6b8eae5b4bd359ce"} err="failed to get container status \"53d685f1941bcdd50ef63c822c8bb7d40c79f09ac01f4e0c6b8eae5b4bd359ce\": rpc error: code = NotFound desc = could not find container \"53d685f1941bcdd50ef63c822c8bb7d40c79f09ac01f4e0c6b8eae5b4bd359ce\": container with ID starting with 53d685f1941bcdd50ef63c822c8bb7d40c79f09ac01f4e0c6b8eae5b4bd359ce not found: ID does not exist" Oct 08 21:07:27 crc kubenswrapper[4988]: I1008 21:07:27.269684 4988 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1bfd72ab-365f-483a-9ecd-3db46219eed9" path="/var/lib/kubelet/pods/1bfd72ab-365f-483a-9ecd-3db46219eed9/volumes" Oct 08 21:07:34 crc kubenswrapper[4988]: I1008 21:07:34.237911 4988 scope.go:117] "RemoveContainer" containerID="6c07ed3a23b9e90ec44e0b3107713fd9bedf485fa1562b3bc3a266ffc1308b7c" Oct 08 21:07:34 crc kubenswrapper[4988]: E1008 21:07:34.238571 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nm54f_openshift-machine-config-operator(f328a20b-5283-407e-8a93-2ab2da599f18)\"" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" Oct 08 21:07:34 crc kubenswrapper[4988]: I1008 21:07:34.780465 4988 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69rf72h_86be9f4e-06b6-4024-a098-820172b202db/util/0.log" Oct 08 21:07:34 crc kubenswrapper[4988]: I1008 21:07:34.954883 4988 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69rf72h_86be9f4e-06b6-4024-a098-820172b202db/util/0.log" Oct 08 21:07:34 crc kubenswrapper[4988]: I1008 21:07:34.977241 4988 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69rf72h_86be9f4e-06b6-4024-a098-820172b202db/pull/0.log" Oct 08 21:07:34 crc kubenswrapper[4988]: I1008 21:07:34.985586 4988 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69rf72h_86be9f4e-06b6-4024-a098-820172b202db/pull/0.log" Oct 08 21:07:35 crc kubenswrapper[4988]: I1008 21:07:35.113442 4988 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69rf72h_86be9f4e-06b6-4024-a098-820172b202db/util/0.log" Oct 08 21:07:35 crc kubenswrapper[4988]: I1008 21:07:35.211868 4988 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69rf72h_86be9f4e-06b6-4024-a098-820172b202db/extract/0.log" Oct 08 21:07:35 crc kubenswrapper[4988]: I1008 21:07:35.223900 4988 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69rf72h_86be9f4e-06b6-4024-a098-820172b202db/pull/0.log" Oct 08 21:07:35 crc kubenswrapper[4988]: I1008 21:07:35.325682 4988 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d28pl68_fdfe229e-0b1f-4af5-8c9b-2d1ed4c64fcb/util/0.log" Oct 08 21:07:35 crc kubenswrapper[4988]: I1008 21:07:35.464275 4988 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d28pl68_fdfe229e-0b1f-4af5-8c9b-2d1ed4c64fcb/util/0.log" Oct 08 21:07:35 crc kubenswrapper[4988]: I1008 21:07:35.483289 4988 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d28pl68_fdfe229e-0b1f-4af5-8c9b-2d1ed4c64fcb/pull/0.log" Oct 08 21:07:35 crc kubenswrapper[4988]: I1008 21:07:35.529980 4988 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d28pl68_fdfe229e-0b1f-4af5-8c9b-2d1ed4c64fcb/pull/0.log" Oct 08 21:07:35 crc kubenswrapper[4988]: I1008 21:07:35.681954 4988 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d28pl68_fdfe229e-0b1f-4af5-8c9b-2d1ed4c64fcb/util/0.log" Oct 08 21:07:35 crc kubenswrapper[4988]: I1008 21:07:35.722157 4988 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d28pl68_fdfe229e-0b1f-4af5-8c9b-2d1ed4c64fcb/pull/0.log" Oct 08 21:07:35 crc kubenswrapper[4988]: I1008 21:07:35.737070 4988 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d28pl68_fdfe229e-0b1f-4af5-8c9b-2d1ed4c64fcb/extract/0.log" Oct 08 21:07:35 crc kubenswrapper[4988]: I1008 21:07:35.904048 4988 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2djwvff_e7b36ee1-53f7-46fc-a92c-bdf893239e77/util/0.log" Oct 08 21:07:36 crc kubenswrapper[4988]: I1008 21:07:36.058195 4988 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2djwvff_e7b36ee1-53f7-46fc-a92c-bdf893239e77/pull/0.log" Oct 08 21:07:36 crc kubenswrapper[4988]: I1008 21:07:36.085093 4988 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2djwvff_e7b36ee1-53f7-46fc-a92c-bdf893239e77/pull/0.log" Oct 08 21:07:36 crc kubenswrapper[4988]: I1008 21:07:36.098040 4988 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2djwvff_e7b36ee1-53f7-46fc-a92c-bdf893239e77/util/0.log" Oct 08 21:07:36 crc kubenswrapper[4988]: I1008 21:07:36.257732 4988 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2djwvff_e7b36ee1-53f7-46fc-a92c-bdf893239e77/extract/0.log" Oct 08 21:07:36 crc kubenswrapper[4988]: I1008 21:07:36.275802 4988 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2djwvff_e7b36ee1-53f7-46fc-a92c-bdf893239e77/pull/0.log" Oct 08 21:07:36 crc kubenswrapper[4988]: I1008 21:07:36.305063 4988 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2djwvff_e7b36ee1-53f7-46fc-a92c-bdf893239e77/util/0.log" Oct 08 21:07:36 crc kubenswrapper[4988]: I1008 21:07:36.886659 4988 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-c8g26_6d4ae11d-cd07-4151-ac5d-70100ef241ec/extract-utilities/0.log" Oct 08 21:07:37 crc kubenswrapper[4988]: I1008 21:07:37.091068 4988 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-c8g26_6d4ae11d-cd07-4151-ac5d-70100ef241ec/extract-utilities/0.log" Oct 08 21:07:37 crc kubenswrapper[4988]: I1008 21:07:37.094739 4988 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-c8g26_6d4ae11d-cd07-4151-ac5d-70100ef241ec/extract-content/0.log" Oct 08 21:07:37 crc kubenswrapper[4988]: I1008 21:07:37.112756 4988 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-c8g26_6d4ae11d-cd07-4151-ac5d-70100ef241ec/extract-content/0.log" Oct 08 21:07:37 crc kubenswrapper[4988]: I1008 21:07:37.255201 4988 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-c8g26_6d4ae11d-cd07-4151-ac5d-70100ef241ec/extract-utilities/0.log" Oct 08 21:07:37 crc kubenswrapper[4988]: I1008 21:07:37.331076 4988 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-c8g26_6d4ae11d-cd07-4151-ac5d-70100ef241ec/extract-content/0.log" Oct 08 21:07:37 crc kubenswrapper[4988]: I1008 21:07:37.465345 4988 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-zmb2k_83e5c288-f6e0-4ce6-a888-c7adbdb3e7e1/extract-utilities/0.log" Oct 08 21:07:37 crc kubenswrapper[4988]: I1008 21:07:37.716901 4988 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-zmb2k_83e5c288-f6e0-4ce6-a888-c7adbdb3e7e1/extract-content/0.log" Oct 08 21:07:37 crc kubenswrapper[4988]: I1008 21:07:37.727980 4988 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-zmb2k_83e5c288-f6e0-4ce6-a888-c7adbdb3e7e1/extract-utilities/0.log" Oct 08 21:07:37 crc kubenswrapper[4988]: I1008 21:07:37.744589 4988 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-zmb2k_83e5c288-f6e0-4ce6-a888-c7adbdb3e7e1/extract-content/0.log" Oct 08 21:07:37 crc kubenswrapper[4988]: I1008 21:07:37.966624 4988 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-zmb2k_83e5c288-f6e0-4ce6-a888-c7adbdb3e7e1/extract-content/0.log" Oct 08 21:07:37 crc kubenswrapper[4988]: I1008 21:07:37.998312 4988 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-zmb2k_83e5c288-f6e0-4ce6-a888-c7adbdb3e7e1/extract-utilities/0.log" Oct 08 21:07:38 crc kubenswrapper[4988]: I1008 21:07:38.260555 4988 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c55jx7_ffc18ede-f7bc-450c-b37a-b5dcf72a1a50/util/0.log" Oct 08 21:07:38 crc kubenswrapper[4988]: I1008 21:07:38.990017 4988 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-c8g26_6d4ae11d-cd07-4151-ac5d-70100ef241ec/registry-server/0.log" Oct 08 21:07:38 crc kubenswrapper[4988]: I1008 21:07:38.995363 4988 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c55jx7_ffc18ede-f7bc-450c-b37a-b5dcf72a1a50/util/0.log" Oct 08 21:07:39 crc kubenswrapper[4988]: I1008 21:07:39.059342 4988 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c55jx7_ffc18ede-f7bc-450c-b37a-b5dcf72a1a50/pull/0.log" Oct 08 21:07:39 crc kubenswrapper[4988]: I1008 21:07:39.127072 4988 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c55jx7_ffc18ede-f7bc-450c-b37a-b5dcf72a1a50/pull/0.log" Oct 08 21:07:39 crc kubenswrapper[4988]: I1008 21:07:39.317806 4988 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c55jx7_ffc18ede-f7bc-450c-b37a-b5dcf72a1a50/util/0.log" Oct 08 21:07:39 crc kubenswrapper[4988]: I1008 21:07:39.384686 4988 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c55jx7_ffc18ede-f7bc-450c-b37a-b5dcf72a1a50/extract/0.log" Oct 08 21:07:39 crc kubenswrapper[4988]: I1008 21:07:39.426909 4988 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c55jx7_ffc18ede-f7bc-450c-b37a-b5dcf72a1a50/pull/0.log" Oct 08 21:07:39 crc kubenswrapper[4988]: I1008 21:07:39.575172 4988 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-kb5xn_590d218c-e870-4304-b2b6-3c0fa7cc7c7a/marketplace-operator/0.log" Oct 08 21:07:39 crc kubenswrapper[4988]: I1008 21:07:39.590908 4988 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-zmb2k_83e5c288-f6e0-4ce6-a888-c7adbdb3e7e1/registry-server/0.log" Oct 08 21:07:39 crc kubenswrapper[4988]: I1008 21:07:39.655644 4988 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-gl4lh_6936c4e6-f11e-47a2-870a-0df3df4ed33c/extract-utilities/0.log" Oct 08 21:07:39 crc kubenswrapper[4988]: I1008 21:07:39.841278 4988 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-gl4lh_6936c4e6-f11e-47a2-870a-0df3df4ed33c/extract-content/0.log" Oct 08 21:07:39 crc kubenswrapper[4988]: I1008 21:07:39.846339 4988 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-gl4lh_6936c4e6-f11e-47a2-870a-0df3df4ed33c/extract-utilities/0.log" Oct 08 21:07:39 crc kubenswrapper[4988]: I1008 21:07:39.846578 4988 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-gl4lh_6936c4e6-f11e-47a2-870a-0df3df4ed33c/extract-content/0.log" Oct 08 21:07:40 crc kubenswrapper[4988]: I1008 21:07:40.064182 4988 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-gl4lh_6936c4e6-f11e-47a2-870a-0df3df4ed33c/extract-utilities/0.log" Oct 08 21:07:40 crc kubenswrapper[4988]: I1008 21:07:40.072435 4988 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-gl4lh_6936c4e6-f11e-47a2-870a-0df3df4ed33c/extract-content/0.log" Oct 08 21:07:40 crc kubenswrapper[4988]: I1008 21:07:40.099671 4988 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-vwv6l_4b10186d-3d69-4998-9162-89df44da5521/extract-utilities/0.log" Oct 08 21:07:40 crc kubenswrapper[4988]: I1008 21:07:40.248824 4988 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-vwv6l_4b10186d-3d69-4998-9162-89df44da5521/extract-content/0.log" Oct 08 21:07:40 crc kubenswrapper[4988]: I1008 21:07:40.282176 4988 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-vwv6l_4b10186d-3d69-4998-9162-89df44da5521/extract-utilities/0.log" Oct 08 21:07:40 crc kubenswrapper[4988]: I1008 21:07:40.334193 4988 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-vwv6l_4b10186d-3d69-4998-9162-89df44da5521/extract-content/0.log" Oct 08 21:07:40 crc kubenswrapper[4988]: I1008 21:07:40.429185 4988 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-vwv6l_4b10186d-3d69-4998-9162-89df44da5521/extract-utilities/0.log" Oct 08 21:07:40 crc kubenswrapper[4988]: I1008 21:07:40.435757 4988 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-gl4lh_6936c4e6-f11e-47a2-870a-0df3df4ed33c/registry-server/0.log" Oct 08 21:07:40 crc kubenswrapper[4988]: I1008 21:07:40.518848 4988 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-vwv6l_4b10186d-3d69-4998-9162-89df44da5521/extract-content/0.log" Oct 08 21:07:41 crc kubenswrapper[4988]: I1008 21:07:41.437537 4988 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-vwv6l_4b10186d-3d69-4998-9162-89df44da5521/registry-server/0.log" Oct 08 21:07:46 crc kubenswrapper[4988]: I1008 21:07:46.238789 4988 scope.go:117] "RemoveContainer" containerID="6c07ed3a23b9e90ec44e0b3107713fd9bedf485fa1562b3bc3a266ffc1308b7c" Oct 08 21:07:46 crc kubenswrapper[4988]: E1008 21:07:46.239869 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nm54f_openshift-machine-config-operator(f328a20b-5283-407e-8a93-2ab2da599f18)\"" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" Oct 08 21:07:54 crc kubenswrapper[4988]: I1008 21:07:54.107780 4988 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-7c8cf85677-pwbvx_38d89d99-8b9b-43e1-85ad-d928aad3d3d1/prometheus-operator/0.log" Oct 08 21:07:54 crc kubenswrapper[4988]: I1008 21:07:54.231593 4988 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-admission-webhook-6b789f9dcf-cs98t_d625ac6d-4570-4a73-8e1f-14ea64510864/prometheus-operator-admission-webhook/0.log" Oct 08 21:07:54 crc kubenswrapper[4988]: I1008 21:07:54.286060 4988 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-admission-webhook-6b789f9dcf-d7jwn_3a9085fc-8d56-486f-a107-3e03f9ae0e64/prometheus-operator-admission-webhook/0.log" Oct 08 21:07:54 crc kubenswrapper[4988]: I1008 21:07:54.453525 4988 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_observability-operator-cc5f78dfc-7qjkj_86cf1f5c-ab13-4b35-ad2f-9529ccdf96db/operator/0.log" Oct 08 21:07:54 crc kubenswrapper[4988]: I1008 21:07:54.538977 4988 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_perses-operator-54bc95c9fb-ln8dz_8cf809bc-9b01-4426-a099-be47b7b711ba/perses-operator/0.log" Oct 08 21:08:00 crc kubenswrapper[4988]: I1008 21:08:00.237906 4988 scope.go:117] "RemoveContainer" containerID="6c07ed3a23b9e90ec44e0b3107713fd9bedf485fa1562b3bc3a266ffc1308b7c" Oct 08 21:08:00 crc kubenswrapper[4988]: E1008 21:08:00.238767 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nm54f_openshift-machine-config-operator(f328a20b-5283-407e-8a93-2ab2da599f18)\"" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" Oct 08 21:08:15 crc kubenswrapper[4988]: I1008 21:08:15.246417 4988 scope.go:117] "RemoveContainer" containerID="6c07ed3a23b9e90ec44e0b3107713fd9bedf485fa1562b3bc3a266ffc1308b7c" Oct 08 21:08:15 crc kubenswrapper[4988]: E1008 21:08:15.247269 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nm54f_openshift-machine-config-operator(f328a20b-5283-407e-8a93-2ab2da599f18)\"" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" Oct 08 21:08:15 crc kubenswrapper[4988]: I1008 21:08:15.279227 4988 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-592hw"] Oct 08 21:08:15 crc kubenswrapper[4988]: E1008 21:08:15.279756 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1bfd72ab-365f-483a-9ecd-3db46219eed9" containerName="extract-content" Oct 08 21:08:15 crc kubenswrapper[4988]: I1008 21:08:15.279778 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="1bfd72ab-365f-483a-9ecd-3db46219eed9" containerName="extract-content" Oct 08 21:08:15 crc kubenswrapper[4988]: E1008 21:08:15.279804 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1bfd72ab-365f-483a-9ecd-3db46219eed9" containerName="extract-utilities" Oct 08 21:08:15 crc kubenswrapper[4988]: I1008 21:08:15.279811 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="1bfd72ab-365f-483a-9ecd-3db46219eed9" containerName="extract-utilities" Oct 08 21:08:15 crc kubenswrapper[4988]: E1008 21:08:15.279823 4988 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1bfd72ab-365f-483a-9ecd-3db46219eed9" containerName="registry-server" Oct 08 21:08:15 crc kubenswrapper[4988]: I1008 21:08:15.279829 4988 state_mem.go:107] "Deleted CPUSet assignment" podUID="1bfd72ab-365f-483a-9ecd-3db46219eed9" containerName="registry-server" Oct 08 21:08:15 crc kubenswrapper[4988]: I1008 21:08:15.280075 4988 memory_manager.go:354] "RemoveStaleState removing state" podUID="1bfd72ab-365f-483a-9ecd-3db46219eed9" containerName="registry-server" Oct 08 21:08:15 crc kubenswrapper[4988]: I1008 21:08:15.281719 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-592hw" Oct 08 21:08:15 crc kubenswrapper[4988]: I1008 21:08:15.292544 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-592hw"] Oct 08 21:08:15 crc kubenswrapper[4988]: I1008 21:08:15.364805 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/41ae311a-b385-4db8-8ab2-fbde36cb60c5-utilities\") pod \"redhat-operators-592hw\" (UID: \"41ae311a-b385-4db8-8ab2-fbde36cb60c5\") " pod="openshift-marketplace/redhat-operators-592hw" Oct 08 21:08:15 crc kubenswrapper[4988]: I1008 21:08:15.364989 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6p4jc\" (UniqueName: \"kubernetes.io/projected/41ae311a-b385-4db8-8ab2-fbde36cb60c5-kube-api-access-6p4jc\") pod \"redhat-operators-592hw\" (UID: \"41ae311a-b385-4db8-8ab2-fbde36cb60c5\") " pod="openshift-marketplace/redhat-operators-592hw" Oct 08 21:08:15 crc kubenswrapper[4988]: I1008 21:08:15.365010 4988 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/41ae311a-b385-4db8-8ab2-fbde36cb60c5-catalog-content\") pod \"redhat-operators-592hw\" (UID: \"41ae311a-b385-4db8-8ab2-fbde36cb60c5\") " pod="openshift-marketplace/redhat-operators-592hw" Oct 08 21:08:15 crc kubenswrapper[4988]: I1008 21:08:15.466559 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6p4jc\" (UniqueName: \"kubernetes.io/projected/41ae311a-b385-4db8-8ab2-fbde36cb60c5-kube-api-access-6p4jc\") pod \"redhat-operators-592hw\" (UID: \"41ae311a-b385-4db8-8ab2-fbde36cb60c5\") " pod="openshift-marketplace/redhat-operators-592hw" Oct 08 21:08:15 crc kubenswrapper[4988]: I1008 21:08:15.466843 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/41ae311a-b385-4db8-8ab2-fbde36cb60c5-catalog-content\") pod \"redhat-operators-592hw\" (UID: \"41ae311a-b385-4db8-8ab2-fbde36cb60c5\") " pod="openshift-marketplace/redhat-operators-592hw" Oct 08 21:08:15 crc kubenswrapper[4988]: I1008 21:08:15.466929 4988 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/41ae311a-b385-4db8-8ab2-fbde36cb60c5-utilities\") pod \"redhat-operators-592hw\" (UID: \"41ae311a-b385-4db8-8ab2-fbde36cb60c5\") " pod="openshift-marketplace/redhat-operators-592hw" Oct 08 21:08:15 crc kubenswrapper[4988]: I1008 21:08:15.467377 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/41ae311a-b385-4db8-8ab2-fbde36cb60c5-catalog-content\") pod \"redhat-operators-592hw\" (UID: \"41ae311a-b385-4db8-8ab2-fbde36cb60c5\") " pod="openshift-marketplace/redhat-operators-592hw" Oct 08 21:08:15 crc kubenswrapper[4988]: I1008 21:08:15.467509 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/41ae311a-b385-4db8-8ab2-fbde36cb60c5-utilities\") pod \"redhat-operators-592hw\" (UID: \"41ae311a-b385-4db8-8ab2-fbde36cb60c5\") " pod="openshift-marketplace/redhat-operators-592hw" Oct 08 21:08:15 crc kubenswrapper[4988]: I1008 21:08:15.487489 4988 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6p4jc\" (UniqueName: \"kubernetes.io/projected/41ae311a-b385-4db8-8ab2-fbde36cb60c5-kube-api-access-6p4jc\") pod \"redhat-operators-592hw\" (UID: \"41ae311a-b385-4db8-8ab2-fbde36cb60c5\") " pod="openshift-marketplace/redhat-operators-592hw" Oct 08 21:08:15 crc kubenswrapper[4988]: I1008 21:08:15.615265 4988 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-592hw" Oct 08 21:08:16 crc kubenswrapper[4988]: I1008 21:08:16.129632 4988 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-592hw"] Oct 08 21:08:16 crc kubenswrapper[4988]: I1008 21:08:16.902757 4988 generic.go:334] "Generic (PLEG): container finished" podID="41ae311a-b385-4db8-8ab2-fbde36cb60c5" containerID="8e8548283df1d9650f3ffec5168982641a5243f247ed87d64f2c5076ccdacde8" exitCode=0 Oct 08 21:08:16 crc kubenswrapper[4988]: I1008 21:08:16.902994 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-592hw" event={"ID":"41ae311a-b385-4db8-8ab2-fbde36cb60c5","Type":"ContainerDied","Data":"8e8548283df1d9650f3ffec5168982641a5243f247ed87d64f2c5076ccdacde8"} Oct 08 21:08:16 crc kubenswrapper[4988]: I1008 21:08:16.903022 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-592hw" event={"ID":"41ae311a-b385-4db8-8ab2-fbde36cb60c5","Type":"ContainerStarted","Data":"178bfddd1136f0b50d8cf59f6bc6508364c7f35e6d94a7b4d2da1fd0dc6ae147"} Oct 08 21:08:18 crc kubenswrapper[4988]: I1008 21:08:18.926102 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-592hw" event={"ID":"41ae311a-b385-4db8-8ab2-fbde36cb60c5","Type":"ContainerStarted","Data":"9c458219d287102802e1a00f1df9bcaf409f38842c3292bb0a612a4bfe8f5fa0"} Oct 08 21:08:22 crc kubenswrapper[4988]: I1008 21:08:22.981582 4988 generic.go:334] "Generic (PLEG): container finished" podID="41ae311a-b385-4db8-8ab2-fbde36cb60c5" containerID="9c458219d287102802e1a00f1df9bcaf409f38842c3292bb0a612a4bfe8f5fa0" exitCode=0 Oct 08 21:08:22 crc kubenswrapper[4988]: I1008 21:08:22.982014 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-592hw" event={"ID":"41ae311a-b385-4db8-8ab2-fbde36cb60c5","Type":"ContainerDied","Data":"9c458219d287102802e1a00f1df9bcaf409f38842c3292bb0a612a4bfe8f5fa0"} Oct 08 21:08:23 crc kubenswrapper[4988]: E1008 21:08:23.686048 4988 upgradeaware.go:427] Error proxying data from client to backend: readfrom tcp 38.102.83.102:54812->38.102.83.102:38495: write tcp 38.102.83.102:54812->38.102.83.102:38495: write: broken pipe Oct 08 21:08:23 crc kubenswrapper[4988]: E1008 21:08:23.893430 4988 upgradeaware.go:441] Error proxying data from backend to client: writeto tcp 38.102.83.102:54890->38.102.83.102:38495: read tcp 38.102.83.102:54890->38.102.83.102:38495: read: connection reset by peer Oct 08 21:08:23 crc kubenswrapper[4988]: I1008 21:08:23.995360 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-592hw" event={"ID":"41ae311a-b385-4db8-8ab2-fbde36cb60c5","Type":"ContainerStarted","Data":"388d43164c4015f3a860febe3c9811c736130530f61ad754db754771e8dd227b"} Oct 08 21:08:25 crc kubenswrapper[4988]: I1008 21:08:25.615517 4988 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-592hw" Oct 08 21:08:25 crc kubenswrapper[4988]: I1008 21:08:25.615869 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-592hw" Oct 08 21:08:26 crc kubenswrapper[4988]: I1008 21:08:26.670114 4988 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-592hw" podUID="41ae311a-b385-4db8-8ab2-fbde36cb60c5" containerName="registry-server" probeResult="failure" output=< Oct 08 21:08:26 crc kubenswrapper[4988]: timeout: failed to connect service ":50051" within 1s Oct 08 21:08:26 crc kubenswrapper[4988]: > Oct 08 21:08:29 crc kubenswrapper[4988]: I1008 21:08:29.237750 4988 scope.go:117] "RemoveContainer" containerID="6c07ed3a23b9e90ec44e0b3107713fd9bedf485fa1562b3bc3a266ffc1308b7c" Oct 08 21:08:29 crc kubenswrapper[4988]: E1008 21:08:29.238573 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nm54f_openshift-machine-config-operator(f328a20b-5283-407e-8a93-2ab2da599f18)\"" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" Oct 08 21:08:35 crc kubenswrapper[4988]: I1008 21:08:35.674698 4988 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-592hw" Oct 08 21:08:35 crc kubenswrapper[4988]: I1008 21:08:35.697213 4988 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-592hw" podStartSLOduration=14.189536695 podStartE2EDuration="20.697194158s" podCreationTimestamp="2025-10-08 21:08:15 +0000 UTC" firstStartedPulling="2025-10-08 21:08:16.904856214 +0000 UTC m=+10642.354698984" lastFinishedPulling="2025-10-08 21:08:23.412513657 +0000 UTC m=+10648.862356447" observedRunningTime="2025-10-08 21:08:24.030350284 +0000 UTC m=+10649.480193054" watchObservedRunningTime="2025-10-08 21:08:35.697194158 +0000 UTC m=+10661.147036938" Oct 08 21:08:35 crc kubenswrapper[4988]: I1008 21:08:35.753378 4988 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-592hw" Oct 08 21:08:35 crc kubenswrapper[4988]: I1008 21:08:35.913679 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-592hw"] Oct 08 21:08:37 crc kubenswrapper[4988]: I1008 21:08:37.155273 4988 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-592hw" podUID="41ae311a-b385-4db8-8ab2-fbde36cb60c5" containerName="registry-server" containerID="cri-o://388d43164c4015f3a860febe3c9811c736130530f61ad754db754771e8dd227b" gracePeriod=2 Oct 08 21:08:37 crc kubenswrapper[4988]: I1008 21:08:37.720929 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-592hw" Oct 08 21:08:37 crc kubenswrapper[4988]: I1008 21:08:37.899978 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/41ae311a-b385-4db8-8ab2-fbde36cb60c5-utilities\") pod \"41ae311a-b385-4db8-8ab2-fbde36cb60c5\" (UID: \"41ae311a-b385-4db8-8ab2-fbde36cb60c5\") " Oct 08 21:08:37 crc kubenswrapper[4988]: I1008 21:08:37.900551 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6p4jc\" (UniqueName: \"kubernetes.io/projected/41ae311a-b385-4db8-8ab2-fbde36cb60c5-kube-api-access-6p4jc\") pod \"41ae311a-b385-4db8-8ab2-fbde36cb60c5\" (UID: \"41ae311a-b385-4db8-8ab2-fbde36cb60c5\") " Oct 08 21:08:37 crc kubenswrapper[4988]: I1008 21:08:37.900807 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/41ae311a-b385-4db8-8ab2-fbde36cb60c5-catalog-content\") pod \"41ae311a-b385-4db8-8ab2-fbde36cb60c5\" (UID: \"41ae311a-b385-4db8-8ab2-fbde36cb60c5\") " Oct 08 21:08:37 crc kubenswrapper[4988]: I1008 21:08:37.901106 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/41ae311a-b385-4db8-8ab2-fbde36cb60c5-utilities" (OuterVolumeSpecName: "utilities") pod "41ae311a-b385-4db8-8ab2-fbde36cb60c5" (UID: "41ae311a-b385-4db8-8ab2-fbde36cb60c5"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 21:08:37 crc kubenswrapper[4988]: I1008 21:08:37.902572 4988 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/41ae311a-b385-4db8-8ab2-fbde36cb60c5-utilities\") on node \"crc\" DevicePath \"\"" Oct 08 21:08:37 crc kubenswrapper[4988]: I1008 21:08:37.912778 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/41ae311a-b385-4db8-8ab2-fbde36cb60c5-kube-api-access-6p4jc" (OuterVolumeSpecName: "kube-api-access-6p4jc") pod "41ae311a-b385-4db8-8ab2-fbde36cb60c5" (UID: "41ae311a-b385-4db8-8ab2-fbde36cb60c5"). InnerVolumeSpecName "kube-api-access-6p4jc". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 21:08:37 crc kubenswrapper[4988]: I1008 21:08:37.994835 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/41ae311a-b385-4db8-8ab2-fbde36cb60c5-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "41ae311a-b385-4db8-8ab2-fbde36cb60c5" (UID: "41ae311a-b385-4db8-8ab2-fbde36cb60c5"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 21:08:38 crc kubenswrapper[4988]: I1008 21:08:38.005037 4988 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6p4jc\" (UniqueName: \"kubernetes.io/projected/41ae311a-b385-4db8-8ab2-fbde36cb60c5-kube-api-access-6p4jc\") on node \"crc\" DevicePath \"\"" Oct 08 21:08:38 crc kubenswrapper[4988]: I1008 21:08:38.005076 4988 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/41ae311a-b385-4db8-8ab2-fbde36cb60c5-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 08 21:08:38 crc kubenswrapper[4988]: I1008 21:08:38.169531 4988 generic.go:334] "Generic (PLEG): container finished" podID="41ae311a-b385-4db8-8ab2-fbde36cb60c5" containerID="388d43164c4015f3a860febe3c9811c736130530f61ad754db754771e8dd227b" exitCode=0 Oct 08 21:08:38 crc kubenswrapper[4988]: I1008 21:08:38.169590 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-592hw" event={"ID":"41ae311a-b385-4db8-8ab2-fbde36cb60c5","Type":"ContainerDied","Data":"388d43164c4015f3a860febe3c9811c736130530f61ad754db754771e8dd227b"} Oct 08 21:08:38 crc kubenswrapper[4988]: I1008 21:08:38.169631 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-592hw" event={"ID":"41ae311a-b385-4db8-8ab2-fbde36cb60c5","Type":"ContainerDied","Data":"178bfddd1136f0b50d8cf59f6bc6508364c7f35e6d94a7b4d2da1fd0dc6ae147"} Oct 08 21:08:38 crc kubenswrapper[4988]: I1008 21:08:38.169653 4988 scope.go:117] "RemoveContainer" containerID="388d43164c4015f3a860febe3c9811c736130530f61ad754db754771e8dd227b" Oct 08 21:08:38 crc kubenswrapper[4988]: I1008 21:08:38.169684 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-592hw" Oct 08 21:08:38 crc kubenswrapper[4988]: I1008 21:08:38.212339 4988 scope.go:117] "RemoveContainer" containerID="9c458219d287102802e1a00f1df9bcaf409f38842c3292bb0a612a4bfe8f5fa0" Oct 08 21:08:38 crc kubenswrapper[4988]: I1008 21:08:38.236039 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-592hw"] Oct 08 21:08:38 crc kubenswrapper[4988]: I1008 21:08:38.247245 4988 scope.go:117] "RemoveContainer" containerID="8e8548283df1d9650f3ffec5168982641a5243f247ed87d64f2c5076ccdacde8" Oct 08 21:08:38 crc kubenswrapper[4988]: I1008 21:08:38.250262 4988 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-592hw"] Oct 08 21:08:38 crc kubenswrapper[4988]: I1008 21:08:38.329917 4988 scope.go:117] "RemoveContainer" containerID="388d43164c4015f3a860febe3c9811c736130530f61ad754db754771e8dd227b" Oct 08 21:08:38 crc kubenswrapper[4988]: E1008 21:08:38.330589 4988 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"388d43164c4015f3a860febe3c9811c736130530f61ad754db754771e8dd227b\": container with ID starting with 388d43164c4015f3a860febe3c9811c736130530f61ad754db754771e8dd227b not found: ID does not exist" containerID="388d43164c4015f3a860febe3c9811c736130530f61ad754db754771e8dd227b" Oct 08 21:08:38 crc kubenswrapper[4988]: I1008 21:08:38.330629 4988 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"388d43164c4015f3a860febe3c9811c736130530f61ad754db754771e8dd227b"} err="failed to get container status \"388d43164c4015f3a860febe3c9811c736130530f61ad754db754771e8dd227b\": rpc error: code = NotFound desc = could not find container \"388d43164c4015f3a860febe3c9811c736130530f61ad754db754771e8dd227b\": container with ID starting with 388d43164c4015f3a860febe3c9811c736130530f61ad754db754771e8dd227b not found: ID does not exist" Oct 08 21:08:38 crc kubenswrapper[4988]: I1008 21:08:38.330657 4988 scope.go:117] "RemoveContainer" containerID="9c458219d287102802e1a00f1df9bcaf409f38842c3292bb0a612a4bfe8f5fa0" Oct 08 21:08:38 crc kubenswrapper[4988]: E1008 21:08:38.331001 4988 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9c458219d287102802e1a00f1df9bcaf409f38842c3292bb0a612a4bfe8f5fa0\": container with ID starting with 9c458219d287102802e1a00f1df9bcaf409f38842c3292bb0a612a4bfe8f5fa0 not found: ID does not exist" containerID="9c458219d287102802e1a00f1df9bcaf409f38842c3292bb0a612a4bfe8f5fa0" Oct 08 21:08:38 crc kubenswrapper[4988]: I1008 21:08:38.331030 4988 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9c458219d287102802e1a00f1df9bcaf409f38842c3292bb0a612a4bfe8f5fa0"} err="failed to get container status \"9c458219d287102802e1a00f1df9bcaf409f38842c3292bb0a612a4bfe8f5fa0\": rpc error: code = NotFound desc = could not find container \"9c458219d287102802e1a00f1df9bcaf409f38842c3292bb0a612a4bfe8f5fa0\": container with ID starting with 9c458219d287102802e1a00f1df9bcaf409f38842c3292bb0a612a4bfe8f5fa0 not found: ID does not exist" Oct 08 21:08:38 crc kubenswrapper[4988]: I1008 21:08:38.331075 4988 scope.go:117] "RemoveContainer" containerID="8e8548283df1d9650f3ffec5168982641a5243f247ed87d64f2c5076ccdacde8" Oct 08 21:08:38 crc kubenswrapper[4988]: E1008 21:08:38.331765 4988 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8e8548283df1d9650f3ffec5168982641a5243f247ed87d64f2c5076ccdacde8\": container with ID starting with 8e8548283df1d9650f3ffec5168982641a5243f247ed87d64f2c5076ccdacde8 not found: ID does not exist" containerID="8e8548283df1d9650f3ffec5168982641a5243f247ed87d64f2c5076ccdacde8" Oct 08 21:08:38 crc kubenswrapper[4988]: I1008 21:08:38.331791 4988 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8e8548283df1d9650f3ffec5168982641a5243f247ed87d64f2c5076ccdacde8"} err="failed to get container status \"8e8548283df1d9650f3ffec5168982641a5243f247ed87d64f2c5076ccdacde8\": rpc error: code = NotFound desc = could not find container \"8e8548283df1d9650f3ffec5168982641a5243f247ed87d64f2c5076ccdacde8\": container with ID starting with 8e8548283df1d9650f3ffec5168982641a5243f247ed87d64f2c5076ccdacde8 not found: ID does not exist" Oct 08 21:08:39 crc kubenswrapper[4988]: I1008 21:08:39.257331 4988 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="41ae311a-b385-4db8-8ab2-fbde36cb60c5" path="/var/lib/kubelet/pods/41ae311a-b385-4db8-8ab2-fbde36cb60c5/volumes" Oct 08 21:08:41 crc kubenswrapper[4988]: I1008 21:08:41.238606 4988 scope.go:117] "RemoveContainer" containerID="6c07ed3a23b9e90ec44e0b3107713fd9bedf485fa1562b3bc3a266ffc1308b7c" Oct 08 21:08:41 crc kubenswrapper[4988]: E1008 21:08:41.239787 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nm54f_openshift-machine-config-operator(f328a20b-5283-407e-8a93-2ab2da599f18)\"" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" Oct 08 21:08:55 crc kubenswrapper[4988]: I1008 21:08:55.252235 4988 scope.go:117] "RemoveContainer" containerID="6c07ed3a23b9e90ec44e0b3107713fd9bedf485fa1562b3bc3a266ffc1308b7c" Oct 08 21:08:55 crc kubenswrapper[4988]: E1008 21:08:55.253042 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nm54f_openshift-machine-config-operator(f328a20b-5283-407e-8a93-2ab2da599f18)\"" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" Oct 08 21:09:06 crc kubenswrapper[4988]: I1008 21:09:06.238460 4988 scope.go:117] "RemoveContainer" containerID="6c07ed3a23b9e90ec44e0b3107713fd9bedf485fa1562b3bc3a266ffc1308b7c" Oct 08 21:09:06 crc kubenswrapper[4988]: E1008 21:09:06.239574 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nm54f_openshift-machine-config-operator(f328a20b-5283-407e-8a93-2ab2da599f18)\"" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" Oct 08 21:09:21 crc kubenswrapper[4988]: I1008 21:09:21.238001 4988 scope.go:117] "RemoveContainer" containerID="6c07ed3a23b9e90ec44e0b3107713fd9bedf485fa1562b3bc3a266ffc1308b7c" Oct 08 21:09:21 crc kubenswrapper[4988]: E1008 21:09:21.239041 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nm54f_openshift-machine-config-operator(f328a20b-5283-407e-8a93-2ab2da599f18)\"" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" Oct 08 21:09:34 crc kubenswrapper[4988]: I1008 21:09:34.238360 4988 scope.go:117] "RemoveContainer" containerID="6c07ed3a23b9e90ec44e0b3107713fd9bedf485fa1562b3bc3a266ffc1308b7c" Oct 08 21:09:34 crc kubenswrapper[4988]: E1008 21:09:34.246751 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nm54f_openshift-machine-config-operator(f328a20b-5283-407e-8a93-2ab2da599f18)\"" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" Oct 08 21:09:46 crc kubenswrapper[4988]: I1008 21:09:46.237865 4988 scope.go:117] "RemoveContainer" containerID="6c07ed3a23b9e90ec44e0b3107713fd9bedf485fa1562b3bc3a266ffc1308b7c" Oct 08 21:09:46 crc kubenswrapper[4988]: E1008 21:09:46.238699 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nm54f_openshift-machine-config-operator(f328a20b-5283-407e-8a93-2ab2da599f18)\"" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" Oct 08 21:09:58 crc kubenswrapper[4988]: I1008 21:09:58.227447 4988 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-dddvr/must-gather-6bgk6" event={"ID":"d157ad3b-3365-4413-a0b0-1da5184fc8ec","Type":"ContainerDied","Data":"27b0386a1f9f60257d0cd7d5e4d6e679c50af8dee65d0626b7ecb0c5d2fd29fb"} Oct 08 21:09:58 crc kubenswrapper[4988]: I1008 21:09:58.227383 4988 generic.go:334] "Generic (PLEG): container finished" podID="d157ad3b-3365-4413-a0b0-1da5184fc8ec" containerID="27b0386a1f9f60257d0cd7d5e4d6e679c50af8dee65d0626b7ecb0c5d2fd29fb" exitCode=0 Oct 08 21:09:58 crc kubenswrapper[4988]: I1008 21:09:58.229029 4988 scope.go:117] "RemoveContainer" containerID="27b0386a1f9f60257d0cd7d5e4d6e679c50af8dee65d0626b7ecb0c5d2fd29fb" Oct 08 21:09:58 crc kubenswrapper[4988]: I1008 21:09:58.931020 4988 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-dddvr_must-gather-6bgk6_d157ad3b-3365-4413-a0b0-1da5184fc8ec/gather/0.log" Oct 08 21:10:01 crc kubenswrapper[4988]: I1008 21:10:01.238711 4988 scope.go:117] "RemoveContainer" containerID="6c07ed3a23b9e90ec44e0b3107713fd9bedf485fa1562b3bc3a266ffc1308b7c" Oct 08 21:10:01 crc kubenswrapper[4988]: E1008 21:10:01.239628 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nm54f_openshift-machine-config-operator(f328a20b-5283-407e-8a93-2ab2da599f18)\"" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" Oct 08 21:10:07 crc kubenswrapper[4988]: I1008 21:10:07.188168 4988 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-dddvr/must-gather-6bgk6"] Oct 08 21:10:07 crc kubenswrapper[4988]: I1008 21:10:07.189211 4988 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-must-gather-dddvr/must-gather-6bgk6" podUID="d157ad3b-3365-4413-a0b0-1da5184fc8ec" containerName="copy" containerID="cri-o://3f0bb832ec6b9f8cd5f93ef06d7767f3fdefb7827677752f6aeb98d3ca730b3d" gracePeriod=2 Oct 08 21:10:07 crc kubenswrapper[4988]: I1008 21:10:07.197241 4988 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-dddvr/must-gather-6bgk6"] Oct 08 21:10:07 crc kubenswrapper[4988]: I1008 21:10:07.341182 4988 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-dddvr_must-gather-6bgk6_d157ad3b-3365-4413-a0b0-1da5184fc8ec/copy/0.log" Oct 08 21:10:07 crc kubenswrapper[4988]: I1008 21:10:07.341690 4988 generic.go:334] "Generic (PLEG): container finished" podID="d157ad3b-3365-4413-a0b0-1da5184fc8ec" containerID="3f0bb832ec6b9f8cd5f93ef06d7767f3fdefb7827677752f6aeb98d3ca730b3d" exitCode=143 Oct 08 21:10:07 crc kubenswrapper[4988]: I1008 21:10:07.814661 4988 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-dddvr_must-gather-6bgk6_d157ad3b-3365-4413-a0b0-1da5184fc8ec/copy/0.log" Oct 08 21:10:07 crc kubenswrapper[4988]: I1008 21:10:07.816001 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-dddvr/must-gather-6bgk6" Oct 08 21:10:07 crc kubenswrapper[4988]: I1008 21:10:07.965499 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-687zg\" (UniqueName: \"kubernetes.io/projected/d157ad3b-3365-4413-a0b0-1da5184fc8ec-kube-api-access-687zg\") pod \"d157ad3b-3365-4413-a0b0-1da5184fc8ec\" (UID: \"d157ad3b-3365-4413-a0b0-1da5184fc8ec\") " Oct 08 21:10:07 crc kubenswrapper[4988]: I1008 21:10:07.965662 4988 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/d157ad3b-3365-4413-a0b0-1da5184fc8ec-must-gather-output\") pod \"d157ad3b-3365-4413-a0b0-1da5184fc8ec\" (UID: \"d157ad3b-3365-4413-a0b0-1da5184fc8ec\") " Oct 08 21:10:07 crc kubenswrapper[4988]: I1008 21:10:07.973622 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d157ad3b-3365-4413-a0b0-1da5184fc8ec-kube-api-access-687zg" (OuterVolumeSpecName: "kube-api-access-687zg") pod "d157ad3b-3365-4413-a0b0-1da5184fc8ec" (UID: "d157ad3b-3365-4413-a0b0-1da5184fc8ec"). InnerVolumeSpecName "kube-api-access-687zg". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 21:10:08 crc kubenswrapper[4988]: I1008 21:10:08.086991 4988 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-687zg\" (UniqueName: \"kubernetes.io/projected/d157ad3b-3365-4413-a0b0-1da5184fc8ec-kube-api-access-687zg\") on node \"crc\" DevicePath \"\"" Oct 08 21:10:08 crc kubenswrapper[4988]: I1008 21:10:08.203728 4988 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d157ad3b-3365-4413-a0b0-1da5184fc8ec-must-gather-output" (OuterVolumeSpecName: "must-gather-output") pod "d157ad3b-3365-4413-a0b0-1da5184fc8ec" (UID: "d157ad3b-3365-4413-a0b0-1da5184fc8ec"). InnerVolumeSpecName "must-gather-output". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 21:10:08 crc kubenswrapper[4988]: I1008 21:10:08.296775 4988 reconciler_common.go:293] "Volume detached for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/d157ad3b-3365-4413-a0b0-1da5184fc8ec-must-gather-output\") on node \"crc\" DevicePath \"\"" Oct 08 21:10:08 crc kubenswrapper[4988]: I1008 21:10:08.355799 4988 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-dddvr_must-gather-6bgk6_d157ad3b-3365-4413-a0b0-1da5184fc8ec/copy/0.log" Oct 08 21:10:08 crc kubenswrapper[4988]: I1008 21:10:08.356515 4988 scope.go:117] "RemoveContainer" containerID="3f0bb832ec6b9f8cd5f93ef06d7767f3fdefb7827677752f6aeb98d3ca730b3d" Oct 08 21:10:08 crc kubenswrapper[4988]: I1008 21:10:08.356589 4988 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-dddvr/must-gather-6bgk6" Oct 08 21:10:08 crc kubenswrapper[4988]: I1008 21:10:08.386229 4988 scope.go:117] "RemoveContainer" containerID="27b0386a1f9f60257d0cd7d5e4d6e679c50af8dee65d0626b7ecb0c5d2fd29fb" Oct 08 21:10:09 crc kubenswrapper[4988]: I1008 21:10:09.264220 4988 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d157ad3b-3365-4413-a0b0-1da5184fc8ec" path="/var/lib/kubelet/pods/d157ad3b-3365-4413-a0b0-1da5184fc8ec/volumes" Oct 08 21:10:15 crc kubenswrapper[4988]: I1008 21:10:15.244421 4988 scope.go:117] "RemoveContainer" containerID="6c07ed3a23b9e90ec44e0b3107713fd9bedf485fa1562b3bc3a266ffc1308b7c" Oct 08 21:10:15 crc kubenswrapper[4988]: E1008 21:10:15.245039 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nm54f_openshift-machine-config-operator(f328a20b-5283-407e-8a93-2ab2da599f18)\"" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" Oct 08 21:10:26 crc kubenswrapper[4988]: I1008 21:10:26.238690 4988 scope.go:117] "RemoveContainer" containerID="6c07ed3a23b9e90ec44e0b3107713fd9bedf485fa1562b3bc3a266ffc1308b7c" Oct 08 21:10:26 crc kubenswrapper[4988]: E1008 21:10:26.239500 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nm54f_openshift-machine-config-operator(f328a20b-5283-407e-8a93-2ab2da599f18)\"" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" Oct 08 21:10:38 crc kubenswrapper[4988]: I1008 21:10:38.238452 4988 scope.go:117] "RemoveContainer" containerID="6c07ed3a23b9e90ec44e0b3107713fd9bedf485fa1562b3bc3a266ffc1308b7c" Oct 08 21:10:38 crc kubenswrapper[4988]: E1008 21:10:38.239244 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nm54f_openshift-machine-config-operator(f328a20b-5283-407e-8a93-2ab2da599f18)\"" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" Oct 08 21:10:53 crc kubenswrapper[4988]: I1008 21:10:53.238266 4988 scope.go:117] "RemoveContainer" containerID="6c07ed3a23b9e90ec44e0b3107713fd9bedf485fa1562b3bc3a266ffc1308b7c" Oct 08 21:10:53 crc kubenswrapper[4988]: E1008 21:10:53.239340 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nm54f_openshift-machine-config-operator(f328a20b-5283-407e-8a93-2ab2da599f18)\"" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" Oct 08 21:11:04 crc kubenswrapper[4988]: I1008 21:11:04.238181 4988 scope.go:117] "RemoveContainer" containerID="6c07ed3a23b9e90ec44e0b3107713fd9bedf485fa1562b3bc3a266ffc1308b7c" Oct 08 21:11:04 crc kubenswrapper[4988]: E1008 21:11:04.239148 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nm54f_openshift-machine-config-operator(f328a20b-5283-407e-8a93-2ab2da599f18)\"" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" Oct 08 21:11:17 crc kubenswrapper[4988]: I1008 21:11:17.238839 4988 scope.go:117] "RemoveContainer" containerID="6c07ed3a23b9e90ec44e0b3107713fd9bedf485fa1562b3bc3a266ffc1308b7c" Oct 08 21:11:17 crc kubenswrapper[4988]: E1008 21:11:17.240020 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nm54f_openshift-machine-config-operator(f328a20b-5283-407e-8a93-2ab2da599f18)\"" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" Oct 08 21:11:31 crc kubenswrapper[4988]: I1008 21:11:31.237748 4988 scope.go:117] "RemoveContainer" containerID="6c07ed3a23b9e90ec44e0b3107713fd9bedf485fa1562b3bc3a266ffc1308b7c" Oct 08 21:11:31 crc kubenswrapper[4988]: E1008 21:11:31.238613 4988 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nm54f_openshift-machine-config-operator(f328a20b-5283-407e-8a93-2ab2da599f18)\"" pod="openshift-machine-config-operator/machine-config-daemon-nm54f" podUID="f328a20b-5283-407e-8a93-2ab2da599f18" var/home/core/zuul-output/logs/crc-cloud-workdir-crc-all-logs.tar.gz0000644000175000000000000000005515071551625024454 0ustar coreroot‹íÁ  ÷Om7 €7šÞ'(var/home/core/zuul-output/logs/crc-cloud/0000755000175000000000000000000015071551626017372 5ustar corerootvar/home/core/zuul-output/artifacts/0000755000175000017500000000000015071524113016504 5ustar corecorevar/home/core/zuul-output/docs/0000755000175000017500000000000015071524113015454 5ustar corecore